248 | Yejin Choi on AI and Common Sense

Over the last year, AI large-language models (LLMs) like ChatGPT have demonstrated a remarkable ability to carry on human-like conversations in a variety of different concepts. But the way these LLMs "learn" is very different from how human beings learn, and the same can be said for how they "reason." It's reasonable to ask, do these AI programs really understand the world they are talking about? Do they possess a common-sense picture of reality, or can they just string together words in convincing ways without any underlying understanding? Computer scientist Yejin Choi is a leader in trying to understand the sense in which AIs are actually intelligent, and why in some ways they're still shockingly stupid.

yejin choi

Support Mindscape on Patreon.

Yejin Choi received a Ph.D. in computer science from Cornell University. She is currently the Wissner-Slivka Professor at the Paul G. Allen School of Computer Science & Engineering at the University of Washington and also a senior research director at AI2 overseeing the project Mosaic. Among her awards are a MacArthur fellowship and a fellow of the Association for Computational Linguistics.

7 thoughts on “248 | Yejin Choi on AI and Common Sense”

  1. Great conclusion to the podcast by ms Yejin Choi ‘And thanks to you for this fun conversation’. I am sure she also spoke on behalf of most of your listners, certainly myself.

    Again I have come to the conclusion that LLMs, like probably most deep learning AI models, have automated the art of forgery. This makes these models comparable to the famous Vermeer forger Han van Meegeren.

    https://en.wikipedia.org/wiki/Han_van_Meegeren

    Who could not have fooled experts for years, without the the paintings of the true creative genius of Vermeer available to him.

  2. We have to be careful in dismissing AI cababilities. Only in the end the crucial point was mentioned. It is just that we only found a solution for one of the subsystem of the mind. It is a solution so good that we forget that the LLM’s do not learn (after the initial training), do not rembember anything (other than the current conversation), do not self reflect, playing conterfactual in their mind like we do. Their wolrld building is rudimental at best. We “just” have to find a way to put subsystems togheter (LLM + memory + ?? ) like they are put togheter in our brain. It could take decades, but it could happens next year.

  3. Pingback: Sean Carroll's Mindscape Podcast: Yejin Choi on AI and Common Sense - 3 Quarks Daily

  4. Maria Fátima Pereira

    Obrigada a ambos por este bom episódio.
    Mais elucidada sobre as capacidades e limitações (até ao momento) da I.A.
    Apreciei o “bom senso”!
    Já é o presente e será o futuro!
    Urgente, legislação nacional, e internacional sobre toda a envolvente, de forma a minimizar “riscos” desnecessarios.

  5. I wish there had been discussion of the often repeated comment from AI developers that they don’t know how their AI programs work. The complexity and the deep levels of processing make it impossible to be transparent about the way an AI like ChatGPT produces its responses. What are the implications of this inability to trace the path that produced the text? Will legal safeguards that are being developed that include the requirement of transparency (as in the U.K.) create a hurdle that will be impossible to clear?

  6. ChatGPT is an excellent demonstration (through extensive training) of selecting the next word after a sequence of words. But often people think in images and then summarize the resulting thoughts with words. I’m looking forward to ImageGPT, however, it is suggested that image sequences require a much larger vector space so that may be quite a bit further out in time.
    I have heard that when asked if machines will ever become sentient, Steve Wozniak had said that it will never happen because, “No machine will ever feel what I feel when I see a dog that is happy or the tears I cry when I see an animal that’s been rescued.”
    As touched on in the podcast, it seems the main part that is absent is something like endorphins (perhaps compudorphins), where the computer is not just trained to dutifully adjust weightings to match probabilities from a huge number of sentences, but also to maximize a utility function (a happiness) based on its homeostasis (to borrow a word from the Antonio Damasio podcast) or its status. Maximizing compudorphins would in effect be a self-reflection.
    I had a chat with ChatGPT about this. It helped me to think a little more broadly about this. Right now, ChatGPT is moderately predictable, but if the response of the computer depends on the existing state of the computer and the direction of that state, one should expect far more diverse results. The level of compudorphins would affect the computer’s decision-making process, nudging it toward self-interest, motivation, perhaps curiosity/reluctance, and computer-motivated changes in direction. The computer may not work when we want it to or to answer the question we ask.
    Most children go through a rebellious phase (display behaviors such as questioning authority, seeking independence, testing boundaries, engaging in risky or defiant actions, and forming their own opinions and beliefs).
    Still, I think that Yejin Choi’s assertion of: How will computer’s know what question to ask or perhaps what experiment to do in order to learn what they don’t know? is a very important one. This may require the broad goal and evaluation of learning to be part of the compudorphin utility function.
    ChatGPT suggests safety constraints be used.

Comments are closed.

Scroll to Top