I have many conversations with people about Large Language Models like ChatGPT and Copilot. The idea that “it makes convincing sentences, but it doesn’t know what it’s talking about” is a difficult concept to convey or wrap your head around. Because the sentences are so convincing.

Any good examples on how to explain this in simple terms?

Edit:some good answers already! I find especially that the emotional barrier is difficult to break. If an AI says something malicious, our brain immediatly jumps to “it has intent”. How can we explain this away?

  • Hucklebee@lemmy.worldOP
    link
    fedilink
    arrow-up
    5
    ·
    6 months ago

    I commented something similair on another post, but this is exactly why I find this phenomenon so hard to describe.

    A teenager in a new group still has some understanding and has a mind. It knows many of the meaning of the words that are said. Sure, some catchphrases might be new, but general topics shouldn’t be too hard to follow.

    This is nothing like genAI. GenAI doesn’t know anything at all. It has (simplified) a list of words that somehow are connected to eachother. But AI has no meaning of a wheel, what round is, what rolling is, what rubber is, what an axle is. NO understanding. Just words that happened to describe all of it. For us humans it is so difficult to understand that something uses language without knowing ANY of the meaning.

    How can we describe this so our brains make sense that you can have language without understanding? The Chinese Room experiment comes close, but is quite complicated to explain as well I think.

    • NeoNachtwaechter@lemmy.world
      link
      fedilink
      arrow-up
      2
      ·
      6 months ago

      How can we describe this so our brains make sense that you can have language without understanding?

      I think it is really impossible to describe in easy and limited words.

    • Turun@feddit.de
      link
      fedilink
      arrow-up
      1
      ·
      6 months ago

      NO understanding. Just words that happened to describe all of it.

      If being able to describe it does not mean understanding, then what is understanding?