Frank Jacksonâs hypothetical black-and-white room-reared color neuroscientist âMaryâ, unlike ChatGPT, sees (but only black/white/gray). So her vocabulary for setting up her explanation of color perception is nevertheless grounded. (Even Helen Kellerâs was, because, lacking sight and hearing since 19 months old, she still had touch and kinesthesia.) ChatGPT has no grounding at all.
So we can get from âhorse, horn, vanish, eyes, observing instrumentsâ to the unobservable âpeekaboo unicornâ purely verbally, using just those prior (grounded) words to define or describe it.
But ChatGPT canât get anywhere, because it has no grounded vocabulary at all. Just words âin contextâ (â100 trillion parameters and 300 billion wordsâŠ: 570 gigabytes of text dataâ).
So much for ChatGPTâs chances of the âphase transitionâ some emergence enthusiasts are imagining — from just harvesting and organizing the words in ChatGPTâs  bloated 570gig belly, to an âemergentâ understanding of them. You, me, Helen Keller or Monochrome Mary, we could, in contrast, understand ChatGPTâs words. And we could mean them if we said them.  So could the dead authors of the words in ChatGPTâs belly, once. Thatâs the difference between (intrinsic) understanding and word-crunching by a well-fed statistical parrot.
Two important, connected, details: (1) Mary would be more than surprised if her screen went from B/W to color: she would become able to DO things that she could not do before (on her own) locked in her B/W room — like tell apart red and green, just as would any daltonian, if their trichromacy were repaired.Â
More important, (2) Mary, or Helen, or any daltonianian could be told any number of words about what green is neurologically and what things are green and what not. But they cannot be told what seeing green FEELS-LIKE (i.e., what green looks-like). And thatâs the point. If it werenât for the âhard problemâ â the fact that it feels-like something to see, hear, touch [and understand and mean] something (none of which is something you DO) â Mary, Helen, and a daltonian would not be missing anything about green. Grounding in transducer/effector know-how alone would be all there was to âmeaningâ and âunderstanding.â But itâs not.Â
Mary, Helen and the daltonian can make do with (imperfect) verbal analogies and extrapolations from the senses they do have to the senses they lack. But ChatGPT canât. Because all it has is a bellyful of ungrounded words (but syntactically structured by their grounded authors â some of whom are perhaps already underground, long dead and buriedâŠ).
So meaning and understanding is not just sensorimotor know-how. Itâs based on sentience too (in fact, first and foremost, though, so far, itâs an unsolved hard-problem to explain how or why). ChatGPT (attention âsingularityâ fans!) lacks that completely.Â
It FEELS-LIKE something to understand what red (or round, or anything) means. And thatâs not just transducer/effector know-how that percolates up from the sensorimotor grounding. Itâs what it feels-like to see, hear, touch, taste and manipulate all those things out there that we are talking about, with our words.
ChatGPT is talking, but ABOUT nothing. The âaboutnessâ is supplied by the original authorâs and the readerâs head. ChatGPT is just producing recombinatory, fill-in-the-blanks echolalia.