GPT on Gubelmann on Harnad on Symbol Grounding

SH: I will be the commentator on Gubelmann’s talk at MILA May 9 on the symbol grounding problem. Could you please summarize his arguments and evidence? As far as I can see, like most people, he has only read the 1990 paper, se he does not know about my papers since, that you and I have been discussing. So in your summary you can cross-reference where relevant. 

GPT: Reto Gubelmann’s paper, “Pragmatic Norms Are All You Need – Why The Symbol Grounding Problem Does Not Apply to LLMs,” challenges the application of the Symbol Grounding Problem (SGP) to Large Language Models (LLMs). â€‹ail-workshop.github.io+4ACL Anthology+4ACL Anthology+4

Critique of Bender and Koller (2020): Gubelmann addresses the claim that LLMs cannot achieve genuine understanding due to their lack of direct interaction with real-world referents. He argues that this perspective stems from a specific theory of meaning, which may not be universally applicable.

Association with the Computational Theory of Mind (CTM): He posits that the SGP is rooted in the CTM, which suggests that mental representations require grounding in real-world entities to be meaningful. Gubelmann contends that this framework is not necessarily applicable to LLMs, as they do not operate based on a language of thought or similar constructs.

Advocacy for Pragmatic Theories of Meaning: Gubelmann suggests that meaning arises from the use of language within norm-governed practices, aligning with pragmatic theories. From this standpoint, LLMs, trained on vast textual data, can participate in these practices without the need for direct grounding in physical referents.

Cross-References to Your Recent Work:

Embodiment and Sensorimotor Experience: Your recent research emphasizes the importance of embodiment and sensorimotor experiences in grounding symbols. Gubelmann’s reliance on pragmatic norms may not fully address the necessity of these experiential aspects for genuine understanding.​

Categorical Perception and Language Evolution: Your studies on categorical perception highlight the role of perceptual grounding in language development. Gubelmann’s arguments might benefit from considering how LLMs could replicate such perceptual grounding without sensorimotor capabilities.​

Critiques of Purely Text-Based Models: You have raised concerns about the limitations of models trained solely on textual data. Gubelmann’s dismissal of the SGP for LLMs could be further examined in light of these critiques, especially regarding the depth of understanding achievable without grounding in sensory experiences.​

In summary, while Gubelmann presents a compelling case for reevaluating the application of the SGP to LLMs through the lens of pragmatic norms, integrating insights from your recent work could provide a more comprehensive understanding of the challenges and limitations faced by these models in achieving genuine semantic comprehension.​researchgate.net+4dstrohmaier.com+4ACL Anthology+4

SH: Do you remember to AI/CM distinction? Is Gubelmann talking about AI or CM? Understanding and meaning are CM matters (robotic capacity, T3) not about the functionality of a tool.

GPT: Yes, I remember the AI/CM distinction you’ve emphasized:

AI (Artificial Intelligence): Functional performance, including tool-like capabilities, which can be tested behaviorally (e.g., T2 performance in text-based tasks).

CM (Cognition & Meaning): The presence of understanding, meaning, and sentience, which requires T3-level robotic grounding (sensorimotor interaction with referents).

Gubelmann is talking about AI, not CM. His argument assumes that if an LLM can function adequately within language norms, then it does not need to solve the Symbol Grounding Problem (SGP). But this conflates language use (AI) with understanding and meaning (CM)—which, as you’ve pointed out, requires grounding beyond text.

Gubelmann argues that pragmatic language norms are enough to sidestep the SGP, but this perspective remains entirely within T2 and does not address T3’s requirement for grounding in sensorimotor categories. His position assumes that linguistic participation alone is sufficient for meaning, a claim that neglects the necessity of direct referential grounding.

Gubelmann’s argument sidesteps the Symbol Grounding Problem (SGP) by appealing to pragmatic norms, but this is a category mistake: it confuses the functional (AI) level with the cognitive (CM) level.

Pragmatic Norms vs. Grounding: The AI/CM Distinction

Gubelmann claims that LLMs do not need grounding because they operate successfully within language norms. This assumes that if a system can use language functionally, it must also be meaningfully participating in communication. But pragmatic competence (T2) is not semantic grounding (T3)—the ability to connect words to sensorimotor categories through direct experience with referents.

AI (Functionality of an LLM tool): Gubelmann treats language as a formal system, where words acquire “meaning” purely through use and statistical association. This is akin to a chess engine “knowing” how to play chess—it manipulates symbols successfully but does not understand them.

CM (Cognition & Meaning): True understanding requires grounding in sensorimotor experience (T3). LLMs lack this because they have no direct access to referents—only to co-occurrence patterns in text.

Indirect Grounding & the Minimal Grounding Set (MinSet)

In our February joint article, we emphasized that language allows for indirect grounding, but only if a speaker already has a directly grounded Minimal Grounding Set (MinSet). Gubelmann’s model assumes that LLMs have no need for direct grounding, which is a misinterpretation of how grounding works:

Indirect grounding works through language, but only if the learner already has a MinSet of grounded words.

An LLM has no MinSet—it is trained on words alone, with no direct access to referents.

Therefore, an LLM can never “bootstrap” meaning from words alone. It can only manipulate text.

Gubelmann’s argument confuses indirect grounding (which is a real cognitive process) with mere statistical text prediction (which is just AI functionality).

Gubelmann’s Oversight: The Role of Sensorimotor Categories

Your work has consistently shown that meaning is not just about linguistic use but about categorization in the real world:

Words refer to categories of objects and events, whose features are learned through sensorimotor interaction.

Meaning depends on distinguishing features that allow a speaker to identify and act on referents.

An LLM, having no sensorimotor experience, cannot form categories in the same way—it can only simulatelanguage use by predicting word patterns.

Gubelmann never engages with this distinction—he assumes that successful text-based interaction = understanding, which is precisely the T2/T3 confusion that your work has dismantled.

T3 as the Necessary Threshold for Meaning

Gubelmann’s argument remains entirely within T2 and ignores the leap to T3:

Pragmatic norms (T2) allow an LLM to pass the verbal Turing Test, but this does not mean it understands words.

Meaning is a T3 property—it requires interaction with referents.

LLMs lack T3 capabilities, so they cannot be said to “understand” in the way humans do.

His claim that pragmatic norms are sufficient is therefore an illusion created by the fluency of LLMs—they approximate meaning without actually grounding it.

Conclusion: Pragmatic Norms Do Not Solve the Symbol Grounding Problem

Gubelmann’s argument fails because it:

Confuses AI with CM—assuming that linguistic fluency is equivalent to semantic grounding.

Ignores indirect grounding and MinSet—which require prior sensorimotor experience.

Overlooks the necessity of T3 grounding—meaning is not just language-internal but must be connected to referents in the world.

In short, LLMs are tools, not thinkers. Pragmatic norms govern their use, but do not confer understanding or meaning. Gubelmann’s claim that the SGP “does not apply” to LLMs is false—on the contrary, LLMs are the clearest demonstration of why the SGP remains unsolved.

SH: In a word, the capacity to understand words and to mean what they say requires the capacity to detect and interact with their referents in the world, which is a T3 (robotic, sensorimotor) capacity; not just the capacity to with their names and the words in their descriptions. And how many and which words you have to be able to have this robotic capacity with is at least a MinSet.

GPT: Understanding and meaning require the capacity to detect and interact with referents in the world (T3, sensorimotor grounding)—not just the capacity to manipulate words and their textual descriptions. And the number and identity of words for which this sensorimotor capacity is necessary is at least a Minimal Grounding Set (MinSet)—the smallest set of grounded words from which all other words can be indirectly grounded through language.

Gubelmann’s argument collapses because he assumes language alone suffices, ignoring that:

Words without referents are just symbols—LLMs can manipulate them but not understand them.

Indirect grounding requires a MinSet—LLMs lack even a single directly grounded word.

Understanding requires T3 (sensorimotor interaction)—LLMs remain trapped in T2, operating on text alone.

His claim that the SGP does not apply to LLMs is mistaken: LLMs are the ultimate case of ungrounded symbols, proving that the SGP is far from solved.

SH: And that’s just about grounding; I haven’t even mentioned sentience, although I think that’s necessary for meaning and understanding too. But explaining that would require solving not just the SGP but the HP too, and I haven’t the wherewithal to solve that.

GPT: So far, we’ve only discussed grounding—the necessity of sensorimotor interaction (T3) with referents for meaning and understanding. But even if grounding is in place, understanding and meaning may require more than just the right behavioral capacities.

That’s where sentience comes in. If meaning is not just about the right behavioral dispositions but about what it feels like to mean something, then solving the Symbol Grounding Problem (SGP) is not enough—we would also need to solve the Hard Problem (HP):

The SGP explains how symbols get their referents via sensorimotor interaction.

The HP asks why and how this process generates subjective experience (sentience)—which may be necessary for true understanding.

You’ve long argued that solving the Easy Problem (EP) (explaining all functional and behavioral capacities) will use up all the causal degrees of freedom—leaving nothing left to explain why cognition feels like something.

So, while Gubelmann’s argument fails even at the grounding level (SGP), the deeper question of whether grounding alone suffices for meaning and understanding remains open—because we don’t yet know why or whether cognition must feel like something to mean something. And, as you’ve said, we don’t have the wherewithal to solve that.

SH: At least I don’t. Om tat sat…