Modern computing has achieved a high level of linguistic fluency. Users interact with chat bots that produce grammatically correct and stylistically diverse text. However, a fundamental question remains: does artificial fluency signify artificial understanding?
Foundational Concepts
To analyze the evolution of AI, we must define three basic terms:
Syntax: The arrangement of words and phrases to create well-formed sentences in a language. It is the “grammar” or the “rules” of the string.
Semantics: The study of meaning in language. It is not about the rules of the sentence, but what the sentence refers to in reality.
Distributional Semantics: A technique in computational linguistics where the meaning of a word is estimated based on the words that frequently appear near it in a large database.
The Barrier of Logical Probability
Most current Large Language Models (LLMs) operate on the principle of logical probability. They analyze patterns in vast datasets to predict the next character or word in a sequence. This is a “syntax-heavy” approach. The machine identifies that the word “Microcontroller” is often followed by “Firmware,” but it does not have a physical or conceptual model of what firmware is. It lacks grounding.
Grounding is the connection between a linguistic symbol (a word) and an external reality (an object or concept). Without grounding, an AI is what researchers call a “Stochastic Parrot.” This term describes a system that haphazardly stitches together sequences of linguistic forms according to probabilistic information about how they combine, without any reference to meaning (Bender et al., 2021).
The Shift to Knowledge-Rich Architecture
How do we move beyond simple probability? The answer lies in Knowledge-Rich architecture. This approach requires two primary components:
- An Ontology: A structured, fact-based map of the world. It is a hierarchy of concepts. In an ontology, the machine does not guess; it “knows” that a Sensor is a physical device that Produces data.
- Text Meaning Representation (TMR): A formal, unambiguous translation of a natural language sentence into a machine-readable format based on the ontology.
When a human says, “The bank is deep,” they are referring to a river. When they say, “The bank is closed,” they are referring to a financial institution. A statistical model might guess the meaning based on surrounding words. A Language-Endowed Intelligent Agent (LEIA) uses its ontology to verify which meaning is logically possible in the current context (McShane, 2023).
Actionability and Transparency
A machine that truly understands language must demonstrate actionability. This means the linguistic input must translate into a precise digital or physical action. If a user tells a system to “Optimize the power consumption of the ESP32,” the system must understand the specific electrical parameters of that hardware. It cannot rely on word associations.
Furthermore, knowledge-rich systems provide transparency. Because the machine uses a formal ontology, it can explain its decision path. It can show the specific concepts it used to interpret a command. Statistical models are often “black boxes” where the reason for a specific output is hidden within millions of mathematical weights.
Conclusion
The evolution of AI requires a return to linguistic fundamentals. Fluency is useful, but it is not a substitute for the structured representation of reality. To move from machines that “talk” to machines that “understand,” we must prioritize the integration of complex ontologies and deep semantic analysis over simple statistical probability.
References
-
Bender, E.M., Gebru, T., McMillan-Major, A. and Shmitchell, S. (2021). “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?”. Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency.
Available at: https://dl.acm.org/doi/10.1145/3442188.3445922 -
Marcus, G. (2020). “The Next Step for AI: Robust Artificial Intelligence”. Communications of the ACM, 63(10), pp.44-46.
Available at: https://cacm.acm.org/magazines/2020/10/247596-the-next-step-for-ai/fulltext -
McShane, M. (2023). “Linguistic Knowledge for Language-Endowed Intelligent Agents”. ArXiv Preprints.
Available at: https://arxiv.org/abs/2301.05603 - McShane, M. and Nirenburg, S. (2021). Linguistics for the Age of AI. Cambridge, MA: MIT Press.
Author: Ing. Anfilov
No commentaries yet. Be the first to share your thoughts.