Key insights of today’s newsletter:
Ludwig Wittgenstein, the Austrian philosopher, wrote extensively about language and its intricacies.
His ideas are especially relevant for anyone working in AI, because his views on language help us better understand the limitations of large language models.
This essay explores those ideas, what you can take away from them, and why learning the meaning of words doesn’t necessary equate to learning something about the world.
↓ Go deeper (10 min read)
This might be a bit of an odd question, but does someone who knows the meaning of all the words know anything about the world? And where do new words come from?
In the story of Genesis, God creates the Garden of Eden and brings all the animals to Adam to see what he will name them. Adam names all the cattle, the birds in the sky, and even the wild animals.
It’s a very human thing to do. We point to things and give it a name. It’s most likely how early language evolved. First came words, then sentences and then whole stories — a process that took us over 100,000 years — which ultimately allowed us to communicate increasingly complex ideas, from religion to human rights to the concept of the nation state. In a way, language is the first and most powerful tool ever created.
In Philosophical Investigations (1953), published posthumously, the philosopher Ludwig Wittgenstein investigates the relationship between words and their meaning. To grasp the meaning of a word, he writes, you have to look at its use. He argues that not only have words meaning in relation to other words, the meaning of a word changes through its use. Language games, Wittgenstein called them, or Sprachspiel in German.
Take the word: nice. It originally comes from the Latin word nescius, which meant “ignorant”. The word entered the English language through Middle French in the 14th century, and was used to describe someone stupid or foolish. By the 18th century, however, the meaning shifted to describe someone who was “agreeable, “thoughtful,” or “kind.” Pretty much the opposite of what it used to mean.
Or what about words that have no equivalent in another language? The word Tsundoku (Japanese), for example, describes the act of acquiring books and letting them pile up unread. Or Ya'aburnee (Arabic), an expression that roughly translates to “You bury me”, which is used to express profound love for someone; a love that is so all-consuming that losing this person would literally bury us.
What these examples show is that words and their meaning aren’t static. Language is alive. It’s actively being shaped by us, and surprisingly, it plays a crucial role in understanding AI.
“You shall know a word by the company it keeps”
In 2019, published her book Artificial Intelligence: A Guide for Thinking Humans, in which she writes:
“The NLP research community has proposed several methods for encoding words in a way that would capture such semantic relationships. All of these methods are based on the same idea, which was expressed beautifully by the linguist John Firth in 1957: “You shall know a word by the company it keeps.”
That is, the meaning of a word can be defined in terms of other words it tends to occur with, and the words that tend to occur with those words, and so on. Abhorred tends to occur in the same contexts as hated. Laughed tends to occur with the same words that humor finds in its company.”
Large language models are the latest and by far the most ambitious attempt at teaching computers how to talk. As Mitchell explains it, it’s inspired by this very simple yet powerful idea: what if we could calculate all the ways in which words appear next to each other?
By ingesting massive amounts of text data, LLMs are trained to do exactly that. It works according to a relatively simple learning algorithm that predicts the next token (i.e. word or part of a word). I’ll spare you the technical details, but after training (a process which can take several months) you end up with a so-called ‘pre-trained model’ that has worked out all the combinations in which words appear next to other words, and how they relate. You can see it as a language map. A map that allows the LLM to navigate and respond to a wide range of queries or ‘prompts’, simply by predicting what’s likely to come next.
A common misconception about LLMs is that they get smarter over time. explained this recently on his blog:
ChatGPT and other similar tools do not directly learn from and memorize everything that you say to them. This can be quite unintuitive: these tools imitate a human conversational partner, and humans constantly update their knowledge based on what you say to to them.
(…)
In the case of a “conversation” with a chatbot such as ChatGPT or Claude or Google Gemini, that function input consists of the current conversation (everything said by both the human and the bot) up to that point, plus the user’s new prompt.
Every time you start a new chat conversation, you clear the slate. Each conversation is an entirely new sequence, carried out entirely independently of previous conversations from both yourself and other users.
Understanding this is key to working effectively with these models. Every time you hit “new chat” you are effectively wiping the short-term memory of the model, starting again from scratch.
The key takeaway here is that any knowledge these models have distilled during their training process is pretty much frozen in time.
“The map is not the territory”
From that, an interesting question arises: when a LLM creates this language map, does it also learn something about the world?
The answer is not obvious. Philosopher Alfred Korzybski famously said “the map is not the territory” and “the word is not the thing.” According to Korzybski, many people confuse maps with territories, that is, confuse conceptual models of reality with reality itself. It shares some resemblance with Wittgensteins ideas, which is not entirely surprising, given that Korzybski has not only has read Wittgenstein’s work, but even admitted he and his wife Mira loved it so much they would read from it to each other in bed.
I say this because some of the most renowned AI engineers that pioneered today’s foundation models — Karpathy, Sutskever and others — are convinced that LLMs have internalized a ‘world model’. It must have, they argue, given what they’re are capable of.
At the same time, we also know LLM-powered assistants like ChatGPT, Gemini and Grok can make very, very silly mistakes. Here’s an example shared by cognitive scientist and author in a recent post:
Here’s another one I was able to replicate myself:
Or how about this one by professor and researcher Subbarao Kambhampati, referenced in a recent New York Times piece:
How should we understand this?
Because of their static nature, LLMs don’t learn from their interactions; or, as Wittgenstein would say, they don’t participate in Sprachspiel. So even if ChatGPT gets it right the second time, it hasn’t learned anything from that back-and-forth. Hit “new chat”, and you’re back to square one.
That’s very different from us. A child can stack a tower of blocks before it learns to speak about it. As humans, we update our knowledge continuously; we are in constant conversation with the world. Both Wittgenstein and Korzybski would probably agree that although the AI has learned all the words in the world, it hasn’t learned anything about how these words relate to the world.
Another way of looking at it is that LLMs are booksmart. They’ve read more books than all of us combined, yet never felt the weight of one in their hand. They can write a ten-page essay about the Golden Retriever, but never felt the joy of coming home to one.
This all ties in beautifully with a concept that I’ve previously described as the Intelligence Paradox. Humans treat language fluency as a proxy for intelligence. As a result of that, we tend to overestimate the capabilities of AI assistants like ChatGPT; because they’re smooth talkers. But we should never forget that to an LLM a truthful statement and a hallucinated response look completely identical. To them, it’s just words. Words that are likely to appear next to other words.
They’re really good map-readers, but never set foot in the garden.
Join the conversation 🗣
Share, like or leave a comment if it resonated with you.
I think the argument that they hallucinate does not in fact mean that they don't build a world model - I mean, look at this:
https://www.anthropic.com/research/mapping-mind-language-model
This is a world model of some sort. Hallucinations just mean that the world model can be wrong, and I think that even in the Claude example, it does some odd things like relating "coding error" to "food poisoning" in conceptual space.
But I think I go with Hinton here that it is grasping some sort of meaning(enough to scare me, obviously) and perhaps this should be seen as a matter of degree, with errors. Of course, there are people who also claim that LLMs are discovering Platonic truth, or at least converging to something(maybe all of the same biases?).
https://arxiv.org/pdf/2405.07987
This makes me wonder if there are ways that this questionable relationship to grounded reality can be used as a feature, rather than a bug.
I ran my own, very small, experiment by writing a few nonsense files and adding them to a knowledge base for a GPT4All instance on my machine. Sure enough, when prompted for ideas whose representations were in proximity to the nonsense files, the AI gave me the same nonsense I had written, rephrased a little bit. But maybe, if I could curate a particular representation of a "world" (fictional or just specialized) - perhaps then using the LLM's limited perspectives as a tool could be helpful, rather than risky? If it was described as, and used on purpose, as a figurative map rather than interpreted as the territory.