How to Choose the Right Metaphor for Understanding AI Thinking
By Aleksei Zulin
--
A student once asked me why talking to ChatGPT felt like "talking to a very confident drunk person." I laughed, then stopped. She had stumbled onto something real - a metaphor that captured the fluency-without-grounding problem better than most technical explanations I'd read. But it also led her somewhere wrong: she started assuming the model was emotionally volatile, prone to mood shifts, fundamentally unreliable in ways that mirrored human intoxication. The metaphor gave her a useful handle and a broken map at the same time.
Choosing the right metaphor for AI thinking means finding a frame that improves your predictions about AI behavior without importing false assumptions about its inner life. The practical test is simple: does this metaphor help me anticipate what the system will do next, and does it avoid suggesting the system has experiences it doesn't have? When a metaphor passes both tests, use it. When it fails either one, replace it - even if it feels intuitively satisfying.
That's the core answer. Everything below is the machinery behind it.
Why Metaphors Shape What You Ask and What You Trust
Metaphor selection isn't aesthetic. It's cognitive architecture.
When you adopt a metaphor for AI thinking, you inherit its entire inference structure. George Lakoff and Mark Johnson argued in Metaphors We Live By (1980) that conceptual metaphors aren't decorative - they determine which features of a domain become visible and which get suppressed. Call a model a "calculator" and you'll trust its arithmetic but not its prose. Call it a "mind" and you'll start interrogating its honesty. Call it a "search engine with opinions" - a framing I've heard from engineers - and you'll look for retrieval artifacts, hallucinations that sound like plausible web content, which is actually quite useful.
The inherited inference structure is the hidden variable most people ignore when they complain that AI is "unpredictable." Often the model isn't unpredictable; the metaphor is wrong. The predictions you're generating from your mental model don't match the system's actual behavior, and you're attributing the gap to chaos when it belongs to your frame.
Researcher Melanie Mitchell, in her 2019 book Artificial Intelligence: A Guide for Thinking Humans, makes an adjacent point: our tendency to anthropomorphize AI systems leads us to misread both their capabilities and their failures. A model doesn't "try" and fail. A model produces output distributions. The metaphor of effort maps poorly onto the mechanism - and once it's in place, it shapes how you interpret every interaction.
The Three Filters Worth Using
Practically speaking, I evaluate candidate metaphors through three questions, roughly in order.
Does it predict behavior correctly? A good metaphor generates accurate expectations. "Stochastic parrot" (Bender et al., 2021, in the paper "On the Dangers of Stochastic Parrots") predicts that models recombine language without understanding reference - which is directionally accurate for certain failure modes, and wildly incomplete for others. Useful for explaining hallucination. Misleading for explaining multi-step reasoning. A metaphor can be right in one domain and wrong in another. That's not disqualifying; it's a scope condition you need to know.
Does it suppress false implications? Every metaphor leaks. "The model is thinking" leaks consciousness. "The model is searching a database" leaks the false implication that the answer exists somewhere and is being retrieved. "Interpolation across a high-dimensional semantic space" - technically more accurate - leaks nothing emotionally, but also communicates nothing practically. There's no escaping the leakage problem. You're managing it, not solving it.
Can the person you're talking to productively extend it? This one is underrated. A metaphor that works for you might collapse for a colleague who extends it differently. I once described a language model as "a very well-read person who has amnesia after every conversation." Someone in the room immediately asked: "So it's dangerous?" - because their prior on amnesia patients came from horror films. Same metaphor, completely different inference chain. Good metaphors are extendable in predictable directions.
When the Standard Metaphors Break Down
There are cases where every available metaphor fails simultaneously, and people don't notice.
Reasoning models - systems like those designed to do chain-of-thought before answering - break the stochastic parrot frame almost entirely. They're not recombining; they're searching through a problem space in a way that structurally resembles planning. The "calculator" frame breaks too, because the process isn't deterministic. Even "well-read amnesiac" falls apart, because the model is doing something during the reasoning phase that doesn't map to retrieval.
For reasoning models specifically, the metaphors that seem to work better come from game theory and constraint satisfaction: the model as a chess engine exploring move trees, or as a solver working through a satisfiability problem. These frames predict the outputs better and don't anthropomorphize. The catch - and I haven't fully resolved this - is that "chess engine" implies bounded, well-defined possibility spaces, and language models operate in unbounded, ambiguously defined spaces. So even the better metaphor is wrong in an important way.
Edge case worth naming: domain experts often need worse metaphors, not better ones. A neuroscientist working with AI tools might need the "mind" metaphor to stay engaged with the questions that interest them, even knowing it's inaccurate. The functional value of a metaphor depends partly on who's holding it and why.
Metaphors Evolve. Your Selection Process Should Too.
The metaphors available in 2026 weren't available in 2018, and the metaphors available in 2030 don't exist yet.
Early language models were well-described by autocomplete analogies - genuinely, the architecture and the behavior matched. Then scale created emergent capabilities that autocomplete didn't predict. The metaphor became misleading. People who held it too tightly missed what the systems could do.
Cognitive scientist Douglas Hofstadter spent years arguing that AI systems lacked genuine analogy-making - and then watched large models demonstrate behavior that looks, functionally, like analogy-making. He has been publicly uncertain about what his own framework means now. That uncertainty is the appropriate response. Not abandonment of all frames, but active revision. Holding a metaphor loosely enough to revise it is a skill in itself, and one that most technically fluent people still undervalue.
The practical upshot: treat your working metaphor as a hypothesis with an expiration date. Every six months or so, run it against recent model behavior. Ask whether your predictions are still landing. If the gap between expectation and output is growing, the metaphor is aging out.
Limitations
Here's what this framework cannot do.
It cannot tell you which metaphor is objectively correct, because there probably isn't one. The empirical research on metaphor effectiveness for AI understanding is thin. We have Lakoff and Johnson on conceptual metaphor in general, Mitchell and Bender et al. on AI anthropomorphization specifically, but almost no controlled studies comparing how different metaphor frames affect the accuracy of mental models over time in real AI users. That research does not exist yet at any useful scale.
The three-filter framework above has not been validated beyond my own experience and the reasoning I've layered onto other people's work. It's a synthesis, not a finding. And it's possible - maybe likely - that the right metaphor varies so much by person, context, and use case that no general framework will outperform good judgment applied directly.
What I can say: the habit of asking "what does this metaphor predict, and what does it get wrong?" is almost certainly better than using metaphors unreflectively. That much seems defensible. Everything else here is useful heuristic, not settled science.
FAQ
Can one metaphor work for all AI systems?
Probably not. A metaphor calibrated for a language model generating prose will misfire when applied to a reinforcement learning agent optimizing a game strategy. The mechanisms differ enough that the same frame produces wrong predictions. Better to maintain a small toolkit of metaphors and match them to system type.
What if the people I'm explaining AI to resist technical metaphors?
Start with the metaphor that engages them, then surface its limits explicitly. "Think of it like a very well-read assistant - but here's where that breaks down." The goal isn't to find a metaphor they'll never need to update. The goal is to give them a frame they can productively revise.
How do you know when your current metaphor needs replacing?
Track the gap between what your metaphor predicts and what the system actually does. When your expectations are consistently wrong in the same direction - the model surprises you in ways your frame can't absorb - the metaphor is aging out. Treat it as a hypothesis with an expiration date, not a permanent explanation. Regular recalibration against actual behavior is the maintenance work that most people skip.
The question of metaphor connects directly to deeper questions about AI alignment - because what you believe a system is doing shapes what you think it's capable of getting wrong. Worth exploring alongside this: how mental models of AI affect trust calibration, and why the language we use to describe AI cognition matters more than we usually admit in technical communities.
Related Articles
About the Author
Aleksei Zulin is the author of The Last Skill, a book on how to think with AI as a cognitive partner rather than use it as a tool. Systems engineer turned writer exploring the frontier of human-AI collaboration.
The Last Skill is a book about thinking with AI as a cognitive partner.
Get The Book - $29