AI Doesn't Hallucinate. It Makes Mistakes.
Calling AI errors “hallucinations” humanizes machines and inflates expectations. Language is the UI for trust; misuse becomes a shipped bug with churn, support cost, and legal risk. Treat wording like code: define terms, show process, and label errors precisely.
Why the Difference is a Multi-Million Dollar Problem.
A Thoughtful Question
A thoughtful comment on a recent post of mine posed a brilliant question:
When we say an AI "hallucinates," is it a helpful metaphor or a dangerous distortion?
I love this question because it’s subtle, and the danger is completely unintentional.
This isn't about AI claiming to be human – it's about us giving it human qualities. And because we're the ones doing it, it feels safer.
It’s not.