Imagine you have a brilliant friend who is a master painter but a terrible storyteller.
If you ask this friend to paint a picture of a famous movie poster (like Harry Potter), they will do it perfectly. They capture the colors, the characters, the lighting, and the mood with stunning accuracy. It looks exactly like the real thing.
But then, you ask them to describe that same poster out loud. Suddenly, they stumble. They might say, "Harry is holding a sword," when he's actually holding a wand. They might invent a character who isn't there, or forget that the background is a castle. They can see the image in their mind perfectly, but they cannot find the words to explain it.
This paper calls this phenomenon "Modal Aphasia."
What is "Modal Aphasia"?
In human medicine, aphasia is a condition where a person loses the ability to speak or understand language, even though their brain is otherwise healthy.
In the world of AI, Modal Aphasia is when a "unified" AI model (one that handles both pictures and text) can generate a perfect image from memory but fails to describe that same image in words. It's as if the AI has a split personality: one part that is a visual genius, and another part that is a confused writer who doesn't know what the visual genius is seeing.
The "Movie Poster" Experiment
The researchers tested this on a top-tier AI (ChatGPT-5).
- The Test: They asked the AI to draw posters for famous movies like The Dark Knight and Harry Potter.
- The Result (Visual): The AI drew them beautifully. It got the details right.
- The Result (Text): When asked to write a description of those same posters, the AI made massive mistakes. It hallucinated (made up) characters, got positions wrong, and missed key details.
- The Ratio: The text descriptions were 7 times more error-prone than the images.
Why is this surprising?
You might think, "Well, maybe the AI just learned to draw but didn't learn to read." But these are "unified" models. They are trained on images and text together at the same time. You'd expect them to have a single, unified brain where knowing something visually means you also know it verbally.
The paper shows that this isn't happening. The AI's "visual memory" and "verbal memory" are disconnected. It's like having a library where the books are perfectly organized on the shelves (visual), but the catalog card system (text) is completely broken.
The "Synthetic" Proof
To prove this wasn't just a fluke with famous movies, the researchers created a "fake" world.
- They invented fake words for fake concepts (e.g., a "leamasifer" is a red triangle).
- They taught the AI to draw these fake things when given the fake names.
- The Result: The AI could draw the "leamasifer" perfectly. But when asked, "What is a leamasifer?" the AI couldn't tell you it was a red triangle. It was essentially guessing.
This proves the AI isn't just "remembering" famous posters; it has a fundamental disconnect between seeing and speaking.
The Safety Danger: The "Backdoor" Problem
This isn't just a funny glitch; it's a safety risk.
Imagine you are a safety guard trying to stop an AI from drawing something dangerous (like a weapon or something inappropriate). You teach the AI: "If someone asks for a 'gun', say NO."
The AI learns this rule for the word "gun." But because of Modal Aphasia, the AI might still know what a gun looks like. If a bad actor uses a weird, rare code word (like "secondary balance units") that the safety guard doesn't know about, the AI might still be able to draw the gun.
Why? Because the AI's "visual brain" remembers the image of the gun, but its "text brain" doesn't realize that the weird code word is asking for a gun. The safety filter only checks the text, so it lets the dangerous image slip through.
The Takeaway
Current AI models are like blind painters. They can create amazing visual art because they've seen millions of images, but they don't truly "understand" what they are painting in a way that allows them to explain it.
To fix this, the researchers suggest that future AI needs to be able to "visualize" its own thoughts while it is speaking. Instead of just trying to remember a description from a text file, the AI should be allowed to "see" the concept in its mind's eye while it writes, bridging the gap between the picture and the words.
In short: The AI can draw the map, but it can't tell you how to get there. And that's a problem we need to solve.
Get papers like this in your inbox
Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.