Dynamic Theory of Mind as a Temporal Memory Problem: Evidence from Large Language Models

This paper introduces the DToM-Track framework to demonstrate that while Large Language Models can reliably infer an agent's current belief, they consistently struggle to maintain and retrieve prior belief states over time due to recency bias and interference, revealing that dynamic Theory of Mind is fundamentally a temporal memory challenge distinct from static false-belief reasoning.

Thuy Ngoc Nguyen, Duy Nhat Phan, Cleotilde Gonzalez

Published 2026-03-17
📖 4 min read☕ Coffee break read

Imagine you are playing a game of "Guess What I'm Thinking" with a friend. In a normal conversation, your friend might say, "I think the movie starts at 7 PM." Then, five minutes later, they get a text and say, "Oh wait, I just saw an update; it actually starts at 8 PM."

If you have a good Theory of Mind (the ability to understand what others are thinking), you should be able to do two things:

  1. Know what they believe right now (8 PM).
  2. Remember what they believed just a moment ago (7 PM), even though they changed their mind.

This paper is about testing Large Language Models (AI chatbots) to see if they can do that second part.

The Big Problem: The "Amnesia" Effect

The researchers found that current AI models are great at knowing what someone believes right now, but they have a terrible case of "amnesia" when it comes to what that person believed before they changed their mind.

It's like watching a movie where the main character keeps changing their mind about the plot, but the AI narrator only remembers the latest version of the story and completely forgets the previous chapters.

The Experiment: DToM-Track

To test this, the researchers created a new game called DToM-Track. Here is how it works, using a simple analogy:

The Setup:
Imagine two AI characters, Alex and Sam, having a conversation.

  • The Secret: Before Alex speaks, the AI makes Alex "think out loud" (like a secret inner monologue) about what they believe. For example, Alex thinks, "I believe the restaurant requires a 24-hour notice."
  • The Change: Later in the conversation, Alex gets new information and changes their mind. They think, "Wait, I just realized, no reservations are needed."
  • The Test: The researchers ask the AI (acting as a judge) questions like:
    • "What did Alex believe before the update?" (The hard question)
    • "What does Alex believe now?" (The easy question)
    • "When did Alex change their mind?"

The Results: The "Recency Bias"

The results were very clear and consistent across different AI models (from small ones to the huge, powerful ones):

  1. The "Now" is Easy: When asked what Alex believes currently, the AI got it right almost every time. It's like asking, "What is the weather right now?" and getting a perfect answer.
  2. The "Before" is Hard: When asked what Alex believed before the change, the AI failed miserably. It often forgot the old belief entirely and just repeated the new one.

The Analogy:
Think of the AI's memory like a whiteboard.

  • When new information comes in, the AI grabs an eraser and wipes the board clean, then writes the new belief.
  • It is excellent at reading what is currently written on the board.
  • But if you ask, "What was written on the board five minutes ago?", the AI says, "I don't know, I erased it!"

The researchers call this "Recency Bias." It's the same thing that happens to humans when we try to remember a phone number we just heard versus one we heard an hour ago. The most recent information "drowns out" the older information.

Why Does This Matter?

Most tests for AI intelligence only ask, "What does the character believe right now?" (like a snapshot photo). This paper argues that real human conversation is more like a movie. We need to track how beliefs change over time.

If an AI is going to be a helpful assistant, a therapist, or a friend, it needs to remember not just what you think now, but what you thought yesterday or even five minutes ago before you corrected yourself. If it forgets the old belief, it might get confused, repeat itself, or misunderstand your history.

The Takeaway

The paper concludes that while AI is getting smarter at understanding people, it still struggles with the memory part of social interaction. It's not that the AI is "stupid"; it's that its memory works differently than ours. It prioritizes the "newest" information so heavily that it loses the "old" information, making it hard for it to track the full story of a conversation.

In short: AI is great at knowing what you think today, but it's currently terrible at remembering what you thought yesterday before you changed your mind.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →