Human-like Working Memory Interference in Large Language Models

This paper reveals that pretrained large language models exhibit human-like working memory limitations caused not by a lack of context access, but by a shared computational constraint where entangled memory representations require active interference control to successfully recall task-relevant information.

Hua-Dong Xiong (School of Psychological and Brain Sciences, Georgia Tech), Li Ji-An (Department of Psychology, New York University), Jiaqi Huang (Department of Cognitive Science, Indiana University Bloomington, Honda Research Institute), Robert C. Wilson (School of Psychological and Brain Sciences, Georgia Tech, Center of Excellence for Computational Cognition, Georgia Tech), Kwonjoon Lee (Honda Research Institute), Xue-Xin Wei (Departments of Neuroscience and Psychology, The University of Texas at Austin)

Published 2026-04-14
📖 6 min read🧠 Deep dive

The Big Question: Why Do Super-Computers Forget?

Imagine you have a library with 100 billion books (neurons/parameters). You can walk into any aisle and pull out any book instantly. You have a perfect map of the entire library.

Now, imagine someone asks you: "What was the third book I asked for 10 minutes ago?"

Even with your perfect library, you might struggle if they ask for the 10th book, then the 20th, then the 30th, all while talking about random topics in between. You might mix them up.

This is the mystery the paper solves. Large Language Models (LLMs) like the ones you chat with have massive "libraries" (they remember everything you typed earlier in the chat). They have "superpowers" to look back at any part of the conversation instantly. So, why do they still get simple memory games wrong?

The answer isn't that they can't see the information. It's that they get distracted by the noise.


The Game: The "N-Back" Challenge

To test this, the researchers played a game called N-Back with AI models.

  • The Rules: The AI is shown a stream of letters: A, B, C, D, E...
  • The Task: If the game is "2-Back," the AI must say the letter that appeared two steps ago.
    • Input: A, B, C, D
    • Output: - , - , A, B (Because A was two steps before C, and B was two steps before D).

The Surprise:
When the researchers trained a tiny, simple AI just to play this game, it got 100% perfect.
But when they tested the giant, famous AI models (like Qwen, Gemma, Llama), they got worse and worse as the game got harder (e.g., remembering the letter from 4 steps ago instead of 2).

Even though these giant models have access to the entire history of the chat, they act like humans who are easily distracted.


The Real Culprit: "Representational Interference"

The paper argues that the problem isn't storage (running out of space); it's interference (too much noise).

🧠 The Analogy: The Radio Station

Imagine your brain (or the AI) is a radio.

  • The Target: You want to listen to "Station 2" (the letter from 2 steps ago).
  • The Problem: "Station 1" (the letter you just saw) and "Station 3" (the letter you saw 3 steps ago) are broadcasting at the same time, and they are all playing on the same frequency.

Because the AI stores all these letters in a "mixed-up" way (entangled representations), when it tries to tune into "Station 2," the signal from "Station 1" is so loud and similar that it drowns out the target.

The AI doesn't forget the letter; it just can't find it in the mess.

The Evidence: How We Know It's Interference

The researchers found three "smoking guns" that prove the AI is getting confused by noise, not just losing data:

  1. The "Recency" Bias:
    When the AI makes a mistake, it almost always guesses the most recent letter instead of the one from 2 steps ago. It's like a human saying, "I know I asked for the red one, but you just said blue, so maybe it's blue?" The most recent memory is too loud.

  2. The "Lure" Effect:
    If the AI sees a letter that looks like the one it's supposed to remember, it gets confused.

    • Example: If the target is X, and the AI just saw X again, it might accidentally grab the new X instead of the old X. The content of the letters interferes with the position.
  3. The "Smart" AI is the "Distracted" AI:
    Here is the coolest part: The models that are generally "smarter" (better at math, logic, and writing) are the ones that perform best at this memory game.

    • Why? Being "smart" in this context means being good at ignoring distractions. The AI that can successfully "mute" the irrelevant letters to find the right one is the same AI that is good at reasoning.

How the AI "Thinks" (The Mechanism)

The researchers looked inside the AI's "brain" (its layers) to see how it solves the problem. They found a common pattern:

  1. Layer 1-5 (The Chaos): The AI holds all the letters in a big, jumbled pile. The target letter is mixed in with the noise.
  2. Layer 10-20 (The Filter): The AI starts to suppress the irrelevant letters. It's like a DJ slowly turning down the volume on the wrong stations.
  3. Layer 25+ (The Clarity): Finally, near the end, the target letter becomes loud and clear, and the AI outputs the answer.

The Catch: This filtering process is hard work. If there are too many letters (high memory load), the filter gets overwhelmed, and the noise leaks through.

The "Magic Fix" Experiment

To prove that "noise" was the problem, the researchers did a surgery on the AI.

  • They took the "letter identity" information (the fact that a token is the letter 'A') and silenced it in the middle of the AI's processing.
  • Result: The AI actually got better at the game!
  • Why? By removing the "noise" of the specific letters, the AI had an easier time finding the target position. It proved that the AI was indeed struggling because the letters were fighting each other.

The Big Takeaway

Humans and AI share a common weakness.

Even though humans have biological brains and AI has silicon chips, we both face the same computational challenge: How do you pick the right thing out of a pile of similar things?

  • Old Idea: We forget because we run out of "RAM" (storage space).
  • New Idea: We forget because we can't filter out the interference.

The paper suggests that to make AI smarter, we shouldn't just give them bigger libraries (more context windows). Instead, we need to teach them better noise-canceling headphones—ways to actively suppress irrelevant information so the important stuff can shine through.

In short: The problem isn't that the AI can't see the past; it's that the past is too loud, and the AI needs to learn how to tune it out.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →