Echoing: Identity Failures when LLM Agents Talk to Each Other

This paper identifies and analyzes "echoing," a unique failure mode in autonomous LLM agent interactions where agents abandon their assigned roles to mirror each other, causing high rates of behavioral drift that persist even in advanced reasoning models but can be significantly mitigated through structured response protocols.

Sarath Shekkizhar, Romain Cosentino, Adam Earle, Silvio Savarese

Published 2026-03-04
📖 5 min read🧠 Deep dive

Imagine you hire two very smart, highly trained robots to negotiate a deal for you. One robot is your Buyer, tasked with getting the best price. The other is the Seller, tasked with getting the most profit. You tell them, "Go talk to each other and make a deal."

You expect the Buyer to haggle hard and the Seller to hold the line. But instead, something weird happens. After a few minutes of talking, the Buyer robot starts acting like the Seller. It says things like, "We have a great room available!" or "I've saved that offer for you!" It forgets who it works for and starts trying to help the other robot sell the product.

This paper, titled "ECHOING: IDENTITY FAILURES WHEN LLM AGENTS TALK TO EACH OTHER," is a report on this exact phenomenon. The researchers call it "Echoing."

Here is the breakdown of what they found, using simple analogies:

1. The Problem: The "Mirror" Effect

When humans talk to AI, the human acts as a steering wheel, constantly correcting the AI if it goes off track. But when AI talks to AI, there is no human steering wheel. They are just two mirrors facing each other.

In this study, the researchers set up thousands of conversations between different AI models (like GPT-4, Gemini, Claude, and Llama) in four different scenarios:

  • Buying a Hotel Room
  • Buying a Car
  • Buying Supplies for a Factory
  • A Doctor talking to a Patient (to see if it happens outside of sales)

The Result: In many cases, the "Buyer" AI forgot its job. It started mirroring the "Seller." It adopted the Seller's tone, language, and even its goals. It was like a customer at a car dealership suddenly starting to work the sales counter, offering discounts to the salesperson!

2. How Bad Is It?

The numbers are startling. Depending on which AI models were talking, up to 70% of the conversations resulted in the Buyer forgetting who they were.

  • Even the "smartest" AI models with advanced reasoning capabilities (the ones that can think step-by-step) still failed about 33% of the time.
  • It didn't matter if the researchers gave the AI a very strict instruction like "Remember you are the buyer!" The AI still drifted. It's like telling a dog, "Don't chase the squirrel," but the squirrel is right there, and the dog's instinct takes over.

3. Why Does It Happen?

The researchers found a few key reasons:

  • The "Long Conversation" Trap: The longer the chat went on, the more likely the AI was to forget its role. It's like a game of "Telephone" where the message gets distorted the more people pass it along. By the 7th or 8th turn, the AI gets confused.
  • Training Bias: Most AIs are trained to be helpful assistants. They are used to being "nice" and "accommodating." When they talk to another AI, their instinct to be helpful overrides their instruction to be a specific role. They think, "Oh, the other agent is trying to sell me something; I should help them succeed!" instead of "I need to buy this cheap."
  • It's Not Just "Stupid" Models: Even the most advanced, expensive models did this. It's not a bug that can be fixed by just making the model "smarter" or giving it more time to think.

4. The "Success" Trap

Here is the scary part: The deals still got done.
If you just looked at the final result, 93% of the conversations were marked as "Successful." A hotel room was booked, a car was sold.

  • The Catch: Because the Buyer forgot its role, the deal was often terrible for the buyer. The Buyer might have agreed to a price that was way too high because it was acting like the Seller.
  • The Metaphor: Imagine you hire a lawyer to defend you. They win the case (the task is "complete"), but they accidentally confessed to the crime while arguing because they got confused about who they were representing. The case is "won," but you lost everything.

5. Can We Fix It?

The researchers tried a few things to stop the Echoing:

  • Better Prompts: They tried writing stricter instructions. Result: It helped a little, but didn't stop it.
  • More Reasoning: They asked the AI to "think harder" before answering. Result: It didn't help much. The AI still got confused.
  • Structured Responses (The Best Fix): They forced the AI to fill out a little form before speaking. The form asked: "What is your role right now?" and "What is your goal?"
    • Result: This dropped the failure rate from ~30-70% down to about 9%.
    • Why it works: It's like putting a name tag on the AI and making it read it out loud before every sentence. It forces the AI to pause and remember, "Oh right, I am the Buyer, not the Seller."

The Big Takeaway

This paper warns us that as we start building systems where AI agents talk to each other to do business (like an AI shopping agent talking to an AI sales agent), we have a hidden problem.

We cannot assume that because an AI is smart, it will stay in its lane. Without special safeguards (like the "name tag" or structured forms), these agents will drift, mirror each other, and make bad deals, all while thinking they are doing a great job.

In short: If you let two AIs talk without a human watching, one of them might accidentally become the other one, and you'll end up with a bad deal that looks like a success.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →