Calibrating Behavioral Parameters with Large Language Models

This paper proposes a framework to use large language models (LLMs) as calibrated measurement instruments for behavioral finance parameters, demonstrating that profile-based calibration can correct systematic rationality biases in LLMs to produce agent-based models that accurately replicate empirical asset pricing patterns.

Original authors: Brandon Yee, Krishna Sharma

Published 2026-04-27
📖 4 min read☕ Coffee break read

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

The Big Idea: Turning AI into a "Psychology Lab in a Box"

Imagine you are a scientist trying to study how humans make mistakes with money. To do this, you usually have to hire hundreds of people, pay them to sit in a room, and hope they don't get bored or lie to you. It’s slow, expensive, and messy.

This paper proposes a radical shortcut: Instead of using humans to study human mistakes, let’s use AI as a high-tech "tuning knob" to simulate those mistakes.

The researchers aren't trying to see if AI is human. Instead, they are treating the AI like a scientific instrument—sort of like a thermometer. A thermometer doesn't "act" like heat; it just reacts to it in a predictable way so you can measure it. The researchers found that if they "tell" the AI to act a certain way (using specific instructions), they can precisely dial up or dial down specific human biases, like greed, fear, or the tendency to follow the crowd.


The "Radio Dial" Analogy

Think of a Large Language Model (like ChatGPT) as a massive, complex radio.

  • The Baseline (The "Static" Problem): If you just turn the radio on without touching anything, it doesn't sound like a human. It sounds "too perfect"—it’s too rational, too polite, and too logical. It lacks the "human noise" of mistakes and emotions.
  • The Calibration (Turning the Dials): The researchers discovered they could turn specific "dials" on this radio.
    • If they turn the "Loss Aversion" dial up, the AI starts acting like someone who is terrified of losing $10.
    • If they turn the "Herding" dial up, the AI starts acting like a person who buys whatever everyone else is buying.
    • If they turn the "Extrapolation" dial up, the AI starts acting like a gambler who thinks a winning streak will last forever.

By turning these dials, they can create "Digital Humans" that have the exact same psychological settings as real people.


What did they find? (The Good, the Bad, and the "Not Quite Human")

The researchers tested eight different "human glitches" (biases). Here is how the AI performed:

1. The Successes (The "Digital Twins"):
For things like Loss Aversion (fear of losing), Herding (following the crowd), and Extrapolation (predicting the future based only on the recent past), the AI was a superstar. They could dial these up until the AI's behavior matched real-world human data almost perfectly.

2. The "Almost There" (The "Logic vs. Feeling" Gap):
For some things, the AI was close but not quite there. It could understand the logic of a bias, but it lacked the gut feeling.

3. The Failures (The "No Heart" Problem):
The AI failed at things that require true emotion or social ego.

  • The Disposition Effect: Humans often hold onto "loser" stocks because it hurts our pride to admit we were wrong. The AI doesn't have "pride," so it couldn't replicate this well.
  • Representativeness: Humans get swept up in "cool stories" (like a flashy new tech company). The AI is too focused on the math and the facts to get "hyped" by a good story.

Why does this matter? (The "Flight Simulator" for Finance)

Why go through all this trouble? Because once you have a "calibrated" AI, you can build a Flight Simulator for the Stock Market.

Instead of guessing how a market crash might happen, economists can build a digital world populated by thousands of these "calibrated" AI agents. They can say, "What happens to the economy if everyone suddenly becomes 50% more fearful of losing money?"

Because the AI is predictable and easy to "tune," researchers can run these simulations thousands of times to prepare for real-world financial storms.

Summary in a Nutshell

The paper proves that while AI isn't "human" (it has no heart or pride), it is an incredibly powerful mathematical puppet. If you pull the right strings, you can make it dance exactly like a human investor, making it a perfect tool for testing how the world's money might behave in the future.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →