Artificial Agency Program: Curiosity, compression, and communication in agents

This paper proposes the Artificial Agency Program (AAP), a research agenda and falsifiable framework for developing resource-bounded AI agents driven by curiosity-as-learning-progress, which unifies concepts like predictive compression, intrinsic motivation, and interface quality to enhance human-tool systems through staged experiments and a concrete multimodal testbed.

Richard Csaky

Published 2026-03-02
📖 6 min read🧠 Deep dive

Imagine you are trying to teach a robot to be smart. Most current AI research is like building a super-fast race car engine and then asking, "How fast can this go?" The Artificial Agency Program (AAP) paper by Richard Csaky argues that we are asking the wrong question.

Instead of just building a faster engine, we should be asking: "How does this car drive in real traffic, with a limited gas tank, a driver who gets tired, and a map that is only half-finished?"

Here is the paper explained through simple analogies and metaphors.

1. The Core Idea: The "Smart Assistant" vs. The "Oracle"

Current AI is often trained like an Oracle: it sits in a library with infinite memory, reads every book ever written, and answers questions perfectly. But in the real world, we don't have infinite time, energy, or sensors.

The AAP proposes building AI as an Explorer.

  • The Explorer has a backpack with limited space (memory).
  • The Explorer has a flashlight with a dying battery (energy).
  • The Explorer can only see what's in front of them, not the whole world (partial observation).

The paper argues that true intelligence isn't about knowing everything; it's about knowing how to manage your limited resources to learn, act, and survive in a messy, real world.

2. The Engine of Curiosity: "The Compression Game"

How does this Explorer get curious? It doesn't just look for "new" things (like a child staring at a flashing light). Instead, it plays a game of Compression.

  • The Analogy: Imagine you are trying to describe a movie to a friend.
    • If the movie is random static (noise), you can't compress it; you have to describe every single frame. That's boring and useless.
    • If the movie is a simple cartoon where a ball bounces up and down, you can compress it into one sentence: "Ball bounces." That's easy.
    • True Curiosity happens in the middle. The Explorer looks for patterns that are just hard enough to be interesting but just simple enough that it can figure out a better way to describe them.

The AI gets a "reward" (like a dopamine hit) not just for seeing something new, but for getting better at predicting the future with less effort. It wants to turn a complex, confusing world into a simple, predictable story.

3. The Budget: The "Three Buckets"

The paper suggests that an intelligent agent has a limited budget of "tokens" (units of effort) to spend every second. It has to decide how to split this budget between three buckets:

  1. Observation (Looking): "Should I spend energy to look at this new object?"
  2. Deliberation (Thinking): "Should I spend energy to think about what I saw?"
  3. Action (Doing): "Should I spend energy to move or talk?"

The Metaphor: Think of this like a video game character with a stamina bar.

  • If you run around wildly (high action) without looking, you get lost.
  • If you stare at a wall (high observation) without moving, you go nowhere.
  • If you think too much without acting, you freeze.
  • The Goal: The AI learns to dynamically shift its stamina. If the situation is confusing, it spends more on thinking. If it's safe, it spends more on moving.

4. The "Interface" Problem: The Glass Wall

A major point of the paper is about the Interface—the glass wall between the AI and the real world.

  • Is the glass foggy? (Bad sensors)
  • Is the glass thick? (Slow reaction time)
  • Is the glass cracked? (Noisy data)

The paper introduces a concept called Unification. Imagine the AI is a person wearing a pair of heavy, foggy goggles and thick gloves.

  • Bad Interface: The AI tries to solve a puzzle but can't see the pieces clearly or can't pick them up.
  • Good Interface (Unification): The AI realizes, "Hey, if I spend some of my energy budget to upgrade my goggles or get thinner gloves, I can solve the puzzle much faster."

The AI should be willing to "spend" energy to improve its own senses and tools if it helps it learn better in the long run.

5. The "Inner Monologue": Talking to Yourself

We often think AI needs to talk out loud (like a chatbot) to be smart. This paper suggests that talking is just one tool, and sometimes a wasteful one.

  • The Analogy: Imagine you are solving a math problem.
    • Option A: You write every step down on a piece of paper (Public Language).
    • Option B: You just think the steps in your head (Latent Deliberation).
    • Option C: You whisper a quick reminder to yourself (Private Tokens).

The paper argues that AI should have a "Private Channel." It should be able to "think" in a secret language or hidden symbols that are faster and more efficient than writing out full sentences. It should only "speak" (output text) when it needs to talk to a human or another machine. This is like a pilot having a private radio channel for their co-pilot, rather than shouting instructions to the whole airport.

6. The Ultimate Goal: The "Human-Tool" Team

The paper concludes that we shouldn't judge AI by how smart it is in isolation. We should judge it by how well it fits into a Human-Tool System.

  • The Metaphor: A hammer is not "smart" on its own. But a hammer in the hands of a carpenter is a powerful tool.
  • The AI is the hammer. The human is the carpenter.
  • If the hammer is too heavy, too slippery, or requires too much energy to swing, it's a bad tool, even if it's made of the strongest steel.
  • The goal of AAP is to build AI that feels like a perfect extension of the human hand—easy to use, efficient, and perfectly tuned to our limitations.

Summary: What does this mean for the future?

This paper is a call to stop building "God-like" AI that knows everything but costs a fortune to run and is hard to control. Instead, it wants us to build "Survivor" AI:

  • Agents that know their limits.
  • Agents that get curious about things they can actually understand.
  • Agents that know when to think, when to act, and when to save energy.
  • Agents that treat "talking" as just one option in a toolbox, not the only way to think.

It's about moving from Raw Power to Smart Efficiency.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →