Sparse autoencoders reveal organized biological knowledge but minimal regulatory logic in single-cell foundation models: a comparative atlas of Geneformer and scGPT

This study applies sparse autoencoders to Geneformer and scGPT to reveal that while these single-cell foundation models effectively encode organized biological knowledge and hierarchical abstraction, they largely fail to capture causal regulatory logic, as evidenced by their minimal response to specific transcription factor perturbations.

Ihor Kendiukhov

Published 2026-03-04
📖 4 min read☕ Coffee break read
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

The Big Picture: Peeking Inside the "Black Box"

Imagine two super-smart AI robots, Geneformer and scGPT, that have read millions of biology textbooks (single-cell data). They are great at guessing what a cell is doing or what happens if you change a gene. But scientists have a nagging question: Do these robots actually understand how biology works (the cause-and-effect rules), or are they just really good at spotting patterns (like noticing that people who buy umbrellas also buy raincoats)?

To answer this, the author built a special tool called a Sparse Autoencoder (SAE). Think of this tool as a high-tech X-ray machine that can look inside the robot's brain while it's thinking.

The Discovery: A Library of Hidden Books

When the author used the X-ray machine, they found something amazing:

  1. Massive Overcrowding (Superposition): The robots' brains have a limited number of "slots" (dimensions) to store information. But they are storing thousands of more concepts than there are slots.

    • The Analogy: Imagine a library with only 1,000 shelves, but the librarian has 80,000 books. To make it work, they stack the books in a way that looks like a solid wall to the naked eye. You can't see the individual books unless you have a special decoder. The AI is doing this with biological concepts.
    • The Result: The author found 82,000+ hidden "features" (concepts) in Geneformer and 24,000+ in scGPT.
  2. Organized Knowledge: These hidden books aren't random. They are organized perfectly.

    • The Analogy: If you open the library, you don't just see a mess. You see a section for "Cell Division," a section for "Immune System," and a section for "Mitochondria."
    • The Result: The AI has learned the "vocabulary" of biology. It knows which genes belong to which pathways, just like a human biologist would.
  3. The "U-Shape" Journey: As the information travels through the robot's layers (from input to output), the knowledge changes.

    • Early Layers: Focus on raw parts (like "ribosomes" or "DNA replication").
    • Middle Layers: Get abstract and messy (hard to label).
    • Late Layers: Re-organize into big-picture goals (like "cell differentiation" or "stress response").

The Twist: The Robot Knows the "What," But Not the "Why"

Here is the most critical finding. The author tested if the robot understood causal logic (the "If I pull this lever, that light turns on" relationship).

  • The Test: They simulated a real-world experiment: turning off specific genes (using CRISPR data) and seeing if the robot's internal "books" changed in a way that matched the specific gene's job.
  • The Result: The robot was bad at this.
    • It noticed that something changed (the cell state shifted).
    • But it didn't know specifically which gene caused which effect.
    • The Analogy: Imagine a detective who sees a crime scene and says, "Oh, a robbery happened! The window is broken, and the safe is open." But if you ask, "Did the butler do it, or the gardener?" the detective just shrugs and says, "I don't know, I just know a robbery happened."

The Stat: Out of 48 specific transcription factors (the "bosses" of genes), the robot only correctly identified the specific cause-and-effect relationship for 3 of them (6.2%).

Why Does This Matter?

  1. It's Not the Tool's Fault: The author tried training the X-ray machine on different types of cells (not just one type) to see if the robot was just confused by the data. It didn't help much. The problem is the robot itself.
  2. The Bottleneck: The current AI models are trained to predict the next word (or gene) based on patterns. They are excellent at memorizing correlations (things that happen together) but terrible at learning the actual rules of the universe (cause and effect).
  3. The Future: To make these robots truly understand biology, we need to train them differently. Instead of just asking them to predict patterns, we need to teach them with experiments where they have to figure out why something happened.

The Gift to the World

The author didn't just write a paper; they built two interactive websites (Feature Atlases).

  • Think of these as Google Maps for the AI's brain.
  • Anyone can go online, search for a gene, and see exactly which "hidden book" in the AI's brain is talking about it, which layer it lives in, and how it connects to other concepts.

Summary in One Sentence

These AI models have memorized the entire encyclopedia of biology and organized it beautifully, but they are still just pattern-matching machines that haven't quite learned the actual rules of how genes control life.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →