LOCUS: A Distribution-Free Loss-Quantile Score for Risk-Aware Predictions

The paper introduces Locus, a distribution-free wrapper that generates interpretable, comparable risk scores for any prediction function by modeling realized loss and applying split-calibration to effectively rank inputs and control large-loss events without assuming a specific data distribution.

Matheus Barreto, Mário de Castro, Thiago R. Ramos, Denis Valle, Rafael Izbicki

Published 2026-03-03
📖 5 min read🧠 Deep dive

Imagine you are a doctor using a new AI to diagnose patients. The AI is generally very good; on average, it gets the diagnosis right 95% of the time. That sounds great, right?

But here's the problem: Average doesn't save lives.

If that AI makes a mistake, it might be a tiny, harmless error for one patient, but a catastrophic, life-threatening error for another. In the real world, we don't just care about the average performance; we care about when the AI is likely to fail so we can stop it before it hurts someone.

This is the problem the paper "Locus" tries to solve.

The Problem: The "Blind Spot" of Standard AI

Most AI models are like a weather forecast that says, "It will rain 50% of the time this week." That's a useful average, but it doesn't tell you if today is the day you need an umbrella or if you're safe.

In machine learning, we usually try to measure "uncertainty" (how unsure the AI is).

  • The Old Way: "I'm not sure about this prediction because the data looks weird."
  • The Flaw: Sometimes the AI is very sure of a wrong answer. Imagine a GPS that confidently tells you to drive off a cliff because it's never seen that road before. The AI thinks it's right, but it's actually dangerous.

The Solution: Locus (The "Damage Meter")

The authors created a tool called Locus. Instead of asking, "How unsure is the AI?", Locus asks a much more practical question: "If we use this prediction, how much damage could it cause?"

Think of Locus as a Damage Meter attached to every single prediction.

The Creative Analogy: The Car Insurance App

Imagine you are an insurance company. You have a driver (the AI model) who drives a car.

  1. The Driver: The AI makes a prediction (e.g., "This house is worth $500,000").
  2. The Risk: Sometimes the driver makes a huge mistake (e.g., "This house is worth $500,000" when it's actually $100,000). That's a $400,000 loss.
  3. The Old Approach: You look at the driver's history. "He's a good driver on average!" But you don't know if he's about to crash right now.
  4. The Locus Approach: Locus puts a Speedometer of Potential Loss on the dashboard.
    • For a safe prediction, the meter reads "Low Risk."
    • For a dangerous prediction, the meter reads "High Risk: Potential Loss of $400,000."

How Does It Work? (The Simple Version)

Locus doesn't try to guess the future or build a complex new brain. It acts like a smart referee that uses a simple trick:

  1. The "Test Drive" (Calibration): Before the AI goes to work, Locus takes a bunch of past data and pretends to be the AI. It makes predictions and sees how much "damage" (error) it actually caused.
  2. The "Scorecard": It learns to recognize patterns. "Oh, when the AI predicts a house price in this specific neighborhood, it tends to be off by $50k. When it predicts a house in that other neighborhood, it's off by $500k."
  3. The "Red Flag": When a new prediction comes in, Locus looks at the scorecard and says, "Based on what we've seen before, there is a 90% chance this prediction will be within $10,000 of the truth. But there's a 10% chance it could be off by $200,000."
  4. The Decision: If your company rule is "We can't afford errors over $50,000," Locus will instantly raise a red flag: "Do not trust this prediction! The potential damage is too high."

Why Is This Special?

The paper highlights three cool things about Locus:

  • It Speaks Your Language: Instead of giving you a confusing math number like "Entropy = 0.45," it gives you a number in dollars (or whatever your unit is). "This prediction might cost you $200,000." That's easy to understand.
  • It's "Distribution-Free": This is a fancy way of saying it works even if the AI is weird. You don't need to know how the AI was built or what kind of data it uses. It works like a universal adapter. You can plug it into any AI, and it will give you a reliable damage estimate.
  • It Catches the "Confident Stupidity": As shown in the paper's examples, sometimes an AI is very confident but completely wrong (like the linear model in the low-variance region). Standard uncertainty tools miss this, but Locus catches it because it looks at the actual loss, not just how "scattered" the data looks.

The Bottom Line

In a world where AI is making decisions about loans, medical diagnoses, and self-driving cars, being "mostly right" isn't enough. We need to know when to hit the brakes.

Locus is a safety wrapper that tells you, for every single decision an AI makes: "This one is safe to trust," or "Stop! This one might cause a disaster." It turns the abstract concept of "uncertainty" into a concrete, actionable "risk score" that anyone can understand.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →