Taming Silent Failures: A Framework for Verifiable AI Reliability

This paper introduces the Formal Assurance and Monitoring Environment (FAME), a novel framework that combines offline formal synthesis and online runtime monitoring to detect silent AI failures and ensure verifiable safety in critical systems like autonomous vehicles, thereby offering a certifiable pathway aligned with ISO standards.

Guan-Yan Yang, Farn Wang

Published 2026-03-03
📖 4 min read☕ Coffee break read

The Problem: The "Confidently Wrong" AI

Imagine you hire a brilliant but slightly unreliable navigator for your car. This navigator (the AI) is amazing at spotting pedestrians and traffic signs 99% of the time. However, when it gets confused—say, because of heavy rain or a strange shadow—it doesn't say, "I'm not sure!" or "I can't see!"

Instead, it confidently points at a mailbox and says, "That's a pedestrian!" and keeps driving. It doesn't crash, it doesn't throw an error message, and it doesn't stop. It just silently fails.

In the world of safety-critical systems (like self-driving cars or medical robots), this is the most dangerous kind of failure. The system looks like it's working, but it's actually making a deadly mistake.

The Solution: FAME (The "Safety Net" and "Contract" System)

The authors, Guan-Yan Yang and Farn Wang, propose a new framework called FAME (Formal Assurance and Monitoring Environment).

Think of FAME not as trying to fix the AI's brain (which is too complex and opaque to fully understand), but as putting a strict, unbreakable safety contract around the AI's behavior.

Here is how FAME works, broken down into three simple steps:

1. The Contract (Design-Time)

Before the AI ever hits the road, safety engineers write a strict "rulebook" using a precise mathematical language (called Signal Temporal Logic).

  • Analogy: Imagine writing a contract for your navigator that says: "If a person is within 30 meters, you must see them clearly with 90% confidence. If you lose sight of them for even one second, you must immediately stop."
  • This isn't a vague suggestion like "be careful." It is a hard, mathematical rule that leaves no room for interpretation.

2. The Watchdog (Run-Time)

Once the car is driving, a tiny, super-fast "watchdog" program runs alongside the AI. This watchdog doesn't try to understand how the AI thinks; it only watches what the AI does.

  • Analogy: Think of the AI as a chef cooking a complex meal, and the watchdog as a strict health inspector standing right next to the stove. The inspector doesn't need to know how to cook; they just check if the chef follows the rules (e.g., "Is the chicken cooked? Is the temperature safe?").
  • If the AI starts hallucinating (e.g., seeing a pedestrian where there is none, or missing a real one), the watchdog instantly spots the violation of the contract.

3. The Emergency Brake (Mitigation)

The moment the watchdog sees a rule broken, it doesn't try to "fix" the AI's thinking. Instead, it triggers a pre-programmed safety action.

  • Analogy: If the health inspector sees the chef trying to serve raw chicken, they don't argue with the chef. They immediately hit the "Stop" button, shut down the kitchen, and switch to a backup plan (like serving a pre-made safe meal or pulling the car over).
  • This ensures that even if the AI is having a "bad day," the system remains safe.

Why This is a Big Deal

The paper tested this system on a self-driving car simulation.

  • The Result: In tricky situations (heavy rain, glare, occluded pedestrians), the AI made mistakes 31% of the time. These were "silent failures"—the car didn't know it was failing.
  • FAME's Performance: The FAME watchdog caught 93.5% of these silent failures. It knew the AI was confused and triggered the safety brakes before a crash could happen.
  • No False Alarms: Crucially, in normal driving, the watchdog never panicked. It didn't stop the car when everything was fine (0% false alarms).

The "Feedback Loop": Learning from Mistakes

FAME isn't just a one-time fix; it's a learning machine.

  • Analogy: Every time the watchdog catches the AI making a mistake, it saves a "video replay" of exactly what happened.
  • Later, engineers use these replays to retrain the AI, teaching it not to make that specific mistake again. They also refine the "contract" to be even smarter. Over time, the system gets safer and smarter.

The Bottom Line

This paper argues that we can't wait for AI to be perfect before we trust it with our lives. Instead, we should build verifiable safety nets around it.

Just as a trapeze artist uses a safety net not because they expect to fall, but because the consequence of falling is too high, FAME provides a provable safety net for AI. It allows us to use powerful, intelligent AI systems while ensuring that if they ever get confused, they fail safely rather than silently.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →