The science and practice of proportionality in AI risk evaluations

This paper explores how the EU AI Act's requirement for general-purpose AI providers to evaluate systemic risks can be balanced with innovation through the principle of proportionality, aiming to develop scientific methods that ensure meaningful risk assessments without imposing excessive burdens.

Carlos Mougan, Lauritz Morlock, Jair Aguirre, James R. M. Black, Jan Brauner, Simeon Campos, Sunishchal Dev, David Fernández Llorca, Alberto Franzin, Mario Fritz, Emilia Gómez, Friederike Grosse-Holz, Eloise Hamilton, Max Hasin, Jose Hernandez-Orallo, Dan Lahav, Luca Massarelli, Vasilios Mavroudis, Malcolm Murray, Patricia Paskov, Jaime Raldua, Wout Schellaert

Published Thu, 12 Ma
📖 5 min read🧠 Deep dive

Imagine you are the head of a massive construction company building a new type of super-brick that can build houses, bridges, and even entire cities. But there's a catch: if you get the recipe wrong, these bricks could also be used to build bombs or collapse buildings.

The European Union (EU) has stepped in and said, "You must test these bricks before you sell them to make sure they aren't dangerous." This is the AI Act.

However, the regulators face a tricky problem: How do you make sure the tests are good enough to keep everyone safe, without forcing the builders to spend so much time and money testing that they never finish building anything?

This paper is about finding the perfect balance. It uses a legal concept called "Proportionality." Think of proportionality like a Goldilocks principle: the testing shouldn't be too easy (useless), too hard (impossible), but just right.

Here is how the authors break this down using simple analogies:

1. The Three Rules of "Just Right"

The paper says a test is only "proportionate" if it passes three checks:

  • Suitability (Does it work?): The test must actually tell you something useful.
    • Analogy: If you want to know if a car can stop in the rain, testing it on a sunny day is useless. The test must match the real-world danger.
  • Necessity (Is it the least painful way?): If you can get the same safety information with a cheaper, easier test, you shouldn't force the expensive one.
    • Analogy: If you want to check if a door is locked, you don't need to hire a SWAT team to break it down. A simple "jiggle the handle" test is enough. Using a SWAT team would be "using a sledgehammer to crack a nut."
  • Balancing (Is the cost worth the safety?): The trouble and cost of the test shouldn't be way bigger than the risk you are trying to stop.
    • Analogy: It doesn't make sense to spend $1 million to install a security system for a shed that only contains a single garden hose.

2. The "Pareto Frontier" (The Efficiency Curve)

The paper includes a graph that looks like a curve. Imagine this as a menu of tests.

  • On one side, you have Low Cost / Low Safety Info (like a quick quiz).
  • On the other side, you have High Cost / High Safety Info (like a full-blown simulation).

The "Pareto Frontier" is the best possible deals on that menu. It represents the tests where you get the maximum amount of safety information for the minimum amount of money. If a test is below this line, it's a bad deal (too expensive for the info it gives). If it's above the line, it's impossible (you can't get that much info that cheaply).

The goal of regulators is to pick a point on this line that matches the danger level of the specific AI model.

3. A Real-World Example: The "Hacker" Test

To explain this, the authors look at a specific risk: What if an AI helps a hacker find holes in computer code? They compare three ways to test this:

  • Method A (The "Hinted" Test): The test gives the AI a hint like, "Look at line 50, there's a bug there."
    • Pros: Very cheap and fast.
    • Cons: Not very realistic. If the AI fails, it might just mean it didn't read the hint, not that it's safe.
    • Verdict: Good for a quick "screening" to see if the AI is totally useless.
  • Method B (The "Realistic" Test): The AI has to find bugs in a realistic environment without hints, like a real hacker would.
    • Pros: Much more accurate.
    • Cons: Takes more computer power and time.
    • Verdict: Good if the first test was scary or inconclusive.
  • Method C (The "Super-Simulation" Test): The AI is put in a massive, complex digital world with thousands of scenarios to find bugs.
    • Pros: Extremely rigorous and detailed.
    • Cons: Very expensive and slow.
    • Verdict: Only necessary if the AI is huge, widely used, and the first two tests suggest it might be dangerous.

The Strategy: The paper suggests an iterative approach. Start with the cheap, easy test (Method A). If the AI passes easily, great! Stop there. If it fails or looks suspicious, then move to the harder test (Method B). Only if it still looks dangerous do you spend the fortune on the super-test (Method C). This saves everyone time and money.

4. Why This Matters

Currently, regulators often feel like they are guessing. They might demand a "Super-Simulation" for a small, harmless AI, which kills innovation because the company goes bankrupt testing it. Or, they might accept a "Hinted" test for a dangerous AI, which leaves the public unsafe.

This paper argues that we need science to fix this. We need better tools to measure exactly how much "safety info" a test gives versus how much it costs.

The Bottom Line:
The goal isn't to stop AI progress. It's to make sure the safety checks are smart. We want to use the right tool for the job, ensuring that the burden on companies is fair, but the protection for society is solid. It's about moving from "guessing" to "measuring" so that we can build a safer future without breaking the bank.