Cyber Threat Intelligence for Artificial Intelligence Systems

This paper investigates the evolution of cyber threat intelligence to address AI-specific security threats by analyzing current gaps, proposing a structured knowledge base with concrete indicators of compromise across the AI supply chain, and outlining techniques for measuring artifact similarity to support a practical, AI-tailored defense framework.

Natalia Krawczyk, Mateusz Szczepkowski, Adrian Brodzik, Krzysztof Bocianiak

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine the world of cybersecurity as a massive, high-tech fortress. For decades, the guards (security experts) have been trained to spot intruders trying to break in through windows, pick locks, or sneak in through the front door. They use a giant "Wanted Poster" database (Cyber Threat Intelligence, or CTI) that lists the faces, fingerprints, and known tricks of thieves.

But now, the fortress has a new, very smart, but slightly naive employee: Artificial Intelligence (AI).

This paper, written by a team from Orange Innovation Poland, argues that our old "Wanted Posters" and security guards don't know how to catch thieves who are specifically trying to trick, poison, or steal this new AI employee. Here is the breakdown of their findings, translated into everyday language.

1. The Problem: The Old Rules Don't Fit the New Game

Think of traditional cybersecurity like a bouncer at a club. He checks IDs, looks for known troublemakers, and checks if you're carrying a weapon. It works great for humans.

But AI is different. It's not a person; it's a giant, invisible brain that learns by reading millions of books (data).

  • The Old Threat: A thief breaks a window to steal a TV.
  • The New AI Threat: A thief doesn't break the window. Instead, they sneak into the library and swap a few pages in the books the AI is reading. Now, the AI thinks "Fire" means "Water," or it believes that a picture of a stop sign is actually a speed limit sign.

The paper says our current security tools are looking for broken windows, but they are blind to the book page swaps. We need a new kind of "Wanted Poster" specifically for AI crimes.

2. The New "Wanted Posters" (Indicators of Compromise)

In the old days, a "clue" (Indicator of Compromise) was a specific file name or a bad IP address (like a bad phone number).

For AI, the clues are weirder. The paper suggests we need to look for:

  • Poisoned Recipes: If a chef (the AI) is making soup, did someone sneak a little bit of poison into the ingredients before the cooking started? (This is called Data Poisoning).
  • Fake Glasses: Did someone put special stickers on a stop sign that only the AI can see, making it think the sign is a "Go" signal? (This is an Adversarial Example).
  • The "Jailbreak" Note: Did someone whisper a secret code to the AI that tricks it into ignoring its safety rules? (This is Prompt Injection).

The authors are trying to build a database that lists these specific "poisoned recipes" and "fake glasses" so security systems can spot them before the AI goes to work.

3. Where Do We Get the Clues? (The Sources)

The paper reviews where we can find information about these AI crimes. They found three main types of sources, like different sections of a library:

  • The "Bug List" (Vulnerability Databases): Like a list of known weak spots in a car (e.g., "The brakes fail if it rains"). Examples: AVID and OWASP.
    • Verdict: Good, but often incomplete. They list the weak spots but don't always show how the bad guys actually used them.
  • The "Crime Scene Reports" (Incident Databases): Like police reports detailing exactly what happened when a bank was robbed. Examples: AI Incident Database (AIID).
    • Verdict: Very useful! These tell real stories, like "An autonomous car hit a pedestrian because it confused a plastic bag for a rock." This helps us understand the consequences.
  • The "Thief's Playbook" (Adversary Tactics): Like a manual written by the thieves explaining their tricks. The big one here is MITRE ATLAS.
    • Verdict: This is the gold standard. It maps out exactly how hackers attack AI, step-by-step, similar to how the famous MITRE ATT&CK framework maps out attacks on regular computers.

4. The Challenge: Finding a Needle in a Haystack

Here is the tricky part. If a hacker steals a model (the AI brain) and changes just one tiny line of code, it looks almost identical to the original. How do you catch them?

The paper suggests using Digital Fingerprints (Hashing).

  • Analogy: Imagine you have a giant library of books. A thief steals a book, changes the font size, and adds a doodle on page 5. A normal search for the title won't find it.
  • The Solution: The authors suggest using "Deep Hashing." This is like taking a photo of the entire vibe of the book. Even if the font changes, the "vibe" (the mathematical fingerprint) stays similar enough that the security system can say, "Hey, this looks 95% like that stolen book we know about!"

This allows security tools to catch modified AI models even if the hackers tried to disguise them.

5. Why This Matters (The Conclusion)

The authors conclude that we are currently flying blind. We have great tools for protecting regular computers, but we are struggling to protect AI.

  • The Gap: We don't have enough "Wanted Posters" for AI-specific crimes.
  • The Fix: We need to build a specialized intelligence system that understands AI's unique weaknesses (like poisoned data or prompt injections).
  • The Goal: To create a system where, if a hacker tries to trick an AI, the security guard doesn't just say "Access Denied," but says, "I see you're trying to use a 'Poisoned Recipe' tactic. I've seen this before, and here is how we stop it."

In short: The paper is a call to action. We need to stop treating AI security like regular computer security. We need new maps, new fingerprints, and new rules of the road to keep our AI employees safe from the clever tricks of modern cybercriminals.