Marking Data-Informativity and Data-Driven Supervisory Control of Discrete-Event Systems

This paper introduces the concept of marking data-informativity and develops corresponding algorithms to design valid nonblocking supervisors for unknown discrete-event systems using available behavioral data, while also addressing cases where data is insufficient through restricted informativity and informatizability frameworks.

Yingying Liu, Kuma Fuchiwaki, Kai Cai

Published Mon, 09 Ma
📖 5 min read🧠 Deep dive

Imagine you are teaching a robot to navigate a maze to reach a treasure chest (the "goal"). In the old way of doing things (Model-Based Control), you would first need a perfect, blueprinted map of the entire maze before you could give the robot any instructions. You'd know exactly where the walls are, where the traps are, and every possible path.

But what if the maze is in a dark room, or it's a brand-new environment you've never seen? You don't have a map. However, you do have a notebook of observations. You've watched the robot move a few times, and you've noted:

  1. What it did: "It went Left, then Right, then Up."
  2. What it achieved: "When it went Left-Right-Up, it found the treasure."
  3. What it definitely didn't do: "It never went Down-Left, because that leads to a bottomless pit (which we know exists from physics, even if we haven't seen it yet)."

This paper is about how to teach the robot to reach the treasure using only that notebook, without ever seeing the full map.

The Core Problem: The "Unknown" Maze

The authors ask: Can we write a set of rules (a supervisor) that guarantees the robot reaches the treasure and never gets stuck, even though we don't know the full layout of the maze?

The answer depends on the quality of your notes. If your notes are too vague, you might give the robot a rule like "Always go Right," but in the real maze, going Right might lead to a dead end that you didn't see in your notes.

The Key Concept: "Marking Data-Informativity"

The authors introduce a fancy term called Marking Data-Informativity. Let's break this down with a metaphor:

Imagine you are a detective trying to solve a crime. You have a list of suspects (all the possible maze layouts that fit your notes).

  • Data-Informativity means your notes are so detailed that every single suspect agrees on the same safe path to the treasure.
  • If your notes are "informative," you can say, "No matter which suspect is the real criminal, if they follow these rules, they will get the treasure and won't get stuck."
  • If your notes are not informative, it means there is a "suspect" (a possible maze layout) where your rules would cause the robot to crash or get stuck.

The "Marking" part is crucial. In robotics, "marking" means reaching a goal state. The paper emphasizes that it's not enough to just keep the robot moving; it must actually finish the job (reach the treasure). Many previous methods ignored this, leading to robots that moved forever in circles without ever finding the goal.

The Three Types of Data

To solve this, the paper says you need three specific types of notes:

  1. Observation (DD): What the robot actually did.
  2. Marked Observation (DmD_m): Which of those actions actually led to the goal.
  3. Negative Knowledge (DD^-): What you know is impossible (e.g., "The robot can't fly," or "It can't go through a wall").

The Magic Trick: The paper shows that having a good list of "impossible things" (DD^-) is just as important as having a long list of "things that happened." If you know what can't happen, you can safely assume that if a path isn't in your notes, it probably leads to a dead end or a wall, so you don't have to worry about it.

What Happens When the Notes Aren't Good Enough?

Sometimes, your notes are just too sparse. You might not have enough "Negative Knowledge" to rule out dangerous paths. In this case, the paper says: "Don't give up! Just aim for a smaller goal."

They introduce a concept called Restricted Marking Data-Informativity.

  • Analogy: You wanted the robot to find the "Golden Treasure," but your notes aren't good enough to guarantee that. However, your notes are good enough to guarantee it can find the "Silver Coin" nearby.
  • The paper provides an algorithm to automatically shrink your goal from "Golden Treasure" to the "Largest Possible Safe Goal" that your notes can support. It's like a GPS that says, "I can't get you to the beach, but I can definitely get you to the park."

The Algorithm: The "Safety Filter"

The authors built a step-by-step recipe (an algorithm) to check your notes:

  1. Build a "Shadow Maze": They create a digital model based only on your notes.
  2. Test the Rules: They simulate every possible move the robot could make.
  3. The "Uncontrollable" Test: They ask, "If the robot hits a wall it can't control (like a sudden wind gust), does it still stay safe?"
    • If the answer is Yes for all possibilities: Great! You have a valid supervisor.
    • If the answer is No: The algorithm automatically trims your goal to the safest, largest possible version that does work.

Why This Matters

This is a big deal for the future of AI and robotics.

  • Old Way: You need a perfect map before you can start. (Slow, expensive, impossible for unknown environments).
  • New Way: You can start controlling a system immediately with just a few observations and a bit of common sense about what's impossible.

In summary: This paper teaches us how to be a smart teacher. Instead of waiting for a perfect textbook (the model) to teach a student (the robot), we can use a few good examples and a list of "don'ts" to guide the student to success, even if we don't know the whole story yet. If the examples aren't enough, we simply adjust the student's goals to something achievable, ensuring they never get stuck or fail.