Understanding Annotation Error Propagation and Learning an Adaptive Policy for Expert Intervention in Barrett's Video Segmentation

This paper proposes Learning-to-Re-Prompt (L2RP), a cost-aware framework that analyzes annotation error propagation in endoscopic video segmentation and dynamically learns an adaptive policy to optimize the trade-off between expert intervention effort and segmentation accuracy for Barrett's esophagus dysplasia.

Lokesha Rasanjalee, Jin Lin Tan, Dileepa Pitawela, Rajvinder Singh, Hsiang-Ting Chen

Published 2026-02-26
📖 4 min read☕ Coffee break read

Imagine you are a doctor trying to draw a map of a very tricky, shape-shifting island (a lesion in the esophagus) on a series of 100 moving photographs. This is what happens when doctors annotate Barrett's esophagus videos to train AI.

Doing this manually for every single photo is exhausting and takes forever. So, doctors use a "smart assistant" (an AI called SAM2). The doctor draws the island on just the first photo, and the AI tries to guess where the island is in all the following 100 photos.

The Problem: The "Drifting" Map
Here's the catch: The island moves, the lighting changes, and the camera shakes. If the AI makes a tiny mistake on photo #5, it carries that mistake to photo #6, then #7, and so on. By photo #50, the AI's map might be completely wrong. It's like a game of "Telephone" where the message gets garbled with every turn.

Usually, the doctor has to stop and fix the map every time it gets slightly off, which is still very time-consuming. Or, they might just guess a random time to fix it, which isn't very efficient.

The Solution: A Smart "Check-In" System (L2RP)
This paper introduces a new system called L2RP (Learning-to-Re-Prompt). Think of L2RP as a super-smart co-pilot sitting next to the doctor.

Instead of the doctor guessing when to fix the map, or the AI blindly guessing, L2RP watches the AI's work in real-time. It asks itself: "Is the AI still doing a good job, or is it starting to drift off course?"

  • If the AI is doing well: The co-pilot says, "Keep going, no need to bother the doctor yet."
  • If the AI is starting to drift: The co-pilot says, "Stop! The map is getting messy. We need the doctor to step in and correct it right now."

The system learns exactly when to ask for help so that the doctor does the minimum amount of work to get the best possible map.

The "Prompt" Choices: Drawing the Map
The paper also tested three different ways the doctor can give the AI the first instruction (the "prompt"):

  1. The "Detailed Sketch" (Mask): The doctor carefully traces the exact outline of the island.
    • Pros: Starts very accurate.
    • Cons: Like a delicate sandcastle, it washes away quickly. The AI gets confused easily if the island moves slightly, leading to big errors later.
  2. The "Rough Box" (Box): The doctor draws a square around the island.
    • Pros: A bit more stable.
    • Cons: Less precise to start with.
  3. The "Finger Point" (Point): The doctor just clicks a few dots on the island.
    • Pros: Surprisingly stable! Even though it's the least detailed, the AI holds onto this instruction the longest without getting confused.
    • Cons: Starts slightly less accurate than the sketch.

The Big Discovery
The researchers found that the "Detailed Sketch" (Mask) is the most tempting because it looks perfect at the start, but it requires the doctor to fix the map constantly. The "Finger Point" (Point) is the most reliable "set it and forget it" option.

However, the real magic of L2RP is that it doesn't matter which method you choose. The co-pilot knows exactly when the AI is struggling and asks the doctor to intervene only at the most critical moments.

The Result
By using this smart co-pilot:

  • The final maps are much more accurate.
  • The doctor spends significantly less time correcting mistakes.
  • It works like a budget: You can tell the system, "I have 10 minutes to help," and it will stretch that 10 minutes to cover the whole video as effectively as possible.

In a Nutshell
This paper teaches us how to stop the AI from "drifting" off course and how to use a smart system to decide exactly when a human expert needs to step in. It turns a tedious, hour-long job into a quick, efficient collaboration between human and machine, ensuring the AI learns the right lessons without burning out the doctor.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →