Prompt Group-Aware Training for Robust Text-Guided Nuclei Segmentation

This paper introduces a prompt group-aware training framework that enhances the robustness and generalization of text-guided nuclei segmentation by enforcing consistency among semantically related prompts through quality-guided regularization and logit-level constraints, achieving significant performance gains without altering model architecture or inference.

Yonghuang Wu, Zhenyang Liang, Wenwen Zeng, Xuan Xie, Jinhua Yu

Published 2026-03-09
📖 4 min read☕ Coffee break read

Imagine you are a master chef (the AI) trying to cook a specific dish (segmenting cell nuclei in a medical image) based on a customer's order (the text prompt).

The Problem: The "Picky Customer" Issue

In the past, if a customer said, "I want the red sauce," the chef made it perfectly. But if they said, "I want the red stuff," or "Put the crimson liquid on the plate," the chef might get confused and serve something slightly different each time.

In the world of medical AI, this is a huge problem. Pathologists (the doctors) might describe the same group of cell nuclei in many different ways:

  • "Find all the nuclei."
  • "Locate the cell centers."
  • "Highlight the round purple dots."

Even though these sentences mean the exact same thing, older AI models would get flustered. One description might make the AI draw a perfect circle around the cells, while a slightly different description might make it draw a messy blob. This inconsistency is dangerous in a hospital; you can't trust a tool that changes its mind just because you rephrased your question.

The Solution: The "Group Hug" Training

The authors of this paper, from Fudan University, came up with a clever way to teach the AI to stop being so sensitive to wording. They call it Prompt Group-Aware Training.

Here is how it works, using a simple analogy:

1. The "Study Group" Concept
Instead of treating every text prompt as a separate, isolated instruction, the AI is taught to see them as a study group.

  • Imagine a teacher gives a student five different ways to ask the same math question: "What is 2+2?", "Calculate the sum of two and two," "Add two to two."
  • In the old way, the student might answer "4" to the first one, "4.1" to the second, and "a square" to the third.
  • In this new method, the teacher tells the student: "Hey, these five questions are all in the same group. They all point to the exact same answer (the ground truth). You need to give the same answer to all of them."

2. The "Quality Coach"
The AI also learns to realize that some prompts are "better" than others.

  • A prompt like "Find the nuclei" is a bit vague (Low Quality).
  • A prompt like "Find all the inflammatory nuclei in the top-left corner" is very specific (High Quality).
  • The AI is trained to pay a little more attention to the specific prompts to learn the rules, but it must still make sure it can answer the vague prompts correctly too. It's like a coach telling a player: "Listen closely to the detailed instructions, but don't forget how to play the game when the instructions are short."

3. The "Consistency Check"
During training, the AI is forced to look at its own answers. If it draws a perfect circle for "Find the nuclei" but a messy square for "Locate the cell centers," the system hits a "consistency alarm." It forces the AI to adjust its brain so that both descriptions result in the exact same perfect circle.

The Results: A Trustworthy Assistant

The researchers tested this on six different medical datasets (like testing the chef in six different restaurants).

  • Before: The AI was great with perfect instructions but fell apart with vague ones.
  • After: The AI became a rock-solid professional. It didn't matter if the doctor asked, "Show me the cells" or "Highlight the nuclear structures." The AI drew the same perfect mask every time.

Why This Matters

In the real world, doctors are busy. They might type quickly, use slang, or be vague. They shouldn't have to be professional "prompt engineers" to get a good result from an AI.

This new method makes medical AI robust. It means the tool is reliable enough to be used in real hospitals, where consistency can literally be a matter of life and death. It turns a fickle, picky AI into a dependable partner that understands the intent behind the words, not just the words themselves.