Tell2Adapt: A Unified Framework for Source Free Unsupervised Domain Adaptation via Vision Foundation Model

Tell2Adapt is a unified Source-Free Unsupervised Domain Adaptation framework that leverages Vision Foundation Models with Context-Aware Prompts Regularization and Visual Plausibility Refinement to achieve state-of-the-art performance across 10 domain shifts and 22 anatomical targets in medical image segmentation.

Yulong Shi, Shijie Li, Ziyi Li, Lin Qi

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine you are a master chef who has spent years perfecting a recipe for Spicy Tofu using ingredients from a specific farm in Japan. You know exactly how that tofu tastes, how it looks, and how to cook it.

Now, you are asked to cook the same dish for a new restaurant in Brazil, but there's a catch: You cannot take your original Japanese ingredients with you. You only have the new Brazilian tofu, which looks slightly different, has a different texture, and comes from a different farm. If you try to cook using your old Japanese instincts alone, the dish might taste terrible.

This is the problem doctors face with Medical AI. They train AI models on data from one hospital (the "Japanese farm"), but when they try to use that AI at a different hospital with different machines (the "Brazilian farm"), the AI gets confused and makes mistakes. Usually, to fix this, they would need to bring the old data along, but patient privacy laws say, "No, you can't take that data." This is called Source-Free Unsupervised Domain Adaptation (SFUDA). It's like trying to cook the new dish without the old ingredients and without the old recipe book.

Enter Tell2Adapt, a new framework that solves this problem using a "Super-Chef" and a "Quality Inspector."

The Three Magic Ingredients of Tell2Adapt

1. The "Super-Chef" (The Vision Foundation Model)

Think of a Vision Foundation Model (VFM) like a super-intelligent, all-knowing chef who has tasted every dish in the world. This chef doesn't need to see the specific ingredients you have; they just need a clear description of what you want to cook.

In the past, other methods tried to ask the Super-Chef for help by showing them the new Brazilian tofu and asking, "What does this look like?" But because the tofu looked weird, the Super-Chef got confused and gave bad advice.

Tell2Adapt changes the game. Instead of showing the tofu, it asks the Super-Chef using words. It says, "Hey, please find the Liver in this Abdominal CT scan."

2. The "Translator" (Context-Aware Prompts Regularization - CAPR)

Here's the problem: Doctors are busy. Sometimes they type messy instructions like: "liver in abd ct" or "spleen... wait, no, liver... CT scan." If you give these messy notes to the Super-Chef, they might get confused.

CAPR is like a super-efficient secretary. Before the doctor's messy note reaches the Super-Chef, the secretary reads it, fixes the typos, figures out the context (e.g., "Oh, they are looking at an Abdominal CT"), and rewrites it into a perfect, professional sentence: "Liver in Abdominal CT."

This ensures the Super-Chef always gets a crystal-clear instruction, no matter how messy the original request was. The Super-Chef then draws a perfect map (a "pseudo-label") of where the liver should be.

3. The "Quality Inspector" (Visual Plausibility Refinement - VPR)

Once the Super-Chef draws the map, a small, lightweight AI model (the "Student Chef") learns from it to do the job itself. But sometimes, even with a good map, the Student Chef might get a little carried away and draw a liver that looks like a potato or floats in the air.

VPR is the Quality Inspector. It knows the "laws of anatomy." It checks the Student Chef's work and asks:

  • "Does this shape look like a real liver?"
  • "Is the color and texture consistent with what a liver usually looks like in a CT scan?"

If the Student Chef drew a liver that looks like a potato, the Inspector says, "Nope, that's fake. Throw it out." This removes mistakes and ensures the final result is medically safe and realistic.

Why This is a Big Deal

Before Tell2Adapt, trying to adapt medical AI to new hospitals was like trying to build a house without blueprints, using only guesswork. It worked for small, simple houses (low-gap shifts) but failed for skyscrapers (complex, different medical machines).

Tell2Adapt is different because:

  1. It's Universal: It works on brains, hearts, livers, and polyps. It's not just for one specific organ.
  2. It's Privacy-Safe: It doesn't need the old data. It just needs the new images and the "Super-Chef's" general knowledge.
  3. It's Reliable: By using the "Translator" to clean up instructions and the "Inspector" to check the work, it produces results that are almost as good as if the AI had been trained on the new data from scratch.

The Bottom Line

Tell2Adapt is a smart system that lets medical AI learn new skills on new machines without needing old data. It uses a "Super-Chef" to guide the learning, a "Translator" to make sure the instructions are clear, and a "Quality Inspector" to ensure the final result is safe and accurate. This means doctors can finally use AI tools in more hospitals, saving time and lives, without worrying about privacy or broken models.