Using Unsupervised Domain Adaptation Semantic Segmentation for Pulmonary Embolism Detection in Computed Tomography Pulmonary Angiogram (CTPA) Images

This paper proposes an unsupervised domain adaptation framework combining a Transformer backbone with a Mean-Teacher architecture and three specialized modules (Prototype Alignment, Global and Local Contrastive Learning, and Attention-based Auxiliary Local Prediction) to significantly enhance pulmonary embolism detection in CTPA images by addressing domain shift and annotation scarcity, achieving substantial performance improvements across cross-center and cross-modality tasks.

Wen-Liang Lin, Yun-Chien Cheng

Published 2026-02-24
📖 5 min read🧠 Deep dive

Imagine you are a master detective trained in New York City to spot a specific type of tiny, dangerous clue (a pulmonary embolism) hidden inside complex maps (CT scans of lungs). You are incredibly good at your job in New York.

But then, you are sent to Tokyo to do the same job. The maps in Tokyo look slightly different: the paper texture is different, the ink colors are slightly off, and the lighting in the room is different. Even though the clue is the same, your New York-trained brain gets confused. You start missing clues or seeing them where they don't exist.

This is the problem doctors face with AI. An AI trained on scans from one hospital often fails when used at another hospital because of these subtle differences (called "Domain Shift"). Furthermore, teaching the AI to recognize the new style usually requires a human expert to manually draw every single clue on thousands of new maps, which is too expensive and slow.

This paper presents a clever solution: An AI that teaches itself how to adapt without needing a human teacher for the new city.

Here is how they did it, explained with simple analogies:

1. The "Mean Teacher" Strategy (The Wise Mentor)

Instead of hiring a new human teacher for the Tokyo hospital, the AI uses a "Mean Teacher" system.

  • The Student: The AI trying to learn the new city.
  • The Teacher: A slightly older, more stable version of the Student.
  • How it works: The Teacher guesses where the clues are in the new maps. These guesses are called "Pseudo-labels." The Student tries to match the Teacher's guesses. Over time, the Teacher becomes smarter, and the Student learns from them. It's like a student learning by mimicking a mentor who is slowly getting better at the job.

2. The Three Secret Weapons

To make sure the Student doesn't get confused by the "noise" in the new maps, the researchers added three special tools:

A. Prototype Alignment (The "Group Hug")

  • The Problem: In the new city, the "clue" might look a bit gray instead of red, and the "background" might look a bit blue instead of white. The AI gets confused about what belongs to which group.
  • The Solution: Imagine the AI creates a "center point" (a prototype) for what a "clue" looks like and a "center point" for what "background" looks like. This tool drags the New York "clue center" and the Tokyo "clue center" closer together until they are hugging. It forces the AI to realize, "Oh, even though the colors are different, this is still the same type of object."

B. Global and Local Contrastive Learning (The "Big Picture vs. The Details")

  • The Problem: The AI needs to understand both the whole map (the layout of the lungs) and the tiny details (the shape of the tiny clot).
  • The Solution:
    • Global: It looks at the whole map to understand the "skeleton" or layout. It learns that "a heart is always in the middle," regardless of the image style.
    • Local: It zooms in on tiny patches to learn the texture of the clot.
    • The Trick: It uses a "Momentum Queue" (like a memory bank) to remember thousands of examples it has seen before. This helps it learn the difference between "clue" and "not clue" without needing a massive computer to hold everything at once.

C. Attention-Based Auxiliary Local Prediction (The "Flashlight" vs. "Random Shuffling")

  • The Problem: This is the most important part for tiny clues. Imagine you are looking for a needle in a haystack. If you randomly grab handfuls of hay to look at, 99% of the time, you'll just grab empty hay. You'll never find the needle. This is what happens when AI randomly crops images; it usually misses the tiny embolisms.
  • The Solution: The researchers gave the AI a Flashlight.
    • Because the AI uses a "Transformer" (a type of AI that pays attention to relationships), it naturally knows where to look. It creates a "heat map" showing where the important stuff is.
    • Instead of randomly grabbing a piece of the image, the AI uses its Flashlight to shine only on the areas where the "clue" is likely hiding. It then studies those specific spots intensely. This ensures the AI never wastes time looking at empty background space.

3. The Results: A Detective Who Adapts

The researchers tested this system in two ways:

  1. Cross-Center: Moving from Hospital A (FUMPE) to Hospital B (CAD-PE).
    • Before: The AI was terrible (IoU score of 0.11). It was like a detective who couldn't read the new maps at all.
    • After: The AI became excellent (IoU score of 0.41). It successfully adapted to the new hospital's style.
  2. Cross-Modality: Moving from CT scans to MRI scans (completely different types of images).
    • Result: The AI achieved a 69.9% success rate without ever seeing a single labeled MRI scan from a human expert.

Why This Matters

Most advanced AI systems today are like expensive supercomputers that need massive power and huge teams of experts to train. This new method is like a smart, self-taught detective.

  • It works on standard hospital computers (not just super-servers).
  • It doesn't need expensive human experts to label every new image.
  • It specifically solves the problem of finding tiny things that are easily missed.

In short, this paper gives us a way to take a smart medical AI trained in one place and instantly make it useful in a completely different place, saving time, money, and potentially lives by catching dangerous blood clots that might otherwise be missed.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →