Ordinal Diffusion Models for Color Fundus Images

This paper proposes an ordinal latent diffusion model that leverages the continuous, ordered nature of diabetic retinopathy severity to generate more realistic and clinically consistent color fundus images compared to standard categorical conditional diffusion models.

Gustav Schmidt, Philipp Berens, Sarah Müller

Published 2026-03-02
📖 4 min read☕ Coffee break read

Imagine you are an artist trying to teach a robot how to paint pictures of the human eye, specifically to show how a disease called Diabetic Retinopathy gets worse over time.

In the real world, this disease doesn't just jump from "healthy" to "sick" in giant leaps. It's a slow, continuous slide. A healthy eye slowly develops tiny spots, then more spots, then bleeding, and finally, new, messy blood vessels.

However, doctors usually label these eyes with simple, separate categories, like steps on a ladder:

  • Step 0: Healthy
  • Step 1: Mild
  • Step 2: Moderate
  • Step 3: Severe
  • Step 4: Proliferative (Very bad)

The Problem with Old AI

Previous AI models treated these steps like completely different languages. If you asked the AI to draw a "Step 1" eye, it learned that as a totally separate concept from a "Step 2" eye. It didn't understand that Step 2 is just Step 1 with a little bit more damage. It was like teaching a child that "one apple" and "two apples" are unrelated concepts, rather than just adding one more apple to the pile.

Because of this, when these old AIs tried to draw the disease getting worse, the changes were often jumpy, unrealistic, or just plain wrong.

The New Solution: The "Ordinal" Diffusion Model

The researchers in this paper built a smarter AI called an Ordinal Diffusion Model. Here is how they did it, using some simple analogies:

1. The Volume Knob instead of Buttons

Instead of giving the AI a set of buttons labeled "0, 1, 2, 3, 4," they gave it a volume knob (a slider).

  • They told the AI: "Turn the knob to 0 for a healthy eye. Turn it to 1 for mild, 2 for moderate, and so on."
  • Because the knob moves smoothly, the AI learned that moving from 1 to 2 is just a small adjustment, not a total reboot. This allows the AI to generate images that show a smooth transition of the disease, just like it happens in real life.

2. The "Skeleton" and the "Skin"

The researchers realized that to make a realistic eye, you need two things:

  • The Skeleton: The unique shape of the blood vessels and the optic disc (the "eye" of the eye). This stays the same person, even if they get sick.
  • The Skin: The disease symptoms (the spots, bleeding, etc.).

They taught the AI to separate these. Imagine a mannequin (the skeleton) that stays exactly the same, while a makeup artist (the disease generator) slowly adds more and more "bruises" and "rashes" as you turn the volume knob up. This ensures that the AI doesn't accidentally change the person's eye shape while trying to make them look sicker.

What Did They Find?

They tested this new AI on thousands of real eye photos. Here is what happened:

  • Better Art: The pictures the AI made looked much more realistic than before. If you looked at them, you could clearly see the disease getting worse step-by-step.
  • Smarter Transitions: When they asked the AI to draw an eye that was "halfway" between Mild and Moderate, it didn't just pick one or the other. It drew an eye with a mix of symptoms, perfectly capturing the messy reality of disease progression.
  • The "Time Travel" Effect: They took a photo of a healthy eye and asked the AI to "turn up the disease knob" to show what that specific person's eye would look like if they got worse. The AI kept the person's unique eye structure but added the correct amount of damage. It was like a "what if" simulator for disease.

Why Does This Matter?

In medicine, we often don't have enough photos of people with severe diseases or from certain ethnic groups to train our AI doctors. This new model acts like a photocopier for reality. It can create thousands of new, realistic, and medically accurate eye images to fill in the gaps.

By understanding that disease is a spectrum (a continuous slide) rather than a list of separate boxes, this AI helps doctors train better diagnostic tools, potentially leading to earlier detection and better care for patients with diabetic retinopathy.

In short: They taught the AI to stop thinking in "steps" and start thinking in "slides," resulting in a much smarter, more realistic medical artist.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →