Gradient based Severity Labeling for Biomarker Classification in OCT

This paper proposes a novel contrastive learning strategy for medical images that replaces arbitrary augmentations with disease severity labels derived from anomaly detection gradients to improve biomarker classification accuracy in OCT scans.

Kiran Kokilepersaud, Mohit Prabhushankar, Ghassan AlRegib, Stephanie Trejo Corona, Charles Wykoff

Published 2026-02-24
📖 4 min read☕ Coffee break read

Imagine you are trying to teach a computer to spot tiny, specific problems in a patient's eye scans (called OCT scans), like finding a single drop of water in a massive ocean or a tiny crack in a windshield. These problems are called biomarkers.

The big challenge? Doctors are busy, and labeling thousands of these scans with "yes, there's a problem here" or "no, it's clean" takes a lot of time and money. So, we have a huge pile of unlabeled scans (we don't know what's in them) and a tiny pile of labeled scans (we know exactly what's in them).

This paper proposes a clever way to use that huge pile of unlabeled scans to help the computer learn better, without needing a doctor to label every single one. Here is how they did it, using some simple analogies:

1. The Problem with "Random" Learning

Usually, when computers learn from unlabeled data, they use a trick called Contrastive Learning. Think of this like a game of "Find the Match."

  • The Old Way: You take one photo, apply random filters (like blurring it or changing the colors), and tell the computer, "These two look the same." Then you show it a totally different photo and say, "This one is different."
  • The Medical Problem: In medical images, those random filters are dangerous. If you blur an eye scan, you might accidentally blur out the tiny biomarker you are trying to find! It's like trying to find a specific crack in a windshield by smearing the glass with Vaseline. You might miss the crack entirely.

2. The New Idea: Grouping by "Sickness Level"

Instead of random filters, the authors asked: What if we group scans based on how "sick" they look?

Imagine a hospital waiting room.

  • Healthy people are sitting in one corner.
  • People with a slight cold are in another.
  • People with a severe flu are in a third.
  • People with a critical condition are in the ICU.

Even if we don't know exactly what disease each person has, we can tell who looks "sicker" than the others. The authors realized that scans with similar "sickness levels" (severity) likely share similar structural features, making them perfect "matches" for the computer to learn from.

3. The Magic Tool: The "Gradient Score"

But how do we know who is "sicker" without a doctor? We can't just ask the computer to guess.

The authors used a clever mathematical trick involving Gradients.

  • The Analogy: Imagine the computer has a "muscle memory" of what a healthy eye looks like.
  • When it looks at a healthy eye, it barely needs to adjust its thinking. It's like walking on a flat path; your muscles stay relaxed.
  • When it looks at a sick eye, it has to "stretch" and "strain" to understand what it's seeing. It has to make a big mental adjustment.

The authors measured exactly how much the computer had to "strain" (the gradient) to understand each scan.

  • Low Strain = Healthy (Low Severity Score).
  • High Strain = Sick (High Severity Score).

This gave them a "Severity Score" for every single unlabeled scan, effectively sorting the unlabeled pile into buckets from "Very Healthy" to "Very Sick."

4. The Training Process

Once they had these "Severity Buckets," they taught the computer in two steps:

  1. Step 1 (The Grouping): They told the computer, "All the scans in the 'Medium Sickness' bucket are similar to each other. All the scans in the 'Severe Sickness' bucket are similar to each other." They used this to build a strong mental map of eye structures.
  2. Step 2 (The Fine-Tuning): After the computer learned the general map, they took the small pile of actually labeled data (where doctors said "This is IRF," "This is DME," etc.) and gave the computer a quick final exam to learn the specific names of the diseases.

The Result

By using this "Severity Sorting" method instead of random blurring, the computer got much better at spotting the tiny biomarkers.

  • It improved accuracy by up to 6% compared to other methods.
  • It proved that you don't need to know the exact disease name to learn from data; you just need to know how "abnormal" the image looks compared to a healthy one.

In a Nutshell

Instead of guessing what's in the dark, the authors built a "sickness meter" using math. They sorted thousands of unlabeled eye scans from "Healthy" to "Sick" based on how hard the computer had to work to understand them. This allowed the computer to learn the patterns of disease much faster and more accurately, helping doctors detect eye diseases earlier.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →