Cluster-First Labelling: An Automated Pipeline for Segmentation and Morphological Clustering in Histology Whole Slide Images

This paper introduces a cloud-native, end-to-end pipeline that automates histology whole slide image annotation by segmenting tissue components, clustering them morphologically, and enabling human annotators to label representative clusters instead of individual objects, thereby reducing effort while achieving 96.8% alignment accuracy with manual labels across diverse tissue types.

Original authors: Muhammad Haseeb Ahmad, Sharmila Rajendran, Damion Young, Jon Mason

Published 2026-04-13
📖 5 min read🧠 Deep dive
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine you are a librarian tasked with organizing a library that contains 100,000 books, but there's a catch: every single book is slightly different, and you have to read the cover of every single one to decide which shelf it belongs on. If you did this one by one, it would take you years.

This is exactly the problem scientists face with Histology Whole Slide Images (WSIs). These are giant, high-resolution digital photos of tissue samples (like skin, bone, or organs). A single slide can contain tens of thousands of tiny "objects" (cells, nuclei, clusters). Traditionally, a human expert has to zoom in, draw a line around every single object, and label it. It's slow, expensive, and exhausting.

This paper introduces a clever new system called "Cluster-First Labelling" that solves this by changing the workflow entirely. Here is how it works, using simple analogies:

1. The Old Way: The "One-by-One" Struggle

The Problem: Imagine trying to sort a massive pile of mixed-up LEGO bricks. The old way is to pick up every single brick, look at it, and decide: "Is this a 2x4 red brick? Is this a 1x2 blue brick?" You do this for 15,000 bricks. It takes forever.

2. The New Way: The "Group First" Strategy

The authors built an automated pipeline that acts like a super-smart, tireless robot assistant. Instead of sorting bricks one by one, it does this:

Step A: The Great Slicing (Tiling)

The giant image is chopped up into smaller, manageable puzzle pieces (tiles), like cutting a large pizza into slices so it's easier to eat.

Step B: The "Trash" Filter (Quality Control)

The robot looks at each slice. If a slice is just empty white space or blurry (like a slice of pizza with no toppings), it throws it away immediately. This saves time.

Step C: The "Shape Shifter" (Segmentation)

Using a tool called Cellpose-SAM, the robot finds every object that looks like a cell or a nucleus.

  • Analogy: Imagine the robot is a magic marker that draws a perfect outline around every single LEGO brick in the pile, regardless of what it is. It doesn't know what the brick is yet, but it knows where it is.

Step D: The "ID Card" Generator (Embedding)

The robot takes a picture of each outlined object and runs it through a neural network (ResNet-50). This creates a unique "ID card" (a mathematical fingerprint) for every object based on its shape and texture.

  • Analogy: It's like scanning every LEGO brick and giving it a barcode that says "Red, 2x4, smooth" or "Blue, 1x2, bumpy."

Step E: The "Sorter" (Clustering)

This is the magic step. The robot uses a clustering algorithm (DBSCAN) to group objects that have similar "ID cards."

  • Analogy: Instead of you sorting the bricks, the robot automatically piles all the "Red 2x4s" into one bin, all the "Blue 1x2s" into another, and all the "weird green pieces" into a third. It does this without being told what the bins are called.

3. The Human's New Job: The "Bin Manager"

Now, the human expert doesn't have to look at 15,000 individual bricks. They only have to look at the bins.

  • If the robot made a bin full of "Red 2x4s," the human just looks at a few samples, says, "Yes, that's a Red 2x4," and labels the entire bin.
  • The computer then instantly applies that label to every single brick in that bin.
  • The Result: Instead of 15,000 tasks, the human might only have 25 tasks (one for each bin). That is a 600x reduction in work.

4. Did It Work? (The Results)

The team tested this on 3,696 objects from 13 different types of tissues (human, rat, and rabbit).

  • The Score: The system matched human labels 96.8% of the time.
  • The Perfect Scores: For 7 out of the 13 tissue types, the robot got 100% agreement with the humans.
  • The Struggles: It had a bit of trouble with "Compact Bone" and "Skeletal Muscle."
    • Why? Bone has very few cells per image (making it hard for the robot to find patterns), and muscle has many different-looking parts that look similar to the robot but are actually different to a human eye. It's like trying to sort a pile of identical-looking rocks; the robot needs a little more help there.

Why This Matters

This system is Open Source (free for everyone to use). It turns a job that used to take days of expert time into a process that takes minutes of computer time and minutes of human review.

In a nutshell:
Instead of asking a human to sort a mountain of sand grain by grain, this system asks the computer to group the sand into piles based on color and texture, and then asks the human to just name the piles. It's faster, cheaper, and surprisingly accurate.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →