Automated echocardiographic measurements for longitudinal monitoring of ATTR cardiomyopathy: agreement and repeatability analysis

This study demonstrates that while fully automated AI-assisted echocardiographic measurements show moderate agreement with expert cardiologists due to systematic biases, their repeatability is comparable to experienced readers, supporting their utility for longitudinal monitoring of ATTR cardiomyopathy.

Walser, A., Clerc, O. F., Mork, C., Flammer, A. J., Myhre, P. L., Schwotzer, R., Graeni, C., Ruschitzka, F., Tanner, F. C., Benz, D. C.

Published 2026-04-07
📖 4 min read☕ Coffee break read
⚕️

This is an AI-generated explanation of a preprint that has not been peer-reviewed. It is not medical advice. Do not make health decisions based on this content. Read full disclaimer

Imagine your heart is a house, and ATTR cardiomyopathy is a slow, creeping mold that hardens the walls and makes the rooms smaller over time. To stop this mold from taking over, doctors need to catch the changes early. The best tool they have to peek inside the house is an echocardiogram (an ultrasound of the heart), but there's a catch: it's like trying to measure a room with a ruler held by a human hand. Depending on who is holding it, how tired they are, or how they angle the ruler, the measurements can change slightly. This "human error" makes it hard to tell if the house is actually getting smaller or if the measurer just had a bad day.

This study asked a big question: Can a robot (Artificial Intelligence) measure the heart as reliably as a human doctor, so we can track the disease over time without the "human shake"?

Here is the breakdown of what they did and what they found, using some everyday analogies:

The Experiment: A "Measuring Contest"

The researchers gathered 62 patients with this heart condition. They looked at 178 different heart scans taken over a year. To test the waters, they had four different "measurers" look at the same scans:

  1. The Expert: A seasoned cardiologist (the gold standard).
  2. The Second Expert: Another experienced doctor.
  3. The Novice: A doctor-in-training (less experienced).
  4. The Robot: An AI algorithm (Us2.ai) that automatically measures the heart without human hands.

The Results: Who Measured Best?

1. The Experts vs. The Robot (Agreement)
When the robot tried to measure the thickness of the heart walls and the size of the heart's main chamber, it didn't always agree perfectly with the top expert.

  • The Analogy: Imagine two people measuring a table. The expert says it's 60 inches; the robot says it's 58 inches. They are close, but there's a consistent difference (a "bias").
  • The Reality: The robot was "moderately" good at agreeing with the expert. It consistently measured the heart walls slightly thinner and the chamber slightly smaller than the human expert did.

2. The Experts vs. Each Other (Reliability)
When the two expert doctors measured the same heart, they agreed very well.

  • The Analogy: Two master carpenters measuring the same table will get almost the exact same number.
  • The Reality: Their measurements were very consistent with each other, showing that experienced humans are a solid benchmark.

3. The "Bad Day" Test (Repeatability)
This is the most important part for tracking disease. The study asked: If the same person measures the same heart twice, how much does the number jump around?

  • The Novice: Like a student learning to use a ruler, their numbers jumped around a lot (high variability). If they measured the heart on Monday and then again on Tuesday, the numbers might change significantly just because they were shaky.
  • The Experts: They were very steady. Their numbers stayed consistent.
  • The Robot: This was the big win. Even though the robot didn't agree perfectly with the expert on the exact number, it was extremely consistent with itself. If the robot measured the heart today and then again tomorrow, it gave the exact same result every time.

The Bottom Line: Why This Matters

Think of tracking a disease like watching a garden grow.

  • The Problem: If you use a shaky hand to measure a plant, you might think it grew 2 inches when it only grew 1 inch, or vice versa. You can't be sure if the fertilizer is working.
  • The Solution: The AI is like a laser ruler. Even if the laser ruler is calibrated to be 1 inch shorter than your tape measure (the systematic bias), it never wobbles.

The Conclusion:
The study found that while the AI isn't a perfect copy of a human doctor's specific numbers, it is just as steady and reliable as an experienced doctor when measuring changes over time.

Because the AI is so consistent, it can act as a "steady hand" for doctors. It can detect the subtle changes in the heart's size that signal the disease is getting worse or better, without the "noise" of human error. This means doctors can use AI to personalize treatment plans with much more confidence, knowing that if the numbers change, it's the heart changing, not the person holding the ruler.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →