Analyzing the Performance of ChatGPT in Cardiology and Vascular Pathologies

This study demonstrates that ChatGPT outperformed two medical students in answering challenging cardiology and vascular pathology questions, achieving a 92.10% accuracy rate on a 190-question dataset, thereby highlighting its potential as a valuable tool for medical education.

Walid Hariri

Published Mon, 09 Ma
📖 3 min read☕ Coffee break read

Imagine a high-stakes cooking competition. On one side, you have two of the most talented, hard-working apprentice chefs in the city (the medical students). On the other side, you have a super-intelligent, all-knowing digital chef who has read every cookbook, recipe blog, and culinary magazine ever written (that's ChatGPT).

The challenge? A massive, difficult tasting menu focused entirely on heart and blood vessel diseases (Cardiology and Vascular Pathologies). The judges handed them 190 specific questions about things like heart attacks, blood pressure meds, and fainting spells.

Here is how the competition played out, broken down simply:

The Setup: The "Heart" of the Matter

The researchers wanted to see if this AI "digital chef" could handle the complex, tricky recipes of heart medicine better than real human students. They didn't just pick easy questions; they grabbed 190 tough questions from a French medical exam platform used by students preparing for their final residency tests.

The topics were like specific ingredients in a complex stew:

  • The "Leaky Pipe" (Aortic Aneurysm): A dangerous bulge in the main body artery.
  • The "Pressure Regulators" (Antihypertensives): Drugs to lower high blood pressure.
  • The "Heart's Electrical Map" (ECG): Reading the heart's rhythm like a code.
  • The "Traffic Jam" (AV Block): When the heart's electrical signals get stuck.
  • The "Swollen Roads" (Varicose Veins): Twisted veins in the legs.
  • The "Breathless Engine" (Chronic Pulmonary Heart Disease): When the heart struggles because the lungs are tired.
  • The "Blackout" (Syncope): Fainting spells.

The Results: Who Ate the Most?

When the scores were tallied, the results were surprising:

  • The Human Apprentices: The two top students did a great job. One got 85.8% right, and the other got 82.6% right. They knew their stuff!
  • The Digital Chef (ChatGPT): The AI didn't just keep up; it soared. It got 92.1% of the answers correct.

The Analogy: If the exam was a 100-question test, the students missed about 14 to 18 questions. ChatGPT only missed about 8. It beat the humans by a solid margin, proving that for this specific subject, the AI is a very sharp study partner.

The "Glitch" in the Recipe

However, the digital chef isn't perfect. The paper noted one specific weakness: Math and Units.

Think of it like this: If you ask the AI, "How much salt is in this soup?" it might know the concept of salt perfectly. But if the question involves tricky math with different measurement units (like converting milligrams to grams while calculating a dosage), the AI sometimes gets confused and picks the wrong answer. It's like a chef who knows the flavor profile perfectly but occasionally messes up the measuring cups.

The Big Takeaway

This study is like a proof-of-concept for the future of medical school. It shows that ChatGPT is a powerful tool for learning about heart and blood vessel diseases. It can act like a super-tutor that is always awake, always ready, and knows more than almost any single student.

But here is the catch: Just because the AI scored higher doesn't mean we should replace the students. The AI still needs to get better at the tricky math parts. The goal isn't to let the robot take the exam for the doctor, but to use the robot to help the doctor study harder and smarter before they ever see a real patient.

In short: The AI is currently the "honor student" of the class, but the human doctors are still the ones who need to learn how to cook the meal safely for real people.