PVminerLLM: Structured Extraction of Patient Voice from Patient-Generated Text using Large Language Models

This paper introduces PVminer, a benchmark for structured extraction of patient voice from patient-generated text, and presents PVminerLLM, a supervised fine-tuned large language model that significantly outperforms prompt-based baselines in extracting codes, sub-codes, and evidence spans to enable scalable analysis of non-clinical health drivers.

Samah Fodeh, Linhai Ma, Ganesh Puthiaraju, Srivani Talakokkul, Afshan Khan, Ashley Hagaman, Sarah Lowe, Aimee Roundtree

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine you are a doctor trying to understand your patients. You have their medical charts, which are like spreadsheets: they tell you exactly what medicine they take, their blood pressure numbers, and their test results. These are easy to read and analyze.

But patients also send you messages, fill out surveys, and tell stories about their lives. These are like journals written in a messy, emotional handwriting. They might say, "I'm scared to take this pill because I can't afford it," or "My landlord is evicting me, so I'm stressed."

This paper is about a new tool called PVminerLLM that helps doctors and researchers read those messy journals and turn them into neat, organized data, just like the spreadsheets.

Here is the story of how they built it, explained simply:

1. The Problem: The "Lost in Translation" Gap

For a long time, computers were great at reading medical charts (the spreadsheets) but terrible at reading patient stories (the journals).

  • The Challenge: A patient's message is full of hidden clues. They might mention "housing instability" (they might lose their home) or "shared decision-making" (they want to help choose their treatment).
  • The Old Way: Humans had to read every single message and manually write down these clues. This is like hiring a team of translators to read a million letters by hand. It's slow, expensive, and impossible to do for everyone.
  • The New Hope: We have powerful AI (Large Language Models) that can read and write like humans. But when the researchers asked these AIs to read the patient messages, the AIs got confused. They would give vague answers, miss the point, or format the data messily. It was like asking a brilliant student to fill out a very strict, complicated government form without any practice—they knew the material, but they kept failing the specific test.

2. The Solution: Two Steps to Success

The researchers tried two approaches to teach the AI how to read these patient voices.

Step A: The "Strict Teacher" (Prompt Engineering)

First, they tried to be very specific with the AI. They wrote a super-detailed set of instructions (a "prompt") telling the AI exactly how to format the answer, what categories to look for, and how to quote the text.

  • The Analogy: Imagine giving a robot a recipe and saying, "If you see a tomato, write 'Tomato' in column 3. If you see a sad face, write 'Sad' in column 4."
  • The Result: It helped a little. The robot got better at following rules, but it still missed the subtle meanings. It was like a robot that could follow the recipe but didn't understand why the ingredients mattered. It was still making mistakes, especially with rare or tricky topics.

Step B: The "Apprentice" (Supervised Fine-Tuning)

This was the big breakthrough. Instead of just giving the AI instructions, they trained it. They showed the AI thousands of examples of patient messages and the correct way to label them. They let the AI practice, make mistakes, and learn from the corrections.

  • The Analogy: Instead of just giving the robot a recipe, they hired a master chef (the human experts) to stand next to the robot for weeks, correcting its chopping, seasoning, and plating until the robot learned the art of cooking, not just the rules.
  • The Result: This worked amazingly well. The AI (now called PVminerLLM) became an expert. It could read a messy patient message and instantly pull out the specific clues: "Ah, this person is worried about money (Social Determinant of Health)" or "This person wants to be part of the decision (Shared Decision Making)."

3. The Surprise: You Don't Need a Giant Brain

Usually, people think you need the biggest, most expensive AI supercomputer to do hard tasks.

  • The Discovery: The researchers found that even smaller, cheaper AI models performed just as well as the giant ones after they were trained (fine-tuned).
  • The Metaphor: It's like realizing you don't need a PhD in physics to fix a leaky faucet; you just need the right tool and a little practice. A small, specialized AI can do this job better than a giant, general AI that hasn't been trained for it. This means hospitals with smaller budgets can use this tool too.

4. Why This Matters (The "So What?")

Why do we care about turning patient stories into data?

  • Seeing the Invisible: Right now, if a patient is struggling to pay for rent, that stress might not show up in their medical chart. But it affects their health! This tool makes those invisible struggles visible to doctors.
  • Better Care: If a doctor knows a patient is stressed about housing, they can connect them with a social worker, not just prescribe more pills.
  • Fairness: It helps ensure that the voices of people from different backgrounds are heard and counted, leading to fairer healthcare for everyone.

Summary

Think of PVminerLLM as a super-powered translator.

  • Before: Patient stories were like a pile of unsorted letters in a language only a few humans could read.
  • After: This tool reads those letters and instantly sorts them into neat folders: "Money Worries," "Emotional Support," "Treatment Choices."
  • The Magic: It doesn't need a super-computer to do it; a trained, smaller AI can do the job perfectly, helping doctors understand their patients' real lives, not just their medical numbers.