From Prompting to Preference Optimization: A Comparative Study of LLM-based Automated Essay Scoring

This paper presents a comprehensive comparative study of four major LLM-based paradigms for Automated Essay Scoring on IELTS Writing Task 2, revealing that a configuration combining k-SFT and RAG achieves the strongest overall performance with a 93% F1-Score while highlighting critical accuracy-cost-robustness trade-offs.

Minh Hoang Nguyen, Vu Hoang Pham, Xuan Thanh Huynh, Phuc Hong Mai, Vinh The Nguyen, Quang Nhut Huynh, Huy Tien Nguyen, Tung Le

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine you are a teacher with a mountain of essays to grade. You have to read every single one, check if the student answered the prompt, if their grammar is good, and if their ideas flow well. It's exhausting, takes forever, and sometimes you might be tired and give a slightly different score than you would have on a fresh day.

This is the problem Automated Essay Scoring (AES) tries to solve: building a robot teacher that can grade essays as fairly and quickly as a human.

For a long time, we tried to build these robots using "old-school" math (looking at word counts and sentence length) and then "deep learning" (teaching computers to recognize patterns). But now, we have Large Language Models (LLMs)—the same super-smart AI brains that power chatbots like the one you're talking to right now.

This paper is a taste test. The researchers wanted to find out: Which way of using these AI brains works best for grading English essays? They didn't just pick one method; they cooked up four different "recipes" and compared them on the same batch of IELTS essays (a tough English test for non-native speakers).

Here is a breakdown of their four recipes, using some fun analogies:

The Four Recipes for the Robot Teacher

1. The "Strict Accountant" (Discriminative Fine-Tuning)

  • How it works: You take a standard AI model and force it to memorize thousands of essays and their scores until it becomes a master of spotting patterns. It's like training a dog to sit by repeating "sit" over and over.
  • The Vibe: It's reliable but rigid. It's good at math but bad at understanding the feeling of the essay. It's like a calculator that can add numbers perfectly but doesn't understand why you're adding them.
  • Result: It was okay, but not great. It missed the nuance.

2. The "Lazy Intern" (Prompting / Zero-Shot)

  • How it works: You don't train the AI at all. You just ask it nicely: "Hey AI, please grade this essay like an IELTS examiner." You might even show it two examples first (Few-Shot).
  • The Vibe: This is like hiring a brilliant intern who has read the whole library but has never graded a paper before. Sometimes they are amazing because they are smart; other times, they get confused by how you asked the question.
  • Result: It was hit-or-miss. Sometimes it got it right, sometimes it was wildly off. Also, using the "smartest" (most expensive) AI models was very costly.

3. The "Specialized Tutor with a Textbook" (Instruction Tuning + RAG)

  • How it works: This was the champion of the study.
    • Instruction Tuning: They taught the AI specifically how to be an IELTS examiner, breaking the job down into four specific tasks (Task Response, Coherence, Vocabulary, Grammar).
    • RAG (Retrieval-Augmented Generation): They gave the AI a "cheat sheet" (a database of perfect essays and scoring rules) to look at while it grades.
  • The Vibe: Imagine a tutor who has studied the official grading rules inside out and has a stack of "model essays" right next to them for reference. They don't guess; they check the rules and compare the student's work to the examples.
  • Result: This was the winner. It got the score right 93% of the time (F1-Score). It was accurate, consistent, and didn't hallucinate (make things up).

4. The "Human-Mimic Coach" (Supervised Fine-Tuning + Preference Optimization)

  • How it works: This is the "Specialized Tutor" from Recipe #3, but with an extra step. After learning the rules, they showed the AI examples of "good feedback" vs. "bad feedback" and taught it to prefer the human-like style.
  • The Vibe: This AI is like a coach who not only knows the rules but knows how to talk to students. If the student made a mistake, this AI doesn't just say "Wrong"; it says, "Your ideas are a bit scattered here, try connecting them more smoothly."
  • Result: It was slightly less accurate at getting the exact number right compared to Recipe #3, but the comments it wrote were much more natural, helpful, and human-like.

The Big Takeaways (The "So What?")

The researchers found a classic trade-off, like choosing between a sports car and a family van:

  • The "Specialized Tutor" (Recipe #3) is the Sports Car. It is fast, incredibly accurate, and great for high-stakes testing (like official exams where you need the exact score). It's the best for getting the number right.
  • The "Human-Mimic Coach" (Recipe #4) is the Family Van. It might be slightly slower or less precise with the exact number, but it's much better at giving a comfortable, helpful ride. It's perfect for students who need to learn and improve, not just get a grade.

The Cost Factor:
They also looked at how much money and time each method cost.

  • The "Lazy Intern" (just asking the AI) was cheap to set up but expensive to run (because you have to pay for the big AI every time).
  • The "Specialized Tutor" took some time to train but was very efficient and accurate.
  • The "Human-Mimic Coach" took the most time and money to train, and the extra accuracy didn't always justify the huge cost.

The Final Verdict

If you are building a system to grade thousands of essays for a big exam, use the "Specialized Tutor" (Recipe #3). It's the most accurate and reliable.

If you are building a tool to help students practice and learn, use the "Human-Mimic Coach" (Recipe #4). It gives feedback that feels like a real teacher talking to you.

The paper proves that we don't have to choose between "smart AI" and "human-like AI." By combining the right training methods with a "cheat sheet" of examples, we can build robot teachers that are both accurate and kind.