Agri-Query: A Case Study on RAG vs. Long-Context LLMs for Cross-Lingual Technical Question Answering

This paper presents a case study demonstrating that Hybrid Retrieval-Augmented Generation (RAG) consistently outperforms direct long-context prompting across English, French, and German technical agricultural manuals, achieving over 85% accuracy with models like Gemini 2.5 Flash and Qwen 2.5 7B in cross-lingual question answering.

Julius Gun, Timo Oksanen

Published 2026-03-09
📖 5 min read🧠 Deep dive

Imagine you are a farmer standing in front of a massive, 165-page instruction manual for a high-tech fertilizer machine. The manual is written in English, French, and German. You have a specific question: "What is the exact torque for the vane lock nuts?"

Now, imagine you have two different ways to find that answer:

  1. The "Super-Reader" (Long-Context LLM): You hand the entire 165-page book to a genius student who can read 128,000 words at once. You ask them to find the answer.
  2. The "Librarian" (RAG): You ask a librarian to quickly scan the book, find the three most relevant pages, and hand only those pages to the student. Then, the student reads just those pages to answer you.

This paper, "Agri-Query," is a race between these two approaches to see who is better at finding the needle in the haystack.

The Setup: A Real-World Test

The researchers from the Technical University of Munich didn't just use random text. They used a real, complex manual for an agricultural machine (the Kverneland Exacta-TLX). They created a test with 108 questions:

  • 54 "Findable" questions: The answer is definitely in the book.
  • 54 "Unanswerable" questions: The answer is not in the book (e.g., "How much extra diesel does the tractor use?"). This was a trap to see if the AI would lie (hallucinate) just to be helpful.

They tested this in three languages: English, French, and German. Crucially, they asked all questions in English, even when the manual was in French or German, to see if the AI could bridge the language gap.

The Contenders

They pitted 9 different AI models (including famous ones like Llama, Qwen, and the proprietary Gemini) against each other. Some were huge, some were small. They tested them in two modes:

  1. Direct Long-Context: Feeding the whole book to the AI.
  2. RAG (Retrieval-Augmented Generation): Using a search system to find the right pages first.

They tested three types of "Librarians" (Retrieval methods):

  • Keyword Search: Like using "Ctrl+F" to find exact words.
  • Semantic Search: Like asking a librarian, "I need the part about tightening nuts," even if the word "nut" isn't used.
  • Hybrid Search: A mix of both, using the best of both worlds.

The Results: The Librarian Wins!

Here is the big takeaway, explained simply:

1. The "Super-Reader" gets overwhelmed.
When the AI tried to read the entire 165-page manual at once, it struggled. It suffered from what researchers call the "Lost in the Middle" effect. Imagine reading a 500-page novel and being asked about a detail on page 250. If you read the whole thing in one go, you might forget the middle parts. The AI got confused by all the "noise" (irrelevant text) and often missed the answer or made things up.

2. The "Hybrid Librarian" is the champion.
The Hybrid RAG approach (Keyword + Semantic search) consistently won. By finding the specific pages first and handing them to the AI, the models performed much better.

  • Accuracy: They got over 85% of the answers right.
  • Honesty: They were much better at admitting, "I don't know," when the answer wasn't in the book, rather than making up a fake answer.
  • Small Models Shine: Surprisingly, smaller, cheaper AI models (like the 7B or 3B parameter versions) performed just as well as the giant models when they had the Librarian helping them. This means you don't need a supercomputer to get great results; you just need a good search system.

3. Language is no barrier.
Even though the questions were in English and the manuals were in French or German, the Hybrid RAG system worked brilliantly. The "Librarian" understood the meaning of the question and found the right French or German pages, then the AI translated the answer back to English.

The "Hallucination" Trap

One of the most important findings was about lying.

  • When the AI had to read the whole book (Long-Context), it often tried to guess the answer to the "Unanswerable" questions, leading to hallucinations (confidently wrong answers).
  • When the AI used the Librarian (RAG), it was much more honest. If the Librarian couldn't find the page, the AI was more likely to say, "Not found," rather than inventing a story.

The Bottom Line

If you are building an AI system to help people read technical manuals (like for farming, medicine, or engineering):

  • Don't just dump the whole book into the AI. It gets confused and loses focus.
  • Use a "Search First" strategy (RAG). Find the relevant pages, then ask the AI to read them.
  • Mix your search methods. Combining keyword search with "meaning" search gives the best results.
  • You don't need the biggest AI. A smaller, cheaper AI works great if it has a good search tool helping it.

In short: Don't ask the AI to memorize the library; ask it to use the card catalog.