Can Linguistically Related Languages Guide LLM Translation in Low-Resource Settings?

This paper investigates whether linguistically related pivot languages and few-shot demonstrations can guide Large Language Models in low-resource machine translation without parameter updates, finding that while such inference-time prompting offers modest improvements for poorly represented languages, its effectiveness is often inconsistent and highly sensitive to example construction.

Aishwarya Ramasethu, Niyathi Allu, Rohin Garg, Harshwardhan Fartale, Dun Li Chan

Published 2026-03-18
📖 4 min read☕ Coffee break read

Imagine you are trying to teach a brilliant, well-traveled chef (the Large Language Model or LLM) how to cook a very specific, rare dish from a remote village. The chef has never been to this village, doesn't speak the local dialect, and has no recipe book for it.

This paper asks a simple question: Can we help this chef cook the dish by showing them a recipe from a neighboring village that speaks a similar language, along with a few sample dishes?

Here is the breakdown of the study using everyday analogies:

1. The Problem: The "Language Gap"

Big AI models are like super-chefs who know how to cook French, Italian, and Chinese cuisine perfectly. But for languages that are rare or "low-resource" (like Konkani in India or Tunisian Arabic), the chef has never seen the ingredients. If you just ask the chef to "Cook this in Konkani," they might guess and serve you Italian food instead because that's what they know best.

Usually, to fix this, you'd need to hire a team of linguists to write thousands of new recipes (training data) and retrain the chef. But that takes too much time and money.

2. The Solution: The "Pivot" and the "Cheat Sheet"

The researchers tried a lighter approach. Instead of retraining the chef, they used two tricks during the cooking process:

  • The Pivot Language (The Bridge): They picked a language the chef does know that is similar to the target.

    • For Konkani, they used Marathi (a neighboring language).
    • For Tunisian Arabic, they used Modern Standard Arabic (the formal version).
    • Analogy: It's like telling the chef, "First, translate this English order into Marathi (which you know), and then use that Marathi version as a bridge to figure out the Konkani."
  • Few-Shot Examples (The Cheat Sheet): They showed the chef 3 to 5 examples of "English → Marathi → Konkani" right before asking them to cook the new dish.

    • Analogy: It's like handing the chef a small notepad that says, "See how we turned 'Hello' into Konkani? Do it like that."

3. The Experiment: Two Different Chefs

They tested this method on two different "chefs" (AI models):

  1. Hermes: A general-purpose chef who is good at many things but not specialized in translation.
  2. Tower: A chef who was specifically trained to be a translator.

They tested this on two languages:

  • Konkani: A language with very few digital resources (like a remote village with no internet).
  • Tunisian Arabic: A dialect that is slightly better represented in the chef's memory because it shares a script with the formal Arabic they know.

4. The Results: It Depends on the Village

The findings were mixed, like a weather report:

  • For the "Remote Village" (Konkani): The method worked! The "Pivot" (Marathi) and the "Cheat Sheet" (examples) helped the chef stop guessing and actually produce Konkani. The translation quality improved, though it wasn't perfect.

    • Takeaway: When the chef knows nothing about the target, a linguistic cousin (pivot) acts as a helpful crutch.
  • For the "Semi-Known Village" (Tunisian Arabic): The method didn't help much. The chef was already pretty good at this because the script and words were similar to what they already knew. Adding the extra steps (the pivot) didn't make the dish taste better; sometimes it even confused the chef.

    • Takeaway: If the chef already has a decent idea of the language, adding a bridge might just be extra noise.
  • The "Too Much Info" Problem: They found that giving the chef more examples (more than 3 or 4) actually made the results worse.

    • Analogy: It's like giving a chef a 50-page manual when they only needed a 3-step sticky note. The chef got overwhelmed and started making mistakes.

5. The Big Conclusion

This paper proves that you don't always need to rebuild the whole kitchen (retrain the AI) to cook a new dish.

  • When it works: If the target language is totally new to the AI, using a "cousin" language as a bridge and showing a few examples can guide the AI to the right answer without needing expensive training.
  • When it fails: If the AI already knows the language well, or if you give it too many confusing examples, this trick doesn't help.

In short: Think of this as a "lightweight" translation hack. It's not a magic wand that solves everything, but for languages that are currently ignored by big tech, it's a clever, low-cost way to get a decent meal on the table.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →