Less is More: Adapting Text Embeddings for Low-Resource Languages with Small Scale Noisy Synthetic Data

This paper demonstrates that fine-tuning multilingual text embedding models on a small amount of noisy synthetic data is surprisingly effective for low-resource languages, achieving performance comparable to models trained on massive datasets while challenging the assumption that large-scale, high-quality data is necessary for semantic alignment.

Zaruhi Navasardyan, Spartak Bughdaryan, Bagrat Minasyan, Hrant Davtyan

Published 2026-03-25
📖 4 min read☕ Coffee break read

Imagine you want to teach a brilliant, multilingual librarian (an AI model) how to understand a specific, rare language like Armenian. Usually, to do this, you'd think you need a massive library of perfect, human-written books in that language, or you'd need to hire a team of expert translators to create millions of perfect sentence pairs. It's expensive, slow, and often impossible for languages with few speakers.

This paper says: "Actually, you don't need a library. You just need a few messy, handwritten notes."

Here is the story of their discovery, explained simply:

1. The Problem: The "Perfect Data" Myth

For years, AI researchers believed that to make a computer understand a low-resource language (like Armenian, Georgian, or others with unique scripts), you needed massive amounts of perfect data. They thought, "If the translation isn't perfect, the AI will get confused."

2. The Experiment: The "Messy Reddit" Approach

The researchers decided to test a different idea. Instead of hiring experts, they took a huge pile of English posts from Reddit (titles and their corresponding stories) and asked a standard AI translator to turn them into Armenian.

The Result? The translation was terrible.

  • The grammar was broken.
  • The punctuation was wrong.
  • Some words were translated literally and made no sense.
  • It was "noisy" data—like trying to learn a language by listening to a radio with static.

They took this "garbage" data, cleaned it up just a tiny bit to make sure the meaning was still there, and used it to teach the AI.

3. The "Less is More" Surprise

Here is the magic part:

  • They trained the AI on just 10,000 of these messy pairs.
  • Then, they tried training it on 1 million pairs (100 times more data).

The Result: The AI trained on just 10,000 messy pairs performed just as well (and sometimes better) than the one trained on 1 million pairs.

The Analogy:
Imagine you are trying to teach someone to recognize a specific type of dog (a Golden Retriever).

  • The Old Way: You show them 1,000,000 perfect, high-definition photos of Golden Retrievers in a studio.
  • The New Way: You show them 10,000 blurry, slightly tilted, poorly lit photos of Golden Retrievers taken by a shaky hand.

Surprisingly, the person learns to recognize the dog just as well from the blurry photos. Why? Because the core shape of the dog is there. Once the brain (or AI) understands the basic shape, showing it more blurry photos doesn't help much. In fact, showing too many might just confuse them.

4. Why This Matters

  • It's Cheap: You don't need to hire expensive human translators. You can use free, open-source AI tools to generate the data.
  • It's Fast: You only need a tiny dataset (10k examples) instead of millions.
  • It Works for Others: They tested this on Georgian (another language with a unique script) and it worked there too.

5. The Big Takeaway

The paper proves that for AI, semantic alignment (teaching the AI that "dog" in English means "dog" in Armenian) doesn't require perfection. It just requires the gist to be correct.

The AI is like a resilient student. It doesn't need a perfect textbook; it just needs enough "rough drafts" to figure out the pattern. Once it figures out the pattern, it stops learning from new data and just gets confused.

In short: If you want to build AI tools for rare languages, stop waiting for perfect data. Grab some messy, imperfect data, use a little bit of it, and you'll get amazing results. Less is truly More.