Meenz bleibt Meenz, but Large Language Models Do Not Speak Its Dialect

This paper introduces the first NLP dataset for the endangered Meenzerisch dialect of Mainz and demonstrates that current large language models struggle significantly to generate or define its words, achieving accuracy rates below 10% even with few-shot learning and rule extraction, thereby highlighting an urgent need for further research and resources to preserve German dialects.

Minh Duc Bui, Manuel Mager, Peter Herbert Kann, Katharina von der Wense

Published 2026-03-05
📖 5 min read🧠 Deep dive

Here is an explanation of the paper, "Meenz bleibt Meenz, but Large Language Models Do Not Speak Its Dialect," translated into simple, everyday language with some creative analogies.

The Big Idea: The "Lost in Translation" Problem

Imagine you have a very old, local recipe book written in a secret code that only your great-grandparents in a specific German town (Mainz) understand. This code is called Meenzerisch. It's the language of their carnival, their jokes, and their history.

Now, imagine you ask a super-smart, world-famous robot chef (a Large Language Model or LLM) to translate that recipe book into modern English (Standard German). You expect the robot to be amazing because it knows everything about the internet.

The bad news: The robot chef is completely lost. It doesn't just make a few mistakes; it barely understands the code at all.

This paper is the first time researchers tried to teach these super-smart robots about the Mainz dialect, and the results were a harsh reality check: The robots are failing miserably.


Step 1: Building the Dictionary (The "Digital Archaeology")

Before they could test the robots, the researchers had to build a dictionary. Since Meenzerisch is mostly spoken and rarely written down in digital formats, they had to do some "digital archaeology."

  • The Source: They found a physical book from 1966 called the Mainzer Wörterbuch.
  • The Process:
    1. They took photos of the pages (Scanning).
    2. They used software to turn the pictures into text (OCR), like a scanner trying to read handwriting.
    3. Humans had to fix the scanner's mistakes (Manual Correction).
    4. They used a smart AI to pull out the words and their meanings (Extraction).
  • The Result: They created a digital list of 2,351 words. Each entry has a word in the Mainz dialect and its meaning in standard German. Think of it as a Rosetta Stone for Mainz.

Step 2: The Two Big Tests

With their new dictionary, the researchers put the robots to work with two specific challenges.

Challenge A: The Translator (Definition Generation)

The Task: The robot is given a weird Mainz word (e.g., "Schwollescheer") and asked, "What does this mean in normal German?"
The Analogy: Imagine showing a robot a picture of a "Glocke" (a bell) but calling it a "Schwollescheer." You ask, "What is this?"
The Result: The robots were terrible.

  • The best robot (Llama-3.3) got it right only 6.27% of the time.
  • The average robot got it right less than 5% of the time.
  • Comparison: When asked to translate normal English words, these same robots got it right 86% of the time. They aren't "dumb"; they just don't know this specific dialect.

Challenge B: The Word Generator (Word Generation)

The Task: The robot is given a description in normal German (e.g., "A person who rides horses in a military unit") and asked, "What is the Mainz word for this?"
The Analogy: You tell the robot, "I need the word for 'hunger' in the Mainz dialect."
The Result: This was even worse.

  • The best robot got it right only 1.51% of the time.
  • Most robots were essentially guessing randomly, getting it right less than 1% of the time.

Step 3: Trying to Help the Robots (The "Cheat Sheets")

The researchers didn't give up. They tried two tricks to help the robots learn faster, like giving a student a cheat sheet before a test.

  1. Few-Shot Learning (The "Example" Trick):

    • They showed the robot a few examples first: "Here is a word and its meaning. Here is another. Now, guess this one."
    • Result: It helped a tiny bit, but accuracy only went up to about 9%. It's like giving a student a few practice questions; they still don't understand the subject.
  2. Rule Extraction (The "Grammar Book" Trick):

    • They asked a super-smart AI to write down the rules of the dialect (e.g., "In Mainz, they often change 'en' to 'ele'"). They fed these rules to the other robots.
    • Result: This helped slightly for translating words to meanings, but actually made things slightly worse for guessing the words from meanings. The robots still couldn't apply the rules correctly.

Why Does This Matter?

You might ask, "So what? Who cares if a robot knows a German dialect?"

  1. Cultural Preservation: Dialects are dying out. If we want to save them, we need technology that understands them. If robots can't understand them, they can't help preserve them.
  2. Bias: If a robot only understands "Standard" German, it treats people who speak dialects like they are "broken" or "wrong." This paper shows that the robots are currently biased against these speakers.
  3. The "Low-Resource" Problem: The internet is full of data for English and Standard German. It is almost empty for dialects. The robots are like chefs who have read every cookbook in the world, except for the one from your hometown.

The Conclusion

The paper ends with a clear message: Current AI is not ready for local dialects.

Even the smartest, most expensive robots on the planet are failing to understand the language of Mainz. They need more data, more research, and a lot more help from the people who actually speak the language.

The Takeaway: Just because a robot is "smart" doesn't mean it knows your language. For the dialect of Mainz, the robot is currently as lost as a tourist without a map. As the title says: "Meenz bleibt Meenz" (Mainz remains Mainz), but the robots haven't figured out how to speak its dialect yet.