MCIF: Multimodal Crosslingual Instruction-Following Benchmark from Scientific Talks

This paper introduces MCIF, the first crosslingual, human-annotated benchmark based on scientific talks that systematically evaluates multimodal large language models across four languages, three modalities, and varying input lengths to address gaps in existing assessments of instruction-following capabilities.

Sara Papi, Maike Züfle, Marco Gaido, Beatrice Savoldi, Danni Liu, Ioannis Douros, Luisa Bentivogli, Jan Niehues

Published 2026-02-20
📖 4 min read☕ Coffee break read

Imagine you are trying to teach a new, super-smart robot assistant how to be a multilingual, multi-sensory tour guide for a global science conference.

Right now, most robot assistants are like monolingual librarians. They are great at reading books (text) in English, but if you hand them a video with a German speaker, or ask them to summarize a 2-hour lecture in Chinese, they often get confused, speak the wrong language, or just make things up.

This paper introduces MCIF (Multimodal Crosslingual Instruction-Following), which is essentially a giant, rigorous "driver's license" test for these new AI assistants.

Here is the breakdown of what they did, using some everyday analogies:

1. The Test Track: "The Science Talk Gauntlet"

Instead of testing the robots with simple questions like "What is the capital of France?", the researchers built a test track using real scientific talks from a major computer science conference (ACL).

  • The Content: They took 21 real presentations (about 2 hours of video) and added 79 more summaries.
  • The Challenge: They didn't just use English. They translated everything into German, Italian, and Chinese.
  • The Formats: They tested the robots on:
    • Text: Just the written transcript.
    • Audio: Just the sound of the speaker.
    • Video: The full video with slides and the speaker.
    • Mix: Audio + Video together.

2. The Four "Obstacle Courses" (Tasks)

The robots had to navigate four different types of challenges, each getting harder:

  • 🗣️ Recognition (The "Hearing Test"): The robot listens to a speaker and must write down exactly what they said.
    • Analogy: Like a court stenographer who must type out a fast-talking lawyer perfectly, even if there is background noise.
  • 🌍 Translation (The "Universal Translator"): The robot hears a speech in English and must write it out in Chinese or German.
    • Analogy: Like a simultaneous interpreter at the UN, but they have to do it instantly and perfectly.
  • ❓ Question Answering (The "Pop Quiz"): The robot watches the video and answers specific questions.
    • Analogy: Like a teacher asking, "What was the main point of the third slide?" or "What is the speaker's name?" Some questions could only be answered by listening, some by looking at the slides, and some by both.
  • 📝 Summarization (The "Executive Brief"): The robot must watch a 20-minute lecture and write a 200-word summary.
    • Analogy: Like a busy CEO who needs the "TL;DR" (Too Long; Didn't Read) version of a meeting before their next appointment.

3. The "Trick Questions" (Prompt Variations)

The researchers didn't just ask the same question twice. They created two versions of the test:

  • MCIF-Fix: The robot gets the exact same instruction every time (e.g., "Translate this").
  • MCIF-Mix: The robot gets the same task but phrased differently (e.g., "Can you put this into German?", "Give me the German version," "Translate the audio").
  • Why? To see if the robot gets confused if you change your wording, just like a human might get confused if you ask a question in a weird way.

4. The Results: "The Reality Check"

The researchers tested 23 different AI models (from big companies like Google, Microsoft, and open-source communities). Here is what they found:

  • The "Short Memory" Problem: Most robots are great at short clips (like a 30-second TikTok) but fall apart when the video gets long (like a 20-minute lecture). They forget the beginning by the time they reach the end.
  • The "Multitasking" Struggle: When you give a robot both audio and video at the same time, it often gets overwhelmed. It's like trying to listen to a podcast while watching a movie; the robot tends to ignore one of them or mix them up.
  • The "Language Switch" Glitch: Even when told to speak Italian, many robots accidentally slipped back into English, especially when the task was hard.
  • The "Hallucination" Issue: When the robots didn't know the answer, instead of saying "I don't know," they often made up facts or described the wrong slide.

The Big Takeaway

Think of MCIF as a stress test for the next generation of AI.

Currently, these AI assistants are like smartphones that are great at texting but terrible at video calls. They can read a book, but they struggle to understand a whole movie, translate a foreign language speech, and summarize it all at once without getting confused.

This paper says: "We have built the ultimate test track. Now, AI developers, go fix your robots so they can actually handle the real, messy, multi-language, long-form world we live in."

The good news? They released the test data for free, so anyone can try to build a better robot to pass the test.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →