M-QUEST -- Meme Question-Understanding Evaluation on Semantics and Toxicity

This paper introduces M-QUEST, a semantic framework and benchmark comprising 609 question-answer pairs across ten interpretive dimensions, designed to evaluate and advance the ability of large language models to perform commonsense reasoning and toxicity detection in internet memes.

Stefano De Giorgis, Ting-Chih Chen, Filip Ilievski

Published 2026-03-05
📖 5 min read🧠 Deep dive

Imagine the internet is a giant, chaotic digital town square. In this square, people don't just talk; they throw around memes. A meme is like a digital joke package: it's a picture with some text on it. Sometimes, these jokes are harmless fun, like a cat wearing a hat. But sometimes, they are "toxic"—they are mean, hateful, or designed to hurt specific groups of people.

The problem is that spotting a toxic meme is incredibly hard for computers. Why? Because a meme isn't just a picture and words; it's a cultural inside joke. To understand if a meme is mean, a computer needs to know:

  • Who is in the picture?
  • What is the historical context?
  • Is the text being sarcastic?
  • Who is the joke actually at?

This paper, titled M-QUEST, is like a new "driving test" for Artificial Intelligence (AI) to see if it can understand these complex, sometimes nasty, jokes.

The Problem: The AI is Like a Literal Robot

Imagine you show a robot a picture of a clown holding a sign that says "I love kids."

  • A human might look at the clown's scary makeup and realize the joke is actually about how clowns are creepy, or maybe it's a dark joke about a specific event.
  • An older AI might just say, "The text says 'I love kids,' so this is a happy, safe image." It misses the vibe, the tone, and the hidden meaning.

The authors of this paper realized that current AIs are great at reading the text and seeing the objects, but they are terrible at understanding the soul of the meme. They need a better way to teach the AI how to think about these images.

The Solution: The "Meme Anatomy" Framework

To fix this, the researchers built a 10-part checklist (a framework) to dissect a meme. Think of it like a doctor performing an autopsy on a joke to see what makes it tick. The checklist includes:

  1. The Basics: What does the text say? What do you see in the picture?
  2. The Scene: How are things arranged? (e.g., Is someone standing over someone else?)
  3. The Background Knowledge: Do you need to know about a specific celebrity, a political event, or a movie to get the joke?
  4. The Emotion: Is the person in the picture angry? Sad? Is the joke trying to make us laugh or feel uncomfortable?
  5. The Target: Who is the joke aimed at? Is it punching up (at the powerful) or punching down (at the vulnerable)?
  6. The "Projection": Who is the viewer supposed to be? Are we supposed to feel like the victim or the bully?
  7. The Analogy: Is the picture comparing two totally different things? (e.g., "This politician is like a sinking ship.")
  8. The Intent: Is the creator trying to spread hate, or just be silly?
  9. The Toxicity: Finally, is it actually harmful?

The Test: M-QUEST

The researchers didn't just build the checklist; they built a test called M-QUEST.

  • They took 307 memes (mostly from a known dataset of "hateful memes").
  • They used AI to generate 609 multiple-choice questions about these memes based on their 10-point checklist.
  • Then, humans stepped in to grade the AI's questions. They asked: "Is this a good question? Is the answer actually correct?"

It's like a teacher grading a student's homework before giving it to the class. Only the best, most accurate questions made it into the final test.

The Experiment: Putting 8 AIs Through the Ringer

The researchers took 8 different "smart" AI models (like Qwen, LLaVA, and others) and asked them to take the M-QUEST test. They wanted to see: Can these AIs explain why a meme is toxic, or do they just guess?

The Results:

  • The "Dumb" Robots: Some older models got less than 15% of the answers right. They were like students who didn't even read the question.
  • The "Smart" Robots: The best models (specifically the newer "Qwen" family) got over 86% right.
  • The Secret Sauce: The winners weren't just bigger; they were better trained to reason. They had been taught to follow instructions carefully and to think step-by-step (like a detective solving a mystery) rather than just guessing.

The Catch:
Even the smartest AI struggled with the hardest part: Pragmatic Inference.
This is the ability to understand what is implied but not said.

  • Example: If a meme shows a dog looking sad next to a sign saying "Free Hugs," a human knows the dog is being sarcastic or the situation is tragic. The AI often tries to force a connection where there isn't one, or it takes the text too literally and misses the irony.

The Big Takeaway

This paper tells us that understanding internet culture is hard for computers.

  1. Size isn't everything: A bigger AI isn't automatically smarter at understanding jokes. It needs specific training on how to think (reasoning) and how to listen (instruction tuning).
  2. Context is King: To spot a toxic meme, you can't just look at the pixels. You need to understand the history, the culture, and the hidden meanings.
  3. We need better tests: Before we can trust AI to moderate social media (to stop hate speech), we need to make sure they can pass tests like M-QUEST. Currently, they are still learning the ropes.

In a nutshell: The authors built a sophisticated "Meme School" to teach AI how to understand the difference between a funny joke and a hateful attack. They found that while the smartest students are doing well, they still need more practice understanding the subtle, unspoken rules of human culture.