SPOT: An Annotated French Corpus and Benchmark for Detecting Critical Interventions in Online Conversations

This paper introduces SPOT, the first annotated French corpus and benchmark for detecting "stopping points"—subtle critical interventions that pause or redirect online discussions—and demonstrates that fine-tuned encoder models outperform prompted LLMs in this task, particularly when enriched with contextual metadata.

Manon Berriche, Célia Nouri, Chloée Clavel, Jean-Philippe Cointet

Published Tue, 10 Ma
📖 5 min read🧠 Deep dive

Imagine you are walking through a giant, noisy digital town square (like Facebook). Most people are shouting, arguing, or sharing wild stories. Researchers usually study the loudest shouts: the hate speech, the blatant lies, or the organized mobs.

But there's another kind of interaction that happens all the time, which is much quieter and harder to spot. It's the "Stop Sign."

This paper introduces a new project called SPOT (Stopping Points in Online Threads) to find and study these "Stop Signs."

What is a "Stopping Point"?

Think of a conversation as a train moving down a track. A "stopping point" isn't necessarily a derailment (a huge fight) or a repair crew (a fact-checker with a clipboard).

A stopping point is a passenger who suddenly yells, "Wait a minute!" or "Is this really true?" or "This is ridiculous!"

  • It's not always a full argument. Sometimes it's just a skeptical look, a sarcastic joke, or a short "Report this."
  • It doesn't have to be right. The person might be wrong, but they are still pausing the train to make everyone think, "Hmm, maybe we shouldn't just keep rolling with this."
  • It's subtle. Unlike a fact-checker who says, "Here are the statistics proving you wrong," a stopping point might just say, "Yeah, and I'm the Queen of England," using irony to show the original claim is absurd.

The Problem: Computers Are Bad at "Wait a Minute"

For a long time, computer programs (AI) have been great at spotting obvious hate speech or clear lies. But they are terrible at spotting these subtle "Stop Signs."

Why? Because computers usually read one sentence at a time, like reading a single line of a book without looking at the rest of the page.

  • The Computer's Mistake: If someone says, "This is completely absurd!" the computer thinks, "Great! They are angry and disagreeing! That's a Stop Sign!"
  • The Reality: In the context of the conversation, that person might actually be agreeing with the post but just using strong words. They aren't stopping the train; they are just cheering it on.

To understand a "Stop Sign," you need to see the whole picture: the original post, the article being shared, the group it's posted in, and what the person replied to.

The Solution: The SPOT Corpus

The researchers built a massive library called SPOT.

  • The Data: They collected over 43,000 comments from French Facebook pages where users had flagged links as "fake news."
  • The Human Touch: Real humans (sociologists and linguists) read these comments and manually labeled them: "Is this a Stop Sign?" or "Is this just noise?"
  • The Context: They didn't just save the comment; they saved the whole "scene" around it (the article, the page name, the previous comments).

The Big Experiment: Humans vs. Robots

The team tested two types of AI to see who could find these Stop Signs better:

  1. The "Instruction-Tuned" Robots (LLMs): These are the fancy, big-brained AI models (like the ones you chat with) that you can just ask, "Find the Stop Signs!" without teaching them specifically.

    • The Result: They struggled. Even when given detailed instructions, they got the answer right only about 55% of the time. They were like a tourist in a foreign country trying to understand local slang; they missed the nuance.
  2. The "Specialized" Robots (Fine-tuned Encoders): These are smaller, specialized models that were "trained" specifically on this dataset, learning from the human examples.

    • The Result: They crushed it, getting it right about 78% of the time.
    • The Lesson: For complex social tasks, especially in languages other than English, you can't just ask a general AI to "figure it out." You have to train a specialist who knows the specific culture and context.

The Secret Sauce: Context

The most important discovery was that context is king.
When the specialized robots were given just the comment, they were okay. But when they were given the whole story (the article title, the page name, the previous comment), their performance jumped up significantly.

It's like trying to guess if a person is joking. If you just hear "You're a genius," you might think they are being sarcastic. But if you know they are talking to their best friend who just won a prize, you know they are being sincere. The AI needed that extra context to get it right.

Why Does This Matter?

This paper teaches us that online moderation and understanding isn't just about deleting bad words. It's about understanding the flow of conversation.

  • For Researchers: It gives us a new tool to study how ordinary people self-regulate their communities without needing a police officer.
  • For AI Developers: It proves that for social media, "bigger isn't always better." A smaller, specialized model that understands the context of a conversation is far more effective than a giant, general-purpose model.

In short: The internet is full of people dropping subtle "Stop Signs" to question what they see. The SPOT project is the first map to find them, proving that to understand human conversation, you need to look at the whole scene, not just the words.