DEBISS: a Corpus of Individual, Semi-structured and Spoken Debates

This paper introduces DEBISS, a novel corpus of spoken and individual semi-structured debates designed to address the scarcity of debate datasets by providing comprehensive NLP annotations for tasks such as speech-to-text, speaker diarization, argument mining, and debater quality assessment.

Klaywert Danillo Ferreira de Souza, David Eduardo Pereira, Cláudio E. C. Campelo, Larissa Lucena Vasconcelos

Published 2026-03-06
📖 4 min read☕ Coffee break read

Imagine you want to teach a robot how to argue like a human. You can't just feed it a dictionary; you need to show it real conversations where people think on their feet, interrupt each other, stumble over words, and try to convince one another.

For a long time, scientists trying to build these "argument robots" have been stuck with a limited menu. They mostly had two options:

  1. The "Political Theater" Menu: Highly scripted, formal debates between politicians (like presidential debates). These are great, but they are too polished and rigid. Real people don't talk like that.
  2. The "Internet Comment Section" Menu: Written posts on Twitter or Reddit. These are messy and real, but they lack the sound of a human voice, the pauses, the "umms," and the energy of a face-to-face conversation.

Enter DEBISS: The "Real-Life Debate" Recipe Book.

This paper introduces a new, special collection of data called DEBISS (Spoken, Individual, and Semi-structured Debates). Think of it as a high-definition recording studio where 67 computer science students from a Brazilian university got together to debate a hot topic: "Generative Artificial Intelligence and its impact on society."

Here is what makes DEBISS special, explained through some everyday analogies:

1. The "Semi-Structured" Sandwich

Most debates are either a rigid script (like a play) or a chaotic free-for-all (like a mosh pit). DEBISS is the perfect sandwich.

  • The Bread (Structure): There are rules. A moderator asks specific questions, and students have to answer them. This keeps the conversation on track.
  • The Filling (Spontaneity): Between the questions, students can speak freely, interrupt (politely), and react to each other. This captures the messy, human "flow" of real conversation that computers usually miss.

2. The "Individual" Spotlight

In many group debates, people speak as a team, blending their voices into one "we." In DEBISS, every student is an individual soloist. Even though they are in a group of 3 to 5, each person defends their own unique viewpoint. This allows researchers to study exactly how one person constructs an argument, rather than how a group blends their ideas.

3. The "Gold Mine" of Annotations

Collecting the audio is just the first step. The real magic of DEBISS is what the researchers did after recording. They didn't just transcribe the words; they added layers of "metadata" (extra information) like a chef adding spices to a dish:

  • The "Who Said What" Map (Speaker Diarization): They labeled exactly which student said which sentence, even when voices overlapped.
  • The "Argument Map" (Argument Mining): They broke down the speech to find the "claims" (what they believe), the "evidence" (proof they used), and the "reasons" (why they believe it).
  • The "Stutter Detector" (Disfluency): They marked where people said "um," "uh," or repeated themselves. This is crucial because real humans aren't perfect speakers, and AI needs to learn to understand us despite our stutters.
  • The "Scorecard" (Evaluation): After the debate, students graded themselves and each other. This gives researchers a way to measure "good debating" based on human judgment, not just computer algorithms.

Why Does This Matter?

Think of DEBISS as a gym for Artificial Intelligence.

  • For Language Models: It helps AI learn to understand spoken Portuguese (a language often underrepresented in tech) and how to handle the messy reality of human speech.
  • For Education: It helps teachers understand what makes a student a good debater. By analyzing the data, we can see what strategies work best and help students improve their critical thinking.
  • For Society: As AI becomes more common, we need to understand how humans argue about AI. This dataset captures a real conversation about AI, made by people studying AI, about the future of AI.

The Bottom Line

The researchers built a library of 9 hours and 35 minutes of real, spoken, Brazilian Portuguese debates. They cleaned it up, labeled every word, and tagged every argument. They are now opening the doors to this library so other scientists can use it to build better, more human-like AI tools.

It's not just a recording; it's a training manual for the next generation of intelligent machines, teaching them how to listen, understand, and perhaps, one day, argue back.