Talk to Your Slides: High-Efficiency Slide Editing via Language-Driven Structured Data Manipulation

This paper introduces "Talk to Your Slides," a high-efficiency slide editing agent that leverages language-driven structured data manipulation instead of visual perception to achieve faster, more accurate, and cost-effective text-centric and formatting modifications compared to Multimodal LLM-based GUI agents, supported by the newly proposed TSBench benchmark.

Kyudan Jung, Hojun Cho, Jooyeol Yun, Soyoung Yang, Jaehyeok Jang, Jaegul Choo

Published 2026-03-04
📖 5 min read🧠 Deep dive

Here is an explanation of the paper "Talk to Your Slides" using simple language and creative analogies.

The Big Problem: The "Pixel" vs. The "Blueprint"

Imagine you have a massive PowerPoint presentation with 50 slides, and your boss asks you to translate all the text from Korean to English, fix the spelling, and change the font color to blue.

  • The Old Way (GUI Agents): Imagine hiring a robot that can only see the screen like a human does. It has to look at every slide, take a picture, read the text using OCR (like a scanner), figure out where the text is, click the mouse, type the new words, and check if it looks right.

    • The Analogy: This is like trying to fix a car engine by looking at a photo of the car and guessing where the bolts are. It's slow, prone to mistakes (the robot might miss a word), and very expensive because the robot has to "think" hard about every single pixel.
  • The New Way (Talk-to-Your-Slides): Instead of looking at the picture, this new agent opens the blueprint of the presentation. It speaks directly to the software's internal code.

    • The Analogy: This is like having a master mechanic who opens the hood and talks directly to the engine's computer. They don't need to guess where the bolts are; they just send a command: "Change bolt #42 to blue." It's instant, precise, and cheap.

What is "Talk-to-Your-Slides"?

The researchers built an AI agent that acts as a translator between human language and computer code. Instead of trying to "see" the slides, it "reads" the structured data (the XML and object code) that makes up the PowerPoint file.

Here is how it works, broken down into four simple steps:

  1. The Planner (Instruction Understanding):
    • Analogy: You tell a smart assistant, "Fix the slides." The assistant breaks this down into a checklist: "Slide 1: Translate text. Slide 2: Change font color." It creates a clear map of what needs to happen.
  2. The Reader (Document Understanding):
    • Analogy: The assistant opens the "backstage" of the PowerPoint file. It doesn't look at the pretty pictures; it reads the raw data. It knows exactly which text box is where, what font is used, and what color it is. It turns this messy data into a neat, organized list (JSON).
  3. The Editor (Document Editing):
    • Analogy: The assistant takes your checklist and the raw data list. It says, "Okay, for Slide 1, I need to change this specific text string to English and make it bold." It updates the list with the new information.
  4. The Builder (Code Generator):
    • Analogy: Finally, the assistant writes a short computer script (Python code) that tells PowerPoint exactly how to apply those changes. It runs the script, and poof—the slides are fixed instantly.

Why is this a Big Deal?

The paper proves that for tasks involving text and formatting, looking at the screen is actually the wrong way to do it.

  • Speed: It's 34% faster. Because it skips the "looking and guessing" part, it flies through 50 slides in minutes instead of hours.
  • Accuracy: It's 34% more accurate. Since it reads the actual text code, it never misses a word or misreads a letter (a common problem for robots that just look at images).
  • Cost: It's 87% cheaper. Processing images requires powerful, expensive AI models. Reading text code is much lighter and cheaper.

The "TSBench" Benchmark

The researchers realized there was no good test to see how well AI could edit slides (most tests only checked if AI could make slides from scratch). So, they built TSBench.

  • The Analogy: Imagine a driving test. Before, we only tested if a car could drive in a straight line. TSBench is a complex obstacle course with 379 different challenges, like "Change the color of the third tire on the left" or "Translate the dashboard text." They even added a "Hard Mode" with tricky questions to see if the AI gets confused or gives up gracefully.

The Catch (Limitations)

The paper is honest about what it can't do yet.

  • The Analogy: The agent is a master mechanic, but it's blind to the "vibe." If you say, "Make this slide look more balanced," the agent might struggle because "balanced" is a feeling, not a specific code instruction. It's great at moving text boxes and changing fonts, but it might need a human (or a visual AI) to check if the final design looks "pretty."

The Bottom Line

Talk-to-Your-Slides is a game-changer for office work. It stops treating PowerPoint like a picture you have to stare at and starts treating it like a database you can talk to. It turns a tedious, day-long chore into a few minutes of automated magic, saving time, money, and a lot of headaches.

In short: Stop looking at the screen; start reading the code.