FLUID: A Fine-Grained Lightweight Urban Signalized-Intersection Dataset of Dense Conflict Trajectories

This paper introduces FLUID, a fine-grained, lightweight dataset and processing framework derived from drone footage that captures dense traffic conflicts and rich behavioral data at urban signalized intersections to support traffic modeling, human preference mining, and autonomous driving research.

Yiyang Chen, Zhigang Wu, Guohong Zheng, Xuesong Wu, Liwen Xu, Haoyuan Tang, Zhaocheng He, Haipeng Zeng

Published 2026-02-24
📖 4 min read☕ Coffee break read

Imagine trying to understand how a bustling city intersection works. You could stand on a corner and watch cars go by, but you'd miss what's happening on the other side of the street. You could sit in a car and see what's right in front of you, but you'd be blind to the chaos happening behind you.

This paper introduces FLUID, a new "super-eye" for traffic researchers. It's a massive dataset created by flying drones over three busy intersections in China, capturing every car, bike, and pedestrian in high definition.

Here is the breakdown of what makes FLUID special, explained with some everyday analogies:

1. The Problem: The "Blind Spot" of Traffic Data

Before FLUID, traffic datasets were like trying to solve a puzzle with missing pieces.

  • Ground cameras are like security guards standing at one corner; they can't see the whole picture and sometimes scare drivers into driving differently.
  • Car sensors (like those in self-driving cars) are like looking through a narrow keyhole; you see what's right in front of you, but you miss the big picture.
  • Existing drone data was often too blurry, missed small details (like pedestrians), or didn't have enough "traffic jams" to study how people react when things get crowded.

2. The Solution: The "Drone Chef"

The researchers didn't just fly a drone and hope for the best. They built a lightweight, step-by-step kitchen to process the raw video into a perfect meal of data.

  • Stabilization: Drones wobble in the wind. The team used a digital "steady-cam" to smooth out the shaky footage, like stabilizing a shaky hand holding a camera.
  • The "All-Star" Detection Team: Instead of using one robot to find cars and bikes, they trained three different AI models (like three different detectives) and let them work together. One is great at spotting tiny mopeds, another at spotting big trucks, and the third at catching fast-moving cars. They combine their findings to ensure nothing is missed.
  • The "Traffic Cop" Filter: Sometimes the AI gets confused and thinks one car is two cars, or sees a shadow as a vehicle. The team built a smart filter (using math called "Time-to-Collision") that acts like a strict traffic cop, removing the fake cars and keeping only the real ones.

3. What's in the Box? (The Dataset)

FLUID isn't just a video file; it's a complete "traffic simulation kit" containing:

  • The Scenes: Three different types of intersections (a standard 4-way, a 4-way with a special right-turn lane, and a T-junction).
  • The Crowd: Over 20,000 traffic participants (cars, trucks, buses, tricycles, mopeds, and pedestrians).
  • The Drama: This is the best part. The dataset is full of conflicts. While other datasets might have 1 or 2 near-misses per minute, FLUID has 2.8 conflicts per minute. It's like capturing a movie scene where the action never stops. About 15% of all the cars in the video were involved in a near-miss or a rule-breaking moment.
  • The Rules: They recorded the traffic lights, the road maps, and even the specific "intentions" of drivers (e.g., "I'm turning left but didn't yield").

4. Why Does This Matter? (The "Why")

Think of FLUID as a gym for self-driving cars and traffic planners.

  • For Self-Driving Cars: To teach a robot to drive, you need to show it dangerous situations, not just empty roads. FLUID provides thousands of "near-accidents" so the AI can learn how to react when a pedestrian darts out or a truck cuts them off.
  • For City Planners: It helps them see exactly where people break the rules. Are people running red lights at the left turn? Are pedestrians jaywalking at the crosswalk? The data pinpoints these "hotspots" so cities can fix them.
  • For Researchers: It's transparent. Unlike some commercial tools that are "black boxes" (you pay, you get data, you don't know how it was made), FLUID gives you the raw video, the code, and the math. It's like giving a chef the recipe, not just the finished cake.

The Bottom Line

FLUID is a high-definition, drone-shot, traffic conflict library that fills the gaps in our understanding of how humans and machines interact at intersections. It's designed to be the "gold standard" for making our roads safer and our self-driving cars smarter, by showing them exactly what happens when traffic gets messy.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →