CoLC: Communication-Efficient Collaborative Perception with LiDAR Completion

The paper proposes CoLC, a communication-efficient collaborative perception framework that leverages LiDAR completion techniques—specifically Foreground-Aware Point Sampling, Completion-Enhanced Early Fusion, and Dense-Guided Dual Alignment—to restore scene completeness from sparse transmissions and achieve superior perception-communication trade-offs while remaining robust to model heterogeneity.

Yushan Han, Hui Zhang, Qiming Xia, Yi Jin, Yidong Li

Published 2026-03-03
📖 4 min read☕ Coffee break read

Imagine you are driving a self-driving car in a heavy fog. You can only see a few meters ahead. Suddenly, a car pulls out from a side street you can't see, or a pedestrian steps out from behind a large truck. Your car's sensors are blind to these dangers.

Collaborative Perception is like giving your car a "superpower": it lets nearby cars share what they see with you. If the car next to you sees the pedestrian, it tells you, and you can stop in time.

However, there's a catch. Sharing raw sensor data (like a high-definition 3D map of everything around you) is like trying to stream a 4K movie over a dial-up internet connection. It takes too much bandwidth, causes lag, and is too expensive for real-world use.

Most current solutions try to solve this by sending only "summarized" data (like "there is a car here"). But this is like sending a text message saying "Car" instead of showing a photo. You lose the details, and if the cars are using different software, they might not understand each other's summaries.

CoLC (Communication-Efficient Collaborative Perception with LiDAR Completion) is a new, smarter way to do this. Think of it as a "Smart Sketch and Fill-In" system.

Here is how it works, broken down into three simple steps:

1. The Smart Sketch (Foreground-Aware Point Sampling)

Instead of sending a massive, heavy file of every single point in the scene, the neighboring cars act like a sketch artist.

  • The Problem: If they only send the "important" things (like the car or pedestrian), the picture looks weird and floating because the background (the road, the trees) is missing.
  • The CoLC Solution: They send a selective sketch. They send all the important objects (the "Foreground") but also a few key "anchor points" from the background (the "Background") to show context.
  • The Analogy: Imagine you are describing a room to a friend over the phone. Instead of listing every single dust mote, you say, "There's a red sofa here, a coffee table there, and a window on the left." You keep the message short but give enough context so the friend can picture the room.

2. The Magic Fill-In (LiDAR Completion)

Now, your car receives this "sparse sketch." It looks a bit empty and patchy. This is where the Magic Fill-In happens.

  • The Problem: Your car needs a full, dense 3D map to drive safely, not a sketch with holes in it.
  • The CoLC Solution: Your car has a built-in AI "Imagination Engine." It looks at the sparse sketch it received and uses its training to "fill in the blanks." It predicts where the missing points should be, turning the sketch back into a solid, dense 3D model.
  • The Analogy: It's like looking at a connect-the-dots puzzle where only 20% of the dots are there. Your brain (the AI) instantly fills in the lines to see the whole picture of a dog, even though you only saw a few dots.

3. The Double-Check (Dense-Guided Dual Alignment)

Sometimes, when the AI "imagines" the missing parts, it might get the shape slightly wrong or the colors (semantics) a bit off.

  • The Solution: During the training phase, the system uses a Double-Check method. It compares its "filled-in" picture against a perfect, real-life picture to make sure the shapes are straight and the objects are identified correctly.
  • The Analogy: It's like an art teacher correcting a student's drawing. "You drew the car's wheel here, but it should be slightly lower to match the ground." This ensures the final result is not just a guess, but a reliable reconstruction.

Why is this a big deal?

  • It's Efficient: It sends much less data (like a sketch) but gives you the full experience (the 3D map).
  • It's Robust: Because it sends raw "dots" rather than complex summaries, it works even if your car uses different software than the neighbor's car. It's like speaking a universal language of "dots" instead of a specific dialect.
  • It's Safe: It recovers the missing details that other methods lose, ensuring the car doesn't miss a pedestrian just because the data was compressed too much.

In short: CoLC allows self-driving cars to share a "rough draft" of their view, which their own AI then instantly turns into a "finished masterpiece," ensuring everyone stays safe without clogging up the communication network.