Locally Adaptive Decay Surfaces for High-Speed Face and Landmark Detection with Event Cameras

This paper introduces Locally Adaptive Decay Surfaces (LADS), a novel event representation that dynamically modulates temporal decay based on local signal dynamics to overcome the limitations of fixed-parameter methods, thereby achieving state-of-the-art face detection and landmark localization accuracy at high frequencies while enabling the use of lighter network architectures.

Paul Kielty, Timothy Hanley, Peter Corcoran

Published 2026-02-27
📖 5 min read🧠 Deep dive

Imagine you are trying to take a photo of a busy street using a camera that doesn't take pictures like normal cameras do. Instead of capturing a full image every second, this special camera (called an Event Camera) only snaps a tiny "pixel" whenever something changes—like a car driving by, a bird flapping its wings, or a person blinking. It's incredibly fast and efficient, but the data it sends back is a chaotic, scattered stream of dots rather than a clear picture.

The challenge for computer scientists is: How do you turn this scattered stream of dots into a clear picture that a computer can understand?

The Old Way: The "Static Timer"

Traditionally, researchers tried to build a picture by grouping these dots into time buckets (like 30 times a second). They used a simple rule: "Keep the dots for a fixed amount of time, then forget them."

Think of this like a bucket with a hole in the bottom.

  • If you pour water (events) into the bucket, it fills up.
  • The hole lets water leak out at a constant speed, no matter what.

The Problem:

  • When things are still: If a person is sitting perfectly still, the "water" leaks out too fast. The computer forgets the shape of their face before it can even recognize it.
  • When things move fast: If a person waves their hand wildly, the bucket fills up so fast that the water overflows and mixes everything together. The hand becomes a blurry mess, and the computer can't tell where the fingers are.

The old method used the same "leak speed" for the whole image. It was a one-size-fits-all approach that failed when the scene was complex.

The New Solution: LADS (The "Smart Sponge")

The authors of this paper introduced a new method called LADS (Locally Adaptive Decay Surfaces).

Imagine replacing that leaky bucket with a smart, magical sponge that changes its behavior based on what it's touching.

  • In quiet areas (like a nose or cheek): The sponge becomes sticky. It holds onto the "dots" (events) for a long time. This preserves the shape of the face so the computer doesn't lose track of it.
  • In busy areas (like a blinking eye or moving hand): The sponge becomes super absorbent and then instantly releases. It grabs the new movement, shows it clearly, and then immediately lets the old movement fade away. This prevents the "blur" that happens when things move too fast.

How Does the Sponge Know What to Do?

The paper tests three different ways for the sponge to decide when to be sticky and when to be fast:

  1. Event Rate: "Are lots of dots hitting this spot?" If yes, fade them out fast. If no, hold them tight.
  2. Edge Detection (LoG): "Are there sharp lines here?" If the dots form a sharp edge (like an eyelid), fade them quickly to keep the edge crisp.
  3. Frequency (FFT): "Is this area vibrating with high-energy activity?" If yes, clear it out fast.

Why This Matters (The Results)

The researchers tested this on a task that is very hard for computers: finding a face and pinpointing exactly where the eyes, nose, and mouth are (facial landmarks) using these event cameras.

  • At normal speed (30 Hz): The new "Smart Sponge" method was better than the old "Leaky Bucket" method. It found faces more accurately and located facial features with less error.
  • At super speed (240 Hz): This is where LADS truly shines. The old methods fell apart when the camera moved fast; the faces became unrecognizable blobs. But LADS kept the faces clear and sharp. It was so good that it outperformed previous records set at much slower speeds.

The Bonus: Lighter Computers

Because the "Smart Sponge" does such a good job of keeping the image clear before the computer even looks at it, the computer doesn't need to be as "smart" or powerful to do the work.

  • Old way: Needed a giant, heavy brain (a massive neural network) to try to fix the blurry images.
  • LADS way: The image is already clear, so the computer can use a tiny, lightweight brain. This means this technology can run on small, battery-powered devices like robots, drones, or cars without needing a supercomputer.

The Bottom Line

This paper is about teaching computers to be context-aware. Instead of treating every part of a scene the same way, LADS looks at what's happening in each tiny corner of the image and adjusts its memory accordingly. It holds onto stillness and lets go of chaos.

This makes event cameras much more useful for real-world applications like driver monitoring (watching if a driver is falling asleep), robotics, and human-computer interaction, allowing them to see fast movements clearly without getting confused or needing expensive hardware.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →