eStonefish-Scenes: A Sim-to-Real Validated and Robot-Centric Event-based Optical Flow Dataset for Underwater Vehicles

This paper introduces eStonefish-Scenes, a synthetic event-based optical flow dataset for underwater vehicles generated via the Stonefish simulator, along with the eWiz processing library, and validates its sim-to-real transferability by demonstrating that a model trained exclusively on this synthetic data achieves high-accuracy optical flow estimation on real-world underwater sequences without fine-tuning.

Jad Mansour, Sebastian Realpe, Hayat Rajani, Michele Grimaldi, Rafael Garcia, Nuno Gracias

Published 2026-02-23
📖 5 min read🧠 Deep dive

The Big Problem: Teaching Robots to "See" Underwater

Imagine you are trying to teach a robot submarine how to swim through a coral reef without crashing into anything. To do this, the robot needs to understand how things are moving around it (a concept called optical flow).

Usually, robots use standard cameras that take pictures like a human does—snapshots of the world. But underwater is tricky: the water is murky, the light changes fast, and things move quickly. Standard cameras get blurry and confused in these conditions.

Enter Event Cameras. Think of these not as cameras, but as super-fast, hyper-sensitive eyes. Instead of taking full pictures, they only "blink" when something changes (like a fish swimming by or a shadow passing). They are incredibly fast, don't get blurry, and work great in the dark.

The Catch: To teach a robot how to use these "blink-eyes," you need a massive library of practice data. But getting real underwater data is a nightmare. It's expensive, dangerous, and you can't easily know exactly where the robot is or how fast it's moving to check if it's doing the right thing. It's like trying to learn to drive a car in a blizzard without ever seeing the road or having a GPS.

The Solution: A "Video Game" for Robots

The authors of this paper decided to build a perfect video game to solve this problem. They created a synthetic dataset called eStonefish-Scenes.

Think of this dataset as a high-tech flight simulator for underwater robots. Instead of going to the ocean, they built a digital ocean inside a computer using a simulator called Stonefish.

Here is what makes their "game" special:

  1. The World: They didn't just make a flat floor. They built a vibrant, 3D coral reef with thousands of different corals and plants.
  2. The Wildlife: They added schools of fish that swim naturally. They used a "Boids" algorithm (think of it as a digital version of how birds flock) so the fish move together, avoid obstacles, and react to the robot, just like real fish.
  3. The Robot: They simulated a real robot submarine (BlueROV2) equipped with the special "blink-eye" camera.
  4. The Secret Sauce: Because this is a computer simulation, they know the perfect truth. They know exactly how fast the robot is moving and exactly how the water is flowing. This gives them the "answer key" that is impossible to get in the real ocean.

The Toolkit: "eWiz"

Building a dataset is hard, but using it should be easy. The authors also built a software toolbox called eWiz.

  • Analogy: If the dataset is a giant warehouse of raw ingredients, eWiz is the kitchen, the recipe book, and the chef's knife all in one. It helps researchers load the data, chop it up (augment it), cook it (train their AI), and taste-test it (evaluate the results). It handles all the messy technical details so scientists can focus on the cooking.

The Big Test: Does the Game Teach Real Skills?

The biggest question is: If a robot learns in a video game, can it actually drive a real car?

To answer this, the team did a "Sim-to-Real" test:

  1. Training: They taught a neural network (an AI brain) to predict motion using only the fake data from their video game. They never showed it a single real underwater photo.
  2. The Real World Test: They took a real robot submarine, put a real "blink-eye" camera on it, and drove it around a swimming pool in a lab.
  3. The Ground Truth: To know if the robot was right, they put a giant, high-resolution poster of a coral reef on the bottom of the pool. They used math to calculate exactly how the robot moved relative to that poster.
  4. The Twist (Uncertainty): Real life is messy. Sometimes the poster is hard to see, or the water is cloudy. The team developed a special way to measure confidence. Imagine the robot saying, "I'm 99% sure I'm moving left here, but I'm only 50% sure about that blurry spot over there." They used this "confidence score" to judge the robot's performance fairly.

The Results: A Home Run

The results were impressive. The AI, which had never seen a real underwater scene, was able to navigate the real pool with high accuracy.

  • It successfully predicted how the robot was moving.
  • It handled the "blurry" and uncertain parts of the real world gracefully.
  • It proved that you don't need to spend millions of dollars and risk expensive equipment to train underwater robots. You can train them in a digital coral reef first.

Summary

In short, this paper is about building a realistic underwater video game to teach robots how to swim.

  • The Problem: Real underwater data is too hard to get.
  • The Fix: A synthetic dataset (eStonefish-Scenes) with realistic fish and coral, plus a software toolkit (eWiz) to make it easy to use.
  • The Proof: They trained an AI on the fake data, tested it in a real pool, and it worked almost perfectly.

This is a huge step forward because it means we can build smarter, safer underwater robots without needing to dive into the deep ocean just to collect practice data.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →