S2R-HDR: A Large-Scale Rendered Dataset for HDR Fusion

This paper introduces S2R-HDR, a large-scale synthetic dataset of 24,000 high-quality HDR images generated via Unreal Engine 5, along with a domain adaptation method called S2R-Adapter, to overcome data scarcity and improve the generalization of learning-based HDR fusion models.

Yujin Wang, Jiarui Wu, Yichen Bian, Fan Zhang, Tianfan Xue

Published 2026-02-17
📖 5 min read🧠 Deep dive

Imagine you are trying to teach a robot how to take the perfect photo of a busy city street at sunset. The street is full of moving cars, people walking, and the sun is blindingly bright in the sky while the shadows are pitch black.

To teach the robot, you need to show it thousands of examples of what a "perfect" photo looks like in these tricky situations. But here's the problem: taking real photos like this is a nightmare. You'd need special cameras, perfect weather, and you'd have to freeze time to get the "perfect" shot. It's too expensive, too slow, and often impossible to do perfectly.

This is where the paper "S2R-HDR" comes in. It's like a master chef who decided, "If we can't get enough real ingredients, let's build the best possible fake kitchen."

Here is the story of their solution, broken down into three simple parts:

1. The "Magic Kitchen": S2R-HDR (The Dataset)

Instead of going out to take 24,000 real photos, the researchers built a massive, hyper-realistic video game world using a powerful engine called Unreal Engine 5.

  • The Analogy: Think of this like a video game level designer. They didn't just build one room; they built 24,000 different scenes. They put in moving cars, running dogs, people walking, and even direct sunlight hitting a shiny car.
  • Why it's special: In the real world, if you want to see what a photo looks like with more light or less light, you have to take a new picture. In their "Magic Kitchen," they can instantly generate every possible version of a photo (bright, dark, super-bright) because they control the "sun" and the "camera" with a computer.
  • The Result: They created a library of 24,000 perfect training examples. This is huge! Previous libraries only had about 100 to 150 examples. It's like going from teaching a student with one textbook to giving them a whole library.

2. The "Translator": S2R-Adapter (The Domain Adaptation)

There is a catch. Even though their video game world looks amazing, it's still a simulation. A robot trained only on video games might look at a real tree and think, "That texture looks too smooth; it's fake." This is called the "Domain Gap."

  • The Analogy: Imagine you learned to drive in a driving simulator. You know the rules, but when you get in a real car, the steering feels different, and the tires make real noise. You might crash because the simulator didn't feel real enough.
  • The Solution: The researchers built a little "translator" tool called S2R-Adapter.
    • The "Share Branch": This part remembers everything the robot learned in the video game (like "cars move fast," "sun is bright"). It makes sure the robot doesn't forget its training.
    • The "Transfer Branch": This part is the translator. It teaches the robot, "Okay, real trees look a bit rougher, and real wind shakes the leaves differently."
  • The Magic: This tool allows the robot to take its "video game brain" and instantly upgrade it to handle "real world" photos without needing thousands of new real-world photos to relearn everything.

3. The "Self-Correcting GPS": Test-Time Adaptation

Sometimes, you don't even have the "answer key" (the perfect photo) to check if the robot is right. You just have the messy real-world photo.

  • The Analogy: Imagine you are driving in a foggy city you've never seen. You don't have a map. But, your car has a smart system that says, "Hey, that looks like a tree, but the fog is making it look weird. Let me adjust my view slightly to see if it's really a tree."
  • How it works: The researchers taught their system to look at a photo, guess what it sees, and then ask itself, "Am I confused?" If the system is confused (high uncertainty), it automatically tweaks its "Translator" (the Adapter) to fit that specific scene better, right on the spot.

The Grand Finale: Why Does This Matter?

The researchers tested their robot on real-world photos.

  • Without their help: Other robots (trained on small, old datasets) made mistakes. They left "ghosts" behind moving cars or blew out the bright sun, turning it into a white blob.
  • With S2R-HDR: Their robot produced crystal-clear photos. It handled moving cars without ghosts and recovered details in the blinding sun that other robots missed.

In a nutshell:
They realized that taking real photos for training is too hard, so they built a giant, perfect video game world to teach the AI. Then, they built a smart translator to help the AI understand that the real world is a little different from the game. The result? A camera system that can take amazing photos in any crazy, moving, bright, or dark situation, even if it was trained mostly on a computer.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →