IntrinsicWeather: Controllable Weather Editing in Intrinsic Space

IntrinsicWeather is a diffusion-based framework that enables controllable weather editing by decomposing images into intrinsic maps for enhanced spatial control and leveraging CLIP-space interpolation for fine-grained weather generation, outperforming existing methods and benefiting downstream tasks like autonomous driving.

Yixin Zhu, Zuo-Liang Zhu, Jian Yang, Miloš Hašan, Jin Xie, Beibei Wang

Published 2026-03-12
📖 4 min read☕ Coffee break read

Imagine you have a photograph of a street on a sunny day. Now, imagine you want to turn that photo into a rainy, snowy, or foggy scene.

Most current AI tools try to do this by painting over the pixels, like a digital artist smudging colors on a canvas. The problem? They often mess up the underlying scene. They might make a car look like a puddle, change the shape of a building, or make the shadows look weird because they are just guessing how the weather affects the image, not the world inside the image.

IntrinsicWeather is a new AI framework that takes a smarter, more "physics-based" approach. Instead of painting over the photo, it acts like a digital stage manager who breaks the scene down into its fundamental parts, swaps out the weather, and rebuilds the scene perfectly.

Here is how it works, using a simple analogy:

1. The "Deconstruction" Phase (The Inverse Renderer)

Imagine you have a complex Lego castle. To change the weather, you don't just spray paint the whole thing. Instead, you take it apart into three specific piles:

  • The Bricks (Materials): The color and texture of the road, the car, and the trees. These don't change whether it's raining or sunny.
  • The Shape (Geometry): The 3D structure of the scene. The road is still flat; the car is still a boxy shape.
  • The Lighting (Irradiance): This is the "weather" part. It includes the sun, the rain, the fog, and the shadows.

IntrinsicWeather uses a special AI "decoder" to look at your photo and separate these three piles. It strips away the rain or snow to reveal the clean, dry "bricks and shape" underneath. This is called Inverse Rendering.

2. The "Reconstruction" Phase (The Forward Renderer)

Now that you have the clean piles of bricks and shape, you want to build a new scene.

  • You keep the Bricks and Shape exactly the same (so the car doesn't turn into a tree).
  • You grab a new Lighting set based on a text prompt you give the AI, like "Make it a heavy snowstorm."
  • The AI "re-builds" the image, applying the new snow and shadows onto the original, unchanged structure.

This is called Forward Rendering. Because the AI knows the difference between the object and the weather, the result looks incredibly realistic. The snow piles up naturally on the car, and the wet road reflects the streetlights correctly.

The Secret Sauce: "The Spotlight" (IMAA)

One of the biggest challenges is that outdoor scenes are huge. A car far away is tiny, while a building nearby is huge. Standard AI often gets confused and misses the small details.

The authors added a clever trick called Intrinsic Map-Aware Attention. Think of this as giving the AI a flashlight.

  • When the AI needs to figure out the shape of a road, the flashlight shines on the road.
  • When it needs to figure out the metal on a car, the flashlight shines on the car.
  • This ensures the AI pays attention to the right details in the right places, preventing it from hallucinating weird shapes or missing small objects like distant pedestrians.

Why Does This Matter?

You might ask, "Why not just use a filter?"

  • For Self-Driving Cars: Self-driving cars need to "see" clearly. If it's raining, their cameras get confused. IntrinsicWeather can take a rainy photo, strip away the rain to see the road clearly, and then test if the car's safety software works. It helps train robots to drive in bad weather without needing to wait for a real storm.
  • For Creativity: It allows photographers and filmmakers to change the weather in a scene without ruining the physics of the image. You can turn a sunny day into a blizzard, and the snow will land exactly where it should.

The "Training Gym"

To teach this AI how to do all this, the researchers built two massive new "training gyms" (datasets):

  1. WeatherSynthetic: 38,000 computer-generated scenes with perfect weather data.
  2. WeatherReal: 18,000 real-world photos where they used their own AI to guess the weather-free version.

In a Nutshell

IntrinsicWeather is like a time-traveling photographer who can strip a scene down to its bare bones, swap the weather like changing a costume, and put it back together so perfectly that the physics, lighting, and geometry remain flawless. It moves beyond simple "photo editing" into "scene reconstruction," making it a powerful tool for both safety technology and creative art.