FengWu: Pushing the Skillful Global Medium-range Weather Forecast beyond 10 Days Lead

FengWu is an advanced AI-driven global weather forecasting system that utilizes a novel multi-modal, multi-task deep learning architecture with a replay buffer to achieve state-of-the-art medium-range predictions, extending skillful forecast lead times beyond 10 days while outperforming GraphCast in accuracy and efficiency.

Kang Chen, Tao Han, Junchao Gong, Lei Bai, Fenghua Ling, Jing-Jia Luo, Xi Chen, Leiming Ma, Tianning Zhang, Rui Su, Yuanzheng Ci, Bin Li, Xiaokang Yang, Wanli Ouyang

Published 2026-03-03
📖 5 min read🧠 Deep dive

Imagine you are trying to predict the weather for the next two weeks. For decades, scientists have used massive, super-complex physics simulations to do this. It's like trying to predict the path of a single drop of water in a raging river by calculating the physics of every single molecule around it. It's accurate, but it's incredibly slow, expensive, and requires supercomputers the size of a building.

Enter FengWu. Think of FengWu not as a physics calculator, but as a super-observant meteorologist who has memorized 39 years of weather history. Instead of calculating physics from scratch every time, FengWu looks at the current weather patterns and says, "I've seen this movie before; I know how the story ends."

Here is a simple breakdown of how FengWu works and why it's a big deal, using some everyday analogies:

1. The "Multi-Modal" Approach: The Orchestra vs. The Soloist

Most previous AI weather models treated the atmosphere like a single, giant soup. They threw all the data (temperature, wind, humidity, pressure) into one blender and tried to guess the result.

FengWu is smarter. It treats the atmosphere like a symphony orchestra.

  • The Old Way: Listening to the whole orchestra at once and trying to guess the next note.
  • The FengWu Way: It has a dedicated "conductor" for each section. One AI "listens" only to the strings (temperature), another only to the brass (wind), and another only to the percussion (humidity).
  • The Magic: After each section practices its part, they all meet in the middle (a "Cross-modal Transformer") to jam together. This allows FengWu to understand how a change in wind speed specifically affects humidity, rather than just guessing based on a blurry mix of everything.

2. The "Multi-Task" Learning: The Fair Coach

In the past, AI models tried to predict everything with the same level of effort. It's like a coach telling a sprinter and a marathon runner to train with the exact same intensity. The sprinter gets bored, and the marathon runner gets exhausted.

FengWu realizes that predicting temperature is easy, but predicting a sudden storm is hard. It uses a "Fair Coach" (Uncertainty Loss).

  • If the model is confident about the temperature, it relaxes a bit.
  • If the model is struggling with a complex storm pattern, the coach says, "Focus harder here!"
  • This automatically adjusts the difficulty for each part of the weather, ensuring the model learns the hard stuff without getting distracted by the easy stuff.

3. The "Replay Buffer": Learning from Mistakes

This is FengWu's secret weapon for long-term predictions.

  • The Problem: If you ask an AI to predict 10 days ahead, it has to predict Day 1, then use Day 1 to predict Day 2, then Day 2 for Day 3, and so on. This is called "chaining." If it makes a tiny mistake on Day 1, that mistake gets bigger on Day 2, and by Day 10, the prediction is total nonsense. It's like playing the game of "Telephone" with 50 people; the message gets garbled.
  • The FengWu Solution: FengWu uses a Replay Buffer. Imagine a student practicing for a test. Instead of just taking the test once, they take a practice test, write down their wrong answers, and then re-take the test using their own wrong answers as the starting point.
  • By forcing the AI to predict the future based on its own previous (slightly wrong) predictions during training, it learns how to correct its own mistakes. It's like a pilot practicing in a simulator where the wind keeps changing based on their previous errors, so they learn to fly better in the real world.

4. The Results: Breaking the 10-Day Barrier

Why does this matter?

  • The Old Limit: Historically, weather forecasts become useless after about 7 to 10 days.
  • FengWu's Breakthrough: FengWu has pushed the "useful" forecast window to 10.75 days.
  • The Comparison: It beats the current state-of-the-art AI model (GraphCast) on 80% of all weather variables.
  • The Speed: While traditional supercomputers might take hours to run a 10-day forecast, FengWu can do it in less than 30 seconds on a standard high-end graphics card. It's like going from a slow, fuel-guzzling steam train to a high-speed electric bullet train.

The Bottom Line

FengWu is a leap forward because it stops trying to simulate the physics of every molecule and starts learning the patterns of the atmosphere like a master detective. By treating different weather elements as distinct "modalities," coaching itself to focus on the hard problems, and learning from its own past mistakes, it can now tell us what the weather will be like more than 10 days in advance with high accuracy.

This means farmers can plan harvests further out, airlines can route flights more efficiently, and cities can prepare for extreme weather events with much more lead time. It's not just a better forecast; it's a smarter way to see the future.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →