From Code to Figure: A FAIR-Aligned Data Provenance Chain for Reproducible Simulation Research in Numerical Physics

This paper presents an integrated, FAIR-aligned workflow that combines version control, automated testing, structured logging, and standardized post-processing to establish a complete data provenance chain ensuring reproducibility from code development to published figures in numerical physics simulations.

Original authors: Markus Uehlein, Tobias Held, Christopher Seibel, Lukas G. Jonda, Baerbel Rethfeld, Sebastian T. Weber

Published 2026-04-30
📖 5 min read🧠 Deep dive

This is an AI-generated explanation of the paper below. It is not written or endorsed by the authors. For technical accuracy, refer to the original paper. Read full disclaimer

Imagine you are a chef who has spent years perfecting a complex recipe for a dish that changes slightly every time you cook it. One day, you publish a photo of the final dish in a cookbook. A year later, someone tries to recreate it, but they can't. Why? Because they don't know exactly which version of the recipe you used, what specific brand of ingredients you had in your pantry that day, or if you tweaked the oven temperature mid-cook.

This paper, written by Markus Uehlein and his team, is about solving that exact problem for scientists who run computer simulations instead of cooking meals. In the world of "numerical physics" (using computers to model how materials behave), the "recipes" are software codes that are constantly being updated, and the "dishes" are massive datasets.

Here is how the authors propose to keep everything traceable, using a simple, four-step workflow they call a Data Provenance Chain.

1. The Recipe Book (Version Control & Code Review)

In the past, if a scientist changed a line of code, they might just save it as simulation_final_v2_real_final.cpp. This is a recipe disaster waiting to happen.

The authors use a system called Git (think of it as a time-traveling recipe book). Every time someone changes the code, it gets a unique timestamp and a "review" from a colleague before it's saved. This ensures that if you look at a simulation from five years ago, you can see the exact version of the code used, down to the specific line of text. It's like having a photo of the chef's hands and the exact ingredients on the counter at the moment the dish was made.

2. The Safety Checks (Automated Testing)

Before a simulation runs, the software performs automatic "safety checks."

  • Unit Checks: The code checks if the math makes sense physically. For example, it won't let you add "meters" to "seconds" (you can't add distance to time!). If you try, the computer stops you before the simulation even starts.
  • Physics Checks: The code runs tiny test simulations to make sure the physics behaves the way it should (e.g., "If I heat this up, does the energy go up?"). If the answer is no, the system knows something is broken.

3. The "Black Box" Recorder (Structured Logging & Metadata)

When the simulation actually runs, it doesn't just spit out a list of numbers. It creates a hierarchical file (a fancy digital folder structure) that acts like a "black box" recorder on an airplane.

Inside this file, the scientists store:

  • The raw data (the results).
  • The exact input settings (the recipe).
  • The "build log" (what version of the code was used).
  • The environment (what kind of computer CPU was used).
  • A diary of the run (any warnings or errors that happened while it was cooking).

They use a standard format called HDF5/NeXus. Think of this as a universal container that keeps the data organized so that even if the original scientist forgets what they did, anyone else can open the box and understand exactly what happened.

4. The Plating (From Data to Figures)

Finally, the scientists turn that raw data into the pretty graphs and pictures you see in a published paper. Usually, this step is messy—scientists might write a one-off script to make a graph and then delete it.

In this workflow, the step to make the picture is also version-controlled. The script used to make the graph is saved, and the graph itself is stamped with a link back to the raw data and the code used to make it.

The Big Picture: The "Chain of Custody"

The main point of this paper is that these four steps shouldn't be separate islands. They need to be a chain.

  • Old Way: You publish a picture. Someone asks, "How did you get this?" You say, "I ran a simulation." They ask, "Which one?" You say, "I think it was the one from last Tuesday." Reproducibility fails.
  • New Way (The Paper's Method): You publish a picture. You click a link, and it shows you the exact code version, the exact input file, the computer it ran on, and the script used to make the picture. Reproducibility succeeds.

The authors tested this on their own long-running simulation software (called monstr), which has been used for many studies over several years. They showed that by linking the code, the data, and the figures together, they created a system where anyone can trace a published result all the way back to the original software state, ensuring that scientific findings remain reliable and reusable for the long term.

In short: They built a system where every scientific result comes with its own "receipt" that proves exactly how it was made, preventing the "it works on my machine" problem from ruining scientific trust.

Drowning in papers in your field?

Get daily digests of the most novel papers matching your research keywords — with technical summaries, in your language.

Try Digest →