Regularizing INR with diffusion prior self-supervised 3D reconstruction of neutron computed tomography data

This paper introduces Diffusive INR (DINR), a novel framework that regularizes implicit neural representations with a diffusion prior trained on synthetic data to achieve high-quality, artifact-reduced 3D reconstructions of concrete microstructures from sparse-view neutron computed tomography, outperforming state-of-the-art methods under extreme data limitations.

Maliha Hossain, Haley Duba-Sullivan, Amirkoushyar Ziabari

Published Thu, 12 Ma
📖 4 min read☕ Coffee break read

Imagine you are trying to solve a massive, 3D jigsaw puzzle, but someone has stolen 90% of the pieces. You only have a few scattered clues left. Your goal is to figure out what the complete picture looks like.

This is exactly the challenge scientists face when using Neutron Computed Tomography (CT) to look inside objects like concrete, batteries, or fuel cells. Neutrons are great at seeing through materials that block X-rays, but they are "lazy" and slow. To get a clear picture, you usually need to take thousands of photos (views) from every angle. But sometimes, you can't wait that long, or the equipment is too weak. So, you end up with very few photos—maybe just 5 or 9 instead of thousands.

When you try to rebuild the 3D image from so few photos using old-school math, the result is a blurry, distorted mess full of "ghosts" and streaks. It's like trying to guess the shape of a cat based on a single blurry paw print.

The New Solution: "DINR" (The Smart Detective)

The authors of this paper created a new tool called DINR (Diffusive Implicit Neural Representation). Think of DINR as a super-smart detective who doesn't just look at the few clues you have; they also have a massive library of "what things usually look like" in their head.

Here is how DINR works, broken down into simple steps:

1. The Two-Brain Approach

DINR has two "brains" working together:

  • Brain A (The Architect): This is an Implicit Neural Representation (INR). Imagine a digital artist who can draw a perfect 3D object using a tiny set of instructions (mathematical weights) instead of storing millions of pixels. This artist is great at creating smooth shapes but sometimes gets confused when the clues are too few.
  • Brain B (The Art Historian): This is a Diffusion Model. Think of this as an AI that has studied millions of 3D images of concrete, rocks, and bubbles. It knows the "texture" and "rules" of how these materials naturally look. It's like an art historian who knows that concrete usually has tiny pores and cracks, not smooth, perfect spheres.

2. The "Restoration" Process

When you give DINR a blurry, incomplete set of photos:

  1. The Architect tries to build a 3D model based only on the few photos you gave it. It's a rough draft.
  2. The Art Historian looks at that rough draft and says, "Hey, that doesn't look right. Real concrete has these tiny details. Let me fix it."
  3. They go back and forth. The Architect adjusts the shape to match the photos, and the Historian adjusts the texture to match reality.
  4. They repeat this dance until the image is sharp, detailed, and matches the few clues you have, without adding fake "ghosts."

Why is this better than the old way?

  • The Old Way (FBP): This is like trying to fill in a crossword puzzle by guessing every letter randomly. If you miss a few clues, the whole word becomes nonsense. In CT, this creates terrible streaks and blurs.
  • The "Middle" Way (MBIR): This is like a very careful puzzle solver who uses strict rules (like "edges must be straight"). It's better, but it often misses the tiny, complex details (like the tiny pores in concrete) because it's too rigid.
  • The DINR Way: This is the best of both worlds. It respects the few photos you have (the data) but uses its "memory" of what real materials look like (the prior) to fill in the missing gaps intelligently.

The Results: Seeing the Invisible

The researchers tested this on concrete. Concrete is full of tiny holes (pores) and cracks that are crucial for safety.

  • When they only had 4 or 5 views (extremely sparse data), the old methods produced a blurry blob where you couldn't see the holes.
  • DINR managed to reconstruct the concrete so clearly that you could see the tiny pores and the texture, almost as if they had taken thousands of photos.

The Big Picture

This is a breakthrough because it means we can scan things much faster or with weaker equipment and still get high-quality, detailed 3D images.

  • For Batteries: We could check if a battery is safe in real-time without waiting hours for a scan.
  • For Construction: We could inspect bridges or dams for hidden cracks much more easily.
  • For Science: We can study how water moves through soil or plants without destroying the sample.

In short, DINR is like giving a blurry, low-resolution photo a "magic upgrade" by teaching the computer to use its imagination (based on real-world training) to fill in the missing pieces perfectly. It turns a "good enough" guess into a "scientifically accurate" picture.