A Learned Proximal Alternating Minimization Algorithm and Its Induced Network for a Class of Two-block Nonconvex and Nonsmooth Optimization

This paper proposes a learned proximal alternating minimization (LPAM) algorithm and its corresponding interpretable network (LPAM-net) for solving two-block nonconvex and nonsmooth optimization problems, proving their convergence to Clarke stationary points and demonstrating superior performance in joint multi-modal MRI reconstruction.

Yunmei Chen, Lezhi Liu, Lei Zhang

Published 2026-03-10
📖 5 min read🧠 Deep dive

Here is an explanation of the paper, translated from academic jargon into a story about building a better puzzle solver.

The Big Picture: Solving the "Impossible" Puzzle

Imagine you are trying to solve a giant, complex jigsaw puzzle, but someone has ripped out 80% of the pieces and scattered the remaining ones on the floor. This is what happens in MRI imaging. To save time, machines often scan only a tiny fraction of the data (the "k-space"). The goal is to reconstruct the full, clear picture of a brain from these missing pieces.

Traditionally, computers try to guess the missing pieces using math. But when the math gets too messy (non-smooth) or the picture is too complex (non-convex), the computer gets stuck, confused, or produces blurry, artifact-ridden images.

This paper introduces a new method called LPAM (Learned Proximal Alternating Minimization) and a neural network built from it, called LPAM-net. Think of it as teaching a computer to be a master puzzle solver that doesn't just guess randomly, but follows a smart, proven strategy while learning from experience.


The Three Magic Ingredients

The authors combined three powerful ideas to create this new solver:

1. The "Sandpaper" Technique (Smoothing)

The Problem: Some parts of the math are "rough" or jagged (nonsmooth). Imagine trying to slide a heavy box across a floor covered in jagged rocks. It's hard to predict how it will move, and the computer gets stuck.
The Solution: The authors use a technique called smoothing. Imagine covering those jagged rocks with a layer of soft sandpaper. The floor is still bumpy, but now it's smooth enough for the computer to slide the box easily.

  • The Twist: As the computer gets better at solving the puzzle, they slowly remove the sandpaper (diminishing smoothing). This allows the computer to eventually handle the original, rough terrain without getting stuck.

2. The "Residual" Shortcut (ResNet)

The Problem: Deep learning networks often struggle to learn when they have to start from scratch every time. It's like trying to climb a mountain by taking one giant, exhausting step at a time.
The Solution: They borrowed an idea from Residual Learning (ResNet). Instead of asking the computer to rebuild the whole image from scratch, they ask it: "What is the small correction needed to fix the current image?"

  • The Analogy: Imagine you are painting a wall. Instead of painting the whole wall again, you just paint over the spots that look wrong. This "residual" approach makes training faster, prevents the computer from getting confused (vanishing gradients), and leads to higher quality results.

3. The "Safety Net" (BCD)

The Problem: Sometimes, the computer's "smart guess" (the ResNet step) might actually make the picture worse or go off-track.
The Solution: They built in a Safety Net using a classic method called Block Coordinate Descent (BCD).

  • The Analogy: Think of the computer as a hiker trying to find the bottom of a valley. The "ResNet" step is a fast, confident stride. But if the hiker feels like they are walking uphill or into a cliff, the "Safety Net" kicks in. It forces the hiker to take a very careful, guaranteed step down. This ensures the computer never loses its way, even if the fancy learning part makes a mistake.

How It Works in Practice: The MRI Example

The authors tested this on Multi-Modal MRI, which means looking at two different types of brain scans (T1 and T2) at the same time.

  • The Old Way: Usually, computers look at the T1 scan and the T2 scan separately, like two people solving two different puzzles in isolation.
  • The LPAM Way: This new method looks at both puzzles simultaneously. It realizes that the T1 and T2 scans share common features (like the shape of the brain or a tumor). By learning these shared features together, it fills in the missing pieces much more accurately.

The Results: Why It Matters

When they tested LPAM-net against other top-tier methods:

  1. Better Quality: The reconstructed images were sharper, with higher "PSNR" (a score for image quality) and better structural similarity. The edges of tumors and tissues were clearer.
  2. Fewer Parameters: Despite being smarter, the network was actually smaller and more efficient. It didn't need millions of extra settings to work; it used the math structure to do the heavy lifting.
  3. Reliability: The authors proved mathematically that the algorithm will converge to a good solution. It's not just a "black box" that works sometimes; it has a guarantee that it will eventually find the right answer.

The Takeaway

This paper is about building a hybrid brain. It combines the raw power of deep learning (learning from data) with the reliability of classical math (proven algorithms).

  • Old Deep Learning: "I've seen millions of images, I'll guess the rest." (Fast, but can be unreliable or unexplainable).
  • Old Math: "I will follow these strict rules to find the answer." (Reliable, but slow and can't handle messy real-world data).
  • LPAM-net: "I will follow the strict rules, but I'll learn how to take the shortcuts along the way."

The result is a tool that is fast, accurate, efficient, and—most importantly—trustworthy for doctors trying to diagnose brain tumors from incomplete scans.