An Updated Assessment of Reinforcement Learning for Macro Placement

This paper presents an updated and rigorous assessment of Google's deep reinforcement learning approach (Circuit Training) for macro placement by introducing stronger baselines, new sub-10nm benchmarks, and commercial-grade evaluations to address reproducibility challenges and identify remaining open questions regarding scalability and pre-training methodologies.

Chung-Kuan Cheng, Andrew B. Kahng, Sayak Kundu, Yucheng Wang, Zhiang Wang

Published 2026-03-12
📖 5 min read🧠 Deep dive

Imagine you are trying to pack a massive, incredibly complex moving truck. The truck is a computer chip, and the items you need to pack are huge, heavy furniture pieces (called "macros") like memory banks and processor cores, mixed with millions of tiny Lego bricks (standard cells).

Your goal is to arrange these items so the truck is as small as possible, the items don't rattle around too much (low power), and you can easily reach everything without tripping over wires (high performance).

For decades, humans and traditional computer algorithms have been the best at solving this packing puzzle. But in 2021, Google announced a new "AI Robot" (called AlphaChip or Circuit Training) that claimed it could pack this truck better than any human, in less than six hours, using deep learning.

This paper is a reality check. A team of researchers decided to test Google's claim again, but this time with a much stricter rulebook, better tools, and a fairer playing field. Here is what they found, explained simply:

1. The "Black Box" Problem

When Google first announced their AI, they said, "Here is our amazing result!" but they didn't give everyone the exact recipe or the exact ingredients. It was like a magician showing a trick but hiding the secret move.

  • The Issue: Other scientists couldn't fully copy the experiment because the code and data were incomplete or "black boxed."
  • The Fix: This new paper built a completely open, transparent lab. They recreated the environment, fixed the missing pieces, and made sure everyone could see exactly how the sausage was made.

2. The "Old Dog" vs. The "New Puppy"

The researchers compared three main contenders:

  • The Human Expert: A master packer who has done this for years.
  • The "Old Dog" (Simulated Annealing): A classic, math-based algorithm that has been around since the 1980s. The researchers upgraded this old dog with a new "super collar" (a technique called "Go-With-The-Winners" and multi-threading) to make it run faster and smarter.
  • The "New Puppy" (Google's AI): The latest version of Google's reinforcement learning robot.

The Result:
Surprisingly, the upgraded "Old Dog" (Simulated Annealing) and the Human Experts consistently packed the truck better than the AI.

  • Better Packing: The traditional methods created layouts with less wire length (fewer tangled wires) and better timing.
  • Cheaper & Faster: The AI required a massive amount of computing power (like a supercomputer farm) and took a long time to train. The traditional methods did the same job using a fraction of the energy and time.

3. The "Practice Makes Perfect" Myth

Google suggested that their AI gets better if you "pre-train" it on other chips first (like a student studying math before taking a physics exam).

  • The Study: The researchers tried this pre-training method.
  • The Finding: While pre-training helped the AI converge on some problems, it didn't magically make it beat the traditional methods. In fact, when the chips got really huge (scaled up), the AI often got confused and failed to find a good solution, whereas the traditional methods kept working steadily.

4. The "Fake Score" vs. The "Real Score"

This is the most critical insight.

  • The Proxy Score: The AI was trained to minimize a "proxy score" (a simulation of how good the chip is). It was like a student trying to get a high score on a practice test.
  • The Real Score: The researchers ran the final designs through a commercial "Place-and-Route" tool (the real factory simulator) to see the actual performance, power, and area.
  • The Disconnect: They found that a low "proxy score" (a good practice test result) did not guarantee a good real-world result. The AI was optimizing for the wrong things. It was like a student who memorized the practice test answers perfectly but failed the actual exam because the questions were slightly different.

5. The "Stability" Issue

The researchers found that the AI is a bit of a "gamble."

  • If you run the AI twice with the same settings, you might get two very different results. Sometimes it works great; sometimes it fails completely.
  • The traditional methods are like a reliable clock: if you set them up the same way, they give you the exact same high-quality result every time.

The Big Takeaway

This paper isn't saying AI is useless. It's saying that hype doesn't equal reality.

  • Reproducibility Matters: Science needs to be open. If you can't copy the experiment, you can't trust the result.
  • Don't Ignore the Classics: Just because a new, flashy technology (AI) arrives, it doesn't mean the old, well-understood tools (Simulated Annealing) are obsolete. In fact, when optimized correctly, the old tools are still winning.
  • Measure What Matters: You can't just optimize for a simulation score; you have to optimize for the final, real-world product.

In short: The AI robot is a promising student, but the veteran human packer and the upgraded classic algorithm are still the ones getting the job done best, faster, and cheaper. The research community needs to be careful not to get swept up in the hype before the data is truly solid.