Mapping Historic Urban Footprints in France: Balancing Quality, Scalability and AI Techniques

This study presents a scalable dual-pass deep learning pipeline that successfully extracts the first open-access, nationwide urban footprint dataset for metropolitan France from historical maps (1925–1950), achieving 73% accuracy by effectively mitigating artifacts like text and contour lines to enable quantitative analysis of pre-1970s urban sprawl.

Walid Rabehi, Marion Le Texier, Rémi Lemoy

Published Wed, 11 Ma
📖 4 min read☕ Coffee break read

Imagine trying to understand how a city grew over 100 years ago, but all you have are old, dusty, hand-drawn maps scattered across a library. Some are faded, some have coffee stains, and the ink is smudged. Trying to figure out exactly where the buildings were just by looking at these maps is like trying to find a needle in a haystack while wearing foggy glasses.

This paper is about a team of researchers who built a super-smart digital detective to solve this problem for all of France.

Here is the story of how they did it, broken down into simple steps:

1. The Problem: The "Foggy Glasses"

For a long time, we have great satellite photos of cities from the 1970s to today. But before that? We have a huge gap. We know France existed, and we know cities grew, but we didn't have a digital map showing exactly where the buildings were between 1925 and 1950.

The only clues were the Scan Histo maps: beautiful, high-resolution scans of old paper maps. But they are tricky. They are full of:

  • Text: City names and labels that look like buildings.
  • Roads: Lines that look like streets but aren't buildings.
  • Shadows and Stains: Old paper gets dirty, and the ink fades differently in different places.

If you tried to teach a computer to read these maps using standard methods, it would get confused. It might think a city name written in big bold letters is a skyscraper, or that a contour line (a line showing a hill) is a wall.

2. The Solution: The "Two-Pass" Detective

The researchers didn't just throw a standard computer program at the problem. They created a two-step training process, kind of like training a new employee.

  • Pass 1: The "Rough Draft" Detective
    First, they taught an AI (a type of computer brain called a U-Net) using a small set of examples. This AI made a first guess at where the cities were.

    • The Result: It was okay, but messy. It found the big black buildings but got confused by big text and roads.
    • The Trick: Instead of giving up, the researchers looked at where the AI made mistakes. They used those mistakes to teach the AI what not to look for.
  • Pass 2: The "Refined" Detective
    They gave the AI a second chance, but this time with a better "textbook." They fed it the messy results from the first pass and showed it specifically: "Hey, this text isn't a building," and "This road isn't a house."

    • The Result: The AI learned to ignore the noise. It stopped thinking the word "Paris" was a building and started focusing only on the actual shapes of the cities.

3. The Scale: Cleaning the Whole Map

They didn't just do this for one town. They applied this method to 941 giant map tiles covering the entire country of France (except for a few islands).

  • Imagine taking a puzzle of the whole country, cutting it into 941 pieces, and having a robot clean up every single piece to find the buildings.
  • They used a supercomputer (a very powerful machine) to do this work, which would have taken a human hundreds of years to finish.

4. The Result: A Time-Travel Map

The final product is a digital map of France from the 1920s–1950s that shows exactly where the urban areas were.

  • Accuracy: It got about 73% right. That's a huge success for such old, messy maps.
  • Why it matters: Before this, we had to guess how cities grew before the 1970s. Now, we have a clear picture. We can see how Paris expanded, how small towns grew into cities, and how World War II destruction and rebuilding changed the landscape.

5. The "Open Source" Gift

The best part? The researchers didn't keep this secret. They released everything to the public:

  • The code (the recipe for the detective).
  • The data (the old maps and the new digital version).
  • The results (the final map).

The Big Picture Analogy

Think of this like restoring an old, damaged photograph.

  • Old Way: You try to guess what's in the photo by squinting. You might miss details or see things that aren't there.
  • This Paper's Way: You use a smart AI tool that learns to "clean" the photo. It learns to ignore the scratches (text and roads) and highlight the people (the buildings). Once it's trained, it can clean up thousands of photos instantly, giving us a crystal-clear view of history that was previously blurry.

This work allows historians, urban planners, and anyone interested in the past to finally "see" the cities of 100 years ago with the same clarity we have for today's cities.