Assessing airborne laser scanning and aerial photogrammetry for deep learning-based stand delineation

This study demonstrates that deep learning-based forest stand delineation achieves comparable accuracy using temporally aligned digital photogrammetry-derived canopy height models and digital terrain models as it does with airborne laser scanning data, suggesting that large-scale, consistent datasets can be assembled without relying on the more complex and temporally misaligned ALS data.

Håkon Næss Sandum, Hans Ole Ørka, Oliver Tomic, Terje Gobakken

Published 2026-02-26
📖 5 min read🧠 Deep dive

Imagine you are a forest manager trying to organize a massive, wild library. In this library, the "books" are trees, and the "shelves" are called forest stands. A forest stand is just a group of trees that look and act the same (same age, same species, same size) so they can be managed together.

Traditionally, humans have had to draw the lines between these shelves by hand, looking at aerial photos and guessing where one group of trees ends and another begins. It's slow, it's subjective (two experts might draw the lines differently), and it's hard to scale up.

This paper is about teaching a computer (using Deep Learning) to draw these lines automatically, and testing if it can do it just as well as a human expert.

Here is the breakdown of their experiment, explained with simple analogies:

1. The Problem: The "Mismatched Puzzle"

To teach the computer, you need to show it pictures of the forest and the "correct" answer (the lines drawn by an expert).

  • The Old Way: They used two different types of data: Aerial Photos (taken by a plane with a camera) and Laser Scans (ALS, where a plane shoots lasers to measure height).
  • The Glitch: The photos and the laser scans were often taken at different times (maybe a year apart). It's like trying to solve a puzzle where the picture on the box is from 2023, but the puzzle pieces are from 2024. Trees might have been cut down or grown, making the "correct" answer look wrong compared to the input data. This limits how much data you can use to train the AI.

2. The Solution: The "All-in-One" Camera

The researchers wanted to know: Can we use a newer technology called Digital Aerial Photogrammetry (DAP) instead of lasers?

  • How DAP works: It takes thousands of overlapping photos and uses math to build a 3D model of the trees, just like the lasers do, but using only the photos.
  • The Benefit: Since the 3D model and the color photos come from the exact same flight, they are perfectly synchronized in time. No more mismatched puzzles!
  • The Fear: Critics worried that DAP might be "smoother" and miss small details (like gaps in the canopy) compared to the super-precise lasers.

3. The Experiment: The "Taste Test"

The researchers set up a "taste test" for their AI (a neural network called U-Net). They trained three different versions of the AI on data from six different towns in Norway:

  1. Team Laser: Used the old-school Laser Scans + Photos.
  2. Team Photo: Used the new DAP 3D models + Photos.
  3. Team Terrain: Used DAP + Photos + a map of the ground elevation (DTM).

They asked: Which team draws the lines most like the human experts?

4. The Results: The "Twin" Effect

The results were surprising and very encouraging:

  • All teams performed equally well. The AI using the new "Photo-only" 3D models (Team Photo) did just as good a job as the one using the expensive Laser scans.
  • The "Twin" Effect: When they compared the AI's drawings to the human expert's drawings, there was some disagreement (which is normal; humans disagree too). BUT, when they compared the three different AIs to each other, they agreed almost perfectly.
    • Analogy: Imagine three different chefs making a cake. They all taste slightly different from the "perfect" recipe (the human expert), but they all taste almost identical to each other. This means the AI has learned a very consistent way of seeing the forest, even if it's not exactly how a human sees it.
  • The Ground Map didn't help: Adding the "Terrain" map (DTM) didn't make the AI any better. In this specific area (which is relatively flat), the AI could already "feel" the hills and valleys just by looking at the trees and shadows.

5. The Catch: "Over-zealous" Cutting

The AI was very good at finding the forest, but it sometimes got a little too excited about drawing lines.

  • It would draw tiny, jagged lines around small patches of trees that a human would ignore.
  • Analogy: It's like a child coloring inside the lines but making the lines so wiggly and detailed that the picture looks messy. The researchers noted that these tiny errors can be fixed later with a simple "smoothing" tool, just like editing a photo.

6. Why This Matters

This study is a big deal for the future of forestry because:

  1. It's Cheaper and Easier: You don't need expensive laser scanners. You just need a plane with a camera.
  2. It's Faster: Because the data is perfectly aligned in time, you can train the AI on much larger areas without worrying about "mismatched" data.
  3. It's Consistent: The AI is less tired and less subjective than a human. It draws the lines the same way every time.

In a nutshell: The researchers proved that a computer can learn to organize a forest just as well using "smart photos" as it does using expensive lasers. This opens the door to automating forest management on a massive scale, saving time and money while keeping our forests healthy.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →