Conformal prediction for high-dimensional functional time series: Applications to subnational mortality

This paper proposes a model-agnostic, distribution-free approach using split and sequential conformal prediction to construct prediction intervals for high-dimensional functional time series, demonstrating their effectiveness in forecasting subnational mortality rates for Japan and Canada compared to traditional methods.

Han Lin Shang

Published Thu, 12 Ma
📖 4 min read☕ Coffee break read

Imagine you are a weather forecaster. You don't just want to say, "It will rain tomorrow." You want to say, "It will rain, and I'm 95% sure the amount will be between 1 and 3 inches." That "between 1 and 3 inches" part is your prediction interval. It's your safety net.

This paper is about building better, safer nets for predicting complex, changing patterns—specifically, how death rates change over time across different regions in Japan and Canada.

Here is the breakdown of the paper's ideas using simple analogies:

1. The Problem: The "Crystal Ball" is Cracked

Traditionally, statisticians build complex mathematical models (like a crystal ball) to predict the future. They assume the world follows specific rules.

  • The Risk: If the rules you assumed are slightly wrong (model misspecification), your crystal ball gives you a false sense of security. Your "safety net" might be too small, and you'll get wet when it rains.
  • The Old Fix: Some people use "bootstrapping" (resampling data over and over), but that's like trying to count every grain of sand on a beach by picking them up one by one. It's accurate but takes forever and requires a supercomputer.

2. The Solution: The "Conformal Prediction" Safety Net

The author, Han Lin Shang, suggests a smarter way called Conformal Prediction. Think of this not as a crystal ball, but as a tailor.

  • Instead of guessing the rules of the universe, the tailor looks at how much the fabric (the data) actually stretches and shrinks in the past.
  • It doesn't care what kind of fabric you have (it's "model-agnostic"). It just measures the real-world wiggles and creates a net that fits those wiggles perfectly.
  • The Goal: To create a prediction interval that is guaranteed to catch the future outcome 95% of the time, no matter how messy the data is.

3. The Challenge: Too Many Threads (High-Dimensional Data)

The data here isn't just one line; it's a massive tapestry.

  • The Data: Death rates for every age group (0 to 100+) in 47 different Japanese prefectures, tracked over 49 years.
  • The Analogy: Imagine trying to predict the weather for 47 different cities simultaneously, where the temperature in each city changes every day in a unique curve. That is a "High-Dimensional Functional Time Series." It's a lot of moving parts.

4. The Two Methods: The "Split" vs. The "Live Update"

The paper compares two ways to make this tailor work:

Method A: Split Conformal Prediction (The "Rehearsal" Approach)

  • How it works: You take your data and cut it into three pieces:
    1. Training: To learn the pattern.
    2. Validation (Rehearsal): To test how wide your net should be.
    3. Test: The real future prediction.
  • The Flaw: It's like a chef tasting a soup before the main event to decide how much salt to add. But once the main event starts, the chef can't taste it again. If the soup changes flavor later (new data arrives), the chef is stuck with the old salt level.
  • Result: In the paper, this method often made the net too small (underestimated risk), especially for long-term predictions.

Method B: Sequential Conformal Prediction (The "Live Update" Approach)

  • How it works: This method doesn't need a separate "rehearsal" phase. As soon as new data arrives (e.g., this year's death rates), it instantly updates the net.
  • The Analogy: Imagine a smart thermostat that doesn't just guess the temperature; it constantly feels the room and adjusts the heating right now. If the room gets colder, the net widens immediately.
  • Result: This method was conservative. It made the net slightly larger than necessary (overestimated risk), but that's a good thing! It meant the actual outcomes were almost always caught inside the net.

5. The Verdict: Better Safe Than Sorry

The author tested these methods on real data from Japan and Canada.

  • The Winner: Sequential Conformal Prediction.
  • Why? It didn't waste data on a "rehearsal" set. It learned on the fly. While it created slightly wider nets (which might look like "less precise" at first glance), those nets were more reliable. They caught the truth more often than the "Split" method.

Summary in One Sentence

Instead of guessing the future based on rigid rules, this paper teaches us to build a flexible, self-updating safety net that gets smarter every time a new piece of data arrives, ensuring we are rarely caught off guard by the unexpected.