Assessment of Spatio-Temporal Predictors in the Presence of Missing and Heterogeneous Data

This paper proposes a residual correlation analysis framework that utilizes tailored spatio-temporal graphs and distribution-free statistics to assess the optimality and localize performance gaps of deep learning models in the presence of missing and heterogeneous spatio-temporal data.

Daniele Zambon, Cesare Alippi

Published 2026-03-02
📖 5 min read🧠 Deep dive

Imagine you are a chef running a massive, high-tech kitchen that predicts exactly what your customers will want to eat every hour of the day, every day of the year. You have a super-smart AI chef (a Deep Learning Model) that looks at the weather, the day of the week, and past orders to guess the next meal.

Usually, to see if your AI chef is doing a good job, you just taste the food and check the score: "Did the customer like it? Yes/No." If the score is high, you think, "Great job!"

But what if the AI chef is secretly struggling in specific corners of the kitchen? What if it's great at predicting lunch but terrible at predicting breakfast, or great at predicting orders for the north side of the city but confused by the south side? Traditional scores might hide these problems because the average score still looks good.

This paper introduces a new tool called AZ-Analysis. Think of it not as a taste test, but as a detective's magnifying glass that looks at the mistakes (the "residuals") the AI makes.

Here is how it works, using simple analogies:

1. The "Whiteness" Test: Is the Noise Random?

Imagine your AI chef makes a mistake.

  • Good Mistakes: If the AI guesses "Pizza" and the customer orders "Burger," and the next time it guesses "Burger" and the customer orders "Salad," these mistakes are random. They are like static on a radio—unpredictable and scattered. This is actually good because it means the AI has learned everything it can; the rest is just pure luck.
  • Bad Mistakes: If the AI guesses "Pizza" and the customer orders "Burger," and every single time it makes that mistake, it keeps making the same mistake in a pattern, that's a problem. It's like a broken record skipping the same spot. This means the AI is missing a hidden rule (e.g., "It's Tuesday, so people always want Tacos").

The paper's method checks if the mistakes are random "static" or a "broken record." If they are a broken record, the AI isn't finished learning yet.

2. The "Spatio-Temporal" Map: Where and When?

The data this AI handles is Spatio-Temporal.

  • Spatial (Space): Think of a map of the city. The "nodes" are different neighborhoods or sensors.
  • Temporal (Time): Think of a timeline.

The paper builds a 3D Map (a graph) connecting these neighborhoods across time.

  • The Problem: Real-world data is messy. Some sensors break (missing data), and some neighborhoods have weird rules (heterogeneous data). Traditional tools crash when data is missing or messy.
  • The Solution: The AZ-Analysis is like a robust drone that can fly over a stormy, broken city. It doesn't care if some buildings are missing or if the streets are different shapes. It can still spot where the "broken record" mistakes are happening.

3. The Three Questions It Answers

The paper says this tool helps answer three specific questions:

  • Q1: Is the AI totally done learning?

    • Analogy: Is the radio static truly random, or is there a hidden song playing underneath?
    • Result: If the tool finds patterns, the AI is not optimal yet.
  • Q2: Which specific neighborhoods are struggling?

    • Analogy: Is the AI bad at predicting orders for Downtown but great for Suburbia?
    • Result: The tool highlights specific "nodes" (sensors or areas) where the mistakes are correlated. You can then go fix the AI's training just for that neighborhood.
  • Q3: Which specific times are the AI failing?

    • Analogy: Does the AI get confused every time the sun rises or sets?
    • Result: The tool points out specific time intervals (like "Dawn" or "Rush Hour") where the model fails, even if the overall error score looks low.

4. Why Is This Special?

Most statistical tools are like fine-dining critics: they require the food to be perfectly plated, served at the right temperature, and the ingredients to be identical. If the data is messy (missing values, weird distributions), these tools refuse to work.

The AZ-Analysis is like a street food inspector. It doesn't care if the data is messy, incomplete, or weird. It just looks at the pattern of the mistakes.

  • It doesn't need to know the "distribution" of the data (it doesn't need to know if the data is a bell curve or a jagged mountain).
  • It only needs the mistakes to be centered around zero (the AI isn't consistently guessing too high or too low).

The Real-World Examples

The authors tested this on two real-world scenarios:

  1. Traffic Flow: They looked at traffic sensors. They found that the AI was making weird, correlated mistakes specifically when data was being "filled in" (imputed) because of missing sensors. The standard error scores didn't catch this, but the AZ-Analysis did.
  2. Solar Energy: They predicted energy production from solar panels. They found that the AI struggled specifically at dawn and dusk. Even though the average error was low, the pattern of mistakes showed the AI didn't understand the transition of light.

The Takeaway

This paper gives us a new way to audit our AI models. Instead of just asking, "How accurate is the average?" it asks, "Where and when is the AI confused, and is that confusion random or a sign of a deeper problem?"

It turns the "black box" of Deep Learning into a transparent map, showing us exactly where to shine a light to improve the model, even when the data is messy and incomplete.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →