Large Language Models for Travel Behavior Prediction

This study demonstrates that large language models, utilized through zero-shot prompting or as embedding generators for supervised learning, offer a flexible and data-efficient alternative to traditional numerical models for predicting travel behavior.

Baichuan Mo, Hanyong Xu, Ruoyun Ma, Jung-Hoon Cho, Dingyi Zhuang, Xiaotong Guo, Jinhua Zhao

Published Thu, 12 Ma
📖 4 min read☕ Coffee break read

Imagine you are trying to predict what a person will do next: Will they drive, take the train, or hop on a bus? Will they be going to work, to a party, or to the grocery store?

For decades, transportation experts have tried to answer this using mathematical calculators. These calculators are like rigid recipe books. If you feed them enough data (thousands of past trips), they can cook up a pretty good prediction. But if you only give them a few ingredients (a small amount of data), the recipe falls apart, and the results taste terrible.

This paper introduces a new chef to the kitchen: Large Language Models (LLMs). Think of these as super-smart, well-read librarians who have read almost every book, article, and blog post on the internet. They don't just crunch numbers; they understand stories, logic, and human common sense.

The researchers tested two ways to use these "librarian chefs" to predict travel behavior:

1. The "Zero-Shot" Oracle (The Instant Expert)

The Analogy: Imagine you walk up to a genius who has never met you before, but you describe your situation: "I'm in a rush, I hate traffic, and I have a monthly train pass." You ask, "How will I get to work?"

The genius doesn't need to study your past data. They just use their vast knowledge of how the world works to say, "You'll probably take the train because it's fast and you already have a pass."

  • How it works in the paper: The researchers wrote a detailed prompt (a set of instructions) describing a traveler's situation and the rules of the road (like "people hate traffic"). They asked the AI to guess the travel mode.
  • The Result: Even without looking at any specific data about the traveler, the AI guessed almost as well as the traditional "recipe book" models, and sometimes even better when data was scarce.

2. The "Smart Translator" (The Feature Booster)

The Analogy: Imagine you have a very simple calculator that is bad at understanding complex human feelings. But, you have a translator who can turn a messy, emotional story into a clean, organized summary.

You give the messy story to the translator (the LLM). The translator writes a short, high-level summary (an "embedding") that captures the essence of the story. You then feed this clean summary to your simple calculator. Suddenly, your calculator becomes a genius because it's working with the "soul" of the data, not just the raw numbers.

  • How it works in the paper: The researchers used the AI to convert travel descriptions into mathematical "vectors" (digital summaries). They then fed these summaries into standard models.
  • The Result: When data was very limited, this "Translator + Calculator" combo outperformed the standard models that tried to learn from scratch.

The Good, The Bad, and The "Hallucinations"

The Superpower: Explainability
Old models are like black boxes. They give you an answer (e.g., "Take the train") but can't tell you why.
The LLM is like a detective. It says, "I think you'll take the train because you have a pass, and it saves you 30 minutes compared to driving." This is huge for planners who need to understand why people make choices.

The Weakness: The "Confident Liar"
Sometimes, the AI gets too creative. This is called a hallucination.

  • Example: If you ask the AI to compare travel times, it might confidently say, "The train is 50% faster!" when the data actually says it's only 10% faster. It's like a student who knows the concept of math but makes a silly calculation error because they are trying to sound smart.

The Takeaway

This paper is a wake-up call for transportation planners.

  • If you have a mountain of data: Stick to your traditional math models; they are still the kings.
  • If you have very little data (a small town, a new city, a rare event): Don't panic. Call the "Librarian" (the LLM). It can use its general knowledge of human behavior to make surprisingly accurate guesses without needing to be taught the specific rules of your town first.

In short, we are moving from a world where computers only count to a world where computers can reason about how we move. It's not perfect yet, but it's a powerful new tool in the toolbox.