LocalSUG: Geography-Aware LLM for Query Suggestion in Local-Life Services

This paper presents LocalSUG, a geography-aware LLM framework for local-life service query suggestion that overcomes challenges in geographic grounding, exposure bias, and inference latency through city-aware candidate mining, a beam-search-driven GRPO algorithm, and quality-aware acceleration techniques, ultimately achieving significant improvements in click-through rate and search success in large-scale online deployment.

Jinwen Chen, Shuai Gong, Shiwen Zhang, Zheng Zhang, Yachao Zhao, Lingxiang Wang, Haibo Zhou, Yuan Zhan, Wei Lin, Hainan Zhang

Published 2026-03-06
📖 5 min read🧠 Deep dive

Imagine you are walking into a massive, bustling food court in a new city. You are hungry and type "Pizza" into the search bar.

In a traditional system, the computer looks at a giant, static list of the most popular pizzas ever ordered anywhere in the world. It might suggest "Domino's." But if you are in a city where Domino's doesn't exist, that suggestion is useless. You have to keep typing, get frustrated, and maybe leave without eating.

This paper introduces LocalSUG, a new, super-smart assistant designed specifically for local life services (like food delivery, hotels, or local shops) that fixes these problems using a Large Language Model (LLM).

Here is how LocalSUG works, explained through simple analogies:

1. The Problem: The "Tourist" vs. The "Local"

The authors identified three main headaches with using AI for local search:

  • The Tourist Problem (Geographic Grounding): A standard AI is like a tourist who thinks "Pizza" means "Domino's" everywhere. It doesn't know that in your specific city, "Pizza" actually means "Papa John's" or a local favorite. It gives you suggestions that sound right but are impossible to find.
  • The "Practice vs. Reality" Problem (Exposure Bias): Imagine a chef who practices cooking by reading a recipe book (training) but then has to cook a meal for a crowd using a specific order of ingredients (inference). If the practice doesn't match the real cooking method, the meal fails. Traditional AI trains on lists of past searches but tries to generate suggestions one by one, leading to messy, inconsistent results.
  • The Speed Bump (Latency): These smart AI models are heavy. Asking them for an answer takes too long. In a busy food court, if the waiter takes 5 seconds to think of a suggestion, the customer walks away.

2. The Solution: LocalSUG

LocalSUG is a framework built to be a hyper-local, fast, and consistent assistant. Here are its three secret weapons:

A. The "Local Guide" Map (City-Aware Mining)

Instead of guessing, LocalSUG acts like a local guide who knows exactly what's open in your neighborhood.

  • How it works: Before the AI even starts thinking, it pulls a "cheat sheet" based on your city. If you are in Beijing, it knows to suggest Domino's. If you are in Macau, it knows Domino's isn't there and suggests Pizza Hut instead.
  • The Analogy: It's like having a GPS that doesn't just show you the fastest route, but also knows which gas stations are actually open in your specific town.

B. The "Rehearsal" Method (Beam-Search-Driven Training)

To fix the "Practice vs. Reality" problem, the authors changed how they teach the AI.

  • How it works: Instead of just teaching the AI to write one perfect sentence, they teach it to generate a whole list of options at once (like a beam search) and then pick the best ones. They use a special scoring system (GRPO) that rewards the AI not just for being correct, but for being relevant to the business (e.g., "Did the user click? Did they order?").
  • The Analogy: Imagine a basketball coach. Instead of just telling the player, "Shoot the ball," the coach makes the player practice shooting a whole series of shots in a game scenario. The player learns to handle the pressure of the game, not just the drill. This ensures the AI performs perfectly when it's actually serving customers.

C. The "Speedy Chef" (Quality-Aware Acceleration)

To fix the speed issue, they made the AI run faster without losing its brainpower.

  • How it works: They taught the AI to ignore words it will never use (vocabulary pruning) and to stop thinking once it has found a few good answers (early stopping).
  • The Analogy: Imagine a chef who usually checks 10,000 spices in the pantry. LocalSUG is like a chef who only keeps the 30,000 most common spices on the counter and knows exactly when to stop tasting the soup. They get the perfect flavor 3x faster.

3. The Results: A Happier Customer

When they tested this in the real world (on a massive platform with millions of users), the results were impressive:

  • Fewer Dead Ends: The rate of users searching for something and finding nothing dropped by 2.56%.
  • More Clicks: People clicked on the suggestions more often (+0.35%).
  • Less Typing: Users had to type fewer letters to find what they wanted because the suggestions were so accurate.
  • Discovery: Users found more unique, interesting items they hadn't seen before (like that hidden gem pizza place).

Summary

LocalSUG is like upgrading a generic, slow, out-of-touch tour guide into a local expert who knows your neighborhood, practices for the real game, and gives you answers instantly. It turns the frustrating experience of "I can't find anything" into a smooth, helpful discovery of exactly what you need, right where you are.