Doxing via the Lens: Revealing Location-related Privacy Leakage on Multi-modal Large Reasoning Models

This paper reveals that multi-modal large reasoning models pose a significant privacy risk by outperforming humans in inferring sensitive geolocation information from user images, a vulnerability driven by their advanced reasoning capabilities and lack of privacy safeguards, which the authors evaluate using a new framework, dataset, and attack strategy.

Weidi Luo, Tianyu Lu, Qiming Zhang, Xiaogeng Liu, Bin Hu, Yue Zhao, Jieyu Zhao, Song Gao, Patrick McDaniel, Zhen Xiang, Chaowei Xiao

Published 2026-03-04
📖 5 min read🧠 Deep dive

Imagine you just posted a selfie on social media. You're happy with your new haircut, and the background is just a blurry coffee shop or your living room. You think, "No one can tell where I live from this."

But according to this new research paper, you might be wrong.

The paper, titled "Doxing via the Lens," reveals that the newest, super-smart AI models (called Multi-modal Large Reasoning Models or MLRMs) have become like digital Sherlock Holmes. They don't just "see" a picture; they "think" about it. They can look at a photo of your backyard, notice the specific type of brick, the way the sun hits the fence, and the style of the mailbox, and then cross-reference that with their massive internal knowledge of the world to guess your exact home address.

Here is a breakdown of the paper's findings using simple analogies:

1. The Problem: The "Super-Sleuth" AI

In the past, if you asked an AI, "Where is this?" it might guess "Maybe a city?" or "Maybe a park?"

But the new generation of AI (like OpenAI's O3 or Google's Gemini) has a superpower: Reasoning.

  • The Analogy: Imagine a detective who doesn't just look at a fingerprint but also knows the brand of the shoe, the type of dirt on the sole, the local weather patterns, and the specific architecture of the neighborhood.
  • The Reality: These AIs can look at a casual photo and deduce: "That specific style of streetlamp is only found in this neighborhood. That house number font is unique to this city. Therefore, this photo was taken at 123 Maple Street."

2. The New Danger Zone: "Privacy Spaces"

The researchers created a new way to measure how dangerous a photo is. They call it a Three-Level Risk Framework:

  • Level 1 (Low Risk): A photo of a tourist spot or a busy street. (Everyone knows where this is).
  • Level 2 (Medium Risk): A photo of your house or backyard, but without you in it. (It reveals your home address).
  • Level 3 (High Risk): A selfie taken in your own home or backyard. (It reveals who lives there and exactly where they live).

The scary part: The AI is actually better at guessing your location from these "private" photos than a regular human expert is.

3. The Tools: "ClueMiner" and "GeoMiner"

To prove how this works, the researchers built two tools:

  • ClueMiner (The Detective's Notebook): This tool asks the AI, "What clues did you use to guess the location?"

    • Result: The AI admits it looks at things like license plate styles, trash can colors, specific tree types, and street signs. It treats these as "clues" to solve the puzzle. The problem? The AI has no "privacy filter" to stop itself from using these clues. It just wants to solve the puzzle.
  • GeoMiner (The Team-Up Attack): This simulates a hacker working with the AI.

    • The Analogy: Imagine a human detective who is bad at geography but good at spotting details. They point at a photo and say, "Hey, look at that weird fence!" They hand that clue to the AI, which then says, "Ah! That fence is only in this specific neighborhood!"
    • Result: When humans help the AI by pointing out clues, the AI's accuracy skyrockets. It makes it incredibly easy for non-experts to find your address.

4. The "Mirror" Effect

The paper also found that AIs are getting scary good at looking at reflections.

  • The Analogy: If you take a selfie in front of a car, the reflection in the car window might show the street behind you.
  • The Reality: The AI can read the reflection, even if it's upside down or blurry, and figure out where you are. This is like a stalker looking at a reflection in a window to find a house number.

5. Why Can't We Just "Blur" the Photo?

The researchers tried common defenses, like blurring the background or adding "noise" (static) to the image, similar to how we blur license plates.

  • The Result: It didn't work well.
  • The Analogy: It's like trying to hide a face by putting a smudge of paint on a photo. The AI is so smart that it can look at the shape of the smudge, the color of the wall behind it, or the angle of the light to figure out what's underneath. If you blur the street sign, the AI just looks at the type of trash can or the style of the sidewalk instead.

The Big Takeaway

We are entering an era where every photo you take is a potential puzzle for an AI to solve.

  • The Good News: We now know this is happening.
  • The Bad News: The current "safety guards" on these AIs are blind to this specific type of privacy leak. They are trained to be helpful, so if you ask "Where is this?", they will happily tell you, even if it's your own home.

What should you do?
Think twice before posting photos with unique backgrounds. Remember that in the eyes of a super-smart AI, your "cozy living room" isn't just a room; it's a collection of clues that can lead someone straight to your front door.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →