ProxyFL: A Proxy-Guided Framework for Federated Semi-Supervised Learning

This paper proposes ProxyFL, a unified proxy-guided framework that simultaneously mitigates external and internal data heterogeneity in Federated Semi-Supervised Learning by optimizing global proxies against outliers and re-incorporating discarded samples via a positive-negative proxy pool.

Duowen Chen, Yan Wang

Published 2026-02-25
📖 4 min read☕ Coffee break read

Imagine a group of detectives (the clients) trying to solve a massive mystery together, but they can't share their secret notebooks (the data) because of privacy rules. They have to send only their theories and conclusions (the model updates) to a central headquarters (the server).

This is Federated Learning. But here's the twist: most detectives only have a few confirmed clues (labeled data) and a huge pile of scribbled notes with no answers (unlabeled data). This is Federated Semi-Supervised Learning (FSSL).

The problem? The detectives are all working in different neighborhoods with different types of crimes (External Heterogeneity), and within their own offices, the confirmed clues don't match the messy scribbles (Internal Heterogeneity).

The paper introduces a new method called ProxyFL to fix this. Here is how it works, using simple analogies:

1. The Old Way: "The Average Vote" (And Why It Fails)

Usually, the headquarters tries to create a "Master Detective" by simply averaging everyone's theories.

  • The Problem: If one detective is an outlier (maybe they only see cat burglaries while everyone else sees bank robberies), their weird theory drags the average off course.
  • The Result: The Master Detective becomes confused and bad at solving crimes.
  • The "Low-Confidence" Issue: To avoid mistakes, detectives usually throw away their messy, uncertain notes. This means they are ignoring a huge amount of potential evidence, making the team smaller and weaker.

2. The New Way: ProxyFL (The "Mental Map" Strategy)

ProxyFL introduces a clever trick: instead of just averaging theories, they create a shared "Mental Map" of categories (called Proxies). Think of these Proxies as the "ideal definition" of a Cat, a Dog, or a Bank Robbery.

Part A: Fixing the "Different Neighborhoods" (External Heterogeneity)

The Old Way: Just take the average of everyone's definition of a "Cat." If one detective thinks a "Cat" looks like a "Hamster," the average becomes a weird "Hamster-Cat."
The ProxyFL Way: The headquarters creates a Global Mental Map. Instead of blindly averaging, it actively adjusts this map to fit the reality of all the detectives, ignoring the weird outliers.

  • Analogy: Imagine a teacher drawing a map of "What a Cat looks like." If one student draws a cat with wings, the teacher doesn't just average the drawings. The teacher looks at all the drawings, sees the wings are an outlier, and draws a perfect, balanced cat that represents the group's true understanding. This map is sent back to everyone to align their thinking.

Part B: Fixing the "Messy Notes" (Internal Heterogeneity)

The Old Way: If a detective isn't 100% sure if a note says "Hamster" or "Mouse," they throw the note in the trash to be safe.
The ProxyFL Way: ProxyFL says, "Don't throw it away! Let's keep it, but be smart about it."

  • The "Indecisive Categories" Trick: Instead of forcing a guess like "It's a Mouse," the system says, "Okay, this note is confusing. It could be a Mouse OR a Hamster."
  • The "Positive-Negative Pool": The system creates a special training zone.
    • Positive: It tells the model, "This note is close to being a Mouse or a Hamster."
    • Negative: It tells the model, "This note is definitely not a Dog or a Car."
  • Analogy: Imagine a student studying for a test. Instead of skipping the hard questions (low-confidence samples), they write down, "This answer is likely A or B, but definitely not C or D." This helps them learn the boundaries of the answers without getting scared of making a mistake.

Why is this a Big Deal?

  1. No Privacy Leaks: The "Mental Map" (Proxies) is just a tiny part of the model's brain, not the actual data. It's like sharing a summary of your thoughts rather than your diary.
  2. More Data, Less Waste: By keeping the "messy notes" and treating them as "maybe this, maybe that," the team uses all the available evidence, not just the easy stuff.
  3. Faster Learning: Because the "Master Detective" (Global Model) has a clearer map and more practice data, the whole team learns much faster and solves the mystery more accurately.

The Bottom Line

ProxyFL is like a smart team leader who:

  1. Refines the group's shared dictionary so everyone agrees on what things look like, even if they come from different places.
  2. Encourages the team to use their uncertain notes by treating them as "possible options" rather than discarding them.

The result? A super-smart, privacy-safe AI that learns faster and better, even when the data is messy and scattered across many different devices.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →