LF2L: Loss Fusion Horizontal Federated Learning Across Heterogeneous Feature Spaces Using External Datasets Effectively: A Case Study in Second Primary Cancer Prediction

This paper proposes a Loss Fusion Horizontal Federated Learning (LF2L) framework that effectively integrates heterogeneous external SEER data with local Taiwanese hospital records to significantly improve the prediction accuracy of second primary lung cancer while preserving patient privacy and addressing feature inconsistencies.

Chia-Fu Lin, Yi-Ju Tseng

Published 2026-03-10
📖 5 min read🧠 Deep dive

Imagine you are a detective trying to solve a very tricky case: predicting when a cancer survivor might develop a completely new, second type of cancer.

This is a life-or-death puzzle. The more data you have, the better your detective work becomes. However, in this story, the detective (the researchers) faces two major problems:

  1. The "Small Notebook" Problem: The local hospital in Taiwan has a notebook with only about 10,000 patient stories. It's a good start, but not enough to see the big picture.
  2. The "Foreign Language" Problem: There is a massive library of patient stories in the US (called SEER) with over 85,000 records. But, the US library uses different categories and labels than the Taiwan notebook. If you try to glue the two notebooks together, the pages don't match up, and the information gets messy or lost.

The Old Ways (And Why They Failed)

The researchers tried a few standard approaches, but they hit dead ends:

  • The "Local Only" Approach: They just used the Taiwan notebook.
    • Analogy: It's like trying to learn to cook a complex dish by only tasting one spoonful of soup. You might get the basic flavor, but you'll miss the subtle spices that make it perfect. The model wasn't smart enough because it didn't have enough examples.
  • The "Naive Merge" Approach: They tried to force the US and Taiwan data into one giant pile, filling in the missing blanks with "unknown."
    • Analogy: Imagine trying to build a house by smashing two different blueprints together. One blueprint says "put a window here," and the other says "put a door there." If you just smash them, you end up with a wall that has a window and a door in the same spot, or a hole where nothing should be. The data gets confused, and the house (the model) becomes unstable.
  • The "Standard Teamwork" Approach (Federated Learning): They tried to let the two hospitals "talk" to each other without sharing the actual patient files.
    • Analogy: This is like two chefs trying to cook a meal together over a video call, but they can only agree on ingredients they both have. If the Taiwan chef has a special spice (like a specific gene mutation) and the US chef doesn't, they have to throw that spice away. They end up cooking a bland meal because they ignored the unique, powerful ingredients.

The New Solution: LF2L (The "Loss Fusion" Framework)

The researchers invented a clever new method called LF2L. Think of it as a Master Chef and an Apprentice working together in a way that respects their different kitchens.

Here is how it works, step-by-step:

  1. The "Common Ground" Chat (Federated Learning):
    First, the two hospitals look at the features they both have (like age, gender, basic blood work). They use these common features to train a "Global Brain."

    • Analogy: The two chefs agree on the basic recipe steps they both know. They create a shared "flavor profile" of the soup.
  2. The "Secret Sauce" (Local Features):
    Instead of throwing away the unique ingredients (like the special Taiwan gene mutations), the local hospital keeps them in its own private kitchen. It trains a "Local Brain" using its own unique data plus the "flavor profile" it got from the Global Brain.

    • Analogy: The Taiwan chef takes the shared flavor profile and adds their secret, special spice. They don't need to tell the US chef what the spice is; they just use it to improve their own dish.
  3. The "Loss Fusion" (The Magic Glue):
    This is the secret sauce of the method. The system uses a special "scorecard" (called Loss) to measure how good the predictions are. It combines the score from the Global Brain and the Local Brain.

    • Analogy: Imagine a judge tasting the soup. The judge gives a score based on how well the soup tastes overall. If the Global Brain says, "This needs more salt," and the Local Brain says, "But I added my special spice, so it needs less," the system learns how to balance them perfectly. It doesn't force them to be the same; it teaches them how to work together to get the best result.

The Result: A Better Prediction

By using this method, the researchers didn't have to throw away any data. They got the size of the US dataset (more examples) and the specificity of the Taiwan dataset (unique medical details).

  • The Outcome: The new model was significantly better at predicting second cancers than any of the old methods. It was more accurate, more reliable, and didn't violate patient privacy (because the actual patient data never left the local hospitals).

The Big Takeaway

This paper teaches us that in the world of medical AI, you don't have to choose between having a lot of data or having the right data.

Instead of forcing everyone to speak the same language (which loses information), we can build a system where different languages are translated into a shared "feeling" or "understanding," allowing everyone to contribute their unique strengths to solve the problem together. It's like a global orchestra where every musician plays a different instrument, but they all follow the same conductor to create a beautiful symphony.