Modular Delta Merging with Orthogonal Constraints: A Scalable Framework for Continual and Reversible Model Composition

This paper introduces Modular Delta Merging with Orthogonal Constraints (MDM-OC), a scalable framework that enables interference-free, reversible, and compliant composition of fine-tuned models by projecting task-specific deltas into orthogonal subspaces and merging them via gradient-based optimization.

Haris Khan, Sadia Asif, Shumaila Asif, Muhammad Zeeshan Karamat, Rajesh Upadhayaya

Published 2026-04-14
📖 4 min read☕ Coffee break read

Imagine you have a master chef, Chef Base, who knows how to cook a perfect, basic soup.

Now, imagine you want to teach this chef new recipes: one for spicy curry, one for a sweet dessert, and one for a vegan salad. In the old way of doing things (traditional machine learning), you'd send the chef to culinary school for each new dish. But there's a catch: every time they learn the curry, they start to forget how to make the soup. By the time they learn the salad, they might have forgotten the dessert entirely. This is called "Catastrophic Forgetting."

Alternatively, you could try to just mix the recipes together by averaging the ingredients. But if you mix a spicy curry recipe with a sweet dessert recipe, you end up with a gross, confusing mess that tastes like neither.

MDM-OC is a new, brilliant kitchen system that solves both problems. Here is how it works, broken down into simple concepts:

1. The "Delta" (The Change, Not the Whole)

Instead of rewriting the chef's entire brain for every new recipe, MDM-OC only writes down the changes (the "deltas").

  • Old Way: "Here is the full recipe for Curry."
  • MDM-OC Way: "Here is a small note: Add 2 extra chili peppers and swap the broth for coconut milk."
    This keeps things light and efficient.

2. The "Orthogonal" Trick (The Invisible Walls)

This is the magic part. In a normal kitchen, if you try to add "Spicy" and "Sweet" to the same pot, they clash.
MDM-OC builds invisible walls (orthogonal subspaces) between the recipes.

  • Imagine the "Spicy" recipe lives in a room on the Left.
  • The "Sweet" recipe lives in a room on the Right.
  • The "Vegan" recipe lives in a room Upstairs.

Because these rooms are perfectly perpendicular (orthogonal) to each other, adding chili peppers to the Left room cannot accidentally turn the dessert sweet or ruin the salad. They don't interfere with each other. The system mathematically forces these new skills to live in their own unique "directions" so they never bump into each other.

3. The "Merging" (The Unified Masterpiece)

When you want the chef to serve a full menu, you don't retrain them. You simply take the Base Chef and gently layer the "Change Notes" from the Left, Right, and Upstairs rooms on top of them.
Because the rooms are separate, the chef can now make the Curry, the Dessert, and the Salad perfectly at the same time, without forgetting the original soup.

4. The "Un-Merging" (The GDPR Superpower)

This is the most unique feature. Imagine a customer says, "I don't want the dessert recipe in my kitchen anymore because of privacy laws."

  • Old Systems: You have to scrub the chef's brain, which is messy and might accidentally delete the soup recipe too.
  • MDM-OC: Because the dessert recipe was stored in its own separate "Upstairs" room, you just take that specific note away. The chef instantly forgets the dessert but remembers everything else perfectly. It's like erasing a single line of code without breaking the whole program. This is crucial for laws like GDPR that require you to "delete" data or skills on demand.

5. The "Stability" (The Safety Net)

To make sure the chef doesn't get confused when switching between these rooms, the system uses two safety nets:

  • Elastic Weight Consolidation: This is like a "memory anchor." It tells the chef, "Don't change the core ingredients of the soup too much."
  • Synthetic Replay: This is like a "practice dummy." The chef occasionally practices the old recipes using fake ingredients to keep the muscle memory sharp.

Why Does This Matter?

  • No More Amnesia: The AI doesn't forget old tasks when learning new ones.
  • No More Messy Mixes: New skills don't ruin old ones.
  • Privacy & Compliance: You can surgically remove specific skills (like learning from a specific user's data) without rebuilding the whole AI.
  • Scalability: You can keep adding hundreds of new skills without the system getting slow or bloated.

In short: MDM-OC is like giving a super-intelligent assistant a set of modular, non-interfering toolkits. You can snap a new tool on to learn a new skill, or snap it off to forget it, all while keeping the original brain perfectly intact and the whole system running smoothly.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →