Joint Optimization of Model Partitioning and Resource Allocation for Anti-Jamming Collaborative Inference Systems

This paper proposes a joint optimization framework for anti-jamming collaborative DNN inference that maximizes a revenue metric of delay and accuracy by simultaneously determining model partitioning, resource allocation, and transmit power through an efficient alternating optimization algorithm.

Mengru Wu, Jiawei Li, Jiaqi Wei, Bin Lyu, Kai-Kit Wong, Hyundong Shin

Published 2026-03-04
📖 5 min read🧠 Deep dive

Imagine you have a very smart, but very tired, robot assistant (the Device) that needs to solve a complex puzzle, like identifying a car in a blurry photo. The robot is great at seeing the picture, but its brain is too small to finish the whole puzzle.

So, the robot calls for help from a super-computer in the cloud (the Edge Server). They decide to split the work: the robot does the first few steps, then sends a "progress report" (the Intermediate Data) over the air to the super-computer, which finishes the rest.

The Problem: The Mean Neighbor with a Megaphone
Now, imagine a mischievous neighbor (the Jammer) standing between the robot and the super-computer. This neighbor is shouting loudly (sending Jamming signals) to drown out the robot's progress report.

  • If the report gets garbled by the shouting, the super-computer gets confused and makes mistakes.
  • If the robot tries to shout louder to be heard, it drains its battery quickly.
  • If the robot tries to do more work itself to avoid sending a report, it gets exhausted and slow.

The goal of this paper is to figure out the perfect balance: How much work should the robot do? How loud should it shout? And how much help should the super-computer give? All while trying to finish the puzzle quickly, accurately, and without running out of battery, even while that mean neighbor is shouting.

The Solution: A Three-Part Dance

The authors propose a smart strategy called "Joint Optimization." Think of it as a dance where three partners adjust their moves simultaneously to avoid the shouting neighbor.

1. The "Split Point" (Where to cut the work)
The robot and the super-computer have to decide exactly where to split the puzzle.

  • Analogy: Imagine a relay race. Do you run the first 100 meters and hand off the baton? Or do you run 200 meters?
  • If you hand off too early, the baton (the data) is huge and hard to send through the shouting.
  • If you run too far, you get tired.
  • The paper uses a special mathematical "recipe" (Data Regression) to predict exactly how much the shouting will mess up the message based on where they split the work.

2. The "Volume Control" (Transmit Power)
The robot needs to decide how loud to shout.

  • Analogy: If the neighbor is shouting softly, you can whisper. If the neighbor is screaming, you have to yell. But if you yell too hard, you lose your voice (battery).
  • The system calculates the perfect volume to be heard clearly without wasting energy.

3. The "Helper's Speed" (Resource Allocation)
The super-computer needs to decide how fast to work on the second half of the puzzle.

  • Analogy: If the super-computer has many other people to help, it can't work too fast on just one robot's puzzle. It has to share its speed fairly.

How They Solve It: The "Quantum Genetic Algorithm"

Solving this puzzle is incredibly hard because there are millions of combinations of "where to split," "how loud to shout," and "how fast to work." It's like trying to find the perfect combination on a lock with a billion dials.

To solve this, the authors use a clever computer trick called an Alternating Optimization algorithm.

  • Step 1: They fix the volume and the split point, then figure out the best speed for the super-computer.
  • Step 2: They fix the speed and the split point, then figure out the best volume for the robot.
  • Step 3: They fix the volume and speed, then use a "Quantum Genetic Algorithm" to find the best split point.

What is a Quantum Genetic Algorithm?
Think of this as a super-smart evolution simulator.

  • Imagine a population of 100 different robots, each trying a different way to split the puzzle.
  • The "fittest" robots (those that finish fast and accurately) get to "reproduce" and mix their strategies.
  • The "Quantum" part is like giving these robots a superpower: they can try many different strategies at the same time (like being in two places at once) before picking the best one. This helps them find the perfect solution much faster than a normal computer could.

The Results: Why It Matters

The authors tested their idea in a simulation (a video game world) with 10 robots and a mean shouting neighbor.

  • The Winner: Their new strategy beat all the old ways of doing things.
  • The Trade-off: They created a score called RDA (Revenue of Delay and Accuracy). It's like a grade that combines "How fast did you finish?" and "How correct was the answer?"
  • The Result: Even when the neighbor was shouting very loudly, their system kept the robots working efficiently. The robots didn't run out of battery, and the super-computer didn't get confused.

In a Nutshell

This paper teaches us how to keep a team of robots and super-computers working together efficiently, even when someone is trying to jam their communication. By mathematically figuring out exactly where to split the work, how loud to talk, and how fast to compute, we can build smarter, more resilient systems for the future of 6G networks. It's about teamwork, smart planning, and not letting the noise win.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →