Learning in the Null Space: Small Singular Values for Continual Learning

This paper introduces NESS, a continual learning method that mitigates catastrophic forgetting by constraining task-specific updates to an approximate null space derived from the smallest singular values of input representations, thereby enabling efficient adaptation while preserving performance on previous tasks.

Cuong Anh Pham, Praneeth Vepakomma, Samuel Horváth

Published 2026-02-26
📖 5 min read🧠 Deep dive

Imagine you are a student trying to learn a new language every week.

  • Week 1: You learn Spanish. You get great at it.
  • Week 2: You start learning French. But as you practice French, you start forgetting Spanish words.
  • Week 3: You learn Italian. Now, when you try to speak Spanish or French, you mix everything up or can't remember anything.

This is called Catastrophic Forgetting. It's the biggest problem in "Continual Learning" (teaching AI to learn new things without forgetting the old stuff).

The paper you shared introduces a clever new method called NESS (Null-space Estimated from Small Singular values) to solve this. Here is how it works, explained with simple analogies.

The Problem: The "Noisy Classroom"

Imagine your brain (the AI) is a classroom.

  • Old Knowledge (Spanish): The students are sitting in the front rows, chatting loudly. They occupy the "loud" space.
  • New Knowledge (French): You want to teach new students to sit in the back.
  • The Mistake: Most AI methods try to teach French by shouting over the Spanish students or rearranging the whole room. This causes chaos, and the Spanish students get confused (forgetting).

The Old Solution: "Gradient Projection" (The Bouncer)

Previous methods tried to fix this by acting like a strict bouncer. Every time the AI tried to learn something new, the bouncer would check: "Is this new idea going to bump into the old Spanish students?" If yes, the bouncer would physically push the new idea away (projecting the gradient) so it didn't hit the old students.

This works, but it's like constantly shoving people around. It's computationally expensive and can be clumsy.

The New Solution: NESS (The "Quiet Corner" Strategy)

The authors of this paper realized something brilliant: Not all parts of the classroom are loud.

If you look at the classroom, the front rows are loud (high energy, big movements). But there are corners, the ceiling, and the floor where no one is sitting. These are the "Quiet Corners" (mathematically, these are the directions with Small Singular Values).

NESS changes the strategy:
Instead of shoving the new students away from the old ones, NESS says: "Let's just teach the new students in the Quiet Corners where the old students aren't sitting."

Here is the step-by-step breakdown:

1. Mapping the Room (The SVD)

Before teaching a new task, the AI looks at all the data from the previous tasks (the Spanish students). It calculates a map of the room to find the "Quiet Corners."

  • Big Singular Values: These are the loud, crowded areas where the old knowledge lives.
  • Small Singular Values: These are the empty, quiet areas where no one is sitting.

2. Building the "Quiet Desk" (The Null Space)

NESS builds a special, tiny desk (a mathematical subspace) specifically in those quiet corners. It locks this desk in place so it never moves.

  • The Frozen Basis: The desk itself is fixed. It represents the "safe zone" where learning won't disturb the old knowledge.
  • The Trainable Matrix: The AI only learns what to write on this desk. It doesn't move the desk; it just fills it with new information.

3. Learning Without Interference

When the AI learns French (Task 2), it only writes on this "Quiet Desk."

  • Because the desk is in a corner where the Spanish students (Task 1) aren't sitting, the French lessons cannot accidentally erase the Spanish notes.
  • The old knowledge stays perfectly safe.
  • The new knowledge is learned efficiently.

Why is this a big deal?

  1. It's Stable: Because the "desk" is locked in the quiet corner, the AI never accidentally bumps into old memories. This means zero or near-zero forgetting.
  2. It's Efficient: The AI doesn't need to remember every single old example or calculate complex "bouncer" moves. It just learns on a small, fixed piece of paper.
  3. It Works: The paper tested this on image recognition tasks (like identifying cats, dogs, and cars). The results showed that NESS learned new things just as well as other methods but forgot much less. In fact, on some tests, learning the new task actually helped the AI remember the old one better (Positive Backward Transfer).

The Takeaway

Think of NESS as a smart librarian.

  • Old methods try to rearrange the whole library every time a new book arrives, hoping the old books don't get knocked over.
  • NESS finds the empty shelf in the back of the library that nobody uses, puts the new book there, and locks the door. The old books stay exactly where they are, safe and sound, while the new book gets its own perfect home.

By using the "quiet corners" (small singular values) of the data, NESS allows AI to learn continuously without the headache of forgetting everything it learned yesterday.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →