Dropping Anchor and Spherical Harmonics for Sparse-view Gaussian Splatting

The paper proposes DropAnSH-GS, a novel sparse-view 3D Gaussian Splatting method that mitigates overfitting by simultaneously dropping anchor Gaussians and their spatial neighbors to disrupt local redundancy, while also randomly truncating high-degree spherical harmonic coefficients to concentrate appearance information and enable model compression.

Shuangkang Fang, I-Chao Shen, Xuanyang Zhang, Zesheng Wang, Yufeng Wang, Wenrui Ding, Gang Yu, Takeo Igarashi

Published 2026-02-25
📖 4 min read☕ Coffee break read

Imagine you are trying to paint a beautiful landscape, but you only have three blurry photos to guide you. This is the challenge of "Sparse-View 3D Gaussian Splatting." The computer tries to build a 3D world from very few pictures.

The problem? The computer gets too confident. It starts "memorizing" the few photos it has instead of learning the actual shape of the world. This is called overfitting. It's like a student who memorizes the answers to three practice tests but fails the real exam because they didn't understand the concepts.

To fix this, previous methods tried a technique called Dropout. Imagine the computer is a choir of thousands of singers (called "Gaussians"). To stop them from memorizing, the conductor randomly tells some singers to go silent. The idea is that the remaining singers must work harder to fill the gaps, learning the song better.

But here's the catch: In this 3D world, the singers stand right next to each other and sing the exact same note. If you silence one, the neighbor immediately sings louder to cover the gap. The silence is never really felt, so the singers don't learn anything new. They just keep memorizing.

The paper "DropAnSH-GS" introduces a smarter way to silence the choir. Here is how they did it, using simple analogies:

1. The "Anchor and Neighborhood" Strategy (Dropping Anchors)

Instead of silencing one random singer, the new method picks a "Leader" (an Anchor) and silences the Leader plus their entire neighborhood.

  • The Analogy: Imagine a crowded room where everyone is whispering the same secret. If you tell one person to stop talking, the person next to them just whispers it louder. But if you tell a whole group of friends standing in a circle to stop talking, you create a big silence.
  • The Result: The remaining singers (Gaussians) can't just lean on their neighbors to fill the gap. They are forced to listen to people far away and learn the whole song structure, not just the local whisper. This forces the computer to build a more robust, accurate 3D model.

2. The "Color Palette" Strategy (Dropping Spherical Harmonics)

The 3D models also use "Spherical Harmonics" (SH) to describe colors. Think of SH as a set of paintbrushes:

  • Low-degree brushes: Big, broad strokes for basic colors (e.g., "sky is blue").
  • High-degree brushes: Tiny, fine-detail brushes for complex patterns (e.g., "the exact texture of a leaf").

In sparse-view training, the computer gets obsessed with the tiny detail brushes. It tries to memorize every tiny speck of dust in the few photos, which ruins the overall picture.

  • The Solution: The new method randomly throws away the fine-detail brushes during training.
  • The Result: The computer is forced to learn the scene using only the big, broad strokes first. It learns the "skeleton" of the color before worrying about the "flesh."
  • The Bonus: Because the computer learned to rely on the big strokes, you can later throw away the fine-detail brushes permanently without losing much quality. This makes the final 3D model much smaller and faster to load, like compressing a high-res photo into a smaller file without it looking blurry.

Why This Matters

  • Better Quality: The 3D scenes look sharper and have fewer weird artifacts (glitches) when viewed from new angles.
  • Smaller Files: You can shrink the file size significantly (sometimes by 75%!) just by cutting off the "fine detail" math after training, and the image still looks great.
  • Fast & Easy: It doesn't slow down the computer much; it just changes how the computer learns.

In summary: The old way was like telling one person in a crowd to be quiet (and they got covered up). The new way is like clearing out a whole block of the city, forcing the remaining people to communicate across the whole town. Plus, it teaches the computer to focus on the "big picture" colors first, making the final result both higher quality and smaller in size.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →