From Volume Rendering to 3D Gaussian Splatting: Theory and Applications

This tutorial provides a comprehensive overview of 3D Gaussian Splatting, detailing its theoretical foundations, addressing key limitations such as memory footprint and lighting baking, and surveying its diverse applications in surface reconstruction, avatar modeling, and content generation.

Vitor Pereira Matias, Daniel Perazzo, Vinicius Silva, Alberto Raposo, Luiz Velho, Afonso Paiva, Tiago Novello

Published 2026-03-02
📖 6 min read🧠 Deep dive

Imagine you want to create a perfect, 3D hologram of a room just by taking a bunch of photos of it from different angles. For a long time, computers tried to do this using "Neural Radiance Fields" (NeRFs). Think of NeRFs like a giant, invisible fog that fills the entire room. To figure out what the room looks like from a new angle, the computer has to ask every single tiny speck of that fog, "Are you there? What color are you?" This is incredibly accurate but also incredibly slow and heavy, like trying to count every grain of sand on a beach just to see a picture of the shore.

Then came 3D Gaussian Splatting (3DGS), the star of this paper. Instead of a foggy cloud, 3DGS builds the scene out of millions of tiny, colorful, fuzzy balloons (Gaussians).

Here is the simple breakdown of how this technology works, why it's a game-changer, and where it's going next, using some everyday analogies.

1. The Core Idea: From Fog to Fuzzy Balloons

In the old way (NeRF), the computer had to simulate light traveling through a continuous fog. It was like trying to paint a picture by mixing every possible shade of paint in a bucket and then guessing which shade goes where.

In 3DGS, the computer says, "Let's just use balloons!"

  • The Setup: You start with a few photos. The computer finds the key points (like the corners of a table or the tip of a nose) and places a "balloon" there.
  • The Balloon: Each balloon isn't just a solid sphere; it's a fuzzy, transparent cloud with a specific color, size, and shape. Some are flat like pancakes, some are long like sausages, and some are round like marbles.
  • The Magic (Splatting): Instead of asking every point in the room, the computer just "splats" (throws) these balloons onto a 2D screen. Because they are fuzzy, they overlap and blend together naturally to form a solid-looking image. It's like throwing thousands of wet paint blobs at a canvas; from a distance, they look like a perfect painting, but up close, you see the individual blobs.

Why is this better?

  • Speed: It's like using a paint roller (splatting) instead of painting every single pixel by hand (fog simulation). It renders in real-time, meaning you can walk around the 3D scene instantly without waiting.
  • Efficiency: It only puts balloons where things actually exist. It doesn't waste time calculating the empty air in the middle of the room.

2. The Training Process: The Sculptor's Workshop

How does the computer learn where to put these balloons?

  1. Initialization: It starts with a rough cloud of points (like a wireframe) and puts a balloon on every point.
  2. The Critique: It looks at the photo it just made and compares it to the real photo you took. "Hmm, this shadow is too dark," or "This edge is too blurry."
  3. The Fix (Adaptation): The computer acts like a sculptor with a magical tool:
    • If a balloon is too big and blurry, it splits it into two smaller ones.
    • If a detail is missing, it clones a balloon and moves it closer.
    • If a balloon is useless (too transparent), it prunes (removes) it.
  4. Repeat: It does this thousands of times until the 3D scene looks exactly like the photos.

3. The Problems: The Balloon House is Heavy

While 3DGS is fast and looks great, the paper points out a few "growing pains":

  • Memory Hog: To make a complex scene look perfect, you might need hundreds of thousands of balloons. This is like trying to store a library of books in a tiny backpack; it takes up a lot of space on your hard drive.
  • Baked-in Lighting: Currently, the balloons "bake" the lighting into their color. If you take a photo in the sun, the balloons are painted yellow. If you try to move the scene to a dark room, the balloons stay yellow. They don't know how to react to new lights (like a real object would).
  • No Reflections: Because the balloons are just "fuzzy clouds," they struggle to show complex reflections or see-through glass, which usually require light to bounce around (secondary rays).

4. The Future: Making the Balloons Smarter

The second half of the paper is a tour of how researchers are fixing these issues and using 3DGS for cool new things:

  • Making them smaller: New methods are teaching the balloons to be more efficient, using fewer of them to get the same quality (like using high-quality paint instead of a million blobs).
  • Adding Physics: Researchers are now giving the balloons "physics brains." They can simulate water flowing, cloth waving, or a character jumping. It's like turning the static balloons into a simulated fluid or solid object.
  • Creating Avatars: You can now build 3D humans that look real and can move. By attaching balloons to a digital skeleton (like a digital mannequin), you can animate a person's face or body instantly.
  • From Text to 3D: Imagine typing "a cat wearing a hat" and having the computer instantly generate a 3D scene of it using these balloons. New AI models are learning to do this, turning text or a single photo into a full 3D world.
  • Fixing the "Wild": Real-world photos are messy (people walking by, changing light). New versions of 3DGS are getting better at ignoring these distractions and building a clean 3D model even from messy, casual photos.

The Bottom Line

3D Gaussian Splatting is a revolution in how we turn 2D photos into 3D worlds. It swapped the slow, heavy "fog" of the past for a fast, colorful "cloud of balloons."

  • The Good: It's incredibly fast, looks amazing, and is easy to use in video games and VR.
  • The Challenge: It uses a lot of memory and doesn't handle complex lighting (like reflections) perfectly yet.
  • The Future: We are moving toward making these balloons smaller, smarter, and capable of simulating real-world physics, allowing us to generate entire 3D movies or interactive worlds from just a few photos or even a text description.

Think of it as the difference between building a house out of heavy, solid concrete blocks (old methods) versus building it out of millions of lightweight, self-assembling LEGO bricks that can be rearranged instantly (3DGS). It's faster, more flexible, and the future of 3D creation.

Get papers like this in your inbox

Personalized daily or weekly digests matching your interests. Gists or technical summaries, in your language.

Try Digest →