Gradient Radio
Gradient Radio
Breaking the Boundaries of Virtual Reality with Dual Gaussian Splatting
0:00
-6:11

Breaking the Boundaries of Virtual Reality with Dual Gaussian Splatting

A Deep Dive into the Future of Immersive Video Technology

Paper: Robust Dual Gaussian Splatting for Immersive Human-centric Volumetric Videos

Stepping into Reality: How Dual Gaussian Splatting is Revolutionizing Immersive Experiences

·
September 17, 2024
Stepping into Reality: How Dual Gaussian Splatting is Revolutionizing Immersive Experiences

Imagine attending a live concert or a high-stakes sporting event—not merely watching from a screen, but virtually present within the venue, free to move and choose your perspective. Such immersive experiences, once confined to science fiction, are becoming increasingly plausible due to advancements in virtual reality technology. A recent development in …

In this episode of Gradient Radio, we explore the groundbreaking technique known as Dual Gaussian Splatting, a major leap forward in volumetric video technology. Discover how this innovation is transforming virtual reality experiences, enabling real-time, high-fidelity playback of immersive performances. Whether you’re a VR enthusiast, developer, or storyteller, this breakthrough is reshaping the future of digital and real-world integration. Tune in to learn how AI is bridging the gap between physical and virtual spaces.

Paper: Robust Dual Gaussian Splatting for Immersive Human-centric Volumetric Videos

Abstract: Volumetric video represents a transformative advancement in visual media, enabling users to freely navigate immersive virtual experiences and narrowing the gap between digital and real worlds. However, the need for extensive manual intervention to stabilize mesh sequences and the generation of excessively large assets in existing workflows impedes broader adoption. In this paper, we present a novel Gaussian-based approach, dubbed \textit{DualGS}, for real-time and high-fidelity playback of complex human performance with excellent compression ratios. Our key idea in DualGS is to separately represent motion and appearance using the corresponding skin and joint Gaussians. Such an explicit disentanglement can significantly reduce motion redundancy and enhance temporal coherence. We begin by initializing the DualGS and anchoring skin Gaussians to joint Gaussians at the first frame. Subsequently, we employ a coarse-to-fine training strategy for frame-by-frame human performance modeling. It includes a coarse alignment phase for overall motion prediction as well as a fine-grained optimization for robust tracking and high-fidelity rendering. To integrate volumetric video seamlessly into VR environments, we efficiently compress motion using entropy encoding and appearance using codec compression coupled with a persistent codebook. Our approach achieves a compression ratio of up to 120 times, only requiring approximately 350KB of storage per frame. We demonstrate the efficacy of our representation through photo-realistic, free-view experiences on VR headsets, enabling users to immersively watch musicians in performance and feel the rhythm of the notes at the performers' fingertips.

Discussion about this episode

User's avatar