Diffusion Based 3DGS Relighting

Michael Rubloff

Michael Rubloff

Jul 2, 2024

Email
Copy Link
Twitter
Linkedin
Reddit
Whatsapp
Diffusion Radiance Field
Diffusion Radiance Field

It was only recently where I was discussing how improving reflections in captures were an active topic of research. I have steadily watched paper after paper targeting relighting pass over. Starting with IllumiNeRF, then LE3D, to DiLightNet, to Neural Gaffer, and most recently GS-ROR, relighting as a topic is truly on fire.

Today we're looking at another relighting method that's using a diffusion prior to relight Radiance Fields. They use 3DGS because of their upfront explicit representation advantage, but this method can be adopted by NeRFs or any subsequent Radiance Field method.

The first step in their method is the creation of a 2D relighting neural network with explicit control over lighting direction. This was achieved by fine-tuning a pre-trained Stable Diffusion (SD) model using a multi-illumination dataset. The dataset, which contained images of various scenes captured under different lighting conditions, enabled the network to generate realistic relit versions of each image. The fine-tuning process involved using ControlNet. By leveraging depth maps and spherical harmonics encoding of lighting directions, the network could accurately control the light direction during the relighting process.

Once the 2D relighting network is established, it's used to augment a multi-view dataset captured under a single lighting condition. This involves generating multiple relit versions of each image in the dataset, effectively transforming the original single-illumination dataset into a comprehensive multi-illumination dataset. This augmentation is crucial as it provides the necessary data diversity to train a robust relightable radiance field.

The final step in the method trains a relightable radiance field using 3D Gaussian Splatting (3DGS). This representation is enhanced with a small Multi-Layer Perceptron (MLP), which replaces the spherical harmonic coefficients, and an auxiliary feature vector to address inconsistencies and inaccuracies in the synthesized relit images. The MLP allowed for direct control of the lighting direction, enabling realistic and interactive relighting of full scenes in real time. By optimizing a per-image auxiliary latent vector, the method ensured multi-view consistency, effectively handling the inherent challenges of varying illumination across different views. The training process started with the unlit images to establish a baseline radiance field, followed by integrating the multi-illumination data to develop a robust model capable of interactive and realistic relighting.

To improve the quality of the relit images, the researchers adopted additional strategies to enhance brightness, contrast, and edge sharpness. These included adjusting the denoising objectives and performing color matching between the input and predicted images to maintain consistency in color and brightness across different lighting conditions. Additionally, to ensure the temporal consistency of the generated relighting, the auxiliary latent vectors were optimized to correct for discrepancies in lighting directions, providing smooth transitions across different views.

Once the training is complete, they are able to render it with relighting capabilities, still at 30 fps.

The ControlNet method they're using specializes in indoor scenes, and this appears to be a result of the research focus rather than an addressable issue. One other limitation is that ControlNet does not perform physics-based reasoning, meaning that occasionally shadows might appear in places they shouldn't be and vice versa. There's still a bit of work to be done on that front, but it seems by no means insurmountable.

In total, it runs for 30K steps, with the first 5K dedicated to 3DGS and the subsequent 25K training the full multi-illumination solution. They are able to run this with similar VRAM consumption compared to the original 3DGS, meaning that it can be done on consumer-grade hardware. The code has not been published yet, but with the paper's acceptance into EGSR 2024, we might be seeing it sooner than later! Their project page can be found here for more information.

Featured

Recents

Featured

Platforms

Postshot V.5 Released

The newest version of Postshot is here!

Michael Rubloff

Dec 23, 2024

Platforms

Postshot V.5 Released

The newest version of Postshot is here!

Michael Rubloff

Dec 23, 2024

Platforms

Postshot V.5 Released

The newest version of Postshot is here!

Michael Rubloff

News

Create a Hyper-Real Holiday Card

Just in time for the holidays comes a way to share hyper real holiday wishes!

Michael Rubloff

Dec 23, 2024

News

Create a Hyper-Real Holiday Card

Just in time for the holidays comes a way to share hyper real holiday wishes!

Michael Rubloff

Dec 23, 2024

News

Create a Hyper-Real Holiday Card

Just in time for the holidays comes a way to share hyper real holiday wishes!

Michael Rubloff

Platforms

GSOPs 2.0: Now Commercially Viable with Houdini Commercial License

The 2.0 release for GSOPs is here with a commercial license!

Michael Rubloff

Dec 20, 2024

Platforms

GSOPs 2.0: Now Commercially Viable with Houdini Commercial License

The 2.0 release for GSOPs is here with a commercial license!

Michael Rubloff

Dec 20, 2024

Platforms

GSOPs 2.0: Now Commercially Viable with Houdini Commercial License

The 2.0 release for GSOPs is here with a commercial license!

Michael Rubloff

Platforms

Odyssey Announces Generative World Model, Explorer

Odyssey shows off their photo real world generator, powered by Radiance Fields.

Michael Rubloff

Dec 18, 2024

Platforms

Odyssey Announces Generative World Model, Explorer

Odyssey shows off their photo real world generator, powered by Radiance Fields.

Michael Rubloff

Dec 18, 2024

Platforms

Odyssey Announces Generative World Model, Explorer

Odyssey shows off their photo real world generator, powered by Radiance Fields.

Michael Rubloff