GaussianAnything: Interactive Point Cloud Latent Diffusion for 3D Generation
Abstract
While 3D content generation has advanced significantly, existing methods still face challenges with input formats, latent space design, and output representations. This paper introduces a novel 3D generation framework that addresses these challenges, offering scalable, high-quality 3D generation with an interactive Point Cloud-structured Latent space. Our framework employs a Variational Autoencoder (VAE) with multi-view posed RGB-D(epth)-N(ormal) renderings as input, using a unique latent space design that preserves 3D shape information, and incorporates a cascaded latent diffusion model for improved shape-texture disentanglement. The proposed method, GaussianAnything, supports multi-modal conditional 3D generation, allowing for point cloud, caption, and single/multi-view image inputs. Notably, the newly proposed latent space naturally enables geometry-texture disentanglement, thus allowing 3D-aware editing. Experimental results demonstrate the effectiveness of our approach on multiple datasets, outperforming existing methods in both text- and image-conditioned 3D generation.
Community
GaussianAnything generates high-quality and editable surfel Gaussians through a cascaded native 3D diffusion pipeline, given single-view images or texts as the conditions.
Hello there! I made the fish in your image example. It's under CC by so I would appreciate a credit. :)
Thanks for your fabulous 3D assets shared online for the public use! I will update the project page and paper accordingly to credit your asset in the later version ;)
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- 3D-Adapter: Geometry-Consistent Multi-View Diffusion for High-Quality 3D Generation (2024)
- L3DG: Latent 3D Gaussian Diffusion (2024)
- LucidFusion: Generating 3D Gaussians with Arbitrary Unposed Images (2024)
- LaGeM: A Large Geometry Model for 3D Representation Learning and Diffusion (2024)
- MvDrag3D: Drag-based Creative 3D Editing via Multi-view Generation-Reconstruction Priors (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper