SelfSplat: Pose-Free and 3D Prior-Free Generalizable 3D Gaussian Splatting
Abstract
We propose SelfSplat, a novel 3D Gaussian Splatting model designed to perform pose-free and 3D prior-free generalizable 3D reconstruction from unposed multi-view images. These settings are inherently ill-posed due to the lack of ground-truth data, learned geometric information, and the need to achieve accurate 3D reconstruction without finetuning, making it difficult for conventional methods to achieve high-quality results. Our model addresses these challenges by effectively integrating explicit 3D representations with self-supervised depth and pose estimation techniques, resulting in reciprocal improvements in both pose accuracy and 3D reconstruction quality. Furthermore, we incorporate a matching-aware pose estimation network and a depth refinement module to enhance geometry consistency across views, ensuring more accurate and stable 3D reconstructions. To present the performance of our method, we evaluated it on large-scale real-world datasets, including RealEstate10K, ACID, and DL3DV. SelfSplat achieves superior results over previous state-of-the-art methods in both appearance and geometry quality, also demonstrates strong cross-dataset generalization capabilities. Extensive ablation studies and analysis also validate the effectiveness of our proposed methods. Code and pretrained models are available at https://gynjn.github.io/selfsplat/
Community
We are excited to share our recent work "SelfSplat:Pose-Free and 3D Prior-Free Generalizable 3D Gaussian Splatting".
TL;DR: We present SelfSplat, enabling 3D reconstruction from unposed images without any 3D priors.
We utilized 3D Gaussian Splatting and unsupervised depth estimation framework to result in reciprocal improvements in both pose accuracy and 3D reconstruction quality. Furthermore, we incorporate a matching-aware pose estimation network and a depth refinement module to enhance geometry consistency across views, ensuring more accurate and stable 3D reconstructions.
Paper: https://arxiv.org/abs/2411.17190
Project Page: https://gynjn.github.io/selfsplat/
Code: https://github.com/Gynjn/selfsplat
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- PF3plat: Pose-Free Feed-Forward 3D Gaussian Splatting (2024)
- No Pose, No Problem: Surprisingly Simple 3D Gaussian Splats from Sparse Unposed Images (2024)
- Epipolar-Free 3D Gaussian Splatting for Generalizable Novel View Synthesis (2024)
- PreF3R: Pose-Free Feed-Forward 3D Gaussian Splatting from Variable-length Image Sequence (2024)
- Large Spatial Model: End-to-end Unposed Images to Semantic 3D (2024)
- GPS-Gaussian+: Generalizable Pixel-wise 3D Gaussian Splatting for Real-Time Human-Scene Rendering from Sparse Views (2024)
- DepthSplat: Connecting Gaussian Splatting and Depth (2024)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper