Introducing LaRa: The Future of 360° Radiance Field Reconstruction! 🌟
LaRa, a groundbreaking feed-forward model, significantly advances 3D reconstruction by unifying local and global reasoning in transformer layers. This innovative approach leverages Gaussian Volumes, an image encoder, and Group Attention Layers to achieve high-quality and rapid 360° radiance field reconstructions. Unlike traditional methods constrained by small baselines, LaRa excels in handling large baselines, offering high fidelity and robustness even in zero-shot and out-of-domain testing scenarios.
LaRa addresses the core challenge of reconstructing objects' shape and appearance from multi-view images, a critical task in computer vision and graphics. By overcoming limitations associated with dense image captures and small camera baselines, LaRa introduces a novel volume transformer. This method progressively and implicitly performs feature matching, efficiently aggregating local and global features to produce photorealistic renderings from just four input images.
LaRa demonstrates exceptional performance, achieving high-resolution 360° novel view synthesis with minimal training resources—only four GPUs over two days. The model's efficiency is enhanced by a coarse-to-fine decoding process and efficient rasterization, enabling high-quality mesh reconstruction with off-the-shelf algorithms. With its robust zero-shot generalization and ability to handle diverse viewpoints, LaRa sets a new standard in 3D reconstruction, pushing the boundaries of what's possible in visual effects, e-commerce, virtual and augmented reality, and robotics.
"LaRa: Efficient Large-Baseline Radiance Fields"
Anpei Chen, Haofei Xu, Stefano Esposito, Siyu Tang, Andreas Geiger
Github Repo: https://lnkd.in/gJ78NjqB
Arxiv Preprint: https://lnkd.in/g5nrJsuK
#Transformers #FewShot #3DReconstruction #ComputerVision #AiResearch
Teaching Assistant @ University of Florida | Construction Management PhD
8moUniversity of Florida College of Design, Construction and Planning