Omnidirectional 3D Scene Reconstruction from Single Image

Ren Yang, Jiahao Li, Yan Lu
Microsoft Research
NeurIPS 2025

Abstract

Reconstruction of 3D scenes from a single image is a crucial step towards enabling next-generation AI-powered immersive experiences. However, existing diffusion-based methods often struggle with reconstructing omnidirectional scenes due to geometric distortions and inconsistencies across the generated novel views, hindering accurate 3D recovery. To overcome this challenge, we propose Omni3D, an approach designed to enhance the geometric fidelity of diffusion-generated views for robust omnidirectional reconstruction. Our method leverages priors from pose estimation techniques, such as MASt3R, to iteratively refine both the generated novel views and their estimated camera poses. Specifically, we minimize the 3D reprojection errors between paired views to optimize the generated images, and simultaneously, correct the pose estimation based on the refined views. This synergistic optimization process yields geometrically consistent views and accurate poses, which are then used to build an explicit 3D Gaussian Splatting representation capable of omnidirectional rendering. Experimental results validate the effectiveness of Omni3D, demonstrating significantly advanced 3D reconstruction quality in the omnidirectional space, compared to previous state-of-the-art methods.

BibTeX

@article{Yang2025Omni3D,
  title={Omnidirectional 3D Scene Reconstruction from Single Image},
  author={Ren Yang, Jiahao Li and Yan Lu},
  journal={the Thirty-Ninth Annual Conference on Neural Information Processing Systems (NeurIPS)},
  year={2025},
}