Volumetric scene reconstruction from a single image is crucial for a broad range of applications like autonomous driving and robotics. Recent volumetric reconstruction methods achieve impressive results, but generally require expensive 3D ground truth or multi–view supervision. We propose to leverage pre–trained 2D diffusion models and depth prediction models to generate synthetic scene geometry from a single image. This can then be used to distill a feed–forward scene reconstruction model. Our experiments on the challenging KITTI–360 and Waymo datasets demonstrate that our method matches or outperforms state–of–the–art baselines that use multi–view supervision, and offers unique advantages, for example regarding dynamic scenes. For more details and code, please check out our project page.
inproceedings WWM+25
SE3 Labs
BibTeXKey: WWM+25