AnyRecon: Arbitrary-View 3D Reconstruction with Video Diffusion Model
Abstract
AnyRecon enables scalable 3D reconstruction from arbitrary sparse inputs using diffusion models with persistent scene memory and geometry-aware conditioning for improved geometric consistency.
Sparse-view 3D reconstruction is essential for modeling scenes from casual captures, but remain challenging for non-generative reconstruction. Existing diffusion-based approaches mitigates this issues by synthesizing novel views, but they often condition on only one or two capture frames, which restricts geometric consistency and limits scalability to large or diverse scenes. We propose AnyRecon, a scalable framework for reconstruction from arbitrary and unordered sparse inputs that preserves explicit geometric control while supporting flexible conditioning cardinality. To support long-range conditioning, our method constructs a persistent global scene memory via a prepended capture view cache, and removes temporal compression to maintain frame-level correspondence under large viewpoint changes. Beyond better generative model, we also find that the interplay between generation and reconstruction is crucial for large-scale 3D scenes. Thus, we introduce a geometry-aware conditioning strategy that couples generation and reconstruction through an explicit 3D geometric memory and geometry-driven capture-view retrieval. To ensure efficiency, we combine 4-step diffusion distillation with context-window sparse attention to reduce quadratic complexity. Extensive experiments demonstrate robust and scalable reconstruction across irregular inputs, large viewpoint gaps, and long trajectories.
Community
Arbitrary views in, consistent 3D out โ AnyRecon finally makes sparse reconstruction scale.
Excited to share our latest work, AnyRecon! ๐
Project Page: https://yutian10.github.io/AnyRecon/
Code: https://github.com/OpenImagingLab/AnyRecon
Sparse-view 3D reconstruction is essential for bringing casual captures to life, yet scaling it to complex, large-scale scenes remains a significant challenge. Many current diffusion-based methods are limited by fixed input cardinality and a lack of explicit geometric grounding.
With AnyRecon, we introduce a scalable framework designed to handle arbitrary, unordered sparse inputs. Key highlights include:
โ
Flexible Conditioning: Unlike traditional models, our framework adapts to an arbitrary number of reference views.
โ
Dual-Memory
Design: We combine explicit 3D geometry memory with implicit scene memory, coupled with geometry-aware view retrieval to select the most informative segments for generation.
โ
Efficient Architecture: By leveraging 4-step diffusion distillation and block sparse attention, we maintain high fidelity without the heavy computational overhead.
Get this paper in your agent:
hf papers read 2604.19747 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 1
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper