Reconstructing dynamic 4D scenes remains challenging due to moving objects that corrupt camera pose estimation. We propose MoRe, a feed-forward 4D reconstruction network that recovers dynamic 3D scenes efficiently.
Our core innovation is the Attention-Forcing strategy, which guides the model to decouple motion cues from background geometry. Additionally, we introduce Grouped Causal Attention for streamable input processing, ensuring temporal consistency across reconstructed frames.
@article{fang2025more,
title={MoRe: Motion-aware Feed-forward 4D Reconstruction Transformer},
author={Fang, Junton and Chen, Zequn and Zhang, Weiqi and Di, Donglin and Zhang, Xuancheng and Yang, Chengmin and Liu, Yu-Shen},
journal={arXiv preprint},
year={2025}
}