Reconstructing dynamic 4D scenes remains challenging due to moving objects that corrupt camera pose estimation. We propose MoRe, a feed-forward 4D reconstruction network that recovers dynamic 3D scenes efficiently.
Our core innovation is the Attention-Forcing strategy, which guides the model to decouple motion cues from background geometry. Additionally, we introduce Grouped Causal Attention for streamable input processing, ensuring temporal consistency across reconstructed frames.
@misc{fang2026moremotionawarefeedforward4d,
title={MoRe: Motion-aware Feed-forward 4D Reconstruction Transformer},
author={Juntong Fang and Zequn Chen and Weiqi Zhang and Donglin Di and Xuancheng Zhang and Chengmin Yang and Yu-Shen Liu},
booktitle={Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition},
year={2026}
}