菜单

关于 🐙 GitHub
arXiv 提交日期: 2026-04-08
📄 Abstract - Fast Spatial Memory with Elastic Test-Time Training

Large Chunk Test-Time Training (LaCT) has shown strong performance on long-context 3D reconstruction, but its fully plastic inference-time updates remain vulnerable to catastrophic forgetting and overfitting. As a result, LaCT is typically instantiated with a single large chunk spanning the full input sequence, falling short of the broader goal of handling arbitrarily long sequences in a single pass. We propose Elastic Test-Time Training inspired by elastic weight consolidation, that stabilizes LaCT fast-weight updates with a Fisher-weighted elastic prior around a maintained anchor state. The anchor evolves as an exponential moving average of past fast weights to balance stability and plasticity. Based on this updated architecture, we introduce Fast Spatial Memory (FSM), an efficient and scalable model for 4D reconstruction that learns spatiotemporal representations from long observation sequences and renders novel view-time combinations. We pre-trained FSM on large-scale curated 3D/4D data to capture the dynamics and semantics of complex spatial environments. Extensive experiments show that FSM supports fast adaptation over long sequences and delivers high-quality 3D/4D reconstruction with smaller chunks and mitigating the camera-interpolation shortcut. Overall, we hope to advance LaCT beyond the bounded single-chunk setting toward robust multi-chunk adaptation, a necessary step for generalization to genuinely longer sequences, while substantially alleviating the activation-memory bottleneck.

顶级标签: computer vision model training systems
详细标签: test-time training 4d reconstruction long-context elastic weight consolidation spatiotemporal representation 或 搜索:

基于弹性测试时训练的高速空间记忆模型 / Fast Spatial Memory with Elastic Test-Time Training


1️⃣ 一句话总结

本文提出了一种结合弹性权重巩固的测试时训练方法,通过稳定模型在长序列数据上的快速更新,有效缓解了灾难性遗忘和过拟合问题,从而构建了一个能高效处理长视频序列并实现高质量3D/4D重建的模型。

源自 arXiv: 2604.07350