Vmrnn: Integrating vision mamba and lstm for efficient and accurate spatiotemporal forecasting

Y Tang, P Dong, Z Tang, X Chu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Combining Convolutional Neural Networks (CNNs) or Vision Transformers (ViTs)
with Recurrent Neural Networks (RNNs) for spatiotemporal forecasting has yielded …

Swinlstm: Improving spatiotemporal prediction accuracy using swin transformer and lstm

S Tang, C Li, P Zhang, RN Tang - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Abstract Integrating CNNs and RNNs to capture spatiotemporal dependencies is a prevalent
strategy for spatiotemporal prediction tasks. However, the property of CNNs to learn local …

MS-RNN: A flexible multi-scale framework for spatiotemporal predictive learning

Z Ma, H Zhang, J Liu - arXiv preprint arXiv:2206.03010, 2022 - arxiv.org
Spatiotemporal predictive learning, which predicts future frames through historical prior
knowledge with the aid of deep learning, is widely used in many fields. Previous work …

Spatial-temporal consistency network for low-latency trajectory forecasting

S Li, Y Zhou, J Yi, J Gall - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Trajectory forecasting is a crucial step for autonomous vehicles and mobile robots in order to
navigate and interact safely. In order to handle the spatial interactions between objects …

Temporal attention unit: Towards efficient spatiotemporal predictive learning

C Tan, Z Gao, L Wu, Y Xu, J Xia… - Proceedings of the …, 2023 - openaccess.thecvf.com
Spatiotemporal predictive learning aims to generate future frames by learning from historical
frames. In this paper, we investigate existing methods and present a general framework of …

Simvp: Towards simple yet powerful spatiotemporal predictive learning

C Tan, Z Gao, S Li, SZ Li - arXiv preprint arXiv:2211.12509, 2022 - arxiv.org
Recent years have witnessed remarkable advances in spatiotemporal predictive learning,
incorporating auxiliary inputs, elaborate neural architectures, and sophisticated training …

Mambamixer: Efficient selective state space models with dual token and channel selection

A Behrouz, M Santacatterina, R Zabih - arXiv preprint arXiv:2403.19888, 2024 - arxiv.org
Recent advances in deep learning have mainly relied on Transformers due to their data
dependency and ability to learn at scale. The attention module in these architectures …

Triplet attention transformer for spatiotemporal predictive learning

X Nie, X Chen, H Jin, Z Zhu… - Proceedings of the …, 2024 - openaccess.thecvf.com
Spatiotemporal predictive learning offers a self-supervised learning paradigm that enables
models to learn both spatial and temporal patterns by predicting future sequences based on …

Hpnet: Dynamic trajectory forecasting with historical prediction attention

X Tang, M Kan, S Shan, Z Ji, J Bai… - Proceedings of the …, 2024 - openaccess.thecvf.com
Predicting the trajectories of road agents is essential for autonomous driving systems. The
recent mainstream methods follow a static paradigm which predicts the future trajectory by …

Muse-vae: Multi-scale vae for environment-aware long term trajectory prediction

M Lee, SS Sohn, S Moon, S Yoon… - Proceedings of the …, 2022 - openaccess.thecvf.com
Accurate long-term trajectory prediction in complex scenes, where multiple agents (eg,
pedestrians or vehicles) interact with each other and the environment while attempting to …