A survey on vision mamba: Models, applications and challenges

R Xu, S Yang, Y Wang, B Du, H Chen - arXiv preprint arXiv:2404.18861, 2024 - arxiv.org
Mamba, a recent selective structured state space model, performs excellently on long
sequence modeling tasks. Mamba mitigates the modeling constraints of convolutional …

Mamba-360: Survey of state space models as transformer alternative for long sequence modelling: Methods, applications, and challenges

BN Patro, VS Agneeswaran - arXiv preprint arXiv:2404.16112, 2024 - arxiv.org
Sequence modeling is a crucial area across various domains, including Natural Language
Processing (NLP), speech recognition, time series forecasting, music generation, and …

Pointmamba: A simple state space model for point cloud analysis

D Liang, X Zhou, X Wang, X Zhu, W Xu, Z Zou… - arXiv preprint arXiv …, 2024 - arxiv.org
Transformers have become one of the foundational architectures in point cloud analysis
tasks due to their excellent global modeling ability. However, the attention mechanism has …

Localmamba: Visual state space model with windowed selective scan

T Huang, X Pei, S You, F Wang, C Qian… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent advancements in state space models, notably Mamba, have demonstrated
significant progress in modeling long sequences for tasks like language understanding. Yet …

Point could mamba: Point cloud learning via state space model

T Zhang, X Li, H Yuan, S Ji, S Yan - arXiv preprint arXiv:2403.00762, 2024 - arxiv.org
In this work, for the first time, we demonstrate that Mamba-based point cloud methods can
outperform point-based methods. Mamba exhibits strong global modeling capabilities and …

Vl-mamba: Exploring state space models for multimodal learning

Y Qiao, Z Yu, L Guo, S Chen, Z Zhao, M Sun… - arXiv preprint arXiv …, 2024 - arxiv.org
Multimodal large language models (MLLMs) have attracted widespread interest and have
rich applications. However, the inherent attention mechanism in its Transformer structure …

The hidden attention of mamba models

A Ali, I Zimerman, L Wolf - arXiv preprint arXiv:2403.01590, 2024 - arxiv.org
The Mamba layer offers an efficient selective state space model (SSM) that is highly effective
in modeling multiple domains including NLP, long-range sequences processing, and …

Vmambair: Visual state space model for image restoration

Y Shi, B Xia, X Jin, X Wang, T Zhao, X Xia… - arXiv preprint arXiv …, 2024 - arxiv.org
Image restoration is a critical task in low-level computer vision, aiming to restore high-quality
images from degraded inputs. Various models, such as convolutional neural networks …

Gamba: Marry gaussian splatting with mamba for single view 3d reconstruction

Q Shen, X Yi, Z Wu, P Zhou, H Zhang, S Yan… - arXiv preprint arXiv …, 2024 - arxiv.org
We tackle the challenge of efficiently reconstructing a 3D asset from a single image with
growing demands for automated 3D content creation pipelines. Previous methods primarily …

Lightm-unet: Mamba assists in lightweight unet for medical image segmentation

W Liao, Y Zhu, X Wang, C Pan, Y Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
UNet and its variants have been widely used in medical image segmentation. However,
these models, especially those based on Transformer architectures, pose challenges due to …