A survey on visual mamba

H Zhang, Y Zhu, D Wang, L Zhang, T Chen, Z Wang… - Applied Sciences, 2024 - mdpi.com
State space models (SSM) with selection mechanisms and hardware-aware architectures,
namely Mamba, have recently shown significant potential in long-sequence modeling. Since …

Vision mamba: Efficient visual representation learning with bidirectional state space model

L Zhu, B Liao, Q Zhang, X Wang, W Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently the state space models (SSMs) with efficient hardware-aware designs, ie, the
Mamba deep learning model, have shown great potential for long sequence modeling …

A survey on vision mamba: Models, applications and challenges

R Xu, S Yang, Y Wang, B Du, H Chen - arXiv preprint arXiv:2404.18861, 2024 - arxiv.org
Mamba, a recent selective structured state space model, performs excellently on long
sequence modeling tasks. Mamba mitigates the modeling constraints of convolutional …

Vm-unet: Vision mamba unet for medical image segmentation

J Ruan, S Xiang - arXiv preprint arXiv:2402.02491, 2024 - arxiv.org
In the realm of medical image segmentation, both CNN-based and Transformer-based
models have been extensively explored. However, CNNs exhibit limitations in long-range …

Graph mamba: Towards learning on graphs with state space models

A Behrouz, F Hashemi - Proceedings of the 30th ACM SIGKDD …, 2024 - dl.acm.org
Graph Neural Networks (GNNs) have shown promising potential in graph representation
learning. The majority of GNNs define a local message-passing mechanism, propagating …

Gated linear attention transformers with hardware-efficient training

S Yang, B Wang, Y Shen, R Panda, Y Kim - arXiv preprint arXiv …, 2023 - arxiv.org
Transformers with linear attention allow for efficient parallel training but can simultaneously
be formulated as an RNN with 2D (matrix-valued) hidden states, thus enjoying linear (with …

Vivim: a video vision mamba for medical video object segmentation

Y Yang, Z Xing, L Zhu - arXiv preprint arXiv:2401.14168, 2024 - arxiv.org
Traditional convolutional neural networks have a limited receptive field while transformer-
based networks are mediocre in constructing long-term dependency from the perspective of …

Pointmamba: A simple state space model for point cloud analysis

D Liang, X Zhou, X Wang, X Zhu, W Xu, Z Zou… - arXiv preprint arXiv …, 2024 - arxiv.org
Transformers have become one of the foundational architectures in point cloud analysis
tasks due to their excellent global modeling ability. However, the attention mechanism has …

Swin-umamba: Mamba-based unet with imagenet-based pretraining

J Liu, H Yang, HY Zhou, Y Xi, L Yu, Y Yu… - arXiv preprint arXiv …, 2024 - arxiv.org
Accurate medical image segmentation demands the integration of multi-scale information,
spanning from local features to global dependencies. However, it is challenging for existing …

Mambair: A simple baseline for image restoration with state-space model

H Guo, J Li, T Dai, Z Ouyang, X Ren, ST Xia - arXiv preprint arXiv …, 2024 - arxiv.org
Recent years have witnessed great progress in image restoration thanks to the
advancements in modern deep neural networks eg Convolutional Neural Network and …