Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives

J Li, J Chen, Y Tang, C Wang, BA Landman… - Medical image …, 2023 - Elsevier
Transformer, one of the latest technological advances of deep learning, has gained
prevalence in natural language processing or computer vision. Since medical imaging bear …

Are we ready for a new paradigm shift? a survey on visual deep mlp

R Liu, Y Li, L Tao, D Liang, HT Zheng - Patterns, 2022 - cell.com
Recently, the proposed deep multilayer perceptron (MLP) models have stirred up a lot of
interest in the vision community. Historically, the availability of larger datasets combined with …

Unext: Mlp-based rapid medical image segmentation network

JMJ Valanarasu, VM Patel - … conference on medical image computing and …, 2022 - Springer
UNet and its latest extensions like TransUNet have been the leading medical image
segmentation methods in recent years. However, these networks cannot be effectively …

Davit: Dual attention vision transformers

M Ding, B Xiao, N Codella, P Luo, J Wang… - European conference on …, 2022 - Springer
In this work, we introduce Dual Attention Vision Transformers (DaViT), a simple yet effective
vision transformer architecture that is able to capture global context while maintaining …

Conv2former: A simple transformer-style convnet for visual recognition

Q Hou, CZ Lu, MM Cheng… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Vision Transformers have been the most popular network architecture in visual recognition
recently due to the strong ability of encode global information. However, its high …

As-mlp: An axial shifted mlp architecture for vision

D Lian, Z Yu, X Sun, S Gao - arXiv preprint arXiv:2107.08391, 2021 - arxiv.org
An Axial Shifted MLP architecture (AS-MLP) is proposed in this paper. Different from MLP-
Mixer, where the global spatial feature is encoded for information flow through matrix …

An image patch is a wave: Phase-aware vision mlp

Y Tang, K Han, J Guo, C Xu, Y Li… - Proceedings of the …, 2022 - openaccess.thecvf.com
In the field of computer vision, recent works show that a pure MLP architecture mainly
stacked by fully-connected layers can achieve competing performance with CNN and …

Hire-mlp: Vision mlp via hierarchical rearrangement

J Guo, Y Tang, K Han, X Chen, H Wu… - Proceedings of the …, 2022 - openaccess.thecvf.com
Previous vision MLPs such as MLP-Mixer and ResMLP accept linearly flattened image
patches as input, making them inflexible for different input sizes and hard to capture spatial …

Pointmixer: Mlp-mixer for point cloud understanding

J Choe, C Park, F Rameau, J Park… - European Conference on …, 2022 - Springer
MLP-Mixer has newly appeared as a new challenger against the realm of CNNs and
Transformer. Despite its simplicity compared to Transformer, the concept of channel-mixing …

Sequencer: Deep lstm for image classification

Y Tatsunami, M Taki - Advances in Neural Information …, 2022 - proceedings.neurips.cc
In recent computer vision research, the advent of the Vision Transformer (ViT) has rapidly
revolutionized various architectural design efforts: ViT achieved state-of-the-art image …