[HTML][HTML] Transformers in medical image analysis

K He, C Gan, Z Li, I Rekik, Z Yin, W Ji, Y Gao, Q Wang… - Intelligent …, 2023 - Elsevier
Transformers have dominated the field of natural language processing and have recently
made an impact in the area of computer vision. In the field of medical image analysis …

Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives

J Li, J Chen, Y Tang, C Wang, BA Landman… - Medical image …, 2023 - Elsevier
Transformer, one of the latest technological advances of deep learning, has gained
prevalence in natural language processing or computer vision. Since medical imaging bear …

Visual attention network

MH Guo, CZ Lu, ZN Liu, MM Cheng, SM Hu - Computational Visual Media, 2023 - Springer
While originally designed for natural language processing tasks, the self-attention
mechanism has recently taken various computer vision areas by storm. However, the 2D …

Rethinking vision transformers for mobilenet size and speed

Y Li, J Hu, Y Wen, G Evangelidis… - Proceedings of the …, 2023 - openaccess.thecvf.com
With the success of Vision Transformers (ViTs) in computer vision tasks, recent arts try to
optimize the performance and complexity of ViTs to enable efficient deployment on mobile …

Global context vision transformers

A Hatamizadeh, H Yin, G Heinrich… - International …, 2023 - proceedings.mlr.press
We propose global context vision transformer (GC ViT), a novel architecture that enhances
parameter and compute utilization for computer vision. Our method leverages global context …

Metaformer baselines for vision

W Yu, C Si, P Zhou, M Luo, Y Zhou… - … on Pattern Analysis …, 2023 - ieeexplore.ieee.org
MetaFormer, the abstracted architecture of Transformer, has been found to play a significant
role in achieving competitive performance. In this paper, we further explore the capacity of …

A generalization of vit/mlp-mixer to graphs

X He, B Hooi, T Laurent, A Perold… - International …, 2023 - proceedings.mlr.press
Abstract Graph Neural Networks (GNNs) have shown great potential in the field of graph
representation learning. Standard GNNs define a local message-passing mechanism which …

Adaptive frequency filters as efficient global token mixers

Z Huang, Z Zhang, C Lan, ZJ Zha… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recent vision transformers, large-kernel CNNs and MLPs have attained remarkable
successes in broad vision tasks thanks to their effective information fusion in the global …

Potter: Pooling attention transformer for efficient human mesh recovery

C Zheng, X Liu, GJ Qi, C Chen - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Transformer architectures have achieved SOTA performance on the human mesh recovery
(HMR) from monocular images. However, the performance gain has come at the cost of …

Rethinking mobile block for efficient attention-based models

J Zhang, X Li, J Li, L Liu, Z Xue, B Zhang… - 2023 IEEE/CVF …, 2023 - computer.org
This paper focuses on developing modern, efficient, lightweight models for dense
predictions while trading off parameters, FLOPs, and performance. Inverted Residual Block …