On efficient training of large-scale deep learning models: A literature review

L Shen, Y Sun, Z Yu, L Ding, X Tian, D Tao - arXiv preprint arXiv …, 2023 - arxiv.org
The field of deep learning has witnessed significant progress, particularly in computer vision
(CV), natural language processing (NLP), and speech. The use of large-scale models …

Gsva: Generalized segmentation via multimodal large language models

Z Xia, D Han, Y Han, X Pan, S Song… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Generalized Referring Expression Segmentation (GRES) extends the scope of
classic RES to refer to multiple objects in one expression or identify the empty targets absent …

Efficient diffusion transformer with step-wise dynamic attention mediators

Y Pu, Z Xia, J Guo, D Han, Q Li, D Li, Y Yuan… - … on Computer Vision, 2025 - Springer
This paper identifies significant redundancy in the query-key interactions within self-attention
mechanisms of diffusion transformer models, particularly during the early stages of …

Efficienttrain: Exploring generalized curriculum learning for training visual backbones

Y Wang, Y Yue, R Lu, T Liu, Z Zhong… - Proceedings of the …, 2023 - openaccess.thecvf.com
The superior performance of modern deep networks usually comes with a costly training
procedure. This paper presents a new curriculum learning approach for the efficient training …

Reusing pretrained models by multi-linear operators for efficient training

Y Pan, Y Yuan, Y Yin, Z Xu, L Shang… - Advances in Neural …, 2023 - proceedings.neurips.cc
Training large models from scratch usually costs a substantial amount of resources. Towards
this problem, recent studies such as bert2BERT and LiGO have reused small pretrained …

On Efficient Training of Large-Scale Deep Learning Models

L Shen, Y Sun, Z Yu, L Ding, X Tian, D Tao - ACM Computing Surveys, 2024 - dl.acm.org
The field of deep learning has witnessed significant progress in recent times, particularly in
areas such as computer vision (CV), natural language processing (NLP), and speech. The …

A General and Efficient Training for Transformer via Token Expansion

W Huang, Y Shen, J Xie, B Zhang… - Proceedings of the …, 2024 - openaccess.thecvf.com
The remarkable performance of Vision Transformers (ViTs) typically requires an extremely
large training cost. Existing methods have attempted to accelerate the training of ViTs yet …

Accelerating Augmentation Invariance Pretraining

J Lin, CE Wu, Y Wei, P Morgado - arXiv preprint arXiv:2410.22364, 2024 - arxiv.org
Our work tackles the computational challenges of contrastive learning methods, particularly
for the pretraining of Vision Transformers (ViTs). Despite the effectiveness of contrastive …

Dynamic Patch Sampling for Efficient Training and Dynamic Inference in Vision Transformers

B McDanel, CPN Huynh - 2023 International Conference on …, 2023 - ieeexplore.ieee.org
We introduce the notion of a Patch Sampling Schedule (PSS), that varies the number of
Vision Transformer (ViT) patches used per batch during training. Since all patches are not …