Visual tuning

BXB Yu, J Chang, H Wang, L Liu, S Wang… - ACM Computing …, 2023 - dl.acm.org
Fine-tuning visual models has been widely shown promising performance on many
downstream visual tasks. With the surprising development of pre-trained visual foundation …

Knowledge distillation from a stronger teacher

T Huang, S You, F Wang, C Qian… - Advances in Neural …, 2022 - proceedings.neurips.cc
Unlike existing knowledge distillation methods focus on the baseline settings, where the
teacher models and training strategies are not that strong and competing as state-of-the-art …

Deep transfer learning for intelligent vehicle perception: A survey

X Liu, J Li, J Ma, H Sun, Z Xu, T Zhang, H Yu - Green Energy and Intelligent …, 2023 - Elsevier
Deep learning-based intelligent vehicle perception has been developing prominently in
recent years to provide a reliable source for motion planning and decision making in …

Automated knowledge distillation via monte carlo tree search

L Li, P Dong, Z Wei, Y Yang - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
In this paper, we present Auto-KD, the first automated search framework for optimal
knowledge distillation design. Traditional distillation techniques typically require handcrafted …

Knowledge diffusion for distillation

T Huang, Y Zhang, M Zheng, S You… - Advances in …, 2024 - proceedings.neurips.cc
The representation gap between teacher and student is an emerging topic in knowledge
distillation (KD). To reduce the gap and improve the performance, current methods often …

Kd-zero: Evolving knowledge distiller for any teacher-student pairs

L Li, P Dong, A Li, Z Wei… - Advances in Neural …, 2023 - proceedings.neurips.cc
Abstract Knowledge distillation (KD) has emerged as an effective technique for compressing
models that can enhance the lightweight model. Conventional KD methods propose various …

Norm: Knowledge distillation via n-to-one representation matching

X Liu, L Li, C Li, A Yao - arXiv preprint arXiv:2305.13803, 2023 - arxiv.org
Existing feature distillation methods commonly adopt the One-to-one Representation
Matching between any pre-selected teacher-student layer pair. In this paper, we present N …

TransKD: Transformer knowledge distillation for efficient semantic segmentation

R Liu, K Yang, A Roitberg, J Zhang, K Peng… - arXiv preprint arXiv …, 2022 - arxiv.org
Large pre-trained transformers are on top of contemporary semantic segmentation
benchmarks, but come with high computational cost and a lengthy training. To lift this …

Directional connectivity-based segmentation of medical images

Z Yang, S Farsiu - Proceedings of the IEEE/CVF conference …, 2023 - openaccess.thecvf.com
Anatomical consistency in biomarker segmentation is crucial for many medical image
analysis tasks. A promising paradigm for achieving anatomically consistent segmentation …

Mixskd: Self-knowledge distillation from mixup for image recognition

C Yang, Z An, H Zhou, L Cai, X Zhi, J Wu, Y Xu… - … on Computer Vision, 2022 - Springer
Abstract Unlike the conventional Knowledge Distillation (KD), Self-KD allows a network to
learn knowledge from itself without any guidance from extra networks. This paper proposes …