Contrastive representation distillation

Y Tian, D Krishnan, P Isola - arXiv preprint arXiv:1910.10699, 2019 - arxiv.org
… the teacher's representation of the data. We formulate this objective as contrastive learning.
Experiments demonstrate that our resulting new objective outperforms knowledge distillation

Wasserstein contrastive representation distillation

L Chen, D Wang, Z Gan, J Liu… - Proceedings of the …, 2021 - openaccess.thecvf.com
… for distillation based on Wasserstein distance, where both global contrastive learning and
… Note that our approach utilizes feature representations at the penultimate layer (before logits…

Deep contrastive representation learning with self-distillation

Z Xiao, H Xing, B Zhao, R Qu, S Luo… - … on Emerging Topics …, 2023 - ieeexplore.ieee.org
contrastive learning, and self distillation. … contrastive learning also considers the contrast
similarity of low-level semantic information between peer residual blocks. Our self distillation

Complementary relation contrastive distillation

J Zhu, S Tang, D Chen, S Yu, Y Liu… - Proceedings of the …, 2021 - openaccess.thecvf.com
… proposes to use contrastive tasks as … contrastive loss, our method is more similar to CRD,
but our objective is the mutual relations of deep representations, instead of the representations

[PDF][PDF] Contrastive Consistent Representation Distillation.

S Fu, H Yang, X Yang - BMVC, 2023 - papers.bmvc2023.org
representation quality affects the distillation performance, we utilize different models to provide
those representations. … -cached keys is referred to as contrastive encoder. The model that …

Distillation with contrast is all you need for self-supervised point cloud representation learning

K Fu, P Gao, R Zhang, H Li, Y Qiao, M Wang - arXiv preprint arXiv …, 2022 - arxiv.org
Representation). For easier understanding of our framework, we will first briefly review knowledge
distillation and contrastive … self-supervised point cloud representation learning. Finally, …

Contrastive distillation on intermediate representations for language model compression

S Sun, Z Gan, Y Cheng, Y Fang, S Wang… - arXiv preprint arXiv …, 2020 - arxiv.org
distillation efficacy, we propose Contrastive Distillationdistillation framework where the
student is trained to distill knowledge through intermediate layers of the teacher via a contrastive

Simcvd: Simple contrastive voxel-wise representation distillation for semi-supervised medical image segmentation

C You, Y Zhou, R Zhao, L Staib… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
… of-the-art voxel-wise representation learning in a more … contrastive voxel-wise representation
distillation framework, which can be utilized to produce superior voxel-wise representations

Contrastive pre-training and representation distillation for medical visual question answering based on radiology images

B Liu, LM Zhan, XM Wu - … , Strasbourg, France, September 27–October 1 …, 2021 - Springer
… Second, to obtain a general and lightweight feature extractor, we distill the three teacher
models into a small student model by contrastive representation distillation. The distilled model …

LRC-BERT: latent-representation contrastive knowledge distillation for natural language understanding

H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
distillation method LRCBERT based on contrastive learning … considered by the existing
distillation methods. Furthermore, … stage training method for the total distillation loss. Finally, by …