Contrastive representation distillation
… the teacher's representation of the data. We formulate this objective as contrastive learning.
Experiments demonstrate that our resulting new objective outperforms knowledge distillation …
Experiments demonstrate that our resulting new objective outperforms knowledge distillation …
Wasserstein contrastive representation distillation
… for distillation based on Wasserstein distance, where both global contrastive learning and
… Note that our approach utilizes feature representations at the penultimate layer (before logits…
… Note that our approach utilizes feature representations at the penultimate layer (before logits…
Deep contrastive representation learning with self-distillation
… contrastive learning, and self distillation. … contrastive learning also considers the contrast
similarity of low-level semantic information between peer residual blocks. Our self distillation …
similarity of low-level semantic information between peer residual blocks. Our self distillation …
Complementary relation contrastive distillation
… proposes to use contrastive tasks as … contrastive loss, our method is more similar to CRD,
but our objective is the mutual relations of deep representations, instead of the representations …
but our objective is the mutual relations of deep representations, instead of the representations …
[PDF][PDF] Contrastive Consistent Representation Distillation.
S Fu, H Yang, X Yang - BMVC, 2023 - papers.bmvc2023.org
… representation quality affects the distillation performance, we utilize different models to provide
those representations. … -cached keys is referred to as contrastive encoder. The model that …
those representations. … -cached keys is referred to as contrastive encoder. The model that …
Distillation with contrast is all you need for self-supervised point cloud representation learning
… Representation). For easier understanding of our framework, we will first briefly review knowledge
distillation and contrastive … self-supervised point cloud representation learning. Finally, …
distillation and contrastive … self-supervised point cloud representation learning. Finally, …
Contrastive distillation on intermediate representations for language model compression
… distillation efficacy, we propose Contrastive Distillation … distillation framework where the
student is trained to distill knowledge through intermediate layers of the teacher via a contrastive …
student is trained to distill knowledge through intermediate layers of the teacher via a contrastive …
Simcvd: Simple contrastive voxel-wise representation distillation for semi-supervised medical image segmentation
… of-the-art voxel-wise representation learning in a more … contrastive voxel-wise representation
distillation framework, which can be utilized to produce superior voxel-wise representations …
distillation framework, which can be utilized to produce superior voxel-wise representations …
Contrastive pre-training and representation distillation for medical visual question answering based on radiology images
… Second, to obtain a general and lightweight feature extractor, we distill the three teacher
models into a small student model by contrastive representation distillation. The distilled model …
models into a small student model by contrastive representation distillation. The distilled model …
LRC-BERT: latent-representation contrastive knowledge distillation for natural language understanding
H Fu, S Zhou, Q Yang, J Tang, G Liu, K Liu… - Proceedings of the AAAI …, 2021 - ojs.aaai.org
… distillation method LRCBERT based on contrastive learning … considered by the existing
distillation methods. Furthermore, … stage training method for the total distillation loss. Finally, by …
distillation methods. Furthermore, … stage training method for the total distillation loss. Finally, by …
相关搜索
- contrastive representation distillation iclr
- contrastive representation distillation wasserstein
- contrastive knowledge distillation
- contrastive distillation complementary relation
- contrastive distillation semi-supervised medical image segmentation
- contrastive distillation language model compression
- knowledge distillation representation matching