Towards a general-purpose foundation model for computational pathology
Quantitative evaluation of tissue images is crucial for computational pathology (CPath) tasks,
requiring the objective characterization of histopathological entities from whole-slide images …
requiring the objective characterization of histopathological entities from whole-slide images …
Emerging properties in self-supervised vision transformers
In this paper, we question if self-supervised learning provides new properties to Vision
Transformer (ViT) that stand out compared to convolutional networks (convnets). Beyond the …
Transformer (ViT) that stand out compared to convolutional networks (convnets). Beyond the …
R-drop: Regularized dropout for neural networks
Dropout is a powerful and widely used technique to regularize the training of deep neural
networks. Though effective and performing well, the randomness introduced by dropout …
networks. Though effective and performing well, the randomness introduced by dropout …
Cross-image relational knowledge distillation for semantic segmentation
Abstract Current Knowledge Distillation (KD) methods for semantic segmentation often
guide the student to mimic the teacher's structured information generated from individual …
guide the student to mimic the teacher's structured information generated from individual …
Co2l: Contrastive continual learning
Recent breakthroughs in self-supervised learning show that such algorithms learn visual
representations that can be transferred better to unseen tasks than cross-entropy based …
representations that can be transferred better to unseen tasks than cross-entropy based …
Distilling large vision-language model with out-of-distribution generalizability
Large vision-language models have achieved outstanding performance, but their size and
computational requirements make their deployment on resource-constrained devices and …
computational requirements make their deployment on resource-constrained devices and …
Masked video distillation: Rethinking masked feature modeling for self-supervised video representation learning
Benefiting from masked visual modeling, self-supervised video representation learning has
achieved remarkable progress. However, existing methods focus on learning …
achieved remarkable progress. However, existing methods focus on learning …
Online prototype learning for online continual learning
Online continual learning (CL) studies the problem of learning continuously from a single-
pass data stream while adapting to new data and mitigating catastrophic forgetting …
pass data stream while adapting to new data and mitigating catastrophic forgetting …
Pyramidclip: Hierarchical feature alignment for vision-language model pretraining
Large-scale vision-language pre-training has achieved promising results on downstream
tasks. Existing methods highly rely on the assumption that the image-text pairs crawled from …
tasks. Existing methods highly rely on the assumption that the image-text pairs crawled from …
Ressl: Relational self-supervised learning with weak augmentation
Self-supervised Learning (SSL) including the mainstream contrastive learning has achieved
great success in learning visual representations without data annotations. However, most of …
great success in learning visual representations without data annotations. However, most of …