Video unsupervised domain adaptation with deep learning: A comprehensive survey
Video analysis tasks such as action recognition have received increasing research interest
with growing applications in fields such as smart healthcare, thanks to the introduction of …
with growing applications in fields such as smart healthcare, thanks to the introduction of …
A sentence speaks a thousand images: Domain generalization through distilling clip with language guidance
Abstract Domain generalization studies the problem of training a model with samples from
several domains (or distributions) and then testing the model with samples from a new …
several domains (or distributions) and then testing the model with samples from a new …
Domain-agnostic mutual prompting for unsupervised domain adaptation
Abstract Conventional Unsupervised Domain Adaptation (UDA) strives to minimize
distribution discrepancy between domains which neglects to harness rich semantics from …
distribution discrepancy between domains which neglects to harness rich semantics from …
Context-aware robust fine-tuning
Contrastive language-image pre-trained (CLIP) models have zero-shot ability of classifying
an image belonging to “[CLASS]” by using similarity between the image and the prompt …
an image belonging to “[CLASS]” by using similarity between the image and the prompt …
Improved Zero-Shot Classification by Adapting VLMs with Text Descriptions
The zero-shot performance of existing vision-language models (VLMs) such as CLIP is
limited by the availability of large-scale aligned image and text datasets in specific domains …
limited by the availability of large-scale aligned image and text datasets in specific domains …
Bdc-adapter: Brownian distance covariance for better vision-language reasoning
Large-scale pre-trained Vision-Language Models (VLMs), such as CLIP and ALIGN, have
introduced a new paradigm for learning transferable visual representations. Recently, there …
introduced a new paradigm for learning transferable visual representations. Recently, there …
Soft prompt generation for domain generalization
Large pre-trained vision language models (VLMs) have shown impressive zero-shot ability
on downstream tasks with manually designed prompt. To further adapt VLMs to downstream …
on downstream tasks with manually designed prompt. To further adapt VLMs to downstream …
Disentangled Prompt Representation for Domain Generalization
Abstract Domain Generalization (DG) aims to develop a versatile model capable of
performing well on unseen target domains. Recent advancements in pre-trained Visual …
performing well on unseen target domains. Recent advancements in pre-trained Visual …
Landa: Language-guided multi-source domain adaptation
Z Wang, L Zhang, L Wang, M Zhu - arXiv preprint arXiv:2401.14148, 2024 - arxiv.org
Multi-Source Domain Adaptation (MSDA) aims to mitigate changes in data distribution when
transferring knowledge from multiple labeled source domains to an unlabeled target …
transferring knowledge from multiple labeled source domains to an unlabeled target …
Adapting to Distribution Shift by Visual Domain Prompt Generation
In this paper, we aim to adapt a model at test-time using a few unlabeled data to address
distribution shifts. To tackle the challenges of extracting domain knowledge from a limited …
distribution shifts. To tackle the challenges of extracting domain knowledge from a limited …