Video unsupervised domain adaptation with deep learning: A comprehensive survey

Y Xu, H Cao, L Xie, X Li, Z Chen, J Yang - ACM Computing Surveys, 2024 - dl.acm.org
Video analysis tasks such as action recognition have received increasing research interest
with growing applications in fields such as smart healthcare, thanks to the introduction of …

A sentence speaks a thousand images: Domain generalization through distilling clip with language guidance

Z Huang, A Zhou, Z Ling, M Cai… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Domain generalization studies the problem of training a model with samples from
several domains (or distributions) and then testing the model with samples from a new …

Domain-agnostic mutual prompting for unsupervised domain adaptation

Z Du, X Li, F Li, K Lu, L Zhu, J Li - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Conventional Unsupervised Domain Adaptation (UDA) strives to minimize
distribution discrepancy between domains which neglects to harness rich semantics from …

Context-aware robust fine-tuning

X Mao, Y Chen, X Jia, R Zhang, H Xue, Z Li - International Journal of …, 2024 - Springer
Contrastive language-image pre-trained (CLIP) models have zero-shot ability of classifying
an image belonging to “[CLASS]” by using similarity between the image and the prompt …

Improved Zero-Shot Classification by Adapting VLMs with Text Descriptions

O Saha, G Van Horn, S Maji - Proceedings of the IEEE/CVF …, 2024 - openaccess.thecvf.com
The zero-shot performance of existing vision-language models (VLMs) such as CLIP is
limited by the availability of large-scale aligned image and text datasets in specific domains …

Bdc-adapter: Brownian distance covariance for better vision-language reasoning

Y Zhang, C Zhang, Z Liao, Y Tang, Z He - arXiv preprint arXiv:2309.01256, 2023 - arxiv.org
Large-scale pre-trained Vision-Language Models (VLMs), such as CLIP and ALIGN, have
introduced a new paradigm for learning transferable visual representations. Recently, there …

Soft prompt generation for domain generalization

S Bai, Y Zhang, W Zhou, Z Luan, B Chen - European Conference on …, 2025 - Springer
Large pre-trained vision language models (VLMs) have shown impressive zero-shot ability
on downstream tasks with manually designed prompt. To further adapt VLMs to downstream …

Disentangled Prompt Representation for Domain Generalization

D Cheng, Z Xu, X Jiang, N Wang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Domain Generalization (DG) aims to develop a versatile model capable of
performing well on unseen target domains. Recent advancements in pre-trained Visual …

Landa: Language-guided multi-source domain adaptation

Z Wang, L Zhang, L Wang, M Zhu - arXiv preprint arXiv:2401.14148, 2024 - arxiv.org
Multi-Source Domain Adaptation (MSDA) aims to mitigate changes in data distribution when
transferring knowledge from multiple labeled source domains to an unlabeled target …

Adapting to Distribution Shift by Visual Domain Prompt Generation

Z Chi, L Gu, T Zhong, H Liu, Y Yu… - arXiv preprint arXiv …, 2024 - arxiv.org
In this paper, we aim to adapt a model at test-time using a few unlabeled data to address
distribution shifts. To tackle the challenges of extracting domain knowledge from a limited …