Stylip: Multi-scale style-conditioned prompt learning for clip-based domain generalization

S Bose, A Jha, E Fini, M Singha… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract arge-scale foundation models, such as CLIP, have demonstrated impressive zero-
shot generalization performance on downstream tasks, leveraging well-designed language …

UCDR-Adapter: Exploring Adaptation of Pre-Trained Vision-Language Models for Universal Cross-Domain Retrieval

H Jiang, ZQ Cheng, G Moreira, J Zhu, J Sun… - arXiv preprint arXiv …, 2024 - arxiv.org
Universal Cross-Domain Retrieval (UCDR) retrieves relevant images from unseen domains
and classes without semantic labels, ensuring robust generalization. Existing methods …

INDIGO: Intrinsic Multimodality for Domain Generalization

P Mangla, S Chandhok, M Aggarwal… - arXiv preprint arXiv …, 2022 - arxiv.org
For models to generalize under unseen domains (aka domain generalization), it is crucial to
learn feature representations that are domain-agnostic and capture the underlying …

Self-supervised few-shot learning for real-time traffic sign classification.

AKT Nguyen, T Tran, PH Nguyen… - International Journal of …, 2024 - search.ebscohost.com
Although supervised approaches for traffic sign classification have demonstrated excellent
performance, they are limited to classifying several traffic signs defined in the training …

Prompt Tuning Is All We Need?

H Yu, H Zheng, Y Zhang, S Xie, X Cao, Z Fang - openreview.net
Recent advances in pre-trained vision-language models, eg, CLIP, have demonstrated
remarkable success in domain generalization (DG) by tuning prompts. To promote DG, one …