A comprehensive survey on test-time adaptation under distribution shifts
Abstract Machine learning methods strive to acquire a robust model during the training
process that can effectively generalize to test samples, even in the presence of distribution …
process that can effectively generalize to test samples, even in the presence of distribution …
Adding conditional control to text-to-image diffusion models
L Zhang, A Rao, M Agrawala - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
We present ControlNet, a neural network architecture to add spatial conditioning controls to
large, pretrained text-to-image diffusion models. ControlNet locks the production-ready large …
large, pretrained text-to-image diffusion models. ControlNet locks the production-ready large …
Multi-concept customization of text-to-image diffusion
While generative models produce high-quality images of concepts learned from a large-
scale database, a user often wishes to synthesize instantiations of their own concepts (for …
scale database, a user often wishes to synthesize instantiations of their own concepts (for …
Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation
Large text-to-image models achieved a remarkable leap in the evolution of AI, enabling high-
quality and diverse synthesis of images from a given text prompt. However, these models …
quality and diverse synthesis of images from a given text prompt. However, these models …
An image is worth one word: Personalizing text-to-image generation using textual inversion
Text-to-image models offer unprecedented freedom to guide creation through natural
language. Yet, it is unclear how such freedom can be exercised to generate images of …
language. Yet, it is unclear how such freedom can be exercised to generate images of …
Stylegan-t: Unlocking the power of gans for fast large-scale text-to-image synthesis
Text-to-image synthesis has recently seen significant progress thanks to large pretrained
language models, large-scale training data, and the introduction of scalable model families …
language models, large-scale training data, and the introduction of scalable model families …
Hyperdreambooth: Hypernetworks for fast personalization of text-to-image models
Personalization has emerged as a prominent aspect within the field of generative AI
enabling the synthesis of individuals in diverse contexts and styles while retaining high …
enabling the synthesis of individuals in diverse contexts and styles while retaining high …
Ablating concepts in text-to-image diffusion models
Large-scale text-to-image diffusion models can generate high-fidelity images with powerful
compositional ability. However, these models are typically trained on an enormous amount …
compositional ability. However, these models are typically trained on an enormous amount …
Test-time training with masked autoencoders
Test-time training adapts to a new test distribution on the fly by optimizing a model for each
test input using self-supervision. In this paper, we use masked autoencoders for this one …
test input using self-supervision. In this paper, we use masked autoencoders for this one …
Encoder-based domain tuning for fast personalization of text-to-image models
Text-to-image personalization aims to teach a pre-trained diffusion model to reason about
novel, user provided concepts, embedding them into new scenes guided by natural …
novel, user provided concepts, embedding them into new scenes guided by natural …