Dreambooth: Fine tuning text-to-image diffusion models for subject-driven generation

N Ruiz, Y Li, V Jampani, Y Pritch… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large text-to-image models achieved a remarkable leap in the evolution of AI, enabling high-
quality and diverse synthesis of images from a given text prompt. However, these models …

Iti-gen: Inclusive text-to-image generation

C Zhang, X Chen, S Chai, CH Wu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Text-to-image generative models often reflect the biases of the training data, leading to
unequal representations of underrepresented groups. This study investigates inclusive text …

Boxdiff: Text-to-image synthesis with training-free box-constrained diffusion

J Xie, Y Li, Y Huang, H Liu, W Zhang… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recent text-to-image diffusion models have demonstrated an astonishing capacity to
generate high-quality images. However, researchers mainly studied the way of synthesizing …

Plug-and-play diffusion features for text-driven image-to-image translation

N Tumanyan, M Geyer, S Bagon… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large-scale text-to-image generative models have been a revolutionary breakthrough in the
evolution of generative AI, synthesizing diverse images with highly complex visual concepts …

ViCo: Plug-and-play Visual Condition for Personalized Text-to-image Generation

S Hao, K Han, S Zhao, KYK Wong - arXiv preprint arXiv:2306.00971, 2023 - arxiv.org
Personalized text-to-image generation using diffusion models has recently emerged and
garnered significant interest. This task learns a novel concept (eg, a unique toy), illustrated …

Harnessing the spatial-temporal attention of diffusion models for high-fidelity text-to-image synthesis

Q Wu, Y Liu, H Zhao, T Bui, Z Lin… - Proceedings of the …, 2023 - openaccess.thecvf.com
Diffusion-based models have achieved state-of-the-art performance on text-to-image
synthesis tasks. However, one critical limitation of these models is the low fidelity of …

Prompt-free diffusion: Taking" text" out of text-to-image diffusion models

X Xu, J Guo, Z Wang, G Huang… - Proceedings of the …, 2024 - openaccess.thecvf.com
Abstract Text-to-image (T2I) research has grown explosively in the past year owing to the
large-scale pre-trained diffusion models and many emerging personalization and editing …

Svdiff: Compact parameter space for diffusion fine-tuning

L Han, Y Li, H Zhang, P Milanfar… - Proceedings of the …, 2023 - openaccess.thecvf.com
Recently, diffusion models have achieved remarkable success in text-to-image generation,
enabling the creation of high-quality images from text prompts and various conditions …

Dense text-to-image generation with attention modulation

Y Kim, J Lee, JH Kim, JW Ha… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Existing text-to-image diffusion models struggle to synthesize realistic images given dense
captions, where each text prompt provides a detailed description for a specific image region …

Training-free consistent text-to-image generation

Y Tewel, O Kaduri, R Gal, Y Kasten, L Wolf… - arXiv preprint arXiv …, 2024 - arxiv.org
Text-to-image models offer a new level of creative flexibility by allowing users to guide the
image generation process through natural language. However, using these models to …