Dragdiffusion: Harnessing diffusion models for interactive point-based image editing

Y Shi, C Xue, JH Liew, J Pan, H Yan… - Proceedings of the …, 2024 - openaccess.thecvf.com
Accurate and controllable image editing is a challenging task that has attracted significant
attention recently. Notably DragGAN developed by Pan et al.(2023) is an interactive point …

Text-to-image diffusion models in generative ai: A survey

C Zhang, C Zhang, M Zhang, IS Kweon - arXiv preprint arXiv:2303.07909, 2023 - arxiv.org
This survey reviews text-to-image diffusion models in the context that diffusion models have
emerged to be popular for a wide range of generative tasks. As a self-contained work, this …

Localizing object-level shape variations with text-to-image diffusion models

O Patashnik, D Garibi, I Azuri… - Proceedings of the …, 2023 - openaccess.thecvf.com
Text-to-image models give rise to workflows which often begin with an exploration step,
where users sift through a large collection of generated images. The global nature of the text …

Hive: Harnessing human feedback for instructional visual editing

S Zhang, X Yang, Y Feng, C Qin… - Proceedings of the …, 2024 - openaccess.thecvf.com
Incorporating human feedback has been shown to be crucial to align text generated by large
language models to human preferences. We hypothesize that state-of-the-art instructional …

Matting anything

J Li, J Jain, H Shi - … of the IEEE/CVF Conference on …, 2024 - openaccess.thecvf.com
In this paper we propose the Matting Anything Model (MAM) an efficient and versatile
framework for estimating the alpha matte of any instance in an image with flexible and …

Dream the impossible: Outlier imagination with diffusion models

X Du, Y Sun, J Zhu, Y Li - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Utilizing auxiliary outlier datasets to regularize the machine learning model has
demonstrated promise for out-of-distribution (OOD) detection and safe prediction. Due to the …

A neural space-time representation for text-to-image personalization

Y Alaluf, E Richardson, G Metzer… - ACM Transactions on …, 2023 - dl.acm.org
A key aspect of text-to-image personalization methods is the manner in which the target
concept is represented within the generative process. This choice greatly affects the visual …

Cross-image attention for zero-shot appearance transfer

Y Alaluf, D Garibi, O Patashnik… - ACM SIGGRAPH 2024 …, 2024 - dl.acm.org
Recent advancements in text-to-image generative models have demonstrated a remarkable
ability to capture a deep semantic understanding of images. In this work, we leverage this …

ReVersion: Diffusion-based relation inversion from images

Z Huang, T Wu, Y Jiang, KCK Chan, Z Liu - SIGGRAPH Asia 2024 …, 2024 - dl.acm.org
Diffusion models gain increasing popularity for their generative capabilities. Recently, there
have been surging needs to generate customized images by inverting diffusion models from …

Promptpaint: Steering text-to-image generation through paint medium-like interactions

JJY Chung, E Adar - Proceedings of the 36th Annual ACM Symposium …, 2023 - dl.acm.org
While diffusion-based text-to-image (T2I) models provide a simple and powerful way to
generate images, guiding this generation remains a challenge. For concepts that are difficult …