Preserve your own correlation: A noise prior for video diffusion models
Despite tremendous progress in generating high-quality images using diffusion models,
synthesizing a sequence of animated frames that are both photorealistic and temporally …
synthesizing a sequence of animated frames that are both photorealistic and temporally …
Grounded text-to-image synthesis with attention refocusing
Driven by the scalable diffusion models trained on large-scale datasets text-to-image
synthesis methods have shown compelling results. However these models still fail to …
synthesis methods have shown compelling results. However these models still fail to …
Freecontrol: Training-free spatial control of any text-to-image diffusion model with any condition
Recent approaches such as ControlNet offer users fine-grained spatial control over text-to-
image (T2I) diffusion models. However auxiliary modules have to be trained for each spatial …
image (T2I) diffusion models. However auxiliary modules have to be trained for each spatial …
Portraitbooth: A versatile portrait model for fast identity-preserved personalization
Recent advancements in personalized image generation using diffusion models have been
noteworthy. However existing methods suffer from inefficiencies due to the requirement for …
noteworthy. However existing methods suffer from inefficiencies due to the requirement for …
Cross-image attention for zero-shot appearance transfer
Recent advancements in text-to-image generative models have demonstrated a remarkable
ability to capture a deep semantic understanding of images. In this work, we leverage this …
ability to capture a deep semantic understanding of images. In this work, we leverage this …
It's All About Your Sketch: Democratising Sketch Control in Diffusion Models
This paper unravels the potential of sketches for diffusion models addressing the deceptive
promise of direct sketch control in generative AI. We importantly democratise the process …
promise of direct sketch control in generative AI. We importantly democratise the process …
Text-guided synthesis of eulerian cinemagraphs
We introduce Text2Cinemagraph, a fully automated method for creating cinemagraphs from
text descriptions---an especially challenging task when prompts feature imaginary elements …
text descriptions---an especially challenging task when prompts feature imaginary elements …
Alchemist: Parametric control of material properties with diffusion models
We propose a method to control material attributes of objects like roughness metallic albedo
and transparency in real images. Our method capitalizes on the generative prior of text-to …
and transparency in real images. Our method capitalizes on the generative prior of text-to …
The chosen one: Consistent characters in text-to-image diffusion models
Recent advances in text-to-image generation models have unlocked vast potential for visual
creativity. However, the users that use these models struggle with the generation of …
creativity. However, the users that use these models struggle with the generation of …
GeneAvatar: Generic Expression-Aware Volumetric Head Avatar Editing from a Single Image
Recently we have witnessed the explosive growth of various volumetric representations in
modeling animatable head avatars. However due to the diversity of frameworks there is no …
modeling animatable head avatars. However due to the diversity of frameworks there is no …