[HTML][HTML] Graph convolutional networks: a comprehensive review

S Zhang, H Tong, J Xu, R Maciejewski - Computational Social Networks, 2019 - Springer
Graphs naturally appear in numerous application domains, ranging from social analysis,
bioinformatics to computer vision. The unique capability of graphs enables capturing the …

Generative adversarial networks for face generation: A survey

A Kammoun, R Slama, H Tabia, T Ouni… - ACM Computing …, 2022 - dl.acm.org
Recently, generative adversarial networks (GANs) have progressed enormously, which
makes them able to learn complex data distributions in particular faces. More and more …

Diffusionclip: Text-guided diffusion models for robust image manipulation

G Kim, T Kwon, JC Ye - … of the IEEE/CVF conference on …, 2022 - openaccess.thecvf.com
Recently, GAN inversion methods combined with Contrastive Language-Image Pretraining
(CLIP) enables zero-shot image manipulation guided by text prompts. However, their …

Sdedit: Guided image synthesis and editing with stochastic differential equations

C Meng, Y He, Y Song, J Song, J Wu, JY Zhu… - arXiv preprint arXiv …, 2021 - arxiv.org
Guided image synthesis enables everyday users to create and edit photo-realistic images
with minimum effort. The key challenge is balancing faithfulness to the user input (eg, hand …

Towards universal fake image detectors that generalize across generative models

U Ojha, Y Li, YJ Lee - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
With generative models proliferating at a rapid rate, there is a growing need for general
purpose fake image detectors. In this work, we first show that the existing paradigm, which …

Dense text-to-image generation with attention modulation

Y Kim, J Lee, JH Kim, JW Ha… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Existing text-to-image diffusion models struggle to synthesize realistic images given dense
captions, where each text prompt provides a detailed description for a specific image region …

Taming transformers for high-resolution image synthesis

P Esser, R Rombach, B Ommer - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Designed to learn long-range interactions on sequential data, transformers continue to show
state-of-the-art results on a wide variety of tasks. In contrast to CNNs, they contain no …

Peco: Perceptual codebook for bert pre-training of vision transformers

X Dong, J Bao, T Zhang, D Chen, W Zhang… - Proceedings of the …, 2023 - ojs.aaai.org
This paper explores a better prediction target for BERT pre-training of vision transformers.
We observe that current prediction targets disagree with human perception judgment. This …

Contrastive learning for unpaired image-to-image translation

T Park, AA Efros, R Zhang, JY Zhu - … , Glasgow, UK, August 23–28, 2020 …, 2020 - Springer
In image-to-image translation, each patch in the output should reflect the content of the
corresponding patch in the input, independent of domain. We propose a straightforward …

De-fake: Detection and attribution of fake images generated by text-to-image generation models

Z Sha, Z Li, N Yu, Y Zhang - Proceedings of the 2023 ACM SIGSAC …, 2023 - dl.acm.org
Text-to-image generation models that generate images based on prompt descriptions have
attracted an increasing amount of attention during the past few months. Despite their …