You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Code for CVPR'2022 paper ✨ "Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language Model"
The code relies on the PaddleGAN (PaddlePaddle implementation of StyleGAN2).
Download the pre-trained StyleGAN2 generator from here.
We provided several pretrained PPE models on here.
Invert real images:
The mapper is trained on latent vectors, so it is necessary to invert images into latent space.
To edit human face, StyleCLIP provides the CelebA-HQ that was inverted by e4e:
test set.
Usage:
Please first put downloaded pretraiend models and data on ckpt folder.
Inference
In PaddlePaddle version, we only provide inference code to generate editing results:
python mapper/evaluate.py
Reference
@article{xu2022ppe,
author = {Zipeng Xu and Tianwei Lin and Hao Tang and Fu Li and Dongliang He and Nicu Sebe and Radu Timofte and Luc Van Gool and Errui Ding},
title = {Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language Model},
journal = {arXiv preprint arXiv:2111.13333},
year = {2021}
}
Code for CVPR'2022 paper ✨ "Predict, Prevent, and Evaluate: Disentangled Text-Driven Image Manipulation Empowered by Pre-Trained Vision-Language Model"