Uncouple Generative Adversarial Networks for Transferring Stylized Portraits to Realistic Faces

Stylized portraits widely exist in artwork or paintings. It is interesting to restore the original identity of portrait artworks, but consider that the rarity and the style diversity of these artworks, it is difficult to pair and obtain sufficient training data to restore their original identities by existing methods. Therefore, it is challenging to explore a method to restore a single stylized portrait to its original identity. Although CycleGAN can convert paintings into realistic photographs in unpaired datasets, it was not developed specifically for portraits, and photo-realistic faces require more accurate structures, thus the visual results obtained with CycleGAN are not satisfactory. In this paper, we propose Uncouple-Generative Adversarial Networks (UncGANs) for transferring stylized portraits to realistic faces. Our UncGANs framework is inspired by CariGANs to tackle the visual problem in CycleGAN for obtaining realistic faces from stylized portraits. In addition, we introduce three losses, namely, the semantic style consistency loss and the cycle consistency loss to effectively guide the training of generators and discriminators on unpaired datasets, the global and local adversarial loss ensure the consistency of appearance characteristics before and after translation, and the location consistency loss to establish the precise correspondence between the source domain and the target domain as well as assist the discriminators. Extensive experimental results and comparisons with state-of-the-art methods including Style, Deep-Image-Analogy, UNIT, MUNIT, CycleGAN, CP-GAN, and PS2-MAN demonstrate that our framework is better at generating realistic faces from stylized portraits with accurate structures and features.

[1]  拓海 杉山,et al.  “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”の学習報告 , 2017 .

[2]  Wen Gao,et al.  Mapping learning in eigenspace for harmonious caricature generation , 2006, MM '06.

[3]  Jonathon Shlens,et al.  A Learned Representation For Artistic Style , 2016, ICLR.

[4]  Xiaogang Wang,et al.  Deep Learning Face Attributes in the Wild , 2014, 2015 IEEE International Conference on Computer Vision (ICCV).

[5]  Qingyao Wu,et al.  Cartoon-to-Photo Facial Translation with Generative Adversarial Networks , 2018, ACML.

[6]  Leon A. Gatys,et al.  A Neural Algorithm of Artistic Style , 2015, ArXiv.

[7]  Jing Liao,et al.  CariGANs , 2018, ACM Trans. Graph..

[8]  Alexei A. Efros,et al.  Image-to-Image Translation with Conditional Adversarial Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[9]  Jan Kautz,et al.  Multimodal Unsupervised Image-to-Image Translation , 2018, ECCV.

[10]  Hyunsoo Kim,et al.  Learning to Discover Cross-Domain Relations with Generative Adversarial Networks , 2017, ICML.

[11]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[12]  Susan E. Brennan,et al.  From the Leonardo Archive , 2007, Leonardo.

[13]  Paul W. H. Chung,et al.  Use of Neural Networks in Automatic Caricature Generation: An Approach Based on Drawing Style Capture , 2005, IbPRIA.

[14]  Jaakko Lehtinen,et al.  Progressive Growing of GANs for Improved Quality, Stability, and Variation , 2017, ICLR.

[15]  Ping Tan,et al.  DualGAN: Unsupervised Dual Learning for Image-to-Image Translation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[16]  Takayuki Fujiwara,et al.  On KANSEI facial image processing for computerized facial caricaturing system PICASSO , 1999, IEEE SMC'99 Conference Proceedings. 1999 IEEE International Conference on Systems, Man, and Cybernetics (Cat. No.99CH37028).

[17]  Vishal M. Patel,et al.  High-Quality Facial Photo-Sketch Synthesis Using Multi-Adversarial Networks , 2017, 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018).

[18]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[19]  Sepp Hochreiter,et al.  GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium , 2017, NIPS.

[20]  Harry Shum,et al.  Example-based caricature generation with exaggeration , 2002, 10th Pacific Conference on Computer Graphics and Applications, 2002. Proceedings..

[21]  Yingtao Tian,et al.  Towards the Automatic Anime Characters Creation with Generative Adversarial Networks , 2017, ArXiv.

[22]  Jan Kautz,et al.  Unsupervised Image-to-Image Translation Networks , 2017, NIPS.

[23]  Ergun Akleman,et al.  Making caricatures with morphing , 1997, SIGGRAPH '97.

[24]  Alexei A. Efros,et al.  Toward Multimodal Image-to-Image Translation , 2017, NIPS.

[25]  Jung-Woo Ha,et al.  StarGAN: Unified Generative Adversarial Networks for Multi-domain Image-to-Image Translation , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[26]  Ergun Akleman,et al.  Making Extreme Caricatures with a New Interactive 2D Deformation Technique with Simplicial Complexes , 2006 .

[27]  Léon Bottou,et al.  Wasserstein GAN , 2017, ArXiv.

[28]  Golam Ashraf,et al.  Shape Stylized Face Caricatures , 2011, MMM.

[29]  Tsai-Yen Li,et al.  Automatic Caricature Generation by Analyzing Facial Features , 2004 .

[30]  Cynthia Rudin,et al.  PULSE: Self-Supervised Photo Upsampling via Latent Space Exploration of Generative Models , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[31]  Erik Reinhard,et al.  Human facial illustrations: Creation and psychophysical evaluation , 2004, TOGS.

[32]  Holger Winnemöller,et al.  Real-time video abstraction , 2006, ACM Trans. Graph..

[33]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[34]  Lin Gao,et al.  DeepFaceDrawing: deep generation of face images from sketches , 2020, ACM Trans. Graph..

[35]  Yuichi Yoshida,et al.  Spectral Normalization for Generative Adversarial Networks , 2018, ICLR.

[36]  Lior Wolf,et al.  Unsupervised Cross-Domain Image Generation , 2016, ICLR.

[37]  Jenn-Jier James Lien,et al.  Synthesis of Exaggerative Caricature with Inter and Intra Correlations , 2007, ACCV.

[38]  Xin Lin,et al.  Style Transfer for Anime Sketches with Enhanced Residual U-net and Auxiliary Classifier GAN , 2017, 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR).

[39]  Lior Wolf,et al.  Unsupervised Creation of Parameterized Avatars , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[40]  Sergey Ioffe,et al.  Batch Renormalization: Towards Reducing Minibatch Dependence in Batch-Normalized Models , 2017, NIPS.

[41]  Leon A. Gatys,et al.  Image Style Transfer Using Convolutional Neural Networks , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[42]  Harry Shum,et al.  PicToon: a personalized image-based cartoon system , 2002, MULTIMEDIA '02.

[43]  Thomas Brox,et al.  U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.

[44]  Jacob Abernethy,et al.  On Convergence and Stability of GANs , 2018 .

[45]  Fei Yang,et al.  Facial expression editing in video using a temporally-smooth factorization , 2012, 2012 IEEE Conference on Computer Vision and Pattern Recognition.

[46]  Inbar Mosseri,et al.  XGAN: Unsupervised Image-to-Image Translation for many-to-many Mappings , 2017, Domain Adaptation for Visual Understanding.

[47]  Cewu Lu,et al.  Combining sketch and tone for pencil drawing production , 2012, NPAR '12.

[48]  Hugo Larochelle,et al.  Modulating early visual processing by language , 2017, NIPS.

[49]  Gang Hua,et al.  Visual attribute transfer through deep image analogy , 2017, ACM Trans. Graph..

[50]  John P. Lewis,et al.  Improved automatic caricature by feature normalization and exaggeration , 2004, SIGGRAPH '04.

[51]  Li Fei-Fei,et al.  Perceptual Losses for Real-Time Style Transfer and Super-Resolution , 2016, ECCV.