CP-VTON+: Clothing Shape and Texture Preserving Image-Based Virtual Try-On

Recently proposed Image-based virtual try-on (VTON) approaches have several challenges regarding diverse human poses and clothing styles. First, clothing warping networks often generate highly distorted and misaligned warped clothes, due to the erroneous clothingagnostic human representations, mismatches in input images for clothing-human matching, and improper regularization transform parameters. Second, blending networks can fail to retain the remaining clothes due to the wrong representation of humans and improper training loss for the composition-mask generation. We propose CP-VTON+ (Clothing shape and texture Preserving VTON) to overcome these issues, which significantly outperforms the state-ofthe-art methods, both quantitatively and qualitatively.

[1]  Liang Lin,et al.  Toward Characteristic-Preserving Image-based Virtual Try-On Network , 2018, ECCV.

[2]  Alexei A. Efros,et al.  The Unreasonable Effectiveness of Deep Features as a Perceptual Metric , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[3]  Larry S. Davis,et al.  VITON: An Image-Based Virtual Try-on Network , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.

[4]  Josef Sivic,et al.  Convolutional Neural Network Architecture for Geometric Matching , 2019, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[5]  Wojciech Zaremba,et al.  Improved Techniques for Training GANs , 2016, NIPS.

[6]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.