Semantic Image Inpainting Vsing Self-Learning Encoder-Decoder and Adversarial Loss

Images are exposed to deterioration over years due to many factors. These factors may include, but not limited to, environmental factors, chemical processing, improper storage, etc. Image inpainting has gained significant attention from researchers to recover the deteriorated parts in images. In this paper, two new techniques for image inpainting techniques using Deep Convolution Neural Networks (CNN) are proposed. In the first technique, a self-tuned Encoder-Decoder architecture based on a Fully Convolution Network (FCN) is used to generate different sized blocks from non-deteriorated image dataset with L2 being used as a loss measure. On the other hand, the second technique is a two-step technique inspired from Context Encoders. In the first step, Context Encoders are trained on non-deteriorated image dataset to select blocks from training images with minimum L2 loss. In the second step, the selected block is applied to Generative Adversarial Networks (GAN) in order to improve the quality of the recovered image. Several simulation examples were made to proof that the performance self-tuned Encoder-Decoder and GAN is the same. Simulations have also shown that the proposed methods have superior performance in recovering missing regions in deteriorated images over other state-of-art techniques. Paris Street View dataset was used for training and validation to validate our results.

[1]  Camille Couprie,et al.  Semantic Segmentation using Adversarial Networks , 2016, NIPS 2016.

[2]  Pascal Frossard,et al.  Image inpainting through neural networks hallucinations , 2016, 2016 IEEE 12th Image, Video, and Multidimensional Signal Processing Workshop (IVMSP).

[3]  Minh N. Do,et al.  Semantic Image Inpainting with Perceptual and Contextual Losses , 2016, ArXiv.

[4]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[5]  Remus Brad,et al.  Image Inpainting Methods Evaluation and Improvement , 2014, TheScientificWorldJournal.

[6]  J. Elman Learning and development in neural networks: the importance of starting small , 1993, Cognition.

[7]  Yoshua. Bengio,et al.  Learning Deep Architectures for AI , 2007, Found. Trends Mach. Learn..

[8]  Yoshua Bengio,et al.  Extracting and composing robust features with denoising autoencoders , 2008, ICML '08.

[9]  Soumith Chintala,et al.  Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks , 2015, ICLR.

[10]  H. Sebastian Seung,et al.  Natural Image Denoising with Convolutional Networks , 2008, NIPS.

[11]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[12]  Alexei A. Efros,et al.  What makes Paris look like Paris? , 2015, Commun. ACM.

[13]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[14]  Gregory Shakhnarovich,et al.  Learning Representations for Automatic Colorization , 2016, ECCV.

[15]  Trevor Darrell,et al.  Caffe: Convolutional Architecture for Fast Feature Embedding , 2014, ACM Multimedia.

[16]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[17]  Andrew L. Maas Rectifier Nonlinearities Improve Neural Network Acoustic Models , 2013 .

[18]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[19]  Kristen Grauman,et al.  On-demand Learning for Deep Image Restoration , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).