Improved Deep Distributed Light Field Coding

Light fields enable increasing the degree of realism and immersion of visual experience by capturing a scene with a higher number of dimensions than conventional 2D imaging. On another side, higher dimensionality entails significant storage and transmission overhead compared to traditional video. Conventional coding schemes achieve high coding gains by employing an asymmetric codec design, where the encoder is significantly more complex than the decoder. However, in the case of light fields, the communication and processing among different cameras could be expensive, and the possibility of trading the complexity between the encoder and the decoder becomes a desirable feature. We leverage the distributed source coding paradigm to effectively reduce the encoder’s complexity at the cost of increased computation at the decoder side. Specifically, we train two deep neural networks to improve the two most critical parts of a distributed source coding scheme: the prediction of side information and the estimation of the uncertainty in the prediction. Experiments show considerable BD-rate gains, above 59% over HEVC-Intra and 17.45% over our previous method DLFC-I.

[1]  Stefan B. Williams,et al.  Decoding, Calibration and Rectification for Lenselet-Based Plenoptic Cameras , 2013, 2013 IEEE Conference on Computer Vision and Pattern Recognition.

[2]  Bernd Girod,et al.  Rate-adaptive codes for distributed source coding , 2006, Signal Process..

[3]  Catarina Brites,et al.  Correlation Noise Modeling for Efficient Pixel and Transform Domain Wyner–Ziv Video Coding , 2008, IEEE Transactions on Circuits and Systems for Video Technology.

[4]  Giuseppe Valenzise,et al.  View Synthesis-based Distributed Light Field Compression , 2020, 2020 IEEE International Conference on Multimedia & Expo Workshops (ICMEW).

[5]  Bernd Girod,et al.  Distributed compression for large camera arrays , 2004, IEEE Workshop on Statistical Signal Processing, 2003.

[6]  Catarina Brites,et al.  A robust fusion method for multiview distributed video coding , 2014, EURASIP J. Adv. Signal Process..

[7]  Pamela C. Cosman,et al.  Correlation noise classification based on matching success for transform domain Wyner-Ziv video coding , 2009, 2009 IEEE International Conference on Acoustics, Speech and Signal Processing.

[8]  Ting-Chun Wang,et al.  Learning-based view synthesis for light field cameras , 2016, ACM Trans. Graph..

[9]  Marco Dalai,et al.  The DISCOVER codec: Architecture, Techniques and Evaluation , 2007, PCS 2007.

[10]  Touradj Ebrahimi,et al.  New Light Field Image Dataset , 2016, QoMEX 2016.

[11]  Warnakulasuriya Anil Chandana Fernando,et al.  The VISNET II DVC codec: Architecture, tools and performance , 2010, 2010 18th European Signal Processing Conference.

[12]  Touradj Ebrahimi,et al.  Distributed Video Coding: Selecting the most promising application scenarios , 2008, Signal Process. Image Commun..

[13]  Anastasis A. Sofokleous,et al.  Review: H.264 and MPEG-4 Video Compression: Video Coding for Next-generation Multimedia , 2005, Comput. J..

[14]  Stefano Tubaro,et al.  Distributed Video Coding: Trends and Perspectives , 2009, EURASIP J. Image Video Process..

[15]  Leon A. Gatys,et al.  Image Style Transfer Using Convolutional Neural Networks , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[16]  Catarina Brites,et al.  Distributed video coding: Assessing the HEVC upgrade , 2015, Signal Process. Image Commun..

[17]  G. Bjontegaard,et al.  Calculation of Average PSNR Differences between RD-curves , 2001 .

[18]  Bernd Girod,et al.  Distributed Video Coding , 2005, Proceedings of the IEEE.

[19]  Bernd Girod,et al.  Wyner-Ziv coding of light fields for random access , 2004, IEEE 6th Workshop on Multimedia Signal Processing, 2004..

[20]  Xin Huang,et al.  Cross-band noise model refinement for transform domain Wyner-Ziv video coding , 2012, Signal Process. Image Commun..

[21]  Peter Schelkens,et al.  Performance analysis of JPEG Pleno light field coding , 2019, Optical Engineering + Applications.

[22]  Peter Schelkens,et al.  JPEG Pleno light field coding technologies , 2019, Optical Engineering + Applications.

[23]  Xin Huang,et al.  Side Information and Noise Learning for Distributed Video Coding Using Optical Flow and Clustering , 2012, IEEE Transactions on Image Processing.

[24]  Xin Huang,et al.  Improved side information generation for Distributed Video Coding , 2008, 2008 IEEE 10th Workshop on Multimedia Signal Processing.

[25]  Cristian Perra,et al.  Data formats for high efficiency coding of Lytro-Illum light fields , 2015, 2015 International Conference on Image Processing Theory, Tools and Applications (IPTA).

[26]  Gary J. Sullivan,et al.  Overview of the High Efficiency Video Coding (HEVC) Standard , 2012, IEEE Transactions on Circuits and Systems for Video Technology.

[27]  Stuart William Perry,et al.  Adaptive Content Frame Skipping for Wyner–Ziv-Based Light Field Image Compression , 2020, Electronics.

[28]  Ravi Ramamoorthi,et al.  Learning to Synthesize a 4D RGBD Light Field from a Single Image , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[29]  Edgar Martínez-Moro,et al.  An Introduction to LDPC Codes , 2013 .

[30]  Wen Gao,et al.  Free viewpoint switching in multi-view video streaming using Wyner-Ziv video coding , 2006, Electronic Imaging.

[31]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[32]  Vladlen Koltun,et al.  Photographic Image Synthesis with Cascaded Refinement Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).

[33]  Thomas M. Cover,et al.  Elements of Information Theory , 2005 .

[34]  Neus Sabater,et al.  Learning occlusion-aware view synthesis for light fields , 2019, Pattern Analysis and Applications.

[35]  Christine Guillemot,et al.  Optimal Reconstruction in Wyner-Ziv Video Coding with Multiple Side Information , 2007, 2007 IEEE 9th Workshop on Multimedia Signal Processing.

[36]  Alex Kendall,et al.  What Uncertainties Do We Need in Bayesian Deep Learning for Computer Vision? , 2017, NIPS.