View Synthesis with Kinect-Based Tracking for Motion Parallax Depth Cue on a 2D Display

Recent advancements in 3D video generation, processing, compression, and rendering increase accessibility to 3D video content. However, the majority of 3D displays available on the market belong to the stereoscopic display class and require users to wear special glasses in order to perceive depth. As an alternative, autostereoscopic displays can render multiple views without any additional equipment. The depth perception on stereoscopic and autostereoscopic displays is realized via a binocular depth cue called stereopsis. Another important depth cue, that is not exploited by autostereoscopic displays, is motion parallax which is a monocular depth cue. To enable the motion parallax effect on a 2D display, we propose to use the Kinect sensor to estimate the pose of the viewer. Based on pose of the viewer the real-time view synthesis software adjusts the view and creates the motion parallax effect on a 2D display. We believe that the proposed solution can enhance the content displayed on digital signature displays, kiosks, and other advertisement media where many users observe the content during move and use of the glasses-based 3D displays is not possible or too expensive.

[1]  A. Boev,et al.  Opengl-Based Control of Semi-Active 3D Display , 2008, 2008 3DTV Conference: The True Vision - Capture, Transmission and Display of 3D Video.

[2]  Jin Liu,et al.  An Experimental Multimedia System Allowing 3-D Visualization and Eye-Controlled Interaction Without User-Worn Devices , 1999, IEEE Trans. Multim..

[3]  Wa Wijnand IJsselsteijn,et al.  Visual discomfort in stereoscopic displays: a review , 2007, Electronic Imaging.

[4]  Lukasz Dabala,et al.  Simulated Holography Based on Stereoscopy and Face Tracking , 2014, ICCVG.

[5]  Atanas P. Gotchev,et al.  Binocular depth perception: Does head parallax help people see better in depth? , 2014, 2014 3DTV-Conference: The True Vision - Capture, Transmission and Display of 3D Video (3DTV-CON).

[6]  Mark Nawrot,et al.  Depth from motion parallax scales with eye movement gain. , 2003, Journal of vision.

[7]  A. Verri,et al.  A compact algorithm for rectification of stereo pairs , 2000 .

[8]  Christian Huitema,et al.  Real-time 3D face tracking based on active appearance model constrained by depth data , 2014, Image Vis. Comput..

[9]  Wijnand A. IJsselsteijn,et al.  Perceptual attributes of crosstalk in 3D images , 2005, Displays.

[10]  Itu-T and Iso Iec Jtc Advanced video coding for generic audiovisual services , 2010 .

[11]  James E. Cutting,et al.  Chapter 3 – Perceiving Layout and Knowing Distances: The Integration, Relative Potency, and Contextual Use of Different Information about Depth* , 1995 .

[12]  Chris Harrison,et al.  Pseudo-3D Video Conferencing with a Generic Webcam , 2008, 2008 Tenth IEEE International Symposium on Multimedia.

[13]  Janusz Sobecki,et al.  Configuration of Complex Interactive Environments , 2014, MISSI.

[14]  Toby Sharp,et al.  Real-time human pose recognition in parts from single depth images , 2011, CVPR.

[15]  Zhaozheng Yin,et al.  Improving depth perception with motion parallax and its application in teleconferencing , 2009, 2009 IEEE International Workshop on Multimedia Signal Processing.

[16]  Atanas Gotchev,et al.  Relative importance of depth cues on portable autostereoscopic display , 2010, MoViD '10.

[17]  Christoph Fehn,et al.  Depth-image-based rendering (DIBR), compression, and transmission for a new approach on 3D-TV , 2004, IS&T/SPIE Electronic Imaging.