Fast gaze-contingent optimal decompositions for multifocal displays

As head-mounted displays (HMDs) commonly present a single, fixed-focus display plane, a conflict can be created between the vergence and accommodation responses of the viewer. Multifocal HMDs have long been investigated as a potential solution in which multiple image planes span the viewer's accommodation range. Such displays require a scene decomposition algorithm to distribute the depiction of objects across image planes, and previous work has shown that simple decompositions can be achieved in real-time. However, recent optimal decompositions further improve image quality, particularly with complex content. Such decompositions are more computationally involved and likely require better alignment of the image planes with the viewer's eyes, which are potential barriers to practical applications. Our goal is to enable interactive optimal decomposition algorithms capable of driving a vergence- and accommodation-tracked multifocal testbed. Ultimately, such a testbed is necessary to establish the requirements for the practical use of multifocal displays, in terms of computational demand and hardware accuracy. To this end, we present an efficient algorithm for optimal decompositions, incorporating insights from vision science. Our method is amenable to GPU implementations and achieves a three-orders-of-magnitude speedup over previous work. We further show that eye tracking can be used for adequate plane alignment with efficient image-based deformations, adjusting for both eye rotation and head movement relative to the display. We also build the first binocular multifocal testbed with integrated eye tracking and accommodation measurement, paving the way to establish practical eye tracking and rendering requirements for this promising class of display. Finally, we report preliminary results from a pilot user study utilizing our testbed, investigating the accommodation response of users to dynamic stimuli presented under optimal decomposition.

[1]  D. A. Owens A comparison of accommodative responsiveness and contrast sensitivity for sinusoidal gratings , 1980, Vision Research.

[2]  David H. Brainard,et al.  Calibration of a computer controlled color monitor , 1989 .

[3]  Christopher D. Saunter,et al.  Dynamic lens and monovision 3D displays to improve viewer comfort , 2015, Optics express.

[4]  James F. O'Brien,et al.  Optimal presentation of imagery with focus cues on multi-plane displays , 2015, ACM Trans. Graph..

[5]  Bahram Javidi,et al.  A 3D integral imaging optical see-through head-mounted display. , 2014, Optics express.

[6]  Yue Liu,et al.  Light field head-mounted display with correct focus cue using micro structure array , 2014 .

[7]  Nikhil Balram,et al.  Design and optimization of a near-eye multifocal display system for augmented reality , 2015 .

[8]  오윤식,et al.  Volumetric Three-dimensional Display Systems , 2001 .

[9]  Wolfgang Heidrich,et al.  HDR-VDP-2: a calibrated visual metric for visibility and quality predictions in all luminance conditions , 2011, ACM Trans. Graph..

[10]  Nikhil Balram,et al.  Content-adaptive focus configuration for near-eye multi-focal displays , 2016, 2016 IEEE International Conference on Multimedia and Expo (ICME).

[11]  Gordon Wetzstein,et al.  Accommodation-invariant computational near-eye displays , 2017, ACM Trans. Graph..

[12]  Kevin J. MacKenzie,et al.  Vergence and accommodation to multiple-image-plane stereoscopic displays: "real world" responses with practical image-plane separations? , 2012, J. Electronic Imaging.

[13]  Sheng Liu,et al.  A systematic method for designing depth-fused multi-focal plane three-dimensional displays. , 2010, Optics express.

[14]  James Gao,et al.  High-speed switchable lens enables the development of a volumetric stereoscopic display. , 2009, Optics express.

[15]  David M. Hoffman,et al.  The zone of comfort: Predicting visual discomfort with stereo displays. , 2011, Journal of vision.

[16]  Martin S. Banks,et al.  Creating effective focus cues in multi-plane 3D displays , 2011, Optics express.

[17]  Gregory Kramida,et al.  Resolving the Vergence-Accommodation Conflict in Head-Mounted Displays , 2016, IEEE Transactions on Visualization and Computer Graphics.

[18]  Johannes Burge,et al.  Optimal defocus estimation in individual natural images , 2011, Proceedings of the National Academy of Sciences.

[19]  王涌天,et al.  Light field head-mounted display with correct focus cue using micro structure array , 2014 .

[20]  Erik Blaser,et al.  Retinal blur and the perception of egocentric distance. , 2010, Journal of vision.

[21]  Karol Myszkowski,et al.  Wide Field Of View Varifocal Near-Eye Display Using See-Through Deformable Membrane Mirrors , 2017, IEEE Transactions on Visualization and Computer Graphics.

[22]  Douglas Lanman,et al.  Near-eye light field displays , 2013, SIGGRAPH '13.

[23]  Kevin J. MacKenzie,et al.  Real-world stereoscopic performance in multiple-focal-plane displays: How far apart should the image planes be? , 2012, Electronic Imaging.

[24]  Marcus A. Magnor,et al.  An Affordable Solution for Binocular Eye Tracking and Calibration in Head-mounted Displays , 2015, ACM Multimedia.

[25]  Kevin J. MacKenzie,et al.  Accommodation to multiple-focal-plane displays: Implications for improving stereoscopic displays and for accommodation control. , 2010, Journal of vision.

[26]  J. C. Kotulak,et al.  A computational model of the error detector of human visual accommodation , 1986, Biological Cybernetics.

[27]  Hakan Urey,et al.  State of the Art in Stereoscopic and Autostereoscopic Displays , 2011, Proceedings of the IEEE.

[28]  Emily A. Cooper,et al.  Blur and Disparity Are Complementary Cues to Depth , 2012, Current Biology.

[29]  Humza J. Tahir,et al.  The impact of higher-order aberrations on the strength of directional signals produced by accommodative microfluctuations. , 2014, Journal of vision.

[30]  Gordon Wetzstein,et al.  The light field stereoscope , 2015, ACM Trans. Graph..

[31]  Hong Hua,et al.  High-resolution optical see-through multi-focal-plane head-mounted display using freeform optics. , 2014, Optics express.

[32]  Douglas Lanman,et al.  Focal surface displays , 2017, ACM Trans. Graph..

[33]  Gordon Wetzstein,et al.  Novel Optical Configurations for Virtual Reality: Evaluating User Preference and Performance with Focus-tunable and Monovision Near-eye Displays , 2016, CHI.

[34]  Sheng Liu,et al.  A Novel Prototype for an Optical See-Through Head-Mounted Display with Addressable Focus Cues , 2010, IEEE Transactions on Visualization and Computer Graphics.

[35]  Michael Potmesil,et al.  A lens and aperture camera model for synthetic image generation , 1981, SIGGRAPH '81.

[36]  Barry G. Blundell,et al.  Volumetric Three-Dimensional Display Systems , 2000 .

[37]  Rahul Narain,et al.  Blur and the perception of depth at occlusions. , 2016, Journal of vision.

[38]  Qiang Ji,et al.  In the Eye of the Beholder: A Survey of Models for Eyes and Gaze , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[39]  Junzhong Liang,et al.  Objective measurement of wave aberrations of the human eye with the use of a Hartmann-Shack wave-front sensor. , 1994, Journal of the Optical Society of America. A, Optics, image science, and vision.

[40]  Pat Hanrahan,et al.  Achieving Near-Correct Focus Cues Using Multiple Image Planes , 2004 .

[41]  Martin S. Banks,et al.  A stereo display prototype with multiple focal distances , 2004, ACM Trans. Graph..

[42]  Alexei A. Goon,et al.  Multifocal planes head-mounted displays. , 2000, Applied optics.

[43]  Douglas Lanman,et al.  Near-eye light field displays , 2013, SIGGRAPH Emerging Technologies.

[44]  Benjamin Watson,et al.  Using texture maps to correct for optical distortion in head-mounted displays , 1995, Proceedings Virtual Reality Annual International Symposium '95.

[45]  David M. Hoffman,et al.  Vergence-accommodation conflicts hinder visual performance and cause visual fatigue. , 2008, Journal of vision.

[46]  M. Alpern,et al.  Variability of accommodation during steady fixation at various levels of illuminance. , 1958, Journal of the Optical Society of America.

[47]  James F. O'Brien,et al.  Using blur to affect perceived distance and size , 2010, TOGS.

[48]  George Drettakis,et al.  Accommodation and Comfort in Head-Mounted Displays , 2018 .

[49]  Andrew W. Fitzgibbon,et al.  An automated calibration method for non-see-through head mounted displays , 2011, Journal of Neuroscience Methods.

[50]  Hans-Peter Seidel,et al.  GazeStereo3D: seamless disparity manipulations , 2016, ACM Trans. Graph..

[51]  Gordon Wetzstein,et al.  Optimizing virtual reality for all users through gaze-contingent and adaptive focus displays , 2017, Proceedings of the National Academy of Sciences.