Supporting Multitasking in Video Conferencing using Gaze Tracking and On-Screen Activity Detection

The use of videoconferencing in the workplace has been steadily growing. While multitasking during video conferencing is often necessary, it is also viewed as impolite and sometimes unacceptable. One potential contributor to negative attitudes towards such multitasking is the disrupted sense of eye contact that occurs when an individual shifts their gaze away to another screen, for example, in a dual-monitor setup, common in office settings. We present an approach to improve a sense of eye contact over videoconferencing in dual-monitor setups. Our approach uses computer vision and desktop activity detection to dynamically choose the camera with the best view of a user's face. We describe two alternative implementations of our solution (RGB-only, and a combination of RGB and RGB-D cameras). We then describe results from an online experiment that shows the potential of our approach to significantly improve perceptions of a person's politeness and engagement in the meeting.

[1]  Kentaro Toyama,et al.  Gaze Awareness for Video-Conferencing: A Software Approach , 2000, IEEE Multim..

[2]  Jean-Charles Bazin,et al.  Gaze correction with a single webcam , 2014, 2014 IEEE International Conference on Multimedia and Expo (ICME).

[3]  Wei Dong,et al.  One piece at a time: why video-based communication is better for negotiation and conflict resolution , 2012, CSCW.

[4]  Jung P. Shim,et al.  Trust in videoconferencing , 2006, CACM.

[5]  Abhishek Ranjan,et al.  Automatic camera control using unobtrusive vision and audio tracking , 2010, Graphics Interface.

[6]  Andrew F. Monk,et al.  Some advantages of video conferencing over high-quality audio conferencing: fluency and awareness of attentional focus , 1998, Int. J. Hum. Comput. Stud..

[7]  J. J. Higgins,et al.  The aligned rank transform for nonparametric factorial analyses using only anova procedures , 2011, CHI.

[8]  Yukiko I. Nakano,et al.  Estimating user's engagement from eye-gaze behaviors in human-agent conversations , 2010, IUI '10.

[9]  Roel Vertegaal,et al.  GAZE-2: conveying eye contact in group video conferencing using eye-controlled camera direction , 2003, CHI '03.

[10]  Kenji Mase,et al.  Human-Machine Cooperative Viewing System for Wide-angle Multi-view Videos , 2015, IUI Companion.

[11]  Jennifer Marlow,et al.  Taking Notes or Playing Games?: Understanding Multitasking in Video Communication , 2016, CSCW.

[12]  Massimo Zancanaro,et al.  Overt or subtle? Supporting group conversations with automatically targeted directives , 2014, IUI.