Shooting a moving target: Motion-prediction-based transmission for 360-degree videos

Enabled by the rapid development of virtual reality hardware and software, 360-degree video content has proliferated. From the network perspective, 360-degree video transmission imposes significant challenges because it consumes 4 6χ the bandwidth of a regular video with the same resolution. To address these challenges, in this paper, we propose a motion-prediction-based transmission mechanism that matches network video transmission to viewer needs. Ideally, if viewer motion is perfectly known in advance, we could reduce bandwidth consumption by 80%. Practically, however, to guarantee the quality of viewing experience, we have to address the random nature of viewer motion. Based on our experimental study of viewer motion (comprising 16 video clips and over 150 subjects), we found the viewer motion can be well predicted in 100∼500ms. We propose a machine learning mechanism that predicts not only viewer motion but also prediction deviation itself. The latter is important because it provides valuable input on the amount of redundancy to be transmitted. Based on such predictions, we propose a targeted transmission mechanism that minimizes overall bandwidth consumption while providing probabilistic performance guarantees. Real-data-based evaluations show that the proposed scheme significantly reduces bandwidth consumption while minimizing performance degradation, typically a 45% bandwidth reduction with less than 0.1% failure ratio.

[1]  Joseph J. LaViola,et al.  Double exponential smoothing: an alternative to Kalman filter-based predictive tracking , 2003, IPT/EGVE.

[2]  Madhukar Budagavi,et al.  360 degrees video coding using region adaptive smoothing , 2015, 2015 IEEE International Conference on Image Processing (ICIP).

[3]  Allan G Farman,et al.  Field of view. , 2009, Oral surgery, oral medicine, oral pathology, oral radiology, and endodontics.

[4]  Parth Rajesh Desai,et al.  A Review Paper on Oculus Rift-A Virtual Reality Headset , 2014, ArXiv.

[5]  Younghui Kim,et al.  Rich360 , 2016, ACM Trans. Graph..

[6]  Steven M. LaValle,et al.  Head tracking for the Oculus Rift , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Mohamed Hefeeda,et al.  Adaptive streaming of interactive free viewpoint videos to heterogeneous clients , 2016, MMSys.

[8]  Leandros Tassiulas,et al.  Resource Allocation and Cross-Layer Control in Wireless Networks , 2006, Found. Trends Netw..

[9]  Pascal Frossard,et al.  Coding and replication co-design for interactive multiview video streaming , 2012, 2012 Proceedings IEEE INFOCOM.

[10]  Niklas Carlsson,et al.  Quality-adaptive Prefetching for Interactive Branched Video using HTTP-based Adaptive Streaming , 2014, ACM Multimedia.

[11]  Xin Liu,et al.  Viewing 360 degree videos: Motion prediction and bandwidth optimization , 2016, 2016 IEEE 24th International Conference on Network Protocols (ICNP).

[12]  Li-Yi Wei,et al.  Mapping virtual and physical reality , 2016, ACM Trans. Graph..

[13]  Brendan Iribe Oculus Rift를 이용한 체감형 게임 구현 , 2014 .

[14]  Antti Toskala,et al.  LTE for UMTS: Evolution to LTE-Advanced , 2011 .