ABSTRACT
Research on visual attention in 360° content is crucial to understand how people perceive and interact with this immersive type of content and to develop efficient techniques for processing, encoding, delivering and rendering. And also to offer a high quality of experience to end users. The availability of public datasets is essential to support and facilitate research activities of the community. Recently, some studies have been presented analyzing exploration behaviors of people watching 360° videos, and a few datasets have been published. However, the majority of these works only consider head movements as proxy for gaze data, despite the importance of eye movements in the exploration of omnidirectional content. Thus, this paper presents a novel dataset of 360° videos with associated eye and head movement data, which is a follow-up to our previous dataset for still images [14]. Head and eye tracking data was obtained from 57 participants during a free-viewing experiment with 19 videos. In addition, guidelines on how to obtain saliency maps and scanpaths from raw data are provided. Also, some statistics related to exploration behaviors are presented, such as the impact of the longitudinal starting position when watching omnidirectional videos was investigated in this test. This dataset and its associated code are made publicly available to support research on visual attention for 360° content.
- Zoya Bylinskii, Tilke Judd, Aude Oliva, Antonio Torralba, and Frédo Durand. 2016. What do different evaluation metrics tell us about saliency models? arXiv preprint arXiv:1604.03605 (2016).Google Scholar
- Xavier Corbillon, Francesca De Simone, and Gwendal Simon. 2017. 360-Degree Video Head Movement Dataset. Proceedings of the 8th ACM on Multimedia Systems Conference - MMSys'17 (June 2017), 199--204. Google ScholarDigital Library
- Xavier Corbillon, Gwendal Simon, Alisa Devlic, and Jacob Chakareski. 2017. Viewport-adaptive navigable 360-degree video delivery. IEEE International Conference on Communications (2017).Google ScholarCross Ref
- Yu Fang, Ryoichi Nakashima, Kazumichi Matsumiya, Ichiro Kuriki, and Satoshi Shioiri. 2015. Eye-head coordination for visual cognitive processing. PloS one 10, 3 (2015), e0121035.Google ScholarCross Ref
- Brian Hu, Ishmael Johnson-Bey, Mansi Sharma, and Ernst Niebur. 2017. Head movements during visual exploration of natural images in virtual reality. In 2017 51st Annual Conference on Information Sciences and Systems (CISS). 1--6.Google ScholarCross Ref
- ITU. 2008. Subjective video quality assessment methods for multimedia applications. (April 2008).Google Scholar
- Halszka Jarodzka, Kenneth Holmqvist, and Marcus Nyström. 2010. A vector-based, multidimensional scanpath similarity measure. In Proceedings of the 2010 symposium on eye-tracking research & applications. ACM, 211--218. Google ScholarDigital Library
- Eileen Kowler. 2011. Eye movements: The past 25 years. Vision research 51, 13 (2011), 1457--1483.Google Scholar
- Benjamin J. Li, Jeremy N. Bailenson, Adam Pines, Walter J. Greenleaf, and Leanne M. Williams. 2017. A Public Database of Immersive VR Videos with Corresponding Ratings of Arousal, Valence, and Correlations between Head Movements and Self Report Measures. Frontiers in Psychology 8, DEC (Dec. 2017).Google Scholar
- Wen-Chih Lo, Ching-Ling Fan, Jean Lee, Chun-Ying Huang, Kuan-Ta Chen, and Cheng-Hsin Hsu. 2017. 360Âř Video Viewing Dataset in Head-Mounted Virtual Reality. Proceedings of the 8th ACM on Multimedia Systems Conference - MMSys'17 (June 2017), 211--216. Google ScholarDigital Library
- G. Marmitt and AT. T Duchowski. 2002. Modeling visual attention in vr: Measuring the accuracy of predicted scanpaths. In Eurographics 2002, Short Presentations. Saarbrücken, Germany, 217--226.Google Scholar
- Gerd Marmitt and Andrew T Duchowski. 2002. Modeling visual attention in VR: Measuring the accuracy of predicted scanpaths. Ph.D. Dissertation. Clemson University.Google Scholar
- Margaret H. Pinson, Lark Kwon Choi, and Alan Conrad Bovik. 2014. Temporal Video Quality Model Accounting for Variable Frame Delay Distortions. IEEE Transactions on Broadcasting 60, 4 (Dec. 2014), 637--649.Google ScholarCross Ref
- Yashas Rai, Jesús Gutiérrez, and Patrick Le Callet. 2017. A dataset of head and eye movements for 360 degree images. In Proceedings of the 8th ACM Multimedia Systems Conference, MMSys 2017. Google ScholarDigital Library
- Yashas Rai, Patrick Le Callet, and Gene Cheung. 2016. Quantifying the relation between perceived interest and visual salience during free viewing using trellis based optimization. In Image, Video, and Multidimensional Signal Processing Workshop (IVMSP), 2016 IEEE 12th. IEEE, 1--5.Google Scholar
- Yashas Rai, Patrick Le Callet, and Philippe Guillotel. 2017. Which saliency weighting for omni directional image quality assessment?. In Quality of Multimedia Experience (QoMEX), 2017 Ninth International Conference on. IEEE, 1--6.Google ScholarCross Ref
- Salient360. 2018. Special Issue. Signal Processing: Image Communication (2018). To appear.Google Scholar
- Dario D Salvucci and Joseph H Goldberg. 2000. Identifying fixations and saccades in eye-tracking protocols. In Proceedings of the 2000 symposium on Eye tracking research & applications. ACM, 71--78. Google ScholarDigital Library
- Ana Serrano, Vincent Sitzmann, Jaime Ruiz-Borau, Gordon Wetzstein, Diego Gutierrez, and Belen Masia. 2017. Movie editing and cognitive event segmentation in virtual reality video. ACM Transactions on Graphics (TOG) 36, 4 (2017), 47. Google ScholarDigital Library
- Vincent Sitzmann, Ana Serrano, Amy Pavel, Maneesh Agrawala, Diego Gutierrez, Belen Masia, and Gordon Wetzstein. 2018. Saliency in VR: How do people explore virtual environments? IEEE Transactions on Visualization and Computer Graphics (2018), 1--1. Google ScholarDigital Library
- Yu-Chuan Su, Dinesh Jayaraman, and Kristen Grauman. 2016. Pano2Vid: Automatic Cinematography for Watching 360° Videos. 1 (Dec. 2016).Google Scholar
- Evgeniy Upenik and Touradj Ebrahimi. 2017. A simple method to obtain visual attention data in head mounted virtual reality. In 2017 IEEE International Conference on Multimedia & Expo Workshops (ICMEW). 73--78.Google ScholarCross Ref
- Chenglei Wu, Zhihao Tan, Zhi Wang, and Shiqiang Yang. 2017. A Dataset for Exploring User Behaviors in VR Spherical Video Streaming. Proceedings of the 8th ACM on Multimedia Systems Conference - MMSys'17 (June 2017), 193--198. Google ScholarDigital Library
- Matt Yu, Haricharan Lakshman, and Bernd Girod. 2015. A Framework to Evaluate Omnidirectional Video Coding Schemes. In 2015 IEEE International Symposium on Mixed and Augmented Reality. 31--36. Google ScholarDigital Library
Index Terms
- A dataset of head and eye movements for 360° videos
Recommendations
360-Degree Video Head Movement Dataset
MMSys'17: Proceedings of the 8th ACM on Multimedia Systems ConferenceWhile Virtual Reality applications are increasingly attracting the attention of developers and business analysts, the behaviour of users watching 360-degree (i.e. omnidirectional) videos has not been thoroughly studied yet. This paper introduces a ...
An eye tracking dataset for point of gaze detection
ETRA '12: Proceedings of the Symposium on Eye Tracking Research and ApplicationsThis paper presents a new, publicly available eye tracking dataset, aimed to be used as a benchmark for Point of Gaze (PoG) detection algorithms. The dataset consists of a set of videos recording the eye motion of human test subjects as they were ...
Compensation of head movements in mobile eye-tracking data using an inertial measurement unit
UbiComp '14 Adjunct: Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct PublicationAnalysis of eye movements recorded with a mobile eye-tracker is difficult since the eye-tracking data are severely affected by simultaneous head and body movements. Automatic analysis methods developed for remote-, and tower-mounted eye-trackers do not ...
Comments