Research Article
On the Consistency of 360 Video Quality Assessment in Repeated Subjective Tests: A Pilot Study
@ARTICLE{10.4108/eetinis.v11i1.4323, author={Majed Elwardy and Hans-Juergen Zepernick and Thi My Chinh Chu and Yan Hu}, title={On the Consistency of 360 Video Quality Assessment in Repeated Subjective Tests: A Pilot Study}, journal={EAI Endorsed Transactions on Industrial Networks and Intelligent Systems}, volume={11}, number={1}, publisher={EAI}, journal_a={INIS}, year={2024}, month={1}, keywords={360 video, subjective tests, quality of experience, quality assessment, pilot study, annotated dataset, opportunity-limited conditions, standing viewing, seated viewing}, doi={10.4108/eetinis.v11i1.4323} }
- Majed Elwardy
Hans-Juergen Zepernick
Thi My Chinh Chu
Yan Hu
Year: 2024
On the Consistency of 360 Video Quality Assessment in Repeated Subjective Tests: A Pilot Study
INIS
EAI
DOI: 10.4108/eetinis.v11i1.4323
Abstract
Immersive media such as virtual reality, augmented reality, and 360◦ video have seen tremendous technological developments in recent years. Furthermore, the advances in head-mounted displays (HMDs) offer the users increased immersive experiences compared to conventional displays. To develop novel immersive media systems and services that satisfy the expectations of the users, it is essential to conduct subjective tests revealing users’ perceived quality of immersive media. However, due to the new viewing dimensions provided by HMDs and the potential of interacting with the content, a wide range of subjective tests are required to understand the many aspects of user behavior in and quality perception of immersive media. The ground truth obtained by such subjective tests enable the development of optimized immersive media systems that fulfill the expectations of the users. This article focuses on the consistency of 360◦ video quality assessment to reveal whether users’ subjective quality assessment of such immersive visual stimuli changes fundamentally over time or is kept consistent with each user having their own behavior signature. A pilot study was conducted under pandemic conditions with participants given the task of rating the quality of 360◦ video stimuli on an HMD in standing and seated viewing. The choice of conducting a pilot study is motivated by the fact that immersive media impose high cognitive load on the participants and the need to keep the number of participants under pandemic conditions as low as possible. To gain insight into the consistency of the participants’ 360◦ video assessment over time, three sessions were held for each participant and each viewing condition with long and short breaks between sessions. In particular, the opinion scores and head movements were recorded for each participant and each session in standing and seated viewing. The statistical analysis of this data leads to the conjecture that the quality rating stays consistent throughout these sessions with each participant having their own quality assessment signature. The head movements, indicating the participants’ scene exploration during the quality assessment task, also remain consistent for each participant according their individual narrower or wider scene exploration signature. These findings are more pronounced for standing viewing than for seated viewing. This work supports the role of pilot studies being a useful approach of conducting pre-tests on immersive media quality under opportunity-limited conditions and for the planning of subsequent full subjective tests with a large panel of participants. The annotated RQA360 dataset containing the data recorded in the repeated subjective tests is made publicly available to the research community.
Copyright © 2024 M. Elwardy et al., licensed to EAI. This is an open access article distributed under the terms of the CC BY-NC-SA 4.0, which permits copying, redistributing, remixing, transformation, and building upon the material in any medium so long as the original work is properly cited.