A Viewport-driven Multi-metric Fusion Approach for 360-Degree Video Quality Assessment
We propose a new viewport-based multi-metric fusion (MMF) approach for visual quality assessment of 360-degree (omnidirectional) videos. Our method is based on computing multiple spatio-temporal objective quality metrics (features) on viewports extracted from 360-degree videos, and learning a model that combines these features into a metric that closely matches subjective quality scores. The main motivations for the proposed method are that: 1) quality metrics computed on viewports better captures the user experience than metrics computed on the projection domain; 2) no individual objective image quality metric always performs the best for all types of visual distortions, while a learned combination of them is able to adapt to different conditions. Experimental results, based on the largest available 360-degree videos quality dataset, demonstrate that the proposed metric outperforms state-of-the-art 360-degree and 2D video quality metrics.
2019_12_Azevedo_ICME_360MMF.pdf
openaccess
4.2 MB
Adobe PDF
1f02a022ed971ec2bfbd6526eee13f9b
2019_12_ICME_360_Multi_metrics.pdf
openaccess
4.2 MB
Adobe PDF
5c86600782aa0e2a50a522857e2f4f02