Literature DB >> 17001990

A convolutional neural network approach for objective video quality assessment.

Patrick Le Callet1, Christian Viard-Gaudin, Dominique Barba.   

Abstract

This paper describes an application of neural networks in the field of objective measurement method designed to automatically assess the perceived quality of digital videos. This challenging issue aims to emulate human judgment and to replace very complex and time consuming subjective quality assessment. Several metrics have been proposed in literature to tackle this issue. They are based on a general framework that combines different stages, each of them addressing complex problems. The ambition of this paper is not to present a global perfect quality metric but rather to focus on an original way to use neural networks in such a framework in the context of reduced reference (RR) quality metric. Especially, we point out the interest of such a tool for combining features and pooling them in order to compute quality scores. The proposed approach solves some problems inherent to objective metrics that should predict subjective quality score obtained using the single stimulus continuous quality evaluation (SSCQE) method. This latter has been adopted by video quality expert group (VQEG) in its recently finalized reduced referenced and no reference (RRNR-TV) test plan. The originality of such approach compared to previous attempts to use neural networks for quality assessment, relies on the use of a convolutional neural network (CNN) that allows a continuous time scoring of the video. Objective features are extracted on a frame-by-frame basis on both the reference and the distorted sequences; they are derived from a perceptual-based representation and integrated along the temporal axis using a time-delay neural network (TDNN). Experiments conducted on different MPEG-2 videos, with bit rates ranging 2-6 Mb/s, show the effectiveness of the proposed approach to get a plausible model of temporal pooling from the human vision system (HVS) point of view. More specifically, a linear correlation criteria, between objective and subjective scoring, up to 0.92 has been obtained on a set of typical TV videos.

Entities:  

Mesh:

Year:  2006        PMID: 17001990     DOI: 10.1109/TNN.2006.879766

Source DB:  PubMed          Journal:  IEEE Trans Neural Netw        ISSN: 1045-9227


  4 in total

1.  Deep Adaptive Log-Demons: Diffeomorphic Image Registration with Very Large Deformations.

Authors:  Liya Zhao; Kebin Jia
Journal:  Comput Math Methods Med       Date:  2015-05-18       Impact factor: 2.238

2.  Objective Video Quality Assessment Based on Machine Learning for Underwater Scientific Applications.

Authors:  José-Miguel Moreno-Roldán; Miguel-Ángel Luque-Nieto; Javier Poncela; Pablo Otero
Journal:  Sensors (Basel)       Date:  2017-03-23       Impact factor: 3.576

3.  Keratoconus detection using deep learning of colour-coded maps with anterior segment optical coherence tomography: a diagnostic accuracy study.

Authors:  Kazutaka Kamiya; Yuji Ayatsuka; Yudai Kato; Fusako Fujimura; Masahide Takahashi; Nobuyuki Shoji; Yosai Mori; Kazunori Miyata
Journal:  BMJ Open       Date:  2019-09-27       Impact factor: 2.692

4.  Testing the ability of unmanned aerial systems and machine learning to map weeds at subfield scales: a test with the weed Alopecurus myosuroides (Huds).

Authors:  James Pt Lambert; Dylan Z Childs; Rob P Freckleton
Journal:  Pest Manag Sci       Date:  2019-05-21       Impact factor: 4.845

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.