Literature DB >> 35735972

No-Reference Quality Assessment of Authentically Distorted Images Based on Local and Global Features.

Domonkos Varga1.   

Abstract

With the development of digital imaging techniques, image quality assessment methods are receiving more attention in the literature. Since distortion-free versions of camera images in many practical, everyday applications are not available, the need for effective no-reference image quality assessment algorithms is growing. Therefore, this paper introduces a novel no-reference image quality assessment algorithm for the objective evaluation of authentically distorted images. Specifically, we apply a broad spectrum of local and global feature vectors to characterize the variety of authentic distortions. Among the employed local features, the statistics of popular local feature descriptors, such as SURF, FAST, BRISK, or KAZE, are proposed for NR-IQA; other features are also introduced to boost the performances of local features. The proposed method was compared to 12 other state-of-the-art algorithms on popular and accepted benchmark datasets containing RGB images with authentic distortions (CLIVE, KonIQ-10k, and SPAQ). The introduced algorithm significantly outperforms the state-of-the-art in terms of correlation with human perceptual quality ratings.

Entities:  

Keywords:  image statistics; no-reference image quality assessment; quality-aware features

Year:  2022        PMID: 35735972      PMCID: PMC9224559          DOI: 10.3390/jimaging8060173

Source DB:  PubMed          Journal:  J Imaging        ISSN: 2313-433X


1. Introduction

With the considerable advancements made in digital imaging and technology and the easy availability of cheap image-capturing devices, a large number of digital images are captured by non-technical users every day. As a consequence, people upload huge amounts of images and videos to the internet and extensively use streaming applications. In addition, visual information represents 85% of the information that is usable for human beings. Therefore, the quality assessment of digital images is of great importance and a hot research topic, along with several practical applications, such as benchmarking computer vision algorithms [1], monitoring the quality of network visual communication applications [2], fingerprint image evaluation [3], medical imaging applications [4], evaluating image compression [5], or denoising [6] algorithms. Image degradations may occur due to various reasons, such as noise, blurring, fading, or blocking artifacts. Further, the mentioned degradations can be introduced in all phases of the imaging process, such as acquisition, compression, transmission, decompression, storage, and display. Traditionally, image quality assessment (IQA) algorithms are divided into three distinct classes in the literature with respect to the accessibility of the reference (distortion-free) images [7], i.e., no-reference (NR), full-reference (FR), and reduced-reference (RR). As the idioms indicate, NR-IQA algorithms have absolutely no access to the reference images, FR-IQA methods have full access to the reference images, while RR-IQA approaches have partial information about the reference images. In the literature, NR-IQA is recognized as a more difficult research task than the other two classes due to the complete lack of reference images [8]. In this paper, we introduce a novel NR-IQA model that relies on the fusion of local and global image features. Namely, many NR-IQA methods [9,10,11,12] characterize digital images with global features that are computed using the whole image. This approach is effective in case of artificial distortions (such as JPEG or JPEG2000 compression noise), since they are distributed uniformly in the image. However, authentically or naturally distorted images are often contaminated by noise locally. To improve the perceptual quality estimation of authentically distorted images, we combined novel local and global feature vectors using the statistics of local feature descriptors and some powerful perceptual features. Specifically, 93 features were employed, from which, 80 were introduced for NR-IQA in this study. It was empirically proved that the statistics of local feature descriptors are quality-aware features (since they interpret local image regions from the viewpoint of a human visual system); combining them with global features results in a higher performance than the current state-of-the-art. Our proposed method is code-named as FLG-IQA, referring to the fact that it is based on the fusion of local and global features. The rest of this study is arranged as follows. In Section 2, related and previous papers are reviewed and outlined. Section 3 provides an overview of the used benchmark databases and evaluation metrics. Moreover, it describes the proposed approach in detail. Section 4 presents experimental results. Specifically, an ablation study is introduced to explain the effects of individual features. Moreover, a comparison to 12 other state-of-the-art methods is presented. Finally, this study is concluded in Section 5.

2. Literature Review

In this section, a review of the existing NR-IQA algorithms is given. For more comprehensive summaries about this topic, interested readers can refer to the studies by Zhai et al. [7], Mohammadi et al. [13], Yang et al. [14], and the book by Xu et al. [15]. NR-IQA methods can be further divided into training-free and machine learning-based classes. Our proposed method falls into the machine learning-based class; thus, this state-of-the-art study mainly focuses on this category. As the terminology indicates, machine learning-based algorithms incorporate some kind of machine learning algorithm to provide an estimation of the perceptual quality of a digital image, while training-free methods do not contain any training steps or rely on machine learning techniques using distorted images. For example, Venkatanath et al. [16] proposed the perception-based image quality evaluator (PIQE), which estimates the image qualities of distorted images using mean subtracted contrast normalized coefficients calculated at all pixel locations. In contrast, the naturalness image quality evaluator (NIQE) proposed by Mittal et al. [17] determines the distance between naturalness features extracted from the distorted image and the features obtained beforehand from distortion-free, pristine images to quantify perceptual image quality. NIQE was further developed in [18,19]. Namely, Zhang et al. [18] improved NIQE by using a Bhattacharyya-like distance [20] between the learned multivariate Gaussian model from the pristine images and those of the distorted images. In contrast, Wu et al. [19] boosted NIQE by more complex features to increase the prediction performance. Recently, Leonardi et al. [21] utilized deep features extracted from a pre-trained convolutional neural network to construct an opinion–unaware method using the correlations through the Gramian matrix between feature maps. In the literature, various types of machine learning-based NR-IQA methods can be found. Namely, many machine learning-based algorithms rely on natural scene statistics (NSS), which is a powerful tool in characterizing image distortions. The main assumption of NSS is that digital images of high quality follow a sort of statistical regularity and distorted images significantly deviate from this pattern [22]. For instance, Saad et al. [23] examined the statistics of discrete cosine transform (DCT) coefficients. Specifically, a generalized Gaussian distribution (GGD) model was fitted on the DCT coefficients and its parameters were utilized as quality-aware features and mapped onto a perceptual quality score with the help of a support vector regressor (SVR). Another line of papers focused on the wavelet transform to extract quality-aware features. For example, Moorthy and Bovik [24] carried out a wavelet transform over three scales and three orientations; similar to [23], GGDs were fitted to the subband coefficients and their parameters were used as quality-aware features. This method was further improved in [25], where the correlations across scales, subbands, and orientations were also used as quality-aware features. In contrast, Tang et al. [26] extracted quality-aware features from complex pyramid wavelet coefficients. In [27], the authors used color statistics, such as NSS. Specifically, mean subtracted contrast normalized (MSCN) coefficients were created from the color channels of different color spaces. Next, GGD was fitted on these coefficients; similar to the previously mentioned methods, its parameters were used to map onto quality scores. Similar to [27], Mittal et al. [9] applied MSCN coefficients to characterize perceptual quality but they extracted them from the spatial domain. Other proposed models [28,29,30] utilized the statistics of local binary patterns [31] to characterize image texture degradation in the presence of noise and distortions. In [32], Freitas et al. presented a performance comparison of a wide range of local binary pattern types and variants for NR-IQA. Ye and Doerman [33,34] applied the first visual codebooks for NR-IQA. More specifically, Gabor filters were applied for feature extraction and visual codebook creation. Subsequently, the perceptual quality of an image was expressed as the weighted average of the codeword quality scores. In [35], an unsupervised feature learning framework was introduced where an unlabeled codebook was compiled from raw image patches using means clustering. With the popularity of deep learning techniques and research, many deep learning-based methods were also proposed in the literature. For instance, Kang et al. [36] trained a convolutional neural network (CNN) with a convolutional layer (using both max and min pooling) and two fully-connected layers. Specifically, the proposed CNN was trained on image patches and the predicted scores of the patches were averaged to obtain the estimated quality of the whole input image. In contrast, Bianco et al. [37] introduced the DeepBIQ architecture, which extracts deep features with the help of a pre-trained CNN from multiple image patches to compile feature vectors that were mapped onto perceptual quality with a SVR. Similarly, Gao et al. [38] relied on a pre-trained CNN but multi-level feature vectors were extracted through global average pooling layers. Sheng et al. [39] utilized the fact that visual saliency is correlated with image quality [40] and trained a saliency-guided deep CNN for NR-IQA from scratch. In contrast, Liu et al. [41] introduced the learning to rank framework [42] for NR-IQA. Specifically, the authors implemented a Siamese CNN to realize this framework. Li et al. [43] implemented a novel loss function (very similar to Pearson’s linear correlation coefficient) to provide the CNN with a shorter convergence time and a better perceptual quality estimation performance. Celona and Schettini [44] trained a novel network, which processes input images at multiple scales, trained jointly by considering NR-IQA as regression, classification, and pairwise ranking, simultaneously.

3. Materials and Methods

3.1. Materials

In this subsection, the applied IQA benchmark databases are described. Further, the applied evaluation methodology and implementation environment are given.

3.1.1. Applied IQA Databases

To evaluate our proposed methods and compare them against the state-of-the-art, three publicly available databases containing RGB images with authentic distortions were utilized—CLIVE [45], KonIQ-10k [46], and SPAQ [47]. The main properties of the applied IQA databases are given in Table 1. The empirical mean opinion score (MOS) distributions of the utilized IQA benchmark databases containing authentic distortions are depicted in Figure 1. In the field of IQA, MOS corresponds to the arithmetic mean of the collected individual quality ratings. Further, those distortions are considered authentic and were introduced to the images during the daily (usually non-expert) usage of imaging devices, such as overexposure, underexposure, camera jitter, motion blur, or noises from camera vibration [48].
Table 1

Summary about the applied benchmark IQA databases with authentic distortions. DSLR: digital single-lens reflex camera. DSC: digital still camera. SPHN: smartphone.

AttributeCLIVE [45]KonIQ-10k [46]SPAQ [47]
#Images116210,07311,125
Resolution 500×500 1024×768 4000×4000
#Subjects81001,467600
#Annotations14001,200,000186,400
Scale of quality scores0–1001–50–100
Subjective methodologyCrowdsourcingCrowdsourcingLaboratory
Types of camerasDSLR/DSC/SPHNDSLR/DSC/SPHNSPHN
Year of publication201720182020
Figure 1

Empirical MOS distributions in the applied IQA databases: (a) CLIVE [45], (b) KonIQ-10k [46], and (c) SPAQ [47].

3.1.2. Evaluation

The evaluation of NR-IQA algorithms relies on the measurements of the correlations between predicted and ground truth perceptual quality scores. In the literature, Pearson’s linear correlation coefficient (PLCC), Spearman’s rank-order correlation coefficient (SROCC), and Kendall’s rank-order correlation coefficient (KROCC) are widely applied and accepted for this end [49]. As recommended by Sheikh et al. [50], non-linear mapping was carried out between the predicted and the ground truth scores before the computation of PLCC using a logistic function with five parameters, where and Q indicate the predicted and mapped scores, respectively. Further, , denote the fitting parameters. As usual, in machine learning, particularly in NR-IQA, 80% of images in a database were used for training and the remaining 20% for testing. Further, median PLCC, SROCC, and KROCC values, which were measured over 100 random train–test splits, are reported in this study. PLCC, SROCC, and KROCC can be defined between vectors and as where where N stands for the length of the vectors, and are the means of the vectors, and indicate the number of concordant and discordant pairs between and , respectively. Our experiments were carried out in MATLAB R2021a, mainly employing the functions of the image processing and machine learning and statistics toolboxes. The main characteristics of the computer configuration applied in our experiments are outlined in Table 2.
Table 2

The applied computer configuration of the experiments.

Computer modelSTRIX Z270H Gaming
Operating systemWindows 10
CPUIntel(R) Core(TM) i7-7700K CPU 4.20 GHz (8 cores)
Memory15 GB
GPUNvidia GeForce GTX 1080

3.2. Methods

A high-level summary of the proposed method is depicted in Figure 2, while Table 3 sums up the introduced and applied quality-aware features. As it can be seen from Figure 2, feature vectors were extracted from the set of training images to obtain a quality model with the help of a machine learning algorithm. Formally, our quality model is defined by , where q is the estimated quality score, , and G is a regression model that can be the Gaussian process regression (GPR) and support vector regressor (SVR). In the training phase, the model is optimized to minimize the distance between the estimated and ground truth quality scores. In the testing phase, the obtained quality model G was applied to estimate the perceptual quality of previously unseen test images. As already mentioned in the previous subsection, the evaluation of the quality model relies on measuring the correlation strength between the predicted and ground truth quality scores. Further, median PLCC, SROCC, and KROCC values measured over 100 random train–test splits are reported in this study.
Figure 2

Workflow of the proposed NR-IQA algorithm.

Table 3

Summary of the applied features. Quality-aware features proposed by this paper are in bold.

Feature NumberInputFeatureNumber of Features
f1-f5 SURF [55], Grayscale imagemean, median, std, skewness, kurtosis5
f6-f10 FAST [53], Grayscale imagemean, median, std, skewness, kurtosis5
f11-f15 BRISK [56], Grayscale imagemean, median, std, skewness, kurtosis5
f16-f20 KAZE [57], Grayscale imagemean, median, std, skewness, kurtosis5
f21-f25 ORB [58], Grayscale imagemean, median, std, skewness, kurtosis5
f26-f30 Harris [54], Grayscale imagemean, median, std, skewness, kurtosis5
f31-f35 Minimum Eigenvalue [59], Grayscale imagemean, median, std, skewness, kurtosis5
f36-f40 SURF [55], Filtered imagemean, median, std, skewness, kurtosis5
f41-f45 FAST [53], Filtered imagemean, median, std, skewness, kurtosis5
f46-f50 BRISK [56], Filtered imagemean, median, std, skewness, kurtosis5
f51-f55 KAZE [57], Filtered imagemean, median, std, skewness, kurtosis5
f56-f60 ORB [58], Filtered imagemean, median, std, skewness, kurtosis5
f61-f65 Harris [54], Filtered imagemean, median, std, skewness, kurtosis5
f66-f70 Minimum Eigenvalue [59], Filtered imagemean, median, std, skewness, kurtosis5
f71-f77 Binary imageHu invariant moments [60]7
f78-f87RGB imagePerceptual features10
f88 GL-GM map (α=0.3)histogram variance1
f89 GL-GM map (α=0.6)histogram variance1
f90 GL-GM map (α=0.9)histogram variance1
f91GM map [61]histogram variance1
f92RO map [61]histogram variance1
f93RM map [61]histogram variance1
The main (novel) contribution of the proposed method is the introduced quality-aware features outlined in Table 3. To tackle the wide variety of authentic distortions, a broad spectrum of statistical local and global features were applied. Artificial distortions (such as JPEG or JPEG2000 compression) are usually uniformly distributed in an image, which can be characterized well by global and homogeneous features [51]. However, authentic distortions often appear locally in a digital image, which can be better captured by local, non-homogeneous image features. Namely, authentic distortions are mainly introduced to the images during daily (non-expert) usage of imaging devices, such as overexposure, underexposure, camera jitter, motion blur, or noises from camera vibration [48]. In this study, the statistics of local feature descriptors supplemented by global features were introduced to construct a novel image quality model. Specifically, local feature descriptors were designed to characterize images from the human visual system’s point of view [52]. For instance, FAST [53] and Harris [54] are suitable for finding small mono-scale keypoints. On the other hand, other local feature descriptors, such as SURF [55], find multi-scale keypoints. Although they are designed to exhibit some kind of invariance against noise and illumination, perfect robustness does not exist. That is why their statistics may help characterize local image quality. Moreover, images with authentic distortions may suffer from overexposure or underexposure, which influence image content globally. The main goal of this study was to empirically prove that the fusion of local and global features can effectively estimate the perceptual quality of digital images with authentic distortions. To be specific, 93 features were applied in this study, of which, 80 were introduced for NR-IQA. The introduced features can be divided into four groups, i.e., the statistics of local feature descriptors measured on the grayscale image (f1-f35 in Table 3) and on the Prewitt-filtered image (f36-f70), Hu invariant moments computed from the binarized Sobel edge map of the input image (f66-f70), perceptual features (f78-f87), and the relative Grünwald–Letnikov derivative and gradient statistics (f88-f93). In the following subsection, each group of the features is introduced in detail. Summary of the applied features. Quality-aware features proposed by this paper are in bold.

3.3. Statistics of Local Feature Descriptors

In contrast to artificially distorted images, where distortions are uniformly distributed, natural or authentic distortions are often present locally in images. This is why the statistics of local feature descriptors are proposed for quality-aware features in this study. In this study, we calculated the statistics of the following local feature detectors: SURF [55], FAST [53], BRISK [56], KAZE [57], ORB [58], Harris [54], and minimum eigenvalue [59]. Specifically, the strongest 250 interest points were detected separately with each of the above-mentioned local feature detectors and the adherent feature vectors were determined. In this study, the following statistics of the obtained feature vectors were considered as local features: mean, median, standard deviation, skewness, and kurtosis. The skewness of a vector containing n elements is calculated as while its kurtosis can be given as where stands for the arithmetic mean of . Specifically, the statistics of the features around the feature points were calculated and their arithmetic means were considered as quality-aware features. Moreover to increase the distinctiveness of local feature detectors [62], the statistics of the local feature detectors were also extracted from the filtered version of the input image obtained by Prewitt operators [63].

3.4. Hu Invariant Moments

The moment is a projection of a function (in image processing an image ) to the polynomial basis. Formally, it can be written as where N and M are the width and the height of image , respectively. The order of the moment is defined as . It can be easily pointed out that is the mass of the image. Further, and are the coordinates of the center of gravity of the image. The central moment is defined as It is evident that . Next, the normalized central moment is given as where Hu [60] proposed seven invariant moments, which were defined using the normalized central moment, such as The human visual system is very sensitive to edge and contour information, since this information gives reliable implications about the structure of an image [64,65]. To characterize the contour information of an image, the Hu invariant moments [60] of the image’s binary Sobel edge map were used as quality-aware features. The horizontal and vertical derivative approximations of input image were determined as where ∗ is the convolution operator. Next, the gradient magnitude was computed as The binary Sobel edge map was obtained from the gradient magnitude by applying a cutoff threshold corresponding to the quadruple of ’s mean.

3.5. Perceptual Features

Some perceptual features, which are proved to be consistent with human quality judgments [66], were also built into our model. In the following, an RGB color image is denoted by I and is a color channel of input image I. Moreover, x stands for the pixel coordinate and we assume that I has N pixels. Blur: It is probably the most dominant source of perceptual image quality deterioration in digital imaging [67]. To quantify the emergence of the blur effect, the blur metric of Crété-Roffet et al. [68], which is based on the measurements of intensity variations between neighboring pixels, was implemented due to its low computational costs. Colorfulness: In [69], Choi et al. pointed out that colorfulness is a critical component in human image quality judgment. We determined colorfulness using the following formula proposed by Hasler and Suesstrunk [70]: where and stand for the standard deviation and the mean of the matrices denoted in the subscripts, respectively. Specifically, these matrices are given as: where R, G, and B are the red, green, and blue color channels, respectively. Chroma: It is one of the relevant image features among a series of color metrics in the CIELAB color space. Moreover, chroma is significantly correlated with haze, blur, or motion blur in the image [70]. It is defined as where a and b are the corresponding color channels of the CIELAB color space. The arithmetic mean of was used as a perceptual feature in our model. Color gradient: The estimated color gradient magnitude (CGM) map is defined as where and stand for the approximate directional derivatives in the horizontal x and vertical y directions of , respectively. In our study, the mean and standard deviations of are utilized as quality-aware features. Dark channel feature (DCF): In the literature, Tang et al. [71] proposed DCF [72] for image quality assessment, since it can effectively identify haze effects in images. A dark channel is defined as where denotes the image patches around the pixel location x. In our implementation, an image patch corresponds to a square. Next, the DCF is defined as where is the size of image I. Michelson contrast: Contrast is one of the most fundamental characteristics of an image, since it influences the ability to distinguish objects from each other in an image [73]. Thus, contrast information is built into our NR-IQA model. The Michelson contrast measures the difference between the maximum and minimum values of an image [74], defined as Root mean square (RMS) contrast is defined as where denotes the mean luminance of . Global contrast factor (GCF): Contrary to Michelson and RMS contrasts, GCF considers multiple resolution levels of an image to estimate human contrast perception [75]. It is defined as where ’s are the average local contrasts and ’s are the weighting factors. The authors examined nine different resolution levels, which is why the number of weighting factors are nine; ’s are defined as which is a result of an optimum approximation from the best fitting [75]. The local contrasts are defined as follows. First, the image of size is rearranged into a one-dimensional vector using row-wise sorting. Next, the local contrast in pixel location i is defined as where denotes the pixel value at location i after gamma correction . Finally, the average local contrast at resolution i (denoted by in Equation (31)) is determined as the average of all ’s over the entire image. Entropy: It is a quantitative measure of the image’s carried information [76]. Typically, an image with better quality is able to transmit more information. This is why entropy was chosen as a quality-aware feature. The entropy of a grayscale image is defined as where consists of the normalized histogram counts of the grayscale image. Table 4 illustrates the average values of the perceptual features in five equal MOS intervals of CLIVE [45] from very low image quality to very high image quality. From these numerical results, it can be observed that the mean values of the applied perceptual features are roughly proportional with the perceptual quality class. For instance, the mean values of several perceptual features (the mean and the standard deviations of the color gradient, Michelson contrast, RMS contrast, GCF, and entropy) monotonically increase with the quality class. Similarly, Table 5 illustrates the standard deviation values of the perceptual features in five equal MOS intervals of CLIVE [45] from very low image quality to very high image quality. It can be seen that the standard deviation values are also roughly proportional to the perceptual quality classes. For instance, the standard deviation values of several perceptual features (color gradient-mean, DCF, Michelson contrast, RMS contrast, and entropy) exhibit a remarkable proportionality with the perceptual quality classes.
Table 4

Mean values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.

0MOS<20 20MOS<40 40MOS<60 60MOS<80 80MOS100
Blur0.4120.3620.3150.2850.329
Colorfulness0.0460.0380.0420.0450.072
Chroma15.51013.68114.99515.40921.977
Color gradient-mean92.801116.884154.651189.795196.287
Color gradient-std132.693163.876207.837244.420235.855
DCF0.2170.2110.1970.2200.192
Michelson contrast2.8042.8322.9112.9372.953
RMS contrast0.2010.2010.2190.2220.223
GCF5.3045.4886.6026.2646.796
Entropy6.8326.9857.1827.4137.583
Table 5

Standard deviation values of perceptual features in CLIVE [45] with respect to five equal MOS intervals.

0MOS<20 20MOS<40 40MOS<60 60MOS<80 80MOS100
Blur0.1090.0960.0750.0670.093
Colorfulness0.0500.0330.0370.0390.049
Chroma12.6988.1439.6808.92711.720
Color gradient-mean45.48066.16489.76296.28399.800
Color gradient-std58.23671.18782.10484.17978.250
DCF0.1410.1220.1170.1150.105
Michelson contrast0.3280.2520.1730.1430.140
RMS contrast0.0800.0680.0650.0560.051
GCF1.9341.6651.7611.8571.746
Entropy1.0190.9660.7480.5320.227

3.6. Relative Grünwald–Letnikov Derivative and Gradient Statistics

In image processing, image gradient is one of the most widely used features [77,78,79,80] and a strong predictive factor for image quality [61,81]. To characterize gradient degradation in the presence of image noise, the idea of gradient magnitude (GM), relative gradient orientation (RO), and relative gradient magnitude (RM) maps were applied from [61], on the one hand. On the other hand, the idea of GM maps was generalized and developed further using the Grünwald–Letnikov (GL) derivative [82]. Once, GM, RO, and RM maps were computed—following the recommendations of [61]—their histogram variances were used as quality-aware features. Given a normalized histogram , the histogram variance is defined as where n is the length of and is the mean of . The Grünwald–Letnikov derivative, introduced by Anton Karl Grünwald and Aleksey Vasilievich Letnikov, enables differentiating a function a non-integer amount of times [83]. Generally, a one-dimensional function can be differentiated for any using the following formula where Grünwald and Letnikov invented an approach, which enables taking the derivative of a function by arbitrary, non-integer -times. Formally, it has been written where is the Gamma-function, is the th order Grünwald–Letnikov derivative, and x and stand for the upper and lower bounds, respectively. For a discrete two-dimensional signal (in the context of image processing: a digital image), , the GL derivative in the x-direction can be defined as Similarly, in the y-direction Similar to the traditional definition, the GL derivate can be given as Although the physical meaning of the GL fractional derivative is not absolutely understandable, it is important to notice that GL derivatives, in contrast to integer derivatives, do not have local characters [84]. Namely, their derivatives depend on the entire functions. In previous work [85], the combination of global and local variations of an image using GL derivatives (global) and image gradients (local) proved to be beneficial for FR-IQA. Note that the full-reference (FR) setting aims to evaluate an image using reference images without distortions. In this study, in order to characterize global variations of images for no reference (NR)-IQA, the computation of the GM [61] map was modified by using the equations of the GL derivative, as in Equations (39)–(41). More specifically, three GL-GM maps were computed with , , and , respectively, and their histogram variances were taken as quality-aware features. The GM map can be given very similarly to Equation (41): where and stand for the approximate directional derivatives in the horizontal x and vertical y directions of , respectively. The definition of the RO map is as follows [61]. First, the gradient orientation needs to be defined: The RO map can be given as where is the local average orientation; the authors define as where the average directional derivatives are defined as and where W stands for the local neighborhood over the values that are computed. Similar to the RO map, the RM map can be given as

4. Results

In this section, our results are presented. Specifically, Section 4.1 consists of an ablation study to analyze the performance of the proposed, individual quality-aware features. Next, Section 4.2 describes the results of a performance comparison to other state-of-the-art NR-IQA methods.

4.1. Ablation Study

In this subsection, an ablation study on CLIVE [45] is presented to reason the design choices and demonstrate the individual performances of the proposed quality-aware features. The results are summarized in Table 6. From these results, it can be observed that the statistics of feature descriptors are quality-aware features and can deliver a rather strong performance. However, the set of the applied perceptual features delivers the strongest performance. It can also be seen that combining the statistics of local feature descriptors with global features results in improved performance. Moreover, GPR with the rational quadratic kernel function outperforms SVR with the Gaussian kernel function as a regressor for all of the proposed quality-aware features. Figure 3 provides detailed results for all local feature descriptors. On their own, all local feature descriptors can provide weak or mediocre performances. However, their concatenations provide rather strong performances. We attribute this result to the ability of local feature descriptors to diversely characterize local image distortions. Based on the above observations, we used GPR with the rational quadratic kernel function as a regressor in the proposed method, which is code-named FLG-IQA, referring to the fact that it is based on the fusion of local and global features.
Table 6

Ablation study on CLIVE [45] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits.

SVRGPR
Method PLCC SROCC KROCC PLCC SROCC KROCC
Feature descriptors, RGB image0.5180.4840.3370.5780.5230.364
Feature descriptors, filtered image0.5290.4880.3380.5820.5270.364
Hu invariant moments0.3020.2950.1990.3280.3200.219
Perceptual features0.6070.5880.4200.6260.5980.425
GL and gradient statistics0.5280.4920.3430.5410.4950.343
All0.6360.6040.4280.6850.6440.466
Figure 3

Comparison of the statistics of local feature descriptors as quality-aware features in CLIVE [45]. Median SROCC values were measured over 100 random train–test splits. (a) RGB image, SVR, (b) RGB image, GPR, (c) filtered image, SVR, (d) filtered image, GPR.

To demonstrate that all features are relevant in FLG-IQA, we present an experiment based on the evaluation protocol described in Section 3.1.2 and using the CLIVE [45] database, in which a given feature is eliminated from the entire feature vector. As demonstrated in Figure 4, all features are important. If one of the applied features is removed, the performance of FLG-IQA falls back. However, the statistics of SURF, KAZE, and minimum eigenvalue local feature descriptors on the Prewitt-filtered image are the most decisive to the performance of FLG-IQA from the statistics of local feature descriptors. Moreover, perceptual features have the most contributing effects to the performance of the proposed method. If we contrast the results in Figure 4 with results in Figure 3d and Table 6, the following interesting fact can be observed. Features that have strong effects on the performance of FLG-IQA, do not have always superior performance individually. For instance, if the statistics of the minimum eigenvalue local feature descriptor in the Prewitt-filtered image are removed, a significant performance drop can be observed. On the other hand, its individual performance lags behind those of the Harris statistics on the filtered image, although Harris has a rather strong individual performance. This indicates that the statistics of local feature descriptors are quality-aware feature vectors and complement each other in NR-IQA. Further, perceptual features have strong individual performances and their removal from the feature vector throw back the performance, indicating that they are very strong predictors of the perceptual image quality.
Figure 4

Performance in terms of SROCC of the proposed FLG-IQA in cases when a given feature is removed from the proposed feature vector. The performance of the entire feature vector is indicated by ‘X’. Median SROCC values were measured on CLIVE [45] after 100 random train–test splits.

To further prove that all entries of the proposed feature vector are important, the rank importance of all predictors was investigated using the RReliefF algorithm [86]. Namely, the main idea behind RReliefF [87] is to estimate the discriminative power of features based on their ability on how well they differentiate between instances that lie near each other in the feature space. To this end, RReliefF penalizes those predictors that provide different values to neighbors with the same response values. On the other hand, it rewards those predictors that give different values to neighbors with different response values. Further, the number of examined neighbors is an input parameter of RReliefF. For all details about RReliefF, we refer to the paper by Robnik-Sikonja and Kononenko [88]. The results of the RReliefF algorithm—using and 7 nearest neighbors—on the features extracted from the images of CLIVE [45], are depicted in Figure 5. From these results, it can be seen that all entries of the proposed feature vector are important since the weights of importance are non-negative in all cases.
Figure 5

Results of the RReliefF algorithm on the features extracted from the images of CLIVE [45]. (a) nearest neighbours, (b) nearest neighbours, (c) nearest neighbours, (d) nearest neighbours.

4.2. Comparison to the State-of-the-Art

In this subsection, the proposed FLG-IQA algorithm is compared to several state-of-the-art methods, such as BLIINDS-II [23], BMPRI [12], BRISQUE [9], CurveletQA [10], DIIVINE [25], ENIQA [89], GRAD-LOG-CP [11], GWH-GLBP [90], NBIQA [91], OG-IQA [61], PIQE [16], and SSEQ [92], whose original MATLAB source codes can be found online. The above-mentioned algorithms were evaluated in the same environment with the same evaluation protocol as ours. Since PIQE [16] is a training-free method without any machine learning technique, it was directly evaluated in the full datasets. The results are summarized in Table 7 and Table 8, where it can be seen that the proposed FLG-IQA is able to outperform all other considered methods in three large IQA databases with authentic distortions, i.e., CLIVE [45], KonIQ-10k [46], and SPAQ [47]. Table 9 illustrates the direct and weighted average performance values obtained from the achieved results on the used IQA benchmark databases. It can be observed that the proposed FLG-IQA outperforms all other examined state-of-the-art algorithms by a large margin in this comparison. Specifically, FLG-IQA surpasses the second best methods by approximately 0.05 in terms of PLCC, SROCC, and KROCC in direct and weighted averages as well. In general, all methods achieved higher values in the case of the weighted average, which implies that the examined methods tend to perform better on larger databases. As an illustration of the results, Figure 6 depicts ground truth versus predicted scores in CLIVE [45] and KonIQ-10k [46] test sets, respectively.
Table 7

Comparison to the state-of-the-art in CLIVE [45] and KonIQ-10k [46] databases. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.

CLIVE [45]KonIQ-10k [46]
Method PLCC SROCC KROCC PLCC SROCC KROCC
BLIINDS-II [23]0.4730.4420.2910.5740.5750.414
BMPRI [12]0.5410.4870.3330.6370.6190.421
BRISQUE [9]0.5240.4970.3450.7070.6770.494
CurveletQA [10] 0.636 0.621 0.4210.7300.7180.495
DIIVINE [25]0.6170.5800.4050.7090.6930.471
ENIQA [89]0.5960.5640.3760.7610.745 0.544
GRAD-LOG-CP [11]0.6070.6040.3830.7050.6960.501
GWH-GLBP [90]0.5840.5590.3950.7230.6980.507
NBIQA [91]0.6290.604 0.427 0.771 0.749 0.515
OG-IQA [61]0.5450.5050.3640.6520.6350.447
PIQE [16]0.1720.1080.0810.2080.2460.172
SSEQ [92]0.4870.4360.3090.5890.5720.423
FLG-IQA 0.685 0.644 0.466 0.806 0.771 0.578
Table 8

Comparison to the state-of-the-art in the SPAQ [47] database. Median PLCC, SROCC, and KROCC values were measured over 100 random train–test splits. The best results are in bold and the second-best results are underlined.

MethodPLCCSROCCKROCC
BLIINDS-II [23]0.6760.6750.486
BMPRI [12]0.7390.7340.506
BRISQUE [9]0.7260.7200.518
CurveletQA [10]0.7930.7740.503
DIIVINE [25]0.7740.7560.514
ENIQA [89] 0.813 0.804 0.603
GRAD-LOG-CP [11]0.7860.7820.572
GWH-GLBP [90]0.8010.7960.542
NBIQA [91]0.8020.7930.539
OG-IQA [61]0.7260.7240.594
PIQE [16]0.2110.1560.091
SSEQ [92]0.7450.7420.549
FLG-IQA 0.850 0.845 0.640
Table 9

Comparison to the state-of-the-art. Direct and weighted average PLCC, SROCC, and KROCC are reported based on the results measured in CLIVE [45], KonIQ-10k [46], and SPAQ [47]. The best results are in bold and the second-best results are underlined.

Direct AverageWeighted Average
Method PLCC SROCC KROCC PLCC SROCC KROCC
BLIINDS-II [23]0.5740.5640.3970.6200.6180.443
BMPRI [12]0.6390.6130.4200.6830.6690.459
BRISQUE [9]0.6520.6310.4520.7070.6890.498
CurveletQA [10]0.7200.7040.4730.7560.7410.495
DIIVINE [25]0.7000.6760.4630.7370.7180.489
ENIQA [89]0.7230.704 0.508 0.778 0.765 0.565
GRAD-LOG-CP [11]0.6990.6940.4850.7400.7340.530
GWH-GLBP [90]0.7030.6840.4810.7550.7400.519
NBIQA [91] 0.734 0.715 0.494 0.779 0.7630.522
OG-IQA [61]0.6410.6210.4680.6830.6730.516
PIQE [16]0.1970.1700.1150.2080.1940.127
SSEQ [92]0.6070.5830.4270.6610.6500.480
FLG-IQA 0.780 0.753 0.561 0.822 0.801 0.603
Figure 6

Ground truth scores versus predicted scores in (a) CLIVE [45] and (b) KonIQ-10k [46] test sets.

To prove that the achieved performance difference against the state-of-the-art in CLIVE [45], KonIQ-10k [46], and SPAQ [47] was significant, significance tests were also carried out. Specifically, one-sided t-tests were applied between the 100 SROCC values provided by the proposed FLG-IQA method and one other examined state-of-the-art algorithm. Further, the null hypothesis was that the mean SROCC values of the two sets were equal to each other at a confidence level of . The results of the significance tests are summarized in Table 10 where symbol denotes that the proposed FLG-IQA is significantly better (worse) than the algorithm represented in the row of the table on the IQA benchmark database represented in the column. From the presented results, it can be seen that FLG-IQA is significantly better than the state-of-the-art in the utilized IQA benchmark databases containing authentic distortions.
Table 10

Results of the significance tests. Symbol denotes that the proposed FLG-IQA algorithm is significantly ( confidence interval) better (worse) than the NR-IQA algorithm in the row on the IQA benchmark database in the column.

MethodCLIVE [45]KonIQ-10k [46]SPAQ [47]
BLIINDS-II [23]111
BMPRI [12]111
BRISQUE [9]111
CurveletQA [10]111
DIIVINE [25]111
ENIQA [89]111
GRAD-LOG-CP [11]111
GWH-GLBP [90]111
NBIQA [91]111
OG-IQA [61]111
PIQE [16]111
SSEQ [92]111
The effectiveness of the proposed FLG-IQA was further proved in a cross-database test where the examined state-of-the-art algorithm and the proposed method were trained on the large KonIQ-10k [46] and tested in CLIVE [45]. The results of this test are summarized in Table 11. From the presented numerical results, it can be seen that the proposed method provides a significantly higher performance than the other methods. Specifically, FLG-IQA performs better than the second-best method bu approximately 0.11 in terms of PLCC and 0.07 in terms of SROCC, respectively. Figure 7 depicts the results of FLG-IQA in the cross database in normalized ground truth scores versus a normalized predicted score scatter plot.
Table 11

Results of the cross-database test. The examined and the proposed methods were trained on KonIQ-10k [46] and tested on CLIVE [45]. The best results are in bold and the second-best results are underlined.

MethodPLCCSROCCKROCC
BLIINDS-II [23]0.1070.0900.063
BMPRI [12]0.4530.3890.298
BRISQUE [9] 0.509 0.4600.310
CurveletQA [10]0.4960.505 0.347
DIIVINE [25]0.4790.4340.299
ENIQA [89]0.4280.3860.272
GRAD-LOG-CP [11]0.4270.3840.261
GWH-GLBP [90]0.4800.4790.328
NBIQA [91]0.503 0.509 0.284
OG-IQA [61]0.4420.4270.289
SSEQ [92]0.2700.2560.170
FLG-IQA 0.613 0.571 0.399
Figure 7

Normalized ground truth scores versus normalized predicted score scatter plot of FLG-IQA in the cross-database test.

5. Conclusions

In this paper, a novel NR-IQA method for authentically distorted images was introduced. Specifically, a diverse set of local and global quality-aware features was proposed and applied with a GPR with the rational quadratic kernel function to obtain a perceptual quality estimator. The main idea behind the usage of local feature descriptor statistics was that these feature descriptors interpret local image regions from the human visual system’s viewpoint. The features were studied by taking into consideration their effects on the performance of perceptual quality estimation. The numerical comparison to 12 other state-of-the-art methods on three popular benchmark databases (CLIVE [45], KonIQ-10k [46], and SPAQ [47]) proved the superior performance of the proposed method. Our future work will involve boosting the quality-aware properties of the local feature descriptors by applying bio-inspired filters.
  21 in total

1.  No-reference image quality assessment using visual codebooks.

Authors:  Peng Ye; David Doermann
Journal:  IEEE Trans Image Process       Date:  2012-03-06       Impact factor: 10.856

2.  Single Image Haze Removal Using Dark Channel Prior.

Authors:  Kaiming He; Jian Sun; Xiaoou Tang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-09-09       Impact factor: 6.226

3.  A statistical evaluation of recent full reference image quality assessment algorithms.

Authors:  Hamid Rahim Sheikh; Muhammad Farooq Sabir; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2006-11       Impact factor: 10.856

4.  Blind image quality assessment: from natural scene statistics to perceptual quality.

Authors:  Anush Krishna Moorthy; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2011-04-25       Impact factor: 10.856

5.  No-reference image quality assessment in the spatial domain.

Authors:  Anish Mittal; Anush Krishna Moorthy; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2012-08-17       Impact factor: 10.856

6.  VSI: a visual saliency-induced index for perceptual image quality assessment.

Authors:  Lin Zhang; Ying Shen; Hongyu Li
Journal:  IEEE Trans Image Process       Date:  2014-08-07       Impact factor: 10.856

7.  Massive Online Crowdsourced Study of Subjective and Objective Picture Quality.

Authors:  Deepti Ghadiyaram; Alan C Bovik
Journal:  IEEE Trans Image Process       Date:  2015-11-11       Impact factor: 10.856

8.  MRIQC: Advancing the automatic prediction of image quality in MRI from unseen sites.

Authors:  Oscar Esteban; Daniel Birman; Marie Schaer; Oluwasanmi O Koyejo; Russell A Poldrack; Krzysztof J Gorgolewski
Journal:  PLoS One       Date:  2017-09-25       Impact factor: 3.240

9.  No Reference, Opinion Unaware Image Quality Assessment by Anomaly Detection.

Authors:  Marco Leonardi; Paolo Napoletano; Raimondo Schettini; Alessandro Rozza
Journal:  Sensors (Basel)       Date:  2021-02-02       Impact factor: 3.576

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.