Literature DB >> 36005467

Full-Reference Image Quality Assessment Based on an Optimal Linear Combination of Quality Measures Selected by Simulated Annealing.

Domonkos Varga1.   

Abstract

Digital images can be distorted or contaminated by noise in various steps of image acquisition, transmission, and storage. Thus, the research of such algorithms, which can evaluate the perceptual quality of digital images consistent with human quality judgement, is a hot topic in the literature. In this study, an image quality assessment (IQA) method is introduced that predicts the perceptual quality of a digital image by optimally combining several IQA metrics. To be more specific, an optimization problem is defined first using the weighted sum of a few IQA metrics. Subsequently, the optimal values of the weights are determined by minimizing the root mean square error between the predicted and ground-truth scores using the simulated annealing algorithm. The resulted optimization-based IQA metrics were assessed and compared to other state-of-the-art methods on four large, widely applied benchmark IQA databases. The numerical results empirically corroborate that the proposed approach is able to surpass other competing IQA methods.

Entities:  

Keywords:  feature selection; full-reference image quality assessment; simulated annealing

Year:  2022        PMID: 36005467      PMCID: PMC9409967          DOI: 10.3390/jimaging8080224

Source DB:  PubMed          Journal:  J Imaging        ISSN: 2313-433X


1. Introduction

Nowadays, people increasingly communicate through media in form of audio, video, and digital images. Therefore, image quality assessment (IQA) has found many applications and become a hot research topic in the research community [1]. Namely, IQA methods evaluate the perceptual quality of digital images and support, among others, image enhancement [2], restoration [3], steganography [4], or denoising algorithms [5]. Further, IQA is also necessary in the benchmarking of many image processing or computer-vision algorithms [6,7,8]. In the literature, IQA is classified into two groups, i.e., subjective and objective IQA. Specifically, subjective IQA deals with the collection of users’ quality ratings for a set of digital images either in a laboratory [1] or in an online crowd-sourcing experiment [9]. Moreover, images’ perceptual quality is expressed as a mean opinion score (MOS), which is the arithmetic mean of individual quality scores. As a result, subjective IQA provides quality labelled images with objective IQA as training or test data [10]. Namely, objective IQA deals with algorithms and mathematical models that are able to predict the quality of a given image. Conventionally, objective IQA is divided into three classes [11]—full-reference (FR) [12], reduced-reference (RR) [13], and no-reference (NR) [14]—with respect to the availability of the reference (distortion-free) images. As the names indicate, FR-IQA methods have full access to the reference images. In contrast, NR-IQA algorithms evaluate image quality without any information about the reference images [15], and RR-IQA algorithms have partial information about them.

1.1. Contribution

The development of objective FR-IQA algorithms can also involve fusion-based strategies that already take existing FR-IQA metrics and try to create a “super evaluator”. Recently, many complex fusion-based approaches have been published in the literature [16,17,18,19]. The main contribution to this paper is also a fusion-based approach. Namely, we demonstrate a solution based on a linear combination of several already existing FR-IQA metrics optimized with a simulated annealing (SA) algorithm using a root mean square error (RMSE) objective, which is able to produce well-performing fusion-based FR-IQA metrics. To be more specific, a linear combination of 16 FR-IQA metrics is used in an optimization problem to select FR-IQA metrics and find their weights via an SA algorithm that minimizes the RMSE of the prediction. Unlike the approach of Oszust [20], we apply simulated annealing instead of a genetic algorithm for performing the fusion of FR-IQA metrics. Namely, simulated annealing usually achieves better results in the case of continuous function approximation than basic genetic algorithms because they choose one or two genes at a given location [21]. The proposed fusion-based metrics was evaluated on large, popular, and widely accepted IQA benchmark databases, such as LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25].

1.2. Organization

The rest of this paper is organized as follows. In Section 2, an overview about the current state of FR-IQA is given. Next, the proposed fusion-based metric is introduced in Section 3. Our experimental results, together with the description of the applied benchmark IQA databases, evaluation environment, and performance indices, are given in Section 4. Finally, a conclusion is drawn in Section 5.

2. Literature Review

In this paper, we follow the classification of FR-IQA algorithms presented in [26]. To be specific, Ding et al. [26] categorized existing FR-IQA algorithms into five distinct classes, i.e., error visibility, structural similarity, information theoretic, learning-based, and fusion-based methods. Error visibility methods measure a distance between the pixels of the distorted and the reference images to quantify perceptual quality degradation. The representative method of this class of FR-IQA is the mean squared error (MSE) method, which measures the average of the squares of the errors. In other words, it is the average squared difference between the reference and the distorted images in the context of FR-IQA [27]. Another well-known example is the peak signal-to-noise ratio (PSNR), which is commonly applied to assess the quality of the reconstruction of lossy compression codecs [28]. Although both MSE and PSNR have low computational costs and their physical meaning is clear and well understood, they often mismatch with subjective perceptions of visual quality. Structural similarity methods measure the similarity between the corresponding regions of the distorted and reference images using sliding-windows in the images and correlation measures. The representative and first published method of this class is the structural similarity index (SSIM) [29], which has become extremely popular in the field with many extensions and applications [30]. The theorem of SSIM has become extremely popular in the research community and inspired many variants. For example, the wavelet domain structural similarity [31] carries out SSIM in the wavelet domain to quantify perceptual quality. This work was extended by Sampat et al. [32] into the complex wavelet domain. In [33], information content was utilized as weights in the pooling process of local image quality scores. In contrast, Wang et al. [34] extended SSIM to multi-scale processing to improve perceptual quality estimation. Li and Bovik [35] elaborated an FR-IQA metric by taking the average of SSIMs computed over three different regions of an image, such as edges, textures, and smooth regions. Kolaman and Yadid-Pecht [36] found an extension of SSIM to color images by representing red, green, and blue color channels with quaternions. Later, SSIM was also extended to hyperspectral images [37]. Information theoretic methods approach the FR-IQA task from the point of view of information communication. For example, Sheikh et al. [38,39] compared the information content of the reference and distorted images. Namely, perceptual quality was quantified by how much information is similar between the reference and distorted images. In contrast, Larson and Chandler [25] classified image distortions as near-threshold and supra-threshold. The authors elaborated two quality indexes for both distortion types. Finally, the overall perceptual quality was determined based on the quality scores of near-threshold and supra-threshold distortions. As the terminology suggests, learning-based methods rely on a specific machine learning algorithm to create a quality model from training images. Next, the obtained quality model is tested on previously unseen images. For instance, Liang et al. [40] implemented a special convolutional neural network containing two paths, one for the reference image and the other for the distorted image. Further, this network was trained on -sized image patches sampled simultaneously from the reference and distorted images. As a consequence, the perceptual quality of a distorted image was estimated by the average score of the considered patches. Kim and Lee [41] devised a similar network, but it predicts a visual sensitivity map that is multiplied by an error map calculated directly from the reference and the distorted images to estimate perceptual image quality. Ahn et al. [42] further improved the idea of Kim and Lee [41] by implementing an end-to-end trained convolutional neural network with three inputs, i.e., reference image, distorted image, and spatial error map. Similar to [41], a distortion-sensitivity map was predicted from the inputs and was later multiplied by the spatial error map to give an estimation for the perceptual image quality. In contrast to the previously mentioned methods, Ding et al. [43] extracted a set of feature maps from the reference and the distorted images using the Sobel operator, log Gabor filter, and local pattern analysis. Subsequently, the extracted feature maps were compared, and from the resulting similarity scores a feature vector was compiled that was mapped onto perceptual quality scores with a trained support vector regressor. Tang et al. [44] took a similar approach, but the authors employed a different set of features (phase congruency maps [45], gradient magnitude maps, and log Gabor maps). Further, the similarity scores of the feature maps were mapped onto perceptual quality with a trained random forest regressor. Fusion-based FR-IQA methods utilize existing FR-IQA metrics to create a new FR-IQA algorithm. First, Okarma [46] suggested the idea of combined methods. Namely, the author proposed a combined metric using the product and power of MS-SSIM [34], VIF [38], and R-SVD [47]. This approach was developed further in [19], where the optimal exponents in the product were determined by using MATLAB’s fminsearch command. In [48], Oszust took a similar approach, but the author applied the scores of traditional FR-IQA metrics as predictor variables in a lasso regression. Instead of lasso regression, Yuan et al. [49] used kernel ridge regression in a similar layout. The work of Lukin et al. [50] exhibits the properties of both learning-based and fusion-based methods. Specifically, the authors created a training and a test set from the images of an IQA benchmark database. Next, the scores of several traditional FR-IQA metrics were used as image features, and a neural network was trained to estimate perceptual image quality. Amirshahi et al. [51] elaborated a special fusion-based FR-IQA metric relying on a pretrained convolutional neural network. Namely, the authors ran a reference-distorted image pair through an AlexNet [52] network and compared the activation maps with the help of a traditional FR-IQA metric. Next, the resulted scores were aggregated to obtain a single score for the perceptual image quality. Bakurov et al. [53] revisited the classical SSIM [29] and MS-SSIM [34] metrics by applying evolutionary and swarm intelligence optimization methods to find optimal hyperparameters for SSIM and MS-SSIM instead of the original settings. Fusion-based metrics were also proposed for remote sensing images [54], stitched panoramic images [55], and 3D image quality assessment [18]. For more detailed studies about FR-IQA, we refer readers to the book of Xu et al.’s [56] and to the study of Pedersen and Hardeberg [57]. Further, Zhang et al. [58] provide an evaluation of several state-of-the-art FR-IQA algorithms on various IQA benchmark databases. Zhai and Min provided an comprehensive overview of classical algorithms in [59]. For the quality assessment of screen content images [60], Min et al. gave an overview in [61].

3. Proposed Method

As already mentioned, an FR-IQA metric should deliver perceptual quality scores consistent with the human judgement using both the distorted and reference images. Let us express the aggregated decision of n different FR-IQA metrics by a weighted sum as: where stands for the quality scores provided by the FR-IQA metrics. Further, is a real vector of weights whose values are found via an optimization procedure to ensure an effective fusion of FR-IQA metrics. Namely, an optimization fusion was carried out in our study using open-source FR-IQA metrics, such as FSIM [62], FSIMc [62], GSM [63], IFC [38], IFS [64], IW-SSIM [33], MAD [25], MS-SSIM [34], NQM [65], PSNR, RFSIM [66], SFF [67], SR-SIM [12], SSIM [29], VIF [39], and VSI [68]. In the literature, Pearson’s linear correlation coefficient (PLCC), Spearman’s rank-order correlation coefficient (SROCC), Kendall’s rank order correlation coefficient (KROCC), and root mean square error (RMSE) are often considered to characterize the consistency between the ground-truth quality scores of an IQA benchmark database and the quality scores predicted by an FR-IQA metric [22]. From these performance indices, RMSE was applied as an objective function in the proposed optimization based metric. Figure 1 and Figure 2 depict flowcharts where the compilation of the proposed fusion-based metrics and its application for FR-IQA are demonstrated.
Figure 1

In the offline optimization stage, the proposed fusion-based metric is obtained by using 20% of the reference with its corresponding distorted counterparts. Next, a simulated annealing (SA) optimization process selects FR-IQA metrics and provides them with weights. The resulting metric is codenamed as LCSA-IQA to refer to the fact that is the linear combination of selected FR-IQA metrics where the weights were assigned using simulated annealing.

Figure 2

The optimal linear combination of the selected FR-IQA metrics is applied to estimate perceptual image quality.

Formally, the optimization problem can be written as where is vector containing the quality scores of a set of images obtained by Equation (1) and contains the corresponding ground-truth scores. Further, prior to the calculation of RMSE, a non-linear regression is also applied [22] since a non-linear relationship exists between the ground-truth and predicted scores. Formally, it can be written where stand for the parameters of the regression model. In addition, Q and are the fitted and predicted scores, respectively. Since we use four large, widely accepted IQA benchmark databases, i.e., LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25], in this paper, four optimization-based fusion FR-IQA metrics are proposed, respectively. To this end, approximately 20% of the reference images were randomly selected from a given benchmark IQA database. More precisely, and were compiled based on those distorted images whose reference counterparts were randomly selected. Although 20% is a common choice for parameter setting in the literature [69,70], there are also researchers who applied 30% [62] or 80% [71] for parameter tuning. However, we evaluate all the fusion based metrics on all the databases to demonstrate results independent from the database. Next, the optimization problem was solved described by Equation (2) to determine the weights for Equation (1). Since the number of possible solutions increases exponentially with number of the considered FR-IQA metrics, simulated annealing (SA) [72,73] was used to solve the above-described optimization task. Namely, SA is a probabilistic optimization technique for estimating the global optimum of a given function. The stochastic nature of this algorithm enables the usage of nonlinear objective functions where many other methods do not operate well. SA was inspired by the physical model of heating a material and then slowly decreasing the temperature to eliminate imperfections from the material. Hence, minimizing the system’s energy is the main goal. More precisely, the SA randomly generates a new point at each iteration. Based on a probability distribution with a scale proportional to the temperature, the new point’s distance from the present point or the size of the search is determined. All new points that reduce the objective are accepted by the algorithm, but points that increase the objective can also be accepted with a pre-defined probability. Due to this property of the method, SA is prevented from being stuck in local minima in early iterations. In our implementation, the SA was performed using MATLAB R2020a with a Global Optimization Toolbox using for as initial point and defining no lower or upper bounds for the method. After 100 runs of SA, the best solution— —was selected, where d denotes the database from which 20% of the reference images was chosen randomly. In the end of the SA optimization processes using LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25] databases, the following FR-IQA metrics can be obtained, which are codenamed LCSA, referring to the fact that they are linear combinations of FR-IQA measures selected by simulated annealing: The corresponding vectors are as follows:

4. Results

In this section, our experimental results are presented. First, the applied IQA benchmark databases and evaluation protocol are described in Section 4.1. Next, Section 4.2 presents a comparison to other competing state-of-the-art methods on four large IQA benchmark databases, i.e., LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25].

4.1. Applied IQA Benchmark Databases and Evaluation Protocol

The main properties of the applied IQA benchmark databases are outlined in Table 1. These databases consist of a set of reference images, whose visual quality are considered perfect and flawless. Further, distorted images are generated artificially from the reference images using different distortion types (i.e., JPEG compression noise, JPEG2000 compression noise, salt and pepper, motion blur, Gaussian, Poisson, etc.) at different distortion levels. Figure 3 depicts the empirical MOS distributions of the applied benchmark databases.
Table 1

Summary of benchmark databases used in this study.

LIVE [22]TID2013 [23]TID2008 [24]CSIQ [25]
No. of reference images29252530
No. of distorted images77930001700866
No. of distortions524176
No. of levels5544-5
No. of observers16191783835
Resolution 768×512 512×384 512×384 500×500
Figure 3

Empirical MOS distributions in the used benchmark IQA databases: (a) LIVE, (b) TID2013, (c) TID2008, and (d) CSIQ.

In the literature, PLCC, SROCC, and KROCC is widely used and accepted to characterize the performance of FR-IQA methods. They are measured between the ground-truth scores of an IQA benchmark database and the predicted scores. Moreover, prior to the calculation of PLCC a non-linear regression is also applied [22] since a non-linear relationship exists between the ground-truth and predicted scores. This non-linear relationship was also defined by Equation (3). Further, Q and are the fitted and predicted scores, respectively. PLCC between vectors x and y with length m is defined as where and are the mean subtracted version of vectors and , respectively. On the other hand, SROCC can be defined as where and are the ith entries of vectors x and y, respectively. In contrast, KROCC uses the number of concordant pairs () and the number of discordant pairs () between vectors x and y and is defined as As already mentioned, the proposed fusion-based metrics were implemented using MATLAB R2020a and its Global Optimization Toolbox. The computer configuration applied in our experiments is summarized in Table 2.
Table 2

Computer configuration applied in our experiments.

Computer modelSTRIX Z270H Gaming
Operating systemWindows 10
Memory15 GB
CPUIntel(R) Core(TM) i7-7700K CPU 4.20 GHz (8 cores)
GPUNvidia GeForce GTX 1080

4.2. Comparison to the State-of-the-Art

In this subsection, the proposed fusion-based metrics are compared to several state-of-the-art FR-IQA whose original source codes were made publicly available by the authors. Moreover, we reimplemented the fusion-based SSIM-CNN [51] method in MATLAB R2020a (available at: https://github.com/Skythianos/SSIM-CNN (accessed on 12 May 2022)). The PLCC, SROCC, and KROCC performance comparisons of the proposed fusion-based FR-IQA metrics with the state-of-the-art are summarized in Table 3 and Table 4. Specifically, Table 3 demonstrates the results on LIVE [22] and TID2013 [23], while Table 4 contains the obtained results for TID2008 [24] and CSIQ [25] databases. The obtained results clearly show that the proposed metrics are able to outperform the state-of-the-art. Specifically, those metrics that were parameter-tuned on database d always deliver the highest correlation values, while another not parameter-tuned on database d usually provides the second-best results.
Table 3

PLCC, SROCC, and KROCC performance comparison of the proposed fusion-based FR-IQA metrics on LIVE and TID2013 databases with the state-of-the-art. The best results are typed in bold, and the second best results are underlined.

LIVE [22]TID2013 [23]
FR-IQA Metric PLCC SROCC KROCC PLCC SROCC KROCC
2stepQA [74]0.9370.9320.8280.7360.7330.550
CSV [75] 0.967 0.9590.8340.8520.8480.657
DISTS [76]0.9540.9540.8110.7590.7110.524
ESSIM [77]0.9630.9620.8400.7400.7970.627
FSIM [62]0.9600.9630.8330.8590.8020.629
FSIMc [62]0.9610.9650.8360.8770.8510.667
GSM [63]0.9440.9550.8310.7890.7870.593
IFC [38]0.9270.9260.7580.5540.5390.394
IFS [64]0.9590.9600.8250.8790.8700.679
IW-SSIM [33]0.9520.9560.8170.8320.7780.598
MAD [25] 0.967 0.9670.8420.8270.7780.600
MS-SSIM [34]0.9410.9510.8040.7940.7850.604
NQM [65]0.9120.9090.7410.6900.6430.474
PSNR0.8720.8760.6870.6160.6460.467
ReSIFT [78]0.9610.9620.8380.6300.6230.471
RFSIM [66]0.9350.9400.7820.8330.7740.595
RVSIM [79]0.6410.6300.4950.7630.6830.520
SFF [67]0.9630.9650.8360.8710.8510.658
SR-SIM [12]0.9550.9620.8290.8590.8000.631
SSIM [29]0.9410.9510.8040.6180.6160.437
SSIM-CNN [51]0.9650.9630.8380.7590.7520.566
SUMMER [80] 0.967 0.9590.8330.6230.6220.472
VIF [39]0.9410.9640.8280.7740.6770.515
VSI [68]0.9480.9520.805 0.900 0.894 0.677
LCSA1 0.974 0.974 0.857 0.8200.7880.607
LCSA2 0.8460.9620.828 0.916 0.903 0.731
LCSA3 0.9470.9690.8430.7700.8210.647
LCSA4 0.967 0.970 0.844 0.8590.8230.649
Table 4

PLCC, SROCC, and KROCC performance comparison of the proposed fusion-based FR-IQA metrics on TID2008 and CSIQ databases with the state-of-the-art. The best results are typed in bold, and the second best results are underlined.

TID2008 [24]CSIQ [25]
FR-IQA Metric PLCC SROCC KROCC PLCC SROCC KROCC
2stepQA [74]0.7570.7690.5740.8410.8490.655
CSV [75]0.8520.8480.6570.9330.9330.766
DISTS [76]0.7050.6680.4880.9300.9300.764
ESSIM [77]0.6580.8760.6960.8140.9330.768
FSIM [62]0.8740.8810.6950.9120.9240.757
FSIMc [62]0.8760.8840.6990.9190.9310.769
GSM [63]0.7820.7810.5780.8960.9110.737
IFC [38]0.5750.5680.4240.8370.7670.590
IFS [64]0.8790.8690.6780.9580.9580.817
IW-SSIM [33]0.8420.8560.6640.8040.9210.753
MAD [25]0.8310.8290.6390.9500.9470.797
MS-SSIM [34]0.8380.8460.6480.8990.9130.739
NQM [65]0.6080.6240.4610.7430.7400.564
PSNR0.4470.4890.3460.8530.8090.599
ReSIFT [78]0.6270.6320.4840.8840.8680.695
RFSIM [66]0.8650.8680.6780.9120.9300.765
RVSIM [79]0.7890.7430.5660.9230.9030.728
SFF [67]0.8710.8510.6580.9640.9600.826
SR-SIM [12]0.8590.7990.6310.9250.9320.773
SSIM [29]0.6690.6750.4850.8120.8120.606
SSIM-CNN [51]0.7700.7370.5510.9520.9460.794
SUMMER [80]0.8170.8230.6230.8260.8300.658
VIF [39]0.8080.7490.5860.9280.9200.754
VSI [68]0.8980.8960.7090.9280.9420.785
LCSA1 0.8860.8740.685 0.966 0.9560.819
LCSA2 0.8960.9060.7270.8970.9490.800
LCSA3 0.923 0.921 0.755 0.964 0.961 0.827
LCSA4 0.906 0.909 0.737 0.977 0.973 0.857
Table 5 illustrates the direct and weighted average of correlation values measured on LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25]. From the results of direct averages, it can be clearly seen that the proposed and provide the best results in two out of three performance indices, while is able to produce second best KROCC value. The results of weighted averages are biased towards those FR-IQA measures that perform well on TID2013 [23] since it is the largest database from the applied benchmarks. Similarly, is the best-performing method in this respect because it provides the best results for SROCC and KROCC. Further, delivers the second best PLCC and KROCC values, while ’s performance is equivalent in terms of SROCC and KROCC to those of .
Table 5

PLCC, SROCC, and KROCC performance comparison of the proposed fusion-based FR-IQA metrics with the state-of-the-art. The best results are typed in bold, the second best results are underlined.

Direct AverageWeighted Average
FR-IQA Metric PLCC SROCC KROCC PLCC SROCC KROCC
2stepQA [74]0.8180.8210.6520.7810.7830.605
CSV [75]0.9010.8970.7290.8770.8730.694
DISTS [76]0.8370.8160.6470.7920.7590.582
ESSIM [77]0.7940.8920.7330.7560.8570.691
FSIM [62]0.9010.8930.7290.8830.8600.689
FSIMc [62]0.9080.9080.7430.8930.8850.710
GSM [63]0.8530.8590.6850.8210.8230.638
IFC [38]0.7230.7000.5420.6440.6250.473
IFS [64]0.9190.9140.7500.9000.8930.715
IW-SSIM [33]0.8570.8780.7080.8460.8400.664
MAD [25]0.8940.8800.7200.8620.8380.667
MS-SSIM [34]0.8680.8740.6990.8380.8390.659
NQM [65]0.7380.7290.5600.7030.6840.516
PSNR0.6970.7050.5250.6340.6540.480
ReSIFT [78]0.7760.7710.6220.7050.7000.550
RFSIM [66]0.8860.8780.7050.8650.8410.663
RVSIM [79]0.7790.7400.5770.7770.7230.558
SFF [67]0.9170.9080.7450.8950.8800.703
SR-SIM [12]0.9000.8730.7160.8800.8380.675
SSIM [29]0.7600.7640.5830.6980.7000.518
SSIM-CNN [51]0.8610.8490.6870.8140.8000.626
SUMMER [80]0.8080.8090.6470.7450.7460.582
VIF [39]0.8630.8280.6710.8250.7650.605
VSI [68] 0.919 0.921 0.744 0.909 0.908 0.716
LCSA1 0.9120.8980.7420.8770.8570.688
LCSA2 0.889 0.930 0.772 0.899 0.917 0.751
LCSA3 0.9010.918 0.768 0.8590.885 0.725
LCSA4 0.927 0.919 0.772 0.901 0.885 0.725
In the following, we examine the performance of the proposed and the other state-of-the-art methods on the individual distortion types of the applied IQA benchmark databases. The distortion types and their abbreviations used by the databases are summarized in Table 6. Further, Table 7, Table 8, Table 9 and Table 10 contain detailed results on the different distortion types of LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25], respectively. To be more specific, the SROCC values are given for each individual distortion types.
Table 6

Distortion types used in the applied benchmark IQA databases (LIVE [22], TID2013 [23], TID2008 [24], and CSIQ [25]).

AbbreviationDescriptionLIVE [22]TID2013 [23]TID2008 [24]CSIQ [25]
AGNadditive Gaussian noise🗸🗸🗸🗸
ANCadditive noise in color components 🗸🗸🗸
SCNspatially correlated noise 🗸🗸
MNmasked noise 🗸🗸
HFNhigh-frequency noise 🗸🗸
INimpulse noise 🗸🗸
QNquantization noise 🗸🗸
FFsimulated fast fading Rayleigh channel🗸
GBGaussian blur🗸🗸🗸
GCDglobal contrast decrement 🗸
DENimage denoising 🗸
JPEGJPEG compression noise🗸🗸🗸🗸
JP2KJPEG2000 compression noise🗸🗸🗸🗸
JGTEJPEG transmission errors 🗸🗸
J2TEJPEG2000 transmission errors 🗸🗸
NEPNnon-eccentricity pattern noise 🗸🗸
BLOCKlocal block-wise distortions of different intensity 🗸🗸
MSmean shift 🗸🗸
CCcontrast change 🗸🗸
CCSchange of color saturation 🗸
MGNmultiplicative Gaussian noise 🗸
CNcomfort noise 🗸
LCNIlossy compression of noisy images 🗸
ICQDimage color quantization with dither 🗸
CAchromatic aberration 🗸
SSRsparse sampling and reconstruction 🗸
Table 7

Comparison on LIVE’s [22] distortion types. SROCC values are given. The highest values are typed in bold, while the second highest ones are underlined.

FSIMFSIMcIFSMS-SSIMSFFVIFVSILCSA1LCSA2LCSA3LCSA4
AGN0.9650.972 0.988 0.973 0.986 0.986 0.9840.9760.9610.9620.965
FF0.9500.9520.9400.9470.9530.9650.943 0.984 0.978 0.988 0.980
GB0.9710.9710.9670.9540.9750.9730.9530.9780.989 0.997 0.996
JPEG0.983 0.984 0.9780.9820.979 0.985 0.9760.9740.9730.9640.965
JP2K 0.972 0.9700.9690.9630.9670.9700.9600.9520.9690.967 0.978
All0.9630.9650.9600.9510.9650.9640.952 0.974 0.9620.969 0.970
Table 8

Comparison on TID2013’s [23] distortion types. SROCC values are given. The highest values are typed in bold, while the second highest ones are underlined.

FSIMFSIMcIFSMS-SSIMSFFVIFVSILCSA1LCSA2LCSA3LCSA4
AGN0.8970.910 0.938 0.8650.9070.899 0.946 0.9080.9320.9250.925
ANC0.8210.8540.8540.7730.8170.830 0.871 0.8460.8540.853 0.857
SCN0.8750.8900.9340.8540.8980.884 0.937 0.908 0.940 0.9330.915
MN0.7940.8090.7960.807 0.819 0.845 0.7700.7920.7690.8110.801
HFN0.8980.904 0.914 0.8600.8980.897 0.920 0.904 0.914 0.9090.903
IN0.8070.8250.8390.7630.787 0.854 0.874 0.5740.7950.7900.728
QN0.872 0.881 0.8340.8710.8610.785 0.875 0.8540.8860.8440.863
GB0.9550.9550.966 0.967 0.968 0.9650.9610.9540.9560.9590.970
DEN0.9300.9330.9180.9270.9090.891 0.948 0.917 0.937 0.913 0.937
JPEG0.932 0.934 0.9290.9270.9270.919 0.954 0.9210.9300.9290.932
JP2K0.9580.9590.9610.9500.9570.952 0.971 0.950 0.965 0.9570.953
JGTE0.8460.8610.8930.8480.8830.841 0.922 0.854 0.891 0.8630.859
J2TE0.8910.8920.9010.8890.8710.876 0.923 0.909 0.916 0.913 0.916
NEPN0.7920.7940.7840.7970.7670.7720.806 0.826 0.8150.815 0.822
BLOCK 0.549 0.553 0.1000.4800.1790.5310.1710.4520.3530.3280.185
MS0.7530.7490.658 0.791 0.6650.628 0.770 0.5540.6780.4550.620
CC0.4690.4680.4470.4630.469 0.839 0.4750.5350.448 0.631 0.423
CCS0.275 0.836 0.8260.4100.8270.3100.8100.712 0.829 0.8130.813
MGN0.8470.8570.8790.7790.8430.847 0.912 0.875 0.900 0.8820.875
CN0.9120.9140.9040.8530.9010.895 0.924 0.911 0.923 0.9040.906
LCNI0.9470.9490.9430.9070.9260.9200.9560.951 0.958 0.945 0.957
ICQD0.8760.882 0.901 0.8560.8800.8410.8840.891 0.903 0.8910.900
CA0.872 0.893 0.8860.8780.8790.885 0.891 0.8620.8730.8700.874
SSR0.9570.9580.9560.9480.9520.935 0.963 0.9480.957 0.965 0.955
All0.8020.8510.8700.7850.8510.677 0.894 0.788 0.903 0.8210.823
Table 9

Comparison on TID2008’s [24] distortion types. SROCC values are given. The highest values are typed in bold, while the second highest ones are underlined.

FSIMFSIMcIFSMS-SSIMSFFVIFVSILCSA1LCSA2LCSA3LCSA4
AGN0.8570.8760.9170.8090.8730.880 0.923 0.887 0.916 0.9060.905
ANC0.8530.893 0.896 0.8050.8630.876 0.912 0.8870.8900.8930.889
SCN0.8480.871 0.931 0.8210.8940.8700.9300.8940.915 0.936 0.918
MN0.8020.8260.8020.8110.837 0.868 0.7730.7820.733 0.857 0.817
HFN0.9090.916 0.922 0.8690.9120.908 0.925 0.9010.909 0.922 0.917
IN0.7450.7720.8140.6910.748 0.833 0.830 0.3960.7290.7520.618
QN0.856 0.873 0.797 0.859 0.8450.797 0.873 0.825 0.859 0.8550.854
GB0.9470.9470.9600.956 0.962 0.9540.9530.9330.9440.953 0.963
DEN0.9600.9620.9490.9580.9380.916 0.969 0.9360.956 0.964 0.963
JPEG0.9280.9290.9280.9320.9320.917 0.962 0.921 0.942 0.9390.937
JP2K0.9770.9780.9780.9700.9770.9710.9850.975 0.991 0.986 0.977
JGTE0.8710.8760.8740.8680.8570.859 0.916 0.886 0.914 0.8930.904
J2TE0.8540.8560.8780.8610.8390.8500.8940.8890.885 0.911 0.901
NEPN0.7490.7510.7040.7380.6970.7620.770 0.831 0.773 0.805 0.796
BLOCK 0.849 0.846 0.0870.7550.5370.8320.6300.8260.6310.7420.672
MS 0.672 0.6550.5220.7340.5230.510 0.671 0.4600.3830.5540.497
CC0.6480.6510.6270.6380.646 0.819 0.6560.6300.604 0.732 0.577
All0.8810.8840.8690.8460.8510.7490.8960.8740.906 0.921 0.909
Table 10

Comparison on CSIQ’s [25] distortion types. SROCC values are given. The highest values are typed in bold, while the second highest ones are underlined.

FSIMFSIMcIFSMS-SSIMSFFVIFVSILCSA1LCSA2LCSA3LCSA4
AGN0.9260.9360.9590.9470.9470.9580.9640.965 0.971 0.967 0.976
ANC0.9230.9370.9530.9330.9550.951 0.964 0.9120.9480.962 0.969
GB0.9730.9730.9620.9710.9750.9750.968 0.983 0.9720.971 0.981
GCD0.9420.9440.9490.9530.9540.9350.950 0.975 0.959 0.972 0.963
JPEG0.9650.9660.9660.9630.9640.9710.9620.967 0.983 0.981 0.979
JP2K0.9680.970 0.971 0.968 0.976 0.9670.9690.9560.9500.9410.950
All0.9240.9310.9580.9130.9600.9200.9420.9560.949 0.961 0.973

5. Conclusions

In this study, we presented a novel fusion-based FR-IQA metric using simulated annealing. Specifically, an optimization problem was solved based on the weighted sum of several FR-IQA metrics by minimizing the root mean squared error between the predicted and ground-truth perceptual quality scores. The evaluation of the proposed fusion-based metrics on four large publicly available and widely accepted IQA benchmark databases empirically corroborated that the proposed metrics are able to produce competitive results compared to the state-of-the-art in terms of various performance indices, such as PLCC, SROCC, and KROCC. Future research could involve other optimization techniques and their combination for improved perceptual quality prediction. Another direction is the generalization of the proposed method for other types of media.
  20 in total

1.  Image quality assessment: from error visibility to structural similarity.

Authors:  Zhou Wang; Alan Conrad Bovik; Hamid Rahim Sheikh; Eero P Simoncelli
Journal:  IEEE Trans Image Process       Date:  2004-04       Impact factor: 10.856

2.  Image quality assessment based on gradient similarity.

Authors:  Anmin Liu; Weisi Lin; Manish Narwaria
Journal:  IEEE Trans Image Process       Date:  2011-11-15       Impact factor: 10.856

3.  Information content weighting for perceptual image quality assessment.

Authors:  Zhou Wang; Qiang Li
Journal:  IEEE Trans Image Process       Date:  2010-11-15       Impact factor: 10.856

4.  An information fidelity criterion for image quality assessment using natural scene statistics.

Authors:  Hamid Rahim Sheikh; Alan Conrad Bovik; Gustavo de Veciana
Journal:  IEEE Trans Image Process       Date:  2005-12       Impact factor: 10.856

5.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

6.  FSIM: a feature similarity index for image quality assessment.

Authors:  Lin Zhang; Lei Zhang; Xuanqin Mou; David Zhang
Journal:  IEEE Trans Image Process       Date:  2011-01-31       Impact factor: 10.856

7.  Unified Blind Quality Assessment of Compressed Natural, Graphic, and Screen Content Images.

Authors: 
Journal:  IEEE Trans Image Process       Date:  2017-08-02       Impact factor: 10.856

8.  VSI: a visual saliency-induced index for perceptual image quality assessment.

Authors:  Lin Zhang; Ying Shen; Hongyu Li
Journal:  IEEE Trans Image Process       Date:  2014-08-07       Impact factor: 10.856

9.  Full-Reference Image Quality Assessment with Linear Combination of Genetically Selected Quality Measures.

Authors:  Mariusz Oszust
Journal:  PLoS One       Date:  2016-06-24       Impact factor: 3.240

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.