Literature DB >> 33043059

Multivariate Statistical Approach to Image Quality Tasks.

Praful Gupta1, Christos G Bampis2, Jack L Glover3, Nicholas G Paulter3, Alan C Bovik1.   

Abstract

Many existing Natural Scene Statistics-based no reference image quality assessment (NR IQA) algorithms employ univariate parametric distributions to capture the statistical inconsistencies of bandpass distorted image coefficients. Here we propose a multivariate model of natural image coefficients expressed in the bandpass spatial domain that has the potential to capture higher-order correlations that may be induced by the presence of distortions. We analyze how the parameters of the multivariate model are affected by different distortion types, and we show their ability to capture distortion-sensitive image quality information. We also demonstrate the violation of Gaussianity assumptions that occur when locally estimating the energies of distorted image coefficients. Thus we propose a generalized Gaussian-based local contrast estimator as a way to implement non-linear local gain control, that facilitates the accurate modeling of both pristine and distorted images. We integrate the novel approach of generalized contrast normalization with multivariate modeling of bandpass image coefficients into a holistic NR IQA model, which we refer to as multivariate generalized contrast normalization (MVGCN). We demonstrate the improved performance of MVGCN on quality relevant tasks on multiple imaging modalities, including visible light image quality prediction and task success prediction on distorted X-ray images.

Entities:  

Keywords:  X-ray images; generalized contrast normalization; image quality assessment; multivariate statistical modeling

Year:  2018        PMID: 33043059      PMCID: PMC7542606     

Source DB:  PubMed          Journal:  J Imaging        ISSN: 2313-433X


Introduction

The perceptual quality assessment of visual media has drawn considerable attention in the recent past owing to the millions of images and videos captured and shared daily on social media websites, such as Facebook, Twitter and Instagram. Large scale video streaming services such as YouTube, Netflix and Hulu contribute heavily to internet traffic, which continues to expand rapidly as consumer demand for content increases. Reliable assessment of picture quality by large groups of human subjects is an inconvenient, time-consuming task that is very difficult to organize at scale. Thus, objective no-reference (NR) image quality assessment (IQA) models, which do not require any additional information beyond the input image, are often deployed in such settings to automatically and accurately predict visual quality as perceived by an average human subject. These models have also been successfully used to perceptually optimize the image capture process to improve the perceptual quality of the acquired visual signals. In addition, ‘quality-aware’ perceptual strategies are used to compress visual media to deliver high quality content to consumers over constrained network bandwidths [1]. Many NR IQA algorithms have been proposed recently, which for increased clarity, we will broadly classify into three categories. 1) Distortion-specific approaches that include algorithms that predict the quality of images afflicted by one or more known distortion types such as blockiness [2], ringing [3] and blur [4,5] artifacts. These models are difficult to generalize to other distortion types. 2) Purely data-driven approaches involve the extraction of low-level image features such as color and texture statistics [6], which are then mapped to subjective image quality scores using regression. More recently, deep learners have been trained to learn large sets of low level image features, which are then used to feed classical regressors that map the features to subjective quality space [7]. The general framework of convolutional neural network-based IQA models involve feeding a pre-processed patch to convolutional layers, which are often followed by pooling layers. The learned features are then fed to a combination of fully connected layers followed by non-linear activation and dropout layers [8,9]. 3) Natural scene statistics (NSS)-based approaches leverage statistical models of natural images and quantify the severity of distortion by measuring the degree of ‘unnaturalness’ caused by the presence of distortions. The perceptual image quality is measured as a distance of the distorted image from the subspace of natural images [10-13]. A number of techniques have been devised for general purpose NR IQA. The generalized Renyi entropy and normalized pseudo-Wigner distribution have been used to model directionality or anisotropicity of the variance of expected entropy to predict image quality [14]. NSS-based models have been designed to extract quality-aware features under natural image models in the wavelet [13], spatial [12] and discrete cosine transform (DCT) domains [15], achieving high correlations with human opinion scores. The divisive normalization transform (DNT), which is used to model the nonlinear response properties of sensory neurons, forms an integral component in the density estimation of natural images [16]. A commonly used parametric form of DNT is: where x denotes a natural image signal that has been processed with a bandpass linear filter, and {α, β, γ} are parameters that can be optimized on an ensemble of natural image data. As shown in [17], when bandpass natural images are subjected to DNT with α = 2 they become Gaussianized with reduced spatial dependencies. The underlying Gaussian Scale Mixture (GSM) [18] model of the marginal and joint statistics of natural (photographic) image wavelet coefficients also implies similar normalization (α = 2) of neighboring coefficients. In our recent work, we developed a generalized Gaussian Scale Mixture (GGSM) model of the wavelet coefficients of photographic images, including distorted ones [19]. This new model factors a local cluster of wavelet coefficients into a product of a generalized Gaussian vector and a positive mixing multiplier. The GGSM model demonstrates the hypothesis that the normalized wavelet-filtered coefficients of distorted images follow a generalized Gaussian behavior, devolving into a Gaussian if distortion is not present. A related approach was adopted in [20], where a finite Generalized Gaussian Mixture Model (GGMM) was used as a prior when modeling image patches in an image restoration task. Here we build on the above ideas, and propose a generalized Gaussian-based local contrast estimator, which we use in conjunction with a multivariate density estimator to extract perceptual quality-rich features in spatial domain. NSS models have been well studied on an increasing variety of natural imaging modalities, including visible light (VL), long wavelength infrared (LWIR) [21], fused VL and LWIR [22] and X-ray images [23]. This kind of statistical modeling of these imaging modalities has led to the development of new and interesting applications, and are of significance to the design of visual interpretation algorithms. In a like vein, here we explore the effectiveness and versatility of multivariate generalized contrast normalization (MVGCN) by deploying it in applications arising in two different imaging modalities. Specifically, blind quality assessment (QA) of VL images and the prediction of technician detection task performance on distorted X-ray images. The rest of the paper is organized as follows. In Section 2, we describe the generalized contrast normalization technique, which forms the core of MVGCN. We detail the multivariate statistical image model in Section 3, and analyze the effects of distortions on the estimated parameters of the multivariate model. Section 5 describes the first application, whereby MVGCN features are used to predict the detection task performance of trained bomb experts on X-ray images. The second application is explained in Section 4, where the MVGCN model is used to drive an NR IQA algorithm. Finally, Section 6 concludes the paper with possible ideas for future work.

Generalized contrast normalization

It is well established in the vision science and image quality literature that processing a natural scene by a linear bandpass operation followed by non-linear local contrast normalization has a decorrelating and Gaussianizing effect on the pixel values of the images of these natural scenes [24-26]. This kind of processing of visual data mirrors efficient representations computed by neuronal processing that takes place along the early visual pathway. These statistical models of natural (photographic) images have been used effectively in applications ranging from low-level tasks such as image denoising [27-29] and image restoration [30,31] as well as higher level processes such as face recognition [32,33], object detection [34,35], and segmentation [36,37]. A number of NSS-based IQA algorithms [12,13] operate under the hypothesis that the divisively normalized bandpass responses of a pristine image follow Gaussian behavior, and that the presence of distortion renders an image statistically unnatural, whereby the characteristic underlying Gaussianity is lost [17], as depicted in Fig. 1, where Gaussianity is a poor fit to the distribution of bandpass, divisively normalized coefficients of a JP2000 (JP2K) compressed image. Here we propose a way of collectively modeling both pristine and distorted images, using a generalized contrast normalization approach that is based on the premise that the divisively normalized bandpass coefficients of both distorted and undistorted images follow a generalized Gaussian distribution. We refer to the processed coefficients as mean subtracted generalized contrast normalized (MSGCN) coefficients.
Figure 1.

Histogram of divisively normalized bandpass coefficients of a JP2K compressed image. The difference between the best generalized Gaussian distribution (GGD) and Gaussian fits indicates that the generalized Gaussian-based contrast estimator is more appropriate for distorted coefficients. The computed Kullback-Leibler divergence values of the Gaussian and GGD fits were found to be KLDgauss = 0.083 and KLDGGD = 0.005 respectively.

Given a M × N grayscale image of intensity I = [I(i, j)], the MSGCN coefficients are computed as: where = [μ(i, j)] and = [σ(i, j)] are the local weighted mean[1] and local contrast fields defined as: where i = 1, 2, …, M, j = 1, 2, …, N are spatial indices, and w = {w|k = −K, …, K, l = −L, …, L} is a 2D isotropic Gaussian kernel normalized to unit volume with K = L = 3 and truncated to 3 standard deviations. C and ϵ are small positive constants used to prevent instabilities. γ is estimated using the popular moment-matching technique detailed in [38]. The generalized Gaussian corresponds to a Gaussian density function when γ = 2, and a Laplacian density function when γ = 1. MSGCN coefficients behave in a similar manner against different distortion types as do mean subtracted contrast normalized (MSCN) coefficients that are generated under the Gaussian model assumption (γ = 2)[12]. Distortions such as white noise tend to increase the variance of MSGCN coefficients, while distortions such as compression and blur, which increase correlations, tend to reduce variance. The MSGCN model is more generic than MSCN model and provides an elegant approach to study the statistics of distorted images.

Multivariate Image Statistics

In this section we use the aforementioned MSGCN coefficients to develop a multivariate NSS model and a way to extract quality-rich features. The generalized contrast normalization (GCN) transform is a form of local gain control mechanism that accounts for the non-linear properties of neurons, resulting from the pooled activity of neighboring sensory neurons [39]. These kinds of perceptually-relevant transformations account for the contrast masking effect, which plays an important role in distortion perception [39]. Although the GCN transform, as with other DNTs, reduces redundancies in visual data, the normalized coefficients of natural images may still exhibit dependencies in some form (depending on the image content), as depicted in Fig 2. Distortions such as compression, upscaling, and blur that reduce the amount of complexity of an image and that induce artificial correlations, tend to affect the MSGCN coefficients in a pronounced way. Increased statistical interdependencies are observed to occur between neighboring coefficients with increased distortion strength.
Figure 2.

Illustration of the dependency of horizontally adjacent MSGCN coefficients of exemplar pristine images from the LIVE database [40]. The degree of these dependencies increases with the distortion severity. The PLCC2(Pearson’s Linear Correlation Coefficient) is used as a dependency measure.

The Multivariate Generalized Gaussian Distribution

Once the MSGCN map of an input image is computed using (2), a 5D Multivariate Generalized Gaussian (MVGG) distribution is used to model the joint distribution of five neighboring coefficients as illustrated in Fig. 3. MVGG distributions have been extensively studied in the literature [41-44]. We utilize the Kotz-type distribution [41], which is a form of zero-mean multivariate elliptical distribution defined as: where s is a shape parameter that determines the exponential fall-off of the distribution (the higher s, the lower the fall-off rate), Σ is the scale parameter (matrix) which controls the spread of the coefficients along different dimensions, d is the dimension of x, and Γ(·) is the gamma function The MVGG distribution becomes a multivariate Laplace distribution when s = 0.5, a multivariate Gaussian distribution when s = 1 and a multivariate Uniform distribution as s → ∞.
Figure 3.

Set of adjacent MSGCN coefficients used to form the joint distribution model. The additional symmetrically placed samples relative to the coordinate (i, j) are not included to reduce the model size, and since it is likely that distortions along the same orientation will be redundant.

This form of MVGG distribution has also been used in a reduced-reference IQA framework [45], and in an RGB color texture model [46] of the joint statistics of color-image wavelet coefficients, a generalized Gaussian scale mixture (GGSM) model of the conditioned density of a GGSM vector [19], and in a no-reference IQA algorithm [15] to model the joint empirical distribution of extracted DCT features and subjective scores, where a bivariate version of the MVGG is used. The moment-matching scheme [41] used to estimate the shape and scale parameters of an MVGG is detailed in the Appendix.

Analysis of the Shape Parameter of the MVGG Distribution

We next analyze how the shape parameter of the MVGG distribution varies when modeling the joint distribution of adjacent MSGCN coefficients of natural[3], photographic images from two widely used databases – the Waterloo exploration database [47] and the Berkeley Segmentation Database (BSD)[48]. The Waterloo exploration IQA Database contains 4,744 pristine natural images reflecting a great diversity of real-world content. The Berkeley Segmentation Database was designed to support research on image segmentation and contains 300 training images and 200 test images. In our analysis, we only used ostensibly pristine images to generate MSGCN response maps, toward modeling a 5-dimensional joint empirical distribution of neighboring MSGCN coefficients using an MVGG density. Fig. 4(b) plots a histogram of the estimated shape parameter values of the MVGG model. The shape parameter peaked at around the same value (s = 1) on both databases, suggesting that the joint distribution of MSGCN coefficients of the pristine images may be reliably modeled as a multivariate Gaussian. This outcome may be viewed as a multivariate extension of the well-established Gaussian property of univariate normalized bandpass coefficients [18,24-26]. There are, however, a few samples within the studied collection of natural images where the estimated shape parameter deviated from s = 1. For example, a few images from the Waterloo exploration database, e.g., those shown in Fig. 4(a), contain predominantly flat, highly correlated regions which yielded peakier MVGG fits where s < 1. Cloudless sky regions (upper left of Fig. 4(a)) are bereft of any objects, and cause this effect. The lower two images of Fig. 4(a) have large saturated over/under-exposed areas, and may be viewed as substantially distorted. Overall, undistorted non-sky images of this type are rare. Conversely, the images shown in Fig. 4(c) are each almost entirely comprised of heavily textured regions, with less peaky fits (s > 1). These kinds of images are also unusual.
Figure 4.

Empirical distribution (b) of the estimated shape parameter s obtained by fitting the joint MSGCN coefficients of pristine images from the Waterloo exploration and the BSD database with an MVGG. Images yielding values of s at extremities of the distribution (s ≤ 0.2 and s > 2.0) are shown in (a) and (c), respectively.

Effect of Distortions on the Shape Parameter

Having established the relevance of the shape parameter of the MVGG and values it assumes on pristine images, we next examine how it behaves in presence of distortions. In this experiment, we degraded 1000 pristine images from the Waterloo exploration database using three common distortions – JPEG compression, Gaussian blur and additive white Gaussian noise (AWGN), each applied at ten different levels. We then followed a similar modeling procedure as that described in previous subsection – we fit the 5D empirical joint distribution of MSGCN coefficients of the distorted images with an MVGG distribution. Figure 5 depicts the way the shape parameter characteristically varies in the presence of the different degradation types and levels. Gaussian blur (Fig. 5(a)) and JPEG (Fig. 5(c)) degradations lead to peaky, heavy-tailed MVGG fits and reduced values of s. This effect becomes more pronounced with increasing distortion strength. Conversely, AWGN (Fig. 5(b)) degradations increase the randomness and entropy of an image, leading to larger values of s.
Figure 5.

Boxplots of the estimated shape parameter s for three different distortions: (a) Gaussian blur, (b) AWGN and (c) JPEG compression[4]. Outliers were removed from the plots for better visualization.

The presence of some degradations deviate the distributions of distorted MSGCN coefficients from multivariate Gaussian behavior. To better understand this effect, we computed the Kullback - Leibler (KL) divergences between the empirical bivariate[5] joint distribution of vertically adjacent MSGCN coefficients and its multivariate Gaussian fit, which are shown in Fig. 6. As shown in Fig. 6(b), increases of the AWGN standard deviation produced a slight decrease in the KL divergence, indicating that the joint distribution of the MSGCN coefficients becomes more similar to Gaussian, which is not unexpected given that the AWGN is Gaussian. Degradations such as blur and JPEG compression, which result in peakier MVGG fits, caused larger KL divergences, which increase with increasing distortion levels.
Figure 6.

Boxplots of the KL divergence between the 2-D empirical distributions of MSGCN coefficients and their multivariate Gaussian fits for three different distortions: (a) Gaussian blur, (b) AWGN and (c) JPEG compression. Outliers were removed from the plots for better visualization.

Feature extraction

Given that the MVGG model can be used to characterize distorted image statistical behavior well, we can build feature-driven image quality prediction tools. As a first set of ‘quality-aware’ features, compute the estimated shape parameter s and the five eigenvalues of the estimated covariance (scale) matrix Σ of the MVGG distribution. The premise behind the choice of these features is that the joint distribution of neighboring MSGCN coefficients of pristine images follow a multivariate Gaussian distribution, but the presence of distortion causes deviation from Gaussianity. Since each distortion affects the coefficient distributions in a characteristic manner, it is possible to predict the type and perceptual severity of distortions, and hence, the perceived image quality. As shown in Fig. 2, even after the application of the GCN transform, the MSGCN responses remain correlated on images degraded by correlation-inducing distortions such as compression and blur. Such distortions lead to more polarized eigenvalues of the estimated covariance matrix than do other distortions (AWGN). In order to demonstrate the effect of distortions on the eigenvalues, we use the ratio of the minimum and maximum eigenvalues (λ/λ) of the estimated scale matrix Σ from the best 2D MVGG fit to the vertically adjacent MSGCN coefficients. We also fit a 5D MVGG to the five neighboring coefficients (as shown in Fig. 3(b)). Figure 7 shows the boxplots of the ratio λ/λ over all images from the LIVE database [40], but classified by distortion type. The pattern of variation of the eigenvalues of the estimated covariance matrix in the presence of different distortion types is indicative of the rich perceptual quality information captured by eigenvalues.
Figure 7.

Boxplots of the ratio of the minimum and maximum eigenvalues of the estimated Covariance matrix Σ over all reference and distorted images from the LIVE database; “WN” is white noise, “GBLUR” is Gaussian blur and “FF” is fast fading Rayleigh channel.

The pairwise products of adjacent MSGCN coefficients, like those of MSCN coefficients, also exhibit statistical regularities on natural, photographic images. We follow a similar modeling approach as that described in [12], and use a zero-mode asymmetric generalized gaussian distribution (AGGD) to fit the pairwise products along four directions whose density is defined as [12]: where and The AGGD parameters (α, β, β) are estimated using the moment-matching technique described in [49]. In addition to (α, β, β), AGGD mean yields a fourth quality-aware feature. Extracting these four parameters along four orientations (H, V, D1 and D2) given by: where i ∈ {1, 2, 3, ..M − 1} and j ∈ {1, 2, 3, ..N − 1} are spatial indices, yields a total of 16 features. In order to capture even higher-order correlations caused by complex distortions, we model the joint paired-product response map along the four directions (H, V, D1 and D2) using a 4-dimensional MVGG distribution. The eigenvalues of the estimated covariance matrix of the 4D MVGG density are extracted as an additional set of four quality relevant features. Since all of the features are extracted at two scales, a total of 26 × 2 = 52 perceptually-relevant quality-aware MVGCN features are computed. A brief summary of all of these features and their methods of computation is laid out in Table 1. In subsequent sections, we study the effectiveness of the MVGCN features by applying them to multiple image quality relevant tasks.
Table 1.

Feature summary of joint MSGCN(m), paired-products (pp) and joint paired-products (j) coefficients. All features are extracted at two scales..

Feature IDFeature DescriptionComputation Procedure
m1shape5-D MVGG fit to MSGCN coefficients
m1 - m5eigenvalues of scale matrix5-D MVGG fit to MSGCN coefficients
pp1 - pp4shape, mean, left variance and right varianceAGGD fit to H pairwise coefficients
pp5 - pp8shape, mean, left variance and right varianceAGGD fit to V pairwise coefficients
pp9 - pp12shape, mean, left variance and right varianceAGGD fit to D1 pairwise coefficients
pp13 - pp16shape, mean, left variance and right varianceAGGD fit to D2 pairwise coefficients
j1 - j4eigenvalues of scale matrix4-D MVGG fit to H, V, D1 and D2 pp coefficients

Quality Assessment of Visible Light images

In order to demonstrate the quality-rich feature extraction capabilities of the MVGCN model, we utilized them for the blind image quality assessment task. We compared the performance of MVGCN against a number of well-known NR IQA algorithms, such as SSEQ [50], CORNIA [51], CNN-IQA [8], BLIINDS [15], NIQE [10], BRISQUE [12] and DIIVINE[6] [13] (all of which are publicly available), and two full reference (FR) IQA algorithms – PSNR and MS-SSIM [52]. We conducted our experiments on four widely used IQA databases namely: LIVE [40], TID08 [53], CSIQ [54] and LIVE in the Wild Challenge [55]. In all of the experiments, each model was trained on 80% of the database while the other 20% was used for testing. A support vector regressor (SVR) was used with radial basis function (RBF) to map quality features to the DMOS (Differential Mean Opinion Scores) after determining its parameters using 5-fold cross validation on the training set. The train-test splits were carried out in a manner to ensure that the training and test sets would not share reference images, so that the performances of the models would reflect their ability to learn distortions, without bias from overfitting on image content. A total of 100 such splits were performed, and the median Spearman’s rank ordered correlation coefficient (SROCC) and Pearson’s Linear correlation coefficient (PLCC) computed between the predicted quality scores and the DMOS are reported in Table 2. The overall results reported in Table 2 were computed by first applying Fisher’s z-transformation [56] given by: and then averaging the transformed correlation scores for each method across each database, and finally applying the inverse Fisher’s z-transform.
Table 2.

Median Spearman’s Rank Ordered Correlation Coefficient (SROCC) and Pearson’s Linear Correlation Coefficient (PLCC) across 100 train-test trials on the LIVE, CSIQ, TID08 and LIVE Challenge databases. The best two NR IQA models are boldfaced.

DBLIVETID08CSIQChallengeOverall
SROCCPLCCSROCCPLCCSROCCPLCCSROCCPLCCSROCCPLCC
PSNR0.8920.8830.5610.5710.8030.800--0.7560.758
MS-SSIM0.9530.9420.8600.8450.9130.896--0.8940.888
SSEQ0.8890.8890.6350.6800.6910.7490.4760.5150.6950.732
CORNIA0.9440.9460.6830.7420.6960.7680.6210.6580.7620.800
BLIINDS0.9270.9300.6620.6970.7390.7840.5030.5380.7380.764
NIQE0.9120.9070.2580.3460.6320.7210.4580.5020.6420.682
CNN-IQA0.9460.9480.7220.7500.8540.8780.5750.5560.8200.832
BRISQUE0.9400.9430.6000.6540.7380.7580.6020.6360.7410.768
DIIVINE0.8970.8970.5940.6360.7370.7510.6000.6230.7130.732
MVGCN0.9460.9470.6880.7350.7350.7750.6220.6460.7710.796
Learning-based algorithms that involve a training stage to learn optimal parameters are sometimes susceptible to overfitting, especially when trained and tested on the same database, due to similar modeling of distortions, similar experimental conditions, and other factors. The main objective of NR IQA algorithms is their ability to generalize well on other datasets. To demonstrate the generalization capabilities, we trained the NR IQA models on one entire database and evaluated their performance on common distortion types from other databases, including: JPEG2000 (JP2K) and JPEG compression, Gaussian blur and AWGN. Table 3 reports the database-independence performance of MVGCN, while Table 4 compares its aggregate performance against other NR IQA models across four leading IQA databases. We used the non-parametric Wilcoxon rank-sum test to conduct the statistical significance analysis (reported in Table 5) between different algorithms across multiple databases. As can be noted from the tables, MVGCN performed better than several leading NSS-based NR IQA algorithms, and competed well against CORNIA [51], which uses raw image patches in an unsupervised manner to learn a dictionary of local descriptors. CORNIA extracts a 20000-D feature vector and is much more computationally expensive than MVGCN, as shown in the time complexity analysis results reported in Table 6. Although CNN-IQA performed better than other models on CSIQ and TID08 databases, it failed to deliver comparable performance on LIVE Challenge database, which consists of authentic real-world distortions. This raises questions on the practical application of such models and limits their use in real-world scenarios.
Table 3.

SROCC for Database-independent experiments on MVGCN across multiple IQA databases. Rows: training dataset; Column: testing dataset. The overall performance is calculated for each training database.

Train/TestLIVETID08CSIQTID13Overall
LIVE-0.9270.9050.9220.919
TID080.841-0.6780.8760.813
CSIQ0.7870.801-0.7650.785
TID130.8390.9550.662-0.862
Table 4.

Aggregate results of Database-independent tests for various IQA models. The best two NR IQA models are boldfaced.

ModelOverall SROCCOverall PLCC
SSEQ0.8100.833
CORNIA0.8650.881
BLIINDS0.8240.840
DIIVINE0.8400.844
BRISQUE0.8090.813
MVGCN0.8540.863
Table 5.

Results of statistical significance test performed between SROCC values of different NR IQA algorithms across four databases. The elements in each cell correspond to the following databases (from left to right): LIVE, CSIQ, TID08 and LIVE Challenge. ‘1’ means that the row algorithm is statistically superior than the column algorithm with a confidence of 95%, ‘0’ signifies statistically worse and ‘-’ means statistical equivalence.

SSEQBRISQUECORNIANIQEDIIVINEBLIINDSMVGCN
SSEQ----00100–000111−−1000000000
BRISQUE1101----010011111---1–01−−00
CORNIA1–111011----11111–1110–1-0--
NIQE100000000000----100000000000
DIIVINE−−010---0–000111----0–010–00
BLIINDS11110–1001–011111–10----0–00
MVGCN1111−−11-1--11111–111–11----
Table 6.

Comparison of median time taken per image to extract features by different NR IQA algorithms on a 4 GHz Quad-Core processor with 32 GBs of RAM. The median is computed over all distorted images from the LIVE database.

AlgorithmTime (in sec.)
SSEQ0.77
CORNIA2.10
BLIINDS27.66
DIIVINE9.28
BRISQUE0.03
MVGCN0.08

Predicting detection performance on X-ray images

In previous work, we studied the natural scene statistics (NSS) of X-ray images and found that the NSS modeling paradigm applies quite well to X-ray image data, although the model is somewhat different from that of visible light (VL) images [23,57]. In prior work, we used a nominal set of X-ray NSS features along with standardized objective image quality indicators (IQIs) to analyze the relationship between X-ray image quality and the task performance of professional bomb technicians who were asked to detect and identify a collection of diverse potential threat objects. To analyze the effects of image quality on task performance, we conducted a human task performance study in which professional bomb technicians were asked to detect and identify improvised explosive device (IED) components[7] in X-ray images that we created, degraded, and presented to them in an interactive viewing environment [58]. The degradations included spatially correlated noise, reduced spatial resolution, and combinations of these. The NIST-LIVE database of ground truth judgments of bomb experts was then used to evaluate the predictive performance of the objective X-ray image quality features. More details regarding the task performance study protocols can be found in [59]. Given that the MVGCN model provides a powerful NSS-based perceptual image quality feature extractor, we examined its performance against other NSS-based models and also against conventional IEEE/ANSI N42.55 [60] metrics. We hypothesized that the presence of degradations would change the characteristic statistical properties of the MSGCN coefficients of X-ray images, which would allow MVGCN model to better capture degradations, and would better correlate with the outcomes of expert detection and identification tasks conducted on degraded X-ray images. The models used for comparison are the QUality Inspectors of X-ray images (QUIX) model [57], the IEEE/ANSI N42.55 standard [60] and combinations of these. QUIX features are a set of simple and efficient NSS-based perceptual quality features that accurately predict human task performance. In [57], QUIX considers only horizontal and vertical correlations while extracting features denoted as ‘pp’ features. In order to be consistent and to have a fair comparison against QUIX, we developed a reduced feature version of MVGCN, which we refer to as MVGCN-X-ray, which does not include the products of diagonal coefficients as part of the paired-product modeling and corresponding MVGG fits. A summary of the MVGCN-X-ray features used and the feature extraction procedure is described in Table 8.
Table 8.

Feature summary for joint MSGCN(m), paired-products (pp) and joint paired-products (j) coefficients for the X-ray application. All features are extracted at two scales.

Feature IDFeature DescriptionComputation Procedure
m1shape3-D MVGG fit to MSGCN coefficients
m1 - m3eigenvalues of scale matrix3-D MVGG fit to MSGCN coefficients
pp1 - pp3shape, mean and right varianceAGGD fit to H pairwise coefficients
pp4 - pp6shape, mean and right varianceAGGD fit to V pairwise coefficients
j1 - j2eigenvalues of scale matrix2-D MVGG fit to H and V pp coefficients
Image quality indicators (IQIs) are a set of standard objective image quality metrics defined in IEEE/ANSI N42.55 [60]. These IQIs are determined by analysis of images of a standard test object under test conditions. In our analysis, we used eight IQIs, including ‘steel penetration’, ‘spatial resolution’, ‘organic material detection’, ‘dynamic range’, ‘noise’, and three other descriptive features that are extracted from the spectral distribution of the measured modulation transfer function (MTF), noise equivalent quanta (NEQ) and noise power spectrum (NPS). Given that CORNIA is among the top performing IQA algorithms, albeit much more computationally expensive, as observed in the previous application, we compared its time complexity against MVGCN on X-ray images. CORNIA required about 50 times more time than MVGCN-X-ray did (as reported in Table 7) to extract features from high spatial-resolution[8] X-ray images.
Table 7.

The time complexity comparison between CORNIA and MVGCN-X-ray to extract features from an X-ray image of size 6329 × 5380 on a 4 GHz Quad-Core processor with 32 GBs of RAM.

AlgorithmTime (in sec.)
CORNIA542.42
MVGCN-X-ray9.61
To evaluate performance, we divided the NIST-LIVE database on the basis of component and clutter combinations. The component categories include IED components: ‘power source’, ‘detonator’, ‘load’, ‘switch’ and ‘metal pipe’, which are labeled by professional bomb technicians, if found in an image, else labeled as not found. Here we consider the task of measuring the accuracy of objective image quality models to predict the detection performance of experts. We further divided each category into four clutter types: clutter (laptop), shielding (steel plate), clutter with shielding, and no clutter. Clutter/Shileding was added to some images to make the detection task more challenging. We then devised a binary classification framework whereby features were mapped to a binary variable indicating whether the component was successfully identified by an expert. We used a logistic regression model to be consistent with [57]. The data from each component-clutter category was divided into an 80 % training set to learn logistic function parameters, which were then used to predict on the remaining 20 % test set. We used a similar performance evaluation methodology as followed in [57] – generated random disjoint train-test splits and computed median log loss and area-under-the-ROC-curve (AUC) scores over 1000 iterations (reported in Table 9).A smaller value of log loss and a larger value of AUC indicates superior classification performance, implying better correlation with human judgments.
Table 9.

Median log loss and AUC scores across 1000 train-test trials on different component-clutter combinations. The best two feature groups for each component-clutter category are boldfaced.

IQIsQUIXMVGCN-X-rayQUIX + IQIsMVGCN + IQIs
ComponentClutter typelog-lossAUClog-lossAUClog-lossAUClog-lossAUClog-lossAUC
Power sourceClutter0.4850.8670.4890.8310.5000.8200.4510.8770.5030.863
Power sourceShield0.5280.8630.5020.9000.5000.9000.5260.8800.5060.888
Power sourceShield with Clutter0.3120.5910.2050.8330.1780.8640.2050.9000.1760.864
Power sourceNo Clutter0.4960.8860.5790.8630.5440.8440.4800.8950.4700.889
DetonatorClutter0.2000.9330.2290.9440.1910.9620.2130.9380.1990.921
DetonatorShield with Clutter0.6240.8750.5260.8750.3370.8750.5120.8750.3400.875
DetonatorNo Clutter0.4550.9440.3950.9440.4070.9440.3460.9440.3440.944
LoadClutter0.1630.9440.1370.9440.1310.9440.1340.9440.1330.944
LoadNo Clutter0.5050.8520.6220.8830.5610.8610.5920.8610.5590.889
SwitchClutter0.3180.9320.2750.9360.2280.9410.2490.9270.2130.932
SwitchShield0.3690.9320.3110.9360.3500.9410.3260.9270.3510.932
SwitchNo Clutter0.4140.9280.5060.8720.4760.8720.4070.9070.4060.920
Metal pipeClutter0.2911.0000.3970.1250.3460.5000.3041.0000.2640.875
Metal pipeShield0.3941.0000.5111.0000.4770.4620.4781.0000.4531.000
Metal pipeShield with Clutter0.4320.9050.4720.9050.4430.9170.3630.9520.3750.952
Weighted Average0.4020.8810.4140.8740.3820.8700.3730.9120.3570.906
We also demonstrated in [57] that QUIX features and IQIs supply complementary information, which when combined into a single predictor performed better than either of them in isolation. Under a similar premise, we augmented MVGCN-X-ray features with IQIs to obtain similar benefits in performance. As shown in Table 9, the combination of MVGCN-X-ray with IQIs yielded better performance than any of the other features in isolation, while competing well against the combination of QUIX and IQIs. The improvement in performance of the combination can be attributed to the capture of different levels of distortion-sensitive higher-order correlations by the MVGCN-X-ray features and by complementary X-ray image quality information supplied by IQIs.

Conclusion

We designed a multivariate approach to NR IQA which uses generalized contrast normalization – a form of DNT that is more suitable to model degraded image coefficients. We investigated the effect of degradations on the estimated shape and eigenvalues of the estimated covariance matrix of MVGG fit to the joint distribution of neighboring MSGCN coefficients. Further, we demonstrated applications of the MVGCN model to the blind QA of visible light images and on the prediction of threat object detection and identification by trained experts on degraded X-ray images, achieving near state-of-the-art performance in both applications. There are a number of possible future directions. It is of interest to utilize the MVGCN model to design a spatio-temporal model of normalized bandpass video coefficients for video QA. The aforementioned multivariate modeling approach is also possibly extensible to other NSS models that utilize univariate parametric distributions of bandpass image coefficients. Furthermore, studying the statistics of other imaging modalities such as millimeter-wave, computed tomography (CT), and multi-view X-ray images are also potential future directions of exploration.
  17 in total

1.  Natural signal statistics and sensory gain control.

Authors:  O Schwartz; E P Simoncelli
Journal:  Nat Neurosci       Date:  2001-08       Impact factor: 24.884

2.  Blind image quality assessment: a natural scene statistics approach in the DCT domain.

Authors:  Michele A Saad; Alan C Bovik; Christophe Charrier
Journal:  IEEE Trans Image Process       Date:  2012-03-21       Impact factor: 10.856

3.  Image denoising via sparse and redundant representations over learned dictionaries.

Authors:  Michael Elad; Michal Aharon
Journal:  IEEE Trans Image Process       Date:  2006-12       Impact factor: 10.856

4.  Blind image quality assessment through anisotropy.

Authors:  Salvador Gabarda; Gabriel Cristóbal
Journal:  J Opt Soc Am A Opt Image Sci Vis       Date:  2007-12       Impact factor: 2.129

5.  Blind image quality assessment: from natural scene statistics to perceptual quality.

Authors:  Anush Krishna Moorthy; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2011-04-25       Impact factor: 10.856

6.  Waterloo Exploration Database: New Challenges for Image Quality Assessment Models.

Authors: 
Journal:  IEEE Trans Image Process       Date:  2016-11-22       Impact factor: 10.856

7.  No-reference image quality assessment in the spatial domain.

Authors:  Anish Mittal; Anush Krishna Moorthy; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2012-08-17       Impact factor: 10.856

8.  Predicting Detection Performance on Security X-Ray Images as a Function of Image Quality.

Authors:  Praful Gupta; Zeina Sinno; Jack L Glover; Nicholas G Paulter; Alan C Bovik
Journal:  IEEE Trans Image Process       Date:  2019-01-31       Impact factor: 10.856

9.  Massive Online Crowdsourced Study of Subjective and Objective Picture Quality.

Authors:  Deepti Ghadiyaram; Alan C Bovik
Journal:  IEEE Trans Image Process       Date:  2015-11-11       Impact factor: 10.856

10.  Robust face recognition via sparse representation.

Authors:  John Wright; Allen Y Yang; Arvind Ganesh; S Shankar Sastry; Yi Ma
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2009-02       Impact factor: 6.226

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.