Literature DB >> 33119656

Single image super-resolution via Image Quality Assessment-Guided Deep Learning Network.

Zhengqiang Xiong1, Manhui Lin2, Zhen Lin3, Tao Sun1, Guangyi Yang1, Zhengxing Wang4.   

Abstract

In recent years, deep learning (DL) networks have been widely used in super-resolution (SR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided single image super-resolution (SISR) method is proposed in DL architecture, in order to achieve a nice tradeoff between perceptual quality and distortion measure of the SR result. Unlike existing DL-based SR algorithms, an IQA net is introduced to extract perception features from SR results, calculate corresponding loss fused with original absolute pixel loss, and guide the adjustment of SR net parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, an interactive training model is established via cascaded network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process. The performance comparison between our proposed method with recent SISR methods shows that the former achieves a better tradeoff between perceptual quality and distortion measure than the latter. Extensive benchmark experiments and analyses also prove that our method provides a promising and opening architecture for SISR, which is not confined to a specific network model.

Entities:  

Mesh:

Year:  2020        PMID: 33119656      PMCID: PMC7595386          DOI: 10.1371/journal.pone.0241313

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


1. Introduction

Despite the rapid development of imaging technology, imaging devices still have limited achievable resolution due to several theoretical and practical restrictions. Super-resolution (SR) technology provides a far promising computational imaging approach to generate high-resolution (HR) images via an existing low-resolution (LR) image or image sequences, which have been widely applied in video surveillance, medical diagnostic imaging, as well as radar imaging systems. SR has two main categories, namely, single image SR (SISR) and multi-frame image SR (MISR) [1]. SISR is practical in many aspects because it has unlimited amount of LR input images [2]. For this reason, this study mainly focuses on the SISR problem. SISR is an underdetermined inverse problem, and, thus, is relatively challenging because a given LR input image may have multiple solutions based on various texture details in its corresponding HR image. Two problems need to be addressed to generate high-quality SR images. The first problem refers the unsatisfactory edge preservation and texture restoration caused by insufficient regularized constraints given that a single image provides very limited feature information. The second problem refers the difficulty in the quantitative evaluation of the estimated parameters because of the inability to measure their ground truth [3-5]. In recent years, deep neural networks (DNN) have been widely used in SR and have demonstrated superior performance [6, 7]. Several quintessential methods, such as non-uniform interpolation, frequency domain, and machine learning-based reconstruction approaches, have been developed for SR. Although these methods can provide optimal or near-optimal images that increase the resolution, they cannot guarantee detail enhancement, such as loss of high-frequency information and edge blur. To solve these problems, deep learning-based SR methods are developed given that the mapping relations of image feature from LR to HR can be fully explored, and the reconstruction results have remarkable robustness and stability in multiple scale spaces [8,9]. According to the form of the loss function, existing deep learning-based SR methods fall largely into two categories: 1) Absolute loss (AL)-based methods [10-13]. AL-based methods mainly focus on improving the quantitative indicator of IQA and generally take the forms of MSE or MAE. 2) Perceptual loss (PL)-based methods [14-17]. PL-based methods are imported to avoid excessive smoothing, and perceived quality has been the primary concern in the subsequent approaches. Blau [18] proved that the distortion and perceptual quality of an image were at odds with each other. Thus, algorithms with less distortion generally suffer from poor perceived quality, and vice versa. To satisfy diverse demands, we cannot simply conclude that one is more important than the other. In practical applications, the excellent performance of one indicator (absolute or perceived quality) or determining a compromise approach is likely to be given attention. Recently, PL has elicited increasing attention given that images have been observed by humans eventually. Nevertheless, perception features are extracted by preliminary trained net among the existing methods, thereby causing the inefficiency of the extracted features. This phenomenon is mainly caused by the different datasets that are used by the training and SR nets. Therefore, a SISR model, which integrates image quality assessment (IQA) to SR is proposed. Specifically, perception features are extracted from SR results by IQA net to calculate corresponding PL. After that, PL is fused with original pixel loss, thereby leading to the remarkable convergence of the SR parameters that are balanced between human perception and distortion measure. In our work, IQA network is injected to guide the training of SR net. On the one hand, the IQA net, which is different from SR generative adversarial network (SRGAN) [19], adaptively adjusts itself to a new distribution of input images that avoid an explicit adversarial training. On the other hand, visual geometry group (VGG) [20] and discriminator networks are designed and incorporated with each other. This strategy has dramatically reduced the computing resource while preventing over-fitting problems. The specific contributions of this study are presented as follows: An IQA-guided SISR model is proposed creatively using DL networks. Unlike existing DL-based SR algorithms, an IQA net is used to guide the adjustment of SR net parameters and prevent the over-fitting problems. We establish an interactive training mechanism via cascaded network, where the IQA net acts as a supervisor when constructing loss function of the SR net, and thus solve the problem of heterogeneous datasets used by IQA and SR networks. The approach in this paper is proved to achieve a nice tradeoff between perceptual quality and distortion measure by extensive qualitative and quantitative experiments, taking into account of both absolute pixel loss and visual effects. We propose a promising and opening architecture, which is not confined to a specific net model. The baseline of SR net is expected to be adjusted flexibly according to practical requirements (e.g. Using a generator with better net architecture to obtain improvements in network performance, or selecting another lightweight net architecture to promote the efficiency of the network). The remainder of this work is organized as follows. Section 2 presents the related works. Section 3 elaborates the net structure of the proposed method. Section 4 presents the experimental results. Section 5 draws the conclusion.

2. Related works

2.1. Super-resolution

Deep learning technique has gradually become the mainstream in SR field with the rapid development of parallel computing. Conventional methods, including interpolation, frequency domain, and machine learning-based reconstruction approaches, have achieved high reconstruction efficiency [9,21]. However, these methods exhibit limitations in predicting detailed, realistic textures. SR results from deep learning methods outperform those from other approaches in PSNR and perceptual evaluation given the inherent ability of SR to extract high-level features [4,22]. Section 1 highlights the two main SR approaches derived from deep learning, namely, AL-based and PL-based methods. To facilitate understanding, we review several major works of these methods briefly.

AL-based SR methods

Although deep networks improve the accuracy of the SR results [10-13], problems, such as over-fitting and large model, appear. Subsequently, deep recursive convolutional network (DRCN) [14], where network depth is increased without introducing new parameters for additional convolutions by using a very deep recursive layer with up to 16 recursions, is proposed to solve these problems. Lim [15] developed an enhanced deep SR network (EDSR) for SISR and achieved a performance that exceeded those of the state-of-the-art SR methods via optimization by removing unnecessary modules in the conventional residual networks (ResNet) [16]. Inspired by ResNet, VDSR, and DRCN, a deep recursive residual network (DRRN) with relatively deep network structure is designed by adjusting the architecture of existing ResNet. In this network, residual learning is adopted globally and locally to mitigate the difficulty of training, and recursive learning is used to control the model parameters. Extensive benchmark assessment shows that DRRN significantly outperforms state-of-the-art methods in SISR. Zhang et al. [17] presented a very deep residual channel attention network (RCAN), where a residual in residual structure is proposed to form a very deep network. The main network focuses on learning high-frequency information by short skip connections, thereby promoting the system performance. The key idea of these methods is to minimize the MSE between SR and HR images. However, MSE essentially focuses on pixel loss. Consequently, dealing with uncertainty in restoring high-frequency details is difficult. Specifically, images generated from MSE face the over-smooth problems, and thus, cannot match up to the human visual system [23]. Considering the subjective image quality, perception-based constraints are introduced to act as the loss functions.

PL-based SR methods

The SRGAN proposed by [19] was considered the opening research of the PL-based SR approach. Perceptual loss was presented for the first time in the field of SISR. Instead of the typically used per-pixel loss, a feature reconstruction loss was introduced to allow the transfer of semantic knowledge from the pre-trained loss network to the SR network. The results show that this work is effective in reconstructing fine details, thereby leading to positive visual results. In [24], the authors enhanced the vein features by adding generative adversarial network (GAN) loss. The idea is to constrain the SR images to satisfy the regulation of natural images. Sajjadi [23] introduced texture matching loss, which generated abundant texture information, on the basis of [25]. Wang [26] proposed a stereo SR network to integrate the information from a stereo image pair for SR, which effectively captures stereo correspondence to improve SR performance. Guo [27] designed an ORDSR network by using DCT, thereby achieving state-of-the-art SR image quality with less parameters than most deep CNN methods. Cao [28] developed a multi-scale residual channel attention network, considering that the increase in network depth resulted in the difficulty of training the network. The network exploited the image feature with conventional kernels of different sizes and used a channel attention mechanism to recalibrate the channel significance of feature mappings adaptively. The experiments on the benchmark dataset show that the proposed method can compete with the state-of-the-art SR methods. Since images are observed by humans eventually, PL has been paid much attention recently. Nevertheless, perception features are extracted by preliminary trained net and different datasets are used between training and testing nets among most existed methods, leading to the ineffectiveness of the extracted features. Besides, few of the methods mentioned above take into account of both perceptual quality and distortion measure of the SR result simultaneously. To address this issue, IQA is introduced to our work. The detailed network architecture will be explained in Section 3.

2.2 Image quality assessment

IQA has become the basic research of the related areas of image processing, the assessment results of which act as the reference indicator of the image processing system [29]. IQA, as a crucial step in image processing, can clearly reflect image distortion, serve as the basis for image automatic filtration, and be considered as the network feedback to guide parameter adjustment. In general, IQA methods fall into two categories, namely, full-reference (FR) and no-reference (NR) methods, depending on whether a reference image is required [30]. FR-IQA methods refer to the quality assessment of distorted images by comparing with the original image, which is an undistorted version of the same image. The simplest approach to measure image quality is by calculating the PSNR and the structural similarity index metric (SSIM) [31]. However, PSNR and SSIM do not always correlate with human visual perception and image quality. Other IQA methods were proposed to address the limitation of PSNR and SSIM. IQA metrics, including visual information fidelity (VIF) [32], Fast SSIM [33], information fidelity criteria (IFC) [34], multi-scale structural similarity (MS-SSIM) [35], and mean deviation similarity index (MDSI) [36], correlate well with human perception. NR-IQA has experienced two phases in its development course, namely, machine learning and deep learning. In general, learning-based methods establish the mapping relations between image features and quality. Moorthy [37] first applied machine learning to IQA, in his work, support vector machine (SVM) and natural scene statistics (NSS) were incorporated with each other to achieve favorable results, and their combination was called BIQI. Subsequently, many different methods, including DIIVINE [38], BRISQUE [39], BLIINDS-II [40], and IL-NIQE [41] were introduced gradually. Given the difficult extraction of high-dimensional features for machine learning as the dataset’s scale increases rapidly, deep learning technique is used for IQA. NR-IQA via deep learning (DLIQA) is an innovative work in this field [42]. The author suggested the extraction of multilevel representations from a very deep DNN model for learning an effective NR-IQA method. Subsequently, some other typical DNN, including CNNIQA [43], DeepIQA [44], BIECON [45], dipIQA [46], RankIQA [47], and Hallucinated-IQA [48], were adopted to develop NR-IQA methods. These networks are designed to extract features automatically, and the corresponding mathematical methods are derived. In addition, an end-to-end training approach maps the input images to image quality score output. NR-IQA methods provide quality images without the need for any reference image or features. The quality measure totally depends on the nature of human visual system because of the absence of reference images [49]. For this reason, NR-IQA is applied in our IQA net.

3. Proposed methods

In this section, we describe the proposed deep networks, and the model architecture of which is displayed in Fig 1. Our method consists of two components, namely, SR and IQA networks. SR network intends to generate SR images, and IQA network is used to assess the quality of those input images (e.g., reconstructed SR images). These networks are creatively cascaded with each other in our work, in order to promote the performance and robustness of SISR network. Specifically, SR images produced by the SR network are imported to IQA network, which outputs the image quality scores. Meanwhile, this image quality indicator is regarded as feedback to SR structure, which guided the training of SR network.
Fig 1

Framework of the proposed method, in which SR images produced by the SR network are imported to IQA network, and IQA network outputs the image quality scores in return.

3.1 Network structure

IQA network

IQA, which serves as feedback, should not only be derivable but also possess appropriate number of network layers. With respect to relatively shallow architectures, such as CNNIQA [44], extracting sufficient feature information is difficult. Relatively deep architectures, such as Hallucinated-IQA [48], face the difficulty of training and consume substantial memory. When coupled with other networks, a very deep architecture would be extremely difficult to design and use. As shown in Fig 2, an improved DeepIQA (DeepIQA has two versions: FR version and NR version, we apply NR version in this work) is used as the IQA net in our work. The figure shows that the feature extraction is realized by cascading convolutional layers in five levels, with increasing channel numbers of 32, 64, 128, 256, and 512. Each level contains two convolutional layers with identical output channel number. The batch normalization layer and the activation function are adopted to connect two adjacent convolutional layers. Strided convolutions are used to reduce the image resolution each time, and the number of features is doubled. Two fully connected (FC) layers are used to map the features obtained from the convolutional layers. The leaky ReLU function [50] is added between the FC layers, and the Sigmoid function [51] is used to limit the output within the range of 0 − 1.
Fig 2

Architecture of the IQA net, where the feature extraction is realized by cascading convolutional layers in five levels.

Compared with the NR version of DeepIQA, the proposed IQA network is advanced in the following seven aspects: All the max pooling layers are substituted by strided convolution layers to avoid gradient instabilities, preventing generation of artifact. To avoid gradient sparseness [24], Leaky ReLU is used to take the place of ReLU, and α is set to 0.2. Since IQA network serves as providing feedback to guide parameters adjustment, the structure of IQA network should not be too complex. To save memory, the first and second down samplings are conducted on the factor of ×4 (the size of the corresponding convolution kernel is 5 × 5), and the others are achieved on the factor of ×2 (the size of the corresponding kernel is 3 × 3). Batch normalization is considered, following each convolution layer to accelerate convergence. The brightness range of input images is normalized to [0, 1] by , fitting the feature extraction. Instead of the patch size of 32 × 32, our method receives patches with the size of 256 × 256 as the input to eliminate the boundary effects caused by image blocking. Sigmoid function is used to limit the output range between 0 and 1, rendering the output value as probability, thereby resulting in the convenient building of loss functions. The improved network structure is shown in Fig 2. Feature maps at different depths correspond to varied abstraction levels, representing features of the input image in different dimensions. For convenience, we consider fli as the feature map out of the last convolution layer prior to the i-th downsampling layer.

SR network

To ensure the feasibility of the proposed method, the same structure as EDSR [15] is adopted in this work. EDSR has reduced unnecessary blocks in vanilla ResNet [16], enhancing the performance with a compact model structure. Residual scaling is adopted for ease of stable training of larger models. As shown in Fig 3, the network consists of 16 residual blocks and 64 filters, which ensure that multi-level features can be shared to boost the performance.
Fig 3

Architecture of the SR net, the network consists of 16 residual blocks and 64 filters, which ensure that multi-level features can be shared to boost the performance.

3.2 Interactive training strategy

Two problems should be tackled in designing a cascade SR network using SR and IQA network, as follows: 1) the existing IQA dataset is extremely small to provide sufficient training samples. Thus, the IQA network should be provided with unsupervised training or self-supervised training. Inspired by RankIQA [47], the pairwise ranking hinge loss is introduced to create training labels dynamically, overcoming the shortcomings of insufficient samples; 2) the IQA network acts as a supervisor in cascaded network when constructing loss function of the SR network. Consequently, the output distribution from the SR net must be covered by the input distribution of the IQA net. To achieve this result, we designed an interactive training mechanism considering feature stability and effectiveness of the IQA features, thereby improving the overall performance. To ensure that the IQA network can deal with all the SR images generated from the SR network, an intuitive idea would be alternate training. Specifically, we initially train the IQA network for k iterations and then use the obtained stable features to guide the training of the SR network for one iteration. Then, the new output of the SR network is fed into the IQA network to calculate losses as well as gradients, accounting for another k iterations of training of the IQA net. Evidently, this solution is time consuming when k is a large number. In addition, the gradient descent direction provided by the SR net tends to be transient because the parameters of the IQA net update almost continuously. To avoid this condition, we propose a new training solution, where the SR network is trained n times after the k-time iterative training of the IQA network. The balance between feature stability and feature effectiveness should be considered. On the one hand, more effective features ensure that the parameters of the SR network converge to an optimum; on the other hand, more stable features guarantee the steady convergence of the SR network. Hence, the selection of k and n is crucial. In this work, they were determined experimentally.

Training of SR network

The parameters of the IQA network are fixed during the SR training. The output from the SR network and the original HR image are placed into the IQA network for perceptive feature extraction. We calculate the similarity between SR and HR images on different feature layers to establish the loss function in the perception level (namely, reconstruction of perceptual loss). In addition, the pixel loss (content loss) between SR and HR images is computed to evaluate the image distortion. Then, a joint loss function is obtained by weighted combination of these losses, which are used to optimize the training performance of the SR network.

Training of IQA network

Provided that the parameters of SR network are fixed, we estimate a margin with an FR-IQA method, considering the reconstructed SR result and the HR labels as inputs. Then, the pairwise ranking hinge loss is figured out with a specified margin, enabling self-supervised learning of the IQA network.

Alternate training

The results from the IQA network are used to guide the training of the SR network. Thus, the features extracted by the IQA network must be reliable to provide accurate reference information for SR network. The parameters of the IQA network are fixed during the training of the SR network. Thus, reckoning the IQA features as stable is reasonable. Under this condition, the convergence direction of SR should also be steady. However, the distribution of generated SR images fluctuates as the training continues, deviating from the original distribution gradually. On the contrary, when the parameters of the SR network are fixed, the assessment ability of the IQA net increases. In other words, the features extracted by the IQA net are more adaptable to the distribution of the input images. Thus, the IQA network can possibly fail to adapt to the newly generated input distributions if the parameters of the SR network are updated rapidly, thereby leading to poor effectiveness. On the contrary, frequent update of the parameters of the IQA network results in difficulty of convergence of the SR network. Each time of alternate training between the SR network and the IQA network is referred to as a training phase in this work. Then, the two networks should be maintained in the transition state between two phases to allow the IQA net to effectively fit to the output distribution of the SR net during the previous phase. Moreover, the output of the SR net should not be excessively far from its most recent distribution.

3.3 Loss functions

To improve the performance of the SR network and ensure that the IQA net effectively guides the training of the SR network, an appropriate loss function should be designed. The loss functions of SR and IQA are discussed as follows.

Loss function for SR net

In accordance with the requirements of joint training, the loss function for SR network is formulated as follows: where I is the SR image, I represents the reference HR image, G denotes the perceptual loss function of IQA net, and wi(i = 0, 1, 2, …, 5) is the weight coefficient of loss function. The first term of this function is essentially the content reconstruction loss. Content loss ensures the consistency of content by minimizing the MAE between the SR and HR images. The second term stands for the perceptual loss constructed from various feature layers of the IQA network. To better capture the texture details, the strategy in [23] is considered, where we regard the texture feature as style feature of local blocks and utilize a patch-wise style reconstruction loss. Following the configuration in [23], the patches with the size of 16 × 16 are used for texture loss calculation. We use the feature maps that resulted from the convolution layers prior to the first three down-sampling layers to compute the texture loss, denoted as fl1, fl2, and fl3, because much more high frequency information lie in shallow and middle level features.

Loss function for IQA net

The deficiency of the training samples is one of the major problems faced by deep learning-based IQA. On the one hand, humans more easily distinguish image quality via comparison of a given image pair than evaluate the absolute quality of a single one, as well as the neural network. On the other hand, labeling a large number of samples with the help of some priori knowledge is easy if quality comparison can be modeled as a binary classification problem. In this case, labels are generated dynamically. Inspired by RankIQA [47], the loss function of the IQA network is formulated as follows: where f (I) and f (I) are predicted values of I and I, respectively. A higher value indicates better image quality. m stands for the margin, which adjusts the degree of punishment for fuzzy points near the boundary. Eq (2) shows that if the prediction results are inconsistent with the actual ranking, then f (I) > f (I) − m, and the punitive steps should be performed. On the contrary, if the predictions truly reflect the actual ranking, then f (I) ≤ f (I) − m, and the parameters shall not be updated. Then, the self-supervised training of the IQA network is realized, solving the problem of label insufficiency. Evidently, m in Eq (2) is a hyper-parameter and generally needs to be set manually. It would be elegant if its value can be adaptively adjusted. For further explanation, we define the discrimination of a network. For a pair of positive and negative samples, the following cases are considered: the outputs of network A are 0.9 and 0.1 and those of network B are 0.6 and 0.4. Then, we deem that A has higher discrimination ability than B. Combining the feature effectiveness and stability, it imposes stricter conditions on zero loss when m increases. The network is, then, forced to study the slight difference among samples, and the network is likely to possess relatively high discrimination when fully converged. The learning task can be difficult in this case, resulting in relatively slow convergence and less feature stability, which indicates that the IQA networks in adjacent phases tend to focus on different characteristics, causing an unsatisfactory jitter of the gradient descent direction. When m is small, the loss easily approaches zero. At this point, the goal is relatively simple and the network converges rapidly. The parameters of the IQA networks in the adjacent phases have a subtle difference from each other, and the given gradient descent direction remains consistent. Although the feature stability is enhanced, the effectiveness and the reference value of the feature decrease. To obtain a tradeoff between the feature stability and feature effectiveness, we propose a dynamic adjustment strategy. The parameter m is dynamically adjusted during training, driving IQA to emphasize on the effectiveness in the early stage while on stability during the subsequent period. Therefore, the FR-IQA indicators that are negatively correlated to image quality are used to assess the margin to adjust the training targets dynamically. The IQA of the existing FR-IQA indicators is unreliable, and the contributions of the target adjustment on network training depend on the accuracy of the FR-IQA indicators to some extent. To weaken the effect caused by inconformity between FR-IQA indicators and human perception and to maintain a steady training, we take the moving average of the FR-IQA score as the margin. To set the initial value of the margin, let For the i-th (i > 0) iteration, m is updated using where m is updated at each iteration (IQA and SR networks) to make a more precise approximation. Thus, the IQA loss function at the i-th iteration can be determined by the following: To better understand Eq (5), two extreme cases are analyzed. If m = 1, then the loss function is summarized as follows: Eq (5) degrades into the loss function for fitting problem, which classifies SR images to 0 and HR images to 1. If m = 0, then the loss function can be rewritten as follows: Then, it can be viewed as a loss function for a binary classification problem. In our work, the root mean square error (RMSE) and the perceptual index (PI) are used to evaluate the image distortion and perceived quality, respectively. The PI is calculated as follows: where Ma and NIQE follow the same definition as in [8] and [52], respectively. For fair comparison, all reported RMSE and PI measures were calculated with the removal of four pixels from each border.

4. Experiments and analysis

To evaluate the performance of the proposed method, the networks were trained on the training set of DIV2K [53]. Then, comprehensive assessments of the proposed SISR model were carried out on several widely used benchmark databases, including DIV2K (we only used the first 16 images of the validation set due to high memory cost), PIRM-self [54], Set5 [55], Set14 [56], BSD100 [57], and Urban100 [58]. The IQA networks were pre-trained on the TID2013 dataset [59] and KonIQ10k dataset [60]. In this section, some basic settings of benchmarks for the experiment are provided, and then the details of the parameter selection are further described. Lastly, extensive comparison experiments, and quantitative analysis are also provided. We further compare our network to several state-of-the-art SISR methods. Our training was implemented in Pytorch 1.0.0 (Python 3.6) under Ubantu 16.04 operating system, and some tests were conducted in the MATLAB R2016a on an Intel i7-6700K (4.0 GHZ), GeForce GTX1080 with 16 GB RAM environment.

4.1 Network setup

SR Network setting

The parameters for the SR network were set as follows: the batch size is set to 16, and the training images are cropped into 48 × 48 non-overlapping patches to form the training samples. The weight coefficient is set as w = {0.3, 1e5, 1e5, 1e5, 0, 0}, and the total number of iterations is 1e5. The ADAM optimizer is used with the learning rate at 1e − 4, for both nets. For the IQA net, all the historical information of the gradient are cleared at the beginning of each phase. We fix the scaling factor at 4, and the FR-IQA metric in [38] is used to estimate the margin (β = 0.99).

IQA Network pre-training

We initially combined KonIQ10k dataset with TID2013 dataset to provide sufficient training samples. We trained our model with ADAM optimizer with the learning rate initialized as 1e − 4. We set the epoch size as 100. Simultaneously, we maintained the parameters of the third down-sampling layer and those ahead fixed, and the learning rate was reduced to 1e − 5 for fine tuning the parameters of the IQA network on Ma dataset [8]. Pre-training is necessary because it tunes the IQA network at an appropriate initial point, which effectively accelerates convergence.

4.2 Impact of SISR network parameters

Determination of k and n

To study the contributions of parameter k and n on the training effects, the number of iterations for a SR network is fixed at 60, 000, thereby ensuring that all SR networks are trained to the same extent. Let r = k/(k + n). When k is sufficiently large, the IQA network is sufficiently trained. Therefore, a larger r indicates that we focus more on feature effectiveness given that a more frequently updated network generally provides more up-to-date information, and a smaller r signifies that more focus is given on feature stability. Although in exhaustive, the five listed data arrays show that: 1) the synchronous increase of k and n improves the assessment quality with fixed r. This finding is mainly caused by the SR net, which was trained more sufficiently, thereby enhancing the effectiveness and reliability of the extracted features. However, this result does not conclude that a larger k implies better performance. We have to reemphasize the influence of r. The comparative result among the 1st, 3rd and 5th sets of parameters provides some proof; 2) The results from the 1st, 3rd and 5th sets of parameters indicate that a noticeable change on the indices with different r values is found with fixed (k + n) and all k values at the same order of magnitude. This finding could possibly be explained by the different importance attached to the feature effectiveness and the feature stability. The 3rd set of parameters clearly shows the very small value of r results in insufficient effectiveness of feature representations, and the 5th set of parameters demonstrates that the large value of r can lead to poor stability of extracted features. Explicitly, the value of r has considerable performance implications. Comparative experiments are conducted on 0812.png in DIV2K dataset to provide corresponding visual effect to elaborate the impact of the status of the IQA network on the training of the SR network intuitively. Quantitative results on single-image DIV2K validation set are exhibited in Fig 4, where the pictures are restored using hyper-parameters from the first three groups in Table 1. The gradient instabilities cause reconstruction image artifacts using the first array of parameters. The reconstruction also suffers from excessive smoothing and color noise when the third group of parameters is selected. This finding is mainly induced by less effective feature representations. Although the result from the 2nd set of parameters executes a better PI performance than the first two sets, much longer time and more iterations (120, 000) are consumed for training than the 3rd set of experiments (100, 000 iterations). The experimental results also reflect that better recognition effect can be obtained through finer hyper-parameter tuning. As a compromise, k = 200 and n = 300 are set to achieve a better balance in terms of speed and accuracy.
Fig 4

The impact of IQA network’s status on the training of the SR network in vision.

(a) is the result of 0812.png in DIV2K dataset using the 1st set of hyper-parameters, (b) is the result of 0812.png in DIV2K dataset using the 2nd set of hyper-parameters, (c) is the result of 0812.png in DIV2K dataset using the 3rd set of hyper-parameters.

Table 1

Effect of hyper-parameter on performance indicators.

NO.Hyper-ParametersPerformance Indicators
KnrRMSEPI
1110.510.71284.1245
25005000.511.36163.7267
31004000.2510.25175.0542
42003000.410.76473.9531
53002000.610.81574.1707

The impact of IQA network’s status on the training of the SR network in vision.

(a) is the result of 0812.png in DIV2K dataset using the 1st set of hyper-parameters, (b) is the result of 0812.png in DIV2K dataset using the 2nd set of hyper-parameters, (c) is the result of 0812.png in DIV2K dataset using the 3rd set of hyper-parameters.

Weight coefficients of losses

A total of 30 groups of parameters are tested to determine the weight coefficients of the different components of the loss function. In accordance with the weight of the perceptual loss, these parameters are again divided into three new groups, each of which contains 10 sets of different parameters. The weight coefficients of the three newly established groups are {1e5, 1e5, 1e5, 0, 0}, {2e5, 2e5, 2e5, 0, 0}, {4e5, 4e5, 4e5, 0, 0}. The weight coefficient of content loss for each group increases from 0.2 to 2 at a rate of 0.2. The testing experiments are designed on PIRM-self [54] validation set that provides smaller image size and faster computing speed because calculating the evaluation index on DIV2K validation set is time consuming, and the result is shown in Fig 5. Nevertheless, model training is still conducted on the DIV2K validation set, which considers the identification of 30 datasets. The training iterations for different weight coefficient sets are considered 2e5 to reconcile with the training extent. Therefore, some results could have been over-fitted.
Fig 5

Comparisons of the performance curve of PI versus the RMSE in different weight coefficients of loss components.

The performance curve of PI versus the RMSE is shown in Fig 5. With RMSE in [11,12] and PI in [4.5, 5.5], the weight coefficients of the loss components slightly influenced the SR performance. This result also shows that our method is insensitive to the weight coefficients, thereby rendering great convenience in adjusting the SISR parameters.

4.3 Determination of margin m

We observed experimentally that the value of m fluctuates from 0.3 to 0.5 as the margin is estimated using MDSI [38]. To further highlight the enhancements of SR performance by margin estimation via FR-IQA metric, m takes on the value 0.3, 0.4, and 0.5. Meanwhile, the comparison experiment for various margins was conducted, and the results are listed in Table 2.
Table 2

Effect on method performance of different margin.

m0.30.40.5ma(MDSI)ma(1-MSSSIM)
RMSE10.592010.472010.481810.764710.3405
PI4.17894.25184.31313.95314.4319
From the summary in Table 2, we can draw a clear conclusion that margin estimation through the moving average of MDSI effectively ameliorates network performance, thereby verifying the discussion in Section 3.3. In addition, comparisons between the last two columns in the table show that PI decreases using MS-SSIM, which demonstrates the significance of the FR-IQA metric selection to the result.

4.4 Comparison with other SR methods

Classical and the state-of-the-art SR methods, including Bicubic interpolation [61], EDSR [15], RCAN [17], EDSR-GAN [19], EDSR-VGG2,2 [24] and EnhanceNet [23], are compared with our model to demonstrate the superiority of the proposed method. To ensure fairness, the results of EDSR, RCAN, and EnhanceNet are obtained from the released codes. It should be pointed out that the architecture of SRGAN [19] is slightly adjusted when its model is replicated. To be specific, EDSR (It has been proved to have much better performance than the original SRResNet) is employed as the generator of SRGAN to exclude the influence of generators. To distinguish the original version from ours, the SRGAN in this work is marked as EDSR-GAN. EDSR net is also trained using MSE and VGG2,2 as the loss function respectively and represented as EDSR-VGG2,2. In addition, our method has been implemented separately by using EDSR and RCAN as the generator (SR net) and represented as Ours (EDSR) and Ours (RCAN), respectively.

4.4.1 Quantitative comparison

On these basis, the, the above methods are quantitatively evaluated on the basis of RMSE and PI indicators, with datasets including PIRM-self, Set5, Set14, DIV2K, BSD100, and Urban100. Table 3 reports the corresponding results.
Table 3

Performance indicators of different methods on SR dataset with scale factor ×4.

MethodPIRM-selfSet5Set14DIV2KBSD100Urban100
(RMSE/PI)(RMSE/PI)(RMSE/PI)(RMSE/PI)(RMSE/PI)(RMSE/PI)
Bicubic16.9700/7.001017.3185/7.466418.5871/7.197013.3126/7.313817.7917/7.188024.8628/7.1584
EDSR11.1643/5.24256.8489/5.770711.1938/5.35308.8092/5.606212.5619/5.370515.3666/5.1618
RCAN10.8312/4.83126.3941/5.916310.9239/5.23328.6054/5.365112.3545/5.139314.2759/4.9794
EDSR-GAN12.6563/3.2168.5848/4.698912.7263/3.911210.0543/4.131013.7665/3.584317.3064/4.2103
EDSR-VGG2,211.8390/5.05937.4269 /6.350711.7883/5.39729.3512/5.281513.1319/5.736016.3068/5.0028
EnhanceNet15.9853/2.68769.9432/3.141014.8162/3.009112.0428/3.617016.1286/2.979419.4939/3.4054
Ours(EDSR)13.9166/3.01178.8144/4.249213.6566/3.689210.7607/3.953215.0210/3.160118.9136/3.6906
Ours(RCAN)13.3740/2.97488.6329/4.757713.2347/3.601210.4531/4.126414.6446/3.121418.4560/3.6458
For convenience, the best two RMSE/PI values of each dataset are marked in bold and the next best RMSE/PI values are underlined. Table 3 clearly shows that RCAN achieves much lower RMSE than Bicubic and EDSR on all given test datasets. In particular, RCAN and EDSR have superior performance to Bicubic in RMSE and PI, thereby indicating less distortion and high perceptual quality. In addition, by comparing the results of RCAN and EDSR, the distortion can be further minimized when a deeper and more complex generator is assembled. However, none of the three methods provides a plausible and satisfactory PI, that is, the reconstruction results have poor or medium perceptual quality. Considering the SR methods that emphasize perceptual quality, the approaches represented by EDSR-GAN, EDSR-VGG2,2, EnhanceNet, Ours(EDSR), and Ours(RCAN) have better performance in PI than the conventional method (e.g., Bicubic) and the MAE loss-based methods (e.g., RCAN and EDSR). Moreover, the ranking of RMSE is contrary to that of PI for perceptual loss-based SR methods that use generators with the same structures. The results suggest that EDSR-VGG2,2 reaches the minimum RMSE together with the maximum PI, whereas an opposite effect could be observed for EnhanceNet. Our method gains satisfying RMSE and PI for both EDSR and RCAN generators. Therefore, the results suggest that our model achieves a good balance of AL and perceived quality. In addition, a horizontal comparison between Ours(EDSR) and Ours(RCAN) has been conducted on PIRM-self dataset. The comparison also shows that the improvement of generator structure will ameliorate the model performance, thereby coinciding with the conclusion in [62]. This finding also indicates the remarkable potential of our method for performance improvement. Table 3 shows that the PI of Ours(EDSR) is superior to that of Ours(RCAN) on Set5 and DIV2K datasets. However, the result is very likely to be influenced by the specific distribution of data because the two datasets only contain 5 and 16 images, respectively. To exhibit the performance of our approach more intuitively, the scatter graph of RMSE versus PI, which presents the experimental results on PIRM-self in Table 3, is drawn in Fig 6.
Fig 6

Scatter graph of RMSE versus PI for our methods and others (Bicubic, EnhanceNet, EDSR, RCAN, EDSR-GAN, EDSR-VGG2,2), where a low PI value indicates better perceived quality and a small RMSE value indicates better absolute quality.

In Fig 6, the horizontal axis refers to PI, where a low PI value indicates better perceived quality; the vertical axis represents the RMSE, and a small RMSE value indicates better absolute quality. Contrary to RCAN, EnhanceNet displays excellent PI but poor performance in RMSE. Bicubic is disadvantaged in PI and RMSE because of the simple interpolation on LR images. However, these indicators of our methods are well situated on the plane, thereby further confirming that our model provides a compromise between perceptual quality and distortion measure.

4.4.2 Visual effects comparison

The quantitative evaluation results of our models are provided on public benchmark datasets. Our models are compared with the state-of-the-art methods, including Bicubic, EDSR, RCAN, EDSR-GAN, EDSR-VGG2,2, EnhanceNet. For comparison, RMSE and PI for images on DIV2K, BSD100, Set5, Set14, and Urban100 datasets are measured with scale factor ×4. Comparative results of visual effects are provided in Fig 7. The testing datasets used in Fig 7 were obtained from Huang [58] (https://github.com/jbhuang0604/SelfExSR).
Fig 7

Visual effects and quantitative comparison (RMSE and Perceptual Index) of our methods with Bicubic, EnhanceNet, RCAN, EDSR-GAN and EDSR-VGG 2,2.

(a) shows the results of 095.png from Urban100 dataset, (b) shows the results of 039.png from BSD100 dataset, (c) shows the results of 001.png from Set5 dataset, (d) shows the results of 011.png from Set14 dataset, (e) shows the results of 804.png from DIV2K dataset, (f) shows the results of 805.png from DIV2K dataset, (g) shows the results of 057.png from Urban100 dataset, (h) shows the results of 080.png from Urban100 dataset, (i) shows the results of 801.png from DIV2K dataset.

Visual effects and quantitative comparison (RMSE and Perceptual Index) of our methods with Bicubic, EnhanceNet, RCAN, EDSR-GAN and EDSR-VGG 2,2.

(a) shows the results of 095.png from Urban100 dataset, (b) shows the results of 039.png from BSD100 dataset, (c) shows the results of 001.png from Set5 dataset, (d) shows the results of 011.png from Set14 dataset, (e) shows the results of 804.png from DIV2K dataset, (f) shows the results of 805.png from DIV2K dataset, (g) shows the results of 057.png from Urban100 dataset, (h) shows the results of 080.png from Urban100 dataset, (i) shows the results of 801.png from DIV2K dataset. Fig 7 visualizes the edge detail of the SR images. The networks trained with MAE Loss experience excessive smoothing and lack of physical realism. RCAN achieves slightly better performance than EDSR because of using an advanced network structure. Evidently, EDSR-VGG2,2 attains better visual experience than EDSR and RCAN. However, magnified images suffer from artifact that displeases visual experience. This finding is mainly caused by the max pooling layers of the VGG net. In addition, EnhanceNet obtains a satisfactory visual effect that embodies high-frequency distinct edge and abundant detail information. However, contrary to the feature distribution of natural images, high-frequency noise was unavoidably generated in the magnified images. This finding is particularly outstanding in Fig 7(b), which shows diminished visual impact in some cases. In summary, EDSR-GAN achieves relatively better visual effect among various algorithms because its textual result approaches closer to the actual circumstance. Even so, EDSR-GAN is also over-smoothing and unable to restore the high-frequency details in some situations (Fig 7(d), 7(g) and 7(h)). Compared with EnhanceNet and EDSR-GAN, our method provides compromising yet competitive results. Our method is significantly better than EnhanceNet in terms of artifact and noise. Meanwhile, our method outperforms EDSR-GAN in terms of detailed information in some situations. In general, extensive benchmark experiments reveal that the proposed model achieves a better tradeoff between perceptual quality and distortion measure given that the performance is comparable to state-of-the-art methods.

4.4.3. Complexity of networks parameters

In order to compare the network’s complexity of different algorithms, the number of network parameters is figured out and reported in Fig 8. Results are evaluated on Set5 with scale factor ×4. In Fig 8(a), the horizontal axis refers to parameter number and the vertical axis represents the RMSE, while the horizontal axis shows parameter number and the vertical axis reflects PI in Fig 8(b).
Fig 8

Scatter graph of the number of parameters versus SR performance for our methods and others (Bicubic, EnhanceNet, EDSR, RCAN, EDSR-GAN, EDSR-VGG2,2).

Results are evaluated on Set5 with scale factor ×4: (a) is the number of parameters versus RMSE, (b) is the number of parameters versus PI.

Scatter graph of the number of parameters versus SR performance for our methods and others (Bicubic, EnhanceNet, EDSR, RCAN, EDSR-GAN, EDSR-VGG2,2).

Results are evaluated on Set5 with scale factor ×4: (a) is the number of parameters versus RMSE, (b) is the number of parameters versus PI. Through comprehensive analysis of Fig 8, it can be seen that the number of parameters of our network is identical to that of EDSR or RCAN since our method is on the basis of them. It has also demonstrated that the architecture proposed is opening, and thus not confined to a specific net model. The baseline of SR net is expected to be adjusted flexibly according to practical requirements. Under the same baseline, Ours(EDSR) achieves lower PI than that of EDSR-GAN with comparable RMSE. Although our method has no advantage in RMSE compared with EDSR-VGG2,2, the PI superiority is definite. The same conclusion can also be drawn on the comparison between Ours(RCAN) and RCAN. Overall, these results coincide with the conclusion of the previous quantitative and visual effects comparison.

5. Conclusion

In this paper, we propose an IQA-guided SISR method using DL networks. Taking into account of both absolute pixel loss and visual effects of the SR results, an improved IQA network is introduced to guide the adjustment of SR network parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, we establish an interactive training mechanism via cascaded network, where the IQA network acts as a supervisor when constructing loss function of the SR network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process and prevent the over-fitting problems at the same time. The performance comparison between our proposed method with recent SISR methods shows that the former achieves a better tradeoff between perceptual quality and distortion measure than the latter. To be specific, our proposed method has better performance in terms of artifact and noise. Meanwhile, our method outperforms others in terms of detailed information in some situations. Extensive benchmark experiments and analyses also prove that our method provides a promising and opening architecture for SISR, which is not confined to a specific network model. Although the proposed method has achieved some good results, this work is just the beginning to implement this IQA-guided approach in SISR problem. In our work, the performance is proved through benchmark datasets, its robustness of different kinds of real data remains to be verified. In further works, we will develop this algorithm and make it applicable to more data types (such as infrared images, remote sensing images, et al.). (RAR) Click here for additional data file. (RAR) Click here for additional data file. (ZIP) Click here for additional data file. (ZIP) Click here for additional data file. 10 Sep 2020 PONE-D-20-25793 Single image super-resolution via Image Quality Assessment-Guided Deep Learning Network PLOS ONE Dear Dr. Sun, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Oct 25 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Zhihan Lv, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. 3. Please upload a new copy of Figure 4a, 4b and 4c as the detail is not clear. Please follow the link for more information: https://blogs.plos.org/plos/2019/06/looking-good-tips-for-creating-your-plos-figures-graphics/" https://blogs.plos.org/plos/2019/06/looking-good-tips-for-creating-your-plos-figures-graphics/ 4.We note that Figure(s) [7] in your submission contain copyrighted images. All PLOS content is published under the Creative Commons Attribution License (CC BY 4.0), which means that the manuscript, images, and Supporting Information files will be freely available online, and any third party is permitted to access, download, copy, distribute, and use these materials in any way, even commercially, with proper attribution. For more information, see our copyright guidelines: http://journals.plos.org/plosone/s/licenses-and-copyright. We require you to either (1) present written permission from the copyright holder to publish these figures specifically under the CC BY 4.0 license, or (2) remove the figures from your submission: 1.    You may seek permission from the original copyright holder of Figure(s) [7] to publish the content specifically under the CC BY 4.0 license. We recommend that you contact the original copyright holder with the Content Permission Form (http://journals.plos.org/plosone/s/file?id=7c09/content-permission-form.pdf) and the following text: “I request permission for the open-access journal PLOS ONE to publish XXX under the Creative Commons Attribution License (CCAL) CC BY 4.0 (http://creativecommons.org/licenses/by/4.0/). Please be aware that this license allows unrestricted use and distribution, even commercially, by third parties. Please reply and provide explicit written permission to publish XXX under a CC BY license and complete the attached form.” Please upload the completed Content Permission Form or other proof of granted permissions as an "Other" file with your submission. In the figure caption of the copyrighted figure, please include the following text: “Reprinted from [ref] under a CC BY license, with permission from [name of publisher], original copyright [original copyright year].” 2.    If you are unable to obtain permission from the original copyright holder to publish these figures under the CC BY 4.0 license or if the copyright holder’s requirements are incompatible with the CC BY 4.0 license, please either i) remove the figure or ii) supply a replacement figure that complies with the CC BY 4.0 license. Please check copyright information on all replacement figures and update the figure caption with source information. If applicable, please specify in the figure caption text when a figure is similar but not identical to the original image and is therefore for illustrative purposes only. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly Reviewer #3: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Under the background that deep learning has excellent characteristics in the field of single image super-resolution (SR) (SISR) and is widely used, a SISR method guided by image quality assessment (IQA) is proposed based on deep learning structure, in order to balance the perceived quality and distortion measurement corresponding to SISR results. In this method, IQA network is introduced to extract the perceptual features in SR, and the corresponding loss fused with the original absolute pixel loss is calculated to guide the adjustment of SR network parameters. In addition, an interactive training model is constructed by cascaded network, and based on the problem of insufficient samples in the training process, the hinge loss method of pairwise sorting is proposed. Combined with the evaluation results of benchmark data sets, it is found that the proposed method can not only guarantee the objective quality score, but also significantly improve the perception effect. I think this paper can be published with a few modifications and adjustments 1: In the abstract part, the author does not explain the professional term "SISR". In order to improve the readability of the article, it is suggested to explain its meaning. In addition, a large part of the abstract is a description of the construction process of the SISR method, while the expression of the research results is only a short sentence. I think it is necessary to add more details about the results in combination with the research content. 2: In the introduction, first, I have doubts about the term "LR". What does it mean? Then, I think the purpose and significance of the research should be highlighted in this part, which is of great significance to reveal the necessity of research work. 3: In Section 2.2 image quality evaluation, the author mentioned that among NR-IQA method and FR-IQA method, the method selected at last is NR-IQA method. However, there are a lot of descriptions about FR-IQA, which is unnecessary in my opinion. It is enough to emphasize the application advantages of NR-IQA method over FR-IQA method. 4: In the first paragraph of “Proposed methods”, it is mentioned that “In general, these networks have been trained interactively and, thus, promote the performance and robustness of SISR network.” I think the conclusion should be supported by some theories or references. Only a brief statement like this is unconvincing. 5: In Section 3.1, it is mentioned that “The leaky ReLu function [23] is added between the FC layers? Why choose the The leaky ReLU and what are its advantages? Since this is part of the innovation contribution of the article, it is necessary to supplement the basis. 6: In Section 3.1, it is mentioned that "To save memory, the first and second down sampling are...". How to determine the original sampling coefficient? In order to highlight the feasibility of coefficient setting, it is necessary to briefly explain. 7: What is the meaning of "EDSR" below the Figure 2? 8: In the next paragraph of equation 1, the author mentioned "the patches with the size of 16 × 16 are used for texture loss calculation". I have doubts about why the “patches” with the size of 16 × 16 are used. 9: In Section 4.2, PIRM self-validation was selected. There are no references or explanations in this part. Please add. 10: In the conclusion part, it is very helpful to optimize the structure of the paper if the author can analyze and summarize the shortcomings of the research work, and put forward the planning and prospect of the future research work. Reviewer #2: In order to achieve a good tradeoff between perceived quality and distortion measurement of service request results, and effectively improve image resolution, an image quality evaluation guided SISR method is proposed in service request architecture. From the results of stochastic resonance, IQA network is introduced to extract perceptual features. Through the cascade network, an interactive training model is established. The results show that the performance of the model is better, which can significantly improve the image resolution. However, the author needs to revise and clarify the following questions before the article is published. 1: In the abstract part, there is no description of the main results, conclusions and research significance of the paper, which makes it difficult for readers to see the contribution of the article to the field. 2: The author's scientific question is incorrect, not because SISR is practical in many aspects and it has high performance, it is necessary to study the algorithm. In this paper, the author should more elaborate on the practical significance of ultra-high resolution image and what is the difficulty of improving image resolution in reality? 3: Why do people need deep learning to improve image resolution? Why can't ordinary support vector method and linear regression method be applied to improve image quality. 4: Please simplify the introduction and elaborate the main research background, research questions, research significance, previous research progress and innovation of this paper. Please reorganize the language according to this format, delete unnecessary background description, and highlight the research focus of the article. 5: In this study, stochastic resonance network and IQA network are used to process images in series. However, in the related reports, the nonlocal residual neural network NR-Net and random forest QA method are used, and finally IQA is carried out on the image. This way of image resolution is also higher, and the accuracy of the model is better. Therefore, in this study, why use the combination of stochastic resonance network and IQA? According to the reference “Liu S, Thung K H, Lin W, et al. Real-Time Quality Assessment of Pediatric MRI via Semi-Supervised Deep Nonlocal Residual Neural Networks[J]. IEEE Transactions on Image Processing, 2020. 6: What are the specific meanings of and G in equation 1 and m in equation 2? Please check the meaning of the equation letter and improve it. 7: When the model is trained, I suggest adding the proportion of data sets and tests used for deep learning to the total data. Details are added so that other readers can repeat the experiment. 8: In the model performance comparison and parameter determination, different iterations are key to these indicators. In this paper, in order to study the contribution of parameters k and n to the training effect, the iteration number of service request network is fixed at 60000, so as to ensure that all service request networks are trained to the same degree. However, the number of iterations is determined according to the previous experiments or according to the relevant references. Because the number of iterations is relatively large and there is no loss function result, I am not sure that your parameter is valid. 9: In part 4.2, it is mentioned that the weight coefficient of the loss component slightly affects the resonant frequency performance. However, the result is the method has good stability. Is there any inevitability between the two? Do you mean that the effect of Pi on the performance of RMSE is small, so the model is stable? 10: I suggest that the contents of Table 3 should be represented by line chart, so that the differences of different methods in different data sets can be clearly seen, and the expression is more intuitive and clearer. 11: The image effect in this paper is obviously different from other algorithms, which proves the effectiveness of the algorithm. But why is it not compared with other deep learning algorithms, such as DCN and DDPG, which have relatively good performance in image sign extraction. Please explain. 12: In the result part of the paper, a large number of methods and references are added. I suggest that the two parts should be described separately, including training of data sets, and comparison of different algorithm models. These are part of the method, while the real results and the contents discussed need to be listed separately. 13: In the conclusion part, more content is to elaborate the significance of this study, but the main research results are not summarized. For example, what are the advantages of the model mentioned in this paper compared with other models? How much is the image resolution improved? In addition, the author is required to add the advantages and disadvantages of this study and put forward the future research direction. Reviewer #3: In this study, a deep learning architecture based single image super-resolution method guided by image quality assessment is proposed to achieve a good compromise between perceived quality and distortion measurement of super-resolution results. However, the method proposed in this paper is different from the super-resolution method of opportunity deep learning. In this paper, an image quality assessment network is introduced to extract perceptual features from super-resolution. By calculating the corresponding loss fused with the original absolute pixel loss, the adjustment of super-resolution network parameters is further guided. At the same time, through the establishment of interactive training network model, a hinge loss method based on acceptance sorting is proposed to overcome the shortcomings in the training process and solve the problem of image quality assessment and heterogeneous data sets used in super-resolution network. However, some contents of this paper need to be modified to meet the requirements of journals. The author is requested to revise it according to the following contents. I hope to see the revised content in the next manuscript. 1. At present, the content of the abstract is a little confused, which needs the author to elaborate according to the research purpose, research methods, research results and conclusion mode, so as to highlight the research content of this paper, and then let readers have a clearer understanding of the content of the article. 2: In the introduction, the fourth paragraph does not conform to the research background. There should not be a large section of the research results here. Moreover, the first three paragraphs for the background of the study is not in-depth, so I cannot understand the content of this paper. It is suggested that the author reinterpret this part. 3: The last part of the introduction is not suitable to be elaborated in the introduction. It is suggested that the contribution of this paper should be put into the conclusion, and the research significance and innovation of this paper should be elaborated at the end of this part. 4: In the related work, more literature review content has been added, which is repetitive with the content described in the introduction. It makes me feel that the author is quoting the results of others and has no own research content. The author has better adjust these contents. Previous scholars' research is only used for reference, not to complete your article by expounding the previous scholars' research content and research results. Otherwise, the research article is more like a summary of the study. 5: In this paper, image quality assessment based on deep learning is mentioned, but I don't see how to apply deep learning to image quality assessment network. In this paper, the author only describes "such as CNNIQA [44], extracting sufficient feature information is difficult. Relatively deep architectures, such as Hallucinated-IQA". Which method of deep learning is used to extract features? If a convolutional neural network is used, it should be directly described, rather than expressed in such a way as this. 6: When the network is trained, the author does not explain the origin of the data set, and the author needs to supplement this part. 7: The content of Section 3.2 is about training, but this part is about the content of the method. It is suggested to put it into the experiment of the fourth part. 8: The fourth chapter belongs to the content of analysis, that is, the results and discussion. It is not appropriate to continue to elaborate the equation in this part, such as equation (8). It is suggested to put this part in the method section. In the fourth chapter, it only necessary to describe the research results and discussion content. 9: Below Figure 4, it is mentioned that “Although the result from the 2nd set of parameters executes a better PI performance than the first two sets, much longer time and more iterations (120, 000) are consumed for training than the 4th set of experiments (100, 000 iterations).”. However, I don't see the results under different iterations in this paper, so the author needs to supplement. 10: What is the meaning of Figure 5? The result shown in Figure 5 at present is beyond my comprehension. The author needs to consider adjusting the coordinates so that the contents in the diagram can be displayed as much as possible, or expressing it in a different way. 11: In the method, the loss function is mentioned, but the comparison of loss function is not given in the paper. The author needs to adjust the content or expression. 12: The tables in this paper can be converted into figures as much as possible, so that the results can be observed more intuitively. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes: Xin Gao [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 29 Sep 2020 Response to Reviewer 1 General Comments: “Under the background that deep learning has excellent characteristics in the field of single image super-resolution (SR) (SISR) and is widely used, a SISR method guided by image quality assessment (IQA) is proposed based on deep learning structure, in order to balance the perceived quality and distortion measurement corresponding to SISR results. In this method, IQA network is introduced to extract the perceptual features in SR, and the corresponding loss fused with the original absolute pixel loss is calculated to guide the adjustment of SR network parameters. In addition, an interactive training model is constructed by cascaded network, and based on the problem of insufficient samples in the training process, the hinge loss method of pairwise sorting is proposed. Combined with the evaluation results of benchmark data sets, it is found that the proposed method can not only guarantee the objective quality score, but also significantly improve the perception effect. I think this paper can be published with a few modifications and adjustments.” Response: We really appreciate the reviewer’s positive comments and recognition of the contribution of the manuscript. We revise the manuscript based on the reviewers’ comments and believe the revised manuscript is substantially improved. 1.Abstract, P1 (answer to Reviewer #1, suggestion #1) Comment: “In the abstract part, the author does not explain the professional term "SISR". In order to improve the readability of the article, it is suggested to explain its meaning. In addition, a large part of the abstract is a description of the construction process of the SISR method, while the expression of the research results is only a short sentence. I think it is necessary to add more details about the results in combination with the research content.” Response: Thanks for pointing out our shortcomings! In this revision, we explain the professional term "SISR". Also we more concise expressions of the research results. In the last manuscript: In recent years, deep learning (DL) networks have been widely used in single image super-resolution (SR) (SISR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided SISR method is proposed in DL architecture…Evaluation on the benchmark datasets indicates that the proposed method can ensure the remarkable performance of both objective quality score and perception effect. In the revised version: In recent years, deep learning (DL) networks have been widely used in super-resolution (SR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided single image super-resolution (SISR) method is proposed in DL architecture... The performance comparison between our proposed method with recent SISR methods shows that the former achieves a better tradeoff between perceptual quality and distortion measure than the latter. Extensive benchmark experiments and analyses also prove that our method provides a promising and opening architecture for SISR, which is not confined to a specific network model. 2.Introduction, P1 (answer to Reviewer #1, suggestion #2) Comment: “In the introduction, first, I have doubts about the term "LR". What does it mean? Then, I think the purpose and significance of the research should be highlighted in this part, which is of great significance to reveal the necessity of research work.” Response: Thanks for your suggestion! LR means low-resolution, which is explained in the first paragraph. According to your suggestion we add more explanation about our research purpose and significance in this revision. In the revised version: Despite the rapid development of imaging technology, imaging devices still have limited achievable resolution due to several theoretical and practical restrictions. Super-resolution (SR) technology provides a far promising computational imaging approach to generate high-resolution (HR) images via an existing low-resolution (LR) image or image sequences, which have been widely applied in video surveillance, medical diagnostic imaging, as well as radar imaging systems… 3.P5, Section 2.2, Paragraph 8 (answer to Reviewer #1, suggestion #3) Comment: “In Section 2.2 image quality evaluation, the author mentioned that among NR-IQA method and FR-IQA method, the method selected at last is NR-IQA method. However, there are a lot of descriptions about FR-IQA, which is unnecessary in my opinion. It is enough to emphasize the application advantages of NR-IQA method over FR-IQA method.” Response: Thanks for this suggestion! In this revision, we remove some descriptions about FR-IQA. In the last manuscript: FR-IQA methods refer to the quality assessment of distorted images by comparing with the original image, which is an undistorted version of the same image. The extent of distortion is calculated by measuring the deviation of distorted image from the reference image. The simplest approach to measure image quality is by calculating the PSNR and the structural similarity index metric (SSIM) [32]. However, PSNR and SSIM do not always correlate with human visual perception and image quality. Other IQA methods were proposed to address the limitation of PSNR and SSIM. IQA metrics, including visual information fidelity (VIF) [33], Fast SSIM [34], information fidelity criteria (IFC) [35], multi-scale structural similarity (MS-SSIM) [36], and mean deviation similarity index (MDSI) [37], correlate well with human perception. Although these methods have improved the accuracy of IQA, they are built on far more complex mathematical models and are mostly non-differentiable. This finding causes a great deal of inconvenience to solve the optimization problems. Therefore, PSNR and SSIM are still widely used by researchers. In the revised version: FR-IQA methods refer to the quality assessment of distorted images by comparing with the original image, which is an undistorted version of the same image. The simplest approach to measure image quality is by calculating the PSNR and the structural similarity index metric (SSIM) [32]. However, PSNR and SSIM do not always correlate with human visual perception and image quality. Other IQA methods were proposed to address the limitation of PSNR and SSIM. IQA metrics, including visual information fidelity (VIF) [33], Fast SSIM [34], information fidelity criteria (IFC) [35], multi-scale structural similarity (MS-SSIM) [36], and mean deviation similarity index (MDSI) [37], correlate well with human perception. 4.P6, Section 3, Paragraph 1 (answer to Reviewer #1, suggestion #4) Comment: “In the first paragraph of “Proposed methods”, it is mentioned that “In general, these networks have been trained interactively and, thus, promote the performance and robustness of SISR network.” I think the conclusion should be supported by some theories or references. Only a brief statement like this is unconvincing.” Response: Thank you for your suggestion. We go through the entire paragraph once again and adjust the organization of this paragraph to make it more clearly explained. In the last manuscript: These networks are creatively cascaded with each other in our work. Specifically, SR images produced by the SR network are imported to IQA network, which outputs the image quality scores. Meanwhile, this image quality indicator is regarded as feedback to SR structure, which guided the training of SR network. In general, these networks have been trained interactively and, thus, promote the performance and robustness of SISR network. In the revised version: These networks are creatively cascaded with each other in our work, in order to promote the performance and robustness of SISR network. Specifically, SR images produced by the SR network are imported to IQA network, which outputs the image quality scores. Meanwhile, this image quality indicator is regarded as feedback to SR structure, which guided the training of SR network. 5.P6, Section 3.1, Paragraph 2 (answer to Reviewer #1, suggestion #5) Comment: In Section 3.1, it is mentioned that “The leaky ReLu function [23] is added between the FC layers? Why choose the leaky ReLU and what are its advantages? Since this is part of the innovation contribution of the article, it is necessary to supplement the basis. Response: Thanks for your friendly reminder! The reason why we choose the leaky ReLU is explained in the following 6th line and we cite the relevant reference. (•‘To avoid gradient sparseness [26], Leaky ReLU is used to take the place of ReLU, and α is set to 0.2.’) 6.P6, Section 3.1, Paragraph 3 (answer to Reviewer #1, suggestion #6) Comment: “In Section 3.1, it is mentioned that "To save memory, the first and second down sampling are...". How to determine the original sampling coefficient? In order to highlight the feasibility of coefficient setting, it is necessary to briefly explain.” Response: Thank you for your suggestion. We add a briefly explanation about this sentence. In the revised version: • Since IQA network serves as providing feedback to guide parameters adjustment, the structure of IQA network should not be too complex. To save memory, the first and second down samplings are conducted on the factor of ×4 (the size of the corresponding convolution kernel is 5 × 5), and the others are achieved on the factor of ×2 (the size of the corresponding kernel is 3 × 3). 7.P6, Section 3.1, Paragraph 3 (answer to Reviewer #1, suggestion #7) Comment: “What is the meaning of "EDSR" below the Figure 2?” Response: EDSR is a famous SISR algorithm proposed by Lim B. et al. in ref [15]. It is explained in Section 2, Para 3(Lim [15] developed an enhanced deep SR network (EDSR)….) 8.P6, Section 3.1, Paragraph 3 (answer to Reviewer #1, suggestion #8) Comment: “In the next paragraph of equation 1, the author mentioned "the patches with the size of 16 × 16 are used for texture loss calculation". I have doubts about why the “patches” with the size of 16 × 16 are used.” Response: Images generated from MSE face the over-smooth problems, and thus, cannot match up to the human visual system. To solve this problem, Sajjadi proposed texture matching loss, which generated abundant texture information in ref [25]. The patches with the size of 16 × 16 is tested in ref [25]. Here we just keep the configuration in [25], hoping to accomplish the same effect in capturing the texture details. 9.P12, Section 4.2, Paragraph 6 (answer to Reviewer #1, suggestion #9) Comment: “In Section 4.2, PIRM self-validation was selected. There are no references or explanations in this part. Please add.” Response: Thanks for your careful reading. We add the corresponding reference in this sentence. In the revised version: • The testing experiments are designed on PIRM-self [53] validation set that provides smaller image size and faster computing speed… 10.P16, Section 5, Paragraph 3 (answer to Reviewer #1, suggestion #10) Comment: “In the conclusion part, it is very helpful to optimize the structure of the paper if the author can analyze and summarize the shortcomings of the research work, and put forward the planning and prospect of the future research work.” Response: Thanks for your suggestion. According to the problems you point out, we rewrite the conclusion to summarize the shortcomings of the research work and propose some future work. In the revised version: We add the following paragraph: Although the proposed method has achieved some good results, this work is just the beginning to implement this IQA-guided approach in SISR problem. In our work, the performance is proved through benchmark datasets, its robustness of different kinds of real data remains to be verified. In further works, we will develop this algorithm and make it applicable to more data types (such as infrared images, remote sensing images, et al.). Response to Reviewer 2 General Comments: “In order to achieve a good tradeoff between perceived quality and distortion measurement of service request results, and effectively improve image resolution, an image quality evaluation guided SISR method is proposed in service request architecture. From the results of stochastic resonance, IQA network is introduced to extract perceptual features. Through the cascade network, an interactive training model is established. The results show that the performance of the model is better, which can significantly improve the image resolution. However, the author needs to revise and clarify the following questions before the article is published.” Response: Thank you very much for your comprehensive comments and constructive suggestions. We read and consider each comment very carefully, and thoroughly revise the manuscript according to your comments and suggestions. We hope that the manuscript reads more convincingly after the revision. 1.Abstract, P1 (answer to Reviewer #2, suggestion #1) Comment: “In the abstract part, there is no description of the main results, conclusions and research significance of the paper, which makes it difficult for readers to see the contribution of the article to the field.” Response: Thanks for your suggestion. According to the problems you point out, we rewrite the abstract to highlight the contributions of our work. In the last manuscript: In recent years, deep learning (DL) networks have been widely used in single image super-resolution (SR) (SISR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided SISR method is proposed in DL architecture, in order to achieve a nice tradeoff between perceptual quality and distortion measure of the SR result. Unlike existing DL-based SR algorithms, an IQA net is introduced to extract perception features from SR results, calculate corresponding loss fused with original absolute pixel loss, and guide the adjustment of SR net parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, an interactive training model is established via cascaded network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process. Evaluation on the benchmark datasets indicates that the proposed method can ensure the remarkable performance of both objective quality score and perception effect. In the revised version: In recent years, deep learning (DL) networks have been widely used in super-resolution (SR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided single image super-resolution (SISR) method is proposed in DL architecture, in order to achieve a nice tradeoff between perceptual quality and distortion measure of the SR result. Unlike existing DL-based SR algorithms, an IQA net is introduced to extract perception features from SR results, calculate corresponding loss fused with original absolute pixel loss, and guide the adjustment of SR net parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, an interactive training model is established via cascaded network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process. The performance comparison between our proposed method with recent SISR methods shows that the former achieves a better tradeoff between perceptual quality and distortion measure than the latter. Extensive benchmark experiments and analyses also prove that our method provides a promising and opening architecture for SISR, which is not confined to a specific network model. 2.Introduction, P1 (answer to Reviewer #2, suggestion #2) Comment: “The author's scientific question is incorrect, not because SISR is practical in many aspects and it has high performance, it is necessary to study the algorithm. In this paper, the author should more elaborate on the practical significance of ultra-high resolution image and what is the difficulty of improving image resolution in reality?” Response: Thank you for your comment. We reorganize the logical structure of ‘Introduction’ according to your suggestion, and add more explanation about our research purpose and significance in this revision. In the revised version: Despite the rapid development of imaging technology, imaging devices still have limited achievable resolution due to several theoretical and practical restrictions. Super-resolution (SR) technology provides a far promising computational imaging approach to generate high-resolution (HR) images via an existing low-resolution (LR) image or image sequences, which have been widely applied in video surveillance, medical diagnostic imaging, as well as radar imaging systems…. SISR is an underdetermined inverse problem, and, thus, is relatively challenging because a given LR input image may have multiple solutions based on various texture details in its corresponding HR image…. 3.Introduction (answer to Reviewer #2, suggestion #3) Comment: “Why do people need deep learning to improve image resolution? Why can't ordinary support vector method and linear regression method be applied to improve image quality.” Response: Conventional methods, including interpolation, frequency domain, support vector method and linear regression method approaches, can achieved high reconstruction efficiency in some cases. However, these methods exhibit limitations in predicting detailed, realistic textures. SR results from deep learning methods outperform those from other approaches in PSNR and perceptual evaluation given the inherent ability of SR to extract high-level features. For these reasons deep learning technique has gradually become the mainstream in SISR field. 4.Introduction, P1~P2 (answer to Reviewer #2, suggestion #4) Comment: Please simplify the introduction and elaborate the main research background, research questions, research significance, previous research progress and innovation of this paper. Please reorganize the language according to this format, delete unnecessary background description, and highlight the research focus of the article.” Response: Thanks for pointing out our shortcomings! In this revision, we reduce some relatively unnecessary content of expressions of the basic theories. Also we replace some explanations of the research basis with shorter and more concise expressions to highlight the content related to the research progress. All changes are marked in our revised Manuscript. 5.Full text (answer to Reviewer #2, suggestion #5) Comment: “In this study, stochastic resonance network and IQA network are used to process images in series. However, in the related reports, the nonlocal residual neural network NR-Net and random forest QA method are used, and finally IQA is carried out on the image. This way of image resolution is also higher, and the accuracy of the model is better. Therefore, in this study, why use the combination of stochastic resonance network and IQA? According to the reference “Liu S, Thung K H, Lin W, et al. Real-Time Quality Assessment of Pediatric MRI via Semi-Supervised Deep Nonlocal Residual Neural Networks[J]. IEEE Transactions on Image Processing, 2020.” Response: Thank you for your comment. According to your suggestion, we cite this reference paper in the ‘Related Work’ part in this revision. This work provides a good real-time quality assessment method of Pediatric MRI images. And it is not contradictory with our work: Firstly, we didn’t propose a stochastic resonance network, but a super-resolution (SR) network. Secondly, this the related report mainly introduces a reliable method IQA. However, our work focuses more on how to improve the resolution of a single image, and IQA functions as a feedback parameter. 6.P9, Section 3.3 (answer to Reviewer #2, suggestion #6) Comment: “What are the specific meanings of and G in equation 1 and m in equation 2? Please check the meaning of the equation letter and improve it.” Response: Thank you for your careful reading. G in equation 1 denotes the perceptual loss function of IQA net, and m in equation 2 denotes the margin used in the loss function, they are pointed out in the revision. In the last manuscript: where ISR is the SR image, IHR represents the reference HR image and wi(i = 0, 1, 2, ..., 5) is the weight coefficient of loss function. In the revised version: where ISR is the SR image, IHR represents the reference HR image, G denotes the perceptual loss function of IQA net, and wi (i=0, 1, 2, …, 5) is the weight coefficient of loss function. 7.P10-12, Section 4 (answer to Reviewer #2, suggestion #7) Comment: “When the model is trained, I suggest adding the proportion of data sets and tests used for deep learning to the total data. Details are added so that other readers can repeat the experiment.” Response: Thank you for your suggestion. Since the proportion of network parameters may need to be changed according to the types of training data. We just show some major parameters in this stage. This work is just the beginning to implement this IQA-guided approach in SISR problem, we will develop the details in the future work. 8.P11, Section 4.2, paragraph 1 (answer to Reviewer #2, suggestion #8) Comment: “In the model performance comparison and parameter determination, different iterations are key to these indicators. In this paper, in order to study the contribution of parameters k and n to the training effect, the iteration number of service request network is fixed at 60000, so as to ensure that all service request networks are trained to the same degree. However, the number of iterations is determined according to the previous experiments or according to the relevant references. Because the number of iterations is relatively large and there is no loss function result, I am not sure that your parameter is valid.” Response: Sorry for not providing more details. The iteration number was set to a very large number at 60000 to ensure that the potential of all model variants could be fully utilized, which is actually an empirical observation from our experiments. 9.P13, Section 4.2, paragraph 7 (answer to Reviewer #2, suggestion #9) Comment: “In part 4.2, it is mentioned that the weight coefficient of the loss component slightly affects the resonant frequency performance. However, the result is the method has good stability. Is there any inevitability between the two? Do you mean that the effect of Pi on the performance of RMSE is small, so the model is stable?” Response: Sorry for bringing you an ambiguity! Our original intention is to express that the performance is insensitive to the specific parameter (weight coefficients of the loss components), therefore brings great convenience in parameters adjustment. In the revision, we remove the ambiguous expressions, according to you suggestion. In the last manuscript: This result also shows that our method has good stability because it is insensitive to the weight coefficients, thereby rendering great convenience in adjusting the SISR parameters. In the revised version: This result also shows that our method is insensitive to the weight coefficients, thereby rendering great convenience in adjusting the SISR parameters. 10.P14, Table 3 (answer to Reviewer #2, suggestion #10) Comment: “I suggest that the contents of Table 3 should be represented by line chart, so that the differences of different methods in different data sets can be clearly seen, and the expression is more intuitive and clearer.” Response: Thanks for this suggestion! We’ve ever tried to show the results by line chart, since the values are too concentrated, it looks unclear and hard to compare the performance of different SR methods. To make the differences of different methods seen more clearly, we provide a scatter graph of RMSE versus PI for our methods and others (seen as Figure 6.), hoping to play a part of same role as the line chart form you have suggested. 11.Section 4 (answer to Reviewer #2, suggestion #11) Comment: “The image effect in this paper is obviously different from other algorithms, which proves the effectiveness of the algorithm. But why is it not compared with other deep learning algorithms, such as DCN and DDPG, which have relatively good performance in image sign extraction. Please explain.” Response: Thanks for this suggestion! We’d say that we have already compared our method with several deep learning-based methods in the experiments, which are EnhanceNet, EDSR, RCAN, EDSR-GAN, and EDSR-VGG2,2. These methods were published in recent years and achieved good performance in different datasets as reported by their papers, for which we think the comparison with them is convincing enough to show the effectiveness of our method. Compared with those methods above, image sign extraction methods like DCN and DDPG are not so irrelevant to the topic of this paper, so we do not include them. 12.P13, Section 4.4 (answer to Reviewer #2, suggestion #12) Comment: “In the result part of the paper, a large number of methods and references are added. I suggest that the two parts should be described separately, including training of data sets, and comparison of different algorithm models. These are part of the method, while the real results and the contents discussed need to be listed separately.” Response: Thank you for your suggestion! According to your suggestion, we separate the method descriptions with the results part. 13.P16, Section 5 (answer to Reviewer #2, suggestion #13) Comment: “In the conclusion part, more content is to elaborate the significance of this study, but the main research results are not summarized. For example, what are the advantages of the model mentioned in this paper compared with other models? How much is the image resolution improved? In addition, the author is required to add the advantages and disadvantages of this study and put forward the future research direction.” Response: Thanks for your suggestion. According to the problems you point out, we rewrite the conclusion part to explain the advantages of the proposed models and summarize the shortcomings of the research work and propose some future work. Response to Reviewer 3 General Comments: “In this study, a deep learning architecture based single image super-resolution method guided by image quality assessment is proposed to achieve a good compromise between perceived quality and distortion measurement of super-resolution results. However, the method proposed in this paper is different from the super-resolution method of opportunity deep learning. In this paper, an image quality assessment network is introduced to extract perceptual features from super-resolution. By calculating the corresponding loss fused with the original absolute pixel loss, the adjustment of super-resolution network parameters is further guided. At the same time, through the establishment of interactive training network model, a hinge loss method based on acceptance sorting is proposed to overcome the shortcomings in the training process and solve the problem of image quality assessment and heterogeneous data sets used in super-resolution network. However, some contents of this paper need to be modified to meet the requirements of journals. The author is requested to revise it according to the following contents. I hope to see the revised content in the next manuscript.” Response: We really appreciate the reviewer’s positive comments and recognition of the contribution of the manuscript. We revised the manuscript again based on the reviewers’ comments and believe that the revised manuscript is substantially improved. 1.Abstract, P1 (answer to Reviewer #3, suggestion #1) Comment: “At present, the content of the abstract is a little confused, which needs the author to elaborate according to the research purpose, research methods, research results and conclusion mode, so as to highlight the research content of this paper, and then let readers have a clearer understanding of the content of the article.” Response: Thanks for your suggestion. According to the problems you point out, we rewrite the abstract to highlight the research content of this paper. In the last manuscript: In recent years, deep learning (DL) networks have been widely used in single image super-resolution (SR) (SISR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided SISR method is proposed in DL architecture, in order to achieve a nice tradeoff between perceptual quality and distortion measure of the SR result. Unlike existing DL-based SR algorithms, an IQA net is introduced to extract perception features from SR results, calculate corresponding loss fused with original absolute pixel loss, and guide the adjustment of SR net parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, an interactive training model is established via cascaded network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process. Evaluation on the benchmark datasets indicates that the proposed method can ensure the remarkable performance of both objective quality score and perception effect. In the revised version: In recent years, deep learning (DL) networks have been widely used in super-resolution (SR) and exhibit improved performance. In this paper, an image quality assessment (IQA)-guided single image super-resolution (SISR) method is proposed in DL architecture, in order to achieve a nice tradeoff between perceptual quality and distortion measure of the SR result. Unlike existing DL-based SR algorithms, an IQA net is introduced to extract perception features from SR results, calculate corresponding loss fused with original absolute pixel loss, and guide the adjustment of SR net parameters. To solve the problem of heterogeneous datasets used by IQA and SR networks, an interactive training model is established via cascaded network. We also propose a pairwise ranking hinge loss method to overcome the shortcomings of insufficient samples during training process. The performance comparison between our proposed method with recent SISR methods shows that the former achieves a better tradeoff between perceptual quality and distortion measure than the latter. Extensive benchmark experiments and analyses also prove that our method provides a promising and opening architecture for SISR, which is not confined to a specific network model. 2.Introduction, P1~P2 (answer to Reviewer #3, suggestion #2) Comment: “In the introduction, the fourth paragraph does not conform to the research background. There should not be a large section of the research results here. Moreover, the first three paragraphs for the background of the study is not in-depth, so I cannot understand the content of this paper. It is suggested that the author reinterpret this part.” Response: Thanks for pointing out our shortcomings! In this revision, we reorganize the logical structure of ‘Introduction’ according to your suggestion, and add more explanation about our research purpose and significance in this revision. All changes are marked in our revised Manuscript. 3.P3, Introduction (answer to Reviewer #3, suggestion #3) Comment: “The last part of the introduction is not suitable to be elaborated in the introduction. It is suggested that the contribution of this paper should be put into the conclusion, and the research significance and innovation of this paper should be elaborated at the end of this part.” Response: Thanks for your suggestion! We summarize the contribution in the conclusion. Yet, after a fully discussion, we think it is also necessary to list our contributions here to help the readers better understand what we have done before they read the conclusion. 4.P4, Section 2 (answer to Reviewer #3, suggestion #4) Comment: “In the related work, more literature review content has been added, which is repetitive with the content described in the introduction. It makes me feel that the author is quoting the results of others and has no own research content. The author has better adjust these contents. Previous scholars' research is only used for reference, not to complete your article by expounding the previous scholars' research content and research results. Otherwise, the research article is more like a summary of the study.” Response: We'd like to thank you for pointing out this problem. We adjust this part and reduce some relatively unnecessary content of expressions of the basic theories. Also we replace some explanations of the research basis with shorter and more concise expressions to highlight the content related to the research progress. 5.P6, Section 3.1, paragraph 1(answer to Reviewer #3, suggestion #5) Comment: “In this paper, image quality assessment based on deep learning is mentioned, but I don't see how to apply deep learning to image quality assessment network. In this paper, the author only describes "such as CNNIQA [44], extracting sufficient feature information is difficult. Relatively deep architectures, such as Hallucinated-IQA". Which method of deep learning is used to extract features? If a convolutional neural network is used, it should be directly described, rather than expressed in such a way as this.” Response: Thanks for your kind suggestion. We explain about which IQA model is used in the next paragraph “As shown in Figure. 2, an improved DeepIQA (DeepIQA has two versions: FR version and R version, we apply NR version in this work) is used as the IQA net in our work. The figure shows that the feature extraction is realized by cascading convolutional layers in five levels, with increasing channel numbers of 32, 64, 128, 256, and 512…” We mention CNNIQA [44] and Hallucinated-IQA here only to demonstrate that not every IQA model is suitable in our SISR problem. 6.P7, Section 3.2 (answer to Reviewer #3, suggestion #6) Comment: “When the network is trained, the author does not explain the origin of the data set, and the author needs to supplement this part.” Response: Thanks for your kind suggestion. We explain the origin of the data set in paragraph Section 4 ( ‘To evaluate the performance of the proposed method, the networks were trained on the training set of DIV2K [52].’). Since Section 3.2 is written only to explain the training process in our work, we haven’t yet explained the specific dataset used in training here. 7.P7, Section 3.2 (answer to Reviewer #3, suggestion #7) Comment: “The content of Section 3.2 is about training, but this part is about the content of the method. It is suggested to put it into the experiment of the fourth part.” Response: Thanks for your kind suggestion. Section 3.2 actually explains the training strategy used in our paper, detailed training parameters is introduced in Section 4 (the experiment part). In order to avoid bringing about ambiguity, we revise the title of this section. 8.P11, Section 4.1 (answer to Reviewer #3, suggestion #8) Comment: “The fourth chapter belongs to the content of analysis, that is, the results and discussion. It is not appropriate to continue to elaborate the equation in this part, such as equation (8). It is suggested to put this part in the method section. In the fourth chapter, it only necessary to describe the research results and discussion content.” Response: Thanks for your reminder. We move this part to the Section 3 ahead, according to your suggestion. 9.P11, Section 4.1 (answer to Reviewer #3, suggestion #8) Comment: “Below Figure 4, it is mentioned that ‘Although the result from the 2nd set of parameters executes a better PI performance than the first two sets, much longer time and more iterations (120, 000) are consumed for training than the 4th set of experiments (100, 000 iterations).’. However, I don't see the results under different iterations in this paper, so the author needs to supplement.” Response: Thanks for your reminder. The paragraph which you point out is an explanation of a small experiment about the process of parameter adjustment. Indeed, it is more perfect to experiment with different iterations according to your suggestion. However, it will take a plenty of time to compare different iterations using different datasets and it is unable to complete in this phase. This work is just the beginning to implement this IQA-guided approach in SISR problem, and we really appreciate your valuable advice. We would like to accomplish the advice your have proposed in the future work. 10.P13, Section 4.2 (answer to Reviewer #3, suggestion #10) Comment: “What is the meaning of Figure 5? The result shown in Figure 5 at present is beyond my comprehension. The author needs to consider adjusting the coordinates so that the contents in the diagram can be displayed as much as possible, or expressing it in a different way.” Response: Figure 5 shows the changing trends about the value of PI and RMSE. From Figure 5, it is clear that the points tend to gather. It is placed here to demonstrate that the performance of the model in terms of PI and RMSE changes a little with varying weight coefficients of loss components, which implies that our model is insensitive to the weight coefficients, thereby rendering great convenience in adjusting the SISR parameters. 11.P8, Section 3.3 (answer to Reviewer #3, suggestion #11) Comment: “In the method, the loss function is mentioned, but the comparison of loss function is not given in the paper. The author needs to adjust the content or expression.” Response: Thanks for your reminder. Actually we compare and analyze the different kinds of loss functions in the ‘Introduction’ and ‘Related Work’ part (see. “According to the loss function, existing deep learningbased SR methods fall largely into two categories: 1) Absolute loss (AL)-based methods [10-13]. AL-based methods mainly focus on improving the quantitative indicator of IQA and generally take the forms of MSE or MAE. Results show that this method can achieve favorable results with high peak signal-to-noise ratio (PSNR) value provided that the…”). In order to avoid repetition, we just show the process of how we design the loss function without displaying the comparison of loss function here. 12.Full text (answer to Reviewer #3, suggestion #12) Comment: “The tables in this paper can be converted into figures as much as possible, so that the results can be observed more intuitively.” Response: Thanks for your suggestion. Since PLOS ONE require for editable original data, we design our tables like this. According to your suggestion, we prepare a copy of tables with figure forms, and we would contact with the editor to ensure a both intuitive and acceptable form. Submitted filename: rebuttal letter.docx Click here for additional data file. 13 Oct 2020 Single image super-resolution via Image Quality Assessment-Guided Deep Learning Network PONE-D-20-25793R1 Dear Dr. Sun, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Zhihan Lv, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed Reviewer #3: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Partly Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Under the background that deep learning has excellent characteristics in the field of single image super-resolution (SR) (SISR) and is widely used, a SISR method guided by image quality assessment (IQA) is proposed based on deep learning structure, in order to balance the perceived quality and distortion measurement corresponding to SISR results. In this method, IQA network is introduced to extract the perceptual features in SR, and the corresponding loss fused with the original absolute pixel loss is calculated to guide the adjustment of SR network parameters. In addition, an interactive training model is constructed by cascaded network, and based on the problem of insufficient samples in the training process, the hinge loss method of pairwise sorting is proposed. In this revision, authors explain and discuss my concerns in details. Therefore, I recommend this paper is mature enough to meet the publication quality. Reviewer #2: In order to achieve a good tradeoff between perceived quality and distortion measurement of service request results, and effectively improve image resolution, an image quality evaluation guided SISR method is proposed in service request architecture. From the results of stochastic resonance, IQA network is introduced to extract perceptual features. Through the cascade network, an interactive training model is established. The results show that the performance of the model is better, which can significantly improve the image resolution. Author improved his paper. The paper can be accepted now. Reviewer #3: In this study, a deep learning architecture based single image super-resolution method guided by image quality assessment is proposed to achieve a good compromise between perceived quality and distortion measurement of super-resolution results. However, the method proposed in this paper is different from the super-resolution method of opportunity deep learning. In this paper, an image quality assessment network is introduced to extract perceptual features from super-resolution. By calculating the corresponding loss fused with the original absolute pixel loss, the adjustment of super-resolution network parameters is further guided. At the same time, through the establishment of interactive training network model, a hinge loss method based on acceptance sorting is proposed to overcome the shortcomings in the training process and solve the problem of image quality assessment and heterogeneous data sets used in super-resolution network. The authors have already addressed all the comments. I think this paper can be published now. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes: Xin Gao 15 Oct 2020 PONE-D-20-25793R1 Single image super-resolution via Image Quality Assessment-Guided Deep Learning Network Dear Dr. Sun: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Zhihan Lv Academic Editor PLOS ONE
  15 in total

1.  Blind image quality assessment: a natural scene statistics approach in the DCT domain.

Authors:  Michele A Saad; Alan C Bovik; Christophe Charrier
Journal:  IEEE Trans Image Process       Date:  2012-03-21       Impact factor: 10.856

2.  Information content weighting for perceptual image quality assessment.

Authors:  Zhou Wang; Qiang Li
Journal:  IEEE Trans Image Process       Date:  2010-11-15       Impact factor: 10.856

3.  An information fidelity criterion for image quality assessment using natural scene statistics.

Authors:  Hamid Rahim Sheikh; Alan Conrad Bovik; Gustavo de Veciana
Journal:  IEEE Trans Image Process       Date:  2005-12       Impact factor: 10.856

4.  Image information and visual quality.

Authors:  Hamid Rahim Sheikh; Alan C Bovik
Journal:  IEEE Trans Image Process       Date:  2006-02       Impact factor: 10.856

5.  A feature-enriched completely blind image quality evaluator.

Authors:  Alan C Bovik
Journal:  IEEE Trans Image Process       Date:  2015-04-24       Impact factor: 10.856

6.  Learning multiple linear mappings for efficient single image super-resolution.

Authors:  Kaibing Zhang; Dacheng Tao; Xinbo Gao; Xuelong Li; Zenggang Xiong
Journal:  IEEE Trans Image Process       Date:  2015-01-07       Impact factor: 10.856

7.  No-reference image quality assessment in the spatial domain.

Authors:  Anish Mittal; Anush Krishna Moorthy; Alan Conrad Bovik
Journal:  IEEE Trans Image Process       Date:  2012-08-17       Impact factor: 10.856

8.  Single image super-resolution with non-local means and steering kernel regression.

Authors:  Kaibing Zhang; Xinbo Gao; Dacheng Tao; Xuelong Li
Journal:  IEEE Trans Image Process       Date:  2012-07-16       Impact factor: 10.856

9.  dipIQ: Blind Image Quality Assessment by Learning-to-Rank Discriminable Image Pairs.

Authors: 
Journal:  IEEE Trans Image Process       Date:  2017-05-26       Impact factor: 10.856

10.  KonIQ-10k: An ecologically valid database for deep learning of blind image quality assessment.

Authors:  Vlad Hosu; Hanhe Lin; Tamas Sziranyi; Dietmar Saupe
Journal:  IEEE Trans Image Process       Date:  2020-01-24       Impact factor: 10.856

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.