Literature DB >> 35582159

Single-image reconstruction using novel super-resolution technique for large-scaled images.

Ramanath Datta1, Sekhar Mandal2, Saiyed Umer3, Ahmad Ali AlZubi4, Abdullah Alharbi4, Jazem Mutared Alanazi4.   

Abstract

A fast and novel method for single-image reconstruction using the super-resolution (SR) technique has been proposed in this paper. The working principle of the proposed scheme has been divided into three components. A low-resolution image is divided into several homogeneous or non-homogeneous regions in the first component. This partition is based on the analysis of texture patterns within that region. Only the non-homogeneous regions undergo the sparse representation for SR image reconstruction in the second component. The obtained reconstructed region from the second component undergoes a statistical-based prediction model to generate its more enhanced version in the third component. The remaining homogeneous regions are bicubic interpolated and reflect the required high-resolution image. The proposed technique is applied to some Large-scale electrical, machine and civil architectural design images. The purpose of using these images is that these images are huge in size, and processing such large images for any application is time-consuming. The proposed SR technique results in a better reconstructed SR image from its lower version with low time complexity. The performance of the proposed system on the electrical, machine and civil architectural design images is compared with the state-of-the-art methods, and it is shown that the proposed scheme outperforms the other competing methods.
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022.

Entities:  

Keywords:  Image reconstruction; Patch; Prediction model; Sparse representation; Super resolution

Year:  2022        PMID: 35582159      PMCID: PMC9099350          DOI: 10.1007/s00500-022-07142-4

Source DB:  PubMed          Journal:  Soft comput        ISSN: 1432-7643            Impact factor:   3.732


Introduction

A high-resolution image is always desirable in most digital image processing applications such as office automation, medical imaging, remote sensing, and video surveillance. A high-resolution image gives a better appearance and better classification of image regions. The image resolution means the scene area is represented by a single pixel, and higher resolution (HR) refers to a small area resulting in more spatial image details. This resolution of an image depends on the camera device used during image acquisition. There are various issues to obtaining a high-resolution image using hardware devices. However, algorithmic image reconstruction techniques are more promising for bringing HR images from low-resolution (LR) ones. We may get the HR images from observed LR images by suppressing the degradation introduced during image acquisition and increasing the high-frequency components. The process is referred to as Super-Resolution (SR) technique. Moreover, during the SR process, the non-redundant information in LR images is combined with domain-specific knowledge to create an HR image. One commonly used approach to SR is single-image interpolation, which is used to increase the image size. In this case, no additional information is provided, and hence, the quality of the HR image is limited due to the nature of the problem. We adopt here the patch-based (or exemplar-based) method for SR, where an LR image is partitioned into several patches, and then each LR patch is replaced by its corresponding HR patch to reconstruct the HR image, say, . Various methods differ mainly due to: (1) How to partition into LR patches; (2) How to generate or select HR patch corresponding to a given LR patch; (3) How to combine HR patches to reconstruct . The objective of this problem domain is to obtain the best possible SR image and accept the SR image as quickly as possible. The solution to this problem depends upon available single LR images or multiple LR images. So, the methods of SR technique are categorized into single-frame SR and multi-frame SR techniques. Moreover, the single-frame SR algorithm is also categorized into two approaches: (a) interpolation-based method and (b) patch-based method (Milanfa 2010). Most of these methods use the single-frame and the patch-based super-resolution techniques to reconstruct HR images from LR images. These methods have achieved better performance in the research area of image super-resolution. From the great success of this single-frame and patch-based SR method (Yang et al. 2010), in this paper, patches are extracted from the input image I. Then, the implementation focuses on the recovering SR version of the given low-resolution version of I. To capture the significant co-occurrence prior and speed up the process, we obtain some representation from the image patch pairs extracted from HR and LR images. These representations are the first- and second-order gradients from each patch. The contributions of the proposed work are as follows:The organization of this paper is as follows: Sect. 2 discusses some related works. Section 3 describes the proposed method with a detailed discussion of each component of the proposed system. The experimental results and discussions have been demonstrated in Sect. 4. Section 5 concludes this paper. To speed up the super-resolution (SR) image reconstruction, the LR patches are classified into homogeneous and non-homogeneous groups based on the features extracted from each patch. If an LR patch belongs to the non-homogeneous group, the proposed SR method is applied to obtain its HR patch; otherwise, the bi-cubic interpolation method is applied on the homogeneous patches. For the classification of homogeneous and non-homogeneous LR patches, an unsupervised clustering technique has been employed on the texture features computed from the grey-level co-occurrence matrices of the image patches. In the training phase, the features extracted from the LR and HR patch pair of and undergo a sparse representation to generate Dictionary. We have adopted this sparse representation using Dictionary, as a recent result suggests that HR signals’ linear relationships are well recovered from their lower-dimensional projection obtained from sparse representation. Moreover, image patch-based sparse representation plays an important role in regularizing SR ill-posed problems with effectiveness and robustness properties. The obtained HR image patches are fed to a prediction model to produce refined SR image patches.

Related work

The interpolation-based methods such as Bilinear or Bicubic interpolation applied on smooth images with some jagged and ringing artefacts to exploit the natural image prior that results better results (Dai et al. 2007; Sun et al. 2008). The patch-based methods locally capture the prior co-occurrences between LR and HR image patch pairs. Then, it employs a large database of LR and HR patch pairs and uses a learning mechanism for the corresponding mapping between the LR and HR image patches that are applied to a new LR image to reconstruct its most likely HR version (Yang et al. 2010; Glasner et al. 2009). Generally, these methods are based on the image edge prior or image gradient prior (Sun et al. 2008). The objective of this kind of method is to magnify the image so that the edge sharpness and the texture details within the image are preserved. Natsui and Nagao (2016) proposed a single-frame SR technique using multiple graph-structured programs based on Cartesian Genetic Programming, which is one of the evolutionary methods like genetic algorithms. Lai et al. (2012) proposed a method for single-frame SR technique, where a total variation regularization was used to minimize the iterative back-projection-based SR reconstruction error by suppressing the chessboard and ringing artefacts at the time of acquiring the high-resolution image. Dang and Radha (2017) proposed a single-image SR technique using Tangent Space Learning of high-resolution patch manifold, which is based on the linear approximation of HR patch space using sparse subspace clustering algorithms. A method for compressed sensing reconstruction method with single plane wave transmission for super-resolution of ultrasound images had been proposed by Shu et al. (2018). A multi-frame SR technique based on the fusion of multiple low-resolution images captured at Visual Wavelength and Near-Infrared lighting conditions with different camera positions for synthesizing HR colour images had been proposed by Honda et al. (2018). A well comprehensive survey for super-resolution of biometric images such as the face (2D+3D), iris, finger-print and gait based on operation domain (spatial and frequency), singe-frame, multi-frame, reconstruction, learning, feature domain and deep-learning-based SR techniques had been discussed by Nguyen et al. (2018). Akyol and GöKmen (2012) have proposed a method for face super-resolution by enhancing the shape and texture information of faces. A multi-channel constraint-based image super-resolution technique that has the capability for collaborative representations, clustering, multilayer mapping relationships to reconstruct HR images from LR images had been proposed by Cao and Li (2018). Zhao et al. (2017) proposed a method for SR technique where the adaptive sparse coding had been employed by establishing the regularization parameters with integrating correlation and sparsity terms in the regularization. Their SR method modulates the collaborative representation and the sparse representation. The CNN-based architectures with a subpixel network have been considered for the resolution enhancement of 2D cone-beam CT image slices of ex vivo teeth in Hatvani et al. (2018). A novel dehazing model for remote sensing images has been proposed by Singh and Kumar (2018a). Deep Learning and Transfer Learning-Based Super-Resolution Reconstruction from Single Medical Image has been proposed by Zhang and An (2017). A deep learning framework based on a generative adversarial network to perform super-resolution incoherent imaging systems has been proposed in  Liu et al. (2019). Metaheuristic-based deep COVID-19 screening model from chest X-ray has been proposed in images (Kaur et al. 2021). Further, a multi-objective differential evolution-based deep neural networks had been employed for multi-modality medical image fusion technique by Kaur and Singh (2021). Singh et al. had built a model for defogging of road images using gain coefficient-based trilateral filter (Singh and Kumar 2018b).

Proposed methodology

The proposed super-resolution (SR) technique is applied mainly to reconstruct the high-resolution (HR) image from the low-resolution (LR) image of electrical circuit drawings using the single-frame and patch-based super-resolution technique. The proposed method can also be applied to other drawings, such as mechanical drawing and architectural drawing. We have done some experiments in this direction to see the performance of the proposed method. The proposed method has three main phases: (1) In the first phase, we partition m-training images s into several patches. Then, the patches are divided into two clusters based on the homogeneous/non-homogeneous texture pattern using the k-means clustering algorithm. Then, classification of an LR patch of an individual LR image is done by comparing its GLCM texture features with that of the cluster centres; (2) In the second phase, the bicubic interpolation (Hwang and Lee 2004) (i.e. no super-resolution is applied on it) is applied on the homogeneous patches, while the non-homogeneous patches undergo for sparse representation based SR method followed by statistical prediction model based algorithms; (3) Here, the colour image I is transformed to its colour channel and only luminance channel (Y) is used in the first and second phase. The remaining channels are interpolated using the bi-cubic interpolation technique and are combined with the obtained HR Y-channel in the third phase. A brief overview of the proposed system is shown in Fig. 1.
Fig. 1

Data flow diagram of the proposed system

Data flow diagram of the proposed system

Preprocessing

During preprocessing, (R, G, B) of each input low-resolution (LR) image is transformed to colour channel, and super-resolution is performed only on Y-channel LR image to obtain the Y-channel HR (High Resolution) image. Here, we have employed the patch-based SR (Super-Resolution) technique. Hence, a patch with overlapping of p pixels both the horizontal and vertical directions is considered from . Then, Gray-level Co-occurrences Matrix (GLCM) texture features (Umer et al. 2016) are computed from each , i.e. gives , where N be the number of patches and d is the number of texture features from each patch. Now, k-means clustering algorithm is applied on GLCM texture features collected from m-training images to obtain a code-book considering two classes (): homogeneous and non-homogeneous. Here, the usefulness of k-means clustering algorithm are: (1) It can well differentiate between the homogeneous and non-homogeneous patterns within the image based on GLCM texture features computed from patches; (2) It takes lesser time for finding the dictionary of discriminate patches; (3) The other clustering algorithm can take more time with better features. At the same time, the k-means provides an excellent pre-clustering technique and reduces the spaces between clusters into disjoint smaller sub-spaces better than the other clustering algorithms. For a given image , the feature vector corresponding to each is compared with using minimum distance classifier to predict whether has homogeneous or non-homogeneous region. The non-homogeneous is considered as LR () and undergoes for super-resolution technique. In the next section, we discuss the super-resolution technique using sparse representation.

SR using sparse representation (Case-1)

Sparse representation is a powerful technique for representing, compressing, and processing high dimensional signals in low dimensional space. The classes of signals such as images and audio can be modelled through sparse representation concerning their fixed bases. Effective and efficient convex optimization or greedy pursuit algorithms are available for computing those representations with high fidelity. These are the main reasons for a successful and wide use of sparse representation (Wright et al. 2010). So, the sparse representation has been extensively used in many computer vision tasks, such as super image resolution, image de-noising, image inpainting, motion and data segmentation, image classification, and face recognition. In most of these applications, using sparsity as prior, sparse representation outperforms the state-of-the-art results. We have employed patch-based sparse representation for single-image super-resolution (SR) in this work. However, the non-overlapping patch-wise operation may cause block artefacts; hence, overlapping patches are employed to overcome this problem. The sparse representation provides high-quality image reconstruction for SR algorithms by using overlapping patches over the image (Toutounchi et al. 2017). On the other hand, overlapping patches leads to more time consumption. To overcome this, we have adopted only a tiny overlap between patches. To reduce the computational cost further, we have employed a technique for patch selection based on its homogeneous or non-homogeneous texture pattern (discussed above) and applied the sparse representation only on those patches that are non-homogeneous. During sparse representation, we consider LR patches () from Y-channel of LR image and then extract features from each . In the training phase, more precisely during Dictionary learning, we take high-resolution (HR) image, say, from the application domain and deliberately form its low-resolution (LR) version, say, through the appropriate down-sampling method. If is a gray-level image, so is . However, if we had to handle colour images, then consider that and are corresponding Y-channel images. Different features can be extracted from the LR patches which are mentioned in the existing literature. Freeman et al. (2000) extracted edge-based information from LR patches by using a high pass filter. Chang et al. (2004) and Yang et al. (2010) have extracted first- and second-order gradients from LR and HR patches. In this work, we also use first () and second-order () gradients of patches. These filters are directly applied to the training images (LR and HR ), which yields four gradient maps at each location. Then, any patch is represented in terms of feature vector corresponding to its gradient map. To track the correspondence between HR (obtained original image) and LR patches (down-sampled image), feature vectors of and are concatenated to form . The above-mentioned feature vectors are used to create two dictionaries and corresponding to LR and HR patches, respectively, which will be subsequently exploited to obtain sparse representations of LR and HR patches, respectively. Since sparse representation is an ill-posed problem, we take the help of constraints to solve this problem. First, the image observation model is considered where a low-resolution image is obtained from the given high-resolution image , such thatwhere is the down-sampling factor and is the blurring filter. Since Eq. (1) represents a many-to-one mapping, for a given low-resolution image infinitely many solutions of can be obtained by solving Eq. (1). To resolve this issue, we consider that each patch can be represented as a sparse linear combination of dictionary . Similar concept is true for also. Or, in general, the vector can be represented by a sparse combination of dictionary , which in turn would be concatenation of and , i.e. . The vector v contains various features including pixel intensities of HR and LR patches concatenated together. Hence,for some coefficient vector , where K is the number of words or elements in the Dictionary and , which is a sparsity constraint. Note that both and are unknown and are learned simultaneously by optimizing the following equation.Thus, the above mentioned [Eq. (3)] sparse representation corresponds to both HR and LR patches together, which have spatial compatibility between the neighbours. So using this for x, the entire HR image is regularized and refined using the reconstruction constant. For this purpose, local modelling for sparsity prior has been introduced for local patches to recover some HR patch x which may be lost during processing. During local modelling, the HR patch is derived for the corresponding LR patch . For these two dictionaries, and is needed separately. Recall that dictionary is generated, through sparse model of the patches, from vector . This suggests that dictionary can straightaway be decomposed into and . Also note that here both these dictionaries and have the same sparse representation (i.e. or are same) for each and patch pairs. And this suggests the method for generating HR patch, given the test LR patch and two dictionaries. For each input LR patch , the corresponding sparse representation is obtained based on . Now, the patch bases of are combined according to to generate high-resolution patch . This solution may be mathematically formulated aswhere is the feature extraction operator that includes intensity mapping of the patch along with first- and second-order spatial derivatives of the patch. Solving Eq. (4), we obtain a to represent y in terms of the dictionary . Now, same is used to generate based on . So we haveThe problem defined in Eq.(4) is NP-hard problem (Aharon et al. 2006), and it says that the obtained sparse representation could be sufficiently sparse. This can be efficiently approximated by introducing -norm in place of norm as follows:Now, the regularization parameter is introduced in Eq. (6) to obtain following loss functionHere, the term balances the solution’s sparsity and the approximation’s fidelity to y. The formulation in Eq. (7) is Lasso (Yang et al. 2012) optimization problem, which is a linear regression regularization with -norm on . In this work, we obtain corresponding HR patches for SR image to only those LR patches which are non-homogeneous regions. So, during processing, each non-homogeneous region is considered as image, and then, a patch slides in a raster scan, i.e. horizontally and then vertically. So, there may be the possibility of overlapping patches for the ambiguous region. So, to check closely the previously computed HR patch from the SR reconstruction of y, we modify Eq.(7) as follows:where w contains the value for the particular reconstructed HR image on overlap and is a matrix for overlap region between the target patch and the previously computed HR patch. The optimization in Eq.(8) is formulated aswhere and . The parameter is the control of trade-off between the matching of LR input y and finding a HR patch x which is compatible with its neighbour. The final optimal solution of Eq.(9) is , and the final HR patch is computed as Some generic HR images Result of the proposed SR technique using sparse representation Result of the proposed SR technique using sparse representation followed by prediction model The HR patch obtained from Eq. (10) has much more better resolution then the LR patch y.

SR using prediction model (Case-2)

Now, the obtained HR patch undergoes a predicted model which generates much higher resolution HR patch considering the previously obtained as LR patch y. Here, also two dictionaries and have been used. The objective of this prediction model is to predict the missing HR details for each LR patch y (here ) via and , having different elements. Then, a statistical prediction model is being utilized for prediction the HR representation vector of each patch from its corresponding . In above Case-1 technique, the sparse coefficient is similar for . To introduce more sparsity in , is computed from which is as follows‘where is a threshold which adaptively changes for each LR patch and is computed based on following criterion for , where is a pre-specified parameter. So, the sparsity patterns for LR patch and for the HR patch capture the relationships between these two patterns and for this a statistical prediction model is required which is described in the following section. Comparative performance for Case-1 and Case-2 SR techniques for generic images using PSNR and SSIM , , , , , , , are large-scaled electrical circuit images, whereas , , , , , , , are the large-scaled machine and civil layout design images To capture the statistical dependencies within the sparsity pattern and , the restricted Boltzmann machine (RBM) (Sutskever et al. 2009) technique is employed which may be described by the conditional probability (Peleg and Elad 2014) as follows:where is the bias vector and is an interaction matrix between and . Thus,where is the sigmoid function and is due to the elements of which are statistically independent of . RBM is an exponential model that works with binary vectors, and it leads to the conditional marginal probability for each element of in such thatThe obtained is employed to compute the HR co-efficient using based on the below criterion:where is the Gaussian distribution, i.e. , where and . This leads to the following conditional expectation:The models defined in Eq.(15) and (16) show the linear mapping of to and it happens when is known. In the case where relation between and is nonlinear, then the final estimate of using , , may be given by an MMSE estimator as follows:From the above MMSE estimator (Eq. (17)), the obtained is not only sparse but also leads to better signal recovery (Peleg and Elad 2014). Moreover, MMSE estimator (Baum and Eagon 1967) can be represented as a product of linear terms with respect to and a nonlinear term concerning . The final predicted HR patch x is given by

Initial validation of proposed methods

Since we are more acquainted with natural images (or scenes), the above-discussed super-resolution (SR) techniques (both Case-1 and Case-2) are applied on generic images like Butterfly, Girl, Pepper, Starfish and Zebra as shown in Fig. 2. The effect of SR technique using sparse representation (Case-1) is shown in Fig. 3; whereas that of SR technique using sparse representation followed by prediction model (Case-2) is shown in Fig. 4. Moreover, the performance is also shown in Table 1 in terms of Peak Signal to Noise Ratio (PSNR) and Structural Similarity Index (SSIM) (Hore and Ziou 2010) indexes. The Table shows that the performance of Case-2 is much better than Case-1.
Fig. 2

Some generic HR images

Fig. 3

Result of the proposed SR technique using sparse representation

Fig. 4

Result of the proposed SR technique using sparse representation followed by prediction model

Table 1

Comparative performance for Case-1 and Case-2 SR techniques for generic images using PSNR and SSIM

ButterflyGirlPepperStarfishZebra
Case-132.9241.9734.0435.3931.02
0.87510.95790.83800.91480.6839
Case-235.7445.0134.8938.8331.48
0.94120.97540.87720.96050.7344
Note that the collection of HR and LR patch pairs, collected from natural image pairs, undergoes a dictionary learning process to obtain and . For this learning, we have employed the parametric dictionary learning discussed in Yang et al. (2010) and Peleg and Elad (2014).

Experimental results

This section presents and evaluates the performance of proposed SR techniques when applied to engineering drawing (i.e. line drawing) images in general and electrical circuit drawings in particular. Description for employed Image names and their sizes

Data used

We demonstrate the results of the proposed single-image SR technique on electrical circuits, machine layout, and civil architectural design images originally captured in a high-resolution camera. These images are huge, and using these images for any application is computationally time-consuming on moderate capacity machines. We keep only its lower version (i.e. down-sampled version) and apply the proposed super-resolution (SR) techniques whenever required to speed up the process. Figure 5 shows the original version of the electrical circuit images (, , , , , , , ), machine layout and civil architectural design images (, , , , , , , ). Table 2 provides the size of these images in terms of (Rows) (Columns) (Channels). We have applied the proposed method to these images and have obtained the performance. These images are large-scaled images and huge. To make the system understandable, convenient and comparable with the other state-of-the-art methods, we have manually cropped a small region from these images (shown in Fig. 5) and considered them as original HR images and then down-sampled each cropped region to its lower resolution images (shown in Fig. 5).
Fig. 5

, , , , , , , are large-scaled electrical circuit images, whereas , , , , , , , are the large-scaled machine and civil layout design images

Table 2

Description for employed Image names and their sizes

NameSizeNameSize
Img1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1700 \times 2400 \times 3$$\end{document}1700×2400×3Img9\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$900 \times 2100 \times 3$$\end{document}900×2100×3
Img2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1650 \times 3050 \times 3$$\end{document}1650×3050×3Img10\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$600 \times 900 \times 3$$\end{document}600×900×3
Img3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1850 \times 3600 \times 3$$\end{document}1850×3600×3Img11\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$600 \times 900 \times 3$$\end{document}600×900×3
Img4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$2200 \times 3300 \times 3$$\end{document}2200×3300×3Img12\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1800 \times 2400 \times 3$$\end{document}1800×2400×3
Img5\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$750 \times 1600 \times 3$$\end{document}750×1600×3Img13\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1200 \times 1500 \times 3$$\end{document}1200×1500×3
Img6\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$1100 \times 1700 \times 3$$\end{document}1100×1700×3Img14\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$750 \times 900 \times 3$$\end{document}750×900×3
Img7\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$3500 \times 4900 \times 3$$\end{document}3500×4900×3Img15\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$2100 \times 3000 \times 3$$\end{document}2100×3000×3
Img8\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$2500 \times 3300 \times 3$$\end{document}2500×3300×3Img16\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$2400 \times 3600 \times 3$$\end{document}2400×3600×3
Performance of the proposed SR technique using the dictionary pair (, ), (,) and (, , respectively

Results and discussion

During experimentation, each image I is converted to its (Y, , ) colour channels. Then, only Y-channel of LR image () is considered to undergo to the proposed algorithm to reconstruct Y-Channel of HR image . Corresponding and are up-sampled and interpolated to be combined with and then, converted to RGB image. Now, a window of size slides over with 50% overlapping of pixels in horizontal and vertical direction picks up the patches of same size. Then, GLCM (Umer et al. 2016) texture features are extracted from each . The collection of these texture features selected randomly from training images are used to obtain a code-book using k-means clustering algorithm for evaluating the homogeneous and non-homogeneous region for each . Performance of the reconstructed SR image using the proposed system with respect to PSNR and SSIM indexes Any homogeneous patch does not undergo the proposed SR technique. It is bicubic interpolated and converted to its RGB version. The up-sampled patches are used to form the high-resolution (HR) image. The non-homogeneous patches undergo the proposed SR technique. Here, during the super-resolution technique, we consider each non-homogeneous patch as an LR image and its reconstructed SR patch is utilized to form the final HR image. Note that for each non-homogeneous LR region , we apply first-order gradient (, ) and second-order gradient (, ) features which yield four gradient maps at each location. Then, a window with 50% overlapping is considered over each feature map of . Now, feature vector with respect to each position of is extracted from four feature map which are concatenated to obtain a feature vector for to obtain its representation as . Then, the joint feature learning technique for the dictionary has been performed on the concatenated HR and LR patch pair features which derive and , dictionary for the corresponding HR and LR patch pairs, respectively. Here, for each dictionary or , items are considered. During dictionary learning, the colour images are transformed from the RGB to Y channels and then, features are extracted from the corresponding Y-channel only. The i-th LR patch corresponding to the non-homogeneous region is used to obtain it’s SR patch by employing the proposed SR technique via sparse representation technique described in Sect. 3.2. The obtained undergoes to the statistical prediction model described in Sect. 3.3. The performance of the proposed SR technique for electrical circuit images , , , , , , , and for machine and civil layout design images , , , , , , and using the dictionary pairs (, ), (, ) and (, ) are shown in Fig. 6, respectively.
Fig. 6

Performance of the proposed SR technique using the dictionary pair (, ), (,) and (, , respectively

Table 3 shows the performance of the proposed system in terms of PSNR (first row for each image) and SSIM (second row for each image) indexes for the images shown in Fig. 6 using the dictionary pairs (,), (, ) and (, ), respectively.
Table 3

Performance of the reconstructed SR image using the proposed system with respect to PSNR and SSIM indexes

\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_1$$\end{document}Img1,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_2$$\end{document}Img2,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_3$$\end{document}Img3,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_4$$\end{document}Img4,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_5$$\end{document}Img5,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_6$$\end{document}Img6,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_7$$\end{document}Img7,\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_8$$\end{document}Img8
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$D_h \in {\mathbb {R}}^{512 \times 100}$$\end{document}DhR512×10033.8240.5232.6431.2231.5631.1732.6532.85
0.87920.91970.89910.92870.91350.91490.91910.9372
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$D_h \in {\mathbb {R}}^{1024 \times 100}$$\end{document}DhR1024×10035.8242.5234.3833.4033.5232.3132.4735.08
0.87910.92980.87840.91810.92960.92030.92900.9191
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$D_h \in {\mathbb {R}}^{2048 \times 100}$$\end{document}DhR2048×10040.8446.5239.8938.6338.8937.5936.5140.11
0.99940.99990.99870.99820.99150.99280.99900.9991
From the performance, as shown in Table 3 using different dictionary pairs in Fig. 6 for electrical circuit, mechanical, civil and architectural design images and also from Table. 3, it is observed that for dictionary pair (, ) produces slightly better reconstructed SR images. For further comparison of the proposed system with other competing methods such as Bicubic-Interpolation, Zhang et al. (2015), Marquina and Osher (2008), Purkait and Chanda (2012) and Yang et al. (2010), we have employed (, ) dictionary pair. Here, both visual perception analysis (Figs. 7 and 8) and the quantitative measures PSNR and SSIM (Table 4) have been employed with the reconstructed SR image obtained from the competing method and the proposed one. The performance of the proposed system with respect to visual information fidelity (VIF) evaluation index [?] is also reported in Table 4 (Third row) along with the PSNR and SSIM indexes. The performance comparison with respect to these indexes shows the superiority of the proposed system.
Fig. 7

Performance comparison of the proposed SR technique with the other existing state-of-the-art methods for , , , , , , and , respectively

Fig. 8

Performance comparison of the proposed SR technique with the other existing state-of-the-art methods for , , , , , , and , respectively

Table 4

Performance comparison of the proposed SR technique with the other existing techniques with respect to PSNR (first-rows), SSIM (second-rows), and VIF (third-rows) indexes

\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_1$$\end{document}Img1\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_2$$\end{document}Img2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_3$$\end{document}Img3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_4$$\end{document}Img4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_5$$\end{document}Img5\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_6$$\end{document}Img6\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_7$$\end{document}Img7\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Img_8$$\end{document}Img8
Bicubic34.8338.0239.3636.3637.8137.2235.8640.29
0.99920.99970.99930.99860.99250.99470.99910.99917
0.47580.44770.43590.46230.44960.45470.46670.4278
Zhang et al. (2015)37.6837.7034.7532.7235.6636.7332.8136.88
0.94220.96560.90600.90210.89270.86520.96870.9769
0.45070.45050.47650.49430.46850.45910.49350.4577
Marquina et al. (2008)41.3645.7743.2538.7638.5538.1635.9239.82
0.99990.99970.99950.99930.99740.99970.99930.9993
0.41840.37960.40170.44120.44310.44650.46620.4319
Purkait et al. (2012)38.0045.0442.9939.0739.1536.5437.2940.90
0.99990.99970.99970.99970.99800.99980.99970.9997
0.44790.3860.4040.43850.43780.46070.45410.4224
Yang et al. (2010)37.3243.5835.7238.5133.7133.8134.2439.69
0.99940.99930.99720.99690.99100.99210.99720.9981
0.45390.39880.46790.44340.48560.48470.48090.433
Proposed40.8446.5239.8938.6338.8937.5936.5140.11
0.99940.99990.99870.99820.99150.99280.99900.9991
0.59070.59570.58990.58880.5890.58790.58690.5901
Performance comparison of the proposed SR technique with the other existing state-of-the-art methods for , , , , , , and , respectively Performance comparison of the proposed SR technique with the other existing state-of-the-art methods for , , , , , , and , respectively Performance comparison of the proposed SR technique with the other existing techniques with respect to PSNR (first-rows), SSIM (second-rows), and VIF (third-rows) indexes Bicubic-Interpolation is widely used for data interpolation on the two-dimensional regular grid. It is a relatively standard technique in image interpolation with good results and low complexity. Zhang et al. (2015) have given empirical studies on the sensitivity of different single-image super-resolution algorithms based on different blurring kernels. Marquina and Osher (2008) proposed a convolutional model that uses the total variation of the signal followed by the Bregman iterative refinement procedure for single-image super-resolution. Purkait and Chanda (2012) had modelled a nonlinear regularization method based on multiscale morphology for reserving the edges for super-resolution (SR) image reconstruction. Finally, Yang et al. (2010) had proposed an image super-resolution method using sparse representation technique only where the sparse representation for each low-resolution patch was used to get its coefficient as representation to obtain the corresponding high-resolution patch. From Figs. 7 and 8, it has been shown that the proposed system applied on images has better reconstructed. Here, Table 5 shows the comparison performance with respect to time in Sec.. Since the proposed system and method in Purkait and Chanda (2012) outperforms other competing methods. Additionally, the performance of the proposed method and method in Purkait and Chanda (2012) is more or less the same even in some circumstances. It has been observed that for some images, the proposed system overcomes (Purkait and Chanda 2012). But due to some experimental setup and might be in some tuning of parameters, the proposed system will take lesser time than Purkait and Chanda (2012). Moreover, the performance reported in Table 4, it has been observed that the proposed system gives outstanding performance in terms of PSNR, SSIM, and VIF as compared to the methods reported in Table 5. Hence, the proposed system outperforms other competing methods.
Table 5

Average performance of the competing methods and the proposed system in terms of time (S)

NameTime (S)
Zhang et al. (2015)10.82
Marquina et al. (2008)08.12
Purkait et al. (2012)07.11
Yang et al. (2010)10.24
Proposed07.31

Conclusions

This paper presents a novel method of single-image super-resolution technique. The proposed scheme has three components. In the first component, to speed up the process, the input image is divided into several regions analysed into homogeneous or non-homogeneous regions based on the texture pattern analysis in those regions. The non-homogeneous region undergoes a sparse representation technique to get a better-reconstructed HR region in the second component. In the third component, the reconstructed HR region from the second component undergoes a prediction model based on the statistical modelling of sparse representation using the Boltzmann machine technique to get a more enhanced reconstructed HR image. The homogeneous regions are bicubic interpolated and reflect the outcome image. Experimental results demonstrate that the proposed method better reconstructed SR images for electrical, machine, and civil design images. The comparison with the existing state-of-the-art methods shows that the proposed system outperforms other methods efficiently. The proposed approach might take some time to generate super-resolution images for the larger-scaled images. In the future, some deep learning models will be employed to solve super-resolution problems for the vastly scaled engineering design images. Average performance of the competing methods and the proposed system in terms of time (S)
  9 in total

1.  Coupled dictionary training for image super-resolution.

Authors:  Jianchao Yang; Zhaowen Wang; Zhe Lin; Scott Cohen; Thomas Huang
Journal:  IEEE Trans Image Process       Date:  2012-04-03       Impact factor: 10.856

2.  Super resolution image reconstruction through Bregman iteration using morphologic regularization.

Authors:  Pulak Purkait; Bhabatosh Chanda
Journal:  IEEE Trans Image Process       Date:  2012-05-25       Impact factor: 10.856

3.  Image super-resolution via sparse representation.

Authors:  Jianchao Yang; John Wright; Thomas S Huang; Yi Ma
Journal:  IEEE Trans Image Process       Date:  2010-05-18       Impact factor: 10.856

4.  Multiple Class Segmentation Using A Unified Framework over Mean-Shift Patches.

Authors:  Lin Yang; Peter Meer; David J Foran
Journal:  Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit       Date:  2007-07-16

5.  A statistical prediction model based on sparse representations for single image super-resolution.

Authors:  Tomer Peleg; Michael Elad
Journal:  IEEE Trans Image Process       Date:  2014-06       Impact factor: 10.856

6.  Multi-modality medical image fusion technique using multi-objective differential evolution based deep neural networks.

Authors:  Manjit Kaur; Dilbag Singh
Journal:  J Ambient Intell Humaniz Comput       Date:  2020-08-08

7.  Deep Learning- and Transfer Learning-Based Super Resolution Reconstruction from Single Medical Image.

Authors:  YiNan Zhang; MingQiang An
Journal:  J Healthc Eng       Date:  2017-07-06       Impact factor: 2.682

8.  Deep learning-based super-resolution in coherent imaging systems.

Authors:  Tairan Liu; Kevin de Haan; Yair Rivenson; Zhensong Wei; Xin Zeng; Yibo Zhang; Aydogan Ozcan
Journal:  Sci Rep       Date:  2019-03-08       Impact factor: 4.379

9.  Metaheuristic-based Deep COVID-19 Screening Model from Chest X-Ray Images.

Authors:  Manjit Kaur; Vijay Kumar; Vaishali Yadav; Dilbag Singh; Naresh Kumar; Nripendra Narayan Das
Journal:  J Healthc Eng       Date:  2021-03-01       Impact factor: 2.682

  9 in total
  1 in total

1.  Image Super-Resolution Reconstruction Method for Lung Cancer CT-Scanned Images Based on Neural Network.

Authors:  Jianming Xu; Weichun Liu; Yang Qin; Guangrong Xu
Journal:  Biomed Res Int       Date:  2022-07-18       Impact factor: 3.246

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.