Literature DB >> 35327830

Self-Adaptive Image Thresholding within Nonextensive Entropy and the Variance of the Gray-Level Distribution.

Qingyu Deng1, Zeyi Shi1, Congjie Ou1.   

Abstract

In order to automatically recognize different kinds of objects from their backgrounds, a self-adaptive segmentation algorithm that can effectively extract the targets from various surroundings is of great importance. Image thresholding is widely adopted in this field because of its simplicity and high efficiency. The entropy-based and variance-based algorithms are two main kinds of image thresholding methods, and have been independently developed for different kinds of images over the years. In this paper, their advantages are combined and a new algorithm is proposed to deal with a more general scope of images, including the long-range correlations among the pixels that can be determined by a nonextensive parameter. In comparison with the other famous entropy-based and variance-based image thresholding algorithms, the new algorithm performs better in terms of correctness and robustness, as quantitatively demonstrated by four quality indices, ME, RAE, MHD, and PSNR. Furthermore, the whole process of the new algorithm has potential application in self-adaptive object recognition.

Entities:  

Keywords:  Otsu-based algorithm; gray-level distribution; image thresholding; nonextensive entropy; self-adaptive algorithm

Year:  2022        PMID: 35327830      PMCID: PMC8947459          DOI: 10.3390/e24030319

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

One of the most important tasks in image segmentation is to precisely extract objects from their backgrounds. Image thresholding has previously been widely adopted because of its simplicity and efficiency [1,2,3]. For different types of images, a large number of thresholding algorithms exist based on the characteristics of images. More specifically, the gray-level distribution, i.e., the histogram of the gray-level image, plays an important role in the image thresholding algorithms. It is obvious that different types of images will show different histogram profiles, which contain information relating to both the objects and their backgrounds. Therefore, it is desirable to identify characteristic functions that can suggest proper thresholds to separate the objects and backgrounds. The Otsu algorithm [4] is widely adopted to deal with images having a bimodal histogram distribution and can be easily extended to multi-level image segmentation [5,6,7,8,9]. The entropy-based algorithm [10,11,12,13,14] is another option for image segmentation since the gray-level histogram can be considered as a kind of probability distribution, and maximization of the corresponding entropies is a nature-inspired means of finding the optimal thresholds. In order to improve the robustness and anti-interference of the thresholding algorithms, two-dimensional histogram distributions [15,16,17] are frequently used to detect the edges and noise of the images, and thus achieve better segmentation results [18,19,20]. It is worth mentioning that, among these entropy-based algorithms, “Tsallis entropy-based thresholding” introduces the concept of nonextensivity into the image segmentation field [21,22]. The nonextensive entropy can be traced from the complex physical systems that have long-range interactions and/or long-duration memories [23,24]. There is a nonextensive parameter that measures the strength of the above mentioned non-local effects. Therefore, it is reasonable to adopt the nonextensive parameter to illustrate the global correlations among all pixels of an image. From the viewpoint of information theory, the nonextensive parameter of an image can be determined by the maximization of the redundancy of the gray-level distribution [25]. Since there are too many categories of images, a unique segmentation algorithm to deal with all of them effectively does not exist. Nevertheless, a stable algorithm that can correctly segment a wider variety of images is one of the important goals in computer vision research [26,27,28,29,30,31,32,33]. The Otsu and Otsu-based algorithms tend to separate the foreground and background equivalently, so they are not suitable for the extraction of tiny objects. Conversely, the entropy-based algorithms are too sensitive to the perturbation in images, and this instability restricts these algorithms from being applied in a more general scope. In this study, based on the explicit mathematical interpretation and numerical evaluation, it was found that the Otsu algorithm and the nonextensive entropy-based algorithm can be properly combined. An effective objective function is thus proposed to overcome both of the above-mentioned deficiencies. Moreover, the effective nonextensive parameter in the proposed algorithm is automatically determined by the information redundancy of an image [25]. Therefore, the proposes approach is a self-adaptive algorithm that can hopefully be applied to a more general scope of scenes. The remainder of this paper is organized as follow: in Section 2, the general properties of the Otsu algorithm are illustrated, and the entropy-based algorithms are briefly introduced; in Section 3, based on mathematical calculation and numerical evaluation, an effective objective function is proposed for self-adaptive image thresholding; the detailed results and analysis are illustrated in Section 4; and the conclusions are presented in Section 5.

2. Image Thresholding Algorithms

Assuming that the size of an image is and the range of its gray-level is considered as , the probability of the gray-level can be defined as: where is the total number of pixels in the image, represents the number of pixels for which the gray-level value is equal to . Therefore, the normalization of the probability distribution is explicitly expressed as Equation (1).

2.1. Otsu Algorithm

Now suppose that the threshold of an image is . The corresponding gray-level histogram can be divided into two classes, and , and the cumulative probability of the above two classes can be written as:The mean values of the gray-level of and are given by:Using the same idea, the mean gray-level value of the image is: Therefore, the variances of , , and the total histogram can be respectively written as: Based on Equation (5), the within-class variance and between-class variance are defined as [4]: and the following relation holds: It can be easily seen that for the arbitrary threshold value , the following relations always hold: The key point of the Otsu algorithm is to maximize the between-class variance by selecting a proper threshold value , i.e., In fact, using Equation (8), the between-class variance can be rewritten as: From Equation (10), it is shown that the between-class variance is dominated by two factors, and . Maximizing the factor means that the gray-level difference between and is tuned to the maximum by a proper threshold , which coincides with the principle of image segmentation. However, maximizing the factor requires finding another threshold to satisfy , which means that the number of pixels in the foreground is equal to that in the background. In general, for a given image, . However, the optimal threshold represents the trade-off between and . Therefore, the Otsu algorithm always has a tendency to equally separate the pixels of an image, demonstrating the deficiency when extracting tiny objects from the background.

2.2. Otsu–Kapur Algorithm

The Otsu algorithm is a classical global thresholding technique based on the clustering theorem. The idea of the entropy-based algorithm is quite different from Otsu’s, although both of them start from the image’s histogram. Shannon entropy is widely adopted in entropy-based image thresholding. It was first proposed by Pun [34] and improved by Kapur in 1985 [35]. By using the a priori entropy of the foreground and background, an objective function is obtained to indicate the optimal threshold under the Maximum Entropy Principle. Based on the gray-level histogram distribution of an image, Shannon entropy is given by: Assuming that the histogram is separated into two parts ( and ) by threshold ,. the corresponding entropies are: The objective function is given by the sum of and : and the optimal threshold of Kapur algorithm is determined by: In practice, the Kapur algorithm has better performance than the Otsu algorithm in extracting tiny targets from their background. However, this algorithm is quite sensitive to the perturbation of pixels. For instance, the value of Equation (13) varies drastically with the threshold , which means that the optimal threshold can be easily disturbed by the variation in gray-level distribution and lead to incorrect segmentation. This instability also restricts the application of the Kapur algorithm to a more general scope. Taking the characteristics of the Otsu algorithm into account, it is possible to increase the stability by combining the Kapur and Otsu algorithms, without losing the accuracy of extracting tiny objects. For a given image, the total gray-level variance is fixed. From Equation (7), we can see that maximizing the between-class variance is equivalent to minimizing the within-class variance . Therefore, Equations (5) and (14) yield the objective function: The optimal threshold is obtained by the following algorithm:

2.3. Two-Dimensional Entropic Algorithm

The above-mentioned thresholding algorithms are based on the one-dimensional(1D) gray-level histogram. In order to improve the accuracy and robustness, Ahmed [36] considered not only the pixel’s gray-level value, but also the spatial correlation of the pixels in an image. Therefore, the mean gray-level value of neighboring pixels is relevant and the one-dimensional (1D) histogram distribution is extended to the two-dimensional (2D) distribution. If a pixel’s gray-level is equal to and the average gray-level of its neighborhood is , the number of this kind of pixel among the image is . The 2D probability distribution can be written as: The total entropy of the 2D histogram is defined as: If the two thresholds are located at and , the 2D gray-level histogram is divided into four regions, as shown in Figure 1.
Figure 1

The distribution of the two-dimensional histogram with the threshold (s, t).

Assume that the pixels are mainly distributed at two regions, and in Figure 1. The cumulative probabilities of and are: The corresponding entropies can be written as: Based on the additivity of Shannon entropy, the total entropy is defined as: which is dependent on threshold . By the same idea of the 1D entropy-based algorithm, maximizing the objective function, i.e., Equation (21), can yield the optimal thresholds: In practice, the above 2D entropic algorithm is effective for images with uneven illumination, noise, missing edges, poor contrast, and other interference from the environment [37]. It is reasonable to consider more correlations between the pixel and its neighborhood, and the histogram distribution can be extended to higher dimensions. However, the increase in the number of dimensions will lead to an exponential increment in computation.

2.4. Tsallis Entropy Algorithm

As mentioned above, Shannon entropy is additive and shows the property of extensivity in image processing. The concept of entropy was first proposed in thermodynamics to describe the physical systems that have a huge number of microstates. Furthermore, the extensivity of entropy is based on the assumption that the microstates among the system are independent of each other. However, for some systems with long-range interactions, long-time memory and fractal-type structures, the extensivity may not hold anymore. Tsallis introduces a kind of non-extensive entropy [23] to describe such systems, expressed as: where is a real number that describes the nonextensivity of the system. In the limit, Tsallis entropy is reduced to Shannon entropy and the extensivity of the system is recovered. The nonextensive generalization of entropy also shed lights on the information theory. In image segmentation, Tsallis entropy shows potential superiority and flexibility in a more general scope of image class [21]. In the Tsallis entropy algorithm, the cumulative probability of foreground and background are: According to Equation (23), the entropy of each part can be defined as [21]: Suppose that and are subsystems of the full image, due to the nonextensivity; the total entropy of the image is expressed as: where the third term on the right-hand side of Equation (26) shows the pseudo-additivity of Tsallis entropy. Maximizing yields the optimal threshold , which is given by:Obviously, the optimal result of Equation (27) depends on the nonextensive parameter , which describes the strength of internal correlation of the image. In other words, for an arbitrary two pixels in the image, their gray-level values may have long-range correlations. More specifically, for an image containing several objects, the pixels of objects will exhibit similar gray-level values, even though they are not adjacent to each other. It is possible to measure this kind of long-range correlation by nonextensive entropy [18,21], and this idea inspired a new algorithm that is discussed below. Since the parameter is an additional index that can tune the optimal threshold, it is of great importance to determine the exact value of for a given image. Recently, Abdiel and coauthors introduce a methodology to evaluate the nonextensive parameter of an image [25]. Based on the information theory, the generalized redundancy of an image that presents nonextensive properties can be expressed as [25]: where is the possible maximum q-entropy of the image that can be achieved at , i.e., equipartition of the gray-level probability. Maximizing Equation (28) by a proper value of means that the gray-level histogram of the given image is renormalized to deviate from the equal probability case (containing zero information) as far as possible. Therefore, the information contained in the image histogram can be strengthened by a particular , which is highly image category dependent.

3. New Algorithm

As mentioned above, the nonextensive entropy algorithm is suitable for describing the long-range correlations within an image. However, like other entropy-based algorithms, it is still very sensitive to the perturbation of signals, so the scope of its application is limited. By comparison, the Otsu algorithm is stable but not accurate for small target extraction. Therefore, it is possible to combine the advantages of the two and develop a new algorithm with a more general scope of application. It is worth mentioning that the nonextensive parameter in Tsallis entropy is now determined by information redundancy and cannot be tuned arbitrarily. Based on Equations (5), (7) and (26), a new objective function can be written as:In order to retain the concavity of Tsallis entropy, should be satisfied [23]. Alternatively, is called superextensivity, which will increase the total entropy of the system in comparison with the extensive case () [38]. In practice, almost all categories of images exhibit the property of superextensivity [25]. Therefore, the proper range of the nonextensive parameter can be . From Equations (9) and (27), we can see that both of the two algorithms are aimed to maximize the objective functions. Taking Equation (7) into account, it can be easily seen that the aim of Equation (29) is to maximize the objective function, i.e., The optimal threshold is obtained from Equation (30) with the above-mentioned range of . For a synthetic image having a bimodal histogram distribution, as shown in Figure 2, the profile of each peak is the normalized q-Gaussian distribution function [39]. From Equations (9) and (27), we can see that both the Otsu algorithm and the Tsallis entropy algorithm indicate the valley gray-level between the two peaks as the optimal threshold, which exactly coincides with the result of Equation (30). For other natural pictures that have an arbitrary histogram distribution, there is no evidence that the result of Equation (9) coincides with that of Equation (27), whereas Equation (29) shows a trade-off between them and Equation (30) may yield a proper suggestion. For the histogram of Figure 2, it should be noted that the magnitude difference between and is very large. As shown in Figure 3, both of them are functions of threshold . However, the values of the Tsallis entropy algorithm are totally suppressed by those of the Otsu algorithm for any possible threshold . Therefore, it is unsuitable to combine and directly.
Figure 2

Normalized histogram distribution.

Figure 3

Objective functions of the Otsu and Tsallis algorithms.

In order to avoid the impact of the magnitude difference, the q-exponential function [40] can be adopted to revise the magnitude of . By definition, Tsallis entropy with a continuous probability distribution function can be expressed as: where represents the probability density of the normalized gray-level value . For a system presenting nonextensive q-entropy, the corresponding probability distribution can be written as the q-Gaussian function [39]: where is the variance of and is the partition function to keep the probability normalization condition, i.e., where is the Gamma function and will reduce to factorial if is an integer. Substituting into Equation (31) yields: where: is the integration constant for a given value of q. If and are two identical q-Gaussian distribution functions, according to the nonextensivity of Tsallis entropy, the total entropy can be written as:Substituting into Equation (36) yields:Therefore, the magnitude of is comparable with at the proper range of q, and the rationality of Equation (29) is shown. The main steps of the present algorithm can be seen in Figure 4:
Figure 4

The procedure of the new algorithm.

The above procedure can also be applied to the segmentation of RGB or other color images. The intense distribution of each color channel can be considered as a gray-level distribution. Therefore, the threshold value of each channel can be obtained directly. It should be mentioned that the intense distributions may differ for different color channels, so the above algorithm cannot yield a unified value, in general. By comparison, both the Otsu algorithm and entropy-based algorithm can be independently adopted for multi-level image thresholding. According to the idea of Equation (29), the advantages of these two kinds of typical thresholding algorithms can be combined by extending Equation (29) to the multi-level case.

4. Analysis of Experimental Results

In order to show the stability and feasibility of the proposed algorithm, we used four quality indices, namely, Misclassification Error (ME), Relative Foreground Area Error (RAE), Modified Hausdorff Distance (MHD), and Peak Signal-to-Noise Ratio (PSNR), to illustrate the performance of Equation (29) and make comparisons with the other algorithms mentioned in Section 2.

4.1. Misclassification Error (ME)

Misclassification error expresses the percentage of wrongly assigned image pixels that represent the object and background images. For the single threshold segmentation, ME can be simply expressed as [41]: where and represent the foreground and background of the ground-truth image, and are the foreground and background pixels in the segmented image, and is the cardinality of the set. The value of ME is between 0 and 1. The lower the value of ME, the better the segmentation result.

4.2. Relative Foreground Area Error (RAE)

RAE is a quality assessment parameter that calculates the area of difference between the segmented image and the ground-truth image, which is defined as [42]: where and are the area of the ground-truth image and the segmented image, respectively. Obviously, for an ideal segmentation in which coincides with , RAE is zero.

4.3. Modified Hausdorff Distance (MHD)

Hausdorff distance is used to determine the degree of similarity between two objects that are overlapped with each other. In order to maintain the symmetry form, the Modified Hausdorff Distance (MHD) is more frequently used and is defined as [43]: where and represent objects belonging to the ground-truth image and the segmented result , respectively, and is the Hausdorff distance. This parameter can objectively describe the distortion degree of the segmented image and the ground-truth image. If perfectly coincides with , then MHD is zero, by definition. Unlike ME and RAE, MHD is not normalized. For failed segmentation, the value of MHD will be much larger than 1.

4.4. Peak Signal-to-Noise Ratio (PSNR)

The Peak Signal-to-Noise Ratio is a measurement algorithm used in the image transmission. First, the concept of Mean Square Error (MSE) is required, which is a measure of the difference between two images. It is defined as [44]: where and are pixels of the ground-truth image and segmented image, respectively. It can be easily seen that if for arbitrary coordinates . Therefore, lower MSE represents better quality of image segmentation. Accordingly, PSNR is defined in terms of MSE:Equation (43) shows that, for ideal segmentation (), PSNR will tend to infinity.

4.5. Experimental Results

First, we applied the proposed algorithm to segment several well-known testing images. The results of the four other algorithms mentioned above are also listed, as shown in Figure 5, Figure 6 and Figure 7.
Figure 5

Lena.

Figure 6

Cameraman.

Figure 7

Baboon.

In Figure 5, compared to the results of the four other algorithms, i.e., Figure 5b–e, the result of the proposed algorithm has more details and edge contours. In Figure 6, we can see that both the 2D histogram algorithm and the Tsallis entropic algorithm failed to extract the objects from the image. However, the result of the proposed algorithm, i.e., Figure 6f is quite acceptable. We can see more detail information in it, in comparison with Figure 6b,c. In Figure 7e, it is shown that the Tsallis entropic algorithm over segments the original image and the detail of the baboon’s face is lost. However, the Otsu, Otsu–Kapur and Shannon 2D thresholds are also not appropriate. As shown in Figure 7b–d, the baboon’s eyes are blurred by too many black pixels. In contrast, the proposed algorithm has a moderate result, as shown in Figure 7f. In order to show the advantages of the proposed algorithm more convincingly, we choose 50 test images from VOC-2012, BSD300, and Ref. [45] to compare the performance of these five algorithms. These images have totally different gray-level histograms. Accordingly, their nonextensive parameters are also very different from each other, as shown in Table 1.
Table 1

The values of q of 50 test images.

Test q Test q Test q
10.6971180.4908350.5385
20.3891190.3971360.5217
30.6992200.6089370.5613
40.4067210.5240380.4409
50.6424220.4844390.5170
60.4933230.4003400.5139
70.4472240.5079410.5189
80.4706250.4895420.4960
90.4846260.4724430.5670
100.4990270.5000440.5107
110.5218280.5730450.5304
120.5159290.5602460.4680
130.4993300.4379470.4757
140.4572310.5881480.5823
150.6161320.5557490.5716
160.5976330.4936500.4884
170.5276340.5479
In order to further illustrate the segmentation results visually, we chose pictures 1–5 of Table 1 as examples, as shown in Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12.
Figure 8

Rice.

Figure 9

Infrared image.

Figure 10

Jet.

Figure 11

Plane.

Figure 12

Hawk.

In Figure 8, the ground-truth image shows that the number of pixels in the foreground is comparable to that of background. Therefore, both the Otsu algorithm and the proposed algorithm achieve acceptable results. However, the entropy-based algorithms cannot yield good results, as shown in Figure 8e,f. In Figure 9, the infrared object is tiny in comparison with the full image size. The Otsu-based algorithm failed to determine the correct results as expected, as shown in Figure 9c,d. By comparison, the infrared image may have a long-range correlation among the pixels, so the Shannon entropy-based algorithm also failed, as shown in Figure 9e. The results of Figure 9f,g are very close to the ground-truth image, which indicates that the value of the nonextensive parameter q can correctly evaluate the long-range correlation in an image. In addition, the value of q is automatically generated by maximizing Equation (27), so the new algorithm is self-adaptive. The results of Figure 10 and Figure 12 are quite similar to that of Figure 8, since there is a large amount of noise in the background and the entropy-based algorithms are very unstable to perturbation, in spite of the increasing dimension of the histogram. However, the new algorithm can still correctly segment the images, which shows the potential application in a more general scope, including tiny object recognition (Figure 9 and Figure 11), background noise suppression (Figure 10 and Figure 12), and detection of long-range correlation. From Figure 8e to Figure 12e, we can see that the 2-D Shannon algorithm, as a well-known entropic thresholding procedure, does not have stable outputs. However, the idea of extending the dimension of the histogram using the correlation of the neighboring pixels is still heuristic. It is of great interest to extend Equation (29) into two, or even higher, dimensions of the histogram because the development of optimization algorithms [15], refs. [19,20] can effectively reduce the computational cost. For each image from the testing set, it should be mentioned that the new algorithm and the Tsallis entropy algorithm share the same value of q, which is determined by maximizing the information redundancy. However, the Tsallis entropy algorithm is very unstable if the image is subject to noise interference, even with a proper value of q. In comparison, the new algorithm is always stable. By comparison, Otsu algorithm is robust but cannot effectively recognize tiny objects. The new algorithm extracts the advantages of both the Otsu and entropy-based algorithms in a proper manner, and this point can be further shown using the detailed quality indices. For 50 images in the testing set, Table 2, Table 3, Table 4 and Table 5 list the above-mentioned four quality indices of the results generated by the five different algorithms, respectively. Due to the variety in the testing set, the new algorithm cannot always ensure the best performance for all images, but its results are still acceptable. Furthermore, the statistical results of Table 2, Table 3, Table 4 and Table 5 clearly show the universality of the proposed algorithm for different kinds of images.
Table 2

The values of ME of 50 testing images segmented by 5 different algorithms.

ImagesOtsuOtsu-KapurShannon2DTsallisProposed
13.628 × 10−38.611 × 10−31.213 × 10−12.205 × 10−14.408 × 10−2
25.453 × 10−15.366 × 10−14.563 × 10−11.487 × 10−31.416 × 10−3
32.946 × 10−32.188 × 10−38.968 × 10−18.965 × 10−13.899 × 10−3
46.011 × 10−16.169 × 10−16.639 × 10−31.514 × 10−32.614 × 10−3
51.083 × 10−29.282 × 10−39.399 × 10−19.401 × 10−14.328 × 10−3
63.580 × 10−13.136 × 10−11.053 × 10−21.247 × 10−22.623 × 10−2
74.384 × 10−11.006 × 10−31.006 × 10−31.822 × 10−31.388 × 10−3
82.930 × 10−15.017 × 10−31.941 × 10−25.503 × 10−35.017 × 10−3
91.168 × 10−31.917 × 10−33.372 × 10−33.196 × 10−32.799 × 10−3
107.516 × 10−31.917 × 10−39.763 × 10−39.532 × 10−37.789 × 10−3
112.305 × 10−23.689 × 10−25.975 × 10−23.689 × 10−23.689 × 10−2
122.034 × 10−23.025 × 10−25.327 × 10−23.943 × 10−23.943 × 10−2
131.950 × 10−26.446 × 10−35.553 × 10−28.635 × 10−11.950 × 10−2
143.745 × 10−11.221 × 10−22.893 × 10−21.317 × 10−21.221 × 10−2
151.002 × 10−21.122 × 10−23.156 × 10−22.614 × 10−21.685 × 10−2
164.359 × 10−34.359 × 10−32.533 × 10−21.671 × 10−26.512 × 10−3
172.238 × 10−22.651 × 10−25.562 × 10−22.866 × 10−22.651 × 10−2
184.041 × 10−13.276 × 10−25.220 × 10−22.166 × 10−23.276 × 10−2
194.014 × 10−14.014 × 10−15.303 × 10−43.409 × 10−42.272 × 10−4
202.714 × 10−16.770 × 10−48.680 × 10−47.118 × 10−26.770 × 10−4
211.126 × 10−21.126 × 10−21.119 × 10−21.126 × 10−29.982 × 10−3
225.111 × 10−12.359 × 10−24.783 × 10−22.476 × 10−22.359 × 10−2
235.150 × 10−15.150 × 10−13.889 × 10−18.214 × 10−34.829 × 10−3
244.171 × 10−14.278 × 10−24.346 × 10−24.391 × 10−24.278 × 10−2
255.128 × 10−13.313 × 10−37.931 × 10−33.313 × 10−33.313 × 10−3
261.737 × 10−36.830 × 10−46.803 × 10−32.590 × 10−32.590 × 10−3
271.998 × 10−22.161 × 10−21.662 × 10−21.998 × 10−22.161 × 10−2
284.378 × 10−25.334 × 10−21.167 × 10−16.683 × 10−25.843 × 10−2
293.967 × 10−12.186 × 10−23.360 × 10−21.654 × 10−22.186 × 10−2
304.126 × 10−16.240 × 10−49.885 × 10−18.053 × 10−41.888 × 10−3
319.050 × 10−31.214 × 10−21.954 × 10−22.059 × 10−21.749 × 10−2
321.881 × 10−11.928 × 10−18.684 × 10−12.020 × 10−11.975 × 10−1
332.676 × 10−12.789 × 10−12.642 × 10−12.921 × 10−12.882 × 10−1
345.120 × 10−35.020 × 10−39.320 × 10−17.235 × 10−36.085 × 10−3
353.980 × 10−28.950 × 10−31.796 × 10−29.851 × 10−39.851 × 10−3
369.672 × 10−28.409 × 10−28.336 × 10−28.409 × 10−28.409 × 10−2
371.425 × 10−21.195 × 10−29.081 × 10−17.309 × 10−36.360 × 10−3
382.487 × 10−13.413 × 10−49.976 × 10−12.453 × 10−43.413 × 10−4
397.105 × 10−34.132 × 10−39.818 × 10−13.855 × 10−33.855 × 10−3
402.120 × 10−11.879 × 10−15.034 × 10−15.036 × 10−11.434 × 10−1
411.169 × 10−11.243 × 10−12.059 × 10−11.661 × 10−11.462 × 10−1
425.753 × 10−32.372 × 10−39.867 × 10−12.372 × 10−33.891 × 10−3
431.121 × 10−11.172 × 10−11.540 × 10−11.158 × 10−11.172 × 10−1
441.081 × 10−42.012 × 10−37.158 × 10−12.792 × 10−33.765 × 10−3
456.593 × 10−27.443 × 10−27.850 × 10−27.610 × 10−27.747 × 10−2
464.632 × 10−12.564 × 10−38.158 × 10−32.913 × 10−32.564 × 10−3
471.810 × 10−11.621 × 10−11.746 × 10−11.447 × 10−11.563 × 10−1
481.727 × 10−21.940 × 10−22.483 × 10−13.504 × 10−22.738 × 10−2
496.357 × 10−34.805 × 10−34.645 × 10−31.680 × 10−31.226 × 10−3
501.773 × 10−11.574 × 10−31.504 × 10−31.875 × 10−31.574 × 10−3
Table 3

The value of RAE of 50 testing images segmented by 5 different algorithms.

ImagesOtsuOtsu-KapurShannon2DTsallisProposed
11.031 × 10−22.776 × 10−24.025 × 10−17.316 × 10−11.462 × 10−1
29.964 × 10−19.963 × 10−19.957 × 10−13.640 × 10−13.428 × 10−1
31.988 × 10−39.748 × 10−49.758 × 10−19.753 × 10−13.780 × 10−3
42.897 × 10−12.653 × 10−19.690 × 10−19.694 × 10−12.158 × 10−1
56.043 × 10−16.202 × 10−16.624 × 10−31.380 × 10−32.511 × 10−3
63.668 × 10−14.187 × 10−11.045 × 10−21.459 × 10−23.068 × 10−2
79.851 × 10−12.551 × 10−22.051 × 10−21.415 × 10−18.173 × 10−2
89.622 × 10−13.009 × 10−16.280 × 10−13.210 × 10−13.009 × 10−1
93.699 × 10−28.886 × 10−21.464 × 10−11.398 × 10−11.246 × 10−1
101.569 × 10−28.886 × 10−21.352 × 10−29.160 × 10−33.171 × 10−3
112.542 × 10−24.068 × 10−26.589 × 10−24.068 × 10−24.068 × 10−2
122.290 × 10−23.416 × 10−26.015 × 10−24.452 × 10−24.452 × 10−2
138.167 × 10−22.120 × 10−14.338 × 10−19.225 × 10−12.120 × 10−1
149.742 × 10−15.519 × 10−17.448 × 10−15.705 × 10−15.519 × 10−1
151.710 × 10−23.738 × 10−21.961 × 10−11.609 × 10−19.227 × 10−2
161.156 × 10−21.156 × 10−26.296 × 10−24.434 × 10−21.727 × 10−2
171.011 × 10−11.175 × 10−12.184 × 10−11.258 × 10−11.175 × 10−1
184.710 × 10−12.131 × 10−24.327 × 10−27.205 × 10−32.131 × 10−2
199.948 × 10−19.948 × 10−12.029 × 10−11.129 × 10−16.779 × 10−2
209.675 × 10−12.416 × 10−23.314 × 10−23.136 × 10−22.416 × 10−2
214.315 × 10−14.315 × 10−14.300 × 10−14.315 × 10−14.021 × 10−1
229.552 × 10−14.787 × 10−16.642 × 10−14.938 × 10−14.787 × 10−1
239.762 × 10−19.762 × 10−19.687 × 10−13.959 × 10−12.781 × 10−1
248.857 × 10−14.430 × 10−14.469 × 10−14.494 × 10−14.430 × 10−1
259.551 × 10−11.143 × 10−12.476 × 10−11.208 × 10−11.143 × 10−1
261.790 × 10−31.692 × 10−56.459 × 10−32.008 × 10−32.008 × 10−3
271.010 × 10−21.261 × 10−26.825 × 10−31.010 × 10−21.261 × 10−2
283.615 × 10−21.110 × 10−13.536 × 10−11.782 × 10−11.395 × 10−1
294.042 × 10−12.227 × 10−23.423 × 10−21.685 × 10−22.227 × 10−2
304.145 × 10−14.988 × 10−49.944 × 10−12.199 × 10−49.230 × 10−4
311.719 × 10−21.781 × 10−29.285 × 10−28.773 × 10−26.442 × 10−2
321.937 × 10−11.988 × 10−19.007 × 10−12.088 × 10−12.038 × 10−1
332.787 × 10−12.904 × 10−12.752 × 10−13.042 × 10−13.001 × 10−1
343.808 × 10−32.893 × 10−39.977 × 10−13.324 × 10−31.986 × 10−3
351.867 × 10−14.910 × 10−26.479 × 10−22.067 × 10−22.067 × 10−2
362.157 × 10−11.399 × 10−19.805 × 10−21.399 × 10−11.399 × 10−1
371.514 × 10−21.264 × 10−29.946 × 10−12.725 × 10−35.820 × 10−3
382.491 × 10−13.419 × 10−49.995 × 10−12.457 × 10−43.419 × 10−4
397.191 × 10−34.195 × 10−39.986 × 10−12.747 × 10−32.747 × 10−3
401.141 × 10−29.791 × 10−39.979 × 10−19.987 × 10−14.589 × 10−3
411.679 × 10−11.796 × 10−13.030 × 10−12.432 × 10−12.133 × 10−1
425.796 × 10−31.452 × 10−39.996 × 10−11.452 × 10−34.138 × 10−4
431.128 × 10−11.190 × 10−11.599 × 10−11.174 × 10−11.190 × 10−1
441.108 × 10−42.063 × 10−37.342 × 10−12.864 × 10−33.862 × 10−3
455.195 × 10−26.429 × 10−27.306 × 10−26.650 × 10−26.842 × 10−2
469.915 × 10−13.928 × 10−16.730 × 10−14.237 × 10−13.928 × 10−1
472.117 × 10−11.897 × 10−12.042 × 10−11.693 × 10−11.829 × 10−1
484.399 × 10−22.698 × 10−25.420 × 10−11.175 × 10−18.164 × 10−2
496.647 × 10−34.985 × 10−34.885 × 10−31.077 × 10−33.077 × 10−4
508.989 × 10−16.968 × 10−24.994 × 10−21.045 × 10−26.968 × 10−2
Table 4

The value of MHD of 50 testing images segmented by 5 different algorithms.

ImagesOtsuOtsu-KapurShannon2DTsallisProposed
10.70291.28364.96636.77153.1038
211.715611.611010.62440.19760.1890
30.59220.502418.457718.45770.6871
414.334714.64050.86460.32460.4621
51.25171.162621.383621.36400.7552
69.230010.24771.07500.93961.4021
79.43870.12410.11570.23320.1869
87.93420.55241.49780.58380.5524
90.11660.22860.34170.35430.3202
101.04531.04821.29701.31171.0921
111.34591.88482.59761.88481.8848
121.13201.52872.30101.90391.9039
130.36940.85071.69098.82870.8507
148.63371.37372.27071.43571.3737
150.86800.95321.61641.56511.2187
160.27100.27101.02970.79900.3800
171.24001.36892.02931.46681.3689
187.97691.92612.40251.58221.9261
196.96896.96890.02190.03390.0294
204.16510.12660.14330.13490.1266
210.85450.85450.85200.85450.7773
229.82651.66022.39061.69951.6602
2311.281711.28179.73910.86200.5391
246.17401.89421.90341.91991.8942
259.19260.49141.01720.49140.4914
260.31060.19671.01520.54960.5496
271.82181.85721.67331.82181.8572
283.20173.45561.86303.49653.4838
294.64760.42810.59620.31060.4281
3010.33850.178021.69960.22720.3698
311.28041.64572.24502.47512.1817
325.63625.740218.57675.97575.8568
335.50166.12905.30756.91986.7114
340.71130.720021.24281.27861.0295
353.62821.29322.05791.63021.6302
365.91865.43925.66585.43925.4392
371.88221.691020.81211.24891.0532
3810.76410.096622.20880.07600.0966
390.63360.474721.80530.53340.5334
400.55250.250020.04220.73320.7332
414.29364.58296.54645.79885.2712
420.43110.348022.12190.34800.5231
437.37377.55728.72837.50837.5572
440.04880.447817.65880.57560.7164
454.36654.82124.90574.89454.9490
468.56510.29030.75850.32220.2903
478.38247.91257.88237.45987.7639
481.48071.80187.05692.45572.1798
490.66110.54770.85800.41130.3190
505.51220.12840.16000.22340.1284
Table 5

The value of PSNR of 50 testing images segmented by 5 different algorithms.

ImagesOtsuOtsu-KapurShannon2DTsallisProposed
124.402820.64949.16136.565813.5576
22.63352.70343.406728.276828.4887
325.306626.59940.47280.474525.0903
42.20992.097321.779028.197225.8268
519.651420.32350.26880.267923.6369
619.481819.481819.508619.481820.0075
73.580929.969929.969927.392328.5733
85.331322.995217.119822.593622.9952
929.324827.172424.720724.953925.5296
1021.239821.239820.104120.208121.0849
1116.372114.330912.236514.330914.3309
1216.916115.191812.734714.041714.0417
1321.906917.098712.55420.637117.0987
144.265019.130915.385718.804119.1309
1519.991219.497715.008415.825417.7334
1623.606223.606215.963117.768221.8623
1716.500115.765312.547115.427015.7653
183.934914.845612.823216.643014.8456
193.96383.963832.754834.673636.4345
205.663831.693630.614531.476431.6936
2121.906917.098712.55420.637117.0987
222.914916.272613.202316.061816.2726
2319.991219.497715.008415.825417.7334
2423.606223.606215.963117.768221.8623
2516.500115.765312.547115.427015.7653
2627.600231.655421.672825.866725.8667
2716.992516.651617.792216.992516.6516
2813.586412.72889.328711.750112.3330
294.014516.603114.735617.812916.6031
303.844232.04820.049930.940227.2400
3120.433219.154517.090516.861917.5713
327.25477.14740.61256.94477.0439
335.72385.54555.77925.34335.4028
3422.907322.99300.305721.405622.1574
3514.000620.481417.455420.065220.0652
3610.144410.752110.790010.752110.7521
3718.461219.22560.418321.361221.9652
386.042634.66820.010136.102434.6682
3921.484323.83830.079724.138824.1388
406.73627.26012.98062.97898.4326
419.32019.05536.86207.79548.3483
4222.400526.24820.058226.248224.0984
439.50239.30828.12339.36109.3082
4439.661426.96371.451625.539624.2415
4511.809011.282111.051011.185811.1083
463.342025.910620.883925.355525.9106
477.42307.90097.57968.39318.0586
4817.625917.12146.049714.553715.6255
4921.967323.182823.329827.746929.1127
507.511428.029728.225727.270028.0297
For 50 testing images, the segmented results and the corresponding ME of the different thresholding algorithms may be very different. However, the performances of these algorithms can be statistically evaluated. Figure 13 shows the average ME values of the 50 images yielded by the above-mentioned five algorithms, and Figure 14 represents the average values of RAE. As shown, the average ME value of the proposed algorithm is the lowest in comparison with those of the other algorithms. Furthermore, the variance in the ME value of the proposed algorithm is much less than those of the other algorithms. Therefore, both accuracy (lower ME value) and stability (lower variance) of the new algorithm are better than those of the algorithms, which means that this algorithm is suitable and more robust for a more general category of images. By the same analysis, we can see that the average value and variance in RAE of the new algorithm were both the lowest among all the results, which indicates that the new algorithm is better than the others in foreground area detection.
Figure 13

The average values of ME of different algorithms and the corresponding variance bars.

Figure 14

The average values of RAE of different algorithms and the corresponding variance bars.

As mentioned above, the values of MHD and PSNR are not normalized. For 50 segmented results of the testing set, the distributions of MHD and PSNR are not at , but at . Therefore, it is possible that their variances are larger than the averages. Figure 15 shows the comparison of MHD among the five algorithms. As we can see, the average MHD of the new algorithm again achieves the lowest value, and the corresponding variance is less than that of the others. This means that the new algorithm can maintain the shape of the objects in a more correct and stable manner. Figure 16 shows the comparison results of PSNR. Unlike the other three quality indices, the larger PSNR value means a better quality of information transmission. Therefore, we can see that the new algorithm still performs better than the others in this quality index (largest PSNR value), with a better robustness (lowest variance).
Figure 15

The average value of MHD of different algorithms and the corresponding variance bars.

Figure 16

The average value of PSNR of different algorithms and the corresponding variance bars.

5. Conclusions

In the task of computer vision, it is of great importance to explore algorithms that can correctly recognize the objects from different kinds of backgrounds in a stable way. The Otsu algorithm is based on the variance in the gray-level distribution of an image. It can yield stable thresholding results but has deficiencies in small target recognition. The entropy-based algorithms are suitable for small target extraction and can even detect the long-range correlation among pixels using a nonextensive parameter. However, the entropy-based objective functions can be easily disturbed by noise. In the present paper, based on the rigorous mathematical and numerical results, we combine the advantages of the Otsu algorithm and nonextensive entropy algorithm to develop a new algorithm that can effectively segment the objects from various kinds of background in a more stable manner. For 50 images chosen from different categories, the quality indices of ME, RAE, MHD, and PSNR were adopted to evaluate the segmentation results. In comparison with the other famous thresholding algorithms, the statistical results show that the proposed algorithm has better performance than the others in each of the four quality indices. In addition, there is no artificial intervention during the whole process. Therefore, the proposed algorithm is an approach to automatic image thresholding that has potential application in self-adaptive object recognition.
  9 in total

1.  An Adaptive and Robust Edge Detection Method based on Edge Proportion Statistics.

Authors:  Yang Liu; Zongwu Xie; Hong Liu
Journal:  IEEE Trans Image Process       Date:  2020-03-18       Impact factor: 10.856

2.  A Robust Parameter-free Thresholding Method for Image Segmentation.

Authors:  Xinhua Cao; Taihao Li; Hongli Li; Shunren Xia; Fuji Ren; Ye Sun; Xiaoyin Xu
Journal:  IEEE Access       Date:  2018-12-20       Impact factor: 3.367

3.  Image segmentation based on gray level and local relative entropy two dimensional histogram.

Authors:  Wei Yang; Lulu Cai; Fei Wu
Journal:  PLoS One       Date:  2020-03-03       Impact factor: 3.240

4.  2D Tsallis Entropy for Image Segmentation Based on Modified Chaotic Bat Algorithm.

Authors:  Zhiwei Ye; Juan Yang; Mingwei Wang; Xinlu Zong; Lingyu Yan; Wei Liu
Journal:  Entropy (Basel)       Date:  2018-03-30       Impact factor: 2.524

5.  A Chaotic Electromagnetic Field Optimization Algorithm Based on Fuzzy Entropy for Multilevel Thresholding Color Image Segmentation.

Authors:  Suhang Song; Heming Jia; Jun Ma
Journal:  Entropy (Basel)       Date:  2019-04-15       Impact factor: 2.524

6.  A Hybrid Preaching Optimization Algorithm Based on Kapur Entropy for Multilevel Thresholding Color Image Segmentation.

Authors:  Bowen Wu; Liangkuan Zhu; Jun Cao; Jingyu Wang
Journal:  Entropy (Basel)       Date:  2021-11-29       Impact factor: 2.524

7.  A Grouping Differential Evolution Algorithm Boosted by Attraction and Repulsion Strategies for Masi Entropy-Based Multi-Level Image Segmentation.

Authors:  Seyed Jalaleddin Mousavirad; Davood Zabihzadeh; Diego Oliva; Marco Perez-Cisneros; Gerald Schaefer
Journal:  Entropy (Basel)       Date:  2021-12-21       Impact factor: 2.524

8.  Enhanced Slime Mould Algorithm for Multilevel Thresholding Image Segmentation Using Entropy Measures.

Authors:  Shanying Lin; Heming Jia; Laith Abualigah; Maryam Altalhi
Journal:  Entropy (Basel)       Date:  2021-12-20       Impact factor: 2.524

  9 in total
  1 in total

1.  Performance and Robustness of Regional Image Segmentation Driven by Selected Evolutionary and Genetic Algorithms: Study on MR Articular Cartilage Images.

Authors:  Jan Kubicek; Alice Varysova; Martin Cerny; Kristyna Hancarova; David Oczka; Martin Augustynek; Marek Penhaker; Ondrej Prokop; Radomir Scurek
Journal:  Sensors (Basel)       Date:  2022-08-23       Impact factor: 3.847

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.