Literature DB >> 33784310

Estimation of different types of entropies for the Kumaraswamy distribution.

Abdulhakim A Al-Babtain1, Ibrahim Elbatal2, Christophe Chesneau3, Mohammed Elgarhy4.   

Abstract

The estimation of the entropy of a random system or process is of interest in many scientific applications. The aim of this article is the analysis of the entropy of the famous Kumaraswamy distribution, an aspect which has not been the subject of particular attention previously as surprising as it may seem. With this in mind, six different entropy measures are considered and expressed analytically via the beta function. A numerical study is performed to discuss the behavior of these measures. Subsequently, we investigate their estimation through a semi-parametric approach combining the obtained expressions and the maximum likelihood estimation approach. Maximum likelihood estimates for the considered entropy measures are thus derived. The convergence properties of these estimates are proved through a simulated data, showing their numerical efficiency. Concrete applications to two real data sets are provided.

Entities:  

Year:  2021        PMID: 33784310      PMCID: PMC8009427          DOI: 10.1371/journal.pone.0249027

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


1 Introduction

Information theory provides natural mathematical tools for measuring the uncertainty of random variables and the information shared by them. In this regard, entropy and mutual information are two fundamental concepts. More precisely, the probability distribution of a random variable is associated with some sort of uncertainty, and entropy is used to quantify it. The concept of entropy was formerly proposed by [1]. Since that publication, many areas of study such as statistics, neurobiology, cryptography, bioinformatics, quantum computer science and linguistics, have developed various entropy-based measures. Modern and exhaustive reviews on the ‘entropy universe’ can be found in [2-6]. In applied probability and statistics, many authors have conducted their studies for diverse and important distributions based on entropy. The essential references in this regard are briefly presented below. Reference [7] used the concept of entropy to communicate on the probability distribution of electric charge between atoms observed in a certain condition. Reference [8] derived the entropy for the Feller-Pareto family and presented the entropy ordering property for some related sample minimum and maximum. Reference [9] estimated the entropy of the Weibull distribution by considering different loss functions based on a generalized progressively hybrid censoring scheme. Reference [10] discussed the entropy for the generalized half-logistic distribution based on the type II censored samples. References [11] and [12] proposed estimates for the entropy of absolutely continuous random variables. Reference [13] presented an indirect method using a decomposition to simplify the entropy’s calculation under the progressive type II censoring. Reference [14] derived a nonparametric kernel estimator for the general Shannon entropy. Reference [15] estimated the entropy for several exponential distributions and extended the results to other circumstances. Reference [16] estimated the Shannon entropy of the Rayleigh model under doubly generalized type-II hybrid censoring, and evaluated its performance by two criteria. Reference [17] derived a nonparametric wavelet estimator for the general Shannon entropy. Reference [18] provided an exact expression for entropy information contained in both types of progressively hybrid censored data and applied it in the setting of the exponential distribution. Reference [19] investigated entropy measures for weighted and truncated weighted exponential distributions. Reference [20] presented the estimation of entropy for inverse Weibull distribution under multiple censored data. Reference [21] introduced estimation of entropy for inverse Lomax distribution under the multiple censored scheme. Reference [22] examined Bayesian and non-Bayesian methods to estimate the dynamic cumulative residual Rényi entropy for the Lomax distribution. Surprisingly, to our knowledge, the entropy of the famous Kumaraswamy distribution has not been studied in depth. In this article, we fill this gap both probabilistically and statistically. The specificities and interests of the Kumaraswamy distribution are described below. First, it was introduced by [23], and was motivated as an alternative to the beta distribution which are (i) mathematically simpler, without special function in particular, and (ii) more suited to the modeling of various hydrological phenomena observed at low frequency (daily rainfall, daily flow of rivers, etc.). Mathematically, the probability density function (pdf) of the Kumaraswamy distribution is specified by with f(x;a, b) = 0 otherwise, where a, b > 0. This pdf is unimodal if a, b > 1, uniantimodal if a, b < 1, increasing if a > 1, b ≤ 1, decreasing if a ≤ 1, b > 1 or constant if a = b = 1, in the same way as the beta distribution. The corresponding cumulative distribution and quantile functions are quite simple; they are defined without special function contrary to those of the beta distribution. Special cases of the Kumaraswamy distribution correspond to the distribution of minimum or maximum of uniform samples. We may refer the reader to [24] for all the known features of this distribution. Also, the kumaraswamy distribution has generated many flexible distributions with various domains and number of parameters through the generalized Kumaraswamy class elaborated by [25]. In a sense, this study complements the work of [24] by investigating the overall concept of entropy of the Kumaraswamy distribution, which has never been studied before. More precisely, we consider six well-referenced entropy measures. We derive their analytical expressions by using the well-known beta function. We compare them numerically by considering different parameter values. Then, we propose an efficient strategy based on the maximum likelihood approach to estimate these entropy measures. A simulation study is done to see how effective our strategy is. Graphical and numerical comparisons are performed. We end the study by two illustrative examples on real data sets, showing how the methodology can be applied in a concrete statistical setting. The following sections make up the document. Section 2 presents a result on a special integral, and shows how it is related to important entropy measures of the Kumaraswamy distribution. Numerical values of these entropy measures with different values of the parameters are also given. Section 3 studies the estimation of these entropy measures. Then, using generated values from the Kumaraswamy distribution, graphical and numerical comparisons are discussed. The entropy of the random characteristics behind two real data sets is investigated. Finally, conclusions are presented in section 4.

2 Entropy of the Kumaraswamy distribution

2.1 An integral result

The following result shows that a certain integral involving the pdf of the Kumaraswamy distribution can be expressed in terms of the classical beta function. The connection between this integral and the considered entropy measures will be developed later. Proposition 1 Let δ > 0, f(x;a, b) be specified by Eq (1) and Then, I(a, b) exists if and only if min(a, b)>max(1 − 1/δ, 0), and it is expressed as where B(u, v) denotes the classical beta function, that is for u, v > 0. Proof. Owing to Eq (1), we have When x tends to 0, we have x(1 − x) ∼ x, which is integrable in the neighborhood of 0 if and only if δ(1 − a)<1 by the Riemann integral criteria. Similarly, when x tends to 1, we have which is integrable in the neighborhood of 1 if and only if δ(1 − b)<1 by the Riemann integral criteria. In summary, I(a, b) exists if and only if δmax(1 − a, 1 − b)<1, which is equivalent to min(a, b)>1 − 1/δ. Now, under this assumption, by applying the change of variables y = x, that is with , we obtain This ends the proof of Proposition 1. In fact, the beta function is implemented in most of the mathematical software (see the function beta of the package stat of R, the Beta function of Mathematica, etc.). Therefore, thanks to Proposition 1, the computation of I(a, b) can be done quite efficiently with little effort. Also, the existing results on the beta functions allow a mathematical control of this integral. Some related results are presented below. Through the use of the standard Euler gamma function given as , one can write Also, assuming that δ(1 − 1/a) + 1/a and δ(b − 1) + 1 are positive integers, the following formula holds: By virtue of the main result in [26], if δ(a − 1) ≥ a − 1 and b ≥ 1, then we have with the best possible constants α = 0 and β = 0.08731…. Therefore, for not too large value of δ, the following numerical approximation seems acceptable: In our study, the interest of Proposition 1 is that I(a, b) is the main ingredient in the definitions of various entropy measures of the Kumaraswamy distribution, as developed in the next part.

2.2 Various entropy measures

The entropy of the Kumaraswamy distribution can be measured in different manners. The most useful entropy measures of the literature are recalled in Table 1 for a general distribution with pdf denoted by f(x;φ), φ representing a possible vector of parameters. Also, we suppose that δ > 0 and δ ≠ 1 as basic assumptions in this general case.
Table 1

Important entropy measures of a distribution with pdf f(x;φ) at δ.

Name of the entropyReferenceNotationExpression
Rényi[28]Rδ(φ) 11δlog[+f(x;φ)δdx]
Havrda and Charvat[29]HCδ(φ) 121δ1[+f(x;φ)δdx1]
Arimoto[30]Aδ(φ) δ1δ{[+f(x;φ)δdx]1δ1}
Tsallis[31]Tδ(φ) 1δ1[1+f(x;φ)δdx]
Awad and Alawneh 1[27]AA1δ(φ) 1δ1log{[supxf(x;φ)]1δ+f(x;φ)δdx}
Awad and Alawneh 2[27]AA2δ(φ) 121δ1[{[supxf(x;φ)]1δ+f(x;φ)δdx}1]
For the two entropy measures proposed by [27], it is supposed that is finite and well identified. From Table 1, we see that the integral is central to determine the considered entropy measures. Now, we present the corresponding entropy measures of the Kumaraswamy distribution. Based on Proposition 1, it is supposed that a, b and δ satisfy min(a, b)>max(1 − 1/δ, 0).

Rényi entropy

Based on Table 1, Eq (1) and Proposition 1, the Rényi entropy of the Kumaraswamy distribution can be expressed as

Havrda and Charvát entropy

From Table 1, Eq (1) and Proposition 1, the Havrda and Charvát entropy of the Kumaraswamy distribution can be expressed as

Arimoto entropy

Again, from Table 1, Eq (1) and Proposition 1, the Arimoto entropy of the Kumaraswamy distribution is specified by

Tsallis entropy

Based on Table 1, Eq (1) and Proposition 1, the Tsallis entropy of the Kumaraswamy distribution can be expressed as

Awad and Alawneh 1 entropy

From Table 1, Eq (1) and Proposition 1, the Awad and Alawneh 1 entropy of the Kumaraswamy distribution is given as Before going further, we need to determine sup0< f(x;a, b). The following lemma provides the necessary in this regard. Lemma 2 Let f(x;a, b) be given as Eq (1). Then, sup0< f(x;a, b) is finite if and only if a ≥ 1 and b ≥ 1 with ab ≠ 1, and in this case, we have Proof. We have Therefore, f′(x*;a, b) = 0 implies that Since f′(x;a, b)>0 for x < x* and f′(x;a, b)<0 for x > x*, x* is a maximum point for f(x;a, b). Hence, Note that, for a = 1, with the convention 00 = 1, we have f(x*;a, b) = b(b − 1)(b − 1)1− = b and for b = 1, we have . This ends the proof of Lemma 2. Based on Lemma 2, if a > 1 and b > 1, Eq (2) becomes

Awad and Alawneh 2 entropy

From Table 1, Eq (1), Proposition 1 and Lemma 2, the Awad and Alawneh 2 entropy of the Kumaraswamy distribution is given as Theoretically, it is complicated to study the behavior of these entropy measures. For this reason, a numerical study is proposed in the next section.

2.3 Numerical values

We now investigate the numerical values for the six entropy measures presented in Subsection 2.2 under the following configuration of the parameters: Configuration 1: a = 2, b ∈ Υ with Υ = {1.5, 2.0, 2.5, 3.0, 3.5, 4.0, 4.5, 5.0, 5.5, 6.0} and δ = 0.5, Configuration 2: a = 2, b ∈ Υ and δ = 1.5, Configuration 3: a = 2, b ∈ Υ and δ = 2.5, Configuration 4: a ∈ Υ, b = 2 and δ = 0.5, Configuration 5: a ∈ Υ, b = 2 and δ = 1.5, and Configuration 6: a ∈ Υ, b = 2 and δ = 2.5. The findings of all the six entropy measures are presented for these configurations in Tables 2–7, respectively.
Table 2

Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 0.5.

bRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.034-0.092-0.075-0.076-0.1420.430
2.0-0.037-0.100-0.081-0.083-0.1510.457
2.5-0.047-0.127-0.103-0.106-0.1630.500
3.0-0.060-0.161-0.129-0.134-0.1750.538
3.5-0.074-0.197-0.157-0.163-0.1840.570
4.0-0.088-0.233-0.183-0.193-0.1920.596
4.5-0.102-0.267-0.209-0.221-0.1980.618
5.0-0.115-0.299-0.232-0.248-0.2030.637
5.5-0.128-0.330-0.255-0.273-0.2080.653
6.0-0.140-0.359-0.275-0.297-0.2120.667
Table 7

Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 2.5.

aRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.070-0.426-0.170-0.184-0.0810.378
2.0-0.095-0.600-0.233-0.259-0.0930.423
2.5-0.137-0.932-0.346-0.402-0.1030.462
3.0-0.181-1.340-0.472-0.578-0.1100.490
3.5-0.223-1.799-0.602-0.775-0.1160.511
4.0-0.264-2.297-0.732-0.990-0.1210.527
4.5-0.301-2.830-0.860-1.220-0.1240.539
5.0-0.336-3.392-0.985-1.462-0.1270.549
5.5-0.369-3.983-1.108-1.717-0.1290.557
6.0-0.399-4.600-1.227-1.982-0.1310.563
In view of Tables 2–7, the following comments can be formulated. First, we recall that Tables 2–4 indicate the values of the entropy measures of the Kumaraswamy distribution for a fixed value of a and different values for b and δ. In this context,
Table 4

Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 2.5.

bRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.087-0.546-0.214-0.235-0.0890.408
2.0-0.095-0.600-0.233-0.259-0.0930.423
2.5-0.114-0.743-0.283-0.320-0.0970.440
3.0-0.134-0.913-0.340-0.393-0.1010.454
3.5-0.155-1.094-0.398-0.471-0.1030.464
4.0-0.174-1.278-0.454-0.551-0.1050.472
4.5-0.193-1.463-0.509-0.631-0.1070.478
5.0-0.210-1.647-0.561-0.710-0.1080.483
5.5-0.226-1.830-0.611-0.789-0.1090.487
6.0-0.241-2.011-0.659-0.867-0.1100.490
the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when b is increasing while the Awad and Alawneh 2 entropy is increasing when b is increasing. the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing. Tables 5–7 show the values of the entropy of the Kumaraswamy distribution for a fixed value of b and different values for a and δ. In this setting,
Table 5

Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 0.5.

aRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.026-0.072-0.059-0.060-0.1250.374
2.0-0.037-0.100-0.081-0.083-0.1510.457
2.5-0.059-0.160-0.128-0.132-0.1800.555
3.0-0.086-0.229-0.180-0.189-0.2050.641
3.5-0.115-0.298-0.232-0.247-0.2250.713
4.0-0.143-0.366-0.280-0.303-0.2410.773
4.5-0.170-0.429-0.324-0.356-0.2550.824
5.0-0.196-0.489-0.364-0.405-0.2670.867
5.5-0.222-0.544-0.400-0.451-0.2760.904
6.0-0.246-0.595-0.432-0.493-0.2850.936
the Rényi, Havrda and Charvat, Arimoto, Tsallis and Awad and Alawneh1 entropy measures are decreasing when a is increasing while the Awad and Alawneh 2 entropy is increasing when a is increasing. the Rényi, Havrda and Charvat, Arimoto and Tsallis entropy measures are decreasing when δ is increasing while the Awad and Alawneh1 entropy is increasing when δ is increasing, but the Awad and Alawneh2 entropy is decreasing and increasing when δ is increasing.

3 Maximum likelihood estimation

The inference on the six considered entropy measures of the Kumaraswamy distribution is now investigated via the maximum likelihood technique. This technique is well-known and has proved itself in various modern studies such as those in [32-34].

3.1 Estimation of the entropy measures

The estimation of the parameters of the Kumaraswamy model through the maximum likelihood technique is well-known and the details can be found in [24]. The minimal theory is recalled below. Based on n values x1, …, x supposed to be observed from a random variable X with the Kumaraswamy distribution with parameters a and b, the maximum likelihood estimates (MLEs) of a and b, say and , are defined by where ℓ(a, b) denotes the log-likelihood function specified by These MLEs are also the solutions of the two following equations according to a and b: That is, and satisfy the following simple relation: Then, the properties of these MLEs follow from the usual maximum likelihood theory. In particular, thanks to the functional invariance of the MLEs, one can deduce easily the MLEs of the entropy measures. More concretely, based on the six entropy measures described in Subsection 2.2, is the MLE of R(a, b), is the MLE of HC(a, b), is the MLE of A(a, b), is the MLE of T(a, b), is the MLE of AA1(a, b), and is the MLE of AA2(a, b).

3.2 Simulation

We now investigate the numerical behavior of the MLEs of the entropy measures via the use of simulated values. That is, we consider N = 5000 samples of values from a random variable X with the Kumaraswamy distribution of parameters a and b with different samples sizes; n = 100, 200, 300 and 1000 are considered. The following configurations on the parameters are considered: Configuration1: a = 3, b = 3 and δ ∈ Ξ with Ξ = {0.5, 1.5, 2.5}, and Configuration 2: a = 3, b = 5 and δ ∈ Ξ. In each configuration, for each sample, the MLEs and are determined. Then, based on the N samples of fixed size, we determine the average of the N MLEs and use it to define the entropy estimates. The corresponding mean squared error (MSE) and mean deviation (MD) defined by the following generic formulas: MSE = sum(exact value—estimate)2 / N and MD = sum abs(exact value—estimate) / N, respectively, are also calculated. These assessment criteria are often used quite effectively to make a full comparison of models. In this regard, we can refer the reader to the useful works of [35-37]. The results on the Rényi entropy under Configurations 1 and 2 are given in Tables 8 and 9, respectively, results on the Havrda and Charvat entropy under Configurations 1 and 2 are indicated in Tables 10 and 11, respectively, results on the Arimoto entropy under Configurations 1 and 2 are presented in Tables 12 and 13, respectively, results on the Tsallis entropy under Configurations 1 and 2 are given in Tables 14 and 15, respectively, results on the Awad and Alawneh 1 entropy under Configurations 1 and 2 are given in Tables 16 and 17, respectively, and results on the Awad and Alawneh 2 entropy under Configurations 1 and 2 are indicated in Tables 18 and 19.
Table 8

Numerical values of the simulation related to the Rényi entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
Rδ(a, b)EstimateMSEMDRδ(a, b)EstimateMSEMDRδ(a, b)EstimateMSEMD
100-0.2107-0.22150.00200.0344-0.3674-0.38120.00390.0487-0.4379-0.45230.00460.0529
200-0.21320.00090.0233-0.37020.00180.0332-0.44060.00210.0362
300-0.21170.00050.0180-0.36830.00100.0257-0.43870.00120.0280
1000-0.21130.00020.0103-0.36800.00030.0147-0.43850.00040.0160
Table 9

Numerical values of the simulation related to the Rényi entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
Rδ(a, b)EstimateMSEMDRδ(a, b)EstimateMSEMDRδ(a, b)EstimateMSEMD
100-0.2753-0.28000.00210.0371-0.4504-0.45530.00370.0488-0.5258-0.53070.00420.0522
200-0.27810.00110.0260-0.45350.00190.0343-0.52890.00210.0366
300-0.28020.00070.0218-0.45640.00130.0287-0.53210.00150.0306
1000-0.27750.00020.0111-0.45320.00030.0146-0.52870.00040.0157
Table 10

Numerical values of the simulation related to the Havrda and Charvat entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
HCδ(a, b)EstimateMSEMDHCδ(a, b)EstimateMSEMDHCδ(a, b)EstimateMSEMD
100-0.2414-0.24940.00210.0357-0.6885-0.71020.01540.0967-1.4364-1.49600.09130.2330
200-0.24520.00130.0280-0.69870.00940.0759-1.46560.05500.1819
300-0.24380.00070.0214-0.69490.00540.0579-1.45440.03110.1382
1000-0.24330.00020.0113-0.69370.00150.0307-1.44960.00850.0731
Table 11

Numerical values of the simulation related to the Havrda and Charvat entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
HCδ(a, b)EstimateMSEMDHCδ(a, b)EstimateMSEMDHCδ(a, b)EstimateMSEMD
100-0.3104-0.31880.00270.0400-0.8623-0.88640.01970.1085-1.8572-1.92990.13730.2816
200-0.31340.00120.0282-0.87080.00870.0761-1.88360.05780.1953
300-0.31510.00080.0227-0.87530.00610.0613-1.89390.04050.1576
1000-0.31250.00030.0127-0.86810.00190.0343-1.87300.01230.0877
Table 12

Numerical values of the simulation related to the Arimoto entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
Aδ(a, b)EstimateMSEMDAδ(a, b)EstimateMSEMDAδ(a, b)EstimateMSEMD
100-0.1900-0.19800.00120.0275-0.3908-0.40730.00510.0553-0.5008-0.52160.00800.0695
200-0.19430.00060.0195-0.39970.00250.0390-0.51200.00390.0489
300-0.19250.00040.0160-0.39590.00160.0321-0.50710.00260.0403
1000-0.19120.00010.0091-0.39330.00050.0181-0.50390.00080.0227
Table 13

Numerical values of the simulation related to the Arimoto entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
Aδ(a, b)EstimateMSEMDAδ(a, b)EstimateMSEMDAδ(a, b)EstimateMSEMD
100-0.2406-0.24760.00140.0297-0.4859-0.50130.00590.0607-0.6182-0.63800.00940.0767
200-0.24260.00060.0197-0.49040.00250.0401-0.62400.00400.0507
300-0.24200.00040.0161-0.48910.00170.0327-0.62220.00270.0413
1000-0.24080.00010.0095-0.48640.00060.0193-0.61880.00090.0244
Table 14

Numerical values of the simulation related to the Tsallis entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
Tδ(a, b)EstimateMSEMDTδ(a, b)EstimateMSEMDTδ(a, b)EstimateMSEMD
100-0.2000-0.20830.00150.0304-0.4033-0.41930.00560.0584-0.6190-0.65080.01810.1038
200-0.20410.00070.0210-0.41110.00260.0402-0.63440.00810.0710
300-0.20430.00050.0180-0.41160.00190.0345-0.63480.00600.0608
1000-0.20040.00010.0093-0.40400.00050.0178-0.62070.00150.0311
Table 15

Numerical values of the simulation related to the Tsallis entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
Tδ(a, b)EstimateMSEMDTδ(a, b)EstimateMSEMDTδ(a, b)EstimateMSEMD
100-0.2572-0.26330.00150.0308-0.5051-0.51730.00560.0589-0.8004-0.82710.02040.1120
200-0.26040.00090.0232-0.51170.00320.0443-0.81500.01150.0838
300-0.25900.00050.0183-0.50880.00190.0348-0.80860.00670.0656
1000-0.25910.00020.0100-0.50870.00060.0189-0.80750.00200.0356
Table 16

Numerical values of the simulation related to the Awad and Alawneh 1 entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
AA1δ(a, b)EstimateMSEMDAA1δ(a, b)EstimateMSEMDAA1δ(a, b)EstimateMSEMD
100-0.48694-0.489870.001020.02573-0.33028-0.331230.000210.01169-0.25981-0.260350.000090.00751
200-0.488300.000570.01950-0.330690.000120.00887-0.260040.000050.00569
300-0.487930.000340.01448-0.330600.000070.00658-0.260000.000030.00422
1000-0.487030.000110.00825-0.330280.000020.00375-0.259800.000010.00241
Table 17

Numerical values of the simulation related to the Awad and Alawneh 1 entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
AA1δ(a, b)EstimateMSEMDAA1δ(a, b)EstimateMSEMDAA1δ(a, b)EstimateMSEMD
100-0.51891-0.521480.000740.02153-0.34379-0.344650.000140.00942-0.26837-0.268860.000060.00598
200-0.520910.000380.01553-0.344560.000070.00677-0.268830.000030.00430
300-0.520430.000230.01221-0.344380.000040.00534-0.268730.000020.00339
1000-0.519510.000070.00660-0.344030.000010.00288-0.268520.000010.00183
Table 18

Numerical values of the simulation related to the Awad and Alawneh 2 entropy for Configuration 1 (a = 3, b = 3).

nδ = 0.5δ = 1.5δ = 2.5
AA2δ(a, b)EstimateMSEMDAA2δ(a, b)EstimateMSEMDAA2δ(a, b)EstimateMSEMD
1000.665530.669620.002260.038160.519720.520740.000410.016280.499270.499830.000200.01135
2000.665020.001270.027920.519090.000230.011970.498740.000110.00835
3000.666680.000840.022840.519940.000150.009770.499370.000080.00681
10000.666210.000270.013010.519930.000050.005560.499400.000020.00387
Table 19

Numerical values of the simulation related to the Awad and Alawneh 2 entropy for Configuration 2 (a = 3, b = 5).

nδ = 0.5δ = 1.5δ = 2.5
AA2δ(a, b)EstimateMSEMDAA2δ(a, b)EstimateMSEMDAA2δ(a, b)EstimateMSEMD
1000.715150.720430.001660.032560.539220.540860.000260.013030.512630.513660.000120.00892
2000.715950.000800.023000.539300.000130.009250.512640.000060.00634
3000.717110.000530.018250.539850.000080.007320.513030.000040.00502
10000.716630.000170.010590.539760.000030.004240.512990.000010.00290
Based on Tables 8–19, in all the situations, we see that the MLEs of the entropy measures are close to the target values and, as anticipated, the MSEs and MDs decrease and approach 0 as n increases. This proves the accuracy of the proposed estimation methods in the context of the Kumaraswamy distribution. Also, one can notice that the MSEs and MDs increase as δ increases. For a visual approach, the behavior of the MSEs and MDs are illustrated in Figs 1–12, for the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 1 and Awad and Alawneh 2 entropy measures following the settings of Tables 8–19, respectively.
Fig 1

Plots of the (a) MSEs and (b) MDs for the Rényi entropy in the setting of Table 8.

Fig 12

Plots of the (a) MSEs and (b) MDs for the Awad and Alawneh 2 entropy in the setting of Table 19.

Figs 1–12 support the claims formulated about the results of Tables 8–19.

3.3 Illustrative examples

In this Section, two real life data sets are used to illustrate the proposed methodology. The considered data sets are described below.

The first data set

The data set consists of 48 rock samples from an oil reservoir. It corresponds to twelve oil tank cores that were sampled by four cross sections. Each core was measured for permeability and each cross section has the following variables: total pore area, total pore perimeter, and shape. We analyze the perimeter of the shape by a squared variable (area). It has been analyzed by [38], among others. Explicitely, the data set is: {0.0903296, 0.2036540, 0.2043140, 0.2808870, 0.1976530, 0.3286410, 0.1486220, 0.1623940, 0.2627270, 0.1794550, 0.3266350, 0.2300810, 0.1833120, 0.1509440, 0.2000710, 0.1918020, 0.1541920, 0.4641250, 0.1170630, 0.1481410, 0.1448100, 0.1330830, 0.2760160, 0.4204770, 0.1224170, 0.2285950, 0.1138520, 0.2252140, 0.1769690, 0.2007440, 0.1670450, 0.2316230, 0.2910290, 0.3412730, 0.4387120, 0.2626510, 0.1896510, 0.1725670, 0.2400770, 0.3116460, 0.1635860, 0.1824530, 0.1641270, 0.1534810, 0.1618650, 0.2760160, 0.2538320, 0.2004470}.

The second data set

This data set contains 20 observations of flood data. It was analyzed by [39]. The data set is listed as follows: {0.265, 0.392, 0.297, 0.3235, 0.402, 0.269, 0.315, 0.654, 0.338, 0.379, 0.418, 0.423, 0.379, 0.412, 0.416, 0.449, 0.484, 0.494, 0.613, 0.74}. In order to check the adequateness of the Kumaraswamy distribution to these data, we apply the Kolmogorov-Smirnov test. We find p-value 0.2092 and p-value = 0.3359 for the first and second data sets, respectively. Since both satisfy p-values >0.05, the two considered data set are in adequateness with the Kumaraswamy distribution. Now, Tables 20 and 21 present the estimations of the six entropy measures considered in Subsection 2.2, following the methodology described in Subsection 3.1, for the first and second data sets, respectively.
Table 20

Estimates of the considered entropy measures with different values of δ for the first data set.

δRδ(a, b)HCδAδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
0.5-0.395-0.881-0.597-0.730-2.6680.807
1.5-0.487-2.570-1.361-1.5051.6680.567
2.5-0.523-7.887-1.768-3.3990.4380.530
Table 21

Estimates of the considered entropy measures with different values of δ for the second data set.

δRδ(a, b)HCδAδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
0.5-0.201-0.499-0.370-0.413-2.6630.801
1.5-0.291-1.359-0.751-0.7961.6660.570
2.5-0.327-3.245-0.953-1.3990.4370.533
We can notice that, under our framework, the Rényi, Havrda and Charvat, Arimoto, Tsallis, Awad and Alawneh 2 entropy measures are decreasing when δ is increasing while the Awad and Alawneh 1 entropy is increasing when δ is increasing. Tour knowledge, it is the first time that the entropy of the uncertainty behind these data sets are evaluated. They can be taken into account for further statistical analysis in the future.

4 Conclusion

For the first time, this article proposed a special focus on the entropy of the Kumaraswamy distribution. Both theoretical and practical aspects were covered, though complementary works. In particular, six different entropy measures were investigated. After determining the closed-form expressions of these measures, an estimation strategy was developed to evaluate them in a practical setting. A simulation study ensured the convergence of the obtained estimates. Two real-life data sets are used to show how the related entropy can be concretely estimated. The finding of this study aims to be applied by the statistician to assess the entropy of diverse data with values on the unit interval, such as modern rate, percentage and proportion type data. The limitation of current research remains on the classicity of the statistical framework considered. Directions for future research include the estimation of the entropy of the Kumaraswamy distribution in more sophisticated statistical schemes with physical motivations, such as the progressive type II censoring scheme, generalized progressively hybrid censoring scheme, etc., or taking into account generalized versions of the Kumaraswamy distribution, such as the one proposed by [40].
Table 3

Numerical values of the considered entropy measures of the Kumaraswamy distribution at a = 2 and δ = 1.5.

bRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.069-0.280-0.162-0.164-0.1080.398
2.0-0.075-0.306-0.177-0.179-0.1130.416
2.5-0.091-0.377-0.217-0.221-0.1200.439
3.0-0.110-0.461-0.264-0.270-0.1250.457
3.5-0.129-0.547-0.312-0.320-0.1290.471
4.0-0.147-0.631-0.359-0.370-0.1320.483
4.5-0.165-0.713-0.404-0.418-0.1350.491
5.0-0.181-0.792-0.447-0.464-0.1370.499
5.5-0.197-0.867-0.488-0.508-0.1390.505
6.0-0.211-0.939-0.528-0.550-0.1400.510
Table 6

Numerical values of the considered entropy measures of the Kumaraswamy distribution at b = 2 and δ = 1.5.

aRδ(a, b)HCδ(a, b)Aδ(a, b)Tδ(a, b)AA1δ(a, b)AA2δ(a, b)
1.5-0.055-0.221-0.128-0.130-0.0970.361
2.0-0.075-0.306-0.177-0.179-0.1130.416
2.5-0.111-0.466-0.267-0.273-0.1280.468
3.0-0.151-0.649-0.369-0.380-0.1400.508
3.5-0.191-0.838-0.473-0.491-0.1490.537
4.0-0.228-1.026-0.574-0.601-0.1560.561
4.5-0.264-1.211-0.673-0.709-0.1610.579
5.0-0.297-1.392-0.768-0.815-0.1660.594
5.5-0.328-1.568-0.860-0.918-0.1700.606
6.0-0.358-1.739-0.948-1.019-0.1730.617
  7 in total

1.  Bayesian spatial-temporal model for the main and interaction effects of roadway and weather characteristics on freeway crash incidence.

Authors:  Huiying Wen; Xuan Zhang; Qiang Zeng; N N Sze
Journal:  Accid Anal Prev       Date:  2019-08-12

2.  Shannon entropy as a new measure of aromaticity, Shannon aromaticity.

Authors:  Siamak Noorizadeh; Ehsan Shakerzadeh
Journal:  Phys Chem Chem Phys       Date:  2010-03-16       Impact factor: 3.676

3.  Injury severities of truck drivers in single- and multi-vehicle accidents on rural highways.

Authors:  Feng Chen; Suren Chen
Journal:  Accid Anal Prev       Date:  2011-04-22

4.  Analysis of hourly crash likelihood using unbalanced panel data mixed logit model and real-time driving environmental big data.

Authors:  Feng Chen; Suren Chen; Xiaoxiang Ma
Journal:  J Safety Res       Date:  2018-04-25

5.  Estimation of Entropy for Inverse Lomax Distribution under Multiple Censored Data.

Authors:  Rashad A R Bantan; Mohammed Elgarhy; Christophe Chesneau; Farrukh Jamal
Journal:  Entropy (Basel)       Date:  2020-05-28       Impact factor: 2.524

Review 6.  A Brief Review of Generalized Entropies.

Authors:  José M Amigó; Sámuel G Balogh; Sergio Hernández
Journal:  Entropy (Basel)       Date:  2018-10-23       Impact factor: 2.524

Review 7.  The Entropy Universe.

Authors:  Maria Ribeiro; Teresa Henriques; Luísa Castro; André Souto; Luís Antunes; Cristina Costa-Santos; Andreia Teixeira
Journal:  Entropy (Basel)       Date:  2021-02-11       Impact factor: 2.524

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.