Literature DB >> 33286463

A New Belief Entropy in Dempster-Shafer Theory Based on Basic Probability Assignment and the Frame of Discernment.

Jiapeng Li1, Qian Pan2.   

Abstract

Dempster-Shafer theory has been widely used in many applications, especially in the measurement of information uncertainty. However, under the D-S theory, how to use the belief entropy to measure the uncertainty is still an open issue. In this paper, we list some significant properties. The main contribution of this paper is to propose a new entropy, for which some properties are discussed. Our new model has two components. The first is Nguyen entropy. The second component is the product of the cardinality of the frame of discernment (FOD) and Dubois entropy. In addition, under certain conditions, the new belief entropy can be transformed into Shannon entropy. Compared with the others, the new entropy considers the impact of FOD. Through some numerical examples and simulation, the proposed belief entropy is proven to be able to measure uncertainty accurately.

Entities:  

Keywords:  Dempster–Shafer theory; Shannon entropy; basic probability assignment; belief entropy; frame of discernment; uncertainty measure

Year:  2020        PMID: 33286463      PMCID: PMC7517227          DOI: 10.3390/e22060691

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


1. Introduction

How to measure uncertainty is a meaningful question to be solved. Our work will also discuss this issue. First of all, we need to know what uncertainty is. The uncertainty problem is still very extensive. It means not certainly known, questionable, and problematic. Uncertainty can mainly divided into three types: vagueness, which is boundary uncertainty; nonspecificity, which is size (cardinality) uncertainty; and discord, which expresses conflict. Correspondingly, there are some theories to solve these problems: fuzzy set theory [1], probability theory [2], evidence theory [3,4], and rough sets [5]. Besides, some extended theories are also presented for the uncertainty measure, e.g., generalized evidence theory [6], complex numbers [7], fuzzy numbers [8,9,10], Z numbers [11,12], D numbers theory [13,14,15,16], and so on [17,18,19,20,21,22]. In this paper, we use evidence theory to study these open issues. In 1960, Dempster [3] proposed upper and lower probabilities to solve the multivalued mapping problem. In 1971, Shafer [4] completed the theory proposed by Dempster and formed evidence theory, also called D-S theory. After years of exploration, D-S theory is a very effective tool for modeling and processing information uncertainty. In 1948, Shannon [23] used the concepts in thermodynamics to define information entropy. Under probability theory, Shannon entropy was very good at measuring the degree of information uncertainty. However, D-S theory is easier than probability theory for getting prior data, and the former has the advantage of fused data. Thus, we introduce D-S theory to replace probability theory in uncertainty. D-S theory uses basic probability assignment (BPA), which is under the frame of discernment (FOD), to represent the degree of support for a focal element. Different FODs may have different BPAs. Besides, the core of D-S theory is Dempster’s combination rule. Dempster’s combination rule provides a way to fuse different BPAs. The proposition of evidence theory provides mathematical support for the establishment of uncertain models. On the other hand, it is well known in information theory that Hartley [24] and Shannon’s measures are both effective ways to deal with information uncertainty. Meanwhile, D-S theory as an extension of probability theory contains much ignorance information. Thus, Höhle [25] was the earliest to combine D-S theory and Shannon entropy, namely Höhle entropy. Subsequently, Nguyen [26], Dubois and Prade [27], Klir [28], Jiroušek and Shenoy [29], Nikhil R. Pal [30,31], Deng [32], Pan and Deng [33], and Wang [34] defined their uncertainty models. Some of them have been successfully applied in real situations [35,36]. However, these models are not effective in some places [37,38]. According to their entropy, most of them are only focused on the BPA of every focal element and the cardinality of an element or use the belief function and plausibility function to measure uncertainty. Therefore, no one focused on FOD. Obviously, the scale of FOD can impact the degree of uncertainty. We combined different models and proposed a new uncertainty measurement, namely B& F entropy, because the uncertainty is determined by both BPA and FOD. In this method, it can well reflect the impact of FOD on uncertainty. In the end, we will give a few examples to compare the new model and others. Besides, we design a simulation to illustrate the feasibility and effectiveness of the proposed model. The outline of the remainder of the paper is as follows. In Section 2, we briefly review the Hartley and Shannon measures and D-S theory. Some essential properties are briefly introduced in Section 3. Section 4 presents and existing uncertainty measures. In Section 5, we discuss some properties and define a new entropy. In Section 6, some significant numerical examples and simulations are carried out to illustrate the feasibility and effectiveness of the proposed belief entropy. In Section 7, we summarize our findings and conclude with some open questions.

2. Preliminaries

The focus of this paper is based on D-S theory and information entropy. We divide this section into two parts. In the D-S theory section, some basic concepts will briefly be introduced. In the information entropy section, we will introduce two typical representatives, the Hartley measure and Shannon entropy.

2.1. D-S Theory

Dempster–Shafer theory, also called evidence reasoning or evidence theory, originated from Dempster [3] and was developed by his student Shafer [4]. Through a series of improvements and reinforcements, a method of uncertainty reasoning using “evidence” and “combination” was formed. In a way, D-S theory is a promotion of Bayesian reasoning. Dempster–Shafer theory is often applied to pattern recognition [39,40,41,42,43,44], fault diagnosis [45,46,47,48], uncertainty modeling [20,49], clustering [50], decision making [51,52], risk analysis [53,54,55,56], and other hot fields [57,58]. The idea of D-S theory is based on the frame of discernment X, which , and the set of all subsets in X is called the power set . In the power set , it contains elements. means the cardinality of X, which is the number of elements in X. Under this frame, Dempster and Shafer defined some basic concepts as follows.

2.1.1. Basic Belief Assignment

Based on the above power set , function m : satisfies: The function is also called a basic probability assignment or mass function. If , then a is a focal element. means the value of trust that the object belongs to a. The larger is, the higher the trust value is. Some definitions of BPA are as follows. The vacuous BPA means entirely unknown for the true result. In contrast, from the Bayesian BPA, we can know to which category the target should belong.

2.1.2. Belief Function

The belief function is the sum of the basic probability assignments for all subsets of a and is given by: It is the lower limit of support for a.

2.1.3. Plausibility Function

The plausibility function is the sum of the basic probability assignments for all subsets that intersect with a and is given by: It is the upper limit of support for a. The value, between the belief function and plausibility function, is the degree of uncertainty for evidence.

2.1.4. Dempster’s Combination Rule

Dempster’s combination rule is the most commonly used method in evidence fusion. This rule takes into account the degree of conflict between the evidence and defined conflict coefficient k to measure the degree of conflict among different evidence. Suppose and are independent BPAs from the different evidence resources, respectively. The fusion result of and under Dempster’s combination is as follows: where k is a conflict coefficient, defined by: Notice that Dempster’s combination rule is invalid, if two bodies of evidence completely conflict . Furthermore, if , Dempster’s combination rule cannot be applied to the two BPAs’ fusion.

2.2. Origin of Information Entropy

Different authors have measured information uncertainty in a variety of ways, and Hartley and Shannon laid the foundation for it. The information entropy and their extended models have been applied to many fields [59]. Next, we will briefly introduce the Hartley measure and Shannon entropy.

2.2.1. Hartley Measure

Suppose X is an FOD and a is a subset of X. Then, the Hartley measure [24] is defined as: where means the cardinality of a. Obviously, the measurement is proportional to the cardinality of a. When a is a singleton of X, , this means there is no conflict. Unfortunately, the measurement method of Hartley does not show the effect of the probability distribution on the degree of uncertainty.

2.2.2. Shannon Entropy

In 1948, Shannon [23] proposed information entropy, namely Shannon entropy. His model uses the concept of entropy from thermodynamics. where is the probability of x and satisfies . As he said in his thesis, the role of information is used to eliminate the uncertainty. Shannon entropy is an excellent way to measure and eliminate uncertainty. It played a crucial role in solving the probability problem. We can conclude from his definition that it is based on the probability distribution. With the emergence of D-S theory, the information entropy was given a new meaning. The format of our new model is also derived from the Shannon entropy.

3. Properties of the Uncertainty Measure in D-S Theory

According to Klir and Wierman [60] and Klir and Folger [61], we introduce some important properties of entropy for D-S theory, including non-negativity, maximum, monotonicity, probability consistency, additivity, sub-additivity, and range. These properties for a measure that captures both discord and non-specificity are defined as follows.

3.1. Non-Negativity

Suppose m is a BPA on FOD X; the entropy must be: where this is equality if and only if and . Only when entropy satisfies the non-negativity property, it provides a standard for measurement uncertainty.

3.2. Maximum Entropy

It makes sense that the vacuous BPA for uncertainty is lager than other normal BPAs . Thus, the maximum entropy property is defined as:

3.3. Monotonicity

As the number of focal elements in FOD increases, so should the degree of uncertainty. The monotonicity property is defined as: where and are the vacuous BPAs for FOD X and FOD Y. Meanwhile, .

3.4. Probability Consistency

Let be a Bayesian BPA, and then, the entropy should be the same as Shannon entropy. Therefore, the probability consistency property follows as: where is the Shannon entropy and is the BPA of X corresponding to .

3.5. Additivity

Let and be independent BPAs on FOD X and FOD Y, respectively. ⊕ means Dempster’s combination rule. Thus, the additivity property is defined as: where is a BPA for FOD . Note that , where m is the and combined by Dempster’s combination rule.

3.6. Sub-Additivity

Let m be a BPA on FOD . Let and be the marginal BPAs of FOD X and FOD Y. Then, define:

3.7. Range

As Klir and Wierman defined, the range of is .

4. The Development of Entropy Based on D-S Theory

In this section, some belief entropies of BPAs in D-S theory proposed by others are reviewed. We also discuss whether or not these models satisfy the properties we list. Yager [62] defined the belief entropy using the conflict coefficient between two focal elements, simplified as follows: where is the plausibility function associated with a under m. The entropy of Yager only measures the degree of conflict between evidence. only satisfies the additivity property. Dubois [27] used a new information measurement method to get the new formula of entropy. From the definition of Dubois, this entropy only answers the question of the non-specific part of the uncertainty. If m is a Bayesian BPA, then . It is noticeable that is clearly a weighted Hartley [24] measure. satisfies the maximum entropy and monotonicity properties. Nguyen [26] defined a new entropy according to Shannon entropy. From the definition format, it only uses the BPA to capture the part of the conflict. This is inaccurate for uncertain measurements. It only satisfies the probabilistic consistency property and the additivity property. Lamata and Moral [63] used the entropy theory proposed by Yager and Dubois. They both have two components: one measures the innate contradiction, while the other measures the imprecision of the information. This definition does not satisfy the maximum entropy and sub-additivity properties. Jiroušek and Shenoy [29] entropy is a combination of the Shannon and Dubois definitions. where is the normalized result of plausibility function . The first part is the measurement of conflict based on Shannon entropy, and the second part is to measure the non-specificity portion of uncertainty. The entropy of satisfies non-negativity, maximum entropy, monotonicity, probability consistency, and additivity. Klir and Ramer [28] defined: Due to the Yager entropy not concluded the broader view of conflict (it only considered the conflict situation of ), Klir and Ramer proposed a new method to solve this problem. It is easy to see that this entropy can measure the conflict of evidential claims within each body of evidence in bits. However, under certain conditions, it is difficult for to express the aspect of uncertainty. It just does not satisfy the maximum entropy property. Nikhil R. Pal [30,31] focused on nonspecificity and randomness under a total uncertainty environment. They summed up the methods proposed by Lamata and Moral and Klir and Ramer. It was pointed out that there would be mistakes against common sense in certain situations. The first part is, in some sense, analogous to Yager’s entropy, and the second part measures the conflict of the body of evidence. It does not satisfy the maximum entropy property. Jousselme [64] entropy is based on pignistic transformation [65]. He finally proved that as the evidence changes, the entropy becomes more sensitive. Deng [32] defined an entropy: As proven by Joaquín Abellán [66], the Deng entropy does not satisfy the monotonicity, additivity, and subadditivity properties. Pan and Deng [33] developed Deng entropy and defined it as follows: where and are the belief function and plausibility function, respectively. uses the interval probability to measure the discord and non-specificity uncertainty of BPA. It does not satisfy the maximum entropy, additivity, sub-additivity, and range properties. W [34] is another modified model based on Deng entropy: where is a constant and , is the function about the cardinality of X. is a change number; it can take different values to represent different entropies. However, as the parameter changes, it has little effect on the value of W entropy [34].

5. A New Belief Entropy Based on Evidence Theory

As introduced at the start of the first chapter of Shafer’s book [4], D-S theory is a theory of evidence. That means using the mathematical form to express the degree of support for evidence. Based on the entropy proposed by previous scholars, for the measurement method of information uncertainty, there remain several aspects of the frame of discernment about which relatively little is known. In D-S theory, if we have the same cardinality of BPA, but different FODs, the results of uncertainty should be changed. However, most of them we listed above only focused on the value of BPA or the cardinality of every BPA, and the effect of FOD was totally ignored. Thus, these definitions cannot measure the degree of uncertainty under different FOD. To improve these deficiencies, we suggest that the FOD is also important for the measurement of uncertainty. Therefore, we introduce the scale of FOD to our new entropy. The new belief entropy based on D-S theory, namely B& F entropy, is defined as follows: where denotes the cardinality of the focal element a and equals the number of elements in FOD. Like some of the definitions we mentioned, the new definition can be represented by a combination of other entropies. Thus, the new entropy also can be expressed as: where is Nguyen’s entropy and is Dubois’ entropy. Obviously, the new entropy is a combination of and times . Similar to most of the belief entropies, the first component in the new belief entropy is designed to measure the discord uncertainty of BPA. At length, the second component is the measure of non-specificity of the mass function among various focal elements [27,32,61]. In addition, it can capture the information about the size of cardinality. When m is a Bayesian BPA or the cardinality of FOD equals one, the new entropy degenerates to Pal’s definition. The most important information about FOD is the quantity of the focal element, namely . If is modified, the accuracy of uncertain measurement will be affected. Here, we use an example to show that is the best way to represent the information of FOD. As shown in Figure 1, it is obvious that and cannot reflect the effect of FOD on entropy very well. When the cardinality of FOD is greater than 10, is almost constant, but is very large. Thus, can well contain the information of the FOD size.
Figure 1

Comparison of different frame of discernment (FOD) information.

The new entropy connects the degree of information uncertainty and the FOD, meanwhile improving the information uncertainty measurement method. According to Section 3, the basic properties of the new belief entropy are proven as follows: (P1) Non-negativity: Let be a cardinality of the focal element and be a cardinality of FOD. It is obvious that ; thus, , if and only if m is the Bayesian BPA and . Therefore, the new definition satisfies the non-negativity property. (P2) Maximum entropy: Let be a Bayesian BPA and be a vacuous BPA, then , . Although, according to our calculations, , it does not mean is the maximum value. Later, we will further explain the max value through simulation. In this part, we just give some simple explanations. From the definition of Nguyen we introduced, this entropy does not satisfy the maximum entropy, as it consists of Nguyen’s entropy and Dubois entropy. Thus, the maximum entropy is not satisfied with the new belief entropy. (P3) Monotonicity: We suppose that denotes the vacuous BPA, then . Obviously, increases with . Therefore, satisfies the monotonicity property. (P4) Probability consistency: When is a Bayesian BPA, then , . From this result, we conclude that the new belief entropy satisfies the probability consistency property. (P5) Additivity and sub-additivity: Let , where a, b, c is a focal element and X, Y means the FOD. Meanwhile, and . According to the definition of the above properties, , where is the marginal BPA for X and is the marginal BPA for Y. We can see from the above proof that the new entropy satisfies the additivity property, if and only if . Otherwise, the new belief entropy neither satisfies the additivity property nor sub-additivity. To be more intuitive, we consider the following example: Let Z be the product of FOD and FOD . We have that BPA on Z is m, and the marginal BPAs on X and Y are and . We suppose the case on Z is shown as follows: where . Thus, the BPAs on X and Y are: The calculation results are as follows: Obviously, . Therefore, the additivity and sub-additivity properties are not satisfied with the new entropy. (P6) Range: As demonstrated by the maximum entropy property, the value of the new entropy , and . Thus, it does not satisfy the range property. From the above results we proved, the new belief entropy satisfies the non-negativity, monotonicity, and probability consistency properties, and does not satisfy the maximum entropy, additivity, subadditivity, and range properties.

6. Numerical Example and Simulation

In the first part of this section, some examples are given to illustrate the effectiveness of the new belief entropy. The influence of different BPAs on entropy is shown in the second section.

6.1. Numerical Example

6.1.1. Example 1

Let FOD , and we get a BPA from the sensor as . Shannon entropy and the new definition proposed by the authors’ calculation results are as follows:

6.1.2. Example 2

Suppose there are three FODs , , . Every Bayesian BPA of these FODs is equal. Their BPAs are as follows: The new belief entropy is calculated as follows: It is obvious that uncertainty increases as the number of focal elements increases. This is reasonable.

6.1.3. Example 3

Using the FOD raised by Example 2 and the vacuous BPAs , , , the new entropy results are calculated as follows: Comparing Example 2 and Example 3, it is easy to get that the results of the vacuous BPA are bigger than the results of the Bayesian BPA.

6.1.4. Example 4

In this example, we compare the difference between Pal entropy and entropy. Let FOD and . Meanwhile, suppose the following two situations exist: Thus, the Pal entropy and entropy results calculated and compared are the following: We can draw the following conclusions: By comparison, we can conclude that the result of is more reasonable. Because of has fewer focal elements and they have the same element in two BPAs, therefore, the uncertainty of should be bigger than the uncertainty of . From an overall view, as long as the focal elements for every BPA are equal, the results of Pal entropy keep constant, even if the number of focal elements on FOD is different. This is unreasonable. However, for the new belief entropy, it reflects the impact of the number of FODs on information uncertainty. Obviously, the degree of information uncertainty is proportional to FOD. Thus, the new definition proposed in this paper is more reasonable for the above Section 6.1.4.

6.1.5. Example 5

In this example, we suppose a FOD that has ten focal elements, and four mass functions, , where is the subset of and i is equal to the cardinality of B. We chose ten subsets of to assignment B and used Dubois entropy, Deng entropy, Pan–Deng entropy, and the new belief entropy for comparison. In Section 4, we already listed these definitions of entropy. When changes, their values can be calculated by MATLAB. The calculation results of these definitions are shown in the following Table 1.
Table 1

The value of different definitions when changes.

CasesDubois EntropyDeng EntropyPan–Deng EntropyNew Entropy
B1={1} 0.41142.662316.14435.1363
B2={1,2} 1.21143.930317.491613.1363
B3={1,2,3} 1.67944.908219.860817.8160
B4={1,24} 2.01145.787820.822921.1363
B5={1,25} 2.26906.625621.831423.7118
B6={1,26} 2.47947.444122.752125.8160
B7={1,27} 2.65738.253224.133127.5952
B8={1,28} 2.81149.057825.068529.1363
B9={1,29} 2.94749.860026.021230.4957
B10={1,210} 3.069010.661227.194731.7118
Table 1 and Figure 2 show that the new belief entropy is larger than Deng entropy and Dubois entropy. On the other hand, the growth trend of the new belief entropy is slower than Deng entropy and Pan–Deng entropy and the same as Dubois entropy. For example, we chose , and , to illustrate the impact of each additional element in on uncertainty, under different cardinality of . From the Table 1, we can get:
Figure 2

Comparison between the new belief entropy and other entropies.

Where the P& D entropy in Figure 2 is the we listed in Section 4. Although the four entropy values in Figure 2 increased, their slopes were different. Deng entropy and Pan–Deng entropy increased linearly, while the slopes of Dubois entropy and the new entropy decreased with the increase of the cardinality of B. We believe that the growth trend of the latter was more reasonable. This was because the scale of B was an important indicator to measure the change of information uncertainty, which should change with the size of cardinality. With the same cardinality of , our new belief entropy was larger than the Dubois entropy. It could well reflect the degree of uncertainty. Therefore, through comprehensive analysis, we considered that the new belief entropy was more accurate. Yager entropy, Pal entropy, Klir and Rammer entropy, and Jiroušek and Shenoy entropy are plotted in Figure 3.
Figure 3

Results’ comparison of other entropies.

From Figure 3, it can be seen that these definitions kept a small value. The degrees of uncertainty measured by Klir and Rammer and Yager decreased visibly with the increasing of the elements in B. This was understandable. The uncertainty measures proposed by Pal and Jiroušek and Shenoy were nearly linear with the cardinality of B. They had the same growth trend as Deng entropy. Where the J& S entropy in Figure 3 is the we listed in Section 4.

6.1.6. Example 6

In recent years, much research has been modified based on Deng entropy theory [33,34,37]. In this example, we chose to compare W entropy and our new model. Although W entropy takes into account the scale of FOD, the effect of the scale of FOD on W entropy is very limited [34]. As Equation (26) shows, the value of our new model would change exponentially with the scale of FOD. As they showed in their examples, when increased from zero to 10, the change trend of W entropy was almost the same as Deng entropy. However, as we demonstrated in Section 6.1.5, the growth trend of entropy was different from Deng entropy. Therefore, we could see the effectiveness and superiority of the proposed entropy.

6.1.7. Example Summary

Based on the examples proposed above, we list some typical cases that may affect the new belief entropy and compare it with other entropies. From Section 6.1.2 and Section 6.1.3, we could see that the new entropy was more sensitive to the vacuous BPA. Section 6.1.4 shows the limitations of the general entropy, and the new entropy could solve the problem caused by the different number of FODs. Section 6.1.5 reflected the change of the new entropy and other entropies as the number of elements increased. In Section 6.1.6, we made a simple comparison between W entropy and entropy.

6.2. Simulation

Here, we use MATLAB to complete the test. This test could more intuitively feel how the new belief entropy changed with the different BPAs. We supposed an FOD . This FOD had three BPAs, , , and . and can take any value from zero to one. However, according to D-S theory in Section 2, we limited the value of these BPAs, where . Obviously, exists only when . The simulation results are as Figure 4 and Figure 5 show, where the x-axis is , the y-axis is , and the z-axis means the value of the new entropy.
Figure 4

The value of the new belief entropy with changes of BPA.

Figure 5

The value of the new belief entropy with changes of BPA.

When , the max value of the new entropy , where . When considering the BPA , , and could get the max value of the new entropy, . From this max value result, we obtained that the new definition did not satisfy the maximum entropy property. Analysis: These simulation results suggested that the main trend of the new entropy was changing with different BPAs. It also indicated that the new entropy increased as the vacuous BPA increased, when , which was reasonable. Therefore, the new entropy could reflect well the degree of measurement of information uncertainty.

7. Conclusions and Discussion

First of all, we reviewed some earlier definitions proposed by Hartley, Shannon, Yager, Nguyen, Lamata and Moral, Jiroušek and Shenoy, Klir and Ramer, Dubois, Nikhil R. Pal, Joussemle, Deng, and Pan–Deng. However, none of them reflected the number of FODs’ effect on uncertainty. We discussed an open issue, which was how to measure information uncertainty. Our principle was to include as much known information as possible under D-S theory. Thus, in this paper, we considered the cardinality of FOD and defined a new model to measure uncertainty. Meanwhile, some properties of the new entropy were discussed. The result of the examples and simulation proved that the new entropy could be more effective and accurate when compared with other entropies. When the target belonged to the set of clusters and the total number of targets could not be determined, our method could get the information uncertainty from the target accurately. Compared with traditional methods, the new entropy was easy to calculate. This meant that in the same time, it could process more data. In future work, we will apply it to solve practical problems and improve it in real applications.
  4 in total

1.  Conflict management based on belief function entropy in sensor fusion.

Authors:  Kaijuan Yuan; Fuyuan Xiao; Liguo Fei; Bingyi Kang; Yong Deng
Journal:  Springerplus       Date:  2016-05-17

2.  Sensor Data Fusion with Z-Numbers and Its Application in Fault Diagnosis.

Authors:  Wen Jiang; Chunhe Xie; Miaoyan Zhuang; Yehang Shou; Yongchuan Tang
Journal:  Sensors (Basel)       Date:  2016-09-15       Impact factor: 3.576

3.  A modified belief entropy in Dempster-Shafer framework.

Authors:  Deyun Zhou; Yongchuan Tang; Wen Jiang
Journal:  PLoS One       Date:  2017-05-08       Impact factor: 3.240

4.  Modeling Sensor Reliability in Fault Diagnosis Based on Evidence Theory.

Authors:  Kaijuan Yuan; Fuyuan Xiao; Liguo Fei; Bingyi Kang; Yong Deng
Journal:  Sensors (Basel)       Date:  2016-01-18       Impact factor: 3.576

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.