Academia.eduAcademia.edu
INFORMATION SCIENCES 61,211-231(1992) 211 zyxwvutsrqp Higher Order Fuzzy Entropy and Hybrid Entropy of a Set N. R. PAL* and S. K. PAL+ zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA Indian Statistical Institute, 203 B.T. Road, Calcutta 700 035, India Communicated by Abraham Kandel ABSTRACT New entropy measures such as higher order fuzzy entropy and hybrid entropy of a set have been introduced. Various properties along with their proofs have been included. Applicability of these new measures to various problems has been highlighted. 1. INTRODUCTION The present work consists of three parts. The existing fuzzy entropy measures have been critically analyzed in the first part. In the second part, a need for a higher order entropy H’ (r a 11, which will give a measure of average uncertainty associated with any arbitrary subset with r-supports, has been established. The new definition has been introduced in such a way that when r = 1, the definition boils down to that of the existing entropy of a fuzzy set. Various attempts, available in the literature, to combine the probabilistic and fuzzy (possibilistic) entropy have been critically reviewed and a new definition of hybrid entropy (Z-Z,,) has been introduced in the third part of the paper. This measure can be regarded as a generalized entropy of a set such that when the fuzziness is removed, the measure turns to be the classical entropy of the set. *With the Computer Science Unit. ‘With the Electronics and Communication Sciences Unit. @Ehevier Science Publishing Co., Inc 1992 655 Avenue of the Americas, New York, NY 10010 zyxwvutsrqponmlkjihgfedcbaZYXWVU 0020- 0255/ 92/ $05.00 N. R. PAL AND S. K. PAL 212 Various properties of these measures have been stated and proved. The higher order entropy zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIH H’ is found to possess some interesting properties which can be used in feature evaluation and image segmentation problems. As expected, H’ conveys more information regarding actual structure of the set than H’ (conventional fuzzy entropy) does. H,,,,, on the other hand, can be considered an objective function for proper defuzzification (enhancement) of a set. These characteristics have been further demonstrated with examples. 2. CLASSICAL ENTROPY Shannon [l, 21 defined the term entropy to represent a quantitative measure of information of a probabilistic system. He quantified information as a logarithmic function of the probability distribution. Let pi be the probability of occurrence of the ith state of an n-state system, 0 G pi Q 1, Epi = 1. The gain in information from the occurrence of the ith state of such a system is defined as logO/pi), and the expected value of the gain function is called the entropy of the system. Thus the entropy is H= (1) zyxwvutsrq 2 Pilog(l/Pi). i=l Recently, Pal and Pal 131 have pointed out some justification in favor of taking exp0 - pi) as the gain in information from the occurrence of the ith state of a probabilistic system. Therefore, a new definition of entropy takes the form H= tPiexp(l-pi). (2) i=l It should be mentioned here that both of the above measures give the average amount of information that would be gained from the occurrence of any arbitrary state of a probabilistic system. 3. ENTROPY MEASURES OF FUZZY SETS Zadeh [4] introduced the concept of fuzzy sets in which imprecise knowledge can be used to define an event. A fuzzy set A is represented as A={xi/pLA(Xi),i=1,2 )...) n}, (3) HIGHER ORDER WZZY ENTROPY 213 where ,u,&J gives the degree of belongingness of the element xi to the set A. In the case of an ordinary set an element either belongs to or does not belong to a set. Thus an ordinary set A is defined by the characteristic function C,: R" *{O,1). On the other hand, a fuzzy set A is defined by a characteristic function pA: Rn -+ [O,l]. The function p, associates with each x E R" a grade of membership to the set A and is known as the membership function. Similarly, a property P defined on an event Xi is a function P(x,), which can have values only in the interval [O,ll. A set of these functions which assigns the degree of possessing some property P by the event xi constitutes what is called a property set. In the areas of pattern recognition, image processing, speech recognition etc., it is often required to get some idea about the degree of ambiguity (fuzziness) present in a fuzzy set. A measure of fuzziness is expected to give the average amount of difficulty that is involved in taking a decision whether an element belongs to the set or not [5-71. It is to be noted that this is not a property of the individual element of the set but a property of the set as a whole. Therefore, if every element of the set has a membership value of 0.5, then the fuzziness in the set should be m~imum. There have been different definitions of entropy for a fuzzy set to provide measures of fuzziness. Zadeh [4] defined the entropy of a fuzzy subset A for the finite set (x*,xz,..., x,,) with respect to the probability distribution {pi, pz, . . . , p,} as zyxwvutsrqpon HP= - i PA(Xi)PilOg(Pi), (4) i=l where pA is the membership function of A, and pi is the probability of occurrence of xi. Zadeh in [4] did not clarify the physical significance of I;r’; even he did not mention whether HP can be used as a measure of fuzziness or not. This NP can be viewed as a weighted Shannon entropy where the membership values are taken as the weight. Let us now see how the use of HP, as a measure of fuzziness, leads to an undesirable situation. Consider a fuzzy set with two elements x1,x2 with probabilities pl,p2 as shown in Table 1. Since probabilities are same for both the cases, the fuzziness is expected to be higher for case 1; but this is not reflected in Hp. This example therefore reveals the difficulty in using HP as a measure of fuzziness. 214 N. R. PAL AND S. K. PAL TABLE 1 Value of Zadeh’s Entropy HP Case zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLK Pi Iii 1 0.5 0.5 0.01 0.99 0.028007 2 0.8 0.9 0.01 0.99 0.44800 Kaufmann [5] defined the entropy of a fuzzy set with n supports as HK={ -l/log(n)) i 4ailog(&), (5) i=l where (ibi=~i/C~=l~i, i=l,Z ,..., n. The drawback of this measure is that it does not depend on the absolute values of pi, but on their relative values. Thus a set with pi = 0.1 or 0.8 or 1 for all of its elements would have same entropy (equal to 1). This is intuitively unappealing. Deluca and Termini 161 have used a different expression, based on Shannon’s function, to define the entropy of a fuzzy set as follows. H=-k 2 (~ilOg(~i)~(l-~i)lOg(l-~i)}~ (6) i-l where k is a normalizing constant. Equation (6) is claimed to express an average amount of fuzziness (ambiguity) present in a set A and it has the following desirable properties. 1: H is minimum iff pi = 0 or 1 for all i. is maximum iff pi = OS for all i. P 3: Ha H*, where H* is the entropy of A*, a sharpened version of A. (A* is a sharpened version of A if CL* Q p for p in [O,OS) and p* >, p for p in [OS, 11.1 zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA P 4: H = H’, where H’ is the entropy of the complement set [for the sake of notational simplicity &xi) has been written as piI. P P 2: H Any measure of fuzziness should possess these four properties. Kosko [7j defined the entropy of a fuzzy set as the ratio of distances between the fuzzy set and its nearest and furthest nonfuzzy neighbours. This measure also satisfies properties P 1-P 4. HIGHER ORDER FUZZY ENTROPY 215 Pal and Pal [3] have also given a measure of fuzziness in a set as zyxwvutsrqponmlk H= k i {Iliexp(I-I-li)+(l-CLi)exP(~i)}. (7) zyxwvutsr i=l It has been proved [3] that Equation (7) satisfies properties P 1-P 4. It is to be noted here that the meaning of the entropy of a fuzzy set is quite different from that of the probabilistic entropy [Equations (1) and (211.The former gives, as mentioned before, the average amount of ambiguity (difficulty) in deciding whether an element belongs to a set or not, while the later gives the average gain in information from the occurrence of an event. Since an ordinary set is a special case of a fuzzy set, a relationship between these two types of information is expected. There have been several attempts in this regard to combine the probabilistic (Shannon’s) and possibilistic (fuzzy) entropy. These are explained below. Deluca and Termini [6] attempted to do so in the following manner. Consider an experiment in which the elements xi, x2,. . . , x, may occur, once and only once in each trial with probabilities pi, pz, . . . , p,, (pi > 0, Cipi = 1). Shannon’s entropy of the probability distribution is given by WP,,P,,..., P,) = - 5 PilOg( (8) i=l This gives the average amount of information gained from the knowledge of the occurrence of an element. Suppose there exists a difficulty in the interpretation of xi, the outcome of a trial, as 0 or 1. The amount of ambiguity involved in the interpretation of xi is given by [6] s(cLi)=-~ilog(cli)-(l-I.Li)log(l-cli). The statistical average (m) of S(CL;)is m = i (9) given, by Pis(cLi). zyxwvutsrqponmlkjihgfedcbaZYXWVUT (10) i=l This m is claimed to give the average amount of difficulty in taking a decision (0 or 1) on the elements xi, i = 1,2,. .., n. Combining Equations (8) and (101, the total entropy is defined as follows: zyxwvutsrqponmlkjihgfedcbaZYXWVUTSR H,, = H(P,,P~,...,P,,)+ m(~9P~,P~~~~~7Pn)- (11) N. R. PAL AND S. K. PAL 216 According to Deluca and Termini, Hi,, gives the total average uncertainty that one may have in making a prevision about the elements of Ix,, x1,. . . , x,) which occur as a result of the experiment, and in taking a decision 0 or 1 on their values. Xie and Bedrosian [8] defined the total entropy of a fuzzy set in a little different way. Consider a set AC containing only 0 and 1 with probabilities pa and pi (pa + pi = 1, 0 G pa, pi G 1). Suppose that, due to some reason, the sharpness in the set has changed and resulted in a fuzzy set A. The membership value of an element is changed to an arbitrary value in the range [O,OSl from 0 and in [0.5,1] from 1. In this way the ordinary set AC has been changed to a fuzzy set A. Thus the fuzzy set A has two types of uncertainties: One is due to the random uncertainty in the ordinary set, and the other is the fuzzy uncertainty arising due to the fuzziness in the set. They defined the total entropy of a fuzzy set as follows: zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPON f4,,=~(bPl)+(l/~) it (12) S(k). zyxwvutsrqponmlkjihgfedcb i=l This entropy reduces to Shannon’s entropy of a two-state fuzziness is removed (i.e., when the second part vanishes). 4. JUSTIFICATION system when the FOR NEW DEFINITIONS In this section we shall be justifying the need for two new definitions namely, hybrid and higher order entropy of fuzzy sets. Before explaining their necessity, let us, first of all, critically analyze the aforementioned measures. Regarding Equation (11) the following points are in order. Deluca and Termini [6] presented Equation (11) as if m in Equation (11) is different from Equation (6). But this is not the case. A critical analysis of the aforesaid example considered to derive Equation (11) shows that m is, in fact, equal to H in Equation (6). Suppose that the said experiment is repeated N times; the entropy of the resulting fuzzy set, as given by Equation (6), would then be H=(-l/N) ; (k%log(cL~)+(l-~;)log(I-/-%)). (13) i=l Let ni be the number of times xi has occurred; then CZini = N. Thus, k zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJI ~i {~i l og(CLi )+(l -CLi )l Og(l -Il i )) H=(-l/N) i=l or H=- 5 i-l (ni /N)(~i l og(CLi )+(l -CLi )l og(l -CLi )} zyxwvutsrqponmlkjihgfedcbaZYXWVUTS HIGHER or ORDER FUZZY ENTROPY 217 ~~j{~~log(lL;)+(l-ILi)log(l-~i)j zyxwvutsrqponmlkjihgfedcbaZYXWV H=- i=l or H= 2 PiS(pi) zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONML i=l or Therefore, zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA m in Equation (11) is not different form zyxwvutsrqponmlkjihgfedcbaZYXWV H in Equation (6). Referring back to the same experiment of Deluca and Termini, there will not be any difficulty in interpreting an outcome xi as either 0 or 1 when the associated fuzziness is removed. In other words, the entropy of the system should boil down to that of a two-state system HCP,, pl), p. + p, = 1. But this is not supported by Equation (10, which, in such a case, reduces to p,). This has also been pointed out by Xie and Bedrosian [8], H(P,,P,,..., whose measure [Equation (1211 is free from this drawback. However, Equation (12) has the following unappealing implications. They established an equivalence between fuzzy information and Shannon information because both of the information measures have the same mathematical form. If pi = pi, they inferred that the average amount of fuzzy information yielded by a fuzzy set with n elements is “equivalent” to the average amount of Shannon information yielded by n independent binary Shannon information sources. Based on this, it has been concluded that fuzzy information can be transferred to Shannon information and inversely [8]. This type of equivalence is physically meaningless except that both of the measures yield same numerical value. Moreover, fuzzy information is conceptually different from the probabilistic information. Their arithmatic sum may not yield any meaningful quantity. In that sense it is difficult to interpret H,,,. Furthermore, if fuzziness is removed, Equation (12) always reduces to H(p,,p,) irrespective of the defuzzification process. For example, consider the following two cases. In the first case, the symbols in the range [O,OS] are converted to zero and the remaining to unity. In the second case, some of the symbols in [0,0.5] are wrongly mapped to 1. In both cases Equation (12) will yield the same entropy. This is not at all a desirable property. It is to be mentioned here that since a fuzzy set is a generalized version of an ordinary set, the entropy of a fuzzy set deserves to be a generalized version of classical entropy by taking into account not only the fuzziness of the set but also the underlying probability structure. In other words, it should be such that the classical entropy becomes its special case when fuzziness is removed. It is also not necessary that, with the removal of fuzziness, the value of generalized entropy decreases. Furthermore, considering the example of 0 and 1 in N. R. PAL AND S. K. PAL 218 providing entropy measure by Xie and Bedrosian, there would be one and only one type of uncertainty (difficulty), which is associated with the interpretation of an incoming symbol as 0 or 1. Of course, this difficulty depends on two factors, namely, the probability distribution and the possibility distribution; but this cannot be the sum of two uncertainties (e.g., probabilistic and fuzzy) as done in Equation (12). Such a measure may be called hybrid enfropy. The entropy of a fuzzy set as given by either Equation (6) or (7) gives the average ambiguity in taking the decision whether an element belongs to the set or not. Consider a fuzzy set “good football players” with 20 members. For any player x, p(x) gives the degree of goodness of the player x. The entropy as given by Equation (6) or (7) for such a set gives an average amount of difficulty in taking a decision whether an individual player is good or not. This type of information is not always sufficient. Often we are interested to know if a team of, say, 11 players is formed by selecting any 11 players from the 20, to what extent, on an average, the team can be called “good.” This raises two important issues: first, how to measure the degree to which a collection of objects, as a whole, possesses the property of “goodness”; second, how to get a measure of average amount of uncertainty related to such collections. Let us now consider the first issue. Suppose xi, xz, . . . , x, are n supporting elements with membership values pi, kz, . . . , p,, with respect to some property zyxwvutsrqp P. Then to what extent does {xi, x2,.. . , x,J as a whole (collectively) possess the property P? This obviously depends on the problem at hand. For example, in a quiz team, if pi is the ability of the ith member, then the ability of the team as a whole would be max(Fi), because if one member succeeds the team succeeds. On the other hand, suppose a group of acrobats are standing in such a fashion that all of them will fall if any one of them falls. Under this situation, if pi is the stability of the ith member, then the stability of the team as a whole would be min(pi). Allen [9] has addressed this issue of properties of sets. In order to get an answer to the second problem, a new definition of entropy of a fuzzy set is required which will give a measure of average uncertainty associated with any arbitrary subset with r supports. Such an entropy may be called the rth order entropy of a fuzzy set. The definition should be such that for r = 1 it corresponds to a measure satisfying the properties P 1-P 4 of Equation (6) or (7). 5. NEW DEFINITIONS In this section we shall define the higher order entropy and the hybrid entropy of a fuzzy set. HIGHER ORDER FUZZY ENTROPY 5.1. 219 HIGHER ORDER ENTROPY Let P be a fuzzy property set with a finite number of supports n, i.e., n}, where pi denotes the degree to which xi possesses P={ILi/Xi, i=l,2,..., the property zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDC P. Out of n elements consider a combination of r elements. Let S,!’denote the ith such combination and &Sir) denote the degree to which the combination Sy, as a whole, possesses the property P. There are (( y)} such combinations. The entropy of order r of the fuzzy set A is defined as (2 H’= (I/( :)) igl (PL($WP{~-P(&?) (14) A similar definition using the logarithmic function can also be given as follows: H&(-1,(;)) (3 c (~(~~)log{~(s~)~ i=l +(l-~(s~)]log(I-cL~~~)~l~ (15) In our subsequent discussion, unless stated explicitly, H’ will refer to Equation (14). Therefore, H’ will give a measure of the average amount of difficulty in taking a decision on any subset of size r with respect to the property P. In the example of acrobats, H’ will denote an average amount of ambiguity (difficulty) in deciding a random team of r acrobats as “stable.” If r = 1, H’ in Equations (14) and (15) reduces to Equations (7) and (61, respectively. H’ has the following properties. 1. H’ attains a maximum if pi = 0.5 for all i = 1,2,. . .,n. 2. H’ attains a minimum if pi = 0 or 1 for all i = 1,2,. . ., n. Pr 3. H’> H*‘, where H*’ is the rth-order entropy of a sharpened version of the fuzzy set. Pr 4. H’ is, in general, not equal to H”, where H” is the rth-order entropy of the complement set. Pr 5. H’> H’+’ when all pi E [0,0.5] and H’ Q H’+l when all pi E [0.5,11. Pr Pr Note that the property P 4 of Equation (6) or (7) is not, in general, valid here. The additional property Pr 5 implies that H’ is a monotonically nonincreasing function of r for pi E [0,0.5] and a monotonically nondecreasing 220 N. R. PAL AND S. K. PAL function of zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDC r for CLiE [OS, l] (when min operator has been used to get the group membership value). Proof of Pr l- Pr 3 for H’ (i.e., for r = 1) are given in [3] when exponential behavior of gain function is considered in the entropy measure. Their proofs for H’ (r > 1) are straightforward. The proof of Pr 5 is given below. Let the property set P be defined from the set of supports Sir as j=1,2 P[=((p;./xj), ,..., r,i=1,2 ,..., (:)), where xi E Sl and p; denotes the degree of possessing the property P by the jth element of Sy. Let &$I = mini{& j = 1,2,. . . , r]. Now there are (n - r) subsets of size (r + 1) such that sr11’ c sr+’ Let us denote these subsets by Sil,:‘, j = 1,2,. . . , n - r; i.e., SLf’ is the jth subset of P of size (r + 1) such that Now sy c si’,,f l * min{$r,& -cL(S[) ,. ..,$J.} 2 min{&j,&j,. j=1,2 >F(S;fr), ,...,n .., j.&(r) - r. (16) Thus, for every Sy, i = 1,2,. . . , , there are (n - r) subsets of size 0. + 1) such that condition (16) is true. us for every &?J), there are (n - r> times &!‘+‘I such that p(S[) >&y), j=1,2 ,...,(n-r). Since from a set of size (r + 1) we can generate size r, we can infer that in the sequence &,fl)Y i= n 1,2 7..., r , ( 1 exactly (r + 1) distinct set of j=1,2 ,...,(n-r), each of /.@r+‘), has occurred exactly (r + 1) times. i= n ( 1 1,2 9*.*, r +1 HIGHER ORDER FUZZY ENTROPY 221 Thus, (,:I> c *(n-r)E>(r+l) P(Sr+‘) i=l n ( r 1 zyxwvutsrqponmlkjihgfedcbaZY - &I dsir+‘) CL, 2 n ( 1 r+l -* CL, 2 Pr+l . (17) Thus, the average value of all #J) is not less than that of &$!‘“). In order to prove Pr 5, let us investigate the behavior of an entropic function with respect to its membership values. Consider the entropic function [Equation (711 zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFE H=(l/n) 5 {CLiexp(l-~i)+(l-CLi)exP(CLi)j i=l or H=(l/n) 2 i=l c(cLi)Y where C(~~)=~~iexp(l-~~i)+(l-_i)exp(~~i). 222 Differentiating N.R.PALANDS.K.PAL C(cLi) with respect to pi, we get zyxwvutsrqponmlkjihgfedcbaZYXW *C’(I_li)>O <o for zyxwvutsrqponmlkjihgfedcbaZYXWVUT 0 < jLi < OS for 0.5 <pi d 1 = zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONM 0 for pi = 0.5. Thus, C(pi) increases monotonically with pi over [O,OS), attains a maximum at 0.5, and then monotonically decreases over zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONML (0.5,11. This fact and inequality (17) prove the property Pr 5. Illustration Let us consider the problem of selecting a team of 5 acrobats from a group of 10. Define a fuzzy set A called “stable acrobats” with 10 supports, i.e., A={pi/xi,i=1,2 ,..., 101, where pi denotes the stability of the ith individual xi. The fuzzy entropy of order 5 gives the ambiguity (difficulty) in deciding any arbitrary team of 5 players from the group of 10 as stable or unstable. The values of H’- H6 presented in Table 2 highlight the key features of higher order entropy as compared to H ‘. Based on Table 2 the following observations can be made. When all the pi values are same, H’- H6, as expected become the same. This is because of the fact that the difficulty in taking a decision regarding possession of a property on an individual is same as that of a group selected therefrom. The value of H’ would, of course, be dependent on the ki values. Since all the ~~ in case 2 are 0.5 (most ambiguous), the corresponding H’, value is higher than that of case 1. Consider cases 3 and 4. In case 3 the difference between H’ and Hz is very high, indicating a higher dissimilarity within the singletons of the fuzzy set with respect to the possession of property P. On the other hand, the case 4 for which H1 is very close to H 2 reflects that the fuzzy singletons of the set are more alike (similar) with respect to property P. It is also seen that the higher the similarity among singletons, the quicker is the convergence to the limiting value of H’. Based on the above observation, let us define an index of similarity of supports of a fuzzy set as S = H ’ / H ’ (when H 2 = 0, H1 is also zero and S is taken as 1). Obviously, when pi E [0.5,11 and the min operator is used to assign the degree of possession of the property by a collection of supports, S TABLE 2 Higher Order Entropy zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJI H’ Case H= H3 H4 H5 H6 s ~1,1,1,1,1,1,1,1,1,1~ 0 0 zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONML 1.0 0 zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA 0 0 0 {0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5,0.5~ {1,1,1,1,1,0.5,0.5,0.5,0.5,0.5) {0.5,0.5,0.5,0.5,0.5,0.6,0.6,0.6,0.6,0.6) {0.6,0.6,0.65,0.9,0.9,0.9,0.9,0.9,0.9,0.915) {0.8,0.8,0.8,0.8,0.8,0.8,0.9,0.9,0.9,0.9~ {0.5,0.5,0.5,0.5,0.5,0.5,0.9,0.9,0.9,0.9~ (0.7,0.7,0.7,0.7,0.7,0.8,0.8,0.8,0.8,0.8) 1 0.5 0.9fw9018 0.5384138 0.5388815 0.7483621 0.7485938 1 0.7777778 0.9915119 0.6785914 0.6135760 0.9161205 0.8028872 1 0.9166667 0.9968170 0.7810661 0.6415336 0.9790324 0.8300339 1 0.9761905 0.9990906 0.8553280 0.6495325 0.9970038 0.8416682 1 0.9960318 0.9998485 0.9059016 0.6508657 1 0.8455463 1 1 1 0.9373114 0.6508657 1 0.8463219 1.0 0.6428570 0.9892983 0.7934287 0.8782636 0.8168817 0.9323772 N. R. PAL AND S. K. PAL 224 will lie in [O,ll, as Zf’g H’+l. Similarly, when pi E [0, OS], S may be defined as H2/H1 so that S lies in zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJI [0,11. zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONM The higher the value of S, the more alike (similar) are the supports of the fuzzy set with respect to the property P. This index of similarity can therefore be regarded as a measure of the degree to which the members of a fuzzy set are alike. Let us now consider cases 5 and 6. In both cases the H1 values are almost same but the index of similarity is lower for case 5 than for case 6. This indicates that case 6 has more similar supports than case 5. Note that cases 7 and 8 also have almost identical values for H ‘ , but these values are much higher than those of cases 5 and 6. Based on this observation, we can infer that fuzzy sets corresponding to cases 7 and 8 have a larger number of supports with a lower degree of possession in [0.5, l] of the property P than those corresponding to cases 5 and 6. Again the index of similarity for case 8 is higher than that for case 7, indicating that the members of the fuzzy set corresponding to case 8 are more alike than those corresponding to case 7. Therefore, the value of conventional fuzzy entropy (H ‘ ) can only indicate whether the fuzziness in a set is low or high. In addition to this, the value of H’ also enables one to infer whether the fuzzy set contains similar supports (or elements) or not. The similarity index thus defined can be successfully used for measuring interclass and intraclass ambiguity (i.e., class homogenity and contrast) in pattern recognition and image processing problems. Let us consider the work of Pal and Chakraborty [lo], who have used the measure H1 for evaluating feature importance in the pattern recognition problem. They used the P-type membership function to represent a pattern class. They made the membership values lie in the range [0.5,1] with 0.5 corresponding to the boundary elements of the pattern class and 1 at the central point (the point corresponding to the average value of the feature). This type of representation is a very natural one. They decided a feature to be of high importance if it has low H1 value. We have already shown (e.g., cases 5 and 6, or 7 and 8 of Table 2) that it is possible to have two fuzzy sets with significantly different compactness, but almost identical values for entropy. It is therefore not appropriate to evaluate the importance of a feature just on the basis of HI. A better evaluation may possibly be done by considering H1 and S together. Obviously, a low value of H’ together with a high value of S will indicate that the feature is more important. Table 2 also reveals the validity of the property Pr 5 i.e., H’ < H’ +l when /Li E [0.5,1]. 5.2. HYBRID ENTROPY In Section 3 we have discussed the various attempts that have been made to combine the probabilistic and fuzzy uncertainties and their associated prob- 225 HIGHER ORDER FUZZY ENTROPY lems. Since an ordinary set is a special case of a fuzzy set, it is logical to think of a generalized definition of entropy of a fuzzy set which reduces to the probabilistic entropy in absence of fuzziness. Let us refer back to the example (mentioned in Section 3) of digital communication over a noisy channel. For this type of example, Xie and Bedrosian [8] assumed two types of uncertainties. But one can visualize that there is one and only one type of difficulty and that is in the interpretation of an incoming symbol as 0 or 1. Of course, this is dependent on two different factors: the probability of generation of 0 and 1 by the source and the transfer function (channel noise), which makes them fuzzy. Let p,, and pi be the probabilities of occurrence of the 0 and 1 symbols, respectively, and let pi denote the membership for the fuzzy set “symbol close to 1.” Let us consider an expression (18) zyxwvutsr i=l Differentiating E, with respect to pi, we get i=1,2 Ei(CL1)=(1/12)(l-_i)exP(l-~i), ,.**, It, (i.e., the rate of change of E, is seen to be nonnegative). Thus E, is a monotonically increasing function of pi for ki E zyxwvutsrqponmlkjih 10,11. zyxwvutsrqponmlkj In other words, as pi increases from 0 to 1, E, also increases. Now, for an incoming zero (0) symbol, if pi increases, the difficulty in the correct interpretation of the symbol increases and hence the difficulty in the interpretation of a “0” as “1” decreases; i.e., a wrong interpretation of a “0” symbol becomes favorable. On the other hand, for an incoming 1 symbol, if ki increases, the difficulty in its correct interpretation decreases. The higher the value of El, the more favorable will be the interpretation of an arbitrary incoming symbol as 1. Therefore, E, can be taken as the average likeliness (possibility1 of interpreting a received symbol as “1.” Similarly, zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA & 1=(1/n) i (l -Pi )eXP(Pi ) (19) i=l can be viewed as the average likeliness of interpreting a received symbol as 0. N. R. PAL AND S. K. PAL 226 Since pa and pr are, respectively, the probabilities of occurrence of 0 and 1, the hybrid entropy of the fuzzy set A (symbols close to 1) may be defined as (20) (using logarithmic behavior) or (21) (using exponential behavior) Hhy has the following properties. PROPERTY1. In the absence of fuzziness when np, received symbols take the value zero (0) and np, symbols take the value of one (0, then E, and E, reduce to p0 and p,, respectively, and Equation (20) or (21) boils down to classical entropy. lhus we see that Hhy boils down to classical entropy zyxwvutsrqponmlkjihgfedcb w hen a proper defiuziification process is applied to detect (restore) correctly the incoming symbols. This property has significant application in automatic image enhancement, segmentation, and the noise reduction process where the difference between the classical entropy and H, can act as an objective criterion. The lower the difference, the lesser is the fuzziness associated with the individual symbol and the higher will be the accuracy in classifying them as 0 or 1. PROPERTY2. Zf pi = 0.5 for all i = 1,2,. . . , n, then E, = E, and Equation (20) takes the form H,,=-polog (l/n) i 0.5*exp(O.5) (l/n) i-l t 0.5*exp(O.5) zyxwvutsrqpo i-l or H,,= -log(o.5*~xp(o.~))~ i.e., Hhy takes a constant value and becomes independent of p0 and pl. PROPERTY 3. E, and E, cannot take any arbitrary values. They satisfy the following conditions : 0.5*exp(0.5)*pIg E,=~0.5*exp(0.5)+(1-0.5*exp(O.5))* PI HIGHER ORDER FUZZY ENTROPY 227 zyxwvutsrq and (1-p1)*0.5*exp(0.5)<E0<l-pi*((l-0.5*exp(O.5)) The proof of this property is given below. Proof. Let us consider Equation (19), which is El = (l/n) i zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGF zyxwvutsrqponmlkjihgfedcbaZYXWVUTS Pied1 - Pi) i=l - &max (l/n)[np,*(0.5)*exp(0.5)+ zyxwvutsrqponmlkjihgfedcbaZ npl] = [since E, is found to be monotonically increasing function of CLiand the expected number of symbol 0 (1) is np, (np,) and each of them can take a maximum value of 0.5 (01 zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIH =pl+(l-pl)*0.5*exp(0.5) =0.5*exp(O.5)+(1-0.5*exp(O.5))Pt (i) and -E ~min=(l/n)[np~*O+np,*0.5*~xptO~5)l = p1 *0.5*exp(O.5) (ii) Combining (i) and (ii), one gets 0.5*exp(0.5)*plg a E,~0.5*exp(0.5)+(1- 0.5*exp(O.5))p1 (cl) 0.82436* p1 Q E, < 0.82436+0.17564* pl. Similarly, from Equation (20) we have E o_=(i/ ~){~po*i+n~l*0.5*e~(0.5)} =l-pi(l-0.5exp(O.5)} (iii) =(l- p1)*0.5*exp(0.5). (3 and N. R. PAL AND S. K. PAL 228 Combining (iii) and (iv), one can write the following condition: (l-p,)*0.5*exp(0.5)GE091-pt(l-0.5*exp(O.5)) (c2) *(l-p,)*0.82436<E,<l-O.l7564*p,. This completes PROPERTY the proof. n 4. zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFE E, + E, attains the maximum value of exp(0.5) when all pi assume a value of 0.5. Thus, E, + E, G exp(0.5). Proof of this property can be found in 131. As mentioned in Property 1, when there is a proper defuzzification process applied to result in the npo times 0 symbol and the np, times 1 symbol, then E, and E, reduce to p. and p,, respectively, to make E, + El = 1. Interpretation to Image Processing Let us consider an example of a digital image in which, say, 70% pixels look white, while the remaining 30% look dark. Thus the probability of a white pixel pW is 0.7 and that of a dark pixel pb is 0.3. Suppose the whiteness of the pixels is not constant, i.e., there is a variation (grayness) and similar is the case with the black pixels. The hybrid entropy of such an image can be written as follows: H,,=-p,lOg((l/n)C~ieXP(l-~i))-~blOB((l/n)C(1-~i)exP(LL,)) i i where pi gives the degree of whiteness of the ith pixel. Now Hhy will be close to the classical entropy if there is not much variation in the grayness within the white and dark portions. Table 3 demonstrates the use of Hhy in acting as an objective criterion for a proper defuzzification process. Each fuzzy set of the Table 3 is a sharpened (defuzzified) version of the previous one such that p value increases (decreases) in the sharpened version for those p > 0.5 ( < 0.5). One can observe that with proper defuzzification E, approaches 0.7 and E, approaches 0.3; in fact, they monotonically decrease to 0.7 and 0.3, respectively, from their initial values. As a result, IIf - II,,,,] decreases to zero. Let us now consider the effect of improper defuzzification on the pattern shown in case 1 of the Table 4. Two types of defuzzifications are considered here. In cases 2-4 all the symbols with p = 0.5 are transformed to zero when some of them were actually generated from symbol “1.” In cases 5 and 6 of HIGHER ORDER FUZZY ENTROPY 229 HIGHER ORDER FWZZY ENTROPY 213 where ,u,,&xJ gives the degree of belongingness of the element xi to the set zyxwvutsrqponmlkjihgfedcbaZYXWVUTSRQPONMLKJIHGFEDCBA A. In the case of an ordinary set an element either belongs to or does not belong to a set. Thus an ordinary set A is defined by the characteristic function C,: R” --f {O,l). On the other hand, a fuzzy set A is defined by a characteristic function /A/,: Rn -+ [O,l]. The function p, associates with each x E R” a grade of membership to the set A and is known as the membership function. Similarly, a property P defined on an event Xi is a function P(x,), which can have values only in the interval [O,ll. A set of these functions which assigns the degree of possessing some property P by the event xi constitutes what is called a property set. In the areas of pattern recognition, image processing, speech recognition etc., it is often required to get some idea about the degree of ambiguity (fuzziness) present in a fuzzy set. A measure of fuzziness is expected to give the average amount of difficulty that is involved in taking a decision whether an element belongs to the set or not 1.571. It is to be noted that this is not a property of the individual element of the set but a property of the set as a whole. Therefore, if every element of the set has a membership value of 0.5, then the fuzziness in the set should be m~imum. There have been different definitions of entropy for a fuzzy set to provide measures of fuzziness. Zadeh [4] defined the entropy of a fuzzy subset A for the finite set x,1 with respect to the probability distribution {pr, pz,. . . , p,} as (+Q,,*.., HP= - i p,&i)PihT(Pi), i=l where pA is the membership function of A, and pi is the probability of occurrence of xi. Zadeh in [4] did not clarify the physical significance of HP; even he did not mention whether HP can be used as a measure of fuzziness or not. This Up can be viewed as a weighted Shannon entropy where the membership values are taken as the weight. Let us now see how the use of HP, as a measure of fuzziness, leads to an undesirable situation. Consider a fuzzy set with two elements x1,x2 with probabilities pl,p2 as shown in Table 1. Since probabilities are same for both the cases, the fuzziness is expected to be higher for case 1; but this is not reflected in HP. This example therefore reveals the difficulty in using HP as a measure of fuzziness. HIGHER ORDER FUZZY ENTROPY 231 Table 4 some of the p values greater than 0.5 which were generated from symbol 1 (or belong to the white portion of the image) are wrongly defuxxified and brought down towards zero (instead of 1). In both situations, it is to be noted that IH - Hi,,,1 does not reduce to zero. Let us now consider column no. 7 in Tables 3 and 4, which indicates that the measure Htot of Xie and Bedrosian does not have the aforesaid properties. Now, in image processing the process of defuzzification can be viewed as a contrast enhancement operation. Therefore, the measure IH - H,I can be regarded as an objective criterion for appropriate enhancement of an image. 6. CONCLUSIONS New entropy measures such as higher-order fuzzy entropy and hybrid entropy of a set have been introduced. The higher order entropy (H’) is found to possess some interesting properties which can be used in feature evaluation and image segmentation problem. It leads to define a measure, called the “index of similarity” of supports of a set. As expected, zyxwvutsrqponmlkjihgfedcbaZ H’ conveys more information about the actual structure of a set than H1 does. Hhy, on the other hand, can be used as an objective measure for proper defuzzification (enhancement) of a set. REFERENCES 1. C. E. Shannon, A mathematical theory of communication, Bell System Tech. J. 27~379-423 (Jul. 1948). 2. C. E. Shannon and W. Weaver, The Mathematical Theory of Communication, University of Illinois Press, 1949. 3. N. R. Pal and S. R. Pal, Object background segmentation using a new definition of entropy, IEE Rot. E, 136:2&t-295 (1989). 4. L. A. Zadeh, Probability measures of fuzzy events, J. Math. Anal. Appl. 23:421-427 (1%8). 5. A. Kaufmann, Introduction to the lkory of Fuzzy Subsets-Fundamental Theoretical Elements, Academic, New York, 1975, Vol. 1. 6. A. Deluca and S. Termini, A definition of nonprobabilistic entropy in the setting of fuzzy sets theory, Inform. Control 20~301-312 (1972). 7. B. Kc&o, Fuzzy entropy and conditioning, Inform. Sci. 40:165-174 (1986). 8. W. X. Xie and S. D. Bedrosian, An information measure for fuzzy sets, IEEE Trans. Syst. Man Cybemet. SMC-14 (l):lSl-156 (Jan./Feb. 1984). 9. A. D. Allen, Measuring the emperical properties of sets, IEEE Trans. Syst. Man Cybemet. SMC-4 (0x56-73 (1974). 10. S. K Pal and B. Chakraborty, Fuzzy set theoretic measure for automatic feature evaluation, IEEE Trans. Syst. Man Cykmet. SMC-16 (5):754-760 fSept./Oct. 1986). Received I4 June 1989; revised 20 November 1989