1. Introduction
In recent years, there has been numerous authors who gave characterizations of entropy measures on fuzzy sets and their generalizations. Most notably, the majority of them had worked on developing entropy measures on intuitionistic fuzzy sets (IFS). Alongside with their introduction of new ways of entropy measures on IFS, these authors have also given some straightforward examples to show how their entropy measures can be applied to various applications including multi-attribute decision making (MADM) problems [
1,
2].
In 2016, Zhu and Li [
3] gave a new definition for entropy measures on IFS. The new definition was subsequently compared against many other previous definitions of entropy measures on IFS. Montes et al. [
4] proposed another new definition for entropy measures on intuitionistic fuzzy sets based on divergence. Both of these research groups [
3,
4] subsequently demonstrated the applications of their definition of IFS onto MADM problems, and both of them deployed examples of IFS, whose data values were not derived from real-life datasets but were predetermined by the authors to justify their new concepts. On the other hand, Farnoosh et al. [
5] also gave their new definition for entropy measures on IFS, but they focused only on discussing its potential application in fault elimination of digital images rather than MADM. Ansari et al. [
6] also gave a new definition of entropy measures on IFS in edge detection of digital images. Both research groups [
5,
6] did not provide examples on how their new definitions for entropy measures on IFS may be applied on MADM.
Some of the definitions of entropy measures defined for IFS were parametric in nature. Gupta et al. [
7] defined an entropy measures on IFS, characterized by a parameter α. Meanwhile, Joshi and Kumar [
8] independently (with respect to [
7]) defined a new entropy measures on IFS, also characterized by a parameter α. An example on MADM was also discussed by Joshi and Kumar [
8], once again involving a small, conceptual IFS like those encountered in the work of Zhu and Li [
3] as well as Montes et al. [
4]. The works by Joshi and Kumar [
8] were subsequently followed by Garg et al. [
9] who defined an entropy measure on IFS characterized by two parameters: (α, β). Like the previous authors, Garg et al. [
9] discussed the application of their proposed entropy measure on MADM using a similar manner. In particular, they compared the effect of different parameters α, β on the results of such decision-making process. Besides, they had also compared the results yielded by the entropy measure on IFS from some other authors. Joshi and Kumar [
10] also defined another entropy measure on IFS, following their own previous work on the classical fuzzy sets in [
11] and also the work by Garg et al. in [
9].
For various generalizations derived from IFS, such as inter-valued intuitionistic fuzzy sets (IVIFS) or generalized intuitionistic fuzzy soft sets (GIFSS), there were also some studies to establish entropy measures on some generalizations, followed by a demonstration on how such entropy measures can be applied to certain MADM problems. Recently, Garg [
12] defined an entropy measure for inter-valued intuitionistic fuzzy sets and discussed the application of such entropy measures on solving MADM problems with unknown attribute weights. In 2018, Rashid et al. [
13] defined another distance-based entropy measure on the inter-valued intuitionistic fuzzy sets. Again, following the conventions of the previous authors, they clarified the applications of their work on MADM problem using a simple, conceptual small dataset. Selvachandran et al. [
14] defined a distance induced intuitionistic entropy for generalized intuitionistic fuzzy soft sets, for which they also clarified the applications of their work on MADM problems using a dataset of the same kind.
As for the Pythagorean fuzzy set (PFS) and its generalizations, an entropy measure was defined by Yang and Hussein in [
15]. Thao and Smarandache [
16] proposed a new entropy measure for Pythagorean fuzzy sets in 2019. Such new definitions of entropy in [
16] discarded the use of natural logarithm as in [
15], which is computationally intensive. Such work was subsequently followed by Athira et.al. [
17,
18], where an entropy measure was given for Pythagorean fuzzy soft sets—a further generalization of Pythagorean fuzzy sets. As for vague set and its generalizations, Feng and Wang [
19] defined an entropy measure considering the hesitancy degree. Later, Selvachandran et al. [
20] defined an entropy measure on complex vague soft sets. In the ever-going effort of establishing entropy measures for other generalizations of fuzzy sets, Thao and Smarandache [
16] and Selvachandran et al. [
20] were among the research groups who justified the applicability of their entropy measures using examples on MADM. Likewise, each of those works involved one or several (if more than one example provided in a work) small and conceptual datasets created by the authors themselves.
Besides IFS, PFS, vague sets and all their derivatives, there were also definitions of entropy established on some other generalizations of fuzzy sets in recent years, some came alongside with examples on MADM involving conceptual datasets as well [
21]. Wei [
22] defined an asymmetrical cross entropy measure for two fuzzy sets, called the fuzzy cross-entropy. Such cross entropy for interval neutrosophic sets was also studied by Sahin in [
23]. Ye and Du [
21] gave four different new ways entropy measures on interval-valued neutrosophic sets. Sulaiman et al. [
24,
25] defined entropy measures for interval-valued fuzzy soft sets and multi-aspect fuzzy soft sets. Hu et al. [
26] gave an entropy measure for hesitant fuzzy sets. Al-Qudah and Hassan [
27] gave an entropy measure for complex multi-fuzzy soft sets. Barukab et al. [
28] gave an entropy measure for spherical fuzzy sets. Piasecki [
29] gave some remarks and characterizations of entropy measures among fuzzy sets. In 2019, Dass and Tomar [
30] further examined the legitimacy of some exponential entropy measures on IFS, such as those defined by Verna and Sharma in [
31], Zhang and Jiang in [
32], and Mishra in [
33]. On the other hand, Kang and Deng [
34] outlined the general patterns for which the formula for entropy measures could be formed, thus also applicable for entropy measures to various generalizations of fuzzy sets. Santos et al. [
35] and Cao and Lin [
36] derived their entropy formulas for data processing based on those for fuzzy entropy with applications in image thresholding and electroencephalogram.
With many entropy measures being defined for various generalizations of fuzzy sets, it calls upon a need to standardize which kind of functions are eligible to be used as entropy measures and which are not. One of the most notable and recent works in this field is accomplished by Majumdar [
1], who established an axiomatic definition on the entropy measure on a single-valued neutrosophic set (SVNS). Such an axiomatic definition of entropy measure, once defined for a particular generalization of fuzzy set, serves as an invaluable tool when choosing a new entropy measure for a particular purpose. Moreover, with the establishment of such an axiomatic definition of entropy measure, it motivates researchers to work on deriving a collection of functions, which all qualify themselves to be used as entropy measures, rather than inventing a single standalone function as an entropy measure for a particular scenario.
In 2017, Smarandache [
2] firstly established a concept of plithogenic sets, intended to serve as a profound and conclusive generalization from most (if not all) of the previous generalizations from fuzzy sets. This obviously includes the IFS, where most works had been done to establish its great variety of entropy measures. However, Smarandache [
2] did not give any definitions on entropy measures for plithogenic sets.
Our work on this paper shall be presented as follows: Firstly, in
Section 2, we mention all the prerequisite definitions needed for the establishment of entropy measures for plithogenic sets. We also derive some generalizations of those previous definitions. Such generalizations are necessary to further widen the scope of our investigation on the set of functions that qualifies as entropy measures for plithogenic sets. Then, in
Section 3, we first propose new entropy measures for plithogenic sets in which requirements for any function to be an entropy measure of plithogenic sets are outlined in the axiomatic definition. Later in
Section 3, several new formulae for the entropy measure of plithogenic sets are also introduced. In
Section 4, we will apply a particular example of our entropy measure onto a MADM problem related to the selection of locations.
Due to the complexity and the novelty of plithogetic sets, as well as the scope constraints of this paper, the plithogenic set involved in the demonstration of MADM will be of a small caliber within 150 data values in total. Those data values contained in the plithogenic set example will be also conceptual in nature (only two to three digits per value). Such presentation, although it may be perceived as simple, is in alliance with the common practice of most renowned works done by the previous authors discussed before, whenever a novel way of entropy measure is invented and first applied on a MADM problem. Hence, such a start-up with a small and conceptual dataset does not hinder the justification on the practicability of the proposed notions. Quite the contrary, it enables even the most unfamiliar readers to focus on the procedure of such novel methods of dealing with MADM problems, rather than being overwhelmed by the immense caliber of computation encountered in dealing with up-to-date real-life datasets.
2. Preliminary
Throughout all the following of this article, let be the universal set.
Definition 1 [
1]
. A single valued neutrosophic sets (SVNS) on is defined to be the collection where and We denoteto be the collection of all SVNS on.
Majumdar [
1] have established the following axiomatic definition for an entropy measure on SVNS.
Definition 2 [
1]
. An entropy measure on SVNS is a function that satisfies the following axioms for all :
ifis a crisp set i.e.,for all.
iffor all.
ifis contained ini.e.,for all.
for all.
In the study of fuzzy entropy, a fuzzy set with membership degree of 0.5 is a very special fuzzy set as it is the fuzzy set with the highest degree of fuzziness. Similarly, in the study of entropy for SVNSs, a SVNS with all its membership degree of 0.5 for all the three membership components is very special as it is the SVNS with the highest degree of uncertainty. Hence, we denote
as the SVNS with
for all
Such axiomatic descriptions in Definition 2 of this paper, defined by Majumdar [
1], serve as the cornerstone for establishing similar axiomatic descriptions for the entropy measures on other generalizations of fuzzy sets, which shall certainly include that for plithogenic sets by Smarandache [
2].
We however, disagree with (iii) of Definition 2. As an illustrative example, let
be empty, then
has zero entropy because it is of
absolute certainty that
“does not contain any element”. Whereas a superset of
, say
, may have higher entropy because it may not be crisp. Thus, we believe that upon establishing (iii) of Definition 2, the authors in [
1] concerned only the case where
and
are very close to the entire
. Thus, on the establishment of entropy measures on plithogenic sets in this article, only axioms (i), (ii) and (iv) of Definition 2 will be considered.
Together with axioms (i), (ii), and (iv) of Definition 2, the following two well-established generalizations of functions serve as our motives of defining the entropies, which allows different users to customize to their respective needs.
Definition 3 [
1]
. Let ,
be a function satisfying the following for all .
(commutativity)
, ifand
(associativity)
Thenis said to be a T-norm function.
Example 1. “minimum” is a T-norm function.
Definition 4 [
1]
. Let ,
be a function satisfying the following for all .
(commutativity)
, ifand
(associativity)
Then,is said to be an S-norm (or a T-conorm) function.
Example 2. “maximum” is an S-norm function.
In the study of fuzzy logic, we also find ourselves seeking functions that measure the central tendencies, as well as a given position of the dataset, besides maximum and minimum. Such measurement often involves more than two entities and those entities can be ordered of otherwise. This is the reason we introduce the concept of M-type function and the S-type function, defined on all finite (not ordered) sets with entries in . Due to the commutativity of S-norm function, S-type function is thus a further generalization of S-norm function as it allows more than two entities. In all the following of this article, let us denote as the collection of all finite sets with entries in . To avoid ending up with too many brackets in an expression, it is convenient to denote the image of under as .
Definition 5 [
1]
. Let ,
be a function satisfying the following:- (i)
- (ii)
Thenis said to be an M-type function.
Remark 1. “maximum”, “minimum”, “mean”, “interpolated inclusive median”, “interpolated exclusive median”, “inclusive first quartile”, “exclusive 55th percentile”, andare some particular examples of M-type functions.
Definition 6 [
1]
. Let ,
be a function satisfying the following:- (i)
.
- (ii)
Ifcontains at least an element, then.
- (iii)
Ifcontains at least an element, then.
- (iv)
For every two sets fromwith the same cardinality:andIffor all, then.
Thenis said to be an S-type function.
Remark 2. “maximum”,andare some particular examples of S-type functions.
Lemma 1. If f is an S-type function, then it is also an M-type function.
Proof. As contains one element which equals to 1, , thus the lemma follows. □
Remark 3. The converse of this lemma is not true however, as it is obvious that “mean” is an M-type function but not an S-type function.
All of these definitions and lemmas suffice for the establishment of our entropy measure for plithogenic sets.
3. Proposed Entropy Measure for Plithogenic Sets
In [
2], Smarandache introduced the concept of plithogenic set. Such a concept is as given in the following definition.
Definition 7 [
2]
. Let be a universal set. Let .
Let be a set of attributes. For each attribute :
Let be the set of all its corresponding attribute values. Take .
Define a function ,
called the attribute value appurtenance degree function. Define a function ,
called the attribute value contradiction (dissimilarity) degree function, which further satisfies:- (i)
, for all.
- (ii)
, for all.
Then:
- (a)
is said to form a plithogenic set on.
- (b)
is said to be the attribute value appurtenance fuzzy degree function (abbr. AFD-function) forin, andis called the appurtenance fuzzy degree ofin.
- (c)
is said to be the attribute value contradiction (dissimilarity) fuzzy degree function (abbr. CFD-function) forin, andis called the contradiction fuzzy degree betweenand.
Remark 4. If, , , and, thenis reduced to a single valued neutrosophic set (SVNS) on
Remark 5. If, for all, is such thatfor alland for all, for all, thenis reduced to a generalized intuitionistic fuzzy soft set (GIFSS) on
Remark 6. If, , , is such that, andall satisfied for all, and, thenis reduced to an inter-valued intuitionistic fuzzy set (IVIFS) on
Remark 7. If, , , is such thatfor all, and, thenis reduced to an intuitionistic fuzzy set (IFS) on
Remark 8. If, , , is such thatfor all, and, thenis reduced to a Pythagorean fuzzy set (PFS) on
Remark 9. If, and, thenis reduced to a fuzzy set on
Remark 10. If, , , and, thenis reduced to a classical crisp set on
In all the following, the collection of all the plithogenic sets on shall be denoted as .
Definition 8. Let ∈
.
The compliment for ,
is defined as where for all .
Remark 11. This definition of compliment follows from page 42 of [
2].
Remark 12. It is clear thatas well.
With the all these definitions established, we now proceed to define a way of measurement of entropy for plithogenic sets. In the establishment of such entropy measures, we must let all the AFD-functions and all the CFD-functions to participate in contributing to the overall entropy measures of .
We now discuss some common traits of how each element from and shall contribute to the overall entropy measures of , all of which are firmly rooted in our conventional understanding of entropy as a quantitative measurement for the amount of disorder.
Firstly, on the elements of : In accordance with Definition 7, each is the appurtenance fuzzy degree of , over the attribute value ( in turn belongs to the attribute ). Note that indicates absolute certainty of membership of in ; whereas indicates absolute certainty of non-membership of in . Hence, any satisfying must be regarded as contributing zero magnitude to the overall entropy measure of , as absolute certainty implies zero amount of disorder. On the other hand, indicates total uncertainty of the membership of in , as 0.5 is in the middle of 0 and 1. Hence, any satisfying must be regarded as contributing the greatest magnitude to the overall entropy measure of , as total uncertainty implies the highest possible amount of disorder.
Secondly, on the elements of : For each attribute , indicates that the attribute values are of identical meaning (synonyms) with each other (e.g., “big” and “large”), whereas indicates that the attribute values are of opposite meaning to each other (e.g., “big” and “small”). Therefore, in the case of and , it implies that is absolutely certain to be inside one among , while outside of the other, even though and carry identical meaning to each other. Such collection of is therefore of the highest possible amount of disorder, because their combined meaning implies an analogy to the statement of “ is very large and not big” or “ is not large and very big”. As a result, such collection of aforementioned must be regarded as contributing the greatest magnitude to the overall entropy measure of . Furthermore, in the case of and , it implies that is absolutely certain to be inside both and (or outside both and ), even though and carry opposite meaning with each other. Likewise, such collection of is of the highest possible amount of disorder, because their combined meaning implies an analogy to the statement of “ something is very big and very small” or “ something is not big and not small”. As a result, such a collection of aforementioned must be regarded as contributing the greatest magnitude to the overall entropy measure of as well.
We now define the three axioms of entropy on plithogenic sets, analogous to the axioms (i), (ii), and (iv) in Definition 2 respectively.
Definition 9. An entropy measure on plithogenic sets, is a functionsatisfying the following three axioms
- (i)
(analogy to (i) in Definition 2): Letsatisfying the following conditions for all:
- (a)
for all.
- (b)
whenever.
- (c)
whenever.
Then.
- (ii)
(analogy to (ii) in Definition 2). Letsatisfyingfor all. Then.
- (iii)
(analogy to (iv) in Definition 2). For all, holds.
The three axioms in Definition 9 thus serve as general rules for which any functions must fulfill to be used as entropy measures on plithogenic sets. However, the existence of such functions satisfying these three axioms needs to be ascertained. To ensure that we do have an abundance of functions satisfying these axioms, we must therefore propose and give characterization to such functions with explicit examples and go to the extent of proving that each one among our proposed examples satisfy all these axioms. Such a procedure of proving the existence of many different entropy functions is indispensable. This is because in practical use, the choices of an entropy measure will fully depend on the type of scenario examined, as well as the amount of computing power available to perform such computations, without jeopardizing the axioms of entropy measures as mentioned. It is only by doing so that users are guaranteed to have plenty of room to customize an entropy measure of plithogenic sets suited for their particular needs. In light of this motivation, a theorem showing a collection of functions satisfying those axioms is presented in this paper.
Theorem 1. Let, , be any M-type functions. Let, be any S-type functions. Letbe any function satisfying the following conditions:
- (i)
, , .
- (ii)
is increasing within [0, 0.5]. In other words,whenever.
Letbe any function satisfying the following conditions:
- (i)
for all [0, 0.5], .
- (ii)
is increasing within [0.5, 1]. In other words,whenever.
Define,
wherefor all.
Define,
where:for all.
Then, any function,
in the form offor all,
are all entropy measures on plithogenic sets.
Proof. + Axiom (i): Taking any arbitrary , and .
As , .
Whenever , it follows that , which implies
.
Thus, .
Since , follows, which further implies that .
- c.
whenever , it implies .
Thus, .
Since , follows, which further implies that .
Hence, follows for all .
+ Axiom (ii): Taking any arbitrary , and .
As for all , we have
for all . This further implies that
for all .
+ Axiom (iii): follows by Definition 8. This will imply the following
- (a)
.
- (b)
Therefore, it follows that
now follows. □
Remark 13. Asand Such a version of the formula serves as an even more explicit representation of.
Remark 14. For instance, the following is one of the many theoretical ways of choosingto form a particular entropy measure on plithogenic sets.
- (a)
, for all .
- (b)
, for all.
- (c)
, for all.
- (d)
, for all.
- (e)
, for all.
- (f)
, for all.
- (g)
, for all.
In practical applications, however, the choices of will depend on the type of scenario examined, as well as the amount of computing power available to perform such computations. Such abundance of choices is a huge advantage, because it allows each user plenty of room of customization suited for their own needs, without jeopardizing the principles of entropy functions.
4. Numerical Example of Plithogenic Sets
In this section, we demonstrate the utility of the proposed entropy functions for plithogenic sets using an illustrative example of a MADM problem involving a property buyer making a decision whether to live in Town or Town .
4.1. Attributes and Attributes Values
Three different addresses within Town
are selected:
. Another four different addresses within Town
are selected as well:
. All the seven addresses are investigated by that person based on 3 attributes as follows:
For each of the 3 attributes, the following attribute values are considered:
4.2. Attribute Value Appurtenance Degree Functions
In light of the limitation of one person doing the investigation, there could possibly be some characteristics of Town left unknown or unsure of. As a result, our example involved in this paper, though small in caliber, shall provide a realistic illustration of such phenomena.
Thus, in our example: Let the attribute value appurtenance degree functions for Town
be given in
Table 1,
Table 2 and
Table 3 (as deduced by the property buyer).
For example:
indicates that schools exist near address in town .
indicates that Grab services are very likely to exist near address in town .
indicates that police stations exist near address in town .
Similarly, let the attribute value appurtenance degree functions for Town
be given in
Table 4,
Table 5 and
Table 6 (as deduced by the property buyer):
4.3. Attribute Value Contradiction Degree Functions
Moreover, each of the attributes of a town may be dependent on one another. For example, in a place where schools are built, clinics should be built near to the schools, whereas factories should be built far from the schools. Moreover, the police force should spread their manpower patrolling across the town away from a police station. As a result, our example involved in this paper, though small in caliber, shall provide a realistic illustration of such phenomena as well.
Thus, as an example, let the attribute value contradiction degree functions for the attributes
be given in
Table 7,
Table 8 and
Table 9: (as deduced by the property buyer), to be used for both towns.
In particular,
indicates that schools and factories should not be in the same place, because it is not healthy to the students.
indicates that schools and clinics should be available together, so that any student who falls ill can visit the clinic.
, because it is very inefficient for police to patrol only nearby a police station itself, instead of places of a significant distance to a police station. This also ensures that police force will be present in all places, as either a station or a patrol unit will be present.
, because all train stations must have buses going to/from it. On the other hand, one must also be able to reach a train station from riding a bus.
due to the conflicting nature of the two businesses.
4.4. Two Plithogenic Sets Representing Two Towns
From all attributes of the two towns given, we thus form two plithogenic sets representing each of them
- (a)
, which describes Town
- (b)
, which describes Town
Intuitively, it is therefore evident that the property buyer should choose Town over Town as his living place. One of the many reasons being, in Town schools and factories are unlikely to appear near one address within the town, whereas in Town there exist addresses where schools and factories are both nearby (so both schools and factories are near to each other). Moreover, in Town , the police force is more efficient as they spread their manpower across the town, rather than merely patrolling near their stations and even leaving some addresses unguarded as in Town . On top of this, there exists places in town where taxi and grab services are near to each other, which can cause conflict or possibly vandalism to each other’s property. Town is thus deemed less “chaotic”, whereas Town is deemed more “chaotic”.
As a result, our entropy measure must be able to give Town as having lower entropy than Town , under certain choices of which are customized for the particular use of the property buyer.
4.5. An Example of Entropy Measure on Two Towns
Choose the following to form in accordance with Theorem 1:
- (a)
, for all .
- (b)
, for all .
- (c)
, for all .
- (d)
, for all .
Then, by the calculation in accordance with Theorem 1 which is subsequently highlighted in
Figure 1.
We have , and .
Town is concluded to have lower entropy, and, therefore, is less “chaotic”, compared to Town .