
The information measure of interval-valued intuitionistic fuzzy sets holds extensive application value in decision-making and pattern recognition. The primary contribution of this paper lied in presenting a functional model for assessing the information content of interval-valued intuitionistic fuzzy sets. Initially, we introduced the concept of closest crisp set associated with interval-valued intuitionistic fuzzy set and explored its pertinent properties. Subsequently, taking into account the distance between interval-valued intuitionistic fuzzy set and its closest crisp set, we derived a comprehensive expression for functions pertaining to similarity measures, knowledge measures and entropy on interval-valued intuitionistic fuzzy set that adhered to specific criteria. Furthermore, we explored the interconversion between these three measures. The advantage of these functional expressions and transformation relationships lied in their ability to generate numerous formulas for defining information measures. Finally, we demonstrated the practical application of knowledge measure in investment case. The practicality of the proposed measures were corroborated through sensitivity analysis and comparative analysis.
Citation: Le Fu, Jingxuan Chen, Xuanchen Li, Chunfeng Suo. Novel information measures considering the closest crisp set on fuzzy multi-attribute decision making[J]. AIMS Mathematics, 2025, 10(2): 2974-2997. doi: 10.3934/math.2025138
[1] | Chunfeng Suo, Yan Wang, Dan Mou . The new construction of knowledge measure on intuitionistic fuzzy sets and interval-valued intuitionistic fuzzy sets. AIMS Mathematics, 2023, 8(11): 27113-27127. doi: 10.3934/math.20231387 |
[2] | Li Li, Xin Wang . Hamming distance-based knowledge measure and entropy for interval-valued Pythagorean fuzzy sets. AIMS Mathematics, 2025, 10(4): 8707-8720. doi: 10.3934/math.2025399 |
[3] | Muhammad Riaz, Maryam Saba, Muhammad Abdullah Khokhar, Muhammad Aslam . Novel concepts of $ m $-polar spherical fuzzy sets and new correlation measures with application to pattern recognition and medical diagnosis. AIMS Mathematics, 2021, 6(10): 11346-11379. doi: 10.3934/math.2021659 |
[4] | Li Li, Mengjing Hao . Interval-valued Pythagorean fuzzy entropy and its application to multi-criterion group decision-making. AIMS Mathematics, 2024, 9(5): 12511-12528. doi: 10.3934/math.2024612 |
[5] | T. M. Athira, Sunil Jacob John, Harish Garg . A novel entropy measure of Pythagorean fuzzy soft sets. AIMS Mathematics, 2020, 5(2): 1050-1061. doi: 10.3934/math.2020073 |
[6] | Atiqe Ur Rahman, Muhammad Saeed, Hamiden Abd El-Wahed Khalifa, Walaa Abdullah Afifi . Decision making algorithmic techniques based on aggregation operations and similarity measures of possibility intuitionistic fuzzy hypersoft sets. AIMS Mathematics, 2022, 7(3): 3866-3895. doi: 10.3934/math.2022214 |
[7] | Changlin Xu, Yaqing Wen . New measure of circular intuitionistic fuzzy sets and its application in decision making. AIMS Mathematics, 2023, 8(10): 24053-24074. doi: 10.3934/math.20231226 |
[8] | Shichao Li, Zeeshan Ali, Peide Liu . Prioritized Hamy mean operators based on Dombi t-norm and t-conorm for the complex interval-valued Atanassov-Intuitionistic fuzzy sets and their applications in strategic decision-making problems. AIMS Mathematics, 2025, 10(3): 6589-6635. doi: 10.3934/math.2025302 |
[9] | Muhammad Arshad, Muhammad Saeed, Khuram Ali Khan, Nehad Ali Shah, Wajaree Weera, Jae Dong Chung . A robust MADM-approach to recruitment-based pattern recognition by using similarity measures of interval-valued fuzzy hypersoft set. AIMS Mathematics, 2023, 8(5): 12321-12341. doi: 10.3934/math.2023620 |
[10] | Ya Qin, Siti Rahayu Mohd. Hashim, Jumat Sulaiman . A new distance measure and corresponding TOPSIS method for interval-valued intuitionistic fuzzy sets in multi-attribute decision-making. AIMS Mathematics, 2023, 8(11): 26459-26483. doi: 10.3934/math.20231351 |
The information measure of interval-valued intuitionistic fuzzy sets holds extensive application value in decision-making and pattern recognition. The primary contribution of this paper lied in presenting a functional model for assessing the information content of interval-valued intuitionistic fuzzy sets. Initially, we introduced the concept of closest crisp set associated with interval-valued intuitionistic fuzzy set and explored its pertinent properties. Subsequently, taking into account the distance between interval-valued intuitionistic fuzzy set and its closest crisp set, we derived a comprehensive expression for functions pertaining to similarity measures, knowledge measures and entropy on interval-valued intuitionistic fuzzy set that adhered to specific criteria. Furthermore, we explored the interconversion between these three measures. The advantage of these functional expressions and transformation relationships lied in their ability to generate numerous formulas for defining information measures. Finally, we demonstrated the practical application of knowledge measure in investment case. The practicality of the proposed measures were corroborated through sensitivity analysis and comparative analysis.
Since Zadeh introduced fuzzy set theory, extensive research and extension related to fuzzy sets have been pursued. Among the various generalized forms that have been proposed, Atanassov [1] introduced the definition of intuitionistic fuzzy sets (IFSs). This was a broader concept of fuzzy sets designed to address uncertain information. It offered a more precise characterization of such information through the triad of degree of membership, degree of nonmembership, and degree of hesitation. As an extension of IFS, Atanassov [2] presented an alternative definition of interval-valued intuitionistic fuzzy sets (IVIFSs), wherein the degree of membership, nonmembership, and hesitation are represented by subintervals within the range of [0,1]. As a result, it can precisely capture the dynamic features. Furthermore, Xu [3] examined certain attributes and average operators of IVIFSs. Currently, IVIFSs have been extensively utilized in diverse areas, including quality evaluation, venture capital, and medical diagnosis [4,5,6,7], owing to their remarkable flexibility in managing uncertainty or ambiguity.
Information measures pertaining to IFSs and IVIFSs have garnered escalating attention due to their exceptional performance in addressing practical applications. Interested readers can to [8,9,10]. Axiomatic definitions of distance measures on IFSs and IVIFSs were respectively provided in [11,12]. Furthermore, a novel distance measure for IFSs was introduced through the utilization of line integrals [13]. Drawing from the relationship between distance measure and similarity measure, similarity measure has also been paid attention to by a considerable number of researchers, and numerous scholars investigated the similarity measure for IVIFSs on the basis of distance measures [14,15,16,29,30,31,32]. Fuzzy entropy was initially introduced by Zadeh, and in recent years, there has been a growing interest among scholars in exploring this concept. Interested readers are encouraged to consult references [17,18,19] for further details. An entropy on IVIFSs constructed based on a distance measure was proposed by Zhang et al. [20], and the relationship between similarity measure and entropy was investigated in the paper. Che et al. [21] presented a method for constructing an entropy on IVIFSs in terms of the distance function. The cross-entropy approach, originating from the information theory presented by Shannon [22], was later applied to decision-making problems by Ye [23] who developed the fuzzy cross-entropy of IVIFSs using an analogy with the intuitionistic fuzzy cross-entropy. In recent years, there has been a notable surge in the interest surrounding knowledge measures within the frameworks of IFSs and IVIFSs, as clearly demonstrated by the references [24,25,26]. Das et al. [27] employed the knowledge measures under IFSs and IVIFSs to compute the criterion weights in the decision problem, and Guo [28] considered both aspects of knowledge related to IVIFSs, i.e., information content and information clarity to construct a model of knowledge measure and apply it to a decision-making problem under unknown weights. In addition, some scholars have also extended IVIFS [33,34].
The structure of this paper is as follows: Section 2 delves into several definitions and axioms that will be utilized in the subsequent research outlined in this paper. Section 3 centers on the construction of similarity measures, knowledge measures, and entropy, all grounded in the axiomatic definition of the closest crisp set and the distance measure. Furthermore, it explores the intricate transformation relationships among these three constructs. Subsequently, we present a range of formulas for information measures, each utilizing distinct functions. In Section 4, we demonstrate the efficacy and versatility of our newly introduced methodologies by applying the proposed knowledge measure to tackle a decision-making problem and the similarity measure to address a pattern recognition problem. Lastly, in Section 5, we offer some concluding remarks on the contents of this paper.
In this section, we will briefly introduce some of fuzzy set theories and concepts which should be used in this paper. The X in the following text is represented as: X=(x1,x2,⋯,xn). Let N denote the set of natural numbers.
Definition 2.1. ( [2]) An IVIFS ˜A on a finite set X can be defined as the following form:
˜A={<x,μ˜A(x),ν˜A(x)>|x∈X}, |
where μ˜A(x)=[μL˜A(x),μU˜A(x)]⊆[0,1] and ν˜A(x)=[νL˜A(x),νU˜A(x)]⊆[0,1], which satisfies 0≤μU˜A(x)+νU˜A(x)≤1, for any x∈X. For the interval hesitation margin π˜A(x)=[πL˜A(x),πU˜A(x)]⊆[0,1], we have πL˜A(x)=1−μU˜A(x)−νU˜A(x) and πU˜A(x)=1−μL˜A(x)−νL˜A(x).
Specifically, if μL˜A(x)=μU˜A(x) and νL˜A(x)=νU˜A(x), then the IVIFS is reduced to an intuitionistic fuzzy set.
From the interval representation of membership degree and nonmembership degree of IVIFSs, it is not difficult to see that the geometric meaning of IVIFSs at a certain point is a rectangle, as shown in Figure 1. The three vertices of the triangle are denoted as (0,0),(1,0),(0,1), and the two sides of the triangle represent the degree of membership μ˜A(x) and nonmembership ν˜A(x), so that (0,0) expresses the case if π˜A(x)=1. (1,0) and (0,1) mean the two cases of the crisp sets. For the hypotenuse of the triangle, that is, the segment from (1,0) to (0,1), it corresponds to ν˜A(x)=1−μ˜A(x). Moreover, there is 0≤μ˜A(x)+ν˜A(x)≤1 for an IVIFS, which also means that none of the individual vertices of the rectangle in Figure 1 can extend beyond the hypotenuse of the triangle.
For convenience, let IVIFS(X) express the set of IVIFSs on X. Besides, we use P(X) to denote the set of all crisp sets defined on X.
Definition 2.2. ( [3]) For any ˜A,˜B∈IVIFS(X), the following relations can be defined:
(1) ˜A⊆˜B if μL˜A(x)≤μL˜B(x), μU˜A(x)≤μU˜B(x) and νL˜A(x)≥νL˜B(x), νU˜A(x)≥νU˜B(x) for any x∈X;
(2) ˜A=˜B if ˜A⊆˜B and ˜A⊇˜B;
(3) The complement set ˜Ac={⟨x,ν˜A(x),μ˜A(x)⟩|x∈X}.
Definition 2.3. ( [12]) For any ˜A,˜B,˜C∈IVIFS(X), a mapping D:IVIFS(X)×IVIFS(X)→[0,1] is called a distance measure on IVIFSs, and it satisfies the following properties:
(D1)0≤D(˜A,˜B)≤1;
(D2)D(˜A,˜B)=0 if ˜A=˜B;
(D3)D(˜A,˜B)=D(˜B,˜A);
(D4)D(˜A,˜C)≤D(˜A,˜B)+D(˜B,˜C);
(D5)If˜A⊆˜B⊆˜C, then D(˜A,˜C)≥max{D(˜A,˜B),D(˜B,˜C)}.
Consider ˜A,˜B∈IVIFS(X), where ˜A={⟨x,[μL˜A(x),μU˜A(x)],[νL˜A(x),νU˜A(x)]⟩|x∈X}, ˜B={⟨x,[μL˜B(x),μU˜B(x)],[νL˜B(x),νU˜B(x)]⟩|x∈X}.
For convenience, we denote:
|μL˜A(xi)−μL˜B(xi)|=αi,|μU˜A(xi)−μU˜B(xi)|=βi,|νL˜A(xi)−νL˜B(xi)|=γi,|νU˜A(xi)−νU˜B(xi)|=δi. |
For arbitrary p, the distance measure between ˜A and ˜B is presented in the following form:
Dp(˜A,˜B)=(1qnn∑i=1(αpi∗βpi⋆γpi∗δpi))1p, | (2.1) |
where the parameter q is determined by the two operations "∗" and "⋆".
When p assumes different values, various distance measures can be obtained, as demonstrated in Table 1.
∗=+ ⋆=+ |
Minkovsky | q=4 | p=1 | DH(˜A,˜B)=(14n∑ni=1(αi+βi+γi+δi)) |
p=2 | DE(˜A,˜B)=(14n∑ni=1(α2i+β2i+γ2i+δ2i))12 | |||
p=3 | DM(˜A,˜B)=(14n∑ni=1(α3i+β3i+γ3i+δ3i))13 | |||
∗=∨,⋆=+ ∗=∨,⋆=∨ |
p=1 | q=2 | Hamming-Hausdorff | DHH(˜A,˜B)=(12n∑ni=1(αi∨βi+γi∨δi)) |
q=1 | Hausdorff | Dh(˜A,˜B)=(1n∑ni=1(αi∨βi∨γi∨δi)) |
Definition 2.4. ( [28]) Let ˜A,˜B∈IVIFS(X), and a mapping K:IVIFS(X)→[0,1] is called a knowledge measure on IVIFS(X) if K satisfies the following conditions:
(K1)K(˜A)=1 if ˜A∈P(X);
(K2)K(˜A)=0 if π˜A(x)=[1,1] for any x∈X;
(K3)K(˜A)≥K(˜B) if ˜A is less fuzzy than ˜B, i.e., ˜A⊆˜B for μL˜B(x)≤νL˜B(x) and μU˜B(x)≤νU˜B(x) or ˜A⊇˜B for μL˜B(x)≥νL˜B(x) and μU˜B(x)≥νU˜B(x), for any x∈X;
(K4)K(˜A)=K(˜Ac).
Definition 2.5. ( [18]) Assume that the entropy is a function E:IVIFS(X)→[0,1], which satisfies the nether four conditions:
(E1)E(˜A)=0 if ˜A∈P(X);
(E2)E(˜A)=1 if μ˜A(x)=ν˜A(x) for any x∈X;
(E3)E(˜A)≤E(˜B) if ˜A is less fuzzy than ˜B, i.e., μ˜A(x)≤μ˜B(x) and ν˜A(x)≥ν˜B(x) for μ˜B(x)≤ν˜B(x) or μ˜A(x)≥μ˜B(x) and ν˜A(x)≤ν˜B(x) for μ˜B(x)≥ν˜B(x);
(E4)E(˜A)=E(˜Ac).
Furthermore, this paper introduces the definition of the closest crisp set within the realm of IVIFSs and derives several equations pertaining to similarity measure, knowledge measure, and entropy, all grounded in the distance measure. Additionally, it delves into the methodology for converting between these measures.
The extent of resemblance between two comparable entities is frequently characterized or elucidated through the utilization of similarity measure. Within this segment, we primarily introduce a novel formulation of similarity measure pertaining to IVIFSs, which relies on distance measure and the proximate crisp set denoted as C˜A.
To this end, inspired by the concept of the closest crisp set considering intuitionistic fuzzy sets in [17], the notion of closest crisp set for IVIFSs is introduced.
Definition 3.1. Let ˜A∈IVIFS(X), and the closest crisp set to ˜A, which is denoted by C˜A, satisfies the following condition: If μ˜A(x)≥ν˜A(x), then x∈C˜A; if μ˜A(x)≱, then x\notin C_{\tilde{A}} .
Since any crisp set is an IVIFS with zero interval hesitation index, the closest crisp set of \tilde{A} can be expressed as:
\begin{eqnarray} \mu_{C_{\tilde{A}}}^{L}(x) = \mu_{C_{\tilde{A}}}^{U}(x) = \begin{cases} 1\; \; \; \; \; \; \; \; \; \; \; \; \; {\rm{if}}\; \mu_{\tilde{A}}^{L}(x)\geq\nu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x)\geq\nu_{\tilde{A}}^{U}(x), \\ 0\; \; \; \; \; \; \; \; \; \; \; \; \; {\rm otherwise}, \end{cases} \end{eqnarray} | (3.1) |
\begin{eqnarray} \nu_{C_{\tilde{A}}}^{L}(x) = \nu_{C_{\tilde{A}}}^{U}(x) = \begin{cases} 0\; \; \; \; \; \; \; \; \; \; \; \; \; {\rm if}\; \mu_{\tilde{A}}^{L}(x)\geq\nu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x)\geq\nu_{\tilde{A}}^{U}(x), \\ 1\; \; \; \; \; \; \; \; \; \; \; \; \; {\rm otherwise}. \end{cases} \end{eqnarray} | (3.2) |
From Figure 2, it can be seen that in the left panel, \mu_{\tilde{A}}^{L}(x) > \nu_{\tilde{A}}^{L}(x), \; \mu_{\tilde{A}}^{U}(x) > \nu_{\tilde{A}}^{U}(x) , so that there is x\in C_{\tilde{A}} , or \mu_{C_{\tilde{A}}}^{L}(x) = \mu_{C_{\tilde{A}}}^{U}(x) = 1 , \nu_{C_{\tilde{A}}}^{L}(x) = \nu_{C_{\tilde{A}}}^{U}(x) = 0 . The opposite is true for the right panel, \mu_{\tilde{A}}^{L}(x) < \nu_{\tilde{A}}^{L}(x), \; \mu_{\tilde{A}}^{U}(x) < \nu_{\tilde{A}}^{U}(x) ; thus, there is x\notin C_{\tilde{A}} , or \mu_{C_{\tilde{A}}}^{L}(x) = \mu_{C_{\tilde{A}}}^{U}(x) = 0 , \nu_{C_{\tilde{A}}}^{L}(x) = \nu_{C_{\tilde{A}}}^{U}(x) = 1 .
Next, a corollary is given below that presents practical properties of the closest crisp set.
Corollary 3.2. Assume that \tilde{A}\in IVIFS(X) , and let C_{\tilde{A}} be its closest crisp set. It has the following properties:
(1) \tilde{A}\in \mathcal{P}(X) if, and only if, \tilde{A} = C_{\tilde{A}} .
(2) If \mu_{\tilde{A}}^{L}(x)\neq\nu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x)\neq\nu_{\tilde{A}}^{U}(x) , then C^{c}_{\tilde{A}}(x) = C_{\tilde{A}^{c}}(x) for any x\in X , where C_{\tilde{A}^{c}} expresses the closest crisp set to A^{c} .
The preceding Corollary 3.2(2) can be elucidated more intuitively by scrutinizing Figure 3, which illustrates that the rectangular positions designated by \tilde{A}(x) and \tilde{A^{c}}(x) are precisely as depicted in the image when the conditions \mu_{\tilde{A}}^{L}(x)\geq\nu_{\tilde{A}}^{L}(x) and \mu_{\tilde{A}}^{U}(x)\geq\nu_{\tilde{A}}^{U}(x) are satisfied. Furthermore, it becomes evident that C_{\tilde{A}}(x) and C_{\tilde{A}^{c}}(x) occupy opposing corners, thereby establishing the equality C^{c}_{\tilde{A}}(x) = C_{\tilde{A}^{c}}(x) . Note here that the conditional requirement of Corollary 3.2(2) is \mu_{\tilde{A}}^{L}(x)\neq\nu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x)\neq\nu_{\tilde{A}}^{U}(x) , the reason for which is that if \mu_{\tilde{A}}^{L}(x) = \nu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x) = \nu_{\tilde{A}}^{U}(x) , as shown in Figure 4, then \tilde{A}(x) and \tilde{A}^{c}(x) are coincident and their closest crisp sets are the same.
Based on the proposed the closest crisp set, information measure functions are explored.
Definition 3.3. A function f:[0, 1]^{2}\rightarrow [0, 1] is called a binary aggregation function, for any x, y\in [0, 1] , it satisfies the following conditions:
(1) f is component-wise increasing;
(2) f(0, 0) = 0 ;
(3) f(1, 1) = 1 ;
(4) f(x, y) = f(y, x) .
Definition 3.4. ( [9]) Let \tilde{A}, \tilde{B}\in IVIFS(X) , and a mapping S : IVIFS(X)\times IVIFS(X)\rightarrow [0, 1] is called a similarity measure on IVIFSs if S satisfies the following properties:
(S1) Boundary: 0\leq S(\tilde{A}, \tilde{B})\leq 1 ;
(S2) Symmetry: S(\tilde{A}, \tilde{B}) = S(\tilde{B}, \tilde{A}) ;
(S3) Reflexivity: S(\tilde{A}, \tilde{B}) = 1 iff \tilde{A} = \tilde{B} ;
(S4) Complementarity: S(\tilde{A}, \tilde{A}^{c}) = 0 iff \tilde{A}\in \mathcal{P}(X) .
Theorem 3.5. Let f:[0, 1]\times[0, 1]\rightarrow [0, 1] be a binary aggregation function and D be an aforementioned distance in Definition 2.3 for IVIFSs. Then, for any \tilde{A}, \tilde{B}\in IVIFS(X) ,
S(\tilde{A}, \tilde{B}) = f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{B}, C_{\tilde{B}})) |
is a similarity measure for IVIFSs.
Proof. In the following, we will prove the proposed formula satisfies four conditions in Definition 3.4.
(S1) : It is clearly established.
(S2) : By the definition of function f , it can be clearly deduced that
S(\tilde{A}, \tilde{B}) = f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{B}, C_{\tilde{B}})) = f(D(\tilde{B}, C_{\tilde{B}}), D(\tilde{A}, C_{\tilde{A}})) = S(\tilde{B}, \tilde{A}). |
(S3) : If S(\tilde{A}, \tilde{B}) = 1 , according to the definition of f and the distance measure, we can get
f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{B}, C_{\tilde{B}})) = 1\Leftrightarrow D(\tilde{A}, C_{\tilde{A}}) = D(\tilde{B}, C_{\tilde{B}}) = 1, |
thus, \tilde{A} = \tilde{B} .
(S4) : If S(\tilde{A}, \tilde{A}^{c}) = 0 , based on the definition of f and the distance measure,
f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{A}^{c}, C_{\tilde{A}^{c}})) = 0\Leftrightarrow D(\tilde{A}, C_{\tilde{A}}) = D(\tilde{A}^{c}, C_{\tilde{A}^{c}}) = 0 |
can be deduced. Then, \tilde{A}\in \mathcal{P}(X) .
With regards to Theorem 3.5, it is feasible to devise diverse formulations for calculating similarity measures on IVIFSs, leveraging various binary aggregation functions f:[0, 1]\times[0, 1]\rightarrow [0, 1] . By incorporating distinct functions f and distance measures D , a range of similarity measures can be derived, as exemplified in Table 2.
Distance measures | Conversion functions | Similarity measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{H1}(\tilde{A}, \tilde{B})= \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{H2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{H3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{H4}(\tilde{A}, \tilde{B})=\log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{H5}(\tilde{A}, \tilde{B})=2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{E1}(\tilde{A}, \tilde{B})= \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{E2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{E3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{E4}(\tilde{A}, \tilde{B})=\log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{E5}(\tilde{A}, \tilde{B})=2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{M1}(\tilde{A}, \tilde{B})= \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{M2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{M3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{M4}(\tilde{A}, \tilde{B})=\log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{M5}(\tilde{A}, \tilde{B})=2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{HH1}(\tilde{A}, \tilde{B})= \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{HH4}(\tilde{A}, \tilde{B})=\log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{HH5}(\tilde{A}, \tilde{B})=2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{h1}(\tilde{A}, \tilde{B})= \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{h2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{h3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{h4}(\tilde{A}, \tilde{B})=\log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{h5}(\tilde{A}, \tilde{B})=2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{B}, C_{\tilde{B}})}-1 |
The knowledge content of IVIFSs is typically characterized through knowledge measures, with entropy serving as a pivotal quantifying the uncertainty associated with fuzzy variables. In this section, we established an approach of developing similarity measures to devise knowledge measures and entropy for IVIFSs.
Theorem 3.6. Let f be a function defined in Definition 3.3 and D be an aforementioned distance in Definition 2.3 for IVIFSs. For any \tilde{A}\in IVIFS(X) ,
K(\tilde{A}) = 1-f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{A^{c}}, C_{\tilde{A}^{c}})) |
is a knowledge measure on IVIFS(X) .
Proof. In the following we will show that the knowledge measure we construct satisfies the four conditions in Definition 2.4.
(KD1) : Let K(\tilde{A}) = 1 , then consider the definition of the function f , that is, D(\tilde{A}, C_{\tilde{A}}) = D(\tilde{A^{c}}, C_{\tilde{A}^{c}}) = 0 . Therefore, by Definition 2.3 and Corollary 3.2(1), it is easily obtained that \tilde{A}\in \mathcal{P}(X) .
(KD2) : It is not difficult to see that if K(\tilde{A}) = 0 , then D(\tilde{A}, C_{\tilde{A}}) = D(\tilde{A^{c}}, C_{\tilde{A}^{c}}) = 1 , and according to the definition of the distance D ,
\mu_{\tilde{A}}^{L}(x) = \mu_{\tilde{A}}^{U}(x) = \nu_{\tilde{A}}^{L}(x) = \nu_{\tilde{A}}^{U}(x) = 0, |
can be obtained. Thus, \pi_{\tilde{A}}(x) = [1, 1] .
(KD3) : Assume that \tilde{A}\subseteq\tilde{B} for \mu_{\tilde{B}}^{L}(x)\leq\nu_{\tilde{B}}^{L}(x) and \mu_{\tilde{B}}^{U}(x)\leq\nu_{\tilde{B}}^{U}(x) . It is certain that
\mu_{\tilde{A}}^{L}(x)\leq\mu_{\tilde{B}}^{L}(x)\leq\nu_{\tilde{B}}^{L}(x)\leq\nu_{\tilde{A}}^{L}(x), \; \mu_{\tilde{A}}^{U}(x)\leq\mu_{\tilde{B}}^{U}(x)\leq\nu_{\tilde{B}}^{U}(x)\leq\nu_{\tilde{A}}^{U}(x), |
and from the notion of the closest crisp set, it can be derived that
\mu_{C_{\tilde{A}}}^{L}(x) = \mu_{C_{\tilde{A}}}^{U}(x) = \mu_{C_{\tilde{B}}}^{L}(x) = \mu_{C_{\tilde{B}}}^{U}(x) = 0, |
\nu_{C_{\tilde{A}}}^{L}(x) = \nu_{C_{\tilde{A}}}^{U}(x) = \nu_{C_{\tilde{B}}}^{L}(x) = \nu_{C_{\tilde{B}}}^{U}(x) = 1. |
Therefore, let \tilde{N} = < [0, 0], [1, 1] > , and if \tilde{N}\subseteq \tilde{A}\subseteq \tilde{B} , then
\begin{equation} \begin{split} \notag D(\tilde{A}, \tilde{N})& = D([\mu_{\tilde{A}}^{L}(x), \mu_{\tilde{A}}^{U}(x)], [\nu_{\tilde{A}}^{L}(x), \nu_{\tilde{A}}^{U}(x)], [0, 0], [1, 1])\\ &\leq D(\tilde{B}, \tilde{N}) = D([\mu_{\tilde{B}}^{L}(x), \mu_{\tilde{B}}^{U}(x)], [\nu_{\tilde{B}}^{L}(x), \nu_{\tilde{B}}^{U}(x)], [0, 0], [1, 1]). \end{split} \end{equation} |
Similarly, it is obvious that D(\tilde{A^{c}}, C_{\tilde{A}^{c}})\leq D(\tilde{B^{c}}, C_{\tilde{B}^{c}}) . Then, by considering the function f to be component-wise increasing, there is
f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{A^{c}}, C_{\tilde{A}^{c}}))\leq f(D(\tilde{B}, C_{\tilde{B}}), D(\tilde{B^{c}}, C_{\tilde{B}^{c}})). |
Hence, it is easy to know that K(\tilde{A})\geq K(\tilde{B}). On the other hand, assume that \tilde{A}\supseteq\tilde{B} for \mu_{\tilde{B}}^{L}(x)\geq\nu_{\tilde{B}}^{L}(x) and \mu_{\tilde{B}}^{U}(x)\geq\nu_{\tilde{B}}^{U}(x) . K(\tilde{A})\geq K(\tilde{B}) can also be obtained.
(KD4) : In order to verify K(\tilde{A}) = K(\tilde{A}^{c}) , simply prove that
f(D(\tilde{A}, C_{\tilde{A}}), D(\tilde{A^{c}}, C_{\tilde{A}^{c}})) = f(D(\tilde{A^{c}}, C_{\tilde{A}^{c}}), D(\tilde{A}, C_{\tilde{A}})), |
which is obvious from the definition of the function f .
Next, a series of knowledge measures can be obtained by using different distance measures and different expressions of function f . They are shown in Table 3.
Distance measures | Conversion functions | Knowledge measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{H1}(\tilde{A})=1- \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{H2}(\tilde{A})=1- \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{H3}(\tilde{A})=1- \sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{H4}(\tilde{A})=1- \log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{H5}(\tilde{A})=2- 2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{E1}(\tilde{A})=1- \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{E2}(\tilde{A})=1- \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{E3}(\tilde{A})=1- \sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{E4}(\tilde{A})=1- \log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{E5}(\tilde{A})=2- 2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{M1}(\tilde{A})=1- \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{M2}(\tilde{A})=1- \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{M3}(\tilde{A})=1- \sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{M4}(\tilde{A})=1- \log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{M5}(\tilde{A})=2- 2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{HH1}(\tilde{A})=1- \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{HH4}(\tilde{A})=1- \log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{HH5}(\tilde{A})=2- 2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{h1}(\tilde{A})=1- \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{h2}(\tilde{A})=1- \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{h3}(\tilde{A})=1- \sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{h4}(\tilde{A})=1- \log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{h5}(\tilde{A})=2- 2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} |
Subsequently, entropy from the perspective of knowledge measures for IVIFSs will be explored.
Corollary 3.7. Let K be a knowledge measure induced by the aforementioned function f for IVIFSs, then for any \tilde{A}\in IVIFS(X) ,
E(\tilde{A}) = 1-K(\tilde{A}) |
is an entropy for IVIFSs.
Based on Corollary 3.7, it is straightforward to deduce the existence of a distinct numerical correlation between entropy and the newly devised knowledge metric. Put simply, within the context of an IVIFS, a system with a lower entropy value may consistently possess a greater quantity of knowledge, provided that it adheres to the prescribed axioms.
By carefully examining the intricate relationship between the recently introduced knowledge measure and entropy, we can effortlessly deduce certain entropies.
Upon scrutinizing the information measures previously discussed for IVIFSs, it becomes evident that they share certain linkages, which we shall subsequently elaborate in the form of corollaries.
Corollary 3.8. Let S be an aforementioned similarity measure in Theorem 3.5 for IVIFSs, and for any \tilde{A}\in IVIFS(X) ,
K(\tilde{A}) = 1-S(\tilde{A}, \tilde{A^{c}}) |
is a knowledge measure for IVIFSs.
Corollary 3.9. Let S be an aforementioned similarity measure in Theorem 3.5 for IVIFSs; assuming that \tilde{A}\in IVIFS(X) ,
E(\tilde{A}) = S(\tilde{A}, \tilde{A^{c}}) |
is an entropy for IVIFSs.
Regarding the construction method and transformation relationship of information measures, there are some important issues to be further discussed below, and these discussions pave the way for the subsequent research.
(1) Drawing upon the axiomatic definition of IVIFSs, the intricate relationship among similarity measures, knowledge measures, and the entropy of IVIFSs is comprehensively formulated in general terms, accompanied by explicit expressions. Notably, several key findings are presented in the form of theorems, which underscore the existence of a profound connection between the information measures associated with IVIFSs.
(2) Among these theorems, through the axiomatic definition of IVIFS, we can easily convert the above information measures, i.e., similarity measure, knowledge measure, and entropy, to each other. Specifically, we consider that the similarity measure and knowledge measure of IVIFS can be converted to each other, and, finally, based on the general relationship between knowledge measure and entropy, the entropy is constructed with the similarity measure, so that the larger the similarity measure, the less the amount of knowledge it contains, and the larger the entropy.
(3) The result of Corollary 3.7 in this paper shows that there is a dyadic relationship between entropy and knowledge measure, which paves the way for continuing to explore whether there is a dyadic relationship between entropy and knowledge measure for fuzzy sets and their extended forms.
This section employs the recently proposed knowledge measure to tackle multi-attribute decision-making (MADM) problems under conditions of unknown weights, while utilizing the novel similarity measure to address pattern recognition issues within the framework of IVIFSs. The derived outcomes are subsequently contrasted with several established measures for comprehensive analysis.
In the complex process of MADM, where the weights of individual attributes remain elusive, this subsection introduces a methodology for determining these weights through the utilization of knowledge measures techniques.
Construct an MADM process that examines m alternatives \tilde{A} = \{\tilde{A}_1, \tilde{A}_2, \ldots, \tilde{A}_m\} using n attributes C = \{C_1, C_2, \ldots, C_n\} . Suppose that w = (w_1, w_2, \ldots, w_n)^{T} is the weight vector of attributes, where 0\leq w_j \leq 1 and \sum_{j = 1}^{n}w_j = 1, j = 1, 2, \ldots, n .
The MADM problem of IVIFSs can be described in the matrix form below:
R= \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ \tilde{A}_1 \\ \tilde{A}_2 \\ \vdots \\ \tilde{A}_m \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} C_1 & \ \ C_2 & \cdots & \ \ C_n \end{array}}\\ {\left( {\begin{array}{*{20}{c}} \tilde{a}_{11} & \tilde{a}_{12} & \cdots & \tilde{a}_{1 n} \\ \tilde{a}_{21} & \tilde{a}_{22} & \cdots & \tilde{a}_{2 n} \\ \vdots & \vdots & \ddots & \vdots \\ \tilde{a}_{m 1} & \tilde{a}_{m 2} & \cdots & \tilde{a}_{m n} \end{array}} \right)} \end{array}}, \end{array} | (4.1) |
where \tilde{a}_{ij} = \langle (x, [\mu^{ij}_{\tilde{a}L}(x), \mu^{ij}_{\tilde{a}U}(x)], [\nu^{ij}_{\tilde{a}L}(x), \nu^{ij}_{\tilde{a}U}(x)])|x\in X\rangle (i = 1, 2, \ldots, m; j = 1, 2, \ldots, n) indicates the assessment of an expert corresponding to attribute C_{j}(j = 1, 2, \ldots, n) to evaluate the alternative \tilde{A}_{i}(i = 1, 2, \ldots, m) .
Step 1. By defining the knowledge measure of IVIFSs using K_{H1} , the knowledge measure based decision matrix Q = (\tilde{k}_{ij})_{m\times n} is generated from the decision matrix R above. The matrix Q = (\tilde{k}_{ij})_{m\times n} is shown below.
Q = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {\tilde{A}_1}\\ {\tilde{A}_2}\\ {\vdots}\\ {\tilde{A}_m } \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} C_1 &\ \ C_2& \cdots&\ \ C_n \end{array}}\\ {\left( {\begin{array}{*{20}{c}} \tilde{k}_{11}&\tilde{k}_{12}&\cdots&\tilde{k}_{1n}\\ \tilde{k}_{21}&\tilde{k}_{22}&\cdots&\tilde{k}_{2n}\\ \vdots&\vdots&\ddots&\vdots \\ \tilde{k}_{m1}&\tilde{k}_{m2}&\cdots&\tilde{k}_{mn} \end{array}} \right)} \end{array}}, \end{array} | (4.2) |
where \tilde{k}_{ij} is the knowledge measure corresponding to each \tilde{a}_{ij} .
Step 2. The following formula can be used to construct the attribute weights based on the knowledge measure when the attribute weights w_j are completely unknown:
\begin{equation} w_j = \frac{\sum\nolimits_{i = 1}^{m}\tilde{k}_{ij}}{\sum\nolimits_{j = 1}^{n}\sum\nolimits_{i = 1}^{m}\tilde{k}_{ij}}, \end{equation} | (4.3) |
where 0\leq w_j \leq 1 and \sum_{j = 1}^{n}w_j = 1 .
Step 3. In an interval-valued intuitionistic fuzzy environment, the interval-valued intuitionistic fuzzy arithmetic mean operator [3] is used to calculate the weighted aggregation value of each alternative \tilde{A}_i after the weights of each attribute have been determined and the operator is expressed as follows:
\begin{equation} \begin{split} Z_{i} = & \langle[\mu^{L}_{i}, \mu^{U}_{i}], [\nu^{L}_{i}, \nu^{U}_{i}]\rangle\\ = &\langle[1-\prod\limits_{j = 1}^{n}\left(1-\mu^{ij}_{\tilde{a}L}(x)\right)^{w_{j}}, 1-\prod\limits_{j = 1}^{n}\left(1-\mu^{ij}_{\tilde{a}U}(x)\right)^{w_{j}}], \\ &[\prod\limits_{j = 1}^{n}\left(\nu^{ij}_{\tilde{a}L}(x)\right)^{w_{j}}, \prod\limits_{j = 1}^{n}\left(\nu^{ij}_{\tilde{a}U}(x)\right)^{w_{j}}]\rangle, \end{split} \end{equation} | (4.4) |
where i = 1, 2, \ldots, m; j = 1, 2, \ldots, n. The calculated Z_{i} is an IVIFS.
Step 4. In order to differentiate between the alternatives, the alternatives will be ranked below by calculating a score function for each alternative, with the following equation:
\begin{equation} \phi\left(Z_i\right) = \frac{\lambda(\mu^{L}_{i}+\mu^{U}_{i})}{(1-\lambda)(\nu^{L}_{i}+\nu^{U}_{i})}, i = 1, 2, ..., m, \lambda\in(0, 1). \end{equation} | (4.5) |
Subsequently, arrange the alternatives in a nonincreasing sequence of score function, the larger the score function of the alternative, the better the alternative is. In order to realize the whole decision-making process, Figure 5 describes a framework.
In this section, a practical case study, adapted from [5], will be presented to demonstrate the application of the newly introduced knowledge measure in addressing MADM problems involving unknown weights.
An investment company is about to make an investment, and the four possible choices for the investment capital are (1) \tilde{A}_1 , a car company; (2) \tilde{A}_2 , a food company; (3) \tilde{A}_3 , a computer company; and (4) \tilde{A}_4 , a branch office. The investment firm is trying to decide which of these is the best option. The following four attributes must also be taken into consideration by the investment company when making its choice: C_1 : Risk analysis; C_2 : Growth analysis; C_3 : Environmental impact analysis; and C_4 : Social and political impact analysis.
Let w = (w_1, w_2, w_3, w_4)^{T} be the weight vector of the attributes, which is completely unknown. Additionally, the ratings of the alternatives over the attributes are listed in the following Table 4.
C_1 | C_2 | C_3 | C_4 | |
\tilde{A}_1 | \langle[0.1, 0.2], [0.1, 0.2]\rangle | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.5, 0.5], [0.5, 0.5]\rangle |
\tilde{A}_2 | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.0, 0.0], [0.25, 0.75]\rangle | \langle[0.3, 0.4], [0.4, 0.6]\rangle |
\tilde{A}_3 | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.2, 0.4], [0.2, 0.4]\rangle | \langle[0.2, 0.3], [0.4, 0.7]\rangle | \langle[0.2, 0.3], [0.5, 0.6]\rangle |
\tilde{A}_4 | \langle[0.2, 0.3], [0.6, 0.7]\rangle | \langle[0.4, 0.7], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.5, 0.7], [0.1, 0.3]\rangle |
Step 1. Based on K_{H1} , the knowledge measure matrix is calculated as follows:
Q = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {\tilde{A}_1}\\ {\tilde{A}_2}\\ {\tilde{A}_3}\\ {\tilde{A}_4 } \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} C_1 & \ \ \ C_2& \ \ \ C_3&\ \ \ C_4 \end{array}}\\ {\left( {\begin{array}{*{20}{c}} 0.5&0.5&0.475&0.5\\ 0.65&0.5&0.75&0.575\\ 0.5&0.5&0.65&0.65\\ 0.7&0.65&0.5&0.7 \end{array}} \right).} \end{array}} \end{array} | (4.6) |
Step 2. Next, the weight of attributes are calculated by applying Eq (4.3):
w_1 = 0.2527, w_2 = 0.2312, w_3 = 0.2554, w_4 = 0.2608. |
Step 3. The interval-valued intuitionistic fuzzy weighted average operator can be computed by utilizing the formula presented in Eq (4.4). As a result, the combined results of the schemes are obtained as follows:
Z_1 = \langle[0.3326, 0.4370], [0.2486, 0.3966]\rangle;Z_2 = \langle[0.2737, 0.4085], [0.2536, 0.5111]\rangle; |
Z_3 = \langle[0.2129, 0.3796], [0.3207, 0.5426]\rangle;Z_4 = \langle[0.3379, 0.5766], [0.2203, 0.4234]\rangle. |
Step 4. The scoring values of the alternatives are calculated with Eq (4.5) when \lambda = \frac{1}{2} as follows:
\phi(Z_{1}) = 1.1922, \phi(Z_{2}) = 0.8921, \phi(Z_{3}) = 0.6863, \phi(Z_{4}) = 1.4207. |
Therefore, the sorting order is \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 , and the optimal alternative is \tilde{A}_4 . Furthermore, to underscore the practicality of the newly introduced knowledge measure in addressing the MADM problem, we contrast the derived outcomes with diverse entropy and knowledge measures. Evidently, the majority of the ranking outcomes closely align with our findings, as evident from the subsequent Table 5 and Figure 6. Notably, \tilde{A}_4 emerges as the optimal solution.
Methods | Weighting vector | Alternatives ranking |
E_{JPCZ}(E1) [29] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{LZX}(E2) [30] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{WWZ}(E3) [31] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{ZJJL}(E4) [32] | w={(0.2366, 0.3159, 0.1859, 0.2616)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Das et al. [27] | w={(0.2506, 0.2211, 0.2472, 0.2811)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Proposed method | w={(0.2527, 0.2312, 0.2554, 0.2608)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
The aforementioned examples demonstrate the remarkable effectiveness of applying knowledge measures to tackle MADM problems that involve unknown weights. Additionally, they underscore the usefulness and practical applicability of the newly introduced knowledge measures in our daily lives.
It is evident that the parameter \lambda in Eq (4.5) can be interpreted as the weight of membership degree, and its magnitude significantly impacts the ultimate ranking of alternatives. Sensitivity analysis is performed by selecting various \lambda values. Table 6 presents the score function values \phi(Z_{i}) and the rankings of each alternative, which are calculated based on different \lambda values.
\phi(Z_{1}) | \phi(Z_{2}) | \phi(Z_{3}) | \phi(Z_{4}) | Alternatives order | |
\lambda=0.1 | 0.1325 | 0.0991 | 0.0763 | 0.1579 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.2 | 0.2980 | 0.2230 | 0.1716 | 0.3552 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.3 | 0.5109 | 0.3823 | 0.2941 | 0.6089 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.4 | 0.7948 | 0.5947 | 0.4576 | 0.9472 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.5 | 1.1922 | 0.8920 | 0.6863 | 1.4207 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.6 | 1.7882 | 1.3380 | 1.0295 | 2.1311 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.7 | 2.7817 | 2.0813 | 1.6015 | 3.3150 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.8 | 4.7686 | 3.5680 | 2.7454 | 5.6829 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.9 | 10.7295 | 8.0280 | 6.1771 | 12.7865 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
We randomly select 9 values for \lambda within the interval (0, 1) , and it is evident that regardless of how \lambda varies, the ranking of alternatives remains consistent throughout Table 6. Additionally, we will visualize the data in Figure 7, which illustrates that the ultimate ranking remains unaltered despite changes in \lambda . As evident from Table 6, when \lambda > 0.5 , the score function values of the alternative solutions exhibit a substantial increase, primarily due to a greater preference for membership degree under such conditions.
To clearly demonstrate the disparities among alternative solutions as \lambda escalates, consider the following definitions: T_{1} = \phi(Z_{4})-\phi(Z_{1}), T_{2} = \phi(Z_{1})-\phi(Z_{2}), T_{3} = \phi(Z_{2})-\phi(Z_{3}). It can be found that with the increase of \lambda , the discrimination between the alternatives increases in Table 7 and Figure 8. That is to say, with the increase of \lambda , the score function \phi(Z_{i}) can better distinguish the alternatives. Therefore, the ranking results should be minimally influenced by subjective factors, which further demonstrates the stability of the method proposed in this paper.
T_{1} | T_{2} | T_{3} | |
\lambda=0.1 | 0.0254 | 0.0334 | 0.0229 |
\lambda=0.2 | 0.0571 | 0.0750 | 0.0514 |
\lambda=0.3 | 0.0980 | 0.1286 | 0.0881 |
\lambda=0.4 | 0.1524 | 0.2001 | 0.1371 |
\lambda=0.5 | 0.2286 | 0.3002 | 0.2057 |
\lambda=0.6 | 0.3428 | 0.4502 | 0.3085 |
\lambda=0.7 | 0.5333 | 0.7004 | 0.4799 |
\lambda=0.8 | 0.9143 | 1.2007 | 0.8226 |
\lambda=0.9 | 2.0571 | 2.7015 | 1.8509 |
Novel coronavirus pneumonia (COVID-19) is an acute respiratory infectious disease caused by the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2). Since its outbreak in Pakistan, the COVID-19 situation there has garnered widespread attention. The epidemic's spread in Pakistan has exhibited certain fluctuations, with a gradual rise in confirmed cases, exerting a profound impact on public health, socioeconomic conditions, and the daily lives of citizens.
This paper refers to the example of healthcare facilities in Khyber Pakhtunkhwa, Pakistan in [33]. There is currently a review panel comprising three experts, e_{1}, e_{2} , and e_{3} , tasked with evaluating four hospitals A_{l} (l = 1, 2, 3, 4) in Khyber Pakhtunkhwa based on four attributes: c_{1} : Reverse Transcription-Polymerase Chain Reaction Facilities (RT-PCR), c_{2} : Personal protective equipment (PPE), c_{3} : Shortage of Isolation Ward (SIW), c_{4} : Hight Qualified Staff (HQS).
Next, the specific steps of the decision-making method are as follows.
Step 1. We construct interval valued intuitionistic fuzzy decision matrices, as shown in Tables 8–10.
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle |
A_2 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle |
A_4 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_3 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_4 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.5], [0.2, 0.3]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.4], [0.2, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_4 | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.5], [0.2, 0.5]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
Step 2. By utilizing K_{h1} to define the knowledge measure of IVIFSs, a decision matrix Q^{k} (k = 1, 2, 3) , grounded in this knowledge measure, is derived from the decision matrix. The matrix Q^{k}(k = 1, 2, 3) is presented as follows.
Q^{1} = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {A_1}\\ {A_2}\\ {A_3}\\ {A_4} \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} c_1 & \ \ c_2& \ \ c_3& \ \ c_4 \end{array}}\\ {\left( {\begin{array}{*{20}{c}} 0.6&0.8&0.6&0.7\\ 0.7&0.7&0.8&0.7\\ 0.7&0.8&0.7&0.7\\ 0.7&0.6&0.8&0.7 \end{array}} \right).} \end{array}} \end{array} |
Q^{2} = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {A_1}\\ {A_2}\\ {A_3}\\ {A_4} \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} c_1 & \ \ c_2& \ \ c_3&\ \ c_4 \end{array}}\\ {\left( {\begin{array}{*{20}{c}} 0.7&0.6&0.8&0.8\\ 0.7&0.8&0.7&0.7\\ 0.7&0.7&0.8&0.8\\ 0.7&0.8&0.6&0.6 \end{array}} \right).} \end{array}} \end{array} |
Q^{3} = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {A_1}\\ {A_2}\\ {A_3}\\ {A_4} \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} c_1 & \ \ c_2& \ \ c_3& \ \ c_4 \end{array}}\\ {\left( {\begin{array}{*{20}{c}} 0.8&0.7&0.7&0.6\\ 0.7&0.8&0.7&0.8\\ 0.8&0.7&0.7&0.7\\ 0.6&0.9&0.7&0.8 \end{array}} \right).} \end{array}} \end{array} |
Step 3. We assign equal weights to the three experts and aggregate their knowledge measure decision matrices into one using the formula below:
\begin{equation} \bar{Q} = \frac{Q^{1}+Q^{2}+Q^{3}}{3}. \end{equation} | (4.7) |
Obtain the following knowledge measure aggregation matrix \bar{Q} :
\bar{Q} = \begin{array}{*{20}{c}} {\begin{array}{*{20}{l}} {}\\ {A_1}\\ {A_2}\\ {A_3}\\ {A_4} \end{array}}&{\begin{array}{*{20}{c}} {\begin{array}{*{20}{c}} c_1 &\ \ \ c_2& \ \ \ c_3& \ \ \ c_4 \end{array}}\\ {\left( {\begin{array}{*{20}{c}} 0.7&0.7&0.7&0.7\\ 0.7&0.77&0.73&0.73\\ 0.73&0.73&0.73&0.73\\ 0.67&0.77&0.7&0.7 \end{array}} \right).} \end{array}} \end{array} |
Step 4. Determine attribute weights by applying Eq (4.3).
w_1 = 0.2437, w_2 = 0.2585, w_3 = 0.2489, w_4 = 0.2489. |
Step 5. Employ Eq (4.4) to apply weights and aggregate the attributes of the alternative (Table 11).
e_1 | e_2 | e_3 | |
z_1 | \langle[0.2577, 0.4316], [0.2067, 0.3869]\rangle | \langle[0.2357, 0.4573], [0.2352, 0.3877]\rangle | \langle[0.2591, 0.4563], [0.2447, 0.3869]\rangle |
z_2 | \langle[0.2306, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2754, 0.4000], [0.1000, 0.2824]\rangle | \langle[0.2278, 0.4000], [0.1196, 0.2833]\rangle |
z_3 | \langle[0.2286, 0.3765], [0.1314, 0.2824]\rangle | \langle[0.2046, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2302, 0.3765], [0.1314, 0.2824]\rangle |
z_4 | \langle[0.2607, 0.4321], [0.2063, 0.3877]\rangle | \langle[0.3289, 0.4773], [0.2067, 0.3869]\rangle | \langle[0.2562, 0.4768], [0.1858, 0.3877]\rangle |
Step 6. When \lambda = \frac{1}{2} , the score values by each expert to the alternative are calculated using Eq (4.5) as shown in Table 12.
\varphi_1 | \varphi_2 | \varphi_3 | \varphi_4 | |
e_1 | 1.1613 | 1.4679 | 1.4623 | 1.1662 |
e_2 | 1.0949 | 1.7662 | 1.4051 | 1.3583 |
e_3 | 1.1328 | 1.5581 | 1.4660 | 1.2781 |
Step 7. Aggregate the score values of three experts to obtain the final ranking of alternative solutions.
\bar{\varphi}_{1} = 1.1297, \bar{\varphi}_{2} = 1.5974, \bar{\varphi}_{3} = 1.4445, \bar{\varphi}_{4} = 1.2675. |
It is not hard to observe that \bar{\varphi}_{2} > \bar{\varphi}_{3} > \bar{\varphi}_{4} > \bar{\varphi}_{1} . Thus, yielding the alternative ranking result: A_2\succ A_3\succ A_4\succ A_1 . This is consistent with the result in [33], which also demonstrates the effectiveness and practicality of the method proposed in this paper.
Suppose X is a finite universe of discourse. There are m patterns which are denoted by IVIFSs
\begin{equation*} \begin{split} \tilde{M}_{j} = &\{\langle x_{1}, [\mu^{L}_{\tilde{M}_{j}}(x_{1}), \mu^{U}_{\tilde{M}_{j}}(x_{1})], [\nu^{L}_{\tilde{M}_{j}}(x_{1}), \nu^{U}_{\tilde{M}_{j}}(x_{1})]\rangle, ..., \\ &\langle x_{n}, [\mu^{L}_{\tilde{M}_{j}}(x_{n}), \mu^{U}_{\tilde{M}_{j}}(x_{n})], [\nu^{L}_{\tilde{M}_{j}}(x_{n}), \nu^{U}_{\tilde{M}_{j}}(x_{n})]\rangle\; |\; x\in X\}\; (j = 1, 2, ..., m) \end{split} \end{equation*} |
and a test sample that needs to be categorized exists, which is denoted by IVIFS
\begin{equation*} \begin{split} \tilde{B} = &\{\langle x_{1}, [\mu^{L}_{\tilde{B}}(x_{1}), \mu^{U}_{\tilde{B}}(x_{1})], [\nu^{L}_{\tilde{B}}(x_{1}), \nu^{U}_{\tilde{B}}(x_{1})]\rangle, ..., \\ &\langle x_{n}, [\mu^{L}_{\tilde{B}}(x_{n}), \mu^{U}_{\tilde{B}}(x_{n})], [\nu^{L}_{\tilde{B}}(x_{n}), \nu^{U}_{\tilde{B}}(x_{n})]\rangle\; |\; x\in X\}. \end{split} \end{equation*} |
The following is the recognition process:
(a) Compute the similarity measure S(\tilde{M}_{j}, \tilde{B}) between \tilde{M}_{j}(j = 1, ..., m) and \tilde{B} with S_{H1} and S_{h1} .
(b) Sort S(\tilde{M}_{j}, \tilde{B}) in ascending order, where a larger value of S(\tilde{M}_{j}, \tilde{B}) indicates a closer proximity of \tilde{B} to category \tilde{M}_{j} .
A diverse array of scholars have proposed numerous approaches to tackle medical diagnosis from varying perspectives. In this section, we employ pattern recognition techniques to address the medical diagnosis challenge: the collection of symptoms serves as the discourse universe, the patient functions as an unidentified test specimen, and the various diseases correspond to multiple modalities. The objective is to categorize patients into distinct groups based on their respective diseases.
Suppose X = \{x_{1}\text{(Temperature)}, \; x_{2}\text{(Cough)}, \; x_{3} \text{(Headache)}, \; x_{4}\text{(Stomach }\; \text{pain)}\} is a group of symptoms and \tilde{M} = \{\tilde{M}_{1}\text{(Viral fever)}, \; \tilde{M}_{2}\text{(Typhoid)}, \; \tilde{M}_{3}{(Pneumonia)}, \; \tilde{M}_{4}\text{(Stomach problem)}\} is a group of the diseases. Then, the disease-symptom matrix expressed by IVIFSs is shown in Table 13 and data is derived from [16].
x_{1} (Temperature) | x_{2} (Cough) | x_{3} (Headache) | x_{4} (Stomach pain) | |
\tilde{M}_{1} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle |
\tilde{M}_{2} | \langle[0.5, 0.6], [0.1, 0.3]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle | \langle[0.4, 0.6], [0.1, 0.2]\rangle |
\tilde{M}_{3} | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.4, 0.6], [0.2, 0.4]\rangle | \langle[0.3, 0.5], [0.2, 0.4]\rangle |
\tilde{M}_{4} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle |
Let the patient B be expressed as:
\begin{equation} \begin{split} \notag B = &\{ \langle x_{1}, [0.4, 0.5], [0.1, 0.2]\rangle, \langle x_{2}, [0.7, 0.8], [0.1, 0.2]\rangle, \\ &\langle x_{3}, [0.9, 0.9], [0.0, 0.1]\rangle, \langle x_{4}, [0.3, 0.5], [0.2, 0.4]\rangle\}. \end{split} \end{equation} |
The aim is to categorize patient B into one of diseases \tilde{M}_{1}, \tilde{M}_{2}, \tilde{M}_{3}, \tilde{M}_{4} . We can then obtain the following results on IVIFSs, as shown in Table 14.
S(\tilde{M}_{1}, B) | S(\tilde{M}_{2}, B) | S(\tilde{M}_{3}, B) | S(\tilde{M}_{4}, B) | Recognition Result | |
S_{1} [11] | 0.73 | 0.80 | 0.78 | 0.73 | \tilde{M}_{2} |
S_{D} [6] | 0.82 | 0.91 | 0.86 | 0.84 | \tilde{M}_{2} |
S_{H1} | 0.32 | 0.36 | 0.33 | 0.28 | \tilde{M}_{2} |
S_{h1} | 0.38 | 0.45 | 0.43 | 0.34 | \tilde{M}_{2} |
We may determine that there is the most similarity between \tilde{M}_{2} and B by taking into account the pattern recognition principle of IVIFSs, but the similarity measure S_{1} is unable to determine which of \tilde{M}_{1} and \tilde{M}_{4} is greater. Furthermore, the newly suggested similarity measures are more practical and capable of resolving this point. In addition, by using the similarity measures S_{H1} and S_{h1} proposed in this paper, the same recognition results can be obtained. Hence, based on the recognition principle, we can classify patient B as suffering from disease \tilde{M}_{2} , thereby conclusively diagnosing them with typhoid illness.
This paper introduces the axiomatic definition of the closest crisp set for IVIFSs, along with its pertinent properties. Subsequently, leveraging the distance measure to the closest crisp set, a novel knowledge measure and entropy for IVIFSs are presented. Furthermore, several theorems are proven, and the interplay between knowledge measures and entropy is examined. Ultimately, the efficacy of the proposed method is validated through two concrete examples: The application of knowledge measures in MADM and similarity measures in pattern recognition.
In the future, we will persist in exploring the application of IVIFS information measures in other domains and strive to develop superior and more rational models for the mutual conversion of these information measures.
Chunfeng Suo: Writing-review & editing, supervision, project administration, funding acquisition; Le Fu: Conceptualization, Methodology, software, validation, writing-original draft, writing-review & editing; Jingxuan Chen: Software, validation, supervision, conceptualization; Xuanchen Li: Validation, writing-review. All authors have read and approved the final version of the manuscript for publication.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This paper was supported by National Science Foundation of China (Grant Nos: 11671244, 12071271), Natural Science Foundation of Jilin Province (Grant Nos: YDZJ202201ZYTS320) and the Sixth Batch of Jilin Province Youth Science and Technology Talent Lifting Project.
The authors declare that there are no conflicts of interest in the publication of this paper.
[1] |
K. Atanassov, Intuitionistic fuzzy sets, Fuzzy Set. Syst., 20 (1986), 87–96. https://doi.org/10.1016/S0165-0114(86)80034-3 doi: 10.1016/S0165-0114(86)80034-3
![]() |
[2] |
K. Atanassov, G. Gargov, Interval-valued intuitionistic fuzzy sets, Fuzzy Set. Syst., 31 (1989), 343–349. https://doi.org/10.1016/0165-0114(89)90205-4 doi: 10.1016/0165-0114(89)90205-4
![]() |
[3] |
Z. Xu, Methods for aggregating interval-valued intuitionistic fuzzy information and their application to decision making, J. Control. Decis., 2 (2007), 215–219. https://doi:10.13195/j.cd.2007.02.97.xuzsh.020. doi: 10.13195/j.cd.2007.02.97.xuzsh.020
![]() |
[4] |
J. Zhang, J. Lin, T. Wu, An interval intuitionistic fuzzy characterization method based on heterogeneous big data and its application in forest land quality assessment, Int. J. Fuzzy Syst., 2024, 1–24. https://doi.org/10.1007/S40815-024-01765-5 doi: 10.1007/S40815-024-01765-5
![]() |
[5] |
F. Herrera, E. Herrera-Viedma, Linguistic decision analysis: Steps for solving decision problems under linguistic information, Fuzzy Set. Syst., 115 (2000), 67–82. https://doi.org/10.1016/S0165-0114(99)00024-X doi: 10.1016/S0165-0114(99)00024-X
![]() |
[6] |
M. Khan, S. Jan, R. Jan, T. Senapati, Complex interval-valued intuitionistic fuzzy decision support system with application to COVID-19 healthcare facilities, Complex Intell. Syst., 9 (2023), 7103–7132. https://doi.org/10.1007/S40747-023-01090-8 doi: 10.1007/S40747-023-01090-8
![]() |
[7] |
J. Zhang, Y. Shan, J. Dong, A threat assessment method based on correlation-similarity information and three-way decisions in interval intuitionistic fuzzy environment, Eng. Appl. Artif. Intell., 135 (2024), 108793. https://doi.org/ 10.1016/J.ENGAPPAI.2024.108793 doi: 10.1016/J.ENGAPPAI.2024.108793
![]() |
[8] |
M. Qayyum, E. Kerre, S. Ashraf, A parametric family of fuzzy similarity measures for intuitionistic fuzzy sets, Mathematics, 11 (2023), 3163. https://doi.org/10.3390/MATH11143163 doi: 10.3390/MATH11143163
![]() |
[9] |
X. Gou, Z. Xu, H. Liao, H. Francisco, Multiple criteria decision making based on distance and similarity measures under double hierarchy hesitant fuzzy linguistic environment, Comput. Ind. Eng., 126 (2018), 516–530. https://doi.org/10.1016/j.cie.2018.10.020 doi: 10.1016/j.cie.2018.10.020
![]() |
[10] |
C. Suo, Y. Wang, D. Mou, The new construction of knowledge measure on intuitionistic fuzzy sets and interval-valued intuitionistic fuzzy sets, AIMS Math., 8 (2023), 27113–27127. https://doi.org/10.3934/math.20231387 doi: 10.3934/math.20231387
![]() |
[11] |
Z. Xu, J. Chen, An overview of distance and simiarity measures of intuitionistic fuzzy sets, Int. J. Uncertain. Fuzz., 16 (2008), 529–555. https://doi.org/ 10.1142/S0218488508005406 doi: 10.1142/S0218488508005406
![]() |
[12] |
Y. Liu, W. Jiang, A new distance measure of interval-valued intuitionistic fuzzy sets and its application in decision making, Soft Comput., 24 (2020), 6987–7003. https://doi.org/10.1007/s00500-019-04332-5 doi: 10.1007/s00500-019-04332-5
![]() |
[13] |
X. Liu, Y. Sun, G. Harish, S. Zhang, Analysis of distance measures in intuitionistic fuzzy set theory: A line integral perspective, Expert Syst. Appl., 226 (2023), 120221. https://doi.org/10.1016/j.eswa.2023.120221 doi: 10.1016/j.eswa.2023.120221
![]() |
[14] |
Z. Xu, Some similarity measures of intuitionistic fuzzy sets and their applica-tions to multiple attribute decision making, Fuzzy Optim. Decis. Mak., 6 (2007), 109–121. https://doi.org/10.1007/s10700-007-9004-z doi: 10.1007/s10700-007-9004-z
![]() |
[15] |
H. Bustince, C. Marco-Detchart, J. Fernandez, C. Wagner, J. Garibaldi, Z. Takac, Similarity between interval-valued fuzzy sets taking into account the width of the intervals and admissible orders, Fuzzy Set. Syst., 390 (2020), 23–47. https://doi.org/10.1016/j.fss.2019.04.002 doi: 10.1016/j.fss.2019.04.002
![]() |
[16] |
M. Luo, J. Liang, A novel similarity measure for interval-valued intuitionistic fuzzy sets and its applications, Symmetry, 10 (2018), 441. https://doi.org/10.3390/sym10100441 doi: 10.3390/sym10100441
![]() |
[17] |
I. Montes, N. Pal, S. Montes, Entropy measures for atanassov intuitionistic fuzzy sets based on divergence, Soft Comput., 22 (2018), 5051–5071. https://doi.org/10.1007/s00500-018-3318-3 doi: 10.1007/s00500-018-3318-3
![]() |
[18] | M. Sun, J. Liu, New entropy and similarity measures for interval-valued intuitionistic fuzzy sets, J. Comput. Sci., 9 (2012), 5799–5806. https://www.researchgate.net/publication/287835113 |
[19] |
R. Tabasam, F. Shahzad, Z. Sohail, Based entropy measure of interval-valued intuitionistic fuzzy sets and its application in multicriteria decision making, Adv. Fuzzy Syst., 2018 (2018), 1–10. https://doi.org/10.1155/2018/3637897 doi: 10.1155/2018/3637897
![]() |
[20] |
Q. Zhang, H. Xing, F. Liu, J. Ye, P. Tang, Some new entropy measures for interval-valued intuitionistic fuzzy sets based on distances and their relationships with similarity and inclusion measure, Inf. Sci., 283 (2014), 55–69. https://doi.org/10.1016/j.ins.2014.06.012 doi: 10.1016/j.ins.2014.06.012
![]() |
[21] |
R. Che, C. Suo, Y. Li, An approach to construct entropies on interval-valued intuitionistic fuzzy sets by their distance functions, Soft Comput., 25 (2021), 1–11. https://doi.org/10.1007/S00500-021-05713-5 doi: 10.1007/S00500-021-05713-5
![]() |
[22] |
C. E. Shannon, A mathematical theory of communication, Bell Syst. Tech. J., 27 (1948), 379–423. https://doi.org/10.1002/j.1538-7305.1948.tb00917.x doi: 10.1002/j.1538-7305.1948.tb00917.x
![]() |
[23] |
Y. Jun, Fuzzy cross entropy of interval-valued intuitionistic fuzzy sets and its optimal decision-making method based on the weights of alternatives, Expert Syst. Appl., 38 (2011), 6179–6183. https://doi.org/10.1016/j.eswa.2010.11.052 doi: 10.1016/j.eswa.2010.11.052
![]() |
[24] |
C. Suo, X. Li, Y. Li, Distance-based knowledge measure and entropy for interval-valued intuitionistic fuzzy sets, Mathematics, 11 (2023), 3468. 10.3390/MATH11163468 doi: 10.3390/MATH11163468
![]() |
[25] |
K. Guo, H. Xu, A unified framework for knowledge measure with application: From fuzzy sets through interval-valued intuitionistic fuzzy sets, Appl. Soft Comput., 109 (2021), 107539. https://doi.org/10.1016/j.asoc.2021.107539 doi: 10.1016/j.asoc.2021.107539
![]() |
[26] |
E. Szmidt, J. Kacprzyk, P. Bujnowski, How to measure the amount of knowledge conveyed by Atanassovs intuitionistic fuzzy sets, Inf. Sci., 257 (2014), 276–285. https://doi.org/10.1016/j.ins.2012.12.046 doi: 10.1016/j.ins.2012.12.046
![]() |
[27] |
S. Das, B. Dutta, D. Guha, Weight computation of criteria in a decision-making problem by knowledge measure with intuitionistic fuzzy set and interval-valued intuitionistic fuzzy set, Soft Comput., 20 (2016), 3421–3442. https://doi.org/10.1007/s00500-015-1813-3 doi: 10.1007/s00500-015-1813-3
![]() |
[28] |
K. Guo, J. Zang, Knowledge measure for interval-valued intuitionistic fuzzy sets and its application to decision making under uncertainty, Soft Comput., 23 (2018), 6967–6978. https://doi.org/10.1007/s00500-018-3334-3 doi: 10.1007/s00500-018-3334-3
![]() |
[29] |
F. Jin, L. Pei, H. Chen, L. Zhou, Interval-valued intuitionistic fuzzy continuous weighted entropy and its application to multi-criteria fuzzy group decision making, Knowl-Based Syst., 59 (2014), 132–141. https://doi.org/10.1016/j.knosys.2014.01.014 doi: 10.1016/j.knosys.2014.01.014
![]() |
[30] | X. Liu, S. Zheng, F. Xiong, Entropy and subsethood for general interval-valued intuitionistic fuzzy sets, Fuzzy systems and knowledge discovery, Second International Conference, FSKD 2005, Changsha, China, 2005, 42–52. https://doi.org/10.1007/11539506-6 |
[31] |
G. Wei, Gray relational analysis method for intuitionistic fuzzy multiple attribute decision making, Expert Syst. Appl., 38 (2011), 11671–11677. https://doi.org/10.1016/j.eswa.2011.03.048 doi: 10.1016/j.eswa.2011.03.048
![]() |
[32] |
Q. Zhang, S. Jiang, B. Jia, S. Luo, Some information measures for interval-valued intuitionistic fuzzy sets, Inf. Sci., 180 (2010), 5130–5145. https://doi.org/10.1016/j.ins.2010.08.038 doi: 10.1016/j.ins.2010.08.038
![]() |
[33] |
M. Khan, S. Jan, R. Jan, T. Senapati, Complex interval-valued intuitionistic fuzzy decision support system with application to COVID-19 healthcare facilities, Complex Intell. Syst., 9 (2023), 7103–7132. https://doi.org/10.1007/S40747-023-01090-8 doi: 10.1007/S40747-023-01090-8
![]() |
[34] |
S. Moslem, A novel parsimonious spherical fuzzy analytic hierarchy process for sustainable urban transport solutions, Eng. Appl. Artif. Intell., 128 (2024), 107447. https://doi.org/10.1016/J.ENGAPPAI.2023.107447 doi: 10.1016/J.ENGAPPAI.2023.107447
![]() |
\ast=+ \star=+ |
Minkovsky | q=4 | p=1 | D_{H}(\tilde{A}, \tilde{B})=\big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}+\beta_{i}+\gamma_{i}+\delta_{i})\big) |
p=2 | D_{E}(\tilde{A}, \tilde{B})= \big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}^{2}+\beta_{i}^{2}+\gamma_{i}^{2}+\delta_{i}^{2})\big)^{\frac{1}{2}} | |||
p=3 | D_{M}(\tilde{A}, \tilde{B})= \big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}^{3}+\beta_{i}^{3}+\gamma_{i}^{3}+\delta_{i}^{3})\big)^{\frac{1}{3}} | |||
\ast=\vee, \star=+ \ast=\vee, \star=\vee |
p=1 | q=2 | Hamming-Hausdorff | D_{HH}(\tilde{A}, \tilde{B})=\big(\frac{1}{2n}\sum_{i=1}^{n} (\alpha_{i}\vee\beta_{i}+\gamma_{i}\vee\delta_{i})\big) |
q=1 | Hausdorff | D_{h}(\tilde{A}, \tilde{B})=\big(\frac{1}{n}\sum_{i=1}^{n} (\alpha_{i}\vee\beta_{i}\vee\gamma_{i}\vee\delta_{i})\big) |
Distance measures | Conversion functions | Similarity measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{H1}(\tilde{A}, \tilde{B})= \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{H2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{H3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{H4}(\tilde{A}, \tilde{B})=\log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{H5}(\tilde{A}, \tilde{B})=2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{E1}(\tilde{A}, \tilde{B})= \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{E2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{E3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{E4}(\tilde{A}, \tilde{B})=\log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{E5}(\tilde{A}, \tilde{B})=2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{M1}(\tilde{A}, \tilde{B})= \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{M2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{M3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{M4}(\tilde{A}, \tilde{B})=\log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{M5}(\tilde{A}, \tilde{B})=2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{HH1}(\tilde{A}, \tilde{B})= \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{HH4}(\tilde{A}, \tilde{B})=\log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{HH5}(\tilde{A}, \tilde{B})=2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{h1}(\tilde{A}, \tilde{B})= \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{h2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{h3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{h4}(\tilde{A}, \tilde{B})=\log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{h5}(\tilde{A}, \tilde{B})=2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{B}, C_{\tilde{B}})}-1 |
Distance measures | Conversion functions | Knowledge measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{H1}(\tilde{A})=1- \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{H2}(\tilde{A})=1- \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{H3}(\tilde{A})=1- \sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{H4}(\tilde{A})=1- \log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{H5}(\tilde{A})=2- 2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{E1}(\tilde{A})=1- \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{E2}(\tilde{A})=1- \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{E3}(\tilde{A})=1- \sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{E4}(\tilde{A})=1- \log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{E5}(\tilde{A})=2- 2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{M1}(\tilde{A})=1- \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{M2}(\tilde{A})=1- \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{M3}(\tilde{A})=1- \sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{M4}(\tilde{A})=1- \log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{M5}(\tilde{A})=2- 2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{HH1}(\tilde{A})=1- \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{HH4}(\tilde{A})=1- \log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{HH5}(\tilde{A})=2- 2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{h1}(\tilde{A})=1- \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{h2}(\tilde{A})=1- \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{h3}(\tilde{A})=1- \sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{h4}(\tilde{A})=1- \log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{h5}(\tilde{A})=2- 2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} |
C_1 | C_2 | C_3 | C_4 | |
\tilde{A}_1 | \langle[0.1, 0.2], [0.1, 0.2]\rangle | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.5, 0.5], [0.5, 0.5]\rangle |
\tilde{A}_2 | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.0, 0.0], [0.25, 0.75]\rangle | \langle[0.3, 0.4], [0.4, 0.6]\rangle |
\tilde{A}_3 | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.2, 0.4], [0.2, 0.4]\rangle | \langle[0.2, 0.3], [0.4, 0.7]\rangle | \langle[0.2, 0.3], [0.5, 0.6]\rangle |
\tilde{A}_4 | \langle[0.2, 0.3], [0.6, 0.7]\rangle | \langle[0.4, 0.7], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.5, 0.7], [0.1, 0.3]\rangle |
Methods | Weighting vector | Alternatives ranking |
E_{JPCZ}(E1) [29] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{LZX}(E2) [30] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{WWZ}(E3) [31] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{ZJJL}(E4) [32] | w={(0.2366, 0.3159, 0.1859, 0.2616)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Das et al. [27] | w={(0.2506, 0.2211, 0.2472, 0.2811)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Proposed method | w={(0.2527, 0.2312, 0.2554, 0.2608)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\phi(Z_{1}) | \phi(Z_{2}) | \phi(Z_{3}) | \phi(Z_{4}) | Alternatives order | |
\lambda=0.1 | 0.1325 | 0.0991 | 0.0763 | 0.1579 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.2 | 0.2980 | 0.2230 | 0.1716 | 0.3552 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.3 | 0.5109 | 0.3823 | 0.2941 | 0.6089 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.4 | 0.7948 | 0.5947 | 0.4576 | 0.9472 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.5 | 1.1922 | 0.8920 | 0.6863 | 1.4207 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.6 | 1.7882 | 1.3380 | 1.0295 | 2.1311 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.7 | 2.7817 | 2.0813 | 1.6015 | 3.3150 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.8 | 4.7686 | 3.5680 | 2.7454 | 5.6829 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.9 | 10.7295 | 8.0280 | 6.1771 | 12.7865 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
T_{1} | T_{2} | T_{3} | |
\lambda=0.1 | 0.0254 | 0.0334 | 0.0229 |
\lambda=0.2 | 0.0571 | 0.0750 | 0.0514 |
\lambda=0.3 | 0.0980 | 0.1286 | 0.0881 |
\lambda=0.4 | 0.1524 | 0.2001 | 0.1371 |
\lambda=0.5 | 0.2286 | 0.3002 | 0.2057 |
\lambda=0.6 | 0.3428 | 0.4502 | 0.3085 |
\lambda=0.7 | 0.5333 | 0.7004 | 0.4799 |
\lambda=0.8 | 0.9143 | 1.2007 | 0.8226 |
\lambda=0.9 | 2.0571 | 2.7015 | 1.8509 |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle |
A_2 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle |
A_4 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_3 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_4 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.5], [0.2, 0.3]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.4], [0.2, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_4 | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.5], [0.2, 0.5]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
e_1 | e_2 | e_3 | |
z_1 | \langle[0.2577, 0.4316], [0.2067, 0.3869]\rangle | \langle[0.2357, 0.4573], [0.2352, 0.3877]\rangle | \langle[0.2591, 0.4563], [0.2447, 0.3869]\rangle |
z_2 | \langle[0.2306, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2754, 0.4000], [0.1000, 0.2824]\rangle | \langle[0.2278, 0.4000], [0.1196, 0.2833]\rangle |
z_3 | \langle[0.2286, 0.3765], [0.1314, 0.2824]\rangle | \langle[0.2046, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2302, 0.3765], [0.1314, 0.2824]\rangle |
z_4 | \langle[0.2607, 0.4321], [0.2063, 0.3877]\rangle | \langle[0.3289, 0.4773], [0.2067, 0.3869]\rangle | \langle[0.2562, 0.4768], [0.1858, 0.3877]\rangle |
\varphi_1 | \varphi_2 | \varphi_3 | \varphi_4 | |
e_1 | 1.1613 | 1.4679 | 1.4623 | 1.1662 |
e_2 | 1.0949 | 1.7662 | 1.4051 | 1.3583 |
e_3 | 1.1328 | 1.5581 | 1.4660 | 1.2781 |
x_{1} (Temperature) | x_{2} (Cough) | x_{3} (Headache) | x_{4} (Stomach pain) | |
\tilde{M}_{1} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle |
\tilde{M}_{2} | \langle[0.5, 0.6], [0.1, 0.3]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle | \langle[0.4, 0.6], [0.1, 0.2]\rangle |
\tilde{M}_{3} | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.4, 0.6], [0.2, 0.4]\rangle | \langle[0.3, 0.5], [0.2, 0.4]\rangle |
\tilde{M}_{4} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle |
S(\tilde{M}_{1}, B) | S(\tilde{M}_{2}, B) | S(\tilde{M}_{3}, B) | S(\tilde{M}_{4}, B) | Recognition Result | |
S_{1} [11] | 0.73 | 0.80 | 0.78 | 0.73 | \tilde{M}_{2} |
S_{D} [6] | 0.82 | 0.91 | 0.86 | 0.84 | \tilde{M}_{2} |
S_{H1} | 0.32 | 0.36 | 0.33 | 0.28 | \tilde{M}_{2} |
S_{h1} | 0.38 | 0.45 | 0.43 | 0.34 | \tilde{M}_{2} |
\ast=+ \star=+ |
Minkovsky | q=4 | p=1 | D_{H}(\tilde{A}, \tilde{B})=\big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}+\beta_{i}+\gamma_{i}+\delta_{i})\big) |
p=2 | D_{E}(\tilde{A}, \tilde{B})= \big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}^{2}+\beta_{i}^{2}+\gamma_{i}^{2}+\delta_{i}^{2})\big)^{\frac{1}{2}} | |||
p=3 | D_{M}(\tilde{A}, \tilde{B})= \big(\frac{1}{4n}\sum_{i=1}^{n} (\alpha_{i}^{3}+\beta_{i}^{3}+\gamma_{i}^{3}+\delta_{i}^{3})\big)^{\frac{1}{3}} | |||
\ast=\vee, \star=+ \ast=\vee, \star=\vee |
p=1 | q=2 | Hamming-Hausdorff | D_{HH}(\tilde{A}, \tilde{B})=\big(\frac{1}{2n}\sum_{i=1}^{n} (\alpha_{i}\vee\beta_{i}+\gamma_{i}\vee\delta_{i})\big) |
q=1 | Hausdorff | D_{h}(\tilde{A}, \tilde{B})=\big(\frac{1}{n}\sum_{i=1}^{n} (\alpha_{i}\vee\beta_{i}\vee\gamma_{i}\vee\delta_{i})\big) |
Distance measures | Conversion functions | Similarity measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{H1}(\tilde{A}, \tilde{B})= \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{H2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{H3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{H4}(\tilde{A}, \tilde{B})=\log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{H5}(\tilde{A}, \tilde{B})=2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{E1}(\tilde{A}, \tilde{B})= \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{E2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{E3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{E4}(\tilde{A}, \tilde{B})=\log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{E5}(\tilde{A}, \tilde{B})=2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{M1}(\tilde{A}, \tilde{B})= \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{M2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{M3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{M4}(\tilde{A}, \tilde{B})=\log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{M5}(\tilde{A}, \tilde{B})=2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{HH1}(\tilde{A}, \tilde{B})= \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{HH4}(\tilde{A}, \tilde{B})=\log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{HH5}(\tilde{A}, \tilde{B})=2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{B}, C_{\tilde{B}})}-1 | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | S_{h1}(\tilde{A}, \tilde{B})= \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | S_{h2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | S_{h3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | S_{h4}(\tilde{A}, \tilde{B})=\log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{B}, C_{\tilde{B}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | S_{h5}(\tilde{A}, \tilde{B})=2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{B}, C_{\tilde{B}})}-1 |
Distance measures | Conversion functions | Knowledge measures |
D_{H}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{H1}(\tilde{A})=1- \frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{H2}(\tilde{A})=1- \sin\frac{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{H3}(\tilde{A})=1- \sqrt[k]{\frac{D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{H4}(\tilde{A})=1- \log^{\big(D_{H}(\tilde{A}, C_{\tilde{A}})+D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{H5}(\tilde{A})=2- 2^{D_{H}(\tilde{A}, C_{\tilde{A}})\cdot D_{H}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{E}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{E1}(\tilde{A})=1- \frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{E2}(\tilde{A})=1- \sin\frac{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{E3}(\tilde{A})=1- \sqrt[k]{\frac{D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{E4}(\tilde{A})=1- \log^{\big(D_{E}(\tilde{A}, C_{\tilde{A}})+D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{E5}(\tilde{A})=2- 2^{D_{E}(\tilde{A}, C_{\tilde{A}})\cdot D_{E}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{M}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{M1}(\tilde{A})=1- \frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{M2}(\tilde{A})=1- \sin\frac{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{M3}(\tilde{A})=1- \sqrt[k]{\frac{D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{M4}(\tilde{A})=1- \log^{\big(D_{M}(\tilde{A}, C_{\tilde{A}})+D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{M5}(\tilde{A})=2- 2^{D_{M}(\tilde{A}, C_{\tilde{A}})\cdot D_{M}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{HH}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{HH1}(\tilde{A})=1- \frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{HH2}(\tilde{A}, \tilde{B})= \sin\frac{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{HH3}(\tilde{A}, \tilde{B})=\sqrt[k]{\frac{D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{B}, C_{\tilde{B}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{HH4}(\tilde{A})=1- \log^{\big(D_{HH}(\tilde{A}, C_{\tilde{A}})+D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{HH5}(\tilde{A})=2- 2^{D_{HH}(\tilde{A}, C_{\tilde{A}})\cdot D_{HH}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} | |
D_{h}(\tilde{A}, \tilde{B}) | f_{1}(x, y)=\frac{x+y}{2} | K_{h1}(\tilde{A})=1- \frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2} |
f_{2}(x, y)=\sin\frac{(x+y)\pi}{4} | K_{h2}(\tilde{A})=1- \sin\frac{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)\pi}{4} | |
f_{3}(x, y)=\sqrt[k]{\frac{x+y}{2}}(k>1) | K_{h3}(\tilde{A})=1- \sqrt[k]{\frac{D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})}{2}}(k>1) | |
f_{4}(x, y)=\log^{x+y+1}_{3} | K_{h4}(\tilde{A})=1- \log^{\big(D_{h}(\tilde{A}, C_{\tilde{A}})+D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})\big)}_{3} | |
f_{5}(x, y)=2^{xy}-1 | K_{h5}(\tilde{A})=2- 2^{D_{h}(\tilde{A}, C_{\tilde{A}})\cdot D_{h}(\tilde{A}^{c}, C_{\tilde{A}^{c}})} |
C_1 | C_2 | C_3 | C_4 | |
\tilde{A}_1 | \langle[0.1, 0.2], [0.1, 0.2]\rangle | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.5, 0.5], [0.5, 0.5]\rangle |
\tilde{A}_2 | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.0, 0.0], [0.25, 0.75]\rangle | \langle[0.3, 0.4], [0.4, 0.6]\rangle |
\tilde{A}_3 | \langle[0.25, 0.5], [0.25, 0.5]\rangle | \langle[0.2, 0.4], [0.2, 0.4]\rangle | \langle[0.2, 0.3], [0.4, 0.7]\rangle | \langle[0.2, 0.3], [0.5, 0.6]\rangle |
\tilde{A}_4 | \langle[0.2, 0.3], [0.6, 0.7]\rangle | \langle[0.4, 0.7], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.5]\rangle | \langle[0.5, 0.7], [0.1, 0.3]\rangle |
Methods | Weighting vector | Alternatives ranking |
E_{JPCZ}(E1) [29] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{LZX}(E2) [30] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{WWZ}(E3) [31] | w={(0.2274, 0.2860, 0.2370, 0.2496)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
E_{ZJJL}(E4) [32] | w={(0.2366, 0.3159, 0.1859, 0.2616)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Das et al. [27] | w={(0.2506, 0.2211, 0.2472, 0.2811)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
Proposed method | w={(0.2527, 0.2312, 0.2554, 0.2608)^{T}} | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\phi(Z_{1}) | \phi(Z_{2}) | \phi(Z_{3}) | \phi(Z_{4}) | Alternatives order | |
\lambda=0.1 | 0.1325 | 0.0991 | 0.0763 | 0.1579 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.2 | 0.2980 | 0.2230 | 0.1716 | 0.3552 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.3 | 0.5109 | 0.3823 | 0.2941 | 0.6089 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.4 | 0.7948 | 0.5947 | 0.4576 | 0.9472 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.5 | 1.1922 | 0.8920 | 0.6863 | 1.4207 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.6 | 1.7882 | 1.3380 | 1.0295 | 2.1311 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.7 | 2.7817 | 2.0813 | 1.6015 | 3.3150 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.8 | 4.7686 | 3.5680 | 2.7454 | 5.6829 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
\lambda=0.9 | 10.7295 | 8.0280 | 6.1771 | 12.7865 | \tilde{A}_4\succ \tilde{A}_1\succ \tilde{A}_2\succ \tilde{A}_3 |
T_{1} | T_{2} | T_{3} | |
\lambda=0.1 | 0.0254 | 0.0334 | 0.0229 |
\lambda=0.2 | 0.0571 | 0.0750 | 0.0514 |
\lambda=0.3 | 0.0980 | 0.1286 | 0.0881 |
\lambda=0.4 | 0.1524 | 0.2001 | 0.1371 |
\lambda=0.5 | 0.2286 | 0.3002 | 0.2057 |
\lambda=0.6 | 0.3428 | 0.4502 | 0.3085 |
\lambda=0.7 | 0.5333 | 0.7004 | 0.4799 |
\lambda=0.8 | 0.9143 | 1.2007 | 0.8226 |
\lambda=0.9 | 2.0571 | 2.7015 | 1.8509 |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle |
A_2 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle |
A_4 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_3 | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_4 | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
c_1 | c_2 | c_3 | c_4 | |
A_1 | \langle[0.2, 0.5], [0.2, 0.3]\rangle | \langle[0.3, 0.5], [0.2, 0.3]\rangle | \langle[0.1, 0.3], [0.3, 0.5]\rangle | \langle[0.4, 0.5], [0.3, 0.5]\rangle |
A_2 | \langle[0.3, 0.4], [0.1, 0.4]\rangle | \langle[0.1, 0.4], [0.2, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.2, 0.4], [0.1, 0.2]\rangle |
A_3 | \langle[0.2, 0.4], [0.1, 0.2]\rangle | \langle[0.3, 0.4], [0.1, 0.2]\rangle | \langle[0.1, 0.3], [0.3, 0.4]\rangle | \langle[0.3, 0.4], [0.1, 0.4]\rangle |
A_4 | \langle[0.4, 0.5], [0.3, 0.5]\rangle | \langle[0.1, 0.5], [0.2, 0.5]\rangle | \langle[0.3, 0.4], [0.1, 0.3]\rangle | \langle[0.2, 0.5], [0.2, 0.3]\rangle |
e_1 | e_2 | e_3 | |
z_1 | \langle[0.2577, 0.4316], [0.2067, 0.3869]\rangle | \langle[0.2357, 0.4573], [0.2352, 0.3877]\rangle | \langle[0.2591, 0.4563], [0.2447, 0.3869]\rangle |
z_2 | \langle[0.2306, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2754, 0.4000], [0.1000, 0.2824]\rangle | \langle[0.2278, 0.4000], [0.1196, 0.2833]\rangle |
z_3 | \langle[0.2286, 0.3765], [0.1314, 0.2824]\rangle | \langle[0.2046, 0.3770], [0.1307, 0.2833]\rangle | \langle[0.2302, 0.3765], [0.1314, 0.2824]\rangle |
z_4 | \langle[0.2607, 0.4321], [0.2063, 0.3877]\rangle | \langle[0.3289, 0.4773], [0.2067, 0.3869]\rangle | \langle[0.2562, 0.4768], [0.1858, 0.3877]\rangle |
\varphi_1 | \varphi_2 | \varphi_3 | \varphi_4 | |
e_1 | 1.1613 | 1.4679 | 1.4623 | 1.1662 |
e_2 | 1.0949 | 1.7662 | 1.4051 | 1.3583 |
e_3 | 1.1328 | 1.5581 | 1.4660 | 1.2781 |
x_{1} (Temperature) | x_{2} (Cough) | x_{3} (Headache) | x_{4} (Stomach pain) | |
\tilde{M}_{1} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.5, 0.6], [0.2, 0.3]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle |
\tilde{M}_{2} | \langle[0.5, 0.6], [0.1, 0.3]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.6, 0.8], [0.1, 0.2]\rangle | \langle[0.4, 0.6], [0.1, 0.2]\rangle |
\tilde{M}_{3} | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.4, 0.6], [0.2, 0.4]\rangle | \langle[0.3, 0.5], [0.2, 0.4]\rangle |
\tilde{M}_{4} | \langle[0.8, 0.9], [0.0, 0.1]\rangle | \langle[0.7, 0.8], [0.1, 0.2]\rangle | \langle[0.7, 0.9], [0.0, 0.1]\rangle | \langle[0.8, 0.9], [0.0, 0.1]\rangle |
S(\tilde{M}_{1}, B) | S(\tilde{M}_{2}, B) | S(\tilde{M}_{3}, B) | S(\tilde{M}_{4}, B) | Recognition Result | |
S_{1} [11] | 0.73 | 0.80 | 0.78 | 0.73 | \tilde{M}_{2} |
S_{D} [6] | 0.82 | 0.91 | 0.86 | 0.84 | \tilde{M}_{2} |
S_{H1} | 0.32 | 0.36 | 0.33 | 0.28 | \tilde{M}_{2} |
S_{h1} | 0.38 | 0.45 | 0.43 | 0.34 | \tilde{M}_{2} |