
During the COVID-19 pandemic, identifying face masks with artificial intelligence was a crucial challenge for decision support systems. To address this challenge, we propose a quadratic Diophantine fuzzy decision-making model to rank artificial intelligence techniques for detecting masks, aiming to prevent the global spread of the disease. Our paper introduces the innovative concept of quadratic Diophantine fuzzy sets (QDFSs), which are advanced tools for modeling the uncertainty inherent in a given phenomenon. We investigate the structural properties of QDFSs and demonstrate that they generalize various fuzzy sets. In addition, we introduce essential algebraic operations, set-theoretical operations, and aggregation operators. Finally, we present a numerical case study that applies our proposed algorithms to select a unique face mask detection method and evaluate the effectiveness of our techniques. Our findings demonstrate the viability of our mask identification methodology during the COVID-19 outbreak.
Citation: Muhammad Danish Zia, Esmail Hassan Abdullatif Al-Sabri, Faisal Yousafzai, Murad-ul-Islam Khan, Rashad Ismail, Mohammed M. Khalaf. A study of quadratic Diophantine fuzzy sets with structural properties and their application in face mask detection during COVID-19[J]. AIMS Mathematics, 2023, 8(6): 14449-14474. doi: 10.3934/math.2023738
[1] | Sohail Ahmad, Ponam Basharat, Saleem Abdullah, Thongchai Botmart, Anuwat Jirawattanapanit . MABAC under non-linear diophantine fuzzy numbers: A new approach for emergency decision support systems. AIMS Mathematics, 2022, 7(10): 17699-17736. doi: 10.3934/math.2022975 |
[2] | Hanan Alohali, Muhammad Bilal Khan, Jorge E. Macías-Díaz, Fahad Sikander . On $ \left(\mathit{p}, \mathit{q}\right) $-fractional linear Diophantine fuzzy sets and their applications via MADM approach. AIMS Mathematics, 2024, 9(12): 35503-35532. doi: 10.3934/math.20241685 |
[3] | Muhammad Naeem, Muhammad Qiyas, Lazim Abdullah, Neelam Khan . Sine hyperbolic fractional orthotriple linear Diophantine fuzzy aggregation operator and its application in decision making. AIMS Mathematics, 2023, 8(5): 11916-11942. doi: 10.3934/math.2023602 |
[4] | Shahzaib Ashraf, Huzaira Razzaque, Muhammad Naeem, Thongchai Botmart . Spherical q-linear Diophantine fuzzy aggregation information: Application in decision support systems. AIMS Mathematics, 2023, 8(3): 6651-6681. doi: 10.3934/math.2023337 |
[5] | Ashraf Al-Quran . T-spherical linear Diophantine fuzzy aggregation operators for multiple attribute decision-making. AIMS Mathematics, 2023, 8(5): 12257-12286. doi: 10.3934/math.2023618 |
[6] | Dongzhihan Wang, Guijin Ma, Xiaorui Liu . An intelligent recognition framework of access control system with anti-spoofing function. AIMS Mathematics, 2022, 7(6): 10495-10512. doi: 10.3934/math.2022585 |
[7] | Rana Muhammad Zulqarnain, Xiao Long Xin, Muhammad Saeed . Extension of TOPSIS method under intuitionistic fuzzy hypersoft environment based on correlation coefficient and aggregation operators to solve decision making problem. AIMS Mathematics, 2021, 6(3): 2732-2755. doi: 10.3934/math.2021167 |
[8] | Muhammad Ali Khan, Saleem Abdullah, Alaa O. Almagrabi . Analysis of deep learning technique using a complex spherical fuzzy rough decision support model. AIMS Mathematics, 2023, 8(10): 23372-23402. doi: 10.3934/math.20231188 |
[9] | Farah Liyana Azizan, Saratha Sathasivam, Nurshazneem Roslan, Ahmad Deedat Ibrahim . Logic mining with hybridized 3-satisfiability fuzzy logic and harmony search algorithm in Hopfield neural network for Covid-19 death cases. AIMS Mathematics, 2024, 9(2): 3150-3173. doi: 10.3934/math.2024153 |
[10] | Muhammad Qiyas, Muhammad Naeem, Muneeza, Arzoo . Logarithmic cubic aggregation operators and their application in online study effect during Covid-19. AIMS Mathematics, 2023, 8(3): 5847-5878. doi: 10.3934/math.2023295 |
During the COVID-19 pandemic, identifying face masks with artificial intelligence was a crucial challenge for decision support systems. To address this challenge, we propose a quadratic Diophantine fuzzy decision-making model to rank artificial intelligence techniques for detecting masks, aiming to prevent the global spread of the disease. Our paper introduces the innovative concept of quadratic Diophantine fuzzy sets (QDFSs), which are advanced tools for modeling the uncertainty inherent in a given phenomenon. We investigate the structural properties of QDFSs and demonstrate that they generalize various fuzzy sets. In addition, we introduce essential algebraic operations, set-theoretical operations, and aggregation operators. Finally, we present a numerical case study that applies our proposed algorithms to select a unique face mask detection method and evaluate the effectiveness of our techniques. Our findings demonstrate the viability of our mask identification methodology during the COVID-19 outbreak.
The first outbreak of coronavirus disease (COVID-19) was initially reported in Wuhan, China, in December 2019 and was later identified as a new type of coronavirus by the World Health Organization (WHO). The rapid spread of the virus worldwide prompted the WHO to declare COVID-19 a public health emergency. As a result, the use of face masks has become increasingly popular in society as a means of protecting against the virus. While some individuals choose not to wear masks to protect themselves from polluted air [15], many believe that using face masks can help prevent the spread of COVID-19.
Computer science-powered deep learning and machine learning techniques have been shown to be valuable tools in the fight against COVID-19 [1]. Machine learning can analyze massive amounts of data, act as an early warning system for potential epidemics, and classify vulnerable groups. Several nations have implemented laws mandating the wearing of facial masks in public places, reflecting the importance of face mask detection as a crucial problem for COVID-19 prevention [12]. Face mask detection is a challenging technique, particularly when dealing with masked face-extracting features versus traditional facial-extracting features.
Various face mask detection strategies are available in the literature [13,21,31], but selecting the most appropriate strategy for a particular circumstance can be challenging due to decision-making in an uncertain environment. A decision-making process, as shown in Figure 1, is often required.
Conventional mathematical methods may not always be the optimal choice for addressing such problems. As a result, the application of fuzzy set (FS) theory is necessary, and a basic understanding of its fundamental concepts is required. The concept of FS was first introduced by Zadeh in 1965 [37] as a means to address ambiguity in natural language. Since its inception, research on FS theory has been extensively conducted in the fields of decision-making and operational research.
However, in many real-world scenarios, the membership function alone may not be sufficient to capture the complexities of the situation. To address this limitation, Atanassov introduced the concept of an intuitionistic fuzzy set (IFS) in his works [7,8,20]. This extension of FSs allows objects to possess nonmembership degrees alongside their membership degrees as long as their sum does not exceed one. Atanassov further proposed IF-relations on IFSs [9], which have gained significant attention and are now utilized in a wide range of practical applications, including decision-making [11,24] and optimization in intuitionistic fuzzy settings. The first lexicographic ordering of intuitionistic fuzzy values and their correlations was established by Feng et al. [16].
In certain real-world situations, the total of membership and nonmembership degrees of objects may exceed 1, posing a problem for traditional FS theory. To overcome this issue, Yager introduced Pythagorean fuzzy sets (PyFSs), where the sum of squares of membership and nonmembership degrees cannot exceed 1 [34,35]. PyFSs have been extensively studied using various methodologies [26,38], and Yager further extended the concept in [36] to introduce q-rung orthopair fuzzy sets (q-ROFSs), which have also been the subject of significant research [4,22]. Atanassov's type 2 IFSs [6] are also well-known in the field. Cuong introduced picture fuzzy sets (PFSs) as a way to handle fuzzy information consisting of hesitance or ignorance [10]. Recent research on PFSs and various aggregation operators can be found in [5,19,27], while [2,3,23,29] explored recent trends in fuzzy decision-making.
Fuzzy sets (FSs), interval-valued fuzzy sets (IFSs), Pythagorean fuzzy sets (PyFSs), and q-rung orthopair fuzzy sets (q-ROFSs) have been widely studied and applied in practice. However, these sets impose strict constraints on membership and nonmembership grades. To overcome these limitations, Riaz and Hashmi [30] proposed a new concept called the linear Diophantine fuzzy set (LDFS), which incorporates reference parameters for membership and nonmembership grades, thereby expanding the scope of these sets. Nevertheless, LDFS still contains acceptance- and rejection-type parameters, which may not fully capture fuzzy information related to denial, ignorance, or confusion. Therefore, the present study aims to introduce the notion of the quadratic Diophantine fuzzy set (QDFS), which includes a third parameter to address such scenarios. The goals of this paper are as follows:
∙ To propose the notion of QDFSs and their operational laws and explain their characteristics and comparison method.
∙ To compare QDFSs with existing FSs, such as IFSs, PyFSs, q-ROFSs, LDFSs, and PFSs, and demonstrate their superiority over them.
∙ To develop aggregation operators, such as the quadratic Diophantine fuzzy weighted averaging aggregation (QDFWAA) operator and the quadratic Diophantine fuzzy weighted geometric aggregation (QDFWGA) operator, and discuss their properties.
∙ To develop a new multiple attribute decision-making (MADM) method based on the proposed operators.
∙ To provide an illustrative example to demonstrate the flexibility and superiority of the proposed method.
This paper is organized as follows. Section 2 provides an introduction to FSs, IFSs, PyFSs, q-ROFSs, and LDFSs. In Section 3, we introduce the concept of QDFSs, compare them with existing FSs, and discuss algebraic and set-theoretical operations. Section 4 proposes aggregation operators, such as QDFWAA and QDFWGA, and examines their properties. Section 5 applies QDFSs to MADM using the proposed aggregation operators. Section 6 presents a comparative analysis of QDFSs with existing sets. Finally, in Section 7, we provide a comprehensive conclusion of the study.
This section presents fundamental concepts related to FS theory. Various types of FSs are explored, each with a distinct approach to handling uncertainty and ambiguity. The discussion commences with an overview of FSs and their extension, IFSs. This extension introduces a second function to describe the degree of nonmembership, in addition to the degree of membership. Next, PyFSs are examined, which provide a more precise method for expressing uncertainty and ambiguity. After that, attention is shifted to q-ROFSs, which offer a wider range of possibilities for communicating ambiguous information. Finally, the definition of LDFSs is presented.
As noted earlier, Lotfi A. Zadeh independently developed FSs in 1965 as an extension of classical set theory. In classical set theory, the membership of elements in a set is evaluated in binary terms; an element is either a member of the set or not. In contrast, FS theory enables a gradual evaluation of an element's membership in a set by utilizing a membership function with a real unit range value between 0 and 1.
Definition 2.1. ([37]) An FS AF on a universal set X is defined as
AF={(x,f(x)):x∈X} |
where f(x) is called membership function with f(x)∈[0,1].
IFSs represent an expansion of Lotfi Zadeh's FSs and were first introduced by Krassimir Atanassov, a Bulgarian mathematician, in 1983. While FSs rely on a membership function f(x) to define the degree of membership of an element in a set, IFSs extend this concept by introducing a second function g(x) that defines the degree of nonmembership of the element in the set. This additional degree of freedom enables the representation of uncertain and indeterminate information in a more effective and comprehensive manner.
Definition 2.2. ([8]) An IFS AIF on a universal set X is defined as
AIF={(x,f(x),g(x)):x∈X} |
where f(x) is called membership function and g(x) is called nonmembership function with f(x),g(x)∈[0,1], satisfying
0≤f(x)+g(x)≤1. |
Compared to IFSs, PyFSs, proposed by Yager, present a distinctive approach for expressing uncertainty and ambiguity with high precision and accuracy. This concept was explicitly designed to offer a structured framework for handling imprecision in real-life scenarios and to represent uncertainty and ambiguity mathematically.
Definition 2.3. ([34]) A PyFS APYF on a universal set X is defined as
APYF={(x,f(x),g(x)):x∈X} |
where f(x) is called membership function and g(x) is called nonmembership function with f(x),g(x)∈[0,1], satisfying
0≤f2(x)+g2(x)≤1. |
Yager introduced q-ROFs as a more powerful tool than IFSs and PyFSs for communicating ambiguous information. q-ROFs offer a broader range of possibilities due to the constraint that the total of the qth powers of the degrees of membership and nonmembership is no greater than one. This property ensures that q-ROFs are capable of conveying more complex and nuanced information than their FS counterparts.
Definition 2.4. ([36]) A q-ROFS Aq−ROF on a universal set X is defined as
Aq−ROF={(x,f(x),g(x)):x∈X} |
where f(x) is called membership function and g(x) is called nonmembership function with f(x),g(x)∈[0,1], satisfying
0≤fq(x)+gq(x)≤1,q≥1. |
LDFSs extend the space of previously specified sets by incorporating reference parameters that correspond to membership and nonmembership grades. In MADM, LDFSs represent an ideal mathematical framework since decision-makers can flexibly choose the degrees.
Definition 2.5. ([30]) An LDFS ALDF on a universal set X is defined as
ALDF={(x,<f(x),g(x)>,<α,β>):x∈X} |
where f(x) is called membership function and g(x) is called nonmembership function and α,β are reference parameters with f(x),g(x)∈[0,1], satisfying
0≤αf(x)+βg(x)≤1, |
0≤α+β≤1. |
In this section, we introduce the concept of QDFS, which is inspired by the idea of a general quadratic Diophantine equation in two variables. Specifically, for two variables x and y, a general second-order Diophantine equation takes the form
αx2+βxy+γy2=k. |
Using this equation as motivation, we define QDFS as follows:
Definition 3.1. Let X be a nonempty universal set. A QDFS AQDF on X is an object of the form
AQDF=(x,<f(x),g(x)>,<α,β,γ>):x∈X |
where f(x),g(x) are, respectively, the membership and nonmembership functions and α,β,γ are reference parameters such that f(x),g(x),α,β,γ∈[0,1]. The functions must satisfy the conditions
0≤αf2(x)+βf(x)g(x)+γg2(x)≤1and0≤α+β+γ≤1. |
A quadratic Diophantine fuzzy number (QDFN) refers to the pair (<f,g>,<α,β,γ>).
The QDFS framework is highly relevant in situations where fuzzy information about acceptance (α), rejection (β), and hesitation or ignorance (γ) is needed. The pair (<f,g>,<α,β,γ>) is a QDFN, and the feasible space for the reference parameters α, β, and γ is illustrated in Figure 2 as a prism with the constraint 0≤α+β+γ≤1.
To provide further insight, Figure 3 shows the feasible regions of QDFNs for specific values of the reference parameters. For instance, Figure 3(a)–(c) illustrates scenarios where the reference parameters of acceptance, rejection, and hesitation are equal to 1, respectively. In contrast, Figure 3(d)–(f) corresponds to situations where any two reference parameters are equal to 0.5.
Now, we demonstrate that the QDFS framework provides a larger feasible space for selecting membership and nonmembership values than IFSs, PyFSs, and q-ROFSs. To support this claim, we present the following theorem and then provide graphical visualizations of these sets (Figure 4).
Theorem 3.2. The space of QDFNs is larger than the spaces of intuitionistic fuzzy numbers (IFNs), Pythagorean fuzzy numbers (PyFNs), and q-rung orthopair fuzzy numbers (q-ROFNs).
Proof. Let (<f,g>,<α,β,γ>) be a QDFN, then the inequality
0≤αf2(x)+βf(x)g(x)+γg2(x)≤1 and 0≤α+β+γ≤1, |
for β=0 and arbitrary choice of α and γ holds for every IFN and PyFN. Hence, every IFN and PyFN is also a QDFN.
A QDFN with a given set of parameters may not necessarily be an IFN or PyFN.
For example, let f(x)=0.72 and g(x)=0.8, then
f(x)+g(x)=1.52>1 |
and
f2(x)+g2(x)=1.16>1 |
but α=0.36, β=0.27, and γ=0.18; we have
αf2(x)+βf(x)g(x)+γg2(x)=0.46<1. |
Similarly, it is easy to check that for a q-ROFS, whenever f(x)≈g(x)→1, then q→∞.
For a special case f(x)=g(x)=1, there does not exist any specific q; for this special case, no q-ROFN exists; however, for any choice of α, β, and γ such that 0≤α+β+γ≤1, we have
αf2(x)+βf(x)g(x)+γg2(x)=α+β+γ≤1. |
For illustration, let f(x) = g(x) = 1 and q≥1, then
f(x)q+g(x)q=2>1. |
Hence, the pair (<1,1>) does not belong to any q-ROFS.
But, for α=0.36, β=0.27, and γ=0.18, we have
αf2(x)+βf(x)g(x)+γg2(x)=0.36(1)2+0.27(1)(1)+0.18(1)2=0.81<1 |
therefore, (<1,1>,<0.36,0.27,0.18>) is a QDFN.
So, it concludes that the space of QDFN consists of more points than the spaces of IFN and PyFN, providing more freedom to assign values to f and g.
From the above arguments, it is evident that the QDFS is comparatively more feasible than other types of FSs.
Figure 5 shows that the feasible spaces of QDFS and LDFS are the same due to the reference parameters. This means that both QDFS and LDFS have reference parameters that enable the complete use of space.
However, QDFS is a more comprehensive tool than LDFS since it incorporates an additional parameter, γ, which caters to the hesitant type, allowing the phenomenon to be ignored. This feature is particularly beneficial in real-world scenarios where decision-makers may be indecisive or uncertain.
For example, in a real estate investment scenario, there may be three possibilities: investing in housing society X (acceptance type), investing in some other housing society (rejection type), or not investing in a housing society at all (ignorance type). As LDFS is unable to deal with such type of information, QDFS is the preferred tool for further investigation. In summary, QDFS offers a more comprehensive solution space than LDFS, making it a more useful tool for decision-making in situations where uncertain or imprecise information is present.
In Subsection 3.1, it is shown that QDFS is advantageous over LDFS because of its ability to tackle hesitation-type scenarios. In the literature, there exists a type of FS named PFS proposed by Cuong [10], which also covers the hesitation-type fuzzy information. Cuong defined the PFS as follows:
Definition 3.3. A PFS APF on a universal set X is defined as
APF={(x,<f(x),h(x),g(x)>):x∈X} |
where f(x) is called membership function, h(x) is called neutral membership function, and g(x) is called nonmembership function with f(x),h(x),g(x)∈[0,1], satisfying 0≤f(x)+h(x)+g(x)≤1.
In Definition 3.3, the PFS imposes a limitation on membership, neutral, and nonmembership functions that their sum must not exceed 1; as a result, the feasible space of PFS becomes restricted. For example, if f(x)=0.43, h(x)=0.52, and g(x)=0.21, then we have
f(x)+h(x)+g(x)=0.43+0.52+0.21=1.16>1. |
Hence, such f,g, and h do not represent the PFN. Now, in QDFS, for α=0.43, β=0.52, and γ=0.21 and the pair F(x)=0.62 and G(x)=0.51, we have
αF2(x)+βF(x)G(x)+γG2(x)=0.38<1. |
So, there exist numbers that are not PFNs but are QDFNs. By following the same arguments as in Theorem 3.2, it is easy to show that every PFN is in fact a QDFN.
From the above comparisons, it is evident that the proposed QDFS is a hybrid type of FS that combines the characteristics of LDFS and PFS. Figure 6 below effectively illustrates how QDFS extends and generalizes other fuzzy set extensions, making it a versatile option for accurate and flexible applications.
Now, we proceed to define the algebraic operations on QDFSs in order to study their structural properties and establish aggregation operators. Specifically, we define the operations as follows:
Let A and B be two QDFSs and λ>0. Then,
● A⊆B iff fA≤fB,gA≥gB,αA≤αB,βA≤βB,γA≥γB.
● A=B iff A⊆B and B⊆A.
● A∪B={(x,<max(fA,fB),min(gA,gB)>,<max(αA,αB),min(βA,βB),min(γA,γB)>)}.
● A∩B={(x,<min(fA,fB),max(gA,gB)>,<min(αA,αB),min(βA,βB),max(γA,γB)>)}.
● Ac={(x,<gA,fA>,<γ,β,α>)}
● A⊕B={(x,<fA+fB−fAfB,gAgB>,<αA+αB−αAαB,βAβB,γAγB+γAβB+γBβA>)}
● A⊗B={(x,<fAfB,gA+gB−gAgB>,<αAαB+αAβB+αBβA,βAβB,γA+γB−γAγB>)}
● λA={(x,<1−(1−fA)λ,gλA>,<1−(1−αA)λ,βλA,(γA+βA)λ−βλA>)}
● Aλ={(x,<fλA,1−(1−gA)λ>,<(αA+βA)λ−βλA,βλA,1−(1−γA)λ>)}.
The proposed algebraic operations can be illustrated with the following example.
Example 3.4. Let A=x,<0.7,0.15>,<0.1,0.18,0.36>,
B=x,<0.44,0.08>,<0.05,0.49,0.14>, and λ=2, then
A∪B={(x,<0.7,0.08>,<0.1,0.18,0.14>)}.
A∩B={(x,<0.44,0.15>,<0.05,0.18,0.36>)}.
Ac={(x,<0.15,0.7>,<0.36,0.18,0.1>)}
A⊕B={(x,<0.83,0.01>,<0.15,0.09,0.25>)}
A⊗B={(x,<0.31,0.22>,<0.43,0.09,0.45>)}
2A={(x,<0.91,0.02>,<0.19,0.03,0.03>)}
Aλ={(x,<0.49,0.28>,<(0.27,0.03,0.59>)}.
The algebra of sets provides a framework for developing and describing the set-theoretic operations of union, intersection, and complementation, as well as the relationships between set equality and set inclusion. Moreover, this algebra furnishes systematic approaches for computing computations utilizing these operations and relations, as well as for evaluating expressions.
For any QDFS A,B, and C, the following set-theoretical laws hold:
a. if A⊆B and B⊆C then A⊆C (law of trichotomy)
b. A∩B=B∩A (commutativity of intersection)
A∪B=B∪A (commutativity of union)
c. (A∩B)∩C=A∩(B∩C) (associativity of intersection)
(A∪B)∪C=A∪(B∪C) (associativity of union)
d. (A∩B)∪C=(A∩B)∪(A∩C) (distributive law of intersection over union)
(A∪B)∩C=(A∪B)∩(A∪C) (distributive law of union over intersection)
e. (A∪B)c=Ac∩Bc (De Morgan's law)
In this section, we introduce the score and accuracy functions, which are essential for developing the ordering between QDFNs. The score function for a QDFN is defined as follows:
Definition 3.5. Let Q=(<fQ,gQ>,<αQ,βQ,γQ>) be a QDFN, then the score function on Q can be defined by the mapping S(Q)→[−1,1]
S(Q)=(fQ−gQ)+(α+β−γ)2. |
Definition 3.6. The accuracy function on Q can be defined by the mapping A(Q)→[0,1]
A(Q)=(fQ+gQ)+(α+β+γ)2. |
To compare two QDFNs Q1 and Q2, we employ score and accuracy function with the following criteria:
● If S(Q1)≤S(Q2), then Q1≤Q2.
● If S(Q1)≥S(Q2), then Q1≥Q2.
● If S(Q1)=S(Q2), then
. if A(Q1)≤A(Q2), then Q1≤Q2.
. if A(Q1)≥A(Q2), then Q1≥Q2.
. if A(Q1)=A(Q2), then Q1=Q2.
Example 3.7. Let Q1=(<0.41,0.17>,<0.1,0.38,0.17>) and Q2=(<0.54,0.23>,<0.21,0.36,0.33>) be two QDFNs, then
S(Q1)=(0.41−0.17)+(0.1+0.38−0.17)2=0.275 |
and
S(Q2)=(0.54−0.23)+(0.21+0.36−0.33)2=0.275 |
as S(Q1)=S(Q2), so check accuracy function, now
A(Q1)=(0.41+0.17)+(0.1+0.38+0.17)2=0.615 |
and
A(Q2)=(0.54+0.23)+(0.21+0.36+0.33)2=0.835 |
since A(Q1)≤A(Q2) hence Q1≤Q2.
This section is dedicated to the introduction of two new aggregation operators, specifically designed for a collection Ω of QDFNs. We refer to these operators as QDFWAA and QDFWGA, respectively. We provide an exhaustive description of the properties of these operators.
Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩) (j=1,2,...,n) be a collection of CLDFNs. Then, the QDFWAA operator can be denoted and defined by
QWA(q1,...,qn)=w1q1⊕w2q2⊕...⊕wnqn, |
where wj is the weight of qj satisfying wj∈[0,1] and n∑j=1wj=1. In particular, if w=(1/n,1/n,1/n,...,1/n)T, then the QWA operator reduces to quadratic Diophantine fuzzy averaging aggregation (QDFAA) operator:
QA(q1,...,qn)=1n(q1⊕q2⊕...⊕qn). |
Theorem 4.1. Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩) (j=1,2,...,n) be a collection of QDFNs, wj is the weight of qj satisfying wj∈[0,1] and n∑j=1wj=1. Then the fused result from QWA operator is also a QDFN and
QWA(q1,...,qn)=(⟨(1−n∏j=1(1−fqj)wj),n∏j=1gwjqj⟩,⟨1−n∏j=1(1−αqj)wj,n∏j=1βwjqj,n∏j=1(γqj+βqj)wj−n∏j=1βwjqj⟩). | (4.1) |
Proof. By employing algebraic operations presented in Subsection 3.3, one can easily verify the first result. Now we prove (4.1) by using mathematical induction on n.
When n=2, we have
QWA(q1,q2)=w1q1⊕w2q2. |
Since
w1q1=(⟨(1−(1−fq1)w1),gw1q1⟩,⟨1−(1−αq1)w1,βw1q1,(γq1+βq1)w1−βw1q1⟩), |
w2q2=(⟨(1−(1−fq2)w2),gw2q2⟩,⟨1−(1−αq2)w2,βw2q2,(γq2+βq2)w2−βw2q2⟩). |
Then
QWA(q1,q2)=w1q1⊕w2q2=(⟨1−(1−fq1)w1(1−fq2)w2,gw1q1gw2q2⟩,⟨1−(1−αq1)w1(1−αq2)w2,βw1q1βw2q2,(γq1+βq1)w1(γq2+βq2)w2−βw1q1βw2q2⟩), |
Suppose (4.1) holds for n=k, that is
QWA(q1,q2,...,qk)=w1q1⊕w2q2⊕...⊕wkqk=(⟨1−k∏j=1(1−fqj)wj,k∏j=1gwjqj⟩,⟨1−k∏j=1(1−αqj)wj,k∏j=1βwjqj,n∏j=1(γqj+βqj)wj−k∏j=1βwjqj⟩). |
Then for n=k+1, we have
1−k∏j=1(1−fqj)wj+(1−(1−fqk+1)wk+1)−(1−k∏j=1(1−fqj)wj)(1−(1−fqk+1)wk+1)=1−k+1∏j=1(1−fqj)wj. |
We have
QWA(q1,...,qk+1)=w1q1⊕w2q2⊕...⊕wkqk⊕wk+1qk+1=(w1q1⊕w2q2⊕...⊕wkqk)⊕wk+1qk+1=(⟨1−k+1∏j=1(1−fqj)wj,k+1∏j=1gwjqj⟩,⟨1−k+1∏j=1(1−αqj)wj,k+1∏j=1βwjqj,k+1∏j=1(γqj+βqj)wj−k+1∏j=1βwjqj⟩). |
Proved.
Theorem 4.2. Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩)
and qj∗=(⟨fqj∗,gqj∗⟩,⟨αqj∗,βqj∗,γqj∗⟩) (j=1,2,...,n) be two collections of QDFNs. Then the QWA operator satisfies the following properties.
i) Idempotency: If all QDFNs qj (j=1,2,...,n) are equal, that is, qj=q for all j, then
QWA(q1,...,qn)=q.
ii) Monotonicity: For qj and qj∗ (j=1,2,...,n), if qj≤qj∗ for all j, then
QWA(q1,...,qn)≤QWA(q1∗,q2∗,...,qn∗). |
iii) Boundedness: Let qj (j=1,2,...,n) be a collection of QDFNs and q−=minj{qj},q+=maxj{qj}, then
q−≤QWA(q1,...,qn)≤q+. |
Proof. i) Idempotency: Let q=(⟨fq,gq⟩,⟨αq,βq,γq⟩). Then qj=q yield
QWA(q1,...,qn)=(⟨1−n∏j=1(1−fqj)wj,n∏j=1gwjqj⟩,⟨1−n∏j=1(1−αqj)wj,n∏j=1βwjqj,n∏j=1(γqj+βqj)wj−n∏j=1βwjqj⟩)=(⟨(1−n∏j=1(1−fq)wj),n∏j=1gwjq⟩,⟨1−n∏j=1(1−αq)wj,n∏j=1βwjq,n∏j=1(γq+βq)wj−n∏j=1βwjq⟩)=(⟨(1−(1−fq)n∑j=1wj),gn∑j=1wjq⟩,⟨1−(1−αq)n∑j=1wj,βn∑j=1wjq,(γq+βq)n∑j=1wj−βn∑j=1wjq⟩)=(⟨(1−(1−fq)),gq⟩,⟨1−(1−αq),βq,(γq+βq)−βq⟩)=(⟨fq,gq⟩,⟨αq,βq,γq⟩)=q. |
Monotonicity: As qj≤qj∗, so fqj≤fqj∗,αqj≤αqj∗ and gqj≥gqj∗,βqj≥βqj∗,γqj≥γqj∗ for any j, we have
(1−fqj)wj≥(1−fqj∗)wj,gwjLj≥gwjqj∗. |
Moreover
1−n∏j=1(1−fqj)wj≤1−n∏j=1(1−fqj∗)wj, n∏j=1gwjqj≥n∏j=1gwjqj∗. |
Therefore
1−n∏j=1(1−fqj)wj−n∏j=1gwjqj≤1−n∏j=1(1−fqj∗)wj−n∏j=1gwjqj∗. | (4.2) |
If q=QWA(q1,q2,,,qn) and q∗=QWA(q1∗,q2∗,,,qn∗). Then by (4.2) and Definition 3.5, we get S(q)≤S(q∗). If S(q)=S(q∗), then it is easy to show that A(q)=A(q∗).
Boundedness: Idempotency and monotonicity of QWA operator implies the following:
For qj≥q−=minj{qj}, we have
QWA(q1,q2,,,qn)≥QWA(q−,q−,,,q−)=q−. |
For qj≤q+=maxj{qj}, we have
QWA(q1,q2,,,qn)≤QWA(q+,q+,,,q+)=q+. |
Hence q−≤QWA(q1,q2,,,qn)≤q+.
Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩) (j=1,2,...,n) be a collection of QDFNs. Then, the QDFWGA operator can be denoted and defined by
QWG(q1,...,qn)=qw11⊗qw22⊗...⊗qwnn, |
where wj is the weight of qj satisfying wj∈[0,1] and n∑j=1wj=1. In particular, if w=(1/n,1/n,1/n,...,1/n)T, then the QWG operator reduces to quadratic Diophantine fuzzy geometric aggregation (QDFGA) operator:
QDFGA(q1,...,qn)=(q1⊗q2⊗...⊗qn)1n. |
Theorem 4.3. Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩) (j=1,2,...,n) be a collection of QDFNs, wj is the weight of qj satisfying wj∈[0,1] and n∑j=1wj=1. Then, the fused result from QWG operator is also a QDFN and
QWG(q1,...,qn)=(⟨n∏j=1fwjqj,(1−n∏j=1(1−gqj)wj)⟩,⟨n∏j=1(αqj+βqj)wj−n∏j=1βwjqj,n∏j=1αwjqj,1−n∏j=1(1−γqj)wj⟩). |
Proof. Straight forward.
Theorem 4.4. Let qj=(⟨fqj,gqj⟩,⟨αqj,βqj,γqj⟩)\\ and qk=(⟨fqj∗,gqj∗⟩,⟨αqj∗,βqj∗,γqj∗⟩) (j=1,2,...,n) be two collections of QDFNs. Then, the QWG operator satisfies the following properties:
i) Idempotency: If all QDFNs qj (j=1,2,...,n) are equal, that is, qj=L for all j, then
QWG(q1,...,qn)=L. |
ii) Monotonicity: For qj and qj∗ (j=1,2,...,n), if qj≤qj∗for all j, then
QWG(q1,...,qn)≤QWG(q1∗,q2∗,...,qn∗). |
iii) Boundedness: Let qj (j=1,2,...,n) be a collection of QDFNs and q−=minj{qj},q+=maxj{qj}, then
q−≤QWG(q1,...,qn)≤q+. |
Proof. The proof is similar to the proof of Theorem 4.2.
In this section, we present a new MADM technique that utilizes quadratic Diophantine fuzzy information and is based on the proposed aggregation operators. To address the challenges associated with complex decision-making problems, we consider a collection of alternatives and a collection of attributes with corresponding weight vectors to evaluate the effectiveness and reliability of the proposed approach.
Let X=X1,X2,...,Xn denote the collection of attributes, Y=Y1,Y2,...,Ym denote the collection of alternatives, and w=w1,w2,...,wn denote the associated attribute weights, where ∑nj=1wj=1. For each alternative Yi and attribute Xj, we use q′ij=(⟨fqij,gqij⟩,⟨αqij,βqij,γqij⟩) to denote the QDFN that represents the characteristic of Yi with respect to Xj. By using a quadratic Diophantine fuzzy decision matrix (QDFDM), we can store the characteristics of all alternatives in relation to the attributes.
In this setting, we propose two aggregation operators, namely, QDFWAA and QDFWGA, to evaluate the overall performance of each alternative based on the given attributes and their corresponding weights. The properties of these operators are described in detail.
The following are the steps involved in the algorithm used for decision-making using QDFS:
Step 1: Using the following transformation, normalize the QDFDM:
qij={q′ij, for benefit attribute Xj¯q′ij, for cost attribute Xj. |
Step 2: Make use of the QWA operator:
∗qi=QWA(qi1,qi2,...,qin), (i=1,2,3,...,m) |
or the QWG operator:
∗∗qi=QWG(qi1,qi2,...,qin), (i=1,2,3,...,m). |
Step 3: Calculate the scores S(∗qi) (or S(∗∗qi)) of the overall attribute value ∗qi (or ∗∗qi) of the alternative Yi using the score function.
Step 4: Sort the alternatives according to their scores. If two scores are identical, the accuracy degree should be used to order the alternatives.
Step 5: Choose the alternative with the highest score (accuracy degree).
Figure 7 shows a flowchart for the aforementioned procedure for more explanation.
Since the discovery of the novel coronavirus, various countries' governments have implemented disciplinary measures to limit its spread. In the face of pandemics, the use of modern technology has become increasingly prevalent in enhancing the operations and performance of numerous organizations. One such technology is face mask detection, which aids in monitoring social distancing and, with artificial intelligence interpretation, identifying the use of face masks.
The WHO has issued several guidelines to mitigate the devastating consequences of COVID-19's spread on both society and the economy. The virus has spread to the point of becoming a global health hazard. These guidelines include the use of face masks, maintaining social distancing, and adopting a virtual workplace culture. Of all these guidelines, face mask detection is a crucial approach that could help determine the number of people wearing masks, maintain an appropriate social distance, and protect a significant population from the virus's severity.
The use of face mask technology often involves computer vision and deep learning. To utilize this technology, the user must subscribe to a face mask alert app on their smartphone or other electronic devices and connect their camera to the app. When someone nearby is not wearing a mask, the user must grant permission to receive warnings or notifications. In addition, users can report violations of mask-wearing rules to the administrator.
It is evident that the most appropriate algorithm selection for face mask identification is the traditional MADM problem. In this subsection, we present an example of algorithm selection for face mask detection with QDF information to demonstrate the approach provided in this work. An IT expert must choose the most suitable face mask detection algorithm from four preexisting algorithms, as follows:
Y1: Convolutional Neural Network (CNN) CNN is a form of neural network. CNN is effective in computer vision applications such as pattern classification due to its cheap computation complexity and ability to retrieve location data. CNN eliminates top-level features by combining convolutional sections with primary pictures. Advanced training algorithms are used in this technique to recognize numerous facial traits and detect whether or not a person is wearing a face mask. Face masks are recognized in real time by the technology, which helps limit the transmission rate. The system also has digital capabilities for receptionists and physicians.
Y2: Region-Based Convolutional Neural Network (R-CNN) The first CNN-focused two-stage object identification approach was R-CNN [25]. R-design CNNs are composed of three distinct blocks. Each source image uses a selective search strategy to produce class-independent regional notions during the first step. To acquire feature vectors of the needed length from each zone recommendation, the second block applies CNN with five convolutional layers and two fully connected layers. Following that, each region suggestion is fed into a distinct CNN to generate feature maps of a predetermined length. The last block uses a linear support vector machine for a specified group to categorize each area suggestion.
Y3: Faster Region-Based Convolutional Neural Network (Faster-R-CNN) Ren et al. [28] presented Faster-R-CNN as a substitute for the suggestion technique; nevertheless, they have since changed to a region proposal network (RPN) [17]. The RPN is just a fully convolutional network (FCN) that accepts any size image and produces a flurry of rectangular-shaped item recommendations. Every object concept has an object class score rate, which determines whether or not the proposal comprises an object.
Y4: Mask Region-Based Convolutional Neural Network (Mask R-CNN) Mask R-CNN, developed by Kaiming He [18], creates region proposals by utilizing the proposal networks of Faster-R-regional CNN. Instead of a pooling layer for regions of interest (RoI), the technique uses an RoI align layer on area suggestions to align recovered features with the input position of the object. The aligned RoIs are then sent through the Mask R-final CNN stage, which produces three outputs: a bounding box offset, a class label, and a binary object mask. To mask each RoI, a simple fully convolutional neural network is utilized.
The attributes considered to evaluate these alternatives are X1=Accuracy, X2= Simplicity, X3=Effectiveness, and X4=Scalability. The parameter α represents the expert's level of belief that an algorithm should have a specific attribute, the parameter β denotes the expert's level of confidence that an algorithm should not have a specific attribute, and the parameter γ represents the expert's degree of doubt of the specific attribute. Face mask detection algorithms are to be evaluated using the QDF information under the above four attributes. (Assume the attribute index's weighting vector is (0.20,0.16,0.37,0.27).)
Now, we make use of the algorithm proposed in Subsection 4.3.
Step 1: The expert information in the form of QDFS is shown in Table 1. As all attributes are of benefit type, there is no need to normalize QDFDM.
X1 | X2 | |
Y1 | (⟨0.58,0.69⟩,⟨0.14,0.49,0.21⟩) | (⟨0.47,0.92⟩,⟨0.28,0.28,0.33⟩) |
Y2 | (⟨0.77,0.30⟩,⟨0.15,0.19,0.54⟩) | (⟨0.80,0.16⟩,⟨0.31,0.15,0.14⟩) |
Y3 | (⟨0.46,0.67⟩,⟨0.26,0.14,0.40⟩) | (⟨0.73,0.94⟩,⟨0.33,0.22,0.43⟩) |
Y4 | (⟨0.31,0.73⟩,⟨0.42,0.16,0.41⟩) | (⟨0.75,0.91⟩,⟨0.19,0.44,0.30⟩) |
X3 | X4 | |
Y1 | (⟨0.62,0.87⟩,⟨0.19,0.12,0.59⟩) | (⟨0.66,0.31⟩,⟨0.60,0.11,0.26⟩) |
Y2 | (⟨0.62,0.26⟩,⟨0.19,0.18,0.50⟩) | (⟨0.49,0.60⟩,⟨0.23,0.21,0.37⟩) |
Y3 | (⟨0.48,0.18⟩,⟨0.63,0.18,0.12⟩) | (⟨0.30,0.11⟩,⟨0.66,0.21,0.10⟩) |
Y4 | (⟨0.90,0.56⟩,⟨0.12,0.15,0.19⟩) | (⟨0.65,0.58⟩,⟨0.64,0.13,0.11⟩) |
Step 2: Now utilize the proposed aggregation operators QWA and QWG to calculate the aggregated QDF data shown in Table 2.
QWA | QWG | |
Y1 | (⟨0.60,0.63⟩,⟨0.34,0.18,0.40⟩) | (⟨0.60,0.78⟩,⟨0.31,0.26,0.41⟩) |
Y2 | (⟨0.66,0.31⟩,⟨0.21,0.18,0.39⟩) | (⟨0.63,0.37⟩,⟨0.21,0.21,0.43⟩) |
Y3 | (⟨0.49,0.27⟩,⟨0.54,0.18,0.20⟩) | (⟨0.45,0.54⟩,⟨0.49,0.48,0.23⟩) |
Y4 | (⟨0.76,0.64⟩,⟨0.37,0.17,0.21⟩) | (⟨0.65,0.69⟩,⟨0.30,0.26,0.24⟩) |
Step 3: The scores of attributes are computed in Table 3 and aggregate scores of alternatives are computed in Table 4.
Score of X1 | Score of X2 | Score of X3 | Score of X4 | |
Y1 | 0.15 | 0.06 | −0.23 | 0.23 |
Y2 | 0.19 | 0.23 | −0.16 | −0.28 |
Y3 | 0.11 (accuracy=0.38) | 0.21 | −0.15 | 0.11 (accuracy=0.54) |
Y4 | 0.22 | −0.03 | −0.16 | 0.07 |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | 0.09 | 4th | 0.14 (accuracy=0.50) | 3rd |
Y2 | 0.13 | 2nd | 0.12 | 4th |
Y3 | 0.15 | 1st | 0.17 | 1st |
Y4 | 0.11 | 3rd | 0.14 (accuracy=0.60) | 2nd |
Step 4: The rankings of alternatives for QWA and QWG are given in Table 4.
Step 5: Since the alternative Y3 has the highest score for both QWA and QWG, the IT expert should select the Faster-R-CNN.
The problem considered above is solved by asymmetric weights based upon personal choice, and the same problem could also be solved for symmetric weights [33] based upon normal distribution; i.e., the attribute index's weighting vector is (0.16,0.35,0.35,0.16)). The changes are highlighted in Tables 5 and 6.
QWA | QWG | |
Y1 | (⟨0.58,0.72⟩,⟨0.30,0.19,0.41⟩) | (⟨0.56,0.84⟩,⟨0.29,0.24,0.42⟩) |
Y2 | (⟨0.71,0.25⟩,⟨0.24,0.17,0.33⟩) | (⟨0.67,0.31⟩,⟨0.23,0.22,0.39⟩) |
Y3 | (⟨0.57,0.35⟩,⟨0.51,0.18,0.24⟩) | (⟨0.50,0.71⟩,⟨0.45,0.44,0.29⟩) |
Y4 | (⟨0.78,0.69⟩,⟨0.31,0.21,0.24⟩) | (⟨0.67,0.77⟩,⟨0.27,0.22,0.26⟩) |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | −0.09 | 4th | −0.085 | 4th |
Y2 | 0.27 | 2nd | 0.21 | 1st |
Y3 | 0.34 | 1st | 0.20 | 2nd |
Y4 | 0.19 | 3rd | 0.07 | 3rd |
} In this section, we present a sensitivity analysis to assess the impact of the aggregation operators and weights used during the decision-making process on the proposed method. As demonstrated in the previous section, the results indicate that the alternative Y3 is consistently ranked as the top choice for all aggregation operators, regardless of the weights employed. However, when QWG with normally distributed weights is utilized, Y3 is ranked as the second-best option. Based on this information, Y3 is designated as the optimal selection.
On the other hand, the alternative Y1 persistently occupies the fourth rank for all aggregation operators, except for QWG with user-defined weights, where it is positioned second to last. Therefore, it can be inferred that Y1 is not a viable choice for this decision-making process and should be excluded from consideration.
In addition, it is noteworthy that the rankings of the alternatives remain unaffected, even when the weights are modified for the operator QWA. The rankings of the alternatives for the various aggregation operators considered are presented in Table 7 for further analysis.
Weights | Rankings | |
QWA | User defined | Y3≤Y2≤Y4≤Y1 |
QWG | User defined | Y3≤Y4≤Y1≤Y2 |
QWA | Normally distributed | Y3≤Y2≤Y4≤Y1 |
QWG | Normally distributed | Y2≤Y3≤Y4≤Y1 |
The proposed method is a versatile tool that has been designed to accommodate a wide range of input data. It is widely acknowledged that data can manifest in diverse shapes and sizes, and finding a model that can effectively handle such variations can be challenging. Nonetheless, the proposed model can handle uncertainties and covers an extensive spectrum of data types, including FSs, IFSs, PFSs, PyFSs, and q-ROFSs.
The proposed model possesses a unique feature of reference parameters, which play a critical role in expanding the membership and nonmembership grades' space. By altering the physical interpretation of these parameters, the model can be applied efficiently in different scenarios, imparting greater flexibility and adaptability to real-life situations.
In the context of MADM problems, we encounter varying types of criteria and input data, depending on the situation. For instance, investment opportunities may have different criteria and input data than recruitment or vendor selection. In such cases, the proposed QDFS model has proven to be highly effective. It is a straightforward and easy-to-understand method that can be applied seamlessly to different alternatives and attributes.
The model's simplicity and accessibility make it user-friendly, even for individuals who lack expertise in data analysis. With the proposed QDFS model, decision-makers can easily analyze various alternatives and make well-informed choices, thus saving time, resources, and costs by avoiding ineffective decisions.
Therefore, the proposed method is a powerful and adaptable tool that can handle different types of input data. Its unique feature of reference parameters and simplicity make it a highly effective tool for solving MADM problems. By employing the proposed QDFS model, organizations can make informed decisions, optimize their resources, and save time. Moreover, since the proposed algorithm yields nearly identical results under various weights and aggregation operators, it is considered valid and recommended for decision-making in uncertain situations. The validation of the algorithm is also illustrated in Figure 8.
In this subsection, we present a qualitative comparison of the proposed QDFS method with existing techniques. Table 8 provides a comparative analysis of the proposed QDFS model and several FS extensions.
Collection of sets | Acceptance Information | Rejection Information | Ignorance Information | Parameterization | Unrestricted Domain |
FS | Yes | No | No | No | No |
IFS | Yes | Yes | No | No | No |
PyFS | Yes | Yes | No | No | No |
q-ROFS | Yes | Yes | No | No | No |
PFS | Yes | Yes | Yes | No | No |
LDFS | Yes | Yes | No | Yes | Yes |
QDFS | Yes | Yes | Yes | Yes | Yes |
It is evident that the QDFS method proposed in this study provides a higher degree of flexibility and autonomy in handling fuzzy information.
The present article introduces QDFS, which is an effective tool for analyzing fuzzy information while taking into account the reference parameters of acceptance, ignorance, and rejection type. The complete framework of QDFS, including its algebraic operations, comparison functions, and aggregation operators, is described in detail. To further illustrate the versatility of the proposed approach, a numerical example is presented. In addition, detailed comparisons between the suggested method and current methods are also provided.
The proposed approach assumes the weights of attributes and experts, which are determined by decision specialists in advance. However, we observed that our approach only considers subjective factors for the weight of experts and does not consider the weight information generated from the decision matrix. Therefore, we recommend exploring the associative weight method to handle actual decision-making problems where the weight of attributes needs to be determined.
Future work will involve adapting the suggested approach to various environments and applying it to the disciplines of similarity measures, VIKOR and TOPSIS methods, and various other aggregation operators. This will enable us to further improve the QDFS method and expand its applicability to a wider range of decision-making scenarios. In addition, we plan to explore the use of QDFS in real-life problems such as site selection for wind power plants [32] and assessment of the innovative ability of universities [14]. By doing so, we hope to gain a deeper understanding of the principles of FSs and how they can be applied to practical decision-making problems.
The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through the Small Groups Project (Grant No. R.G.P.1/109/44).
The authors declare that the publication of this article does not involve any conflicts of interest.
[1] | A. S. Abd-Alzhra, M. S. H. Al-Tamimi, Lossy image compression using hybrid deep learning autoencoder based on k-mean clustering, Design Engin., 2021, 7848–7861. |
[2] |
M. Akram, G. Ali, J. C. R. Alcantud, A novel group decision-making framework under Pythagorean fuzzy n-soft expert knowledge, Eng. Appl. Artif. Intell., 120 (2023), 105879. https://doi.org/10.1016/j.engappai.2023.105879 doi: 10.1016/j.engappai.2023.105879
![]() |
[3] |
M. Akram, R. Bibi, M. Deveci, An outranking approach with 2-tuple linguistic fermatean fuzzy sets for multi-attribute group decision-making, Eng. Appl. Artif. Intell., 121 (2023), 105992. https://doi.org/10.1016/j.engappai.2023.105992 doi: 10.1016/j.engappai.2023.105992
![]() |
[4] |
M. I. Ali, Another view on q-rung orthopair fuzzy sets, Int. J. Intell. Syst., 33 (2018), 2139–2153. https://doi.org/10.1002/int.22007 doi: 10.1002/int.22007
![]() |
[5] |
A. Ashraf, K. Ullah, A.Hussain, M. Bari, Interval-valued picture fuzzy maclaurin symmetric mean operator with application in multiple attribute decision-making, Rep. Mech. Eng., 3 (2022), 210–226. https://doi.org/10.22481/intermaths.v3i1.10721 doi: 10.22481/intermaths.v3i1.10721
![]() |
[6] | K. T. Atanassov, Geometrical interpretation of the elements of the intuitionistic fuzzy objects, Int. J. Bio. Autom., 20 (2016), S27–S42. |
[7] |
K. T. Atanassov, More on intuitionistic fuzzy sets, Fuzzy Set. Syst., 33 (1989), 37–45. https://doi.org/10.1016/0165-0114(89)90215-7 doi: 10.1016/0165-0114(89)90215-7
![]() |
[8] | K. T. Atanassov, Intuitionistic fuzzy sets, In: Intuitionistic fuzzy sets, Studies in Fuzziness and Soft Computing, Physica, Heidelberg, 1999. https://doi.org/10.1007/978-3-7908-1870-3-1 |
[9] | K. T. Atanassov, On intuitionistic fuzzy sets theory, In: Studies in Fuzziness and Soft Computing (STUDFUZZ, volume 283), Springer, 2012. |
[10] |
B. Cuong, V. Kreinovich, Picture fuzzy sets, J. Comput. Sci. Cyb., 30 (2014), 409–420. http://dx.doi.org/10.15625/1813-9663/30/4/5032 doi: 10.15625/1813-9663/30/4/5032
![]() |
[11] |
A. K. Das, C. Granados, IFP-intuitionistic multi fuzzy N-soft set and its induced IFP-hesitant N-soft set in decision-making, J. Amb. Intell. Hum. Comput., 2022, 1–10. https://doi.org/10.1007/s12652-021-03677-w doi: 10.1007/s12652-021-03677-w
![]() |
[12] | M. Dasgupta, O. Bandyopadhyay, S. Chatterji, Automated helmet detection for multiple motorcycle riders using CNN, In: 2019 IEEE Conference on Information and Communication Technology, IEEE, 2019, 1–4. https://doi.org/10.1109/CICT48419.2019.9066191 |
[13] | P. Deval, A. Chaudhari, R. Wagh, A. Auto, M. Parma, CNN based face mask detection integrated with digital hospital facilities, Int. J. Adv. Res. Sci. Commun. Tech., 4 (2021), 492–497. |
[14] |
L. Dong, X. Gu, X. Wu, H. Liao, An improved multimoora method with combined weights and its application in assessing the innovative ability of universities, Expert Syst., 36 (2019), e12362. https://doi.org/10.1111/exsy.12362 doi: 10.1111/exsy.12362
![]() |
[15] | M. S. Ejaz, M. R. Islam, M. Sifatullah, A. Sarker, Implementation of principal component analysis on masked and non-masked face recognition, In: 2019 1st international conference on advances in science, engineering and robotics technology (ICASERT), IEEE, 2019, 1–5. http://dx.doi.org/10.1109/ICASERT.2019.8934543 |
[16] |
F. Feng, M. Liang, H. Fujita, R. R. Yager, X. Liu, Lexicographic orders of intuitionistic fuzzy values and their relationships, Mathematics, 7 (2019), 166. https://doi.org/10.3390/math7020166 doi: 10.3390/math7020166
![]() |
[17] | R. Girshick, Fast R-CNN, In: Proceedings of the IEEE international conference on computer vision, 2015, 1440–1448. |
[18] | K. He, G. Gkioxari, P. Dollár, R. Girshick, Mask R-CNN, In: Proceedings of the IEEE international conference on computer vision, 2017, 2961–2969. |
[19] |
M. R. Khan, K. Ullah, Q. Khan, Multi-attribute decision-making using archimedean aggregation operator in t-spherical fuzzy environment, Rep. Mech. Eng., 4 (2023), 18–38. https://doi.org/10.31181/rme20031012023k doi: 10.31181/rme20031012023k
![]() |
[20] |
K. T. Atanassov, Intuitionistic fuzzy sets, Fuzzy Set. Syst., 20 (1986), 87–96. https://doi.org/10.1016/S0165-0114(86)80034-3 doi: 10.1016/S0165-0114(86)80034-3
![]() |
[21] |
K. Lin, H. Zhao, J. Lv, C. Li, X. Liu, R. Chen, et al., Face detection and segmentation based on improved mask R-CNN, Discrete Dyn. Nat. Soc., 2020 (2020). https://doi.org/10.1155/2020/9242917 doi: 10.1155/2020/9242917
![]() |
[22] |
P. D. Liu, P. Wang, Some q-rung orthopair fuzzy aggregation operators and their applications to multiple-attribute decision making, Int. J. Intell. Syst., 33 (2018), 259–280. https://doi.org/10.1002/int.21927 doi: 10.1002/int.21927
![]() |
[23] |
T. Mahmood, J. Ahmmad, J. Gwak, N. Jan, Prioritization of thermal energy techniques by employing picture fuzzy soft power average and geometric aggregation operators, Sci. Rep., 13 (2023), 1707. https://doi.org/10.1038/s41598-023-27387-9 doi: 10.1038/s41598-023-27387-9
![]() |
[24] |
T. Mahmood, K. Ullah, Q. Khan, N. Jan, An approach toward decision-making and medical diagnosis problems using the concept of spherical fuzzy sets, Neural Comput. Appl., 31 (2019), 7041–7053. https://doi.org/10.1007/s00521-018-3521-2 doi: 10.1007/s00521-018-3521-2
![]() |
[25] | N. A. Mohamed, M. S. H. Al-Tamimi, Image fusion using a convolutional neural network, Solid State Technol., 63 (2020). |
[26] |
K. Naeem, M. Riaz, F. Karaaslan, Some novel features of pythagorean m-polar fuzzy sets with applications, Complex Intell. Syst., 7 (2021), 459–475. https://doi.org/10.1007/s40747-020-00219-3 doi: 10.1007/s40747-020-00219-3
![]() |
[27] |
M. Narang, M. C. Joshi, K. Bisht, A. Pal, Stock portfolio selection using a new decision-making approach based on the integration of fuzzy cocoso with heronian mean operator, Decis. Making Appl. Manag. Eng., 5 (2022), 90–112. https://doi.org/10.31181/dmame0310022022n doi: 10.31181/dmame0310022022n
![]() |
[28] | M. Rezaei, E. Ravanbakhsh, E. Namjoo, M. Haghighat, Assessing the effect of image quality on SSD and faster R-CNN networks for face detection, In: 2019 27th Iranian Conference on Electrical Engineering (ICEE), IEEE, 2019, 1589–1594. http://dx.doi.org/10.1109/IranianCEE.2019.8786526 |
[29] |
M. Riaz, H. M. A. Farid, F. Karaaslan, Linear Diophantine fuzzy aggregation operators with multi-criteria decision-making, J. Comput. Cogn. Eng., 2022. https://doi.org/10.47852/bonviewJCCE3202420 doi: 10.47852/bonviewJCCE3202420
![]() |
[30] |
M. Riaz, M. R. Hashmi, Linear diophantine fuzzy set and its applications towards multi-attribute decision-making problems, J. Intell. Fuzzy Syst., 37 (2019), 5417–5439. https://doi.org/10.3233/JIFS-190550 doi: 10.3233/JIFS-190550
![]() |
[31] |
J. R. S. Veluswami, S. Prakash, N. Parekh, Face mask detection using SSDNET and lightweight custom CNN, SSRN Electron. J., 2021. http://dx.doi.org/10.2139/ssrn.3882472 doi: 10.2139/ssrn.3882472
![]() |
[32] |
X. L. Wu, C. Zhang, L. S. Jiang, H. C. Liao, An integrated method with promethee and conflict analysis for qualitative and quantitative decision-making: Case study of site selection for wind power plants, Cogn. Comput., 12 (2020), 100–114. https://doi.org/10.1007/s12559-019-09675-7 doi: 10.1007/s12559-019-09675-7
![]() |
[33] |
Z. S. Xu, An overview of methods for determining owa weights, Int. J. Intell. Syst., 20 (2005), 843–865. https://doi.org/10.1002/int.20097 doi: 10.1002/int.20097
![]() |
[34] | R. R. Yager, Pythagorean fuzzy subsets, In: 2013 joint IFSA world congress and NAFIPS annual meeting (IFSA/NAFIPS), IEEE, 2013, 57–61. https://doi.org/10.1109/IFSA-NAFIPS.2013.6608375 |
[35] |
R. R. Yager, Pythagorean membership grades in multicriteria decision making, IEEE T. Fuzzy Syst., 22 (2013), 958–965. https://doi.org/10.1109/TFUZZ.2013.2278989 doi: 10.1109/TFUZZ.2013.2278989
![]() |
[36] |
R. R. Yager, Generalized orthopair fuzzy sets, IEEE T. Fuzzy Syst., 25 (2016), 1222–1230. https://doi.org/10.1109/TFUZZ.2016.2604005 doi: 10.1109/TFUZZ.2016.2604005
![]() |
[37] |
L. A. Zadeh, Fuzzy sets, Inform. Control, 8 (1965), 338–353. https://doi.org/10.1016/S0019-9958(65)90241-X doi: 10.1016/S0019-9958(65)90241-X
![]() |
[38] |
Q. Zhang, J. H. Hu, J. F. Feng, A. Liu, Y. L. Li, New similarity measures of pythagorean fuzzy sets and their applications, IEEE Access, 7 (2019), 138192–138202. https://doi.org/10.1109/ACCESS.2019.2942766 doi: 10.1109/ACCESS.2019.2942766
![]() |
1. | Faisal Yousafzai, Muhammad Danish Zia, Murad-ul-Islam Khan, Mohammed. M. Khalaf, Rashad Ismail, Linear Diophantine fuzzy sets over complex fuzzy information with applications in information theory, 2024, 15, 20904479, 102327, 10.1016/j.asej.2023.102327 | |
2. | Xiaoying Wang, Muhammad Danish Zia, Faisal Yousafzai, Sarfraz Ahmed, Muwen Wang, Complex fuzzy intelligent decision modeling for optimizing economic sustainability in transportation sector, 2024, 10, 2199-4536, 3833, 10.1007/s40747-024-01372-9 | |
3. | Muhammad Danish Zia, Faisal Yousafzai, Saleem Abdullah, Kostaq Hila, Complex linear Diophantine fuzzy sets and their applications in multi-attribute decision making, 2024, 132, 09521976, 107953, 10.1016/j.engappai.2024.107953 |
X1 | X2 | |
Y1 | (⟨0.58,0.69⟩,⟨0.14,0.49,0.21⟩) | (⟨0.47,0.92⟩,⟨0.28,0.28,0.33⟩) |
Y2 | (⟨0.77,0.30⟩,⟨0.15,0.19,0.54⟩) | (⟨0.80,0.16⟩,⟨0.31,0.15,0.14⟩) |
Y3 | (⟨0.46,0.67⟩,⟨0.26,0.14,0.40⟩) | (⟨0.73,0.94⟩,⟨0.33,0.22,0.43⟩) |
Y4 | (⟨0.31,0.73⟩,⟨0.42,0.16,0.41⟩) | (⟨0.75,0.91⟩,⟨0.19,0.44,0.30⟩) |
X3 | X4 | |
Y1 | (⟨0.62,0.87⟩,⟨0.19,0.12,0.59⟩) | (⟨0.66,0.31⟩,⟨0.60,0.11,0.26⟩) |
Y2 | (⟨0.62,0.26⟩,⟨0.19,0.18,0.50⟩) | (⟨0.49,0.60⟩,⟨0.23,0.21,0.37⟩) |
Y3 | (⟨0.48,0.18⟩,⟨0.63,0.18,0.12⟩) | (⟨0.30,0.11⟩,⟨0.66,0.21,0.10⟩) |
Y4 | (⟨0.90,0.56⟩,⟨0.12,0.15,0.19⟩) | (⟨0.65,0.58⟩,⟨0.64,0.13,0.11⟩) |
QWA | QWG | |
Y1 | (⟨0.60,0.63⟩,⟨0.34,0.18,0.40⟩) | (⟨0.60,0.78⟩,⟨0.31,0.26,0.41⟩) |
Y2 | (⟨0.66,0.31⟩,⟨0.21,0.18,0.39⟩) | (⟨0.63,0.37⟩,⟨0.21,0.21,0.43⟩) |
Y3 | (⟨0.49,0.27⟩,⟨0.54,0.18,0.20⟩) | (⟨0.45,0.54⟩,⟨0.49,0.48,0.23⟩) |
Y4 | (⟨0.76,0.64⟩,⟨0.37,0.17,0.21⟩) | (⟨0.65,0.69⟩,⟨0.30,0.26,0.24⟩) |
Score of X1 | Score of X2 | Score of X3 | Score of X4 | |
Y1 | 0.15 | 0.06 | −0.23 | 0.23 |
Y2 | 0.19 | 0.23 | −0.16 | −0.28 |
Y3 | 0.11 (accuracy=0.38) | 0.21 | −0.15 | 0.11 (accuracy=0.54) |
Y4 | 0.22 | −0.03 | −0.16 | 0.07 |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | 0.09 | 4th | 0.14 (accuracy=0.50) | 3rd |
Y2 | 0.13 | 2nd | 0.12 | 4th |
Y3 | 0.15 | 1st | 0.17 | 1st |
Y4 | 0.11 | 3rd | 0.14 (accuracy=0.60) | 2nd |
QWA | QWG | |
Y1 | (⟨0.58,0.72⟩,⟨0.30,0.19,0.41⟩) | (⟨0.56,0.84⟩,⟨0.29,0.24,0.42⟩) |
Y2 | (⟨0.71,0.25⟩,⟨0.24,0.17,0.33⟩) | (⟨0.67,0.31⟩,⟨0.23,0.22,0.39⟩) |
Y3 | (⟨0.57,0.35⟩,⟨0.51,0.18,0.24⟩) | (⟨0.50,0.71⟩,⟨0.45,0.44,0.29⟩) |
Y4 | (⟨0.78,0.69⟩,⟨0.31,0.21,0.24⟩) | (⟨0.67,0.77⟩,⟨0.27,0.22,0.26⟩) |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | −0.09 | 4th | −0.085 | 4th |
Y2 | 0.27 | 2nd | 0.21 | 1st |
Y3 | 0.34 | 1st | 0.20 | 2nd |
Y4 | 0.19 | 3rd | 0.07 | 3rd |
Weights | Rankings | |
QWA | User defined | Y3≤Y2≤Y4≤Y1 |
QWG | User defined | Y3≤Y4≤Y1≤Y2 |
QWA | Normally distributed | Y3≤Y2≤Y4≤Y1 |
QWG | Normally distributed | Y2≤Y3≤Y4≤Y1 |
Collection of sets | Acceptance Information | Rejection Information | Ignorance Information | Parameterization | Unrestricted Domain |
FS | Yes | No | No | No | No |
IFS | Yes | Yes | No | No | No |
PyFS | Yes | Yes | No | No | No |
q-ROFS | Yes | Yes | No | No | No |
PFS | Yes | Yes | Yes | No | No |
LDFS | Yes | Yes | No | Yes | Yes |
QDFS | Yes | Yes | Yes | Yes | Yes |
X1 | X2 | |
Y1 | (⟨0.58,0.69⟩,⟨0.14,0.49,0.21⟩) | (⟨0.47,0.92⟩,⟨0.28,0.28,0.33⟩) |
Y2 | (⟨0.77,0.30⟩,⟨0.15,0.19,0.54⟩) | (⟨0.80,0.16⟩,⟨0.31,0.15,0.14⟩) |
Y3 | (⟨0.46,0.67⟩,⟨0.26,0.14,0.40⟩) | (⟨0.73,0.94⟩,⟨0.33,0.22,0.43⟩) |
Y4 | (⟨0.31,0.73⟩,⟨0.42,0.16,0.41⟩) | (⟨0.75,0.91⟩,⟨0.19,0.44,0.30⟩) |
X3 | X4 | |
Y1 | (⟨0.62,0.87⟩,⟨0.19,0.12,0.59⟩) | (⟨0.66,0.31⟩,⟨0.60,0.11,0.26⟩) |
Y2 | (⟨0.62,0.26⟩,⟨0.19,0.18,0.50⟩) | (⟨0.49,0.60⟩,⟨0.23,0.21,0.37⟩) |
Y3 | (⟨0.48,0.18⟩,⟨0.63,0.18,0.12⟩) | (⟨0.30,0.11⟩,⟨0.66,0.21,0.10⟩) |
Y4 | (⟨0.90,0.56⟩,⟨0.12,0.15,0.19⟩) | (⟨0.65,0.58⟩,⟨0.64,0.13,0.11⟩) |
QWA | QWG | |
Y1 | (⟨0.60,0.63⟩,⟨0.34,0.18,0.40⟩) | (⟨0.60,0.78⟩,⟨0.31,0.26,0.41⟩) |
Y2 | (⟨0.66,0.31⟩,⟨0.21,0.18,0.39⟩) | (⟨0.63,0.37⟩,⟨0.21,0.21,0.43⟩) |
Y3 | (⟨0.49,0.27⟩,⟨0.54,0.18,0.20⟩) | (⟨0.45,0.54⟩,⟨0.49,0.48,0.23⟩) |
Y4 | (⟨0.76,0.64⟩,⟨0.37,0.17,0.21⟩) | (⟨0.65,0.69⟩,⟨0.30,0.26,0.24⟩) |
Score of X1 | Score of X2 | Score of X3 | Score of X4 | |
Y1 | 0.15 | 0.06 | −0.23 | 0.23 |
Y2 | 0.19 | 0.23 | −0.16 | −0.28 |
Y3 | 0.11 (accuracy=0.38) | 0.21 | −0.15 | 0.11 (accuracy=0.54) |
Y4 | 0.22 | −0.03 | −0.16 | 0.07 |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | 0.09 | 4th | 0.14 (accuracy=0.50) | 3rd |
Y2 | 0.13 | 2nd | 0.12 | 4th |
Y3 | 0.15 | 1st | 0.17 | 1st |
Y4 | 0.11 | 3rd | 0.14 (accuracy=0.60) | 2nd |
QWA | QWG | |
Y1 | (⟨0.58,0.72⟩,⟨0.30,0.19,0.41⟩) | (⟨0.56,0.84⟩,⟨0.29,0.24,0.42⟩) |
Y2 | (⟨0.71,0.25⟩,⟨0.24,0.17,0.33⟩) | (⟨0.67,0.31⟩,⟨0.23,0.22,0.39⟩) |
Y3 | (⟨0.57,0.35⟩,⟨0.51,0.18,0.24⟩) | (⟨0.50,0.71⟩,⟨0.45,0.44,0.29⟩) |
Y4 | (⟨0.78,0.69⟩,⟨0.31,0.21,0.24⟩) | (⟨0.67,0.77⟩,⟨0.27,0.22,0.26⟩) |
QWA | QWG | |||
Score function | Ranking | Score function | Ranking | |
Y1 | −0.09 | 4th | −0.085 | 4th |
Y2 | 0.27 | 2nd | 0.21 | 1st |
Y3 | 0.34 | 1st | 0.20 | 2nd |
Y4 | 0.19 | 3rd | 0.07 | 3rd |
Weights | Rankings | |
QWA | User defined | Y3≤Y2≤Y4≤Y1 |
QWG | User defined | Y3≤Y4≤Y1≤Y2 |
QWA | Normally distributed | Y3≤Y2≤Y4≤Y1 |
QWG | Normally distributed | Y2≤Y3≤Y4≤Y1 |
Collection of sets | Acceptance Information | Rejection Information | Ignorance Information | Parameterization | Unrestricted Domain |
FS | Yes | No | No | No | No |
IFS | Yes | Yes | No | No | No |
PyFS | Yes | Yes | No | No | No |
q-ROFS | Yes | Yes | No | No | No |
PFS | Yes | Yes | Yes | No | No |
LDFS | Yes | Yes | No | Yes | Yes |
QDFS | Yes | Yes | Yes | Yes | Yes |