Processing math: 97%
Research article Special Issues

The dual fuzzy matrix equations: Extended solution, algebraic solution and solution

  • Received: 14 September 2022 Revised: 10 December 2022 Accepted: 25 December 2022 Published: 12 January 2023
  • MSC : 03E72, 08A72, 26E50

  • In this paper, we propose a direct method to solve the dual fuzzy matrix equation of the form A˜X+˜B=C˜X+˜D with A, C matrices of crisp coefficients and ˜B, ˜D fuzzy number matrices. Extended solution and algebraic solution of the dual fuzzy matrix equations are defined and the relationship between them is investigated. This article focuses on the algebraic solution and a necessary and sufficient condition for the unique algebraic solution existence is given. By algebraic methods we not need to transform a dual fuzzy matrix equation into two crisp matrix equations to solve. In addition, the general dual fuzzy matrix equations and dual fuzzy linear systems are investigated based on the generalized inverses of the matrices. Especially, the solution formula and calculation method of the dual fuzzy matrix equation with triangular fuzzy number matrices are given and discussed. The effectiveness of the proposed method is illustrated with examples.

    Citation: Zengtai Gong, Jun Wu, Kun Liu. The dual fuzzy matrix equations: Extended solution, algebraic solution and solution[J]. AIMS Mathematics, 2023, 8(3): 7310-7328. doi: 10.3934/math.2023368

    Related Papers:

    [1] Xinfeng Liang, Mengya Zhang . Triangular algebras with nonlinear higher Lie n-derivation by local actions. AIMS Mathematics, 2024, 9(2): 2549-2583. doi: 10.3934/math.2024126
    [2] He Yuan, Zhuo Liu . Lie n-centralizers of generalized matrix algebras. AIMS Mathematics, 2023, 8(6): 14609-14622. doi: 10.3934/math.2023747
    [3] Shan Li, Kaijia Luo, Jiankui Li . Generalized Lie n-derivations on generalized matrix algebras. AIMS Mathematics, 2024, 9(10): 29386-29403. doi: 10.3934/math.20241424
    [4] Guangyu An, Xueli Zhang, Jun He, Wenhua Qian . Characterizations of local Lie derivations on von Neumann algebras. AIMS Mathematics, 2022, 7(5): 7519-7527. doi: 10.3934/math.2022422
    [5] Junyuan Huang, Xueqing Chen, Zhiqi Chen, Ming Ding . On a conjecture on transposed Poisson n-Lie algebras. AIMS Mathematics, 2024, 9(3): 6709-6733. doi: 10.3934/math.2024327
    [6] Baiying He, Siyu Gao . The nonisospectral integrable hierarchies of three generalized Lie algebras. AIMS Mathematics, 2024, 9(10): 27361-27387. doi: 10.3934/math.20241329
    [7] Mohd Arif Raza, Huda Eid Almehmadi . Lie (Jordan) σcentralizer at the zero products on generalized matrix algebra. AIMS Mathematics, 2024, 9(10): 26631-26648. doi: 10.3934/math.20241295
    [8] He Yuan, Qian Zhang, Zhendi Gu . Characterizations of generalized Lie n-higher derivations on certain triangular algebras. AIMS Mathematics, 2024, 9(11): 29916-29941. doi: 10.3934/math.20241446
    [9] Huizhang Yang, Wei Liu, Yunmei Zhao . Lie symmetry reductions and exact solutions to a generalized two-component Hunter-Saxton system. AIMS Mathematics, 2021, 6(2): 1087-1100. doi: 10.3934/math.2021065
    [10] Senrong Xu, Wei Wang, Jia Zhao . Twisted Rota-Baxter operators on Hom-Lie algebras. AIMS Mathematics, 2024, 9(2): 2619-2640. doi: 10.3934/math.2024129
  • In this paper, we propose a direct method to solve the dual fuzzy matrix equation of the form A˜X+˜B=C˜X+˜D with A, C matrices of crisp coefficients and ˜B, ˜D fuzzy number matrices. Extended solution and algebraic solution of the dual fuzzy matrix equations are defined and the relationship between them is investigated. This article focuses on the algebraic solution and a necessary and sufficient condition for the unique algebraic solution existence is given. By algebraic methods we not need to transform a dual fuzzy matrix equation into two crisp matrix equations to solve. In addition, the general dual fuzzy matrix equations and dual fuzzy linear systems are investigated based on the generalized inverses of the matrices. Especially, the solution formula and calculation method of the dual fuzzy matrix equation with triangular fuzzy number matrices are given and discussed. The effectiveness of the proposed method is illustrated with examples.



    The variational inequality problem (VIP) was introduced by Stampacchia [1] and provided a very useful tool for researching a large variety of interesting problems arising in physics, economics, finance, elasticity, optimization, network analysis, medical images, water resources, and structural analysis, see for example ([2,3,4,5,6,7,8,9,10,11,12,13,14,15]) and references therein.

    Let H be a real Hilbert space with inner product , and norm , respectively. Let C be a nonempty closed convex subset of H. Let B:CH be an operator.

    In this article, our study is related to a classical variational inequality problem (VIP) which aims to find an element xC such that

    Bx,xx0,  xC. (1.1)

    It is well known that xVI(B,C) if and only if x=PC(xζBx), where ζ>0, in other words, the VIP is equivalent to the fixed point problem (see [16]). Supposing that B is η-strongly monotone and L-Lipschitz continuous with 0<ζ<2ηL2, the following sequence {xn} of Picard iterates:

    xn+1=PC(xnζBxn), (1.2)

    converges strongly to a point xVI(B,C) due to the fact that PC(IζB) is a contraction on C. However, in general, the algorithm (1.2) fails when B is monotone and L-Lipschitz continuous (see [17]). In [7], Korpelevich put forward an extragradient method which provided an important idea for solving monotone variational inequality:

    yn=PC(xnλfxn),xn+1=PC(xnλfyn), (1.3)

    where f is monotone, L-Lipschitz continuous in the finite dimensional Euclidean space Rn and λ(0,1L).

    The another motivation of this article is the split common fixed point problem which aims to find a point uH1 such that

    uFix(T)   and   AuFix(S). (1.4)

    The split common fixed point problem can be regarded as a generalization of the split feasibility problem. Recall that the split feasibility problem is to find a point satisfying

    uC   and   AuQ, (1.5)

    where C and Q are two nonempty closed convex subsets of real Hilbert spaces H1 and H2, respectively and A:H1H2 is a bounded linear operator. Inverse problems in various disciplines can be expressed as the split feasibility problem and the split common fixed point problem. Problem (1.4) was firstly introduced by Censor and Segal [18]. Note that solving (1.4) can be translated to solve the fixed point equation:

    u=S(uτA(IT)Au), τ>0.

    Whereafter, Censor and Segal proposed an algorithm for directed operators. Since then, there has been growing interest in the split common fixed point problem (see [19,20,21,22]).

    Censor et al. [23] first proposed split variational inequality problems by combining the variational inequality problem and the split feasibility problem. Very recently, in 2017, Tian and Jiang [24] considered the following split variational inequality problem: finding an element u such that

    uVI(A,C) and BuFix(T), (1.6)

    where T:H2H2 is nonexpansive, B:H1H2 is a bounded linear operator with its adjoint B, and A:CH1 is a monotone and L-Lipschitz continuous mapping. Then they presented the following iteration method by combining the extragradient method with CQ algorithm for solving the (1.6):

    Algorithm 1.1. Choose an arbitrary initial value x1C. Assume xn has been constructed. Compute

    yn=PC(xnτnA(IT)Axn),zn=PC(ynςnFyn),xn+1=PC(ynςnFzn). (1.7)

    They proved that the iterative sequence {xn} defined by Eq (1.7) converges weakly to an element zΓ, where Γ is the set of solutions of the problem (1.6). However, Algorithm 1.1 fails, in general, to converge strongly in the setting of infinite-dimensional Hilbert spaces. We also notice that Algorithm 1.1 is involved with three metric projections in each iteration, which might seriously affect the efficiency of the method.

    Motivated and inspired by the above works, in the present paper, we consider variational inequality problems and split common fixed point problems for finding an element u such that

    ˆxVI(A,C) and Bˆxn=1Fix(Tn), (1.8)

    where {Tn}n=1:H2H2 is an infinite family of nonexpansive mappings, B:H1H2 is a bounded linear operator with its adjoint B, and A:H1H1 is a monotone and L-Lipschitz continuous mapping. In contrast to Tian and Jiang [24], we consider the common fixed points of an infinite family of nonexpansive mappings instead of only the fixed points of a nonexpansive mapping. The efficiency of the algorithm is also improved by removing the projection operator in the first iteration which might affect the efficiency of the method to a certain extent. Finally, we present a very simple modification to extragradient method, which makes our algorithm have the strong convergence. It is well known that the strong convergence theorem is always more convenient to use.

    This paper is organized as follows: In Section 2, we give some definitions and key lemmas which are used in this paper. Section 3 consists of our algorithms and provides the strong convergence theorems. In Section 4, numerical examples are provided for illustration. Finally, this paper is concluded in Section 5.

    Let H be a real Hilbert space with inner product , and norm , respectively. Let C be a nonempty closed convex subset of H. Let T:CC be an operator. We use Fix(T) to denote the set of fixed points of T, that is, Fix(T)={x|x=Tx,xC}. First, we give some definitions and lemmas related to the involved operators.

    Definition 2.1. An operator T:CC is said to be nonexpansive if TuTvuv for all u,vC.

    Definition 2.2. An operator A:CH is said to be monotone if AxAy,xy0 for all x,yC.

    A monotone operator R:H2H is called maximal monotone if the graph of R is a maximal monotone set.

    Definition 2.3. An operator T:CH is said to be L-Lipschitzian if there exists L>0 such that TxTyLxy for all x,yC.

    Usually, the convergence of fixed point algorithms requires some additional smoothness properties of the mapping T such as demi-closedness.

    Definition 2.4. An operator T is said to be demiclosed if, for any sequence {un} which weakly converges to u, and if Tunw, then Tu=w.

    Recall that the (nearest point or metric) projection from H onto C, denoted by PC, assigns to each uH, the unique point PCuC with the property:

    uPCu=inf{uv:vC}.

    The metric projection PC of H onto C is characterized by

    uPCu,vPCu0or uv2uPCu2+vPCu2 (2.1)

    for all uH,vC. It is well known that the metric projection PC:HC is firmly nonexpansive, that is,

    uv,PCuPCvPCuPCv2or PCuPCv2uv2(IPC)u(IPC)v2 (2.2)

    for all u,vH. More information on the metric projection can be found, for example, in Section 3 of the book by Goebel et al. (see [25]).

    For all u,vH, the following conclusions hold:

    tu+(1t)v2=tu2+(1t)v2t(1t)uv2,  t[0,1], (2.3)
    u+v2=u2+2u,v+v2 (2.4)

    and

    u+v2u2+2v,u+v. (2.5)

    Let {Tn}n=1:HH be an infinite family of nonexpansive mappings and λ1,λ2,... be real numbers such that 0λi1 for each iN. For any nN, define a mapping Wn of C into H as follows:

    Un,n+1=I,Un,n=λnTnUn,n+1+(1λn)I,Un,n1=λn1Tn1Un,n+(1λn1)I,                      Un,k=λkTkUn,k+1+(1λk)I,Un,k1=λk1Tk1Un,k+(1λk1)I,Un,2=λ2T2Un,3+(1λ2)I,Wn=Un,1=λ1T1Un,2+(1λ1)I. (2.6)

    Such a mapping Wn is called the W-mapping generated by T1,T2,...,Tn and λ1,λ2,...,λn. We have the following crucial Lemma concerning Wn:

    Lemma 2.1. [26] Let H be a real Hilbert space. Let {Tn}n=1:HH be an infinite family of nonexpansive mappings such that n=1Fix(Tn). Let λ1,λ2,... be real numbers such that 0λib<1 for each i1. Then we have the following:

    (1) For any xH and k1, the limit limnUn,kx exists;

    (2) Fix(W)=n=1Fix(Tn), where Wx=limnWnx=limnUn,1x, xC;

    (3) For any bounded sequence {xn}H, limnWxn=limnWnxn.

    Lemma 2.2. [27] Assume that {αn} is a sequence of nonnegative real numbers such that

    αn+1(1γn)αn+δn,  nN,

    where {γn} is a sequence in (0,1) and {δn} is a sequence such that

    (1) n=1γn=;

    (2) lim supnδnγn0 or n=1|δn|<. Then limnαn=0.

    Lemma 2.3. [28] Let {ϖn} be a sequence of real numbers. Assume there exists at least a subsequence {ϖnk} of {ϖn} such that ϖnkϖnk+1 for all k0. For every nN0, define an integer sequence {τ(n)} as:

    τ(n)=max{in:ϖni<ϖni+1}.

    Then, τ(n) as n and for all nN0, we have max{ϖτ(n),ϖn}ϖτ(n)+1.

    In this section, we introduce our algorithm and prove its strong convergence. Some assumptions on the underlying spaces and involved operators are listed below.

    (R1) H1 and H2 are two real Hilbert spaces and CH1 is a nonempty closed convex subset.

    (R2) B:H1H2 is a bounded linear operator with its adjoint B.

    (R3) A:H1H1 is a monotone and L-Lipschitz continuous mapping.

    (R4) Ω={ˆx|ˆxVI(A,C) and Bˆxn=1Fix(Tn)}, where Ω is the set of solutions of the problem (1.8).

    Next, we present the following iterative algorithm to find a point ˆxΩ.

    Algorithm 3.1. Choose an arbitrary initial value x1H. Assume xn has been constructed. Compute

    yn=xnτnB(IWn)Bxn,zn=PC(ynςnAyn),xn+1=PC((1αn)(ynςnAzn)), (3.1)

    where {αn} is a sequence in (0,1), ςn is a sequence in (0,1L), and τn is a sequence in (0,1B2).

    Theorem 3.1. If Ω and the following conditions are satisfied:

    (C1) limnαn=0 and n=0αn=;

    (C2) 0<lim infnςnlim supnςn<1L;

    (C3) 0<lim infnτnlim supnτn<1B2.

    Then, the iterative sequence {xn} defined by Eq (3.1) strongly converges to the minimum-norm solution ˆx(=PΩθ).

    Proof. Set z=PΩθ. We can obtain that

    ynz2=xnzτnB(IWn)Bxn2=xnz22τnxnz,B(IWn)Bxn+τnB(IWn)Bxn2=xnz22τnBxnBz,(IWn)Bxn+τnB(IWn)Bxn2xnz2τn(IWn)Bxn2+τ2nB2(IWn)Bxn2xnz2τn(1τnB2)(IWn)Bxn2xnz2. (3.2)

    It follows from (2.1) that

    xn+1z2=PC((1αn)(ynςnAzn))z2(1αn)(ynςnAzn)z2(1αn)(ynςnAzn)xn+12(1αn)(ynςnAznz)+αn(z)2(1αn)(ynςnAznxn+1)+αn(xn+1)2(1αn)ynςnAznz2+αnz2(1αn)αnynςnAzn2((1αn)ynςnAznxn+12+αnxn+12(1αn)αnynςnAzn2)=(1αn)(ynςnAznz2ynςnAznxn+12)+αn(z2xn+12). (3.3)

    We also observe that

    ynςnAznz2ynςnAznxn+12=ynz2ynxn+12+2ςnAzn,zxn+1=ynz2ynxn+12+2ςnAzn,zzn+2ςnAzn,znxn+1=ynz2ynxn+12+2ςnAznAz,zzn+2ςnAz,zzn+2ςnAzn,znxn+1ynz2ynxn+12+2ςnAzn,znxn+1=ynz2ynzn2znxn+12+2ynςnAznzn,xn+1zn. (3.4)

    On the other hand, we have that

    ynςnAznzn,xn+1zn=ynςnAynzn,xn+1zn+ςnAynAzn,xn+1znςnAynAzn,xn+1znςnAynAzn×xn+1znςnLynzn×xn+1zn. (3.5)

    Hence, we can derive that

    xn+1z2=(1αn)(ynςnAznz2ynςnAznxn+12)+αn(z2xn+12),(by(3.4))(1αn)(ynz2ynzn2znxn+12+2ynςnAznzn,xn+1zn)+αn(z2xn+12),(by(3.5))(1αn)(ynz2ynzn2znxn+12+2ςnLynzn×xn+1zn)+αn(z2xn+12)(1αn)(ynz2ynzn2znxn+12+ς2nL2ynzn2+xn+1zn2)+αn(z2xn+12)(1αn)(ynz2+(ς2nL21)ynzn2)+αn(z2xn+12),(by(3.2))(1αn)(xnz2+(ς2nL21)ynzn2)+αn(z2xn+12). (3.6)

    Owing to the assumption (C2), it follows from (3.6) that

    xn+1z2(1αn)ynz2+αn(z2xn+12),(by(3.2))(1αn)xnz2+αn(z2xn+12)(1αn)xnz2+αnz2max{xnz2,z2} (3.7)

    and so

    xnz2max{x1z2,z2}, (3.8)

    which implies that the sequence {xn} is bounded. In view of (3.2) and (3.7), we obtain that

    τn(1τnB2)(IWn)Bxn2xnz2ynz2xnz2xn+1z2+αn(z2xn+12ynz2). (3.9)

    CASE I. Suppose that there exists m>0 such that the sequence {xnz} is decreasing when nm. Then, limnxnz exists. Consequently, according to the assumptions (C1) and (C3), we deduce that

    limn(IWn)Bxn=0. (3.10)

    In virtue of the boundedness of the sequence {Bxn} and Lemma 2.1, we get that

    limnWBxnWnBxn=0. (3.11)

    This together with (3.24) implies that

    limn(IW)Bxn=0. (3.12)

    It follows from (3.6) that

    (1αn)(1ς2nL2)ynzn2(1αn)xnz2xn+1z2+αn(z2xn+12)xnz2xn+1z2+αn(z2xn+12xnz2). (3.13)

    Thanks to the boundedness of the sequence {xn}, we derive that

    limnynzn=0. (3.14)

    In view of (3.30), we can also get that

    limnynxn=limnτnB(IWn)Bxn=0(by(3.10)). (3.15)

    Combining (3.14) and (3.15), we obtain that

    limnznxn=0. (3.16)

    On the other hand, we get that

    xn+1zn=PC((1αn)(ynςnAzn))PC(ynςnAyn)(1αn)(ynςnAzn)(ynςnAyn)(ynςnAzn)(ynςnAyn)+αnynςnAzn)ςnAznςnAyn+αnynςnAznςnAznAyn+αnynςnAznςnLznyn+αnynςnAzn. (3.17)

    Hence, by (3.14), it turns out that

    limnxn+1zn=0 (3.18)

    and consequently, according to (3.16), we have that

    limnxn+1xn=0. (3.19)

    Next, we can take a subsequence {ni} such that

    lim supn(z2xn+12)=limi(z2xni+12). (3.20)

    By the boundedness of the real sequence {xni+1}, we may assume that xni+1x. Since W is nonexpansive, we can derive that Bx=WBx(see Corollary 4.28 in [29]), that is, BxFix(W)=n=1Fix(Tn).

    Now, we show that xVI(A,C). Let

    R(v)={Av+NC(v), vC,  vC, (3.21)

    where NC(v) is the normal cone to C at v. According to Reference [30], we can easily derive that R is maximal monotone. Let (v,w)G(R). Since wAvNC(v) and xnC, we have that

    vxn,wAv0.

    Noting that, due to vC, we get

    vxn+1,xn+1(1αn)(ynςnAzn))0.

    It follows that

    vxn+1,xn+1ynςn+Azn+αnςn(ynςnAzn)0.

    Thus, we can deduce that

    vxni+1,wvxni+1,Avvxni+1,xni+1yniςni+Azni+αniςni(yniςniAzni)+vxni+1,Avvxni+1,AvAznivxni+1,xni+1yniςnivxni+1,αniςni(yniςniAzni)vxni+1,AvAxni+1+vxni+1,Axni+1Aznivxni+1,xni+1yniςnivxni+1,αniςni(yniςniAzni)vxni+1,xni+1yniςnivxni+1,αniςni(yniςniAzni)+vxni+1,Axni+1Azni. (3.22)

    As i, we obtain that

    vx,w0.

    By the maximal monotonicity of R, we derive that xR10. Hence, xVI(A,C). Therefore, xΩ. Since the norm of the Hilbert space H1 is weakly lower semicontinuous(see Lemma 2.42 in [29]), we have the following inequality:

    xlim infixni+1

    and therefore

    xlim supi(xni+1).

    From (3.7), we observe that

    xn+1z2(1αn)xnz2+αn(z2xn+12). (3.23)

    Thanks to z=PΩθ and xΩ, we can deduce that

    lim supn(z2xni+12)=z2+lim supn(xni+12)z2x20.

    Applying Lemma 2.2 to (3.23), we derive that limnxnz=0, which implies that the sequence {xn} converges strongly to z.

    CASE II. For any n0, there exists an integer mn0 such that xmzxm+1z. At this case, we set ϖn=xnz. For nn0, we define a sequence {τn} by

    τ(n)=max{lN|n0ln,ϖlϖl+1}.

    It is easy to show that τ(n) is a non-decreasing sequence such that

    limnτ(n)=+

    and

    ϖτ(n)ϖτ(n)+1.

    This together with (3.9) implies that

    limn(IWτ(n))Bxτ(n)2=0. (3.24)

    Employing techniques similar to CASE I, we have

    lim supn(z2xτ(n)+12)0 (3.25)

    and

    ϖ2τ(n)+1(1ατ(n))ϖ2τ(n)+ατ(n)(z2xτ(n)+12). (3.26)

    Since ϖτ(n)ϖτ(n)+1, we have

    ϖ2τ(n)z2xτ(n)+12. (3.27)

    By (3.25), we obtain that

    lim supnϖτ(n)0

    and so

    limnϖτ(n)=0. (3.28)

    By Eq (3.26), we also obtain

    lim supnϖτ(n)+1lim supnϖτ(n).

    In the light of the last inequality and Eq (3.28), we derive that

    limnϖτ(n)+1=0.

    Applying Lemma 2.3, we obtain

    ϖnϖτ(n)+1.

    Therefore, we get that ϖn0, that is, xnz. This completes the proof.

    Algorithm 3.2. Choose an arbitrary initial value x1C. Assume xn has been constructed. Compute

    yn=xnτnB(IT)Bxn,zn=PC(ynςnAyn),xn+1=PC((1αn)(ynςnAzn)), (3.29)

    where {αn} is a sequence in (0,1), ςn is a sequence in (0,1L), and τn is a sequence in (0,1B2).

    Theorem 3.2. If ˆΩ and the following conditions are satisfied:

    (C1) limnαn=0 and n=0αn=;

    (C2) 0<lim infnςnlim supnςn<1L;

    (C3) 0<lim infnτnlim supnτn<1B2.

    Then, the iterative sequence {xn} defined by Eq (3.29) strongly converges to the minimum-norm solution ˆx(=PˆΩθ), where

    ˆΩ={ˆx|ˆxVI(A,C) and BˆxFix(T)}.

    Algorithm 3.3. Choose an arbitrary initial value x1C. Assume xn has been constructed. Compute

    zn=PC(xnςnAxn),xn+1=PC((1αn)(xnςnAzn)), (3.30)

    where {αn} is a sequence in (0,1) and ςn is a sequence in (0,1L).

    Theorem 3.3. If ˆΩ and the following conditions are satisfied:

    (C1) limnαn=0 and n=0αn=;

    (C2) 0<lim infnςnlim supnςn<1L;

    Then, the iterative sequence {xn} defined by Eq (3.30) strongly converges to the minimum-norm solution ˆx(=PΩθ), where ˆΩ={ˆx|ˆxVI(A,C)}.

    In this section, we present some numerical examples to illustrate our main results. The MATLAB codes run in MATLAB version 9.5 (R2018b) on a PC Intel(R) Core(TM)i5-6200 CPU @ 2.30 GHz 2.40 GHz, RAM 8.00 GB. In all examples y-axes shows the value of xn+1xn while the x-axis indicates to the number of iterations.

    Example 4.1. Let H1=H2=Rn. The feasible set is defined as:

    C:={xRn:x1}.

    Let G:RnRn is a linear operator defined by:

    Ax:=Gx

    for all xRn, where G=(gij)1i,jn is a matrix in Rn×n whose terms are given by:

    gij={1, if j=n+1i and j>i,1, if j=n+1i and j<i,0, otherwise. (4.1)

    It is obvious that A is G-Lipschitz continuous. By a direct calculation, we also have that Ax,x=Gx,x=0 and so, A is monotone. Let B be a matrix in Rn×n which is randomly generated.

    Taking cognizance of the difference of the problems handled by Algorithm 3.1 and Algorithm in Tian and Jiang [24], in order to comparing these two algorithms, we make a very small modification to the one in [24] such that it can also solve the problem (1.8). The modified algorithm can be written as follows:

    Algorithm 4.1.

    yn=xnτnB(IWn)Bxn,zn=PC(ynςnAyn),xn+1=PC((1αn)(ynςnAzn)), (4.2)

    According to the proof of Theorem 3.1, we can easily verify that this modified algorithm works for solving (1.8). The values of control parameters in these two Algorithms are ςn=12G, τn=12B2, α1=12, αn=1n(for all n2), λn=1n+1 and x1=(1,,1)T, and the infinite family of nonexpansive mappings {Tk}k=1:RnRn is defined by:

    Tkx:=Mkx,

    for all xRn, where {Mk} is a sequence of diagonal matrixes in Rn×n:

    Mk=[11k+2     11k+2          11k+2     11k+3]. (4.3)

    The numerical results of the Example 4.1 are reported in Table 1 and Figures 14 by using the stopping criterion xn+1xn1010.

    Table 1.  Example 4.1: Comparison of Algorithm 3.1 with Algorithm 4.1.
    No. of Iter. Time
    n Alg. 3.1 Alg. 4.1 Alg. 3.1 Alg. 4.1
    2 120 154 0.288s 1.133s
    10 156 201 0.691s 2.516s
    50 157 202 4.641s 13.853s
    100 157 203 12.333s 53.145s

     | Show Table
    DownLoad: CSV
    Figure 1.  Example 4.1: Comparison of Algorithm 3.1 with Algorithm 4.1 when n=2.
    Figure 2.  Example 4.1: Comparison of Algorithm 3.1 with Algorithm 4.1 when n=10.
    Figure 3.  Example 4.1: Comparison of Algorithm 3.1 with Algorithm 4.1 when n=50.
    Figure 4.  Example 4.1: Comparison of Algorithm 3.1 with Algorithm 4.1 when n=100.

    Example 4.2. Let H1=H2=L2([0,1]) with the inner product:

    x,y=10x(t)y(t)dt

    and the induced norm:

    x:=(10x2(t)dt)12.

    The feasible set is defined as:

    C:={xRn:x1}.

    The mapping A:L2([0,1])L2([0,1]) is defined by:

    Ax(t):=(1+t)max{0,x(t)}=(1+t)x(t)+|x(t)|2, xL2([0,1]).

    It is easy to see that

    AxAy,xy=10(Ax(t)Ay(t))(x(t)y(t))dt=10(1+t)x(t)y(t)+|x(t)||y(t)|2(x(t)y(t))dt=1012(1+t)((x(t)y(t))2+(|x(t)||y(t)|)(x(t)y(t)))dt0 (4.4)

    and

    AxAy2=10(Ax(t)Ay(t))2dt=10(1+t)2(x(t)y(t)+|x(t)||y(t)|)24dt=10(1+t)2(x(t)y(t))2dt4xy2. (4.5)

    Therefore, the operator A is monotone and 2-Lipschitz continuous. Let Wn=I(Identity mapping). The values of control parameters for Algorithm 4.1 and Algorithm 3.1 are ςn=14, α1=12, αn=1n(for all n2), λn=1n+1 and x1=8t2. It can be seen easily that {xn} strongly converges to the zero vector θ(L2([0,1])). The numerical results of the Example 4.2 are reported in Table 2 and Figures 5 by using the stopping criterion xn+1xnε=0.01.

    Table 2.  Example 4.2: Comparison of Algorithm 3.1 with Algorithm 4.1 when ε=0.01.
    No. of Iter. Time
    ε Alg. 3.1 Alg. 4.1 Alg. 3.1 Alg. 4.1
    0.01 8 13 0.678s 79.280s

     | Show Table
    DownLoad: CSV
    Figure 5.  Example 4.2: Comparison of Algorithm 3.1 with Algorithm 4.1 when \varepsilon = 0.01 .

    Remark 4.1. The numerical results of Example 4.1 and Example 4.2 show that the performance of Algorithm 3.1 is better than Algorithm 4.1 both in CPU time and the number of iterations. Algorithm 3.1 is more effective in both finite and infinite dimensional spaces and especially in conditions involving complex projection calculations, see Tables 1, 2 and Figures 15. In Example 4.1, we observe that the number of iterations tends to be stable, while the CPU time increases, as n increasing.

    In the present paper, we consider variational inequality problems and split common fixed point problems. We construct an iterative algorithm for solving Eq (1.8) which can be regard as a modification and generalization of Algorithm 1.1 with fewer metric projection operators. Under some mild restrictions, we demonstrate the strong convergence analysis of the presented algorithm. We also give some numerical examples to illustrate our main results. Noticeably, in our article, \mathcal{A} is assumed to a monotone and L -Lipschitz continuous mapping. A natural question arises: how to weaken this assumption?

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported by National Natural Science Foundation of China under Grant (No. 62103289), National Natural Science Foundation of China (No. 11761074), Project of Jilin Science and Technology Development for Leading Talent of Science and Technology Innovation in Middle and Young and Team Project (No. 20200301053RQ), and Natural Science Foundation of Jilin Province(No. 2020122336JC).

    The authors declare no conflict of interest.



    [1] S. Abbasbandy, E. Babolian, M. Alavi, Numerical method for solving linear Fredholm fuzzy integral equations of the second kind, Chaos, Soliton. Fract., 31 (2007), 138–146. https://doi.org/10.1016/j.chaos.2005.09.036 doi: 10.1016/j.chaos.2005.09.036
    [2] T. Allahviranloo, M. Ghanbari, On the algebraic solution of fuzzy linear systems based on interval theory, Appl. Math. Model., 36 (2012), 5360–5379. https://doi.org/10.1016/j.apm.2012.01.002 doi: 10.1016/j.apm.2012.01.002
    [3] T. Allahviranloo, M. Ghanbar, A. A. Hosseinzadeh, E. Haghi, R. Nuraei, A note on "Fuzzy linear systems", Fuzzy Set. Syst., 177 (2011), 87–92. https://doi.org/10.1016/j.fss.2011.02.010 doi: 10.1016/j.fss.2011.02.010
    [4] T. Allahviranloo, E. Haghi, M. Ghanbari, The nearest symmetric fuzzy solution for a symmetric fuzzy linear system, An. Stiint. Univ. Ovidius Constanta, Ser. Mat., 20 (2013), 151–172. https://doi.org/10.2478/v10309-012-0011-x doi: 10.2478/v10309-012-0011-x
    [5] T. Allahviranloo, R. Nuraei, M. Ghanbari, E. Haghi, A. A. Hosseinzadeh, A new metric for L-R fuzzy numbers and its application in fuzzy linear systems, Soft Comput., 16 (2012), 1743–1754. https://doi.org/10.1007/s00500-012-0858-9 doi: 10.1007/s00500-012-0858-9
    [6] S. Abbasbandy, M. Otadi, M. Mosleh, Minimal solution of general dual fuzzy linear systems, Chaos, Soliton. Fract., 37 (2008), 1113–1124. https://doi.org/10.1016/j.chaos.2006.10.045 doi: 10.1016/j.chaos.2006.10.045
    [7] T. Allahviranloo, Uncertain information and linear systems, Cham: Springer, 2020. https://doi.org/10.1007/978-3-030-31324-1
    [8] G. Bojadziev, M. Bojadziev, Fuzzy logic control for business, finance, and management, Singapore: World Scientific, 2007.
    [9] M. Chehlabi, Solving fuzzy dual complex linear systems, J. Appl. Math. Comput., 60 (2019), 87–112. https://doi.org/10.1007/s12190-018-1204-x doi: 10.1007/s12190-018-1204-x
    [10] D. Driankov, H. Hellendoorn, M. Reinfrank, An introduction to fuzzy control, Berlin, Heidelberg: Springer, 1996.
    [11] D. Dubois, H. Prade, Fuzzy sets and systems: Theory and applications, New York: Academic Press, 1980.
    [12] R. Ezzati, A method for solving dual fuzzy general linear systems, Appl. Comput. Math., 7 (2008), 235–241.
    [13] M. A. Fariborzi Araghi, M. M. Hoseinzadeh, Solution of general dual fuzzy linear systems using ABS algorithm, Appl. Math. Sci., 6 (2012), 163–171.
    [14] M. Friedman, M. Ma, A. Kandel, Fuzzy linear systems, Fuzzy Set. Syst., 96 (1998), 201–209. https://doi.org/10.1016/S0165-0114(96)00270-9
    [15] M. Ghanbari, T. Allahviranloo, W. Pedrycz, A straightforward approach for solving dual fuzzy linear systems, Fuzzy Set. Syst., 435 (2022), 89–106. https://doi.org/10.1016/j.fss.2021.04.007 doi: 10.1016/j.fss.2021.04.007
    [16] M. Ghanbari, T. Allahviranloo, W. Pedrycz, On the rectangular fuzzy complex linear systems, Appl. Soft Comput., 91 (2020), 106196. https://doi.org/10.1016/j.asoc.2020.106196 doi: 10.1016/j.asoc.2020.106196
    [17] Z. T. Gong, X. B. Guo, K. Liu, Approximate solution of dual fuzzy matrix equations, Inform. Sci., 266 (2014), 112–133. https://doi.org/10.1016/j.ins.2013.12.054 doi: 10.1016/j.ins.2013.12.054
    [18] M. Ghanbari, R. Nuraei, Convergence of a semi-analytical method on the fuzzy linear systems, Iran. J. Fuzzy Syst., 11 (2014), 45–60. https://doi.org/10.22111/IJFS.2014.1623 doi: 10.22111/IJFS.2014.1623
    [19] O. Kaleva, S. Seikkala, On fuzzy metric spaces, Fuzzy Set. Syst., 12 (1984), 215–229. https://doi.org/10.1016/0165-0114(84)90069-1
    [20] M. Ma, M. Friedman, A. Kandel, Duality in fuzzy linear systems, Fuzzy Set. Syst., 109 (2000), 55–58. https://doi.org/10.1016/S0165-0114(98)00102-X doi: 10.1016/S0165-0114(98)00102-X
    [21] S. Muzzioli, H. Reynaerts, Fuzzy linear systems of the form A_{1}x+b_{1} = A_{2}x+b_{2}, Fuzzy Set. Syst., 157 (2006), 939–951. https://doi.org/10.1016/j.fss.2005.09.005 doi: 10.1016/j.fss.2005.09.005
    [22] R. Nuraei, T. Allahviranloo, M. Ghanbari, Finding an inner estimation of the solution set of a fuzzy linear system, Appl. Math. Model., 37 (2013), 5148–5161. https://doi.org/10.1016/j.apm.2012.10.020 doi: 10.1016/j.apm.2012.10.020
    [23] M. Otadi, A New method for solving general dual fuzzy linear systems, J. Math. Ext., 7 (2013), 63–75.
    [24] X. D. Sun, S. Z. Guo, Solution to general fuzzy linear system and its necessary and sufficient condition, Fuzzy Inf. Eng., 1 (2009), 317–327.
    [25] A. Sarkar, G. Sahoo, U. C. Sahoo, Application of fuzzy logic in transport planning, Int. J. Soft Comput., 3 (2012), 1–21.
    [26] Z. F. Tian, X. B. Wu, Iterative method for dual fuzzy linear systems, In: Fuzzy information and engineering. Advances in soft computing, Berlin, Heidelberg: Springer, 2009.
    [27] C. X. Wu, M. Ma, Embedding problem of fuzzy number space: Part I, Fuzzy Set. Syst., 44 (1991), 33–38.
    [28] X. Z. Wang, Z. M. Zhong, M. H. Ha, Iteration algorithms for solving a system of fuzzy linear equations, Fuzzy Set. Syst., 119 (2005), 121–128. https://doi.org/10.1016/S0165-0114(98)00284-X doi: 10.1016/S0165-0114(98)00284-X
    [29] E. A. Youness, I. M. Mekawy, A study on fuzzy complex linear programming problems, Int. J. Contemp. Math. Sci., 7 (2012), 897–908.
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1554) PDF downloads(87) Cited by(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog