Research article Special Issues

Gradient-enhanced fractional physics-informed neural networks for solving forward and inverse problems of the multiterm time-fractional Burger-type equation

  • In this paper, we introduced the gradient-enhanced fractional physics-informed neural networks (gfPINNs) for solving the forward and inverse problems of the multiterm time-fractional Burger-type equation. The gfPINNs leverage gradient information derived from the residual of the fractional partial differential equation and embed the gradient into the loss function. Since the standard chain rule in integer calculus is invalid in fractional calculus, the automatic differentiation of neural networks does not apply to fractional operators. The automatic differentiation for the integer order operators and the finite difference discretization for the fractional operators were used to construct the residual in the loss function. The numerical results demonstrate the effectiveness of gfPINNs in solving the multiterm time-fractional Burger-type equation. By comparing the experimental results of fractional physics-informed neural networks (fPINNs) and gfPINNs, it can be seen that the training performance of gfPINNs is better than fPINNs.

    Citation: Shanhao Yuan, Yanqin Liu, Yibin Xu, Qiuping Li, Chao Guo, Yanfeng Shen. Gradient-enhanced fractional physics-informed neural networks for solving forward and inverse problems of the multiterm time-fractional Burger-type equation[J]. AIMS Mathematics, 2024, 9(10): 27418-27437. doi: 10.3934/math.20241332

    Related Papers:

    [1] Xiaoli Wang, Lizhen Wang . Traveling wave solutions of conformable time fractional Burgers type equations. AIMS Mathematics, 2021, 6(7): 7266-7284. doi: 10.3934/math.2021426
    [2] Zui-Cha Deng, Fan-Li Liu, Liu Yang . Numerical simulations for initial value inversion problem in a two-dimensional degenerate parabolic equation. AIMS Mathematics, 2021, 6(4): 3080-3104. doi: 10.3934/math.2021187
    [3] Mohammad Partohaghighi, Ali Akgül, Jihad Asad, Rania Wannan . Solving the time-fractional inverse Burger equation involving fractional Heydari-Hosseininia derivative. AIMS Mathematics, 2022, 7(9): 17403-17417. doi: 10.3934/math.2022959
    [4] M. J. Huntul . Inverse source problems for multi-parameter space-time fractional differential equations with bi-fractional Laplacian operators. AIMS Mathematics, 2024, 9(11): 32734-32756. doi: 10.3934/math.20241566
    [5] Humaira Yasmin, Aljawhara H. Almuqrin . Analytical study of time-fractional heat, diffusion, and Burger's equations using Aboodh residual power series and transform iterative methodologies. AIMS Mathematics, 2024, 9(6): 16721-16752. doi: 10.3934/math.2024811
    [6] Jian-Gen Liu, Jian Zhang . A new approximate method to the time fractional damped Burger equation. AIMS Mathematics, 2023, 8(6): 13317-13324. doi: 10.3934/math.2023674
    [7] Farman Ali Shah, Kamran, Zareen A Khan, Fatima Azmi, Nabil Mlaiki . A hybrid collocation method for the approximation of 2D time fractional diffusion-wave equation. AIMS Mathematics, 2024, 9(10): 27122-27149. doi: 10.3934/math.20241319
    [8] Asif Khan, Tayyaba Akram, Arshad Khan, Shabir Ahmad, Kamsing Nonlaopon . Investigation of time fractional nonlinear KdV-Burgers equation under fractional operators with nonsingular kernels. AIMS Mathematics, 2023, 8(1): 1251-1268. doi: 10.3934/math.2023063
    [9] Xiangtuan Xiong, Wanxia Shi, Xuemin Xue . Determination of three parameters in a time-space fractional diffusion equation. AIMS Mathematics, 2021, 6(6): 5909-5923. doi: 10.3934/math.2021350
    [10] Shuang-Shuang Zhou, Saima Rashid, Asia Rauf, Khadija Tul Kubra, Abdullah M. Alsharif . Initial boundary value problems for a multi-term time fractional diffusion equation with generalized fractional derivatives in time. AIMS Mathematics, 2021, 6(11): 12114-12132. doi: 10.3934/math.2021703
  • In this paper, we introduced the gradient-enhanced fractional physics-informed neural networks (gfPINNs) for solving the forward and inverse problems of the multiterm time-fractional Burger-type equation. The gfPINNs leverage gradient information derived from the residual of the fractional partial differential equation and embed the gradient into the loss function. Since the standard chain rule in integer calculus is invalid in fractional calculus, the automatic differentiation of neural networks does not apply to fractional operators. The automatic differentiation for the integer order operators and the finite difference discretization for the fractional operators were used to construct the residual in the loss function. The numerical results demonstrate the effectiveness of gfPINNs in solving the multiterm time-fractional Burger-type equation. By comparing the experimental results of fractional physics-informed neural networks (fPINNs) and gfPINNs, it can be seen that the training performance of gfPINNs is better than fPINNs.



    In recent years, fractional partial differential equations (FPDEs) have been widely used in natural science and engineering technology [1,2,3,4]. The advantage of FPDEs lies in their ability to better describe materials and processes that exhibit memory and genetic properties [5,6]. However, the solutions of FPDEs are much more complex. Many researchers have exploited diverse techniques for the investigation of FPDEs such as the finite difference method (FDM) [7], finite element method [8], spectral method [9], virtual element method [10], etc. The development of effective numerical methods to approximate FPDEs has been the goal of some researchers.

    In recent years, neural networks (NNs) have been successfully applied to solve problems in various fields [11,12,13]. Due to the high expressiveness of NNs in functional approximation [14,15,16], using NNs to solve differential and integral equations has become an active and important research field. Physics-informed neural networks (PINNs) [17,18,19,20] are machine learning models that combine deep learning with physical knowledge. PINNs embed PDEs into the loss function of the NNs, enabling the NNs to learn solutions to PDEs. The PINNs algorithm is meshless and simple, and can be applied to various types of PDEs, including integral differential equations, FPDEs, and random partial differential equations. Moreover, PINNs solved the inverse problem of PDEs just as easily as they solved the forward problem [17]. PINNs have been successfully applied to solve various problems in scientific computing [21,22,23]. Pang et al. [24] used the FDM to approximate the fractional derivatives that cannot be automatically differentiated, thus extending the PINNs to fPINNs for solving FPDEs.

    Despite the success of deep learning in the past, solving a wide range of PDEs is theoretically and practically challenging as complexity increases. Therefore, many aspects of PINNs need to be further improved to achieve more accurate predictions, higher computational efficiency, and robustness of training. Lu et al. [25] proposed DeepXDE, a deep learning library for solving PDEs, introduced a new residual-based adaptive refinement method to improve the training efficiency of PINNs, and new residual points were added at the position where the residuals of the PDEs were large, so that the discontinuities of PDEs could be captured well. Zhang et al. [26] combined fPINNs with the spectral method to solve the time-fractional phase field models. It had the characteristics of reducing the approximate number of discrete fractional operators, thus improving the training efficiency and obtaining higher error accuracy. Wu et al. [27] conducted a comprehensive study on two types of sampling of PINNs, including non-adaptive uniform sampling and adaptive non-uniform sampling, and the research results could also be used as a practical guide for selecting sampling methods. Zhang et al. [28] removed the soft constraints of PDEs in the loss function, and used the Lie symmetry group to generate the labeled data of PDEs to build a supervised learning model, thus effectively predicting the large amplitude and high frequency solutions of the Klein-Gordon equation. Zhang et al. [29] introduced the symmetry-enhanced physics-informed neural network (SPINN), which incorporated the invariant surface conditions derived from Lie symmetries or non-classical symmetries of PDEs into the loss function of PINNs, aiming to improve accuracy of PINNs. Lu et al. [30] and Xie et al. [31] introduced gradient-enhanced physics-informed neural networks (gPINNs) to solve PDEs and the idea of embedding the gradient information from the residuals of PDEs into the loss functions has also proven to be effective in other methods such as Gaussian process regression [32].

    In this paper, inspired by the above works, gfPINNs are applied to solve the forward and inverse problems of the multiterm time-fractional Burger-type equation. The integer order derivatives are handled using the automatic differentiation capability of the NNs, while the fractional derivatives of the equation are approximated using finite difference discretization [33,34]. Subsequently, the residual information of the equation is then incorporated into the loss function of NNs and optimized to yield optimal parameters. For the inverse problems of the multiterm time-fractional Burger-type equation, their overall form are known but the coefficient and the orders of time-fractional derivatives are unknown. The gfPINNs explicitly incorporate information from the equation by including the differential operators of the equation directly into the optimization loss function. The parameters to be identified appear in the differential operators, which are then optimized by minimizing the loss function associated with those parameters. A numerical comparison between fPINNs and gfPINNs is conducted using numerical examples. The numerical results demonstrate the effectiveness of gfPINNs in solving the multiterm time-fractional Burger-type equation.

    The structure of this paper is as follows. In Section 2, we define forward and inverse problems for the multiterm time-fractional Burger-type equation. In Section 3, we introduce fPINNs and gfPINNs and give the finite difference discretization to approximate the time-fractional derivatives. In Section 4, we demonstrate the effectiveness of gfPINNs in solving the forward and inverse problems of the multiterm time-fractional Burger-type equation by numerical examples, and compare the experimental results of fPINNs and gfPINNs. Finally, we give the conclusions of this paper in Section 5.

    We consider the following multiterm time-fractional Burger-type equation defined on the bounded domain Ω:

    c1C0Dαtu(x,t)+c2C0Dγtu(x,t)+u(x,t)u(x,t)x=v2u(x,t)x2+f(x,t), (2.1)

    where (x,t)Ω×[0,T] and the initial and boundary conditions are given as

    {u(x,t)=0,xΩ,u(x,0)=g(x),xΩ, (2.2)

    where u(x,t) is the solution of the equation, f(x,t) is the forcing term whose values are only known at scattered spatio-temporal coordinates, v is the kinematic viscosity of fluid, g(x) is a sufficiently smooth function, the fractional orders α and γ have been restricted to (0, 1) and (1, 2), respectively, C0Dθtu(x,t) is the Caputo time-fractional derivative of order θ (θ>0,n1θ<n) of u(x,t) with respect to t [35,36]:

    C0Dθtu(x,t)={1Γ(nθ)tα(ts)n1θnu(x,s)snds,θz+,θu(x,t)tθ,θz+, (2.3)

    where Γ() is the gamma function.

    The forward and inverse problems of solving the multiterm time-fractional Burger-type equation are described as follows. For the forward problem, under the given preconditions of the fractional orders α and γ, the forcing term f, and the initial and boundary conditions, the solution u(x,t) is solved. For the inverse problem, under the given preconditions of the initial and boundary conditions, the forcing term f, and additional concentration measurements at the final time u(x,t)=h(x,t), the fractional orders α and γ, the flow velocity v, and the solution u(x,t) are solved.

    This subsection introduces the idea of fPINNs and we consider both the forward and inverse problems, along with their corresponding NNs. We first consider the forward problem of the multiterm time-fractional Burger-type equation in the following form:

    {L{u(x,t)}=f(x,t),(x,t)Ω×[0,T],u(x,t)=0,xΩ,u(x,0)=g(x),xΩ, (3.1)

    where L{} is a nonlinear operator and L{u(x,t)}=c1C0Dαtu(x,t)+c2C0Dγtu(x,t)+u(x,t)u(x,t)xv2u(x,t)x2. We divide the nonlinear operator L{} into two parts, L=LAD+LnonAD. The first part is an integer derivative operator, which can be automatically differentiated (AD) using the chain rule. We have

    LAD{}={u(x,t)u(x,t)xv2u(x,t)x2,α(0,1),γ(1,2),c22u(x,t)t2+u(x,t)u(x,t)xv2u(x,t)x2,α(0,1),γ=2,c1u(x,t)t+u(x,t)u(x,t)xv2u(x,t)x2,α=1,γ(1,2), (3.2)

    and the second category consists of operators that lack automatic differentiation capabilities:

    LnonAD{}={c1C0Dαtu(x,t)+c2C0Dγtu(x,t),α(0,1),γ(1,2),c1C0Dαtu(x,t),α(0,1),γ=2,c2C0Dγtu(x,t),α=1,γ(1,2). (3.3)

    For LnonAD, we can discretize it using FDM and denote by LFDM the discretization version of LnonAD.

    During the NNs training process, our goal is to optimize its parameters in order to ensure that the approximate solution of the equation closely satisfies the initial and boundary conditions. The approximate solution is chosen as

    ˜u(x,t)=tρ(x)uNN(x,t)+g(x), (3.4)

    where uNN represents the output of the NNs. The NNs acts as a surrogate model, approximating the relationship between spatio-temporal coordinates and the solution of the equation. It is defined by its weights and biases, forming the parameter vector μ; see Figure 1 for a simple NN. This is fully connected with a single hidden layer consisting of three neurons. In this network, x and t are two inputs, which go through a linear transformation to obtain x1=w1x+w4t+b1, x2=w2x+w5t+b2, and x3=w3x+w6t+b3 in the hidden layer, and then, they go through a nonlinear transformation to get Yi=f(xi) for i=1,2,3. We choose the hyperbolic tangent function tanh(). Yi to go through a linear transformation to obtain the output of the NNs, uNN(x,t;μ)=w7Y1+w8Y2+w9Y3+b4. The vector of parameters μ is comprised of the weights wi and biases bi. ρ(0)=ρ(1)=0 and the auxiliary function ρ(x) is preselected. g(x) is the initial condition function such that it satisfies the initial and boundary conditions automatically.

    Figure 1.  A simple NN.

    The loss function of fPINNs for the forward problem with the approximate solution is defined as the mean-squared error of the equation residual

    LFW=1|SF|(x,t)S[LFDM{˜u(x,t)}+LAD{˜u(x,t)}f(x,t)]2, (3.5)

    where SFΩ×[0,T] and |SF| represents the number of training points. Then, we train the NNs to optimize the loss function of the forward problem with respect to the NNs parameters μ, thus obtaining the optimal parameters μbest. Finally, we specify a set of arbitrary test points to test the trained NNs and observe the training performance.

    The codes for solving the forward and inverse problems of the equation using NNs is similar. We only need to incorporate the parameters to be identified in the inverse problem into the loss function to be optimized in the forward problem, and no other changes are necessary. Next, we consider the following form of the inverse problem:

    {Lξ={α,γ,v}{u(x,t)}=f(x,t),(x,t)Ω×[0,T],u(x,t)=0,xΩ,u(x,0)=g(x),xΩ,u(x,t)=h(x,t),(x,t)Ω×[0,T], (3.6)

    where ξ is the parameter of the equation, so the loss function LIV for the inverse problem under consideration is

    LIV{μ,ξ={α,γ,v}}=WI11|SI1|(x,t)SI1[L{α,γ}FDM{˜u(x,t)}+LvAD{˜u(x,t)}f(x,t)]2+WI21|SI2|(x,t)SI2[˜u(x,t)h(x,t)]2, (3.7)

    where α(0,1) and γ(1,2), SI1Ω×[0,T] and SI2Ω×[0,T] are two sets of different training points, and WI1 and WI2 are preselected weight coefficients. We train the NNs to minimize the loss function, thereby obtaining αbest and γbest, the flow velocity vbest, and the optimal parameters μbest of the NNs.

    We incorporate the residual information of the equation into the loss function of NNs and train the NNs to minimize this loss function, thus obtaining the optimal parameters of NNs. If the residuals in the PDEs are zero, then the gradient of the residuals in the PDEs should also be zero. Therefore, adding gradient information to the loss function is a necessary condition for training NNs. One motivation behind gfPINNs is that the residual in the loss function often fluctuates near zero. Penalizing the slope of the residual can reduce these fluctuations, making the residual closer to zero. In this section, we continue to consider the formulation of the forward and inverse problems of the equation discussed in the previous section.

    We first consider the forward problem in the form of (3.1) and provide the loss function of gfPINNs for this form:

    LgFW=WFLFW+Wg1FLg1FW+Wg2FLg2FW, (3.8)

    where

    Lg1FW=1|Sg1F|(x,t)Sg1F[LFDM{˜u(x,t)}x+LAD{˜u(x,t)}xf(x,t)x]2, (3.9)
    Lg2FW=1|Sg2F|(x,t)Sg2F[LFDM{˜u(x,t)}t+LAD{˜u(x,t)}tf(x,t)t]2, (3.10)

    and the approximate solution of the equation is the same as Eq (3.4): ˜u(x,t)=ρ(x)uNN(x,t)+g(x). The expression LFW as shown in Eq (3.5), where WF, Wg1F, and Wg2F are preselected weighting coefficients, Sg1FΩ×[0,T] and Sg2FΩ×[0,T] are two sets of different training points.

    Next, we consider the inverse problem in the form of (3.6) and provide the loss function of gfPINNs for this form. The approach for the inverse problem of gfPINNs is similar to that of fPINNs. We provide the loss function for the inverse problem of gfPINNs.

    LgIV=WILIV{μ,ξ={α,γ,v}}+Wg1ILg1IV+Wg2ILg2IV, (3.11)

    where

    Lg1IV=Wg1I11|Sg1I1|(x,t)Sg1I1[L{α,γ}FDM{˜u(x,t)}x+LvAD{˜u(x,t)}xf(x,t)x]2+Wg1I21|Sg1I2|(x,t)Sg1I2[˜u(x,t)xh(x,t)x]2, (3.12)
    Lg2IV=Wg2I11|Sg2I1|(x,t)Sg2I1[L{α,γ}FDM{˜u(x,t)}t+LvAD{˜u(x,t)}tf(x,t)t]2+Wg2I21|Sg2I2|(x,t)Sg2I2[˜u(x,t)th(x,t)t]2, (3.13)

    and the expression LIV{μ,ξ={α,γ,v}} as shown in Eq (3.7), where WI, Wg1I, Wg2I, Wg1I1, Wg1I2, Wg2I1, and Wg2I2 are preselected weighting coefficients, Sg1I1,Sg2I1Ω×[0,T], Sg1I2,andSg2I2Ω×[0,T] are four sets of different training points.

    This defines the loss function of gfPINNs, which is exactly the same as discussed above for fPINNs. We train the NNs to obtain the optimal parameters of the NNs.

    In the x direction [0,M], we take the mesh points xp=ihx,i=0,1,2,...,M1, and in the t direction [0,T], we take the mesh points tn=nτ,n=0,1,...,N, where hx=MM1 and τ=TN are the uniform spatial step size and temporal step size, respectively. Denote Ωh{0iM1}, Ωτ{0nN}. Suppose uni = u(xi,tn) is a grid function on Ωh×Ωτ.

    We approximate the fractional derivatives of the equation using the finite difference discretization [33,34].

    For α(0,1), we have C0Dαtu(x,t)(xi,tn)=Dατ˜uni+R1(˜uni),

    Dατ˜uni:=ταΓ(2α)[aα0˜uni+n1k=1(aαnkaαnk1)˜ukiaαn1˜u0i], (3.14)

    where ˜uni=˜u(xi,tn), R1C(τ2α), and aαk=(k+1)1αk1α.

    Lemma 3.1. [33] α(0,1), aαl=(l+1)1αl1α, l=0,1,2,,

    (1) 1=aα0>aα1>aα2>>aαl>0,limlaαl0,

    (2) (1α)lα<a(α)l1<(1α)(l1)α,l1.

    For γ(1,2), C0Dγtu(x,t)(xi,tn)=Dγτ˜uni+R2(˜uni),

    Dγτ˜uni:=τ1γΓ(3γ)[bγ0δt˜uni+n1k=1(bγnkbγnk1)δt˜ukibγn1δt˜u0i], (3.15)

    where δtu(x,t)=u(x,t)t, R2C(τ3γ), and bγk=(k+1)2γk2γ.

    Given the spatial position x, it can be seen from the finite difference discretization that the time-fractional derivative of ˜u(x,t) evaluated at time t depends on the value of ˜u(x,t) calculated at all previous times 0, τ, 2τ, , t. We call the current time and the previous time the training points and the auxiliary points, respectively.

    In this section, we demonstrate the effectiveness of gfPINNs in solving forward and inverse problems of the multiterm time-fractional Burger-type equation and we compared fPINNs with gfPINNs. We solve the forward problems of the equation and present the experimental results in Section 4.1. We solve the inverse problems and present the experimental results in Section 4.2.

    We give a fabricated solution to the problem u(x,t)=tpsin(πx). In the given approximate solution (3.4), the auxiliary function ρ() is defined as ρ()=1x22. We use the following form of L2 relative error:

    {k[u(xtest,k,ttest,k)˜u(xtest,k,ttest,k)]2}12{k[u(xtest,k,ttest,k)]2}12 (4.1)

    to measure the performance of the NNs, where ˜u denotes the approximated solution, u is the exact solution, and (xk,tk) denotes the k-th test point.

    We wrote the code in Python and took advantage of the automatic differentiation capability of TensorFlow [37]. The stochastic gradient descent Adam algorithm [38] was used to optimize the loss function. We initialized the NNs parameters using normalized Glorot initialization [39]. Otherwise, when training a neural network, we set the learning rate, the number of neurons, the number of hidden layers, and the activation function as 1×103, 20, 4, and tanh(x), respectively.

    In this section, we consider the the multiterm time-fractional Burger-type equation of the form (2.1) with initial and boundary conditions (2.2). We let v=1, (x,t)[0,1]×[0,1], and g(x)=0, considering the smooth fabricated solution u(x,t)=tpsin(πx) and the forcing term

    f(x,t)=Γ(p+1)Γ(p+1α)t(pα)(pα)sin(πx)+Γ(p+1)Γ(p+1γ)t(pγ)(pγ)sin(πx)+t2psin(πx)cos(πx)+π2tpsin(πx). (4.2)

    Case 1: We choose c1=1, c2=0, and α=0.5, considering the smooth fabricated solution u(x,t)=t4sin(πx) and the forcing term f(x,t)=3.5Γ(5)Γ(4.5)t3.5sin(πx)+t8sin(πx)cos(πx)+π2t4sin(πx). We consider M11 training points of the spatial domain: xi=ihx for i=1,2,,M11 and N training points of the time domain: tn=nτ for n=1,2,,N. We do not need to place training points on the initial and boundary since the approximate solution ˜u(x,t)=tx(1x)uNN(x,t;μ) satisfies the initial and boundary conditions automatically. For fPINNs, the loss function can be written as

    LFW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[a0.50˜u(xi,tn)+n1k=1(a0.5nka0.5nk1)˜u(xi,tk)]+˜u(xi,tn)˜u(xi,tn)xi2˜u(xi,tn)x2if(xi,tn)}2. (4.3)

    The loss function of gfPINNs can be given as

    Lg2FW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[a0.50˜u(xi,tn)xi+n1k=1(a0.5nka0.5nk1)˜u(xi,tk)xi]+˜u(xi,tn)2˜u(xi,tn)x2i+(˜u(xi,tn)xi)23˜u(xi,tn)x3if(xi,tn)xi}2, (4.4)
    Lg2FW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[a0.50˜u(xi,tn)tn+n1k=1(a0.5nka0.5nk1)˜u(xi,tk)tk]+˜u(xi,tn)2˜u(xi,tn)xitn+˜u(xi,tn)xi˜u(xi,tn)tn3˜u(xi,tn)x2itnf(xi,tn)tn}2. (4.5)

    By substituting Eqs (4.3)–(4.5) into Eq (3.8), we get the gfPINNs loss function LgFW with WF=1, Wg1F=1, and Wg2F=1. Next, we selected 2000 training points to train fPINNs and gfPINNs and other parameters of the NNs are set to those described at the beginning of this section. Figures 24 present a comparison between the predicted solutions from the fPINNs and gfPINNs models and the exact solution of the equation, demonstrating that gfPINNs can effectively solve the equation. Figure 5 shows the absolute errors between the exact solution and the solutions predicted by fPINNs and gfPINNs, and it can be seen that the prediction performance of gfPINNs is better than that of fPINNs. Figure 6 illustrates the L2 relative errors of both fPINNs and gfPINNs models for a single experiment as the iteration count varies, showing that while both can achieve errors as low as 104, gfPINNs exhibits comparatively lower error and reduced oscillation.

    Figure 2.  The exact solution and predicted solutions of the equation.
    Figure 3.  The exact solution and numerical solutions' profiles of velocity u(x,t) with α=0.5.
    Figure 4.  Predicted cross-sectional views of the equation using fPINNs and gfPINNs.
    Figure 5.  The absolute errors for solutions predicted by fPINNs and gfPINNs.
    Figure 6.  The L2 relative error of the problem with the number of iterations.

    Case 2: We choose c1=0, c2=1, and γ=1.5, considering the smooth fabricated solution u(x,t)=t4sin(πx) and the forcing term f(x,t)=2.5Γ(5)Γ(3.5)t2.5sin(πx)+t8sin(πx)cos(πx)+π2t4sin(πx). Similarly, we give the loss function of fPINNs as

    LFW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[b1.50˜u(xi,tn)tn+n1k=1(b1.5nkb1.5nk1)˜u(xi,tn)tk]+˜u(xi,tn)˜u(xi,tn)xi2˜u(xi,tn)x2if(xi,tn)}2. (4.6)

    For gfPINNs, the loss function can be written as

    Lg1FW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[b1.502˜u(xi,tn)tnxi+n1k=1(b1.5nkb1.5nk1)2˜u(xi,tk)tkxi]+˜u(xi,tn)2˜u(xi,tn)x2i+(˜u(xi,tn)xi)23˜u(xi,tn)x3if(xi,tn)xi}2, (4.7)
    Lg2FW=1(M11)NM1i=1Nn=1{τ0.5Γ(1.5)[b1.502˜u(xi,tn)t2n+n1k=1(b1.5nkb1.5nk1)2˜u(xi,tk)t2k]+˜u(xi,tn)2˜u(xi,tn)xitn+˜u(xi,tn)xi˜u(xi,tn)tn3˜u(xi,tn)x2itnf(xi,tn)tn}2. (4.8)

    By substituting Eqs (4.6)–(4.8) into Eq (3.8), we get the gfPINNs loss function LgFW with WF=1, Wg1F=0.16, and Wg2F=0.16. Next, we selected 2000 training points to train fPINNs and gfPINNs and other parameters of the NNs are set to those described at the beginning of this section. Figures 79 present a comparison between the predicted solutions from the fPINNs and gfPINNs models and the exact solution of the equation, demonstrating that gfPINNs can effectively solve the equation. Figure 10 illustrates the absolute errors between the exact solution and the solutions predicted by both fPINNs and gfPINNs, revealing that the gfPINNs exhibit a relatively smaller absolute error. Figure 11 presents the iteration convergence curves for both the fPINNs and gfPINNs models for a single experiment, revealing that while both can achieve L2 relative errors of 104 with increasing iterations, the prediction errors of gfPINNs are relatively low and more stable, resulting in superior prediction performance compared to fPINNs.

    Figure 7.  The exact solution and predicted solutions of the equation.
    Figure 8.  The exact solution and numerical solutions' profiles of velocity u(x,t) with γ=1.5.
    Figure 9.  Predicted cross-sectional views of the equation using fPINNs and gfPINNs.
    Figure 10.  The absolute errors for solutions predicted by fPINNs and gfPINNs.
    Figure 11.  The L2 relative error of the problem with the number of iterations.

    We use the code that solves the forward problem to solve the inverse problem. We simply add the parameters to be identified in the inverse problem to the list of parameters to be optimized in the forward problem, without changing anything else. In this section, gfPINNs are applied to solve the inverse problems of the multiterm time-fractional Burger-type equation of the form (3.6). We let v=1, (x,t)[0,1]×[0,1], g(x)=0, and considering additional concentration measurements at the final time u(x,1)=h(x,1). Here, we still consider the smooth fabricated solution u(x,t)=tpsin(πx) and the forcing term of formula (4.2).

    Case 1: We choose c1=1 and c2=0. Similarly, we get the gfPINNs loss function LgFW with WI=1, Wg1I=0.25, and Wg2I=0.25. We set the fractional derivative to be 0.6. We selected 470 training points to train fPINNs and gfPINNs and other parameters of the NNs are set to those described at the beginning of this section. Figures 1214 display a comparison between the predicted solutions from the fPINNs and gfPINNs models and the exact solution of the equation, demonstrating that gfPINNs can effectively solve the problem. Figure 15 illustrates the absolute errors between the exact solution and the solutions predicted by both fPINNs and gfPINNs, revealing that the gfPINNs exhibit a relatively smaller and more stable absolute error. Figure 16 illustrates the iteration convergence curves for the fPINNs and gfPINNs for a single experiment, indicating that although gfPINNs incur a higher computational cost for solving the inverse problem due to an additional loss term, both models can achieve L2 relative errors of 104 as iterations progress, with gfPINNs showing a lower and more stable error curve compared to fPINNs.

    Figure 12.  The exact solution and predicted solutions of the equation.
    Figure 13.  The exact solution and numerical solutions' profiles of velocity u(x,t).
    Figure 14.  Predicted cross-sectional views of the equation using fPINNs and gfPINNs.
    Figure 15.  The absolute errors for solutions predicted by fPINNs and gfPINNs.
    Figure 16.  The L2 relative error of the problem with the number of iterations.

    Case 2: We choose c1=0 and c2=1. Similarly, we get the gfPINNs loss function LgFW with WI=1, Wg1I=0.16, and Wg2I=0.0001. We set the fractional derivative to be 1.6. We selected 400 training points to train fPINNs and gfPINNs and other parameters of the NNs are set to those described at the beginning of this section. For fPINNs and gfPINNs, we get the similar conclusion as Case 1 by training the NNs and observing the experimental results. Figures 1719 display a comparison between the predicted solutions from the fPINNs and gfPINNs models and the exact solution of the equation, demonstrating that gfPINNs can effectively solve the problem. Figure 20 illustrates the absolute errors between the exact solution and the solutions predicted by both fPINNs and gfPINNs, revealing that the gfPINNs exhibit a relatively smaller absolute error. Figure 21 compares the L2 relative errors of fPINNs and gfPINNs for a single experiment as iterations progress, revealing that while gfPINNs incurs a higher computational cost due to an additional loss term, both models can achieve an L2 relative error of 103, with gfPINNs demonstrating a lower and more stable error curve than fPINNs.

    Figure 17.  The exact solution and predicted solutions of the equation.
    Figure 18.  The exact solution and numerical solutions' profiles of velocity u(x,t).
    Figure 19.  Predicted cross-sectional views of the equation using fPINNs and gfPINNs.
    Figure 20.  The absolute errors for solutions predicted by fPINNs and gfPINNs.
    Figure 21.  The L2 relative error of the problem with the number of iterations.

    In this paper, the effectiveness of gfPINNs in solving the forward and inverse problems of the multiterm time-fractional Burger-type equation is verified through numerical examples. The L2 relative errors for solutions predicted by both fPINNs and gfPINNs can achieve 104 for forward problems and 103 or even 104 for inverse problems. The experimental results indicate that gfPINNs demonstrate relatively lower and more stable errors with the increase of training iterations, thereby enhancing prediction performance. Nonetheless, the inclusion of an additional loss term in gfPINNs may result in a higher computational cost, such as when solving inverse problems, fPINNs exhibit faster convergence compared to gfPINNs.

    Shanhao Yuan, Yanqin Liu, Qiuping Li and Chao Guo: Conceptualization, Methodology; Yibin Xu, Shanhao Yuan and Yanfeng Shen: Software, Visualization, Validation; Shanhao Yuan: Writing–Original draft preparation; Yanqin Liu: Writing–Reviewing & editing. All authors have read and approved the final version of the manuscript for publication.

    We appreciated the support by the Natural Science Foundation of Shandong Province (ZR2023MA062), the National Science Foundation of China (62103079), the Belt and Road Special Foundation of The National Key Laboratory of Water Disaster Prevention (2023491911), and the Open Research Fund Program of the Data Recovery Key Laboratory of Sichuan Province (DRN19020).

    The authors declare that they have no conflicts of interest.



    [1] L. Cristofaro, R. Garra, E. Scalas, I. Spassiani, A fractional approach to study the pure-temporal epidemic type aftershock sequence (ETAS) process for earthquakes modeling, Fract. Calc. Appl. Anal., 26 (2023), 461–479. https://doi.org/10.1007/s13540-023-00144-5 doi: 10.1007/s13540-023-00144-5
    [2] Y. Zhang, H. G. Sun, H. H. Stowell, M. Zayernouri, S. E. Hansen, A review of applications of fractional calculus in earth system dynamics, Chaos Solitons Fract., 102 (2017), 29–46. https://doi.org/10.1016/j.chaos.2017.03.051 doi: 10.1016/j.chaos.2017.03.051
    [3] M. I. Molina, Fractional electrical impurity, New J. Phys., 26 (2024), 013020. https://doi.org/10.1088/1367-2630/ad19f8 doi: 10.1088/1367-2630/ad19f8
    [4] Y. Q. Yang, Q. W. Qi, J. Y. Hu, J. S. Dai, C. D. Yang, Adaptive fault-tolerant control for consensus of nonlinear fractional-order multi-agent systems with diffusion, Fractal Fract., 7 (2023), 1–20. https://doi.org/10.3390/fractalfract7100760 doi: 10.3390/fractalfract7100760
    [5] P. Baliarsingh, L. Nayak, Fractional derivatives with variable memory, Iran. J. Sci. Technol. Trans. A Sci., 46 (2022), 849–857. https://doi.org/10.1007/s40995-022-01296-4 doi: 10.1007/s40995-022-01296-4
    [6] J. B. Hu, Studying the memory property and event-triggered control of fractional systems, Inform. Sci., 662 (2024), 120218. https://doi.org/10.1016/j.ins.2024.120218 doi: 10.1016/j.ins.2024.120218
    [7] J. Guo, D. Xu, W. L. Qiu, A finite difference scheme for the nonlinear time‐fractional partial integro‐differential equation, Math. Methods Appl. Sci., 43 (2020), 3392–3412. https://doi.org/10.1002/mma.6128 doi: 10.1002/mma.6128
    [8] H. Z. Hu, Y. P. Chen, J. W. Zhou, Two-grid finite element method for time-fractional nonlinear schrodinger equation, J. Comp. Math., 42 (2024), 1124–1144. https://doi.org/10.4208/jcm.2302-m2022-0033 doi: 10.4208/jcm.2302-m2022-0033
    [9] W. Zhang, C. X. Wu, Z. S. Ruan, S. F. Qiu, A Jacobi spectral method for calculating fractional derivative based on mollification regularization, Asymptot. Anal., 136 (2024), 61–77. https://doi.org/10.3233/ASY-231869 doi: 10.3233/ASY-231869
    [10] Q. L. Gu, Y. P. Chen, J. W. Zhou, J. Huang, A fast linearized virtual element method on graded meshes for nonlinear time-fractional diffusion equations, Numer. Algorithms, 2024. https://doi.org/10.1007/s11075-023-01744-1
    [11] S. S. Yu, M. Guo, X. Y. Chen, J. L. Qiu, J. Q. Sun, Personalized movie recommendations based on a multi-feature attention mechanism with neural networks, Mathematics, 11 (2023), 1–22. https://doi.org/10.3390/math11061355 doi: 10.3390/math11061355
    [12] X. Y. Ding, J. Q. Lu, X. Y. Chen, Lyapunov-based stability of time-triggered impulsive logical dynamic networks, Nonlinear Analy. Hybrid Syst., 51 (2024), 101417. https://doi.org/10.1016/j.nahs.2023.101417 doi: 10.1016/j.nahs.2023.101417
    [13] T. G. Yang, G. C. Li, T. Y. Wang, S. Y. Yuan, X. Y. Yang, X. G. Yu, et al., A novel 1D-convolutional spatial-time fusion strategy for data-driven fault diagnosis of aero-hydraulic pipeline systems, Mathematics, 11 (2023), 1–21. https://doi.org/10.3390/math11143113 doi: 10.3390/math11143113
    [14] L. Lu, Y. H. Su, G. E. Karniadakis, Collapse of deep and narrow neural nets, 2018, arXiv: 1808.04947.
    [15] Y. Q. Liu, T. Mao, D. X. Zhou, Approximation of functions from Korobov spaces by shallow neural networks, Inform. Sci., 670 (2024), 120573. https://doi.org/10.1016/j.ins.2024.120573 doi: 10.1016/j.ins.2024.120573
    [16] G. A. Anastassiou, D. Kouloumpou, Neural network approximation for time splitting random functions, Mathematics, 11 (2023), 1–25. https://doi.org/10.3390/math11092183 doi: 10.3390/math11092183
    [17] M. Raissi, P. Perdikaris, G. E. Karniadakis, Physics-informed neural networks: a deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations, J. Comput. Phys., 378 (2019), 686–707. https://doi.org/10.1016/j.jcp.2018.10.045 doi: 10.1016/j.jcp.2018.10.045
    [18] Q. Z. Hou, Y. X. Li, V. P. Singh, Z. W. Sun, Physics-informed neural network for diffusive wave model, J. Hydrology, 637 (2024), 131261. https://doi.org/10.1016/j.jhydrol.2024.131261 doi: 10.1016/j.jhydrol.2024.131261
    [19] S. M. Sivalingam, P. Kumar, V. Govindaraj, A neural networks-based numerical method for the generalized Caputo-type fractional differential equations, Math. Comput. Simul., 213 (2023), 302–323. https://doi.org/10.1016/j.matcom.2023.06.012 doi: 10.1016/j.matcom.2023.06.012
    [20] Q. Z. Hou, Y. X. Li, V. P. Singh, Z. W. Sun, J. G. Wei, Physics-informed neural network for solution of forward and inverse kinematic wave problems, J. Hydrology, 633 (2024), 130934. https://doi.org/10.1016/j.jhydrol.2024.130934 doi: 10.1016/j.jhydrol.2024.130934
    [21] H. Bararnia, M. Esmaeilpour, On the application of physics informed neural networks (PINN) to solve boundary layer thermal-fluid problems, Int. Commun. Heat Mass Transfer, 132 (2022), 105890. https://doi.org/10.1016/j.icheatmasstransfer.2022.105890 doi: 10.1016/j.icheatmasstransfer.2022.105890
    [22] X. P. Zhang, Y. Zhu, J. Wang, L. L. Ju, Y. Z. Qian, M. Ye, et al., GW-PINN: a deep learning algorithm for solving groundwater flow equations, Adv. Water Resour., 165 (2022), 104243. https://doi.org/10.1016/j.advwatres.2022.104243 doi: 10.1016/j.advwatres.2022.104243
    [23] S. P. Zheng, Y. Y. Lin, J. H. Feng, F. Jin, Viscous regularization PINN algorithm for shallow water equations (Chinese), Chinese J. Comput. Phys., 40 (2023), 314–324.
    [24] G. F. Pang, L. Lu, G. E. Karniadakis, fPINNs: fractional physics-informed neural networks, SIAM J. Sci. Comput, 41 (2019), A2603–A2626. https://doi.org/10.1137/18M1229845 doi: 10.1137/18M1229845
    [25] L. Lu, X. H. Meng, Z. P. Mao, G. E. Karniadakis, DeepXDE: a deep learning library for solving differential equations, SIAM Rev., 63 (2021), 208–228. https://doi.org/10.1137/19M1274067 doi: 10.1137/19M1274067
    [26] S. P. Wang, H. Zhang, X. Y. Jiang, Fractional physics-informed neural networks for time-fractional phase field models, Nonlinear Dyn., 110 (2022), 2715–2739. https://doi.org/10.1007/s11071-022-07746-3 doi: 10.1007/s11071-022-07746-3
    [27] C. X. Wu, M. Zhu, Q. Y. Tan, Y. Kartha, L. Lu, A comprehensive study of non-adaptive and residual-based adaptive sampling for physics-informed neural networks, Comput. Methods Appl. Mech. Eng., 403 (2023), 115671. https://doi.org/10.1016/j.cma.2022.115671 doi: 10.1016/j.cma.2022.115671
    [28] Z. Y. Zhang, S. J. Cai, H. Zhang, A symmetry group based supervised learning method for solving partial differential equations, Comput. Methods Appl. Mech. Eng., 414 (2023), 116181. https://doi.org/10.1016/j.cma.2023.116181 doi: 10.1016/j.cma.2023.116181
    [29] Z. Y. Zhang, H. Zhang, L. S. Zhang, L. L. Guo, Enforcing continuous symmetries in physics-informed neural network for solving forward and inverse problems of partial differential equations, J. Comput. Phys., 492 (2023), 112415. https://doi.org/10.1016/j.jcp.2023.112415 doi: 10.1016/j.jcp.2023.112415
    [30] J. Yu, L. Lu, X. H. Meng, G. E. Karniadakis, Gradient-enhanced physics-informed neural networks for forward and inverse PDE problems, Comput. Methods Appl. Mech. Eng., 393 (2022), 114823. https://doi.org/10.1016/j.cma.2022.114823 doi: 10.1016/j.cma.2022.114823
    [31] G. Z. Xie, B. B. Fu, H. Li, W. L. Du, Y. D. Zhong, L. W. Wang, et al., A gradient-enhanced physics-informed neural networks method for the wave equation, Eng. Anal. Bound. Elem., 166 (2024), 105802. https://doi.org/10.1016/j.enganabound.2024.105802 doi: 10.1016/j.enganabound.2024.105802
    [32] Y. X. Deng, G. Lin, X. Yang, Multifidelity data fusion via gradient-enhanced Gaussian process regression, Commun. Comput. Phys., 28 (2020), 1812–1837. https://doi.org/10.4208/cicp.OA-2020-0151 doi: 10.4208/cicp.OA-2020-0151
    [33] Z. Z. Sun, X. N. Wu, A fully discrete difference scheme for a diffusion-wave system, Appl. Numer. Math., 56 (2006), 193–209. https://doi.org/10.1016/j.apnum.2005.03.003 doi: 10.1016/j.apnum.2005.03.003
    [34] B. T. Jin, R. Lazarovl, Z. Zhou, An analysis of the L1 scheme for the subdiffusion equation with nonsmooth data, IMA J. Numer. Anal., 36 (2016), 197–221. https://doi.org/10.1093/imanum/dru063 doi: 10.1093/imanum/dru063
    [35] A. A. Kilbas, H. M. Srivastava, J. J. Trujillo, Theory and applications of fractional differential equations, Amsterdam: Elsevier, 2006.
    [36] I. Podlubny, Fractional differential equations, Academic Press, 1999.
    [37] M. Ramchandani, H. Khandare, P. Singh, P. Rajak, N. Suryawanshi, A. S. Jangde, et al., Survey: Tensorflow in machine learning, J. Phys. Conf. Ser., 2273 (2022), 012008. https://doi.org/10.1088/1742-6596/2273/1/012008 doi: 10.1088/1742-6596/2273/1/012008
    [38] D. P. Kingma, J. Ba, Adam: a method for stochastic optimization, In: Proceedings of the 3rd International Conference on Learning Representations (ICLR), San Diego, 2015.
    [39] M. V. Narkhede, P. P. Bartakke, M. S. Sutaone, A review on weight initialization strategies for neural networks, Artif. Intell. Rev., 55 (2022), 291–322. https://doi.org/10.1007/s10462-021-10033-z doi: 10.1007/s10462-021-10033-z
  • This article has been cited by:

    1. Jiawei Wang, Yanqin Liu, Limei Yan, Kunling Han, Libo Feng, Runfa Zhang, Fractional sub-equation neural networks (fSENNs) method for exact solutions of space–time fractional partial differential equations, 2025, 35, 1054-1500, 10.1063/5.0259937
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1127) PDF downloads(87) Cited by(1)

Figures and Tables

Figures(21)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog