Processing math: 100%
Research article Special Issues

A weighted online regularization for a fully nonparametric model with heteroscedasticity

  • In this paper, combining B-spline function and Tikhonov regularization, we propose an online identification approach for reconstructing a smooth function and its derivative from scattered data with heteroscedasticity. Our methodology offers the unique advantage of enabling real-time updates based on new input data, eliminating the reliance on historical information. First, to address the challenge of heteroscedasticity and computation cost, we employ weight coefficients along with a judiciously chosen set of knots for interpolation. Second, a reasonable approach is provided to select weight coefficients and the regularization parameter in objective functional. Finally, We substantiate the efficacy of our approach through a numerical example and demonstrate its applicability in solving inverse problems. It is worth mentioning that the algorithm not only ensures the calculation efficiency, but also trades the data accuracy through the data volume.

    Citation: Lei Hu. A weighted online regularization for a fully nonparametric model with heteroscedasticity[J]. AIMS Mathematics, 2023, 8(11): 26991-27008. doi: 10.3934/math.20231381

    Related Papers:

    [1] Cuiping Wang, Xiaoshuang Zhou, Peixin Zhao . Empirical likelihood based heteroscedasticity diagnostics for varying coefficient partially nonlinear models. AIMS Mathematics, 2024, 9(12): 34705-34719. doi: 10.3934/math.20241652
    [2] Junke Kou, Hao Zhang . Wavelet estimations of the derivatives of variance function in heteroscedastic model. AIMS Mathematics, 2023, 8(6): 14340-14361. doi: 10.3934/math.2023734
    [3] Liqi Xia, Xiuli Wang, Peixin Zhao, Yunquan Song . Empirical likelihood for varying coefficient partially nonlinear model with missing responses. AIMS Mathematics, 2021, 6(7): 7125-7152. doi: 10.3934/math.2021418
    [4] Gaosheng Liu, Yang Bai . Statistical inference in functional semiparametric spatial autoregressive model. AIMS Mathematics, 2021, 6(10): 10890-10906. doi: 10.3934/math.2021633
    [5] Bin Yang, Min Chen, Jianjun Zhou . Forecasting the monthly retail sales of electricity based on the semi-functional linear model with autoregressive errors. AIMS Mathematics, 2025, 10(1): 1602-1627. doi: 10.3934/math.2025074
    [6] Xin Liang, Xingfa Zhang, Yuan Li, Chunliang Deng . Daily nonparametric ARCH(1) model estimation using intraday high frequency data. AIMS Mathematics, 2021, 6(4): 3455-3464. doi: 10.3934/math.2021206
    [7] Fatimah Alshahrani, Wahiba Bouabsa, Ibrahim M. Almanjahie, Mohammed Kadi Attouch . Robust kernel regression function with uncertain scale parameter for high dimensional ergodic data using $ k $-nearest neighbor estimation. AIMS Mathematics, 2023, 8(6): 13000-13023. doi: 10.3934/math.2023655
    [8] Khanittha Talordphop, Yupaporn Areepong, Saowanit Sukparungsee . An empirical assessment of Tukey combined extended exponentially weighted moving average control chart. AIMS Mathematics, 2025, 10(2): 3945-3960. doi: 10.3934/math.2025184
    [9] Samuel Asante Gyamerah, Philip Ngare, Dennis Ikpe . Mitigating geographical basis risk of weather derivatives using spatial-temporal regime-switching temperature model. AIMS Mathematics, 2019, 4(4): 1274-1290. doi: 10.3934/math.2019.4.1274
    [10] Xiaowei Zhang, Junliang Li . Model averaging with causal effects for partially linear models. AIMS Mathematics, 2024, 9(6): 16392-16421. doi: 10.3934/math.2024794
  • In this paper, combining B-spline function and Tikhonov regularization, we propose an online identification approach for reconstructing a smooth function and its derivative from scattered data with heteroscedasticity. Our methodology offers the unique advantage of enabling real-time updates based on new input data, eliminating the reliance on historical information. First, to address the challenge of heteroscedasticity and computation cost, we employ weight coefficients along with a judiciously chosen set of knots for interpolation. Second, a reasonable approach is provided to select weight coefficients and the regularization parameter in objective functional. Finally, We substantiate the efficacy of our approach through a numerical example and demonstrate its applicability in solving inverse problems. It is worth mentioning that the algorithm not only ensures the calculation efficiency, but also trades the data accuracy through the data volume.



    The underlying model of this article can be formulated quite easily: given some real valued variables x and y fulfill the heteroscedastic model:

    y=f(x)+σ(x)ε, (1.1)

    where f(x) belongs to Sobolev space W2,2(0,1), variance function σ(x) is a positive function defined on (0,+) and the random error term ε is independent of x and satisfies E(ε)=0 and Var(ε)=1.

    Such a problem is widely applied in various fields such as Computerized Tomography (CT) and the inverse problem of option pricing (IPOP) in [1,2,3]. Moreover, this is a classical ill-posed problem. There are plenty of regularization methods for this ill-posed problem in one dimension or higher dimensions [4,5]. Researchers [6,7,8,9] have proposed some statistical methods to solve this ill-posed problem. Zhang [10] employs a relatively small number of knots for interpolation to reduce computation cost. These classical approaches are based on the equal variance of the error term and, when N is large enough, needs large amount of calculating. However, the variance σ(x) in (1.1) is typically unequal, a phenomenon known as heteroscedasticity [11,12].

    Heteroscedasticity, which is an econometric term, is the variances of perturbations σ(x) in the model which are not exactly equal[13]. Heteroscedasticity will lead to large errors in the model. In order to reduce the error caused by heteroscedasticity, the principal methods are mainly divided into three parts. The most common method is to take the logarithm of the data[14]. However, not all data must be logarithmic. The second method is Robust Standard Error Regression, which is the most popular and effective treatment method at present[15,16]. The main idea is to modify the objective function in the classical least squares regression, which is very sensitive to outliers. Robust method is only applicable to heteroscedasticity and independent observation. Another method is FGLS regression[17]. For points with larger residual value, the smaller weight is given to solve the heteroscedasticity problem. It is still the least square method in essence. However, the convergence speed of this method is slow, that is, the deviation of limited samples will be large.

    B-spline functions are crucial elements in various fields, especially in computer graphics, computer-aided design (CAD), and numerical analysis[18,19,20]. They play a significant role in curve and surface representation, interpolation, approximation, and modeling[21,22,23]. B-spline functions offer a versatile mathematical framework for representing complex shapes and data, enabling efficient and accurate solutions in various industries and scientific disciplines. Their ability to balance smoothness, flexibility, and local control makes them a cornerstone of computational modeling and design processes. Among them, cubic splines are the most representative. Cubic B-spline functions strike a balance between simplicity and expressiveness. They are more flexible than linear or quadratic B-splines, allowing for smoother curves and surfaces, while still being relatively straightforward to manipulate.

    In this paper, based on B-splines function and FGLS regression, we propose a weighted online regularization method which apply weight coefficients and a relatively small number of knots for interpolation to reduce the effect of heteroscedasticity and computation cost, and exchange the amount of data for the accuracy of reconstruction. Moreover, this algorithm has the characteristics of smaller memory occupation and can process online data, which means that when new data is added, there is no need to re-process the processed data.

    The paper is organized as follows. In Section 2, we introduce the issue which is to be studied and propose an online reconstruction algorithm. Error estimation is carried out in Section 3. Section 4 provides a principle of optimal selection for weight coefficients and the regularization parameter in objective function. The performance of the proposed algorithm is illustrated in Section 5. In Section 6, we present two applications of our reconstruction algorithm in inverse problems. Section 7 concludes main results.

    In this section, we consider the following problem: for a lager positive integer N, we try to reconstruct function f(x) and its derivative f(x) in the model (1.1) from observation data {(xi,yi)}Ni=1 and function σ(x).

    First, we define the Tikhonov objective functional as follows:

    J(g)=1NNi=1wi(g(xi)yi)2+α||g(x)||2L2(0,1), (2.1)

    where g(x)W2,2(0,1), {wi}Ni=1 are weight coefficients and α is a regularization parameter. The minimization problem is reformulated as follows:

    fN=argmingJ(g)=argming(1NNi=1wi(g(xi)yi)2+α||g(x)||2L2(0,1)). (2.2)

    fN is defined as the approximated solution of f(x) in (1.1).

    In this paper, we reconstruct not only function f(x) but also its derivative f(x). This is a classical ill-posed problem. It is computationally costly to solve the problem with local regression methods such as kernel regression and local polynomial regression. This problem is usually solved by regularization method and spline functions. Moreover, B-spline basis functions have compact support which makes it possible to speed up calculations. Thus, we choose B-spline function.

    Then we solve suitable g(x) in the finite dimensional function space of cubic B-spline function instead of the infinite dimensional space. Inspired by the method in [25], we just select some equidistant points not sample points as interpolation node.

    Let M be a positive integer and mesh size d=M1. Equidistant knots {pj}Mj=0 are defined as

    pj=jd,j=1,2,,M.

    The vector space of all cubic B-spline functions with knots {pj}Mj=0 is called space Vm. Assume function g(x) belongs to space Vm, g(x) can be written as

    g(x)=M+1j=1λjϕj(x), (2.3)

    where {λj}M+1j=1 are constants and ϕj(x)=ϕ(xpjd). ϕ(x) is standard cubic B-spline function defined in [24].

    Denote the weight coefficient W=diag(w1,w2,,wN), the noisy sample y=(y1,y2,,yN)T, the function parameter λ=(λ1,λ0,,λM+1)T and the row vector Hx=(ϕ1(x),ϕ0(x),,ϕM+1(x)). Through the Eq (2.3) and the definition above, the Eq (2.2) can be rewritten as

    λN=argminλJ1(λ):=argminλ(1N(WHλWy)T(WHλWy)+αλTPλ), (2.4)

    where H:=(Hx1,Hx2,,HxN)T and PR(M+3)×(M+3) is given by

    P=(pij)M+1i,j=1,pij=10ϕi(x)ϕj(x)dx.

    When λN is known, the approximated solution fN(x)=HxλN.

    Theorem 1. Suppose N>2, and the observation points {xi}Ni=1 are not identical. Then, the minimization problem (2.4) has a unique minimizer λN which satisfies:

    (1NHTWTWH+αP)λN=1NHTWTWy. (2.5)

    Proof. Since J1(λ) in (2.4) is a quadratic form with respect to λ, the derivative of J1(λ) takes zero only at λ=λN, that is

    (1NHTWTWH+αP)λN=1NHTWTWy.

    Moreover, P is positive definite and HTWTWH is positive semidefinite. Thus, λN is unique.

    Through the theorem above, we propose a online Tikhonov regularization algorithm in Algorithm 1. In "Require" sentence, we input the observation data {(xi,yi)}Ni=1 and the necessary parameters–the number of knots M, weight coefficients W in (4.6) and regularization parameter α in Table 1. The algorithm proceeds from Line 1 to Line 8. In Line 1–2, we initialize matrix A0=0R(M+3)×(M+3) and b0=0R and generate the matrix PR(M+3)×(M+3). Line 3–6 provides a method to update A and b in (2.5) according to the newly entered data. Through solving linear system in Line 7, we have λN and the approximated solution fN satisfies fN(x)=HxλN.

    Algorithm 1 The Online Tikhonov regularization.
    Require: The number of knots M, mesh size d=1/M, the number of sample N, the observation data {(xi,yi)}Ni=1, weight coefficients W and regularization parameter α.
    Ensure: The approximated solution fNVm.
    1: Initialize A0=0R(M+3)×(M+3) and b0=0R;
    2: Generate the matrix PR(M+3)×(M+3)
            P=(pij)M+1i,j=1,pij=10ϕi(x)ϕj(x)dx;
    3: for i1,2,,N do
    4:  Generate row vector Hxi=(ϕ1(xi),ϕ0(xi),,ϕM+1(xi));
    5:  Upgrade Ai by Aii1iAi1+w2iiHTxiHxi;
    6:  Upgrade bi by bii1ibi1+w2iiHTxiyi;
    7: Solve linear system (αP+AN)λN=bN for λN;
    8: return fN(x)=HxλN.

    Table 1.  Parameter selection.
    Parameter M α wi
    Value N15 σ2N45 1/(k+σ(xi))2

     | Show Table
    DownLoad: CSV

    Remark 1. When a new data is inputed, one just run the algorithm from Line 4 to Line 6. The computational complexity at Line 4, 5 and 6 are O(1) and the computational complexity at Line 7 is O(M). Moreover, the total data storage of Algorithm 1 is O(M).

    Weight coefficients W are introduced to reduce the effect of heteroscedasticity. The selection of weight coefficients W will be thoroughly discussed in Section 3.

    Note that, N continues to increase with the influx of new data, when N becomes so large that MN15, we need to increase M to MN15 and restart the algorithm.

    In this chapter, we analyze the reconstruction error of function f(x). Some assumptions are necessary before proving. Suppose that the matrix P is positive definited. Without losing generality, it is advisable to set the domain of variable x to [0,1]. The weighting coefficient matrix W can be redefined as

    W=diag(w1,w2,,wM)T.

    Make wmin and wmax are the minimum and maximum values of parameters {wj}Nj=1 respectively. Then for the observation data {(xi,yi)}Ni=1, yi can be expressed as

    yj=f(xj)+σ(xj)εj,j=1,2,,N,

    where εj are independent with expectation 0 and variance 1.

    Recording error function as

    eN(x)=fN(x)f(x).

    Next, consider the influence of regularization function on function itself and random noise.

    When the data is divided into deterministic part and random part, the reconstruction result can also be divided into two parts

    fN=fabs+fnoise,

    where the deterministic part fabs satisfies

    fabs=argmingVmJ(g;α,yε),

    here ε is a column vector of the observed noise

    ε=(σ1ε1,σ2ε2,,σNεN)TRN,

    let σ2=maxxjσ2(xj). And the random part fnoise satisfies

    fnoise=argmingVmJ(g;α,ε).

    Then, the error function can be rewritten as

    eN(x)=(fabs(x)f(x))+fnoise(x):=eabs(x)+fnoise(x).

    Two important lemmas are introduced before error estimation.

    Firstly, the indicator function is defined to describe the distribution of observation points.

    Definition 1. Divide the interval [0,1] into M disjoint subspaces Ii, 1iM,

    I1=[0,d],Ii=(idd,id],

    where d=1/M is the grid spacing. There are Ni observation data on each subspace Ii, 1 iM. Define the indicator function ρM(x) as

    ρN(x)={ρN,i=Ni(Md)1,zIi,0,x[0,1]. (3.1)

    Lemma 1. ([10], Lemma 3.3) Assume that the indicator function ρM(x) has an upper bound ρu in the interval [0,1],

    supx[0,1]ρN(x)ρu.

    Then, for the first order continuous differentiable function f(x),x[0,1], the mean square sum of values at all observation points satisfies:

    1NNj=1f2(xj)2ρu(||f||2L(0,1)+d2||f||2L2(0,1)).

    Next, the error estimation of cubic spline interpolation is introduced.

    Lemma 2. ([24], Theorem 1.56) Assume the objective function f(x)H2(a,b), x0=a, xM=b, the grid spacing is d. Let sf be a cubic spline interpolation function of f with natural or fixed boundary conditions, then the interpolation error can be estimated as

    ||sff||L2(a,b)d24||sff||L2(a,b)d24||f||L2(a,b),
    ||sff||L2(a,b)d2||sff||L2(a,b)d2||f||L2(a,b).

    In addition, if sf has natural boundary conditions, then the relation between the second derivatives of f and sf is

    ||sf||2L2(a,b)+||sff||2L2(a,b)=||f||2L2(a,b).

    Now, the error estimate of the deterministic part is analyzed.

    Theorem 2. Assume that ρM is an indicator function on the interval [0,1] and has an upper bound ρu. f(x)H2(0,1). Then, the mean square sum of eabs(x) at the observation points is estimated as follows:

    ||eabs||2L2(0,1)=1NNj=1(fabs(xj)f(xj))298wmaxwminρud4||f||2L2(0,1)+αwmin||f||2L2(0,1).

    Proof. The deterministic part. Note

    EN=1NNj=1(sf(xj)f(xj))2,

    where sf(x) is a cubic spline interpolation function satisfies natural or fixed boundary conditions on the interval [0,1]. By Lemmas 1 and 2, we have

    EN2ρu(||fsf||2L2(0,1)+d2||fsf||2L2(0,1))2ρu(d416||f||2L2(0,1)+d42||f||2L2(0,1))98ρud4||f||2L2(0,1). (3.2)

    And

    1NNj=1(fabs(xj)f(xj))21wmin(1NNj=1wj(fabs(xj)f(xj))2+α||fabs||2L2(0,1)). (3.3)

    For fabs is the smallest element of the weighted regular functional J(;α,yε),

    1NNj=1wj(fabs(xj)f(xj))2+α||fabs||2L2(0,1)1NNj=1wj(sf(xj)f(xj))2+α||sf||2L2(0,1). (3.4)

    Then,

    1NNj=1wj(fabsf(xj))2+α||fabs||2L2(0,1)wmax(EN+αwmax||sf||2L2(0,1)). (3.5)

    From Eqs (3.3)–(3.5), we have

    1NNj=1(fabs(xj)f(xj))2wmaxwminEN+αwmin||sf||2L2(0,1) (3.6)

    On the other hand, by Lemma 2,

    ||sf||2L2(0,1)||f||2L2(0,1).

    Combining (3.2) and (3.6),

    1NNj=1(fabs(xj)f(xj))298wmaxwminρud4||f||2L2(0,1)+αwmin||f||2L2(0,1). (3.7)

    Remark 2. According to formula (3.7), in order to control the error effectively, it is necessary to select the same order of regularization parameter α as d4 to control the interpolation error.

    Next, consider the error estimation of the random part. The reconstruction result of the random part is

    fnoise=argmingVmJ(g;α,ε).

    By the conclusion in chapter 4, we have

    fnoise=Hλε,

    where λε satisfies

    λε=1N(1NHTWTWH+αP)1HTWTWε.

    Theorem 3. Under the above assumptions, the mean square sum of the random part fnoise on the observation points is satisfied with the probability of 1β

    ||fnoise||2L2(0,1)=1NNj=1f2noise(xj)4σ2MwminβN.

    Proof. The proof is divided into three steps.

    Step 1: Rewrite λε as

    λε=(HTH+αNP)1HTε,H:=WH,ε:=Wε=P1HT(HP1HT+αNI)1ε=P1HT(S+αNI)1ε,S:=HP1HT. (3.8)

    where I is an identity matrix of N×N.

    Step 2: Feature decomposition S. For S satisfying:

    S:=HP1HT.

    Obviously, it is semi-positive definite, thus S can be rewritten as

    S=UTUT,

    where U is an orthogonal matrix and T is a diagonal matrix composed of the eigenvalues of S. For

    rank(S)rank(P1)=M+3.

    T can be rewritten as

    T=diag(t1,t2,,tM+3,0,,0).

    where {tj}M+3j=1 is a monotonically decreasing sequence,

    t1t2tM+30.

    Step 3: Obtaining estimates in the form of confidence intervals.

    For the mean square sum of fnoise on observation points satisfies:

    1NNj=1f2noise=1N||Hλ||22.

    And

    E||Hλ||22=E[εT(S+αNI)1HP1HTHP1HT(S+αNI)1ε]=E[tr((WTW)1S2(S+αNI)2εεT)]=tr((WTW)1S2(S+αNI)2E[εεT])=σ2tr((WTW)1S2(S+αNI)2)σ2wminNj=1t2j(tj+αN)2σ2wmin(M+3). (3.9)

    Combined with Markov inequality: for non-negative variable X, for any given a>0, the following formula holds

    P(Xa)E[|X|]a.

    Therefore, the estimation of noise mean square sum error of noise under the probability of 1β satisfies:

    1NNj=1f2noise(xj)E||Hλε||22βNσ2(M+3)wminβN.

    Finally, we estimate the error ||eN||2L2(0,1). Through Theorems 2 and 3, the errors of the deterministic and random parts of the reconstruction function are analyzed. Moreover, we can obtain the mean square error of f(x) under the probability of 1β.

    Theorem 4. Assume that ρM is an indicator function on the interval [0,1] and has an upper bound ρu. Let f(x)H2(0,1), for any β[0,1], such that the error eN under the probability of 1β satisfies:

    ||eN||2L2(0,1)98wmaxwminρud4||f||2L2(0,1)+αwmin||f||2L2(0,1)+σ2(M+3)wminβN. (3.10)

    Proof. By Theorems 2 and 3, the conclusion is easy to get.

    From inequality (3.10), the estimation error is determined by the model parameters. Thus, it is a very important problem to select appropriate model parameters. In this chapter, we are going to determine the regularization parameter α, the number of knots M, and weight coefficients W in Algorithm 1.

    First, we consider the regularization parameter α and the number of knots M. Through Theorem 4, the three formulas

    98wmaxwminρud4||f||2L2(0,1),αwmin||f||2L2(0,1)andσ2(M+3)wminβN

    must be in the same order of magnitude. Otherwise, it will result in oversize error ||eN||2L2(0,1) or over-fitting. Assume the value of ρu and ||f||2L2(0,1) do not effect the order of magnitude. Thus, we have

    wmaxd4ασ2M/N,

    where d=1/M. Through the equation above, the regularization parameter α and the number of knots M satisfy:

    M(Nwmax/σ2)15,αw15max(σ2/N)45. (4.1)

    Next, we consider the optimal choice of weight cofficients W. For weighted linear regression, researchers provide a method to choose weight cofficients. Thus, based on the method of weighted linear regression and the linear form of the reconstruction results, we provide Theorem 3.1 to choose weight coefficients.

    Let eN=fNf be the error function of the proposed regularization algorithm. From error function eN and Eq (2.5), the error comes from two aspects: the interpolation error caused by pre selected interpolation nodes and the random error caused by observation noise. The error eN can not be reduced easily. But we can select suitable weight coefficients W to reduce the variance of the error eN. Denote f(x) as a best approximated solution of f(x) in space Vm as follows:

    f=argmingVm(1NNi=1(g(xi)f(xi))2+α||g(x)||2L2(0,1)), (4.2)

    where the error e=ff just comes from the interpolation error. Similar the process of Theorem 1, f can also be written as f(x)=HxλN and λN is given by

    (αP+1NHTH)λN=1NHTf, (4.3)

    where f:=(f(x1),f(x2),,f(xN))T.

    Theorem 5. Assume regularization parameter α1, M is large enough and variance function σ(x) satisfies σ:[0,1](0,+). If W is given by

    W=diag(1σ(x1),1σ(x2),,1σ(xN)),

    then, λN(W) is the approximated minimum empirical variance unbiased estimation of λN, that is, λNE[λN(W)] and, for any WRN×N,

    Var(λN(W))minWVar(λN(W)).

    Proof. f can be written as

    f=HλN+ε1,ε1RN, (4.4)

    where ε1 is the interpolation error which tends to zero, when M large enough. From the equation above, Eqs (2.5) and (1.1), we have

    E[λN(W)]=(αP+1NHTWTWH)11NHTWTW(HλN+ε1)(αP+1NHTWTWH)11NHTWTWHλN,ε10(αN(HTWTWH)1P+I(M+3)×(M+3))1λN,α0λN. (4.5)

    Denote ε=(σ(x1)ε,σ(x2)ε,,σ(xN)ε), diagonal matrix D:=WTW and diagonal matrix V:=diag(σ2(x1),σ2(x2),,σ2(xN)). Through the definition above and Eqs (2.5), (1.1) and (4.5), for any WRN×N+, if ε10 and α0, we have

    Var[λN(W)]=E(λN(W)λ)(λN(W)λ)TE((αP+1NHTDH)11NHTDε)((αP+1NHTDH)11NHTDε)T(HTDH)1HTDE[εεT]DH(HTDH)1=(HTDH)1HTDVDH(HTDH)1(HTVH)1.

    Moreover, when W=W,

    Var[λN(W)](HTVH)1minWVar(λN(W)).

    From Theorem 2, when α0 and M is large enough, we can choose W to reduce the variance of estimator λN(W). Moreover, the value {1/σ(xi)}Ni=1 in W can not be infinity. When some weight coefficients 1/σ(xi) are too large, the reconstruction results are only affected by the data corresponding to these large weight coefficients. Thus, we suggest to choose the suitable weight coefficients W as follows:

    Wsuit=diag(1k+σ(x1),1k+σ(x2),,1k+σ(xN)), (4.6)

    where k is a small positive constant such as 0.1 and 0.01 and wi satisfies:

    wi=1/(k+σ(xi))2,i=1,2,,N. (4.7)

    Combining Eqs (4.1) and (4.7), we have the parameter selection Table 1.

    In this section, we provide a illustrative example to show the feasibility and advantages of the weighted regularization method. Assume that the heteroscedastic model satisfies:

    y=f(x)+σ(x)ε,

    where f(x):=40(x0.5)2 and σ(x):=3sin(150πx)+3.5. We consider following problem: for a lager positive integer N=300, given observation data {(xi,yi)}Ni=1 and function σ(x), we try to reconstruct function f(x) and its derivative f(x). We select the number of knots M=10, mesh size d=1/M=0.1, regularization parameter α=104 and weight coefficients W satisfying (4.6) with k=0.1.

    Figure 1 compares the reconstruction results-"improve" in this paper with the results-"tradition" in [10]. When function σ(x) changes sharply, our results are closer to exact function than that those in [10].

    Figure 1.  Observation points, noise data and reconstruction results "improve" and "tradition" in this paper and [10], respectively.

    In this section, we will give two applications of our numerical differentiation method in two kinds of inverse problems.

    We consider a nonlinear ill-posed problem–the identification of the coefficient c in the boundary value problem

    f(x)+c(x)f(x)=p(x),f(0)=h0,

    from the solution f(x). We can get c(x) directly from

    c(x)=p(x)f(x)f(x). (6.1)

    The example we take here is that f(x)=10esin(πx), p(x)=0 and h0=10. In our computation, {(xi,yi)}Ni=1 is denoted as the observation data where y satisfies:

    y=f(x)+σ(x)ε,

    where σ(x)=5sin(150πx)+5.1. From the observation data above and the improved method in this paper, we reconstruct f(x) and f(x). The results are present in Figure 2.

    Figure 2.  Observation points, noise data and reconstruction results "improve" and "tradition" in this paper and [10], respectively.

    Through the reconstruction results of f(x) and f(x) and Eq (6.1), we have the estimator of c(x) in Figure 3, where our method is better than the traditional.

    Figure 3.  Reconstruction c-"improve" and "tradition" in this paper and [10], respectively.

    In this subsection, we apply the method in this paper to a simple but interesting problem in Computerized Tomography (CT).

    We assume that the attenuation coefficient of an object with respect to an X-ray at x is p(x). We scan the cross-section by an X-ray beam L of unit intensity. The intensity past the object is eLp(x)dx. We denote function f as follows:

    f(L):=Lp(x)dx. (6.2)

    The main problem in CT is to recover the function p from f. However, in many cases, it will be enough if one can reconstruct the interface of the different mediums, that is, the discontinuous points of p which is related directly with the nondifferentiable points of f. Thus we just need to determine the nondifferentiable points in f(x).

    D is denoted as the triangular cross section of a object. The attenuation coefficient is 0 outside the object and 1 inside the object, that is:

    p(x)={1,xD;0,xD, (6.3)

    see Figure 4.

    Figure 4.  picture of D.

    From Figure 4 and Eq (6.2), the function f(x) can be calculated directly:

    f(x)={10x2,0.2x<0.5,810x,0.5x0.8,0,           elsewhere. (6.4)

    In our computation, {(xi,yi)}Ni=1 is denoted as the observation data where y satisfies Eq (1.1) and σ(x)=5sin(150πx)+5. Let the number of data N=300, the number of knots M=50, mesh size d=1/M=0.02, regularization parameter α104 and the parameter k=0.01 in (4.6).

    In Figure 5, we present observation data, noise data and the reconstructions of f(x) and f(x). Compared with the traditional method in [10], through the improved approach in this paper, we can reconstruct f(x) better and find three nondifferentiable points of f(x): 0.2, 0.5 and 0.8.

    Figure 5.  Observation points, noise data and reconstruction results-"improve" and "tradition" in this paper and [10], respectively.

    In this paper, we propose a weighted online regularization method–Algorithm 1, which updates reconstruction results through new input data without old data again. The error estimation in Section 3 illustrates that the algorithm can reach the best convergence order when the relationship between the amount of observed data N and the dimension M of linear space is MN15. Compared with the classical B-spline method, weight coefficients and a relatively small number of knots for interpolation are used to reduce the effect of heteroscedasticity and computation cost. Moreover, a feasible selection method of the regularization parameter, the number of knots, and weight coefficients are provided in Table 1.

    The biggest advantage of this algorithm is to solve the problem of low data accuracy caused by heteroscedasticity by processing a large amount of data. It can significantly reduce the error of observation data while ensuring the calculation efficiency.

    In the next work, we will consider the higher dimensional heteroscedasticity problem, which requires higher regularity of the function and analyze other radial basis functions, such as Sobolev radial basis functions, to solve high-dimensional problems.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported by Doctoral Research Initiation Fund of Shenzhen Institute of Information Technology with grant (No. SZIIT2023KJ003).

    The author declare that have no competing interest.



    [1] J. Cheng, X. Z. Jia, Y. B. Wang, Numerical differentiation and its applications, Inverse Probl. Sci. En., 15 (2007), 339–357. https://doi.org/10.1080/17415970600839093 doi: 10.1080/17415970600839093
    [2] M. Hanke, O. Scherzer, Inverse problems light: numerical differentiation, Am. Math. Mon., 108 (2001), 512–521. https://doi.org/10.1080/00029890.2001.11919778 doi: 10.1080/00029890.2001.11919778
    [3] D. H. Xu, Y. H. Xu, M. B. Ge, Q. F. Zhang, Models and numerics for differential equations and inverse problems, Beijing: Science Press, 2021.
    [4] P. Craven, G. Wahba, Smoothing noisy data with spline functions, Numer. Math., 31 (1978), 377–403. https://doi.org/10.1007/BF01404567 doi: 10.1007/BF01404567
    [5] D. L. Ragozin, Error bounds for derivative estimates based on spline smoothing of exact or noisy data, J. Approx. Theory, 37 (1983), 335–355. https://doi.org/10.1016/0021-9045(83)90042-4 doi: 10.1016/0021-9045(83)90042-4
    [6] J. P. Kaipio, E. Somersalo, Statistical and computational inverse problems, New York: Springer, 2005. https://doi.org/10.1007/b138659
    [7] L. Wasserman, All of nonparametric statistics, New York: Springer, 2006. https://doi.org/10.1007/0-387-30623-4
    [8] G. Claeskens, T. Krivobokova, J. D. Opsomer, Asymptotic properties of penalized spline estimators, Biometrika, 96 (2009), 529–544. https://doi.org/10.1093/biomet/asp035 doi: 10.1093/biomet/asp035
    [9] P. H. C. Eilers, B. D. Marx, Flexible smoothing with b-splines and penalties, Statist. Sci., 11 (1996), 89–121. https://doi.org/10.1214/ss/1038425655 doi: 10.1214/ss/1038425655
    [10] J. Zhang, J. Cheng, M. Zhong, A tikhonov regularization based algorithm for scattered data with random noise, arXiv: 2105.00747. https://doi.org/10.48550/arXiv.2105.00747
    [11] J. A. Fessler, Penalized weighted least-squares image reconstruction for positron emission tomography, IEEE T. Med. Imaging, 13 (1994), 290–300. https://doi.org/10.1109/42.293921 doi: 10.1109/42.293921
    [12] K. R. Ridgway, J. R. Dunn, J. L. Wilkin, Ocean interpolation by four dimensional weighted least squares-application to the waters around Australasia, J. Atmos. Ocean. Tech., 19 (2002), 1357–1375. https://doi.org/10.1175/1520-0426(2002)019 doi: 10.1175/1520-0426(2002)019
    [13] J. M. Wooldridge, Introductory econometrics: a modern approach, Boston: Cengage Learning, 2012.
    [14] Q. Feng, J. Hannig, J. S. Marron. A note on automatic data transformation, Stat., 5 (2016), 82–87. https://doi.org/10.1002/sta4.104 doi: 10.1002/sta4.104
    [15] J. Kalina, On heteroscedasticity in robust regression, International Days of Statistics and Economics, 41 (2011), 228–237. https://doi.org/10.1111/j.1467-9310.2011.00660.x doi: 10.1111/j.1467-9310.2011.00660.x
    [16] B. Sun, L. Ma, T. Shen, R. Geng, Y. Zhou, Y. Tian, A robust data-driven method for multiseasonality and heteroscedasticity in time series preprocessing, Wirel. Commun. Mob. Com., 2021 (2021), 6692390. https://doi.org/10.1155/2021/6692390 doi: 10.1155/2021/6692390
    [17] M. Marzjarani, A comparison of a general linear model and the ratio estimator, International Journal of Statistics and Probability, 9 (2020), 54–65. https://doi.org/10.5539/ijsp.v9n3p54 doi: 10.5539/ijsp.v9n3p54
    [18] A. Bashan, N. M. Yagmurlu, Y. Ucar, A. Esen, A new perspective for the numerical solution of the modified equal width wave equation, Math. Method. Appl. Sci., 44 (2021), 8925–8939. https://doi.org/10.1002/mma.7322 doi: 10.1002/mma.7322
    [19] A. Bashan, N. M. Yagmurlu, Y. Ucar, A. Esen, Finite difference method combined with differential quadrature method for numerical computation of the modified equal width wave equation, Numer. Meth. Part. D. E., 37 (2021), 690–706. https://doi.org/10.1002/num.22547 doi: 10.1002/num.22547
    [20] A. Bashan, A. Esen, Single soliton and double soliton solutions of the quadratic-nonlinear Korteweg-de Vries equation for small and long-times, Numer. Meth. Part. D. E., 37 (2021), 1561–1582. https://doi.org/10.1002/num.22597 doi: 10.1002/num.22597
    [21] Y. Ucar, N. M. Yagmurlu, A. Bashan, Numerical solutions and stability analysis of modified burgers equation via modified cubic b-spline differential quadrature methods, Sigma J. Eng. Nat. Sci., 37 (2019), 129–142.
    [22] A. Bashan, N. M. Yagmurlu, A mixed method approach to the solitary wave, undular bore and boundary-forced solutions of the regularized long wave equation, Comp. Appl. Math., 41 (2022), 169. https://doi.org/10.1007/s40314-022-01882-7 doi: 10.1007/s40314-022-01882-7
    [23] A. Bashan, N. M. Yagmurlu, Y. Ucar, A. Esen, Numerical approximation to the MEW equation for the single solitary wave and different types of interactions of the solitary waves, J. Differ. Equ. Appl., 28 (2022), 1193–1213. https://doi.org/10.1080/10236198.2022.2132154 doi: 10.1080/10236198.2022.2132154
    [24] G. Micula, S. Micula, Handbook of splines, Dordrecht: Springer, 1999. https://doi.org/10.1007/978-94-011-5338-6
    [25] A. Koppel, G. Warnell, E. Stump, A. Ribeiro, Parsimonious online learning with kernels via sparse projections in function space, J. Mach. Learn. Res., 20 (2019), 83–126.
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1348) PDF downloads(51) Cited by(0)

Figures and Tables

Figures(5)  /  Tables(1)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog