Research article Special Issues

On the randomised stability constant for inverse problems

  • In this paper we introduce the randomised stability constant for abstract inverse problems, as a generalisation of the randomised observability constant, which was studied in the context of observability inequalities for the linear wave equation. We study the main properties of the randomised stability constant and discuss the implications for the practical inversion, which are not straightforward.

    Citation: Giovanni S. Alberti, Yves Capdeboscq, Yannick Privat. On the randomised stability constant for inverse problems[J]. Mathematics in Engineering, 2020, 2(2): 264-286. doi: 10.3934/mine.2020013

    Related Papers:

    [1] Lauri Oksanen, Mikko Salo . Inverse problems in imaging and engineering science. Mathematics in Engineering, 2020, 2(2): 287-289. doi: 10.3934/mine.2020014
    [2] Valentina Candiani, Matteo Santacesaria . Neural networks for classification of strokes in electrical impedance tomography on a 3D head model. Mathematics in Engineering, 2022, 4(4): 1-22. doi: 10.3934/mine.2022029
    [3] Michiaki Onodera . Linear stability analysis of overdetermined problems with non-constant data. Mathematics in Engineering, 2023, 5(3): 1-18. doi: 10.3934/mine.2023048
    [4] Valentina Volpini, Lorenzo Bardella . Asymptotic analysis of compression sensing in ionic polymer metal composites: The role of interphase regions with variable properties. Mathematics in Engineering, 2021, 3(2): 1-31. doi: 10.3934/mine.2021014
    [5] Elena Beretta, M. Cristina Cerutti, Luca Ratti . Lipschitz stable determination of small conductivity inclusions in a semilinear equation from boundary data. Mathematics in Engineering, 2021, 3(1): 1-10. doi: 10.3934/mine.2021003
    [6] William R. B. Lionheart . Histogram tomography. Mathematics in Engineering, 2020, 2(1): 55-74. doi: 10.3934/mine.2020004
    [7] Plamen Stefanov . Conditionally stable unique continuation and applications to thermoacoustic tomography. Mathematics in Engineering, 2019, 1(4): 789-799. doi: 10.3934/mine.2019.4.789
    [8] Yu Chen, Jin Cheng, Giuseppe Floridia, Youichiro Wada, Masahiro Yamamoto . Conditional stability for an inverse source problem and an application to the estimation of air dose rate of radioactive substances by drone data. Mathematics in Engineering, 2020, 2(1): 26-33. doi: 10.3934/mine.2020002
    [9] Rolando Magnanini, Giorgio Poggesi . Interpolating estimates with applications to some quantitative symmetry results. Mathematics in Engineering, 2023, 5(1): 1-21. doi: 10.3934/mine.2023002
    [10] Jinghong Li, Hongyu Liu, Wing-Yan Tsui, Xianchao Wang . An inverse scattering approach for geometric body generation: a machine learning perspective. Mathematics in Engineering, 2019, 1(4): 800-823. doi: 10.3934/mine.2019.4.800
  • In this paper we introduce the randomised stability constant for abstract inverse problems, as a generalisation of the randomised observability constant, which was studied in the context of observability inequalities for the linear wave equation. We study the main properties of the randomised stability constant and discuss the implications for the practical inversion, which are not straightforward.


    Inverse problems are the key to all experimental setups where the physical quantity of interest is not directly observable and must be recovered from indirect measurements. They appear in many different contexts including medical imaging, non-destructive testing, seismic imaging or signal processing. In mathematical terms, an inverse problem consists in the inversion of a linear or nonlinear map

    T:XY,xT(x),

    which models how the quantity of interest x belonging to a space X is related to the measurements y=T(x) in the space Y. The reader is referred to the many books on inverse problems for a comprehensive exposition (see, e.g., [2,3,4,24,27,30,44,47]).

    Inverse problems can be ill-posed: the map T may not be injective (i.e., two different x1 and x2 may correspond to the same measurement T(x1)=T(x2)) or, when injective, T1:ranTYX may not be continuous (i.e., two different and not close x1 and x2 may correspond to almost identical measurements T(x1)T(x2)). Various strategies have been introduced to tackle the issue of inversion in this setting, Tykhonov regularisation being the most famous method [17,29].

    Our purpose is to investigate the role of randomisation in the resolution of inverse problems. By randomisation, we refer to the use of random measurements, or to the case of random unknowns. We do not mean the well-established statistical approaches in inverse problems, as in the Bayesian framework, where probability is used to assess the reliability of the reconstruction.

    Even with this distinction, the wording "randomisation" may refer to many different concepts in the framework of inverse problems. A huge literature is devoted to the issue of randomisation in the measuring process: the unknown x in X is fixed and deterministic, and we choose the measurements randomly according to some suitable distribution. For example, compressed sensing [19] and passive imaging with ambient noise [20] belong to this class. Another very popular idea consists in randomising the unknown: in this scenario, we try to recover most unknowns x in X, according to some distribution. For example, this is typically the situation when deep learning [21] is applied to inverse problems. We briefly review these instances in Appendix A.

    This article deals with a more recent approach, introduced in the framework of observability or control theory: the so-called randomised observability constants [38,39,40]. We argue that, in contrast with the initial motivation for its introduction, the randomised observability constant is not necessarily indicative of the likelihood for randomised unknowns to be observable.

    In the next section, we recall the notion of randomised observability constant and comment on its use in optimal design problems. In section 3, we reformulate the randomised observability constant in an abstract setting as randomised stability constant. In section 4 we show that when the classical (deterministic) stability constant is null, a positive randomised stability constant need not imply, as one could hope, that the inverse problem can be solved for most unknowns. In the course of our study, we make several observations on the properties of the randomised stability constant. Section 5 contains several concluding remarks and discusses possible future directions.

    In this section we briefly review the randomised observability constant introduced in [38,39,40] and its main properties. This was motivated by the original idea of randomisation by Paley and Zygmund, which we now briefly discuss.

    In order to understand the principle of randomisation that will be at the core of this paper, it is useful to recall the historical result by Paley and Zygmund on Fourier series. Let (cn)nZ be an element of 2(C) and f be the Fourier series given by

    f:TθnZcneinθ,

    where T denotes the torus R/(2π). According to the so-called Parseval identity, the function f belongs to L2(T); furthermore, the coefficients cn can be chosen in such a way that f does not belong to any Lq(T) for q>2. Some of the results obtained by Paley and Zygmund (see [35,36,37]) address the regularity of the Fourier series f. They show that if one changes randomly and independently the signs of the Fourier coefficients cn, then the resulting random Fourier series belongs almost surely to any Lq(T) for q>2. More precisely, introducing a sequence (βνn)nZ of independent Bernoulli random variables on a probability space (A,A,P) such that

    P(βνn=±1)=12,

    then, the Fourier series fν given by

    fν:TθnZβνncneinθ

    belongs almost surely to Lq(T) for all q<+.

    In [13], the effect of the randomisation on the initial data of solutions to dispersive equations was investigated. In that case, initial data and solutions of the PDE considered are expanded in a Hilbert basis of L2(Ω) made of eigenfunctions (ϕj)j1 of the Laplace operator. The randomisation procedure consists hence in multiplying all the terms of the series decomposition by well-chosen independent random variables. In particular, regarding for instance the homogeneous wave equation with Dirichlet boundary conditions, one can show that for all initial data (y0,y1)H10(Ω)×L2(Ω), the Bernoulli randomisation keeps the H10×L2 norm constant. It is observed that many other choices of randomisation are possible. For instance a positive effect of the randomisation can be observed by considering independent centred Gaussian random variables with variance 1 (Gaussian randomisation). Indeed, it allows to generate a dense subset of the space of initial data H10(Ω)×L2(Ω) through the mapping

    R(y0,y1):AH10(Ω)×L2(Ω),ν(y0ν,y1ν),

    where (y0ν,y1ν) denotes the pair of randomised initial data, provided that all the coefficients in the series expansion of (y0,y1) are nonzero. Several other properties of these randomisation procedures are also established in [13].

    We now review how randomisation appeared in the framework of inverse problems involving an observability inequality. The property of observability of a system is related to the following issue: How to recover the solutions of a PDE from the knowledge of partial measurements of the solutions. In what follows, we will concentrate on wave models, having in particular photoacoustic/thermoacoustic tomography imaging in mind.

    Let T>0 and ΩRd be a bounded Lipschitz domain with outer unit normal ν. We consider the homogeneous wave equation with Dirichlet boundary conditions

    {tty(t,x)Δy(t,x)=0(t,x)[0,T]×Ω,y(t,x)=0(t,x)[0,T]×Ω. (2.1)

    It is well known that, for all (y0,y1)H10(Ω)×L2(Ω), there exists a unique solution yC0([0,T],H10(Ω))C1((0,T),L2(Ω)) of (2.1) such that y(0,x)=y0(x) and ty(0,x)=y1(x) for almost every xΩ. Let Γ be a measurable subset of Ω, representing the domain occupied by some sensors, which take some measurements over a time horizon [0,T].

    The inverse problem under consideration reads as follows.

    Inverse problem: Reconstruct the initial condition (y0,y1) from the knowledge of the partial boundary measurements

    1Γ(x)yν(t,x),(t,x)[0,T]×Ω.

    To solve this problem, we introduce the so-called observability constant: CT(Γ) is defined as the largest non-negative constant C such that

    C(y(0,),ty(0,))2H10(Ω)×L2(Ω)T0Γ|yν(t,x)|2dHd1dt, (2.2)

    for any solution y of (2.1), where H10(Ω) is equipped with the norm uH10(Ω)=uL2(Ω). Then, the aforementioned inverse problem is well-posed if and only if CT(Γ)>0. In such a case, we will say that observability holds true in time T. Moreover, observability holds true within the class of C domains Ω if (Γ,T) satisfies the Geometric Control Condition (GCC) (see [8]), and this sufficient condition is almost necessary.

    Let us express the observability constant more explicitly. Fix an orthonormal basis (ONB) (ϕj)j1 of L2(Ω) consisting of (real-valued) eigenfunctions of the Dirichlet-Laplacian operator on Ω, associated with the negative eigenvalues (λ2j)j1. Then, any solution y of (2.1) can be expanded as

    y(t,x)=+j=1yj(t)ϕj(x)=12+j=1(ajλjeiλjt+bjλjeiλjt)ϕj(x), (2.3)

    where the coefficients aj and bj account for initial data. More precisely, we consider the ONB of H10(Ω)×L2(Ω) given by {ψ+j,ψj:j1}, where

    ψ+j=12(ϕjλj,iϕj),ψj=12(ϕjλj,iϕj). (2.4)

    Expanding now the initial data with respect to this basis we can write (y0,y1)=+j=1ajψ+j+bjψj, namely,

    y0=+j=1aj+bj2λjϕj,y1=+j=1iajbj2ϕj.

    The corresponding solution to (2.1) is given by (2.3). In addition, Parseval's identity yields (y0,y1)2H10(Ω)×L2(Ω)=+j=1|aj|2+|bj|2.

    Then, the constant CT(Γ) rewrites

    CT(Γ)=inf(aj),(bj)2(C)+j=1(|aj|2+|bj|2)=1T0Γ|yν(t,x)|2dHd1dt,

    where y(t,x) is given by (2.3).

    The constant CT(Γ) is deterministic and takes into account any (aj),(bj)2(C), including the worst possible cases. Interpreting CT(Γ) as a quantitative measure of the well-posed character of the aforementioned inverse problem, one could expect that such worst cases do not occur too often; thus it would appear desirable to consider a notion of observation in average.

    Motivated by the findings of Paley and Zygmund (see §2.1) and its recent use in another context [11,13], making a random selection of all possible initial data for the wave equation (2.1) consists in replacing CT(Γ) with the so-called randomised observability constant defined by

    CT,rand(Γ)=inf(aj),(bj)2(C)+j=1(|aj|2+|bj|2)=1E(T0Γ|yνν(t,x)|2dHd1dt), (2.5)

    where

    yν(t,x)=12+j=1(βν1,jajλjeiλjt+βν2,jbjλjeiλjt)ϕj(x) (2.6)

    and (βν1,j)jN and (βν2,j)jN are two sequences of independent random variables of Bernoulli or Gaussian type, on a probability space (A,A,P) with mean 0 and variance 1. Here, E is the expectation in the probability space, and runs over all possible events ν. In other words, we are randomising the Fourier coefficients {aj,bj}j1 of the initial data (y0,y1) with respect to the basis {ψ±j}j1.

    The randomised observability constant was introduced in [38,39,40,41,42]. It can be expressed in terms of deterministic quantities (see [40,Theorem 2.2]).

    Proposition 1. Let ΓΩ be measurable. We have

    CT,rand(Γ)=T2infjN1λ2jΓ(ϕjν(x))2dHd1. (2.7)

    Proof. In view of Lemma 1 (see below), we have that

    CT,rand(Γ)=inf{1Γνy+j2L2([0,T]×Ω),1Γνyj2L2([0,T]×Ω):j1},

    where y±j is the solution to (2.1) with initial condition ψ±j. Thus

    y±j(t,x)=12λje±iλjtϕj(x),

    and in turn

    1Γνy±j2L2([0,T]×Ω)=12λ2j[0,T]×Γ|e±iλjtνϕj(x)|2dtdx,

    which leads to our thesis.

    We have CT,rand(Γ)CT(Γ) (see Proposition 4 below). It has been noted in [41] that the observability inequality defining CT,rand(Γ) is associated to a deterministic control problem for the wave equation (2.1), where the control has a particular form but acts in the whole domain Ω.

    Regarding CT,rand(Γ), we refer to [42,Section 4] for a discussion on the positivity of this constant. The authors show that if Ω is either a hypercube or a disk, then CT,rand(Γ)>0 for every relatively non-empty open subset Γ of Ω. In particular, in some cases CT(Γ)=0 while CT,rand(Γ)>0. This raised hopes that, even if recovering all unknowns is an unstable process, recovering most unknowns could be feasible, since apparently most unknowns are observable. This heuristic argument, mentioned amongst the motivations for the study of the optimisation of the randomised observability constant, was not investigated further in the aforementioned papers. This matter will be studied in the following sections, dedicated more generally on the possible use of such a constant for investigating the well-posed character of general inverse problems.

    A larger observability constant CT(Γ) in (2.2) leads to a smaller Lipschitz norm bound of the inverse map. Therefore CT(Γ) can be used as the quantity to maximise when searching for optimal sensors' positions. However, this turns out to be somewhat impractical. When implementing a reconstruction process, one has to carry out in general a very large number of measures; likewise, when implementing a control procedure, the control strategy is expected to be efficient in general, but possibly not for all cases. Thus, one aims at exhibiting an observation domain designed to be the best possible in average, that is, over a large number of experiments. Adopting this point of view, it appeared relevant to consider an average over random initial data. In [38,39,40], the best observation is modelled in terms of maximising a randomised observability constant, which coincides with CT,rand(Γ) when dealing with the boundary observation of the wave equation.

    When dealing with internal observation of the wave equation on a closed manifold, it has been shown in [26] that the related observability constant reads as the minimum of two quantities: The infimum of the randomised observability constants over every orthonormal eigenbasis and a purely geometric criterion standing for the minimal average time spent by a geodesic in the observation set.

    However, one should keep in mind that a large randomised constant may not be associated with a reconstruction method (see section 4).

    It is convenient to generalise the construction of the previous section to an abstract setting. In what follows, none of the arguments require the precise form of the forward operator related to the wave equation, as they rely solely on the structure of the randomised constant.

    For the remainder of this paper, we let X and Y be separable infinite-dimensional Hilbert Spaces, and P:XY be an injective bounded linear operator.

    If P1:ranPX is a bounded operator, the inverse problem of finding x from P(x) can be solved in a stable way for all xX, without the need for randomisation. This can be measured quantitatively by the constant

    Cdet=infxX{0}Px2Yx2X>0.

    The smaller Cdet is, the more ill-conditioned the inversion becomes.

    On the other hand, when P1 is unbounded, the situation is different, and the inverse problem is ill-posed [24,30]. In this case, although the kernel of P reduces to {0}, we have

    Cdet=infxX{0}Px2Yx2X=0. (3.1)

    Examples of such maps abound. The example that motivated our study was that introduced in section 2.2, with X=H10(Ω)×L2(Ω), Y=L2([0,T]×Ω) and

    P(y0,y1)=y|[0,T]×Γ,

    where ΓΩ and y is the solution to (2.1) with initial condition (y0,y1): If Γ is not large enough, P is still injective but P1 is unbounded [31]. Any injective compact linear operator satisfies (3.1).

    Let us now introduce the randomised stability constant, which generalises the randomised observability constant to this general setting. We consider the class of random variables introduced in the last section. Choose an ONB e={ek}kN of X and write x=k=1xkekX. We consider random variables of the form

    xν=k=1βνkxkek,

    where βνk are i.i.d. complex-valued random variables on a probability space (A,A,P) with vanishing mean and variance 1, so that E(|βνk|2)=1 for every k. These include the Bernoulli and Gaussian random variables considered in the previous section. It is worth observing that, in the case of Bernoulli random variables, we have |βνk|2=1 for every k, so that xνX=xX.

    Definition 1. The randomised stability constant is defined as

    Crand(e)=infxX{0}E(P(xν)2Yx2X).

    As in the previous section, this constant should represent stability of the inverse problem for most unknowns. By definition, we have Crand(e)Cdet. In general, this is a strict inequality: We will provide examples in section 4. This can be heuristically seen also by the following deterministic expression for the randomised stability constant.

    Lemma 1. There holds

    Crand(e)=infkP(ek)2Y. (3.2)

    Proof. Since (Y,Y) is also a Hilbert space, we find

    P(xν)2Y=k=1βνkxkP(ek),l=1βνlxlP(ek)Y=k=1|βνk|2|xk|2P(ek)2Y+k,lklβνk¯βνlxk¯xlP(ek),P(ek)Y. (3.3)

    Since βνk are i.i.d. with vanishing mean and such that E(|βνk|2)=1, we obtain

    E(P(xν)2Yx2X)=E(P(xν)2Y)k=1|xk|2=k=1|xk|2P(ek)2Yk=1|xk|2infkP(ek)2Y, (3.4)

    which means Crand(e)infkP(ek)2Y. Choosing x=ek, we obtain (3.2).

    The aim of this section is to discuss the impact of the randomised observability constant in inverse problems. We would like to address the following question: How can the positivity of Crand be exploited in the solution of an inverse ill-posed problem? We will not fully address this issue, but rather provide a few positive and negative partial results.

    We remind the reader that the randomisation introduced in the last section when (3.1) holds, was based on the point of view that the ratio in (3.1) is not "usually" small, and that, hopefully, in most cases inversion "should" be possible. It is worthwhile to observe that the subset of the xX such that PxYcxX for some fixed c>0 is never generic in X, since {xX:PxY<cxX} is open and non empty by (3.1). This caveat in mind, we nevertheless wish to test if some evidence can be given to support our optimistic approach that in most cases, inversion should be possible.

    Proposition 2. For every ϵ>0 and xX, there exists c>0 such that

    P(PxνYcxX)>1ϵ. (4.1)

    Proof. Take xX. Define the real-valued map g(c)=P(PxνYcxX). Take a sequence cn0. It is enough to show that

    limn+g(cn)=1.

    We write

    g(cn)=Afn(ν)dP(ν), where fn(ν)={1ifPxνYcnxX,0otherwise.

    Note that fn is monotone increasing and, since kerP={0}, limnfn(ν)=1 for every ν. Thus by the Monotone Convergence Theorem,

    limnAfn(ν)dP(ν)=A(limnfn(ν))dP(ν)=1.

    This thus shows that, for a fixed x, our intuition is vindicated: in the vast majority of cases, the inequality PxνYcxX holds true. This is true independently of Crand(e); we now investigate whether the positivity of Crand(e) may yield a stronger estimate.

    The next step is to estimate the probability that, for a given xX{0}, the square of the ratio P(xν)YxX used in Definition 1 is close to its mean value. The large deviation result we could derive describes the deviation from an upper bound to Crand(e), namely the constant Krand(e) defined by

    Krand(e)=supkP(ek)2Y. (4.2)

    Theorem 1 (large deviation estimate). Assume that Y=L2(Σ,μ), where (Σ,S,μ) is a measure space. Let (βνk)kN be a sequence of independent random variables of Bernoulli type, on a probability space (A,A,P) with mean 0 and variance 1. Let xX{0} and xν=k=1βνkekxk. Then, for every δ>0 we have

    P(PxνYδxνX)exp(21eδKrand(e)).

    The proof of Theorem 1 is postponed to Appendix B. The argument follows the same lines as the one of [11,Theorem 2.1] and the general method introduced in [13].

    Remark 1 (Application to a wave system). Considering the wave equation (2.1) and adopting the framework of section 2.2 leads to choose {ψ±j}j1 defined by (2.4) as the orthonormal basis e. In that case, X=H10(Ω)×L2(Ω), Σ=[0,T]×Ω, dμ=dtdHd1 and Y=L2(Σ). Following the discussion in section 2.2, the map P is given by P(y0,y1)=1Γyν, where y is the unique solution of (2.1). Further, we have

    Crand(e)=T2infjN1λ2jΓ(ϕjν)2dHd1,Krand(e)=T2supjN1λ2jΓ(ϕjν)2dHd1,

    where the first equality is given in Proposition 1, and the second one follows by applying the same argument.

    Note that according to the so-called Rellich identity*, we have 0<Krand(e)T2diam(Ω), under additional mild assumptions on the domain Ω.

    This identity, discovered by Rellich in 1940 [43], reads

    2λ2=Ωx,ν(ϕν)2dHd1

    for every eigenpair (λ,ϕ) of the Laplacian-Dirichlet operator, Ω being a bounded connected domain of Rn either convex or with a C1,1 boundary.

    The estimate given in Theorem 1 is on the "wrong side", since we show that the ratio related to the inversion is much bigger than Krand(e) with low probability. The issue is not the boundedness of P, which is given a priori, but of its inverse. This would correspond to a result of the type

    P(P(xν)2Yx2X<Crand(e)δ)small constant, (4.3)

    namely, a quantification of the estimate given in Proposition 2, uniform in x. If such a bound held, it would show that Crand(e) is a reliable estimator of the behaviour of the ratio P(x)2Yx2X in general. Notice that, in the favourable case when P1 is bounded, there exists δ0[0,Crand(e)) such that

    P(P(xν)2Yx2X<Crand(e)δ)=0

    for all δ[δ0,Crand(e)).

    In this general framework, estimate (4.3) does not hold, see Example 2. Using a concentration inequality, a weaker bound can be derived.

    Proposition 3. Assume that Y=L2(Σ,μ), where (Σ,S,μ) is a measure space. Let (βνk)kN be a sequence of independent random variables of Bernoulli type, on a probability space (A,A,P) with mean 0 and variance 1. Let xX{0} and xν=k=1βνkekxk. Then, for every δ>0 we have

    P(P(xν)2Yx2XE(P(xν)2Yx2X)<δ)<exp(δ24Krand(e)2). (4.4)

    This result is based on an appropriate Hoeffding inequality; its proof is postponed to Appendix B. Note that (4.4) is not a large deviation result: the quantity under consideration is bounded between 0 and 1, and the upper bound obtained is not small. This is unavoidable, see Example 2.

    We collect here several observations that suggest that the positivity of the randomised stability constant may not be helpful for solving the inverse problems, not even for most unknowns.

    We remind the reader why (3.1) renders inversion unstable. Hypothesis (3.1) implies that there exists a sequence (xn)nN such that

    xnX=1 and PxnY<1n for all nN. (4.5)

    Suppose that our measurements are not perfect, and are affected by a low level of noise δ, δYϵ, with ϵ>0. Then, for every n such that nϵ>1, we have

    P(x+xn)P(x)Y<ϵ,

    hence x and x+xn correspond to the same measured data, even if (x+xn)xX=1. This is an unavoidable consequence of the unboundedness of P1, and is true for every xX, even if the randomised stability constant were positive (and possibly large).

    Lemma 1 shows that the ONB used to randomise our input plays a role, as it appears explicitly in the formula (3.2). The following proposition underscores that point. Namely, if we consider all possible randomisations with respect to all ONB of X we recover the deterministic stability constant Cdet.

    Proposition 4. We have

    infeCrand(e)=Cdet,

    where the infimum is taken over all ONB of X. In particular, if P1 is unbounded then infeCrand(e)=0.

    Proof. By definition of Crand(e), we have that Crand(e)Cdet for every ONB e, and so it remains to prove that

    infeCrand(e)Cdet.

    By definition of Cdet, we can find a sequence xnX such that xnX=1 for every n and Pxn2YCdet. For every n, complete xn to an ONB of X, which we call e(n). By Lemma 1 we have Crand(e(n))Pxn2Y, and so

    infeCrand(e)infnCrand(e(n))infnPxn2Ylimn+Pxn2Y=Cdet.

    This result shows that, in general, the randomised stability constant strongly depends on the choice of the basis. There will always be bases for which it becomes arbitrarily small when P1 is unbounded.

    It is also worth observing that for compact operators, which arise frequently in inverse problems, the randomised stability constant is always zero.

    Lemma 2. If P is compact then Crand(e)=0 for every ONB e of X.

    Proof. Since ek tends to zero weakly in X, by the compactness of P we deduce that P(ek) tends to zero strongly in Y. Thus, by Lemma 1 we have

    Crand(e)=infkP(ek)2Ylimk+P(ek)2Y=0,

    as desired.

    Let us now consider some examples. The first example is finite dimensional and the kernel of the operator is not trivial. Note that the definition of Crand(e) and all results above, except Proposition 2 and Lemma 2, are valid also in this case, with suitable changes due to the finiteness of the ONB.

    Example 1. Choose X=R2 and Y=R, and consider the map

    S:R2R(x,y)x+y.

    The associated inverse problem can be phrased: Find the two numbers whose sum is given. This problem is ill-posed and impossible to solve. The deterministic stability constant vanishes

    inf(x1,x2)R2{(0,0)}|x1+x2|2x21+x22=0,

    and S1 does not exist. However, the randomised constant obtained using the canonical basis is positive. Indeed, |P(1,0)|=|P(0,1)|=1, therefore

    Crand({(1,0),(0,1)})=inf{1,1}=1.

    The positivity of this constant does not imply the existence of any useful method to perform the reconstruction of x and y from x+y, even for most (x,y)R2.

    Had we chosen as orthonormal vectors 12(1,1) and 12(1,1), since |P(1,1)|=0, we would have found

    Crand(12(1,1),12(1,1))=0.

    One may wonder whether the features highlighted above are due to the fact that the kernel is not trivial. That is not the case, as the following infinite-dimensional generalisation with trivial kernel shows.

    Example 2. Consider the case when X=Y=2, equipped with the canonical euclidean norm. Let e={ek}+k=0 denote the canonical ONB of 2. Take a sequence (ηn)nN0 such that ηn>0 for all n, and limnηn=0. We consider the operator P defined by

    P(e2n)=e2n+e2n+1,P(e2n+1)=e2n+(1+ηn)e2n+1.

    The operator P may be represented with respect to the canonical basis e by the block-diagonal matrix

    P=[110011+η00000110011+η1].

    In other words, P may be expressed as

    P(x)=(x0+x1,x0+(1+η0)x1,x2+x3,x2+(1+η1)x3,),x2.

    We note that kerT={0} and that its inverse is given by

    P1(y)=((1+η10)y0η10y1,η10(y1y0),(1+η11)y2η11y3,η11(y3y2),),

    which is an unbounded operator since η1n+. Given the block diagonal structure of this map, the inversion consists of solving countably many inverse problems (i.e., linear systems) of the form

    {x2n+x2n+1=y2n,x2n+(1+ηn)x2n+1=y2n+1.

    As soon as ˜n is such that η˜n becomes smaller than the noise level, all the following inverse problems for n˜n are impossible to be solved, since they reduce to the "sum of two numbers" discussed in Example 1.

    Note that

    P(e2n)22=2,P(e2n+1)22=1+(1+ηn)2,

    therefore

    Crand(e)=2.

    If we choose instead the rotated orthonormal basis,

    v2n=12(e2n+e2n+1),v2n+1=12(e2ne2n+1),

    then P(v2n+1)=ηn12e2n+1, and so

    Crand({vk}k)=infkP(vk)22limnP(v2n+1)22=0.

    We now turn to (4.3) and (4.4). For some k0, consider x=e2k. Then P(xν)=βν2k(e2k+e2k+1) and therefore P(xν)Y=P(x)Y: there is no deviation as

    P(xν)2Yx2X=E(P(xν)2Yx2X)=Crand(e)=2.

    Thus, the probabilities in (4.3) and (4.4) are equal to 0, and the inequalities are trivial.

    However, alternatively, consider x=e2k+e2k+1. Then P(xν)2Y=4+(2+η2k)2 if βν1βν2=1 and P(xν)2Y=η22k if βν1βν2=1. Therefore

    P(xν)2Yx2XCrand(e)={(2+η2k)22with probability12,η22k22with probability12.

    As a consequence, (4.3) cannot be true in general for every x. Similarly, we have

    P(xν)2Yx2XE(P(xν)2Yx2X)={2+η2kwith probability12,2η2kwith probability12,

    and the left-hand side of (4.4) can indeed be large for some x.

    It is worth observing that a very similar example is considered in [32] to show that particular complications may arise when using neural networks for solving some inverse problems, even naive and small scale (cfr. §A.3).

    These examples show that considering the observability constant for a particular basis sheds little light on a potential stable inversion of the problem in average, and that considering all possible randomisations leads to the same conclusion as the deterministic case (confirming Proposition 4).

    The pitfalls we encountered when we tried to make use of the randomised stability constant all stem from the linearity of the problems we are considering. The seminal work of Burq and Tzvetkov [12], which showed existence of solutions in super-critical regimes for a semilinear problem did not involve tinkering with associated linear operator (the wave equation); it is the nonlinearity that controlled the critical threshold. In both compressed sensing and passive imaging with random noise sources, nonlinearity plays a key role; further, deep networks are nonlinear maps (cfr. Appendix A).

    The naive intuition we discussed earlier, namely, that extreme situations do not occur often, is more plausible for nonlinear maps where pathological behaviour is local.

    Example 3. As a toy finite-dimensional example, consider the map

    T:RR,T(x)=x(xϵ)(x+ϵ),

    for some small ϵ>0. Then T can be stably inverted outside of a region with size of order ϵ, since there the inverse is continuous. Thus, a random initial data has little chance of falling precisely in the problematic region. Such a case cannot happen with linear maps.

    Example 4. Let A:HH be an unbounded linear operator on a Hilbert space with compact resolvent, so that the spectrum of A is discrete. Define the nonlinear map

    T:H×[0,1]H×[0,1],T(x,λ)=(Ax+λx,λ).

    Note that A+λI is invertible with probability 1 if λ is chosen uniformly in [0,1]. Thus, if H×[0,1] is equipped with a product probability measure whose factor on [0,1] is the uniform probability, then x may be reconstructed from T(x) with probability 1.

    In this paper we focused on the randomised stability constant for linear inverse problems, which we introduced as a generalisation of the randomised observability constant.

    We argue that, despite its intuitive and simple definition, the randomised stability constant has no implications in the practical solution of inverse problems, even for most unknowns. As the examples provided show, this may be due to the linearity of the problem. With nonlinear problems, the situation is expected to be completely different. It could be that the randomised stability constant is meaningful in the context of a nonlinear inversion process, involving for example a hierarchical decomposition [34,46], but we do not know of results in that direction: this is left for future research.

    The authors declare no conflict of interest.

    In this appendix we briefly review three different techniques for solving inverse problems where randomisation plays a crucial role. We do not aim to provide an exhaustive overview, or to report on the most recent advances, or to discuss the many variants that have been studied. The examples we present are used to contrast possible different approaches, and the level of mathematical understanding associated with them.

    Since the seminal works [15,16], compressed sensing (CS) has provided a theoretical and numerical framework to overcome Nyqvist criterion in sampling theory for the reconstruction of sparse signals. In other words, sparse signals in Cn may be reconstructed from k discrete Fourier measurements, with k smaller than n and directly proportional to the sparsity of the signal (up to log factors, see eq. (A.2) below). Let us give a quick overview of the main aspects of CS, in order to show how it fits in the general framework of section 1. For additional details, the reader is referred to the book [19], and to the references therein.

    Given sN={1,2,}, let X be the set of s-sparse signals in Cn, namely

    X={xCn:#suppxs}.

    Let F:CnCn denote the discrete Fourier transform. In fact, any unitary map may be considered, by means of the notion of incoherence [14]. In any case, the Fourier transform is a key example for the applications to Magnetic Resonance Imaging and Computerised Tomography (via the Fourier Slice Theorem). In order to subsample the Fourier measurements, we consider subsets Sa of cardinality k of {1,2,,n} and parametrise them with a{1,2,,(nk)}. Let Y=Ck and Pa:CnCk be the projection selecting the entries corresponding to Sa. We then define the measurement map

    Ta=PaF:XCk.

    In other words, Ta is the partial Fourier transform, since only the frequencies in Sa are measured, and #Sa=kn.

    Given an unknown signal x0X, we need to reconstruct it from the partial knowledge of its Fourier measurements represented by y:=Ta(x0). The sparsity of x0 has to play a crucial role in the reconstruction, since as soon as k<n the map PaF:CnCk necessarily has a non-trivial kernel. It is worth observing that sparsity is a nonlinear condition: If X were a linear subspace of Cn, the problem would be either trivial or impossible, depending on ker(PaF)X. Thus nonlinearity plays a crucial role here.

    The simplest reconstruction algorithm is to look for the sparsest solution to Tax=y, namely to solve the minimisation problem

    minxCnx0subject to Tax=y,

    where x0=#suppx. However, this problem is NP complex, and its direct resolution impractical. Considering the convex relaxation 1 of 0 leads to a well-defined minimisation problem

    minxCnx1subject toTax=y, (A.1)

    whose solution may be easily found by convex optimisation (in fact, by linear programming).

    The theory of CS guarantees exact reconstruction. More precisely, if ˜x is a minimiser of (A.1), then ˜x=x0 with high probability, provided that

    kCslogn, (A.2)

    and that a is chosen uniformly at random in {1,2,,(nk)} (namely, the subset Sa is chosen uniformly at random among all the subsets of cardinality k of {1,2,,n}) [15]. In addition, in the noisy case, with measurements of the form y=Ta(x0)+η where η2ϵ, by relaxing the equality "Tax=y" to the inequality "Taxy2ϵ" in (A.1), one obtains the linear convergence rate x0˜x2Cϵ, namely, the solution is stable.

    In summary, CS allows for the stable reconstruction of all sparse signals from partial Fourier measurements, for most choices of the measured frequencies. The corresponding forward map Ta:XCk is nonlinear, simply because X is not a vector space.

    The material presented in this part is taken from [20], to which the reader is referred for more detailed discussion on this topic.

    A typical multistatic imaging problem is the recovery of some properties of a medium with velocity of propagation c(x)>0 from some measurements at locations xjR3 of the solution u(t,x) of the wave equation

    2tu(t,x)c(x)2Δu(t,x)=f(t)δ(xy),(t,x)R×R3,

    where f(t) is the source pulse located at y. One of the major applications of this setup is geophysical imaging, where one wants to recover properties of the structure of the earth from measurements taken on the surface. Generating sources in this context is expensive and disruptive. Earthquakes are often used as sources, but they are rare and isolated events. Yet, noisy signals, as they may be recorded by seismographers, even if low in amplitude, may be relevant and useful even in absence of important events.

    The key idea is to consider the data generated by random sources (e.g., in seismology, those related to the waves of the sea). The equation becomes

    2tu(t,x)c(x)2Δu(t,x)=n(t,x),(t,x)R×R3,

    where the source term n(t,x) is a zero-mean stationary random process that models the ambient noise sources. We assume that its autocorrelation function is

    E(n(t1,y1)n(t2,y2))=F(t2t1)K(y1)δ(y1y2),

    where F is the time correlation function (normalised so that F(0)=1) and K characterises the spatial support of the sources. The presence of δ(y1y2) makes the process n delta-correlated in space.

    The reconstruction is based on the calculation of the empirical cross correlation of the signals recorded at x1 and x2 up to time T:

    CT(τ,x1,x2)=1TT0u(t,x1)u(t+τ,x2)dt.

    Its expectation is the statistical cross correlation

    E(CT(τ,x1,x2))=C(1)(τ,x1,x2),

    which is given by

    C(1)(τ,x1,x2)=12πR×R3ˆF(ω)K(y)¯ˆG(ω,x1,y)ˆG(ω,x2,y)eiωτdtdy, (A.3)

    where ˆ denotes the Fourier transform in time and G(t,x,y) is the time-dependent Green's function. Moreover, CT is a self-averaging quantity, namely

    limT+CT(τ,x1,x2)=C(1)(τ,x1,x2)

    in probability.

    The role of randomised sources is now clear: from the measured empirical cross correlation CT with large values of T it is possible to estimate, with high probability, the statistical cross correlation C(1). Using (A.3), from C(1)(τ,x1,x2) it is possible to recover (some properties of) the Green function G, which yield useful information about the medium, such as travel times.

    Convolutional Neural Networks have recently been used for a variety of imaging and parameter reconstruction problems [22], including Electrical Impedance Tomography (EIT) [23,33,48], optical tomography [18], inverse problems with internal data [9], diffusion problems in imaging [7], computerised tomography [10,28], photoacoustic tomography [5,25] and magnetic resonance imaging [49,50]. In the following brief discussion, we decided to focus on inverse problems for partial differential equations (PDE), and in particular on EIT, but similar considerations are valid for most methods cited above.

    Significant improvement has been observed in EIT with deep learning compared to previous imaging approaches. Let ΩRd, d2, be a bounded Lipschitz domain with outer unit normal ν. The data in EIT is (a part of) of the Dirichlet-to-Neumann map

    Λσ:H12(Ω)/RH12(Ω)/Rvσuν|Ω

    where u(x) denotes the unique solution of the elliptic problem

    {div(σ(x)u(x))=0xΩ,u(x)=v(x)xΩ,

    and σ(x)>0 is the unknown conductivity. The experimental data is usually part of the inverse map, namely the Neumann-to-Dirichlet map Λ1σ. In two dimensions, provided that the electrodes are equally separated on the unit disk, the data may be modelled by

    TNΛ1σTN,

    where TN is the L2 projection on span{θcos(nθ):1nN}, where N is the number of electrodes: It is the partial Fourier transform limited to the first N coefficients.

    Direct neural network inversion approaches suffer from drawbacks alike direct non-regularised inversion attempts: the output is very sensitive to measurement errors and small variations. Successful approaches to Deep Learning EIT [23,33,48], and to other parameter identification problems in PDE, often involve two steps.

    The first step consists in the derivation of an approximate conductivity σ by a stable, albeit blurry, regularised inversion method. For instance, in [23] the "D-bar" equation is used, while in [33] a one-step Gauss-Newton method is used. In both cases, the output of this step is a representation of the conductivity coefficient, which depends on the inversion method used. This first step is deterministic and its analysis is well understood. The forward problem, relating the conductivity to the Dirichlet-to-Neumann map, is nonlinear, independently of the inversion algorithm used. Indeed, the map Λσ is a linear operator, but σΛσ is nonlinear.

    The second, post-processing, step uses a neural network to "deblur" the image, and in fact restores details that were not identifiable after the first step.

    The second step is not unlike other successful usage of deep-learning approaches for image classification; in general they are known to be successful only with very high probability (and in turn for random unknowns). More precisely, since the findings of [45], deep networks are known to be vulnerable to so-called "adversarial perturbations" (see the review article [1] and the references therein). Given an image x that is correctly classified by the network with high confidence, an adversarial perturbation is a small perturbation p such that the images x and y=x+p are visually indistinguishable but the perturbed image y is misclassified by the network, possibly with high confidence too. State-of-the-art classification networks are successful for the vast majority of natural images, but are very often vulnerable to such perturbations.

    These instabilities are not specific to image classification problems; they appear in the same way in image reconstruction [6]. In this case, given an image that is well-reconstructed by the network, it is possible to create another image that is visually indistinguishable from the original one, but that is not well-reconstructed by the network.

    A full mathematical understanding of deep networks is still lacking, and the reasons of this phenomenon are not fully known. However, the large Lipschitz constant of the network certainly plays a role, since it is a sign of potential instability: in order for the network to be effective, the weights of its linear steps need to be chosen large enough, and the composition of several layers yields an exponentially large constant.

    The proofs of Theorem 1 and Proposition 3 rest upon a classical large deviation estimate, the so-called Hoeffding inequality, see e.g. [11,Prop. 2.2], whose proof is recalled for the convenience of the reader.

    Proposition 5. Let (ανn)n1 be a sequence of sequence of independent random variables of Bernoulli type, on a probability space (A,A,P) with mean 0 and variance 1. Then, for any t>0 and any sequence (un)n12(C), we have

    P(+n=1ανnvn<t)=P(+n=1ανnvn>t)exp(12t2+n=1|vn|2).

    Proof. There holds E(exp(ανnvn))=E(k=0(ανnvn)kk!)=k=01k!E((ανnvn)k). All odd powers of k vanish as ανn has zero mean and is symmetrical. Therefore for any λ>0

    E(exp(λανnvn))=k=01(2k)!E((λανnvn)2k)k=0λ2kv2kn2k(k!)=exp(λ22v2n).

    Applying Chernoff's inequality, we obtain for any λ>0

    P(+n=1ανnvn>t)E[exp(λ(+n=1ανnvn))]exp(λt)exp(λ212+n=1v2nλt).

    Choosing λ=tv2n this yields

    P(+n=1ανnvn<t)=P(+n=1ανnvn>t)exp(t22v2n),

    as desired.

    We are now ready to prove Theorem 1.

    Proof of Theorem 1. Fix r2 and set Yν=P(xν)xνX. Markov's inequality yields

    P(YνYδ)=P(YνrYδr)1δrE(YνrY). (B.1)

    Let us denote by Lrν the standard Lebesgue space with respect to the probability measure dP. Recall that Y=L2(Σ,μ). To provide an estimate of the right-hand side, notice that

    E(YνrY)=AYνrYdP(ν)=A(Σ|Yν(s)|2dμ(s))r/2dP(ν)=Σ|Yν(s)|2dμ(s)r/2Lr/2ν(Σ|Yν(s)|2Lr/2νdμ(s))r/2=(ΣYν(s)2Lrνdμ(s))r/2=sYν(s)LrνrY (B.2)

    by using Jensen's inequality.

    Furthermore, for a.e. sΣ, we have

    Yν(s)rLrν=A|Yν(s)|rdP(ν)=+0rur1P(|Yν(s)|>u)du

    Here, we use that if X denotes a non-negative random variable and φ:R+R+, then

    E(φ(X))=+0φ(u)P(X>u)du.

    and by using Proposition 5 and the fact that Yν(s) reads

    Yν(s)=k=1βνkxk(Pek)(s)k=1|xk|2,

    one gets

    Yν(s)rLrν2+0rur1exp(12k=1|xk|2u2+k=1|(Pek)(s)|2|xk|2)du.

    As a consequence, by using the change of variable

    v=k=1|xk|2u+k=1|(Pek)(s)|2|xk|2

    we get

    Yν(s)rLrνC(r)(+k=1|(Pek)(s)|2|xk|2k=1|xk|2)r/2

    with

    C(r)=2+0rvr1e12v2dv.

    An elementary computation yields C(r)<rr. Therefore,

    Yν(s)Lrν<(r2+k=1|(Pek)(s)|2|xk|2k=1|xk|2)1/2.

    According to (B.2), we infer that

    E(YνrY)<(r2+k=1|(Pek)()|2|xk|2k=1|xk|2)1/2rY=r2+k=1|(Pek)()|2|xk|2k=1|xk|2r/2L1(Σ).

    From (4.2) and estimate (B.1), we get

    P(YνYδ)<1δr(r2+k=1Pek2Y|xk|2k=1|xk|2)r/2(Krand(e)δ2r2)r/2,

    using the triangle inequality. Minimising the upper bound with respect to r, that is, choosing

    r2=e2δ2Krand(e)

    in the inequality above, one finally obtains

    P(YνYδ)exp(e1δKrand(e)). (B.3)

    Note that we have assumed r2, thus implicitly posited that e2δ24Krand(e); we multiply the bound by exp(2) to cover the other case.

    We conclude by proving Proposition 3.

    Proof of Proposition 3. As in (3.3), we have

    P(xν)2Y=k=1x2kP(ek)2Y+2k,lk<lβνkβνlxkxlP(ek),P(el)Y. (B.4)

    Observing that the family ανk,l={βνkβνl}k,lk<l is made of independent Bernoulli variables, equal almost surely to 1 or 1, with 1/2 as probability of success, we apply Proposition 5 with

    vk,l=2xkxXxlxXP(ek),P(el)Y

    and obtain for all δ>0

    P(P(xν)2Yx2XE(P(xν)2Yx2X)<δ)<exp(δ22K),

    with

    K=k,lk<l(2xkxlx2XP(ek),P(el)Y)22k,l(|xk||xl|x2XPekYPelY)2=2(k|xk|2x2XPek2Y)22Krand(e)2.


    [1] Akhtar N, Mian A (2018) Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access 6: 14410-14430. doi: 10.1109/ACCESS.2018.2807385
    [2] Alberti GS, Capdeboscq Y (2018) Lectures on Elliptic Methods for Hybrid Inverse Problems, Paris: Société Mathématique de France.
    [3] Ammari H (2008) An Introduction to Mathematics of Emerging Biomedical Imaging, Berlin: Springer.
    [4] Ammari H, Garnier J, Kang H, et al. (2017) Multi-Wave Medical Imaging: Mathematical Modelling & Imaging Reconstruction, World Scientific.
    [5] Antholzer S, Haltmeier M, Schwab J (2018) Deep learning for photoacoustic tomography from sparse data. Inverse Probl Sci Eng 27: 987-1005.
    [6] Antun V, Renna F, Poon C, et al. (2019) On instabilities of deep learning in image reconstruction-Does AI come at a cost?. arXiv preprint arXiv:1902.05300.
    [7] Arridge S, Hauptmann A (2019) Networks for nonlinear diffusion problems in imaging. J Math Imaging Vis DOI: https://xs.scihub.ltd/https://doi.org/10.1007/s10851-019-00901-3.
    [8] Bardos C, Lebeau G, Rauch J (1992) Sharp sufficient conditions for the observation, control, and stabilization of waves from the boundary. SIAM J Control Optim 30: 1024-1065. doi: 10.1137/0330055
    [9] Berg J, Nyström K (2017) Neural network augmented inverse problems for PDEs. arXiv preprint arXiv:1712.09685.
    [10] Bubba TA, Kutyniok G, Lassas M, et al. (2019) Learning the invisible: A hybrid deep learning-shearlet framework for limited angle computed tomography. Inverse Probl 35: 064002. doi: 10.1088/1361-6420/ab10ca
    [11] Burq N (2010) Random data Cauchy theory for dispersive partial differential equations. In: Proceedings of the International Congress of Mathematicians, New Delhi: Hindustan Book Agency, 3: 1862-1883.
    [12] Burq N, Tzvetkov N (2008) Random data Cauchy theory for supercritical wave equations II: A global existence result. Invent Math 173: 477-496. doi: 10.1007/s00222-008-0123-0
    [13] Burq N, Tzvetkov N (2014) Probabilistic well-posedness for the cubic wave equation. J Eur Math Soc 16: 1-30. doi: 10.4171/JEMS/426
    [14] Candès EJ, Romberg J (2007) Sparsity and incoherence in compressive sampling. Inverse Probl 23: 969-985. doi: 10.1088/0266-5611/23/3/008
    [15] Candès EJ, Romberg J, Tao T (2006) Robust uncertainty principles: Exact signal reconstruction from highly incomplete frequency information. IEEE Trans Inform Theory 52: 489-509. doi: 10.1109/TIT.2005.862083
    [16] Donoho DL (2006) Compressed sensing. IEEE Trans Inform Theory 52: 1289-1306. doi: 10.1109/TIT.2006.871582
    [17] Engl HW, Hanke M, Neubauer A (1996) Regularization of Inverse Problems, Dordrecht: Kluwer Academic Publishers Group.
    [18] Feng J, Sun Q, Li Z, et al. (2018) Back-propagation neural network-based reconstruction algorithm for diffuse optical tomography. J Biomed Opt 24: 051407.
    [19] Foucart S, Rauhut H (2013) A Mathematical Introduction to Compressive Sensing, New York: Birkhäuser/Springer.
    [20] Garnier J, Papanicolaou G (2016) Passive Imaging with Ambient Noise, Cambridge: Cambridge University Press.
    [21] Goodfellow I, Bengio Y, Courville A (2016) Deep Learning, MIT Press.
    [22] Haltmeier M, Antholzer S, Schwab J (2019) Deep Learning for Image Reconstruction, World Scientific.
    [23] Hamilton SJ, Hauptmann A (2018) Deep d-bar: Real-time electrical impedance tomography imaging with deep neural networks. IEEE Trans Med Imaging 37: 2367-2377. doi: 10.1109/TMI.2018.2828303
    [24] Hasanoğlu AH, Romanov VG (2017) Introduction to Inverse Problems for Differential Equations, Cham: Springer.
    [25] Hauptmann A, Lucka F, Betcke M, et al. (2018) Model-based learning for accelerated, limitedview 3-d photoacoustic tomography. IEEE Trans Med Imaging 37: 1382-1393. doi: 10.1109/TMI.2018.2820382
    [26] Humbert E, Privat Y, Trélat E (2019) Observability properties of the homogeneous wave equation on a closed manifold. Commun Part Diff Eq, 44: 749-772. doi: 10.1080/03605302.2019.1581799
    [27] Isakov V (2006) Inverse Problems for Partial Differential Equations, New York: Springer.
    [28] Jin KH, McCann MT, Froustey E, et al. (2017) Deep convolutional neural network for inverse problems in imaging. IEEE Trans Image Process 26: 4509-4522. doi: 10.1109/TIP.2017.2713099
    [29] Kaltenbacher B, Neubauer A, Scherzer O (2008) Iterative Regularization Methods for Nonlinear ill-posed Problems, Berlin: Walter de Gruyter.
    [30] Kirsch A (2011) An Introduction to the Mathematical Theory of Inverse Problems, New York: Springer.
    [31] Laurent C, Léautaud M (2019) Quantitative unique continuation for operators with partially analytic coefficients. Application to approximate control for waves. J Eur Math Soc 21: 957-1069.
    [32] Maass P (2019) Deep learning for trivial inverse problems. In: Compressed Sensing and Its Applications, Cham: Springer International Publishing, 195-209.
    [33] Martin S, Choi CTM (2017) A post-processing method for three-dimensional electrical impedance tomography. Sci Rep 7: 7212. doi: 10.1038/s41598-017-07727-2
    [34] Modin K, Nachman A, Rondi L (2019) A multiscale theory for image registration and nonlinear inverse problems. Adv Math 346: 1009-1066. doi: 10.1016/j.aim.2019.02.014
    [35] Paley R (1930) On some series of functions. Math Proc Cambridge 26: 458-474. doi: 10.1017/S0305004100016212
    [36] Paley R, Zygmund A (1930) On some series of functions, (1). Math Proc Cambridge, 26: 337-357. doi: 10.1017/S0305004100016078
    [37] Paley R, Zygmund A (1932) On some series of functions, (3). Math Proc Cambridge, 28: 190-205. doi: 10.1017/S0305004100010860
    [38] Privat Y, Trélat E, Zuazua E (2013) Optimal observation of the one-dimensional wave equation. J Fourier Anal Appl 19: 514-544. doi: 10.1007/s00041-013-9267-4
    [39] Privat Y, Trélat E, Zuazua E (2015) Optimal shape and location of sensors for parabolic equations with random initial data. Arch Ration Mech An 216: 921-981. doi: 10.1007/s00205-014-0823-0
    [40] Privat Y, Trélat E, Zuazua E (2016) Optimal observability of the multi-dimensional wave and Schrödinger equations in quantum ergodic domains. J Eur Math Soc 18: 1043-1111. doi: 10.4171/JEMS/608
    [41] Privat Y, Trélat E, Zuazua E (2016) Randomised observation, control and stabilization of waves [Based on the plenary lecture presented at the 86th Annual GAMM Conference, Lecce, Italy, March 24, 2015]. ZAMM Z Angew Math Mech 96: 538-549. doi: 10.1002/zamm.201500181
    [42] Privat Y, Trélat E, Zuazua E (2019) Spectral shape optimization for the neumann traces of the dirichlet-laplacian eigenfunctions. Calc Var Partial Dif 58: 64. doi: 10.1007/s00526-019-1522-3
    [43] Rellich F (1940) Darstellung der eigenwerte von δu+λu= 0 durch ein randintegral. Math Z 46: 635-636. doi: 10.1007/BF01181459
    [44] Otmar Scherzer (2015) Handbook of Mathematical Methods in Imaging. Vol. 1, 2, 3. New York: Springer.
    [45] Szegedy C, Zaremba W, Sutskever I, et al. (2013) Intriguing properties of neural networks. arXiv preprint arXiv:1312.6199.
    [46] Tadmor E, Nezzar S, Vese L (2004) A multiscale image representation using hierarchical (BV, L2) decompositions. Multiscale Model Simul 2: 554-579. doi: 10.1137/030600448
    [47] Tarantola A (2005) Inverse Problem Theory and Methods for Model Parameter Estimation. Philadelphia: Society for Industrial and Applied Mathematics.
    [48] Wei Z, Liu D, Chen X (2019) Dominant-current deep learning scheme for electrical impedance tomography. IEEE Trans Biomedical Eng 66: 2546-2555. doi: 10.1109/TBME.2019.2891676
    [49] Yang G, Yu S, Dong H, et al. (2018) Dagan: Deep de-aliasing generative adversarial networks for fast compressed sensing mri reconstruction. IEEE Trans Med Imaging 37: 1310-1321. doi: 10.1109/TMI.2017.2785879
    [50] Zhu B, Liu JZ, Cauley SF, et al. (2018) Image reconstruction by domain-transform manifold learning. Nature 555: 487-492. doi: 10.1038/nature25988
  • This article has been cited by:

    1. Giovanni S. Alberti, Paolo Campodonico, Matteo Santacesaria, Compressed Sensing Photoacoustic Tomography Reduces to Compressed Sensing for Undersampled Fourier Measurements, 2021, 14, 1936-4954, 1039, 10.1137/20M1375152
  • Reader Comments
  • © 2020 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(3808) PDF downloads(358) Cited by(1)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog