
Cancer is a manifestation of disorders caused by the changes in the body's cells that go far beyond healthy development as well as stabilization. Breast cancer is a common disease. According to the stats given by the World Health Organization (WHO), 7.8 million women are diagnosed with breast cancer. Breast cancer is the name of the malignant tumor which is normally developed by the cells in the breast. Machine learning (ML) approaches, on the other hand, provide a variety of probabilistic and statistical ways for intelligent systems to learn from prior experiences to recognize patterns in a dataset that can be used, in the future, for decision making. This endeavor aims to build a deep learning-based model for the prediction of breast cancer with a better accuracy. A novel deep extreme gradient descent optimization (DEGDO) has been developed for the breast cancer detection. The proposed model consists of two stages of training and validation. The training phase, in turn, consists of three major layers data acquisition layer, preprocessing layer, and application layer. The data acquisition layer takes the data and passes it to preprocessing layer. In the preprocessing layer, noise and missing values are converted to the normalized which is then fed to the application layer. In application layer, the model is trained with a deep extreme gradient descent optimization technique. The trained model is stored on the server. In the validation phase, it is imported to process the actual data to diagnose. This study has used Wisconsin Breast Cancer Diagnostic dataset to train and test the model. The results obtained by the proposed model outperform many other approaches by attaining 98.73 % accuracy, 99.60% specificity, 99.43% sensitivity, and 99.48% precision.
Citation: Muhammad Bilal Shoaib Khan, Atta-ur-Rahman, Muhammad Saqib Nawaz, Rashad Ahmed, Muhammad Adnan Khan, Amir Mosavi. Intelligent breast cancer diagnostic system empowered by deep extreme gradient descent optimization[J]. Mathematical Biosciences and Engineering, 2022, 19(8): 7978-8002. doi: 10.3934/mbe.2022373
[1] | Karl Hajjar, Lénaïc Chizat . On the symmetries in the dynamics of wide two-layer neural networks. Electronic Research Archive, 2023, 31(4): 2175-2212. doi: 10.3934/era.2023112 |
[2] | Eray Önler . Feature fusion based artificial neural network model for disease detection of bean leaves. Electronic Research Archive, 2023, 31(5): 2409-2427. doi: 10.3934/era.2023122 |
[3] | Dong-hyeon Kim, Se-woon Choe, Sung-Uk Zhang . Recognition of adherent polychaetes on oysters and scallops using Microsoft Azure Custom Vision. Electronic Research Archive, 2023, 31(3): 1691-1709. doi: 10.3934/era.2023088 |
[4] | Ziqing Yang, Ruiping Niu, Miaomiao Chen, Hongen Jia, Shengli Li . Adaptive fractional physical information neural network based on PQI scheme for solving time-fractional partial differential equations. Electronic Research Archive, 2024, 32(4): 2699-2727. doi: 10.3934/era.2024122 |
[5] | Ilyоs Abdullaev, Natalia Prodanova, Mohammed Altaf Ahmed, E. Laxmi Lydia, Bhanu Shrestha, Gyanendra Prasad Joshi, Woong Cho . Leveraging metaheuristics with artificial intelligence for customer churn prediction in telecom industries. Electronic Research Archive, 2023, 31(8): 4443-4458. doi: 10.3934/era.2023227 |
[6] | Kai Huang, Chang Jiang, Pei Li, Ali Shan, Jian Wan, Wenhu Qin . A systematic framework for urban smart transportation towards traffic management and parking. Electronic Research Archive, 2022, 30(11): 4191-4208. doi: 10.3934/era.2022212 |
[7] | Ruyu Yan, Jiafei Jin, Kun Han . Reinforcement learning for deep portfolio optimization. Electronic Research Archive, 2024, 32(9): 5176-5200. doi: 10.3934/era.2024239 |
[8] | Mohd. Rehan Ghazi, N. S. Raghava . Securing cloud-enabled smart cities by detecting intrusion using spark-based stacking ensemble of machine learning algorithms. Electronic Research Archive, 2024, 32(2): 1268-1307. doi: 10.3934/era.2024060 |
[9] | Manal Abdullah Alohali, Mashael Maashi, Raji Faqih, Hany Mahgoub, Abdullah Mohamed, Mohammed Assiri, Suhanda Drar . Spotted hyena optimizer with deep learning enabled vehicle counting and classification model for intelligent transportation systems. Electronic Research Archive, 2023, 31(7): 3704-3721. doi: 10.3934/era.2023188 |
[10] | Jiaxin Zhang, Hoang Tran, Guannan Zhang . Accelerating reinforcement learning with a Directional-Gaussian-Smoothing evolution strategy. Electronic Research Archive, 2021, 29(6): 4119-4135. doi: 10.3934/era.2021075 |
Cancer is a manifestation of disorders caused by the changes in the body's cells that go far beyond healthy development as well as stabilization. Breast cancer is a common disease. According to the stats given by the World Health Organization (WHO), 7.8 million women are diagnosed with breast cancer. Breast cancer is the name of the malignant tumor which is normally developed by the cells in the breast. Machine learning (ML) approaches, on the other hand, provide a variety of probabilistic and statistical ways for intelligent systems to learn from prior experiences to recognize patterns in a dataset that can be used, in the future, for decision making. This endeavor aims to build a deep learning-based model for the prediction of breast cancer with a better accuracy. A novel deep extreme gradient descent optimization (DEGDO) has been developed for the breast cancer detection. The proposed model consists of two stages of training and validation. The training phase, in turn, consists of three major layers data acquisition layer, preprocessing layer, and application layer. The data acquisition layer takes the data and passes it to preprocessing layer. In the preprocessing layer, noise and missing values are converted to the normalized which is then fed to the application layer. In application layer, the model is trained with a deep extreme gradient descent optimization technique. The trained model is stored on the server. In the validation phase, it is imported to process the actual data to diagnose. This study has used Wisconsin Breast Cancer Diagnostic dataset to train and test the model. The results obtained by the proposed model outperform many other approaches by attaining 98.73 % accuracy, 99.60% specificity, 99.43% sensitivity, and 99.48% precision.
The training of artificial neural networks (ANNs) with rectified linear unit (ReLU) activation via gradient descent (GD) type optimization schemes is nowadays a common industrially relevant procedure which appears, for instance, in the context of natural language processing, face recognition, fraud detection, and game intelligence. Although there exist a large number of numerical simulations in which GD type optimization schemes are effectively used to train ANNs with ReLU activation, till this day in the scientific literature there is in general no mathematical convergence analysis which explains the success of GD type optimization schemes in the training of such ANNs.
GD type optimization schemes can be regarded as temporal discretization methods for the gradient flow (GF) differential equations associated to the considered optimization problem and, in view of this, it seems to be a natural direction of research to first aim to develop a mathematical convergence theory for time-continuous GF differential equations and, thereafter, to aim to extend such a time-continuous convergence theory to implementable time-discrete GD type optimization methods.
Although there is in general no theoretical analysis which explains the success of GD type optimization schemes in the training of ANNs in the literature, there are several auspicious analysis approaches as well as several promising partial error analyses regarding the training of ANNs via GD type optimization schemes and GFs, respectively, in the literature. For convex objective functions, the convergence of GF and GD processes to the global minimum in different settings has been proved, e.g., in [1,2,3,4,5]. For general non-convex objective functions, even under smoothness assumptions GF and GD processes can show wild oscillations and admit infinitely many limit points, cf., e.g., [6]. A standard condition which excludes this undesirable behavior is the Kurdyka-Łojasiewicz inequality and we point to [7,8,9,10,11,12,13,14,15,16] for convergence results for GF and GD processes under Łojasiewicz type assumptions. It is in fact one of the main contributions of this work to demonstrate that the objective functions occurring in the training of ANNs with ReLU activation satisfy an appropriate Kurdyka-Łojasiewicz inequality, provided that both the target function and the density of the probability distribution of the input data are piecewise polynomial. For further abstract convergence results for GF and GD processes in the non-convex setting we refer, e.g., to [17,18,19,20,21] and the references mentioned therein.
In the overparametrized regime, where the number of training parameters is much larger than the number of training data points, GF and GD processes can be shown to converge to global minima in the training of ANNs with high probability, cf., e.g., [22,23,24,25,26,27,28]. As the number of neurons increases to infinity, the corresponding GF processes converge (with appropriate rescaling) to a measure-valued process which is known in the scientific literature as Wasserstein GF. For results on the convergence behavior of Wasserstein GFs in the training of ANNs we point, e.g., to [29,30,31], [32, Section 5.1], and the references mentioned therein.
A different approach is to consider only very special target functions and we refer, in particular, to [33,34] for a convergence analysis for GF and GD processes in the case of constant target functions and to [35] for a convergence analysis for GF and GD processes in the training of ANNs with piecewise linear target functions. In the case of linear target functions, a complete characterization of the non-global local minima and the saddle points of the risk function has been obtained in [36].
In this article we establish two basic results for GF differential equations in the training of fully-connected feedforward ANNs with one hidden layer and ReLU activation. Specifically, in the first main result of this article, see Theorem 1.1 below, we establish in the training of such ANNs under the assumption that the probability distribution of the input data of the considered supervised learning problem is absolutely continuous with a bounded density function that every GF differential equation possesses for every initial value a solution which is also unique among a suitable class of solutions (see (1.6) in Theorem 1.1 for details). In the second main result of this article, see Theorem 1.2 below, we prove in the training of such ANNs under the assumption that the target function and the density function are piecewise polynomial (see (1.8) below for details) that every non-divergent GF trajectory converges with an appropriate speed of convergence (see (1.11) below) to a critical point.
In Theorems 1.1 and 1.2 we consider ANNs with d∈N={1,2,3,…} neurons on the input layer (d-dimensional input), H∈N neurons on the hidden layer (H-dimensional hidden layer), and 1 neuron on the output layer (1-dimensional output). There are thus Hd scalar real weight parameters and H scalar real bias parameters to describe the affine linear transformation between d-dimensional input layer and the H-dimensional hidden layer and there are thus H scalar real weight parameters and 1 scalar real bias parameter to describe the affine linear transformation between the H-dimensional hidden layer and the 1-dimensional output layer. Altogether there are thus
d=Hd+H+H+1=Hd+2H+1 | (1.1) |
real numbers to describe the ANNs in Theorems 1.1 and 1.2. We also refer to Figure 1 for a graphical illustration of the architecture of an example ANN with d=4 neurons on the input layer and H=5 neurons on the hidden layer.
The real numbers a∈R, b∈(a,∞) in Theorems 1.1 and 1.2 are used to specify the set [a,b]d in which the input data of the considered supervised learning problem takes values in and the function f:[a,b]d→R in Theorem 1.1 specifies the target function of the considered supervised learning problem.
In Theorem 1.1 we assume that the target function is an element of the set C([a,b]d,R) of continuous functions from [a,b]d to R but beside this continuity hypothesis we do not impose further regularity assumptions on the target function.
The function p:[a,b]d→[0,∞) in Theorems 1.1 and 1.2 is an unnormalized density function of the probability distribution of the input data of the considered supervised learning problem and in Theorem 1.1 we impose that this unnormalized density function is bounded and measurable.
In Theorems 1.1 and 1.2 we consider ANNs with the ReLU activation function
R∋x↦max{x,0}∈R. | (1.2) |
The ReLU activation function fails to be differentiable and this lack of regularity also transfers to the risk function of the considered supervised learning problem; cf. (1.5) below. We thus need to employ appropriately generalized gradients of the risk function to specify the dynamics of the GFs. As in [34, Setting 2.1 and Proposition 2.3] (cf. also [33,37]), we accomplish this, first, by approximating the ReLU activation function through continuously differentiable functions which converge pointwise to the ReLU activation function and whose derivatives converge pointwise to the left derivative of the ReLU activation function and, thereafter, by specifying the generalized gradient function as the limit of the gradients of the approximated risk functions; see (1.3) and (1.5) in Theorem 1.1 and (1.9) and (1.10) in Theorem 1.2 for details.
We now present the precise statement of Theorem 1.1 and, thereafter, provide further comments regarding Theorem 1.2.
Theorem 1.1 (Existence and uniqueness of solutions of GFs in the training of ANNs). Let d,H,d∈N, a∈R, b∈(a,∞), f∈C([a,b]d,R) satisfy d=dH+2H+1, let p:[a,b]d→[0,∞) be bounded and measurable, let Rr∈C(R,R), r∈N∪{∞}, satisfy for all x∈R that (∪r∈N{Rr})⊆C1(R,R), R∞(x)=max{x,0}, supr∈Nsupy∈[−|x|,|x|]|(Rr)′(y)|<∞, and
lim supr→∞(|Rr(x)−R∞(x)|+|(Rr)′(x)−1(0,∞)(x)|)=0, | (1.3) |
for every θ=(θ1,…,θd)∈Rd let Dθ⊆N satisfy
Dθ={i∈{1,2,…,H}:|θHd+i|+∑dj=1|θ(i−1)d+j|=0}, | (1.4) |
for every r∈N∪{∞} let Lr:Rd→R satisfy for all θ=(θ1,…,θd)∈Rd that
Lr(θ)=∫[a,b]d(f(x1,…,xd)−θd−∑Hi=1θH(d+1)+i[Rr(θHd+i+∑dj=1θ(i−1)d+jxj)])2p(x)d(x1,…,xd), | (1.5) |
let θ∈Rd, and let G:Rd→Rd satisfy for all ϑ∈{v∈Rd:((∇Lr)(v))r∈Nis convergent} that G(ϑ)=limr→∞(∇Lr)(ϑ). Then
(i) it holds that G is locally bounded and measurable and
(ii) there exists a unique Θ∈C([0,∞),Rd) which satisfies for all t∈[0,∞), s∈[t,∞) that DΘt⊆DΘs and
Θt=θ−∫t0G(Θu)du. | (1.6) |
Theorem 1.1 is a direct consequence of Theorem 3.3 below. In Theorem 1.2 we also assume that the target function f:[a,b]d→R is continuous but additionally assume that, roughly speaking, both the target function f:[a,b]d→R and the unnormalized density function p:[a,b]d→[0,∞) coincide with polynomial functions on suitable subsets of their domain of definition [a,b]d. In Theorem 1.2 the (n×d)-matrices αki∈Rn×d, i∈{1,2,…,n}, k∈{0,1}, and the n-dimensional vectors βki∈Rn, i∈{1,2,…,n}, k∈{0,1}, are used to describe these subsets and the functions Pki:Rd→R, i∈{1,2,…,n}, k∈{0,1}, constitute the polynomials with which the target function and the unnormalized density function should partially coincide. More formally, in (1.8) in Theorem 1.2 we assume that for every x∈[a,b]d we have that
p(x)=∑i∈{1,2,…,n},α0ix+β0i∈[0,∞)nP0i(x)andf(x)=∑i∈{1,2,…,n},α1ix+β1i∈[0,∞)nP1i(x). | (1.7) |
In (1.11) in Theorem 1.2 we prove that there exists a strictly positive real number β∈(0,∞) such that for every GF trajectory Θ:[0,∞)→Rd which does not diverge to infinity in the sense* that lim inft→∞||Θt||<∞ we have that Θt∈Rd, t∈[0,∞), converges with order β to a critical point ϑ∈G−1({0})={θ∈Rd:G(θ)=0} and we have that the risk L(Θt)∈R, t∈[0,∞), converges with order 1 to the risk L(ϑ) of the critical point ϑ. We now present the precise statement of Theorem 1.2.
*Note that the functions ||⋅||:(∪n∈NRn)→R and ⟨⋅,⋅⟩:(∪n∈N(Rn×Rn))→R satisfy for all n∈N, x=(x1,…,xn), y=(y1,…,yn)∈Rn that ||x||=[∑ni=1|xi|2]1/2 and ⟨x,y⟩=∑di=1xiyi.
Theorem 1.2 (Convergence rates for GFs trajectories in the training of ANNs). Let d,H,d,n∈N, a∈R, b∈(a,∞), f∈C([a,b]d,R) satisfy d=dH+2H+1, for every i∈{1,2,…,n}, k∈{0,1} let αki∈Rn×d, let βki∈Rn, and let Pki:Rd→R be a polynomial, let p:[a,b]d→[0,∞) satisfy for all k∈{0,1}, x∈[a,b]d that
kf(x)+(1−k)p(x)=∑ni=1[Pki(x)1[0,∞)n(αkix+βki)], | (1.8) |
let Rr∈C(R,R), r∈N∪{∞}, satisfy for all x∈R that (∪r∈N{Rr})⊆C1(R,R), R∞(x)=max{x,0}, supr∈Nsupy∈[−|x|,|x|]|(Rr)′(y)|<∞, and
lim supr→∞(|Rr(x)−R∞(x)|+|(Rr)′(x)−1(0,∞)(x)|)=0, | (1.9) |
for every r∈N∪{∞} let Lr:Rd→R satisfy for all θ=(θ1,…,θd)∈Rd that
Lr(θ)=∫[a,b]d(f(x1,…,xd)−θd−∑Hi=1θH(d+1)+i[Rr(θHd+i+∑dj=1θ(i−1)d+jxj)])2p(x)d(x1,…,xd), | (1.10) |
let G:Rd→Rd satisfy for all θ∈{ϑ∈Rd:((∇Lr)(ϑ))r∈Nis convergent} that G(θ)=limr→∞(∇Lr)(θ), and let Θ∈C([0,∞),Rd) satisfy lim inft→∞||Θt||<∞ and ∀t∈[0,∞):Θt=Θ0−∫t0G(Θs)ds. Then there exist ϑ∈G−1({0}), C,β∈(0,∞) which satisfy for all t∈[0,∞) that
||Θt−ϑ||≤C(1+t)−βand|L∞(Θt)−L∞(ϑ)|≤C(1+t)−1. | (1.11) |
Theorem 1.2 above is an immediate consequence of Theorem 5.4 in Subsection 5.3 below. Theorem 1.2 is related to Theorem 1.1 in our previous article [37]. In particular, [37, Theorem 1.1] uses weaker assumptions than Theorem 1.2 above but Theorem 1.2 above establishes a stronger statement when compared to [37, Theorem 1.1]. Specifically, on the one hand in [37, Theorem 1.1] the target function is only assumed to be a continuous function and the unnormalized density is only assumed to be measurable and integrable while in Theorem 1.2 it is additionally assumed that both the target function and the unnormalized density are piecewise polynomial in the sense of (1.8) above. On the other hand [37, Theorem 1.1] only asserts that the risk of every bounded GF trajectory converges to the risk of critical point while Theorem 1.2 assures that every non-divergent GF trajectory converges with a strictly positive rate of convergence to a critical point (the rate of convergence is given through the strictly positive real number β∈(0,∞) appearing in the exponent on the left inequality in (Eq 1.11) in Theorem 1.2) and also assures that the risk of the non-divergent GF trajectory converges with rate 1 to the risk of the critical point (the convergence rate 1 is ensured through the 1 appearing in the exponent on the right inequality in (Eq 1.11) in Theorem 1.2).
We also point out that Theorem 1.2 assumes that the GF trajectory is non-divergent in the sense that lim inft→∞||Θt||<∞. In general, it remains an open problem to establish sufficient conditions which ensure that the GF trajectory has this non-divergence property. In this aspect we also refer to Gallon et al. [38] for counterexamples for which it has been proved that every GF trajectory with sufficiently small initial risk does in the training of ANNs diverge to ∞ in the sense that lim inft→∞||Θt||=∞.
The remainder of this article is organized in the following way. In Section 2 we establish several regularity properties for the risk function of the considered supervised learning problem and its generalized gradient function. In Subsection 3 we employ the findings from Section 2 to establish existence and uniqueness properties for solutions of GF differential equations. In particular, in Subsection 3 we present the proof of Theorem 1.1 above. In Subsection 4 we establish under the assumption that both the target function f:[a,b]d→R and the unnormalized density function p:[a,b]d→[0,∞) are piecewise polynomial that the risk function is semialgebraic in the sense of Definition 4.3 in Subsection 4 (see Corollary 4.10 in Subsection 4 for details). In Subsection 5 we engage the results from Sections 2 and 4 to establish several convergence rate results for solutions of GF differential equations and, thereby, we also prove Theorem 1.2 above.
In this section we establish several regularity properties for the risk function L:Rd→R and its generalized gradient function G:Rd→Rd. In particular, in Proposition 2.12 in Subsection 2.5 below we prove for every parameter vector θ∈Rd in the ANN parameter space Rd=RdH+2H+1 that the generalized gradient G(θ) is a limiting subdifferential of the risk function L:Rd→R at θ. In Definition 2.8 in Subsection 2.5 we recall the notion of subdifferentials (which are sometimes also referred to as Fréchet subdifferentials in the scientific literature) and in Definition 2.9 in Subsection 2.5 we recall the notion of limiting subdifferentials. In the scientific literature Definitions 2.8 and 2.9 can in a slightly different presentational form, e.g., be found in Rockafellar & Wets [39, Definition 8.3] and Bolte et al. [9, Definition 2.10], respectively.
Our proof of Proposition 2.12 uses the continuously differentiability result for the risk function in Proposition 2.3 in Subsection 2.2 and the local Lipschitz continuity result for the generalized gradient function in Corollary 2.7 in Subsection 2.4. Corollary 2.7 will also be employed in Subsection 3 below to establish existence and uniqueness results for solutions of GF differential equations. Proposition 2.3 follows directly from [37, Proposition 2.10, Lemmas 2.11 and 2.12]. Our proof of Corollary 2.7, in turn, employs the known representation result for the generalized gradient function in Proposition 2.2 in Subsection 2.2 below and the local Lipschitz continuity result for certain parameter integrals in Corollary 2.6 in Subsection 2.4. Statements related to Proposition 2.2 can, e.g., be found in [37, Proposition 2.2], [33, Proposition 2.3], and [34, Proposition 2.3].
Our proof of Corollary 2.6 uses the elementary abstract local Lipschitz continuity result for certain parameter integrals in Lemma 2.5 in Subsection 2.4 and the local Lipschitz continuity result for active neuron regions in Lemma 2.4 in Subsection 2.3 below. Lemma 2.4 is a generalization of [35, Lemma 7], Lemma 2.5 is a slight generalization of [35, Lemma 6], and Corollary 2.6 is a generalization of [37, Lemma 2.12] and [35, Corollary 9]. The proof of Lemma 2.5 is therefore omitted.
In Setting 2.1 in Subsection 2.1 below we present the mathematical setup to describe ANNs with ReLU activation, the risk function L:Rd→R, and its generalized gradient function G:Rd→Rd. Moreover, in (2.6) in Setting 2.1 we define for a given parameter vector θ∈Rd the set of hidden neurons which have all input parameters equal to zero. Such neurons are sometimes called degenerate (cf. Cheridito et al. [36]) and can cause problems with the differentiability of the risk function, which is why we exclude degenerate neurons in Proposition 2.3 and Corollary 2.7 below.
In this subsection we present in Setting 2.1 below the mathematical setup that we employ to state most of the mathematical results of this work. We also refer to Figure 2 below for a table in which we briefly list the mathematical objects introduced in Setting 2.1.
Setting 2.1. Let d,H,d∈N, a∈R, b∈(a,∞), f∈C([a,b]d,R) satisfy d=dH+2H+1, let w=((wθi,j)(i,j)∈{1,…,H}×{1,…,d})θ∈Rd:Rd→RH×d, b=((bθ1,…,bθH))θ∈Rd:Rd→RH, v=((vθ1,…,vθH))θ∈Rd:Rd→RH, and c=(cθ)θ∈Rd:Rd→R satisfy for all θ=(θ1,…,θd)∈Rd, i∈{1,2,…,H}, j∈{1,2,…,d} that
wθi,j=θ(i−1)d+j,bθi=θHd+i,vθi=θH(d+1)+i,andcθ=θd, | (2.1) |
let Rr∈C1(R,R), r∈N, satisfy for all x∈R that
lim supr→∞(|Rr(x)−max{x,0}|+|(Rr)′(x)−1(0,∞)(x)|)=0 | (2.2) |
and supr∈Nsupy∈[−|x|,|x|]|(Rr)′(y)|<∞, let λ:B(Rd)→[0,∞] be the Lebesgue–Borel measure on Rd, let p:[a,b]d→[0,∞) be bounded and measurable, let N=(Nθ)θ∈Rd:Rd→C(Rd,R) and L:Rd→R satisfy for all θ∈Rd, x=(x1,…,xd)∈Rd that
Nθ(x)=cθ+∑Hi=1vθi{bθi+∑dj=1wθi,jxj,0} | (2.3) |
and L(θ)=∫[a,b]d(f(y)−Nθ(y))2p(y)λ(dy), for every r∈N let Lr:Rd→R satisfy for all θ∈Rd that
Lr(θ)=∫[a,b]d(f(y)−cθ−∑Hi=1vθi[Rr(bθi+∑dj=1wθi,jyj)])2p(y)λ(dy), | (2.4) |
for every ε∈(0,∞), θ∈Rd let Bε(θ)⊆Rd satisfy Bε(θ)={ϑ∈Rd:||θ−ϑ||<ε}, for every θ∈Rd, i∈{1,2,…,H} let Iθi⊆Rd satisfy
Iθi={x=(x1,…,xd)∈[a,b]d:bθi+∑dj=1wθi,jxd>0}, | (2.5) |
for every θ∈Rd let Dθ⊆N satisfy
Dθ={i∈{1,2,…,H}:|bθi|+∑dj=1|wθi,j|=0}, | (2.6) |
and let G=(G1,…,Gd):Rd→Rd satisfy for all θ∈{ϑ∈Rd:((∇Lr)(ϑ))r∈Nis convergent} that G(θ)=limr→∞(∇Lr)(θ).
Next we add some explanations regarding the mathematical framework presented in Setting 2.1 above. In Setting 2.1
● the natural number d∈N represents the number of neurons on the input layer of the considered ANNs,
● the natural number H∈N represents the number of neurons on the hidden layer of the considered ANNs, and
● the natural number d∈N measures the overall number of parameters of the considered ANNs
(cf. (1.1) and Figure 1 above). The real numbers a∈R, b∈(a,∞) in Setting 2.1 are employed to specify the d-dimensional set [a,b]d⊆Rd in which the input data of the supervised learning problem considered in Setting 2.1 takes values in and which, thereby, also serves as the domain of definition of the target function of the considered supervised learning problem.
In Setting 2.1 the function f:[a,b]d→R represents the target function of the considered supervised learning problem. In Setting 2.1 the target function f is assumed to be an element of the set C([a,b]d,R) of continuous functions from the d-dimensional set [a,b]d to the reals R (first line in Setting 2.1).
The matrix valued function w=((wθi,j)(i,j)∈{1,…,H}×{1,…,d})θ∈Rd:Rd→RH×d in Setting 2.1 is used to represent the inner weight parameters of the ANNs considered in Setting 2.1. In particular, in Setting 2.1 we have for every θ∈Rd that the H×d-matrix wθ=(wθi,j)(i,j)∈{1,…,H}×{1,…,d}∈RH×d represents the weight parameter matrix for the affine linear transformation from the d-dimensional input layer to the H-dimensional hidden layer of the ANN associated to the ANN parameter vector θ∈Rd (cf. (2.1), (2.3), and Figure 1).
The vector valued function b=((bθ1,…,bθH))θ∈Rd:Rd→RH in Setting 2.1 is used to represent the inner bias parameters of the ANNs considered in Setting 2.1. In particular, in Setting 2.1 we have for every θ∈Rd that the d-dimensional vector bθ=(bθ1,…,bθH)∈RH represents the bias parameter vector for the affine linear transformation from the d-dimensional input layer to the H-dimensional hidden layer of the ANN associated to the ANN parameter vector θ∈Rd (cf. (2.1), (2.3), and Figure 1).
The vector valued function v=((vθ1,…,vθH))θ∈Rd:Rd→RH in Setting 2.1 is used to describe the outer weight parameters of the ANNs considered in Setting 2.1. In particular, in Setting 2.1 we have for every θ∈Rd that the transpose of the H-dimensional vector vθ=(vθ1,…,vθH)∈RH represents the weight parameter matrix for the affine linear transformation from the H-dimensional hidden layer to the 1-dimensional output layer of the ANN associated to the ANN parameter vector θ∈Rd (cf. (2.1), (2.3), and Figure 1).
The real valued function c=(cθ)θ∈Rd:Rd→R in Setting 2.1 is used to represent the outer bias parameters of the ANNs considered in Setting 2.1. In particular, in Setting 2.1 we have for every θ∈Rd that the real number cθ∈R describes the bias parameter for the affine linear transformation from the H-dimensional hidden layer to the 1-dimensional output layer of the ANN associated to the ANN parameter vector θ∈Rd (cf. (2.1), (2.3), and Figure 1).
In Setting 2.1 we consider ANNs with the ReLU activation function R∋x↦max{x,0}∈R (cf. (1.2)). The ReLU activation function fails to be differentiable and this lack of differentiability typically transfers from the activation function to the realization functions Nθ:Rd→R, θ∈Rd, of the considered ANNs and the risk function L:Rd→R of the considered supervised learning problem, both, introduced in (2.3) in Setting 2.1. In general, there thus do not exist standard derivatives and standard gradients of the risk function and, in view of this, we need to introduce suitably generalized gradients of the risk function to specify the GF dynamics. As in [34, Setting 2.1 and Proposition 2.3] (cf. also [33,37]), we accomplish this,
● first, by approximating the ReLU activation function through appropriate continuously differentiable functions which converge pointwise to the ReLU activation function and whose derivatives converge pointwise to the left derivative of the ReLU activation function,
● then, by using these continuously differentiable approximations of the ReLU activation function to specify approximated risk functions, and,
● finally, by specifying the generalized gradient function as the pointwise limit of the standard gradients of the approximated risk functions.
In Setting 2.1 the functions Rr:R→R, r∈N, serves as such appropriate continuously differentiable approximations of the ReLU activation function and the hypothesis in (2.2) ensures that these functions converge pointwise to the ReLU activation function and that the derivatives of these functions converge pointwise to the left derivative of the ReLU activation function (cf. also (1.3) in Theorem 1.1 and (1.9) in Theorem 1.2). These continuously differentiable approximations of the ReLU activation function are then used in (2.4) in Setting 2.1 (cf. also (1.5) in Theorem 1.1 and (1.10) in Theorem 1.2) to introduce continuously differentiable approximated risk functions Lr:Rd→R, r∈N, which converge pointwise to the risk function L:Rd→R (cf., e.g., [37, Proposition 2.2]). Finally, the standard gradients of the approximated risk functions Lr:Rd→R, r∈N, are then used to introduce the generalized gradient function G=(G1,…,Gd):Rd→Rd in Setting 2.1. In this regard we also note that Proposition 2.2 in Subsection 2.2 below, in particular, ensures that the function G=(G1,…,Gd):Rd→Rd in Setting 2.1 is indeed uniquely defined.
Proposition 2.2. Assume Setting 2.1. Then it holds for all θ∈Rd, i∈{1,2,…,H}, j∈{1,2,…,d} that
G(i−1)d+j(θ)=2vθi∫Iθixj(Nθ(x)−f(x))p(x)λ(dx),GHd+i(θ)=2vθi∫Iθi(Nθ(x)−f(x))p(x)λ(dx),GH(d+1)+i(θ)=2∫[a,b]d[max{bθi+∑dj=1wθi,jxj,0}](Nθ(x)−f(x))p(x)λ(dx),andGd(θ)=2∫[a,b]d(Nθ(x)−f(x))p(x)λ(dx). | (2.7) |
Proof of Proposition 2.2. Observe that, e.g., [37, Proposition 2.2] establishes 2.7. The proof of Proposition 2.2 is thus complete.
Proposition 2.3. Assume Setting 2.1 and let U⊆Rd satisfy U={θ∈Rd:Dθ=∅}. Then
(i) it holds that U⊆Rd is open,
(ii) it holds that L|U∈C1(U,R), and
(iii) it holds that ∇(L|U)=G|U.
Proof of Proposition 2.3. Note that [37, Proposition 2.10,Lemmas 2.11 and 2.12] establish items (i), (ii), and (iii). The proof of Proposition 2.3 is thus complete.
Lemma 2.4. Let d∈N, a∈R, b∈(a,∞), for every v=(v1,…,vd+1)∈Rd+1 let Iv⊆[a,b]d satisfy Iv={x∈[a,b]d:vd+1+∑di=1vixi>0}, for every n∈N let λn:B(Rn)→[0,∞] be the Lebesgue–Borel measure on Rn, let p:[a,b]d→[0,∞) be bounded and measurable, and let u∈Rd+1∖{0}. Then there exist ε,C∈(0,∞) such that for all v,w∈Rd+1 with max{||u−v||,||u−w||}≤ε it holds that
∫IvΔIwp(x)λd(dx)≤C||v−w||. | (2.8) |
Proof of Lemma 2.4. Observe that for all v,w∈Rd+1 we have that
∫IvΔIwp(x)λd(dx)≤(supx∈[a,b]dp(x))λd(IvΔIw). | (2.9) |
Moreover, note that the fact that for all y∈R it holds that y≥−|y| ensures that for all v=(v1,…,vd+1)∈Rd+1, i∈{1,2,…,d+1} with ||u−v||<|ui| it holds that
uivi=(ui)2+(vi−ui)ui≥|ui|2−|ui−vi||ui|≥|ui|2−||u−v|||ui|>0. | (2.10) |
Next observe that for all v1,v2,w1,w2∈R with min{|v1|,|w1|}>0 it holds that
|v2v1−w2w1|=|v2w1−w2v1||v1w1|=|v2(w1−v1)+v1(v2−w2)||v1w1|≤[|v2|+|v1||v1w1|][|v1−w1|+|v2−w2|]. | (2.11) |
Combining this and 2.10 demonstrates for all v=(v1,…,vd+1), w=(w1,…,wd+1)∈Rd+1, i∈{1,2,…,d} with max{||v−u||,||w−u||}<|u1| that v1w1>0 and
|viv1−wiw1|≤[2||v|||v1w1|][2||v−w||]≤[4||v−u||+4||u|||v1w1|]||v−w||. | (2.12) |
Hence, we obtain for all v=(v1,…,vd+1), w=(w1,…,wd+1)∈Rd+1, i∈{1,2,…,d} with max{||v−u||,||w−u||}≤|u1|2 and |u1|>0 that v1w1>0 and
|viv1−wiw1|≤(2|u1|+4||u||)||v−w|||u1+(v1−u1)||u1+(w1−u1)|≤6||u||||v−w||(|u1|−||v−u||)(|u1|−||w−u||)≤[24||u|||u1|2]||v−w||. | (2.13) |
In the following we distinguish between the case maxi∈{1,2,…,d}|ui|=0, the case (maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×[2,∞), and the case (maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×{1}. We first prove 2.8 in the case
maxi∈{1,2,…,d}|ui|=0. | (2.14) |
Note that (2.14) and the assumption that u∈Rd+1∖{0} imply that |ud+1|>0. Moreover, observe that (2.14) shows that for all v=(v1,…,vd+1)∈Rd+1, x=(x1,…,xd)∈IuΔIv we have that
|([∑di=1vixi]+vd+1)−([∑di=1uixi]+ud+1)|=|[∑di=1vixi]+vd+1|+|[∑di=1uixi]+ud+1|≥|[∑di=1uixi]+ud+1|=|ud+1|. | (2.15) |
In addition, note that for all v=(v1,…,vd+1)∈Rd+1, x=(x1,…,xd)∈[a,b]d it holds that
|([∑di=1vixi]+vd+1)−([∑di=1uixi]+ud+1)|≤[∑di=1|vi−ui||xi|]+|vd+1−ud+1|≤max{|a|,|b|}[∑di=1|vi−ui|]+|vd+1−ud+1|≤(1+dmax{|a,b|})||v−u||. | (2.16) |
This and (2.15) prove that for all v∈Rd+1 with ||u−v||≤|ud+1|2+dmax{|a,b|} we have that IuΔIv=∅, i.e., Iu=Iv. Therefore, we get for all v,w∈Rd+1 with max{||u−v||,||u−w||}≤|ud+1|2+dmax{|a,b|} that Iv=Iw=Iu. Hence, we obtain for all v,w∈Rd+1 with max{||u−v||,||u−w||}≤|ud+1|2+dmax{|a,b|} that λd(IvΔIw)=0. This establishes (2.8) in the case maxi∈{1,2,…,d}|ui|=0. In the next step we prove 2.8 in the case
(maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×[2,∞). | (2.17) |
For this we assume without loss of generality that |u1|>0. In the following let Jv,wx⊆R, x∈[a,b]d−1, v,w∈Rd+1, satisfy for all x=(x2,…,xd)∈[a,b]d−1, v,w∈Rd+1 that Jv,wx={y∈[a,b]:(y,x2,…,xd)∈Iv∖Iw}. Next observe that Fubini's theorem and the fact that for all v∈Rd+1 it holds that Iv is measurable show that for all v,w∈Rd+1 we have that
λd(IvΔIw)=∫[a,b]d1IvΔIw(x)λd(dx)=∫[a,b]d(1Iv∖Iw(x)+1Iw∖Iv(x))λd(dx)=∫[a,b]d−1∫[a,b](1Iv∖Iw(y,x2,…,xd)+1Iw∖Iv(y,x2,…,xd))λ1(dy)λd−1(d(x2,…,xd))=∫[a,b]d−1∫[a,b](1Jv,wx(y)+1Jw,vx(y))λ1(dy)λd−1(dx)=∫[a,b]d−1(λ1(Jv,wx)+λ1(Jw,vx))λd−1(dx). | (2.18) |
Furthermore, note that for all x=(x2,…,xd)∈[a,b]d−1, v=(v1,…,vd+1), w=(w1,…,wd+1)∈Rd+1, s∈{−1,1} with min{sv1,sw1}>0 it holds that
Jv,wx={y∈[a,b]:(y,x2,…,xd)∈Iv∖Iw}={y∈[a,b]:v1y+[∑di=2vixi]+vd+1>0≥w1y+[∑di=2wixi]+wd+1}={y∈[a,b]:−sv1([∑di=2vixi]+vd+1)<sy≤−sw1([∑di=2wixi]+wd+1)}. | (2.19) |
Hence, we obtain for all x=(x2,…,xd)∈[a,b]d−1, v=(v1,…,vd+1), w=(w1,…,wd+1)∈Rd+1, s∈{−1,1} with min{sv1,sw1}>0 that
λ1(Jv,wx)≤|sv1([∑di=2vixi]+vd+1)−sw1([∑di=2wixi]+wd+1)|≤[∑di=2|viv1−wiw1||xi|]+|vd+1v1−wd+1w1|≤max{|a|,|b|}[∑di=2|viv1−wiw1|]+|vd+1v1−wd+1w1|. | (2.20) |
Furthermore, observe that (2.10) demonstrates for all v=(v1,…,vd+1)∈Rd+1 with ||u−v||<|u1| that u1v1>0. This implies that for all v=(v1,…,vd+1), w=(w1,…,wd+1)∈Rd+1 with max{||u−v||,||u−w||}<|u1| there exists s∈{−1,1} such that min{sv1,sw1}>0. Combining this and (2.13) with (2.20) proves that there exists C∈R such that for all x∈[a,b]d−1, v,w∈Rd+1 with max{||u−v||,||u−w||}≤|u1|2 we have that λ1(Jv,wx)+λ1(Jw,vx)≤C||v−w||. This, (2.18), and (2.9) establish (2.8) in the case (maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×[2,∞). Finally, we prove (2.8) in the case
(maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×{1}. | (2.21) |
Note that (2.21) demonstrates that |u1|>0. In addition, observe that for all v=(v1,v2), w=(w1,w2)∈R2, s∈{−1,1} with min{sv1,sw1}>0 it holds that
Iv∖Iw={y∈[a,b]:v1y+v2>0≥w1y+w2}={y∈[a,b]:−sv2v1<sy≤−sw2w1}⊆{y∈R:−sv2v1<sy≤−sw2w1}. | (2.22) |
Therefore, we get for all v=(v1,v2), w=(w1,w2)∈R2, s∈{−1,1} with min{sv1,sw1}>0 that
λ1(Iv∖Iw)≤|(−sv2v1)−(−sw2w1)|=|v2v1−w2w1|. | (2.23) |
Furthermore, note that (2.10) ensures for all v=(v1,v2)∈R2 with ||u−v||<|u1| that u1v1>0. This proves that for all v=(v1,v2), w=(w1,w2)∈R2 with max{||u−v||,||u−w||}<|u1| there exists s∈{−1,1} such that min{sv1,sw1}>0. Combining this with (2.23) demonstrates for all v=(v1,v2), w=(w1,w2)∈R2 with max{||u−v||,||u−w||}<|u1| that min{|v1|,|w1|}>0 and
λ1(IvΔIw)=λ1(Iv∖Iw)+λ1(Iw∖Iv)≤2|v2v1−w2w1|. | (2.24) |
This, (2.13), and (2.9) establish (2.8) in the case (maxi∈{1,2,…,d}|ui|,d)∈(0,∞)×{1}. The proof of Lemma 2.4 is thus complete.
Lemma 2.5. Let d,n∈N, a∈R, b∈(a,∞), x∈Rn, C,ε∈(0,∞), let ϕ:Rn×[a,b]d→R be locally bounded and measurable, assume for all r∈(0,∞) that
supy,z∈Rn,||y||+||z||≤r,y≠zsups∈[a,b]d|ϕ(y,s)−ϕ(z,s)|||y−z||<∞, | (2.25) |
let μ:B([a,b]d)→[0,∞) be a finite measure, let Iy∈B([a,b]d), y∈Rn, satisfy for all y,z∈{v∈Rn:||x−v||≤ε} that μ(IyΔIz)≤C||y−z||, and let Φ:Rn→R satisfy for all y∈Rn that
Φ(y)=∫Iyϕ(y,s)μ(ds). | (2.26) |
Then there exists C∈R such that for all y,z∈{v∈Rn:||x−v||≤ε} it holds that |Φ(y)−Φ(z)|≤C||y−z||.
Proof of Lemma 2.5. The proof is analogous to the proof of [35, Lemma 6].
Corollary 2.6. Assume Setting 2.1, let ϕ:Rd×[a,b]d→R be locally bounded and measurable, and assume for all r∈(0,∞) that
supθ,ϑ∈Rd,||θ||+||ϑ||≤r,θ≠ϑsupx∈[a,b]d|ϕ(θ,x)−ϕ(ϑ,x)|||θ−ϑ||<∞. | (2.27) |
Then
(i) it holds that
Rd∋θ↦∫[a,b]dϕ(θ,x)p(x)λ(dx)∈R | (2.28) |
is locally Lipschitz continuous and
(ii) it holds for all i∈{1,2,…,H} that
{ϑ∈Rd:i∉Dϑ}∋θ↦∫Iθiϕ(θ,x)p(x)λ(dx)∈R | (2.29) |
is locally Lipschitz continuous.
Proof of Corollary 2.6. First observe that Lemma 2.5 (applied for every θ∈Rd with n↶d, x↶θ, μ↶(B([a,b]d)∋A↦∫Ap(x)λ(dx)∈[0,∞)), (Iy)y∈Rn↶([a,b]d)y∈Rd in the notation of Lemma 2.5) establishes item (i). In the following let i∈{1,2,…,H}, θ∈{ϑ∈Rd:i∉Dϑ}. Note that Lemma 2.4 shows that there exist ε,C∈(0,∞) which satisfy for all ϑ1,ϑ2∈Rd with max{||θ−ϑ1||,||θ−ϑ2||}≤ε that
∫Iϑ1iΔIϑ2ip(x)λ(dx)≤C||ϑ1−ϑ2||. | (2.30) |
Combining this with Lemma 2.5 (applied for every θ∈Rd with n↶d, x↶θ, μ↶(B([a,b]d)∋A↦∫Ap(x)λ(dx)∈[0,∞)), (Iy)y∈Rn↶(Iyi)y∈Rd in the notation of Lemma 2.5) demonstrates that there exists C∈R such that for all ϑ1,ϑ2∈Rd with max{||θ−ϑ1||,||θ−ϑ2||}≤ε it holds that
|∫Iϑ1iϕ(ϑ1,x)p(x)λ(dx)−∫Iϑ2iϕ(ϑ2,x)p(x)λ(dx)|≤C||ϑ1−ϑ2||. | (2.31) |
This establishes item (ii). The proof of Corollary 2.6 is thus complete.
Corollary 2.7. Assume Setting 2.1. Then
(i) it holds for all k∈N∩(Hd+H,d] that
Rd∋θ↦Gk(θ)∈R | (2.32) |
is locally Lipschitz continuous,
(ii) it holds for all i∈{1,2,…,H}, j∈{1,2,…,d} that
{ϑ∈Rd:i∉Dϑ}∋θ↦G(i−1)d+j(θ)∈R | (2.33) |
is locally Lipschitz continuous, and
(iii) it holds for all i∈{1,2,…,H} that
{ϑ∈Rd:i∉Dϑ}∋θ↦GHd+i(θ)∈R | (2.34) |
is locally Lipschitz continuous.
Proof of Corollary 2.7. Observe that (2.7) and Corollary 2.6 establish items (i), (ii), and (iii). The proof of Corollary 2.7 is thus complete.
Definition 2.8 (Subdifferential). Let n∈N, f∈C(Rn,R), x∈Rn. Then we denote by ˆ∂f(x)⊆Rn the set given by
ˆ∂f(x)={y∈Rn:lim infRn∖{0}∋h→0(f(x+h)−f(x)−⟨y,h⟩||h||)≥0}. | (2.35) |
Definition 2.9 (Limiting subdifferential). Let n∈N, f∈C(Rn,R), x∈Rn. Then we denote by ∂f(x)⊆Rn the set given by
∂f(x)=⋂ε∈(0,∞)¯[⋃y∈{z∈Rn:||x−z||<ε}ˆ∂f(y)] | (2.36) |
(cf. Definition 2.8).
Lemma 2.10. Let n∈N, f∈C(Rn,R), x∈Rn. Then
∂f(x)={y∈Rn:∃z=(z1,z2):N→Rn×Rn:([∀k∈N:z2(k)∈ˆ∂f(z1(k))],[lim supk→∞(||z1(k)−x||+||z2(k)−y||)=0])} | (2.37) |
(cf. Definitions 2.8 and 2.9).
Proof of Lemma 2.10. Note that (2.36) establishes (2.37). The proof of Lemma 2.10 is thus complete.
Lemma 2.11. Let n∈N, f∈C(Rn,R), let U⊆Rn be open, assume f|U∈C1(U,R), and let x∈U. Then ˆ∂f(x)=∂f(x)={(∇f)(x)} (cf. Definitions 2.8 and 2.9).
Proof of Lemma 2.11. This is a direct consequence of, e.g., Rockafellar & Wets [39, Exercise 8.8]. The proof of Lemma 2.11 is thus complete.
Proposition 2.12. Assume Setting 2.1 and let θ∈Rd. Then G(θ)∈∂L(θ) (cf. Definition 2.9).
Proof of Proposition 2.12. Throughout this proof let ϑ=(ϑn)n∈N:N→Rd satisfy for all n∈N, i∈{1,2,…,H}, j∈{1,2,…,d} that wϑni,j=wθi,j, bϑni=bθi−1n1Dθ(i), vϑni=vθi, and cϑn=cθ. We prove Proposition 2.12 through an application of Lemma 2.10. Observe that for all n∈N, i∈{1,2,…,H}∖Dθ it holds that bϑni=bθi. This implies for all n∈N, i∈{1,2,…,H}∖Dθ that
i∉Dϑn. | (2.38) |
In addition, note that for all n∈N, i∈Dθ it holds that bϑni=−1n<0. This shows for all n∈N, i∈Dθ that
i∉Dϑn. | (2.39) |
Hence, we obtain for all n∈N that Dϑn=∅. Combining this with Proposition 2.3 and Lemma 2.11 demonstrates that for all n∈N it holds that ˆ∂L(ϑn)={(∇L)(ϑn)}={G(ϑn)} (cf. Definition 2.8). Moreover, observe that limn→∞ϑn=θ. It thus remains to show that G(ϑn), n∈N, converges to G(θ). Note that Corollary 2.7 ensures that for all k∈N∩(Hd+H,d] it holds that
limn→∞Gk(ϑn)=Gk(θ). | (2.40) |
Furthermore, observe that Corollary 2.7, (2.38) and (2.39) assure that for all i∈{1,2,…,H}∖Dθ, j∈{1,2,…,d} it holds that
limn→∞G(i−1)d+j(ϑn)=G(i−1)d+j(θ)andlimn→∞GHd+i(ϑn)=GHd+i(θ). | (2.41) |
In addition, note that for all n∈N, i∈Dθ we have that Iϑni=Iθi=∅. Hence, we obtain for all i∈Dθ, j∈{1,2,…,d} that
limn→∞G(i−1)d+j(ϑn)=0=G(i−1)d+j(θ)andlimn→∞GHd+i(ϑn)=0=GHd+i(θ). | (2.42) |
Combining this, (2.40) and (2.41) demonstrates that limn→∞G(ϑn)=G(θ). This and Lemma 2.10 assure that G(θ)∈∂L(θ). The proof of Proposition 2.12 is thus complete.
In this section we employ the local Lipschitz continuity result for the generalized gradient function in Corollary 2.7 from Section 2 to establish existence and uniqueness results for solutions of GF differential equations. Specifically, in Proposition 3.1 in Subsection 3.1 below we prove the existence of solutions GF differential equations, in Lemma 3.2 in Subsection 3.2 below we establish the uniqueness of solutions of GF differential equations among a suitable class of GF solutions, and in Theorem 3.3 in Subsection 3.3 below we combine Proposition 3.1 and Lemma 3.2 to establish the unique existence of solutions of GF differential equations among a suitable class of GF solutions. Theorem 1.1 in the introduction is an immediate consequence of Theorem 3.3.
Roughly speaking, we show in Theorem 3.3 the unique existence of solutions of GF differential equations among the class of GF solutions which satisfy that the set of all degenerate neurons of the GF solution at time t∈[0,∞) is non-decreasing in the time variable t∈[0,∞). In other words, in Theorem 3.3 we prove the unique existence of GF solutions with the property that once a neuron has become degenerate it will remain degenerate for subsequent times.
Our strategy of the proof of Theorem 3.3 and Proposition 3.1, respectively, can, loosely speaking, be described as follows. Corollary 2.7 above implies that the components of the generalized gradient function G:Rd→Rd corresponding to non-degenerate neurons are locally Lipschitz continuous so that the classical Picard-Lindelöf local existence and uniqueness theorem for ordinary differential equations can be brought into play for those components. On the other hand, if at some time t∈[0,∞) the i-th neuron is degenerate, then Proposition 2.2 above shows that the corresponding components of the generalized gradient function G:Rd→Rd vanish. The GF differential equation is thus satisfied if the neuron remains degenerate at all subsequent times s∈[t,∞). Using these arguments we prove in Proposition 3.1 the existence of GF solutions by induction on the number of non-degenerate neurons of the initial value.
Proposition 3.1. Assume Setting 2.1 and let θ∈Rd. Then there exists Θ∈C([0,∞),Rd) which satisfies for all t∈[0,∞), s∈[t,∞) that
Θt=θ−∫t0G(Θu)duandDΘt⊆DΘs. | (3.1) |
Proof of Proposition 3.1. We prove the statement by induction on the quantity H−#(Dθ)∈N∩[0,H]. Assume first that H−#(Dθ)=0, i.e., Dθ={1,2,…,H}. Observe that this implies that wθ=0 and bθ=0. In the following let κ∈R satisfy
κ=∫[a,b]df(x)p(x)λ(dx). | (3.2) |
Note that the Picard–Lindelöf Theorem shows that there exists a unique c∈C([0,∞),R) which satisfies for all t∈[0,∞) that
c(0)=cθandc(t)=c(0)+2κt−2(∫[a,b]dp(x)λ(dx))(∫t0c(s)ds). | (3.3) |
Next let Θ∈C([0,∞),Rd) satisfy for all t∈[0,∞), i∈{1,2,…,H}, j∈{1,2,…,d} that
wΘti,j=wθi,j=bΘti=bθi=0,vΘti=vθi,andcΘt=c(t). | (3.4) |
Observe that (2.7), (3.3), and (3.4) ensure for all t∈[0,∞) that
cΘt=cθ+2κt−2(∫[a,b]dp(x)λ(dx))(∫t0cΘsds)=cθ−2∫t0(−κ+∫[a,b]dcΘsp(x)λ(dx))ds=cθ−2∫t0∫[a,b]d(cΘs+∑Hi=1[vΘsimax{bΘsi+∑dj=1wΘsi,jxj,0}]−f(x))p(x)λ(dx)ds=cθ−2∫t0∫[a,b]d(NΘs(x)−f(x))p(x)λ(dx)ds=cθ−∫t0Gd(Θs)ds. | (3.5) |
Next note that (3.4) and (2.7) show for all t∈[0,∞), i∈N∩[1,d) that DΘt={1,2,…,H} and Gi(Θt)=0. Combining this with (3.4) and (3.5) proves that Θ satisfies 3.1. This establishes the claim in the case #(Dθ)=H.
For the induction step assume that #(Dθ)<H and assume that for all ϑ∈Rd with #(Dϑ)>#(Dθ) there exists Θ∈C([0,∞),Rd) which satisfies for all t∈[0,∞), s∈[t,∞) that Θt=ϑ−∫t0G(Θu)du and DΘt⊆DΘs. In the following let U⊆Rd satisfy
U={ϑ∈Rd:Dϑ⊆Dθ} | (3.6) |
and let G:U→Rd satisfy for all ϑ∈U, i∈{1,2,…,d} that
Gi(ϑ)={0:i∈{(ℓ−1)d+j:ℓ∈Dθ,j∈N∩[1,d]}∪{Hd+ℓ:ℓ∈Dθ}Gi(ϑ):else. | (3.7) |
Observe that (3.6) assures that U⊆Rd is open. In addition, note that Corollary 2.7 implies that G is locally Lipschitz continuous. Combining this with the Picard–Lindelöf Theorem demonstrates that there exist a unique maximal τ∈(0,∞] and Ψ∈C([0,τ),U) which satisfy for all t∈[0,τ) that
Ψt=θ−∫t0G(Ψu)du. | (3.8) |
Next observe that (3.7) ensures that for all t∈[0,τ), i∈Dθ, j∈{1,2,…,d} we have that
wΨti,j=wθi,j=bΨti=bθi=0andvΨti=vθi. | (3.9) |
This, (3.7), and (2.7) demonstrate for all t∈[0,τ) that G(Ψt)=G(Ψt). In addition, note that (3.6) and (3.9) imply for all t∈[0,τ) that DΨt=Dθ. Hence, if τ=∞ then Ψ satisfies (3.1). Next assume that τ<∞. Observe that the Cauchy-Schwarz inequality and [37, Lemma 3.1] prove for all s,t∈[0,τ) with s≤t that
||Ψt−Ψs||≤∫ts||G(Ψu)||du≤(t−s)1/2[∫ts||G(Ψu)||2du]1/2≤(t−s)1/2[∫t0||G(Ψu)||2du]1/2=(t−s)1/2(L(Ψ0)−L(Ψt))1/2≤(t−s)1/2(L(Ψ0))1/2. | (3.10) |
Hence, we obtain for all (tn)n∈N⊆[0,τ) with lim infn→∞tn=τ that (Ψtn) is a Cauchy sequence. This implies that ϑ:=limt↑τΨt∈Rd exists. Furthermore, note that the fact that τ is maximal proves that ϑ∉U. Therefore, we have that Dϑ∖Dθ≠∅. Moreover, observe that (3.9) shows that for all i∈Dθ, j∈{1,2,…,d} it holds that wϑi,j=bϑi=0 and, therefore, i∈Dϑ. This demonstrates that #(Dϑ)>#(Dθ). Combining this with the induction hypothesis ensures that there exists Φ∈C([0,∞),Rd) which satisfies for all t∈[0,∞), s∈[t,∞) that
Φt=ϑ−∫t0G(Φu)duandDΦt⊆DΦs. | (3.11) |
In the following let Θ:[0,∞)→Rd satisfy for all t∈[0,∞) that
Θt={Ψt:t∈[0,τ)Φt−τ:t∈[τ,∞). | (3.12) |
Note that the fact that ϑ=limt↑τΨt and the fact that Φ0=ϑ imply that Θ is continuous. Furthermore, observe that the fact that G is locally bounded and (3.8) ensure that
Θτ=ϑ=limt↑τΨt=limt↑τ[θ−∫t0G(Ψs)ds]=θ−∫τ0G(Ψs)ds=θ−∫τ0G(Θs)ds. | (3.13) |
Hence, we obtain for all t∈[τ,∞) that
Θt=(Θt−Θτ)+Θτ=(Φt−τ−Φ0)+Θτ=−∫t−τ0G(Φs)ds+θ−∫τ0G(Θs)ds=−∫τtG(Θs)+θ−∫τ0G(Θs)ds=θ−∫t0G(Θs)ds. | (3.14) |
This shows that Θ satisfies (3.1). The proof of Proposition 3.1 is thus complete.
Lemma 3.2. Assume Setting 2.1 and let θ∈Rd, Θ1,Θ2∈C([0,∞),Rd) satisfy for all t∈[0,∞), s∈[t,∞), k∈{1,2} that
Θkt=θ−∫t0G(Θku)duandDΘkt⊆DΘks. | (3.15) |
Then it holds for all t∈[0,∞) that Θ1t=Θ2t.
Proof of Lemma 3.2. Assume for the sake of contradiction that there exists t∈[0,∞) such that Θ1t≠Θ2t. By translating the variable t if necessary, we may assume without loss of generality that inf{t∈[0,∞):Θ1t≠Θ2t}=0. Next note that the fact that Θ1 and Θ2 are continuous implies that there exists δ∈(0,∞) which satisfies for all t∈[0,δ], k∈{1,2} that DΘkt⊆Dθ. Furthermore, observe that 3.15 ensures for all t∈[0,∞), i∈Dθ, k∈{1,2} that i∈DΘkt. Hence, we obtain for all t∈[0,∞), i∈Dθ, j∈{1,2,…,d}, k∈{1,2} that
G(i−1)d+j(Θkt)=GHd+i(Θkt)=GH(d+1)+i(Θkt)=0. | (3.16) |
In addition, note that the fact that Θ1 and Θ2 are continuous implies that there exists a compact K⊆{ϑ∈Rd:Dϑ⊆Dθ} which satisfies for all t∈[0,δ], k∈{1,2} that Θkt∈K. Moreover, observe that Corollary 2.7 proves that for all i∈{1,2,…,H}∖Dθ, j∈{1,2,…,d} it holds that G(i−1)d+j,GHd+i,GH(d+1)+i,Gd:K→R are Lipschitz continuous. This and (3.16) show that there exists L∈(0,∞) such that for all t∈[0,δ] we have that
||G(Θ1t)−G(Θ2t)||≤L||Θ1t−Θ2t||. | (3.17) |
In the following let M:[0,∞)→[0,∞) satisfy for all t∈[0,∞) that Mt=sups∈(0,t]||Θ1s−Θ2s||. Note that the fact that inf{t∈[0,∞):Θ1t≠Θ2t}=0 proves for all t∈(0,∞) that Mt>0. Moreover, observe that (3.17) ensures for all t∈(0,δ) that
||Θ1t−Θ2t||=‖ | (3.18) |
Combining this with the fact that M is non-decreasing shows for all t \in (0, \delta) , s \in (0, t] that
\begin{equation} ||\Theta_s^1 - \Theta_s^2|| \leq L s M_s \leq L t M_t. \end{equation} | (3.19) |
This demonstrates for all t \in (0, \min \left\{ {{L^{-1}, \delta }} \right\}) that
\begin{equation} 0 < M_t \leq Lt M_t < M_t, \end{equation} | (3.20) |
which is a contradiction. The proof of Lemma 3.2 is thus complete.
Theorem 3.3. Assume Setting 2.1 and let \theta \in \mathbb{R}^ \mathfrak{d} . Then there exists a unique \Theta \in C([0, \infty), \mathbb{R}^ \mathfrak{d}) which satisfies for all t \in [0, \infty) , s \in [t, \infty) that
\begin{equation} \Theta_t = \theta - \int_0^t \mathcal{G} ( \Theta_u ) \, \mathrm{d} u \qquad\mathit{\text{and}}\qquad \mathbf{D}^{\Theta_t} \subseteq \mathbf{D}^{\Theta_s }. \end{equation} | (3.21) |
Proof of Theorem 3.3. Proposition 3.1 establishes the existence and Lemma 3.2 establishes the uniqueness. The proof of Theorem 3.3 is thus complete.
In this section we establish in Corollary 4.10 in Subsection 4.3 below that under the assumption that both the target function f \colon [\mathscr{a}, \mathscr{b}] ^d \to \mathbb{R} and the unnormalized density function \mathfrak{p} \colon [\mathscr{a}, \mathscr{b}]^d \to [0, \infty) are piecewise polynomial in the sense of Definition 4.9 in Subsection 4.3 we have that the risk function \mathcal{L} \colon \mathbb{R}^{ \mathfrak{d} } \to \mathbb{R} is a semialgebraic function in the sense of Definition 4.3 in Subsection 4.1. In Definition 4.9 we specify precisely what we mean by a piecewise polynomial function, in Definition 4.2 in Subsection 4.1 we recall the notion of a semialgebraic set, and in Definition 4.3 we recall the notion of a semialgebraic function. In the scientific literature Definitions 4.2 and 4.3 can in a slightly different presentational form, e.g., be found in Bierstone & Milman [40, Definitions 1.1 and 1.2] and Attouch et al. [8, Definition 2.1].
Note that the risk function \mathcal{L} \colon \mathbb{R}^{ \mathfrak{d} } \to \mathbb{R} is given through a parametric integral in the sense that for all \theta \in \mathbb{R}^{ \mathfrak{d} } we have that
\begin{equation} \mathcal{L}( \theta ) = \int_{[ \mathscr{a} , \mathscr{b} ] ^d} ( f ( y ) - \mathscr{N} ^{ {\theta} } (y) )^2 \, \mathfrak{p} ( y ) \, \lambda ( \mathrm{d} y ) . \end{equation} | (4.1) |
In general, parametric integrals of semialgebraic functions are no longer semialgebraic functions and the characterization of functions that can occur as such integrals is quite involved (cf. Kaiser [41]). This is the reason why we introduce in Definition 4.6 in Subsection 4.2 below a suitable subclass of the class of semialgebraic functions which is rich enough to contain the realization functions of ANNs with ReLU activation (cf. (4.30) in Subsection 4.2 below) and which can be shown to be closed under integration (cf. Proposition 4.8 in Subsection 4.2 below for the precise statement).
Definition 4.1 (Set of polynomials). Let n \in \mathbb{N}_0 . Then we denote by \mathscr{P}_n \subseteq C(\mathbb{R}^n, \mathbb{R}) the set† of all polynomials from \mathbb{R}^n to \mathbb{R} .
†Note that \mathbb{R}^0 = \{ 0 \} , C(\mathbb{R}^0, \mathbb{R}) = C(\{ 0 \}, \mathbb{R}) , and \#(C(\mathbb{R}^0, \mathbb{R})) = \#(C(\{ 0 \}, \mathbb{R})) = \infty . In particular, this shows for all n \in \mathbb{N}_0 that \operatorname{dim}(\mathbb{R}^n) = n and \#(C(\mathbb{R}^n, \mathbb{R})) = \infty .
Definition 4.2 (Semialgebraic sets). Let n \in \mathbb{N} and let A \subseteq \mathbb{R}^n be a set. Then we say that A is a semialgebraic set if and only if there exist k \in \mathbb{N} and (P_{i, j, \ell })_{ (i, j, \ell) \in \left\{ {{1, 2, \ldots, k}} \right\} ^2 \times \left\{ {{0, 1}} \right\}} \subseteq \mathscr{P}_n such that
\begin{equation} A = \bigcup\limits_{i = 1}^k \bigcap\limits_{j = 1}^k \bigl\{ x \in \mathbb{R}^n \colon P_{i, j, 0} ( x ) = 0 < P_{i , j , 1} ( x ) \bigr\} \end{equation} | (4.2) |
(cf. Definition 4.1).
Definition 4.3 (Semialgebraic functions). Let m, n \in \mathbb{N} and let f \colon \mathbb{R}^n \to \mathbb{R}^m be a function. Then we say that f is a semialgebraic function if and only if it holds that \left\{ {{ (x, f (x)) \colon x \in \mathbb{R}^n }} \right\} \subseteq \mathbb{R}^{m+n} is a semialgebraic set (cf. Definition 4.2).
Lemma 4.4. Let n \in \mathbb{N} and let f, g \colon \mathbb{R}^n \to \mathbb{R} be semialgebraic functions (cf. Definition 4.3). Then
(i) it holds that \mathbb{R}^n \ni x \mapsto f(x) + g(x) \in \mathbb{R} is semialgebraic and
(ii) it holds that \mathbb{R}^n \ni x \mapsto f(x) g(x) \in \mathbb{R} is semialgebraic.
Proof of Lemma 4.4. Note that, e.g., Coste [42, Corollary 2.9] (see, e.g., also Bierstone & Milman [40, Section 1]) establishes items (i) and (ii). The proof of Lemma 4.4 is thus complete.
Definition 4.5 (Set of rational functions). Let n \in \mathbb{N} . Then we denote by \mathscr{R}_n the set given by
\begin{equation} \mathscr{R}_n = \left\{ {{R \colon \mathbb{R}^n \to \mathbb{R} \colon \left[ {{ \exists \, P, Q \in \mathscr{P}_n \colon \forall \, x \in \mathbb{R}^n \colon R(x) = \begin{cases} \frac{P(x) }{ Q ( x ) } & \colon Q ( x ) \not = 0 \\[0.5ex] 0 & \colon Q ( x ) = 0 \end{cases} }} \right]}} \right\} \end{equation} | (4.3) |
(cf. Definition 4.1).
Definition 4.6. Let m \in \mathbb{N} , n \in \mathbb{N}_0 . Then we denote by \mathscr{A}_{m, n} the \mathbb{R} -vector space given by
\begin{eqnarray} \mathscr{A}_{m, n} = \operatorname{span} \Bigl( \Bigl\{ f \colon \mathbb{R}^m \times \mathbb{R}^n \to \mathbb{R} \colon \Bigl[ \exists \, r \in \mathbb{N}, \, A_1, A_2, \ldots, A_r \in \left\{ {{ \left\{ {{0}} \right\}, [0, \infty ), (0, \infty )}} \right\}, \\ R \in \mathscr{R}_m , \, Q \in \mathscr{P}_n, \, P = (P_{i, j})_{ (i, j) \in \left\{ {{1, 2, \ldots, r }} \right\} \times \left\{ {{0, 1, \ldots, n }} \right\}} \subseteq \mathscr{P}_m \colon \forall \, \theta \in \mathbb{R}^m , \, x = (x_1, \ldots, x_n) \in \mathbb{R}^n \colon \\ f ( \theta , x ) = R ( \theta ) Q ( x ) \left[ {{ \prod\nolimits_{i = 1}^r \mathbf{1}_{\smash{{A_i}}} \left( {{ P_{i, 0} ( \theta ) + \sum\nolimits_{j = 1}^n P_{i, j} ( \theta ) x_j }} \right) }} \right] \Bigr] \Bigr\} \Bigr) \end{eqnarray} | (4.4) |
(cf. Definitions 4.1 and 4.5).
Lemma 4.7. Let m \in \mathbb{N} , f \in \mathscr{A}_{m, 0 } (cf. Definition 4.6). Then f is semialgebraic (cf. Definition 4.3).
Proof of Lemma 4.7. Throughout this proof let r \in \mathbb{N} , A_1, A_2, \ldots, A_r \in \left\{ {{ \left\{ {{0}} \right\}, [0, \infty), (0, \infty)}} \right\} , R \in \mathscr{R}_m , P = (P_i)_{ i \in \left\{ {{1, 2, \ldots, r }} \right\}} \subseteq \mathscr{P}_m , and let g \colon \mathbb{R}^m \to \mathbb{R} satisfy for all \theta \in \mathbb{R}^m that
\begin{equation} g(\theta) = R ( \theta ) \prod _{i = 1}^r \mathbf{1}_{\smash{{A_i}}} \left( {{ P_i ( \theta ) }} \right) \end{equation} | (4.5) |
(cf. Definitions 4.1 and 4.5). Due to the fact that sums of semialgebraic functions are again semialgebraic (cf. Lemma 4.4), it suffices to show that g is semialgebraic. Furthermore, observe that for all y \in \mathbb{R} it holds that \mathbf{1}_{\smash{{(0, \infty)}}} (y) = 1 - \mathbf{1}_{\smash{{[0, \infty) }}} (- y) and \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} (y) = \mathbf{1}_{\smash{{[0, \infty) }}} (y) \mathbf{1}_{\smash{{[0, \infty) }}} (- y) . Hence, by linearity we may assume for all i \in \left\{ {{1, 2, \ldots, r }} \right\} that A_i = [0, \infty) . Next let Q_1, Q_2 \in \mathscr{P}_m satisfy for all x \in \mathbb{R}^m that
\begin{equation} R(x) = \begin{cases} \frac{Q_1 ( x ) }{ Q_2 ( x ) } & \colon Q_2 ( x ) \not = 0 \\ 0 & \colon Q_2 ( x ) = 0. \end{cases} \end{equation} | (4.6) |
Note that the graph of \mathbb{R}^m \ni \theta \mapsto R(\theta) \in \mathbb{R} is given by
\begin{equation} \left\{ {{(\theta , y ) \in \mathbb{R}^m \times \mathbb{R} \colon Q_2(\theta ) = 0 , \, y = 0 }} \right\} \cup \left\{ {{(\theta , y ) \in \mathbb{R}^m \times \mathbb{R} \colon Q_2 (\theta ) \not = 0 , \, Q_2 ( \theta ) y - Q_1 ( \theta ) = 0}} \right\}. \end{equation} | (4.7) |
Since both of these sets are described by polynomial equations and inequalities, it follows that \mathbb{R}^m \ni \theta \mapsto R(\theta) \in \mathbb{R} is semialgebraic. In addition, observe that for all i \in \left\{ {{1, 2, \ldots, r}} \right\} the graph of \mathbb{R}^m \ni \theta \mapsto \mathbf{1}_{\smash{{[0, \infty) }}} \left({{ P_i (\theta) }} \right) \in \mathbb{R} is given by
\begin{equation} \left\{ {{(\theta , y ) \in \mathbb{R}^m \times \mathbb{R} \colon P_i ( \theta ) < 0 , \, y = 0 }} \right\} \cup \left\{ {{(\theta , y ) \in \mathbb{R}^m \times \mathbb{R} \colon P_i ( \theta ) \geq 0 , \, y = 1 }} \right\}. \end{equation} | (4.8) |
This demonstrates for all i \in \left\{ {{1, 2, \ldots, r}} \right\} that \mathbb{R}^m \ni \theta \mapsto \mathbf{1}_{\smash{{[0, \infty) }}} \left({{ P_i (\theta) }} \right) \in \mathbb{R} is semialgebraic. Combining this and (4.5) with Lemma 4.4 demonstrates that g is semialgebraic. The proof of Lemma 4.7 is thus complete.
Proposition 4.8. Let m, n \in \mathbb{N} , \mathscr{a} \in \mathbb{R} , \mathscr{b} \in (\mathscr{a}, \infty) , f \in \mathscr{A}_{m, n} (cf. Definition 4.6). Then
\begin{equation} \left[ {{ \mathbb{R}^m \times \mathbb{R}^{n-1} \ni (\theta, x_1, \ldots, x_{n-1} ) \mapsto \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n \in \mathbb{R} }} \right] \in \mathscr{A}_{m, n-1}. \end{equation} | (4.9) |
Proof of Proposition 4.8. By linearity of the integral it suffices to consider a function f of the form
\begin{equation} f(\theta , x ) = R ( \theta ) Q ( x ) \prod\limits_{i = 1}^r \mathbf{1}_{\smash{{A_i}}} \left( {{ P_{i, 0} ( \theta ) + \sum\nolimits_{j = 1}^n P_{i , j} ( \theta ) x_j }} \right) \end{equation} | (4.10) |
where r \in \mathbb{N} , \left({{P_{i, j}}} \right)_{(i, j) \in \left\{ {{1, 2, \ldots, r}} \right\} \times \left\{ {{0, 1, \ldots, n }} \right\} } \subseteq \mathscr{P}_m , A_1, A_2, \ldots, A_r \in \left\{ {{ \left\{ {{0}} \right\}, (0, \infty), [0, \infty) }} \right\} , Q \in \mathscr{P}_n , and R \in \mathscr{R}_m (cf. Definitions 4.1 and 4.5). Moreover, note that for all y \in \mathbb{R} it holds that \mathbf{1}_{\smash{{(0, \infty)}}} (y) = 1 - \mathbf{1}_{\smash{{[0, \infty) }}} (- y) and \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} (y) = \mathbf{1}_{\smash{{[0, \infty) }}} (y) \mathbf{1}_{\smash{{[0, \infty) }}} (- y) . Hence, by linearity we may assume that A_i = [0, \infty) for all i \in \left\{ {{1, 2, \ldots, r }} \right\} . Furthermore, by linearity we may assume that Q is of the form
\begin{equation} Q(x_1, \ldots, x_n) = \prod _{\ell = 1}^n ( x_\ell ) ^{i_\ell} \end{equation} | (4.11) |
with i_1, i_2, \ldots, i_n \in \mathbb{N}_0 . In the following let \mathfrak{s} \colon \mathbb{R} \to \mathbb{R} satisfy for all x \in \mathbb{R} that \mathfrak{s} (x) = \mathbf{1}_{\smash{{(0, \infty)}}} (x) - \mathbf{1}_{\smash{{(0, \infty) }}} (- x) , for every \theta \in \mathbb{R}^m , k \in \left\{ {{-1, 0, 1}} \right\} let \mathcal{S}_k^\theta \subseteq \left\{ {{1, 2, \ldots, r}} \right\} satisfy \mathcal{S}_k^\theta = \left\{ {{ i \in \left\{ {{1, 2, \ldots, r }} \right\} \colon \mathfrak{s} (P_{i, n} (\theta)) = k }} \right\} , and for every i \in \left\{ {{1, 2, \ldots, r}} \right\} let Z_i \colon \mathbb{R}^m \times \mathbb{R}^n \to \mathbb{R} satisfy for all (\theta, x) \in \mathbb{R}^m \times \mathbb{R}^n that
\begin{equation} Z_i ( \theta , x ) = - P_{i, 0} ( \theta ) - \sum\nolimits_{j = 1}^{n-1} P_{i, j} ( \theta ) x_j . \end{equation} | (4.12) |
Observe that (4.10), (4.11), and (4.12) imply for all \theta \in \mathbb{R}^m , x = (x_1, \ldots, x_n) \in \mathbb{R}^n that
\begin{equation} f ( \theta , x ) = R ( \theta ) \left( {{ \prod _{\ell = 1}^n ( x_\ell ) ^{i_\ell} }} \right) \left( {{ \prod _{i = 1}^r \mathbf{1}_{\smash{{ [ 0 , \infty ) }}} \left( {{ P_{i, n } ( \theta ) x_n - Z_i ( \theta , x ) }} \right) }} \right) . \end{equation} | (4.13) |
This shows that f(\theta, x) can only be nonzero if
\begin{equation} \begin{split} \forall \, i \in \mathcal{S}^\theta_1 &\colon x_n \geq \frac{Z_i ( \theta , x )}{ P_{i , n} ( \theta ) }, \\ \forall \, i \in \mathcal{S}^\theta_{-1} &\colon x_n \leq \frac{Z_i ( \theta , x )}{ P_{i , n} ( \theta ) }, \\ \forall \, i \in \mathcal{S}^\theta_0 &\colon - Z_i ( \theta , x ) \ge 0. \end{split} \end{equation} | (4.14) |
Hence, if for given \theta \in \mathbb{R}^m , (x_1, \ldots, x_{n-1}) \in \mathbb{R}^{n-1} there exists x_n \in [\mathscr{a}, \mathscr{b}] which satisfies these conditions then (4.13) and the fact that \int y ^{i_ n } \, \mathrm{d} y = \frac{1}{i_n + 1 } y^{i_n + 1 } imply that
\begin{equation} \begin{split} & \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n \\ & = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod\nolimits_{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ \left( {{ \min \left\{ {{ \mathscr{b} , \min\limits_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) } }} \right\} }} \right)^{\! i_n + 1} \! \! - \left( {{ \max \left\{ {{ \mathscr{a}, \max\limits_{j \in \mathcal{S}_1^\theta} \frac{ Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) } }} \right\} }} \right)^{\! i_n + 1} }} \right]. \end{split} \end{equation} | (4.15) |
Otherwise, we have that \int_ \mathscr{a}^ \mathscr{b} f (\theta, x_1, \ldots, x_n) \, \mathrm{d} x_n = 0 . It remains to write these expressions in the different cases as a sum of functions of the required form in Definition 4.6 by introducing suitable indicator functions. Note that there are four possible cases where the integral is nonzero:
● It holds that \mathscr{a} < \max_{j \in \mathcal{S}_1^\theta} \frac{ Z_j (\theta, x) }{ P_{j, n} (\theta) } < \min_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j (\theta, x) }{ P_{j, n} (\theta) } < \mathscr{b} . In this case, we have
\begin{equation} \begin{split} & \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n \\ & = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod\nolimits_{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ \left( {{ \min\limits_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) }}} \right)^{i_n + 1} - \left( {{ \max\limits_{j \in \mathcal{S}_1^\theta} \frac{ Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) }}} \right)^{i_n + 1} }} \right]. \end{split} \end{equation} | (4.16) |
● It holds that \mathscr{a} < \max_{j \in \mathcal{S}_1^\theta} \frac{ Z_j (\theta, x) }{ P_{j, n} (\theta) } < \mathscr{b} \le \min_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j (\theta, x) }{ P_{j, n} (\theta) } . In this case, we have
\begin{equation} \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod _{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ \mathscr{b}^{i_n + 1} - \left( {{ \max\limits_{j \in \mathcal{S}_1^\theta} \frac{ Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) }}} \right)^{i_n + 1} }} \right]. \end{equation} | (4.17) |
● It holds that \max_{j \in \mathcal{S}_1^\theta} \frac{ Z_j (\theta, x) }{ P_{j, n} (\theta) } \le \mathscr{a} < \min_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j (\theta, x) }{ P_{j, n} (\theta) } < \mathscr{b} . In this case, we have
\begin{equation} \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod _{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ \left( {{ \min\limits_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j ( \theta , x ) }{ P_{j , n} ( \theta ) }}} \right)^{i_n + 1} - \mathscr{a}^{i_n + 1} }} \right]. \end{equation} | (4.18) |
● It holds that \max_{j \in \mathcal{S}_1^\theta} \frac{ Z_j (\theta, x) }{ P_{j, n} (\theta) } \le \mathscr{a} < \mathscr{b} \le \min_{j \in \mathcal{S}_{-1}^\theta} \frac{Z_j (\theta, x) }{ P_{j, n} (\theta) } . In this case, we have
\begin{equation} \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod _{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ \mathscr{b}^{i_n + 1} - \mathscr{a}^{i_n + 1} }} \right]. \end{equation} | (4.19) |
Since these four cases are disjoint, by summing over all possible choices A, B, C \subseteq \left\{ {{1, 2, \ldots, r}} \right\} of the sets \mathcal{S}^\theta_k , k \in \left\{ {{-1, 0, 1}} \right\} , and all choices of (non-empty) subsets \mathcal{I}, \mathcal{J} of \mathcal{S}^\theta_1 , \mathcal{S}_{-1}^\theta where the maximal/minimal values are achieved, we can write
\begin{equation} \int_ \mathscr{a}^ \mathscr{b} f ( \theta , x_1, \ldots, x_n ) \, \mathrm{d} x_n = \frac{ R ( \theta )}{i_n + 1 } \left( {{ \prod _{\ell = 1}^{n-1} x_\ell^{i_\ell} }} \right) \left[ {{ (I) + (II) + (III) + (IV) }} \right], \end{equation} | (4.20) |
where (I), (II), (III), (IV) denote the functions of \theta \in \mathbb{R}^m and (x_1, \ldots, x_{n-1}) \in \mathbb{R}^{ n - 1 } given by
\begin{equation} \begin{split} (I) & = \sum\limits_{ A \dot{\cup} B \dot{\cup} C = \left\{ {{1, \ldots, r }} \right\} } \left[ {{ \prod\limits_{j \in A} \mathbf{1}_{\smash{{(0, \infty ) }}} ( P_{ j , n} ( \theta ) ) \prod\limits_{j \in B} \mathbf{1}_{\smash{{(0, \infty ) }}} ( - P_{j , n} ( \theta ) ) \prod\limits_{j \in C} \left( {{ \mathbf{1}_{\smash{{ \left\{ {{0 }} \right\} }}} ( P_{j , n} ( \theta ) ) \mathbf{1}_{\smash{{[0, \infty ) }}} ( - Z_j ( \theta , x ) }} \right) }} \right] \\ & \sum\limits_{ \varnothing \not = \mathcal{I} \subseteq A} \sum\limits_{ \varnothing \not = \mathcal{J} \subseteq B} \Biggl[ \Biggl[ \prod\limits_{i \in \mathcal{I} } \left( {{ \mathbf{1}_{\smash{{( \mathscr{a} , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{i , n } ( \theta ) }}} \right) \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} \left( {{ \frac{ Z_i ( \theta , x )}{P_{i , n} ( \theta ) } - \frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } }} \right) }} \right) \Biggr. \Biggr. \\ & \times \prod\limits_{j \in A \backslash \mathcal{I} } \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } - \frac{ Z_j ( \theta , x )}{P_{j , n} ( \theta ) } }} \right) \prod\limits_{i \in \mathcal{J} } \left( {{ \mathbf{1}_{\smash{{( \mathscr{a} , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{i , n } ( \theta ) }}} \right) \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} \left( {{ \frac{ Z_i ( \theta , x )}{P_{i , n} ( \theta ) } - \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n} ( \theta ) } }} \right) }} \right) \\ & \Biggl. \times \prod\limits_{j \in B \backslash \mathcal{J} } \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \frac{ Z_j ( \theta , x )}{P_{j , n} ( \theta ) } - \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n} ( \theta ) } }} \right) \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n} ( \theta ) } - \frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } }} \right) \Biggr] \\ & \Biggl. \times \left[ {{ \left( {{ \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n } ( \theta ) } }} \right)^{i_n + 1 } - \left( {{\frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } }} \right) ^{i_n + 1 } }} \right] \Biggr], \end{split} \end{equation} | (4.21) |
\begin{equation} \begin{split} (II) & = \sum\limits_{ A \dot{\cup} B \dot{\cup} C = \left\{ {{1, \ldots, r }} \right\} } \left[ {{ \prod\limits_{j \in A} \mathbf{1}_{\smash{{(0, \infty ) }}} ( P_{j , n} ( \theta ) ) \prod\limits_{j \in B} \mathbf{1}_{\smash{{(0, \infty ) }}} ( - P_{j , n} ( \theta ) ) \prod\limits_{j \in C} \left( {{ \mathbf{1}_{\smash{{ \left\{ {{0 }} \right\} }}} ( P_{j , n} ( \theta ) ) \mathbf{1}_{\smash{{[0, \infty ) }}} ( - Z_j ( \theta , x ) }} \right) }} \right] \\ & \sum\limits_{ \varnothing \not = \mathcal{I} \subseteq A} \Biggl[ \Biggl[ \prod\limits_{i \in \mathcal{I} } \left( {{ \mathbf{1}_{\smash{{( \mathscr{a} , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) }}} \right) \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} \left( {{ \frac{ Z_i ( \theta , x )}{P_{i , n} ( \theta ) } - \frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } }} \right) }} \right) \Biggr. \Biggr. \\ & \times \prod\limits_{j \in A \backslash \mathcal{I} } \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } - \frac{ Z_j ( \theta , x )}{P_{j , n} ( \theta ) } }} \right) \prod\limits_{i \in B } \left( {{ \mathbf{1}_{\smash{{[ \mathscr{b} , \infty ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{i , n} ( \theta ) }}} \right) }} \right) \\ & \Biggl. \times \left[ {{ \mathscr{b}^{i_n + 1 } - \left( {{\frac{ Z_{\min \mathcal{I}} ( \theta , x ) }{P_{ \min \mathcal{I} , n} ( \theta ) } }} \right) ^{i_n + 1 } }} \right] \Biggr], \end{split} \end{equation} | (4.22) |
\begin{equation} \begin{split} (III) & = \sum\limits_{ A \dot{\cup} B \dot{\cup} C = \left\{ {{1, \ldots, r }} \right\} } \left[ {{ \prod\limits_{j \in A} \mathbf{1}_{\smash{{(0, \infty ) }}} ( P_{j , n} ( \theta ) ) \prod\limits_{j \in B} \mathbf{1}_{\smash{{(0, \infty ) }}} ( - P_{j , n} ( \theta ) ) \prod\limits_{j \in C} \left( {{ \mathbf{1}_{\smash{{ \left\{ {{0 }} \right\} }}} ( P_{j , n} ( \theta ) ) \mathbf{1}_{\smash{{[0, \infty ) }}} ( - Z_j ( \theta , x ) }} \right) }} \right] \\ & \sum\limits_{ \varnothing \not = \mathcal{J} \subseteq B} \Biggl[ \Biggl[ \prod\limits_{i \in A } \left( {{ \mathbf{1}_{\smash{{(- \infty , \mathscr{a} ] }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) }}} \right) }} \right) \prod\limits_{i \in \mathcal{J} } \left( {{ \mathbf{1}_{\smash{{( \mathscr{a} , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) }}} \right) \mathbf{1}_{\smash{{\left\{ {{0}} \right\}}}} \left( {{ \frac{ Z_i ( \theta , x )}{P_{ i , n} ( \theta ) } - \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n } ( \theta ) } }} \right) }} \right) \\ & \Biggl. \times \prod\limits_{j \in B \backslash \mathcal{J} } \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \frac{ Z_j ( \theta , x )}{P_{j , n} ( \theta ) } - \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n} ( \theta ) } }} \right) \Biggr] \times \left[ {{ \left( {{ \frac{ Z_{\min \mathcal{J}} ( \theta , x ) }{P_{ \min \mathcal{J} , n} ( \theta ) } }} \right)^{i_n + 1 } - \mathscr{a} ^{i_n + 1 } }} \right] \Biggr], \end{split} \end{equation} | (4.23) |
and
\begin{equation} \begin{split} (IV) & = \sum\limits_{ A \dot{\cup} B \dot{\cup} C = \left\{ {{1, \ldots, r }} \right\} } \left[ {{ \prod\limits_{j \in A} \mathbf{1}_{\smash{{(0, \infty ) }}} ( P_{j , n} ( \theta ) ) \prod\limits_{j \in B} \mathbf{1}_{\smash{{(0, \infty ) }}} ( - P_{j , n} ( \theta ) ) \prod\limits_{j \in C} \left( {{ \mathbf{1}_{\smash{{ \left\{ {{0 }} \right\} }}} ( P_{j , n} ( \theta ) ) \mathbf{1}_{\smash{{[0, \infty ) }}} ( - Z_j ( \theta , x ) }} \right) }} \right] \\ & \times \left( {{ \prod\limits_{i \in A } \mathbf{1}_{\smash{{(- \infty , \mathscr{a} ] }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) } }} \right)\prod\limits_{i \in B } \mathbf{1}_{\smash{{[ \mathscr{b} , \infty ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{i , n } ( \theta ) } }} \right) }} \right) \left[ {{ \mathscr{b}^{i_n + 1 } - \mathscr{a} ^{i_n + 1 } }} \right] . \end{split} \end{equation} | (4.24) |
Note that the first products over all elements of A, B, C precisely describe the conditions that \mathcal{S}_1^\theta = A , \mathcal{S}_{ - 1 }^\theta = B , \mathcal{S}_0^\theta = C , and \forall \, j \in \mathcal{S}_0^\theta \colon - Z_j (\theta, x) \ge 0 . Furthermore, observe that, e.g., in (I) we we must have for all i \in \mathcal{I} , j \in A \backslash \mathcal{I} that \frac{Z_j (\theta, x)}{ P_{j, n } (\theta) } < \frac{ Z_{\min \mathcal{I}} (\theta, x) }{P_{ \min \mathcal{I}, n} (\theta) } = \frac{Z_i (\theta, x)}{ P_{i, n } (\theta) } \in (\mathscr{a}, \mathscr{b}) in order to obtain a non-zero value. In other words, the maximal value of \frac{Z_i (\theta, x)}{ P_{i, n } (\theta) } , i \in A , is achieved exactly for i \in \mathcal{I} , and similarly the minimal value of \frac{Z_j (\theta, x)}{ P_{j, n } (\theta) } , j \in B , is achieved exactly for j \in \mathcal{J} (and analogously in (II), (III) ). Moreover, note that we have for all i \in \mathcal{I} \subseteq A that
\begin{equation} \begin{split} \mathbf{1}_{\smash{{( \mathscr{a} , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) }}} \right) & = \mathbf{1}_{\smash{{( \mathscr{a} , \infty ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{ i , n } ( \theta ) }}} \right) \mathbf{1}_{\smash{{(- \infty , \mathscr{b} ) }}} \left( {{ \frac{Z_i ( \theta , x ) }{ P_{i , n } ( \theta ) }}} \right) \\ & = \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ Z_i ( \theta , x ) - \mathscr{a} P_{ i , n} ( \theta ) }} \right) \mathbf{1}_{\smash{{(0, \infty ) }}} \left( {{ \mathscr{b} P_{ i , n} ( \theta ) - Z_i ( \theta , x ) }} \right). \end{split} \end{equation} | (4.25) |
Here Z_i (\theta, x) is polynomial in \theta and linear in x_1, \ldots, x_{n-1} , and thus of the form required by Definition 4.6. Similarly, the other indicator functions can be brought into the correct form, taking into account the different signs of P_{j, n} (\theta) for j \in A and j \in B . Moreover, observe that the remaining terms can be written as linear combinations of rational functions in \theta and polynomials in x . Hence, we obtain that the functions defined by (I), (II), (III), (IV) are elements of \mathscr{A}_{m, n-1} . The proof of Proposition 4.8 is thus complete.
Definition 4.9. Let d \in \mathbb{N} , let A \subseteq \mathbb{R}^d be a set, and let f \colon A \to \mathbb{R} be a function. Then we say that f is piecewise polynomial if and only if there exist n \in \mathbb{N} , \alpha_1, \alpha_2, \ldots, \alpha_n \in \mathbb{R}^{n \times d} , \beta_{1}, \beta_2, \ldots, \beta_n \in \mathbb{R}^n , P_1, P_2, \ldots, P_n \in \mathscr{P}_d such that for all x \in A it holds that
\begin{equation} f(x) = \sum\nolimits_{i = 1}^n \left[ {{ P_i(x) \mathbf{1}_{\smash{{[0, \infty )^n}}} \left( {{ \alpha_i x + \beta_i }} \right) }} \right] \end{equation} | (4.26) |
(cf. Definition 4.1).
Corollary 4.10. Assume Setting 2.1 and assume that f and \mathfrak{p} are piecewise polynomial (cf. Definition 4.9). Then \mathcal{L} is semialgebraic (cf. Definition 4.3).
Proof of Corollary 4.10. Throughout this proof let F \colon \mathbb{R}^d \to \mathbb{R} and \mathfrak{P} \colon \mathbb{R}^d \to \mathbb{R} satisfy for all x \in \mathbb{R}^d that
\begin{equation} F ( x ) = \begin{cases} f(x) & \colon x \in [ \mathscr{a} , \mathscr{b} ] ^d \\ 0 & \colon x \notin [ \mathscr{a} , \mathscr{b} ] ^d \end{cases} \qquad\text{and}\qquad \mathfrak{P} ( x ) = \begin{cases} \mathfrak{p} ( x ) & \colon x \in [ \mathscr{a} , \mathscr{b} ] ^d \\ 0 & \colon x \notin [ \mathscr{a} , \mathscr{b} ] ^d. \end{cases} \end{equation} | (4.27) |
Note that (4.27) and the assumption that f and \mathfrak{p} are piecewise polynomial assure that
\begin{equation} \left[ {{ \mathbb{R}^ \mathfrak{d} \times \mathbb{R}^d \ni (\theta , x ) \mapsto F ( x ) \in \mathbb{R} }} \right] \in \mathscr{A}_{ \mathfrak{d} , d} \quad\text{and}\quad \left[ {{ \mathbb{R}^ \mathfrak{d} \times \mathbb{R}^d \ni (\theta , x ) \mapsto \mathfrak{P} ( x ) \in \mathbb{R} }} \right] \in \mathscr{A}_{ \mathfrak{d} , d} \end{equation} | (4.28) |
(cf. Definition 4.6). In addition, observe that the fact that for all \theta \in \mathbb{R}^ \mathfrak{d} , x \in \mathbb{R}^d we have that
\begin{equation} \begin{split} \mathscr{N} ^{ {\theta} } ( x ) & = \mathfrak{c}^{{\theta}} + \sum\limits_{i = 1}^ H \mathfrak{v}^{{\theta}}_i \max \{\sum\limits_{\ell = 1}^d \mathfrak{w}^{{\theta}}_{i , \ell} x_\ell + \mathfrak{b}^{{\theta}}_i , 0\} \\ & = \mathfrak{c}^{{\theta}} + \sum\limits_{i = 1}^ H \mathfrak{v}^{{\theta}}_i \left( {\sum\nolimits_{\ell = 1}^d \mathfrak{w}^{{\theta}}_{i , \ell} x_\ell + \mathfrak{b}^{{\theta}}_i} \right) \mathbf{1}_{\smash{{[0, \infty ) }}} \left( {\sum\nolimits_{\ell = 1}^d \mathfrak{w}^{{\theta}}_{i , \ell} x_\ell + \mathfrak{b}^{{\theta}}_i } \right) \end{split} \end{equation} | (4.29) |
demonstrates that
\begin{equation} \left[ {{ \mathbb{R}^ \mathfrak{d} \times \mathbb{R}^d \ni (\theta , x ) \mapsto \mathscr{N} ^{ {\theta} } ( x ) \in \mathbb{R} }} \right] \in \mathscr{A}_{ \mathfrak{d} , d} . \end{equation} | (4.30) |
Combining this with (4.28) and the fact that \mathscr{A}_{ \mathfrak{d}, d} is an algebra proves that
\begin{equation} \left[ {{ \mathbb{R}^ \mathfrak{d} \times \mathbb{R}^d \ni (\theta , x ) \mapsto ( \mathscr{N} ^{ {\theta} } ( x ) - F ( x ) ) ^2 \mathfrak{P} ( x ) \in \mathbb{R} }} \right] \in \mathscr{A}_{ \mathfrak{d} , d } . \end{equation} | (4.31) |
This, Proposition 4.8, and induction demonstrate that
\begin{equation} \left[ {{ \mathbb{R}^ \mathfrak{d} \ni \theta \mapsto \int_ \mathscr{a}^ \mathscr{b} \int_ \mathscr{a}^ \mathscr{b} \cdots \int_ \mathscr{a}^ \mathscr{b} ( \mathscr{N} ^{ {\theta} } ( x ) - F ( x ) ) ^2 \mathfrak{P} ( x ) \, \mathrm{d} x_d \cdots \, \mathrm{d} x_2 \, \mathrm{d} x_1 \in \mathbb{R} }} \right] \in \mathscr{A}_{ \mathfrak{d} , 0}. \end{equation} | (4.32) |
Fubini's theorem hence implies that \mathcal{L} \in \mathscr{A}_{ \mathfrak{d}, 0 } . Combining this and Lemma 4.7 shows that \mathcal{L} is semialgebraic. The proof of Corollary 4.10 is thus complete.
In this section we employ the findings from Sections 2 and 4 to establish in Proposition 5.2 in Subsection 5.2 below, in Proposition 5.3 in Subsection 5.2, and in Theorem 5.4 in Subsection 5.3 below several convergence rate results for solutions of GF differential equations. Theorem 1.2 in the introduction is a direct consequence of Theorem 5.4. Our proof of Theorem 5.4 is based on an application of Proposition 5.3 and our proof of Proposition 5.3 uses Proposition 5.2. Our proof of Proposition 5.2, in turn, employs Proposition 5.1 in Subsection 5.1 below. In Proposition 5.1 we establish that under the assumption that the target function f \colon [\mathscr{a}, \mathscr{b}] ^d \to \mathbb{R} and the unnormalized density function \mathfrak{p} \colon [\mathscr{a}, \mathscr{b}] ^d \to [0, \infty) are piecewise polynomial (see Definition 4.9 in Subsection 4.3) we have that the risk function \mathcal{L} \colon \mathbb{R}^ \mathfrak{d} \to \mathbb{R} satisfies an appropriately generalized Kurdyka-Łojasiewicz inequality.
In the proof of Proposition 5.1 the classical Łojasiewicz inequality for semialgebraic or subanalytic functions (cf., e.g., Bierstone & Milman [40]) is not directly applicable since the generalized gradient function \mathcal{G} \colon \mathbb{R}^{ \mathfrak{d} } \to \mathbb{R}^{ \mathfrak{d} } is not continuous. We will employ the more general results from Bolte et al. [9] which also apply to not necessarily continuously differentiable functions.
The arguments used in the proof of Proposition 5.2 are slight adaptions of well-known arguments in the literature; see, e.g., Kurdyka et al. [12, Section 1], Bolte et al. [9, Theorem 4.5], or Absil et al. [6, Theorem 2.2]. On the one hand, in Kurdyka et al. [12, Section 1] and Absil et al. [6, Theorem 2.2] it is assumed that the object function of the considered optimization problem is analytic and in Bolte et al. [9, Theorem 4.5] it is assumed that the objective function of the considered optimization problem is convex or lower C^2 and Proposition 5.2 does not require these assumptions. On the other hand, Bolte et al. [9, Theorem 4.5] consider more general differential dynamics and the considered gradients are allowed to be more general than the specific generalized gradient function \mathcal{G} \colon \mathbb{R}^{ \mathfrak{d} } \to \mathbb{R}^{ \mathfrak{d} } which is considered in Proposition 5.2.
Proposition 5.1 (Generalized Kurdyka-Łojasiewicz inequality). Assume Setting 2.1, assume that \mathfrak{p} and f are piecewise polynomial, and let \vartheta \in \mathbb{R}^ \mathfrak{d} (cf. Definition 4.9). Then there exist \varepsilon, \mathfrak{D} \in (0, \infty) , \alpha \in (0, 1) such that for all \theta \in B_\varepsilon (\vartheta) it holds that
\begin{equation} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^\alpha \leq \mathfrak{D} || \mathcal{G} ( \theta ) ||. \end{equation} | (5.1) |
Proof of Proposition 5.1. Throughout this proof let \mathbf{M} \colon \mathbb{R}^ \mathfrak{d} \to [0, \infty] satisfy for all \theta \in \mathbb{R}^ \mathfrak{d} that
\begin{equation} \mathbf{M}( \theta ) = \inf \left( {{ \left\{ {{||h|| \colon h \in \partial \mathcal{L} ( \theta ) }} \right\} \cup \left\{ {{ \infty }} \right\} }} \right). \end{equation} | (5.2) |
Note that Proposition 2.12 implies for all \theta \in \mathbb{R}^ \mathfrak{d} that
\begin{equation} \mathbf{M}( \theta ) \leq || \mathcal{G} ( \theta ) || . \end{equation} | (5.3) |
Furthermore, observe that Corollary 4.10, the fact that semialgebraic functions are subanalytic, and Bolte et al. [9, Theorem 3.1 and Remark 3.2] ensure that there exist \varepsilon, \mathfrak{D} \in (0, \infty) , \mathfrak{a} \in [0, 1) which satisfy for all \theta \in B_\varepsilon (\vartheta) that
\begin{equation} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^ \mathfrak{a} \leq \mathfrak{D} \mathbf{M} ( \theta ). \end{equation} | (5.4) |
Combining this and (5.3) with the fact that \sup_{\theta \in B_\varepsilon (\vartheta) } | \mathcal{L} (\theta) - \mathcal{L} (\vartheta) | < \infty demonstrates that for all \theta \in B_\varepsilon (\vartheta) , \alpha \in (\mathfrak{a}, 1) we have that
\begin{equation} \begin{split} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^{\alpha} & \le | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^ \mathfrak{a} \left( {{ \sup\nolimits_{\psi \in B_\varepsilon ( \vartheta ) } | \mathcal{L} ( \psi ) - \mathcal{L} ( \vartheta ) | ^{\alpha - \mathfrak{a}} }} \right) \\ &\le \left( {{ \mathfrak{D} \sup\nolimits_{\psi \in B_\varepsilon ( \vartheta ) } | \mathcal{L} ( \psi ) - \mathcal{L} ( \vartheta ) | ^{\alpha - \mathfrak{a}} }} \right) || \mathcal{G} ( \theta ) ||. \end{split} \end{equation} | (5.5) |
This completes the proof of Proposition 5.1.
Proposition 5.2. Assume Setting 2.1 and let \vartheta \in \mathbb{R}^ \mathfrak{d} , \varepsilon, \mathfrak{D} \in (0, \infty) , \alpha \in (0, 1) satisfy for all \theta \in B_\varepsilon (\vartheta) that
\begin{equation} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^\alpha \leq \mathfrak{D} || \mathcal{G}( \theta ) || . \end{equation} | (5.6) |
Then there exists \delta \in (0, \varepsilon) such that for all \Theta \in C([0, \infty), \mathbb{R}^ \mathfrak{d}) with \Theta_0 \in B_\delta (\vartheta) , \forall \, t \in [0, \infty) \colon \Theta_t = \Theta_0 - \int_0^t \mathcal{G} (\Theta_s) \, \mathrm{d} s , and \inf_{t \in \left\{ {{ s \in [0, \infty) \colon \Theta_s \in B_\varepsilon (\vartheta) }} \right\} } \mathcal{L} (\Theta_t) \geq \mathcal{L} (\vartheta) there exists \psi \in \mathcal{L}^{ - 1 }(\left\{ {{ \mathcal{L}(\vartheta) }} \right\}) such that for all t \in [0, \infty) it holds that
\begin{equation} \Theta_t \in B_{ \varepsilon }( \vartheta ) , \qquad \int_0^\infty || \mathcal{G} ( \Theta_s ) || \, \mathrm{d} s \leq \varepsilon , \qquad | \mathcal{L}( \Theta_t ) - \mathcal{L}( \psi ) | \leq ( 1 + \mathfrak{D}^{ - 2 } t )^{ - 1 } , \end{equation} | (5.7) |
\begin{equation} \mathit{\text{and}} \qquad ||\Theta_t - \psi || \leq \left[ {{ 1 + \left( {{ \mathfrak{D}^{-1 / \alpha} ( 1 - \alpha ) }} \right)^{\frac{\alpha}{1 - \alpha } } t }} \right]^ { - \min \left\{ {{1, \frac{1 - \alpha}{ \alpha } }} \right\} } . \end{equation} | (5.8) |
Proof of Proposition 5.2. Note that the fact that \mathcal{L} is continuous implies that there exists \delta \in (0, \varepsilon / 3) which satisfies for all \theta \in B_\delta (\vartheta) that
\begin{equation} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) | ^{1 - \alpha } \leq \min \left\{ {{ \frac{\varepsilon ( 1 - \alpha ) }{3 \mathfrak{D} }, \frac{1 - \alpha}{ \mathfrak{D}} , 1 }} \right\}. \end{equation} | (5.9) |
In the following let \Theta \in C([0, \infty), \mathbb{R}^ \mathfrak{d}) satisfy \forall \, t \in [0, \infty) \colon \Theta_t = \Theta_0 - \int_0^t \mathcal{G} (\Theta_s) \, \mathrm{d} s , \Theta_0 \in B_\delta (\vartheta) , and
\begin{equation} \inf\nolimits_{t \in \left\{ {{ s \in [0, \infty ) \colon \Theta_s \in B_\varepsilon ( \vartheta ) }} \right\} } \mathcal{L} ( \Theta_t ) \geq \mathcal{L} ( \vartheta ). \end{equation} | (5.10) |
In the first step we show that for all t \in [0, \infty) it holds that
\begin{equation} \Theta_t \in B_\varepsilon ( \vartheta ). \end{equation} | (5.11) |
Observe that, e.g., [37, Lemma 3.1] ensures for all t \in [0, \infty) that
\begin{equation} \mathcal{L} ( \Theta_t ) = \mathcal{L} ( \Theta_0 ) - \int_0^t || \mathcal{G} ( \Theta_s ) || ^2 \, \mathrm{d} s. \end{equation} | (5.12) |
This implies that [0, \infty) \ni t \mapsto \mathcal{L} (\Theta_t) \in [0, \infty) is non-increasing. Next let L \colon [0, \infty) \to \mathbb{R} satisfy for all t \in [0, \infty) that
\begin{equation} L(t) = \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) \end{equation} | (5.13) |
and let T \in [0, \infty] satisfy
\begin{equation} T = \inf \left( {{ \left\{ {{t \in [0, \infty ) \colon ||\Theta_t - \vartheta || \geq \varepsilon }} \right\} \cup \left\{ {{\infty}} \right\} }} \right). \end{equation} | (5.14) |
We intend to show that T = \infty . Note that (5.10) assures for all t \in [0, T) that L(t) \geq 0 . Moreover, observe that (5.12) and (5.13) ensure that for almost all t \in [0, T) it holds that L is differentiable at t and satisfies L ' (t) = \frac{ \mathrm{d}}{ \mathrm{d} t} (\mathcal{L} (\Theta_t)) = - || \mathcal{G} (\Theta_t) || ^2 . In the following let \tau \in [0, T] satisfy
\begin{equation} \tau = \inf \left( {{ \left\{ {{t \in [0, T) \colon L ( t ) = 0 }} \right\} \cup \left\{ {{T }} \right\} }} \right). \end{equation} | (5.15) |
Note that the fact that L is non-increasing implies that for all s \in [\tau, T) it holds that L(s) = 0 . Combining this with (5.12) demonstrates for almost all s \in (\tau, T) that \mathcal{G} (\Theta_s) = 0 . This proves for all s \in [\tau, T) that \Theta_s = \Theta_\tau . Next observe that (5.6) ensures that for all t \in [0, \tau) it holds that
\begin{equation} 0 < [ L ( t ) ] ^\alpha = | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | ^\alpha \leq \mathfrak{D} || \mathcal{G} ( \Theta_t ) ||. \end{equation} | (5.16) |
Combining this with the chain rule proves for almost all t \in [0, \tau) that
\begin{equation} \begin{split} \frac{ \mathrm{d} }{ \mathrm{d} t} ([ L ( t ) ]^{1 - \alpha } ) & = (1-\alpha) [ L ( t ) ]^{-\alpha} \left( {{ - || \mathcal{G}(\Theta_t)||^2 }} \right) \\ &\leq - ( 1 - \alpha ) \mathfrak{D}^{-1} || \mathcal{G} ( \Theta_t ) ||^{-1} || \mathcal{G} ( \Theta_t ) || ^2 = - \mathfrak{D}^{-1} (1 - \alpha ) || \mathcal{G}(\Theta_t)||. \end{split} \end{equation} | (5.17) |
In addition, note that the fact that [0, \infty) \ni t \mapsto L(t) \in \mathbb{R} is absolutely continuous and the fact that for all r \in (0, \infty) it holds that (r, \infty) \ni y \mapsto y^{1 - \alpha } \in \mathbb{R} is Lipschitz continuous demonstrate for all t \in [0, \tau) that [0, t] \ni s \mapsto [L (s)]^{ 1 - \alpha } \in \mathbb{R} is absolutely continuous. Integrating (5.17) hence shows for all s, t \in [0, \tau) with t \le s that
\begin{equation} \int _t ^s || \mathcal{G}(\Theta_u ) || \, \mathrm{d} u \leq - \mathfrak{D} \left( {{1 - \alpha }} \right) ^{-1} \left( {{ [ L ( s ) ] ^{1 - \alpha } - [ L ( t ) ] ^{ 1 - \alpha } }} \right) \leq \mathfrak{D} \left( {{1 - \alpha}} \right)^{-1} [ L ( t ) ]^{1 - \alpha} . \end{equation} | (5.18) |
This and the fact that for almost all s \in (\tau, T) it holds that \mathcal{G} (\Theta_s) = 0 ensure that for all s, t \in [0, T) with t \le s we have that
\begin{equation} \int_t^s || \mathcal{G}(\Theta_u ) || \, \mathrm{d} u \leq \mathfrak{D} \left( {{ 1 - \alpha}} \right)^{-1} [ L ( t ) ]^{1 - \alpha} . \end{equation} | (5.19) |
Combining this with (5.9) demonstrates for all t \in [0, T) that
\begin{equation} ||\Theta_t - \Theta_{0} || = \left\| {{\int_{0}^t \mathcal{G} ( \Theta_s ) \, \mathrm{d} s}} \right\| \leq \int _{0} ^t || \mathcal{G}(\Theta_s ) || \, \mathrm{d} s \leq \frac{ \mathfrak{D} | \mathcal{L} ( \Theta_0 ) - \mathcal{L} ( \vartheta ) |^{1-\alpha} }{1 - \alpha} \leq \min \left\{ {{ \frac{\varepsilon}{3} , 1 }} \right\}. \end{equation} | (5.20) |
This, the fact that \delta < \varepsilon / 3 , and the triangle inequality assure for all t \in [0, T) that
\begin{equation} ||\Theta_t - \vartheta || \leq ||\Theta_t - \Theta_{0} || + ||\Theta_{0} - \vartheta|| \leq \frac{\varepsilon}{3} + \delta \leq \frac{\varepsilon}{3} + \frac{\varepsilon}{3} = \frac{2 \varepsilon }{3 }. \end{equation} | (5.21) |
Combining this with (5.14) proves that T = \infty . This establishes (5.11).
Next observe that the fact that T = \infty and (5.20) prove that
\begin{equation} \int_0^\infty || \mathcal{G} ( \Theta_s ) || \, \mathrm{d} s \leq \min \left\{ {{ \frac{\varepsilon}{3} , 1 }} \right\} \le \varepsilon < \infty. \end{equation} | (5.22) |
In the following let \sigma \colon [0, \infty) \to [0, \infty) satisfy for all t \in [0, \infty) that
\begin{equation} \sigma ( t ) = \int_t^\infty || \mathcal{G} ( \Theta_s )|| \, \mathrm{d} s. \end{equation} | (5.23) |
Note that (5.22) proves that \limsup_{t \to \infty} \sigma (t) = 0 . In addition, observe that (5.22) assures that there exists \psi \in \mathbb{R}^ \mathfrak{d} such that
\begin{equation} \limsup \nolimits_{t \to \infty} ||\Theta_t - \psi || = 0. \end{equation} | (5.24) |
In the next step we combine the weak chain rule for the risk function in (5.12) with (5.11) and (5.6) to obtain that for almost all t \in [0, \infty) we have that
\begin{equation} L ' ( t ) = - || \mathcal{G} ( \Theta_t ) || ^2 \leq - \mathfrak{D}^{-2} [ L ( t ) ] ^{ 2 \alpha}. \end{equation} | (5.25) |
In addition, note that the fact that L is non-increasing and (5.9) ensure that for all t \in [0, \infty) it holds that L (t) \leq L (0) \leq 1 . Therefore, we get for almost all t \in [0, \infty) that
\begin{equation} L ' ( t ) \leq - \mathfrak{D}^{-2} [ L ( t ) ] ^{ 2 }. \end{equation} | (5.26) |
Combining this with the fact that for all t \in [0, \tau) it holds that L (t) > 0 establishes for almost all t \in [0, \tau) that
\begin{equation} \frac{ \mathrm{d}}{ \mathrm{d} t} \left( {{ \frac{ \mathfrak{D}^2}{L ( t ) } }} \right) = - \frac{ \mathfrak{D}^2 L ' ( t )}{[ L ( t ) ] ^2} \geq 1. \end{equation} | (5.27) |
The fact that for all t \in [0, \tau) it holds that [0, t] \ni s \mapsto L (s) \in (0, \infty) is absolutely continuous hence demonstrates for all t \in [0, \tau) that
\begin{equation} \frac{ \mathfrak{D}^2}{L(t)} \geq \frac{ \mathfrak{D}^2}{L ( 0 )} + t \geq \mathfrak{D}^2 + t. \end{equation} | (5.28) |
Therefore, we infer for all t \in [0, \tau) that
\begin{equation} L ( t ) \leq \mathfrak{D}^2 \left( {{ \mathfrak{D}^2 + t }} \right)^{-1} = \left( {{ 1 + \mathfrak{D}^{-2}t }} \right)^{-1}. \end{equation} | (5.29) |
This and the fact that for all t \in [\tau, \infty) it holds that L(t) = 0 prove that for all t \in [0, \infty) we have that
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | = L ( t ) \leq \left( {{ 1 + \mathfrak{D}^{-2}t }} \right)^{-1}. \end{equation} | (5.30) |
Furthermore, observe that (5.24) and the fact that \mathcal{L} is continuous imply that \limsup_{t \to \infty} | \mathcal{L} (\Theta_t) - \mathcal{L} (\psi) | = 0 . Hence, we obtain that \mathcal{L} (\psi) = \mathcal{L} (\vartheta) . This shows for all t \in [0, \infty) that
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \psi ) | \leq \left( {{ 1 + \mathfrak{D}^{-2}t }} \right)^{-1}. \end{equation} | (5.31) |
In the next step we establish a convergence rate for the quantity ||\Theta_t - \psi|| , t \in [0, \infty) . We accomplish this by employing an upper bound for the tail length of the curve \Theta_t \in \mathbb{R}^ \mathfrak{d} , t \in [0, \infty) . More formally, note that (5.19), (5.11), and (5.6) demonstrate for all t \in [0, \infty) that
\begin{equation} \begin{split} \sigma(t) & = \int_t^\infty || \mathcal{G} ( \Theta_u )|| \, \mathrm{d} u = \lim\limits_{s \to \infty} \left[ {{ \int_t^s || \mathcal{G} ( \Theta_u ) || \, \mathrm{d} u }} \right] \\ & \leq \mathfrak{D} \left( {{1 - \alpha}} \right)^{-1} \left( L ( t ) \right)^{1-\alpha} \leq \mathfrak{D} \left( {{1-\alpha}} \right)^{-1} \left( {{ \mathfrak{D} || \mathcal{G}(\Theta_t) || }} \right)^{\frac{1-\alpha}{\alpha}}. \end{split} \end{equation} | (5.32) |
Next observe that the fact that for all t \in [0, \infty) it holds that \sigma (t) = \int_0^\infty || \mathcal{G} (\Theta_s)|| \, \mathrm{d} s - \int_0^t || \mathcal{G} (\Theta_s) || \, \mathrm{d} s shows that for almost all t \in [0, \infty) we have that \sigma' (t) = - || \mathcal{G} (\Theta_t) || . This and (5.32) yield for almost all t \in [0, \infty) that \sigma(t) \leq \mathfrak{D}^{1 / \alpha} \left({{1-\alpha}} \right)^{-1} \left[{{- \sigma ' (t) }} \right]^{\frac{1-\alpha}{\alpha}} . Therefore, we obtain for almost all t \in [0, \infty) that
\begin{equation} \sigma ' ( t ) \leq - \left[ {{ ( 1-\alpha ) \mathfrak{D}^{-1 / \alpha} \sigma ( t ) }} \right]^{\frac{\alpha}{1-\alpha}}. \end{equation} | (5.33) |
Combining this with the fact that \sigma is absolutely continuous implies for all t \in [0, \infty) that
\begin{equation} \sigma(t) - \sigma(0) \leq - \left[ {{ ( 1-\alpha ) \mathfrak{D}^{-1 / \alpha} }} \right]^{\frac{\alpha}{1-\alpha}} \int _{0}^t [ \sigma ( s ) ]^{\frac{\alpha}{1-\alpha}} \, \mathrm{d} s. \end{equation} | (5.34) |
In the following let \beta, \mathfrak{C} \in (0, \infty) satisfy \beta = \max \left\{ {{ 1, \frac{\alpha}{1-\alpha} }} \right\} and \mathfrak{C} = \left({{ (1-\alpha) \mathfrak{D} ^{ -1 / \alpha} }} \right)^{\frac{\alpha}{1-\alpha}} . Note that (5.34) and the fact that for all t \in [0, \infty) it holds that \sigma (t) \leq \sigma (0) \leq 1 ensure that for all t \in [0, \infty) it holds that
\begin{equation} \sigma(t) \leq \sigma(0) - \mathfrak{C} \int _{0}^t [ \sigma ( s ) ]^\beta \, \mathrm{d} s. \end{equation} | (5.35) |
This, the fact that \sigma is non-increasing, and the fact that for all t \in [0, \infty) it holds that 0 \leq \sigma(t) \leq 1 prove that for all t \in [0, \infty) we have that
\begin{equation} \left( \sigma ( t ) \right)^\beta \le \sigma ( t ) \leq \sigma(0) - \mathfrak{C} [ \sigma ( t ) ]^\beta t \leq 1 - \mathfrak{C} t [ \sigma ( t ) ]^\beta . \end{equation} | (5.36) |
Hence, we obtain for all t \in [0, \infty) that \sigma(t) \leq \left({{ 1 + \mathfrak{C} t }} \right)^{-\frac{1}{\beta}} . Combining this with the fact that for all t \in [0, \infty) it holds that
\begin{equation} \begin{split} ||\Theta_t - \psi|| &\leq \limsup\limits_{s \to \infty} ||\Theta_t - \Theta_s || = \limsup\limits_{s \to \infty} \left\| {{ \int_t^s \mathcal{G} ( \Theta_u ) \, \mathrm{d} u }} \right\| \leq \limsup\limits_{s \to \infty} \left[ {{ \int_t^s || \mathcal{G} ( \Theta_u ) || \, \mathrm{d} u }} \right] \\ & = \int_t^\infty || \mathcal{G} ( \Theta_u ) || \, \mathrm{d} u = \sigma ( t ) \end{split} \end{equation} | (5.37) |
shows that for all t \in [0, \infty) we have that ||\Theta_t - \psi || \le (1 + \mathfrak{C} t) ^{- 1 / \beta} . This, (5.11), (5.22), and (5.31) establish (5.8). The proof of Proposition 5.2 is thus complete.
Proposition 5.3. Assume Setting 2.1, assume that \mathfrak{p} and f are piecewise polynomial, and let \Theta \in C ([0, \infty), \mathbb{R}^ \mathfrak{d}) satisfy
\begin{equation} \liminf\limits_{t \to \infty } ||\Theta_t || < \infty \qquad\mathit{\text{and}}\qquad \forall \, t \in [0, \infty ) \colon \Theta_t = \Theta_0 - \int_0^t \mathcal{G} ( \Theta_s ) \, \mathrm{d} s \end{equation} | (5.38) |
(cf. Definition 4.9). Then there exist \vartheta \in \mathcal{G}^{ - 1 } (\left\{ {{ 0 }} \right\}) , \mathfrak{C}, \tau, \beta \in (0, \infty) which satisfy for all t \in [\tau, \infty) that
\begin{equation} ||\Theta_t - \vartheta|| \leq \left( {{ 1 + \mathfrak{C} ( t - \tau ) }} \right)^{- \beta} \qquad\mathit{\text{and}}\qquad | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | \leq \left( {{1 + \mathfrak{C} ( t - \tau ) }} \right) ^{-1}. \end{equation} | (5.39) |
Proof of Proposition 5.3. First observe that [37, Lemma 3.1] ensures that for all t \in [0, \infty) it holds that
\begin{equation} \mathcal{L}( \Theta_t ) = \mathcal{L}( \Theta_0 ) - \int_0^t || \mathcal{G} ( \Theta_s ) || ^2 \, \mathrm{d} s . \end{equation} | (5.40) |
This implies that [0, \infty) \ni t \mapsto \mathcal{L} (\Theta_t) \in [0, \infty) is non-increasing. Hence, we obtain that there exists \mathbf{m} \in [0, \infty) which satisfies that
\begin{equation} \mathbf{m} = \limsup\nolimits_{t \to \infty} \mathcal{L} ( \Theta_t ) = \liminf\nolimits_{t \to \infty} \mathcal{L} ( \Theta_t ) = \inf\nolimits_{t \in [0, \infty )} \mathcal{L} ( \Theta_t ). \end{equation} | (5.41) |
Moreover, note that the assumption that \liminf_{t \to \infty } ||\Theta_t || < \infty ensures that there exist \vartheta \in \mathbb{R}^ \mathfrak{d} and \tau = (\tau_n)_{n \in \mathbb{N}} \colon \mathbb{N} \to [0, \infty) which satisfy \liminf_{n \to \infty} \tau_n = \infty and
\begin{equation} \limsup\nolimits_{n \to \infty} ||\Theta_{\tau_n} - \vartheta || = 0. \end{equation} | (5.42) |
Combining this with (5.41) and the fact that \mathcal{L} is continuous shows that
\begin{equation} \mathcal{L} ( \vartheta ) = \mathbf{m} \qquad\text{and}\qquad \forall \, t \in [0, \infty ) \colon \mathcal{L} ( \Theta_t ) \geq \mathcal{L} ( \vartheta ). \end{equation} | (5.43) |
Next observe that Proposition 5.1 demonstrates that there exist \varepsilon, \mathfrak{D} \in (0, \infty) , \alpha \in (0, 1) such that for all \theta \in B_\varepsilon (\vartheta) we have that
\begin{equation} | \mathcal{L} ( \theta ) - \mathcal{L} ( \vartheta ) |^\alpha \leq \mathfrak{D} || \mathcal{G} ( \theta ) || . \end{equation} | (5.44) |
Combining this and (5.42) with Proposition 5.2 demonstrates that there exists \delta \in (0, \varepsilon) which satisfies for all \Phi \in C([0, \infty), \mathbb{R}^ \mathfrak{d}) with \Phi_0 \in B_\delta (\vartheta) , \forall \, t \in [0, \infty) \colon \Phi_t = \Phi_0 - \int_0^t \mathcal{G} (\Phi_s) \, \mathrm{d} s , and \inf_{ t \in \left\{ {{ s \in [0, \infty) \colon \Phi_s \in B_{ \varepsilon }(\vartheta) }} \right\} } \mathcal{L}(\Phi_t) \ge \mathcal{L}(\vartheta) that it holds for all t \in [0, \infty) that
\begin{equation} \Phi_t \in B_{ \varepsilon }( \vartheta ) , \qquad | \mathcal{L}( \Phi_t ) - \mathcal{L}( \vartheta ) | \leq ( 1 + \mathfrak{D}^{ - 2 } t )^{ - 1 } , \end{equation} | (5.45) |
\begin{equation} \text{and} \qquad ||\Phi_t - \vartheta || \leq \left[ {{ 1 + \left( {{ \mathfrak{D}^{-1 / \alpha} ( 1 - \alpha ) }} \right)^{\frac{\alpha}{1 - \alpha } } t }} \right]^ { - \min \left\{ {{1, \frac{1 - \alpha}{ \alpha } }} \right\} } . \end{equation} | (5.46) |
Moreover, note that (5.42) ensures that there exists n \in \mathbb{N} which satisfies \Theta_{\tau_n } \in B_\delta (\vartheta) . Next let \Phi \in C([0, \infty), \mathbb{R}^ \mathfrak{d}) satisfy for all t \in [0, \infty) that
\begin{equation} \Phi_t = \Theta_{t + \tau_n }. \end{equation} | (5.47) |
Observe that (5.43) and (5.47) assure that
\begin{equation} \Phi_0 \in B_\delta ( \vartheta ) , \qquad \inf\nolimits_{ t \in [0, \infty) } \mathcal{L}( \Phi_t ) \ge \mathcal{L}( \vartheta ) , \qquad\text{and}\qquad \forall \, t \in [0, \infty ) \colon \Phi_t = \Phi_0 - \int_0^t \mathcal{G} ( \Phi_s ) \, \mathrm{d} s . \end{equation} | (5.48) |
Combining this with (5.46) proves for all t \in [\tau_n, \infty) that
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | \le \left( {{ 1 + \mathfrak{D}^{-2} ( t - \tau_n ) }} \right)^{-1} \end{equation} | (5.49) |
and
\begin{equation} ||\Theta_t - \vartheta || \leq \left[ {{ 1 + \left( {{ \mathfrak{D}^{-1 / \alpha} ( 1 - \alpha ) }} \right)^{\frac{\alpha}{1 - \alpha } } ( t - \tau_n ) }} \right]^ { - \min \left\{ {{1, \frac{1 - \alpha}{ \alpha } }} \right\} } . \end{equation} | (5.50) |
Next note that [37, Corollary 2.15] shows that \mathbb{R}^ \mathfrak{d} \ni \theta \mapsto || \mathcal{G} (\theta) || \in [0, \infty) is lower semicontinuous. The fact that \liminf_{s \to \infty} || \mathcal{G} (\Theta_s) || = 0 and the fact that \limsup_{t \to \infty} ||\Theta_t - \vartheta || = 0 hence imply that \mathcal{G} (\vartheta) = 0 . Combining this with (5.49) and (5.50) establishes (5.39). The proof of Proposition 5.3 is thus complete.
By choosing a sufficiently large \mathscr{C} \in (0, \infty) we can conclude a simplified version of Proposition 5.3. This is precisely the subject of the next result, Theorem 5.4 below. Theorem 1.2 in the introduction is a direct consequence of Theorem 5.4.
Theorem 5.4. Assume Setting 2.1, assume that \mathfrak{p} and f are piecewise polynomial, and let \Theta \in C ([0, \infty), \mathbb{R}^ \mathfrak{d}) satisfy \liminf_{t \to \infty } ||\Theta_t || < \infty and \forall \, t \in [0, \infty) \colon \Theta_t = \Theta_0 - \int_0^t \mathcal{G} (\Theta_s) \, \mathrm{d} s (cf. Definition 4.9). Then there exist \vartheta \in \mathcal{G}^{ - 1 } (\left\{ {{ 0 }} \right\}) , \mathscr{C}, \beta \in (0, \infty) which satisfy for all t \in [0, \infty) that
\begin{equation} ||\Theta_t - \vartheta|| \leq \mathscr{C} ( 1 + t ) ^{ - \beta } \qquad\mathit{\text{and}}\qquad | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | \leq \mathscr{C} ( 1 + t ) ^{ - 1 } . \end{equation} | (5.51) |
Proof of Theorem 5.4. Observe that Proposition 5.3 assures that there exist \vartheta \in \mathcal{G}^{ - 1 } (\left\{ {{ 0 }} \right\}) , \mathfrak{C}, \tau, \beta \in (0, \infty) which satisfy for all t \in [\tau, \infty) that
\begin{equation} ||\Theta_t - \vartheta|| \leq \left( {{ 1 + \mathfrak{C} ( t - \tau ) }} \right)^{- \beta} \end{equation} | (5.52) |
and
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | \leq \left( {{1 + \mathfrak{C} ( t - \tau ) }} \right) ^{-1}. \end{equation} | (5.53) |
In the following let \mathscr{C} \in (0, \infty) satisfy
\begin{equation} \mathscr{C} = \max \left\{ {{ \mathfrak{C}^{-1} , 1 + \tau , \mathfrak{C}^{- \beta}, (1 + \tau )^\beta , ( 1 + \tau ) ^\beta \left[ {{ \sup\nolimits_{s \in [0, \tau ] } ||\Theta_s - \vartheta|| }} \right] , (1 + \tau ) \mathcal{L} ( \Theta_0 ) }} \right\}. \end{equation} | (5.54) |
Note that (5.53), (5.54), and the fact that [0, \infty) \ni t \mapsto \mathcal{L} (\Theta_t) \in [0, \infty) is non-increasing show for all t \in [0, \tau] that
\begin{equation} ||\Theta_t - \vartheta || \le \sup\nolimits_{s \in [0, \tau ]} ||\Theta_s - \vartheta|| \le \mathscr{C} ( 1 + \tau ) ^{- \beta } \le \mathscr{C} ( 1 + t ) ^{-\beta } \end{equation} | (5.55) |
and
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | = \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta) \le \mathcal{L} ( \Theta_t ) \le \mathcal{L} ( \Theta_0 ) \le \mathscr{C} ( 1 + \tau )^{-1} \le \mathscr{C} ( 1 + t ) ^{-1}. \end{equation} | (5.56) |
Moreover, observe that (5.52) and (5.54) imply for all t \in [\tau, \infty) that
\begin{equation} ||\Theta_t - \vartheta || \leq \mathscr{C} \left( {{ \mathscr{C}^{1 / \beta} + \mathfrak{C} \mathscr{C}^{1 / \beta} ( t - \tau ) }} \right) ^{ - \beta } \le \mathscr{C} \left( {{ \mathscr{C}^{1 / \beta} - \tau + t }} \right)^{-\beta} \le \mathscr{C} (1 + t ) ^{-\beta } . \end{equation} | (5.57) |
In addition, note that (5.53) and (5.54) demonstrate for all t \in [\tau, \infty) that
\begin{equation} | \mathcal{L} ( \Theta_t ) - \mathcal{L} ( \vartheta ) | \le \mathscr{C} \left( {{ \mathscr{C} + \mathfrak{C} \mathscr{C} ( t - \tau ) }} \right) ^{-1} \le \mathscr{C} \left( {{ \mathscr{C} - \tau + t }} \right)^{-1} \le \mathscr{C} ( 1 + t ) ^{-1} . \end{equation} | (5.58) |
This completes the proof of Theorem 5.4.
This project has been partially funded by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany's Excellence Strategy EXC 2044-390685587, Mathematics Münster: Dynamics-Geometry-Structure.
The authors declare that there are no conflicts of interest.
[1] |
E. Aličković, A. Subasi, Breast cancer diagnosis using GA feature selection and Rotation Forest, Neural Comput. Appl., 28 (2015), 753–763. https://doi.org/10.1007/s00521-015-2103-9 doi: 10.1007/s00521-015-2103-9
![]() |
[2] | World Health Organization, Breast cancer 2021, 2021. Available from: https://www.who.int/news-room/fact-sheets/detail/breast-cancer. |
[3] |
Y. S. Sun, Z. Zhao, Z. N. Yang, F. Xu, H. J. Lu, Z. Y. Zhu, et al., Risk factors and preventions of breast cancer, Int. J. Biol. Sci., 13 (2017), 1387–1397. https://doi.org/10.7150/ijbs.21635 doi: 10.7150/ijbs.21635
![]() |
[4] |
J. B. Harford, Breast-cancer early detection in low-income and middle-income countries: Do what you can versus one size fits all, Lancet Oncol., 12 (2011), 306–312. https://doi.org/10.1016/s1470-2045(10)70273-4 doi: 10.1016/s1470-2045(10)70273-4
![]() |
[5] |
C. Lerman, M. Daly, C. Sands, A. Balshem, E. Lustbader, T. Heggan, et al., Mammography adherence and psychological distress among women at risk for breast cancer, J. Natl. Cancer Inst., 85 (1993), 1074–1080. https://doi.org/10.1093/jnci/85.13.1074 doi: 10.1093/jnci/85.13.1074
![]() |
[6] |
P. T. Huynh, A. M. Jarolimek, S. Daye, The false-negative mammogram, Radiographics, 18 (1998), 1137–1154. https://doi.org/10.1148/radiographics.18.5.9747612 doi: 10.1148/radiographics.18.5.9747612
![]() |
[7] | M. G. Ertosun, D. L. Rubin, Probabilistic Visual Search for Masses within mammography images using Deep Learning, in 2015 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), (2015), 1310–1315. https://doi.org/10.1109/bibm.2015.7359868 |
[8] | Y. Lu, J. Y. Li, Y. T. Su, A. A. Liu, A review of breast cancer detection in medical images, in 2018 IEEE Visual Communications and Image Processing, (2018), 1–4. https://doi.org/10.1109/vcip.2018.8698732 |
[9] |
J. Ferlay, I. Soerjomataram, R. Dikshit, S. Eser, C. Mathers, M. Rebelo, et al., Cancer incidence and mortality worldwide: Sources, methods and major patterns in Globocan 2012, Int. J. Cancer, 136 (2014), E359–E386. https://doi.org/10.1002/ijc.29210 doi: 10.1002/ijc.29210
![]() |
[10] |
N. Mao, P. Yin, Q. Wang, M. Liu, J. Dong, X. Zhang, et al., Added value of Radiomics on mammography for breast cancer diagnosis: A feasibility study, J. Am. Coll. Radiol., 16 (2019), 485–491. https://doi.org/10.1016/j.jacr.2018.09.041 doi: 10.1016/j.jacr.2018.09.041
![]() |
[11] |
H. Wang, J. Feng, Q. Bu, F. Liu, M. Zhang, Y. Ren, et al., Breast mass detection in digital mammogram based on Gestalt Psychology, J. Healthc. Eng., 2018 (2018), 1–13. https://doi.org/10.1155/2018/4015613 doi: 10.1155/2018/4015613
![]() |
[12] |
S. McGuire, World cancer report 2014, Switzerland: World Health Organization, international agency for research on cancer, Adv. Nutrit. Int. Rev., 7 (2016), 418–419. https://doi.org/10.3945/an.116.012211 doi: 10.3945/an.116.012211
![]() |
[13] |
M. K. Gupta, P. Chandra, A comprehensive survey of Data Mining, Int. J. Comput. Technol., 12 (2020), 1243–1257. https://doi.org/10.1007/s41870-020-00427-7 doi: 10.1007/s41870-020-00427-7
![]() |
[14] | T. Zou, T. Sugihara, Fast identification of a human skeleton-marker model for motion capture system using stochastic gradient descent method, in 2020 8th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics (BioRob)., (2020), 181–186. https://doi.org/10.1109/biorob49111.2020.9224442 |
[15] |
A. Reisizadeh, A. Mokhtari, H. Hassani, R. Pedarsani, An exact quantized decentralized gradient descent algorithm, IEEE Trans. Signal Process., 67 (2019), 4934–4947. https://doi.org/10.1109/tsp.2019.2932876 doi: 10.1109/tsp.2019.2932876
![]() |
[16] |
D. Maulud, A. M. Abdulazeez, A review on linear regression comprehensive in machine learning, J. Appl. Sci. Technol. Trends, 1 (2020), 140–147. https://doi.org/10.38094/jastt1457 doi: 10.38094/jastt1457
![]() |
[17] |
D. R. Wilson, T. R. Martinez, The general inefficiency of batch training for gradient descent learning, Neural Networks, 16 (2003) 1429–1451. https://doi.org/10.1016/s0893-6080(03)00138-2 doi: 10.1016/s0893-6080(03)00138-2
![]() |
[18] |
D. Yi, S. Ji, S. Bu, An enhanced optimization scheme based on gradient descent methods for machine learning, Symmetry, 11 (2019), 942. https://doi.org/10.3390/sym11070942 doi: 10.3390/sym11070942
![]() |
[19] |
D. A. Zebari, D. Q. Zeebaree, A. M. Abdulazeez, H. Haron, H. N. Hamed, Improved threshold based and trainable fully automated segmentation for breast cancer boundary and pectoral muscle in mammogram images, IEEE Access, 8 (2020), 203097–203116. https://doi.org/10.1109/access.2020.3036072 doi: 10.1109/access.2020.3036072
![]() |
[20] | D. Q. Zeebaree, H. Haron, A. M. Abdulazeez, D. A. Zebari, Trainable model based on new uniform LBP feature to identify the risk of the breast cancer, in 2019 International Conference on Advanced Science and Engineering (ICOASE), 2019. https://doi.org/10.1109/icoase.2019.8723827 |
[21] |
D. Q. Zeebaree, A. M. Abdulazeez, L. M. Abdullrhman, D. A. Hasan, O. S. Kareem, The prediction process based on deep recurrent neural networks: A Review, Asian J. Comput. Inf. Syst., 10 (2021), 29–45. https://doi.org/10.9734/ajrcos/2021/v11i230259 doi: 10.9734/ajrcos/2021/v11i230259
![]() |
[22] |
D. Q. Zeebaree, A. M. Abdulazeez, D. A. Zebari, H. Haron, H. N. A. Hamed, Multi-level fusion in ultrasound for cancer detection based on uniform LBP features, Comput. Matern. Contin., 66 (2021), 3363–3382. https://doi.org/10.32604/cmc.2021.013314 doi: 10.32604/cmc.2021.013314
![]() |
[23] |
M. Muhammad, D. Zeebaree, A. M. Brifcani, J. Saeed, D. A. Zebari, A review on region of interest segmentation based on clustering techniques for breast cancer ultrasound images, J. Appl. Sci. Technol. Trends, 1 (2020), 78–91. https://doi.org/10.38094/jastt1328 doi: 10.38094/jastt1328
![]() |
[24] |
P. Kamsing, P. Torteeka, S. Yooyen, An enhanced learning algorithm with a particle filter-based gradient descent optimizer method, Neural Comput. Appl., 32 (2020), 12789–12800. https://doi.org/10.1007/s00521-020-04726-9 doi: 10.1007/s00521-020-04726-9
![]() |
[25] |
Y. Hamid, L. Journaux, J. A. Lee, M. Sugumaran, A novel method for network intrusion detection based on nonlinear SNE and SVM, J. Artif. Intell. Soft Comput. Res., 6 (2018), 265. https://doi.org/10.1504/ijaisc.2018.097280 doi: 10.1504/ijaisc.2018.097280
![]() |
[26] | H. Sadeeq, A. M. Abdulazeez, Hardware implementation of firefly optimization algorithm using fpgas, in 2018 International Conference on Advanced Science and Engineering, (2018), 30–35. https://doi.org/10.1109/icoase.2018.8548822 |
[27] | D. P. Hapsari, I. Utoyo, S. W. Purnami, Fractional gradient descent optimizer for linear classifier support vector machine, in 2020 Third International Conference on Vocational Education and Electrical Engineering (ICVEE), (2020), 1–5. |
[28] |
M. S. Nawaz, B. Shoaib, M. A. Ashraf, Intelligent cardiovascular disease prediction empowered with gradient descent optimization, Heliyon, 7 (2021), 1–10. https://doi.org/10.1016/j.heliyon.2021.e06948 doi: 10.1016/j.heliyon.2021.e06948
![]() |
[29] |
Y. Qian, Exploration of machine algorithms based on deep learning model and feature extraction, J. Math. Biosci. Eng., 18 (2021), 7602–7618. https://doi.org/10.3934/mbe.2021376 doi: 10.3934/mbe.2021376
![]() |
[30] |
Z. Wang, M. Li, H. Wang, H. Jiang, Y. Yao, H. Zhang, et al., Breast cancer detection using extreme learning machine based on feature fusion with CNN deep features, IEEE Access, 7 (2019), 105146–105158. https://doi.org/10.1109/access.2019.2892795 doi: 10.1109/access.2019.2892795
![]() |
[31] | UCI Machine Learning Repository, Breast Cancer Wisconsin (Diagnostic) Data Set. Available from: https://archive.ics.uci.edu/ml/datasets/Breast+ Cancer + Wisconsin + (Diagnostic). |
[32] |
R. V. Anji, B. Soni, R. K. Sudheer, Breast cancer detection by leveraging machine learning, ICT Express, 6 (2020), 320–324. https://doi.org/10.1016/j.icte.2020.04.009 doi: 10.1016/j.icte.2020.04.009
![]() |
[33] |
Z. Salod, Y. Singh, Comparison of the performance of machine learning algorithms in breast cancer screening and detection: A Protocol, J. Public Health Res., 8 (2019). https://doi.org/10.4081/jphr.2019.1677 doi: 10.4081/jphr.2019.1677
![]() |
[34] |
Y. Lin, H. Luo, D. Wang, H. Guo, K. Zhu, An ensemble model based on machine learning methods and data preprocessing for short-term electric load forecasting, Energies, 10 (2017), 1186. https://doi.org/10.3390/en10081186 doi: 10.3390/en10081186
![]() |
[35] | M. Amrane, S. Oukid, I. Gagaoua, T. Ensari, Breast cancer classification using machine learning, in 2018 Electric Electronics, Computer Science, Biomedical Engineerings' Meeting (EBBT), (2018), 1–4. https://doi.org/10.1109/ebbt.2018.8391453 |
[36] |
R. Sumbaly, N. Vishnusri, S. Jeyalatha, Diagnosis of breast cancer using decision tree data mining technique, Int. J. Comput. Appl., 98 (2014), 16–24. https://doi.org/10.5120/17219-7456 doi: 10.5120/17219-7456
![]() |
[37] |
B. Zheng, S. W. Yoon, S. S. Lam, Breast cancer diagnosis based on feature extraction using a hybrid of k-means and support vector machine algorithms, Expert Syst. Appl., 41 (2014), 1476–1482. https://doi.org/10.1016/j.eswa.2013.08.044 doi: 10.1016/j.eswa.2013.08.044
![]() |
[38] |
T. Araújo, G. Aresta, E. Castro, J. Rouco, P. Aguiar, C. Eloy, et al., Classification of breast cancer histology images using convolutional neural networks, Plos One, 12 (2017), e0177544. https://doi.org/10.1371/journal.pone.0177544 doi: 10.1371/journal.pone.0177544
![]() |
[39] | S. P. Rajamohana, A. Dharani, P. Anushree, B. Santhiya, K. Umamaheswari, Machine learning techniques for healthcare applications: early autism detection using ensemble approach and breast cancer prediction using SMO and IBK, in Cognitive Social Mining Applications in Data Analytics and Forensics, (2019), 236–251. https://doi.org/10.4018/978-1-5225-7522-1.ch012 |
[40] |
L. G. Ahmad, Using three machine learning techniques for predicting breast cancer recurrence, J. Health Med. Inf., 4 (2013), 10–15. https://doi.org/10.4172/2157-7420.1000124 doi: 10.4172/2157-7420.1000124
![]() |
[41] |
B. Padmapriya, T. Velmurugan, Classification algorithm based analysis of breast cancer data, Int. J. Data Min. Tech. Appl., 5 (2016), 43–49. https://doi.org/10.20894/ijdmta.102.005.001.010 doi: 10.20894/ijdmta.102.005.001.010
![]() |
[42] | S. Bharati, M. A. Rahman, P. Podder, Breast cancer prediction applying different classification algorithm with comparative analysis using Weka, in 2018 4th International Conference on Electrical Engineering and Information & Communication Technology (ICEEiCT), (2018), 581–584. https://doi.org/10.1109/ceeict.2018.8628084 |
[43] |
K. Williams, P. A. Idowu, J. A. Balogun, A. I. Oluwaranti, Breast cancer risk prediction using data mining classification techniques, Trans. Networks Commun., 3 (2015), 17–23. https://doi.org/10.14738/tnc.32.662 doi: 10.14738/tnc.32.662
![]() |
[44] | P. Mekha, N. Teeyasuksaet, Deep learning algorithms for predicting breast cancer based on tumor cells, in 2019 Joint International Conference on Digital Arts, Media and Technology with ECTI Northern Section Conference on Electrical, Electronics, Computer and Telecommunications Engineering (ECTI DAMT-NCON), 2019. https://doi.org/10.1109/ecti-ncon.2019.8692297 |
[45] | C. Shah, A. G. Jivani, Comparison of data mining classification algorithms for breast cancer prediction, in 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT), 2013. https://doi.org/10.1109/icccnt.2013.6726477 |
[46] |
A. A. Bataineh, A comparative analysis of nonlinear machine learning algorithms for breast cancer detection, Int. J. Mach. Learn. Comput., 9 (2019), 248–254. https://doi.org/10.18178/ijmlc.2019.9.3.794 doi: 10.18178/ijmlc.2019.9.3.794
![]() |
[47] | M. S. M. Prince, A. Hasan, F. M. Shah, An efficient ensemble method for cancer detection, in 2019 1st International Conference on Advances in Science, Engineering and Robotics Technology (ICASERT), 2019. https://doi.org/10.1109/icasert.2019.8934817 |
[48] |
S. Aruna, A novel SVM based CSSFFS feature selection algorithm for Detecting Breast Cancer, Int. J. Comput., 31 (2011), 14–20. https://doi.org/10.5120/3844-5346 doi: 10.5120/3844-5346
![]() |
[49] |
G. Carneiro, J. Nascimento, A. P. Bradley, Automated analysis of unregistered Multi-View Mammograms with deep learning, IEEE Trans. Med. Imaging, 36 (2017), 2355–2365. https://doi.org/10.1109/tmi.2017.2751523 doi: 10.1109/tmi.2017.2751523
![]() |
[50] |
Z. Sha, L. Hu, B. D. Rouyendegh, Deep learning and optimization algorithms for Automatic Breast Cancer Detection, Int. J. Imaging Syst. Technol., 30 (2020), 495–506. https://doi.org/10.1002/ima.22400 doi: 10.1002/ima.22400
![]() |
[51] |
M. Mahmoud, Breast cancer classification in histopathological images using convolutional neural network, Int. J. Comput. Sci. Appl., 9 (2018), 12–15. https://doi.org/10.14569/ijacsa.2018.090310 doi: 10.14569/ijacsa.2018.090310
![]() |
[52] |
Z. Jiao, X. Gao, Y. Wang, J. Li, A deep feature based framework for Breast Masses classification, Neurocomputing, 197 (2016), 221–231. https://doi.org/10.1016/j.neucom.2016.02.060 doi: 10.1016/j.neucom.2016.02.060
![]() |
[53] |
M. H. Yap, G. Pons, J. Marti, S. Ganau, M. Sentis, R. Zwiggelaar, et al., Automated breast ultrasound lesions detection using convolutional neural networks, IEEE. J. Biomed. Health Inf., 22 (2018), 1218–1226. https://doi.org/10.1109/jbhi.2017.2731873 doi: 10.1109/jbhi.2017.2731873
![]() |
[54] |
N. Wahab, A. Khan, Y. S. Lee, Transfer learning based deep CNN for segmentation and detection of mitoses in breast cancer histopathological images, Microscopy, 68 (2019), 216–233. https://doi.org/10.1093/jmicro/dfz002 doi: 10.1093/jmicro/dfz002
![]() |
[55] |
Z. Wang, G. Yu, Y. Kang, Y. Zhao, Q. Qu, Breast tumor detection in digital mammography based on Extreme Learning Machine, Neurocomputing, 128 (2014), 175–184. https://doi.org/10.1016/j.neucom.2013.05.053 doi: 10.1016/j.neucom.2013.05.053
![]() |
[56] |
Y. Qiu, Y. Wang, S. Yan, M. Tan, S. Cheng, H. Liu, et al., An initial investigation on developing a new method to predict short-term breast cancer risk based on Deep Learning Technology, Comput. Aided. Des., 2016. https://doi.org/10.1117/12.2216275 doi: 10.1117/12.2216275
![]() |
[57] |
X. W. Chen, X. Lin, Big data deep learning: Challenges and perspectives, IEEE Access, 2 (2014), 514–525. https://doi.org/10.1109/access.2014.2325029 doi: 10.1109/access.2014.2325029
![]() |
[58] |
J. Arevalo, F. A. González, R. R. Pollán, J. L. Oliveira, M. A. G. Lopez, Representation learning for mammography mass lesion classification with convolutional neural networks, Comput. Methods Programs Biomed., 127 (2016), 248–257. https://doi.org/10.1016/j.cmpb.2015.12.014 doi: 10.1016/j.cmpb.2015.12.014
![]() |
[59] |
Y. Kumar, A. Aggarwal, S. Tiwari, K. Singh, An efficient and robust approach for biomedical image retrieval using zernike moments, Biomed. Signal Process. Control, 39 (2018), 459–473. https://doi.org/10.1016/j.bspc.2017.08.018 doi: 10.1016/j.bspc.2017.08.018
![]() |
[60] |
K. Kalaiarasi, R. Soundaria, N. Kausar, P. Agarwal, H. Aydi, H. Alsamir, Optimization of the average monthly cost of an EOQ inventory model for deteriorating items in machine learning using Python, Therm. Sci., 25 (2021), 347–358. https://doi.org/10.2298/tsci21s2347k doi: 10.2298/tsci21s2347k
![]() |
[61] |
M. Franulović, K. Marković, A. Trajkovski, Calibration of material models for the human cervical spine ligament behaviour using a genetic algorithm, Facta Univ., Series: Mechan. Eng., 19 (2021) 751. https://doi.org/10.22190/fume201029023f doi: 10.22190/fume201029023f
![]() |
[62] |
M. Fayaz, D. H. Kim, A prediction methodology of energy consumption based on Deep Extreme Learning Machine and comparative analysis in residential buildings, Electronics, 7 (2018), 222. https://doi.org/10.3390/electronics7100222 doi: 10.3390/electronics7100222
![]() |
[63] |
G. B. Huang, D. H. Wang, Y. Lan, Extreme learning machines: A survey, Int. J. Mach. Learn. Cybern., 2 (2011), 107–122. https://doi.org/10.1007/s13042-011-0019-y doi: 10.1007/s13042-011-0019-y
![]() |
[64] |
H. Tang, S. Gao, L. Wang, X. Li, B. Li, S. Pang, A novel intelligent fault diagnosis method for rolling bearings based on Wasserstein generative adversarial network and Convolutional Neural Network under Unbalanced Dataset, Sensors, 21 (2021), 6754. https://doi.org/10.3390/s21206754 doi: 10.3390/s21206754
![]() |
[65] |
J. Wei, H. Liu, G. Yan, F. Sun, Multi-modal deep extreme learning machine for robotic grasping recognition, Proceed. Adapt., Learn. Optim., (2016), 223–233. https://doi.org/10.1007/978-3-319-28373-9_19 doi: 10.1007/978-3-319-28373-9_19
![]() |
[66] |
N. S. Naz, M. A. Khan, S. Abbas, A. Ather, S. Saqib, Intelligent routing between capsules empowered with deep extreme machine learning technique, SN Appl. Sci., 2 (2019), 1–14. https://doi.org/10.1007/s42452-019-1873-6 doi: 10.1007/s42452-019-1873-6
![]() |
[67] |
J. Cai, J. Luo, S. Wang, S. Yang, Feature selection in Machine Learning: A new perspective, Neurocomputing, 300 (2018), 70–79. https://doi.org/10.1016/j.neucom.2017.11.077 doi: 10.1016/j.neucom.2017.11.077
![]() |
[68] |
L. M. Abualigah, A. T. Khader, E. S. Hanandeh, A new feature selection method to improve the document clustering using particle swarm optimization algorithm, J. Comput. Sci., 25 (2018), 456–466. https://doi.org/10.1016/j.jocs.2017.07.018 doi: 10.1016/j.jocs.2017.07.018
![]() |
[69] |
P. A. Flach, ROC analysis, encyclopedia of machine learning and data mining, Encycl. Mach. Learn. Data Min., (2016), 1–8. https://doi.org/10.1007/978-1-4899-7502-7_739-1 doi: 10.1007/978-1-4899-7502-7_739-1
![]() |
[70] |
Q. Wuniri, W. Huangfu, Y. Liu, X. Lin, L. Liu, Z. Yu, A generic-driven wrapper embedded with feature-type-aware hybrid bayesian classifier for breast cancer classification, IEEE Access, 7 (2019), 119931–119942. https://doi.org/10.1109/access.2019.2932505 doi: 10.1109/access.2019.2932505
![]() |
[71] |
J. Zheng, D. Lin, Z. Gao, S. Wang, M. He, J. Fan, Deep Learning assisted efficient ADABOOST algorithm for breast cancer detection and early diagnosis, IEEE Access, 8 (2020), 96946–96954. https://doi.org/10.1109/access.2020.2993536 doi: 10.1109/access.2020.2993536
![]() |
[72] |
X. Zhang, D. He, Y. Zheng, H. Huo, S. Li, R. Chai, et al., Deep learning based analysis of breast cancer using advanced ensemble classifier and linear discriminant analysis, IEEE Access, 8 (2020), 120208–120217. https://doi.org/10.1109/access.2020.3005228 doi: 10.1109/access.2020.3005228
![]() |
[73] |
Y. Yari, T. V. Nguyen, H. T. Nguyen, Deep learning applied for histological diagnosis of breast cancer, IEEE Access, 8 (2020), 162432–162448. https://doi.org/10.1109/access.2020.3021557 doi: 10.1109/access.2020.3021557
![]() |
[74] |
A. H. Osman, H. M. Aljahdali, An effective of ensemble boosting learning method for breast cancer virtual screening using neural network model, IEEE Access, 8 (2020), 39165–39174. https://doi.org/10.1109/access.2020.2976149 doi: 10.1109/access.2020.2976149
![]() |
[75] |
Y. Li, J. Wu, Q. Wu, Classification of breast cancer histology images using multi-size and discriminative patches based on Deep Learning, IEEE Access, 7 (2019), 21400–21408. https://doi.org/10.1109/access.2019.2898044 doi: 10.1109/access.2019.2898044
![]() |
[76] |
D. M. Vo, N. Q. Nguyen, S. W. Lee, Classification of breast cancer histology images using incremental boosting convolution networks, Inf. Sci., 482 (2019), 123–138. https://doi.org/10.1016/j.ins.2018.12.089 doi: 10.1016/j.ins.2018.12.089
![]() |
[77] |
S. Y. Siddiqui, M. A. Khan, S. Abbas, F. Khan, Smart occupancy detection for road traffic parking using deep extreme learning machine, J. K.S.U. Comput. Inf. Sci., 34 (2022), 727–733. https://doi.org/10.1016/j.jksuci.2020.01.016 doi: 10.1016/j.jksuci.2020.01.016
![]() |
[78] |
M. A. Khan, S. Abbas, K. M. Khan, M. A. A. Ghamdi, A. Rehman, Intelligent forecasting model of covid-19 novel coronavirus outbreak empowered with deep extreme learning machine, Comput. Matern. Contin., 64 (2020), 1329–1342. https://doi.org/10.32604/cmc.2020.011155 doi: 10.32604/cmc.2020.011155
![]() |
[79] |
S. Abbas, M. A. Khan, L. E. F. Morales, A. Rehman, Y. Saeed, Modelling, simulation and optimization of power plant energy sustainability for IoT enabled smart cities empowered with deep extreme learning machine, IEEE Access, 8 (2020), 39982–39997. https://doi.org/10.1109/ACCESS.2020.2976452 doi: 10.1109/ACCESS.2020.2976452
![]() |
[80] |
A. Rehman, A. Athar, M. A. Khan, S. Abbas, A. Fatima, M. Zareei, et al., Modelling, simulation, and optimization of diabetes type ii prediction using deep extreme learning machine, J. Ambient Intell. Smart Environ., 12 (2020), 125–138. https://doi.org/10.3233/AIS-200554 doi: 10.3233/AIS-200554
![]() |
[81] |
A. Haider, M. A. Khan, A. Rehman, H. S. Kim, A real-time sequential deep extreme learning machine cybersecurity intrusion detection system, Comput. Matern. Contin., 66 (2021), 1785–1798. https://doi.org/10.32604/cmc.2020.013910 doi: 10.32604/cmc.2020.013910
![]() |
[82] |
M. A. Khan, A. Rehman, K. M. Khan, M. A. A. Ghamdi, S. H. Almotiri, Enhance intrusion detection in computer networks based on deep extreme learning machine, Comput. Matern. Contin., 66 (2021), 467–480. https://doi.org/10.32604/cmc.2020.013121 doi: 10.32604/cmc.2020.013121
![]() |
[83] |
U. Ahmed, G. F. Issa, M. A. Khan, S. Aftab, M. F. Khan, R. A. T. Said, et al., Prediction of diabetes empowered with fused machine learning, IEEE Access, 10 (2022), 8529–8538. https://doi.org/10.1109/ACCESS.2022.3142097 doi: 10.1109/ACCESS.2022.3142097
![]() |
[84] |
S. Y. Siddiqui, A. Haider, T. M. Ghazal, M. A. Khan, I. Naseer, S. Abbas, et al., IoMT cloud-based intelligent prediction of breast cancer stages empowered with deep learning, IEEE Access, 9 (2021), 146478–146491. https://doi.org/10.1109/ACCESS.2021.3123472 doi: 10.1109/ACCESS.2021.3123472
![]() |
[85] |
M. Ahmad, M. Alfayad, S. Aftab, M. A. Khan, A. Fatima, B. Shoaib, et.al., Data and machine learning fusion architecture for cardiovascular disease prediction, Comput. Matern. Contin., 69 (2021), 2717–2731. https://doi.org/10.32604/cmc.2021.019013 doi: 10.32604/cmc.2021.019013
![]() |
1. | W. Jung, C.A. Morales, Training neural networks from an ergodic perspective, 2023, 0233-1934, 1, 10.1080/02331934.2023.2239852 | |
2. | Steffen Dereich, Arnulf Jentzen, Sebastian Kassing, On the Existence of Minimizers in Shallow Residual ReLU Neural Network Optimization Landscapes, 2024, 62, 0036-1429, 2640, 10.1137/23M1556241 |