
Reduced order modelling relies on representing complex dynamical systems using simplified modes, which can be achieved through the Koopman operator(KO) analysis. However, computing Koopman eigenpairs for high-dimensional observable data can be inefficient. This paper proposes using deep autoencoders(AE), a type of deep learning technique, to perform nonlinear geometric transformations on raw data before computing Koopman eigenvectors. The encoded data produced by the deep AE is diffeomorphic to a manifold of the dynamical system and has a significantly lower dimension than the raw data. To handle high-dimensional time series data, Takens' time delay embedding is presented as a preprocessing technique. The paper concludes by presenting examples of these techniques in action.
Citation: Neranjaka Jayarathne, Erik M. Bollt. Autoencoding for the 'Good Dictionary' of eigenpairs of the Koopman operator[J]. AIMS Mathematics, 2024, 9(1): 998-1022. doi: 10.3934/math.2024050
[1] | Ravi Agarwal, Snezhana Hristova, Donal O'Regan . Integral presentations of the solution of a boundary value problem for impulsive fractional integro-differential equations with Riemann-Liouville derivatives. AIMS Mathematics, 2022, 7(2): 2973-2988. doi: 10.3934/math.2022164 |
[2] | Hasanen A. Hammad, Hassen Aydi, Manuel De la Sen . The existence and stability results of multi-order boundary value problems involving Riemann-Liouville fractional operators. AIMS Mathematics, 2023, 8(5): 11325-11349. doi: 10.3934/math.2023574 |
[3] | Ymnah Alruwaily, Lamya Almaghamsi, Kulandhaivel Karthikeyan, El-sayed El-hady . Existence and uniqueness for a coupled system of fractional equations involving Riemann-Liouville and Caputo derivatives with coupled Riemann-Stieltjes integro-multipoint boundary conditions. AIMS Mathematics, 2023, 8(5): 10067-10094. doi: 10.3934/math.2023510 |
[4] | Md. Asaduzzaman, Md. Zulfikar Ali . Existence of positive solution to the boundary value problems for coupled system of nonlinear fractional differential equations. AIMS Mathematics, 2019, 4(3): 880-895. doi: 10.3934/math.2019.3.880 |
[5] | Khalid K. Ali, K. R. Raslan, Amira Abd-Elall Ibrahim, Mohamed S. Mohamed . On study the fractional Caputo-Fabrizio integro differential equation including the fractional q-integral of the Riemann-Liouville type. AIMS Mathematics, 2023, 8(8): 18206-18222. doi: 10.3934/math.2023925 |
[6] | Veliappan Vijayaraj, Chokkalingam Ravichandran, Thongchai Botmart, Kottakkaran Sooppy Nisar, Kasthurisamy Jothimani . Existence and data dependence results for neutral fractional order integro-differential equations. AIMS Mathematics, 2023, 8(1): 1055-1071. doi: 10.3934/math.2023052 |
[7] | Ugyen Samdrup Tshering, Ekkarath Thailert, Sotiris K. Ntouyas . Existence and stability results for a coupled system of Hilfer-Hadamard sequential fractional differential equations with multi-point fractional integral boundary conditions. AIMS Mathematics, 2024, 9(9): 25849-25878. doi: 10.3934/math.20241263 |
[8] | Thabet Abdeljawad, Sabri T. M. Thabet, Imed Kedim, Miguel Vivas-Cortez . On a new structure of multi-term Hilfer fractional impulsive neutral Levin-Nohel integrodifferential system with variable time delay. AIMS Mathematics, 2024, 9(3): 7372-7395. doi: 10.3934/math.2024357 |
[9] | Pinghua Yang, Caixia Yang . The new general solution for a class of fractional-order impulsive differential equations involving the Riemann-Liouville type Hadamard fractional derivative. AIMS Mathematics, 2023, 8(5): 11837-11850. doi: 10.3934/math.2023599 |
[10] | Bashir Ahmad, Ahmed Alsaedi, Sotiris K. Ntouyas . Nonlinear Langevin equations and inclusions involving mixed fractional order derivatives and variable coefficient with fractional nonlocal-terminal conditions. AIMS Mathematics, 2019, 4(3): 626-647. doi: 10.3934/math.2019.3.626 |
Reduced order modelling relies on representing complex dynamical systems using simplified modes, which can be achieved through the Koopman operator(KO) analysis. However, computing Koopman eigenpairs for high-dimensional observable data can be inefficient. This paper proposes using deep autoencoders(AE), a type of deep learning technique, to perform nonlinear geometric transformations on raw data before computing Koopman eigenvectors. The encoded data produced by the deep AE is diffeomorphic to a manifold of the dynamical system and has a significantly lower dimension than the raw data. To handle high-dimensional time series data, Takens' time delay embedding is presented as a preprocessing technique. The paper concludes by presenting examples of these techniques in action.
In the modern fields of scientific research and medical diagnostics [1,2,3], there is an increasing reliance on image restoration techniques [4,5,6,7], which are particularly prominent in the field of medical imaging. Ensuring image quality is paramount for the authenticity of data [8]. By processing X-ray projection data obtained from CT scans, we can reconstruct clear tomographic images with a resolution of n×n pixels. This technology encompasses two main schools: mathematical theoretical analysis [9] and iterative methods [10]. The former, like the filtered back projection method [11], is widely used in fields such as CT imaging, while the latter excels in dealing with noise interference and data loss. With technological advancements, image restoration is moving toward greater precision and efficiency.
To reconstruct images, we must solve a large-scale system of linear equations with multiple righthand sides, which is presented as follows:
AX=B | (1) |
In contemporary mathematics and engineering, solving systems of equations stands as a pivotal task, particularly in scientific research and industrial applications where efficiency and precision are of paramount importance. A variety of numerical methods have been widely adopted [12,13,14,15], among which the Kaczmarz algorithm [16] is renowned for its iterative projection approach in approximating the true solution. The algorithm's simplicity has facilitated its application across various domains, including image reconstruction [12,17], medical imaging [11,18], and signal processing [19,20]. Advances in technology have given rise to multiple enhanced versions of the Kaczmarz algorithm [21,22,23,24,25,26,27,28,29,30,31], improving performance in large-scale parallel computing and noisy data environments. Notably, with the development of free pseudo-inverse techniques, Du and Sun [24] further extended the randomized extended average block Kaczmar (REABK) method, proposing a class of pseudo-inverse-free stochastic block iterative methods for solving both consistent and inconsistent systems of linear equations, Free pseudo-inverse can accelerate convergence speed. Pseudo-inverse approximation is a specific case of generalized inverse techniques. For more theoretical analysis and applications, refer to literature [32]. Inspired by references [33,34], this paper introduces a faster lazy free greedy block Kaczmarz (LFGBK) method, exploring matrix sketching techniques as a key tool for accelerating matrix operations. This method employs sampling based on an approximate maximum distance criterion, excelling at selecting small, representative samples from large datasets for more efficient computation. The adaptive randomized block Kaczmarz (ARBK) [35] algorithm integrates adaptive and randomized block selection strategies. However, under certain specific matrix structures, ARBK may experience slow convergence or even fail to converge. Additionally, when dealing with large-scale sparse matrices, the ARBK algorithm incurs significant memory overhead. Our improved algorithm successfully overcomes these drawbacks, offering a more stable and efficient solution.
Additionally, it not only addresses single righthand side linear equations but also extends to multiple righthand side linear equations, solving the memory overflow issues that traditional algorithms face when dealing with image processing vectors. A comprehensive theoretical framework supports the convergence of these methods. Numerical experiments validate its effectiveness, demonstrating improved computational efficiency and laying a solid foundation for further optimization of matrix sketching techniques.
The structure of this paper is as follows: Section 2 introduces the necessary background knowledge. Section 3 presents the faster free pseudo-inverse GBK method for solving single righthand side linear equation systems. Section 4 proposes the leverage score sampling free pseudo-inverse GBK method for solving multiple righthand side linear equation systems. Section 5 details the numerical experiments, and Section 6 concludes the paper.
In this article, we adopt the same notation as in reference [27]. For example, A(i),AT,A†,||A||,||A||F, and [n], respectively, represent the i-th row of the coefficient matrix A, transpose, generalized inverse, spectral norm, F-norm, and the set {1,2,…,n}.
Recently, Niu and Zheng combined greedy strategy with the block Kaczmarz method, proposing the GBK [29] method for solving large-scale consistent linear equation systems. See Algorithm 1 for the specific process.
Algorithm1 GBK method |
1: Input: A,b,l,x0∈range(AT)andη∈(0,1). 2: Output: xl. 3: for k=0,1,…l−1 do 4: Compute εk=ηmax1≤i≤m{|bi−A(i)xk|2||A(i)||22}. 5: Determine the sequence of indicator sets. Tk={ik:|bik−A(ik)xk|2≥εk||A(ik)||22}. 6: Compute xk+1=xk+A†Tk(bTk−ATkxk). 7: end for |
Convergence analysis of the GBK method is described as follows:
Theorem 1 ([29]) If the linear system of Eq (1) is consistent, then the iterative sequence {xk}∞k=0 generated by algorithm 1 converges to the minimum norm solution x∗=A†b of the system of equations, and satisfies for any k≥0,
||xk+1−x∗||22≤(1−γk(η)σ2min(A)||A||2F)k+1||x0−x∗||22. |
The formula for γk(η) is defined as follows: γk(η)=η||A||2F||A||2F−||ATk−1||2F||ATk||2Fσ2max(ATk). Here, γ0(η) is defined as η||AT0||2Fσ2max(AT0), where η is in the range (0,1], and σmin(A) and σmax(A) represent the nonzero minimum singular value and maximum singular value of matrix A.
In the matrix sketching technique, as described in leverage score sampling [36,37], we select samples based on the leverage score of each row. Specifically, we will choose each row with a probability proportional to its leverage score. Therefore, rows with higher scores (i.e., rows with greater influence in the dataset) will have a greater chance of being selected.
Algorithm2 Leverage score sampling method based on manifold |
1: Input:A∈Rm×n. 2: Initialize C as a d×n zero matrix. 3: Initialize the variable sum to 0. 4: Calculate the singular value decomposition of A as U, S, and V. 5: for k=1,…,m 6: Calculate the sum of the squares of each row in U and add it to the sum. 7: Calculate the probability of each data point being sampled: Calculate the sum of squares for each row of U, then divide by the total sum. 8: Based on the calculated probabilities, sampling is conducted to obtain the sampling indices. 9: Utilize C=[indices,:] for indexing. 10: end for 11: Return C∈Rd×n. |
The advantage of this sampling method lies in its ability to select a small, representative subset of samples from a large dataset, enabling more efficient computation. The resulting sample set S∈Rd×m, where d is the chosen number of samples, can be utilized to estimate various properties of the original matrix A, such as singular value decomposition, principal component analysis, and so on. The LFGBK algorithm is similar to the derivation process in reference [38].
Remark: In addition to the method proposed in this paper, there are other sampling techniques such as random Gaussian matrices, subsampled randomized Hadamard transform (SRHT), and uniform sampling. In previous experiments, we also tried other methods like random sparse sampling. Through comparison, we found that the method proposed in this paper excels in extracting the main features of matrices. Other methods not only have higher computational complexity but may also extract all-zero vectors during actual image processing, rendering calculations impossible. Our method effectively avoids these issues, which is why we chose to adopt it.
This chapter introduces an algorithm, namely, the leverage score sampling free pseudo-inverse GBK method for single righthand side, and provides a proof of the corresponding convergence theory for the algorithm.
First, each step follows the approximate maximum illustration principle.
|bik−A(ik)xk|2||A(ik)||22≥ηmax1≤i≤m{|bi−A(i)xk|2||A(i)||22}, |
Second, select the index set Tk for the block matrix ATk; then, project the current estimate onto each row forming the block matrix ATk; finally, calculate the average of the projections to determine the next iteration.
xk+1=xk+(∑i∈Tkwibi−A(i)xk||A(i)||22AT(i)). |
Before presenting the convergence theory for Algorithm 3, let us first introduce a lemma.
Algorithm 3 SLFGBK method |
1: Let A,b,l,x0 be within the range of (AT), parameter η within (0,1), and consider the sequences of step sizes (αk)k≥0 and weights (wk)k≥0. 2: Output:xl. 3: Initialization: Leverage score sampling enables the selection of a small, representative subset from large datasets, facilitating more efficient computation. Algorithm 2 generates a leverage score sampling transformation matrix ˜A=SA∈Rd×n and vector ˜b=Sb, where d≪m. 4: for k=0,1,…l−1 do 5: Calculation ˜εk=ηmax1≤i≤m{|˜bi−˜A(i)xk|2||˜A(i)||22}. 6: Define the index set sequence Tk={ik:|˜biK−˜A(ik)xk|2≥εk||˜A(ik)||22}. 7: Computation xk+1=xk+(∑i∈Tkwi˜bi−˜A(i)xk||˜A(i)||22˜AT(i)). 8: end for |
Lemma 1 ([27]): If any vector u belongs to the range of AT, then
||Au||22≥λmin(ATA)||u||22 |
Theorem 2: The leverage score transformation S satisfies d=O(n2/δθ2), and x∗=A†b is the minimum norm solution of the single righthand side linear equation systems, for any k ≥ 0.
||xk+1−x∗||22≤(1−˜ψk(η)σ2min(˜A)||˜A||2F)||xk−x∗||22 | (2) |
The function ˜ψk(η)=ηt(2t−1t2σ2max(ˆATTk))σ2min,η∈(0,1], where range(A),σmin(A), and σmax(A) represent the range, nonzero minimum singular value, and maximum singular value of matrix A, respectively.
Proof. With algorithm 3 and ˜rk=˜b−˜Axk, we can obtain
xk+1=xk+(∑i∈Tk1|Tk|˜rik˜AT(i)||˜A(i)||22). | (3) |
Expand Eq (3) with the set |Tk|=t and Tk={jk1,…,jkt}.
xk+1=xk+∑i∈Tk1t˜AT(i)eTi˜rk||˜A(i)||22=xk+1t(˜AT(jk1)eTjk1˜rk||˜A(jk1)||22+⋯+˜AT(jkt)eTjkt˜rk||˜A(jkt)||22)=xk+1tˆATTkˆITk˜rk. | (4) |
Among them
ˆATTk=[˜AT(jk1)||˜A(jk1)||2,˜AT(jk2)||˜A(jk2)||2,…,˜AT(jkt)||˜A(jkt)||2]∈Rn×t, | (5) |
and
ˆITk=[e(jk1)||˜A(jk1)||2,e(jk2)||˜A(jk2)||2,…,e(jkt)||˜A(jkt)||2]T∈Rt×d. | (6) |
Subtracting x∗ from Eq (4) simultaneously yields
xk+1−x∗=xk−x∗−1tˆATTkˆITk˜A(xk−x∗)=(I−1tˆATTkˆATk)(xk−x∗) | (7) |
Taking the spectral norm and squaring both sides of Eq (7), and for any positive semi-definite matrix Q, satisfying Q2≼λmax(Q)Q, we can obtain
||xk+1−x∗||22=||(xk−x∗)−1tˆATTkˆATk(xk−x∗)||22=||(xk−x∗)||22−2t||ˆATk(xk−x∗)||22+1t2||ˆATTkˆATk((xk−x∗)||22≤||(xk−x∗)||22−(2t−1t2σ2max(ˆATTk))||ˆATk(xk−x∗)||22 | (8) |
Using Eq (5) and the inequality |˜bjk−˜A(jk)xk|2≥˜εk||˜A(jk)||22, a straightforward calculation yields.
||ˆATk(xk−x∗)||22=∑jk∈Tk1||˜A(jk)||22|˜rjkk|2 | (9) |
≥∑jk∈Tk1||˜A(jk)||22˜εk||˜A(jk)||2 | (10) |
Substituting ˜εk=ηmax1≤i≤d{|˜bi−˜A(i)xk|2||˜A(i)||22} into the above equation yields the following result:
||ˆATk(xk−x∗)||22≥∑jk∈Tkηmax1≤i≤d{|˜rik|2||˜A(i)||22}≥ηtd∑i=1|˜rik|2||A(i)||22||˜A(i)||22||A||2F=ηt||˜rk||2||˜A||2F | (11) |
Given x0∈range(AT) and x∗=A†b, we have xk−x∗∈range(AT). Therefore, by Lemma 1, we can conclude that
||˜rk||2=||˜A(xk−x∗)||22 |
≥λ2min(˜AT˜A)||xk−x∗||22 Combining the above equation, we can obtain the following
||ˆATk(xk−x∗)||22≥ηtσ2min(˜A)||˜A||2F||xk−x∗||22 | (12) |
From Eqs (8) and (12), we can obtain Eq (2), and the iterative sequence {xk}∞k=0 converges to the minimum norm solution x∗=A†b of the system of equations. Hence, Theorem 2 is proved.
In this chapter, we introduce an algorithm known as the multi-righthand-side leverage score sampling free pseudo-inverse GBK method, and we provide a proof for the corresponding convergence theory of the algorithm.
For most image reconstruction, the system of equations is formulated as follows:
AX=B | (13) |
In the context, A∈Rm×n,X=[x1,x2,…,xd]∈Rn×d,B=[b1,b2,…,bd]∈Rm×d,d>1. This paper addresses the solution of multiple righthand side linear equation systems and initially presents the index set for each iteration selection in the multiple righthand side linear equation systems.
max1≤i≤m{|b(i)−A(i)xk|||A(i)||22} | (14) |
Iterate through the following steps:
xk+1=xk+b(i)−A(i)xk||A(i)||22AT(i) |
To solve systems of linear equations with multiple righthand sides, our approach involves working with the j righthand side vector bj by selecting the working row Tkj according to the criterion established in Eq (15).
max1≤i≤m{|B(i,j)−A(i)x(k)j|||A(i)||22},j=1,2,…,kb. | (15) |
The iterative formula corresponding to the solution at the j right endpoint is as follows:
x(k+1)j=xkj+˜B(Tkj,j)−˜A(Tkj)x(k)j||˜A(Tkj)||22˜ATTkj | (16) |
According to the criterion for selecting the index set in Eq (15), each term on the righthand side corresponds to the selection of a working row. For the system of linear equations with multiple righthand terms as in Eq (13), the Kaczmarz method can be extended to an iterative format that simultaneously solves for the righthand side system of linear equations.
In this context, X(k+1)=(x(k+1)1,x(k+1)2,…,x(k+1)d) represents the approximate solution at the k+1 iteration.
Xk+1=Xk+(˜ATjk1,…,˜ATjkd)Tdiag(˜B(jk1,1)−˜Ajk1x(k)1||˜Ajk1||22,…,˜B(jkd,d)−˜Ajkdx(k)d||˜Ajkd||22) |
To reduce computational effort, Algorithm 2 is employed to extract certain rows from matrix A corresponding to the index set Tkj. The sub-matrix formed by these rows is denoted as ˜A,, while the sub-matrix of B corresponding to the index set Tk is denoted as ˜B. The specific configurations of ˜A and ˜B are as follows.
˜BTTkj=(BTTkj1,BTTkj2,…,BTTkjt) | (17) |
and
˜ATTkj=(ATTk1,ATTk2,…,ATTkt) | (18) |
The transposed matrix BTTkji is represented as (B(Tkji,1),B(Tkji,2),…,B(Tkji,d)), where Tkji is an element of the index set Tkj. Here, B(Tkji,t) denotes the element in the i row and t column of the sub-matrix ˜B, and |Tkj| denotes the number of row indices contained in the index set Tkj with |Tkj| equal to t.
The multi-righthand-side method proposed in this paper is a specialized block approach for solving large-scale righthand side linear equations using a leverage-based pseudo-inverse GBK method. We provide a detailed framework for this method, which involves simultaneous computation of multiple righthand sides, with each iteration involving a matrix B∈Rm×t containing t righthand sides, resulting in the selection of t row indices.
This article exclusively discusses the scenario where αk equals 1. The convergence theory for solving large-scale righthand side linear equation groups with the LFGBK method is presented below:
Theorem 3: Assume that the linear equation system (13) is consistent, and so are its extracted subsystems. The iterative sequence X(k) from k=0 to infinity, generated by Algorithm 4 starting from the initial value x0, converges to the least squares solution X∗=A†B. The expected error for the iterative sequence solutions X(k) from k=0 to infinity is:
E||X(k+1)−X∗||2F≤(1−σ2min(˜A(Tkj))||˜A(Tkj)||2F)E||Xk−X∗||2F | (19) |
Algorithm 4 MLFGBK Method |
1: Let A,b,l,x0 be within the range of (AT), parameter η within (0,1), and consider the sequences of step sizes (αk)k≥0 and weights (wk)k≥0. 2: Output:x. 3: Initialization: Algorithm 2 generates a leverage score sampling transformation matrix ˜A=SA∈Rd×n and matrix ˜b=Sb, where d≪m. 4: for k=0,1,…l−1 do 5: Calculation ˜εkj=ηmax1≤i≤m{|B(i,j)−A(i)X(k)j|||A(i)||22},j=1,2,…,d. 6: Establish the sequence of indicator sets. Tkj={ikj:|˜bikj−˜A(ikj)xkj|2≥˜εkj||˜A(ikj)||22}. 7: Calculation Xk+1=Xk+(˜ATjk1,…,˜ATjkd)Tdiag{˜B(jk1,1)−˜Ajk1x(k)1||˜Ajk1||22,…,˜B(jkd,d)−˜Ajkdx(k)d||˜Ajkd||22}. 8: end for |
The tilde-decorated A and B, denoted as ˜A and ˜B, respectively, represent the sub-matrices formed by the rows corresponding to the index set Tkj within matrices A and B.
Proof. According to Algorithm 4 and reference [39], we have
x(k+1)j=xkj+˜B(Tkj,j)−˜A(Tkj)x(k)j||˜A(Tkj)||22˜ATTkj |
Since X∗=A†B is the least squares solution, we infer that x∗j=A†B(:,j) for j=1,2,…,d, where B(:,j) denotes the j column of matrix B.
Further,
E||X(k+1)−X∗||2F=E∑tj=1||x(k+1)j−x∗j||22=∑tj=1E||x(k+1)j−x∗j||22 | (20) |
For j=1,2,…,d, it can be deduced that
||x(k+1)j−x∗j||22=||x(k)j−x∗j||22−E||x(k+1)j−xkj||22 |
Then
E||x(k+1)j−xkj||22=|˜B(Tkj,j)−˜A(Tkj)x(k)j|2||˜A(Tkj)||22=maxTkji∈Tkj{|˜B(Tkji,j)−˜A(Tkji)x(k)j|2||˜A(Tkji)||22} |
The specific details of the term E||x(k+1)j−xkj||22 are elaborated in the proof of Theorem 3.2 in [39].
Due to
maxTkji∈Tkj{|˜B(Tkji,j)−˜A(Tkji)x(k)j|2||˜A(Tkji)||22}=maxTkji∈Tkj{|˜B(Tkji,j)−˜A(Tkji)x(k)j|2||˜A(Tkji)||22}||Akx(k)j−˜B(:,j)||22||Akx(k)j−˜B(:,j)||22=maxTkji∈Tkj{|˜B(Tkji,j)−˜A(Tkji)x(k)j|2||˜A(Tkji)||22}||Akx(k)j−˜B(:,j)||22∑ti=1||˜A(Tkji)||22||˜B(Tkji,j)−ATkjix(k)j||22||˜A(Tkji)||22≥||˜A(Tkj)x(k)j−˜B(:,j)||22∑ti=1||˜A(Tkji)||22≥σ2min(˜A(Tkj))||˜A(Tkj)||22||x(k)j−x∗j||22 | (21) |
In this context, ˜B(:,j) denotes the j column of the matrix ˜B.
E||X(k+1)−Xkj||22=||x(k)j−x∗j||22−E||x(k+1)j−xkj||22≤(1−σ2min(˜A(Tkj))||˜A(Tkj)||22)||x(k)j−x∗j||22 | (22) |
Thus, by combining Eqs (20) and (22), we can derive inequality (19).
The proof of Theorem 3 reveals that the convergence rate of Algorithm 4 is related to the sub-matrix ˜A sampled during each leveraged iteration.
In this section, we evaluate the effectiveness of large-scale image restoration equations using several algorithms through comparative examples: the GBK method [29], the semi-stochastic block Kaczmarz method with simple random sampling for single righthand sides [40], the LFGBK method for single righthand sides, the semi-stochastic block Kaczmarz method with simple random sampling for multiple righthand sides [40], and the LFGBK method for multiple righthand sides. All experiments are implemented in MATLAB, with metrics such as peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), mean squared error (MSE), and computational time (CPU in seconds) reported. Following the approach of [27], the PSNR, SSIM, MSE, and CPU metrics reflect the average iterations and computational time needed for 50 repeated calculations. During all computations, we initialize the matrix x=zeros(n,m) and set the righthand side B=AX, where X represents images from MATLAB's Cameraman, Phantom, and Mri datasets, each sized 100×100. Matrix A is constructed with elements corresponding to the product of the number of X-ray beams (4), the range of scanning angles t (0° to 179°), and the image pixels. The stopping criterion is set as RSE=||Xk−X∗||22||X∗||22≤10−6. In practice, the condition O(n2/δθ2) is quite stringent, but in many real-world computations, a sketching factor of d=n2 yields satisfactory results. We consider matrices A∈R(4∗180∗n,m) of two types: type a, generated by radon(), and type b, generated by randn().
Based on the numerical results from Tables 1 and 2, when variable A is generated using radon, we can draw the following conclusions: ⅰ) The GBK method, single righthand side SRBK (single RHS SRBK) method, single RHS LFGBK method, multiple righthand side SRBK (multiple RHS SRBK) method, and multiple RHS LFGBK method all demonstrate effectiveness in solving equation systems for image restoration. ⅱ) The GBK method, single RHS SRBK method, single RHS LFGBK method, multiple RHS SRBK method, and multiple RHS LFGBK method show similar restoration quality; however, the single RHS LFGBK and multiple RHS LFGBK methods significantly outperform the GBK, Single RHS SRBK, and Multiple RHS SRBK methods in terms of time efficiency. ⅲ) The LFGBK method exhibits superior acceleration effects at η = 0.9 compared to η = 0.8. ⅳ) The MLFGBK method displays more pronounced acceleration effects at η = 0.9 than at η = 0.8.
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 19.595 | 17.996 | 19.076 | 17.640 | 17.595 | 17.519 |
SSIM | 0.999 | 0.998 | 0.999 | 0.998 | 0.999 | 0.999 | |
MSE | 0.011 | 0.015 | 0.012 | 0.017 | 0.017 | 0.017 | |
CPU | 234.78 | 335.37 | 55.00 | 94.755 | 11.581 | 7.543 | |
Phantom | PSNR | 73.636 | 72.694 | 73.639 | 73.655 | 72.377 | 72.430 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 266.67 | 279.43 | 56.186 | 75.168 | 12.496 | 8.556 | |
Mri | PSNR | 29.319 | 28.936 | 28.068 | 29.023 | 28.057 | 28.112 |
SSIM | 0.969 | 0.975 | 0.924 | 0.919 | 0.962 | 0.978 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 259.86 | 375.48 | 60.963 | 162.06 | 19.340 | 15.563 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.595 | 17.519 | 17.611 | 17.732 | 17.894 | 17.876 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | |
CPU | 11.581 | 7.543 | 18.027 | 14.207 | 3.196 | 2.880 | |
Phantom | PSNR | 72.377 | 72.430 | 72.492 | 72.532 | 72.486 | 72.760 |
SSIM | 1.000 | 1.000 | 1.000 | 1.0000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 12.496 | 8.556 | 43.106 | 38.075 | 20.874 | 20.023 | |
Mri | PSNR | 28.057 | 28.112 | 28.081 | 28.234 | 28.161 | 28.506 |
SSIM | 0.962 | 0.978 | 0.901 | 0.899 | 0.895 | 0.891 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 19.340 | 15.563 | 42.535 | 37.575 | 18.237 | 17.539 |
Based on the numerical results from Tables 3 and 4, when variable A is generated using randn, we can draw the following conclusions: ⅰ) The GBK method, single RHS SRBK method, single RHS LFGBK method, multiple RHS SRBK method, and multiple RHS LFGBK method all demonstrate effectiveness in solving equation systems for image restoration. ⅱ) The GBK method, single RHS SRBK method, single RHS LFGBK method, multiple RHS SRBK method, and multiple RHS LFGBK method show similar restoration quality; however, the Single RHS LFGBK and Multiple RHS LFGBK methods significantly outperform the GBK, single RHS SRBK, and multiple RHS SRBK methods in terms of time efficiency. ⅲ) The LFGBK method exhibits superior acceleration effects at η = 0.9 compared to η = 0.8. ⅳ) The MLFGBK method displays more pronounced acceleration effects at η = 0.9 than at η = 0.8.
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 20.122 | 19.195 | 18.419 | 17.594 | 17.519 | 17.552 |
SSIM | 0.999 | 0.998 | 0.998 | 0.998 | 0.999 | 0.999 | |
MSE | 0.009 | 0.012 | 0.014 | 0.017 | 0.017 | 0.017 | |
CPU | 3.651 | 6.916 | 1.433 | 3.596 | 1.698 | 0.723 | |
Phantom | PSNR | 73.904 | 72.475 | 74.040 | 73.704 | 72.619 | 72.464 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 68.434 | 129.151 | 19.739 | 35.465 | 7.776 | 6.878 | |
Mri | PSNR | 30.615 | 29.072 | 29.289 | 28.899 | 28.080 | 28.169 |
SSIM | 0.912 | 0.900 | 0.893 | 0.884 | 0.892 | 0.905 | |
MSE | 0 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 63.758 | 122.983 | 28.132 | 31.338 | 9.451 | 8.641 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.519 | 17.552 | 17.748 | 17.762 | 17.860 | 17.805 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | 0.016 | |
CPU | 1.698 | 0.723 | 14.582 | 11.879 | 2.005 | 1.672 | |
Phantom | PSNR | 72.619 | 72.464 | 72.398 | 72.629 | 72.695 | 72.353 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 7.776 | 6.878 | 47.513 | 45.431 | 13.824 | 13.757 | |
Mri | PSNR | 28.080 | 28.169 | 28.241 | 28.093 | 28.241 | 28.328 |
SSIM | 0.892 | 0.905 | 0.904 | 0.904 | 0.902 | 0.897 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 9.451 | 8.641 | 54.802 | 38.773 | 11.533 | 10.332 |
This study aims to explore a new image reconstruction algorithm, the LFGBK method, which utilizes a greedy strategy. The core of this method lies in transforming the image reconstruction problem into solving a linear system problem with multiple righthand sides. In traditional Kaczmarz algorithms, the iterative process of gradually approaching the solution is often inefficient and susceptible to the influence of initial value selection. However, the LFGBK method introduces leverage score sampling and extends from solving single righthand side linear equations to solving multiple righthand side linear equations, which is crucial for handling large-scale problems due to the time-consuming and resource-intensive nature of pseudo-inverse computation on large matrices. By avoiding pseudo-inverse calculation, the LFGBK method significantly improves the algorithm's computational efficiency. To validate the effectiveness of the proposed algorithm, this study conducts in-depth research through theoretical analysis and simulation experiments. The theoretical analysis confirms the convergence of the LFGBK method, ensuring the reliability and stability of the algorithm. The simulation experiments, compared with traditional Filtered Back Projection (FBP) methods, demonstrate the advantages of LFGBK in terms of image reconstruction quality. The experimental results show that the LFGBK method significantly preserves image details and reduces noise, proving its practicality and superiority in image reconstruction tasks.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
The authors declare there is no conflict of interest.
[1] | S. Brunton, J. Kutz, Data-driven science and engineering: Machine learning, dynamical systems, and control, Cambridge University Press, 2022. https://doi.org/10.1017/9781009089517 |
[2] | E. Bollt, N. Santitissadeekorn, Applied and computational measurable dynamics, SIAM, 2013. https://doi.org/10.1137/1.9781611972641 |
[3] |
E. Bollt, Geometric considerations of a good dictionary for Koopman analysis of dynamical systems: Cardinality, "primary eigenfunction, " and efficient representation, Commun. Nonlinear Sci., 100 (2021), 105833. https://doi.org/10.1016/j.cnsns.2021.105833 doi: 10.1016/j.cnsns.2021.105833
![]() |
[4] | M. Budišić, R. Mohr, I. Mezić, Applied koopmanism, Chaos: An Interdisciplinary J. Nonlinear Sci., 22 (2012), 047510. https://doi.org/10.1063/1.4772195 |
[5] | J. Kutz, S. Brunton, B. Brunton, J. Proctor, Dynamic mode decomposition: data-driven modeling of complex systems, SIAM, 2016. |
[6] |
Y. Lan, I. Mezić, Linearization in the large of nonlinear systems and Koopman operator spectrum, Physica D: Nonlinear Phenomena, 242 (2013), 42–53. https://doi.org/10.1016/j.physd.2012.08.017 doi: 10.1016/j.physd.2012.08.017
![]() |
[7] |
A. Avila, I. Mezić, Data-driven analysis and forecasting of highway traffic dynamics, Nat. Commun., 11 (2020), 1–16. https://doi.org/10.1038/s41467-020-15582-5 doi: 10.1038/s41467-020-15582-5
![]() |
[8] |
I. Mezić, Spectral properties of dynamical systems, model reduction and decompositions, Nonlin. Dynam., 41 (2005), 309–325. https://doi.org/10.1007/s11071-005-2824-x doi: 10.1007/s11071-005-2824-x
![]() |
[9] |
I. Mezić, Spectrum of the Koopman operator, spectral expansions in functional spaces, and state-space geometry, J. Nonlinear Sci., 30 (2020), 2091–2145. https://doi.org/10.1007/s00332-019-09598-5 doi: 10.1007/s00332-019-09598-5
![]() |
[10] |
I. Mezić, A. Banaszuk, Comparison of systems with complex behavior, Physica D, 197 (2004), 101–133. https://doi.org/10.1016/j.physd.2004.06.015 doi: 10.1016/j.physd.2004.06.015
![]() |
[11] |
C. Rowley, I. Mezić, S. Bagheri, P. Schlatter, D. Henningson, Spectral analysis of nonlinear flows, J. Fluid Mech., 641 (2009), 115–127. https://doi.org/10.1017/S0022112009992059 doi: 10.1017/S0022112009992059
![]() |
[12] |
P. Schmid, Dynamic mode decomposition of numerical and experimental data, J. Fluid Mech., 656 (2010), 5–28. https://doi.org/10.1017/S0022112010001217 doi: 10.1017/S0022112010001217
![]() |
[13] | M. Jovanovic, P. Schmid, J. Nichols, Low-rank and sparse dynamic mode decomposition, Center Turbulence Res. Annual Res. Briefs, 2012 (2012), 139–152. |
[14] | I. Kevrekidis, C. Rowley, M. Williams, A kernel-based method for data-driven Koopman spectral analysis, J. Comput. Dynam., 2 (2016), 247–265. |
[15] |
M. Williams, I. Kevrekidis, C. Rowley, A data–driven approximation of the koopman operator: Extending dynamic mode decomposition, J. Nonlinear Sci., 25 (2015), 1307–1346. https://doi.org/10.1007/s00332-015-9258-5 doi: 10.1007/s00332-015-9258-5
![]() |
[16] |
Q. Li, F. Dietrich, E. Bollt, I. Kevrekidis, Extended dynamic mode decomposition with dictionary learning: A data-driven adaptive spectral decomposition of the Koopman operator, Chaos: An Interdisciplinary J. Nonlinear Sci., 27 (2017), 103111. https://doi.org/10.1063/1.4993854 doi: 10.1063/1.4993854
![]() |
[17] | E. Kaiser, J. Kutz, S. Brunton, Data-driven approximations of dynamical systems operators for control, The Koopman Operator In Systems And Control: Concepts, Methodologies, And Applications, (2020), 197–234. https://doi.org/10.1007/978-3-030-35713-9_8 |
[18] |
I. Mezić, Analysis of fluid flows via spectral properties of the Koopman operator, Annual Rev. Fluid Mech., 45 (2013), 357–378. https://doi.org/10.1146/annurev-fluid-011212-140652 doi: 10.1146/annurev-fluid-011212-140652
![]() |
[19] | P. Gaspard, Chaos, scattering and statistical mechanics, Chaos, 2005. |
[20] | R. Abraham, J. Marsden, Foundations of mechanics, American Mathematical Soc., 2008. https://doi.org/10.1090/chel/364 |
[21] | A. Ackleh, E. Allen, R. Kearfott, P. Seshaiyer, Classical and modern numerical analysis: Theory, methods and practice, Crc Press, 2009. https://doi.org/10.1201/b12332 |
[22] | D. Floryan, M. Graham, Charts and atlases for nonlinear data-driven models of dynamics on manifolds, arXiv Preprint arXiv: 2108.05928, (2021). |
[23] |
C. Fefferman, S. Mitter, H. Narayanan, Testing the manifold hypothesis, J. Am. Math. Soc., 29 (2016), 983–1049. https://doi.org/10.1090/jams/852 doi: 10.1090/jams/852
![]() |
[24] | H. Narayanan, S. Mitter, Sample complexity of testing the manifold hypothesis, Adv. Neural Inf. Process. Syst., 23 (2010). |
[25] |
A. Izenman, Introduction to manifold learning, Wires. Comput. Stat., 4 (2012), 439–446. https://doi.org/10.1002/wics.1222 doi: 10.1002/wics.1222
![]() |
[26] |
J. Tenenbaum, V. Silva, J. Langford, A global geometric framework for nonlinear dimensionality reduction, Science, 290 (2000), 2319–2323. https://doi.org/10.1126/science.290.5500.2319 doi: 10.1126/science.290.5500.2319
![]() |
[27] |
S. Roweis, L. Saul, Nonlinear dimensionality reduction by locally linear embedding, Science, 290 (2000), 2323–2326. https://doi.org/10.1126/science.290.5500.2323 doi: 10.1126/science.290.5500.2323
![]() |
[28] |
M. Balasubramanian, E. Schwartz, The isomap algorithm and topological stability, Science, 295 (2002), 7. https://doi.org/10.1126/science.295.5552.7a doi: 10.1126/science.295.5552.7a
![]() |
[29] | M. Belkin, P. Niyogi, Laplacian eigenmaps and spectral techniques for embedding and clustering, Adv. Neural Inf. Process. Syst., 14, 2001. https://doi.org/10.7551/mitpress/1120.003.0080 |
[30] |
Z. Ma, Z. Zhan, Z. Feng, J. Guo, Manifold learning based on straight-like geodesics and local coordinates, IEEE T. Neural Net. Lear., 32 (2020), 4956–4970. https://doi.org/10.1109/TNNLS.2020.3026426 doi: 10.1109/TNNLS.2020.3026426
![]() |
[31] | W. Boothby, W. Boothby, An introduction to differentiable manifolds and Riemannian geometry, Revised, Gulf Professional Publishing, 2003. |
[32] |
X. Chen, J. Weng, W. Lu, J. Xu, J. Weng, Deep manifold learning combined with convolutional neural networks for action recognition, IEEE T. Neural Net. Lear., 29 (2017), 3938–3952. https://doi.org/10.1109/TNNLS.2017.2740318 doi: 10.1109/TNNLS.2017.2740318
![]() |
[33] |
R. Wang, X. Wu, J. Kittler, Symnet: A simple symmetric positive definite manifold deep learning method for image set classification, IEEE T. Neural Net. Lear., 33 (2021), 2208–2222. https://doi.org/10.1109/TNNLS.2020.3044176 doi: 10.1109/TNNLS.2020.3044176
![]() |
[34] |
K. Lee, K. Carlberg, Model reduction of dynamical systems on nonlinear manifolds using deep convolutional autoencoders, J. Comput. Phys., 404 (2020), 108973. https://doi.org/10.1016/j.jcp.2019.108973 doi: 10.1016/j.jcp.2019.108973
![]() |
[35] |
J. Bakarji, K. Champion, J. Nathan Kutz, S. L. Brunton, Discovering governing equations from partial measurements with deep delay autoencoders, P Royal Soc. A, 479 (2023), 20230422. https://doi.org/10.1098/rspa.2023.0422 doi: 10.1098/rspa.2023.0422
![]() |
[36] | Y. LeCun, PhD thesis: Modeles connexionnistes de l'apprentissage (connectionist learning models), (Universite P. et M. Curie (Paris 6), 1987. |
[37] | J. Zhai, S. Zhang, J. Chen, Q. He, Autoencoder and its various variants, 2018 IEEE International Conference On Systems, Man, And Cybernetics (SMC), (2018), 415–419. https://doi.org/10.1109/SMC.2018.00080 |
[38] |
S. Gu, B. Kelly, D. Xiu, Autoencoder asset pricing models, J. Econometrics, 222 (2021), 429–450. https://doi.org/10.1016/j.jeconom.2020.07.009 doi: 10.1016/j.jeconom.2020.07.009
![]() |
[39] | C. Bishop, N. Nasrabadi, Pattern recognition and machine learning, Springer, 2006. |
[40] | B. Karlik, A. Olgac, Performance analysis of various activation functions in generalized MLP architectures of neural networks, Int. J. Artif. Intell. Expert Syst., 1 (2011), 111–122. |
[41] |
P. Pant, R. Doshi, P. Bahl, A. Barati Farimani, Deep learning for reduced order modelling and efficient temporal evolution of fluid simulations, Phys. Fluids, 33 (2021), 107101. https://doi.org/10.1063/5.0062546 doi: 10.1063/5.0062546
![]() |
[42] |
Z. Bai, Krylov subspace techniques for reduced-order modeling of large-scale dynamical systems, Appl. Numer. Math., 43 (2002), 9–44. https://doi.org/10.1016/S0168-9274(02)00116-2 doi: 10.1016/S0168-9274(02)00116-2
![]() |
[43] |
D. Lucia, P. Beran, W. Silva, Reduced-order modeling: new approaches for computational physics, Prog. Aerosp. Sci., 40 (2004), 51–117. https://doi.org/10.1016/j.paerosci.2003.12.001 doi: 10.1016/j.paerosci.2003.12.001
![]() |
[44] |
N. Kazantzis, C. Kravaris, L. Syrou, A new model reduction method for nonlinear dynamical systems, Nonlinear Dynam., 59 (2010), 183–194. https://doi.org/10.1007/s11071-009-9531-y doi: 10.1007/s11071-009-9531-y
![]() |
[45] |
O. San, R. Maulik, Neural network closures for nonlinear model order reduction, Adv. Comput. Math., 44 (2018), 1717–1750. https://doi.org/10.1007/s10444-018-9590-z doi: 10.1007/s10444-018-9590-z
![]() |
[46] | R. Fu, D. Xiao, I. Navon, F. Fang, L. Yang, C. Wang, et al., A non-linear non-intrusive reduced order model of fluid flow by auto-encoder and self-attention deep learning methods, Int. J. Numer. Meth. Eng., (2023). https://doi.org/10.1002/nme.7240 |
[47] |
N. Aubry, P. Holmes, J. Lumley, E. Stone, The dynamics of coherent structures in the wall region of a turbulent boundary layer, J. Fluid Mech., 192 (1988), 115–173. https://doi.org/10.1017/S0022112088001818 doi: 10.1017/S0022112088001818
![]() |
[48] |
G. Berkooz, P. Holmes, J. Lumley, The proper orthogonal decomposition in the analysis of turbulent flows, Annu. Rev. Fluid Mech., 25 (1993), 539–575. https://doi.org/10.1146/annurev.fl.25.010193.002543 doi: 10.1146/annurev.fl.25.010193.002543
![]() |
[49] | P. Holmes, J. Lumley, G. Berkooz, C. Rowley, Turbulence, coherent structures, dynamical systems and symmetry, Cambridge university press, 2012. https://doi.org/10.1017/CBO9780511919701 |
[50] |
H. Hotelling, Analysis of a complex of statistical variables into principal components, J. Educ. Psychol., 24 (1933), 417–441. https://psycnet.apa.org/doi/10.1037/h0071325 doi: 10.1037/h0071325
![]() |
[51] | E. Lorenz, Empirical orthogonal functions and statistical weather prediction, Massachusetts Institute of Technology, Department of Meteorology Cambridge, 1956. |
[52] | M. Loeve, Probability theory: foundations, random sequences, New York, NY: Van Nostrand, 1955. |
[53] |
K. Taira, S. Brunton, S. Dawson, C. Rowley, T. Colonius, B. McKeon, et al., Modal analysis of fluid flows: An overview, Aiaa J., 55 (2017), 4013–4041. https://doi.org/10.2514/1.J056060 doi: 10.2514/1.J056060
![]() |
[54] |
P. Schmid, L. Li, M. Juniper, O. Pust, Applications of the dynamic mode decomposition, Theor. Comp. Fluid Dyn., 25 (2011), 249–259. https://doi.org/10.1007/s00162-010-0203-9 doi: 10.1007/s00162-010-0203-9
![]() |
[55] |
B. Brunton, L. Johnson, J. Ojemann, J. Kutz, Extracting spatial–temporal coherent patterns in large-scale neural recordings using dynamic mode decomposition, J. Neurosci. Meth., 258 (2016), 1–15. https://doi.org/10.1016/j.jneumeth.2015.10.010 doi: 10.1016/j.jneumeth.2015.10.010
![]() |
[56] | E. Berger, M. Sastuba, D. Vogt, B. Jung, H. Amor, Dynamic mode decomposition for perturbation estimation in human robot interaction, The 23rd IEEE International Symposium On Robot And Human Interactive Communication, (2014), 593–600. https://doi.org/10.1109/ROMAN.2014.6926317 |
[57] |
B. Koopman, Hamiltonian systems and transformation in Hilbert space, P. Natl. Acad. Sci., 17 (1931), 315–318. https://doi.org/10.1073/pnas.17.5.315 doi: 10.1073/pnas.17.5.315
![]() |
[58] |
E. Bollt, Q. Li, F. Dietrich, I. Kevrekidis, On matching, and even rectifying, dynamical systems through Koopman operator eigenfunctions, SIAM J. Appl. Dyn. Syst., 17 (2018), 1925–1960. https://doi.org/10.1137/17M116207X doi: 10.1137/17M116207X
![]() |
[59] | T. Kanamaru, Van der Pol oscillator, Scholarpedia, 2007. Available from: http://www.scholarpedia.org/article/Van_der_Pol_oscillator |
[60] |
I. Triandaf, I. Schwartz, Karhunen-Loeve mode control of chaos in a reaction-diffusion process, Phys. Rev. E, 56 (1997), 204–212. https://doi.org/10.1103/PhysRevE.56.204 doi: 10.1103/PhysRevE.56.204
![]() |
[61] | H. Goldstein, C. Poole, J. Safko, Classical mechanics, American Association of Physics Teachers, 2002. |
[62] | F. Takens, Detecting strange attractors in turbulence, Dynamical Systems And Turbulence, Warwick 1980: Proceedings Of A Symposium Held At The University Of Warwick 1979/80, (2006), 366–381. https://doi.org/10.1007/BFb00919 |
[63] | D. Ruelle, F. Takens, On the nature of turbulence, Les Rencontres Physiciens-mathématiciens De Strasbourg-RCP25, 12 (1971), 1–44. |
[64] | K. Falconer, Fractal geometry: Mathematical foundations and applications, John Wiley & Sons, 2004. 10.1002/0470013850 |
[65] | M. Adachi, Embeddings and immersions, American Mathematical Soc., 2012. https://doi.org/10.1090/mmono/124 |
[66] |
A. Skopenkov, Embedding and knotting of manifolds in Euclidean spaces, London Math. Soc. Lecture Note Series, 347 (2008), 248. https://doi.org/10.1017/CBO9780511666315.008 doi: 10.1017/CBO9780511666315.008
![]() |
1. | Narges Peykrayegan, Mehdi Ghovatmand, Mohammad Hadi Noori Skandari, Dumitru Baleanu, An approximate approach for fractional singular delay integro-differential equations, 2022, 7, 2473-6988, 9156, 10.3934/math.2022507 | |
2. | Mehboob Alam, Akbar Zada, Implementation of q-calculus on q-integro-differential equation involving anti-periodic boundary conditions with three criteria, 2022, 154, 09600779, 111625, 10.1016/j.chaos.2021.111625 | |
3. | Binlin Zhang, Rafia Majeed, Mehboob Alam, On Fractional Langevin Equations with Stieltjes Integral Conditions, 2022, 10, 2227-7390, 3877, 10.3390/math10203877 | |
4. | Mehboob Alam, Akbar Zada, Usman Riaz, On a Coupled Impulsive Fractional Integrodifferential System with Hadamard Derivatives, 2022, 21, 1575-5460, 10.1007/s12346-021-00535-0 | |
5. | Mehboob Alam, Dildar Shah, Hyers–Ulam stability of coupled implicit fractional integro-differential equations with Riemann–Liouville derivatives, 2021, 150, 09600779, 111122, 10.1016/j.chaos.2021.111122 | |
6. | Akbar Zada, Mehboob Alam, Khansa Hina Khalid, Ramsha Iqbal, Ioan-Lucian Popa, Analysis of Q-Fractional Implicit Differential Equation with Nonlocal Riemann–Liouville and Erdélyi-Kober Q-Fractional Integral Conditions, 2022, 21, 1575-5460, 10.1007/s12346-022-00623-9 | |
7. | Ravi Agarwal, Snezhana Hristova, Donal O'Regan, Integral presentations of the solution of a boundary value problem for impulsive fractional integro-differential equations with Riemann-Liouville derivatives, 2022, 7, 2473-6988, 2973, 10.3934/math.2022164 | |
8. | Mehboob Alam, Aftab Khan, Muhammad Asif, Analysis of implicit system of fractional order via generalized boundary conditions, 2023, 0170-4214, 10.1002/mma.9139 | |
9. | Mehboob Alam, Khansa Hina Khalid, Analysis of q ‐fractional coupled implicit systems involving the nonlocal Riemann–Liouville and Erdélyi–Kober q ‐fractional integral conditions , 2023, 0170-4214, 10.1002/mma.9208 | |
10. | Rafia Majeed, Binlin Zhang, Mehboob Alam, Fractional Langevin Coupled System with Stieltjes Integral Conditions, 2023, 11, 2227-7390, 2278, 10.3390/math11102278 | |
11. | Mehboob Alam, Akbar Zada, Sumbel Begum, Usman Riaz, Analysis of Fractional Integro-differential System with Impulses, 2023, 9, 2349-5103, 10.1007/s40819-023-01584-6 | |
12. | Akbar Zada, Usman Riaz, Junaid Jamshed, Mehboob Alam, Afef Kallekh, Analysis of impulsive Caputo fractional integro‐differential equations with delay, 2024, 0170-4214, 10.1002/mma.10426 |
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 19.595 | 17.996 | 19.076 | 17.640 | 17.595 | 17.519 |
SSIM | 0.999 | 0.998 | 0.999 | 0.998 | 0.999 | 0.999 | |
MSE | 0.011 | 0.015 | 0.012 | 0.017 | 0.017 | 0.017 | |
CPU | 234.78 | 335.37 | 55.00 | 94.755 | 11.581 | 7.543 | |
Phantom | PSNR | 73.636 | 72.694 | 73.639 | 73.655 | 72.377 | 72.430 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 266.67 | 279.43 | 56.186 | 75.168 | 12.496 | 8.556 | |
Mri | PSNR | 29.319 | 28.936 | 28.068 | 29.023 | 28.057 | 28.112 |
SSIM | 0.969 | 0.975 | 0.924 | 0.919 | 0.962 | 0.978 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 259.86 | 375.48 | 60.963 | 162.06 | 19.340 | 15.563 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.595 | 17.519 | 17.611 | 17.732 | 17.894 | 17.876 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | |
CPU | 11.581 | 7.543 | 18.027 | 14.207 | 3.196 | 2.880 | |
Phantom | PSNR | 72.377 | 72.430 | 72.492 | 72.532 | 72.486 | 72.760 |
SSIM | 1.000 | 1.000 | 1.000 | 1.0000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 12.496 | 8.556 | 43.106 | 38.075 | 20.874 | 20.023 | |
Mri | PSNR | 28.057 | 28.112 | 28.081 | 28.234 | 28.161 | 28.506 |
SSIM | 0.962 | 0.978 | 0.901 | 0.899 | 0.895 | 0.891 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 19.340 | 15.563 | 42.535 | 37.575 | 18.237 | 17.539 |
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 20.122 | 19.195 | 18.419 | 17.594 | 17.519 | 17.552 |
SSIM | 0.999 | 0.998 | 0.998 | 0.998 | 0.999 | 0.999 | |
MSE | 0.009 | 0.012 | 0.014 | 0.017 | 0.017 | 0.017 | |
CPU | 3.651 | 6.916 | 1.433 | 3.596 | 1.698 | 0.723 | |
Phantom | PSNR | 73.904 | 72.475 | 74.040 | 73.704 | 72.619 | 72.464 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 68.434 | 129.151 | 19.739 | 35.465 | 7.776 | 6.878 | |
Mri | PSNR | 30.615 | 29.072 | 29.289 | 28.899 | 28.080 | 28.169 |
SSIM | 0.912 | 0.900 | 0.893 | 0.884 | 0.892 | 0.905 | |
MSE | 0 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 63.758 | 122.983 | 28.132 | 31.338 | 9.451 | 8.641 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.519 | 17.552 | 17.748 | 17.762 | 17.860 | 17.805 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | 0.016 | |
CPU | 1.698 | 0.723 | 14.582 | 11.879 | 2.005 | 1.672 | |
Phantom | PSNR | 72.619 | 72.464 | 72.398 | 72.629 | 72.695 | 72.353 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 7.776 | 6.878 | 47.513 | 45.431 | 13.824 | 13.757 | |
Mri | PSNR | 28.080 | 28.169 | 28.241 | 28.093 | 28.241 | 28.328 |
SSIM | 0.892 | 0.905 | 0.904 | 0.904 | 0.902 | 0.897 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 9.451 | 8.641 | 54.802 | 38.773 | 11.533 | 10.332 |
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 19.595 | 17.996 | 19.076 | 17.640 | 17.595 | 17.519 |
SSIM | 0.999 | 0.998 | 0.999 | 0.998 | 0.999 | 0.999 | |
MSE | 0.011 | 0.015 | 0.012 | 0.017 | 0.017 | 0.017 | |
CPU | 234.78 | 335.37 | 55.00 | 94.755 | 11.581 | 7.543 | |
Phantom | PSNR | 73.636 | 72.694 | 73.639 | 73.655 | 72.377 | 72.430 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 266.67 | 279.43 | 56.186 | 75.168 | 12.496 | 8.556 | |
Mri | PSNR | 29.319 | 28.936 | 28.068 | 29.023 | 28.057 | 28.112 |
SSIM | 0.969 | 0.975 | 0.924 | 0.919 | 0.962 | 0.978 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 259.86 | 375.48 | 60.963 | 162.06 | 19.340 | 15.563 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.595 | 17.519 | 17.611 | 17.732 | 17.894 | 17.876 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | |
CPU | 11.581 | 7.543 | 18.027 | 14.207 | 3.196 | 2.880 | |
Phantom | PSNR | 72.377 | 72.430 | 72.492 | 72.532 | 72.486 | 72.760 |
SSIM | 1.000 | 1.000 | 1.000 | 1.0000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 12.496 | 8.556 | 43.106 | 38.075 | 20.874 | 20.023 | |
Mri | PSNR | 28.057 | 28.112 | 28.081 | 28.234 | 28.161 | 28.506 |
SSIM | 0.962 | 0.978 | 0.901 | 0.899 | 0.895 | 0.891 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 19.340 | 15.563 | 42.535 | 37.575 | 18.237 | 17.539 |
GBK | SSRBK | SLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 20.122 | 19.195 | 18.419 | 17.594 | 17.519 | 17.552 |
SSIM | 0.999 | 0.998 | 0.998 | 0.998 | 0.999 | 0.999 | |
MSE | 0.009 | 0.012 | 0.014 | 0.017 | 0.017 | 0.017 | |
CPU | 3.651 | 6.916 | 1.433 | 3.596 | 1.698 | 0.723 | |
Phantom | PSNR | 73.904 | 72.475 | 74.040 | 73.704 | 72.619 | 72.464 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 68.434 | 129.151 | 19.739 | 35.465 | 7.776 | 6.878 | |
Mri | PSNR | 30.615 | 29.072 | 29.289 | 28.899 | 28.080 | 28.169 |
SSIM | 0.912 | 0.900 | 0.893 | 0.884 | 0.892 | 0.905 | |
MSE | 0 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 63.758 | 122.983 | 28.132 | 31.338 | 9.451 | 8.641 |
SLFGBK | MSRBK | MLFGBK | |||||
data | Metrics | η=0.8 | η=0.9 | η=0.8 | η=0.9 | η=0.8 | η=0.9 |
Cameraman | PSNR | 17.519 | 17.552 | 17.748 | 17.762 | 17.860 | 17.805 |
SSIM | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | 0.999 | |
MSE | 0.017 | 0.017 | 0.016 | 0.016 | 0.016 | 0.016 | |
CPU | 1.698 | 0.723 | 14.582 | 11.879 | 2.005 | 1.672 | |
Phantom | PSNR | 72.619 | 72.464 | 72.398 | 72.629 | 72.695 | 72.353 |
SSIM | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 | |
MSE | 0 | 0 | 0 | 0 | 0 | 0 | |
CPU | 7.776 | 6.878 | 47.513 | 45.431 | 13.824 | 13.757 | |
Mri | PSNR | 28.080 | 28.169 | 28.241 | 28.093 | 28.241 | 28.328 |
SSIM | 0.892 | 0.905 | 0.904 | 0.904 | 0.902 | 0.897 | |
MSE | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | 0.001 | |
CPU | 9.451 | 8.641 | 54.802 | 38.773 | 11.533 | 10.332 |