In this paper, a new hybrid singular value thresholding with diagonal-modify algorithm based on the augmented Lagrange multiplier (ALM) method was proposed for low-rank matrix recovery, in which only part singular values were treated by a hybrid threshold operator with diagonal-update, and which allowed the algorithm to make use of simple arithmetic operation and keep the computational cost of each iteration low. The new algorithm decreased the complexity of the singular value decomposition and shortened the computing time. The convergence of the new algorithm was discussed. Finally, numerical experiments shown that the new algorithm greatly improved the solving efficiency of a matrix recovery problem and saved the calculation cost, and its effect was obviously better than that of the other algorithms mentioned in experiments.
Citation: Ruiping Wen, Liang Zhang, Yalei Pei. A hybrid singular value thresholding algorithm with diagonal-modify for low-rank matrix recovery[J]. Electronic Research Archive, 2024, 32(11): 5926-5942. doi: 10.3934/era.2024274
Related Papers:
[1]
Yang Song, Beiyan Yang, Jimin Wang .
Stability analysis and security control of nonlinear singular semi-Markov jump systems. Electronic Research Archive, 2025, 33(1): 1-25.
doi: 10.3934/era.2025001
[2]
Ruini Zhao .
Nanocrystalline SEM image restoration based on fractional-order TV and nuclear norm. Electronic Research Archive, 2024, 32(8): 4954-4968.
doi: 10.3934/era.2024228
[3]
Natália Bebiano, João da Providência, Wei-Ru Xu .
Approximations for the von Neumann and Rényi entropies of graphs with circulant type Laplacians. Electronic Research Archive, 2022, 30(5): 1864-1880.
doi: 10.3934/era.2022094
Guanrong Li, Yanping Chen, Yunqing Huang .
A hybridized weak Galerkin finite element scheme for general second-order elliptic problems. Electronic Research Archive, 2020, 28(2): 821-836.
doi: 10.3934/era.2020042
[6]
Zhengyu Liu, Yufei Bao, Changhai Wang, Xiaoxiao Chen, Qing Liu .
A fast matrix completion method based on truncated $ {\mathit{L}}_{2, 1} $ norm minimization. Electronic Research Archive, 2024, 32(3): 2099-2119.
doi: 10.3934/era.2024095
[7]
Yingpin Chen, Kaiwei Chen .
Four mathematical modeling forms for correlation filter object tracking algorithms and the fast calculation for the filter. Electronic Research Archive, 2024, 32(7): 4684-4714.
doi: 10.3934/era.2024213
[8]
Yongge Tian .
Characterizations of matrix equalities involving the sums and products of multiple matrices and their generalized inverse. Electronic Research Archive, 2023, 31(9): 5866-5893.
doi: 10.3934/era.2023298
[9]
Yan Xia, Songhua Wang .
Global convergence in a modified RMIL-type conjugate gradient algorithm for nonlinear systems of equations and signal recovery. Electronic Research Archive, 2024, 32(11): 6153-6174.
doi: 10.3934/era.2024286
[10]
Tianbao Liu, Lingling Yang, Yue Li, Xiwen Qin .
An improved dung beetle optimizer based on Padé approximation strategy for global optimization and feature selection. Electronic Research Archive, 2025, 33(3): 1693-1762.
doi: 10.3934/era.2025079
Abstract
In this paper, a new hybrid singular value thresholding with diagonal-modify algorithm based on the augmented Lagrange multiplier (ALM) method was proposed for low-rank matrix recovery, in which only part singular values were treated by a hybrid threshold operator with diagonal-update, and which allowed the algorithm to make use of simple arithmetic operation and keep the computational cost of each iteration low. The new algorithm decreased the complexity of the singular value decomposition and shortened the computing time. The convergence of the new algorithm was discussed. Finally, numerical experiments shown that the new algorithm greatly improved the solving efficiency of a matrix recovery problem and saved the calculation cost, and its effect was obviously better than that of the other algorithms mentioned in experiments.
1.
Introduction
The low-rank matrix recovery, also known as robust PCA or sparse and low-rank matrix decomposition, means to find the lowest rank matrices based on fewer linear measurements, arise in many fields such as collaborative filtering [1,2,3], machine learning [4], picture alignment [5,6], signal processing [7], quantum state tomography [8] and more.
The problem of a low-rank matrix recovery was first introduced in [9,10,11,12], which can be regarded as a matrix-analogue of compressed sensing [13] refers to automatically identify the corrupted elements and recover the original matrix when some elements of the matrix were seriously missing. Wright et al. [14] had proved that the low-rank matrix A can be exactly recovered from matrix D by solving the following convex optimization problem,
minA,E‖A‖∗+λ‖E‖1,s.t.D=A+E,
(1.1)
where ‖A‖∗:=∑rk=1σk(A),σk(A) denotes the k-th largest singular value of A∈Rn1×n2 with rank(A)=r, was called the nuclear norm of A. The sum of the absolute values of the matrix E entries was denoted as ‖E‖1, and λ was a positive weighting parameter.
As for the solution of the model (1.1), there are a lot of studies both from theoretical and algorithmic aspects. For example, the iterative threshold (IT) algorithm (see [14]) had been put forward, which used a soft threshold operator and linear Bremgan iteration to solve model (1.1). The de-capitalize iterative hard threshold (IHT) algorithm (see [15]) had been proposed based on the influence of hard threshold operator on compressed sensing. In summary, the accelerated proximal gradient (APG) method (see [16]), the singular value thresholding (SVT) algorithm (see [17]), the dual algorithm (see [18]), the augmented Lagrangian multiplier (ALM) algorithm (see [19]), and the hybrid augmented Lagrange multiplier (HALM) algorithm (see [20]) had been later presented to deal with the convex problem (1.1). Numerous details and derivations on a low-rank matrix recovery problem can be referred to the references given therein.
This paper aims mainly at establishing a new hybrid thresholding with a diagonal-modify iterative algorithm based on the ALM algorithm for recovering a low-rank matrix. By using a diagonal-modify technique, the sequence matrices generated by the new algorithm were approximated in the true solution well, which saves significant computational cost.
The rest of the paper was organized as follows: In Section 2, the notations used throughout this paper and related preliminaries studies were introduced. In Section 3, we presented the proposed matrix recovery algorithm in detail. The convergence analysis of the new algorithm was given in Section 4. Then, numerical experiments were shown in Section 5. Finally, we ended the paper with the concluding remarks in Section 6.
2.
Notations and preliminaries
In this position, we provide some basic notations and preliminaries that were used in our analysis. Rn1×n2 denotes the set of n1×n2 real matrices; AT is used to express the transpose of a matrix A. diag(d11,d22,…,dnn) stands for a diagonal matrix with the diagonal elements d11,d22,…,dnn. ‖A‖1 denotes the sum of the absolute values of matrix A entries. ‖A‖2 represents the spectral norm, the square root of the maximum eigenvalue of ATA, and the Frobenius norm ‖A‖F=√n2∑j=1n1∑i=1a2ij. The tr(A) represents the trace of A, and the standard inner product between two matrices is denoted by ⟨X,Y⟩=tr(XTY).
Definition 1. (Singular Value Decomposition (SVD) [21]) The singular value decomposition (SVD) of a r-rank matrix A∈Rn1×n2 is defined by
A=UΣrVT,Σr=diag(σ1,σ2,…,σr),
where U∈Rn1×r and V∈Rn2×r are both column orthogonal matrices, and σ1≥σ2≥…≥σr>0 are all positive singular values of A. By the way, ‖A‖∗:=∑rk=1σk(A) denotes the nuclear norm of A.
Definition 2. For each τ≥0, the soft thresholding (shrinkage) operator (see [17]) Sτ was defined by
Sτ[σ]={σ−τ,if|σ|≥τ0,if|σ|<τ,
the hard thresholding operator (see [22]) ητ was defined by
ητ[σ]={σ,if|σ|≥τ0,if|σ|<τ,
where σ∈R, and τ is called a thresholding.
Definition 3. (Hybrid Singular Value Thresholding Operator [20]) Let X=UΣrVT∈Rn1×n2 be the SVD of X mentioned above. For each τ≥0,z>1, the hybrid thresholding operator Hτ,z is defined by
Moreover, Sτ(A):=USτ(Σ)VT and ητ(A):=Uητ(Σ)VT based on definition 2.
3.
Description of algorithms
In this section, a new fast algorithm was proposed after introducing a related algorithm for solving the problem (1.1).
3.1. The augmented Lagrange multipliers (ALM) algorithm
It is well-known that the partial augmented Lagrange function of (1.1) is
L(A,E,Y,μ)=‖A‖∗+λ‖E‖1+⟨Y,D−A−E⟩+μ2‖D−A−E‖2F,
where A,E,Y∈Rn1×n2, μ>0 is the penalty parameter. It is reported that the algorithm of augmented Lagrange multipliers has been applied to the low-rank matrix recovery problem. It is of much better numerical behavior, and it is also of much higher accuracy. Then the augmented Lagrange multipliers algorithm is summarized in the following Algorithm 3.1.
Step 0: Given a sampled matrix D=A+E, parameters μ0>0, ρ>1. Given also two initial matrices Y0=0,E0=0,k:=0;
Step 1: Solve Ak=argminAL(A,Ek,Yk,μk), compute the SVD of the matrix (D−Ek+μ−1kYk),
[Uk,Σk,Vk]=svd(D−Ek+μ−1kYk);
Step 2: Set
Ak+1=UkSμ−1k(Σk)VTk.
Solves Ek+1=argminL(Ak+1,E,Yk,μk),
Ek+1=Sμ−1k(D−Ak+1+μ−1kYk);
Step 3: If ‖D−Ak+1−Ek+1‖F/‖D‖F<ϵ1 and μk‖Ek+1−Ek‖F/‖D‖F<ϵ2, stop; otherwise, go to next Step;
Step 4 : Set Yk+1=Yk+μk(D−Ak+1−Ek+1),
If μk‖Ek+1−Ek‖F/‖D‖F<ϵ2, set μk+1=ρμk; otherwise, go to Step 1.
For convenience, [Uk,Σk,Vk]=svd(⋅) denotes the SVD of the corresponding matrix.
3.2. The hybrid augmented Lagrange multiplier algorithm with diagonal-modify
Because the soft thresholding operator compressed the thresholding in a constant way and maybe lost some effective large thresholdings, while the hard thresholding operator was discontinuous, which reduced the smoothness of the matrix. To complement the advantages of the two operators, a hybrid singular value threshold operator had been designed. Based on the augmented Lagrange multiplier algorithm, the new algorithm employs the hybrid singular threshold operator with diagonal-modify H to deal with the singular value in each iteration.
The details of the so-called hybrid singular value threshold operator: keep some large singular values unchanged, compress some middle size singular values in a constant way, and take some small singular values as 0. This operator can make up for the deficiency of a single soft or hard threshold operator, and improve the accuracy of the recovered matrix partially. Further, a diagonal-modify W(A)=AW was used to improve its recovery efficiency at the kth iteration, where the diagonal matrix Wk=diag(w11(k),w22(k),…,wnn(k))∈Rn×n can be obtained by
Wk=argmin‖A−AkWk‖F.
(3.1)
In fact, Eq (3.1) is easy to compute since it is so simple just some arithmetic operation required, without extra cost. In fact, the exact solution of (3.1) is given by
wjj(k)=⟨A(:,j),Ak(:,j)⟩⟨Ak(:,j),Ak(:,j)⟩,j=1,…,n.
The new algorithm can be designed as follows:
The difference with the classical ALM algorithm focuses on Steps 2 and 3. Moreover, Agorithm 3.2 includes the HALM algorithm in [20] as a special case when W=I.
4.
Convergence analysis
In this section, we presented briefly the convergence of Algorithm 3.2 by analyzing the properties of the sequences {Ak},{Ek}, and {Yk}. Since the D-HALM algorithm was a progression of the classical ALM algorithm, its proof is a trivial generalization. For details of proofs and techniques, one can refer to [19] and references given therein.
Algorithm 3.2. (Hybrid augmented Lagrange multiplier algorithm with diagonal-modify, denoted by D-HALM)
Step 0: Given a sampled matrix D=A+E. Set parameters k>0,ρ>1,ϵ1,ϵ2, the initial matrices Y0=0,E0=0,k:=0;
Step 1: Solve Ak=argminAL(A,Ek,Yk,μk), compute the SVD of the matrix (D−Ek+μ−1kYk),
[Uk,Σk,Vk]=svd(D−Ek+μ−1kYk);
Step 2: Set
˜Ak+1=UkHμ−1k,z(Σk)VTk,Wk=argmin‖A−˜Ak+1Wk‖F
Step 3: Set
Ak+1=Ak+1Wk.
Solves Ek+1=argminL(Ak+1,E,Yk,μk),
Ek+1=Hμ−1k,z(D−Ak+1+μ−1kYk);
Step 4: If ‖D−Ak+1−Ek+1‖F/‖D‖F<ϵ1 and μk‖Ek+1−Ek‖F/‖D‖F<ϵ2, stop, otherwise, go to next step;
Step 5: Set Yk+1=Yk+μk(D−Ak+1−Ek+1), if μk‖Ek+1−Ek‖F/‖D‖F<ϵ2, set μk+1=ρμk,
otherwise, go to Step 1.
Lemma 4.1. (see [17]) Let A∈Rn1×n2 be a matrix and UΣVT be its SVD. Then the subgradients set of the nuclear norm of A is given by
is nonnegative, and the series is convergent provided that μk is nondecreasing.
In Algorithm 3.2, we do not have to solve the sub-problem
(Ak,Ek)=argminA,EL(A,Ek,Yk,μk)
exactly. Rather, updating them once when solving this sub-problem is sufficient for Ak and Ek to convergence the local optimal solution of the problem (1). And this leads to an inexact ALM with diagonal-modify, similar to the IALM introduced in [19]. After the analysis on Ak,Ek, and Yk via these Lemmas above, the validity and optimality of Algorithm 3.2 are guaranteed by the following theorem (see [19]).
Theorem 4.1. Suppose that μk is nondecreasing and the series ∞∑i=1μ−1k diverges, then the sequence {(Ak,Ek)} converges to the optimal solution (A∗,E∗) of problem (1.1).
Proof. Because of Y∗∈∂‖A∗‖∗,Y∗∈∂(‖λE∗‖1),Yk+1∈∂(‖λEk+1‖1), and ¯Yk+1=∂‖Ak+1‖∗−∂AL(Ak+1,Ek,Yk,μk), we have
Thus, ‖Ek−E∗‖2+μ−2k‖Yk−Y∗‖2 is non-increasing, as shown in [19]. From the Theorem 2 in [19], it holds
limk→∞Ak=A∗,limk→∞Ek=E∗,
we obtain that (A∗,E∗) is the solution of the low-rank matrix recovery problem (1.1), which completes the proof.
5.
Numerical experiments
In this section, some random data and video sequences were used to demonstrate the proposed algorithm was feasible and effective to solve a low-rank matrix recovery problem (1.1). All the experiments are performed under Windows 11 and MATLAB R2019a running on a DELL laptop.
5.1. Random data
Some original results of four algorithms (APG, ALM, HALM, and D-HALM) were provided for the n×n matrices with different ranks for the problem (1.1). We conduct numerical experiments on the same workstation. By analyzing and comparing iteration numbers (denoted by IT), computing time in seconds (denoted by CPU(s)), and deviation (Error 1, Error 2), defined by
Error1=‖¯A−A∗‖F‖A∗‖F,Error2=‖¯E−E∗‖F‖E∗‖F.
In the experiments, p=mn2 represents the sampling density, where m is the number of observed entries. We denote the optimal solution by the ordered pair (A∗,E∗), and the output by (¯A,¯E). For Algorithm 3.2 (D-HALM), we choose a fixed weighting parameter λ=1√n, set Y0=DJ(D)(J(D)=max(‖D‖2,λ−1‖D‖∞)[19], and we empirically set the parameter μ0=1.25‖D‖2,τ0=0.5‖D‖2,ϵ1=1×10−8,ϵ2=1×10−7,z=1.4. By the way, the ALM, HALM, and D-HALM algorithms share the same parameters.
Tests were conducted for n1=n2,p∈{0.1,0.2,0.3,0.4}, and the rank of the underlying matrix to be reconstructed, r=0.005n. The specific comparison results were shown in Tables 1–4, and the compared curve of the computing time for different algorithms was mapped in Figures 1 and 2.
Table 1.
Compared results of four algorithms for p=0.1.
From Figures 1 and 2, it can be seen that the time of the D-HALM algorithm was always less than other algorithms. The larger the matrix scale, the more obvious the effect.
As can be shown in Tables 1–4, the proposed algorithm, ALM, and HALM algorithms achieve almost the same iterations and relative errors. However, we can see the CPU of our algorithm was obviously less than that of APG, ALM, and HALM algorithms.
By introducing the speed-up (denoted by SP) as follows,
SP=time of the other algorithmtime of the D-HALM algorithm,
it is noted that the maximum SP values of APG, ALM and HALM algorithms were approximately 7.3, 2.6, and 1.6 when the matrix size n≤10000; however, those of them were approximately 6.1, 5.9 and 1.7. The computational cost of the proposed algorithm was significantly lower than the other algorithms.
5.2. Video sequences
In this subsection, we use four surveillance video sequences (video 1: Hall of a business building; video 2: Mall; video 3: Bootstrap; video 4: Fountain) with different resolutions, respectively, to compare the separated effects of three ADMM algorithms (ALM, HALM, and D-HALM). One of 200 images of four video sequences was randomly selected for testing. Figures 3–5 shown the results of foreground and background separation of four surveillance video sequences under the D=A+E model. The error and running time were reported in Table 5, where
Error=‖A+E−D‖F‖D‖F.
Figure 3.
Experiment results of ALM method on videos 1-4, respectively.
From Table 5, it can be seen that three algorithms shared similar separation accuracy while the running time of D-HALM was cost-effective.
6.
Conclusions
To solve the low-rank matrix recovery problem, we have proposed a hybrid singular value thresholding algorithm with diagonal-modify (D-HALM) based on the classical ALM algorithm. In the proposed algorithm, a hybrid singular value thresholding was used and a diagonal-modify was carried out in each iteration. The iterative sequence generated by Algorithm 3.2 for solving the optimization models converges to the optimal solution, which can be guaranteed by the traditional convergence theory. In experiments, compared with APG, ALM, and HALM algorithms introduced in the past three years. The experiments based on randomly numerical simulation data and video sequence separation have shown both that the proposed algorithm in this paper takes less time and gets more precision for solving the low-rank matrix recovery problem. It was found that the D-HALM algorithm was more efficient for solving low-rank matrix recovery problems than the other algorithms. In addition, it is worth mentioning that the new algorithm can further develop to solve the tensor completion problems.
Use of AI tools declaration
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Acknowledgments
The authors are very much indebted to the anonymous referees for their helpful comments and suggestions, which greatly improved the original manuscript of this paper. The authors are so thankful for the support from the NSF of China (12371381), the special fund for science and technology innovation teams of Shanxi province (202204051002018), and the Shanxi higher education science and technology innovation project (2023L243).
Conflict of interest
The authors declare that they have no conflict of interests.
References
[1]
A. Mongia, A. Majumdar, Matrix completion on multiple graphs: Application in collaborative filter- ing, Signal Process., 165 (2019), 144–148. https://doi.org/10.1016/j.sigpro.2019.07.002 doi: 10.1016/j.sigpro.2019.07.002
[2]
S. Daei, F. Haddadi, A. Amini, Exploitation prior information in block sprase signal, IEEE Trans. Signal Process., 99 (2018). https://doi.org/10.1109/TSP.2019.2931209 doi: 10.1109/TSP.2019.2931209
[3]
Z. Zhang, K. Zhao, H. Zha, Inducible regularization for low-rank matrix factorizations for collaborative filtering, Neurocomputing, 97 (2012), 52–62. https://doi.org/10.1016/j.neucom.2012.05.010 doi: 10.1016/j.neucom.2012.05.010
[4]
S. Li, Q. Li, Z. Zhu, G. Tang, M. B. Wakin, The global geometry of centralized and distributed low- rank matrix recovery without regularization, IEEE Signal Process. Lett., 27 (2020), 1400–1404. https://doi.org/10.1109/LSP.2020.3008876 doi: 10.1109/LSP.2020.3008876
[5]
B. J. Han, J. Y. Simg, Glass reflection removal using co-saliency based image alignment and low- rank matrix completion in gradient domain, IEEE Trans. Image Process., 27 (2018), 1–1. https://doi.org/10.1109/TIP.2018.2849880. doi: 10.1109/TIP.2018.2849880
[6]
L. I. Jie, F. Chen, Y. Wang, J. I. Fei, Y. U. Hua, Spatial spectrum estimation of incoherently dis tributed sources based on low-rank matrix recovery, IEEE Trans. Veh. Technol., 99 (2020). https://doi.org/10.1109/TVT.2020.2986783 doi: 10.1109/TVT.2020.2986783
[7]
Z. H. Zhu, Q. W. Tang, G. G. Wakin, B. Michael, Global optimality in low-rank matrix optimization, IEEE Trans. Signal Process., 66 (2018), 3614–3628. https://doi.org/10.1109/TSP.2018.2835403 doi: 10.1109/TSP.2018.2835403
[8]
M. Kech, M. Wolf, Constrained quantum tomography of semi-algebraic sets with applications to low-rank matrix recovery, Mathematics, 6 (2017), 171–195. https://doi.org/10.1093/imaiai/iaw019 doi: 10.1093/imaiai/iaw019
[9]
E. J. Candès, B. Recht, Exact low-rank matrix completion via convex optimization, in 2008 46th Annual Allerton Conference on Communication, Control, and Computing, (2009), 806–812. https://doi.org/10.1109/ALLERTON.2008.4797640
[10]
E. J. Candès, T. Tao, The power of convex relaxation: Near-optimal matrix completion, IEEE Trans. Inf. Theory, 56 (2010), 2053–2080. https://doi.org/10.1109/TIT.2010.2044061 doi: 10.1109/TIT.2010.2044061
[11]
B. Recht, A simpler approach to matrix completion, J. Mach. Learn. Res., 12 (2009), 3413–3430. https://doi.org/10.1177/0959651810397461 doi: 10.1177/0959651810397461
[12]
B. Recht, M. Fazel, P. A. Parrilo, Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization, SIAM Rev., 52 (2010), 471–501. https://doi.org/10.1137/070697835 doi: 10.1137/070697835
[13]
D. Gross, Recovering low-rank matrices from few coefficients in any basis, IEEE Trans Inf. Theory, 57 (2011), 1548–1566. https://doi.org/10.1109/TIT.2011.2104999 doi: 10.1109/TIT.2011.2104999
[14]
J. Wright, A. Ganesh, S. Rao, Y. Ma, Robust principal component analysis: exact recovery of corrupted low-rank matrices, in Neural Networks for Signal Processing X. Proceedings of the 2000 IEEE Signal Processing Society Workshop, 2009. https://doi.org/10.1109/NNSP.2000.889420
A. Ganesh, Z. Lin, J. Wright, L. Wu, M. Chen, Y. Ma, Fast algorithms for recovering a corrupted low-rank matrix, in 2009 3rd IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP), 2009. https://doi.org/10.1109/CAMSAP.2009.5413299.
[17]
J. F. Cai, E. J. Candès, Z. Shen, A singular value thresholding algorithm for matrix completion, SIAM J. Optim., 20 (2010), 1956–1982. https://doi.org/10.1137/080738970 doi: 10.1137/080738970
[18]
Z. Lin, A. Ganesh, J. Wright, L. Wu, M. Chen, Y. Ma, Fast convex optimization algorithms for exact recovery of a corrupted low-rank matrix, Coordinated Science Laboratory Report no. UILU-ENG-09-2214, DC-246, 2009. https://doi.org/10.1017/S0025315400020749
[19]
Z. Lin, M. Chen, Y. Ma, The augmented lagrange multiplier method for exact recovery of corrupted low-rank matrices, preprint, arXiv: 1009.5055. https://doi.org/10.48550/arXiv.1009.5055
[20]
J. Guo, R. P. Wen, Hybrid augmented lagrange multiplier algorithm for low-rank matrix com- pletion, Numer. Math.: Theory, Methods Appl., 44 (2022), 187–201.
[21]
G. H. Golub, C. F. Van Loan, Matrix Computations, 4th edition, Johns Hopkins Studies in Mathematical Sciences, 2013.
[22]
D. Goldfarb, S. Ma, Convergence of fixed-point continuation algorithms for matrix rank minimization, Found. Comput. Math., 11 (2011), 183–210. https://doi.org/10.1007/s10208-011-9084-6 doi: 10.1007/s10208-011-9084-6