A Review on Low-Rank Models in Data Analysis

  • Received: 01 October 2015 Revised: 01 December 2016 Published: 01 July 2016
  • Nowadays we are in the big data era. The high-dimensionality of data imposes big challenge on how to process them effectively and efficiently. Fortunately, in practice data are not unstructured. Their samples usually lie around low-dimensional manifolds and have high correlation among them. Such characteristics can be effectively depicted by low rankness. As an extension to the sparsity of first order data, such as voices, low rankness is also an effective measure for the sparsity of second order data, such as images. In this paper, I review the representative theories, algorithms and applications of the low rank subspace recovery models in data processing.

    Citation: Zhouchen Lin. A Review on Low-Rank Models in Data Analysis[J]. Big Data and Information Analytics, 2016, 1(2): 139-161. doi: 10.3934/bdia.2016001

    Related Papers:

    [1] Tao Wu, Yu Lei, Jiao Shi, Maoguo Gong . An evolutionary multiobjective method for low-rank and sparse matrix decomposition. Big Data and Information Analytics, 2017, 2(1): 23-37. doi: 10.3934/bdia.2017006
    [2] Guojun Gan, Kun Chen . A Soft Subspace Clustering Algorithm with Log-Transformed Distances. Big Data and Information Analytics, 2016, 1(1): 93-109. doi: 10.3934/bdia.2016.1.93
    [3] Ming Yang, Dunren Che, Wen Liu, Zhao Kang, Chong Peng, Mingqing Xiao, Qiang Cheng . On identifiability of 3-tensors of multilinear rank (1; Lr; Lr). Big Data and Information Analytics, 2016, 1(4): 391-401. doi: 10.3934/bdia.2016017
    [4] John A. Doucette, Robin Cohen . A testbed to enable comparisons between competing approaches for computational social choice. Big Data and Information Analytics, 2016, 1(4): 309-340. doi: 10.3934/bdia.2016013
    [5] Ugo Avila-Ponce de León, Ángel G. C. Pérez, Eric Avila-Vales . A data driven analysis and forecast of an SEIARD epidemic model for COVID-19 in Mexico. Big Data and Information Analytics, 2020, 5(1): 14-28. doi: 10.3934/bdia.2020002
    [6] Marco Tosato, Jianhong Wu . An application of PART to the Football Manager data for players clusters analyses to inform club team formation. Big Data and Information Analytics, 2018, 3(1): 43-54. doi: 10.3934/bdia.2018002
    [7] Jinyuan Zhang, Aimin Zhou, Guixu Zhang, Hu Zhang . A clustering based mate selection for evolutionary optimization. Big Data and Information Analytics, 2017, 2(1): 77-85. doi: 10.3934/bdia.2017010
    [8] Weidong Bao, Haoran Ji, Xiaomin Zhu, Ji Wang, Wenhua Xiao, Jianhong Wu . ACO-based solution for computation offloading in mobile cloud computing. Big Data and Information Analytics, 2016, 1(1): 1-13. doi: 10.3934/bdia.2016.1.1
    [9] Maria Gabriella Mosquera, Vlado Keselj . Identifying electronic gaming machine gambling personae through unsupervised session classification. Big Data and Information Analytics, 2017, 2(2): 141-175. doi: 10.3934/bdia.2017015
    [10] Wenxue Huang, Yuanyi Pan . On Balancing between Optimal and Proportional categorical predictions. Big Data and Information Analytics, 2016, 1(1): 129-137. doi: 10.3934/bdia.2016.1.129
  • Nowadays we are in the big data era. The high-dimensionality of data imposes big challenge on how to process them effectively and efficiently. Fortunately, in practice data are not unstructured. Their samples usually lie around low-dimensional manifolds and have high correlation among them. Such characteristics can be effectively depicted by low rankness. As an extension to the sparsity of first order data, such as voices, low rankness is also an effective measure for the sparsity of second order data, such as images. In this paper, I review the representative theories, algorithms and applications of the low rank subspace recovery models in data processing.


    1. Introduction

    Sparse representation and compressed sensing has achieved tre-mendous success in practice. They naturally fit for order-one data, such as voices and feature vectors. However, in applications we are often faced with various types of data, such as images, videos, and genetic microarrays. They are inherently matrices or even tensors. Then we are naturally faced with a question: how to measure the sparsity of matrices and tensors?

    Low-rank models are recent new tools that can robustly and efficiently handle high-dimensional data. Although rank has been used in statistics as a regularizer of matrices, e.g., reduced rank regression (RRR) [61], and in three-dimensional stereo vision [50], rank constraints are ubiquitous, the surge of low-rank models in recent years was inspired by sparse representation and compressed sensing. There has been systematic development on new theories and applications. In this background, rank is interpreted as the measure of the second order (i.e., matrix) sparsity1, rather than merely a mathematical concept. To illustrate this, we take image and video compression as an example. To achieve effective compression, we have to fully utilize the spatial and temporal correlation in images or videos. Take the Netflix challenge2 (Figure 1) as another example, to infer the unknown user ratings on videos, one has to consider both the correlation between users and the correlation between videos. Since the correlation among columns and rows is closely connected to matrix rank, it is natural to use rank as a measure of the second order sparsity.

    Figure 1. The Netflix challenge is to predict the unknown ratings of users on videos.

    1 The first order sparsity is the sparsity of vectors, whose measure is the number of nonzeros, i.e., the 0 norm 0.

    2 Netflix is a video-renting company, which owns a lot of users' ratings on videos. The user/video rating matrix is very sparse. The Netflix company offered one million US dollars to encourage improving the prediction on the user ratings on videos by 10%. See https://en.wikipedia.org/wiki/Netflix_Prize

    In the following, I review the recent development on low-rank models3. I first introduce linear models in Section 2, then nonlinear ones in Section 3, where the former are classified as single subspace models and multi-subspace ones. Theoretical analysis on some linear models, including exact recovery, closed-form solutions, and block-diagonality structure, is also provided in Section 2. Then I introduce commonly used optimization algorithms for solving low-rank models in Section 4, which can be classified as convex, non-convex, and randomized ones. Next, I review representative applications in Section 5. Finally, I conclude the paper in Section 6.

    3 There has been an excellent review on low-rank models in image analysis by Zhou et al. [82]. However, my review differs significantly from [82]. My review introduces much more low-rank models, e.g., tensor completion and recovery, multi-subspace models, and nonlinear models, while [82] mainly focuses on matrix completion and Robust PCA. My review also provides theoretical analysis and randomized algorithms.


    2. Linear models

    The recent boom of low-rank models started from the matrix completion (MC) problem [6] proposed by E. Candès in 2009. We introduce linear models first. Although they look simple, theoretical analysis show that linear models are very robust to strong noises and missing values. In real applications, they also have sufficient data representation power.


    2.1. Single subspace models

    Single subspace models are to extract one overall subspace from data. The most famous one may be the MC problem, proposed by E. Candès. It is as follows. Given the values of a matrix D at some locations, whether we can recover the whole matrix? This is a very general mathematical model for various problems, such as the above-mentioned Netflix challenge and the measurement of genetic microarrays. Obviously, the answer to this question is non-unique. Observing that we should consider the correlation among matrix columns and rows, E. Candès suggested to choose the solution A with the lowest rank:

    minArank(A),s.t.πΩ(A)=πΩ(D), (1)

    where Ω is the set of indices where the entries are known, πΩ is the projection operator that keeps the values of entries in Ω while filling the remaining entries with zeros. The MC problem is to recover the low-rank structure in the case of missing values. Shortly, E. Candès further considered MC with noise [5]:

    minArank(A),s.t.πΩ(A)πΩ(D)2Fε, (2)

    in order to handle the case when the observed data are noisy, where F is the Frobenius norm.

    When considering the low-rank recovery problem in the case of strong noises, it seems that this problem is well solvable by the traditional Principal Component Analysis (PCA). However, the traditional PCA is effective in accurately recovering the underlying low-rank structure only when the noises are Gaussian. If the noises are non-Gaussian and strong, even a few outliers can make PCA fail. Due to the great importance of PCA in applications, many scholars spent a lot effort on robustifying PCA, proposing many so-called "robust PCAs." However, none of them has a theoretical guarantee that under certain conditions the underlying low-rank structure can be exactly recovered. In 2009, Chandrasekaran et al.[7] and Wright et al. [68] proposed Robust PCA (RPCA) simultaneously. The problem they considered is how to recover the low-rank structure when the data have sparse and large outliers:

    minA,Erank(A)+λE0,s.t.A+E=D, (3)

    where E0 stands for the number of nonzeros in E. Shortly, E. Candè joined J. Wright et al.'s work and obtained stronger results. Namely, the matrix can have missing values. The generalized model is [4]:

    minA,Erank(A)+λE0,s.t.πΩ(A+E)=πΩ(D). (4)

    In their paper, they also discussed a generalized RPCA model which involves dense Gaussian noises [4]:

    minA,Erank(A)+λE0,s.t.πΩ(A+E)πΩ(D)2Fε. (5)

    Chen et al. [9] considered the case that noises cluster in sparse columns and proposed the Outlier Pursuit model, which replaces E0 in the RPCA model with E2,0, i.e., counting how many 2 norms of columns of E are zeros.

    When the data are tensor-like, Liu et al. [42] generalized matrix completion to tensor completion. Although tensors have a mathematical definition of rank, which is based on the CP decomposition [31], it is not computable. So Liu et al. proposed a new rank for tensors, which is defined as the sum of the ranks of matrices unfolded from the tensor in different modes. Their tensor completion model is thus: given the values of a tensor at some entries, recover the missing values by minimizing this new tensor rank. Also using the same new tensor rank, Tan et al. [60] generalized RPCA to tensor recovery. Namely, given a tensor, decompose it as a sum of two tensors, one having a low new tensor rank, the other being sparse.

    There are also matrix factorization based models, such as nonnegative matrix factorization [34]. Such models could be casted as low-rank models. However, they are better viewed as optimization techniques, as mentioned at the end of Section 4.2. So I will not elaborate them here. Interested readers may refer to several excellent reviews on matrix factorization based methods, e.g., [11,59,65].

    To sum up, single-subspace models could be viewed as extensions of the traditional PCA, which is mainly for denoising data and finding common components.


    2.2. Multi-subspace models

    MC and RPCA can only extract one subspace from data. They cannot describe finer details of data within this subspace. The simplest case of finer structure is the multi-subspace model, i.e., data distribute around some subspaces. We need to find these subspaces. This problem is called the Generalized PCA (GPCA) problem [62] or subspace clustering [63], which has a lot of solution methods, such as the algebraic method and RANSAC [63], but none of them have a theoretical guarantee. The emergence of sparse representation offers a new way to this problem. In 2009, E. Elhamifar and R. Vidal proposed the key idea of self-representation, i.e., using other samples to represent every sample. Based on self-representation, they proposed the Sparse Subspace Clustering (SSC) model [14,15] such that the representation matrix is sparse:

    minZ,Erank(Z)+λE0,s.t.D=DZ+E,diag(Z)=0, (6)

    where the constraint diag(Z)=0 is to prevent using the sample itself to represent a sample. Inspired by their work, Liu et al. proposed the Low-Rank Representation (LRR) model [38,39]:

    minZ,Erank(Z)+λE2,0,s.t.D=DZ+E. (7)

    The reason of enforcing the low-rankness of Z is to enhance the correlation among the columns of Z so as to boost the robustness against noise. The optimal representation matrix Z of SSC and LRR could be used as a measure of similarity between samples. Utilizing (|Z|+|Z,T|)/24 to define the similarity between samples (|Z| is the matrix whose entries are the absolute values of those of Z), one can cluster the data into several subspaces via spectral clustering. Zhuang et al. further required Z to be nonnegative and sparse, and applied Z to semi-supervised learning [84].

    4 In their later work [38], Liu et al. changed to use |UZUTZ| as the similarity matrix, where the columns of UZ are the left singular vectors of the skinny SVD of Z. For the reason, please refer to Section 2.3.2.

    LRR requires that the samples are sufficient. In the case of insufficient samples, Liu and Yan [41] proposed the Latent LRR model:

    minZ,L,Erank(Z)+rank(L)+λE0,s.t.D=DZ+LD+E. (8)

    They call DZ as the Principal Feature and LD the Salient Feature. Z is used for subspace clustering and L is used for extracting discriminant features for recognition. As an alternative way, Liu et al. [44] proposed the Fixed Rank Representation (FRR) model:

    minZ,˜Z,EZ˜Z2F+λE2,0,s.t.D=DZ+E,rank(˜Z)r, (9)

    where ˜Z is used for measuring the similarity between samples.

    To further improve the accuracy of subspace clustering, Lu et al. [45] proposed using Trace Lasso to regularize the representation vector:

    minZ,EDdiag(Zi)+λEi0,s.t.Di=DZi+Ei,i=1,,n, (10)

    where Zi is the ith column of Z, Ddiag(Zi) is called the Trace Lasso of Zi, and is the nuclear norm of a matrix (sum of singular values). When the columns of D are normalized in the 2-norm, Trace Lasso has an appealing interpolation property:

    Zi2Ddiag(Zi)Zi1.

    Moreover, the left hand side is achieved when the data are completely correlated (the columns being the same vector or the negative of the vector), while the right hand side is achieved when the data are completely uncorrelated (the columns being orthogonal). Therefore, Trace Lasso has the characteristic of being adaptive to the correlation among samples. This model is called Correlation Adaptive Subspace Segmentation (CASS).

    For better clustering of tensor data, Fu et al. proposed the Tensor LRR model [19], so as to fully utilize the information of tensor in different modes.

    In summary, multi-subspace models can model the data structure much better than the single-subspace ones. Their main purpose is to cluster data, drastically in contrast to that of single-subspace ones, i.e., to denoise data.


    2.3. Theoretical analysis

    The theoretical analysis on low-rank models is relatively rich. It consists of the following three parts.


    2.3.1. Exact recovery

    The above-mentioned low-rank models are all discrete optimization problems, most of which are NP-hard, which incurs great difficulty in efficient solution. To overcome this difficulty, a common way is to approximate discrete low-rank models as convex programs. Roughly speaking, the convex function (over the unit ball of norm) "closest" to the 0 pseudo-norm 0 is the 1 norm 1, i.e., the sum of absolute values of entries, and the convex function (over the unit ball of matrix spectral norm) "closest" to rank is the nuclear norm . Thus, all the above discrete problems can be converted into convex programs, which can be solved much more efficiently. However, this naturally brings a question: can solving a convex program result in the ground truth solution? For most low-rank models targeting on a single subspace, such as MC [6], RPCA [4], RPCA with missing values [4], and Outlier Pursuit [9,74], the answer is affirmative. Briefly speaking, if the outlier is sparse and uniformly random and the ground truth matrix is of low rank, then the ground truth matrix can be exactly recovered. What is surprising is that the exact recoverability is independent on the magnitude of outliers. Instead, it depends on the sparsity of outliers. Such results ensure that the low-rank models for single subspace recovery are very robust. This characteristic is unique when compared with the traditional PCA. Unfortunately, for multi-subspace low-rank models, only LRR has relatively thorough analysis [40]. However, Liu et al. only proved that when the proportion of outliers does not exceed a threshold, the row space of Z0 and which samples are outliers can be exactly known, where Z0 is given by UZUTZ, in which UZΣZVTZ is the skinny SVD of Z. The analysis did not answer whether Z0 and E0 themselves can be exactly recovered. Fortunately, when applying LRR to subspace clustering, we only need the row space of Z0.

    When data are noisy, it is inappropriate to use the noisy data to represent the data themselves. A more reasonable way is to denoise the data first and then apply self-representation on the denoised data, resulting in modified LRR and Latent LRR models:

    minZ,A,EZ+λE2,1,s.t.D=A+E,A=AZ, (11)

    and

    minZ,L,A,EZ+L+λE1,s.t.D=A+E,A=AZ+LA. (12)

    By utilizing the closed-form solutions discovered in the following subsection, Zhang et al. [76] proved that the solutions of modified LRR and Latent LRR can be expressed as that of corresponding RPCA models:

    minA,Erank(A)+λE2,1,s.t.D=A+E, (13)

    and

    minA,Erank(A)+λE1,s.t.D=A+E, (14)

    respectively. So the exact recovery results of RPCA [4] and Outlier pursuit [9,74] can be applied to the modified LRR and Latent LRR models, where again only the column space of D and which samples are outliers can be recovered.


    2.3.2. Closed-form solutions

    An interesting property of low-rank models is that they may have closed-form solutions when the data are noiseless. In comparison, sparse models do not have such a property. Wei and Lin [66] analyzed the mathematical properties of LRR. They first found that the noiseless LRR model:

    minZZ,s.t.D=DZ, (15)

    has a unique closed-form solution. Let the skinny SVD of D be UDΣDVTD, then the solution is VDVTD, which is called the Shape Interaction Matrix in structure from motion. Liu et al. [38] further found that the LRR with a general dictionary:

    minZZ,s.t.D=BZ, (16)

    also has a unique closed-form solution: Z=B+D, where B+ is the Moore-Penrose pseudo-inverse of B. This result is generalized by Yu and Schuurmans [72] to general unitary invariant norms, in which they found mode low-rank models with closed-form solution. Favaro et al. also found some low-rank models which are related to subspace clustering and have closed-form solutions [16]. Zhang et al. [73] further found that the solution to noiseless Latent LRR (both discrete and convex approximation) is non-unique and gave the complete closed-form solutions. In the paper, they also found that discrete noise-less LRR (the E in (7) being 0) is actually not NP-hard and further gave the complete closed-form solutions. To remedy this issue of Latent LRR, based on their analysis, Zhang et al. [75] further proposed to find the sparsest solution among the solution set of Latent LRR.


    2.3.3. Block-diagonal structure

    Multi-subspace clustering models all result in a representation matrix Z. For SSC and LRR, it can be proven that under the ideal conditions, i.e., the data are noiseless and the subspaces are independent (i.e., none of the subspaces can be represented by other subspaces), the optimal representation matrix Z is block-diagonal. As each block corresponds to one subspace, the block-structure of Z is critical to subspace clustering. Surprisingly, Lu et al. [47] proved that if Z is regularized by the squared Frobenius norm (the corresponding model is called Least Squared Representation (LSR)), then under ideal conditions the optimal representation matrix Z is also block-diagonal. Lu et al. further proposed the Enforced Block-Diagonal (EBD) Conditions. As long as the regularizer for Z satisfies the EBD conditions, the optimal representation matrix under the ideal conditions is block-diagonal [47]. The EBD conditions greatly extended the range of possible choices of Z, which is no longer limited to sparsity or low-rankness constraints. For subspace clustering models whose representation matrix Z is solved column-wise, e.g., Trace-Lasso-based CASS (10), Lu et al. also proposed the Enforced Block-Sparse (EBS) Conditions. As long as the regularizer on the columns of Z satisfies the EBS conditions, the optimal representation matrix under the ideal conditions is also block-diagonal [45]. However, all the above results are obtained under the ideal conditions. If the ideal conditions do not hold, i.e., when the data are noisy or when the subspaces are not independent, the optimal Z will not be exactly block-diagonal, which may cause difficulty in the subsequent subspace pursuit. To address this issue, based on the basic result in the spectral graph theory that the algebraic multiplicity of the eigenvalue zero of the Laplacian matrix equals the number of diagonal blocks in the weight matrix, Feng et al. [17] proposed the block-diagonal prior. Adding the block-diagonal prior to the subspace clustering models, an exactly block-diagonal representation matrix Z can be ensured even under non-ideal conditions, thus significantly improved the robustness against noise.

    The grouping effect among the representation coefficients, i.e., when the samples are similar their representation coefficient vectors should also be similar, is also helpful for maintaining the block-diagonal structure of the representation matrix Z when the data are noisy. SSC, LRR, LSR, and CASS are all proven to have the grouping effect. Hu et al. proposed general Enforced Grouping Effect (EGE) Conditions [24], with which one can easily verify whether a regularizer has the grouping effect.

    To conclude, linear models are relatively simple yet powerful enough to model complex data distributions. They can also have good mathematical properties and theoretical guarantees.


    3. Nonlinear models

    Linear models assume that the data distribute near some low-dimensional subspaces. Such assumption can be easily violated in real applications. So developing nonlinear models is necessary. However, low-rank models for clustering nonlinear manifolds are relatively few. A natural idea is to utilize the kernel trick, proposed by Wang et al. [64]. The idea is as follows. Suppose that via a nonlinear mapping ϕ, the set X of samples is mapped to linear subspaces in a high dimensional space. Then the LRR model can be applied to the mapped sample set. Suppose that the noises are Gaussian, the model is:

    minZϕ(X)ϕ(X)Z2F+λZ.

    Since ϕ(X)ϕ(X)Z2F=tr[(ϕ(X)ϕ(X)Z)T(ϕ(X)ϕ(X)Z)], we obtain inner products ϕ(X)Tϕ(X). So we can introduce a kernel function K(x,y), such that K(x,y)=ϕT(x)ϕ(y). Therefore, the above model can be written in a kernalized form without introducing the nonlinear mapping ϕ explicitly. However, when the noises are not Gaussian, the above kernel trick does not apply.

    The other heuristic approach is to add Laplacian or hyper-Laplacian to the corresponding linear models. It is claimed that Laplacian or hyper-Laplacian can capture the nonlinear geometry of the data distribution. For example, Lu et al. [49] added the Laplacian regularization tr(ZLWZT) to the objective function of LRR, where LW is the Laplacian matrix of the weight matrix W in which Wij=exp(xixj2/σ). Zheng et al. [81] added another form of Laplacian regularization tr(DL˜ZDT) to the objective function of LRR, where D is the data matrix and L˜Z is the Laplacian matrix of ˜Z=(|Z|+|ZT|)/2. Yin et al. considered both Laplacian and hyper-Laplacian regularization in the nonnegative low-rank and sparse LRR model [71].

    Although the modifications on linear models result in more powerful nonlinear models, it is hard to analyze their properties. So their performance may heavily depend on the choice of parameters.


    4. Optimization algorithms

    Once we have a mathematical model, we need to solve it efficiently. The discrete low-rank models in Section 2 are mostly NP-hard. So most of the time they could only be solved approximately. A common way is to convert them into continuous optimization problems. There are two ways to do so. The first way is to convert them into convex programs. For example, as mentioned above, one may replace the 0 pseudo-norm 0 with the 1 norm 1 and replace rank with the nuclear norm . Another way is to convert to non-convex programs. More specifically, it is to use a non-convex continuous function to approximate the 0 pseudo-norm 0 (e.g., using the p pseudo-norm p(0<p<1)) and rank (e.g., using the Schatten-p pseudo-norm (the p pseudo-norm of the vector of singular values)). There is still another way. It is to represent the low-rank matrix as a product of two matrices, the number of columns of the first matrix and the number of rows of the second matrix both being the expected rank, and then update the two matrices alternately until they do not change. This special type of algorithm does not appear in the sparsity based models. The advantage of convex programs is that their global optimal solutions can be relatively easy obtained. The disadvantages include that the solution may not be sufficiently low-rank or sparse. In contrast, the advantage of non-convex optimization is that lower-rank or sparser solutions can be obtained. However, their global optimal solution may not be obtained. The quality of solution may heavily depend on the initialization. So the convex and non-convex algorithms complement each other. By fully utilizing the characteristics of problems, it is also possible to design randomized algorithms so that the computation complexity can be greatly reduced.


    4.1. Convex algorithms

    Convex optimization is a relatively mature field. There are a lot of polynomial complexity algorithms, such as interior point methods. However, for large scale or high dimensional data, we often need O(npolylog(n)) complexity, where n is the number or the dimensionality of samples. Even O(n2) complexity is unacceptable. Take the RPCA problem as example, if the matrix size is n×n, then the problem has 2n2 unknowns. Even if n=1000, which corresponds to a relatively small matrix, the number of unknowns already reaches two millions. If we solve RPCA with the interior point method, e.g., using the CVX package [22] by Stanford University, then the time complexity of each iteration is O(n6), while the storage complexity is O(n4). If solved on a PC with 4GB memory, the size of matrix will be limited to 80×80. So to make low-rank models practical, we have to design efficient optimization algorithms.

    Currently, all the optimization methods for large scale computing are first order methods. Representative algorithms include Accelerated Proximal Gradient (APG) [2,54], the Frank-Wolfe algorithm [18,26], and the Alternating Direction Method (ADM) [36,37].

    APG is basically for unconstrained problems:

    minxf(x), (17)

    where the objective function is convex and C1,1, i.e., differentiable and its gradient is Lipschitz continuous:

    f(x)f(y)Lfxy,x,y. (18)

    The convergence rate of traditional gradient descent can only be O(k1), where k is the number of iterations. However, Nesterov constructed an algorithm [54]:

    xk=ykL1ff(yk),tk+1=1+1+4t2k2,yk+1=xk+tk1tk+1(xkxk1), (19)

    where x0=y1=0 and t1=1. whose convergence rate can achieve O(k2). Later, Beck and Teboulle generalized Nesterov's algorithm for the following problem:

    minxg(x)+f(x), (20)

    where g is convex, whose proximity operator minxg(x)+α2xw2 is easily solvable, and f is a C1,1 convex function [2], thus greatly extended the applicable range of Nesterov's method. APG needs to estimate the Lipschitz coefficient Lf of the gradient of the objective function. If the Lipschitz coefficient is estimated too conservatively (too large), the convergence speed will be affected. So Beck and Teboulle further proposed a back-tracking strategy to estimate the Lipschitz coefficient adaptively, so as to speed up convergence [2]. For some problems with special structures, APG can be generalized (Generalized APG, GAPG) [85], such that different Lipschitz coefficients could be chosen for different variables, thus the convergence can be made faster. For problems with linear constraints:

    minxf(x),s.t.A(x)=b, (21)

    where f is convex and C1,1 and A is a linear operator, one may add the squared constraint to the objective function as a penalty, converting the problem to an unconstrained one:

    minxf(x)+β2A(x)b2, (22)

    then solve (22) by APG. To speed up, the penalty parameter β should increase gradually along with iteration, rather than being set at a large value from the beginning. This important trick is called the continuation technique [20].

    For problems with a convex set constraint:

    minxf(x),s.t.xC, (23)

    where f is convex and continuously differentiable and C is a compact convex set, Frank-Wolfe-type algorithms [18,26]:

    gk=argmingCg,f(xk),xk+1=(1γk)xk+γkgk,where γk=2k+2, (24)

    can be used to solve (23). In particular, when the constraint set C is a ball of bounded nuclear norm, gk can be relatively easily computed by finding the singular vectors associated to the leading singular values of f(xk) [26]. Such a particular problem can also be efficiently solved by transforming it into a positive semi-definite program [27], where only the eigenvector corresponding to the largest eigenvalue of a matrix is needed.

    ADM fits for convex problems with separable objective functions and linear or convex-set constraints:

    minx,yf(x)+g(y),s.t.A(x)+B(y)=c, (25)

    where f and g are convex functions and A and B are linear operators. ADM is a variant of the Lagrange Multiplier method. ADM first constructs an augmented Lagrangian function [37]:

    L(x,y,λ)=f(x)+g(y)+ <λ,A(x)+B(y)c+β2A(x)+B(y)c2, (26)

    where λ is the Lagrange multiplier and β>0 is the penalty parameter, then updates the two variables alternately by minimizing the augmented Lagrangian function with the other variable fixed [37]:

    xk+1=argminxL(x,yk,λk),yk+1=argminyL(xk+1,y,λk). (27)

    Finally, ADM updates the Lagrange multiplier [37]:

    λk+1=λk+β(A(xk+1)+B(yk+1)c). (28)

    The advantage of ADM is that its subproblems are simpler than the original problem. They may even have closed-form solutions. When the subproblems are not easily solvable, one may consider approximating the squared constraint β2A(x)+B(y)c2 in the augmented Lagrangian function with its first order Taylor expansion plus a proximal term, to make the subproblem even simpler. This technique is called the Linearized Alternating Direction Method (LADM) [37]. If after linearizing the squared constraint in the augmented Lagrangian function the subproblem is still not easily solvable, one may further linearize the C1,1 component of the objective function [36]. For multi-block (the number of blocks of variables is greater than 2) convex programs, a naive generalization of the two-block ADM may not converge [8]. However, if we change the serial update with parallel update and choose some parameters appropriately, the convergence can still be guaranteed, even if linearization is used [36]. In all the above-mentioned ADM algorithms, the penalty parameter β is allowed to increase dynamically so that the convergence can be accelerated and the difficulty in tuning an optimal penalty parameter can be overcome [36,37].

    When solving the low-rank models with convex surrogates, we often face with the following subproblem:

    minXX+α2XW2F,

    which has a closed-form solution [3]. Suppose that the SVD of W is W=UΣVT, then the optimal solution is X=UΘα1(Σ)VT, where

    Θε(x)={xε,if x>ε,x+ε,if x<ε,0,if εxε. (29)

    So solving low-rank models with nuclear norm, SVD is often indispensable. For m×n matrices, the time complexity of full SVD is O(mnmin(m,n)). So in general the computation cost for solving low-rank models with nuclear norm is large. This issue is more critical when m and n is large. Fortunately, from (29) one can see that it is unnecessary to compute the singular values not exceeding α1 and their associated singular vectors, because these singular values will be shrunk to zeros, thus do not contribute to X. So we only need to compute singular values greater than α1 and their corresponding singular vectors. Such partial SVD computation can be achieved by PROPACK [33] and accordingly the computation cost reduces to O(rmn), where r is the expected rank of the optimal Z. It is worth noting that PROPACK can only provide expected number of leading singular values and their singular vectors. So we have to dynamically predict the value of r when calling PROPACK [37]. When the solution is not sufficiently low-rank, such as Transform Invariant Low-Rank Textures (TILT) [78] ((30) and Section 5.5) which has wide applications in image processing and computer vision, one can use incremental SVD [58] for acceleration.

    Convex algorithms have the advantage of being independent of initialization. However, the quality of their solutions may not be good enough. So exploring nonconvex algorithms is another hot topic in low-rank models.


    4.2. Nonconvex optimization algorithms

    Nonconvex algorithms trade the initialization independency for better solution quality and possibly faster speed as well. For unconstrained problems which use the Schatten-p norm to approximate rank and the p norm to approximate the 0 norm, an effective way is the Iteratively Reweighted Least Squares (IRLS) [46]. To be more precise, approximate tr((XXT)p/2) with tr((XkXTk)(p/2)1(XXT)) and |xi|p with |x(k)i|p2x2i, where Xk is the value of low-rank matrix X at the kth iteration and x(k)i is the ith component of the sparse vector x at the kth iteration. So each time to update X, a matrix equation needs to be solved, while updating x needs solving a linear system. Another way is to apply the idea of APG. To be specific, linearize the C1,1 component of the objective function. Then in each iteration one only needs to solve the following subproblem:

    minXni=1g(σi(X))+α2XW2F,

    where g is a non-decreasing concave function on {x0}, such as xp(0<p<1). Lu et al. [48] provided an algorithm to solve the above subproblem. Another function that approximates rank is the Truncated Nuclear Norm (TNN) [25]: Xr=min(m,n)i=r+1σi(X). TNN does not involve the largest r singular values. So it is not a convex function. The intuition behind TNN is obvious. By minimizing TNN, the tailing singular values will be encouraged to be small, while the magnitudes of the first r singular values are unaffected. So a solution closer to a rank r matrix can be obtained. TNN can be generalized by adding larger weights to smaller singular values, obtaining the Weighted Nuclear Norm (WNN) [23]: Xw,=min(m,n)i=1wiσi(X). However, in this case in general the subproblem

    minXni=1Xw,+α2XW2F,

    does not have a closed-form solution. Instead, a small-scale optimization w.r.t. the singular values need to be solved numerically.

    The third kind of methods for low-rank problems is to represent the expected low-rank matrix X as X=ABT, where A and B both have r columns. Then A and B can be updated alternately until they do not change [67]. The advantage of this kind of methods is its simplicity. However, we have to estimate the rank of low-rank matrix apriori and A and B may easily get stuck.

    Nonconvex algorithms for low-rank models are much richer than convex ones. The price paid is that their performance may heavily depend on initialization. In this case, prior knowledge is important for proposing a good initialization.


    4.3. Randomized algorithms

    All the above-mentioned methods, no matter for convex or non-convex problems, their computation complexity is at least O(rmn), where m×n is the size of the low-rank matrix that we want to compute. This is not fast enough when m and n are both very large. To break this bottleneck, we have to resort to randomized algorithms. However, we cannot reduce the whole computation complexity simply by randomizing each step of a deterministic algorithm, e.g., simply replacing SVD with linear-time SVD [13], because some randomized algorithms are very inaccurate. So we have to design randomized algorithms based on the characteristics of low-rank models. As a result, currently there is limited work on this aspect. For RPCA, Liu et al. proposed the 1-filtering method [43]. It first randomly samples a submatrix Ds, with an appropriate size, from the data matrix D. Then it solves a small-scale RPCA on Ds, obtaining a low-rank As and a sparse Es, Next, it processes the sub-columns and sub-rows of D that Ds resides on, using As, as they should belong to the subspaces spanned by the columns or rows of As up to sparse errors. Finally, the low-rank matrix A that corresponds to the original matrix D can be represented by the Nyström trick, without explicit computing. The complexity of the whole algorithm is O(r3)+O(r2(m+n)), which is linear with respect to the matrix size. For LRR and Latent LRR, Zhang et al. found that if we denoise the data first with RPCA and then apply LRR or Latent LRR on the denoised data, then their solutions can be expressed by the solution of RPCA and vice versa. So the solutions of LRR and Latent LRR can be greatly accelerated by reducing to RPCA [76].

    Randomized algorithms could bring down the order of computation complexity. However, designing randomized algorithms often needs to consider the characteristic of the problems.


    5. Representative applications

    Low-rank models have found wide applications in data analysis and machine learning. For example, there have been a lot of papers on NIPS 2011 which discussed low-rank models. Below I introduce some representative applications.


    5.1. Video denoising [29]

    Image and video denoising can be conveniently formulated as a matrix completion problem. In [29], Ji et al. first broke each of the video frames into patches, then grouped the similar patches. For each group, the patches are reshaped into vectors and assembled into a matrix. Next, the unreliable (noisy) pixels are detected as those whose values deviate from the means of their corresponding row vectors and the remaining pixels are considered reliable (noiseless). The unreliable pixel values can be estimated by applying the matrix completion model (2) to the matrix by marking them as missing values. After denoising all the patches, each frame is restored by averaging the pixel values in overlapping patches. Part of the results of video denoising are shown in Figure 2.

    Figure 2. Video denoising results, adapted from [29]. (a) The results of VBM3D, without preprocessing the impulsive noise. (b) The results of PCA, without preprocessing the impulsive noise. (c) The results of VBM3D, with preprocessing the impulsive noise. (d) The results of PCA, with preprocessing the impulsive noise. (e) The results of Matrix Completion.

    5.2. Keyword extraction [51]

    In document analysis, it is important to extract keywords from documents. Let D be the unnormalized term frequency matrix, where the row indices are the document IDs and the column indices are the term IDs and the (i,j)-th entry is the frequency of the j-th term in the i-th document. Then for documents of similar topics, many of the words are common, forming the "background" topic, and each document should have its unique keywords to discriminate it from others. This phenomenon makes keyword extraction naturally fit for the RPCA model (3), where the sparse error E identifies keywords in each document. One example of keyword extraction is shown in Figure 3.

    Figure 3. An example of keyword extraction by PRCA, adapted from [63]. The sample document is from the Reuters dataset. The blue words are keywords indicated by the sparse term E.

    5.3. Background modeling [4]

    Background modeling is to separate the background and the foreground from a video. The simplest case is that the video is taken by a fixed video camera. It is easy to see that the background hardly changes. So if putting each frame of the background as a column of a matrix, then the matrix should be of low rank. As the foreground consists of moving objects, it often occupies only a small portion of pixels. So the foreground corresponds to the sparse "noise" in the video. So we can obtain the RPCA model (3) for background modeling, where each column of D, A, and E is a frame of the video, the background, and the foreground, respectively, rearranged into a vector. Part of the results of background modeling is shown in Figure 4.

    Figure 4. Background modeling results, adapted from [4]. The first column are frames of surveillance video, the second are the background video, and the third are the foreground video (in absolute value).

    5.4. Robust Alignment by Sparse and Low-Rank decomposition (RASL) [56]

    The RPCA model for background modeling has to assume that the background has been aligned so as to obtain a low-rank background video. In the case of misalignment, we may consider aligning the frames via appropriate geometric transformation. So the mathematical model is:

    minτ,A,EA+λE1,s.t.Dτ=A+E, (30)

    where Dτ represents applying frame-wise geometric deformation τ to each frame, which is a column of D. Now (30) is a nonconvex optimization problem. For efficient solution, Peng et al. [56] proposed to linearize τ locally and update the increment of τ iteratively. That is to say, first solve Δτk from:

    minΔτk,A,EA+λE1,s.t.Dτk+JΔτk=A+E, (31)

    then add Δτk to τk as τk+1, where J is the Jacobian of Dτ with respect to the parameters of transformation τ. Under the affine transformation, part of the results of facial image alignment are shown in Figure 5.

    Figure 5. The iterative process of facial image alignment, adapted from [56].

    5.5. Transform Invariant Low-rank Textures (TILT) [78]

    The purpose of Transform Invariant Low-rank Textures (TILT) is to rectify an image patch D with a geometric transformation τ, such that the content in patch becomes regular, such as being rectilinear or symmetric. Such regularity could be depicted by low-rankness. The mathematical formulation of TILT is the same as that of RASL (30). The solution method is also identical. The difference resides in the interpretation on the matrix D, which is now a rectangular image patch in a single image. Figure 6 gives examples of rectifying image patches under the perspective transform.

    Figure 6. Example of using TILT for image rectification, adapted from [78]. The first row are the original image patches (in rectangles) and their respective rectification transformations (in quadrilaterals. The transformations are to map the quadrilaterals into rectangles). The second row are the rectified image patches.

    In principle, TILT should work for any parameterized transformations. Zhang et al. [79] further considered TILT under generalized cylindrical transformations, which can be used for texture unwarping from buildings. Some examples are shown in Figure 7.

    Figure 7. Texture unwarping from buildings, using TILT under generalized cylindrical transformations. Images are adapted from [79].

    TILT is also widely applied to geometric modeling of buildings [78], camera self-calibration, and lens distortion auto-correction [80]. Due to its importance in applications, Ren and Lin [58] proposed a fast algorithm for TILT to speed up its solution by more than five times.


    5.6. Motion segmentation [38,39]

    Motion segmentation means to cluster the feature points on moving objects in a video, such that each cluster corresponds to an independent object. Then an object can be identified and tracked. For each feature point, its feature vector consists of its image coordinate in each frame and is a column of the data matrix D. Then subspace clustering models, such as those in Section 2.2, could be applied to cluster the feature vectors and hence the corresponding feature points. LRR (7) is regarded as one of the best algorithms for segmenting the motion of rigid bodies [1]. Some of the examples of motion segmentation are shown in Figure 8.

    Figure 8. Examples of motion segmentation, adapted from [63].

    5.7. Image segmentation [10]

    Image segmentation is to partition an image into homogenous regions. It can be viewed as a special clustering problem. Cheng et al. [10] proposed to oversegment the image into superpixels, then extract usual features from the superpixels. Next, they fused multiple features via an integrated LRR model, where basically each feature corresponds to an LRR model. After obtaining the global representation matrix Z, they applied normalized cut to a graph whose weights are given by the similarity matrix (|Z|+|ZT|)/2 to cluster the superpixels into clusters, each corresponding to an image region. Part of the results of image segmentation are shown in Figure 9.

    Figure 9. Examples of image segmentation by LRR, adapted from [10].

    5.8. Gene clustering [12]

    Gene clustering is to group genes with similar functionality. Identifying gene clusters from the gene expression data is helpful for the discovery of novel functional gene interactions. Let D be the transposed gene expression data matrix, whose columns contain the expression levels of a gene in all the samples and whose rows are the expression levels of all the genes in one sample. Cui et al. [12] then applied the LRR model to D to cluster the genes. Two examples of gene clustering are shown in Figure 10.

    Figure 10. Clustering of genes from the yeast dataset by LRR. (A) A heatmap of expression values of genes in Cluster C17. It shows similar expression patterns of genes in different samples. (B) A heatmap of expression values of genes in Cluster C14. It shows different expression patterns of genes in different samples (marked as a and b). The images are adapted from [12].

    5.9. Image saliency detection [32]

    Saliency detection is to detect the visually salient regions in an image without understanding the content of the image. Motion segmentation, image segmentation, and gene clustering all utilize the representation matrix Z in LRR. In contrast, Lang et al. [32] proposed to utilize the sparse "noise" E in LRR for image saliency detection. Note that salient regions in an image is the "larruping" region. So if using other regions to "predict" salient regions, there will be relatively large errors. Therefore, by breaking an image into patches and extracting their features, the salient regions should correspond to those with large sparse "noise" E in the LRR model, where the data matrix D consists of the feature vectors. Part of the examples of saliency detection are shown in Figure 11.

    Figure 11. Examples of image saliency detection, adapted from [32]. The first column are the input images. The second to fifth columns are the detection results of different methods. The last column are the results of LRR-based detection method.

    5.10. Other applications

    There have been many other applications of low-rank models, such as partial duplicate image search [70], face recognition [57], structured texture repairing [35], man-made object upright orientation [30], photometric stereo [69], image tag refinement [83], robust visual domain adaption [28], robust visual tracking [77], feature extraction from 3D faces [52], ghost image removal in computed tomography [21], semi-supervised image classification [84], image set co-segmentation [53], and even audio analysis [53,55], protein-gene correlation analysis, network flow abnormality detection, robust filtering and system identification. Due to the space limit, I omit their introductions.


    6. Conclusions

    Low-rank models have found wide applications in many fields, including signal processing, machine learning, and computer vision. In a few years, there has been rapid development in theories, algorithms, and applications on low-rank models. This review is only a sketchy introduction to this dynamic research topic. Many real problems, if combining the characteristic of problem with proper low-rankness constraints, very often we could obtain better results. In some problems, the raw data may not have a low-rank property. However, the low-rankness could be enhanced by incorporating appropriate transforms (like the improvement of RASL/TILT over RPCA). Some scholars did not check whether the data have low-rank property or do proper pre-processing before claiming that low-rank constraints do not work well. This should be avoided. From the above review, we can see that low-rank models still lack research in the following aspects: generalization from matrices to tensors, nonlinear manifold clustering, and low-complexity (polylog(n)) randomized algorithms, etc. Hope this review can attract more research in these aspects.


    Acknowledgments

    Z. Lin is supported by NSF China (grant nos. 61272341 and 61231002), 973 Program of China (grant no. 2015CB352502), and Microsoft Research Asia Collaborative Research Program.


    [1] [ A. Adler, M. Elad and Y. Hel-Or, Probabilistic subspace clustering via sparse representations, IEEE Signal Processing Letters, 20(2013), 63-66.
    [2] [ A. Beck and M. Teboulle, A fast iterative shrinkage-thresholding algorithm for linear inverse problems, SIAM Journal on Imaging Sciences, 2(2009), 183-202.
    [3] [ J. Cai, E. Candès and Z. Shen, A singular value thresholding algorithm for matrix completion, SIAM Journal on Optimization, 20(2010), 1956-1982.
    [4] [ E. Candès, X. Li, Y. Ma and J. Wright, Robust principal component analysis?, Journal of the ACM, 58(2011), Art. 11, 37 pp.
    [5] [ E. Candès and Y. Plan, Matrix completion with noise, Proceedings of the IEEE, 98(2010), 925-936.
    [6] [ E. Candès and B. Recht, Exact matrix completion via convex optimization, Foundations of Computational Mathematics, 9(2009), 717-772.
    [7] [ V. Chandrasekaran, S. Sanghavi, P. Parrilo and A. Willsky, Sparse and low-rank matrix decompositions, Annual Allerton Conference on Communication, Control, and Computing, 2009, 962-967.
    [8] [ C. Chen, B. He, Y. Ye and X. Yuan, The direct extension of ADMM for multi-block convex minimization problems is not necessarily convergent, Mathematical Programming, 155(2016), 57-79.
    [9] [ Y. Chen, H. Xu, C. Caramanis and S. Sanghavi, Robust matrix completion with corrupted columns, International Conference on Machine Learning, 2011, 873-880.
    [10] [ B. Cheng, G. Liu, J. Wang, Z. Huang and S. Yan, Multi-task low-rank affinity pursuit for image segmentation, International Conference on Computer Vision, 2011, 2439-2446.
    [11] [ A. Cichocki, R. Zdunek, A. H. Phan and S. Ichi Amari, Nonnegative Matrix and Tensor Factorizations:Applications to Exploratory Multi-way Data Analysis and Blind Source Separation, 1st edition, Wiley, 2009.
    [12] [ Y. Cui, C.-H. Zheng and J. Yang, Identifying subspace gene clusters from microarray data using low-rank representation, PLoS One, 8(2013), e59377.
    [13] [ P. Drineas, R. Kannan and M. Mahoney, Fast Monte Carlo algorithms for matrices Ⅱ:Computing a low rank approximation to a matrix, SIAM Journal on Computing, 36(2006), 158-183.
    [14] [ E. Elhamifar and R. Vidal, Sparse subspace clustering, in IEEE International Conference on Computer Vision and Pattern Recognition, 2009, 2790-2797.
    [15] [ E. Elhamifar and R. Vidal, Sparse subspace clustering:Algorithm, theory, and applications, IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(2013), 2765-2781.
    [16] [ P. Favaro, R. Vidal and A. Ravichandran, A closed form solution to robust subspace estimation and clustering, IEEE Conference on Computer Vision and Pattern Recognition, 2011, 1801-1807.
    [17] [ J. Feng, Z. Lin, H. Xu and S. Yan, Robust subspace segmentation with block-diagonal prior, IEEE Conference on Computer Vision and Pattern Recognition, 2014, 3818-3825.
    [18] [ M. Frank and P. Wolfe, An algorithm for quadratic programming, Naval Research Logistics Quarterly, 3(1956), 95-110.
    [19] [ Y. Fu, J. Gao, D. Tien and Z. Lin, Tensor LRR based subspace clustering, International Joint Conference on Neural Networks, 2014, 1877-1884.
    [20] [ A. Ganesh, Z. Lin, J. Wright, L. Wu, M. Chen and Y. Ma, Fast algorithms for recovering a corrupted low-rank matrix, International Workshop on Computational Advances in MultiSensor Adaptive Processing, 2009, 213-216.
    [21] [ H. Gao, J.-F. Cai, Z. Shen and H. Zhao, Robust principal component analysis-based four-dimensional computed tomography, Physics in Medicine and Biology, 56(2011), 3181-3198.
    [22] [ M. Grant and S. Boyd, CVX:Matlab software for disciplined convex programming (web page and software), http://cvxr.com/cvx/, 2009.
    [23] [ S. Gu, L. Zhang, W. Zuo and X. Feng, Weighted nuclear norm minimization with application to image denoising, IEEE Conference on Computer Vision and Pattern Recognition, 2014, 2862-2869.
    [24] [ H. Hu, Z. Lin, J. Feng and J. Zhou, Smooth representation clustering, IEEE Conference on Computer Vision and Pattern Recognition, 2014, 3834-3841.
    [25] [ Y. Hu, D. Zhang, J. Ye, X. Li and X. He, Fast and accurate matrix completion via truncated nuclear norm regularization, IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(2013), 2117-2130.
    [26] [ M. Jaggi, Revisiting Frank-Wolfe:Projection-free sparse convex optimization, in International Conference on Machine Learning, 2013, 427-435.
    [27] [ M. Jaggi and M. Sulovský, A simple algorithm for nuclear norm regularized problems, in International Conference on Machine Learning, 2010, 471-478.
    [28] [ I. Jhuo, D. Liu, D. Lee and S. Chang, Robust visual domain adaptation with low-rank reconstruction, IEEE Conference on Computer Vision and Pattern Recognition, 2012, 2168-2175.
    [29] [ H. Ji, C. Liu, Z. Shen and Y. Xu, Robust video denoising using low rank matrix completion, IEEE Conference on Computer Vision and Pattern Recognition, 2010, 1791-1798.
    [30] [ Y. Jin, Q. Wu and L. Liu, Unsupervised upright orientation of man-made models, Graphical Models, 74(2012), 99-108.
    [31] [ T. G. Kolda and B. W. Bader, Tensor decompositions and applications, SIAM Review, 51(2009), 455-500.
    [32] [ C. Lang, G. Liu, J. Yu and S. Yan, Saliency detection by multitask sparsity pursuit, IEEE Transactions on Image Processing, 21(2012), 1327-1338.
    [33] [ R. M. Larsen, http://sun.stanford.edu/~rmunk/PROPACK/, 2004.
    [34] [ D. Lee and H. Seung, Learning the parts of objects by non-negative matrix factorization, Nature, 401(1999), 788.
    [35] [ X. Liang, X. Ren, Z. Zhang and Y. Ma, Repairing sparse low-rank texture, in European Conference on Computer Vision, 7576(2012), 482-495.
    [36] [ Z. Lin, R. Liu and H. Li, Linearized alternating direction method with parallel splitting and adaptive penality for separable convex programs in machine learning, Machine Learning, 99(2015), 287-325.
    [37] [ Z. Lin, R. Liu and Z. Su, Linearized alternating direction method with adaptive penalty for low-rank representation, Advances in Neural Information Processing Systems, 2011, 612-620.
    [38] [ G. Liu, Z. Lin, S. Yan, J. Sun and Y. Ma, Robust recovery of subspace structures by low-rank representation, IEEE Transactions Pattern Analysis and Machine Intelligence, 35(2013), 171-184.
    [39] [ G. Liu, Z. Lin and Y. Yu, Robust subspace segmentation by low-rank representation, in International Conference on Machine Learning, 2010, 663-670.
    [40] [ G. Liu, H. Xu and S. Yan, Exact subspace segmentation and outlier detection by low-rank representation, International Conference on Artificial Intelligence and Statistics, 2012, 703-711.
    [41] [ G. Liu and S. Yan, Latent low-rank representation for subspace segmentation and feature extraction, in IEEE International Conference on Computer Vision, IEEE, 2011, 1615-1622.
    [42] [ J. Liu, P. Musialski, P. Wonka and J. Ye, Tensor completion for estimating missing values in visual data, IEEE Transactions on Pattern Analysis and Machine Intelligence, 35(2013), 208-220.
    [43] [ R. Liu, Z. Lin, Z. Su and J. Gao, Linear time principal component pursuit and its extensions using l1 filtering, Neurocomputing, 142(2014), 529-541.
    [44] [ R. Liu, Z. Lin, F. Torre and Z. Su, Fixed-rank representation for unsupervised visual learning, IEEE Conference on Computer Vision and Pattern Recognition, 2012, 598-605.
    [45] [ C. Lu, J. Feng, Z. Lin and S. Yan, Correlation adaptive subspace segmentation by trace lasso, International Conference on Computer Vision, 2013, 1345-1352.
    [46] [ C. Lu, Z. Lin and S. Yan, Smoothed low rank and sparse matrix recovery by iteratively reweighted least squared minimization, IEEE Transactions on Image Processing, 24(2015), 646-654.
    [47] [ C. Lu, H. Min, Z. Zhao, L. Zhu, D. Huang and S. Yan, Robust and efficient subspace segmentation via least squares regression, European Conference on Computer Vision, 7578(2012), 347-360.
    [48] [ C. Lu, C. Zhu, C. Xu, S. Yan and Z. Lin, Generalized singular value thresholding, AAAI Conference on Artificial Intelligence, 2015, 1805-1811.
    [49] [ X. Lu, Y. Wang and Y. Yuan, Graph-regularized low-rank representation for destriping of hyperspectral images, IEEE Transactions on Geoscience and Remote Sensing, 51(2013), 4009-4018.
    [50] [ Y. Ma, S. Soatto, J. Kosecka and S. Sastry, An Invitation to 3-D Vision:From Images to Geometric Models, 1st edition, Springer, 2004.
    [51] [ K. Min, Z. Zhang, J. Wright and Y. Ma, Decomposing background topics from keywords by principal component pursuit, in ACM International Conference on Information and Knowledge Management, 2010, 269-278.
    [52] [ Y. Ming and Q. Ruan, Robust sparse bounding sphere for 3D face recognition, Image and Vision Computing, 30(2012), 524-534.
    [53] [ L. Mukherjee, V. Singh, J. Xu and M. Collins, Analyzing the subspace structure of related images:Concurrent segmentation of image sets, European Conference on Computer Vision, 7575(2012), 128-142.
    [54] [ Y. Nesterov, A method of solving a convex programming problem with convergence rate O(1/k2), (Russian) Dokl. Akad. Nauk SSSR, 269(1983), 543-547.
    [55] [ Y. Panagakis and C. Kotropoulos, Automatic music tagging by low-rank representation, International Conference on Acoustics, Speech, and Signal Processing, 2012, 497-500.
    [56] [ Y. Peng, A. Ganesh, J. Wright, W. Xu and Y. Ma, RASL:Robust alignment by sparse and low-rank decomposition for linearly correlated images, IEEE Transactions on Pattern Analysis and Machine Intelligence, 34(2012), 2233-2246.
    [57] [ J. Qian, J. Yang, F. Zhang and Z. Lin, Robust low-rank regularized regression for face recognition with occlusion, The Workshop of IEEE Conference on Computer Vision and Pattern Recognition, 2014, 21-26.
    [58] [ X. Ren and Z. Lin, Linearized alternating direction method with adaptive penalty and warm starts for fast solving transform invariant low-rank textures, International Journal of Computer Vision, 104(2013), 1-14.
    [59] [ A. P. Singh and G. J. Gordon, A unified view of matrix factorization models, in Proceedings of Machine Learning and Knowledge Discovery in Databases, 5212(2008), 358-373.
    [60] [ H. Tan, J. Feng, G. Feng, W. Wang and Y. Zhang, Traffic volume data outlier recovery via tensor model, Mathematical Problems in Engineering, 2013(2013), 164810.
    [61] [ M. Tso, Reduced-rank regression and canonical analysis, Journal of the Royal Statistical Society, Series B (Methodological), 43(1981), 183-189.
    [62] [ R. Vidal, Y. Ma and S. Sastry, Generalized principal component analysis, IEEE Transactions on Pattern Analysis and Machine Intelligence, 27(2005), 1945-1959.
    [63] [ R. Vidal, Subspace clustering, IEEE Signal Processing Magazine, 28(2011), 52-68.
    [64] [ J. Wang, V. Saligrama and D. Castanon, Structural similarity and distance in learning, Annual Allerton Conf. Communication, Control and Computing, 2011, 744-751.
    [65] [ Y.-X. Wang and Y.-J. Zhang, Nonnegative matrix factorization:A comprehensive review, IEEE Transactions on Knowledge and Data Engineering, 25(2013), 1336-1353.
    [66] [ S. Wei and Z. Lin, Analysis and improvement of low rank representation for subspace segmentation, arXiv:1107.1561.
    [67] [ Z. Wen, W. Yin and Y. Zhang, Solving a low-rank factorization model for matrix completion by a nonlinear successive over-relaxation algorithm, Mathematical Programming Computation, 4(2012), 333-361.
    [68] [ J. Wright, A. Ganesh, S. Rao, Y. Peng and Y. Ma, Robust principal component analysis:Exact recovery of corrupted low-rank matrices via convex optimization, Advances in Neural Information Processing Systems, 2009, 2080-2088.
    [69] [ L. Wu, A. Ganesh, B. Shi, Y. Matsushita, Y. Wang and Y. Ma, Robust photometric stereo via low-rank matrix completion and recovery, Asian Conference on Computer Vision, 2010, 703-717.
    [70] [ L. Yang, Y. Lin, Z. Lin and H. Zha, Low rank global geometric consistency for partial-duplicate image search, International Conference on Pattern Recognition, 2014, 3939-3944.
    [71] [ M. Yin, J. Gao and Z. Lin, Laplacian regularized low-rank representation and its applications, IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(2016), 504-517.
    [72] [ Y. Yu and D. Schuurmans, Rank/norm regularization with closed-form solutions:Application to subspace clustering, Uncertainty in Artificial Intelligence, 2011, 778-785.
    [73] [ H. Zhang, Z. Lin and C. Zhang, A counterexample for the validity of using nuclear norm as a convex surrogate of rank, European Conference on Machine Learning, 8189(2013), 226-241.
    [74] [ H. Zhang, Z. Lin, C. Zhang and E. Chang, Exact recoverability of robust PCA via outlier pursuit with tight recovery bounds, AAAI Conference on Artificial Intelligence, 2015, 3143-3149.
    [75] [ H. Zhang, Z. Lin, C. Zhang and J. Gao, Robust latent low rank representation for subspace clustering, Neurocomputing, 145(2014), 369-373.
    [76] [ H. Zhang, Z. Lin, C. Zhang and J. Gao, Relation among some low rank subspace recovery models, Neural Computation, 27(2015), 1915-1950.
    [77] [ T. Zhang, B. Ghanem, S. Liu and N. Ahuja, Low-rank sparse learning for robust visual tracking, European Conference on Computer Vision, 7577(2012), 470-484.
    [78] [ Z. Zhang, A. Ganesh, X. Liang and Y. Ma, TILT:Transform invariant low-rank textures, International Journal of Computer Vision, 99(2012), 1-24.
    [79] [ Z. Zhang, X. Liang and Y. Ma, Unwrapping low-rank textures on generalized cylindrical surfaces, International Conference on Computer Vision, 2011, 1347-1354.
    [80] [ Z. Zhang, Y. Matsushita and Y. Ma, Camera calibration with lens distortion from low-rank textures, IEEE Conference on Computer Vision and Pattern Recognition, 2011, 2321-2328.
    [81] [ Y. Zheng, X. Zhang, S. Yang and L. Jiao, Low-rank representation with local constraint for graph construction, Neurocomputing, 122(2013), 398-405.
    [82] [ X. Zhou, C. Yang, H. Zhao and W. Yu, Low-rank modeling and its applications in image analysis, ACM Computing Surveys, 47(2014), p36.
    [83] [ G. Zhu, S. Yan and Y. Ma, Image tag refinement towards low-rank, content-tag prior and error sparsity, in International conference on Multimedia, 2010, 461-470.
    [84] [ L. Zhuang, H. Gao, Z. Lin, Y. Ma, X. Zhang and N. Yu, Non-negative low rank and sparse graph for semi-supervised learning, IEEE International Conference on Computer Vision and Pattern Recognition, 2012, 2328-2335.
    [85] [ W. Zuo and Z. Lin, A generalized accelerated proximal gradient approach for total-variation-based image restoration, IEEE Transactions on Image Processing, 20(2011), 2748-2759.
  • This article has been cited by:

    1. Xuya Liu, Shumei Wang, Shujun Fu, Li Yuliang, Shouyi Liu, Weifeng Zhou, An Efficient Collaborative Filtering Method for Image Noise and Artifact Removal, 2020, 8, 2169-3536, 124158, 10.1109/ACCESS.2020.3005024
    2. Yunfang Fu, Qiuqi Ruan, Ziyan Luo, Yi Jin, Gaoyun An, Jun Wan, FERLrTc: 2D+3D facial expression recognition via low-rank tensor completion, 2019, 161, 01651684, 74, 10.1016/j.sigpro.2019.03.015
    3. Sanyang Liu, Chong Zhang, 2018, Randomized Method for Robust Principal Component Analysis, 9781450365123, 1, 10.1145/3207677.3278070
    4. Ting Ge, Ning Mu, Tianming Zhan, Zhi Chen, Wanrong Gao, Shanxiang Mu, Brain Lesion Segmentation Based on Joint Constraints of Low-Rank Representation and Sparse Representation, 2019, 2019, 1687-5265, 1, 10.1155/2019/9378014
    5. Fengling Wang, Image Restoration Based on Gradual Reweighted Regularization and Low Rank prior, 2020, 2020, 1024-123X, 1, 10.1155/2020/9365405
    6. Zhanxuan Hu, Feiping Nie, Rong Wang, Xuelong Li, Low Rank Regularization: A review, 2021, 136, 08936080, 218, 10.1016/j.neunet.2020.09.021
    7. Regev Cohen, Yi Zhang, Oren Solomon, Daniel Toberman, Liran Taieb, Ruud JG van Sloun, Yonina C. Eldar, 2019, Deep Convolutional Robust PCA with Application to Ultrasound Imaging, 978-1-4799-8131-1, 3212, 10.1109/ICASSP.2019.8683030
    8. Thierry Bouwmans, Sajid Javed, Hongyang Zhang, Zhouchen Lin, Ricardo Otazo, On the Applications of Robust PCA in Image and Video Processing, 2018, 106, 0018-9219, 1427, 10.1109/JPROC.2018.2853589
    9. Rongrong Liu, Yassine Ruichek, Mohammed El Bagdouri, Multispectral background subtraction with deep learning, 2021, 80, 10473203, 103267, 10.1016/j.jvcir.2021.103267
    10. Fan Li, Shaoquan Zhang, Bingkun Liang, Chengzhi Deng, Chenguang Xu, Shengqian Wang, Hyperspectral Sparse Unmixing With Spectral-Spatial Low-Rank Constraint, 2021, 14, 1939-1404, 6119, 10.1109/JSTARS.2021.3086631
    11. Touseef Ahmad, Soumyendu Raha, Rosly B. Lyngdoh, Anand S. Sahadevan, Praveen K. Gupta, Arundhati Misra, Robust generalized bilinear model with weighted low-rank representation for hyperspectral image unmixing, 2022, 16, 1931-3195, 10.1117/1.JRS.16.024524
    12. Dian Jin, Xin Bing, Yuqian Zhang, Unique Sparse Decomposition of Low Rank Matrices, 2023, 69, 0018-9448, 2452, 10.1109/TIT.2022.3223230
    13. Thierry Bouwmans, Andrews Sobral, Sajid Javed, Soon Ki Jung, El-Hadi Zahzah, Decomposition into low-rank plus additive matrices for background/foreground separation: A review for a comparative evaluation with a large-scale dataset, 2017, 23, 15740137, 1, 10.1016/j.cosrev.2016.11.001
    14. Tongtong Zhang, Yu Zhou, Yuanxiang Li, Xian Wei, SatensoRF: Fast Satellite Tensorial Radiance Field for Multidate Satellite Imagery of Large Size, 2024, 62, 0196-2892, 1, 10.1109/TGRS.2024.3382632
    15. Junbiao Pang, Anjing Hu, Qingming Huang, Bundle fragments into a whole: Mining more complete clusters via submodular selection of interesting webpages for web topic detection, 2024, 09574174, 125125, 10.1016/j.eswa.2024.125125
  • Reader Comments
  • © 2016 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(5316) PDF downloads(701) Cited by(14)

Article outline

Figures and Tables

Figures(16)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog