Literature DB >> 35057728

A truncated nuclear norm and graph-Laplacian regularized low-rank representation method for tumor clustering and gene selection.

Qi Liu1.   

Abstract

BACKGROUND: Clustering and feature selection act major roles in many communities. As a matrix factorization, Low-Rank Representation (LRR) has attracted lots of attentions in clustering and feature selection, but sometimes its performance is frustrated when the data samples are insufficient or contain a lot of noise.
RESULTS: To address this drawback, a novel LRR model named TGLRR is proposed by integrating the truncated nuclear norm with graph-Laplacian. Different from the nuclear norm minimizing all singular values, the truncated nuclear norm only minimizes some smallest singular values, which can dispel the harm of shrinkage of the leading singular values. Finally, an efficient algorithm based on Linearized Alternating Direction with Adaptive Penalty is applied to resolving the optimization problem.
CONCLUSIONS: The results show that the TGLRR method exceeds the existing state-of-the-art methods in aspect of tumor clustering and gene selection on integrated gene expression data.
© 2021. The Author(s).

Entities:  

Keywords:  Clustering; Gene selection; Graph-Laplacian; Low-rank representation; Truncated nuclear norm

Mesh:

Year:  2022        PMID: 35057728      PMCID: PMC8772046          DOI: 10.1186/s12859-021-04333-y

Source DB:  PubMed          Journal:  BMC Bioinformatics        ISSN: 1471-2105            Impact factor:   3.169


Background

In most countries, cancer is the first or second cause of death [1]]. Thus, it is a hot topic to prevent and cure cancer effectively in the medical field. Genes can regulate critical movements of organisms, even including the emergence of cancer [2]. As the improvement of gene sequencing technology, plenty of genomic data are available, which is conducive to researching the pathogenesis of cancer [3]. However, the majority of genomic data have the features of high dimension and small sample, which hinders the advances in medicine studies [4, 5]. Evidently, data dimension reduction acts a momentous role in the process of genomic data analysis. Data dimension reduction aims to get a significant low-dimensional representation of high-dimensional data, remove redundant features and prevent overfitting. Therefore, it has achieved great successes in many areas, such as characteristic gene selection [6], image analysis [7], and text documents [8]. Principal Component Analysis (PCA) is one of the most classic linear dimension reduction methods [9]. Based on the high efficiency of PCA, it has been widely used on different kinds of data and developed in many fields [5, 10, 11]. To boost the robustness of PCA, Candès et al. developed a new PCA in [10], called Robust PCA (RPCA), which is exploited for background modeling from video and analyzing face images. Moreover, in [5], RPCA is exploited for discovering differentially expressed genes by Liu et al. Though, the above PCA methods all have obtained excellent results, the performance of these methods is corrupted with the noisy observation data. In [12], Wright et al. introduced a low-rank matrices recovery approach for removing the noise of data. Then, PCA is applied to the low-rank matrix. The robustness of data processing is enhanced significantly through the approach in [12]. The experimental data X are usually obtained from a union of multiple subspaces rather than a single space, where indicates low-dimensional space hidden in high-dimensional space [13-15]. Since these methods related to PCA prefer to research the data obtained from a single low-dimensional space, Liu et al. proposed a low-rank representation (LRR) model that can excavate the global distribution between data points to study X [16]. LRR strives to look for the lowest rank matrix representation about original data and has got brilliant results in several applications [16, 17]. However, LRR still exists a few shortages, for instance it cannot reveal the local manifold structures of data obtained from a non-linear low-dimensional manifold. Joyfully, various manifold learning models have been put forward, such as ISOMAP [18], Laplacian Eigenmap (LE) [19], Locally Linear Embedding (LLE) [20], and graph-Laplacian regularization [21]. A graph-Laplacian regularized LRR (LLRR) model [14] was developed, which introduces the graph regularization into LRR. In LLRR model, the useful rules hiding among the data points including the global geometric structure and the internal similarity information are all seized. LLRR only exploits one view of data, i.e. data manifold for data analysis. Contrasted with LLRR, Latent LRR (LatLRR) model adds another view, i.e. feature manifold to do image processing [22]. For solving these minimization problems of LRR, LLRR and LatLRR, the common point is to use the nuclear norm to approximate the rank operator. Given a data matrix X, the nuclear norm means that the sum of all singular values belonging to X. Since the nuclear norm minimizes the sum of all singular values for accomplishing the minimization problem, all non-zero singular values have different influences for the rank [23]. Thus, the nuclear norm maybe not the best way to approximate the rank of the matrix. To better approximate the rank and handle the non-convex optimizing problems, the truncated nuclear norm (TNN) was proposed in [24] and attracted much attention [13, 23, 25, 26]. The TNN that is the sum of few smallest singular values of a matrix can dispel the harm of shrinkage of the leading singular values, so it may be a more robust regularization to get the rank of a matrix than the nuclear norm. To strengthen the efficiency and robustness of the model, in our paper, a novel LRR method is developed, named Truncated nuclear norm and graph-Laplacian regularized Low-Rank Representation model (TGLRR). In the objective function of TGLRR, the nuclear norm is replaced by the TNN for reaching the robust approximation of rank function, a graph-Laplacian regularization is imposed to find the local manifold structure, and the L1-norm is used for realizing the sparse constraints of outliers. The main contributions of our paper are showed as follows. Firstly, compared with the popular LRR model regularized by the nuclear norm, our TGLRR method can obtain a better performance by the TNN, and solve the non-convex and discontinuous issues. Secondly, the TGLRR method can seize the valuable information lying in data manifold and feature manifold simultaneously. Finally, the TGLRR method can capture the internal similarity information and some underlying affinity among data points by incorporating a graph regularization term and utilizing a linear association of some bases to represent each data point. The remainder of this article is organized as follows. In the Results section, TGLRR is exploited for clustering and feature selection on integrated gene expression data. In Conclusions section, conclusions and the future work are given. In Methods section, our TGLRR method is put forward and the optimization problem is resolved through an efficient framework based on LADMAP [27].

Results

Integrative gene expression datasets

To validate the performance of TGLRR model, six clustering experiments and a feature selection experiment are conducted. The experimental data are integrative cancer gene expression data instead of single cancer data for avoiding sample imbalance problem. Seven different datasets are produced via integrating five different single gene expression data downloaded from The Cancer Genome Atlas (TCGA, https://www.cancer.gov/about-nci/organization/ccg/research/structural-genomics/tcga). The pertinent information about the seven integrative datasets is listed in Table 1.
Table 1

Description about seven integrative gene expression datasets

DatasetsGenesSamplesSamples classes
PAAD-COAD20502176-2622
HNSC-ESCA20502398-1832
CHOL-HNSC-ESCA2050236-398-1833
COAD-PAAD-ESCA20502262-176-1833
PAAD-ESCA-HNSC20502180-192-4183
HNSC-PAAD-CHOL-ESCA20502398-176-36-1834
ESCA-COAD-CHOL-PAAD20502183-262-36-1764
Description about seven integrative gene expression datasets PAAD, ESCA, COAD, CHOL and HNSC are the abbreviations of Pancreatic Ductal Adenocarcinoma, Esophageal Carcinoma, Colorectal Adenocarcinoma, Cholangiocarcinoma, Head and Neck Squamous Cell Carcinoma, respectively. Taking PAAD-COAD dataset for example, it is only composed of the tumor samples of PAAD and COAD data, in which PAAD data are made up of 176 tumor samples and 4 normal samples, and COAD data consist of 262 tumor samples and 19 normal samples. HNSC-ESCA, CHOL-HNSC-ESCA, COAD-PAAD-ESCA, HNSC-PAAD-CHOL-ESCA and ESCA-COAD-CHOL-PAAD datasets are also made in the production way of PAAD-COAD dataset. But PAAD-ESCA-HNSC dataset is composed of the whole samples of PAAD, ESCA and HNSC. To eliminate redundant features and avoid over-fitting, the dimension of data matrix X is reduced before clustering and feature selection experiment, which can also greatly abate the computational cost. PCA is chosen for dimension reduction experiments in our paper. In addition, 2000-dimensional data X are obtained after dimension reduction.

Parameters selection

There are three important parameters, i.e. regularization terms , and r of the TNN in our TGLRR model. The grid search is used to pick up the values of , and r. Figure 1 shows that clustering results are varied with the parameters and on three distinct integrative datasets of tumor gene expression.
Fig. 1

The clustering performance of TGLRR model versus parameter and . a the clustering results on PAAD-COAD dataset, b the results on COAD-PAAD-ESCA dataset, c the results on HNSC-PAAD-CHOL-ESCA dataset

The clustering performance of TGLRR model versus parameter and . a the clustering results on PAAD-COAD dataset, b the results on COAD-PAAD-ESCA dataset, c the results on HNSC-PAAD-CHOL-ESCA dataset X-axis represents the values range of , Y-axis represents the values range of , and Z-axis represents the clustering accuracy in Fig. 1. It can be distinctly found that the effect of clustering accuracy stems from that is greater than the effect from , especially in HNSC-PAAD-CHOL-ESCA dataset. Finally, it can be found that TGLRR performs well when and on PAAD-COAD dataset, and on HNSC-ESCA dataset, and on CHOL-HNSC-ESCA dataset, and on COAD-PAAD-ESCA dataset, and on HNSC-PAAD-CHOL-ESCA dataset, and and on ESCA-COAD-CHOL-PAAD dataset, respectively. Different from the method in [25] that tries all the possible values to seek the optimal value of r, the method in [28] is used to choose the optimal value of r. A curve graph showing the singular values needs to be drawn. Figure 2 shows the summary curve graph on six datasets applied in clustering experiments.
Fig. 2

The singular values on six distinct matrices

The singular values on six distinct matrices X-axis indicates the number and Y-axis denotes the singular values in Fig. 2. The value of the first inflection point in each curve is chosen as the value of r corresponding to each dataset. The principle of selecting r is that the singular values before inflection point are bigger than the singular values after inflection point. Therefore, the values of r on PAAD-COAD, HNSC-ESCA, CHOL-HNSC-ESCA, COAD-PAAD-ESCA, HNSC-PAAD-CHOL-ESCA and ESCA-COAD-CHOL-PAAD datasets are set as 2, 3, 3, 2, 2 and 3, respectively.

Convergence analysis

Since Algorithm 1 is a practical application based on LADMAP framework whose convergence has been proved in [27], Algorithm 1 should also be convergent. Many approaches are able to demonstrate the convergence property of algorithms [23, 29]. In our paper, an efficient approach in [29] by means of auxiliary function is exploited to validate the convergence property of TGLRR method. The results are exhibited in Fig. 3.
Fig. 3

Convergence curves of TGLRR on gene expression data

Convergence curves of TGLRR on gene expression data The abscissa in Fig. 3 indicates the iteration number and the ordinate denotes the loss function value. As shown in Fig. 3, our model is convergent. The TGLRR method begins to converge after 30 iterations on two datasets, such as HNSC-ESCA and CHOL-HNSC-ESCA. On other four datasets, the TGLRR method converges in 40 iterations. Here, the HNSC-ESCA dataset may be easily addressed, so our method begins to converge on the two datasets after 30 iterations while it needs 40 iterations on the other four datasets.

Clustering results

In this subsection, the TGLRR is applied for clustering, and compared with K-means, LLRR [14], LRR [30], RPCA [5], DGLRR [31], and LatLRR [22]. In respect of the dictionary matrix X, the optimal solution to TGLRR is able to symbolize “the minimum rank representation” of the data matrix X. What’s more, the i-th column about Z could be regarded as a “better” reflection of the i-th column about X so as to make the subspace structure more easily detectable [31]. Namely, the optimal solution could include almost all the sample information about integrative gene expression data X. Therefore, can be used for clustering experiments by K-means. To measure the performance of our approach, three quantity metrics are adopted in this paper, i.e., accuracy (ACC), normalized mutual information (NMI) and F-measure. As a widely used metric in machine learning field, ACC can be defined as follows: where n indicates the total number of tumor samples in an integrated data, is a delta function set to 1 only when and 0 otherwise, denotes true class label of the i-th sample, and represents the cluster label produced by the algorithms. is a mapping function permuting every to match real sample label. The second index of NMI is defined by where T and denote two different tumor index sets separately. H(T) and represent the entropy in T and , respectively. And where P(t) is the marginal probability distribution function, namely, the probabilities that a tumor sample arbitrarily chosen from an integrated dataset belongs to cluster T. In addition, indicates the joint probabilities that a tumor sample belongs to the two clusters T and simultaneously. F-measure is the comprehensive evaluation index considering both precision and recall, and written as: where and . TP, FP, TN and FN indicate the true-positive, false-positive, true-negative and false-negative, respectively. To prove the effectiveness of TGLRR, the detailed clustering results of these methods on integrative tumor gene expression data are listed by three tables. In tables, the values about ACC, NMI and F-measure are the average of 100 clustering results of each approach, and the values on the right of ± are the variance of 100 results. Table 2 reports the clustering results on PAAD-COAD and HNSC-ESCA datasets. Obviously, our TGLRR method exceeds other six comparison methods on PAAD-COAD dataset. TGLRR is more robust than other six methods on PAAD-COAD dataset from the point of the variance values. HNSC-ESCA data are extraordinary and may be easily addressed, in which the clustering results about all algorithms are good and particularly the evaluation indices of RPCA and TGLRR are 1.
Table 2

The clustering results on PAAD-COAD and HNSC-ESCA integrative data

PAAD-COADHNSC-ESCA
ACC(%)NMI(%)F-measure(%)ACC(%)NMI(%)F-measure(%)
K-means91.57 ± 0.8968.77 ± 4.2491.62 ± 1.0199.36 ± 0.0598.00 ± 0.5098.81 ± 0.18
LLRR93.95 ± 0.2971.59 ± 1.2993.83 ± 0.2699.83 ± 0.0098.07 ± 0.0099.80 ± 0.00
LRR93.63 ± 0.5770.70 ± 2.5293.64 ± 0.5199.83 ± 0.0098.07 ± 0.0099.80 ± 0.00
RPCA93.81 ± 0.4671.09 ± 2.2793.81 ± 0.42100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
DGLRR94.14 ± 0.4071.98 ± 1.8094.13 ± 0.4799.83 ± 0.0098.07 ± 0.0099.80 ± 0.00
LatLRR93.76 ± 0.3371.46 ± 1.5093.77 ± 0.2999.83 ± 0.0098.07 ± 0.0099.80 ± 0.00
TGLRR95.15 ± 0.0074.44 ± 0.0095.10 ± 0.00100.00 ± 0.00100.00 ± 0.00100.00 ± 0.00
The clustering results on PAAD-COAD and HNSC-ESCA integrative data The clustering results on two integrated datasets containing three types of tumors are exhibited in Table 3. It can be seen that the clustering performance of TGLRR model outperforms other models on COAD-PAAD-ESCA dataset. On CHOL-HNSC-ESCA dataset, TGLRR’s ACC, NMI and F-measure values are higher than values obtained via other five models except for LRR. Consequently, it still can be said that the TGLRR method outstrips other methods on CHOL-HNSC-ESCA dataset.
Table 3

The clustering results on CHOL-HNSC-ESCA and COAD-PAAD-ESCA data

CHOL-HNSC-ESCACOAD-PAAD-ESCA
ACC (%)NMI (%)F-measure (%)ACC (%)NMI (%)F-measure (%)
K-means83.49 ± 1.7776.23 ± 3.4777.05 ± 3.4283.61 ± 3.2576.11 ± 2.8581.95 ± 4.28
LLRR96.73 ± 0.8094.80 ± 1.2694.69 ± 2.0987.07 ± 2.2379.50 ± 1.4686.19 ± 2.94
LRR97.13 ± 0.4295.32 ± 0.6196.16 ± 0.8288.16 ± 1.9880.27 ± 1.4887.47 ± 2.53
RPCA85.40 ± 2.6481.43 ± 4.1681.26 ± 4.5985.59 ± 2.7278.85 ± 2.3983.98 ± 3.49
DGLRR94.70 ± 1.0392.33 ± 1.7891.63 ± 2.6286.14 ± 2.3878.67 ± 1.8284.93 ± 3.11
LatLRR93.94 ± 1.5791.37 ± 2.4691.57 ± 3.3287.16 ± 2.5279.33 ± 1.9386.16 ± 3.30
TGLRR98.37 ± 0.0090.58 ± 0.0396.09 ± 0.0192.82 ± 0.7779.51 ± 0.9392.62 ± 0.91
The clustering results on CHOL-HNSC-ESCA and COAD-PAAD-ESCA data From Table 4, our TGLRR method outmatches other six methods on HNSC-PAAD-CHOL-ESCA and ESCA-COAD-CHOL-PAAD datasets.
Table 4

The clustering results on HNSC-PAAD-CHOL-ESCA and ESCA-COAD-CHOL-PAAD data

HNSC-PAAD-CHOL-ESCAESCA-COAD-CHOL-PAAD
ACC (%)NMI (%)F-measure (%)ACC (%)NMI (%)F-measure (%)
K-means78.42 ± 0.9471.34 ± 1.0372.19 ± 1.9282.49 ± 2.1577.01 ± 1.7675.71 ± 3.30
LLRR87.66 ± 0.9475.56 ± 0.4086.90 ± 2.0484.41 ± 2.0780.24 ± 1.2282.60 ± 2.73
LRR88.63 ± 0.3975.89 ± 0.2189.16 ± 0.8187.40 ± 2.0582.52 ± 1.2087.62 ± 2.17
RPCA84.85 ± 1.5480.29 ± 1.5081.72 ± 2.5783.39 ± 1.7379.28 ± 1.3176.86 ± 3.10
DGLRR86.68 ± 0.8575.22 ± 0.4184.99 ± 1.9485.99 ± 2.2681.52 ± 1.3984.01 ± 3.13
LatLRR85.14 ± 1.0273.96 ± 0.4384.26 ± 2.2186.04 ± 1.8581.49 ± 1.1482.37 ± 1.94
TGLRR93.46 ± 0.9382.83 ± 0.7590.90 ± 1.2090.62 ± 1.5379.87 ± 1.4990.34 ± 1.64
The clustering results on HNSC-PAAD-CHOL-ESCA and ESCA-COAD-CHOL-PAAD data

Feature selection

Cancers are commonly relevant to gene mutation or abnormal expression of genes. Thus, in this subsection, the TGLRR method is used to identify co-feature genes of PAAD, ESCA and HNSC from PAAD-ESCA-HNSC dataset. From the formula (14), a minimum solution can be got from an integrative gene expression data X via TGLRR scheme. can obtain the feature manifold structure lying in data. As a result, it can be applied in feature gene extraction. From the view of cancer, its pathogenesis may be related to gene mutation [32]. It is extremely meaningful to find out the feature genes inducing cancers from gene expression data. Similar to the subsection of Parameters Selection, 10−2, 103 and 4 are assigned to , and r. Table 5 exhibits the top 10 co-feature genes with the mean of highest relevance score distinguished by the TGLRR method from PAAD-ESCA-HNSC dataset. The related diseases, related pathways and coded proteins about these genes are gotten from GeneCards (https://www.genecards.org/). These genes are most likely to lead to PAAD, ESCA and HNSC simultaneously.
Table 5

The top 10 genes selected via TGLRR on PAAD-ESCA-HNSC

Gene EDRelevance scoreRelated diseasesCoded proteins
CDH1101.03, 96.95, 124.3, 107.43Gastric, breast, colorectal, thyroid and ovarian cancerCadherin superfamily
TGFB173.21, 44.14, 76.66, 64.67Camurati-Engelmann disease, Encephalopathy, Inflammatory Bowel Disease and ImmunodeficiencyTransforming Growth Factor-Beta Superfamily of Proteins
RELA27.63, 11.33, 41.36, 26.77Mucocutaneous Ulceration, Chronic and EpendymomaTranscription Factor
ANXA526.80, 10.31, 42.30, 26.47Pregnancy Loss, Recurrent 3 and Antiphospholipid SyndromeCalcium-Dependent Phospholipid Binding Proteins
RHOA27.48, 11.81, 31.46, 23.58Adenocarcinoma and Peripheral T-Cell LymphomaRho Family of Small GTPases
PTPN1113.04, 13.56, 43.23, 23.28Noonan Syndrome 1 and Juvenile Myelomonocytic LeukemiaProtein Tyrosine Phosphatase
CTNNA120.94, 19.40, 24.80, 21.71Macular Dystrophy, Patterned, 2 and Butterfly-Shaped Pigment DystrophyCell Adhesion Process Protein
IGF2R13.40, 19.07, 25.26, 19.24Hepatocellular Carcinoma and Inclusion-Cell DiseaseReceptor for Both Insulin-Like Growth Factor 2 and Mannose 6-Phosphate
RUNX110.85, 12.97, 25.61, 16.48Platelet Disorder, Familial, with Associated Myeloid Malignancy, leukemia and Isolated Delta-Storage Pool DiseaseTranscription Factor
EWSR112.55, 9.19, 27.33, 16.36Ewing Sarcoma and Desmoplastic Small Round Cell TumorMultifunctional Protein

Take the contents in the second column of the second row as an example, the first, second and third numeral are the relevance score of CDH1 gene to PAAD, ESCA and HNSC, respectively, and the fourth is the mean

The top 10 genes selected via TGLRR on PAAD-ESCA-HNSC Take the contents in the second column of the second row as an example, the first, second and third numeral are the relevance score of CDH1 gene to PAAD, ESCA and HNSC, respectively, and the fourth is the mean From Table 5, clearly, CDH1 gene with the highest relevance score can result in a host of cancers, which indicates that CDH1 may be a dangerous co-feature gene. What's more, PAAD, ESCA and HNSC are all correlative with CDH1 and RHOA, which can be affirmed from [33-38]. It is a verifiable fact that TGFB1 and RELA all serve as a predictor for PAAD and ESCA via consulting some literatures. Some data show that PTPN11 may induce HNSC and PAAD. From [39, 40], it can be seen that ESCA is relevant to IGF2R and RUNX1. In addition, the related pathways of RUNX1 and EWSR1 include transcriptional misregulation in cancer. So, RUNX1 and EWSR1 may be co-characteristic genes of PAAD, ESCA and HNSC. All in all, the TGLRR method is successful in identifying co-characteristic genes on the integrative gene expression datasets.

Discussions

The TGLRR method is applied to the tumor clustering and gene selection, and superior to the other methods. Based on above results, it can be affirmed that the TNN could capture more valuable information existed in data than the nuclear norm from data. By comparing the results of DGLRR, a conclusion can be drawn that the graph Laplacian regularization imposed on feature manifold may cause adverse effects for clustering on our integrative datasets. The TGLRR method has some limitations. For example, on HNSC-PAAD-CHOL-ESCA dataset, the variance values of TGLRR is larger than LRR. It may be caused by the integrated datasets and its stability needs to be improved in future. In a word, these improvements to the prevenient LRR model can help TGLRR catch more useful information concealed in the low-dimensional manifold structure.

Conclusions

The paper proposes a Low-Rank Representation approach called TGLRR. It can capture the global and local geometric structures in data manifold via using the raw data matrix as the dictionary matrix and introducing the graph-Laplacian regularization term. Furthermore, TGLRR can gain a better approximation to the rank operator than the approaches regularized by the nuclear norm. The objective function of TGLRR is perfectly resolved through an iterative algorithm based on LADMAP framework. The efficiency and robustness of our TGLRR method are testified through the encouraging experimental results.

Methods

Related LRR methods

Based on the assumption that the observation data X are sampled from a union of several low-dimensional subspaces located in a high-dimensional spaces, LRR was raised in [16]. If data are noiseless, the rank minimization problem of LRR is written into where is the original data matrix and is a low-rank matrix recovered from X via LRR. is a basis matrix (or named dictionary matrix), which spans the whole data space linearly. The observation data generally exist more or less noise in real life, so the optimization problem (4) may be impracticable. The LRR model with noise is where is the reconstruction errors matrix (or called noise matrix). is a penalty parameter aiming to adjust the sparsity of matrix P and the reconstruction fidelity of data matrix X damaged by errors matrix P. is the L0-norm of matrix P, which indicates the number of non-zero elements in matrix P. Since the rank function is discrete, the problem (5) may have multiple solutions and the L0-minimization is non-convexity and intractable. Usually, solving the problem (5) is NP-hard [41]. To better solve the above rank minimization problem, the nuclear norm is imposed on the low-rank matrix, and the L0-norm is replaced with the L2,1-norm [17]. The convex optimization problem about LRR model is written as follows: where ( is the i-th largest singular value of Z) denotes the nuclear norm of matrix Z, and denotes the L2,1-norm of matrix P. To get a self-expression model, the observation data X are generally installed as the dictionary matrix [13, 14, 22]. The final LRR model becomes For low-rank matrix , its each element can reflect the manifold information, i.e. the similarity between the data point and the data point . Therefore, matrix Z can be seen as an affinity matrix [14]. LRR is devoted to seek the lowest rank representation of the observation data. With the help of an appropriate dictionary matrix, the underlying row space can be recovered via the lowest rank representation such that the true segmentation of data can be correctly revealed. Thus, LRR method can manage the data extracted from a union of multiple subspaces well [17]. Nevertheless, LRR method has to face two issues owing to the raw data X that are used as the basis. First, LRR method requires that the basis contains adequate data samples from the subspaces so as to possess the capacity of representing the underlying subspaces. Second, LRR method demands that noise of data X is little, i.e. only a part of X is corrupted. To remedy these two shortcomings of LRR, Liu et al. proposed the following convex optimization LRR problem [22]: where is the L1-norm of matrix P and G is the feature matrix separated from the original X. Equation (8) is a state-of-the-art LRR-based subspace learning model, named LatLRR. By means of LatLRR model, the observed sampling can be expressed via many unobserved sampling effectively [42]. In practical application, Z and G are applied in cluster analysis and feature selection, respectively.

Truncated nuclear norm (TNN)

The TNN is the summation of a few smaller singular values, i.e. the sum of some largest singular values is subtracted from the nuclear norm [24]. As an approximation of a rank operator, the largest r-th singular values could produce minor amount of information, meanwhile, the minimal -th singular values act a crucial role [23]. Compared to the nuclear norm, the TNN may be a better approximation to the rank operator. Its mathematical formula is where denotes the i-th largest singular value belongs to Z and r is a nonnegative integer and . Since the minimization of Eq. (9) is not convex, it cannot be directly resolved through the approaches. For overcoming this issue, Hu et al. come up with a theorem [25]. According to the Theorem, the equivalent transformation of Eq. (9) is achieved.

Graph-Laplacian regularization

Graph-Laplacian regularization is an outstanding manifold learning method, which can uncover the internal geometrical structures among the data points. As a result, naturally, appears a number of LRR models regularized by graph embedding manifold regularization [13, 43]. Given a k-nearest-neighbor graph G, suppose it has n vertices, and each vertex denotes a data point hidden in an underlying sub-manifold M [11]. Then, a symmetric weight matrix is constructed, where expresses the i weight of the edge linking vertices i and j. The value of every can be calculated via where indicates the k-nearest-neighbors of data point . Next, a diagonal matrix O, termed a degree matrix, need to be established. The value of the i-th member of O can be calculated by the sum of all the similarities associated with vertex , i.e. . The graph-Laplacian matrix L can be obtained by Finally, the graph embedding regularization term can be formulated by

Truncated nuclear norm and graph-Laplacian regularized low-rank representation method

Motivated by strengthening the robustness of LRR, our method (TGLRR) is put forward. Considering that some data may exist nonlinear geometric structure [14] and the disadvantages about the nuclear norm, the TNN and graph embedding manifold learning are introduced into our rank minimization problem to extract more essential information hidden in data. The objective function of TGLRR is formulated as follows: where and are the regularization parameters for balancing the contribution of each term. Essentially, TGLRR can get a more precise approximation to the rank function with the help of the TNN than the nuclear norm. And the underlying low-dimensional structures of data could be captured by the aid of the graph-Laplacian regularization and the basis matrix X.

Optimization solution

To correctly solve the optimization problem of (14), an efficient iterative algorithm based on LADMAP framework is designed. The algorithm (Algorithm 1) is implemented via alternating two iterative procedures till Eq. (14) converges to the minimum. The first step is to determine matrix A and B. Step 1: Given (k indicates the k-th updating), the SVD (Singular Value Decomposition) of need to be conducted. , where , and . and are calculated via and . The second step is to resolve the following convex optimization problem: Step 2: To achieve the separation of objective function (15), an auxiliary variable is introduced. Equation (15) is rewritten as follows: Equation (16) can be solved through LADMAP method, which introduces two Lagrangian multipliers Y1 and Y2. Thus, the augmented Lagrangian function can be defined as where is the penalty parameter and denotes the Frobenius norm of a matrix that is . Next, the alternating minimization strategy is adopted to compute Z, F, G and P. In the iterative procedure, Z, F, G or P is updated when the other three variables are fixed, respectively. Updating Z To get the solution of Z, the below minimization objective w.r.t. Z needs to be solved. Equation (18) has a closed-form solution: where indicates the Singular Value Thresholding operator (SVT),  +  and Updating G Similar to the solution of Z, the SVT operator is employed in computing G. The optimal solution is where and . Updating F The below sub-problem w.r.t. F is Equation (21) is the smooth convex planning problem. Different to the solving rules of Z and G, we can differentiate Eq. (21) and set it to zero to gain the answer of F. Its optimal solution is Updating P Calculating P has to optimize the following objective: The optimal solution to the above sub-problem w.r.t. P can be formulated by is the shrinkage operator defined as , . Updating , and After computing the above variables, two Lagrange multipliers and are given by The iteration rule about is The detailed algorithm about TGLRR model is showed in Algorithm 1.

Time complexity

In this subsection, the time complexity about TGLRR is discussed. Clearly, the main running time of TGLRR is expended on calculating the matrices Z, F, G and P. For the input data matrix X, it has m genes and n samples. The time complexity of SVD method with respect to Z is ( is the lowest rank of Z decided by algorithm 1). For the same activity, the time complexity of SVD decomposition of G is . The optimal solution of F can be obtained in . In the resolving procedure of P, also needs to be updated. The computational cost of P and needs and , respectively. Since, in our dataset, the total time cost of algorithm 1 is .
  16 in total

1.  A global geometric framework for nonlinear dimensionality reduction.

Authors:  J B Tenenbaum; V de Silva; J C Langford
Journal:  Science       Date:  2000-12-22       Impact factor: 47.728

2.  Robust recovery of subspace structures by low-rank representation.

Authors:  Guangcan Liu; Zhouchen Lin; Shuicheng Yan; Ju Sun; Yong Yu; Yi Ma
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2013-01       Impact factor: 6.226

3.  Dual Graph Regularized Latent Low-Rank Representation for Subspace Clustering.

Authors:  Ming Yin; Junbin Gao; Zhouchen Lin; Qinfeng Shi; Yi Guo
Journal:  IEEE Trans Image Process       Date:  2015-08-24       Impact factor: 10.856

4.  Hessian eigenmaps: locally linear embedding techniques for high-dimensional data.

Authors:  David L Donoho; Carrie Grimes
Journal:  Proc Natl Acad Sci U S A       Date:  2003-04-30       Impact factor: 11.205

5.  Supervised Discriminative Sparse PCA for Com-Characteristic Gene Selection and Tumor Classification on Multiview Biological Data.

Authors:  Chun-Mei Feng; Yong Xu; Jin-Xing Liu; Ying-Lian Gao; Chun-Hou Zheng
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-02-22       Impact factor: 10.451

6.  Recovering low-rank and sparse matrix based on the truncated nuclear norm.

Authors:  Feilong Cao; Jiaying Chen; Hailiang Ye; Jianwei Zhao; Zhenghua Zhou
Journal:  Neural Netw       Date:  2016-10-03

7.  Laplacian Regularized Low-Rank Representation and Its Applications.

Authors:  Ming Yin; Junbin Gao; Zhouchen Lin
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2016-03       Impact factor: 6.226

8.  Cancer genomics: Signature analysis suggests cancer origins.

Authors:  Isabel Lokody
Journal:  Nat Rev Genet       Date:  2013-08-28       Impact factor: 53.242

9.  Hyper-Graph Regularized Constrained NMF for Selecting Differentially Expressed Genes and Tumor Classification.

Authors:  Cui-Na Jiao; Ying-Lian Gao; Na Yu; Jin-Xing Liu; Lian-Yong Qi
Journal:  IEEE J Biomed Health Inform       Date:  2020-02-20       Impact factor: 5.772

10.  Graph Regularized Nonnegative Matrix Factorization for Data Representation.

Authors:  Deng Cai; Xiaofei He; Jiawei Han; Thomas S Huang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-12-23       Impact factor: 6.226

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.