Literature DB >> 28827693

Link predication based on matrix factorization by fusion of multi class organizations of the network.

Pengfei Jiao1, Fei Cai2,3, Yiding Feng2, Wenjun Wang2.   

Abstract

Link predication aims at forecasting the latent or unobserved edges in the complex networks and has a wide range of applications in reality. Almost existing methods and models only take advantage of one class organization of the networks, which always lose important information hidden in other organizations of the network. In this paper, we propose a link predication framework which makes the best of the structure of networks in different level of organizations based on nonnegative matrix factorization, which is called NMF 3 here. We first map the observed network into another space by kernel functions, which could get the different order organizations. Then we combine the adjacency matrix of the network with one of other organizations, which makes us obtain the objective function of our framework for link predication based on the nonnegative matrix factorization. Third, we derive an iterative algorithm to optimize the objective function, which converges to a local optimum, and we propose a fast optimization strategy for large networks. Lastly, we test the proposed framework based on two kernel functions on a series of real world networks under different sizes of training set, and the experimental results show the feasibility, effectiveness, and competitiveness of the proposed framework.

Entities:  

Year:  2017        PMID: 28827693      PMCID: PMC5566345          DOI: 10.1038/s41598-017-09081-9

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.379


Introduction

Many real world systems such as social, biological, computer, physical, can be modeled as complex networks[1]. Learning the structure, function and dynamic can help us to understand the formation mechanism, explore the evolution, and forecast the changes of the complex networks[2]. Lots of interesting research hotspots have been proposed, such as community detection[3], spreading dynamics[4], cascading reactions[5], network synchronization[6] and control[7]. Meanwhile, link predication, has a closeness relation to other research topics and a wide range of applications in reality[8], which devotes to estimate and predicate the unobserved or latent existent edges between pairs of nodes in the networks based on the observed linked structure. Link predication has been successfully applied to recommendation system[9], evaluation of network models[10], analysis of network evolution[11, 12], the predication of interactions between proteins in biological networks[13] and so on. The basic and important evidence is that two nodes are more likely linked if they are more similar[14]. There are a growing number of models and methods for link predication proposing recently[14]. These methods can be divided into three categories in general. The first class and classic methods are similarity-based methods, the hypothesis of which are that nodes are similar only if they are linked similar nodes or close to each other based on the distances denoted on the networks in various ways[8]. Such as the common neighbors (CN) index[15] and Jaccard index[8], which are based on the local similarity in the networks, the former denotes the number of common neighbors between the two nodes, the other denotes the ratio of the number of common neighbors and the number of the complete set of neighbors for two nodes. In addition, there are a lot of similarity-based methods, such as global similarity index, Katz[16], and Quasi-Local index, Local Path Index. The second class methods are probabilistic and statistical approaches, which assume that there are generative mechanisms for the network and these methods build a model to fit the observed structure and estimate model parameters and then compute the linked probability of all the unobserved links in candidate set, such as the hierarchical structure model[17] and the stochastic block model[18]. The third class are the algorithmic methods, which usually benefit the link predication as a supervised learning or optimization problem. Such as the matrix factorization model[19], which is usually used in link predication by extract features in the network and is also the foundation of our framework. However, all most of current link predication methods just take advantage of one class organization of the network. For example, the similarity-based methods just take advantage of one of specific similarity structures, such as common neighbors or Jaccard indexes; the hierarchical structure model, as a classic and popular statistical approach, infers hierarchical structure from the observed network for link predication based on the hierarchical rand graph model; the nonnegative matrix factorization (NMF) method extracts the basis matrix and coefficient matrix based on the observed network, which assumes that each pair nodes are independent, although Krishna et al.[19] considers the similarity-based index as a penalty term adding to the objective function of the NMF, there is no determined interpretations for that. A perturbation-based framework based on NMF[20] is proposed, which also could join some similarity-based index for link prediction. As we discussed above, whether the similarity-based methods, the probabilistic and statistical approaches, or the algorithmic methods, do not take full advantage of multi class organization of the observed networks in a simple, intuitive, principled, and interpreted way. How can we construct the different organizations of the complex networks in a principled way? A simple selection is the kernel function[21], which could enable the data to operate in a high-dimensional, implicit feature space, and it has been successfully applied to neural network and support vector machine. For that in a complex network, we can get various of organization structures of it by mapping the network with different kernel functions, called organization structures, which will help us to explore the structure and promote the performance of link predication in the complex network. In this paper, we propose a link predication framework which makes the best of the structure of networks in different level of organizations by use of kernel functions. Based on nonnegative matrix factorization, we proposed a framework that combine the adjacency matrix and one class of organization structure in a principled and effective way, which we called NMF 3 (Nonnegative Matrix Factorization based Fusion Framework). In detail, we first map the observed network into another space by kernel functions. Then we combine the adjacency matrix of the network and one of other organization structures, which makes us obtain the objective function of our framework for link predication based on the nonnegative matrix factorization. Thirdly, we derive an iterative algorithm to optimize the objective function, which converges to a local optimum, and we propose a fast optimization strategy for large networks. Lastly, we test the proposed framework based on two kernel functions on a series of real world networks under different sizes of training set, and the experimental results based on the prediction accuracy show the feasibility, effectiveness, and competitiveness of the proposed framework.

Results

In this section, we introduce the mathematical definition of link predication, the formation of proposed NMF 3, evaluation index and experimental results on a series of real world networks.

Definition of the link predication problem

As most of works about link predication denoted, we consider an unweight and undirected network G = (V, E), V and E represent the sets of nodes and edges in the network, respectively. n = |V| and m = |E| are the number of nodes and edges of this network, respectively. The adjacency matrix of the network is denoted as A, if nodes i and j has a link, then A  = A  = 1 and A  = A  = 0, otherwise. As demanded of link predication, we divide the edges of the network into training set and test set, denoted as E 1 and E 2, it is obvious that E = E 1 ∪ E 2 and E 1 ∩ E 2 = Ø. We use A 1 and A 2 denoting the matrix formation of E 1 and E 2 with all the nodes in V, respectively, and both of them are symmetric with 1 or 0 as the elements and A 1 + A 2 = A. We let L = |E 2|/2 be the number of edges in test set and it is easy to know |E 1| = 2(m − L), the number of all the possible edges in the network but out of the training set, we denote it as candidate set, is . Then we need to learn one model from the training set E 1, compute likelihood scores for each edge in the candidate set, select the edges with top L values, and validate that on the testing set E 2 based on some evaluation indexes.

Formation of proposed NMF3

Here, we will introduce the formation of our proposed link predication framework, including how to map the network into another space to get the other classes of organization structure of the network based on kernel functions and how to construct our proposed model.

Kernel function

Kernel functions have been widely applied to pattern recognition and machine learning, which are based on a fixed nonlinear feature space mapping ϕ(x), the kernel function is generally given by the relation For a given network, we can regard each column of the adjacency matrix, the first order link of one node, as the feature vector of the node. So a series of kernel functions can be applied to the network to get different classes of organization structure. Such as the polynomial kernel, a non-stationary kernel and for problems where all the training data is normalized, the gaussian kernel and exponential kernel, which are examples of radial basis functions, and so on. Without loss of generality, in this paper, we introduce two classic kernel functions as our instances in the proposed framework, called linear kernel[22] and covariance kernel[23]. They are denoted on the network asandwhere, X is the adjacency matrix of the observed network (or training set), X · is the i-th column of X and μ is the average value of all the X ·. It is obvious that both K 1(X) and K 2(X) are the symmetric positive matrices and there is no additional parameter in both them. The linear kernel K 1(X) could extract the local structure informations of the network, yet the covariance kernel K 2(X) could extract the global structure information. Although we just take advantage of the two kernel functions in our paper, we believe that the proposed framework can be easily extended and scaled.

Detail of the framework

Before introducing our proposed framework, we simply review the nonnegative matrix factorization for link predication. Based on the adjacency matrix X of observed network or training set, the objective function can be written aswhere D(X|WH) represents the distance between X and WH, such as the quadratic loss function or K − L divergence. W and H are the latent feature matrices(basis matrix and coefficient matrix), with the size of n × C and C × n, respectively, C is the number of latent features or the inner rank of X. f(W, H) is the penalty function about W and H, such as L 1 or L 2 norm[24]. Without loss of generality, we consider a simple case with the quadratic loss function, and rewrite the objective function asor in a matrix form as How to fuse the other organization structure with equation 5 in a principled way? Motivated by the nonnegative matrix factorization for recommendation system, we propose the objective function of the NMF 3 as followswhere R is the organization structure obtained by the kernel function, which has the same size with X. Parameter γ is used to scale the strength of R. After optimizing the equation 6, we can compute the similarity of all the edges in candidate set by WH. The setting of parameters γ and λ is in the experimental results, and the how to optimize the objective function, the detail algorithm and how to scale it to large networks can be seen in section methods.

Evaluation index

To quantify the performance of the link predication methods, we introduce three evaluation metric, area under the receiver operating characteristic curve (AUC)[25], Precision[26] and Prediction-Power[27]. In fact, link predication methods give an order list of all the edges in candidate set according the computing similarity values. Based on the rank of edges in , the AUC value is the probability that we randomly select an edge in test set E 2 with a higher rank score than a randomly selecting an edge in candidate set . The AUC can always be approximately calculated aswhere t, t′ and t′′ are the number of times of randomly independent comparisons, a higher score of an edge in test set E 2 and both two having a same score. If we select the edges with the top L similarity values, denoted as E , then the Precision can be computed aswhich represents the accuracy of link predication methods. As discussed in ref. 27, the Prediction-Power (PP) is denoted aswhere Precision is the performance of random-predictor can be computed by L/(n(n − 1)/2 − (m − L)), and this metric can assess deviation from the mean random-predictor performance.

Baseline methods

We compare our method with several well known methods, including CN, AA, RA, Salton, Jaccard, ACT and CRA indices and some widely used global methods, which are denoted as following, respectively. Common Neighbors (CN)[15], which is denoted between nodes x and y aswhere Γ(x) denotes the set of neighbors of node x. Adamic-Adar (AA)[28], which is denoted aswhere k is the degree of node z. This index considers the information about the degree of the common neighbors of the two nodes, and assigns the less-connected neighbors more weight. Resource Allocation (RA)[29], which is denoted asthe RA index assigns the different weight to the common neighbors. Salton Index[30], which is denoted asthe index is also based on the number of common neighbors yet with another normalization methods. Jaccard index[8], which is denoted aswhich is the ratio between the number of the intersection of Γ(x) and Γ(y) and the number of the union of that. Average Commute Time (ACT)[31], it is denoted aswhich means that two nodes are more similar if they have a smaller average commute time, and this similarity between the nodes x and y can be defined as the reciprocal of average commute time between x and y. Where represents the elements of matrix L +, the pseudo inverse of the Laplacian matrix of the network. CRA, which is a extend similar index based on the RA denoted by Carlo Vittorio Cannistraci in ref. 27. It is denoted aswhere α refers to the sub-set of neighbors of z that are also common neighbors of nodes x and y. SPM, which is called structural perturbation method[32] for link predication by assuming that the regularity of a network is reflected in the consistency of structural features before and after a random removal of a small set of links. HSM, the hierarchical structure model proposed by Aaron Clauset in ref. 17, which can infer hierarchical structure from network data and predict the missing links. SBM, which can deal the data reliability in complex networks and infer missing and spurious links based on the stochastic block model[18]. LR, a robust principal component analysis based[33] for estimate the missing links in complex networks and we set the weighting parameter to balance the low-rank property and sparsity as 0.1. LOOP, which is an algorithmic framework of probability by denoting a predefined structural Hamiltonian[34] based on the network organizing, and predict each non-observed link by computing the conditional probability of adding the link to the observed network. As far as we know, the methods LOOP and SPM have nearly best performance on link predication recently, however, both of two methods are time-consuming, especially the LOOP. In this paper, we denote NMF 3 − 1 and NMF 3 − 2 indicating the linear kernel and and covariance kernel based on the proposed framework.

Experimental results

Datasets

We evaluate the performance of our proposed framework by using ten quality networks from various areas, including social, biological, and technological network. The networks used in the experiment are described as follows and the basic statistical features are shown in Table 1. Directed links are treated as undirected, multiple links are treated as a single unweighted link and self loops are removed.
Table 1

The basic topological features of ten real networks studied in this paper, where |V| and |E| are the numbers of nodes and links, 〈k〉 is the average degree, CC is the clustering coefficient and 〈d〉 is the average shortest distance. H is the degree heterogeneity, as , and r is the assortative coefficient.

Networks|V||E|k CC d H r
Jazz198274227.7000.6182.2351.3950.020
USAir332212612.8100.7492.7403.4600.208
NetSci3799144.8200.7986.0401.6600.082
PB12221671427.3600.3602.7402.9700.221
Router502262582.4900.0336.4505.5000.138
C. elegans297214814.4700.3082.4601.8000.163
Yeast2375116939.8500.3885.1003.4800.454
Metabolic45320258.9400.6472.6644.4850.226
FWFD128207532.4420.3351.7761.2370.112
FWMW97144629.8140.4681.6931.2660.151
Jazz[35]: A collaboration network of jazz musicians consists of 198 nodes and 2742 interactions. USAir[32]: The air transportation network of USA consists of 332 nodes and 2126 links. The nodes of the network are airports, and each edge represents one airline. NetScience[36]: A coauthor-ship network of scientists working on network theory and experiment consists of 379 nodes and 914 links. The nodes of the network are the scientists, and each edge represents the cooperative relationship between the scientists. Politicablogs[37]: The network of American political blogosphere consists of 1222 nodes and 19021 links. The nodes of the network are blog page, and each edge represents the hyperlinks between the blog pages. Router[38]: A snapshot of the structure of the Internet at the level of autonomous systems consists of 5022 nodes and 6258 links. Celegans[39]: Neural network of elegans consists of 297 nodes and 2148 links. The nodes of the network are neurons, and each edge represents the gap junction between neurons. Yeas[40]: A protein-protein interaction network in budding yeast consists of 2375 nodes and 11693 interactions. The node of the network is protein, and the link represents its interactions. Metabolic[41]: A metabolic network of C. elegans consists of 453 nodes and 2025 interactions. FWFD http://vlado.fmf.uni-lj.si/pub/networks/data/bio/foodweb/foodweb.html: A food web in Florida Bay during the day season. The network contains 128 species of dry season and 2137 interactions. FWMW http://vlado.fmf.uni-lj.si/pub/networks/data/bio/foodweb/foodweb.html: A food web in Mangrove Estuary during the wet season consists of 97 nodes and 1493 interactions. The basic topological features of ten real networks studied in this paper, where |V| and |E| are the numbers of nodes and links, 〈k〉 is the average degree, CC is the clustering coefficient and 〈d〉 is the average shortest distance. H is the degree heterogeneity, as , and r is the assortative coefficient.

Results and analysis

Parameters setting: we select six networks including FWFD, FWMW, Jazz, Metabolic, USAir and Celegans from the all ten networks, and analyze the experimental sensitivity of γ and λ in our framework with the performance of link predication. As represented in Fig. 1, we set the proportion of training set as 0.9, and take the widely used evaluation index Precision for link predication as evidence. It is obvious that the performances on FWFD, Jazz, Metabolic, USAir and Celegans are gradual stable. Although the different settings of γ and λ have significant influence on the predict results, we also know that our framework has equally better performance than other baseline methods. Without losing generality, we set γ = 0.1 and λ = 2 in subsequent experiments.
Figure 1

Parameter sensitivity: we conducted the experiments of parameter sensitivity on six networks. We vary the number of γ and the λ to determine their impact on the network link prediction. The Each data point is averaged over 100 independent runs.

Parameter sensitivity: we conducted the experiments of parameter sensitivity on six networks. We vary the number of γ and the λ to determine their impact on the network link prediction. The Each data point is averaged over 100 independent runs. As represented in Tables 2, 3 and 4, we show the performance on the ten real world networks with the proportion of training set 0.9 based on AUC, Precision and PP, respectively. The black texts represent the largest value in each column, each row in the table represents the experimental results of one method including the average value and standard deviation over 100 times of dividing the network into training set and test set. From Table 2, our method NMF 3, SPM and LOOP have more competitive performance based on precision compared with other methods. As represented in Table 3, our method and SPM have better performance based on AUC, and the RA index is a best select in similarity-based index. In Table 4, we also show the mean value of PP of each method at last column. Also of note, the N/A in the tables represents that the value could not be computed for the corresponding method not applicable to large-scale networks. In general, our methods have nearly equal performance to LOOP and SPM based on the three evaluation index, however, the both two methods are more time-consuming compared with the proposed method in this paper and we analyze the computational complexity in section methods.
Table 2

Link prediction accuracy is measured by precision on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the precision is returned with an average run of over 100 times. For each data set, the presented links are partitioned into training set (90%) and test set (10%).

Precision(0.9)CelegansFWMWFWFDJazzmetabolicUSAirNetSciencePoliticalblogsRouterYeast
NMF 3 − 10.1520.660 0.581 0.6200.343 0.469 0.3270.171 0.174 0.537
0.0220.1140.0210.0080.0210.0130.0240.0070.0080.015
NMF 3 − 20.131 0.700 0.5300.5600.3160.3980.3380.1190.1600.408
0.0250.1220.0260.0190.0230.0100.0190.0090.0240.014
CRA0.1440.0330.0830.5590.2040.4060.3210.1790.0330.119
0.0080.0580.0180.0230.0210.0230.0390.0070.0130.015
CN0.1080.0000.0760.5080.1330.3830.3300.1710.0240.121
0.0110.0000.0050.0460.0210.0420.0460.0050.0000.018
AA0.1250.0000.0810.5320.1940.4150.5420.1680.0260.104
0.0140.0000.0050.0350.0080.0360.0210.0040.0030.015
RA0.1040.0000.0820.5430.281 0.469 0.736 0.1450.0110.115
0.0190.0000.0050.0280.0230.0330.0130.0040.0000.006
Salton0.0420.0000.0090.5370.0490.0590.3200.0070.0360.063
0.0050.0000.0130.0470.0130.0160.0070.0050.0030.001
Jaccard0.0630.0000.0080.5220.0490.0780.3010.0160.0180.031
0.0030.0000.0140.0460.0130.0150.0150.0030.0010.003
ACT0.0630.0000.1530.1670.0820.3290.1900.0700.0260.124
0.0150.0000.0050.0560.0230.0190.0120.0040.0090.016
SPM0.1330.5450.5700.6670.3150.4540.596 0.233 0.004 0.788
0.0250.1220.0260.0190.0230.0100.0190.0090.0240.014
HSM0.0850.4400.2610.3250.1090.1420.2990.1070.0640.081
0.0050.0020.0020.0260.0190.0110.0150.0030.0020.012
SBM0.1450.6010.4170.4100.1970.3350.1780.1100.1560.122
0.0040.0010.0030.0310.0150.0120.0090.0040.0030.015
LR0.1380.0500.5370.5590.2080.3990.0690.0740.0540.468
0.0060.0020.0020.0260.0190.0110.0150.0030.0020.012
LOOP 0.181 0.2000.564 0.685 0.394 0.466N/AN/AN/AN/A
0.0030.0010.0020.0300.0140.015N/AN/AN/AN/A
Table 3

Link prediction accuracy measured by AUC on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the AUC are returned with an average run of over 100 times. For each data set, the presented links are partitioned into training set (90%) and test set (10%).

AUC(0.9)CelegansFWMWFWFDJazzmetabolicUSAirNetSciencePoliticalblogsRouterYeast
NMF 3 − 1 0.908 0.996 0.9560.9600.9180.9560.7910.9510.703 0.972
0.0240.0050.0170.0140.0350.0320.0320.0150.0170.009
NMF 3 − 20.8940.984 0.960 0.9560.9100.9440.8210.9380.7510.969
0.0210.0360.0140.0300.0220.0230.0240.0180.0210.012
CRA0.7820.5000.6450.9820.8670.9350.8270.9000.5330.872
0.0520.0000.0610.0030.0190.0200.0080.0180.0130.019
CN0.8230.3750.5820.9400.9200.9600.9830.9320.5270.880
0.0160.0090.0580.0090.0230.0150.0080.0150.0090.012
AA0.8900.3700.6070.967 0.967 0.9650.9880.9100.5340.879
0.0350.0130.0670.0150.0200.0130.0030.0460.0150.018
RA0.8720.3900.583 0.990 0.9520.9750.9930.9180.5290.884
0.0080.0100.0290.0100.0080.0250.0060.0210.0130.016
Salton0.8020.3830.547 0.990 0.8050.9220.9950.8870.5400.870
0.0320.0320.1040.0000.0260.0370.0050.0400.0170.030
Jaccard0.7930.4000.5100.9700.7700.8820.9950.8720.5260.885
0.0080.0180.1150.0300.0310.0320.0050.0280.0100.022
ACT0.7500.4830.7000.7870.7570.9000.6130.903 0.918 0.910
0.0460.0600.0440.0120.0210.0400.0510.0120.0340.023
SPM0.833 0.996 0.8670.967 0.967 0.967 0.998 0.997 0.5850.930
0.0160.0090.0580.0090.0230.0150.0080.0150.0090.012
HSM0.8500.9400.8210.9120.8150.8550.8100.8510.7090.674
0.0250.0100.0250.0560.0230.0190.0120.0240.0190.016
SBM0.8600.9840.9410.9330.9080.9450.8990.8910.9100.770
0.0310.0090.0350.1520.0300.0210.0180.0180.0210.023
LR0.5730.5500.9060.8860.5850.8000.5700.5150.5350.800
0.0060.0020.0020.0260.0190.0110.0150.0030.0020.012
LOOP0.9010.8150.9550.9780.965 0.976 N/AN/AN/AN/A
0.0040.0010.0030.0310.0150.012N/AN/AN/AN/A
Table 4

Link prediction accuracy measured by Prediction-Power on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the AUC are returned with an average run of over 100 times.

PPCelegansFWMWFWFDJazzmetabolicUSAirNetSciencePoliticalblogsRouterYeast mean
NMF 3 − 11.4721.1841.2351.5852.2302.0672.4011.8743.5443.1102.070
NMF 3 − 21.4091.2091.1951.5402.1951.9962.4151.7163.5082.9912.017
CRA1.357−0.1170.4171.5532.0411.9852.3931.8942.8222.4561.680
CN1.172−1.6360.3501.4981.8191.9802.4051.8742.6842.4631.461
AA1.267−1.6360.3771.5181.9822.0142.6201.8662.7192.3971.512
RA1.320−1.6360.3851.5272.1432.0682.7531.8022.3452.4411.515
Salton0.737−1.636−0.5531.5221.3871.1702.3910.4862.8602.1801.055
Jaccard0.760−1.636−0.6321.5101.3871.2902.3650.8452.5591.8721.032
ACT1.061−1.6360.6561.0151.6091.9132.1651.4862.7192.4741.346
SPM1.4161.1011.2071.6162.1942.0532.6622.0081.9063.2771.944
HSM1.2201.0080.8871.3041.7321.5482.3621.6703.1102.2891.713
SBM1.4521.1431.0911.4051.7241.9212.1371.6823.4972.4671.852
LR1.4310.0631.2001.5392.0131.9971.7251.5103.0363.0511.757
LOOP1.5490.6651.2221.6282.2902.065N/AN/AN/AN/A1.570

For each data set, the presented links are partitioned into training set (90%) and test set (10%).

Link prediction accuracy is measured by precision on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the precision is returned with an average run of over 100 times. For each data set, the presented links are partitioned into training set (90%) and test set (10%). Link prediction accuracy measured by AUC on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the AUC are returned with an average run of over 100 times. For each data set, the presented links are partitioned into training set (90%) and test set (10%). Link prediction accuracy measured by Prediction-Power on the 10 real networks. We compared our methods (NMF 3 − 1, NMF 3 − 2) with other methods on the 10 network data sets and the AUC are returned with an average run of over 100 times. For each data set, the presented links are partitioned into training set (90%) and test set (10%). Furthermore, we analyze the experimental results on the networks with different fraction of training set from 0.9 to 0.2. As reported in Figs 2, 3 and 4, we show the results of Celegans, Jazz, USAir, Metabolic, FWFD and FWMW based on AUC, Precision and PP, respectively (For that it is time-consuming for global methods for larger networks, especially for the SPM and LOOP). The black lines represent the performance of the proposed NMF 3 − 1 and NMF 3 − 2 methods, the purple lines correspond to the SPM and LOOP methods, the rest lines are the other global methods (HSM, SBM, and LR) and similarity-based methods. From the results, it is obvious that SPM, LOOP and our methods have better and competitive performance that others. There are two different expressions in the figures, they are the FWFD and FWMW networks, on which our proposed framework has super performance than other methods. In fact, our methods have competitive and stable performance.
Figure 2

The comparison of AUC of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs.

Figure 3

The comparison of precision of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs.

Figure 4

Legend (The comparison of Prediction-Power of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs.

The comparison of AUC of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs. The comparison of precision of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs. Legend (The comparison of Prediction-Power of six networks under different fractions of missing links. Besides our kernel framework (NMF 3 − 1, NMF 3 − 2), we further compare our methods with eight well-known methods (AA, RA, CN, Salton, ACT, Jaccard, CRA, SPM, HSM, SBM, LP, LOOP). Each data point is averaged over 100 independent runs.

Discussion

In this paper, we have proposed a framework of link predication which could make multi class organizations of the network. We take two kernel functions as the special cases of the proposed framework and experiments show the feasibility, effectiveness, and competitiveness of the framework. As an extension to the nonnegative matrix factorization, our proposed framework for link predication not only inherits the advantages of which, but also take full advantage of multi organizations of the network based on kernel function. Furthermore, we proposed a gradient descent algorithm to optimize the object function and extend it to large networks. Other more, our framework is easy to be extended to directed and weighted networks, for that it is based on nonnegative matrix factorization, just by letting the X be directed and weighted. And we believe that this proposed method highlights the research in which taking different structure information for link predication. There are some limitations and improved studies for our proposed framework in future. One of which is how to set parameters γ and λ to be adaptive on different networks. For our framework only taking the adjacency matrix and one of other organization of the network, making the best of more classes of structure information of the network in a principled and effective way is our next work.

Methods

In this section, we introduce how to optimize the objective function 6 with a gradient descent algorithm, give a simple operation process for the algorithm and propose a strategy to scale the algorithm for larger networks.

Parameter learning

The determination of the number of latent features C is a very important and difficult problem in the matrix factorization. Here, for it is not our primary attention, we take an easy and effective method for automatic determination of C, Colibri [42], which seeks a nonorthogonal basis by sampling the columns of the input matrix. Because of the non convex of objective function 6, we alternate update W with fixed H and update H with fixed W under the Majorization-Minimization framework[43]. We rewrite the objective function 6 asHere, the ⋅ represents the element wise multiplication. To enforce the non-negativity constraints of W and H, we introduce the Lagrangian and write the equation 6 aswhere Φ and Ψ are the Lagrange multipliers, following the Karush- Kuhn-Tucker (KKT) optimality conditions[44], we set , and getand Then the KKT complimentary slackness conditions yieldand Following the works[45, 46], we can easy get the update rules of W and H asandwhich makes the objective function 6 converge to a local minimum.

Algorithm for NMF3

Here, we summed the algorithm for proposed NMF 3 based on the procedure of link predication in 1. Algorithm for the NMF 3 framework based on the procedure of link predication.

Complexity analysis and discussion

Here, we give a simple complexity analysis of the proposed algorithm. The most time-consuming parts are updating W and H, for each iteration, the time cost of (γR · X)H is O(N 2 C + N 2), the time cost of ((1 + γR) · (WHH ) + W) is NC 2 + N 2, so the total time cost of of the algorithm is O(N (N 2 C + NC 2 + N 2 + NC)) ~ O(n (N 2 C)), where n is the number of iterations. If we consider the sparse of real world networks, the time cost can be as O(n (mC)), where m is the number of the edge of the network. The most confused problem of the algorithm is that it just converges to a local minimum, so we need run the algorithm many times and chose a best one which has the least value of the objective function.

Scale to large networks

In order to deal the large networks, we rewritten the object function 6 ashere, the i ~ j indicates there existence an edge between nodes i and j, then we could only compute the observed links in the training set of the network. The optimization process of function 25 is similar to the algorithm 1.
  15 in total

1.  Clustering and preferential attachment in growing networks.

Authors:  M E Newman
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2001-07-26

2.  Constrained Clustering With Nonnegative Matrix Factorization.

Authors:  Xianchao Zhang; Linlin Zong; Xinyue Liu; Jiebo Luo
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2015-07-27       Impact factor: 10.451

3.  Community detection in complex networks using extremal optimization.

Authors:  Jordi Duch; Alex Arenas
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2005-08-24

4.  Finding community structure in networks using the eigenvectors of matrices.

Authors:  M E J Newman
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2006-09-11

5.  Hierarchical structure and the prediction of missing links in networks.

Authors:  Aaron Clauset; Cristopher Moore; M E J Newman
Journal:  Nature       Date:  2008-05-01       Impact factor: 49.962

6.  The meaning and use of the area under a receiver operating characteristic (ROC) curve.

Authors:  J A Hanley; B J McNeil
Journal:  Radiology       Date:  1982-04       Impact factor: 11.105

7.  Graph Regularized Nonnegative Matrix Factorization for Data Representation.

Authors:  Deng Cai; Xiaofei He; Jiawei Han; Thomas S Huang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-12-23       Impact factor: 6.226

8.  Measuring multiple evolution mechanisms of complex networks.

Authors:  Qian-Ming Zhang; Xiao-Ke Xu; Yu-Xiao Zhu; Tao Zhou
Journal:  Sci Rep       Date:  2015-06-11       Impact factor: 4.379

9.  Predicting missing links and identifying spurious links via likelihood analysis.

Authors:  Liming Pan; Tao Zhou; Linyuan Lü; Chin-Kun Hu
Journal:  Sci Rep       Date:  2016-03-10       Impact factor: 4.379

10.  From link-prediction in brain connectomes and protein interactomes to the local-community-paradigm in complex networks.

Authors:  Carlo Vittorio Cannistraci; Gregorio Alanis-Lobato; Timothy Ravasi
Journal:  Sci Rep       Date:  2013       Impact factor: 4.379

View more
  1 in total

1.  Integrative Analyses of Genes Associated With Otologic Disorders in Turner Syndrome.

Authors:  Ruoyan Xue; Qi Tang; Yongli Zhang; Mengyao Xie; Chen Li; Shu Wang; Hua Yang
Journal:  Front Genet       Date:  2022-02-22       Impact factor: 4.599

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.