| Literature DB >> 32038151 |
Yu Xie1, Peixuan Jin1, Maoguo Gong2, Chen Zhang1, Bin Yu1.
Abstract
Networks, such as social networks, biochemical networks, and protein-protein interaction networks are ubiquitous in the real world. Network representation learning aims to embed nodes in a network as low-dimensional, dense, real-valued vectors, and facilitate downstream network analysis. The existing embedding methods commonly endeavor to capture structure information in a network, but lack of consideration of subsequent tasks and synergies between these tasks, which are of equal importance for learning desirable network representations. To address this issue, we propose a novel multi-task network representation learning (MTNRL) framework, which is end-to-end and more effective for underlying tasks. The original network and the incomplete network share a unified embedding layer followed by node classification and link prediction tasks that simultaneously perform on the embedding vectors. By optimizing the multi-task loss function, our framework jointly learns task-oriented embedding representations for each node. Besides, our framework is suitable for all network embedding methods, and the experiment results on several benchmark datasets demonstrate the effectiveness of the proposed framework compared with state-of-the-art methods.Entities:
Keywords: graph neural network; link prediction; multi-task learning; node classification; representation learning
Year: 2020 PMID: 32038151 PMCID: PMC6989613 DOI: 10.3389/fnins.2020.00001
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Notations and their descriptions.
| The given network | |
| Set of nodes in the given network | |
| Set of edges in the given network | |
| A node | |
| An edge between nodes | |
| Number of nodes in the given network | |
| The number of class labels for nodes in | |
| The adjacency matrix of | |
| The dimension of learned node representations | |
| The initial feature matrix of nodes | |
| The embedding representation matrix of nodes |
Figure 1Graphical illustrations of our proposed multi-task network representation learning framework.
Figure 2Schematic depiction of implementation of the proposed framework on graph attention networks.
Statistics of benchmark datasets used in our experiments.
| Nodes | 2,708 | 3,327 | 19,717 |
| Edges | 5,429 | 4,732 | 44,338 |
| Text feature dimension | 1,433 | 3,703 | 500 |
| Classes | 7 | 6 | 3 |
Accuracy of semi-supervised node classification on Cora.
| GCN | 0.842 | 0.842 | 0.828 | 0.828 | 0.821 | 0.821 | 0.807 | 0.807 | 0.800 |
| αLoNGAE | 0.803 | 0.793 | 0.790 | 0.783 | 0.780 | 0.777 | 0.770 | 0.767 | 0.763 |
| GAT | 0.824 | 0.822 | 0.816 | 0.808 | 0.806 | 0.804 | 0.798 | 0.796 | 0.794 |
| MT-GAT (ours) |
The best results are shown in bold, and our MT-GAT with significant improvements over the baselines is shown with underlines.
Accuracy of semi-supervised node classification on Pubmed.
| GCN | 0.838 | 0.838 | 0.806 | 0.806 | 0.774 | 0.774 | 0.741 | 0.741 | |
| αLoNGAE | 0.807 | 0.803 | 0.800 | 0.797 | 0.796 | 0.793 | 0.790 | 0.787 | 0.786 |
| GAT | 0.794 | 0.792 | 0.790 | 0.788 | 0.786 | 0.784 | 0.782 | 0.780 | 0.788 |
| MT-GAT (ours) | 0.854 |
The best results are shown in bold, and our MT-GAT with significant improvements over the baselines is shown with underlines.
AUC and AP performance of different methods on link prediction.
| GAE | 0.910 | 0.920 | 0.895 | 0.899 | 0.964 | 0.965 |
| VGAE | 0.914 | 0.926 | 0.908 | 0.920 | 0.944 | 0.947 |
| LoNGAE | 0.896 | 0.915 | 0.860 | 0.892 | 0.926 | 0.930 |
| αLoNGAE | 0.952 | 0.960 | 0.963 | |||
| GCN | 0.809 | 0.811 | 0.811 | 0.822 | 0.828 | 0.834 |
| MT-GAT (ours) | 0.930 | 0.931 | 0.963 | |||
The best results are shown in bold.
Figure 3The effect of different hyperparameters α on the Cora dataset. We choose the AUC and AP scores of link prediction and the classification accuracy of node classification to demonstrate the effect of different hyperparameters for the experiments.
Accuracy of semi-supervised node classification on Citeseer.
| GCN | 0.846 | 0.824 | 0.824 | 0.824 | 0.813 | 0.802 | 0.802 | 0.780 | |
| αLoNGAE | 0.733 | 0.727 | 0.723 | 0.716 | 0.710 | 0.706 | 0.697 | 0.690 | 0.683 |
| GAT | 0.718 | 0.716 | 0.710 | 0.708 | 0.706 | 0.704 | 0.700 | 0.698 | 0.696 |
| MT-GAT (ours) |
The best results are shown in bold, and our MT-GAT with significant improvements over the baselines is shown with underlines.