| Literature DB >> 35783353 |
Zhen Xu1, Lanning Wei1,2, Huan Zhao1, Rex Ying3, Quanming Yao4, Wei-Wei Tu1, Isabelle Guyon5,6.
Abstract
Graph structured data is ubiquitous in daily life and scientific areas and has attracted increasing attention. Graph Neural Networks (GNNs) have been proved to be effective in modeling graph structured data and many variants of GNN architectures have been proposed. However, much human effort is often needed to tune the architecture depending on different datasets. Researchers naturally adopt Automated Machine Learning on Graph Learning, aiming to reduce human effort and achieve generally top-performing GNNs, but their methods focus more on the architecture search. To understand GNN practitioners' automated solutions, we organized AutoGraph Challenge at KDD Cup 2020, emphasizing automated graph neural networks for node classification. We received top solutions, especially from industrial technology companies like Meituan, Alibaba, and Twitter, which are already open sourced on GitHub. After detailed comparisons with solutions from academia, we quantify the gaps between academia and industry on modeling scope, effectiveness, and efficiency, and show that (1) academic AutoML for Graph solutions focus on GNN architecture search while industrial solutions, especially the winning ones in the KDD Cup, tend to obtain an overall solution (2) with only neural architecture search, academic solutions achieve on average 97.3% accuracy of industrial solutions (3) academic solutions are cheap to obtain with several GPU hours while industrial solutions take a few months' labors. Academic solutions also contain much fewer parameters.Entities:
Keywords: Automated Machine Learning; Graph Neural Networks; data challenge; graph machine learning; node classification
Year: 2022 PMID: 35783353 PMCID: PMC9245425 DOI: 10.3389/frai.2022.905104
Source DB: PubMed Journal: Front Artif Intell ISSN: 2624-8212
General information about winning teams.
|
|
|
|
|
|
| |
|---|---|---|---|---|---|---|
| 1st |
| Meituan Dianping | 6th |
| Tsinghua University | |
| 2nd |
| Nanjing University | 6th |
| Beijing University of Posts | |
| 3rd |
| Ant Financial | 8th |
| Self-employed | |
| 4th |
| Alibaba Inc. | 9th |
| Ant Financial | |
| 5th |
| Zhejiang University | 10th |
| Nanyang Tech. University | |
| 10th |
| Hikvision Inc. |
Two teams tie in the 6th and 10th place. We list them both.
Statistics of all datasets.
|
|
|
|
|
|
|
|
|
|
|
|
|---|---|---|---|---|---|---|---|---|---|---|
| a | Public | Citation | 2.7K | 5.3K | 1.4K | 7 | 1.9 | F | F | 5 |
| b | Public | Citation | 3.3K | 4.6K | 3.7K | 6 | 1.4 | F | F | 3 |
| c | Public | Social | 10K | 733K | 0.6K | 41 | 73.3 | F | F | 81 |
| d | Public | News | 10K | 2,917K | 0.3K | 20 | 291.7 | T | T | 467 |
| e | Public | Finance | 7.5K | 7.8K | 0 | 3 | 1.0 | F | F | 111 |
| f | Feedback | Sales | 10K | 194K | 0.7K | 10 | 19.4 | F | F | 18 |
| g | Feedback | Citation | 10K | 41K | 8K | 5 | 4.1 | F | F | 6 |
| h | Feedback | Medicine | 10K | 2,461K | 0.3K | 23 | 246.1 | T | T | 1,773 |
| i | Feedback | Finance | 15K | 16K | 0 | 3 | 1.1 | F | F | 213 |
| j | Feedback | Medicine | 11K | 22K | 0 | 9 | 2.0 | F | F | 227 |
| k | Private | Sales | 8K | 119K | 0.7K | 8 | 14.9 | F | F | 6 |
| l | Private | Citation | 10K | 40K | 7K | 15 | 4 | F | F | 34 |
| m | Private | News | 10K | 1,425K | 0.3K | 8 | 142.5 | T | T | 360 |
| n | Private | Finance | 14K | 22K | 0 | 10 | 1.6 | F | F | 61 |
| o | Private | Social | 12K | 19K | 0 | 19 | 1.6 | F | F | 62 |
“Avg Deg” is the average number of edges per node. “Directed” and “Weighted” indicate the two properties of a graph. “Skewness” here is calculated by the number of nodes in the largest class divided by the number of nodes in the smallest class.
Figure 1Illustration of AutoGraph scope. Industrial people provide a full pipeline solution that covers data preprocessing to evaluation. Academic researchers focus mainly on model architecture and hyperparameter optimization.
Accuracy and Balanced accuracy of top methods on all datasets (%).
|
|
|
|
|
|
| ||||
|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
| ||
| a | Public | 85.7 | 84.9 |
|
| 88.2 | 87.2 | 87.2 | 85.5 |
| b | Public | 71.4 | 67.8 | 75.2 |
|
|
| 75.6 | 69.0 |
| c | Public | 86.5 | 72.0 | 94.3 | 87.5 | 94.2 | 90.9 |
|
|
| d | Public | 93.7 | 6.1 |
|
| 95.1 | 28.8 | 94.6 | 21.0 |
| e | Public | 59.6 | 38.8 | 88.7 |
| 88.5 | 90.7 |
|
|
| f | Feedback | 86.6 | 78.2 |
|
| 92.3 |
| 92.4 | 91.4 |
| g | Feedback | 94.7 | 92.8 | 95.3 | 93.5 | 95.6 | 93.8 |
|
|
| h | Feedback | 90.4 | 8.8 |
|
| 92.2 | 17.6 | 92.1 | 16.6 |
| i | Feedback | 88.2 | 59.2 | 88.4 | 87.5 | 88.4 |
|
| 91.1 |
| j | Feedback | 90.7 | 68.1 | 95.9 | 89.0 | 96.1 |
|
| 93.3 |
| k | Private | 93.5 | 92.2 |
|
|
|
| 94.8 | 93.1 |
| l | Private | 90.9 | 84.5 |
|
| 94.7 | 91.8 | 94.5 |
|
| m | Private | 85.5 | 24.5 |
|
| 95.7 | 69.0 | 98.0 | 79.4 |
| n | Private | 85.6 | 47.3 |
| 97.3 |
|
| 98.9 | 97.0 |
| o | Private | 49.6 | 15.6 | 91.0 | 84.6 | 91.3 |
|
| 88.5 |
The baseline is a two layer GCN. Bold values are best in comparison with other methods.
Accuracy comparison of GCN baselines, F2GCN, and industrial best solution (%).
|
|
|
|
|
|
|---|---|---|---|---|
| a | 85.7 | 84.4 | 84.4 (95.4) | 88.5 (100) |
| b | 71.4 | 70.5 | 71.3 (94.8) | 75.2 (100) |
| c | 86.5 | 82.3 | 92.8 (98.4) | 94.3 (100) |
| d | 93.7 | 93.6 | 93.9 (97.3) | 96.5 (100) |
| e | 59.6 | 87.5 | 88.4 (99.7) | 88.7 (100) |
| f | 86.6 | 87.6 | 92.1 (99.2) | 92.8 (100) |
| g | 94.7 | 93.4 | 95.3 (100) | 95.3 (100) |
| h | 90.4 | 90.3 | 90.1 (96.4) | 93.5 (100) |
| i | 88.2 | 87.6 | 88.3 (99.9) | 88.4 (100) |
| j | 90.7 | 83.6 | 95.3 (99.4) | 95.9 (100) |
| k | 93.5 | 93.2 | 93.4 (97.9) | 95.5 (100) |
| l | 90.9 | 89.1 | 92.9 (97.9) | 94.9 (100) |
| m | 85.5 | 86.1 | 86.1 (87.8) | 98.1 (100) |
| n | 85.6 | 95.2 | 96.7 (97.7) | 99.0 (100) |
| o | 49.6 | 71.8 | 88.8 (97.6) | 91.0 (100) |
| Avg | − (97.3) | − (100) |
L2, L4 mean 2, and 4 layers for the GNN architecture. Numbers in parentheses are the relative accuracy with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage.
Figure 2Accuracy improvement with respect to baseline.
Number of parameters of baseline, 1st solution and F2GCN (Unit: Millions).
|
|
|
|
|
|---|---|---|---|
| a | 0.023 | 0.908 (75.7) | 1.199 (100) |
| b | 0.059 | 0.700 (44.2) | 1.583 (100) |
| c | 0.011 | 1.598 (98.0) | 1.631 (100) |
| d | 0.006 | 0.042 (3.20) | 1.296 (100) |
| e | 0.121 | 0.354 (31.8) | 1.114 (100) |
| f | 0.013 | 0.039 (2.30) | 1.688 (100) |
| g | 0.134 | 0.313 (13.1) | 2.389 (100) |
| h | 0.006 | 0.271 (20.9) | 1.294 (100) |
| i | 0.241 | 2.269 (113.0) | 2.013 (100) |
| j | 0.171 | 0.834 (60.6) | 1.376 (100) |
| k | 0.012 | 1.478 (108.0) | 1.395 (100) |
| l | 0.108 | 0.614 (25.6) | 2.395 (100) |
| m | 0.005 | 0.010 (0.80) | 1.278 (100) |
| n | 0.218 | 0.488 (27.8) | 1.756 (100) |
| o | 0.192 | 0.822 (52.5) | 1.565 (100) |
| Avg | − (45.1) | − (100) |
Numbers in parentheses are the relative number of parameters with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage.
Figure 3Comparison of the number of parameters of baseline, 1st solution, and F2GCN (log scale).