Literature DB >> 35783353

Bridging the Gap of AutoGraph Between Academia and Industry: Analyzing AutoGraph Challenge at KDD Cup 2020.

Zhen Xu1, Lanning Wei1,2, Huan Zhao1, Rex Ying3, Quanming Yao4, Wei-Wei Tu1, Isabelle Guyon5,6.   

Abstract

Graph structured data is ubiquitous in daily life and scientific areas and has attracted increasing attention. Graph Neural Networks (GNNs) have been proved to be effective in modeling graph structured data and many variants of GNN architectures have been proposed. However, much human effort is often needed to tune the architecture depending on different datasets. Researchers naturally adopt Automated Machine Learning on Graph Learning, aiming to reduce human effort and achieve generally top-performing GNNs, but their methods focus more on the architecture search. To understand GNN practitioners' automated solutions, we organized AutoGraph Challenge at KDD Cup 2020, emphasizing automated graph neural networks for node classification. We received top solutions, especially from industrial technology companies like Meituan, Alibaba, and Twitter, which are already open sourced on GitHub. After detailed comparisons with solutions from academia, we quantify the gaps between academia and industry on modeling scope, effectiveness, and efficiency, and show that (1) academic AutoML for Graph solutions focus on GNN architecture search while industrial solutions, especially the winning ones in the KDD Cup, tend to obtain an overall solution (2) with only neural architecture search, academic solutions achieve on average 97.3% accuracy of industrial solutions (3) academic solutions are cheap to obtain with several GPU hours while industrial solutions take a few months' labors. Academic solutions also contain much fewer parameters.
Copyright © 2022 Xu, Wei, Zhao, Ying, Yao, Tu and Guyon.

Entities:  

Keywords:  Automated Machine Learning; Graph Neural Networks; data challenge; graph machine learning; node classification

Year:  2022        PMID: 35783353      PMCID: PMC9245425          DOI: 10.3389/frai.2022.905104

Source DB:  PubMed          Journal:  Front Artif Intell        ISSN: 2624-8212


1. Introduction

Graph structured data has been prominent in our lives and various tasks are studied based upon, including a recommendation on Social Networks (Fan et al., 2019), traffic forecasting on road networks (Li et al., 2018), drug discovery on molecule graphs (Torng and Altman, 2019), and link prediction on the knowledge graph (Zhang et al., 2020). Graph Neural Networks (GNN) (Kipf and Welling, 2017) have been proved to be effective in modeling graph data and tremendous GNN architectures are proposed every year (Hamilton et al., 2017; Veličković et al., 2018; Wu et al., 2019; Xu et al., 2019). When applying GNN on graph structured data, expertise and domain knowledge are often required and numerous human effort is needed to adapt to new datasets. Automated Machine Learning (AutoML) (Yao et al., 2018; Hutter et al., 2019) aims to reduce human efforts in deploying various applications. AutoML, especially Neural Architecture Search (NAS), has been successfully explored in tremendous applications, including Image Classification (Tan and Le, 2019), Object Detection (Tan et al., 2020), Semantic Segmentation (Nekrasov et al., 2019), Language Modeling (Jiang et al., 2019), and Time Series Forecasting (Chen et al., 2021). As a result, researchers have started to explore Automated Graph Neural Networks (AutoGraph). AutoGraph researchers focus mainly on automatically designing GNN architectures by NAS. The majority of these methods focus on designing the aggregation functions/layers in GNNs with different search algorithms (Zhou et al., 2019; Gao et al., 2020; Peng et al., 2020; Yoon et al., 2020; Li et al., 2021). Other works, SANE (Zhao et al., 2021) and AutoGraph (Li and King, 2020), provide the extra dimension of layer-wise skip connections design; GNAS (Cai et al., 2021), DeepGNAS (Feng et al., 2021), and Policy-GNN (Lai et al., 2020) learn to design the depth of GNNs. DiffMG (Ding et al., 2021) proposed to use NAS to search data-specific meta-graphs in the heterogeneous graph, and PAS (Wei et al., 2021) is proposed to search data-specific pooling architectures for graph classification. The recently proposed F2GNN (Wei et al., 2022) method decouples the design of aggregation operations with architecture topology, which is not considered before. Despite the rich literature from academia, we ask the question of how AutoGraph is used by industrial practitioners. Toward this end, we organized the first AutoGraph challenge at KDD Cup 2020 and collaborated with 4Paradigm, ChaLearn, and Stanford University. This challenge asks participants to provide AutoGraph solutions for the node classification task. The code is executed by the platform on various graph datasets without any human intervention. Through the AutoGraph challenge, we wish to push forward the limit of AutoGraph as well as to understand the gap between industrial solutions and academic ones. In this article, we first introduce the AutoGraph challenge setting. Then, we present the winning solutions which are open sourced for everyone to use. Finally, we experiment further and compare with NAS for GNN methods and quantify empirically the gap with respect to top solutions. We conclude three gaps of AutoGraph between academia and industry: Modeling scope, Effectiveness, and Efficiency.

2. Challenge Background

2.1. General Statistics

The AutoGraph challenge lasted for 2 months. We received over 2200 submissions and more than 140 teams from both high-tech companies (Ant Financial, Bytedance, Criteo, Meituan Dianping, Twitter, NTT DOCOMO, etc.) and universities (MIT, UCLA, Tsinghua University, Peking University, Nanyang Technological University, National University of Singapore, IIT Kanpur, George Washington University, etc.), coming from various countries. The top three teams are aister, PASA_NJU, and qqerret. The top 10 winners' information is shown in Table 1. The 1st winner aister comes from Meituan Dianping, a company on location-based shopping and retailing service. This makes the challenge particularly valuable since we can compare academic solutions with industrial best AutoGraph practices.
Table 1

General information about winning teams.

Place Team name Institute Place Team name Institute
1st aister Meituan Dianping6th SmartMN-THU Tsinghua University
2nd PASA_NJU Nanjing University6th JunweiSun Beijing University of Postsand Telecommunications
3rd qqerret Ant Financial8th u1234x1234 Self-employed
4th common Alibaba Inc.9th AML Ant Financial
5th PostDawn Zhejiang University10th supergx Nanyang Tech. University
10th daydayup Hikvision Inc.

Two teams tie in the 6th and 10th place. We list them both.

General information about winning teams. Two teams tie in the 6th and 10th place. We list them both.

2.2. Problem Formulation

The task of the AutoGraph challenge is node classification under the transductive setting. Formally speaking, consider a graph , where is the set of nodes, i.e., and is the set of edges, which is usually encoded by an adjacency matrix A∈[0, 1]. A is positive if there is an edge connecting from node v to node v. Additionally, a matrix X∈ℝ represents the features of each node. Each node v has a class label , resulting in the label vector . In the transductive semi-supervised node classification task, part of the labels are available during training and the goal is to learn a mapping and predict classes of unlabeled nodes.

2.3. Protocol

The protocol of the AutoGraph challenge is straightforward. Participants should submit a python file containing a Model class with the required fit and predict methods. We prepare an ingestion program reading the dataset and instantiate the class and call fit and predict method until the prediction finishes or the running time has reached the budget limit. The ingestion program outputs the model's prediction on test data and saves it to shared space. Then, another scoring program reads the prediction and ground truth and outputs evaluation scores. The execution of the program is always on the challenge platform. When developing locally, we provide a script to call model.py file methods directly.

2.4. Metric

We use Accuracy (Acc) and Balanced Accuracy (BalAcc) as evaluation metrics, defined as where Ω is the set of test nodes indexes, y is the ground truth label for node v, ŷ is the predicted label for node v, C is the set of classes, and Recall is the recall score for class i. Accuracy (Acc) is used in the challenge to rank participants and Balanced Accuracy (BalAcc) is applied for additional analyses since it takes into account the imbalanced label distribution of datasets.

2.5. Datasets

A total of 15 graph datasets were used during the competition: Five public datasets were directly downloadable by the participants so they could develop their solutions offline. Five feedback datasets were made available on the platform during the feedback phase to evaluate AutoGraph algorithms on the public leaderboard. Finally, the AutoGraph algorithms were evaluated with 5 private datasets, without human intervention. These datasets are quite diverse in domains, shapes, density, and other graph properties because we expect AutoGraph solutions to have good generalization ability. On the other hand, we intentionally keep the characteristics of 5 feedback datasets and 5 private datasets similar to enable transferability. We summarize dataset statistics in Table 2. The licenses and original sources of these datasets are also provided.
Table 2

Statistics of all datasets.

Dataset Phase Domain #Node #Edge #Feature #Class Avg Deg Directed? Weighted? Skewness
aPublicCitation2.7K5.3K1.4K71.9FF5
bPublicCitation3.3K4.6K3.7K61.4FF3
cPublicSocial10K733K0.6K4173.3FF81
dPublicNews10K2,917K0.3K20291.7TT467
ePublicFinance7.5K7.8K031.0FF111
fFeedbackSales10K194K0.7K1019.4FF18
gFeedbackCitation10K41K8K54.1FF6
hFeedbackMedicine10K2,461K0.3K23246.1TT1,773
iFeedbackFinance15K16K031.1FF213
jFeedbackMedicine11K22K092.0FF227
kPrivateSales8K119K0.7K814.9FF6
lPrivateCitation10K40K7K154FF34
mPrivateNews10K1,425K0.3K8142.5TT360
nPrivateFinance14K22K0101.6FF61
oPrivateSocial12K19K0191.6FF62

“Avg Deg” is the average number of edges per node. “Directed” and “Weighted” indicate the two properties of a graph. “Skewness” here is calculated by the number of nodes in the largest class divided by the number of nodes in the smallest class.

Statistics of all datasets. “Avg Deg” is the average number of edges per node. “Directed” and “Weighted” indicate the two properties of a graph. “Skewness” here is calculated by the number of nodes in the largest class divided by the number of nodes in the smallest class.

3. Solutions

In this part, we introduce various methods suitable for the AutoGraph challenge, including the provided challenge baseline and solutions from top-3 winners.

3.1. Baseline (GCN(L2))

In the provided baseline, there is no feature engineering except for using the raw node features. For graphs without node features (e.g., dataset i and j), one hot encoding is used to unroll the node lists to a dummy feature table. During model training, an MLP is first used to map node features to the same embedding dimension. Then a two layer vanilla GCN is applied for learning node embeddings. Another MLP with softmax outputs the final classification. Dropout is used. All the hyperparameters are fixed by our experience. There is no time management since the model is simple and one full training will not cost more time than the allowed budget.

3.2. First Placed Winner

The 1st winner is from team aister. Their code is open source here. The authors use four GNN models, two spatial ones: GraphSage (Hamilton et al., 2017) and GAT (Veličković et al., 2018), and two spectral ones: GCN (Kipf and Welling, 2017) and TAGConv (Du et al., 2017) to process node features collectively. For each GNN model, a heavy search is applied offline to determine the important hyperparameters as well as the boundaries. In the online stage, they use a smaller search space to determine the hyperparameters. In order to accelerate the search, they do not fully train each configuration but instead early stop at 16 epochs if the validation loss is not satisfactory. Additional features are used: node degrees, distribution of 1-hop and 2-hop neighbor nodes' features, etc.

3.3. Second Place Winner

The 2nd winner is from team PASA_NJU. Their code is open source here. They also split the solution into two stages: offline stage and online stage. In the offline stage, the authors train a decision tree based on public data and other self collected datasets to classify graph types into one of three classes. Then they use GraphNAS (Gao et al., 2020) to search massively optimal GNN architectures including aggregation function, activation, number of heads in attention, and hidden units. In the online stage, the authors rapidly classify the dataset and fine tune the offline searched model.

3.4. Third Place Winner

The 3rd winner is from team qqerret. Their code is open source here. The core model is a variant of spatial based GNN, which aggregates 2-hop neighbors of a node with additional linear parts for the node itself. Basically, the new embedding of node i is . Additionally, in the GNN output layer, a few features per node are concatenated for the final fully connected layer, including the number of edges, whether this node connects to a central node that has a lot of edges, the label distribution of 1-hop neighbor nodes, and the label distribution of 2-hop neighboring nodes.

4. Results

We conduct additional experiments after the AutoGraph challenge to further analyze the results. We first reproduce winning solutions and then we compare them with academic solutions. Three gaps are concluded. The first gap is presented as follows and two other gaps are concluded in section 4.2. Gap #1: Modeling Scope Is the First Gap of AutoGraph Between Academia and Industry. In academia, researchers focus mainly on Neural Architecture Search methods to find better GNN architectures. Their contributions differ in their search space, search strategy, and evaluation methods. However, industrial solutions, e.g., 1st solution, focus more on feature engineering and model ensemble. For GNN architectures, they prefer the existing ones with little modification. In other words, industrial people provide a full pipeline solution including data preprocessing, feature engineering, model architecture, hyperparameter optimization, and model ensemble, while academic researchers focus on the model architecture part only. The gap is also illustrated in Figure 1. It might be an interesting direction for both groups to merge, i.e., AutoGraph researchers could explore the automated feature engineering and automated ensemble, and AutoGraph practitioners could adopt NAS methods for GNN.
Figure 1

Illustration of AutoGraph scope. Industrial people provide a full pipeline solution that covers data preprocessing to evaluation. Academic researchers focus mainly on model architecture and hyperparameter optimization.

Illustration of AutoGraph scope. Industrial people provide a full pipeline solution that covers data preprocessing to evaluation. Academic researchers focus mainly on model architecture and hyperparameter optimization.

4.1. Reproducing Winning Solutions

We reproduce all winning methods on all the datasets and include their results in Table 3. We observe that all three winning solutions are close in performance and all significantly beat the GCN baseline. On the other hand, in the AutoGraph challenge, due to the nature of the competition, we rank methods based on their accuracy. We state that this is not sufficient to evaluate solutions comprehensively from the scientific perspective. We add the balanced accuracy here just to show that for some methods that show close performance in accuracy, they could diverge a lot in balanced accuracy. Regarding both accuracy and balanced accuracy, we conclude that 1st solution, which comes from Meituan Dianping Company, is indeed the best among the top winners. Thus, we will later use their solutions for comparison with academic solutions. These winning solutions are already open sourced, which are reproducible and lower the barriers to using AutoGraph.
Table 3

Accuracy and Balanced accuracy of top methods on all datasets (%).

Dataset Phase Baseline (GCN(L2)) 1st solution 2nd solution 3rd solution
Acc BalAcc Acc BalAcc Acc BalAcc Acc BalAcc
aPublic85.784.9 88.5 87.8 88.287.287.285.5
bPublic71.467.875.2 71.2 75.8 71.2 75.669.0
cPublic86.572.094.387.594.290.9 95.4 91.3
dPublic93.76.1 96.5 48.7 95.128.894.621.0
ePublic59.638.888.7 92.8 88.590.7 88.8 92.8
fFeedback86.678.2 92.8 92.1 92.3 92.1 92.491.4
gFeedback94.792.895.393.595.693.8 95.8 94.2
hFeedback90.48.8 93.5 26.3 92.217.692.116.6
iFeedback88.259.288.487.588.4 92.6 88.5 91.1
jFeedback90.768.195.989.096.1 93.7 96.6 93.3
kPrivate93.592.2 95.5 94.4 95.5 94.4 94.893.1
lPrivate90.984.5 94.9 92.6 94.791.894.5 92.6
mPrivate85.524.5 98.1 79.7 95.769.098.079.4
nPrivate85.647.3 99.0 97.3 99.0 98.4 98.997.0
oPrivate49.615.691.084.691.3 90.6 91.4 88.5

The baseline is a two layer GCN. Bold values are best in comparison with other methods.

Accuracy and Balanced accuracy of top methods on all datasets (%). The baseline is a two layer GCN. Bold values are best in comparison with other methods.

4.2. Neural Architecture Search for GNN

We further adopt NAS methods for GNN and compare with the baseline and 1st solution coming from the industry. We choose the recent F2GNN (Wei et al., 2022) in our experiment, which searches for data-specific GNN topology. To compare fairly with GCN baselines, we fix the aggregation to GCN and search only the GNN topology, which we call F2GCN. Since F2GCN requires at least 4 layers, we also run a 4 layer GCN baseline for better comparison. The results are given in Table 4.
Table 4

Accuracy comparison of GCN baselines, F2GCN, and industrial best solution (%).

Dataset GCN(L2) GCN(L4) F2GCN(L4) 1st solution
a85.784.484.4 (95.4)88.5 (100)
b71.470.571.3 (94.8)75.2 (100)
c86.582.392.8 (98.4)94.3 (100)
d93.793.693.9 (97.3)96.5 (100)
e59.687.588.4 (99.7)88.7 (100)
f86.687.692.1 (99.2)92.8 (100)
g94.793.495.3 (100)95.3 (100)
h90.490.390.1 (96.4)93.5 (100)
i88.287.688.3 (99.9)88.4 (100)
j90.783.695.3 (99.4)95.9 (100)
k93.593.293.4 (97.9)95.5 (100)
l90.989.192.9 (97.9)94.9 (100)
m85.586.186.1 (87.8)98.1 (100)
n85.695.296.7 (97.7)99.0 (100)
o49.671.888.8 (97.6)91.0 (100)
Avg− (97.3)− (100)

L2, L4 mean 2, and 4 layers for the GNN architecture. Numbers in parentheses are the relative accuracy with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage.

Accuracy comparison of GCN baselines, F2GCN, and industrial best solution (%). L2, L4 mean 2, and 4 layers for the GNN architecture. Numbers in parentheses are the relative accuracy with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage. Gap #2: Effectiveness Is the Second Gap of AutoGraph Between Academia and Industry. We observe from Figure 2 and Table 4 that all baselines and F2GCN methods are not as good as 1st winning solution. However, for many datasets, e.g., e, f, g, i, j, F2GCN is very close to the best industrial solution. On average, F2GCN which focuses only on architecture search reaches 97.3% of the best solution. Note that the 1st solution constructs additional node features and uses multiple GNN architectures for ensemble while F2GCN does not use any feature engineering or model ensemble. This shows the effectiveness of the winner's engineering practices and F2GCN's adaptive architecture search. The winning teams also have access to public datasets and the public leaderboard to iteratively fine tune their methods. F2GCN does not assume any prior knowledge of the datasets, which shows further its effectiveness.
Figure 2

Accuracy improvement with respect to baseline.

Accuracy improvement with respect to baseline. To better understand the solutions, we calculate the number of parameters of the baseline, F2GCN, and the 1st solution, as shown in Table 5.
Table 5

Number of parameters of baseline, 1st solution and F2GCN (Unit: Millions).

Dataset GCN(L2) F2GCN(L4) 1st solution
a0.0230.908 (75.7)1.199 (100)
b0.0590.700 (44.2)1.583 (100)
c0.0111.598 (98.0)1.631 (100)
d0.0060.042 (3.20)1.296 (100)
e0.1210.354 (31.8)1.114 (100)
f0.0130.039 (2.30)1.688 (100)
g0.1340.313 (13.1)2.389 (100)
h0.0060.271 (20.9)1.294 (100)
i0.2412.269 (113.0)2.013 (100)
j0.1710.834 (60.6)1.376 (100)
k0.0121.478 (108.0)1.395 (100)
l0.1080.614 (25.6)2.395 (100)
m0.0050.010 (0.80)1.278 (100)
n0.2180.488 (27.8)1.756 (100)
o0.1920.822 (52.5)1.565 (100)
Avg− (45.1)− (100)

Numbers in parentheses are the relative number of parameters with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage.

Number of parameters of baseline, 1st solution and F2GCN (Unit: Millions). Numbers in parentheses are the relative number of parameters with respect to 1st solution. We regard 1st solution as 100%. The last line is the average percentage. Gap #3: Efficiency Is the Third Gap of AutoGraph Between Academia and Industry. From Figure 3 and Table 5, F2GCN uses significantly fewer parameters than the best industrial solution on most datasets (13 out of 15). On average, F2GCN consumes 45.1% of the 1st solution in terms of parameter size, which is quite resource-efficient. Note that feature engineering and ensemble do not contain additional parameters and basically, F2GCN searches one GNN model to compete with the ensemble of 4 types of GNN models in the 1st solution. As for time devotion, the winning solutions come from a team's months of work, which consists of 5 or more members. F2GCN only runs for a few GPU hours per dataset, demonstrating its time efficiency compared to industrial solutions.
Figure 3

Comparison of the number of parameters of baseline, 1st solution, and F2GCN (log scale).

Comparison of the number of parameters of baseline, 1st solution, and F2GCN (log scale).

5. Conclusion

We organized the first Automated Graph Learning (AutoGraph) Challenge at KDD Cup 2020. We presented in this article its settings, dataset, and solutions, which are all open sourced. Furthermore, we ran additional post-challenge experiments to compare the baseline [Graph Convolution Network (GCN)], the winning solution (feature engineering-based ensemble of various Graph Neural Networks), and a recent and efficient Neural Architecture Search (NAS) for the GNN method called F2GCN. This article provides results that could bridge the gap between academic research and industrial practices, by correcting the bias favoring certain approaches. This gap is currently at 3 aspects: Gap #1 modeling scope. (academia focuses more on model-centric approaches, emphasizing NAS; industry emphasizes data centric approaches and feature engineering); Gap #2 effectiveness. (academic solutions are perceived by the industry to be less effective than their own counterpart); Gap #3 efficiency. (academic solutions are perceived to be parsimonious or slower than industrial solutions). Our results indicate that the “academic” NAS-based approach that we applied attains performances closely matching those of the winning industrial solution while being both faster and more parsimonious in the number of parameters, therefore, closing Gap #2 and #3. Moreover, we hope that these results will help reduce Gap #1, by encouraging industrial practitioners to apply NAS methods (and particularly F2GCN), eventually combining the best of both approaches. We believe the results we obtained are significant since they involve a benchmark on 15 datasets.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://github.com/AutoML-Research/AutoGraph-KDDCup2020.

Author Contributions

ZX, LW, and HZ contributed to post-challenge experiments and the first draft of the manuscript. RY, WWT, and IG contributed to further discussions. All authors contributed to the application, organization of the AutoGraph challenge at KDD Cup 2020, contributed to manuscript revision, read, and approved the submitted version.

Funding

Funding and support have been received by several research grants, including 4Paradigm, Big Data Chair of Excellence FDS Paris-Saclay, Paris Région Ile-de-France, and ANR Chair of Artificial Intelligence HUMANIA ANR-19-CHIA-0022, ChaLearn, Microsoft, Google. Microsoft and Google provide cloud computing resources for hosting the challenge and post-challenge experiments.

Conflict of Interest

ZX, LW, HZ, and WWT are employed by 4Paradigm, China. The funder 4Paradigm had the following involvement in the study: prepared the challenge, post-challenge analysis and paper writing. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
  1 in total

1.  Graph Convolutional Neural Networks for Predicting Drug-Target Interactions.

Authors:  Wen Torng; Russ B Altman
Journal:  J Chem Inf Model       Date:  2019-10-16       Impact factor: 4.956

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.