Literature DB >> 33237964

A thorough analysis of the contribution of experimental, derived and sequence-based predicted protein-protein interactions for functional annotation of proteins.

Stavros Makrodimitris1,2, Marcel Reinders1,3, Roeland van Ham1,2.   

Abstract

Physical interaction between two proteins is strong evidence that the proteins are involved in the same biological process, making Protein-Protein Interaction (PPI) networks a valuable data resource for predicting the cellular functions of proteins. However, PPI networks are largely incomplete for non-model species. Here, we tested to what extent these incomplete networks are still useful for genome-wide function prediction. We used two network-based classifiers to predict Biological Process Gene Ontology terms from protein interaction data in four species: Saccharomyces cerevisiae, Escherichia coli, Arabidopsis thaliana and Solanum lycopersicum (tomato). The classifiers had reasonable performance in the well-studied yeast, but performed poorly in the other species. We showed that this poor performance can be considerably improved by adding edges predicted from various data sources, such as text mining, and that associations from the STRING database are more useful than interactions predicted by a neural network from sequence-based features.

Entities:  

Mesh:

Substances:

Year:  2020        PMID: 33237964      PMCID: PMC7688180          DOI: 10.1371/journal.pone.0242723

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

One of the main challenges of the postgenomic era is how to extract functional information from the vast amount of sequence data that are available. As the number of known protein sequences grows at a very fast pace (currently >185 million in UniProtKB), experimentally determining the functions of all proteins has become practically infeasible. This creates the need for accurate Automatic Function Prediction (AFP) methods, which can predict a protein’s function(s) using the knowledge that has been accumulated in the past. To this end, the Gene Ontology (GO) is a very valuable resource that provides a systematic representation of function in the form of three ontologies: Biological Process (BP), Molecular Function (MF) and Cell Component (CC) [1]. The Critical Assessment of Functional Annotation (CAFA) is a community-driven benchmark study that compares a large number of available AFP methods in an independent and systematic way [2-4]. One of the main conclusions that one can draw from the several editions of CAFA is that top-performing methods tend to use a combination of different data sources and not only the amino acid sequence. For example, MS-kNN, one of the best methods in CAFA2, combined sequence similarity with human gene co-expression and protein-protein interaction (PPI) data [5]. GOLabeler, which was the best in CAFA3, combined six different data sources with a powerful algorithm that predicts how suitable a GO term is for the input protein [6]. More recently, the authors of GOLabeler introduced an extension named NetGO which also uses PPI networks as an extra data source, reporting even better performance than GOLabeler on the CAFA3 dataset [7]. These observations show that PPI networks are informative data sources for AFP, which can be understood, since if two proteins physically interact, they are likely to be involved in the same biological process or pathway. However, almost all PPI networks are incomplete. The best-characterized model species, Saccharomyces cerevisiae (baker’s yeast), has one of the densest PPI networks, with 116,209 experimentally-derived, physical interactions in the BIOGRID database [8]. Given the fact that S. cerevisiae has about 6,000 protein-coding genes [9], this means that roughly 0.6% of all possible pairs of proteins are known to interact. The human interactome is also quite well characterized, with 424,074 experimental interactions in BIOGRID (about 0.2% of all possible interactions). Moreover, a recent study identified an additional 52,569 high-quality interactions of 8,275 human proteins [10]. On the other hand, in Arabidopsis thaliana, the most well-studied plant species, there are about 27,000 protein coding genes and 48,786 experimentally-derived physical interactions in BIOGRID, i.e. only 0.01% of the possible interactions are known. This is not likely due to protein interactions being less common in A. thaliana, but rather because it is not as well-studied as yeast. The number of known edges is orders of magnitude smaller in other plant species, even in important crops. For example, in tomato (Solanum lycopersicum), there are only 107 interactions in BIOGRID as of June 2019 (<<0.01% of the total number of possible interactions). In rice (Oryza sativa japonica), there are 330 and in corn (Zea mays) 13. This phenomenon is not restricted to plants, but is also true for non-model animal species, such as economically important species like cow (Bos taurus, 529) and pig (Sus scrofa, 88 interactions). Most methods that employ PPI networks in AFP predict functions by propagating the GO annotations through the network [5, 7]. The simplest of such methods transfers the annotations of a protein to its immediate neighbors. This is also known as Guilt-By-Association (GBA). Fig 1a illustrates the GBA method in an example network with 6 proteins: Proteins 1 and 2 are annotated with a GO term, while protein 6 is not. We are asked to predict whether proteins 3-5 should be annotated with that GO term. As seen in Fig 1a, for all three of these proteins we are at least 66.6% certain that they should be assigned that GO term. Fig 1b shows the same example network, assuming that some of its edges are missing. In this case, protein 5 has no known interacting partners, so it is impossible to determine its function. Similarly, protein 1 has a known function, but is disconnected from the rest of the network, so its function cannot be propagated to other proteins. This example shows that when interactions in a PPI network are missing, function prediction cannot benefit from PPI information (as most proteins will have few or no connections to other proteins).
Fig 1

Toy PPI network with 6 nodes.

Nodes annotated with a GO term are shown in blue and nodes not annotated in red. Unlabeled (test) nodes are shown in white. In (a) the entire network is known and the posterior probabilities for each unlabeled node can be calculated accurately. In (b) some of the edges are missing (signified by the dashed lines), making the calculation of posterior probabilities either erroneous or even impossible (e.g. node 5).

Toy PPI network with 6 nodes.

Nodes annotated with a GO term are shown in blue and nodes not annotated in red. Unlabeled (test) nodes are shown in white. In (a) the entire network is known and the posterior probabilities for each unlabeled node can be calculated accurately. In (b) some of the edges are missing (signified by the dashed lines), making the calculation of posterior probabilities either erroneous or even impossible (e.g. node 5). A way to counter the lack of edges is to predict them using other data sources. The STRING database contains a large collection of protein associations predicted using different sources, such as gene co-expression and text mining [11]. Moreover, the recent rise in popularity of deep learning has caused an increase in methods that attempt to predict protein-protein interactions purely from protein sequence. One of the first examples was from Sun et al. [12], followed by DPPI [13], PIPR [14] and the work of Richoux et al. [15]. The advantage of predicting edges from sequence is that it is—at least in theory—not biased towards previous experiments. In contrast to, for example, predictions within the STRING database that still require other people to have previously studied a specific protein or its orthologues. Having an accurate sequence-based predictor of PPIs means that for all possible pairs of proteins we can obtain a score for how probable an interaction between each pair of proteins is. This would enable us to find possible interacting partners for proteins that have not been previously studied at all. In this study, we are interested in quantifying the influence of missing edges in a PPI network on protein function prediction. Moreover, we are interested in how well (deep learning based) sequence-based PPI predictors can recuperate this missing information, and how that translates in improvements of the function prediction. We hypothesize that using such a model to predict interactions would be more effective than STRING in the downstream task of network-based protein function prediction.

Materials and methods

Protein-protein interaction networks

We compared PPI networks in S. cerevisiae, Escherichia coli, A. thaliana and S. lycopersicum using three types of PPIs: 1) Physical interactions that have been experimentally derived. 2) Predicted interactions based on non-experimental protein association data from the STRING database, and 3) Sequence-based predicted interactions based on the amino acid sequence of two proteins using PIPR.

Physical interactions

For the experimental interactions we used the BIOGRID (version 3.5.171) [8] and STRING databases [11]. We only used physical interactions and ignored the genetic interactions. Of note, the STRING database contains a collection of experimental protein-protein interactions from different databases, including BIOGRID (marked with the “experiments” data source code) and we found edges in BIOGRID that were not present in STRING. From STRING, we only chose experimental protein-protein interactions with association scores larger than the median score over the non-zero scores for each species individually. The node degree distributions of these networks are shown in S1 Fig in S1 File.

Predicted interactions

Besides the experimental evidence, STRING contains protein associations from 12 data sources in total: “neighborhood”, “neighborhood transferred”, “co-occurrence”, “database”, “database transferred”, “experiments transferred”, “fusion”, “homology”, “co-expression”, “co-expression transferred”, “text mining” and “text mining transferred”. We use these data as features predictive of two proteins interacting and/or being functionally associated to add edges to the experimental network. We refer to these edges as “predicted edges”. S1 Table in S1 File shows the number of interactions per species and per data type. In each species, we ignored data sources that did not add any new edges. We also removed “database”, as it includes protein associations that were identified by using the GO annotations of proteins and these edges would cause circular reasoning if used to predict GO terms, leading to a biased evaluation. This left us with 9 data sources from which we could infer PPIs in yeast, E. coli and A. thaliana and 8 in tomato (S1 Table in S1 File). The interaction scores have different distributions in different data sources. Therefore, instead of applying a fixed threshold, we selected the protein pairs with the 50% highest non-zero scores for each data source and species individually. Next to individually using the data sources as proxies for the protein-protein interactions, we also combined data sources. This was done by first integrating the STRING scores from different sources as described in [16] (see S1 File for more information) and then keeping the 50% top non-zero scores for every combination, as before. To combine a binary STRING network with the experimental one, we applied an element-wise logical OR to the corresponding adjacency matrices, so an interaction is added to the combined network if it is present in at least one of the original networks. We also examined the possibility of using all STRING edges by creating weighted graphs whose edge weights correspond to the STRING interaction scores. We then added these weighted graphs to the binary experimental network.

Sequence-based predicted interactions

We used PIPR [14] to predict PPIs from protein sequence. It uses a Siamese twin architecture with both convolutional and recurrent units and three fully connected layers at the end. PIPR also makes use of predefined amino acid embeddings, obtained from both chemical properties of amino acids and their co-occurence in protein sequences. PIPR had an accuracy of about 97% in predicting yeast PPIs when trained on a large, balanced dataset from the DIP database. After having trained the model, we feed it all pairs of proteins. For each pair we get a score in the range [0, 1] denoting the probability that these two proteins interact. We add an edge to our predicted PPI network if the score for that edge is greater than or equal to 0.5.

GO annotations

We obtained GO annotations from the GOA website [17] and only used the experimental annotations and curated annotations (evidence codes “EXP”, “IDA”, “IPI”, “IMP”, “IGI”, “IEP”, “IBA”, “IBD”, “IKR”, “IRD” and “TAS”). We used the entire GO graph (not the smaller GO slim versions). Annotations were propagated towards the ontology root, so that when a protein is annotated with a term, it is also annotated with all its ancestors in the GO graph. We focused on the Biological Process Ontology (BPO), as it is the most difficult ontology to predict [3] and also is the most commonly used in further analyses such as gene set enrichment. Table 1 gives an overview of the different dataset sizes for the four species.
Table 1

Number of proteins and known PPIs per species in BIOGRID. (version 3.5.171).

YeastE. coliArabidopsisTomato
approximate #protein-coding genes6,000 [9]4,400 [18]27,029 [19]34,727 [20]
#proteins with BPO annotations (N)4,9972,86910,648651
#BIOGRID edges between proteins with BPO annotations149,65917,54023,37157
#pairs of proteins with BPO annotations (N(N − 1)/2)12,482,5064,114,14656,684,628211,575
% annotated protein pairs interacting1.200.430.040.03
% disconnected proteins0.423.143.496.9

Function prediction methods

We represent the protein-protein interactions as a network with the proteins as nodes and the interactions as binary, undirected edges. Using this network, we can make predictions about the functions of unannotated proteins using the proteins with known function. To do so, we used a simple Guilt-By Assosciation (GBA) method and a more complicated one that uses node embeddings learned using node2vec [21]. We compared these methods to the BLAST and naive baselines, which are commonly used in the CAFA challenges [2, 3]. Each method computes the probability P(p, t) that a GO term t should annotate protein p. Below we provide details about how each method makes this computation. When P(p, t) is undefined, e.g. because a protein has no neighbors in a PPI network or no significant BLAST hits, we set it to zero to indicate that this term cannot be assigned to this protein.

Guilt-By-Association (GBA)

This method assigns a GO term to a protein with posterior probability equal to the fraction of the protein’s interacting partners annotated with that term. More formally, let A be the network’s adjacency matrix, V a set of training proteins and V a set of test proteins. Moreover, let T(p) be the set of GO terms assigned to p ∈ V. For a protein p ∈ V, we define its neighborhood N(p) as all its interacting partners that are in the training set: For a GO term t, the probability it is assigned to test protein p is given by Eq 2: Where I(x) = 1 iff x is a true statement and |S| denotes the number of elements in set S. For weighted graphs, Eq 2 was adapted so that each neighbor transfers its annotations with a weight equal to the edge weight and we divide by the total sum of the weights instead of the number of neighbors.

node2vec

The node2vec algorithm learns a fixed-length embedding for every node, such that the similarity in the embedding space reflects the similarity of neighborhoods in the graph, as defined by random walks [21]. We used these embeddings as feature vectors on which we applied standard machine learning methods; specifically the k-Nearest Neighbors (kNN) and the ridge classifiers. For kNN, we look for the k training proteins with the most similar feature vectors to a query protein p and set P(p, t) equal to the fraction of these k proteins annotated with t. The ridge classifier models protein function prediction as a multi-output regression problem and learns a linear mapping from the feature space to the label space. We use to denote the node2vec feature matrix, where each row contains the feature vector of one protein, and Y ∈ {−1, 1} to denote the label matrix, where each row represents the GO annotations of each protein and a value of 1 in the matrix denotes that the corresponding protein is annotated with the corresponding GO term. The ridge classifier tries to find a linear mapping , such that Y ≈ XW. We also add L2 regularization to the model with coefficient λ which leads to the optimal solution W* = (X X + λI)−1 X Y. To bring the predictions (XW*) in the range [0, 1], we apply a sigmoid function s(a) = (1 + e−)−1 to each predicted value a. We did not post-process the predictions of the ridge method so it is possible that it makes predictions that are inconsistent with the GO hierarchy.

Naive

The naive method of CAFA [2] assigns a GO term to a protein with probability equal to the fraction of training proteins annotated with that term (Eq 3). This means that all test proteins get the same annotation using this method (making it a quite weak baseline).

BLAST

We ran BLAST with default settings and set P(p, t) equal to the maximum sequence identity between p and its hits annotated with t.

Combining two classifiers

Given the posterior probabilities of two classifiers P1(p, t) and P2(p, t) we combined them using Eq 4, which gives a high score for a protein-term pair if at least one of the two methods gives a high score.

Experimental set-up

Evaluation metrics

To compare function prediction across the differently constructed protein-protein interaction networks, we applied a 5-fold cross-validation. As evaluation metrics we used the protein-centric Fmax and Smin that are extensively used in the CAFA challenges. Definitions for these metrics are provided S1 File. We also measured the coverage of each algorithm, defined as the fraction of test proteins for which at least one term has a non-zero posterior probability. As the GO term distributions and frequencies are different in each species, directly comparing the performances across species is not trivial. To counter the effect of GO term frequencies, we use the concept of Prediction Advantage (PA) [22], which is defined as the improvement on the classification loss of a classifier c (L) with respect to the naive classifier (L). The PA, which is defined in Eq 5, can be calculated for any classification loss, so here we used L = 1 − F. In each fold, we discarded the GO terms that had no positive examples in either the training or the test set.

Experimental PPI (EXP)

We started from the experimental PPI network of a given species. This network includes as nodes all proteins that have at least 1 functional annotation, even if they have no interacting partners. Proteins without functional annotations were removed, even if they had known interactions. node2vec is an unsupervised feature extraction step that only depends on the network and not the functional annotations. We additionally tested whether also including the unannotated proteins as nodes in the network would possibly lead to better features in the first step of the node2vec procedure, as it leads to a better neighborhood estimation. To this end, we ran node2vec on the entire EXP network (including unannotated proteins) and then used the extracted (unsupervised) features of the annotated proteins only in the supervised phase. We repeated this experiment for all four species and compared the performance with that of the original node2vec which learned the (unsupervised) features on a network of only annotated proteins.

Combined experimental and predicted PPI (EXP+STRING)

We added predicted edges to the experimental network from the different data sources in STRING. We evaluated all possible combinations of the 9 STRING data sources (8 for tomato): First, we added each data source individually. Then, we tested all combinations of 2 data sources (36 possibilities), all combinations of 3 (84 possibilities) and so on, until we have included all 9 data sources. So, in total, we tested combinations of data sources (255 for tomato) along with the experimental network.

Sequence-based predicted PPI (EXP+SEQ)

We used edges predicted by PIPR for predicting function. We tested the performance of a network with the experimental edges combined with the PIPR predictions.

Optimization of node2vec classification

node2vec has hyperparameters that can have a large influence on the learned features. We tuned these hyperparameters on the experimental PPI network of each species, by splitting the training set of each cross-validation fold into a new training (80% of initial training set) and a validation set (20% of intial training set). For each hyperparameter combination, we generated node features which we fed to the kNN and ridge classifiers for different values of their parameters (k and λ respectively). Finally, for each cross-validation fold, we identified the combination of hyperparameters, classifier and classifier parameter that maximized the Fmax, trained it on the whole training set and used the trained model to make predictions on the test set. Details about the hyperparameters that were tuned and the values considered are provided in S1 File. When running node2vec on all proteins with known interactions (and not only the ones with functional annotations), we again used 5-fold cross-validation as before. The training, validation and test splits in each fold were kept identical. We also repeated the hyperparameter optimization step, as changes in the network topology might call for different hyperparameter values.

Results

Only the yeast experimental PPI network has acceptable function prediction performance

Fig 2a–2d compare the Fmax achieved by the GBA method on the EXP network to the baseline performances in four species using 5-fold cross-validation. In yeast, this simple approach significantly outperforms both naive (p-value < 10−5, paired t-test, FDR-corrected) and BLAST (p-value = 0.5 ⋅ 10−3, paired t-test, FDR-corrected). In E. coli, A. thaliana and tomato, the picture is quite the opposite, with even the naive method largely outperforming GBA (p-values = 0.026, 0.3 ⋅ 10−5, 0.2 ⋅ 10−3 respectively, paired t-test, FDR-corrected). In tomato, the network is so sparse and disconnected that the maximum F1 score is achieved by assigning all GO terms to all proteins. The Prediction Advantage (PA, see Methods) between GBA and naive classifier follows a linear trend with respect to the fraction of existing edges. The calculation was based on only four points, but it still lies under the statistical significance threshold of 0.05 (Fig 2e, Pearson’s ρ = 0.98, p-value = 0.016).
Fig 2

Function prediction performance of PPI networks in four species.

(a-d): On the x-axis, are the different PPI networks. The height of the bars denotes the Fmax in each species. The naive and BLAST baselines are shown as a red and a black horizontal line respectively, with dashed lines showing the corresponding standard deviations. EXP, GBA is shown in blue, EXP+STRING, GBA in cyan and EXP+SEQ in green. The improvement of node2vec on EXP and EXP+STRING is shown as an orange bar. Absence of an orange bar denotes that the two algorithms performed equally. The combinations of EXP, GBA and EXP+STRING, GBA with BLAST are shown in gray and yellow respectively. The error bars denote the standard deviation over the 5 cross-validation folds. e) Prediction Advantage (PA) of F. Each species is shown as a blue dot and red line shows the least squares linear fit. PA values calculated by downsampling the original yeast network at different levels of missing edges are shown as green dots. f) The fraction of annotated proteins for which each method can make predictions (y-axis) for each species (x-axis). On top, the number of total proteins is shown. Different methods are shown in the same colors as in a-d. Note that the naive method has a coverage of 100% by design.

Function prediction performance of PPI networks in four species.

(a-d): On the x-axis, are the different PPI networks. The height of the bars denotes the Fmax in each species. The naive and BLAST baselines are shown as a red and a black horizontal line respectively, with dashed lines showing the corresponding standard deviations. EXP, GBA is shown in blue, EXP+STRING, GBA in cyan and EXP+SEQ in green. The improvement of node2vec on EXP and EXP+STRING is shown as an orange bar. Absence of an orange bar denotes that the two algorithms performed equally. The combinations of EXP, GBA and EXP+STRING, GBA with BLAST are shown in gray and yellow respectively. The error bars denote the standard deviation over the 5 cross-validation folds. e) Prediction Advantage (PA) of F. Each species is shown as a blue dot and red line shows the least squares linear fit. PA values calculated by downsampling the original yeast network at different levels of missing edges are shown as green dots. f) The fraction of annotated proteins for which each method can make predictions (y-axis) for each species (x-axis). On top, the number of total proteins is shown. Different methods are shown in the same colors as in a-d. Note that the naive method has a coverage of 100% by design. To better characterize the effect of missing edges, we simulated the phenomenon in yeast by removing edges either uniformly at random or by an approach that makes nodes with the lowest degree more likely to lose their edges first (S1 File). We found that the Fmax is relatively robust to uniform edge removal up to 40-50%, but Smin deteriorates more quickly (S2 Fig in S1 File), meaning that predicting more specific terms suffers even under this simplified missing edges scenario. The coverage also drops very slowly (at least initially), which implies that most edges are removed from “dense” parts of the network so that the remaining edges can partly make up for this loss. In the degree-based sampling strategy, which is more realistic, we observed a much steeper drop for all three metrics. In this case, poorly-studied proteins lose their connections very quickly making it impossible to make predictions for them, as indicated by the steep decline in coverage. As a result, the average performance also reduces very fast. The PA values calculated from the degree-based downsampling did not confirm the linear relationship between PA and fraction of known edges (green dots in Fig 2e).

Combining PPI networks with homology

In many function prediction pipelines, PPI networks are combined with other data sources and used in ensemble algorithms. Experiments with a simple method that fuses the posterior probabilities of BLAST with those of the PPI classifier (Eq 4) showed minimal performance gains (2-6%) with respect to stand-alone BLAST, for all species except for S. cerevisiae (43%, Fig 2). The difference with respect to BLAST was found statistically significant using the paired t-test. However, after correcting for multiple testing using the False Discovery Rate method, the p-values for E. coli, A. thaliana and tomato lie just below the 5% significance threshold (0.0468, 0.0468 and 0.0486 respectively), whereas for yeast the corrected p-value is 1.5 ⋅ 10−5. These results confirm that using experimental PPI networks with many missing edges is not helpful for function prediction.

node2vec results

The GBA method is very simple and therefore unlikely to be able to capture all the functional signal present in complicated biological networks. We therefore tested whether a more complicated classifier based on node2vec could outperform it. In the same cross-validation loop, we used a validation set to tune the hyperparameters of node2vec and used the same unseen test set as before to evaluate the model. The optimal hyperparameter values varied per cross-validation fold and per species. The 1NN classifier was the optimal choice in yeast and tomato, while the ridge with moderate regularization in E. coli and A. thaliana. More importantly, node2vec performed better than GBA on the EXP network in all species except for tomato, where assigning all terms to all proteins still maximizes the Fmax (Fig 2a–2d, S2 Table in S1 File). Evaluation based on Smin gave similar results (S3 Table in S1 File). We also tested whether including proteins with known interactions but no functional annotations during the feature learning step could improve the performance of node2vec. We used the t-test to compare the Fmax, Smin and coverage of these networks to the ones that consist of only annotated proteins. We found that doing so lead to a small but significant increase in coverage in E. coli and A. thaliana (paired t-test, corrected for the FDR), but there was no significant difference in Fmax or Smin in any of the four species (FDR > 0.05, S4 Table in S1 File). This means that although we can make predictions for more proteins the predictions become less accurate when including these edges. Therefore, for the rest of our experiments we only refer to node2vec trained on the proteins that have GO annotations.

Performance per protein

Comparing the performance for each individual protein, we observed a large non-linear dependency between the performance and the number of annotated neighbors. This dependency was consistently smaller for node2vec (a Spearman correlation of 0.30, 0.60 and 0.81 for yeast, E. coli and A. thaliana respectively) than for GBA (0.41, 0.65 and 0.85 for yeast, E. coli and A. thaliana respectively). We also found that node2vec consistently outperforms GBA regardless of the number of annotated (training) neighbors in E. coli and A. thaliana (Wilcoxon rank sum test, FDR < 0.05, Fig 3 and S4 Table in S1 File). In S. cerevisiae, node2vec is significantly better than GBA for 6 out of 9 bins and significantly worse in 1 bin, while for two bins there were no significant differences (Wilcoxon rank sum test, FDR < 0.05, Fig 3 and S5 Table in S1 File). Finally, node2vec can make predictions for proteins that do not have any training neighbors as long as they are not completely disconnected, as its feature vectors are learned in an unsupervised way using the entire network. This means that, for not too sparse networks, node2vec is the preferred option compared to GBA.
Fig 3

Performance per protein.

Fmax achieved per protein (y-axis) as a function of the number of training neighbors in the EXP network (x-axis) for EXP, GBA (blue), EXP, node2vec (orange) and EXP+STRING, GBA (cyan). The median of each group is denoted by a horizontal line and the 5th and 95th percentiles by the whiskers. The number of proteins in each group is shown at the top of each group and an asterisk (*) next to the number signifies that the difference between EXP, GBA and EXP, node2vec is statistically significant at a False Discovery Rate of 5%. For the EXP+STRING network, we show the performance of the combination of data sources that had the best performance in each species.

Performance per protein.

Fmax achieved per protein (y-axis) as a function of the number of training neighbors in the EXP network (x-axis) for EXP, GBA (blue), EXP, node2vec (orange) and EXP+STRING, GBA (cyan). The median of each group is denoted by a horizontal line and the 5th and 95th percentiles by the whiskers. The number of proteins in each group is shown at the top of each group and an asterisk (*) next to the number signifies that the difference between EXP, GBA and EXP, node2vec is statistically significant at a False Discovery Rate of 5%. For the EXP+STRING network, we show the performance of the combination of data sources that had the best performance in each species.

Adding predicted edges is more useful than using a complex classifier

We then tested to what extent predicted interactions from STRING can improve upon the protein function prediction performance of the EXP networks. As we can see in Fig 2a–2d, the GBA classifier performed considerably better on the EXP+STRING network than on EXP for all species. It also significantly outperformed the naive and BLAST baselines. As shown in Fig 3, the STRING edges offer a performance boost for both nodes that have and nodes that do not have annotated neighbors in the experimental network for all species. However, for hub yeast proteins with more than 20 experimental edges, applying node2vec on the EXP network was more effective than adding predicted edges (Fig 3a). The fraction of proteins that can be annotated by the STRING networks approaches 100% for E. coli and A. thaliana and 80% for tomato (Fig 2f). Using a weighted STRING network with all available interactions instead of a binary one lead to small performance improvements, but mainly for the combinations that performed less well (S3–S6 Figs in S1 File). The effect sizes were rather small for the top-performing combinations (S6 Table in S1 File). This shows that STRING edges possibly contain useful functional signal even at confidence levels lower than those we considered here.

Combining STRING edges with homology

Moreover, combining the predictions of the GBA classifier on this network with BLAST predictions (see Methods) leads to significant improvement (28-76%) over BLAST for all species (Fig 2). The combined model gave significant improvements (10-26%) over its PPI component in yeast, E. coli and A. thaliana and performed equally well in tomato (Fig 2). Smin results show similar trends, with the exception that in yeast, the optimal Smin is achieved by GBA on the EXP+STRING network and not by the combination with BLAST (S3 Table in S1 File). These show that adding predicted edges is very beneficial for all tested PPI networks.

node2vec on STRING edges

Similar to the EXP network, we compared the GBA classifier to the one based on node2vec on EXP+STRING. We again observed that the more complex classifier achieved higher Fmax in yeast, E. coli and A. thaliana (Fig 2a–2d), but in terms of Smin only yeast showed an improvement (S3 Table in S1 File). In addition, Fig 2b–2d show that in not so well-studied species, using a more complicated classifier on the EXP network performs considerably worse than a simple classifier on a more complete network with predicted edges.

Effect of individual STRING data sources

We also examined which STRING data sources were responsible for the observed increase in performance. As shown in Fig 4 and S7–S9 Figs in S1 File, the vast majority of data sources when individually added to the EXP network lead to better function prediction in terms of both Fmax and Smin, with the exception of “experiments transferred” in yeast. Fig 4 and S7–S9 Figs in S1 File also show that “text mining” (in S. cerevisiae and A. thaliana), “text mining transferred” (in E. coli and S. lycopersicum) and “homology” (in all four) were by far the most useful sources. A more in-depth analysis of the results showed that these three data sources alone are actually enough to obtain the maximum performance of the GBA method on the EXP+STRING network (S7–S14 Tables in S1 File) and that removing all of them leads a to significant performance drop (S15 and S16 Tables in S1 File). Moreover, including all nine data sources (eight for tomato) lead to worse Fmax and Smin in all species (Fig 4 and S7–S9 Figs in S1 File).
Fig 4

Performance of STRING edges in A. thaliana.

Fmax (left) and Smin (right) (y-axis) as a function of the number of STRING data sources included (x-axis). Each dot corresponds to one combination of data sources added to the experimental network. Combinations that include “text mining” and/or “text mining transferred” are shown in yellow, combinations that include “homology” in black and combinations that include both in black with yellow border. The rest of the combinations are shown in blue. To ease visibility, we added a random number in the range [-0.5, 0.5] to each combination of the same number of sources. Zero data sources corresponds to the EXP network and the orange line shows the average performance for a specific number of data sources. Horizontal lines denote the performance of the naive (black), BLAST (red) and the combination of BLAST with the EXP PPI network (dashed green).

Performance of STRING edges in A. thaliana.

Fmax (left) and Smin (right) (y-axis) as a function of the number of STRING data sources included (x-axis). Each dot corresponds to one combination of data sources added to the experimental network. Combinations that include “text mining” and/or “text mining transferred” are shown in yellow, combinations that include “homology” in black and combinations that include both in black with yellow border. The rest of the combinations are shown in blue. To ease visibility, we added a random number in the range [-0.5, 0.5] to each combination of the same number of sources. Zero data sources corresponds to the EXP network and the orange line shows the average performance for a specific number of data sources. Horizontal lines denote the performance of the naive (black), BLAST (red) and the combination of BLAST with the EXP PPI network (dashed green).

Edges predicted from protein sequences by a neural network are less useful than STRING edges

The PIPR model for predicting protein-protein interactions from sequence was reported to have 97% cross-validation accuracy on a balanced dataset with about 11,200 data points from S. cerevisiae proteins from the DIP database, a result that we also replicated. This model, however, was not able to generalize to predict BIOGRID edges in yeast, as it achieved an accuracy of 0.59 on a balanced dataset. We also measured the model’s recall, i.e. its ability to identify true interacting pairs, and it was comparable to random guessing (0.51). We therefore set out to train PIPR for predicting BIOGRID edges, keeping the architecture and the training procedure the same. As positive training examples, we used all yeast protein pairs reported to be physically interacting in BIOGRID and as negative examples, an equal-sized set of randomly selected protein pairs that are not reported as interacting. This proved to be a more challenging task for PIPR, as the best validation accuracy achieved was 0.77 (S17 Table in S1 File). The sequence-based predicted PPI network combined with the experimental one (EXP+SEQ) hampers the AFP performance in yeast as compared to EXP (Fig 2a). This is probably due to the addition of many false positive edges, as it predicts that more than 41% of all possible protein pairs are interacting, which is about 10 times more than expected [16]. In contrast, in E. coli, A. thaliana and tomato the EXP+SEQ PPI network seems to be more useful, providing significant improvements over EXP (Fig 2b–2d). However, these improvements are not enough to surpass even the BLAST baseline in E. coli and A. thaliana. Contrary to our expectation, the EXP+STRING network performed significantly better than EXP+SEQ for all species (Fig 2a–2d). This was true even when we removed edges from text mining from the STRING networks. In tomato, the EXP+STRING network cannot make predictions for roughly one fifth of the proteins (Fig 2f). Adding the SEQ edges only for these proteins improved the overall Fmax from 0.61 to 0.67. This shows that SEQ edges are useful, but they are surpassed by the higher quality of STRING edges. Finally, we trained PIPR on A. thaliana edges from BIOGRID and obtained new networks in A. thaliana and S. lycopersicum. Although this network worked slightly better in tomato than the one trained in yeast data, it was still worse than BLAST and EXP+STRING (S18 Table in S1 File).

Discussion

The aim of this work was to investigate ways of addressing the problem of missing edges in experimental protein-protein interaction networks for the downstream task of genome-wide function prediction. Our main hypothesis was that a deep learning model that can identify interacting proteins from sequence with very high accuracy would be a good solution to this issue. We demonstrated how the sparsity of experimental PPI networks leads to poor function prediction performance, using the simple GBA classifier. We did not compare this classifier to any state-of-the-art methods, such as GOLabeler [6] or INGA [23], but rather to the naive and BLAST baselines from the CAFA challenges. The naive classifier, as its name suggests, does not use any information to relate specific proteins to GO terms, rather it only uses the frequency of each GO term in the training set. In the machine learning literature, this classifier is also called the “Bayesian Marginal Predictor” [22] and is the optimal classifier when the distributions of the classes (P(y)) are known, but information about the relationship between the data and the classes (p(x|y)) is missing. This means that any classifier that uses any kind of (informative) data is expected to outperform the naive one. However, we clearly demonstrated the failure of the GBA classifier in predicting BPO terms in E. coli, A. thaliana and tomato, as it performed considerably worse than the naive method. This was not the case in yeast, where the GBA classifier outperformed both baselines. When examining the performance for individual proteins, we found a high correlation between the number of known interacting partners and the prediction accuracy. The GBA method has proven to be very useful in function prediction [5], but it is a very simple approach and therefore heavily relies on the correctness of the given network. We thus expected that using a more complicated approach that captures broader network patterns might (partly) overcome the sparsity. Several such node classification methods exist [24]. Recently, Graph Convolutional Networks (GCNs) have been shown to be effective in such tasks [25]. We chose to use node2vec to generate node features, as it has been successfully applied to protein-protein interaction networks [21] and used these features to train standard classifiers for function prediction. Although, we observed a clear improvement in A. thaliana and E. coli with respect to GBA, the performance remained below that of the baselines, meaning that these models can only partly compensate for missing edges. To make matters worse, we did not observe any improvement in the even sparser tomato network. This difference can be explained by the fact that when tuning the node2vec hyperparameters we rely on the performance on a validation set, which in tomato is very small and only includes “easy” proteins, leading to an apparent high performance for a large number of hyperparameter combinations. This makes it hard to select the optimal hyperparameters for node2vec in tomato, but it is still possible that an improvement could be observed if the correct parameters were known. Using the optimal hyperparameters from another species with a more complete network, e.g. A. thaliana, might be an alternative. However, since the topologies of the two networks are vastly different, the optimal hyperparameters for one species are not necessarily good for the other. Taken together, these observations validated our hypothesis that a sparse PPI network is detrimental to genome-wide AFP. It is worth noting that node2vec can make predictions for nodes that have no annotated neighbors, as opposed to GBA, which helps increase the coverage. Nevertheless, including unannotated proteins with known interactions during the node2vec feature learning step did not lead to better function prediction performance. This hints that -apart from the lack of known interactions- the lack of GO annotations for training proteins also has a considerable negative effect on the accuracy of function prediction algorithms. Many methods have been proposed that try to complete a network by predicting edges. Reviews of such methods can be found in [26] for social and in [27] for biomedical networks. More specifically, the computational prediction of protein-protein interactions has been an active research area for many years [28, 29]. Our work is the first to evaluate the contribution of predicted edges in protein function prediction in a species-specific way. We used the STRING database as a proxy for predicting interaction using omics data such as genome features, homology, co-expression and text mining. In sparse experimental PPI networks, the STRING-derived edges contribute a great deal, increasing the performance of the GBA classifier 1.8-fold in E. coli, more than 2.5-fold in A. thaliana and about 30-fold in tomato. They also outperformed the node2vec method on the EXP network. This is because these extra edges connect proteins that were previously disconnected from the rest of the graph, but also because they can discover new functions for already connected proteins, leading to a performance boost regardless of the number of neighbors. Using these edges was enough to significantly outperform the naive and BLAST baselines. In the case of yeast, which has a more “complete” network, the STRING-derived edges also improved the prediction performance, but to a lesser extent. In fact, in yeast, node2vec on the EXP network and GBA on the EXP+STRING network performed similarly on average, with node2vec being more useful for hub proteins that have a complicated neighborhood. As expected, combining a better network (EXP+STRING) with a better classifier (node2vec) lead to even better performance, though this was not observed in the small tomato dataset. To combine the different STRING data sources, we used the simple algorithm described in [16]. This algorithm (also described in S1 File) assumes independence between the data sources and applies a Bayesian framework to join them into a final score for each protein-protein association. Some more advanced methods have been proposed to perform this integration, such as Mashup [30] and deepNF [31]. Both of these approaches, which are conceptually similar to each other and to node2vec, perform a number of random walks separately for each network derived by each data source to estimate the neighborhood similarity of each node to all other nodes. Then, they learn a feature vector for every node (protein) in order to approximate this similarity as closely as possible. The main difference between the two methods is that Mashup learns these vectors using matrix factorization [30], while deepNF using an autoencoder neural network [31]. Both of these methods outperformed the simple integration strategy in yeast and human PPI networks [31], which means that the performance of the EXP+STRING network could be enhanced by using one of these two methods. On the other hand, these methods—and especially deepNF that has many parameters to be learned—are not guaranteed to work well in a small dataset such as the tomato one. Furthermore, as STRING networks have weighted edges, instead of using thresholds to make them binary, it might be more helpful to employ algorithms that classify nodes directly on weighted graphs, such as those described in [32] and [33]. Our small-scale experiments in that direction gave mixed results, so more research is needed on this issue. Notably, text mining of scientific literature and homology were the most informative STRING data sources for all species. Although removing the text mining edges did lead to a decrease in the maximum performance of EXP+STRING networks, we showed that it did not change the main conclusions of this study. Moreover, we found that edges from “text mining transferred”, i.e. associations that have been discovered through text mining in other species and then transferred based on sequence homology, are very useful in E. coli and tomato. Given that we did not consider GO annotations inferred automatically due to sequence similarity, it is likely that text mining indeed captures true functional information that is conserved across species. This perhaps means that text mining is an underrated data source for functional annotation. We hypothesize that since scientific knowledge is mainly disseminated by publishing articles, text mining on these articles compiles all of this information into one resource. This would explain why otherwise very informative resources such as gene co-expression or operons (in bacteria) are individually useful when added to the EXP network, but are rendered redundant in the presence of text mining edges. Although homology is the most commonly used data source for function prediction, from the descriptions of the methods submitted to the CAFA challenges, we know that only a small minority of them make use of text mining [4]. Two of these methods are described in [34, 35]. A more recent study showed that integrating homology-based predictors with neural-network-based text models leads to a significant performance boost [36], so we expect the role of text mining in function prediction research to be expanded in the future. We also applied a sequence-based neural network model (PIPR) for PPI edge prediction. Firstly, we noticed that although PIPR was very accurate in predicting edges in one yeast dataset, it did not immediately generalize to another dataset from the same species, performing very close to random guessing. Richoux et al. have reported that overfitting and information leaks from the validation set are common when training protein-protein interaction predictors [15]. Although a certain protein pair from the test set cannot be present in the training set too, the two individual proteins can be in the training set in other pairs. This can have an effect for hub proteins with many interacting partners, as in an extreme case the network could learn to always predict this protein to interact with any other protein [15]. The result of these findings as well as ours is that caution is required when using these deep models, despite their high accuracy in one dataset. Nevertheless, PPIs predicted from the PIPR model can be useful for the downstream task of network-based function prediction, as they outperformed the naive baseline. However, our hypothesis that such a model could accurately produce the entire or a big part of the interactome of a species leading to very accurate predicted annotations was not validated, as STRING edges proved more useful. Our experiments in tomato showed that for proteins that were disconnected in the EXP+STRING network, adding SEQ edges gave a significant performance increase, while this was not the case for combining the EXP+STRING network with BLAST. This implies that SEQ can be a useful resource for species with very few protein associations known in STRING. Another limitation of our study is that except for the variable degree of unknown PPIs among the tested species, there is also a large variability in the amount of missing experimental annotations, with yeast being the most well-characterized species and tomato by far the least. This means that it is much more likely that a correctly predicted protein-GO term pair is flagged as a false positive in tomato than in yeast, simply because that annotation has not been discovered yet. Moreover, the GO terms have different frequencies in the four species, meaning that is virtually impossible to compare performances across species. For example, yeast contains a lot more specific annotations than e.g. tomato. This is also demonstrated by the large differences in Smin of the naive method, which means that the total information content of the terms present in each species is vastly different. Calculating the Prediction Advantage with respect to the naive method [22] can correct for differences in term frequencies, but the different degree of missing annotations is harder to correct for while only using experimental annotations. This is not a big issue in our analyses because we did not focus on the exact performance values, but rather on how the performances of different networks (i.e. networks with different edge types) compare to each other within a species. Also, we have shown that the same conclusions can be drawn when evaluation is done using the semantic distance [37], which punishes shallow predictions. Although Fmax and Smin are the most widely-used evaluation metrics for function prediction, a recent study has raised concerns about them [38]. The concerns, which were based on artificially generated predicted annotations, mainly have to do with these metrics being overly lenient to false positive predictions. This might not be a big problem, as due to missing annotations most proteins are likely to be under-annotated. The same study showed that both metrics correlate highly with the signal to noise ratio of the predictions [38]. Based on that we argue that our conclusions do not rely on the choice of evaluation measures, but we believe that proper evaluation of function prediction algorithms is a pressing issue that requires further research.

Conclusion

Our work highlights the difficulty of applying PPI networks in AFP for less well-studied species. We show that predicted PPIs can partially compensate for the sparsity of the networks, with STRING-predicted edges to be the most useful, especially text mining and homology, and sequence-based deep learned predictions mostly to be useful when nodes are still not connected when combining experimental and STRING based PPI edges. (PDF) Click here for additional data file. 10 Aug 2020 PONE-D-20-21508 A thorough analysis of the contribution of experimental, derived and sequence-based predicted protein-protein interactions for functional annotation of proteins PLOS ONE Dear Dr. Makrodimitris, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Specifically, the computational experiments to address the extend to which Protein Protein Interaction Networks are useful in predicting Gene Ontology terms in different species are well thought and designed. Moreover, the work is well organized and presented in an intelligible manner. Both expert reviewers provide useful suggestions (see below for their detailed reports) which can strengthen the conclusions derived from this work and/or improve presentation. I would also like to stress that you should make sure that all figures and tables are appropriately cited in the text (main text and/or supplement). For example, Tables S12 and S13 in Supplementary Text S1 are not cited either in the main manuscript or in Text S1. Please submit your revised manuscript by Sep 24 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Vasilis J Promponas Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The paper aims to perform an analysis of how PPI networks can help in the functional annotation of proteins. The authors also want to study the impact of the quality of a PPI (how well known they are). To do so, they first compare naive algorithms to annotate function with guilt-by-association (GBA) and node2vec (n2v) algorithms using PPI. Then they compare GBA on different PPI: they combine the PPI with different STRING networks. Finally they compare those algorithms with a deep-learning method based on sequences. They do those comparisons on four organisms: yeast, with a well-known PPI, and A. thaliana, E. coli and tomato which do not have a well-known PPI. The paper is well written and nice to read. I have one major concern: 1) The way the "EXP" PPI is made: you chose to remove the nodes that were connected but without functional annotations. You say it yourself in the Discussion: this might have negatively influenced the performance. But you don't explain your choice of getting rid of those nodes. You could either let the nodes in the "EXP" PPI, or make another PPI with those nodes to see the difference. You use algorithms that use neighborhood, then if you modify the neighborhood you can't expect them to perform well. Also you consider STRING networks as "predicted interactions", but all of them are not: - "neighborhood" just states if the genes occur repeatedly in close neighborhood in genomes (mostly prokaryotic). So this one is not a prediction, and might be useful only for E.coli. - "co-occurence" shows the occurence of two genes across species, here again mostly prokaryotic species, and no prediction. - "homology" is a score that is not used "as is" in STRING scoring schemes, it is likely a BLAST used in other channels. - "text-mining" can not predict anything because it just extract information from published articles. You do prediction by using this data as input to one of the algorithm you use, but STRING networks are not predicted. I have several other comments: 2) You have a typo in the abstract: "Here, we tested to what extened" -> "to what extent" 3) You cite STRING v10, but on your github you say you used STRING v11. 4) You chose to select only the 50% highest non-zero scores, have you tried other percent? Why 50%? 5) In Table S2, you might have swap the "EXP, GBA" and "EXP, node2vec" tomato values, because on the text and on Figure 2, you say and we can see that GBA performed better, but the values show otherwise. 6) Paragraph "Combining STRING edges with homology" from "Results" Section: You make a reference of Fig2 after speaking about Smin, but Fig2 shows only Fmax values. 7) Paragraph "Effect of individual STRING data sources" from "Results" Section: You make a reference of Fig S5-8, they do not exist, Fig S2-4 do. Reviewer #2: The authors have performed a very thorough evaluation of impact of PPI network sparseness on the performance of various GO term inference in PPI networks of 4 organisms. They have evaluated both the impact of the addition of edges from different sources to the network (both experimental as well as computationally inferred), and the usage of different strategies for inferring GO annotations from these networks (Guilt by association, sequence similarity, …). The analysis has been carefully performed, and represents a considerable amount of work to evaluate the impact on the prediction performance. I would have some questions and minor comments, as well as a suggestion for a more extensive evaluation which would in my opinion add an additional layer to this analysis Suggestion for major improvement: As the authors state, the available PPIs represent a tiny proportion of the full set of unknown interactions between proteins. However, this sampling is not an unbiased, random sampling, but is likely influenced by the fact that some proteins have been more studied than others. For example in human, oncogenes/proteins are much more likely to appear as hub proteins than other proteins, only because they have been the focus of more in-depth studies. Hence sparsity in one aspect, but biased sparsity is another important one. Hence, I would suggest to add to the study an analysis to evaluate this effect. More precisely, I would suggest to take a relatively dense network like the yeast PPI, and through sub-sampling, obtain more and more sparse networks and evaluate the effect of this down-sampling on the prediction accuracy. This subsampling could be done either by (1) random, unbiased sub-sampling, or (2) by a procedure that would remove edges with a probability that is inversely proportional to the connectivity of the nodes. Hence, highly connected nodes would be more likely to keep their edges, while less connected proteins would be more likely to loose edges, simulating a situation in which the network contains more hubs. It would be interesting then to follow the decrease in prediction accuracy as more and more edges are removed by either of these 2 procedures. Questions/minor points - How is the GO hierarchy dealt with in this study? As the GO ontology contains a high number of terms, very often 2 proteins might be annotated to different terms, which are however very closely related in the hierarchy. Would they count as mismatches in this case? Did the authors use a simplified version of the GO terms (GO slim)? This should be more carefully explained! - The authors should explain the definition of Fmax and Smin, as the readers might not be familiar with these evaluation metrics. - Related to this, the use of Fmax and Smin has been questioned in a recent paper (Plyusnin et al., PLOS Comp. Biology 2019); could the authors comment on this? I understand that they used Fmax and Smin as these were the metrics used in the CAFA assessment, however I would like to have some comments on the performance of these metrics and the possible biases. - The authors have evaluated the effect of using the node2vec procedure instead of the naïve GBA procedure, which should have the advantage of using a larger neighborhood compared to GBA. They state that node2vec is the preferred method compared to GBA (line 268). However, even if the trend shows an increase in performance in Fig 3, the improvement seems hardly significant. Could the authors quantify the improvement of node2vec compared to GBA in Figure 3? - Typo in line 139. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Carl Herrmann [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 17 Sep 2020 First, we would like to thank the editor and reviewers for their constructive comments. They helped us to improve the manuscript considerably. We provide a point-by-point answer to each of them. We present answers to reviewer’s comments (in blue) in black and report changes to the manuscript in italic. Editor Specifically, the computational experiments to address the extend to which Protein Protein Interaction Networks are useful in predicting Gene Ontology terms in different species are well thought and designed. Moreover, the work is well organized and presented in an intelligible manner. Both expert reviewers provide useful suggestions (see below for their detailed reports) which can strengthen the conclusions derived from this work and/or improve presentation. I would also like to stress that you should make sure that all figures and tables are appropriately cited in the text (main text and/or supplement). For example, Tables S12 and S13 in Supplementary Text S1 are not cited either in the main manuscript or in Text S1. Changes in manuscript: We made sure that all supplementary tables and figures, including those added after revision are appropriately cited in both the main document and S1 Text. Reviewer #1: The paper aims to perform an analysis of how PPI networks can help in the functional annotation of proteins. The authors also want to study the impact of the quality of a PPI (how well known they are). To do so, they first compare naive algorithms to annotate function with guilt-by-association (GBA) and node2vec (n2v) algorithms using PPI. Then they compare GBA on different PPI: they combine the PPI with different STRING networks. Finally they compare those algorithms with a deep-learning method based on sequences. They do those comparisons on four organisms: yeast, with a well-known PPI, and A. thaliana, E. coli and tomato which do not have a well-known PPI. The paper is well written and nice to read. Major 1) The way the "EXP" PPI is made: you chose to remove the nodes that were connected but without functional annotations. You say it yourself in the Discussion: this might have negatively influenced the performance. But you don't explain your choice of getting rid of those nodes. You could either let the nodes in the "EXP" PPI, or make another PPI with those nodes to see the difference. You use algorithms that use neighborhood, then if you modify the neighborhood you can't expect them to perform well. Thanks for this remark and allowing us to elaborate on this point. First, do realize that if nodes do not have a functional annotation, it is not possible to evaluate predictions made for these nodes as there is no ground truth. Hence, all predictions are deemed false positives and, moreover, certain measures, such as the recall and F1 scores, are undefined. Consequently, we should not include these unannotated proteins in the test set. Second, take into account that these unannotated proteins do have some functions but that these have not yet been discovered. Now, let’s consider what happens when including these unannotated proteins in the training set: For the GBA and kNN methods these proteins simply lower the posterior probability of their test neighbors having any function (by transferring a label vector of only zeroes). It is possible that these unannotated proteins will have some of the neighboring functions (but undiscovered up till now). Therefore, we believe that we should not dilute the prediction signal by including them in the training set. A similar argument holds for our linear classifier. For the node2vec method this is indeed different as node2vec is an unsupervised feature extraction step that only depends on the network and not the functional annotations. In the discussion of the previous version of the manuscript we referred to this aspect. Together with the reviewer we indeed suspected that including the unannotated proteins also as nodes in the network would possibly lead to better features in the first step of the node2vec procedure, as it leads to a better neighborhood estimation. However, at the 2nd supervised training phase, these nodes should not be used for the same reasons as mentioned before. We do agree with the reviewer, we could have substantiated more the influence on a possibly improved 1st phase on the node2vec procedure. Therefore, we performed a new experiment where we ran node2vec on the entire experimental (EXP) network (including unannotated proteins) and then used the extracted (unsupervised) features of the annotated proteins only in the 2nd phase using a 5-fold cross-validation step as before. The training, validation and test splits in each fold were kept identical and we also repeated the hyperparameter optimization steps of node2vec as big changes in topology might call for different hyperparameter values. We repeated this experiment for all four species and compared the performance with that of the original node2vec which trained the (unsupervised) features on a network of only annotated proteins. The results are shown in Table R1. We found that the network becomes (indeed) less disconnected when including the unannotated proteins as nodes in the network. Also, we could make predictions for more proteins, leading to an increased coverage. However, the increase was only statistically significant for E. coli and A. thaliana (paired t-test over the five cross-validation folds). More importantly, we did not observe any significant differences in either performance metric. The latter observation hints to that these new predictions are not as accurate as for the rest of network. Together, this highlights that a lack of ground truth annotation data hinders function prediction algorithms. Table R1: Fmax, Smin and coverage of the node2vec method on the EXP PPI network in four species when excluding and including proteins without any functional annotations during the feature learning step. Fmax Smin Coverage Annotated Proteins Only All Proteins Annotated Proteins Only All Proteins Annotated Proteins Only All Proteins S. cerevisiae 0.50 ± 0.012 0.50 ± 0.010 30.05 ± 0.54 30.13 ± 0.32 0.99 ± 0.002 0.99 ± 0.001 E. coli 0.28 ± 0.011 0.29 ± 0.008 20.37 ± 0.41 20.47 ± 0.31 0.77 ± 0.016 0.79 ± 0.017 A. thaliana 0.23 ± 0.008 0.23 ± 0.009 28.69 ± 0.49 28.94 ± 0.49 0.57 ± 0.010 0.58 ± 0.011 S. lycopersicum 0.08 ± 0.007 0.08 ± 0.007 19.37 ± 0.89 19.36 ± 0.89 0.03 ± 0.017 0.03 ± 0.017 Changes in manuscript: We have added a description of this experiment in the Methods section. In the Experimental set-up section (subsection “Optimization of node2vec classification”) we now mention that the optimization was done separately for the networks that include unannotated nodes. In the results (subsection “Only the yeast experimental PPI network has acceptable function prediction performance”) we have added a paragraph describing the results above and saying that including those extra nodes does not lead to performance improvement along with a reference to Table R1 (S4 in S1 Text). We have changed the discussion where we originally stated this hypothesis to: “It is worth noting that node2vec can make predictions for nodes that have no annotated neighbors, as opposed to GBA, which helps to increase the coverage. Nevertheless, including unannotated proteins with known interactions during the node2vec feature learning step did not lead to better function prediction performance. This hints that -apart from the lack of known interactions- the lack of GO annotations for training proteins also has a considerable negative effect on the accuracy of function prediction algorithms.” Also you consider STRING networks as "predicted interactions", but all of them are not: - "neighborhood" just states if the genes occur repeatedly in close neighborhood in genomes (mostly prokaryotic). So this one is not a prediction, and might be useful only for E.coli. - "co-occurence" shows the occurence of two genes across species, here again mostly prokaryotic species, and no prediction. - "homology" is a score that is not used "as is" in STRING scoring schemes, it is likely a BLAST used in other channels. - "text-mining" can not predict anything because it just extract information from published articles. You do prediction by using this data as input to one of the algorithm you use, but STRING networks are not predicted. The reviewer is indeed right, these features are not predicted. We were addressing that these features are interpreted by STRING as that a pair of proteins interacts physically or is involved in the same task. But, given the purpose of our manuscript we do understand that this might be confusing. Changes in manuscript: We changed the wording in subsection “predicted interactions” (from line 97 onwards) to clarify this: “Besides the experimental evidence, STRING contains protein associations from 12 data sources in total: "neighborhood", "neighborhood transferred", "co-occurrence", "database", "database transferred", "experiments transferred", "fusion", "homology", "co-expression", "co-expression transferred", "text mining" and "text mining transferred". We use these data as features predictive of two proteins interacting and/or being functionally associated to add edges to the experimental network. We refer to these edges as "predicted edges".” Minor 2) You have a typo in the abstract: "Here, we tested to what extened" -> "to what extent" Changes in manuscript: We corrected the spelling error. 3) You cite STRING v10, but on your github you say you used STRING v11. Thanks for catching this error! We do indeed use version 11. Changes in manuscript: We now refer to the appropriate publication. 4) You chose to select only the 50% highest non-zero scores, have you tried other percent? Why 50%? The threshold of 50% was chosen arbitrarily without trying other values. This threshold can be seen as a hyperparameter of the algorithm that can be tuned using a validation set. Alternatively, all non-zero edges can be kept and we could put a weight to each edge equal to the certainty of that edge as indicated by STRING. Then each node can transfer its functions to its neighbors with a weight equal to the weight of the interaction, resulting in a weighted GBA. As this is more universal than the thresholding method, we performed new experiments with this weighted GBA, i.e. we repeated the experiment of testing all possible combinations of STRING data sources and replicated those across the four species. In general, the weighted network tended to perform slightly better for most combinations. Figure R1 shows this for E. coli. Figure R1: Fmax (left), Smin (middle) and coverage (right) of the binary STRING networks when performing GBA on the 50% top edges (x-axis) vs the weighted GBA (y-axis) for E. coli. Each dot corresponds to one of the 511 combinations of the 9 STRING data sources added to the binary experimental network. The black dashed line shows the y=x line to ease comparison. However, the effect sizes were rather small for the top-performing combinations (Table R2). In yeast, there was barely any difference and the performances of the best combinations of each approach varied at the fourth significant digit. In E. coli and A. thaliana, there is a consistent improvement in both Fmax and Smin, while the tomato results are mixed. Together, these results, although mixed, hint that STRING edges can indeed contain functional signal at lower certainty scores too. We have now included this additional experiment in the main text to highlight the effect of the threshold. Table R2: Fmax, Smin and coverage of the top-performing combination of the EXP+STRING network when using either a binary network with the 50% most probable edges or a weighted one with all edges. Statistically significant differences (paired t-test, FDR < 0.05) are shown in bold. Fmax Smin Coverage Binary Weighted Binary Weighted Binary Weighted Yeast 0.49 ± 0.005 0.49 ± 0.004 27.43 ± 0.57 27.40 ± 0.61 0.99 ± 0.001 0.99 ± 0.001 E. coli 0.46 ± 0.008 0.48 ± 0.009 17.07 ± 0.22 16.61 ± 0.35 0.99 ± 0.004 0.99 ± 0.003 Arabidopsis 0.48 ± 0.004 0.49 ± 0.005 24.12 ± 0.50 23.54 ± 0.50 0.98 ± 0.003 0.99 ± 0.001 Tomato 0.61 ± 0.045 0.64 ± 0.043 9.08 ± 0.89 9.23 ± 1.00 0.86 ± 0.020 0.91 ± 0.018 Changes in manuscript: We have added a supplemental section in S1 Text that shows the results in the form of figures like Figure R1 for all four species and Table R2 (Figures S3-S6 and Table S6). In the main manuscript’s methods section (subsection “Predicted interactions”) we also mention this additional experiment and the weighted GBA classifier. In the Results (subsection “Adding predicted edges is more useful than using a complex classifier”), we added the following text: “Using a weighted STRING network with all available interactions instead of a binary one lead to small performance improvements, but mainly for the combinations that performed less well. The effect sizes were rather small for the top-performing combinations. This shows that STRING edges possibly contain useful functional signal even at confidence levels lower than those we considered here.” 5) In Table S2, you might have swap the "EXP, GBA" and "EXP, node2vec" tomato values, because on the text and on Figure 2, you say and we can see that GBA performed better, but the values show otherwise. Thanks for spotting this mistake! The performance of both GBA and node2vec in tomato is 0.08. This is essentially the lowest possible value, as it arises from assigning all terms to all proteins and signifies that both methods fail completely in this case. The absence of an orange bar in Figure 2d was caused by the fact that the orange and blue bars are at the same height, which was not made clear. The Smin performance of the two methods was also nearly identical in tomato (19.39±0.91 and 19.37±0.89, Table S3). Changes in manuscript: We have corrected the value in Table S2 in S1 Text and adapted the caption of Figure 2 to explain that the bars are at the same height. 6) Paragraph "Combining STRING edges with homology" from "Results" Section: You make a reference of Fig2 after speaking about Smin, but Fig2 shows only Fmax values. Changes in manuscript: We changed the reference to Table S3 which contains the Smin values. 7) Paragraph "Effect of individual STRING data sources" from "Results" Section: You make a reference of Fig S5-8, they do not exist, Fig S2-4 do. Changes in manuscript: We now refer to the correct figures S2-S4. Reviewer #2 The authors have performed a very thorough evaluation of impact of PPI network sparseness on the performance of various GO term inference in PPI networks of 4 organisms. They have evaluated both the impact of the addition of edges from different sources to the network (both experimental as well as computationally inferred), and the usage of different strategies for inferring GO annotations from these networks (Guilt by association, sequence similarity, …). The analysis has been carefully performed, and represents a considerable amount of work to evaluate the impact on the prediction performance. I would have some questions and minor comments, as well as a suggestion for a more extensive evaluation which would in my opinion add an additional layer to this analysis. Major 1) As the authors state, the available PPIs represent a tiny proportion of the full set of unknown interactions between proteins. However, this sampling is not an unbiased, random sampling, but is likely influenced by the fact that some proteins have been more studied than others. For example in human, oncogenes/proteins are much more likely to appear as hub proteins than other proteins, only because they have been the focus of more in-depth studies. Hence sparsity in one aspect, but biased sparsity is another important one. Hence, I would suggest to add to the study an analysis to evaluate this effect. More precisely, I would suggest to take a relatively dense network like the yeast PPI, and through sub-sampling, obtain more and more sparse networks and evaluate the effect of this down-sampling on the prediction accuracy. This subsampling could be done either by (1) random, unbiased sub-sampling, or (2) by a procedure that would remove edges with a probability that is inversely proportional to the connectivity of the nodes. Hence, highly connected nodes would be more likely to keep their edges, while less connected proteins would be more likely to loose edges, simulating a situation in which the network contains more hubs. It would be interesting then to follow the decrease in prediction accuracy as more and more edges are removed by either of these 2 procedures. Thanks for this interesting insight. We fully agree with the reviewer that besides the sparsity aspect, the PPI networks bear a research-based bias. The reviewer does suggest a very interesting additional experiment to study the influence of this bias, which we were happy to include in the new manuscript. As suggested, we started from the full experimental PPI network of yeast and step-wise removed 10%, 20%, 30%, …, 90% and 99% of its edges. The edges were removed at random, using the two strategies suggested by the reviewer. In the first strategy (“uniform”), the probability of removing an edge was uniform over all present edges. In the second strategy (“degree”), the probability of edge removal was inversely proportional to the smaller degree of the two nodes that the edge connects, so that nodes with the fewest edges are most likely to lose them. This procedure gave us 11 down-sampled networks per sampling strategy on which we ran the GBA classifier. This was repeated 5 times with different random seeds to obtain variability estimates. We found that the Fmax is relatively robust to uniform edge removal up to 40-50%, but Smin deteriorates more quickly, meaning that predicting more specific terms suffers even under these simplified circumstances (Figure R2). The coverage also drops very slowly (at least initially), which implies that -as expected- most edges are removed from “dense” parts of the network so that the remaining edges can largely make up for this loss. In the “degree” sampling strategy, which is more realistic, we observed a much steeper drop for all three metrics. In this case, poorly-studied proteins lose their connections very quickly making it impossible to make predictions for them, as indicated by the steep decline in coverage. As a result, the average performance also reduces very fast. Figure R2: Fmax (left), Smin (middle) and coverage (right) of the GBA method on the yeast experimental PPI network (y-axis) as a function of the fraction of missing edges (x-axis). Edges where removed at random, either 1) uniformly (yellow), or 2) inversely proportional to the node degree, so that least connected nodes are more likely to lose their edges (purple). Error bars show standard deviation of 5 rounds of random sampling. This experiment relates to Figure 2e of the main text, where we showed the performance of the GBA classifier (corrected for the naïve performance, which we refer to as Prediction Advantage (PA)) as a function of the fraction of known edges. We used the degree-based down-sampling of edges to test whether we could validate the linear relationship between the two quantities that we had observed previously. The new figure is shown below as Figure R3. Because each network is a superset of all networks with fewer edges due to the step-wise edge removal, these observations are largely dependent, making it difficult to fit a curve to them. However, from the Figure we can see that the linear trend was not validated by this simulation. Figure R3: Prediction Advantage (PA) of Fmax as a function of the fraction of known interactions. Each species is shown as a blue dot and red line shows the least squares linear fit. PA values calculated by downsampling the original yeast network at different levels of missing edges are shown as green dots. Changes in manuscript: We included this new experiment and figure R2 (as Figure S2) in S1 Text. We have added a paragraph in the Results (subsection “Only the yeast experimental PPI network has acceptable function prediction performance”) describing this experiment and the results: “To better characterize the effect of missing edges, we simulated the phenomenon in yeast by removing edges either uniformly at random or by an approach that makes nodes with the lowest degree more likely to lose their edges first (S1 Text). We found that the Fmax is relatively robust to uniform edge removal up to 40-50%, but Smin deteriorates more quickly, meaning that predicting more specific terms suffers even under this simplified missing edges scenario. The coverage also drops slowly (at least initially), which implies that most edges are removed from “dense” parts of the network so that the remaining edges can partly make up for this loss. In the degree-based sampling strategy, which is more realistic, we observed a much steeper drop for all three metrics. In this case, poorly-studied proteins lose their connections very quickly making it impossible to make predictions for them, as indicated by the steep decline in coverage. As a result, the average performance also reduces very fast. The PA values calculated from the degree-based down-sampling did not confirm the linear relationship between PA and fraction of known edges (green dots in Figure 2e).” We also added the Prediction Advantage (PA) values calculated in this down-sampling experiment in Figure 2e that shows the PA with respect to the naïve approach as a function of known edges. In doing so, we also discovered a bug in the code that draws Figure 2e which lead to inverted values, which we corrected. This led to a slightly different slope (0.98 instead of 0.95) and p-value (0.016 instead of 0.049), but did not affect the conclusions drawn before from this figure. Minor 2) How is the GO hierarchy dealt with in this study? As the GO ontology contains a high number of terms, very often 2 proteins might be annotated to different terms, which are however very closely related in the hierarchy. Would they count as mismatches in this case? Did the authors use a simplified version of the GO terms (GO slim)? This should be more carefully explained! We did not use GO slim, but the whole GO version. GO annotations from the gaf file were propagated upwards to also include ancestral terms. This means that if a protein is annotated with a term and a classifier predicts a different but related term, it is likely that the two terms will have common ancestral terms. These will be counted as correct predictions. For example, if a protein is annotated “DNA metabolic process”, which has 13 ancestors excluding the root, but is predicted to be involved in “RNA metabolic process”, which has the exact same ancestors, we will have 13 True Positives, 1 False Positive and 1 False Negative. For the calculation of the F1 score, TPs, FPs and FNs are all treated equally, whereas for the Semantic Distance the weight of each term is determined by its information content. Note that the guarantee of making predictions consistent with the GO graph holds only for the GBA method and k-NN classifier in the case of node2vec. It does not necessarily hold for the ridge classifier. Although we did not correct the predictions of the ridge classifier to take the hierarchy into account, it outperformed k-NN for 2 out of 4 species, so we believe that this did not severely impact its performance. Changes in manuscript: In subsection “GO annotations” (lines 136-138), we now make this explicit: “We used the entire GO graph (not restricted GO slim). Annotations were propagated towards the ontology root, so that when a protein is annotated with a term, it is also annotated with all its ancestors in the GO graph.”. We also mention the possible inconsistencies produced by the ridge classifier (lines 183-185): “We did not post-process the predictions of the ridge method so it is possible that it makes predictions that are inconsistent with the GO hierarchy”. 3) The authors should explain the definition of Fmax and Smin, as the readers might not be familiar with these evaluation metrics. Changes in manuscript: We have added the definitions in S1 Text as well as a reference to those definitions in the methods section of the main document. 4) Related to this, the use of Fmax and Smin has been questioned in a recent paper (Plyusnin et al., PLOS Comp. Biology 2019); could the authors comment on this? I understand that they used Fmax and Smin as these were the metrics used in the CAFA assessment, however I would like to have some comments on the performance of these metrics and the possible biases. The concerns raised by Plyusnin et al. about Fmax and Smin mainly have to do with these metrics being overly lenient to false positive predictions, which might be acceptable since many of the annotations are missing, meaning that most proteins are likely to be under-annotated. On the other hand, the same study showed that both metrics correlate very highly with the signal to noise ratio of the predictions. Based on that we argue that the limitations of the metrics probably will not alter the relative ranking of the tested methods, so our conclusions do not rely on the choice of evaluation measures. Changes in manuscript: We have added a paragraph in the discussion pointing out this potential limitation of our work: “Although Fmax and Smin are the most widely-used evaluation metrics for function prediction, a recent study has raised concerns about them. The concerns, which were based on artificially generated predicted annotations, mainly have to do with these metrics being overly lenient to false positive predictions. This might not be a big problem, as due to missing annotations most proteins are likely to be under-annotated. The same study showed that both metrics correlate highly with the signal to noise ratio of the predictions. Based on that we argue that our conclusions do not rely on the choice of evaluation measures, but we believe that proper evaluation of function prediction algorithms is a pressing issue that requires further research.” 5) The authors have evaluated the effect of using the node2vec procedure instead of the naïve GBA procedure, which should have the advantage of using a larger neighborhood compared to GBA. They state that node2vec is the preferred method compared to GBA (line 268). However, even if the trend shows an increase in performance in Fig 3, the improvement seems hardly significant. Could the authors quantify the improvement of node2vec compared to GBA in Figure 3? Following the suggestion by the reviewer, we quantified the significance of the performance improvement of node2vec with respect to GBA. To do so, we binned proteins from each species based on their degree, as previously, and applied the Wilcoxon rank sum test when comparing the performance the two methods on the proteins of each bin. The resulting p-values were controlled for the False Discovery Rate (FDR) using the Benjamini-Hochberg method. We did that separately for each species. We used an FDR threshold of 0.05 and found that the differences between GBA and node2vec were statistically significant for all bins in A. thaliana and E. coli (with node2vec being the superior method). In yeast, node2vec is significantly better in 6 out of 9 bins, GBA in 1 out of 9 and in 2 out of 9 bins there are no statistically significant differences. Overall, these results indeed show that node2vec is the preferred method even when statistically compared. Changes in manuscript: We have updated Figure 3 to denote significant bins with a star (*) next to the number of proteins in the bin. We have also changed section “Performance per protein” of the results to describe the statistical tests and the results. Finally, we have added a supplementary table (Table S4 in S1 Text) which lists the raw and corrected p-values with an appropriate reference from the main manuscript. 6) Typo in line 139. Changes in manuscript: Typo has been corrected. Submitted filename: Response2Reviewers.pdf Click here for additional data file. 9 Nov 2020 A thorough analysis of the contribution of experimental, derived and sequence-based predicted protein-protein interactions for functional annotation of proteins PONE-D-20-21508R1 Dear Dr. Makrodimitris, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Baldo Oliva Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: (No Response) ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: (No Response) ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: (No Response) ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: (No Response) ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I would like to thank the authors who provided an enhanced version of the manuscript. The authors answered all my concerns, including my major one. I am a bit surprised that the difference between all the nodes and only the annotated ones is that small. I suppose it depends a lot on the topology of the graph. I would have expect that on a sparse graph it could have improve the scoring because it would have create new links between annotated nodes (and not only "dilute" the signal). Reviewer #2: (No Response) ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Carl Herrmann 13 Nov 2020 PONE-D-20-21508R1 A thorough analysis of the contribution of experimental, derived and sequence-based predicted protein-protein interactions for functional annotation of proteins Dear Dr. Makrodimitris: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Prof. Baldo Oliva Academic Editor PLOS ONE
  30 in total

1.  Gene ontology: tool for the unification of biology. The Gene Ontology Consortium.

Authors:  M Ashburner; C A Ball; J A Blake; D Botstein; H Butler; J M Cherry; A P Davis; K Dolinski; S S Dwight; J T Eppig; M A Harris; D P Hill; L Issel-Tarver; A Kasarskis; S Lewis; J C Matese; J E Richardson; M Ringwald; G M Rubin; G Sherlock
Journal:  Nat Genet       Date:  2000-05       Impact factor: 38.330

2.  Compact Integration of Multi-Network Topology for Functional Analysis of Genes.

Authors:  Hyunghoon Cho; Bonnie Berger; Jian Peng
Journal:  Cell Syst       Date:  2016-11-23       Impact factor: 10.304

3.  The Arabidopsis Information Resource (TAIR): improved gene annotation and new tools.

Authors:  Philippe Lamesch; Tanya Z Berardini; Donghui Li; David Swarbreck; Christopher Wilks; Rajkumar Sasidharan; Robert Muller; Kate Dreher; Debbie L Alexander; Margarita Garcia-Hernandez; Athikkattuvalasu S Karthikeyan; Cynthia H Lee; William D Nelson; Larry Ploetz; Shanker Singh; April Wensel; Eva Huala
Journal:  Nucleic Acids Res       Date:  2011-12-02       Impact factor: 16.971

4.  Integrating protein-protein interactions and text mining for protein function prediction.

Authors:  Samira Jaeger; Sylvain Gaudan; Ulf Leser; Dietrich Rebholz-Schuhmann
Journal:  BMC Bioinformatics       Date:  2008-07-22       Impact factor: 3.169

5.  The GOA database: gene Ontology annotation updates for 2015.

Authors:  Rachael P Huntley; Tony Sawford; Prudence Mutowo-Meullenet; Aleksandra Shypitsyna; Carlos Bonilla; Maria J Martin; Claire O'Donovan
Journal:  Nucleic Acids Res       Date:  2014-11-06       Impact factor: 19.160

6.  Neural networks for link prediction in realistic biomedical graphs: a multi-dimensional evaluation of graph embedding-based approaches.

Authors:  Gamal Crichton; Yufan Guo; Sampo Pyysalo; Anna Korhonen
Journal:  BMC Bioinformatics       Date:  2018-05-21       Impact factor: 3.169

7.  Novel comparison of evaluation metrics for gene ontology classifiers reveals drastic performance differences.

Authors:  Ilya Plyusnin; Liisa Holm; Petri Törönen
Journal:  PLoS Comput Biol       Date:  2019-11-04       Impact factor: 4.475

8.  Information-theoretic evaluation of predicted ontological annotations.

Authors:  Wyatt T Clark; Predrag Radivojac
Journal:  Bioinformatics       Date:  2013-07-01       Impact factor: 6.937

9.  MS-kNN: protein function prediction by integrating multiple data sources.

Authors:  Liang Lan; Nemanja Djuric; Yuhong Guo; Slobodan Vucetic
Journal:  BMC Bioinformatics       Date:  2013-02-28       Impact factor: 3.169

10.  deepNF: deep network fusion for protein function prediction.

Authors:  Vladimir Gligorijevic; Meet Barot; Richard Bonneau
Journal:  Bioinformatics       Date:  2018-11-15       Impact factor: 6.937

View more
  1 in total

1.  In silico and gene expression analysis of the acute inflammatory response of gilthead seabream (Sparus aurata) after subcutaneous administration of carrageenin.

Authors:  Jose Carlos Campos-Sánchez; Javier Mayor-Lafuente; Francisco A Guardiola; María Ángeles Esteban
Journal:  Fish Physiol Biochem       Date:  2021-08-26       Impact factor: 2.794

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.