| Literature DB >> 23514099 |
Domenico Cozzetto1, Daniel W A Buchan, Kevin Bryson, David T Jones.
Abstract
BACKGROUND: Accurate protein function annotation is a severe bottleneck when utilizing the deluge of high-throughput, next generation sequencing data. Keeping database annotations up-to-date has become a major scientific challenge that requires the development of reliable automatic predictors of protein function. The CAFA experiment provided a unique opportunity to undertake comprehensive 'blind testing' of many diverse approaches for automated function prediction. We report on the methodology we used for this challenge and on the lessons we learnt.Entities:
Mesh:
Substances:
Year: 2013 PMID: 23514099 PMCID: PMC3584902 DOI: 10.1186/1471-2105-14-S3-S1
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Coverage statistics of the CAFA test set.
| Method | Coverage of Targets | |||
|---|---|---|---|---|
| Released | Assessed | MF | BP | |
| Orthologous Groups | 49.15% | 60.84% | 61.75% | 61.01% |
| PSI-BLAST | 83.83% | 90.42% | 89.89% | 89.45% |
| Profile-Profile Comparison | 99.87% | 99.50% | 99.73% | 99.31% |
| Amino Acid Trigram Mining | 49.64% | 47.73% | 50.00% | 47.25% |
| Swiss-Prot Text Mining | 89.19% | 94.96% | 95.08% | 94.50% |
| FFPRED | 72.40% | 56.30% | 50.82% | 56.65% |
| FunctionSpace | 11.27% | 7.56% | 6.83% | 8.03% |
For each method, the columns report the percent of targets that were assigned GO terms relative to the entire benchmark (Released), the subset officially evaluated (Assessed), the evaluated molecular function predictions (MF), and the assessed biological process assignments (BP).
Figure 1Performance of Jones-UCL in comparison with the component methods. Comparison of the COGIC score distributions achieved by the individual predictors on the MF (upper panel) and BP (lower panel) benchmark sets. Each box spans from the first to the third quartile of the corresponding distribution; the median value is highlighted as a thick line, while putative outliers are plotted as empty circles. The width of each box is proportional to the number of targets predicted and officially assessed. Some names have been shortened ("Orthologues" for Orthologous Groups, "Prof-Prof Aln" for Profile-Profile Alignment, "AA 3-gram Mining" for amino acid trigram mining and "Text Mining" for Swiss-Prot Text Mining).
Figure 2Average precision against average recall for the Jones-UCL and the baseline methods. The graph shows the performance of our aggregate method in comparison with the simple annotation strategies calculated by the CAFA assessors Priors and BLAST. Each data point represents the average precision and recall across the entire set of 595 targets with functional annotations as of July 2011.
Figure 3Performance comparison based on COGIC scores. The boxplot recapitulates the target-based distributions for the Jones-UCL prediction team as well as for the naïve algorithms Priors and BLAST.
Statistical comparisons of prediction performance.
| Method | Method | Set | n | p-value |
|---|---|---|---|---|
| Jones-UCL | Priors | MF | 366 | 7.75E-06 |
| Jones-UCL | BLAST | MF | 351 | 3.37E-23 |
| Jones-UCL | Priors | BP | 436 | 3.43E-15 |
| Jones-UCL | BLAST | BP | 387 | 6.62E-12 |
The performance of pairs of methods was compared separately on the two benchmark subsets MF and BP. We tested the null hypothesis that the distributions of the COGIC scores over the set of common targets were indistinguishable with a paired two-sided Wilcoxon sum rank test. The table reports the number n of common targets and the p-value.
Figure 4Relationship between confidence scores and GO term specificity. Positive values indicate that on average the assignments made by the team Jones-UCL at a given confidence level were more specific than the corresponding experimental annotations. Negative values indicate the converse. Molecular Function (MF) and Biological Process (BP) data were analysed and plotted separately.