| Literature DB >> 29039790 |
Renzhi Cao1, Colton Freitas1, Leong Chan2, Miao Sun3, Haiqing Jiang4, Zhangxin Chen5.
Abstract
With the development of next generation sequencing techniques, it is fast and cheap to determine protein sequences but relatively slow and expensive to extract useful information from protein sequences because of limitations of traditional biological experimental techniques. Protein function prediction has been a long standing challenge to fill the gap between the huge amount of protein sequences and the known function. In this paper, we propose a novel method to convert the protein function problem into a language translation problem by the new proposed protein sequence language "ProLan" to the protein function language "GOLan", and build a neural machine translation model based on recurrent neural networks to translate "ProLan" language to "GOLan" language. We blindly tested our method by attending the latest third Critical Assessment of Function Annotation (CAFA 3) in 2016, and also evaluate the performance of our methods on selected proteins whose function was released after CAFA competition. The good performance on the training and testing datasets demonstrates that our new proposed method is a promising direction for protein function prediction. In summary, we first time propose a method which converts the protein function prediction problem to a language translation problem and applies a neural machine translation model for protein function prediction.Entities:
Keywords: machine learning; neural machine translation; protein function prediction
Mesh:
Substances:
Year: 2017 PMID: 29039790 PMCID: PMC6151571 DOI: 10.3390/molecules22101732
Source DB: PubMed Journal: Molecules ISSN: 1420-3049 Impact factor: 4.411
Figure 1The flowchart of our method for protein function prediction. Three layers of RNN is used for the NMT model.
Figure 2(A) The accuracy during training process for NMT model; (B) The accuracy during training process for extended NMT model. The x-axis represents the steps during the training, we scaled it by dividing 10,000, and the first 10,000 steps are not shown in the figure since the perplexity is very big.
Figure 3Precision and recall of different methods on selected proteins. The top n metric is used, and n is set to be 20. The area under curve (AUC) for NMT, extended NMT, ProLanGO, and Probability based model is 0.286, 0.305, 0.333, and 0.390, respectively.
Figure 4Average GO similarities of ProLanGO and other methods on threshold metric.
Figure 5Average GO similarities of ProLanGO and other methods on top n metric.