Literature DB >> 33601085

FAD-BERT: Improved prediction of FAD binding sites using pre-training of deep bidirectional transformers.

Quang-Thai Ho1, Trinh-Trung-Duong Nguyen2, Nguyen Quoc Khanh Le3, Yu-Yen Ou4.   

Abstract

The electron transport chain is a series of protein complexes embedded in the process of cellular respiration, which is an important process to transfer electrons and other macromolecules throughout the cell. Identifying Flavin Adenine Dinucleotide (FAD) binding sites in the electron transport chain is vital since it helps biological researchers precisely understand how electrons are produced and are transported in cells. This study distills and analyzes the contextualized word embedding from pre-trained BERT models to explore similarities in natural language and protein sequences. Thereby, we propose a new approach based on Pre-training of Bidirectional Encoder Representations from Transformers (BERT), Position-specific Scoring Matrix profiles (PSSM), Amino Acid Index database (AAIndex) to predict FAD-binding sites from the transport proteins which are found in nature recently. Our proposed approach archives 85.14% accuracy and improves accuracy by 11%, with Matthew's correlation coefficient of 0.39 compared to the previous method on the same independent set. We also deploy a web server that identifies FAD-binding sites in electron transporters available for academics at http://140.138.155.216/fadbert/.
Copyright © 2021 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  BERT; Deep learning; Electron transport chain; FAD binding Site; Natural language processing; Position specific scoring matrix

Year:  2021        PMID: 33601085     DOI: 10.1016/j.compbiomed.2021.104258

Source DB:  PubMed          Journal:  Comput Biol Med        ISSN: 0010-4825            Impact factor:   4.589


  4 in total

1.  ProtTrans-Glutar: Incorporating Features From Pre-trained Transformer-Based Models for Predicting Glutarylation Sites.

Authors:  Fatma Indriani; Kunti Robiatul Mahmudah; Bedy Purnama; Kenji Satou
Journal:  Front Genet       Date:  2022-05-31       Impact factor: 4.772

2.  BERT-m7G: A Transformer Architecture Based on BERT and Stacking Ensemble to Identify RNA N7-Methylguanosine Sites from Sequence Information.

Authors:  Lu Zhang; Xinyi Qin; Min Liu; Guangzhong Liu; Yuxiao Ren
Journal:  Comput Math Methods Med       Date:  2021-08-25       Impact factor: 2.238

3.  Generation of functional oligopeptides that promote osteogenesis based on unsupervised deep learning of protein IDRs.

Authors:  Mingxiang Cai; Baichuan Xiao; Fujun Jin; Xiaopeng Xu; Yuwei Hua; Junhui Li; Pingping Niu; Meijing Liu; Jiaqi Wu; Rui Yue; Yong Zhang; Zuolin Wang; Yongbiao Zhang; Xiaogang Wang; Yao Sun
Journal:  Bone Res       Date:  2022-03-01       Impact factor: 13.567

4.  BERT-PPII: The Polyproline Type II Helix Structure Prediction Model Based on BERT and Multichannel CNN.

Authors:  Chuang Feng; Zhen Wang; Guokun Li; Xiaohan Yang; Nannan Wu; Lei Wang
Journal:  Biomed Res Int       Date:  2022-08-24       Impact factor: 3.246

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.