Literature DB >> 33711547

Reservoir hosts prediction for COVID-19 by hybrid transfer learning model.

Yun Yang1, Jing Guo2, Pei Wang2, Yaowei Wang3, Minghao Yu3, Xiang Wang3, Po Yang4, Liang Sun5.   

Abstract

The recent outbreak of COVID-19 has infected millions of people around the world, which is leading to the global emergency. In the event of the virus outbreak, it is crucial to get the carriers of the virus timely and precisely, then the animal origins can be isolated for further infection. Traditional identifications rely on fields and laboratory researches that lag the responses to emerging epidemic prevention. With the development of machine learning, the efficiency of predicting the viral hosts has been demonstrated by recent researchers. However, the problems of the limited annotated virus data and imbalanced hosts information restrict these approaches to obtain a better result. To assure the high reliability of predicting the animal origins on COVID-19, we extend transfer learning and ensemble learning to present a hybrid transfer learning model. When predicting the hosts of newly discovered virus, our model provides a novel solution to utilize the related virus domain as auxiliary to help building a robust model for target virus domain. The simulation results on several UCI benchmarks and viral genome datasets demonstrate that our model outperforms the general classical methods under the condition of limited target training sets and class-imbalance problems. By setting the coronavirus as target domain and other related virus as source domain, the feasibility of our approach is evaluated. Finally, we show the animal reservoirs prediction of the COVID-19 for further analysing.
Copyright © 2021 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  COVID-19; Ensemble learning; Hosts prediction; Machine learning; Transfer learning; Virus origins

Mesh:

Year:  2021        PMID: 33711547      PMCID: PMC7942058          DOI: 10.1016/j.jbi.2021.103736

Source DB:  PubMed          Journal:  J Biomed Inform        ISSN: 1532-0464            Impact factor:   8.000


Introduction

With the rapid growing number of infected and dead people all over the world, COVID-19 is becoming one of the most serious epidemic situations for the humankind. The emerging viral infection, such as COVID-19, Ebola, SARS, MERS and Zika, is a kind of zoonosis which is caused by the cross-species transmission of virus [1]. Due to the pathogenicity, lethality and difficulty of prevention, these diseases have made a grievous threat to the health of mankind and livestock [2]. Virus is an acellular form, usually contains a long chain of nucleic acids and a protein shell, which has no metabolic machinery or enzyme system. Unlike most living things, virus survives mainly depending on the host cells that provide places for virus to replicate, transcribe and transfer to other organisms [3]. Once the zoonotic disease outbreak or new pathogenic virus is found, origin tracing has a great significance not only to figure out how the pathogens infect humans and develop the vaccine for human beings, but also can isolate the infectious animal hosts for further spreading. In the history of mankind, the explorations of the virus origins have made many progresses, for example, Hu et al. [4] take 5-year surveillance of SARSr-CoVs in a cave inhabited by multiple species of horseshoe bats in Yunnan Province, China, to identify the origin of SARS-CoV. Reusken et al. [5] collect 349 serum samples from all over the world to demonstrate the reservoir of MERS-CoV has high relationship with Omani camels. The researchers have made many studies for COVID-19 recently [6], [7], [8], [9]. By the conclusion that the similar phylogenetic characteristics of genomes may have the similar hosts information, the researchers seek the similar genome sequences from the previous virus database and find that the COVID-19 and the virus (Bat CoV RaTG13) have a full gene level consistency up to 96%, and the viruses (bat-SL-CoVZC45 and bat-SL-CoVZXC21) have the consistency nearly 90% [9]. According to these researches, the mainstream of hypothesizes present that the most probable host for COVID-19 comes from the Chinese Horseshoe Bat. However, traditional methods of virus origin need huge evidences from field surveillance, phylogenetic, laboratory experiments, and real-world interventions [10] which consume lots of time and labor. Thus, it is crucial to develop a more precise, celerity and comprehensive method not just for COVID-19, but also for other newly discovered viruses in the future. Recent studies have made a great development in discovering the information of genome sequences via machine learning [11] which offer a new research orientation for virus origins. Through building a mathematical model based on training datasets, machine learning can dig out the latent information from dataset and make predictions for specific tasks [12]. Since a large amount of genetic information is recorded in genome sequences, knowledge about the viral hosts can be mined through machine learning [13], [14], [15], [16]. Compared with other application fields in machine learning, although the overall base of biological data is large, the scarcity of annotated virus and imbalanced hosts information restrict the efficiency of traditional machine learning in hosts prediction. Specifically, due to the difficulty of tracing the virus origins, the number of annotated virus sequences are limited, especially for the newly discovered virus. The variances of data distribution among different virus species will further constrain the performance of traditional machine learning methods in origin predictions. In addition, some hosts categories, like bat and poultry, are the most frequently discovered carriers of virus, which lead to the prediction results bias to the majority categories. Transfer learning breaks the assumption of traditional machine learning that the distribution of training data and testing data must be consistent [17], [18]. It can effectively use the knowledge obtained from the related virus domain to improve the capability for the task of hosts prediction when there is limited data in specific virus species. Transfer learning can be divided into four categories [17]: instance-based transfer learning [19], [20], [21], feature-based transfer learning [22], [23], [24], [25], model-based transfer learning [26], [27] and relation-based transfer learning [28], [29]. Although transfer learning can remarkably decrease the distribution variance between different virus species, it cannot fix the imbalance problem and lead to low generalization when applying into newly discovered virus. While ensemble learning completes classification tasks by constructing and combining multiple weak classifiers to achieve better results [30], [31]. To obtain a superior performance in hosts prediction, ensemble learning can be used for three main purpose: (1) Fix the imbalanced class; (2) Utilize the tri-training strategy to give the predictions from different views; (3) Ensemble the results in each iteration to acquire better generalization. Our previous studies have illustrated the advantages of ensemble learning algorithms [32], [33], [34], [35], [36]. In this paper, we provide an integrated workflow to make a fast, accurate and reliable host tracing for newly discovered virus (e.g., COVID-19). All the adopted genome sequences and the hosts information are downloaded from GenBank. We preprocess and annotate the collected sequences to construct a benchmark dataset of reference sequences (refseqes) [37]. To overcome the fundamental weakness in traditional machine learning methods in hosts prediction, we extend transfer learning and ensemble learning to present a novel hybrid transfer learning model (HTL) that takes the advantages of the information gained from the related virus domain to enhance the prediction reliability on target virus domain. Specifically, considering the scarcity of annotated virus dataset, we utilize transfer learning to decrease the distribution variance between different virus species and leverage the well annotated virus domain as auxiliary to improve the prediction reliability for newly discovered virus. A bootstrap-based ensemble strategy is proposed to unravel the class imbalance problem and reconcile the outputs to obtain a final transfer result. In general, the main contributions of our work are summarized as follows: Considering the traditional virus origin methods consume lots of time and labor, our approach provides a new research reference for hosts tracing from the perspective of data mining and offers an intelligent workflow to accurately and quickly predict the host of newly discovered virus. By collecting the genome sequences and annotating the hosts information, we build a benchmark of viral reference sequences to help researchers discover and develop the fight against virus. To solve the problems of few annotated data and class-imbalance in virus datasets, we utilize the advantages of transfer learning and ensemble learning to conduct a HTL model. Being compared with other classical machine learning algorithms, we demonstrate the superiority of HTL on UCI benchmarks and virus datasets. We utilize HTL to give the host prediction for newly discovered virus (e.g., COVID-19). Being compared with traditional virus origin methods such as sequence identity and phylogenetic tree, we demonstrate the reliability of our method. The remainder of this paper is organized as follows. Section 2 presents the related works. Section 3 describes the workflow of predicting the origins of virus. Section 4 reports the simulation results on UCI benchmarks and viral dataset, furthermore, gives the prediction results of COVID-19. Then we provide further discussion of the results and the future work. Conclusions are drawn in the last section.

Related work

Several previous works about viral origins prediction via machine learning methods are illustrated in this section. Christine et al. [13] propose a methodology to construct the computational models for 11 influenza proteins. By adopting the machine learning algorithm random forest [38], high accurate prediction models are trained to predict the host tropism of influenza A virus. Being compared with other 5 classical machine learning methods, random forest reveals the better performance than others. Babayan et al. [14] utilize gradient boosting machines (GBMs) [39] to demonstrate the superiority of machine learning in predicting the virus hosts and arthropod vectors from viral genome sequences. With the hypothesis of related viruses have closely related hosts, they design a phylogenetic neighbourhood (PN) model to predict host associations from viral phylogenetic connection [40] and combine these phylogenetic information with algorithm to maximize the prediction accuracy. Furthermore, Babayan provides a suitable way to quantify the genome sequences into 4229 traits, including codon pair, dinucleotide, codon, and amino acid biases. Through this way, the latent information of sequences can be extracted into a more interpretative form for machine learning to construct the classification models. Zhang et al. [15] present a method which uses the word frequencies of viral sequences to predict the virus-host infectious associations. Under the hypothesis of the word pattern between viruses and their hosts have more similarities [41], Zhang conducts four different feature vector representations [42], [43], [44] and evaluates them by various supervised machine learning methods. The latest work of viral host prediction, Mock et al. [16] provide a general workflow consists of several steps, including data collection, subsets creation, model construction and output the final prediction. Unlike the work illustrated before, they directly use one hot encoding to alter the sequences into a list of numerical vectors rather than make a sequence quantification. The deep neural network architecture is constructed by the CNN layers [45], LSTM layers [46] and two dense layers. To the best of our knowledge, our work is among the first attempt to utilize the transfer learning to solve the host origin problems. In such way, the fundamental weakness of traditional methods can be reduced, hence a satisfactory result can be achieved.

Methods

In this section, we first describe the proposed workflow of predicting the hosts for newly discovered virus, and then present the way of data preprocessing and viral datasets construction. The hybrid transfer learning model is described as the main component of our proposed method.

AI-based viral origins prediction

The AI-based viral origins prediction method consists of four components: (I) Use the detection technology to discover the new virus; (II) Detect the genome sequences of the viruses by using nucleic acid testing, such as high-throughput sequencing and reverse transcription-polymerase chain reaction; (III) Collect the genome sequences and annotate the datasets with the hosts information; (IV) Train machine learning based classification model with the collected datasets, then utilize the model to make the final prediction. The workflow of AI-based viral origins prediction is illustrated in Fig. 1 .
Fig. 1

An overview of the proposed AI-based viral origins prediction.

An overview of the proposed AI-based viral origins prediction. In this paper, we focus on the last two processes and summarize them into two major perspectives. First, for the data pre-processing and virus dataset construction, we collect the genome sequences and annotate the sequences with the information from GenBank, which is an online gene sequence database collects and annotates all published nucleic acid and protein sequences. Then the feature quantification methods are used to change the collected genome sequences into a numerical form. We build a phylogenetic tree to further demonstrate the variances exist among different virus species and show the similarity between COVID-19 and coronavirus in Fig. 2 . Thus, the collected datasets are divided into the target domain with coronavirus and COVID-19 and the source domain with other related virus sequences. Second, we construct the hybrid transfer learning model to make the final prediction. We utilize the domain adaptation method [23] to map the source domain and target domain into a Reproducing Kernel Hilbert Space (RKHS) [47] in which the variances between the source and target domain should be as similar as possible. After that, ensemble learning methods are used to construct the classification model. With the undersampling [48] and the tri-training strategy [31], our method can refine the class-imbalance problems and get a better performance on target domains.
Fig. 2

The phylogenetic tree of the viral genome sequences. By proportional sampling from the collected datasets, the phylogenetic tree is built to demonstrate the variances among different virus species. The blue leaf nodes denote the other related virus species, the red denote the coronavirus and we highlight the COVID-19 and colored with orange.

The phylogenetic tree of the viral genome sequences. By proportional sampling from the collected datasets, the phylogenetic tree is built to demonstrate the variances among different virus species. The blue leaf nodes denote the other related virus species, the red denote the coronavirus and we highlight the COVID-19 and colored with orange.

Data Pre-processing and virus dataset construction

For predicting the hosts for newly discovered virus (e.g., COVID-19), we focus on the RNA viruses since they are the primary pathogen group responsible for emerging human diseases. The detail information of concerning viral datasets are shown in Table 1 . We download all the viral genome sequences from NCBI viral genome database, which consists of 308 samples of COVID-19, including 307 complete genome sequences and 1 refseq, and 775 annotated refseqes with 11 taxonomic groups (Arenavirus, Astrovirus, Bunyavirus, Calicivirus, Coronavirus, Filovirus, Flavivirus, Hepevirus, Paramyxovirus, Picornavirus, Rhabdovirus). The selected viral species cover most of the virus species that can infect human diseases. The reason for we adopt the refseqes to construct the training datasets is that the complete genome sequences contain a large number of analogous sequences from same virus types, resulting in the overfitting when building the classification model. For each genomic sequence, corresponding coding sequence (CDS) encodes the specific protein products for the virus which provides a significant amount of important information about gene expression [49]. Therefore, we only consider the sequences that encodes the protein product and extract the corresponding CDS to form the final data collection. With regard to the multi coding sequences exist in each genomic sequence caused by the different synthesis of proteins, we quantify each coding sequence and superimpose the results to obtain the characteristic quantization for this sequence. After that, the relevant hosts information is annotated from GenBank. According to the authoritative literatures of viruses, we annotate most of the hosts information based on organisms that have long helped the virus survive. On account of the dense granularity of hosts information given in GenBank, we map the collected hosts information into 10 categories, including Artiodactyl, Carnivore, Fish, Galloanserae, Neoaves, Plant, Primate, Pterobat, Rodent and Vespbat. It is worth noting that the bats (order Chiroptera) were split into Pteropodiformes (families Pteropodidae, Rhinolophidae, Hipposideridae, Megadermatidae, and Rhinopomatidae, here abbreviated “Pterobats”) and Vespertilioniformes (remaining microbat families, here abbreviated “Vespbats”). These hosts categories include mammals, rodents and birds, which are common linked to the virus that can infect humans and spread from species.
Table 1

The descriptions of the virus datasets.

Reservoir groupsSamplesSource domainTarget domain
Other related virusCoronavirusCOVID-19
Artiodactyl79736
Carnivore36333
Fish35323
Galloanserae78726
Neoaves47407
Plant1111074
Primate1461406
Pterobat62539
Rodent1471416
Vespbat34259
Total77571659308
The descriptions of the virus datasets. Inspired by [9], which provides a Blastn research for the complete genomes of COVID-19 and shows the similarities with other coronavirus in sequence identity and query coverage, we speculate that coronavirus is the most suitable virus group to construct the classification model for predicting the hosts for COVID-19 and the discrepancy among different virus species will restrain the accurate of the final prediction. To further demonstrate the variances between different viral species and provide a better visualization, we randomly sample from the collected viral refseqes and the genome sequences of COVID-19 to construct a phylogenetic tree. The MAFFT is a multiple sequence alignment program which is used to align the viral sequences [50]. With the processed sequences, we create the tree by iTOL [51] which is an online tool for the display and manipulation of phylogenetic trees. As shown in Fig. 2, we variegate the leaf nodes with three types of colors to represent the coronavirus group, the COVID-19 virus group and other related virus group, the inside circle denotes the GenBank accession of the viruses. We use 10 kinds of colors to denote 10 reservoir hosts and blank for COVID-19. The branch in the radial direction represents the viral evolution of the evolutionary lineage with time, which means that the nearer the viral species on the branch, the closer the kinship of virus will be and the genome sequences will have the similar expressions. Note a fact that, the COVID-19 virus has the similar branch length with most of coronavirus in depth. In other words, the COVID-19 viruses have the large dissimilarities with other related virus groups, but the similar depth and breadth with most of the coronavirus. However, as shown in Table 1, the annotated training samples are limited and the classes of reservoir are imbalanced, only using coronavirus dataset is insufficient to build a highly generalized model for COVID-19. Consequently, the other related 716 viral sequences with well annotated hosts information can be used as an auxiliary to optimize the task of predicting the hosts of COVID-19. We divide the datasets into target domain with 59 coronavirus refseqes and COVID-19 sequences, and source domain with other related virus sequences as illustrated in Table 1.

Quantification of viral genomic traits

Following the method in [14], we quantify each viral genomic sequences into 4229 genomic traits, including all possible codon pair, dinucleotide, codon, and amino acid biases. The detail calculation methods are defined as follows:

Dinucleotide bias

The dinucleotide bias is calculated across all coding genomic sequences by Eq. (1).in which, the denote the count number, the ,, , and represent the number of nucleotides X and Y, the total nucleotides numbers, the total count of the dinucleotide XY, and the total number of dinucleotides across the whole sequences. According the [52], dinucleotide bias on viral fitness are reported to be strongest at the bridge between neighbouring codons, an extra calculation is made for “bridge” and “non-bridge” codon positions. With each of the 16 possible dinucleotides for nucleotide, it generates 64 traits in this way.

Codon pair score (CPS)

Codon pair bias was measured as the codon pair score (CPS) for each of the 4096 (64 × 64) possible codon-codon pairs. Codons are the three adjacent bases that determine amino acids on mRNA. There are 64 codons, including 61 amino acid codons (including the start codon) and 3 stop codons. In this work, the stop codons are included when count the codon pairs. CPS is calculated by Eq. (2).where A and B represent the number of codons A and B, respectively, and AB denotes the codon pair which is consisted by codons A and B. Similarly, X and Y represent the number of the corresponding amino acids X and Y, and XY as the amino acid pair across the genome sequences. To count the number of the codon pairs and the amino acids, the CPS determines if a given codon pair is over-represented or under-represented. For the given codons A and B, if the number of codon pair AB is larger than the amino acid pair XY, the CPS is a positive number represents the over-represented, on the contrast, a negative number as under-represented.

Codon bias

Codon bias is calculated for the 64 codons separately by dividing the number of each codon by the total number of all codons across the complete sequences. The formulation is presented as Eq. (3)

Amino acid bias

Same as the calculate way of codon bias, the amino acid bias for 21 amino acids (stop is also considered as an amino acid here) is calculated by dividing each amino acid by the total number of amino acids in the selected sequences. According to the quantification methods illustrated above, the genome sequences are transformed into 4229 traits (CPS = 4096, dinucleotide biases = 48, codon biases = 64, amino acid biases = 21).

Hybrid transfer learning model

In this section, we will give a description of our hybrid transfer learning model. As illustrated in Fig. 3 , the proposed method can be further divided into two modules: domain adaptation and ensemble learning model.
Fig. 3

The flow chart of hybrid transfer learning method.

The flow chart of hybrid transfer learning method.

Domain adaptation

In domain adaptation module, domain invariant feature is extracted by transfer learning method. Domain Adaptation is an important branch of feature-based transfer learning that improves the performance of classifier by adopting one or more source domains for the purpose of transferring information [18] and it is one of the most actively researched transfer learning methods [22], [24], [25]. The Maximum Mean Difference (MMD) [53] is a commonly used measurement to evaluate the distribution discrepancy across domains. By minimizing the MMD, the optimal mapping space can be found. Assume that source domain , where is the feature of the source domain and is the corresponding label. Similarly, the target domain can be denoted as , where the feature of the labelled data is and is the corresponding label, is the feature of the unlabelled data. The distribution issues that occur between the source and target domains can be largely summarized in two different aspects: marginal distribution, namely and conditional distribution, namely [18]. Regarding the complex variances between different viral species, it is rare to tell the variance types for the given viral datasets. Therefore, Joint Domain Adaptation (JDA) [23] is adopted to reduce the variance through the perspective of the joint probability distribution, which takes both the marginal and conditional distributions into consideration. In addition, when mapping the source and target domain into the new feature space, JDA provides a method of dimensionality reduction which can effectively reduce the redundancy in our datasets. Unlike the traditional method which the target domain has no annotated datasets in JDA, we can use the limited label information from the target domain to give the pseudo labels for target testing sets and provide a refine adaptation result. The analysis is based on the conceptual framework of MMDE [27], which first takes advantage of kernel tricks to change the way of learning the mapping function into an optimal adaptation matrix A. The detail computation methods for JDA is reported in [23]. After mapping into the new feature space, the new dataset can be notated as in the source domain, and in the target domain.

Ensemble learning model

Although the distribution discrepancy between the source domain and target domain are substantially reduced, the class-imbalance problem and low generalization of model still constrains the performance of classifier on the target domain. To improve the performance on the target domain, we construct the ensemble learning model in three steps: First, during each iteration, the training sets is obtained using undersampling [48]. Specifically, for each process, we keep the target training sets remained and use a subset of the majority class in source domain. In this way, the class-imbalanced problem will be fixed, and the model can be robust to noisy samples. Second, based on the assumption that if two heterogeneous classifiers have consistency for labelling the testing sample, it can be considered as high confidence, we propose a tri-training strategy to exploit unlabelled data in target testing sets. These two heterogeneous classifiers are trained with training sets . If the testing sample fits the condition , then the is considered as the high confidence sample and added into the training sets with its corresponding pseudo label to form the new training set . Then, the new training sets are utilized to train the classifier which is used to give the prediction for all testing sets. We record the accuracy on the target training samples for the final weighting strategy. Repeat the former two steps until the number of iterations is reached. Third, all the outputs of from each iteration are ensemble into a final output. The differences of each output mainly come from the sampling result and the quality of the high confidence samples in each iteration. To cut down the impact in each iteration and give a promising final output, we measure the performance of on the accuracy of the target training sets and define the weights as follow:where and . The final output of hybrid transfer learning method is constructed by a linear combination of the output in each iteration which is shown in Eq. (6).

Simulation

Before giving the prediction for the hosts origins of COVID-19, some simulations need to be made for validation. We apply our hybrid transfer learning method to a collection of UCI medical benchmarks and our collected viral datasets. The experimental settings and the results are reported in Section A and B. After demonstrate the effectiveness of our method, the prediction results for COVID-19 are illustrated in Section C.

UCI medical benchmarks

To evaluate the properties of our approach in handling the large variances datasets, 13 representative medical datasets from the UCI repository [54] are selected to conduct the experiments. For the UCI datasets do not have any hierarchy, when splitting the data to generate the source and target domains, we use the methods in [19] and apply binary feature values in the datasets, such as age group, area to split the datasets and simulate inconsistent data distributions. In the last three groups of datasets, the attributes are all continuous variables. One way to separate the source domain and target domain is to select a given attribute and use K-means to cluster them into two partitions. Intuitively, these two partitions will have different data distributions. The information of concerning datasets is shown in Table 2 in detail.
Table 2

The descriptions of the UCI medical benchmarks.

DatasetsFeatureSampleSource domain
Target domain
Pos.Neg.Pos.Neg.
Autism203461261226236
Heart Disease12270831006720
Brest Cancer8277323016451
Diabetic181151194193417347
Sani54303129408747
Thoracic154705526815132
Lung73651491005660
Colic163681601107228
Cervical3266837535888
Sick272643163313179881
Parkinson22195311911629
Mammography58302344380383
wdbc1456918593172119
The descriptions of the UCI medical benchmarks. For the main parameters in the domain adaptation method, the adopted experiments settings are showed as follows. We simulate the parameter adjustment methods in JDA [23] and seek the optimal parameters through an empirical approach. The iteration number of JDA is set as 10. The optimal is obtained from . The domain adaptation methods involve dimensionality reduction and we select of the initial data dimensions for dimensionality reduction and choose the best parameter for each model. Regarding kernel selection, the RBF kernel is chosen for the datasets Autism, Lung and Cervical, while the linear kernel is used for the rest of the datasets. We set as Logistic Regression (LR), as SVM, as Multi-layer Perception (MLP), and the iteration number of the ensemble learning model is N = 10. The average accuracy and standard deviation of 10 times of experiments are used for evaluation, furthermore, we select 10% target samples as target training sets in each experiment and set the same random seed to guarantee the fairness. To demonstrate the interpretability of HTL and figure out which part makes a major contribution in our method, we split every component of HTL and make a detail comparation for each output results. Firstly, we utilize the domain adaptation method (e.g., JDA) to decrease the variance between source and target domain, and use the data after feature mapping as input to train three heterogenous classifiers (e.g., LR, SVM, MLP). Secondly, the main purpose of constructing the ensemble learning model is to solve the class imbalance problems and enhance the reliability on testing datasets. Therefore, we compare the performance of HTL with the ensemble learning model without undersampling. Table 3 collectively lists all the experiment results. As shown in Table 3, our approach achieves significantly better performance in terms of both classification and standard deviations and wins on 10 out of 13. More specifically, being compared with the results of only adopting the domain adaptation method, the performances of our ensemble strategy show superior results which means by adding the pseudo labels, the classifier can be iteratively influenced to reach a better fitting for testing datasets. What’s more, without the processing of undersampling, the prediction results may bias to the majority class which may cause the accuracy value higher than HTL. However, the bootstrap strategy could enhance the generalization of the classifier on the testing data.
Table 3

Classification accuracy of different components of HTL on benchmarks.

DatasetsDomain adaptation
Ensemble learning modelHTL
LRSVMMLPWithout undersampling
Autism81.5 ± 3.579.1 ± 2.379.3 ± 3.887.3 ± 2.189.1 ± 1.3
Heart Disease80.3 ± 3.179.8 ± 2.881.4 ± 3.682.1 ± 2.383.8 ± 2.8
Brest Cancer72.3 ± 1.274.1 ± 2.373.1 ± 3.578.1 ± 2.279.4 ± 1.2
Diabetic66.1 ± 0.665.2 ± 0.768.2 ± 1.369.4 ± 1.070.6 ± 0.9
Sani70.2 ± 2.372.4 ± 1.971.5 ± 2.174.9 ± 1.875.8 ± 2.4
Thoracic85.1 ± 1.183.2 ± 1.284.5 ± 2.386.7 ± 1.580.1 ± 1.7
Lung72.1 ± 2.475.9 ± 3.576.3 ± 2.176.2 ± 1.479.4 ± 2.2
Colic69.5 ± 1.970.8 ± 1.372.1 ± 1.576.3 ± 2.178.7 ± 2.2
Cervical88.4 ± 2.287.5 ± 2.187.1 ± 3.689.6 ± 2.787.6 ± 1.2
Sick93.2 ± 1.392.4 ± 2.593.2 ± 1.694.1 ± 2.388.1 ± 0.3
Parkinson76.8 ± 3.178.5 ± 4.377.9 ± 3.579.3 ± 3.980.5 ± 3.4
Mammography71.3 ± 2.372.4 ± 3.272.1 ± 2.773.5 ± 2.874.1 ± 1.8
wdbc90.1 ± 0.491.6 ± 0.891.8 ± 1.291.3 ± 0.892.3 ± 0.6
Classification accuracy of different components of HTL on benchmarks. After demonstrate the effectiveness of HTL, we compare our method with other three state-of-the-art transfer learning algorithms for medical classification problems, i.e., TrAdaBoost [19], MTLF [55] and CODA [56]. TrAdaBoost is an instance-based transfer learning method which conducts a new mechanism to automatically adjust the weights of training samples. The other two algorithms belonging to feature-based transfer learning methods. The MTLF takes advantage of the Mahalanobis distance [57] to evaluate the distribution differences between the source and target domains instead of MMD in domain adaptation. The CODA conduct a feature selection strategy based on the Pearson correlation coefficient [58], then co-training is used to improve the classifier. We can easily see their effectiveness of narrowing the data variances when the instance-based methods loss its power, especially in datasets Cervical and Sick. Table 4 shows that TrAdaBoost wins on two datasets and the MTLF algorithm wins on two datasets, while our HTL approach achieves the best results on the other 9 datasets. While in datasets Cervical and Sick, both of the datasets have imbalanced class problem which may cause the high accuracy. In general, our method can reveal the superiority when compared with other transfer learning methods on benchmarks.
Table 4

Classification accuracy of different transfer learning methods on benchmarks

DatasetsTrAdaBoostMTLFCODAHTL
Autism83.6 ± 3.387.6 ± 2.680.7 ± 2.389.1 ± 1.3
Heart Disease70.8 ± 1.378.9 ± 1.175.4 ± 1.283.8 ± 2.8
Brest Cancer69.1 ± 3.271.4 ± 3.673.1 ± 2.379.4 ± 1.2
Diabetic62.8 ± 1.364.4 ± 1.265.7 ± 1.370.6 ± 0.9
Sani72.5 ± 2.874.8 ± 1.773.4 ± 2.975.8 ± 2.4
Thoracic86.1 ± 1.880.2 ± 2.482.1 ± 1.780.1 ± 1.7
Lung67.2 ± 2.968.8 ± 1.777.4 ± 2.179.4 ± 2.2
Colic74.7 ± 3.476.1 ± 1.075.7 ± 1.678.7 ± 2.2
Cervical91.4 ± 0.993.3 ± 0.590.3 ± 0.787.6 ± 1.2
Sick92.5 ± 0.795.9 ± 0.190.6 ± 0.288.1 ± 0.3
Parkinson83.4 ± 1.781.5 ± 0.581.2 ± 1.280.5 ± 3.4
Mammography70.4 ± 1.371.2 ± 1.669.8 ± 2.374.1 ± 1.8
Wdbc89.8 ± 1.590.7 ± 0.990.5 ± 1.592.3 ± 0.6
Classification accuracy of different transfer learning methods on benchmarks

Experiment on virus datasets

In this part, we evaluate the effectiveness of HTL on viral origin tracing problems and compare it with other transfer learning methods and traditional machine learning methods. Since the discrepancy exists among different virus species, when predicting the newly discovered virus (e.g., COVID-19), the results obtained from the same virus species will be more reliable. However, in the case of limited training samples in coronavirus, only using coronavirus samples are insufficient to construct a robust classification model for COVID-19. Thus, the purpose of HTL is to construct a high generalized model for target domain (coronavirus) by leveraging the source domain (other related virus). The detail of the viral genome datasets and the domain partitions are reported in Table 1. By applying the quantification method described before, the collected refseqes are transformed into the numerical datasets with 4229 dimensions. The best parameters of JDA are set as {kernel = linear, dimension = 200,  = 0.1, gamma = 0.01}, the other parameter settings of HTL are same as Section A. The SVM and XGBoost are used as the base classifiers for comparison, since these two algorithms outperform than others on our datasets. The target training sets are randomly sampling from the target domain, while the rests are used for testing. To demonstrate the purposes of HTL and simulate the performances with other traditional machine learning methods, we project the experiment mainly from three perspectives. First, due to the information redundancy problem under the high-dimensional genetic datasets, we use 4 different feature inputs for comparison. The primitive feature represents the quantification results with 4229 traits and our HTL also uses the primitive feature for training. The other three feature processing methods comes from [14]. The Gradient Boosted Tree is used to make feature selection, which selects 50 most important features for training. The PN model [14] is adopted as a compared feature extraction method which uses the Blastn to analyse each genome sequence, then finds the top 5 similar virus from the training sets and uses their hosts information to generate the features for each sample. In [14], by combining selected features with viral PN, the classifier can get a better performance. Thus, these combined features are also used for comparison. Second, by setting the training sets with different compositions, we want to prove the constrains of the traditional machine learning methods in the conditions of limited training sets and unbalanced class problems. Furthermore, directly using the source domain may cause a negative effect on the classification model because of the discrepancies exist among different virus species. The compared baseline methods are implemented by SVM and XGBoost, Table 5 shows the different compositions of training datasets. Third, we conduct different amounts of target training datasets to prove the superiorities of our model in few sample problems. What’s more, transfer learning methods are also adopted to make a further comparation. For each set of experiments, we randomly select a corresponding number (1, 2, 3) of target domain from each category to form 3 comparative experiments. We perform 10 epochs for all experiments and calculate the mean accuracy and standard deviation for them.
Table 5

The compositions of datasets

BaselineTraining dataset LTesting dataset
SVMXS,YSXT,YTXTU
tar-SVMXT,YTXTU
XGBoostXS,YSXT,YTXTU
tar-XGBoostXT,YTXTU
HTLXS,YSXT,YTXTU
The compositions of datasets The final experiment results are collectively illustrated in Table 6 . We can see from the Table 6 that HTL achieves significantly better performance than both the traditional machine learning methods and transfer learning methods. While analysing from the vertical direction of the Table 6, none of the adopted traditional machine learning methods can obtain a satisfactory prediction accurate on our datasets. For the four adopted feature inputs, the combination of feature selection with PN have a better performance than others especially when the target training samples reach to 30.
Table 6

Classification accuracy on virus datasets

Methods10 samples20 samples30 samples
ACC ± stdACC ± stdACC ± std
Primitive Feature
SVM22.6%±5.624.4%±4.828.9%±5.3
XGBoost21.3%±2.428.6%±2.629.1%±5.5
tar-SVM28.9%±4.136.7%±4.539.3%±6.2
tar-XGBoost34.4%±3.931.5%±4.2
Feature SelectionSVM25.2%±1.926.7%±3.324.5%±3.6
XGBoost24.7%±7.428.6%±5.832.3%±5.7
tar-SVM29.2%±4.528.1%±5.423.3%±4.7
tar-XGBoost32.2%±6.535.4%±7.6
PN ModelSVM19.0%±2.125.5%±4.627.2%±5.5
XGBoost17.1%±3.929.0%±5.923.3%±6.6
tar-SVM10.3%±5.922.7%±5.627.8%±4.6
tar-XGBoost19.5%±5.525.1%±7.0
Feature Selection + PN ModelSVM15.4%±4.422.0%±4.729.6%±6.4
XGBoost19.2%±2.933.9%±6.336.3%±7.7
tar-SVM27.1%±5.026.9%±6.633.9%±5.5
tar-XGBoost30.6%±5.338.4%±9.0
TrAdaBoost30.1%±2.335.4%±3.344.7%±3.6
MTLF34.7%±3.936.1%±4.548.1%±4.2
CODA28.3%±5.930.4%±5.135.1%±4.8
Our Method:HTL36.5%±4.340.7%±5.351.5%±3.2
Classification accuracy on virus datasets The genome feature quantification method calculates the intricate information for the sequence, which will lead to the information redundancy when training the classifiers. In the cases of small amounts of target training datasets, traditional methods don’t have the abilities to build a highly generalized models and show a poor performance on target testing datasets. Note that, when there is only one training sample for each class, XGBoost fails to build a model for classification. For each leaf node contains different classes of samples except the last classification node, makes it hard to select a low entropy subset. Furthermore, the feature redundancy further aggravates the emergence of this phenomenon. The results of directly training the classifiers by source domain cannot achieve a better performance than only using the target training sets, even worse than the latter. This evidence means the source domain will cause a negative impact on learning the classifier for target testing sets. However, by conducting the transfer learning methods to decrease the variances among virus species, the accuracy can have a distinct improvement. MTLF achieves a closer result with HTL, which means the domain adaptation methods have a strong ability to decrease the variances and alleviate the feature redundancy. Compared from the horizontal of the Table 6, with the increase of target domain training data set, the accuracy of the algorithm increases gradually, especially for MTLF, which means the target training sets can bring more information to predict the hosts for testing sets. To further demonstrate the performance of HTL, we select the Top-5 methods in each experiment group and compare their ROC curve with HTL. The experiment results are reported in Fig. 4 . When the training samples in the target domain is lower than 30, only using the target training sets is insufficient to build a robust model. By leveraging the transfer learning methods, the source domain can be regarded as a better supplementary and improve the generalization on target domain. In our experiment, both TrAdaBoost and MTLF perform better than most of the traditional machine learning methods. Although, transfer learning can effectively decrease the discrepancy between source and target domain, class imbalance problem still constrains the conventional methods to get a better performance. We can discover from the Fig. 4 that the AUC of HTL always appears the best performance than other compared methods. Furthermore, the AUC value of our method is between 0.7 and 0.9, which indicates the prediction results of HTL has a high confidence.
Fig. 4

The ROC curve of the different amounts of target training datasets. In each experiment group, Top-5 best performance methods are selected to compare the AUC with HTL.

The ROC curve of the different amounts of target training datasets. In each experiment group, Top-5 best performance methods are selected to compare the AUC with HTL. In general, by decreasing the variance between target virus species and related virus species, our method outperforms than other compared traditional machine learning methods and transfer learning methods. The ensemble learning model guarantees the generalization and the robustness for the target testing set. What’s more, the final ensemble strategy decreases the influence by the resample method which can be demonstrated in the standard deviation.

Prediction results for COVID-19

After evaluating the effectiveness and performance of the HTL on virus datasets, we conduct our method to provide the prediction results for COVID-19. All 307 complete genome sequences and 1 refseq of COVID-19 are pre-processed with the same methods described before. We keep the same experiment settings in Section B, but change the target training set into all coronavirus datasets. The experiment results of HTL get the similar prediction results like most mainstreams for COVID-19 that the reservoir hosts of COVID-19 may come from bat. More specifically, the reservoir host of the COVID-19 is Petrobat. While for the complete genome sequences, there are 303 prediction results for Petrobat and 4 results for Vespbat. As shown in Fig. 5 (a), we transform the output probabilities of HTL into a heat map to visualize the prediction information for the complete genome sequences of COVID-19. All the prediction results for Petrobat get high probabilities, which reveals the prediction results for Petrobat have high confidence. We can obviously find in Fig. 5 (c) that the Petrobat gets the highest mean confidence nearly 89.3%, the Rodent and the Vespbat get 3.6% and 3.2% prediction confidence for average. Looking back to the prediction results of refseq, the results are similar with the average probabilities of complete genome sequences, which the Petrobat gets 90.1%, the Rodent and the Vespbat get 3.7% and 2.1% respectively. The complete genome sequences contain many similar samples of COVID-19, which means Petrobat may be the most probable hosts.
Fig. 5

The statistics results of the prediction results and the heat map of the hosts probabilities for COVID-19. In (a), the depth of the color represents the hosts probability of each virus; (b) illustrates the prediction results of complete genome sequences of COVID-19; (c) records the Top-3 prediction confidence of refseq and average of complete genome sequences

The statistics results of the prediction results and the heat map of the hosts probabilities for COVID-19. In (a), the depth of the color represents the hosts probability of each virus; (b) illustrates the prediction results of complete genome sequences of COVID-19; (c) records the Top-3 prediction confidence of refseq and average of complete genome sequences To make our prediction results convincible, we use Blastsn to search the Top-5 closely related viruses from our datasets and show the results in Table 7 . Compared with the refseq of COVID-19 (NC_045512.2), the SARS coronavirus has the highest sequence identity of 80.3% and other Bat coronavirus with 78% and 67.4% sequence identities. Beyond our expectations, there are two sequences of coronavirus with the hosts of Neoaves have the high identities with COVID-19. These two sequences are submitted in [59], which demonstrates the avian and mammalian coronaviruses have similar genome characteristics and structures. In their work, the bats and birds, the warm-blooded flying vertebrates, are ideal hosts for the coronavirus gene source. This conclusion presents another conjecture that whether avian could be the intermediate host or have some relationships with the COVID-19, it needs further studies. However, judging from the prediction results of our method, the Neoaves only has 0.3% probability to be the hosts for COVID-19.
Table 7

The sequence identities with the refseq of COVID-19 (NC_045512.2).

GenBank accessionVirus speciesStrainIdentity (%)Hosts
NC_004718.3CoronavirusSARS coronavirus80.3%Pterobat
NC_014470.1CoronavirusBat coronavirus78.0%Pterobat
NC_016991.1CoronavirusWhite-eye coronavirus HKU1668.9%Neoaves
NC_016992.1CoronavirusSparrow coronavirus HKU1767.8%Neoaves
NC_030886.1CoronavirusRousettus bat coronavirus67.4%Pterobat
The sequence identities with the refseq of COVID-19 (NC_045512.2). Phylogenetic analysis of COVID-19 revealed that the hosts of the closely related genomes can also support our prediction results. For a better visualization, we only adopt 20 refseqes of the closely related sequence in Fig. 6 . All the bootstrap values are 100, which means the credibility when building the branches of the tree. We can see form the Fig. 6., the SARS coronavirus (NC_004718.3) [60] and the Bat coronavirus (NC_014470.1) [61] have the closest clades with COVID-19, the Rousettus bat coronavirus (NC_030886.1)[62] forms one clade close to the above three viruses. The hosts in the branch which constructed by these three sequences all comes from Pterobat. Thus, according to the phylogenetic tree, COVID-19 has the highest homology with the SARS coronavirus (NC_004718.3) which reveals the host of COVID-19 may come from Pterobat. The evidence from phylogenetic analyses and the sequence identities analyses have the high consistencies with the prediction result from our method, which indicates the Petrobat is the most probable hosts for COVID-19.
Fig. 6

The phylogenetic tree constructed by the refseq of COVID-19 (NC_045512.2) and Top-20 related sequences.

The phylogenetic tree constructed by the refseq of COVID-19 (NC_045512.2) and Top-20 related sequences.

Discussion

As demonstrated in the reported experiments, our approach obtains better classification results not only for UCI benchmarks, but also the viral datasets. It provides a promising yet easy-to-use method for addressing the origin tracing problems for the newly discovered virus. Based on both the overall experimental results and the virology analyses, we summarize our approach as follows. Traditional origin tracing methods consume a large number of times and labours, which delay the isolation for the hosts animals and the pathology research of the virus. Machine learning provides a precise way to discover the latent information from annotated genome sequences and make the prediction for specific tasks. According to our experiments, when predicting the hosts for newly discovered virus, the classification model trained from the same viral species will get more reliable results. However, under the conditions of limited annotated virus and class-imbalance problem, directly train the classifiers from other viral datasets will lose their efficiencies by the discrepancies among different viral species. By conducting a combination of transfer learning and ensemble learning, the fundamental weakness of traditional methods can be well solved. From Section 4, the simulation results indicate that our method provides a promising way to improve the accuracy and generalization on UCI medical benchmarks, especially the virus datasets. Although HTL achieves a better performance than all the compared methods, some problems still need to be further studied: 1) For the variances exist among the different virus species, confusing the related virus as one domain group is worthy of discussion. Evaluating the relationships among different virus species and separating the related domain into multiple domains seems more convinced. 2) Though the JDA can significantly reduce the variances among different virus species, the huge calculated quantity consumes enormous computing resources. The large-scale genome datasets with high dimensions become hard to obtain the best transform matrix. As PN model does, the evaluating for the variances among different virus species may dig out from the perspective the sequence genomes. 3) In this paper, we adopt the CDS of the complete sequences for quantification. While for each CDS, some coding regions may have the low relationship with the hosts information, but some regions have more discriminative information which will benefit for hosts prediction. So, finding the CDS regions related to the host is of great significance for improving the accuracy of the algorithm. And other sequence quantification methods can be used for pre-processing in the future work. To predict the hosts for the newly discovered virus, we provide a complete work flow from collecting the virus sequences, genome sequence processing, feature quantification and construction the classification model. The final output of HTL gets a high prediction convince that the Petrobat is the hosts for COVID-19. The sequence identity and phylogenetic analysis of the refseq of COVID-19 get the same results with our method, which convince the prediction results from HTL. However, despite the prediction results of the Petrobat, several facts suggest that another animal is acting as an intermediate host between bats and humans. No traces of bats were found in the Huanan seafood market where the COVID-19 first broke out. The latest research [63] reveal that Malayan pangolins could be the intermediate host for COVID-19, for they detected and found the associated coronavirus in pangolin samples, belonging to two subgenus of the COVID-19, in which a receptor binding domain has a closely relationship with COVID-19. In [64], the researchers found mink viruses have a closer infectivity pattern to COVID-19 by deep learning method. These consequences of infectivity pattern analysis illustrate that bat and mink could be the candidate reservoirs of COVID-19. From the above, on the basis of the prediction results from HTL, it seems likely that the Petrobat may be the initial hosts of COVID-19, and might have another intermediate host, like pangolins, mink or other wild animal, to transmit the COVID-19 to humans.

Conclusion

In this paper, we present a workflow of predicting the hosts for the newly discovered virus, including data collection, data processing, feature quantification and construct the classification model. We construct a virus benchmarks which can be used to proceeding further studies for virus. Considering the limited training samples and class-imbalance problems, we extend transfer learning and ensemble learning to construct a hybrid transfer learning model for virus hosts prediction. We achieve a promising simulation results with other compared methods for UCI benchmarks and virus datasets. Finally, the prediction results reveal that the Petrobat might be the reservoir host of COVID-19, which provides a research interests for virologists to do more investigates.

CRediT authorship contribution statement

Yun Yang: Conceptualization, Funding acquisition, Project administration. Jing Guo: Software, Methodology, Writing - original draft. Pei Wang: Supervision, Writing - review & editing. Yaowei Wang: Formal analysis. Minghao Yu: Data curation, Visualization. Xiang Wang: Investigation. Po Yang: Validation. Liang Sun: Funding acquisition, Resources.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  33 in total

1.  Domain adaptation via transfer component analysis.

Authors:  Sinno Jialin Pan; Ivor W Tsang; James T Kwok; Qiang Yang
Journal:  IEEE Trans Neural Netw       Date:  2010-11-18

2.  Alignment-free sequence comparison (I): statistics and power.

Authors:  Gesine Reinert; David Chew; Fengzhu Sun; Michael S Waterman
Journal:  J Comput Biol       Date:  2009-12       Impact factor: 1.479

3.  Genomic characterization of severe acute respiratory syndrome-related coronavirus in European bats and classification of coronaviruses based on partial RNA-dependent RNA polymerase gene sequences.

Authors:  Jan Felix Drexler; Florian Gloza-Rausch; Jörg Glende; Victor Max Corman; Doreen Muth; Matthias Goettsche; Antje Seebens; Matthias Niedrig; Susanne Pfefferle; Stoian Yordanov; Lyubomir Zhelyazkov; Uwe Hermanns; Peter Vallo; Alexander Lukashev; Marcel Alexander Müller; Hongkui Deng; Georg Herrler; Christian Drosten
Journal:  J Virol       Date:  2010-08-04       Impact factor: 5.103

4.  Codon Pair Bias Is a Direct Consequence of Dinucleotide Bias.

Authors:  Dusan Kunec; Nikolaus Osterrieder
Journal:  Cell Rep       Date:  2015-12-24       Impact factor: 9.423

Review 5.  Ecological origins of novel human pathogens.

Authors:  Mark Woolhouse; Eleanor Gaunt
Journal:  Crit Rev Microbiol       Date:  2007       Impact factor: 7.624

6.  Comparative analysis estimates the relative frequencies of co-divergence and cross-species transmission within viral families.

Authors:  Jemma L Geoghegan; Sebastián Duchêne; Edward C Holmes
Journal:  PLoS Pathog       Date:  2017-02-08       Impact factor: 6.823

7.  Prediction of virus-host infectious association by supervised learning methods.

Authors:  Mengge Zhang; Lianping Yang; Jie Ren; Nathan A Ahlgren; Jed A Fuhrman; Fengzhu Sun
Journal:  BMC Bioinformatics       Date:  2017-03-14       Impact factor: 3.169

8.  Middle East respiratory syndrome coronavirus neutralising serum antibodies in dromedary camels: a comparative serological study.

Authors:  Chantal B E M Reusken; Bart L Haagmans; Marcel A Müller; Carlos Gutierrez; Gert-Jan Godeke; Benjamin Meyer; Doreen Muth; V Stalin Raj; Laura Smits-De Vries; Victor M Corman; Jan-Felix Drexler; Saskia L Smits; Yasmin E El Tahir; Rita De Sousa; Janko van Beek; Norbert Nowotny; Kees van Maanen; Ezequiel Hidalgo-Hermoso; Berend-Jan Bosch; Peter Rottier; Albert Osterhaus; Christian Gortázar-Schmidt; Christian Drosten; Marion P G Koopmans
Journal:  Lancet Infect Dis       Date:  2013-08-09       Impact factor: 25.071

9.  A familial cluster of pneumonia associated with the 2019 novel coronavirus indicating person-to-person transmission: a study of a family cluster.

Authors:  Jasper Fuk-Woo Chan; Shuofeng Yuan; Kin-Hang Kok; Kelvin Kai-Wang To; Hin Chu; Jin Yang; Fanfan Xing; Jieling Liu; Cyril Chik-Yan Yip; Rosana Wing-Shan Poon; Hoi-Wah Tsoi; Simon Kam-Fai Lo; Kwok-Hung Chan; Vincent Kwok-Man Poon; Wan-Mui Chan; Jonathan Daniel Ip; Jian-Piao Cai; Vincent Chi-Chung Cheng; Honglin Chen; Christopher Kim-Ming Hui; Kwok-Yung Yuen
Journal:  Lancet       Date:  2020-01-24       Impact factor: 79.321

10.  MAFFT online service: multiple sequence alignment, interactive sequence choice and visualization.

Authors:  Kazutaka Katoh; John Rozewicki; Kazunori D Yamada
Journal:  Brief Bioinform       Date:  2019-07-19       Impact factor: 11.622

View more
  1 in total

1.  Novel informatics approaches to COVID-19 Research: From methods to applications.

Authors:  Hua Xu; David L Buckeridge; Fei Wang; Peter Tarczy-Hornoch
Journal:  J Biomed Inform       Date:  2022-02-16       Impact factor: 8.000

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.