Literature DB >> 32946529

Short k-mer abundance profiles yield robust machine learning features and accurate classifiers for RNA viruses.

Md Nafis Ul Alam1, Umar Faruq Chowdhury1.   

Abstract

High-throughput sequencing technologies have greatly enabled the study of genomics, transcriptomics and metagenomics. Automated annotation and classification of the vast amounts of generated sequence data has become paramount for facilitating biological sciences. Genomes of viruses can be radically different from all life, both in terms of molecular structure and primary sequence. Alignment-based and profile-based searches are commonly employed for characterization of assembled viral contigs from high-throughput sequencing data. Recent attempts have highlighted the use of machine learning models for the task, but these models rely entirely on DNA genomes and owing to the intrinsic genomic complexity of viruses, RNA viruses have gone completely overlooked. Here, we present a novel short k-mer based sequence scoring method that generates robust sequence information for training machine learning classifiers. We trained 18 classifiers for the task of distinguishing viral RNA from human transcripts. We challenged our models with very stringent testing protocols across different species and evaluated performance against BLASTn, BLASTx and HMMER3 searches. For clean sequence data retrieved from curated databases, our models display near perfect accuracy, outperforming all similar attempts previously reported. On de novo assemblies of raw RNA-Seq data from cells subjected to Ebola virus, the area under the ROC curve varied from 0.6 to 0.86 depending on the software used for assembly. Our classifier was able to properly classify the majority of the false hits generated by BLAST and HMMER3 searches on the same data. The outstanding performance metrics of our model lays the groundwork for robust machine learning methods for the automated annotation of sequence data.

Entities:  

Mesh:

Year:  2020        PMID: 32946529      PMCID: PMC7500682          DOI: 10.1371/journal.pone.0239381

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Viruses are numerous [1] and only a handful have been thoroughly characterized thus far [2]. As we have entered and slowly progress through the age of automated, high-throughput genomics, generation of sequence data itself is no longer of much concern, but rather accurate annotation of these sequences is the bottleneck in the expansion of biological knowledge. Remnants of viral genetics are continually integrated into host DNA where they reside as proviruses. Additionally, a copious number of viruses serendipitously accommodate our bodies as part of our microbiome (often called the virome [3]), deftly contaminate our working samples in the laboratory, and profoundly enrich the ecology of our environment. The scientific and pathological significance of viruses is astounding. The relevance of automated and reliable identification and characterization of highly divergent and novel viruses in laboratory samples and metagenomic data is now greater than ever. The bacterial 16sRNA gene tremendously aids the study of bacterial phylogeny [4]. The lack of an omnipresent gene or genome segment that can be ubiquitously used to extract phylogenetically relevant data about viruses, complicates the study of virus evolution [5, 6]. There are additional levels to the structural complexity of viral genotypes. Viral genomes stand out from those of all other genomic entities as they can be either single, double or gapped DNA or RNA molecules with positive, negative or ambi-sense mechanism of genomic encoding, whereas all known realms of life exclusively rely on DNA as the genetic material. These categories of classification are built on top of the original Baltimore scheme [7] that sought to characterize viral genomes with regard to expression of genetic information [8, 9]. In the search for viruses in metagenomic data, researchers have noted that the usefulness of homology-based search tools are quickly exhausted [6, 10]. Most assembled contigs, likely to be from viral origins, are short and fragmented with no guarantee of containing coding regions that some of these tools rely on [11]. A number of software programs have been developed for the purpose of identifying viral sequences that have integrated into host genomes [12-15]. All machine learning models built for identification and characterization of viral sequences from cell samples or metagenomic data thus far rely on DNA sequences [16, 17]. RNA viruses comprise a major group having great clinical and scientific importance [18]. Presently made even more evident by the global pandemic caused by the 2019 novel SARS Coronavirus [19]. It has been demonstrated that RNA-Seq data can be a very promising avenue for improving knowledge on RNA viruses when leveraged by tactful algorithms [20]. Here, we present a novel short k-mer based scoring function that can be applied to sequences of different types to achieve impressive discriminatory capacity from classic machine learning models. With our scoring function, we were able to calculate the profiles of short genetic elements (k-mers) for RNA virus genomes and human RNA transcripts. We trained a number of classical machine learning classifiers on our sequence-derived numerical data in an attempt to classify assembled RNA sequences of unknown origin as either viral or cellular RNA. We tested the models rigorously using stringent cutoff parameters and many variable test sets including noisy de novo assembly data of cells cultured with Ebola virus. Furthermore, we demonstrated the flexibility of this approach by attempting to classify an RNA virus genome sequence from NCBI as either a positive-strand (ssRNA(+)) or a negative-strand (ssRNA(-)) RNA virus. We evaluated the performance of our selected model against nucleotide blast, protein blast and protein family HMMER3 search. In spite of the stiff testing protocols, our k-mer based numerical features stood out to be informative and robust datapoints for the classification of biological sequences by machine learning algorithms.

Results

Feature selection and scaling

A total of 5460 features were contrived using the sequence scoring formula for k-mer elements of length 1 to 6. The feature columns were tested for data normality by several methods. Q-Q plots of the data columns were suggestive of a normal distribution (not shown but can be found on git repository), but more robust statistical normality tests such as the Kolmogorov-Smirnov test and Shapiro-Wilk test strongly delineated a non-normal distribution for all feature columns. Our features were therefore not suitable for univariate statistical analysis for feature selection. The in-built, tree-based feature importance values of the scikit-learn library derived from our trained Random Forest classifier were applied for feature selection on two levels. Similarly, the scikit-learn optimized Lasso model applied to our trained Logistic Regression classifier with L1 penalty was also used for feature selection on two succeeding levels. Using a combination of both our tree-based selections and Lasso, we obtained a total of 6 levels of feature importance categories (4 from the trees and Lasso themselves and 2 more for each method leading into the other on a second level), each assigning a value of 1 to selected features and 0 to omitted ones (Fig 1). The top 3 feature sets ranked by importance from these 6 combined levels of selection are presented in Table 1. The complete table can be viewed in S1 Table. To simplify the model, make it more computationally scalable and remove noisy features, we selected 68 features that scored above 4 in our feature selection table. For comparison, models were trained on the 194 Features scoring above 3 as well, which includes the entirety of the features in Table 1.
Fig 1

Feature selection algorithms applied on succeeding levels to scale the model.

Tree-based and Lasso regularization-based feature selection applied on two levels successively yielded 6 total sets of features on the corresponding criteria of selection.

Table 1

Top 3 k-meric features sets when ranked by feature importance instated by our feature selection flow-chart.

Summed Feature ImportanceFeatures
6TTGACG, TAGCGT, ATACGC, CAAT, AAAAA, CCAAT, CATACG, TGAT, CGATA, AAAAAA, TCAATC, CAATTG, CGGTAA, GTTGA, GTTGAC, CGATAG, CCAATT, GGTTGA, ACCAAT, TCAA, TTGTCG, CGTCAA, CGGTTA, CAATCG, CGTTGA, TGTCGA, CGCAAT, CAAC, GTCGAA, GGTA, GTTG, GGTT
5CAATC, TCCAAT, GCAATC, TTGAC, CGTAGT, ACGGTT, TTGA, GTTGGT, CATCAA, ATACGG, ATCGTA, CGATAA, CGATTA, GTTGAT, CGAT, CGATC, CGGTT, GCGTAC, CGCGTT, TTGCG, CGTCGA, CAATCT, ATCAAC, ATTGG, ATAGCG, GATC, TTGCGC, GGTTG, TCGGTT, GTCAAT, ATCAAT, TTCGAC, CCGTTA, CCGATA, TCGA, GCGTTA
4ATCATA, ACAATC, CGTA, TCGT, ATGGTA, GGTAGT, TGGT, CTGT, GTCGA, AGCTG, GTCAA, AAATAA, TCCTG, CAGC, TATCCG, CTGGA, CAATT, TCGGTA, AGGAA, ATCAA, ATAACG, TCCAAC, AATTGG, AAATG, TGGTTC, AACC, TTGATG, GGCGTA, TGGTT, CGTTGC, GTCGAC, TCAAC, CAGA, TCAATT, CAGAG, TCAAT, AGGTTG, TCAACC, CGCGTA, ATCCAA, GGTACG, GTCATA, ATTGGT, GCATAC, GGTATG, GATTGG, GTACGC, AACCGA, TCGCAA, TTTTTA, ACATCG, GCGTAA, TTTTTT, ACCGGT, GTTGAG, TAACAC, ATAGGG, CGCAA, CCGCAA, CGGTTG, TATGGT, AACGGT, TAGGGT, CTAACG, CGGTA, GTTGT, GGTTC, ACGATA, TTCGCA, CTCGAT, TCGAGT, AAAATG, TCCAT, GTCAAA, AATAAA, TTGGCG, GTTGC, GGTCAA, CTGTA, ACGCAA

The summed feature score is obtained by summing the total number of times the feature appears in the six filtered feature sets we obtained from our selection algorithm. The remainder of the table can be found in S1 Table.

Feature selection algorithms applied on succeeding levels to scale the model.

Tree-based and Lasso regularization-based feature selection applied on two levels successively yielded 6 total sets of features on the corresponding criteria of selection. The summed feature score is obtained by summing the total number of times the feature appears in the six filtered feature sets we obtained from our selection algorithm. The remainder of the table can be found in S1 Table.

Performance of different classification algorithms

A total of six different types of classifier models were built- Logistic Regression Classifiers, Linear Support Vector Classifiers, Kernel based Support Vector Classifiers, Decision Tree Classifiers, Random Forest Classifiers and Gaussian Naïve Bayes Classifiers. For performance evaluation, all classifiers were trained thrice using 5460, 194 and 68 features accordingly. The train-test split was 25% and for cross-validation we applied 10-fold cross-validation on 9:1 fractions of the training set. The train-test split on each iteration was random. However, multiple iterations with new splits were carried out for all training events to ameliorate any random split biases. For NCBI-viruses and human transcriptome data, all models displayed robust performance with high discriminatory capacity among the two classes (Table 2). The effects of feature reduction on the models can be viewed in the AUROC curves in Fig 2 and the S1 Fig. The poorest performing model was the Gaussian Naïve Bayes classifier trained on all 5460 features. The macro-averaged f1-score for the model was 88% and the binary f1-score was 81.83%. Among the others, the decision tree models with 5460 and 194 features and the Naïve Bayes model trained with 68 features were the only models to obtain binary f1-scores below 95%. All other models displayed resounding accuracy on both their cross-validation and test sets.
Table 2

Performance metrics of 18 models fitted and the selected model.

Part 1: Performance of 18 fitted models on cross-validation and test sets
ModelFeaturesCorrect/TPFalse PositivesTrue NegativesFalse Negativescv_bina meancv_bin stdecv_macb meancv_mac stdef1bincf1macd
Logistic Regression546010494384410.9975590.0016680.9984570.0010540.9976220.998486
Logistic Regression19410200387800.9991940.0014850.9994890.00094211
Logistic Regression681025103851120.9837590.0056040.9897270.0035490.9893820.993267
Linear SVM546010102388510.9958250.0007840.9973510.0004980.9985170.999066
Linear SVM19410810381700.998850.0012810.9992750.00080811
Linear SVM6895413392290.9852450.0051370.9906080.0032670.9886010.992902
RBF Kernel SVM54609951389750.9962930.001780.997650.0011270.9969940.998112
RBF Kernel SVM19410490384810.996560.0015520.9978290.0009780.9995240.999697
RBF Kernel SVM6810120388330.9975720.0016610.9984630.0010510.998520.999067
Decision Tree5460964513817660.9398090.0078240.9619540.0059770.9427870.963846
Decision Tree1941006673775500.9473260.0062550.9664030.0043450.9450450.964892
Decision Tree681013543786450.9433880.0093830.9640250.0067270.9534120.970253
Random Forest5460103203847190.9894370.0035180.9919850.0024450.9908790.994208
Random Forest19497313911130.9912930.0035750.9944840.0020690.9928570.995535
Random Forest6898733886220.9889480.0035250.9932170.0016240.9874940.992144
Naïve Bayes54609733773493550.814590.0139530.8772610.0095960.8183350.880049
Naïve Bayes194997233867110.9821570.0054340.9886550.0034680.9832350.989429
Naïve Bayes68970953812210.9424880.0125540.9630990.00810.943580.965126
Part 2: Performance of the Selected Model and our Random Forest Model on Divergent Data
ModelFeaturesTrue PositivesFalse PositivesTrue NegativesFalse Negativesf1binf1mac
rbf-SVM6868185592112596770.828309890.897644
RF6868181432113197180.82868890.898311

Part 1 lists the performance metrics on the test splits of the training data and Part 2 lists the performance on the divergent data comprised of viruses of other genotypes and mouse RNA transcripts.

a binary f1 score in cross-validation sets.

b macro averaged f1 score in cross-validation sets.

c binary f1 scores in test sets.

d macro f1 scores in test sets.

e standard deviation.

Fig 2

AUROC for 18 fitted models trained while varying the number of features selected.

(a) Performance of all classifiers with all features. (b) Performance with 194 features after first round of selection. (c) Performance with 68 selected features. The weakest performing model was the Gaussian Naïve Bayes model displaying 0.96 AUC. No significant decline in model performance was noted when feature numbers were shrunk from 194 to 68.

AUROC for 18 fitted models trained while varying the number of features selected.

(a) Performance of all classifiers with all features. (b) Performance with 194 features after first round of selection. (c) Performance with 68 selected features. The weakest performing model was the Gaussian Naïve Bayes model displaying 0.96 AUC. No significant decline in model performance was noted when feature numbers were shrunk from 194 to 68. Part 1 lists the performance metrics on the test splits of the training data and Part 2 lists the performance on the divergent data comprised of viruses of other genotypes and mouse RNA transcripts. a binary f1 score in cross-validation sets. b macro averaged f1 score in cross-validation sets. c binary f1 scores in test sets. d macro f1 scores in test sets. e standard deviation.

Model selection

The best performing models were the logistic regression model with 194 features, the linear SVM model with 194 features and Kernel SVM models with 194 and 68 features, all displaying perfect classification accuracy on many random train-split-evaluate iterations (Table 2). The virus data set consisted of a total of 56 ssRNA(+) and ssRNA(-) families. It is common practice with such classifiers to carry out a separate cross-validation where the classifier is trained on the data set leaving out viruses belonging to a family to later use that family of viruses as an exclusive test set for evaluation [16, 21, 22]. We performed this leave-one-out cross-validation procedure for all 18 built classifiers. The complete results can be viewed visually on Fig 3 and numerically on S2 Table. Out of the four best-performing models previously mentioned, the logistic regression model with 194 features and the Kernel SVM model with 68 features displayed the best evaluation scores on the leave-one-family-out cross-validation test. Because the performance of the 194-feature logistic regression model is only slightly improved than the 68 feature Kernel SVM model, and because it is much more computationally demanding to calculate 194 feature scores than to calculate 68, we select our 68 feature Kernel SVM model as the final model for ultimate evaluation on real data.
Fig 3

Leave-one-family-out cross-validation results heatmap.

For every RNA virus family on the horizontal axis, there is a performance score for each model in the vertical axis. Other than the Decision Tree and Naïve Bayes models, performance was reasonably uniform all throughout the map.

Leave-one-family-out cross-validation results heatmap.

For every RNA virus family on the horizontal axis, there is a performance score for each model in the vertical axis. Other than the Decision Tree and Naïve Bayes models, performance was reasonably uniform all throughout the map.

Ability to generalize across species

The performance of the classifiers at this point far exceeded the outcomes of other classifiers built for the same task in previous studies [11, 16, 17, 21]. Given that the focus of these previous classifier were genomic DNA data, in order to test whether this performance amplification was entirely attributable to the greater homogeneity RNA sequences and their shorter lengths compared to genomic DNA sequences, we studied the performance of the classifier on systematically divergent data sets. We scored the sequences of total 77 other families (S3 Table) of viruses having DNA, cRNA and dsRNA genomes of similar length, and 11351 putative-expressed RNA sequences of similar length from the Mus musculus whole transcriptome (S3 Table). The 68 feature rbf-SVM model was able to accurately classify 13114 of 13883 test sequences with 1855 true positives, 92 false positives, 11259 true negatives and 677 false negatives with an AUROC of 0.95 (Fig 4). The f1-macro score was 0.89 and the f1-binary score was 0.83 (Table 2 Part 2). Performance of this magnitude on novel divergent data concerns one to be critical of the evaluation procedure and the presence of possible internal biases within the data set. However, this leaves out any possibility of overfitting of the model itself. Tree-ensemble models are known to stand out to be good discriminators in biological data [15, 23], particularly because of their propensity against overfitting. Our 68-feature random forest model displayed impressive classification accuracy and an AUROC of 0.96 on this divergent test set (Fig 4).
Fig 4

Performance of the rbf-SVM68 and RF68 models on divergent data set.

To assess the capacity of our models to generalize to new examples, we attempted to confound the model with sequence data from viruses of other genotypes (having other than plus and minus RNA genomes) and RNA transcripts from mice (Listed in S3 Table). The robustness of the model performance metrics was still found to be intact.

Performance of the rbf-SVM68 and RF68 models on divergent data set.

To assess the capacity of our models to generalize to new examples, we attempted to confound the model with sequence data from viruses of other genotypes (having other than plus and minus RNA genomes) and RNA transcripts from mice (Listed in S3 Table). The robustness of the model performance metrics was still found to be intact.

Evaluation of performance against nucleotide BLAST

The unfaltering performance of the model in tests carried out thus far led us to suspect systematic biases present in the available RNA sequence data and however stringent, question our testing protocols. For final evaluation it was indispensable that the model be put to the test against noisy real-word RNA-seq data lacking the meticulous curation that our previous data sets had. We collected de novo transcriptome assembly data for cultured human (H. sapiens), mice (M. musculus) and human cell lines incubated with an ssRNA(-) virus, Ebola virus from a previous cross-study on de novo assembly of RNA-seq data [24]. The best current standard for sequence-based taxonomic identification is the BLAST alignment algorithm [25]. Many correlated evolutionary forces govern both the conservation and divergence of sequence data, and as an artifact of these forces, even across the most divergent species, many gene families are shared and homologous sequences are found. This compromises the ability of sequence alignment protocols to properly label sequences that might have originated from highly divergent sources that are yet to be characterized and annotated in the current data repositories. In the Homo sapiens transcriptome assembly, carried out on GM12878 human gammaherpes virus 4 transformed B-lymphocyte cell lines, nucleotide blast query against our sequence database of ssRNA(+) and ssRNA(-) viruses yielded 135 false positives (S4 Table). These false positives were attributed to sequence regions of bovine viral diarrhea virus, giraffe pestivirus and a betacoronavirus that shared homology with some human ubiquitin gene transcripts and heat-shock protein 40 gene transcripts. Our 68 feature rbf kernel SVM model correctly classified 101 of these false positives giving us an accuracy of 75% on the unary data. When feeding the entirety of the human cell line assembly into our classification pipeline we achieved an accuracy of 63% for transcripts of all sizes and a much-improved value of 95% when taking transcripts of the same length range as our training data (Table 3). Similarly, nucleotide blast had given 130 false positives on the Mus musculus transcriptome assemblies (S5 Table), whereas our model accurate classified 93 examples with 72% accuracy on the unary class.
Table 3

Performance metrics of rbf-SVM68 model on real data.

data setTranscript/Software FiltersTotal Number of TranscriptsTrue PositivesFalse PositivesTrue NegativesFalse Negativesf1binaryf1macroAUROCUnary Accuracya
Unary Homo Sapiens Transcriptome AssemblyAll6303331N/A23270823976249N/AN/A0.38681N/A63%
Unary Homo Sapiens Transcriptome AssemblyLength Filtered146382N/A7067139315N/AN/A0.487632N/A95%
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome AssemblyAll1179617531833184954062218280.030890.4326440.799N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome AssemblyBoth Class Length Filtered142803493799138906490.0248350.5055870.776N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome AssemblyOnly Human Length Filtered1498515318379913890618280.6840.8170740.951N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblybridger83744162025263452240.0015760.4319060.6N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblyidba-tran78116301986058182440.0030050.428470.604N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblysoap-trans-default13171119441478753870.00860.3997270.793N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblyrna-spades206133372715431340891290.0102740.399690.775N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblyspades938144488256456255911220.2511260.5374470.843N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblytrinity97849322293674839420.0027780.4348460.611N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblyoases178498180347211433732240.0101970.4507840.691N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblytrans-abyss213005171682681443462200.0049690.4066110.583N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblybinpacker12291780011468160.0168670.4912560.657N/A
Homo Sapiens Cell Incubate with Ebola Virus Transcriptome Assemblyshannon844563236776077600.002530.4186110.862N/A

a Unary accuracy is used to assess performance for test sets containing only one class of examples and thus rendering f1 scores uninformative.

a Unary accuracy is used to assess performance for test sets containing only one class of examples and thus rendering f1 scores uninformative. In the 23-hour Ebola virus with cells incubate, local nucleotide BLAST against our in-house RNA virus database had 10675 unique hits. Upon a second local nucleotide blast of these 10675 results against our human transcriptome training data, we obtained 3483 unique hits. Since we previously found only three virus sequences that originally had blastn hits against the whole human transcriptome, we labelled these second hits as belonging to the human class. This left us with a test set of 7146 positive class transcript examples for RNA viruses and 1172471 negative class examples of human transcripts in the whole 23-hour cell incubate with Ebola virus. The area under the ROC curve was 0.807 and 0.777 for all transcript sets and transcript sets of lengths resembling the training data respectively (Fig 5). The AUC improves more drastically when only human transcripts are filtered leaving all the virus transcripts still in the test set, owing to the minimized skew of the separate classes (Fig 5). The classification metrics from all instances are presented in Table 3.
Fig 5

Performance on RNA-Seq assembly data from human cells cultured with Ebola virus.

The model performed the best with AUC of 0.95 when human transcripts below the training range length were left out. Filtering the short virus transcripts out as well as the human transcripts resulted in a drop in the curve owing to the skew in the number of examples belonging to the human class.

Performance on RNA-Seq assembly data from human cells cultured with Ebola virus.

The model performed the best with AUC of 0.95 when human transcripts below the training range length were left out. Filtering the short virus transcripts out as well as the human transcripts resulted in a drop in the curve owing to the skew in the number of examples belonging to the human class. Given the reported generalization prowess of the classifier from the original test set split and leave-one-out cross-validation, the classifier was expected to perform better on the real transcript data. Discordant with the previously observed success rate of the model, it performed better on the human samples compared to those of viral origin (Table 3). To further investigate the despondence of the model, we assessed the performance of the 10 assembly software separately. This revealed that performance varied significantly between the different assembly software that joined the contigs (Fig 6). Table 3 lists the performance metrics on the 10 assembly software based on virus/human labels instated from nucleotide BLAST results.
Fig 6

Performance on RNA-Seq assembly of human cells cultured with Ebola virus based on the software for assembly.

There is significant variation in performance across the different assembly software. Spades had the most uniform performance where performance was modest whereas more popular assembly tools such as trinity, oases and trans-abyss had poorer but similarly uniform performance.

Performance on RNA-Seq assembly of human cells cultured with Ebola virus based on the software for assembly.

There is significant variation in performance across the different assembly software. Spades had the most uniform performance where performance was modest whereas more popular assembly tools such as trinity, oases and trans-abyss had poorer but similarly uniform performance.

Evaluation of performance against protein BLAST and HMMER3

For characterization of viral genomic data, protein BLAST or NCBI blastx is often used [26, 27]. Hidden Markov Model based multiple-alignment-profiling methods have also drawn the interest of researchers for the task [22, 27, 28]. These methods rely on primary protein sequences for distinction between virus and non-virus data. Although protein primary sequences are more reliable and informative because they lack noisy repeats and ‘junk’ that is characteristic of nucleotide sequence data, the dichotomy for their use in taxonomic classification lies in the fact that they share more homology across species than general sequence data. For these protein-based detection methods, it is common procedure to first weed out the profiles/protein-families against a negative control as they would otherwise display a great propensity towards false positives hits [20]. BLASTx employs a built-in 6 frame-translation algorithm. In order to gain more control over the protocol, we built a Python based in-house six-frame ORF translator with biopython that uses the standard genetic code table (ncbi-translation table ID 1). Short ORFs that are not part of any genes, do not code for proteins and occur entirely due to chance are dispersed throughout genomes. By plotting the frequency of proteins discovered against the protein length cutoff threshold, it was evident that at a length of about 100 amino acids, non-genic random ORF numbers significantly subside (Fig 7). Using this 100-amino-acid cutoff for 6-frame in-silico translation, we translated our entire ssRNA sequence data set and all human de novo transcriptome assemblies to proteins. Querying the complete human proteome for these translated protein sets yielded vast numbers of false positives with very low expect values. We found 1083 in-silico translated viral proteins to have hits with human proteins with a 0.01 e-value cutoff and 27023 proteins with no cutoff. We sifted out these 1083 proteins showing multiple uninformative hits and created a new in-silico viral protein database of only informative proteins. We were left with 38483 predicted proteins in the database. blastp searches on our informative in-silico viral protein database yielded 505794, 3392 and 34 false positives with cutoffs of 10, 0.01 and 0.0001 respectively. Our rbf-SVM68 classifier showed a modest 62%, 64.2% and 70.6% recall in classifying these false hits as human transcripts in the unary data set (Table 4).
Fig 7

Visualizing count of detected ORFs in RNA virus sequences changing in relation to the minimum length cutoff that was set.

We would expect all viral genomes to encode a minimal set of proteins necessary for successful replication and the number of these proteins to deviate minimally across viruses of different types. The graph depicts that at a cutoff range between 100 and 150, we achieve the optimal cutoff point where only important ORFs that have greater probability of being involved in the information pathway of the virus are considered.

Table 4

Performance evaluation of rbf-SVM68 against BLAST and HMMER3.

ToolCutoffQuerySubjectFalse HitsOur ModelFalse PositivesTrue NegativesAccuracy
BLASTnNoneHuman Transcriptome De novo AssemblyRNA Virus Sequence Database135rbf-SMV683410175%
BLASTnNoneMouse Transcriptome De novo AssemblyRNA Virus Sequence Database130rbf-SMV68379372%
BLASTxNoneHuman Transcriptome De novo AssemblyInformative Proteins Only505794rbf-SMV6819195831552862%
BLASTx0.01Human Transcriptome De novo AssemblyInformative Proteins Only3392rbf-SMV681214217864.20%
BLASTx0.0001Human Transcriptome De novo AssemblyInformative Proteins Only34rbf-SMV68102470.60%
hmmseach on informative vFamsNoneHuman Transcriptome De novo AssemblyInformative vFam Profiles Only26003rbf-SMV68119141408954%

Visualizing count of detected ORFs in RNA virus sequences changing in relation to the minimum length cutoff that was set.

We would expect all viral genomes to encode a minimal set of proteins necessary for successful replication and the number of these proteins to deviate minimally across viruses of different types. The graph depicts that at a cutoff range between 100 and 150, we achieve the optimal cutoff point where only important ORFs that have greater probability of being involved in the information pathway of the virus are considered. In the same manner, we sorted out 2346 uninformative vFams from the high performance vFams vFam A available at http://derisilab.ucsf.edu/software/vFam/. Using hmmsearch on our informative vFams against the de novo assembly data set yielded 26003 false hits. Our model was able to correctly classify 14089 of these false hits as human transcripts as noted in Table 4.

Binary classification of plus and minus-strand RNA viruses

To demonstrate the flexibility of our computational pipeline for sequence classification, we separately evaluated the selected model’s ability to distinguish positive and negative sense RNA viruses with 5460, 194 and 68 features respectively. The models were able to very precisely classify positive and negative sense RNA viruses even when trained on a minimalist data set with a train-test divide of 1:5 (Fig 8). The macro averaged f1scores were above .95 in all test cases and the macro scores for the cross-validation set were also convincingly high with low standard deviation. The classification reports are delineated in Table 5. Fig 9 delineates the complete computational pipeline utilized in this study.
Fig 8

Performance of the rbf-SVM model trained with the same computational pipeline for classifying positive and negative sense RNA viruses.

(a) Performance on different train-test splits when trained with all 5460 features. (b) Performance when trained with 194 features. (c) Performance when trained with 68 selected features. All models exhibit a great ability to specify the different sequence classes demonstrating the flexibility of our pipeline.

Table 5

Performance of rbf-SVM in classifying positive and negative sense RNA viruses.

Number of FeaturesTrain-Test SplitTrue PositivesFalse PositivesTrue NegativesFalse Negativescv_binary meancv_binary stdcv_macro meancv_macro stdf1binaryf1macro
54604:11450142260.9727150.0122290.9850830.0066810.979730.988812
54602:330803445210.9668490.0288730.9817880.0158340.9670330.981997
54601:541604587290.9502750.0560180.972740.030450.9663180.981584
1944:112501435130.97590.0133350.9867880.0073030.950570.973031
1942:329803453230.9684340.0104780.9825720.0057860.9628430.979762
1941:538904575680.9240520.0432040.957890.0233250.9196220.956122
684:112531430150.9455880.0268940.9702440.0146630.9328360.963291
682:328163444430.9247080.0317190.9586410.017330.9198040.95637
681:538394575650.8968270.0658650.9436380.0357870.9119050.951941
Fig 9

Diagrammatic representation of our complete experimental design.

Performance of the rbf-SVM model trained with the same computational pipeline for classifying positive and negative sense RNA viruses.

(a) Performance on different train-test splits when trained with all 5460 features. (b) Performance when trained with 194 features. (c) Performance when trained with 68 selected features. All models exhibit a great ability to specify the different sequence classes demonstrating the flexibility of our pipeline.

Discussion

The concept behind this study originated from similar k-mer based prototypes we built for classifying positive and negative sense RNA viruses. We later found that in the past, Random Forest classifiers for 16S rRNA classification based on similar k-mer based principles had displayed outstanding performance metrics [29]. Machine learning attempts at sequence classification have been sparse in the past. But as of late, there is more activity in the field primarily due to advances brought about by next-gen sequencing and growing interest in metagenomics [21, 30, 31], symbionts [11] and microbial colonies [28]. Bzhalava et al. [21] attempted to detect viral sequences in human metagenomic data sets and although their classifier found very promising contrast on GenBank data, the model performed very poorly on a real metagenomic data set. Our models trained on NCBI data displayed 0.8 to 0.95 AUROC on real test data (Fig 5), which is a big improvement form their AUROC of 0.51. k-mer based Random Forest models by the same team using k-mers of lengths 1 to 7 achieved the best AUROC of 0.875 with k-mers of length 6 [16], which is significantly lower than the metrics of our Random Forest models. VirFinder is a k-mer based machine learning tool that can identify prokaryotic viral contigs from metagenomic data [11]. The performance of VirFinder was evaluated with varying k-mer lengths, class skews and contig lengths. Our models trained on varying numbers of features and large class discrepancies display more consistent and robust performance metrics on average across all tests concerning NCBI data. The AUROC for our models on the divergent data set comprised of mouse transcripts and DNA viruses was 0.95 to 0.96 compared to their 0.90 to 0.98 on an analogous simulated data set. The features we constructed from short k-mers displayed strong discriminatory capacity among the sequence classes. This has stood out to be the case in all tests carried out in this study including those that were left out from the final paper. Both positive and negative sense RNA viruses must produce complementary strands of reverse polarity in order to replicate. In positive sense viruses, the reverse complements are necessary for making new copies of their genome since all polymerases are unidirectional from the 5’ end to the 3’ end and in negative sense viruses the necessity of making reverse complements goes beyond DNA replication as their primary genetic information pathway relies on the synthesis of the reverse complement that codes for viral proteins. It felt necessary to include the reverse complements of the viral genomes as we would undoubtedly encounter remnants of them in whatever virus infected cell we sought to explore. The increase in training examples did not hurt the model by relegating bias upon it, as evident from the generalization tests. The scikit-learn library offers feature importance scores from the Random Forest tree ensemble. It is calculated from the selection criterion based on the gain of information by the split at the nodes. We selected the entropy criterion as our measure of information gain because of the heterogeneous nature of our features. The disadvantage of tree-based feature selection is the decrease in weight assigned to correlated features. The smaller k-mers are all slices from the larger k-meric words and hence the are highly correlated. We thus sensed the need for going beyond solely tree-based selection approaches. The only drawback of our model stood out to be the moderate computational power necessary to calculate all the features. For short RNA virus sequences, it took us a few minutes on a 4 core third-generation i5 computer. Based on our results, the design of the model proved to be highly flexible and we make a strong case for the use of similar k-mer scoring profiles for classification of other classes of sequence data. It was thus necessary to scale down the model by reducing the number of features so it could comfortably adapt to much longer sequence classes. In order to scale the model, we went beyond tree-based methods and employed Lasso regularization. Feature selection by Lasso stems from linear regression models by penalizing the coefficients of the variables. Using a combination of these selection methods on multiple levels, we were able to shrink the model down into only few features while only minimally compromising performance. Overfitting is always a concern when applying ML models to biological data. The first impressions on the high accuracy of the classifiers suggested a case of overfitting in the model. The bootstrap aggregating or bagging meta-algorithm employed in Random Forest classification is particularly resistant to high variance and overfitting. Although Decision Trees employ the same bagging principle, because of their larger depths, they are more prone to higher variance and overfitting. Many random sampling events generate subsets of data known as bootstrap samples. Isolated Decision Trees are trained on these bootstrap samples to construct the Random Forest ensemble. A majority vote from the members of the ensemble generates predicted classifications for test sets. Since our random forest models performed noticeably better than the decision tree model, it would lead us to believe that overfitting does not account for the fidelity of our models. Stronger evidence against a case for high variance comes from the model’s impeccable performance when tested against viruses not belonging to the RNA virus groups and mRNA from mouse transcriptomes. These test sets substantially varied from the training data and a highly biased model would have displayed a moderate decrease in performance. High performance metrics on very small training sets in our positive and negative sense RNA classification tests are also good evidence against high bias in the model. Our Gaussian Naïve Bayes models had the poorest performance metrics. This is easily explained by heterogeneous and nonuniform nature of our features. When tested for gaussian normalcy, all features failed to show a distribution resembling the normal distribution. Hence, the performance of the Gaussian Naïve Bayes models was poor owing to the non-continuous datapoints in our feature sets. The improvement in its performance after the reduction in the number of features is suggestive of a non-linear nature of the numerical values and the presence of random noise, which is customary to biological data sets. The superior performance of Logistic Regression models and both Support Vector Machine models suggest a discontinuous polar segregation for the feature values in their respective classes. When progressing to real RNA-Seq assembly data, the previously robust, infallible performance of the model began to wane. In the assemblies from the human cell line, filtering the transcripts to conform to lengths resembling the original training set solved the issue and the robustness of the classifier became apparent again. But yet, performance on the 23-hour Ebola virus cell incubate was less impressive. This can be easily explained by the incongruity of the assembly software, the unwanted mixing of contigs from either source during assembly and the irregularities in the lengths of the assembled contigs. The false negative results were also higher than we expected. It would have been a matter of concern if these misclassifications were a result of the very high mutation rates characteristic of RNA viruses [32], but the evidence from the cross-validation and generalization studies and error analysis within the real data set (not discussed) lead us to believe that such is not the case. We postulate that as de novo assembly tools improve, the classification performances of our models would tend towards that observed in curated sequence data. Since all human training data was composed of gene transcripts, whereas the sequences in the virus training set were mostly comprised of non-genic regions, we hypothesized that this could be a cause for greater numbers of false negatives because the majority of virus sequences in the cell-virus incubate would also inevitably be viral gene transcripts as a consequence of the infectious activity of the Ebola virus, rather than the whole genome or sequencing with long non-genic regions. However, including RNA virus transcripts separately downloaded from NCBI into the training set only marginally improved performance on the Ebola virus and cell incubate sample (not discussed). This was more evidence to support our claim that the poor performance is greatly justified by shortcomings in the assembly of the RNA-Seq data itself.

Conclusion

In this study, the use of short k-mer profiles as machine learning features proved to be very resourceful for training sequence classification models. Machine learning and deep learning models that train on sequence data more commonly treat biological sequences as lifeless vectors of binary information. The sequence of nucleotide bases in nucleic acids does not occur randomly and independently with identical probability. Rather, a confluence of biophysical and biochemical forces actively dictates the nature of this evolution. Our results question this approach for training sequence classifiers and sheds light on the elegance in the emergence and evolution of genomes centered around short k-meric repeats. The evaluation metrics from the classification results in the leave-one-family-out experiments, the performance on the divergent sequences and the ability to differentiate plus and minus strand viruses even when trained on very small data sets, is good evidence that is greatly suggestive of a robust set of numerical machine learning features. The pipeline depicted on Fig 9 can be applied to any two classes of sequence data having considerable taxonomic distance to train machine learning models that can accurately report the presence of either class in unknown sequence data. Nucleotide blast is still the best tool at our disposal for characterization of unknown sequence data. But due to the homology omnipresent in biological sequences even between distinct orders of life, false positives and ambiguous results are always prevalent. The drawbacks of this approach include, the very limited number of viral genomic diversity that has been so far catalogued, the uncertainty of there being a known homolog or close ancestor in the database, and the prevalence of many viral genomic remnants in the genomes of host organisms. We can also reasonably conclude that protein-based methods such as protein blast or pfams are not suited to the task of taxonomic classification of greatly divergent RNA sequences. Pertaining to the nonuniformity of the overlap between the results from our classifier and the tools it was compared to, we propose that a combinatorial approach employing multiple tools simultaneously is currently the best option for segregating sequence data by taxonomy (Fig 10).
Fig 10

Suggested approach for acquiring the best automated annotations.

Our explorations indicate that there is no single annotation tool that can produce the best results without requiring further curation. Perhaps the best way to use these tools is by stacking the results on top of each other to remove maximum number of false positives while consolidating the true positive results.

Suggested approach for acquiring the best automated annotations.

Our explorations indicate that there is no single annotation tool that can produce the best results without requiring further curation. Perhaps the best way to use these tools is by stacking the results on top of each other to remove maximum number of false positives while consolidating the true positive results. De novo assembly of RNA-Seq data facilitates the discovery of novel transcripts and splice variants. It can offer great promise in the discovery of novel RNA viruses without prior sequence information. Software packages that carry out de novo assembly of RNA-Seq data are still improving. The 10 assembly tools we tested our model on were still highly variable in spite of assembling the same raw sequence data. For the identification of divergent viruses, the use of robust and highly specific methods such as PCR, immunoassays or microarrays are not suitable [33]. Deep sequencing of metagenomes and metatranscriptomes broadens the landscape available for novel virus detection by many orders of magnitude. With the promise of third-gen sequencing [34], this uncharted landscape of sequence data will keep expanding, necessitating the need for novel tools for mining and annotating sequence data. We envision that as the accuracy of RNA sequencing technologies advance, the software that are used to assemble them will also upgrade and models such as ours that facilitate the annotation of the assembled data will be expected to catch up to these advancements.

Methods

Contriving features using k-mer abundance profiles

For the DNA letters ‘A’, ‘C’, ‘G’ and ‘T’, we use all possible k-mers of length 1 to 6 to derive feature scores for each sequence. We use DNA letters instead of RNA letters where Thymidine (T) is replaced by Uracil (U) because by convention most sequence data formats are encoded using DNA letters and this is also true for RNA transcripts and RNA genomes. The number of features for any sequence is thus the total number of possible k-mers, or in our case 5460, as shown below. Through dynamic programming, our algorithm simulates the task of sequentially matching a k-mer against a sequence from start to end, going along the sequence like a sliding window, one ribonucleotide base at a time. For any sequence q and k-mer e, we obtain a total count when allowing exactly j total mismatches or gaps. Our final sequence score W(e) for element e, and sequence q, is calculated by the formula given below where n is the length of sequence q. This final score is used as the pre-scaled value for feature e in sequence q in all constructed machine learning models in this study.

Data preprocessing

We sourced all available virus sequences from NCBI [35]. Taxonomical data related to virus genomic architecture was retrieved from the ICTV website [36]. We found 3145 unique viruses having either ssRNA(+) and ssRNA(-) genomes, with 2862 belonging to the plus-strand group and 283 to the minus-strand. The complete human reference transcriptome was obtained from NCBI. All predicted and non-putative RNA sequences from the reference transcriptome were omitted. In compliance with the general length of animal viruses, to build a human/virus binary RNA classifier, only putative and expressed human RNA ranging from 2000 to 35000 in length were selected for the study. For sequences sharing greater than 85% similarity, only the longest sequence was selected for use while omitting all others. We were left with 4114 unique viral genomes and 15476 human RNA transcripts (S6 Table). The reverse complement of the RNA(+) strand in ssRNA(+) viruses is a mandatory intermediate in its replication and the reverse complement of the RNA(-) strand in ssRNA(-) viruses is a mandatory intermediate in both replication and the information pathway. We therefore included the reverse complements of the working RNA virus genomes in our data set.

Libraries and parameters

Python’s biopython module was used for all sequence manipulation tasks. For all statistical analysis we used the scipy.stats module in Python and scikit-learn was used for all machine learning implementations. Plots were drawn using matplotlib’s pyplot. Train-test splits were always 4:1 for training set and test set respectively, unless specifically mentioned in the results. Standard scaling was employed to scale the features. 10-fold cross-validation was carried out within the training sets and both mean and standard deviation of the macro average f1 score and binary f1 score was reported. Similarly, for the test set, both macro averaged f1 scores and binary f1 scores were reported. For skewed classes it is classic convention to label the smaller class as the positive example. Hence, in all training and test examples in this study, viruses were set as the positive binary class. In the classification of positive and negative sense RNA viruses, there were 2862 examples in the positive RNA class and 283 examples in the negative RNA virus class. The negative sense RNA viruses were considered the positive binary class. Logistic Regression models were trained with L2 penalty and 2000 max iterations. For the Support Vector Machine classifiers, both linear and RBF penalization were separately employed. The probability parameter for the SVC class of sklearn was set to True in order to acquire probability scores for drawing ROC curves where needed. The tree-based models, Decision Trees and Random Forest were trained using the default information entropy function as criterion. The number of estimators was set at 200 for Random Forest. For any parameters not explicitly stated, the library defaults were used.

Testing and performance evaluation

For cross-validation and generalization tests 11351 putative, expressed mRNA transcripts between 2000 and 35000 sequence length and sharing no more than 85% similarity with any other transcript were selected from the Mus musculus whole transcriptome available in NCBI. As divergent viruses, 2532 viruses between the specified sequence length range not having positive or negative strand RNA genomes were selected. All information about the 77 virus families and mouse transcripts can be found in S3 Table. De novo RNA-Seq assembly data was sourced from experiments by Hölzer et al. [24] available at https://doi.org/10.17605/OSF.IO/5ZDX4. The following single-number evaluation metrics were used as a measure of model performance.

F1-score

A measure of balance between precision and recall defined as .

F1-binary

The F1 score where only the predictions of the positive class are considered.

F1-macro

The macro-averaged F1 score where the unweighted mean for each class label is taken into account to adjust for an imbalance of class size.

CV_bin and CV_macro

Set of F1-binary and F1-macro scores for cross-validation tests respectively. Both the mean and standard deviation are reported.

AUROC

An abbreviation of Area Under the Receiver Operating Characteristic Curve. A graphical representation of the discriminatory capacity of a binary classifier. A random predictor would have an AUROC of 0.5.

Accuracy

Accuracy in this paper is specifically reserved for cases where the proportion of correct class predictions is reported. Hence, .

Unary accuracy

Accuracy for a data set where all relevant datapoints belong to the same class. Command-line BLAST+ [25] was used for sequence alignments with blastn and blastp. All alignment tasks were carried out locally. For both blastn and blastp, false hits were first isolated by query against a negative control database (i.e. human transcriptome for RNA viruses and our RNA virus sequence database for the de novo assembly data) and only informative hits were considered in the final counts. For blastp, cutoff for minimum ORF length was set at 100 bases. The command-line HMMER3 [37] application was employed for Hidden Markov Profile searches. The hmmsearch command was first carried out on virus protein family alignments known as vFams [22] against our negative control human transcriptome, to remove any alignment profiles that might be uninformative. Later, only the filtered informative protein profiles were used for the final search in metagenomic data. Informative vFams are deposited in the git repository as a.hmm file. The ten assembly tools that have been evaluated are BinPacker [38], Bridger [39], Shannon [40], rnaSPAdes [41], SPAdes [42], idba-tran [43], SOAPdenovo-Trans [44], Trans-Abyss [45], Trinity [46] and Oases [47]. All computational tasks were carried out on a third-generation Intel core-i5 machine having 8GB of RAM. All source-code, data and executed commands for this project including the trained classifiers are available on github at https://github.com/DeadlineWasYesterday/V-Classifier.

Performance of the 18 classifiers when trained on different numbers of features.

(a) Logistic Regression Classifier on 3 Feature Sets. (b) Linear SVM Classifier on 3 Feature Sets. (c) RBF-Kernel SVM Classifier on 3 Feature Sets. (d) Decision Tree Classifier on 3 Feature Sets. (e) Random Forest Classifier on 3 Feature Sets. (f) Gaussian Naïve Bayes Classifier on 3 Feature Sets. (JPG) Click here for additional data file.

All k-meric features ranked by feature importance instated by our feature selection flow-chart.

(XLSX) Click here for additional data file.

Complete results for leave-one-family-out cross-validation procedure.

(XLSX) Click here for additional data file.

Information of sequences used to test the model's ability to generalize.

(CSV) Click here for additional data file.

False nucleotide blast hits across whole human transcriptome when queried against the RNA virus sequence database.

(CSV) Click here for additional data file.

False nucleotide blast hits across whole mouse transcriptome when queried against the RNA virus sequence database.

(CSV) Click here for additional data file.

Detailed information on training set sequences.

(CSV) Click here for additional data file.

Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present. 18 Aug 2020 PONE-D-20-22563 Short k-mer Abundance Profiles Yield Robust Machine Learning Features and Accurate Classifiers for RNA Viruses PLOS ONE Dear Dr. Alam, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Authors need to prepare a new version of the manuscript in accordance with the recommendations and significant comments from the reviewers. Otherwise, the manuscript will not be published. Please submit your revised manuscript by Oct 02 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Ruslan Kalendar, PhD Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? Reviewer #1: Partly Reviewer #2: No ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: N/A Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Reviewer #1: This manuscript addresses the problem of recognizing sequences of RNA viral origin within a data set of sequences. Their solution is a machine learning one, in which they derive from a sequence a set of features, each of which is a short $k$-mer, either its presence or its frequency, which is not clear. They use and evaluate several machine learning methods built using a Python machine learning package. The evaluation efforts are extensive and sufficient. The authors make their Python software available through Github. Always capitalize Python. There are two major problems with the exposition. First, the authors never clearly define the classification problem they are trying to solve. After reading the entire manuscript, it is clearly a binary classification, but what exactly are the classes? Ask a computer scientist for help with properly defining a computational problem. Second, the mathematical exposition on page 34 is the weakest part of the manuscript. The sentences, starting with the first two, make no mathematical sense. What do "mismatches and gaps" have to do with $k$-mers? Again, consult a computer scientist, preferably a theoretical computer scientist. Also, never use the blackboard abbreviation $\\forall$ in technical writing. That and similar notation is reserved for formal logic expressions. Write it out in English and do not put that in the displayed equation. The entire section needs a complete rewrite with more careful mathematical details. Some minor issues follow. Page 2. "De novo" is not hyphenated. Page 3 and throughout. "numerous[1]" should be "numerous [1]". In general, there must be a space before a citation. Page 3, line 48. "Often completely unbeknownst to us." is not a sentence. Make sure every sentence has a subject and a verb. Page 3, line 49. "are astounding" should be "is astounding". Page 3, line 57. "encoding;" should be "encoding," Page 4, lines 65-67. I verified this claim with a search of Web of Science. Page 5, line 92. Do not use the possessive with inanimate things. So, "Scikit-learn's" should be rephrased "The scikit-learn". Make changes throughout. Page 35, line 465. "dataset" should be "data set". Change throughout. Page 37, lines 508-509. Capitalize Intel. "8GBs" should be "8GB". Figure 1, "Layers of Feature Selection", suffers from poor resolution and poor readability. There is no need for the color. Just use black text in black boxes with white background for maximum readability. Reviewer #2: Summary This study applies a short k-mer sequence scores based machine-learning method for classifying RNA viruses from human transcripts. The short k-mer based features are selected and tested with six different classification models: Logistic Regression, SVM, Kernel SVM, Decision Tree, Random Forest and Gaussian Naïve Bayes Classifiers. The methods were mainly tested on (1) RNA viral genomes and human transcripts (2) de-novo assembled contigs from RNA-seq data, and achieved good AUC results on the first data set and reasonable results on the second one. Major issues 1. The main issue is the experiment results lack comparisons. (1) All the results in this study are not compared with other machine learning based viral sequences classifiers, while there are a lot of those kinds of tools, especially those that are also based on k-mer features, like VirFinder. (2) For the results on RNA viral genomes and human transcripts (pager 7, line 113), while the tested models have good AUCs, the results are not compared with blast results or other machine-learning based methods, which does not prove the superiority of the proposed methods 2. For the models training and testing, while as stated a 25% split of training and testing data set is performed (page 7, line 118), no test metrics are reported. For Figure 3, why report the results of cross-validation but not the testing results? 3. The results on real RNA-seq assemblies seem not to be better than blast. On page 15, line 208, blast “yielded 135 false positives” on Human transcriptome assemblies. However, comparing to the results in Table 3, rbf-SVM68 has much more false positives comparing to blast (7067 even after length filtered) 4. How ssRNA(+) and ssRNA(-) are distinguished is not well explained. Usually contigs assembled from RNA-seq data contain both plus and minus strand sequences for a genome, can these contigs be successfully classified as ssRNA(+) or ssRNA(-)? Minor issues: 1. In page 15, line 208, blast “yielded 135 false positives” on Human transcriptome assemblies, while in Supplementary Table 4, there are actually 243 sequences. 2. In page 8, line 130, AUROC is already used before the text “Area under the receiver operating characteristic curve” 3. In Figure 5, typo “Only humam_filtered” 4. Evaluation metrics need some explanation, like macro-averaged f1-score, binary f1-score, ROC AUC etc. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: Lenwood S. Heath Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 18 Aug 2020 Response to reviewer #1 Dear Dr. Heath, Thank you for your valuable feedback. In this revision, we have tried our best to satisfy all previous shortcomings that had been pointed out. Reviewer #1: The authors make their Python software available through Github. Always capitalize Python. Response: All occurrences of the word Python were capitalized. Reviewer #1: There are two major problems with the exposition. First, he authors never clearly define the classification problem they are trying to solve. After reading the entire manuscript, it is clearly a binary classification, but what exactly are the classes? Ask a computer scientist for help with properly defining a computational problem. Response: We acknowledge this concern by adding the following changes in lines 73 and 79 of the introduction: “With our scoring function, we were able to calculate the profiles of short genetic elements (k-mers) for RNA virus genomes and human RNA transcripts. We trained a number of classical machine learning classifiers on our sequence-derived numerical data in an attempt to classify an assembled RNA sequence of unknown origin as either viral or cellular RNA.” “Furthermore, we demonstrated the flexibility of this approach by attempting to classify an RNA virus genome sequence from NCBI as either a positive-strand (ssRNA(+) ) or a negative-strand (ssRNA(-) ) RNA virus.” Reviewer #1: Second, the mathematical exposition on page 34 is the weakest part of the manuscript. The sentences, starting with the first two, make no mathematical sense. What do "mismatches and gaps" have to do with $k$-mers? Again, consult a computer scientist, preferably a theoretical computer scientist. Also, never use the blackboard abbreviation $\\forall$ in technical writing. That and similar notation is reserved for formal logic expressions. Write it out in English and do not put that in the displayed equation. The entire section needs a complete rewrite with more careful mathematical details. Response: We have consulted with our colleagues that have a background in computer science and re-written the mentioned section hoping to make it more mathematically sound and comprehensible. The section is now as follows. Contriving features using k-mer abundance profiles For the DNA letters ‘A’, ’C’, ‘G’ and ‘T’, we use all possible k-mers of length 1 to 6 to derive feature scores for each sequence. We use DNA letters instead of RNA letters where Thymidine (T) is replaced by Uracil (U) because by convention most sequence data formats are encoded using DNA letters and this is also true for RNA transcripts and RNA genomes. The number of features for any sequence is thus the total number of possible k-mers, or in our case 5460, as shown below. ∑_(i=1)^6▒4^i =5460 Through dynamic programming, our algorithm simulates the task of sequentially matching a k-mer against a sequence from start to end, going along the sequence like a sliding window, one ribonucleotide base at a time. For any sequence q and k-mer e, we obtain a total count C_j^q (e) when allowing exactly j total mismatches or gaps. Our final sequence score W_q (e) for element e, and sequence q, is calculated by the formula given below where n is the length of sequence q. W_q (e)=(∑_(j=1)^3▒(¬C_j^q (e))/2^j )⁄n This final score is used as the pre-scaled value for feature e in sequence q in all constructed machine learning models in this study. Reviewer #1: Page 2. "De novo" is not hyphenated. Response: The hyphen was removed. Reviewer #1: Page 3 and throughout. "numerous[1]" should be "numerous [1]". In general, there must be a space before a citation. Response: This (no space) is the default behavior of the citation manager. Nonetheless, the advised spaces were added. Reviewer #1: Page 3, line 48. "Often completely unbeknownst to us." is not a sentence. Make sure every sentence has a subject and a verb. Response: We have omitted the sentence. Reviewer #1: Page 3, line 49. "are astounding" should be "is astounding". Response: Corrected. Reviewer #1: Page 3, line 57. "encoding;" should be "encoding," Response: Corrected. Reviewer #1: Page 4, lines 65-67. I verified this claim with a search of Web of Science. Response: We are yet to notice otherwise as well. Reviewer #1: Page 5, line 92. Do not use the possessive with inanimate things. So, "Scikit-learn's" should be rephrased "The scikit-learn". Make changes throughout. Response: Changes were made throughout the manuscript so as to not anthropomorphize the sklearn library. Reviewer #1: Page 35, line 465. "dataset" should be "data set". Change throughout. Response: We believe they can be used interchangeably. However, we have changed throughout as instructed. Reviewer #1: Page 37, lines 508-509. Capitalize Intel. "8GBs" should be "8GB". Response: Corrected. Reviewer #1: Figure 1, "Layers of Feature Selection", suffers from poor resolution and poor readability. There is no need for the color. Just use black text in black boxes with white background for maximum readability. Response: The figure has been modified to meet the mentioned criteria. We have maximized the contrast by using very light colors and black fonts instead of going completely black-and-white. Response to reviewer #2 We appreciate the concerns and criticisms raised by reviewer #2. We have made some significant changes to address some issues, and will provide a very comprehensive clarification here on matters of rebuttal. Reviewer #2: 1. The main issue is the experiment results lack comparisons. (1) All the results in this study are not compared with other machine learning based viral sequences classifiers, while there are a lot of those kinds of tools, especially those that are also based on k-mer features, like VirFinder. (2) For the results on RNA viral genomes and human transcripts (pager 7, line 113), while the tested models have good AUCs, the results are not compared with blast results or other machine-learning based methods, which does not prove the superiority of the proposed methods Response: The task of making comparisons to other published models had been on the back of our minds throughout the duration of the study. Towards the end, we felt that it became apparent that our models show superior performance. Because there were so few similar studies on the subject and because our manuscript was getting quite long for the topic, we decided not to include tabular or graphical comparisons. We had gone over the VirFinder paper and it had also been cited in the manuscript. The reported performance of VirFinder varies significantly as a function of the lengths of k-mers used, the proportion of viral to bacterial contigs and the size of the contigs. We reported our performance metrics for 5460, 194 and 68 features respectively and got consistent results (Fig. 1 and Supplementary Fig. 1). We also tested the classifiers with varying proportions of sequence data (E.g. when cross-validating by leaving large families out, when testing on mouse transcripts and when testing our PN classifier on very small test sets). The AUROC for our models under varying parameters are notably better and more consistent than those reported in the VirFinder paper. However, our model performance does fluctuate similarly to VirFinder with varying sequence lengths (as evident from Fig. 5). Additionally, they (VirFinder) never test their model’s sequence discriminatory capacity on real data. They have used real data to try and predict healthy/disease state instead, which employs a very different workflow. The simulation dataset they test on, and report 0.90 to 0.97% AUROC curve is built on top of downloaded reference sequences, which is more analogous to our Mouse transcript and DNA virus test set, where we report 0.95 to 0.96% AUROC. It is important to note the following sentence from line 31 in the abstract: “For clean sequence data retrieved from curated databases, our models display near perfect accuracy, outperforming all similar attempts previously reported.” Coming back to VirFinder- we can accept these metrics because our experiment is substantially different from the VirFinder study (they manually cleave the genomes to sample them, they classify bacteria against prokaryotic viruses, most prokaryotic viruses are DNA viruses[1], they use contigs whereas we use assemblies, they use reference-based tools whereas we use de novo assemblies, they use prediction categories, etc.) Our manuscript only claims to outperform ‘similar attempts from the past’. Programs like ViraMiner[2] and Bzhalava et al. 2018 [3] are more ‘similar’ in methodology as well as the goal to accomplish. In spite of this rebuttal, we have made significant and perhaps necessary changes to the manuscript. The following was added to the discussion section in line 324. Machine learning attempts at sequence classification have been sparse in the past. But as of late, there is more activity in the field primarily due to advances brought about by next-gen sequencing and growing interest in metagenomics [21, 31, 32], symbionts [11] and microbial colonies [28]. Bzhalava et al. 2018 [21] attempted to detect viral sequences in human metagenomic data sets and although their classifier found very promising contrast on GenBank data, the model performed very poorly on a real metagenomic data set. Our models trained on NCBI data displayed 0.8 to 0.95 AUROC on real test data (Fig. 5), which is a big improvement form their AUROC of 0.51. k-mer based Random Forest models by the same team using k-mers of lengths 1 to 7 achieved the best AUROC of 0.875 with k-mers of length 6 [16], which is significantly lower than the metrics of our Random Forest models. VirFinder is a k-mer based machine learning tool that can identify prokaryotic viral contigs from metagenomic data [11]. The performance of VirFinder was evaluated with varying k-mer lengths, class skews and contig lengths. Our models trained on varying numbers of features and large class discrepancies display more consistent and robust performance metrics on average across all tests concerning NCBI data. The AUROC for our models on the divergent data set comprised of mouse transcripts and DNA viruses was 0.95 to 0.96 compared to their 0.90 to 0.98 on an analogous simulated data set. Reviewer #2: 2. For the models training and testing, while as stated a 25% split of training and testing data set is performed (page 7, line 118), no test metrics are reported. For Figure 3, why report the results of cross-validation but not the testing results? Response: All results had been reported in the original manuscript. An unfortunate misunderstanding has occurred here. The f1bin and f1mac columns in Table 2 are the metrics for the test sets. These column names had also been explained on the footnotes of Table 2 in the original manuscript. However, to avoid any misunderstanding by the audience, we have added the following section to our Methods section in line 523 to explain our single number evaluation metrics. The following single-number evaluation metrics were used as a measure of model performance. F1-score: A measure of balance between precision and recall defined as (2×(precision × recall))/((precision + recall)) . F1-binary: The F1 score where only the predictions of the positive class are considered. F1-macro: The macro-averaged F1 score where the unweighted mean for each class label is taken into account to adjust for an imbalance of class size. CV_bin and CV_macro: Set of F1-binary and F1-macro scores for cross-validation tests respectively. Both the mean and standard deviation are reported. AUROC: An abbreviation of Area Under the Receiver Operating Characteristic Curve. A graphical representation of the discriminatory capacity of a binary classifier. A random predictor would have an AUROC of 0.5. Accuracy: Accuracy in this paper is specifically reserved for cases where the proportion of correct class predictions is reported. Hence, accuracy= (correct classifications)/(total number of examples) . Unary accuracy: Accuracy for a data set where all relevant datapoints belong to the same class. Additionally: We have tried our best to be as transparent as possible with all our results. For example, we expected our model to perform better when we separately included the protein coding transcripts from viruses into the training data because coding regions are much more informative than other parts of the genome. However, it resulted in a decline in overall performance. This observation was mentioned in the discussion of the original transcript (now line 407) as follows: “Since all human training data was composed of gene transcripts, whereas the sequences in the virus training set were mostly comprised of non-genic regions, we hypothesized that this could be a cause for greater numbers of false negatives because the majority of virus sequences in the cell-virus incubate would also inevitably be viral gene transcripts as a consequence of the infectious activity of the Ebola virus, rather than the whole genome or sequencing with long non-genic regions. However, including RNA virus transcripts separately downloaded from NCBI into the training set only marginally improved performance on the Ebola virus and cell incubate sample (not discussed).” Reviewer #2: 3. The results on real RNA-seq assemblies seem not to be better than blast. On page 15, line 208, blast “yielded 135 false positives” on Human transcriptome assemblies. However, comparing to the results in Table 3, rbf-SVM68 has much more false positives comparing to blast (7067 even after length filtered) Response: The manuscript never claimed to have better performance than BLAST. We only mentioned that our classifiers outperform other similar attempts made in previous studies. In the concluding parts, our goal was to explain that even though BLAST is the gold standard for sequence classification and annotations, the tool is not resistant to false positives. Our classifier had much more false positives than BLAST, as apparent from our reported results. We explained in the results section that our classifier, in spite of having more false positives than BLAST, could correctly classify a large portion of the false positives that we acquire from BLAST search. The following excerpts are from lines 211, 216 and 220 of the results section. “In the Homo sapiens transcriptome assembly, carried out on GM12878 human gammaherpes virus 4 transformed B-lymphocyte cell lines, nucleotide blast query against our sequence database of ssRNA(+) and ssRNA(-) viruses yielded 135 false positives (Supplementary Table 4)…” “…Our 68 feature rbf kernel SVM model correctly classified 101 of these false positives giving us an accuracy of 75% on the unary data.” “Similarly, nucleotide blast had given 130 false positives on the Mus musculus transcriptome assemblies (Supplementary Table 5), whereas our model accurate classified 93 examples with 72% accuracy on the unary class.” This observation has also been shared in the abstract in line 35: “Our classifier was able to properly classify the majority of the false hits generated by BLAST and HMMER3 searches on the same data.” This is what led to the idea behind Figure 10, that in order to acquire the best annotations, we must integrate the tools from different disciplines because any tool based on a single principle will have its shortcomings and yield some false positive results. This was explained in the conclusions of the original manuscript as follows (lines 432 and 439 in the revised manuscript). “Nucleotide blast is still the best tool at our disposal for characterization of unknown sequence data. But due to the homology omnipresent in biological sequences even between distinct orders of life, false positives and ambiguous results are always prevalent…” ”… Pertaining to the nonuniformity of the overlap between the results from our classifier and the tools it was compared to, we propose that a combinatorial approach employing multiple tools simultaneously is currently the best option for segregating sequence data by taxonomy (Fig 10).” Reviewer #2: 4. How ssRNA(+) and ssRNA(-) are distinguished is not well explained. Usually contigs assembled from RNA-seq data contain both plus and minus strand sequences for a genome, can these contigs be successfully classified as ssRNA(+) or ssRNA(-)? Response: Yes. For curated data (as is done in the manuscript), absolutely. The sole purpose of our PN classifier was to demonstrate the flexibility of our machine learning pipeline, which it competently does. We have not done any classification tasks for classifying ssRNA(+) and ssRNA(-) sequences from the de novo assembly data. The PN classifier explained in the results sections was trained and tested entirely with well-annotated viral genome sequences that are deposited in NCBI. The information about the genome types of these viruses were tactfully sourced from the data available in the ICTV website[4] using computer algorithms in Python. This entire process is documented in the git repository. This was also mentioned in the methods section (now line 481 in the revised manuscript). Reviewer #2: Minor issues: 1. In page 15, line 208, blast “yielded 135 false positives” on Human transcriptome assemblies, while in Supplementary Table 4, there are actually 243 sequences. Response: The table has been fixed. In actuality, there were 135 false positives on the table. But the number of rows were 243 because of repeated entries. Simply dropping the duplicates gave us 135 data rows. Which we will now upload as an updated table. Reviewer #2: 2. In page 8, line 130, AUROC is already used before the text “Area under the receiver operating characteristic curve” Response: Changed to AUROC. Reviewer #2: 3. In Figure 5, typo “Only humam_filtered” Response: Fixed. Reviewer #2: 4. Evaluation metrics need some explanation, like macro-averaged f1-score, binary f1-score, ROC AUC etc. Response: Already addressed in the response to the second major point raised by reviewer #2. We have added a new section to the methods briefly describing the evaluation metrics. 1. McGrath, S. and D.v. Sinderen, Bacteriophage: genetics and molecular biology. 2007: Caister Academic Press. 2. Tampuu, A., et al., ViraMiner: Deep learning on raw DNA sequences for identifying viral genomes in human samples. PLOS ONE, 2019. 14(9): p. e0222271. 3. Bzhalava, Z., et al., Machine Learning for detection of viral sequences in human metagenomic datasets. BMC Bioinformatics, 2018. 19(1): p. 336. 4. Lefkowitz, E.J., et al., Virus taxonomy: the database of the International Committee on Taxonomy of Viruses (ICTV). Nucleic Acids Research, 2017. 46(D1): p. D708-D717. Submitted filename: Response to Reviewers.docx Click here for additional data file. 7 Sep 2020 Short k-mer Abundance Profiles Yield Robust Machine Learning Features and Accurate Classifiers for RNA Viruses PONE-D-20-22563R1 Dear Dr. Alam, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Ruslan Kalendar, PhD Academic Editor PLOS ONE 10 Sep 2020 PONE-D-20-22563R1 Short k-mer Abundance Profiles Yield Robust Machine Learning Features and Accurate Classifiers for RNA Viruses Dear Dr. Alam: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Ruslan Kalendar Academic Editor PLOS ONE
  43 in total

1.  De novo assembly and analysis of RNA-seq data.

Authors:  Gordon Robertson; Jacqueline Schein; Readman Chiu; Richard Corbett; Matthew Field; Shaun D Jackman; Karen Mungall; Sam Lee; Hisanaga Mark Okada; Jenny Q Qian; Malachi Griffith; Anthony Raymond; Nina Thiessen; Timothee Cezard; Yaron S Butterfield; Richard Newsome; Simon K Chan; Rong She; Richard Varhol; Baljit Kamoh; Anna-Liisa Prabhu; Angela Tam; YongJun Zhao; Richard A Moore; Martin Hirst; Marco A Marra; Steven J M Jones; Pamela A Hoodless; Inanc Birol
Journal:  Nat Methods       Date:  2010-10-10       Impact factor: 28.547

Review 2.  Origins and challenges of viral dark matter.

Authors:  Siddharth R Krishnamurthy; David Wang
Journal:  Virus Res       Date:  2017-02-09       Impact factor: 3.303

3.  Expression of animal virus genomes.

Authors:  D Baltimore
Journal:  Bacteriol Rev       Date:  1971-09

Review 4.  Emerging view of the human virome.

Authors:  Kristine M Wylie; George M Weinstock; Gregory A Storch
Journal:  Transl Res       Date:  2012-04-24       Impact factor: 7.012

5.  Bridger: a new framework for de novo transcriptome assembly using RNA-seq data.

Authors:  Zheng Chang; Guojun Li; Juntao Liu; Yu Zhang; Cody Ashby; Deli Liu; Carole L Cramer; Xiuzhen Huang
Journal:  Genome Biol       Date:  2015-02-11       Impact factor: 13.583

Review 6.  Viral metagenomics.

Authors:  Eric L Delwart
Journal:  Rev Med Virol       Date:  2007 Mar-Apr       Impact factor: 6.989

7.  Full-length transcriptome assembly from RNA-Seq data without a reference genome.

Authors:  Manfred G Grabherr; Brian J Haas; Moran Yassour; Joshua Z Levin; Dawn A Thompson; Ido Amit; Xian Adiconis; Lin Fan; Raktima Raychowdhury; Qiandong Zeng; Zehua Chen; Evan Mauceli; Nir Hacohen; Andreas Gnirke; Nicholas Rhind; Federica di Palma; Bruce W Birren; Chad Nusbaum; Kerstin Lindblad-Toh; Nir Friedman; Aviv Regev
Journal:  Nat Biotechnol       Date:  2011-05-15       Impact factor: 54.908

8.  Gene prediction in metagenomic fragments: a large scale machine learning approach.

Authors:  Katharina J Hoff; Maike Tech; Thomas Lingner; Rolf Daniel; Burkhard Morgenstern; Peter Meinicke
Journal:  BMC Bioinformatics       Date:  2008-04-28       Impact factor: 3.169

9.  Database resources of the National Center for Biotechnology Information.

Authors: 
Journal:  Nucleic Acids Res       Date:  2015-11-28       Impact factor: 16.971

10.  Why are RNA virus mutation rates so damn high?

Authors:  Siobain Duffy
Journal:  PLoS Biol       Date:  2018-08-13       Impact factor: 8.029

View more
  3 in total

1.  Mapping Data to Deep Understanding: Making the Most of the Deluge of SARS-CoV-2 Genome Sequences.

Authors:  Bahrad A Sokhansanj; Gail L Rosen
Journal:  mSystems       Date:  2022-03-21       Impact factor: 7.324

2.  Predicting Tissue-Specific mRNA and Protein Abundance in Maize: A Machine Learning Approach.

Authors:  Kyoung Tak Cho; Taner Z Sen; Carson M Andorf
Journal:  Front Artif Intell       Date:  2022-05-26

3.  Whole-genome sequencing and gene sharing network analysis powered by machine learning identifies antibiotic resistance sharing between animals, humans and environment in livestock farming.

Authors:  Zixin Peng; Alexandre Maciel-Guerra; Michelle Baker; Xibin Zhang; Yue Hu; Wei Wang; Jia Rong; Jing Zhang; Ning Xue; Paul Barrow; David Renney; Dov Stekel; Paul Williams; Longhai Liu; Junshi Chen; Fengqin Li; Tania Dottorini
Journal:  PLoS Comput Biol       Date:  2022-03-25       Impact factor: 4.475

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.