Literature DB >> 31846471

Menagerie: A text-mining tool to support animal-human translation in neurodegeneration research.

Caroline J Zeiss1, Dongwook Shin2, Brent Vander Wyk3, Amanda P Beck4, Natalie Zatz5, Charles A Sneiderman2, Halil Kilicoglu2.   

Abstract

Discovery studies in animals constitute a cornerstone of biomedical research, but suffer from lack of generalizability to human populations. We propose that large-scale interrogation of these data could reveal patterns of animal use that could narrow the translational divide. We describe a text-mining approach that extracts translationally useful data from PubMed abstracts. These comprise six modules: species, model, genes, interventions/disease modifiers, overall outcome and functional outcome measures. Existing National Library of Medicine natural language processing tools (SemRep, GNormPlus and the Chemical annotator) underpin the program and are further augmented by various rules, term lists, and machine learning models. Evaluation of the program using a 98-abstract test set achieved F1 scores ranging from 0.75-0.95 across all modules, and exceeded F1 scores obtained from comparable baseline programs. Next, the program was applied to a larger 14,481 abstract data set (2008-2017). Expected and previously identified patterns of species and model use for the field were obtained. As previously noted, the majority of studies reported promising outcomes. Longitudinal patterns of intervention type or gene mentions were demonstrated, and patterns of animal model use characteristic of the Parkinson's disease field were confirmed. The primary function of the program is to overcome low external validity of animal model systems by aggregating evidence across a diversity of models that capture different aspects of a multifaceted cellular process. Some aspects of the tool are generalizable, whereas others are field-specific. In the initial version presented here, we demonstrate proof of concept within a single disease area, Parkinson's disease. However, the program can be expanded in modular fashion to support a wider range of neurodegenerative diseases.

Entities:  

Mesh:

Year:  2019        PMID: 31846471      PMCID: PMC6917268          DOI: 10.1371/journal.pone.0226176

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Despite high rates of success reported in animal studies, many promising interventions for neurodegenerative and other complex diseases do not translate to effective therapies in humans [1-3]. Reasons for this translational gap are complex, [4] and occur at all stages of the preclinical [5, 6] and clinical [7, 8] continuum. In clinical trials, failed efficacy remains a primary translational roadblock, particularly in phase II and III trials [7, 9]. In contrast to this reality, discovery literature is heavily weighted toward publication of promising studies in animals [10-13]. In the area of neurodegeneration, this translational gap has fueled skepticism regarding the validity [3, 14, 15] and cost [16] of preclinical animal studies. Several factors that undermine the reliability of animal studies have been identified. These include insufficient rigor in animal study design and reporting [6, 17], publication bias [10, 12], over-reporting of significance [18, 19], over-reliance on non-clinical outcome measures [20], and entrenched use of certain model systems [3, 21, 22]. Together, these issues contribute to poor reproducibility of animal studies, and certainly worsen the translational gap [23]. To address this, reporting and design guidelines [24] have been adopted by numerous journals [25], and by major funding agencies [26]. These, once implemented, should improve rigor and reproducibility of preclinical studies, although widespread evidence for this is not yet available [27]. Regarding publication bias, greater reporting of negative studies [28, 29] would provide a more realistic view of actual preclinical efficacy. Whether these changes alone will improve translatability is unclear. In the case of neurodegenerative disease, the biological complexity of the research problem is a major hurdle. To understand mechanistic phenomena that may be obscured by this complexity, the prevailing approach is to apply hypothesis-based and reductionist methodology to genetically altered animal systems [30]. This has allowed insights that would not have otherwise been possible. However, a drawback of these systems is that their approaches are so specific that their results do not to generalize to other more complex situations i.e. there is limited external validity [2]. The challenges of extending findings from reductionist model systems to patient populations [2, 30–32] are immense. Profound differences in animal and human physiology [33] are a critical source of poor translation. Additionally, study design choices that influence generalizability extend beyond those needed to ensure unbiased study design. These are more contextually defined and concern the relationship between variables such as model choice [34] and mechanism of the intervention [22], integration of biomarker data with clinically relevant outcome measures, and use of progressive disease models and longitudinal study designs if neuroprotection is claimed. Addressing the potential generalizability of interventional animal studies is the impetus for the methods described in this paper. PubMed (https://www.ncbi.nlm.nih.gov/pubmed/) is the largest global public repository of biomedical literature [35]. It is a rich source of animal model discovery data that, if harnessed on a large scale using automated methods, could conceivably be used to inform translational potential of a given therapeutic mechanism or approach. Many examples of automated methods to more effectively search [36], curate [37] or generate new knowledge using literature-based discovery methods [38] from this resource have been described. However, none focuses specifically on the issue of animal-human translation, while accommodating the unfortunate realities listed above that undermine the reliability of published data, even in prestigious journals [39]. In this paper, we describe a text-mining approach that aggregates abstract level data to support subsequent manual evaluation of the generalizability, or external validity, of animal studies in an area that is particularly difficult to model effectively—neurodegeneration. Because we cannot recapitulate human neurodegenerative diseases with any single animal model [40], an alternate approach may be to aggregate evidence across a diversity of models [41] that capture different aspects of a multifaceted cellular process. Consistent results across such a diversity of systems may clarify whether an approach has translational potential [42, 43], particularly if these results extend across comparable clinically relevant outcome measures. In the initial version presented here, we describe the program design and its capacity to accurately extract data of potential translational relevance within a single disease area, Parkinson’s disease (PD). Animal models have been instrumental to development of symptomatic therapies for PD [44, 45]. However, PD confronts the same roadblocks as other neurodegenerative diseases in development of disease altering therapies [46, 47]. We demonstrate retrieval of the expected patterns of animal species and model selection [48, 49] for PD, and confirm previously identified [22] patterns of animal model use. This study paves the way for large scale evaluation of the entire PD corpus to identify those therapeutic approaches that have potential for clinical translation. Components of the tool are modular and readily adaptable to other neurodegenerative disease areas such as Alzheimer’s disease.

Materials and methods

We used a text mining approach to extract characteristics that we have previously determined, using manual curation, to be reliably present in abstracts, and to be translationally useful [22, 50]. These comprise information regarding therapeutic intervention, molecular target (s) or genes, species, model, overall outcome of the study and whether functionally relevant outcome measures [20] were reported. A dataset of 504 PubMed abstracts was manually annotated to develop/refine our approach and to validate it. To assess the utility of the approach, we also applied it to a larger dataset of 14,481 Parkinson’s disease abstracts. Below, we first present the data collection process. Next, the text mining components are described. We conclude this section by describing the validation of the approach on the manually annotated dataset, and on the 14,481abstract dataset. A pictorial summary of our approach is provided (Fig 1).
Fig 1

A system overview of the text-mining tool.

Modules (species, models, genes, interventions and outcomes) defining types of data returned (far right) from an abstract (far left) are shown on the first row. This is achieved through a combination of rule-based and machine learning methods utilizing established terminological resources (Unified Medical Language System (UMLS) Metathesaurus, NCBI Gene and Medical Subject Headings (MeSH)), additional text mining tools (SemRep, GNormPlus and the chemical annotator) and series of signal lists and rules (outcome classifiers and signal expression lists supplemented from databases such as OMIM).

A system overview of the text-mining tool.

Modules (species, models, genes, interventions and outcomes) defining types of data returned (far right) from an abstract (far left) are shown on the first row. This is achieved through a combination of rule-based and machine learning methods utilizing established terminological resources (Unified Medical Language System (UMLS) Metathesaurus, NCBI Gene and Medical Subject Headings (MeSH)), additional text mining tools (SemRep, GNormPlus and the chemical annotator) and series of signal lists and rules (outcome classifiers and signal expression lists supplemented from databases such as OMIM).

1. Data collection

a. Evaluation dataset

A total of 504 PubMed abstracts were used to develop and evaluate the text mining components. This dataset consists of two parts. The training set (406 abstracts), developed in an earlier study [22], spans years 2015–2016 and was used for development and refinement of our text mining components. The test set (98 abstracts), collected for this study, was drawn from year 2017 and was used to validate the components. Criteria used to select PubMed IDs (PMIDs) have been previously described [22]. Only interventional studies were included in training and test sets—these were identified by manual evaluation of the title and were defined as those in which the effect of an intervention (pharmaceutical, phytochemical, physical, genetic, behavioral or environmental) on the PD phenotype was examined. The annotation process was carried out using primarily Excel spreadsheets. Brat annotation tool [51] was used to verify intervention and gene annotations, which tended to be more complex. The annotation process differed in some respects between the characteristics considered: Species and model annotations were carried out in tandem at the abstract level. In addition to individual species and models relevant to the study (e.g., Mouse and MPTP, respectively), species/model combinations that were reported in the abstract were also annotated (e.g., Mouse/MPTP). Gene names were annotated at the mention level; that is, all occurrences of these terms in the title and abstract were annotated. They were also mapped to identifiers in the NCBI Gene database (https://www.ncbi.nlm.nih.gov/gene) and to concept identifiers in the UMLS Metathesaurus (https://uts.nlm.nih.gov/metathesaurus.html). For example, both dopaminergic D1 receptor and D1 receptor were mapped to NCBI Gene identifier 24316. Interventions/disease modifiers were annotated at the abstract level. For meaningful evaluation, they were also mapped to standard identifiers in the MeSH vocabulary (https://meshb.nlm.nih.gov/search), the NCBI Gene database, the UMLS Metathesaurus. For example, both selegiline and deprenyl were mapped to the MeSH descriptor D012642. Overall outcome was annotated at the abstract level, and the outcome value (promising, negative, mixed, and other) was determined based on the title and the last two sentences of the abstract. These terms are defined more fully in the section below. While individual outcome measures (e.g., dyskinesia), polarity-signaling expressions (e.g. reduced), and sentence-level outcomes were annotated in the training set, these were ultimately not used. We did not perform specific annotation for functional outcomes, though a list of functional outcome terms was compiled from the outcome measures annotated in the training set. In the test set, only overall outcome was annotated. Two of the authors (CJZ and CS) annotated the overall outcome in 54 abstracts independently to establish guidelines for annotating this element. The interannotator agreement was found to be 0.55 (Cohen’s κ, moderate agreement). CJZ later adjudicated these annotations and annotated the rest of the abstracts.

b. Text mining components

Text mining components were developed in Java programming language. Existing natural language processing (NLP) tools underpin the components and are further augmented by various rules, term lists, and machine learning models. NLP tools used are the following: SemRep [52] is a rule-based system that maps text in PubMed abstracts to UMLS Metathesaurus concepts with broad semantic classes, such as Pharmacologic Substance or Mammal (and relationships between concepts, which were not used in this study). GNormPlus [53] is a machine learning-based system that identifies gene/protein terms in text and maps them to NCBI Gene identifiers. Chemical annotator is a dictionary-based method developed at the U.S. National Library of Medicine that extracts chemical entities from text and normalizes them to MeSH identifiers. We implemented our approach as six individual modules: a) species, b) model, c) genes, d) interventions/disease modifiers, e) overall outcome and f) functional outcomes. Each module extracts relevant information from a list of abstracts and the results are stored in a relational database (MySQL) for various types of analyses. Details of each module are provided below. Species: Humans and animal species considered in this study included those most commonly used in research: non-human primates, rodents, larger less commonly used mammals (e.g., dog, cat, rabbit) and a selection of commonly used non-mammalian species (e.g., fly, worm, frog and fish). To identify these species, we used rules that map UMLS concepts identified by SemRep and other signal expressions to species terms. For example, the UMLS concept C0008976:Clinical Trials was mapped to species Human and C0006764:Callithrix to Marmoset. These rules account for synonyms describing the same species (e.g., cat and feline for Cat), mapping them to a single species term to facilitate subsequent analysis. Terms identified at the sentence level were consolidated to summarize the spectrum of species mentions at the abstract level. Two rules were used to exclude species terms that, while mentioned in the abstract, may not be relevant to the study under examination: Species terms identified only in the background section of the abstract (if structured) or in the first sentence of the abstract (if unstructured) were excluded. We found this approach was needed to avoid extraneous collection of species terms used in background text summarizing previous findings across species. Human was excluded as species if it was found only in the conclusion section of the abstract (if structured) or within the final 15% of the sentences of the abstract (if unstructured). This was done to ensure that abstracts for animal studies that included discussion about potential in human populations were assigned as animal studies, not human studies. Additionally, we established a hierarchy of terms so that those referring to specific species (e.g., Macaque) were returned in favor of more general terms (e.g., Non-human primate) occurring in the same abstract. In the event that no specific species terms (e.g., Macaque) were used in the abstract, more general terms can be extracted. Rules to collect species data are generalizable to any area of study. Model: In contrast, rules used to collect model data were partly Parkinson’s disease-specific. The methodology used for models was largely the same as that used for species. A series of signal expressions were used to capture commonly used models (e.g., 1-methyl-4-phenyl-1,2,3,6 tetrahydropyridineMPTP and 6-hydroxydopamine6-OHDA, and others) that are largely specific to PD. In addition (and generalizable to any disease area), signal expressions were included to identify the category of genetically altered animal models. These can capture genetically altered models regardless of disease area, and were thus PD-independent. These signal expressions also identified human studies of patients with mutations in PD associated genes. As for species, model was identified at the sentence level, but consolidated at the abstract level to provide a list of the spectrum of model mentions as the final readout. Species and models were identified in tandem. If a species term and a model term are identified within a pre-specified window of each other in the text (3 phrases), they were also paired at the abstract level to make the species/model link explicit. Genes: When considered together with data collected from other modules, gene data allows the user to assess the extent to which molecular pathways or potential targets associated with a given intervention are preserved across model systems. Extraction of gene terms is underpinned by GNormPlus [53] and SemRep [52]. SemRep results were filtered to concepts with one of three UMLS semantic types: Gene or Genome; Amino Acid, Peptide, or Protein; and Enzyme. We augmented GNormPlus and SemRep with a set of gene names with mutations/polymorphisms known to be associated with PD in the Online Mendelian Inheritance in Man (OMIM) database. This dictionary contains 28 such gene names and 818 synonyms associated with them (e.g., parkin for PRKN and alcohol dehydrogenase 3 for ADH1C). This list was designed to address potential misses by GNormPlus and SemRep, that can be considered essential for an automated tool to extract. The module searches the abstract for the presence of synonyms and records the gene name associated with the synonym, if found. This list can be supplemented and updated by code that can extract a list for a given disease from OMIM. Additional rules were used to address several kinds of false positive errors that were identified during the training process. These rules filter out ordinal numbers (e.g., 2nd), confidence intervals (e.g., CI 0.95), gene names that map to common English words (e.g., all, impact), non-specific genetic terms (e.g., transcription factor, protein, candidate gene), and terms relevant to other modules (e.g., dopa, mptp(+), and liraglutide). Other terms were excluded because they were not annotated in the training set, even though they seemed legitimate from an extraction point of view (e.g., neurotrophin, glutamate, acetylcysteine). To facilitate subsequent analysis of returned data, NCBI gene identifiers (for gene terms returned using GNormPlus or the OMIM dictionary) or CUIs (for gene terms returned using SemRep) were also collected. In this way, synonyms associated with the most current gene nomenclature mapped to a single identifier. Interventions/Disease Modifiers: This module aims to capture pharmaceutical, phytochemical, physical, genetic, behavioral or environmental entities that could alter a disease phenotype. The chemical annotator, in addition to SemRep and GNormPlus, provided the basis for this module, augmented with a list of essential interventions. To address the spurious entities identified with the NLP tools, we also use a list of exclusion terms. The chemical annotator maps identified interventions to MeSH identifiers. UMLS concepts identified with SemRep were filtered based on their semantic types. The filter included the following types: Amino Acid, Peptide, or Protein; Biologically Active Substance; Chemical; Food; Hazardous or Poisonous Substance; Hormone; Inorganic Chemical; Organic Chemical; Substance; and Vitamin. GNormPlus was used to identify genetic interventions from the titles only. The list of essential terms was identified in manually annotated sets from two previous publications [22, 54] and was supplemented by interventions given in Alzforum therapeutics (https://www.alzforum.org/therapeutics; searched September 10, 2018) and the Michael J Fox Foundation (https://www.michaeljfox.org; searched September 10, 2018). Interventions for both PD and Alzheimer’s disease (AD) were included, recognizing that some interventions (particularly for neuropsychiatric complications) are shared across different neurodegenerative diseases [55]. The exclusion list was based on our observations on the training set and included generic terms like “treatments”. Database identifiers were used to account for synonyms and consolidate sentence-level terms to summarize the spectrum of interventions identified at the abstract level. Gene mentions in the title only are extracted in this module. Because many genes are mentioned in the body of the abstract, using the title only this enriches data for those molecular entities that are thought to influence the PD phenotype significantly and are potential therapeutic targets. Overall outcome: Our intent with this module was to distill the conclusion of the study (as defined by the authors) regarding the overall potential of the study to influence trajectory of the disease. Four final categories were created: those in which outcomes held therapeutic promise (PROMISING), those with adverse effects (negative), those with both promising and adverse effects (mixed) and those in which outcomes were indeterminate (other). This module is implemented as an ensemble of machine learning models. Specifically, support vector machine (SVM) models were developed for positive and negative outcomes. The positive outcome model classified each abstract as positive or not-positive. The negative outcome model classified each abstract as negative or not- negative. The final decision on overall outcome of a study was made based on the predictions of the two models: positive + not-negative → PROMISING positive + negative → mixed not-positive + negative→ negative not-positive + not-negative→ other Features extracted from the title and the last two sentences (the context) were used for classification. The features were adapted from Niu et al.[56], who used a classification approach to determine polarity of clinical outcomes (no outcome, positive, negative, neutral outcome). These features are the following: n-grams: unigrams and bigrams from the context, stemmed using Porter stemmer [57]. Change word features: Two sets of words are defined for these features: MORE (15 terms, e.g., increase) and LESS (34 terms, e.g., alleviate). The tag MORE is added to all words between a MORE word and the next punctuation, and the tag LESS to the words after a LESS word, aiming to capture the scope of these words. Change/polarity word co-occurrence: These capture co-occurrence of change words and polarity words, which can indicate positive/negative assessment. To extract these features, two additional set of terms were created for GOOD (49 terms, e.g., ameliorate) and BAD (12 terms, e.g., exacerbate) terms. 4 features were generated by combining the four classes: MORE GOOD, MORE BAD, LESS GOOD, LESS BAD, and if terms from respective categories appeared within a pre-specified window (4 words), that feature was set to 1. We also used terms that are categorized as Disorders in the UMLS semantic groups [58] as BAD terms. Therefore, a sentence with the fragment alleviate Parkinson’s disease would be encoded by setting the LESS BAD feature to 1. Negation: All words modified by the negation no in a sentence are appended with NEG, and used as additional features (e.g., no significant change → {significant_NEG, change_NEG}). Semantic type: We encode each UMLS semantic type as a feature and set this feature to 1, if a concept with a given semantic type exists in the context. LIBLINEAR implementation of linear SVM [59] was used for training the models. Functional outcomes: Functional outcomes were defined as those that reflected the effect of an intervention on measurable variables in a living organism (e.g., physical movement or survival). A set of such outcome terms was collected manually from the training set. These include terms common in human clinical trials (e.g., the Unified Parkinson’s Disease Rating Scale), general terms reflecting neurologic function (e.g., bradykinesia) and terms specific to animal model studies (e.g., rotarod performance or turning behavior). We then simply checked whether an abstract contained any of these terms, and categorized the abstract as YES, if it did, and as NO, if no functional outcome measure was detected. Additional data: We collected metadata related to abstracts from PubMed records for subsequent analysis. These included year of publication, journal, and publication type.

2. Evaluation

a. Intrinsic evaluation

After developing and refining the modules using the training data, we validated them on the test set (98 abstracts). We used standard information extraction evaluation metrics: precision (or positive predictive value), recall (sensitivity), and F1 score (the harmonic mean of precision or recall). All modules were evaluated at the abstract level. Gene/protein and intervention modules were also evaluated at the mention level. For the abstract level evaluation, multiple extractions of the same term are consolidated into a single extraction (i.e., only unique terms are considered). For the mention level evaluation, all instances of a term are considered separately. Additionally, the intervention module was evaluated based on title only (i.e., only interventions extracted from the title and the title annotations were compared). For evaluating gene and intervention modules, we also used both exact matching and approximate matching. In exact matching (stricter evaluation), a term extracted by the module should exactly match an annotated term with respect to term boundaries to count as a true positive. In approximate matching, overlap of the terms is acceptable. Given that the annotated dataset is relatively small (504 examples for overall outcomes) and it is standard to evaluate machine learning models that are trained on small datasets using cross-validation, we evaluated the overall outcome module using 10-fold cross-validation on the full dataset, instead of using the training-test split. We repeated the cross-validation experiment 50 times and reported the average of results. Overall outcome models that were used in the large scale evaluation (below) were trained on the full dataset. We also compared evaluation results from our program, where applicable, with results from comparable programs applied to the same 98-abstract test set. These baseline systems include PubTator [18], GNormPlus [44], and SemRep [43] with semantic filtering.

b. Large scale evaluation

To assess the utility of our approach on a larger scale, we collected a dataset of 14,481 abstracts using “Parkinson’s disease” as the sole search criterion in PubMed (searched 11/14/2017). These spanned three time points across 10 years (2008; 3433 PMIDs, 2012; 4727 PMIDs and 2017; 6321 PMIDs; S1 File). In contrast to the evaluation dataset, these abstracts were not enriched for interventional studies by subsequent manual selection of studies with titles implying use of an intervention. This allowed us to assess utility of the program to extract interventional studies from the broadest search possible. Data was analyzed manually using Excel or using a series of queries to aggregate studies by their features (species, model, intervention, outcome, functional outcome, and genes) within and across publication years. Specific queries that implemented logical conjunctions of feature sets (e.g. Species = ‘Mouse’ AND outcome = ‘PROMISING’) were used to generate lists of publications that met the criteria. For each query the list of qualifying publications was outputted along with a set of summary statistics such as the count and proportion of publications per year to facilitate further investigation and data visualization. Using standardized query methods affords the opportunity to specify arbitrarily complex queries to support user-driven exploration of the database. Specific comparisons are given by subheading in the results section. Data access: All supporting data is available as Supplementary Data or on request from CJZ or HK.

Results

1. Intrinsic evaluation of the program

Precision, recall and F1 scores for each module were assessed using exact matching (all modules) as well as approximate matching (Genes and Interventions modules) at the title, abstract and mention level as shown in Table 1.
Table 1

Results of program evaluation on a 98-abstract test set.

ModulePrecisionRecallF1Accuracy
A. Species and model module
Species (exact matching)0.840.900.87-
Model (exact matching)0.870.910.89-
Species/Model (exact matching)0.720.790.75-
B. Genes module
Mention level (approximate matching)0.880.850.87-
Mention level (exact matching)0.790.770.78-
Abstract level (approximate)0.800.830.82-
Abstract level (exact)0.740.760.75-
C. Interventions/disease modifiers module
Abstract level (approximate matching)0.960.430.59-
Abstract level (exact matching)0.850.370.52-
Title only (approximate)0.950.830.89-
Title only (exact)0.830.730.78-
D. Overall outcome evaluation (10-fold cross-validation)
Promising0.911.00.95-
Negative0.960.700.81-
Mixed0.860.590.70-
Other0.790.530.64-
0.90
E. Functional outcome measure
Functional outcome (yes)0.870.900.89-
0.86

Species and model modules

Using exact matching, we obtained higher recall compared to precision; results were similar for species and models. The performance for species/model combination was lower, as expected, since both the species and model needed to be correct and the pre-specified window size (3 phrases) can cause additional errors. Genes module: We obtained similar precision and recall with the genes module. When approximate matching was used for evaluation, the results were better both at the abstract and the mention level, indicating that identifying precise gene/protein term boundaries in text is a challenge.

Intervention/disease modifier module

The results were obtained by considering interventions extracted from the title only using semantic type filtering. Performance at the abstract level was significantly lower, because even though the evaluation was performed at the abstract level, the module still extracted interventions only from the title. The difference between the performance of the system at the abstract level and title only evaluation (0.52 vs. 0.78 F1 score with exact matching) indicates that a significant number of terms defined in our program as interventions/disease modifiers were discussed in the abstract only.

Overall outcome and functional outcome measure

The results of 10-fold cross-validation for overall outcome prediction are provided in Table 1D. Note that the results are the mean average of 50 cross-validation experiments. In contrast to overall outcome, functional outcome (yes/no) results were based on the test set, since the functional outcome term list was derived from the training set. Next, we compared the results of our program to those achieved by comparable baseline programs PubTator [37], GNormPlus [53] and SemRep [43] with UMLS semantic type filtering (Table 2).
Table 2

Comparison with evaluation results from comparable baseline programs (98-abstract test set).

ModulePrecisionRecallF1Accuracy
A. Species and model module
Species (Menagerie; exact matching)0.840.900.87-
Species (PubTator)0.740.660.70-
B. Genes module
Mention level (Menagerie; exact matching)0.790.770.78-
Mention level (GNormPlus; exact matching)0.930.540.69-
C. Interventions/disease modifiers module
Abstract level (Menagerie; exact matching)0.850.370.52-
Abstract level (Chemical and Substances in UMLS; exact matching)0.480.260.34-
D. Overall outcome evaluation
Promising (Menagerie; 10-fold cross-validation)0.911.00.95-
Overall accuracy (Menagerie)0.90
Promising (majority vote)0.761.00.86-
Overall accuracy (majority vote)0.76
E. Functional outcome measure
Functional outcome (Menagerie; yes)0.870.900.890.86
Functional outcome (Finding and Sign or Symptoms UMLS; yes)0.670.870.760.64
Our program achieved improved precision, recall and F1 scores for Species when compared to PubTator. For genes, GNormPlus output was used as-is, with exact matching. Our program achieved improved recall and F1 scores, at the cost of some precision loss. For interventions, concepts within the semantic group Chemicals and Drugs in UMLS were used- our program achieved improved values by all criteria. As our program utilizes annotated signal terms in addition to UMLS concepts, this was expected. For outcomes, a majority vote, which assigned PROMISING overall outcome to all abstracts, was used, and achieved lower performance than our program. For functional outcomes, concepts with the UMLS semantic types Finding and Sign or Symptoms were used as functional outcomes. Our program achieved significantly improved values by all criteria, again most likely due to annotated signal terms specifically directed at PD. Next, in the absence of a comparable baseline program for animal models or functional outcome measures, we assessed this aspect in a larger scale evaluation.

2. Large-scale evaluation results

The evaluation goals for this dataset were to assess those modules that could not be assessed using other programs (models, overall outcome and functional outcome measure) and to demonstrate how data could be aggregated to evaluate the diversity of systems across which a therapeutic approach or cellular mechanism had been assessed.

a. Descriptive characteristics

Overall outcome by publication types: Each abstract (defined by a unique PMID) was assigned primary or secondary data status on the basis of its associated publication type. In all three years, primary data constituted between 70% and 75% of all publications (S1 Table). Because publications are biased towards those reporting promising outcomes [10], we hypothesized that this tendency would be amplified by including secondary publication types in our comparisons. To test this hypothesis, we compared the proportions of outcome types in primary data sources alone, to those including both primary and secondary data sources. In general, approximately 70% of studies report a promising outcome with proportions in primary data sources being slightly lower than in all data sources (S1 Table). In both datasets, a slightly increasing trend in promising reports across 10 years was noted. Primary data was used for subsequent comparisons. Species and animal model use over 10 years: Overall, studies in humans prevailed (41%), followed by studies using mice and rats (approximately 10%), and non-human primates (mentioned in 1% or fewer of studies; Table 3). Predominance of rodents and use of non-human primates as experimental models in PD is well-established [41, 60]. Increasing trends of species mentions were strongest for zebrafish, drosophila and worm studies, reflecting increased use of lower species in mechanistic studies [61-63]. With the exception of marmosets, non-human primate use was flat or declining over time. 21% of studies over the 10-year period did not report species, while an approximately further 13% of studies mentioned animals in generic terms (animal model, rodent, non-human primate).
Table 3

Species mentions in Parkinson’s disease research (2008–2017).

2008–2017200820122017
Human4944 (0.41)103716412266
Species not reported2615 (0.21)6268951094
Animal model1336 (0.11)347471518
Mouse1271 (0.10)261439571
Rat1107 (0.09)313373421
Rodent, unspecified187 (0.02)506968
Non-human primate155 (0.01)475652
Macaque70 (0.006)173221
Marmoset24 (0.002)8412
Vervet monkey5 (0.0004)401
Baboon2(0.0002)200
Cat38 (0.003)71714
Dog10 (0.0008)343
Rabbit4 (0.0003)004
Frog5 (0.0004)122
Zebrafish30 (0.003)3720
Drosophila132 (0.01)274857
Worm61 (0.005)122722
Yeast64 (0.005)222715
Total12060278741125161

Total numbers of publications (defined by unique PMID) are given by time period. Proportions are included in parentheses for the 2008–2017 period. Primary data only are used.

Total numbers of publications (defined by unique PMID) are given by time period. Proportions are included in parentheses for the 2008–2017 period. Primary data only are used. Model co-mentions for animals reported at the species level are given in Table 4. Rat and mouse models were by far the most heavily cited—of these the expected preference for (6-hydroxydopamine) 6-OHDA use in the rat and for (1-methyl-4-phenyl-1,2,3,6-tetrahydropyridine) MPTP use in mice and non-human primates was captured [48, 64]. Use of these two toxins and their variant applications, including hemi-parkinsonism, levodopa induced dyskinesia (LID) and 1-methyl-4-phenylpyridinium (MPP) toxicity constituted the majority of model choices. Chronic toxic models such as MPTP/probenecid, rotenone, paraquat, lacatcystin and maneb [64], traditional pharmacologic models (reserpine, haloperidol and galantamine) [65] and inflammatory models [66] were less common, but consistently used. As expected, mice [67] and lower species were the most heavily utilized genetically altered species.
Table 4

Species/model mentions in Parkinson’s disease research (2008–2017).

Species and model200820122017Total PMIDs/ Species (% toxic models)
Macaque
MPTP7137
Hemi-parkinsonian11029 (100)
Marmoset
MPTP636
1BnTIQ20017 (88)
Rat
6-OHDA93128119
Hemi-parkinsonian262231
MPP975
Lipopolysaccharide induced model859
Genetically altered model669
Levodopa-induced dyskinesia61010
Rotenone61735
Haloperidol456
MPTP2109
Reserpine252
Tremulous jaw movement model200
Galantamine100
Lactacystin103
Maneb100
Paraquat112624 (78)
Mouse
MPTP6998138
Genetically altered model398688
6-OHDA102036
Lipopolysaccharide induced model4514
Paraquat454
Lactacystin313
Reserpine311
Rotenone2215
Levodopa-induced dyskinesia115
Maneb132
MPP156
MPTPp101678 (58)
Zebrafish
Genetically altered model101
MPTP1025 (60)
Drosophila
Genetically altered model689
Paraquat11429 (0)
Worm
Genetically altered model451
Rotenone10011 (0)
Yeast
Genetically altered model2103 (0)

Models are listed by species in which they are mentioned, with 6-OHDA and MPTP models and their variants shaded in gray (column 1). Total numbers of publications in which models and species are co-mentioned in black in the last column. Of these, proportions of 6-OHDA and MPTP models and their variants (toxic models) are shown in parentheses.

Models are listed by species in which they are mentioned, with 6-OHDA and MPTP models and their variants shaded in gray (column 1). Total numbers of publications in which models and species are co-mentioned in black in the last column. Of these, proportions of 6-OHDA and MPTP models and their variants (toxic models) are shown in parentheses. Results confirm that the expected spectrum of vertebrate and invertebrate animal species and models used in PD research were captured [22, 68–71]. Additionally, the previously noted dominance of 6-OHDA and MPTP-associated toxic models [22] was reiterated. These results confirmed that our tool identifies expected patterns of species and model use for the field. Next, we assessed its capacity to extract large-scale patterns regarding therapeutic interventions and gene mentions. Interventions/disease modifier mentions over 10 years: 1862 unique entities (derived from UMLS or MeSH terms) were identified in the entire dataset. These were ranked by the number of associated publications at each time point, allowing identification of the most highly studied entities, as well as the trend of study over time (Fig 2, S2 Table).
Fig 2

Number of publications, by intervention or disease modifier (2008–2017).

Pharmacologic and non-pharmacologic interventions, as well as genes or co-morbidities that influence the Parkinson’s disease phenotype are extracted from the title of the abstract only. Gene mentions in the title enrich data for those molecular entities that are thought to influence the Parkinson’s disease phenotype significantly and are potential therapeutic targets. Bars: Most frequently mentioned 50 of 1862 unique modifiers. Line: Fold change over 5 years, red line = 1. The five most mentioned entities in 2017 are indicated in green boxes; those with highest fold change are shown in white boxes.

Number of publications, by intervention or disease modifier (2008–2017).

Pharmacologic and non-pharmacologic interventions, as well as genes or co-morbidities that influence the Parkinson’s disease phenotype are extracted from the title of the abstract only. Gene mentions in the title enrich data for those molecular entities that are thought to influence the Parkinson’s disease phenotype significantly and are potential therapeutic targets. Bars: Most frequently mentioned 50 of 1862 unique modifiers. Line: Fold change over 5 years, red line = 1. The five most mentioned entities in 2017 are indicated in green boxes; those with highest fold change are shown in white boxes. Results are consistent with published data describing established treatments [72, 73] and recent approaches [74, 75]. Gene mentions over 10 years: This module extracts gene or molecular concept mentions in the entire abstract body, and thus identifies a larger group of genes and mechanisms than in the previous module (Fig 3, S3 Table). 3311 unique entities (derived from UMLS or NCBI gene terms) were identified in the entire dataset. Multiple synonyms describing the same entity mapped to a common UMLS or NCBI identifier. UMLS derived terms exhibit some redundancy with NCBI origin terms (e.g. dopamine transporter and SLC6A3) and capture groups of functionally related entities (e.g. D2 receptors) or molecular concepts (e.g. proteome). Proteins used as routine immunohistochemical markers or standard molecular reagents (e.g. tyrosine hydroxylase and GFAP) are captured. The expected dominance of alpha-synuclein was evident [76].
Fig 3

Number of publications, by gene/molecular concept mention (2008–2017).

Gene mentions in the entire abstract body were extracted to identify a larger group of genes and mechanisms that may not be entities of primary therapeutic interest (extracted in the Interventions module), but that contribute in some way to an understanding of the disease. Bars: Most frequently mentioned 50 of 3311 entities defined by NCBI gene terms or UMLS identifiers (primary data sources). Line: Fold change over 5 years, red line = 1. The five most mentioned entities in 2017 are indicated in green; those with highest fold change in white.

Number of publications, by gene/molecular concept mention (2008–2017).

Gene mentions in the entire abstract body were extracted to identify a larger group of genes and mechanisms that may not be entities of primary therapeutic interest (extracted in the Interventions module), but that contribute in some way to an understanding of the disease. Bars: Most frequently mentioned 50 of 3311 entities defined by NCBI gene terms or UMLS identifiers (primary data sources). Line: Fold change over 5 years, red line = 1. The five most mentioned entities in 2017 are indicated in green; those with highest fold change in white.

b. Integrating information across datasets

The Interventions/Disease Modifier and Genes modules allow the user to rapidly rank these variables by frequency of mention and assess how these fluctuate over time. Next, we integrated additional information regarding species, model, outcome and gene mention data to demonstrate how studies can be aggregated across a diversity of model systems. Two highly mentioned entities extracted by the Interventions/Disease Modifiers module, L-DOPA and alpha synuclein (Fig 2) were selected to illustrate this (Fig 4).
Fig 4

A schematic comparison of species, model, outcome and functional measure characteristics for an established intervention (L-DOPA) and a target of experimental therapeutic interest (alpha-synuclein).

While both interventions were associated with a majority of promising outcomes, these prevailed in studies mentioning L-DOPA (65% vs 56%). In contrast, the proportion of functional measure reporting was far higher in L-DOPA associated studies (90%) than in alpha-synuclein associated studies (27%). Patterns of species and model use confirmed the expected utility of toxic models in L-DOPA associated studies compared to genetic models in alpha-synuclein associated models. Successive filtering by any of these variables can produce a subset of PMIDs for manual inspection (PMIDs not shown in this schematic). Image produced using RAW graphs (http://app.rawgraphs.io/).

A schematic comparison of species, model, outcome and functional measure characteristics for an established intervention (L-DOPA) and a target of experimental therapeutic interest (alpha-synuclein).

While both interventions were associated with a majority of promising outcomes, these prevailed in studies mentioning L-DOPA (65% vs 56%). In contrast, the proportion of functional measure reporting was far higher in L-DOPA associated studies (90%) than in alpha-synuclein associated studies (27%). Patterns of species and model use confirmed the expected utility of toxic models in L-DOPA associated studies compared to genetic models in alpha-synuclein associated models. Successive filtering by any of these variables can produce a subset of PMIDs for manual inspection (PMIDs not shown in this schematic). Image produced using RAW graphs (http://app.rawgraphs.io/). Patterns of species and model use: Identified trends were consistent with symptomatic treatment of striatal denervation (L-DOPA associated studies) compared to the mechanistic approaches inherent in understanding the role of alpha-synuclein in PD, and its potential as a therapeutic disease-altering target [77]. In humans, L-DOPA is used as a primary or comparator treatment in humans or animals [78], as well as in human and animal studies of levodopa induced dyskinesia (LID) [79]. Animal studies in which L-DOPA was mentioned relied heavily upon MPTP and 6-OHDA induced models of striatal denervation. Non-human primates are used more frequently in L-DOPA associated studies, consistent with their prominent role in development of treatments for motor symptoms of PD [60]. This trend was consistent with reported literature describing symptomatic models of PD [48]. Studies examining the role of alpha synuclein in animals were dominated by those in mice [67], followed by rats. Because alpha-synuclein is the major component of hallmark Lewy bodies in Parkinson’s disease, human reports were also highly represented. Consistent with the literature [61–63, 68] alpha synuclein focused studies were quite well represented in fish, invertebrate models and yeast. Patterns of overall outcome and functional outcome reporting: The proportion of promising or mixed overall outcomes (65% and 27% respectively) was higher for L-DOPA associated studies than for alpha-synuclein associated studies (56% and 16% respectively). Conversely, the proportion of negative outcomes was higher in alpha-synuclein associated studies (14% vs 5%). This reflects a previously reported trend in animal studies in which mechanistic insight is gained through genetic interventions that worsen the phenotype [22]. L-DOPA associated studies reported functional outcome measures at a much higher rate (90%) than those assessing the role of alpha-synuclein (27%). We have noted previously that reporting of functional outcome measures appears to associate with those interventions that are approved for use in PD [22]. Gene mentions by intervention: Because a user may wish to explore potential molecular pathways associated with a given intervention, we compared mentions of genes (e.g. ERK1) and gene-related terms (e.g. ERK1-2 Pathway) in L-DOPA or alpha synuclein associated abstracts. We extracted 388 and 689 unique entities (defined by NCBI or UMLS identifiers) respectively (S4 Table). All studies in which alpha-synuclein appeared in the title reported genes in the abstract, whereas 130/353 studies with L-DOPA in the title did not, consistent with the often functional rather than molecular nature of L-DOPA associated studies. The most 15 significantly enriched pathways for L-DOPA and alpha-synuclein associated studies (https://reactome.org/; [80]) are shown in S5 Table. As expected, pathways describing dopaminergic, serotonergic and NMDA receptor signaling prevail in L-DOPA associated studies[81], whereas those describing a broad range of cellular events including protein degradation and trafficking, apoptosis and transcriptional control prevail in alpha synuclein associated publications [82]. Gene mention by species: Defining the distribution of individual gene mentions across species would provide a starting point to assess how generalizable its associated cellular mechanisms could potentially be to humans. To illustrate this, the species distribution of individual genes co- mentioned with alpha-synuclein associated studies is given in S6 Table. Apart from SNCA itself, other genes are mentioned in 4 or fewer species. Fig 5 illustrates those pathways (https://reactome.org/; [80]) that are shared across 3 or more species in L-DOPA and alpha synuclein associated studies.
Fig 5

Pathways identified in three or more species for L-DOPA and alpha synuclein associated studies.

Gene lists identified in S4 Table were organized by species, then submitted to the Reactome (https://reactome.org/) to obtain pathways shared by species. Publications associated with an established and approved intervention such as L-DOPA identify pathways that are shared across a greater number of species than those identified by alpha synuclein related studies.

Pathways identified in three or more species for L-DOPA and alpha synuclein associated studies.

Gene lists identified in S4 Table were organized by species, then submitted to the Reactome (https://reactome.org/) to obtain pathways shared by species. Publications associated with an established and approved intervention such as L-DOPA identify pathways that are shared across a greater number of species than those identified by alpha synuclein related studies. Functional outcome reporting by intervention type: We noted that the difference in functional outcome reporting between L-DOPA and alpha synuclein related studies was marked. To assess whether the pattern of functional outcomes reporting was limited to these two entities, or extended in a more general fashion to approved or experimental therapeutic approaches, we manually classified entities extracted by the Interventions/Disease Modifiers module as Established or Experimental (S7 Table). Those interventions (and their associated targets, e.g., dopamine receptors) that are already approved in the United States for PD and its complications, or clinically utilized supportive therapies such as exercise or physical therapy, were classified under Established. These were defined according to literature reviews [72, 73, 83] and all achieve symptomatic relief rather than slowing of disease progression. The remainder were classified as Experimental and include a heterogeneous group of terms, including gene terms. Overall, studies mentioning Established interventions reported a functional outcome measure in 79% of studies, compared to 45% of those studying Experimental interventions (Table 5). When this finding was broken down by species, functional outcome measures were reported across most species used to test Established interventions. A similar finding was noted in far fewer species in which Experimental interventions were co-mentioned. These data are consistent with previous observations[20, 22].
Table 5

Reporting of functional outcome measure by intervention type.

Functional outcome measure reported
Total PMIDsYES% YES
ESTABLISHED2882270.79
EXPERIMENTAL294513380.45
ESTABLISHED, by speciesTotal PMIDsYES% YES
Human67610.91
Macaque13110.85
Marmoset551.00
Baboon100.00
Dog100.00
Cat210.50
Rat1611290.80
Mouse78520.67
Zebrafish111.00
Worm111.00
EXPERIMENTAL, by speciesTotal PMIDsYES% YES
Human13406720.50
Macaque23140.61
Marmoset12100.83
Vervet monkey520.40
Baboon100.00
Dog520.40
Cat2080.40
Rat7523580.48
Mouse8893390.38
Rabbit310.33
Frog441.00
Zebrafish20130.65
Drosophila87340.39
Worm3680.22
Yeast4640.09

Discussion

Novel discoveries, models and methods using animals constitute a significant portion of federally funded research [16] and are viewed as a means to improve drug discovery rates in the pharmaceutical industry[84]. The Investigative New Drug application (IND) represents a bridge between preclinical and clinical stages, however data in support of an IND focuses primarily on pharmacokinetic/pharmacodynamic data and toxicology, rather than efficacy [85]. Evidence for the latter often has its roots in academic discovery literature, which is heavily weighted toward publication of promising studies in animals [10, 11]. This underscores the need for accurate and realistic assessment of discovery data emerging from academia. Generalizability, or external validity, is the extent to which research findings derived in one experimental context can be reliably applied to another. Most experimental animal systems tend to be reductionistic (defined as minimizing experimental variables to isolate the phenomenon of interest) [30]. Therefore, attempting to extrapolate from these to complex human systems represents a major translational hurdle [31, 32]. Because common neurodegenerative diseases appear to be a uniquely human phenomenon, it is likely that ideal translational animal models will not materialize. Instead, identifying which mechanisms or approaches have demonstrated consistent results across different animal systems that reflect aspects of the human disease may have a greater chance of translating to humans [42]. This paradigm is central to the Food and Drug Administration Animal Rule, a mechanism through which a product may be approved when human testing is not feasible for ethical reasons. In this scenario, efficacy must be demonstrated across more than one species using animal study endpoints that are clearly related to the desired outcome in humans [43]. Using this argument as a basis, our program allows the user to assess the diversity of animal (including human) systems across which these have been studied, and to filter these by various criteria, including whether functional measures [20] have been used to determine efficacy. This allows the user to rapidly organize abstract data for an entire disease corpus to yield a smaller subset of papers that can be manually assessed for potential generalizability of the mechanism from animals to humans.

Comparison with existing resources

Our tool shares some features with those utilized by PubTator [37]. Both our program and PubTator are able to retrieve species (using distinct algorithms) and gene (using the same program GNormPlus). Models, outcomes and reporting of functional outcome measures are unique to our program. A conceptually similar tool is described by Zwierzyna and Overington, 2017[86], in which descriptions of drug screening-related assays in rodents can be screened for mentions of genetic and experimental disease models, treatments, phenotypic readouts and disease indications. Datasets are retrieved from ChEMBL, an open-source manually curated database of bioactive molecules utilized in preclinical drug discovery.

Novel aspects of our program

PubMed represents a major source of preclinical data that has embedded within it, substantial efficacy data. Our program is uniquely able to organize, filter and compare large scale human and animal model abstract data by translationally relevant criteria. We have demonstrated that we are able to recapitulate expected patterns of mechanistic discovery, species distribution and animal model use in the PD field. The Interventions/Disease Modifier and Genes modules allow the user to rapidly rank individual genes or therapeutic approaches by frequency of mention, and determine how these have changed over time. This approach identifies established approaches, as well as those that are recently emerging. It is this latter group that is of most interest, as accurate assessment of translational potential at this stage could accelerate therapeutic development. Within this group, over-reporting of promising results presents the first hurdle in assessing potential efficacy. Consistent with previous reports [11, 12], our program identified promising outcomes in the majority of studies, confirming that use of this criterion alone is not useful in identifying promising therapies. To overcome this, we are able collect additional species, model and outcome data that can be used to assess potential generalizability of a given gene or therapeutic approach. In agreement with previous observations[20, 22], functional outcome measures were more highly reported in established therapies compared to those that were experimental. Because discovery of useful cellular mechanisms can precede approval of related drugs by decades, testing the hypothesis that reporting of this variable, as well as other patterns extracted by our program, can predict which approaches are likely to generalize successfully to humans will require much larger longitudinal datasets.

Challenges and limitations

Because animal models are quite specific to various disease fields, our approach was to focus within a specific field (PD) for which we had already identified translational patterns [22]. The program was able to return expected patterns of species and model use [41]. A limitation of our tool is that we did not define rodent strains, or specific genetic models. This capacity would be useful in discerning the diversity of genetic systems used to explore responses to a given intervention and would be an improvement to consider when extending this tool to fields in which genetic models are primarily used (e.g. Alzheimer’s disease). Data collection for the Interventions/Disease Modifier module is limited to title only. This approach was chosen because we had encountered an unacceptably high false positive rate when the entire abstract was used. A common source of false positive intervention data were gene names. Because we wished to capture genetic interventions central to many rodent studies, we limited collection of gene terms within this module to title only. In this way, our program is biased toward collection of Interventions/Disease Modifier terms, including genetic modifications, that the authors deem worthy of inclusion in the title. Using the genes module, all gene mentions (and thus associated cellular pathways) in an abstract can be tracked and associated with a given intervention, species or model. This represents an essential step in linking mechanisms of a potential therapy with cellular disease mechanisms. Inherent in both the Interventions/Disease Modifier and Genes module is the need to collapse related terms to a common identifier to support subsequent analyses. In the Genes module, NCBI Gene IDs were annotated in the gold standard, whereas SemRep primarily extracts UMLS Metathesaurus concept identifiers (in addition to, some NCBI Gene identifiers). Second, the NCBI Gene database has different identifiers corresponding to the same gene in different organisms. GNormPlus takes this into account, and tries to extract the appropriate identifier for the species discussed in the abstract, whereas SemRep only considers the identifiers for the Homo Sapiens taxonomy. For the Interventions/Disease Modifier module, manual mapping of outlier terms that did not achieve a match with a MeSH or UMLS identifier was required. This process requires updating as the program is extended to accommodate other neurodegenerative diseases.

The scope of the program

Study design quality comprises aspects that promote external validity (an example would be use of outcome measures that are shared across humans and animals), and those that promote internal validity/reduce bias (criteria described in the ARRIVE guidelines). Both are important if results are to achieve translation. Our tool is intended to support assessment of external validity or potential generalizability using criteria other than the stated conclusion of efficacy. The tool is not designed to automatically assess well-studied [87] aspects of study design in which full text data is mined to determine whether reporting according to standardized guidelines [24] has occurred. However, the user can very rapidly (and in unbiased fashion) collect and filter the entire published corpus of a disease area to a relevant subset that can then be read to assess measures used to reduce bias, as well as design choices that influence generalizability. The latter are context driven (e.g. the relationship between model choice [34] and mechanism of the intervention [22], or integration of biomarker data with clinically relevant outcome measures) and difficult to accurately address using automatic methods. These, and study design methods aimed at avoiding bias must still be assessed by an individual with implicit knowledge of the field and an understanding of biologic differences across species that influence translation. Design of the program is modular, and readily adaptable to other disease areas such as Alzheimer’s disease. Because neurodegenerative diseases exhibit molecular, clinical and therapeutic overlap, the tool will facilitate integrated evaluation of this group of conditions, and may reveal patterns that result in more efficient animal use.

Signal lists for species, models, functional outcome measure and interventions with full list of abstracts included in the large-scale evaluation dataset.

Each of these is categorized by Primary (1) or Secondary (2) data type. (XLSX) Click here for additional data file.

Proportions of outcome reporting by outcome type in primary and secondary data sources.

Primary data sources included clinical trials and journal articles. Meta-analyses were included in this category as they yield new insights through statistical analysis of multiple studies. Reviews, case reports, editorials and letters comprised secondary data. In general, approximately 70% of studies report a promising outcome (POS) with proportions in primary data sources being slightly lower than in all data sources. In both datasets, a slightly increasing trend in promising reports across 10 years was noted. (DOCX) Click here for additional data file.

Interventions/disease modifiers extracted over 10 years.

This module extracts pharmacologic and non-pharmacologic interventions, as well as genes or co-morbidities that influence the Parkinson’s disease phenotype. These are extracted from the title of the abstract only. (DOCX) Click here for additional data file.

Gene mentions extracted over 10 years.

These are extracted from the entire abstract. Multiple synonyms describing the same entity map to a common UMLS or NCBI identifier. (DOCX) Click here for additional data file.

Genes extracted from entire abstract in papers in which L-DOPA or alpha-synuclein were extracted by the interventions module.

Genes are listed by text term and corresponding NCBI or UMLS identifier and are ranked in descending order. (DOCX) Click here for additional data file.

The most 15 significantly enriched pathways for L-DOPA and alpha-synuclein associated studies are shown.

(DOCX) Click here for additional data file.

Gene mentions across species in papers in which alpha synuclein has been identified using the Interventions module.

The number of studies mentioning each gene by species is colored coded (green = least, red = most). (DOCX) Click here for additional data file.

Interventions classified as established.

We manually classified entities extracted by the Interventions/disease modifiers module as Established or Experimental. The former list is presented below—all remaining Interventions were included in the Experimental category. (DOCX) Click here for additional data file. 4 Oct 2019 PONE-D-19-18323 Menagerie: A text-mining tool to support animal-human translation in neurodegeneration research PLOS ONE Dear Dr. Zeiss, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses all  the points raised during the review process. We would appreciate receiving your revised manuscript by Nov 18 2019 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). This letter should be uploaded as separate file and labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. This file should be uploaded as separate file and labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. This file should be uploaded as separate file and labeled 'Manuscript'. Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. We look forward to receiving your revised manuscript. Kind regards, Gianluigi Forloni Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and http://www.journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 1. Thank you for including your competing interests statement; "No" Please complete your Competing Interests on the online submission form to state any Competing Interests. If you have no competing interests, please state "The authors have declared that no competing interests exist.", as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now This information should be included in your cover letter; we will change the online submission form on your behalf. Please know it is PLOS ONE policy for corresponding authors to declare, on behalf of all authors, all potential competing interests for the purposes of transparency. PLOS defines a competing interest as anything that interferes with, or could reasonably be perceived as interfering with, the full and objective presentation, peer review, editorial decision-making, or publication of research or non-research articles submitted to one of the journals. Competing interests can be financial or non-financial, professional, or personal. Competing interests can arise in relationship to an organization or another person. Please follow this link to our website for more details on competing interests: http://journals.plos.org/plosone/s/competing-interests [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: I Don't Know ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: PONE-D-19-18323 This paper is well written, but it is long and complex. Some editing and clarification, especially in the methodology and results sections, would be helpful. The results section seems to contain some elements that should be within the methodology section. I found the Results section difficult to understand. The key findings need to be brought out; for example it seems that reporting of functional outcomes is a key predictor of study efficacy, but this wasn’t emphasised (or perhaps I have misunderstood). The authors aim to improve external validity by aggregating evidence across a range of models, arguing that single animal models are unable to recapitulate human neurodegenerative diseases. They argue that consistent results across diverse animal models may clarify whether an approach has translational potential. I think they are correct in highlighting the problems of animal-human translation and in focusing on efficacy. Moreover, text mining approaches have the advantage of being able to offer insights about animal models without using further animals. However, I’m unclear whether their approach will improve external validity, as they claim. They note (pp36-37) that translation is difficult because animal models are reductionist, but suggest this can be overcome by identifying mechanisms that show consistent results across diverse animal models that reflect different aspects of the human disease, noting that FDA requires testing in more than one species. The problem is, this does not ensure external validity. Even if some aspects of the human disease are reflected in diverse animal models, not all aspects of the human disease are going to be accurately modelled. Furthermore, external validity is always going to be compromised because of the problem of animal-human species differences. Although mice and humans have genetic, biochemical and physiological similarities, our lineages diverged around 85 million years ago and since then, have become adapted to very different environments. Mice and other animals are unlikely to be useful for understanding modern, non-communicable diseases in humans, because the causes of these arise from our unique, evolved life histories. I think it would help if the authors acknowledged this. P3, line 20: needs to expand on the reasons for translational failure, i.e. not just state that they are complex but tease out the different factors that impact translation, i.e. publication bias, poor reporting, poor study design, poor conduct of studies etc. (some of which are considered on p4). This will help contextualise the study so that when the authors state that they are addressing external validity, readers will understand where this fits into the translational picture. P3, lines 22-23: this isn’t quite true – the discovery literature is heavily weighted towards publication of studies that claim to be promising, but that if took measures to prevent bias would not achieve such positive results, or that if set in the context of published negative studies would not skew entire fields towards excess significance. P4, line 1: is dichotomy the right word? Line 9, page 5: too many ‘are’s. P41, lines 6-7: Incorrect definition of publication bias in brackets. The tool doesn’t allow users to overcome the problem of publication bias (after all, it cannot take into account unpublished negative studies); rather, it allows users to ignore authors’ claims about their studies’ ‘promising’ results, which is a different phenomenon. Is the tool compromised by being unable to take study quality into account? ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step. 12 Nov 2019 Nov 11, 2019 We thank the reviewers for their thoughtful comments which we feel have improved readability and helped clarify several issues for the reader. All have been addressed as described below. Reviewers comments are underlined, followed by our responses. The authors have confirmed that no competing interests exist 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: I Don't Know The paper does not contain statistical analyses. Intrinsic evaluation was performed using standard information extraction evaluation metrics: precision (or positive predictive value), recall (sensitivity), and F1 score (the harmonic mean of precision or recall). All tables and graphs shown for large scale evaluation are descriptive data (predominantly describing proportions). 5. Review Comments to the Author Reviewer #1: PONE-D-19-18323 1. This paper is well written, but it is long and complex. Some editing and clarification, especially in the methodology and results sections, would be helpful. The results section seems to contain some elements that should be within the methodology section. I found the Results section difficult to understand. The manuscript has been shortened (currently 8973 words vs the original 9783 words (incl abstract, summary, tables and figure legends) and edited as follows: a. The introduction has been edited to address points raised in point 4 below. b. The methods section has been left intact, as removing text here would remove content, and we feel this is a critical portion of the paper. However, some material from the results section has been moved here. c. The results section has been shortened and edited for clarity. The various types of comparisons that could conceivably go in the methods section have been left here to avoid having to repeat these in the results section. The following sentence has been added to clarify this: Page 18, line 4:” Specific comparisons are given by subheading in the results section.” d. The final portion (outcomes by publication type) has been moved to the beginning, with subsequent elimination of one Supp Table through combining the original S1 Table and S8 Table into a new S1 Table. e. The discussion section has been substantially shortened, primarily through removal of redundant text already mentioned in the methods or results section. Additionally, subheading that highlight main points are used. A subheading to clarify limitations and challenges has been used to address the last point raised by the reviewer. 2. The key findings need to be brought out; for example it seems that reporting of functional outcomes is a key predictor of study efficacy, but this wasn’t emphasised (or perhaps I have misunderstood). It appears from this study and a previous one (Zeiss et al, 2017) that use of functional outcome measures do indeed have the potential to predict efficacy that translates to humans. We are careful to mention that actually proving this point requires confirmation in a larger dataset as follows Page 34, line 11-17: “In agreement with previous observations [26, 28], functional outcome measures were more highly reported in established therapies compared to those that were experimental. Because discovery of useful cellular mechanisms can precede approval of related drugs by decades, testing the hypothesis that reporting of this variable, as well as other patterns extracted by our program, can predict which approaches are likely to generalize successfully to humans will require much larger longitudinal datasets” Zeiss CJ, Allore HG, Beck AP. Established patterns of animal study design undermine translation of disease-modifying therapies for Parkinson's disease. PLoS One. 2017 Feb 9;12(2):e0171790. (PS The effort described above has just started using a 80K-abstract PD dataset spanning 1950-2019) 3. The authors aim to improve external validity by aggregating evidence across a range of models, arguing that single animal models are unable to recapitulate human neurodegenerative diseases. They argue that consistent results across diverse animal models may clarify whether an approach has translational potential. I think they are correct in highlighting the problems of animal-human translation and in focusing on efficacy. Moreover, text mining approaches have the advantage of being able to offer insights about animal models without using further animals. However, I’m unclear whether their approach will improve external validity, as they claim. They note (pp36-37) that translation is difficult because animal models are reductionist, but suggest this can be overcome by identifying mechanisms that show consistent results across diverse animal models that reflect different aspects of the human disease, noting that FDA requires testing in more than one species. The problem is, this does not ensure external validity. Even if some aspects of the human disease are reflected in diverse animal models, not all aspects of the human disease are going to be accurately modelled. Furthermore, external validity is always going to be compromised because of the problem of animal-human species differences. Although mice and humans have genetic, biochemical and physiological similarities, our lineages diverged around 85 million years ago and since then, have become adapted to very different environments. Mice and other animals are unlikely to be useful for understanding modern, non-communicable diseases in humans, because the causes of these arise from our unique, evolved life histories. I think it would help if the authors acknowledged this. The fundamental differences between human and animal cell biology that can derail translation are extremely important and have now been alluded to in the introduction and discussion. Page 5, line 6-7: “Profound differences in animal and human physiology[12] are a critical source of poor translation” Page 36, line 20-21:” These, and study design methods aimed at avoiding bias must still be assessed by an individual with implicit knowledge of the field and an understanding of biologic differences across species that influence translation”. There are many days when I share the reviewer’s view on the poor chances of mice/other animals to understand complex diseases in humans. However, the issue is extraordinarily complex. For example, key steps in the use of LDOPA in PD involved animals. These included aspects of catecholamine biochemistry (PMID: 16746563), the universality of catecholamine distribution in the brain (PMID: 13451690); depletion of these stores by reserpine in the rabbit (PMID: 13552704) and restoration of resulting functional deficits by dihydroxyphenylalanine in reserpinized mice (PMID: 13483658) prior to treatment in PD patients (PMID: 13869404). Modulation of striatal neurochemistry by adenosine A2A receptors was demonstrated in rats (PMID: 7812630, 8780014) almost 2 decades before istradefylline’s first approval in Japan. We fully recognize that these drugs achieve symptomatic relief through a highly conserved nigrostriatal circuitry, and that the molecular underpinnings of dopaminergic cell death are more difficult to address. It is very difficult to know which discoveries eventually have clinical application. Given the financial and intellectual resources directed at animal research, the capacity to assess the generalizability of interventions across different systems in an unbiased manner may be useful. We have made it clear that what is presented here is the program design and accuracy as evidenced by its ability to extract known patterns from PD literature and that testing whether the program can predict translational success will require its application to a much larger dataset. 4. P3, line 20: needs to expand on the reasons for translational failure, i.e. not just state that they are complex but tease out the different factors that impact translation, i.e. publication bias, poor reporting, poor study design, poor conduct of studies etc. (some of which are considered on p4). This will help contextualise the study so that when the authors state that they are addressing external validity, readers will understand where this fits into the translational picture. The introduction has been edited to clarify aspects of internal and external validity in animal study design that together impede translation (Page 4 and 5). It has been made clear that the intent of the program is to address some aspects of external validity such as diversity of model systems and use of functional outcome measures, and that what is presented here is the program design and accuracy as evidenced by its ability to extract known patterns from PD literature. P3, lines 22-23: this isn’t quite true – the discovery literature is heavily weighted towards publication of studies that claim to be promising, but that if took measures to prevent bias would not achieve such positive results, or that if set in the context of published negative studies would not skew entire fields towards excess significance. This sentence has been followed in the next paragraph by an explanation of factors that skew published animal research in the direction of excess significance, as well as a summary of measures the research community is /should be taking to address this, including publication of negative results. Page 4, lines 10-19:” Several factors that undermine the reliability of animal studies have been identified. These include insufficient rigor in animal study design and reporting [6, 23], publication bias [16, 18], over-reporting of significance [24, 25], over-reliance on non-clinical outcome measures [26], and entrenched use of certain model systems [3, 27, 28]. Together, these issues contribute to poor reproducibility of animal studies, and certainly worsen the translational gap[29]. To address this, reporting and design guidelines [9] have been adopted by numerous journals[10], and by major funding agencies[11]. These, once implemented, should improve rigor and reproducibility of preclinical studies, although widespread evidence for this is not yet available [30]. Regarding publication bias, greater reporting of negative studies [31, 32] would provide a more realistic view of actual preclinical efficacy. “ P4, line 1: is dichotomy the right word? Dichotomy has been replaced with “this translational gap” Line 9, page 5: too many ‘are’s. Thank you “are” has been replaced with “and” P41, lines 6-7: Incorrect definition of publication bias in brackets. The tool doesn’t allow users to overcome the problem of publication bias (after all, it cannot take into account unpublished negative studies); rather, it allows users to ignore authors’ claims about their studies’ ‘promising’ results, which is a different phenomenon. Yes, this is a good point. The sentence referring to publication bias has been removed and replaced with the following: Page 36, lines 8-9: “Our tool is intended to support assessment of external validity using criteria other than the stated conclusion of efficacy. “ Is the tool compromised by being unable to take study quality into account? We agree that study quality is crucial, and must be evaluated. Design quality comprises aspects that promote external validity (an example would be use of outcome measures that are shared across humans and animals, use of a progressive chronic model if neuroprotective therapies are studied or replication across different species), and those that promote internal validity/reduce bias (e.g reporting criteria in the ARRIVE guidelines). Our program is designed to support assessing the former using abstract data on a large scale, and builds upon concepts introduced in our previous publications (PMID: 25448761; 28182759). In a previous publication (PMID: 28182759), we compared reporting of various criteria in support of external and internal validity, and found that entities such as animal number, statistical tests used, blinding, housing etc were variably reported even for interventions that went on to be approved. These measures were reported in various locations (methods, figure legends, Supp data) of the full text article, thus complicating automatic analysis. In contrast, features supporting external validity (e.g species and model diversity, and use of functional measures) were reported more frequently in approved interventions, and are often mentioned in abstracts, and were thus chosen as aspects of our program. We feel that because studies are often complex, and contextual, automatic exclusion based on failure to report a given ARRIVE guideline could exclude useful papers. Further, it would require us to develop an entirely different program that searches full text papers that are more publicly restricted than abstracts). Having a human with implicit knowledge of the field is a critical aspect of final decision making on study quality – our program very rapidly filters numbers of studies (PMID: 25448761) to a readily manageable subset for manual inspection of design factors that support both internal and external validity. While ARRIVE guidelines have not been broadly implemented yet (PMID: 29795636) there is at least growing awareness of the need to design studies to avoid bias. There is far less awareness of design factors needed to promote generalizability - our program is focused on that problem. We have made the abilities and limits of the program clear to the reader in the last part of the discussion: page 36-37, under the subheading: Scope of the program. Submitted filename: Response to Reviewers R1.docx Click here for additional data file. 22 Nov 2019 Menagerie: A text-mining tool to support animal-human translation in neurodegeneration research PONE-D-19-18323R1 Dear Dr. Zeiss, We are pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it complies with all outstanding technical requirements. Within one week, you will receive an e-mail containing information on the amendments required prior to publication. When all required modifications have been addressed, you will receive a formal acceptance letter and your manuscript will proceed to our production department and be scheduled for publication. Shortly after the formal acceptance letter is sent, an invoice for payment will follow. To ensure an efficient production and billing process, please log into Editorial Manager at https://www.editorialmanager.com/pone/, click the "Update My Information" link at the top of the page, and update your user information. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, you must inform our press team as soon as possible and no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. With kind regards, Gianluigi Forloni Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: N/A ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: (No Response) ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No 9 Dec 2019 PONE-D-19-18323R1 Menagerie: A text-mining tool to support animal-human translation in neurodegeneration research Dear Dr. Zeiss: I am pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximize its impact. If they will be preparing press materials for this manuscript, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. For any other questions or concerns, please email plosone@plos.org. Thank you for submitting your work to PLOS ONE. With kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Gianluigi Forloni Academic Editor PLOS ONE
  82 in total

1.  Aggregating UMLS semantic types for reducing conceptual complexity.

Authors:  A T McCray; A Burgun; O Bodenreider
Journal:  Stud Health Technol Inform       Date:  2001

2.  Analysis of polarity information in medical text.

Authors:  Yun Niu; Xiaodan Zhu; Jianhua Li; Graeme Hirst
Journal:  AMIA Annu Symp Proc       Date:  2005

3.  Trial watch: phase II and phase III attrition rates 2011-2012.

Authors:  John Arrowsmith; Philip Miller
Journal:  Nat Rev Drug Discov       Date:  2013-08       Impact factor: 84.694

Review 4.  Pharmacological treatment of Parkinson disease: a review.

Authors:  Barbara S Connolly; Anthony E Lang
Journal:  JAMA       Date:  2014 Apr 23-30       Impact factor: 56.272

Review 5.  Reproducibility in science: improving the standard for basic and preclinical research.

Authors:  C Glenn Begley; John P A Ioannidis
Journal:  Circ Res       Date:  2015-01-02       Impact factor: 17.367

6.  Model organisms: There's more to life than rats and flies.

Authors:  Jessica Bolker
Journal:  Nature       Date:  2012-11-01       Impact factor: 49.962

Review 7.  Nanoformulation: A Useful Therapeutic Strategy for Improving Neuroprotection and the Neurorestorative Potential in Experimental Models of Parkinson's Disease.

Authors:  Jose V Lafuente; Catalina Requejo; Alejandro Carrasco; Harkaitz Bengoetxea
Journal:  Int Rev Neurobiol       Date:  2017-10-20       Impact factor: 3.230

Review 8.  Yeast models of Parkinson's disease-associated molecular pathologies.

Authors:  Sandra Tenreiro; Vanessa Franssens; Joris Winderickx; Tiago Fleming Outeiro
Journal:  Curr Opin Genet Dev       Date:  2017-06       Impact factor: 5.578

9.  Design, power, and interpretation of studies in the standard murine model of ALS.

Authors:  Sean Scott; Janice E Kranz; Jeff Cole; John M Lincecum; Kenneth Thompson; Nancy Kelly; Alan Bostrom; Jill Theodoss; Bashar M Al-Nakhala; Fernando G Vieira; Jeyanthi Ramasubbu; James A Heywood
Journal:  Amyotroph Lateral Scler       Date:  2008

10.  Conserved cell types with divergent features in human versus mouse cortex.

Authors:  Rebecca D Hodge; Trygve E Bakken; Jeremy A Miller; Kimberly A Smith; Eliza R Barkan; Lucas T Graybuck; Jennie L Close; Brian Long; Nelson Johansen; Osnat Penn; Zizhen Yao; Jeroen Eggermont; Thomas Höllt; Boaz P Levi; Soraya I Shehata; Brian Aevermann; Allison Beller; Darren Bertagnolli; Krissy Brouner; Tamara Casper; Charles Cobbs; Rachel Dalley; Nick Dee; Song-Lin Ding; Richard G Ellenbogen; Olivia Fong; Emma Garren; Jeff Goldy; Ryder P Gwinn; Daniel Hirschstein; C Dirk Keene; Mohamed Keshk; Andrew L Ko; Kanan Lathia; Ahmed Mahfouz; Zoe Maltzer; Medea McGraw; Thuc Nghi Nguyen; Julie Nyhus; Jeffrey G Ojemann; Aaron Oldre; Sheana Parry; Shannon Reynolds; Christine Rimorin; Nadiya V Shapovalova; Saroja Somasundaram; Aaron Szafer; Elliot R Thomsen; Michael Tieu; Gerald Quon; Richard H Scheuermann; Rafael Yuste; Susan M Sunkin; Boudewijn Lelieveldt; David Feng; Lydia Ng; Amy Bernard; Michael Hawrylycz; John W Phillips; Bosiljka Tasic; Hongkui Zeng; Allan R Jones; Christof Koch; Ed S Lein
Journal:  Nature       Date:  2019-08-21       Impact factor: 49.962

View more
  1 in total

1.  Toward assessing clinical trial publications for reporting transparency.

Authors:  Halil Kilicoglu; Graciela Rosemblat; Linh Hoang; Sahil Wadhwa; Zeshan Peng; Mario Malički; Jodi Schneider; Gerben Ter Riet
Journal:  J Biomed Inform       Date:  2021-02-26       Impact factor: 6.317

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.