Literature DB >> 26069732

How difficult is the validation of clinical biomarkers?

Jan Voskuil1.   

Abstract

Recent developments of introducing stratified medicine/personal health care have led to an increased demand for specific biomarkers. However, despite the myriads of biomarkers claimed to be fit for all sorts of diseases and applications, the scientific integrity of the claims and therefore their credibility is far from satisfactory. Biomarker databases are met with scepticism. The reasons for this lack of faith come from different directions: lack of integrity of the biospecimen and meta-analysis of data derived from biospecimen prepared in various ways cause incoherence and false indications. Although the trend for antibody-independent assays is on the rise, demand for consistent performance of antibodies (both in choice of antibody and how to apply it in the correct dilution where applicable) in immune assays remains unmet in too many cases. Quantitative assays suffer from a lack of world-wide accepted criteria when the immune assay is not ELISA-based. Finally, statistical analysis suffer from coherence both in the way software packages are being scrutinized for mistakes in the script and remaining invisible after small-scale analysis, and in the way appropriate queries are fed into the packages in search for output that is fit for the types of data put in. Wrong queries would lead to wrong statistical conclusions, for example when data from a cohort of patients with different backgrounds are being analysed, or when one seeks an answer from software that was not designed for such query.

Entities:  

Keywords:  antibodies; biomarkers; validation

Year:  2015        PMID: 26069732      PMCID: PMC4431379          DOI: 10.12688/f1000research.6395.1

Source DB:  PubMed          Journal:  F1000Res        ISSN: 2046-1402


Introduction

Clinical biomarkers have been around for a long time now, and the field is moving rapidly. In addition to genetic and protein markers, we now also have microRNAs, epigenetic markers, lipids, metabolites, and imaging markers. Some are extremely useful as a (companion-) diagnostic; others may serve as a mere indicator. However, there are problems. There is confusion on the nomenclature and on the way how biomarkers are meant to be validated and used. A proposal published in 2006 was meant to create some clarity and consistency in the matter [1]. The biggest obstacle by far is that Biomarker validation and qualification depend on confirmation at different locations (different labs). There are issues with consistency in the preparation of the biological material used in the different studies, and with consistency in the choice of antibody when required. It should also be noted that in quantitative immunohistochemistry (IHC) one needs a standard in the quantification method [2]. A recent opinion paper reveals yet another layer of complexity: The statistical analysis is prone to wrong conclusions down to coding errors in the software [3]. It may not be a surprise then that one another led to the observation that only about 11% of preclinical research papers demonstrated reproducible results [4]. It is time to take stock and to address the different levels of disturbance complicating the process of biomarker validation and qualification.

Biological material

The integrity of the tissue specimens will determine the quality of the biomarker’s measurements, especially when biomarkers are instable. Post-mortem samples in particular will never represent samples from living individuals because of the post-mortem delay. As the post-mortem delay will differ from individual to individual, the level of decay will vary dramatically per sample. For this reason, post-mortem samples are best fit for qualitative analysis. Quantification of any biomarker in post-mortem samples should be interpreted with extra care [5]. Plasma samples can be prepared in different ways: they can be prepared either by citrate, by ethylenediaminetetraacetic acid (EDTA) or by heparin. In addition, biomarkers can be tested in serum and in whole blood. It is clear that levels of biomarkers will need to be compared between equally treated samples in order to avoid variations in noise from the different ways the samples were prepared [6]. Since this principle is universal, it will be true for any other tissue types. For microscopy, tissue slides and cell suspensions have to be prepared in line with the required assay before they can be investigated. Fixatives (alcohols, aldehydes), embedding materials (paraffin, LR White, etc) and temperatures (frozen vs heated) have profound effects on the integrity of the tissues and cells and they will determine the success of the assay. Again, consistency in the tissue preparation, tissue sections and cells to be analysed is paramount [7, 8]. Mega-data analysis may get skewed when data are collated from samples treated in different ways. A systematic approach to record and keep biospecimen has been proposed and is aimed to become the new standard: Biospecimen Reporting for Improved Study Quality (BRISQ) guidelines provide a tool to improve consistency and to standardize information on the biological samples [9].

Antibody choice

Mass-spec and RT-PCR quantifications will be robust by the consistency of the assay material. However, the robustness of immune assays depends highly on the choice of antibodies used in the assay. Once an antibody has been successfully validated in one assay, this assay is defined by this antibody. Change of antibody will potentially change the outcome altogether as demonstrated in the past [10, 11]. When an antibody needs changing, the assay is no longer validated and the validation procedure will have to be repeated with the new antibody. For this reason the preference goes to monoclonal antibodies. The rationale behind this preference is that the clone number of the antibody would define its characteristics: the expectation then is that the assay will remain validated because the antibodies remain identical when using antibodies from the same clone number, no matter which vendor they are from. Unfortunately this is a myth. Depending on the vendor (and sometimes depending on the catalogue number) the formulations, all with the same clone number, will differ: the antibody may be purified from ascitic fluid, from culture media, or not purified at all (just ascitic fluid or just culture supernatant). These different formulations will have an effect on the way the antibody needs to be diluted to avoid non-specific background [12]. Therefore, the monoclonal antibody needs to be revalidated in the same assay when the original formulation is no longer available. But even subsequent batches from the same formulation show some level of differences, thus undermining the main argument of preference to use monoclonal antibodies in standard assays. A peptide-generated polyclonal antibody from a larger animal than rabbit (for large size batches) may serve as a cost-effective alternative because the batch-to-batch variation of such antibody is limited by the size of the immunizing peptide unlike other polyclonal antibodies [12].

Assay development

When a new assay is being developed a monoclonal antibody may not be always readily available. Then a peptide-generated polyclonal antibody may serve as a good and cost-effective alternative. However, peptide polyclonal antibodies need a new round of validation when a new batch from a different animal arrives, just like different formulated monoclonal antibodies. During assay development it is essential to dilute the antibody far enough to avoid non-specific background, but it needs to be strong enough to allow measuring a dynamic range, especially when the assay is quantitative. When the assay is dependent on a secondary antibody, this antibody needs validation as well (with and without primary) so to assess its non-specific signals (noise) [12]. Specificity needs to be addressed by comparing specimen spiked and un-spiked with the intended protein of interest (analyte) at various quantities. The signals need to be proportionate to the spiked quantities. In addition, specimen known not to have any of the analyte needs to be compared with specimen known to have the analyte at natural levels [13].

Detection and cut-off values

Sensitivity is commonly attributed to the antibody used in an assay, but this is a misunderstanding. Sensitivity is determined by the detection method of which the antibody/or primary and secondary antibodies may take part in. If levels of the analyte are low, a higher sensitivity is required. This increased sensitivity is usually not accomplished by increasing the antibody concentration, although using an antibody with higher affinity will help to some extent. But in general the change of detection method (fluorophore, isotope, PCR, etc.) is the appropriate step to take. Together with the increase of sensitivity, the noise and background will also increase. When a change to a higher sensitivity is required, the validation should focus on a more stringent regime for keeping noise and background at bay [12]. When quantification is a requirement, cut-off values need to be put in place. Both the Lowest Levels Of Quantification (LLOQ) and Highest Levels Of Quantification (HLOQ) must be determined. Often the detection limits are determined as well, but this is only relevant for qualitative work. In IHC these values become tricky, because the intensity of signal is not just a number generated by a detector; the density of signal is combined with the location in the tissue. In addition, the surface area of quantification needs well defined boundaries. And even when all these measures are in place, the quality of the tissue and the quality of the slides can potentially jeopardize these measures and skew the results [14]. Diagnostics by IHC is therefore prone to misinterpretation when for one specific test consistency at all levels (same antibody at same dilution, identically prepared tissue samples, identical area surface, identical staining analysed, etc.) is not followed in all laboratories in the world.

Statistics and jumping to conclusions

Statistical analysis is notoriously used to provide the convenient evidence required by the author(s). No matter what method of statistics is used, when the input data have been selected from a larger set, any outcome will be biased and flawed by default. Only analysis of ALL data (non-selected) would yield proper results, but then they might be inconclusive or inconvenient. The pressure to publish in peer-reviewed papers force authors to present statistics in the most incomprehensible way possible, knowing that their peers will not admit their confusion and likely take the author’s word for it [15]. Even when the statistic results are sound, they may get over-interpreted. Thus original claims were made based on prejudice and weak statistics and only over time, when more scientific details become available, a more complex picture emerged. For example how cholesterol levels are linked to cardiovascular disease [16, 17], how cancer is not merely caused by mutations [18, 19], how obesity is not a choice of lifestyle [20, 21] etc. Simplified claims can be (and has been) driven by apparent conflicts of interest as suggested in a study [22]. The reputation of biomarkers has suffered dramatically from lack of scientific integrity and as a result many scientists lost faith in the usefulness of biomarker databases. New guidelines have been introduced by publishers in order to introduce a new standard on how statistics are presented [23]. There are several statistical packages on the market for scientists and clinicians to use. However, these packages are quite advanced and need expertise handling, very much like a driver’s licence is required in order to safely use a motorised vehicle on the public road. Vendors of such packages admit that their products are not always properly used (personal communications). The chosen algorithms need to be appropriate for the type of data to be analysed: some algorithms are designed for decision making, and they are not necessarily fit for scientific fact finding. In addition, the same data entered in the same system may result in different output on different occasions simply because the wrong type of results is being asked for (personal communications with statistic analysts). Finally, subtle coding errors in the software cannot always be identified in small tests on script integrity, only to skew results when large scale data are being processed [3].

Project design and personalized medical care/stratified approaches

When all the above hurdles have been successfully taken, we are not quite there yet. Each individual is different from the next, and therefore each individual has different tolerance or sensitivity to toxins and medicines. This makes the assessment of biomarkers to follow the progress of a disease, or to follow the efficacy of a therapy, difficult to analyse when a group of patients have been treated all in the same way but the individuals in the groups are so diverse in genetic and/or ethnic background that the data can still be all over the place. Only when a group is defined by a certain genetic or environmental background, would there be sufficient homogeny to assess a biomarker for this particular defined group. For example, only recently it was found that HER2-type breast cancer patients do not benefit as well from therapies when they carry PICK3CA mutations compared to those who do not [24]. It is like the chicken-egg (catch-22) paradigm: one has to start clinical trials in order to identify the non-responsive patients and only then one can leave them out for proper validation of a new biomarker. However, proper validation demands positive and negative controls and not allowing to select the convenient data only. Although this paradox can be dealt with properly, it is no surprise that the search for proper clinical biomarkers remains very challenging for some time to come. The paper “How difficult is the validation of clinical biomarkers?” by Jan Voskuil is timely in that as the author points out the burgeoning growth of biomarker assays particularly in chronic diseases and notably in cancer has led to the misinterpretation of both research and clinical data for all of the reasons pointed out by the author. As correctly noted by the author, there are huge variations in sample collection, storage, preparation, assay used, antibodies involved, and analysis—all of which can lead to wide variability in results and confound the end user of such data. The author correctly points out steps that can be taken early on in the process that can minimize or preclude the accuracy of the results so obtained. There is a need for standardization. Furthermore, the statistical analyses of biomarker data need to be stipulated before the data are collected and not afterwards, using statistical software in a black-box manner to see what results might show statistical significance, even though the significance may be serendipitous or not clinically relevant ( Zapf ). The slavish adherence to standard software packages by many physicians and biomedical scientists would be amusing if the wider implications of their innumeracy were not so dire ( Partin ; Tuppin ). In order for a biomarker to be useful, it must reflect a change in concentration in the media sampled with a change in disease status. It is frequently assumed that serum or blood are the best media for the study of biomarkers but because of the number of potentially confounding variables in serum or blood, tears and saliva, because they reflect intracellular fluids, might serve as better indicators of intracellular events long before these are reflected in the blood ( Pieragostino ; Salvisberg ). I recommend acceptance of this very interesting paper for indexation. I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. Thank you for inviting me to review the paper from Jan Voskuil. I enjoyed reading the manuscript “ How difficult is the validation of clinical biomarkers?” by Jan Voskuil and I agreed to review because of the following three aspects: I appreciate the open review process of F1000Research and that such are published, as reviewers should stop hiding behind anonymity. There is – at least to me – no criterion justifying such if science increasingly wants transparency which we all know science needs. The article provided is a must read. I would assume especially within the Biotech community but also scientists and clinicians should do so. After I read the article the 2 nd time, I decided to write a review but not in the usual way of reviewing a manuscript for a journal, because the necessary aspects have already been included in a comprehensive manner. My intent was including comments as well as additional aspects which may be of importance from the aspect of a clinician, surgeon and scientist for helping to see the subject biomarker from additional and different aspects. General The author provides important aspects by critically evaluating the use of biomarkers. This is necessary as the author reminds us about reality and wishes in science as well as in clinical practice.The different headlines are well thought through and chosen critically while questioning issues surrounding standardization. This is even more important as reports published for biomarkers under investigation are not standardized and make the same mistakes which had been made in the 70s and 80s in terms of tumor markers. Despite the necessity for being critical, it should not be viewed as being a synonymous with negative behavior. The author takes the responsibility by addressing major obstacles and missing data and that is without doubt highly appreciated and needed. Biomarkers in diagnosis and treatment of diseases are measured characteristics and reflect a biological state of a disease. In terms of cancer, there is hope that biomarkers will provide a detection and screening tool for diagnosis, treatment with an influence on outcome orientated patient stratification as well as on predicting and monitoring multimodal treatment. The ideal biomarker is objectively measured in a comprehensible way independent of which laboratory it is investigated in, is easily measurable, cost-effective, and evaluated as an indicator for pathological and/or biological processes and of consistent value across differences in age, gender or ethnicity. Therefore, there is a necessity to remind us to not repeat history by including nearly any protein as a biomarker. Where are we? Many biomarkers are already declared by many companies to determine or diagnose a disease, although no data of half-life, metabolism or different interaction by different pathways are known nor provided. Again, it seems that history repeats itself as we get into the same discussions and situations as during the 70s and 80s in regard to tumor markers and cancer. Therefore, it is of importance that the author attempts to structure this theme into the headlines biological material, antibody choice, assay development, detection of cut-off values, statistics and project design. Project Design Logistically I would have thought that having the sub-headline Project Design earlier in the paper as this would indicate the direction. Biological material Of course there is hope that biomarkers help detecting a disease or serving as a screening tool for making a diagnosis. So far, it is not clear which biological material should be used and also it is not clear if this changes occur during different stages of diseases. Further, there is no standardization on how which biological material is stored and which variables influence the quality of the assay. There is another underestimated variable which needs to be taken into account and using cancer as an example may reveal further problems: Cancer is not one disease and contains a heterogeneous set of dysfunctions such that the information available for biomarkers are also heterogeneous. This gets worse if we remind ourselves about the following: Igarashi, et al., observed in 93 specimens investigated for  tumor microvessel density (MVD) and thymidine phosphorylase (dThdPase), that tumor cells strongly stained for TP were “…often observed as a rim in the periphery of the tumor nest” [1]. On the other hand, biomarkers are not just expressed within tumor cell nests. Takebayashi, et al., revealed in 1998 that normal esophageal tissue showed a TP expression rate of 12.3% compared to 50.9% in the tumor cell area of 163 investigated resected ESCC, although the percentage rate of cells expressing TP was less than 5% in 85.9% of non-neoplastic tissues [2]. Additionally even histomorphological tumor-negative lymph nodes (pN0) showed a TP expression rate of 27.9%. These examples illustrate that use of biomarkers needs a standardization in many aspects and by this scientists and clinicians need to be involved both trying to bring together the necessary aspects for future use of biomarkers, because the examples used could also mean that the gene expression is different in terms where a biopsy is taken as well as where apart of the specimen is cut for investigating biomarkers. Now, this may even be more complicated using fluids and under which condition they had been sampled, stored: was it during an operation? Do we know if drugs influenced those biomarkers of investigation with a short half-lives? Methods (antibody choice, ..) To my knowledge there is no standardization of methods in use for different biomarkers. Again, we had that during the 70s and 80s in terms of tumor markers in use and it took long to resolve. Is there a standard protocol in use for determination of microRNA? Is it not that some use RT-PCR, Northern blotting, oligo-based arrays, in situ hybridization, different assays (together with different samples)? Are these different techniques comparable? Statistics (detection of cut-off values, statistics, …) Cut-offs can be determined of course. How many papers do you know in which a group determines all available necessary variables, such as sensitivity, specificity, positive predictive value, negative predictive value and overall accuracy? The decision-making for a cut-off is ultimately a clinical decision, as the clinician determines what is most important to know. For example, if we do not want to overlook a patient then it can be assumed that it makes sense having a high sensitivity in terms of using response. How is this in terms using a ROC analysis (receiving operator curve analysis)? Using a ROC analysis for better determination of a threshold, we need to compare the different measurements of a biomarker against a gold standard. As addressed below, it gets even more complicated: if there is a standard in use, but we know that it is not justified declaring it as a gold standard, what should we do? Furthermore,as the author points out clearly how correct observations during the past resulted in wrong conclusions, which even increased dogma behavior in cholesterol levels associated with cardiovascular disease, lifestyle and obesity and also in terms of vitamin intake and health, as well as the somatic mutation theory being the cause for cancer. Critical thinking and re-thinking is continuously needed for excellence in science as well as for useful approaches in daily clinical work. Someone may enjoy reading the recent critical remarks about the wear and tear of guidelines [3]. I would argue that these views are also a must reading for the future implications of biomarkers. Clinical Response Classification This section is of course a huge one and has multiple aspects and due to this we cannot expect that all aspects are addressed by the author. However, this aspect is extremely important, as response is in daily clinical use and therefore I would assume, there is a must as a scientific and clinical reviewer addressing some points. One major question is: Can tumor response to therapy be predicted, thereby improving the selection of patients for cancer treatment? This is a major problem now because if a biomarker is measured against the gold standard. But, is it justified declaring clinical response evaluation serving as gold-standard? The readers need to make up their own minds about the following which have recently been addressed as well [4]. The response classification is in use since 1971 since the publication by Miller [5]. What no-one wants to see is the fact, that this response classification is based on one experiment only which was conducted by Moertel and Hanley in 1976 [6]. Experiment 16 experienced oncologists (be aware that, in 1976, there was no definition of an oncologist and there wasn’t one of an expe­rienced oncologist either) in which they cov­ered solid wooden spheres with a layer of rubber foam and placed them on a soft mattress. The colleagues had to measure the diameter of these spheres in a random order using rulers or cali­pers. The analysis showed that there was an error of 25% in the measurement of the size of identical spheres in 25% of the measure­ments, and that an error of at least 50% occurred in 6.8% of the measurements. This means that the clinical response classification in use was based on a single experiment, and is still in use since some 35 years. Only some minor modifications were done: the US National Cancer Institute, together with the European Association for Research and Treatment of Cancer, proposed ‘new response criteria’ for solid tumors; a replacement of 2D measurement with measurement of one dimen­sion was made [7]. Tumor response was defined as a decrease in the largest tumor diameter by 30%, which would translate into a 50% decrease for a spherical lesion [8]. However, no subsequent standardized of this recommendation was carried out, and 35 years after the primary experiment, no additional studies with a struc­tured logistical way of accurate objective assess­ment of treatment response have been conducted or proposed. This opens an important question: If a biomarker is measured and analyzed according to the clinical response classification above, will this reflect biology as needed? Immune Response Another important variable to address contains immunological biomarkers in terms of response: Immunologists might just declare a response if immune-competent cells have been decreased and, possibly, without clinical signs of improvement of patient condition, or decrease of tumor size. But is it appropriate to declare a decrease of an immune-competent cell as a biomarker of any utility? Disease stage The 5 year survival rates for non-metastasized localized esophageal carcinoma according to the American Cancer Society (ACS) in stage I (IA and IB) range between 71 and 57& while they drop in Stage IIA on 46% and range in Stage III (IIIA, IIIB, IIIC) between 20 and 9% [9]. Do we know if different disease stages in cancer are associated with different metabolism of different biomarkers? Influence of paths which may influence quality of biomarker measurements Let us take an example growth factor-beta (TGF-beta). It is known that inhibitors of this pathway, especially TGF-β1, block the proliferation and trigger apoptosis in malignant as well as in benign tumor cells, but with increasing tumor growth development as TGF-β1 resistance to targeted therapy develops [10] - [12]. So, does this mean, that we are not aware that the function of a biomarker quality is independent from the disease, its stage how strong or weak the biomarker under investigation influence different paths? Do different situations in which biomarkers are measured mean different necessary views and if so, how can we judge those? Another example: Epidermal growth factor receptor (EGFR; ErbB-1; HER1 in humans) is the cell-surface receptor for members of the epidermal growth factor family (EGF-family) of extracellular protein ligands [13]. EGFR plays a critical role in tumor progression by stimulating cell cycle progression, invasion, and metastasis [14]. Response measured by the EGFR-antibody Cetuximab in anticancer-treated patients does not correlate with the observed degree of EGFR expression in tumor tissue in patients with metastatic colorectal cancer [15] , [16]. How many of the declarations of so-called breakthroughs in measuring a biomarker are justified? I have no doubt, that it is important to measure different biomarkers, but the marketing and promotional one should not be goal of scientists. This may be seen as an ethical aspect as well, but it is unfortunate but necessary as increasingly observations are reported and determined correctly within publications, but afterwards statements with journalists implicate having a breakthrough result, as nearly every scientific finding these days are declared as such, which from my perspective is very unfortunate. This inflationary marketing way should not be followed. Conclusion Taken the aspects reviewed above together, I repeat my statement: . Many more aspects are necessary taking into account for future evaluation and standardization of biomarkers under investigation. I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.
  38 in total

1.  Poor antibody validation is a challenge in biomedical research: a case study for detection of c-FLIP.

Authors:  Octavian Bucur; Bodvael Pennarun; Andreea Lucia Stancu; Monica Nadler; Maria Sinziana Muraru; Thierry Bertomeu; Roya Khosravi-Far
Journal:  Apoptosis       Date:  2013-10       Impact factor: 4.677

2.  Measuring response in solid tumors: unidimensional versus bidimensional measurement.

Authors:  K James; E Eisenhauer; M Christian; M Terenziani; D Vena; A Muldal; P Therasse
Journal:  J Natl Cancer Inst       Date:  1999-03-17       Impact factor: 13.506

3.  The impact of pre-analytical processing on staining quality for H&E, dual hapten, dual color in situ hybridization and fluorescent in situ hybridization assays.

Authors:  Andrea Babic; Isabell R Loftin; Stacey Stanislaw; Maria Wang; Rachel Miller; Stephanie M Warren; Wenjun Zhang; Alexandria Lau; Melanie Miller; Ping Wu; Mary Padilla; Thomas M Grogan; Lidija Pestic-Dragovich; Abigail S McElhinny
Journal:  Methods       Date:  2010-08-31       Impact factor: 3.608

Review 4.  Review of epidermal growth factor receptor biology.

Authors:  Roy S Herbst
Journal:  Int J Radiat Oncol Biol Phys       Date:  2004       Impact factor: 7.038

5.  Role of transforming growth factor beta in breast carcinogenesis.

Authors:  John R Benson
Journal:  Lancet Oncol       Date:  2004-04       Impact factor: 41.316

6.  Cetuximab monotherapy and cetuximab plus irinotecan in irinotecan-refractory metastatic colorectal cancer.

Authors:  David Cunningham; Yves Humblet; Salvatore Siena; David Khayat; Harry Bleiberg; Armando Santoro; Danny Bets; Matthias Mueser; Andreas Harstrick; Chris Verslype; Ian Chau; Eric Van Cutsem
Journal:  N Engl J Med       Date:  2004-07-22       Impact factor: 91.245

7.  Phase II trial of cetuximab in patients with refractory colorectal cancer that expresses the epidermal growth factor receptor.

Authors:  Leonard B Saltz; Neal J Meropol; Patrick J Loehrer; Michael N Needle; Justin Kopit; Robert J Mayer
Journal:  J Clin Oncol       Date:  2004-03-01       Impact factor: 44.544

8.  Genetic studies of body mass index yield new insights for obesity biology.

Authors:  Adam E Locke; Bratati Kahali; Sonja I Berndt; Anne E Justice; Tune H Pers; Felix R Day; Corey Powell; Sailaja Vedantam; Martin L Buchkovich; Jian Yang; Damien C Croteau-Chonka; Tonu Esko; Tove Fall; Teresa Ferreira; Stefan Gustafsson; Zoltán Kutalik; Jian'an Luan; Reedik Mägi; Joshua C Randall; Thomas W Winkler; Andrew R Wood; Tsegaselassie Workalemahu; Jessica D Faul; Jennifer A Smith; Jing Hua Zhao; Wei Zhao; Jin Chen; Rudolf Fehrmann; Åsa K Hedman; Juha Karjalainen; Ellen M Schmidt; Devin Absher; Najaf Amin; Denise Anderson; Marian Beekman; Jennifer L Bolton; Jennifer L Bragg-Gresham; Steven Buyske; Ayse Demirkan; Guohong Deng; Georg B Ehret; Bjarke Feenstra; Mary F Feitosa; Krista Fischer; Anuj Goel; Jian Gong; Anne U Jackson; Stavroula Kanoni; Marcus E Kleber; Kati Kristiansson; Unhee Lim; Vaneet Lotay; Massimo Mangino; Irene Mateo Leach; Carolina Medina-Gomez; Sarah E Medland; Michael A Nalls; Cameron D Palmer; Dorota Pasko; Sonali Pechlivanis; Marjolein J Peters; Inga Prokopenko; Dmitry Shungin; Alena Stančáková; Rona J Strawbridge; Yun Ju Sung; Toshiko Tanaka; Alexander Teumer; Stella Trompet; Sander W van der Laan; Jessica van Setten; Jana V Van Vliet-Ostaptchouk; Zhaoming Wang; Loïc Yengo; Weihua Zhang; Aaron Isaacs; Eva Albrecht; Johan Ärnlöv; Gillian M Arscott; Antony P Attwood; Stefania Bandinelli; Amy Barrett; Isabelita N Bas; Claire Bellis; Amanda J Bennett; Christian Berne; Roza Blagieva; Matthias Blüher; Stefan Böhringer; Lori L Bonnycastle; Yvonne Böttcher; Heather A Boyd; Marcel Bruinenberg; Ida H Caspersen; Yii-Der Ida Chen; Robert Clarke; E Warwick Daw; Anton J M de Craen; Graciela Delgado; Maria Dimitriou; Alex S F Doney; Niina Eklund; Karol Estrada; Elodie Eury; Lasse Folkersen; Ross M Fraser; Melissa E Garcia; Frank Geller; Vilmantas Giedraitis; Bruna Gigante; Alan S Go; Alain Golay; Alison H Goodall; Scott D Gordon; Mathias Gorski; Hans-Jörgen Grabe; Harald Grallert; Tanja B Grammer; Jürgen Gräßler; Henrik Grönberg; Christopher J Groves; Gaëlle Gusto; Jeffrey Haessler; Per Hall; Toomas Haller; Goran Hallmans; Catharina A Hartman; Maija Hassinen; Caroline Hayward; Nancy L Heard-Costa; Quinta Helmer; Christian Hengstenberg; Oddgeir Holmen; Jouke-Jan Hottenga; Alan L James; Janina M Jeff; Åsa Johansson; Jennifer Jolley; Thorhildur Juliusdottir; Leena Kinnunen; Wolfgang Koenig; Markku Koskenvuo; Wolfgang Kratzer; Jaana Laitinen; Claudia Lamina; Karin Leander; Nanette R Lee; Peter Lichtner; Lars Lind; Jaana Lindström; Ken Sin Lo; Stéphane Lobbens; Roberto Lorbeer; Yingchang Lu; François Mach; Patrik K E Magnusson; Anubha Mahajan; Wendy L McArdle; Stela McLachlan; Cristina Menni; Sigrun Merger; Evelin Mihailov; Lili Milani; Alireza Moayyeri; Keri L Monda; Mario A Morken; Antonella Mulas; Gabriele Müller; Martina Müller-Nurasyid; Arthur W Musk; Ramaiah Nagaraja; Markus M Nöthen; Ilja M Nolte; Stefan Pilz; Nigel W Rayner; Frida Renstrom; Rainer Rettig; Janina S Ried; Stephan Ripke; Neil R Robertson; Lynda M Rose; Serena Sanna; Hubert Scharnagl; Salome Scholtens; Fredrick R Schumacher; William R Scott; Thomas Seufferlein; Jianxin Shi; Albert Vernon Smith; Joanna Smolonska; Alice V Stanton; Valgerdur Steinthorsdottir; Kathleen Stirrups; Heather M Stringham; Johan Sundström; Morris A Swertz; Amy J Swift; Ann-Christine Syvänen; Sian-Tsung Tan; Bamidele O Tayo; Barbara Thorand; Gudmar Thorleifsson; Jonathan P Tyrer; Hae-Won Uh; Liesbeth Vandenput; Frank C Verhulst; Sita H Vermeulen; Niek Verweij; Judith M Vonk; Lindsay L Waite; Helen R Warren; Dawn Waterworth; Michael N Weedon; Lynne R Wilkens; Christina Willenborg; Tom Wilsgaard; Mary K Wojczynski; Andrew Wong; Alan F Wright; Qunyuan Zhang; Eoin P Brennan; Murim Choi; Zari Dastani; Alexander W Drong; Per Eriksson; Anders Franco-Cereceda; Jesper R Gådin; Ali G Gharavi; Michael E Goddard; Robert E Handsaker; Jinyan Huang; Fredrik Karpe; Sekar Kathiresan; Sarah Keildson; Krzysztof Kiryluk; Michiaki Kubo; Jong-Young Lee; Liming Liang; Richard P Lifton; Baoshan Ma; Steven A McCarroll; Amy J McKnight; Josine L Min; Miriam F Moffatt; Grant W Montgomery; Joanne M Murabito; George Nicholson; Dale R Nyholt; Yukinori Okada; John R B Perry; Rajkumar Dorajoo; Eva Reinmaa; Rany M Salem; Niina Sandholm; Robert A Scott; Lisette Stolk; Atsushi Takahashi; Toshihiro Tanaka; Ferdinand M van 't Hooft; Anna A E Vinkhuyzen; Harm-Jan Westra; Wei Zheng; Krina T Zondervan; Andrew C Heath; Dominique Arveiler; Stephan J L Bakker; John Beilby; Richard N Bergman; John Blangero; Pascal Bovet; Harry Campbell; Mark J Caulfield; Giancarlo Cesana; Aravinda Chakravarti; Daniel I Chasman; Peter S Chines; Francis S Collins; Dana C Crawford; L Adrienne Cupples; Daniele Cusi; John Danesh; Ulf de Faire; Hester M den Ruijter; Anna F Dominiczak; Raimund Erbel; Jeanette Erdmann; Johan G Eriksson; Martin Farrall; Stephan B Felix; Ele Ferrannini; Jean Ferrières; Ian Ford; Nita G Forouhi; Terrence Forrester; Oscar H Franco; Ron T Gansevoort; Pablo V Gejman; Christian Gieger; Omri Gottesman; Vilmundur Gudnason; Ulf Gyllensten; Alistair S Hall; Tamara B Harris; Andrew T Hattersley; Andrew A Hicks; Lucia A Hindorff; Aroon D Hingorani; Albert Hofman; Georg Homuth; G Kees Hovingh; Steve E Humphries; Steven C Hunt; Elina Hyppönen; Thomas Illig; Kevin B Jacobs; Marjo-Riitta Jarvelin; Karl-Heinz Jöckel; Berit Johansen; Pekka Jousilahti; J Wouter Jukema; Antti M Jula; Jaakko Kaprio; John J P Kastelein; Sirkka M Keinanen-Kiukaanniemi; Lambertus A Kiemeney; Paul Knekt; Jaspal S Kooner; Charles Kooperberg; Peter Kovacs; Aldi T Kraja; Meena Kumari; Johanna Kuusisto; Timo A Lakka; Claudia Langenberg; Loic Le Marchand; Terho Lehtimäki; Valeriya Lyssenko; Satu Männistö; André Marette; Tara C Matise; Colin A McKenzie; Barbara McKnight; Frans L Moll; Andrew D Morris; Andrew P Morris; Jeffrey C Murray; Mari Nelis; Claes Ohlsson; Albertine J Oldehinkel; Ken K Ong; Pamela A F Madden; Gerard Pasterkamp; John F Peden; Annette Peters; Dirkje S Postma; Peter P Pramstaller; Jackie F Price; Lu Qi; Olli T Raitakari; Tuomo Rankinen; D C Rao; Treva K Rice; Paul M Ridker; John D Rioux; Marylyn D Ritchie; Igor Rudan; Veikko Salomaa; Nilesh J Samani; Jouko Saramies; Mark A Sarzynski; Heribert Schunkert; Peter E H Schwarz; Peter Sever; Alan R Shuldiner; Juha Sinisalo; Ronald P Stolk; Konstantin Strauch; Anke Tönjes; David-Alexandre Trégouët; Angelo Tremblay; Elena Tremoli; Jarmo Virtamo; Marie-Claude Vohl; Uwe Völker; Gérard Waeber; Gonneke Willemsen; Jacqueline C Witteman; M Carola Zillikens; Linda S Adair; Philippe Amouyel; Folkert W Asselbergs; Themistocles L Assimes; Murielle Bochud; Bernhard O Boehm; Eric Boerwinkle; Stefan R Bornstein; Erwin P Bottinger; Claude Bouchard; Stéphane Cauchi; John C Chambers; Stephen J Chanock; Richard S Cooper; Paul I W de Bakker; George Dedoussis; Luigi Ferrucci; Paul W Franks; Philippe Froguel; Leif C Groop; Christopher A Haiman; Anders Hamsten; Jennie Hui; David J Hunter; Kristian Hveem; Robert C Kaplan; Mika Kivimaki; Diana Kuh; Markku Laakso; Yongmei Liu; Nicholas G Martin; Winfried März; Mads Melbye; Andres Metspalu; Susanne Moebus; Patricia B Munroe; Inger Njølstad; Ben A Oostra; Colin N A Palmer; Nancy L Pedersen; Markus Perola; Louis Pérusse; Ulrike Peters; Chris Power; Thomas Quertermous; Rainer Rauramaa; Fernando Rivadeneira; Timo E Saaristo; Danish Saleheen; Naveed Sattar; Eric E Schadt; David Schlessinger; P Eline Slagboom; Harold Snieder; Tim D Spector; Unnur Thorsteinsdottir; Michael Stumvoll; Jaakko Tuomilehto; André G Uitterlinden; Matti Uusitupa; Pim van der Harst; Mark Walker; Henri Wallaschofski; Nicholas J Wareham; Hugh Watkins; David R Weir; H-Erich Wichmann; James F Wilson; Pieter Zanen; Ingrid B Borecki; Panos Deloukas; Caroline S Fox; Iris M Heid; Jeffrey R O'Connell; David P Strachan; Kari Stefansson; Cornelia M van Duijn; Gonçalo R Abecasis; Lude Franke; Timothy M Frayling; Mark I McCarthy; Peter M Visscher; André Scherag; Cristen J Willer; Michael Boehnke; Karen L Mohlke; Cecilia M Lindgren; Jacques S Beckmann; Inês Barroso; Kari E North; Erik Ingelsson; Joel N Hirschhorn; Ruth J F Loos; Elizabeth K Speliotes
Journal:  Nature       Date:  2015-02-12       Impact factor: 49.962

9.  A phospholipidomic analysis of all defined human plasma lipoproteins.

Authors:  Monireh Dashti; Willem Kulik; Frans Hoek; Enno C Veerman; Maikel P Peppelenbosch; Farhad Rezaee
Journal:  Sci Rep       Date:  2011-11-07       Impact factor: 4.379

10.  Commercial antibodies and their validation.

Authors:  Jla Voskuil
Journal:  F1000Res       Date:  2014-10-02
View more
  4 in total

Review 1.  Genomics, microRNA, epigenetics, and proteomics for future diagnosis, treatment and monitoring response in upper GI cancers.

Authors:  Björn L D M Brücher; Yan Li; Philipp Schnabel; Martin Daumer; Timothy J Wallace; Rainer Kube; Bruno Zilberstein; Scott Steele; Jan L A Voskuil; Ijaz S Jamall
Journal:  Clin Transl Med       Date:  2016-04-06

Review 2.  Metabolomic-Based Methods in Diagnosis and Monitoring Infection Progression.

Authors:  Miguel Fernández-García; David Rojo; Fernanda Rey-Stolle; Antonia García; Coral Barbas
Journal:  Exp Suppl       Date:  2018

3.  The Use of Antibody Arrays in the Discovery of New Plasma Biomarkers for Endometriosis.

Authors:  Dorien O; Etienne Waelkens; Arne Vanhie; Daniëlle Peterse; Amelie Fassbender; Thomas D'Hooghe
Journal:  Reprod Sci       Date:  2020-02-03       Impact factor: 3.060

Review 4.  Challenges and Opportunities for Biomarkers of Clinical Response to AHSCT in Autoimmunity.

Authors:  Kristina M Harris; Tingting Lu; Noha Lim; Laurence A Turka
Journal:  Front Immunol       Date:  2018-02-02       Impact factor: 7.561

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.