Literature DB >> 28520858

Rapid Tests for Influenza, Respiratory Syncytial Virus, and Other Respiratory Viruses: A Systematic Review and Meta-analysis.

Andrea H L Bruning1, Mariska M G Leeflang2, Johanna M B W Vos1, Rene Spijker3, Menno D de Jong4, Katja C Wolthers4, Dasja Pajkrt1.   

Abstract

Rapid diagnosis of respiratory virus infections contributes to patient care. This systematic review evaluates the diagnostic accuracy of rapid tests for the detection of respiratory viruses. We searched Medline and EMBASE for studies evaluating these tests against polymerase chain reaction as the reference standard. Of 179 studies included, 134 evaluated rapid tests for influenza viruses, 32 for respiratory syncytial virus (RSV), and 13 for other respiratory viruses. We used the bivariate random effects model for quantitative meta-analysis of the results. Most tests detected only influenza viruses or RSV. Summary sensitivity and specificity estimates of tests for influenza were 61.1% and 98.9%. For RSV, summary sensitivity was 75.3%, and specificity, 98.7%. We assessed the quality of studies using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) checklist. Because of incomplete reporting, the risk of bias was often unclear. Despite their intended use at the point of care, 26.3% of tests were evaluated in a laboratory setting. Although newly developed tests seem more sensitive, high-quality evaluations of these tests are lacking.
© The Author 2017. Published by Oxford University Press for the Infectious Diseases Society of America. All rights reserved. For permissions, e-mail: journals.permissions@oup.com.

Entities:  

Keywords:  influenza; meta-analysis; rapid test; respiratory syncytial virus; respiratory virus

Mesh:

Year:  2017        PMID: 28520858      PMCID: PMC7108103          DOI: 10.1093/cid/cix461

Source DB:  PubMed          Journal:  Clin Infect Dis        ISSN: 1058-4838            Impact factor:   9.079


Acute respiratory tract infections (RTIs) are a leading cause of disease and death worldwide [1]. Recent studies suggest more than half of RTIs are caused by viruses, even in severely ill patients [2-4]. Commonly detected viruses include influenza viruses, respiratory syncytial virus (RSV), adenovirus, human metapneumovirus (hMPV), human parainfluenza viruses, coronaviruses, and rhinoviruses [5, 6]. Clinical signs and symptoms of viral RTIs overlap with those of bacterial infections. It is challenging to clinically distinguish bacterial from viral infections and different viral pathogens [7]. This diagnostic uncertainty leads to overprescription of antibiotics and extra diagnostic testing (along with costs) to rule out bacterial infections [8, 9]. Rapid detection of viral pathogens could overcome these disadvantages. Besides, prompt viral diagnosis may lead to rapid implementation of infection control measures, early administration of antiviral medication, if available, and shorter hospital stays, resulting in reduced healthcare costs [10-12]. For this reason, rapid diagnostic or point-of-care tests have been developed. Compared with other diagnostic modalities—culture, polymerase chain reaction (PCR), or immunofluorescence testing—point-of-care tests are often faster, less expensive, easier to use and accessible to staff without laboratory training. They have the potential to be carried out at or near the point of care. There is a clear trend toward point-of-care testing, and the number and quality of rapid diagnostic tests for respiratory viruses has rapidly increased [13]. For clinicians, it is important to be aware of the diagnostic accuracy of the different rapid viral tests, the factors that affect their accuracies, and test performances in daily practice. Previous systematic reviews addressing the diagnostic accuracy of respiratory tests either evaluated only a single respiratory virus (influenza or RSV) or were conducted in specific populations [14-16]. The aim of our review was to provide a state-of-the-art overview of all available rapid tests for the detection of respiratory viruses in patients of all ages with RTIs. We systematically summarized the available evidence on their diagnostic accuracy for virus detection compared with PCR testing. We assessed the quality of included studies and quantitatively summarized the results in a meta-analysis.

MATERIALS AND METHODS

Protocol

This systematic review was built on a protocol based on the Preferred Reporting Items for Systematic Review and Meta-Analyses for Protocols 2015 (PRISMA-P 2015), which is registered in the Prospero database (registration No. CRD42015024581).

Search Strategy

A systematic literature review was conducted by searching Medline and EMBASE electronic databases, through the Ovid interface, from inception to 18 January 2016. The reference lists of all articles were hand searched for additional studies. The search strategy was developed in collaboration with a medical information specialist (R. S.) and contained search terms for the most common respiratory viruses (influenza or respiratory syncytial virus or metapneumovirus or parainfluenza virus or human adenovirus or human rhinovirus or human bocavirus or human coronavirus) combined with search terms for rapid diagnostic tests (diagnostic kit or antigen [test or detection] or reagent or immunotest or point-of-care systems or rapid, simple, easy, or quick test), including brand names for the most common commercial rapid tests. The complete search strategy is shown in Appendix A. After removal of duplicates, the search results were imported into Covidence, a Web-based software platform that streamlines the production of systematic reviews (https://www.covidence.org/).

Selection Criteria

Studies were considered for inclusion if they were written in English or Dutch and reported original data regarding the accuracy of a rapid test for ≥1 respiratory virus compared with PCR. Although standardization across PCR methods is currently lacking, we selected PCR as the appropriate reference standard because of its current status as the reference standard method for detecting respiratory viruses [17]. Rapid tests were defined (adapted from the World Health Organization simple/rapid tests definition [18]) as any commercially available quick (up to 2 hours) and easy-to-use test requiring little or no additional equipment or technological skills. Only original studies were included. Studies evaluating in-house tests and precommercial versions of rapid tests were excluded, as well as studies using the rapid test result as part of a composite reference standard (to avoid incorporation bias) or performing PCR solely on samples in which rapid test results were negative (to avoid partial verification bias). Case-control studies (use of the rapid test on previously tested known positive or negative samples), case reports, conference abstracts, reviews, and veterinary studies were excluded. Two reviewers (A. H. L. B. and D. P.) independently assessed inclusion eligibility. Initial selection was based on screening of title and abstract. Full-text versions were further assessed for eligibility, and reasons for exclusion were documented.

Data Extraction

A data-extraction-form was developed with a clinical epidemiologist (M. M. G. L.). Two authors (A. H. L. B. and J. M. B. W. V.) independently extracted data. Information collected included author, year of study and publication, country, study population, specimen type, setting, performance of the test in a laboratory or at the point of care, brand name, and data required to calculate sensitivity and specificity. With these data we constructed 2×2 contingency tables. If 2×2 tables could not be extracted or calculated from the published data, we contacted authors for additional information. In some articles, the same sample sets were tested with different rapid tests. Each test comparison was considered a separate analysis, resulting in a total number of included studies higher than the actual number of included articles.

Quality Assessment

The methodological quality of studies was independently reviewed by 2 authors (A. H. L. B. and J. M. B. W. V) using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria, as recommended by the Cochrane Collaboration [19]. Disagreement about inclusion, data extraction, and quality assessment was solved by consensus.

Statistical Analysis

With the 2 × 2 tables, we estimated the sensitivity and specificity of the rapid tests. Each respiratory virus can be regarded as a separate target condition. Therefore, we analyzed the accuracy of the tests for each virus separately. Studies evaluating rapid tests for influenza viruses sometimes reported multiple 2×2 tables for influenza A and B separately. However, most of the included tests detected both influenza subtypes and made no distinction between subtypes. We therefore used “any influenza” as our main analyses and evaluated test accuracy in subtypes in subgroup analyses. Sensitivity and specificity estimates were pooled into receiver operating characteristic plots. Because sensitivity and specificity were negatively correlated, the meta-analysis should be performed on both outcome measures simultaneously. For this, we used a bivariate random effects meta-regression model [20]. To investigate the effect of potential sources of heterogeneity, we added the following covariates to the model: study population (children vs adults), commercial brand of the rapid test, and whether or not the test was performed at the point of care. We calculated summary sensitivity and specificity estimates for each covariate. To assess the effect of study quality, we performed sensitivity analyses. All analyses were performed using SAS software, version 9.4.

RESULTS

Search Results

Our literature searches identified 3527 articles. After screening, 383 articles were eligible for full-text review, of which 258 were excluded (Figure 1). The main reasons for exclusion were that the rapid test was evaluated in a case-control study, the test could not be performed rapid, the study did not include original data, a reference test other than PCR was used, or the study was not reported in English or Dutch. Because several reports evaluated >1 test, 179 separate studies were included in the systematic review and meta-analysis. Of these, 134 studies (74.9%) evaluated rapid tests for influenza viruses, 32 (17.9%) for RSV, and the remaining 13 (7.3%) for other respiratory viruses (hMPV, adenovirus, and parainfluenza virus; Table 1).
Figure 1.

Preferred Reporting Items for Systematic Review and Meta-Analyses (PRISMA) flow diagram.

Table 1.

Characteristics of Included Studies

CharacteristicStudies, No. (%)
Virus evaluated
 Influenza virus (A, B, H1N1, A, and B)134 (74.9)
 RSV32 (17.9)
 Adenovirus6 (3.4)
 Human metapneumovirus5 (2.8)
 Parainfluenza virus type 1–32 (1.1)
Population
 Children74 (41.3)
 Adults14 (7.8)
 Mixed/not reported91 (50.8)
Most frequently studied tests (company)a
 Influenza virus
  QuickVue Influenza A+B (Quidel)26 (14.5)
  BinaxNOW Influenza A&B (Alere Scarborough)19 (10.6)
  Directigen EZ FluA+B (Directigen-EZ, Becton Dickinson)16 (8.9)
  Sofia Influenza A+B (Quidel)9 (5.0)
 RSV
  BinaxNOW RSV (Alere Scarborough)6 (3.3)
  BD Veritor RSV (BD Diagnostics)5 (2.8)
  Sofia RSV (Quidel)4 (2.2)
  Directigen EZ RSV (Becton Dickinson)3 (1.7)
 Multiplex test
  mariPOC (ArcDia International)15 (8.4)
 Other rapid tests77 (43.0)
Point-of-care testing47 (26.3)

Abbreviation: RSV, respiratory syncytial virus.

Preferred Reporting Items for Systematic Review and Meta-Analyses (PRISMA) flow diagram. Characteristics of Included Studies Abbreviation: RSV, respiratory syncytial virus.

Characteristics of Included Studies

Supplementary Table 1 provides the main characteristics of included studies. Supplementary Figure 1 displays the accompanying forest plots of the estimated sensitivities and specificities. Table 1 shows that a total of 50 different rapid tests were evaluated. The most frequently studied tests for influenza were Alere BinaxNOW Influenza A&B Test, Quidel QuickVue Influenza A+B, BD Directigen EZ Flu A+B, and Quidel Sofia Influenza A+B FIA. For RSV, BD Veritor System, Alere BinaxNOW RSV Card, and Quidel Sofia RSV FIA were evaluated most frequently. Most studies were conducted in either children or populations including both adults and children. Of the studied rapid tests, 26.3% were evaluated at the point of care.

Quality of Included Studies

An overview of the risk of bias and concerns regarding applicability of included articles is presented in Figure 2. Because of our inclusion and exclusion criteria, studies were free of spectrum, incorporation, and partial verification bias, and all included studies used an appropriate reference standard. Nevertheless, several study characteristics might incorporate a risk of bias. Because inclusion and exclusion criteria were often defined unclearly or not at all, the risk of bias regarding patient selection was difficult to assess. In at least 37 articles (29.6%), the index test was performed at the point of care, suggesting that index test results were interpreted without knowledge of reference standard results. In 8 of these articles, additional information was available on the clinical feasibility or logistical organization of the test. Owing to frequent incomplete description of study characteristics that might result in a risk of bias, these items were difficult to assess, resulting in an unclear risk. The concerns regarding applicability were generally low.
Figure 2.

Quality assessment of included studies using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria.

Quality assessment of included studies using Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) criteria.

Overall Accuracy of Rapid Tests for Influenza and Respiratory Syncytial Virus

For influenza, sensitivity ranged from 4.4% to 100.0%. For all rapid tests that can detect influenza, that summary estimate was 61.1% for sensitivity (95% confidence interval [CI], 53.3%–68.3%) and 98.9% for specificity (98.4%–99.3%). Operating test characteristics of the studies regarding influenza are shown in Figure 3A.
Figure 3.

Hierarchical summary receiver operating characteristic curve plots of studies evaluating rapid tests for influenza (A) and respiratory syncytial virus (RSV) (B). Each circle indicates a single included study, with the size of the circle proportionate to the size of the study.

Hierarchical summary receiver operating characteristic curve plots of studies evaluating rapid tests for influenza (A) and respiratory syncytial virus (RSV) (B). Each circle indicates a single included study, with the size of the circle proportionate to the size of the study. For RSV, sensitivity had more variation than specificity, but sensitivity estimates were higher than for influenza, ranging from 41.2% to 88.6% (Figure 3B). The summary sensitivity and specificity for all studies evaluating RSV were 75.3% (95% CI, 72.6%–77.8%) and 98.7% (97.3%–99.4%), respectively. Owing to the small number of included studies for adenovirus, hMPV, and parainfluenza virus types 1–3, meta-analysis was not performed.

Investigation of Heterogeneity

The sensitivities of rapid tests for influenza A (68.1%; 95% CI, 58.9%–76.0%) and influenza B (71.0%; 56.8%–82.1%) were comparable. Several studies (n=58) only evaluated the diagnostic accuracy of rapid tests for detecting H1N1. Most of these rapid tests were not developed to detect H1N1 specifically. The sensitivity for H1N1 was generally lower (54.0%, 95% CI 47.6–60.3) than for the other virus subtypes (Table 2).
Table 2.

Subgroup Analyses: Accuracy Estimates

Pooled Sensitivity (95% CI)Pooled Specificity (95% CI)
Influenza
Virus type
 Influenza A68.1 (58.9–76.0)99.2 (98.5–99.6)
 H1N154.0 (47.6–60.3)99.1 (98.5–99.5)
 Influenza B71.0 (56.8–82.1)99.6 (99.2–99.8)
 Influenza A+B61.1 (53.3–68.3)98.9 (98.4–99.3)
Population
 Children66.1 (52.9–79.3)98.3 (97.2–99.5)
 Adults34.1 (14.0–54.1)99.2 (98.2–100.0)
Point-of-care testing62.1 (47.6–74.7)98.4 (96.7–99.2)
Rapid test
 QuickVue Influenza A+B44.6 (29.1–60.0)99.3 (98.8–99.9)
 Sofia Influenza A+B75.3 (59.2–91.5)95.3 (91.5–99.2)
 BinaxNow Influenza A&B44.1 (23.3–64.9)99.4 (98.6–100.0)
 Directigen Flu A+B35.8 (11.8–59.7)99.2 (98.0–99.4)
 mariPOC76.1 (53.5–98.7)99.4 (98.3–100.0)
RSV
Population
 Children75.9 (73.1–78.5)98.5 (968–99.4)
 Mixed70.9 (63.0–77.8)99.1 (95.9–99.8)
Point-of-care testing76.0 (69.8–81.2)99.1 (95.5–99.8)
Rapid test
 BD Veritor RSV76.9 (71.0–82.8)98.9 (97.1–100.0)
 BinaxNOW RSV72.2 (65.2–79.1)98.6 (96.5–100.0)
 Sofia RSV80.0 (73.0–86.9)97.8 (93.8–100.0)

Abbreviations: CI, confidence interval; RSV, respiratory syncytial virus.

Subgroup Analyses: Accuracy Estimates Abbreviations: CI, confidence interval; RSV, respiratory syncytial virus. The diagnostic test accuracy for influenza was significantly decreased when tests were performed in adults (sensitivity, 34.1%; 95% CI, 14.0%–54.1%), compared with performance in children or a mixed population (P < .01). For RSV, age did not significantly influence performance characteristics (P = .20). To identify the rapid test with the best performance characteristics, we evaluated pooled summary estimates for each test separately. This was only possible for tests with ≥4 studies. For influenza, mariPOC and Sofia Influenza A+B FIA had the best overall performance, with summary sensitivities of 76.1% and 75.3%, respectively, and summary specificities of 99.4% and 95.3%, although in some studies [21] the specificity for Sofia Influenza A+B FIA was lower. For RSV, an estimated sensitivity and specificity per test could be calculated only for BD Veritor RSV, Sofia RSV FIA, and BinaxNOW RSV. Sofia RSV FIA had the best overall performance. For both influenza and RSV, diagnostic test accuracy was not influenced by setting—that is, by whether the test was performed at the point of care or in the laboratory. The risk of bias for studies evaluating influenza was generally low. We assessed the difference between studies with low risk of bias for the patient domain and those with an unclear risk of bias in this domain. Sensitivity was generally lower in studies with a low risk of bias (53.3%; 95% CI, 42.1%–64.1%). For RSV, the investigated quality criteria did not have a statistically significant effect on diagnostic test accuracy.

DISCUSSION

Main Findings of the Study

In this systematic review and meta-analysis we provided an overview of the rapid tests that are available for the detection of respiratory viruses in patients with RTIs. The sensitivity of these tests varied considerably, but specificity was high. With the result of the rapid test it is thus possible to rule in a respiratory viral infection, but false-negative results are common. The performance of rapid tests for RSV was generally superior to the performance for influenza. A major advantage of rapid tests is their potential to be performed in a nonlaboratory setting, but only 26.3% were evaluated at the point of care. Although diagnosing RTIs requires a syndromic approach because symptoms of respiratory viruses infection overlap, only few rapid tests currently available can simultaneously detect multiple viruses. Most rapid tests solely detect influenza viruses or RSV. This is historically understandable, because these viruses were considered the most important respiratory viruses. Furthermore, availability of antivirals for influenza renders rapid diagnostic testing of influenza of high priority. However, recent studies indicate that other respiratory viruses, such as rhinoviruses and hMPV, can also cause severe respiratory illness and are sometimes detected at higher frequencies than influenza and RSV [22, 23]. The results of our heterogeneity investigation show that rapid test performance is comparable for influenza A and influenza B, which is in line with findings of a previous meta-analysis [14]. Most of the included studies were performed in children, especially those evaluating rapid tests for RSV, which at least partly explains the lack of influence of age on the diagnostic accuracy estimates for RSV. For influenza, however, rapid test performance was significantly better in children, as reported elsewhere [24, 25]. Age is inversely associated with viral load, which may explain better test results in children. Testing at the point of care did not influence diagnostic test accuracy, although this finding should be interpreted with caution. In many studies, the test was not evaluated at the point of care, nor was the setting or personnel described. Direct head-to-head comparisons between tests performed at the point of care and in the laboratory were limited. Besides, only 8 of the 37 articles describing tests performed at the point of care included information on the clinical feasibility of the test—that is, whether the test was easy to use, how rapid test results were communicated, or how the personnel performing the rapid tests evaluated its usefulness. By estimating the pooled sensitivities of the rapid tests, we aimed to determine which test performed best. The conclusions should be interpreted cautiously. Because of our random-effects model, we could not evaluate all rapid tests separately, and the newly developed molecular testing devices were evaluated only sporadically in prospective cross-sectional studies [26, 27]. In addition, rapid test kits from the same company might undergo manufacturing changes. When we analyzed only high-quality studies (ie, with a low risk of bias), the overall sensitivity decreased, implying an even lower diagnostic accuracy for rapid tests.

Strengths and Limitations

Diagnostic accuracy studies are described in different ways, and there is no standard terminology available [28]. To include all available evidence, we constructed our search strategy as broadly as possible. Data were extracted, and the quality of included studies was assessed by 2 independent reviewers, decreasing the risk of subjectivity to a minimum. In the event of methodological shortcomings—for example, incorporation bias or partial verification bias—diagnostic test performance might be underestimated or overestimated, which we tried to avoid by applying strict exclusion criteria. We deliberately chose to include only studies in which PCR was used as the reference standard, because of its status as the reference standard method for detecting respiratory viruses [17]. This important strength of our systematic review provides a more realistic accuracy estimate for rapid tests. In previous studies using viral culture or immunofluorescence as a reference, pooled sensitivities were 9%–14% higher than in studies that used PCR, and test accuracy was overestimated [15]. We did not assess publication bias because no accurate reporting methods for diagnostic test accuracy studies exist [28]. The accuracy of a systematic review depends on the quality of the studies included. As demonstrated in our quality assessment, many items regarding the risk of bias were unclear. Sensitivity and specificity calculations are extremely sensitive to the design of a study and influenced by many factors, such as virus prevalence, predominant circulating virus strain, time from illness onset to sample collection, type and quality of respiratory specimen, age, and disease severity [11, 29]. In many studies that evaluated specific rapid tests, these items were lacking or unclear. This is a major limitation in diagnostic accuracy studies in general. Comparisons between tests should be ideally performed in the same study, against the same reference standard, and in the same patients.

Future Perspectives

In the current era of emerging novel respiratory viruses, there is a growing need for rapid, sensitive, and specific identification of viral pathogens to allow effective prompt antimicrobial therapy, decrease extra diagnostic testing, and implement pathogen-specific infection control measures. Rapid tests have the potential to fulfill these needs, but one should be aware of their current limitations in diagnostic performance and range of pathogens identified. More sensitive and specific rapid multiplex molecular assays are in development. They have the potential to rapidly and accurately identify not only respiratory viruses but also bacteria. Fully automated molecular methods are commercially available and presented as designed to be operated at the point of care [30]. Although these newer tests seem to perform comparably to laboratory-operated PCRs, comprehensive state-of-the-art clinical evaluations are lacking. Besides, in addition to high costs and low sample throughput, major drawbacks of these newer diagnostic devices are the technical complexity of the tests and their dependence on electronic devices [31, 32], limiting their opportunities for direct point-of-care use. In the upcoming years, the nonmolecular rapid tests will still have a role in practical patient care. It is important for clinicians to be aware of their availability and performance characteristics. Our systematic review provides a helpful tool in this understanding. Although some studies have already evaluated the impact of rapid diagnostics on patient management [33-35], randomized controlled trials are needed to assess the clinical relevance of rapid tests in terms of clinically relevant outcomes, such as antibiotic use, length of hospital stay, and cost efficiency. In addition, point-of-care testing requires novel strategies for logistics and organization [36]. For successful implementation of rapid tests in the clinic, a laboratory test validation is not sufficient and a high-quality evaluation at the point of care is required.

Supplementary Data

Supplementary materials are available at Clinical Infectious Diseases online. Consisting of data provided by the authors to benefit the reader, the posted materials are not copyedited and are the sole responsibility of the authors, so questions or comments should be addressed to the corresponding author. Click here for additional data file. Click here for additional data file. Click here for additional data file.
  34 in total

1.  Community-Acquired Pneumonia Requiring Hospitalization among U.S. Adults.

Authors:  Seema Jain; Wesley H Self; Richard G Wunderink; Sherene Fakhran; Robert Balk; Anna M Bramley; Carrie Reed; Carlos G Grijalva; Evan J Anderson; D Mark Courtney; James D Chappell; Chao Qi; Eric M Hart; Frank Carroll; Christopher Trabue; Helen K Donnelly; Derek J Williams; Yuwei Zhu; Sandra R Arnold; Krow Ampofo; Grant W Waterer; Min Levine; Stephen Lindstrom; Jonas M Winchell; Jacqueline M Katz; Dean Erdman; Eileen Schneider; Lauri A Hicks; Jonathan A McCullers; Andrew T Pavia; Kathryn M Edwards; Lyn Finelli
Journal:  N Engl J Med       Date:  2015-07-14       Impact factor: 91.245

2.  Economic analysis of rapid and sensitive polymerase chain reaction testing in the emergency department for influenza infections in children.

Authors:  Richard E Nelson; Chris Stockmann; Adam L Hersh; Andrew T Pavia; Kent Korgenksi; Judy A Daly; Marc R Couturier; Krow Ampofo; Emily A Thorell; Elizabeth H Doby; Jeff A Robison; Anne J Blaschke
Journal:  Pediatr Infect Dis J       Date:  2015-06       Impact factor: 2.129

3.  Comparison of the Biofire FilmArray RP, Genmark eSensor RVP, Luminex xTAG RVPv1, and Luminex xTAG RVP fast multiplex assays for detection of respiratory viruses.

Authors:  Elena B Popowitch; Stacey S O'Neill; Melissa B Miller
Journal:  J Clin Microbiol       Date:  2013-03-13       Impact factor: 5.948

Review 4.  Influenza diagnosis and treatment in children: a review of studies on clinically useful tests and antiviral treatment for influenza.

Authors:  Timothy M Uyeki
Journal:  Pediatr Infect Dis J       Date:  2003-02       Impact factor: 2.129

5.  Community-acquired pneumonia requiring hospitalization among U.S. children.

Authors:  Seema Jain; Derek J Williams; Sandra R Arnold; Krow Ampofo; Anna M Bramley; Carrie Reed; Chris Stockmann; Evan J Anderson; Carlos G Grijalva; Wesley H Self; Yuwei Zhu; Anami Patel; Weston Hymas; James D Chappell; Robert A Kaufman; J Herman Kan; David Dansie; Noel Lenny; David R Hillyard; Lia M Haynes; Min Levine; Stephen Lindstrom; Jonas M Winchell; Jacqueline M Katz; Dean Erdman; Eileen Schneider; Lauri A Hicks; Richard G Wunderink; Kathryn M Edwards; Andrew T Pavia; Jonathan A McCullers; Lyn Finelli
Journal:  N Engl J Med       Date:  2015-02-26       Impact factor: 91.245

6.  Performance of rapid influenza diagnostic tests (QuickVue) for influenza A and B Infection in India.

Authors:  P A Koul; H Mir; M A Bhat; U H Khan; M M Khan; M S Chadha; R B Lal
Journal:  Indian J Med Microbiol       Date:  2015-02       Impact factor: 0.985

7.  Evaluation of a rapid antigen detection point-of-care test for respiratory syncytial virus and influenza in a pediatric hospitalized population in the Netherlands.

Authors:  Andrea H L Bruning; Karin van Dijk; Hetty W M van Eijk; Gerrit Koen; Job B M van Woensel; Frea H Kruisinga; Dasja Pajkrt; Katja C Wolthers
Journal:  Diagn Microbiol Infect Dis       Date:  2014-08-30       Impact factor: 2.803

Review 8.  Rapid viral diagnosis for acute febrile respiratory illness in children in the Emergency Department.

Authors:  Quynh Doan; Paul Enarson; Niranjan Kissoon; Terry P Klassen; David W Johnson
Journal:  Cochrane Database Syst Rev       Date:  2014-09-15

Review 9.  Systematic reviews and meta-analyses of diagnostic test accuracy.

Authors:  M M G Leeflang
Journal:  Clin Microbiol Infect       Date:  2014-02       Impact factor: 8.067

10.  Lower respiratory tract virus findings in mechanically ventilated patients with severe community-acquired pneumonia.

Authors:  J Karhu; T I Ala-Kokko; T Vuorinen; P Ohtonen; H Syrjälä
Journal:  Clin Infect Dis       Date:  2014-04-11       Impact factor: 9.079

View more
  54 in total

1.  Current Challenges for the Effective Management of the COVID-19 Pandemic.

Authors:  Vibha Sinha; Saurabh Saxena; Sanjay Kumar Pandey; Sudhakar Dwivedi; Suresh Thakur; Alexzander Asea; Ashwini Kumar Dixit; Vineeta Dixit; Naveen Kumar Vishvakarma; Dhananjay Shukla
Journal:  Adv Exp Med Biol       Date:  2021       Impact factor: 2.622

2.  Application of Nanoscale Materials and Nanotechnology Against Viral Infection: A Special Focus on Coronaviruses.

Authors:  Prathap Somu; Sonali Mohanty; Srishti Chakraborty; Subhankar Paul
Journal:  Adv Exp Med Biol       Date:  2021       Impact factor: 2.622

3.  Performance differences among commercially available antigen rapid tests for COVID-19 in Brazil.

Authors:  Mariana Lourenço Freire; Lindicy Leidicy Alves; Carolina Senra de Souza; Juliana Wilke Saliba; Verônica Faria; Mariana Junqueira Pedras; Nara de Oliveira Carvalho; Gláucia Queiroz Andrade; Ana Rabello; Daniel Moreira Avelar; Gláucia Cota
Journal:  PLoS One       Date:  2022-06-16       Impact factor: 3.752

Review 4.  Detection of Influenza A and B Viruses and Respiratory Syncytial Virus by Use of Clinical Laboratory Improvement Amendments of 1988 (CLIA)-Waived Point-of-Care Assays: a Paradigm Shift to Molecular Tests.

Authors:  Marwan M Azar; Marie L Landry
Journal:  J Clin Microbiol       Date:  2018-06-25       Impact factor: 5.948

5.  Typical RSV cough: myth or reality? A diagnostic accuracy study.

Authors:  Mirjam Binnekamp; Karlijn J van Stralen; Larissa den Boer; Marlies A van Houten
Journal:  Eur J Pediatr       Date:  2020-06-13       Impact factor: 3.183

6.  Comparison of the Quidel Sofia SARS FIA Test to the Hologic Aptima SARS-CoV-2 TMA Test for Diagnosis of COVID-19 in Symptomatic Outpatients.

Authors:  Eric T Beck; Wendy Paar; Lara Fojut; Jordan Serwe; Renee R Jahnke
Journal:  J Clin Microbiol       Date:  2021-01-21       Impact factor: 5.948

Review 7.  An Update on Advances in COVID-19 Laboratory Diagnosis and Testing Guidelines in India.

Authors:  K S Rajesh Kumar; Suhail Sayeed Mufti; Vinu Sarathy; Diganta Hazarika; Radheshyam Naik
Journal:  Front Public Health       Date:  2021-03-04

8.  Point-of-care COVID-19 antigen testing in German emergency rooms - a cost-benefit analysis.

Authors:  R Diel; A Nienhaus
Journal:  Pulmonology       Date:  2021-07-06

9.  Burden of influenza among hospitalized febrile children in Ghana.

Authors:  Benedikt Hogan; Luise Ammer; Marlow Zimmermann; Tabea Binger; Ralf Krumkamp; Nimako Sarpong; Theresa Rettig; Denise Dekker; Benno Kreuels; Lisa Reigl; Kennedy G Boahen; Charity Wiafe; Yaw Adu-Sarkodie; Ellis Owusu-Dabo; Jürgen May; Daniel Eibach
Journal:  Influenza Other Respir Viruses       Date:  2017-10-27       Impact factor: 4.380

10.  Rapid antigen detection kit for the diagnosis of SARS-CoV-2 - are we missing asymptomatic patients?

Authors:  Rimjhim Kanaujia; Arnab Ghosh; Ritin Mohindra; Vidhi Singla; Kapil Goyal; Rajendra Gudisa; Vikrant Sharma; Lalit Mohan; Navpreet Kaur; Gursimran Kaur Mohi; Ishani Bora; Radha Kanta Ratho; Roop Kishor Soni; Ashish Bhalla; Mini P Singh
Journal:  Indian J Med Microbiol       Date:  2021-07-20       Impact factor: 0.985

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.