| Literature DB >> 22515596 |
Aaron M Cohen1, Kyle Ambert, Marian McDonagh.
Abstract
BACKGROUND: Systematic Reviews (SRs) are an essential part of evidence-based medicine, providing support for clinical practice and policy on a wide range of medical topics. However, producing SRs is resource-intensive, and progress in the research they review leads to SRs becoming outdated, requiring updates. Although the question of how and when to update SRs has been studied, the best method for determining when to update is still unclear, necessitating further research.Entities:
Mesh:
Year: 2012 PMID: 22515596 PMCID: PMC3420236 DOI: 10.1186/1472-6947-12-33
Source DB: PubMed Journal: BMC Med Inform Decis Mak ISSN: 1472-6947 Impact factor: 2.796
Definition of temporal events and periods of a systematic review update relevant to this study
| End of the Report Cycle | Event | A report has had its peer review completed and is published on the Internet. |
| Pre-update Period | Time Period | Between the End of the Report Cycle and the Report Search for a report update. Little work on the topic, beyond a yearly literature search, is conducted. |
| Report Update Search Begins | Event | Date on which literature search for a report update begins. |
| Report Update Period | Time Period | Between the Report Search Begins and the End of the Report Cycle for a report update. Most of the work of a report update is conducted during this period. |
Figure 1Timeline plot of the most important studies in the inter-update period for each of the nine topics. Black markers are publications that were correctly identified by the classification system, white markers are those that were missed. The shape of the marker designates the type of the important study as defined in the methods section.
Data sets used in this study
| | |||||||
|---|---|---|---|---|---|---|---|
| 285 | 2542 | 2827 | 43 | 245 | 288 | Yes | |
| 96 | 1436 | 1532 | 39 | 476 | 515 | Yes | |
| 133 | 2364 | 2497 | 29 | 334 | 363 | Yes | |
| 20 | 0 | 20 | 5 | 0 | 5 | No | |
| 659 | 3829 | 4488 | 67 | 785 | 852 | Yes | |
| 138 | 1774 | 1912 | 38 | 601 | 639 | Yes | |
| 13 | 82 | 95 | 15 | 62 | 77 | Yes | |
| 79 | 106 | 185 | 1 | 3 | 4 | No | |
| 103 | 823 | 926 | 25 | 229 | 254 | Yes | |
| 204 | 2333 | 2537 | 50 | 375 | 425 | Yes | |
| 145 | 1657 | 1802 | 26 | 206 | 232 | Yes | |
Eleven systematic review topics had both a prior report and an update completed within our data collection window. Included articles are those included in the final systematic review report, while excluded articles are those not included in the report. Drug Effectiveness Review Project (DERP) review inclusion judgments for articles with MEDLINE entry dates prior to the End of the Report Cycle (for the prior report) were used as training data for that topic. DERP review inclusion judgments for articles indexed in MEDLINE during the Pre-Update Period (after the End of the Report Cycle date and prior to the Report Search Begins for the next update) as the testing data for that topic.
Annotation guide for articles that were deemed to potentially motivate a review update on their topic
| A | Study includes evidence on new or serious adverse events relevant to this topic. |
| P | Study includes new patient subgroup, new indication, or evidence specific to new comorbidity. |
| B | Study is notably better designed, or uses novel methods, compared to prior studies. |
| L | Study uses a significantly larger sample size than prior studies for this topic. |
| M | Study includes other significant evidence that may motivate a review update, when taken in combination with other studies. |
Each article included in the actual systematic review update was analyzed and assigned either the single most descriptive annotation, or no annotation, if the article was not deemed to be potentially motivating for a review update.
Annotation counts by type and systematic review topic
| | ||||||
|---|---|---|---|---|---|---|
| 1 | - | 1 | - | 10 | 12 | |
| 2 | - | 3 | - | - | 5 | |
| - | - | 1 | 1 | 15 | 17 | |
| 1 | 8 | 2 | - | - | 11 | |
| 1 | - | 3 | 1 | - | 5 | |
| - | - | - | - | 3 | 3 | |
| - | 1 | - | - | - | 1 | |
| 7 | 1 | 2 | - | 16 | 26 | |
| - | - | 3 | - | - | 3 | |
| 12 | 10 | 12 | 2 | 44 | ||
Overall, correct, and incorrect alert rates as well as recall of important publications for each topic
| 15 | 2.67 | 1.33 | 1.33 | 1.00 | |
| 20 | 1.60 | 0.65 | 0.95 | 0.40 | |
| 41 | 0.59 | 0.24 | 0.34 | 0.88 | |
| 16 | 1.19 | 0.56 | 0.63 | 0.72 | |
| 29 | 0.34 | 0.21 | 0.14 | 0.80 | |
| 19 | 2.58 | 0.79 | 1.79 | 0.33 | |
| 31 | 1.48 | 0.68 | 0.81 | 1.00 | |
| 31 | 0.48 | 0.39 | 0.10 | 0.46 | |
| 19 | 1.16 | 0.68 | 0.47 | 1.00 | |
| 24.56 | 1.34 | 0.61 | 0.73 | 0.73 |
Classifier performance on the training and test sets at the closest threshold to a recall of 0.55 on the training set for each topic
| | | |||||||
|---|---|---|---|---|---|---|---|---|
| −0.2466 | 155 | 2423 | 119 | 130 | 0.566 | 0.544 | 0.555 | |
| −0.3302 | 57 | 1361 | 75 | 39 | 0.432 | 0.594 | 0.500 | |
| −0.3550 | 70 | 2196 | 168 | 63 | 0.294 | 0.526 | 0.377 | |
| −0.1696 | 375 | 3544 | 285 | 284 | 0.568 | 0.569 | 0.569 | |
| −0.2390 | 77 | 1699 | 75 | 61 | 0.507 | 0.558 | 0.531 | |
| −0.3943 | 8 | 61 | 21 | 5 | 0.276 | 0.615 | 0.381 | |
| −0.3461 | 57 | 695 | 128 | 46 | 0.308 | 0.553 | 0.396 | |
| −0.2532 | 117 | 2235 | 98 | 87 | 0.544 | 0.574 | 0.558 | |
| −0.2617 | 79 | 1583 | 74 | 66 | 0.516 | 0.545 | 0.530 | |
True positives (TP), true negatives (TN), false positives (FP), false negatives (FN), F1 measure (F1, the harmonic mean of precision and recall).
Classifier performance on the training and test sets at the closest threshold to a recall of 0.55 on the training set for each topic
| | | |||||||
|---|---|---|---|---|---|---|---|---|
| −0.2466 | 20 | 225 | 20 | 23 | 0.500 | 0.465 | 0.482 | |
| −0.3302 | 13 | 457 | 19 | 26 | 0.406 | 0.333 | 0.366 | |
| −0.3550 | 10 | 320 | 14 | 19 | 0.417 | 0.345 | 0.377 | |
| −0.1696 | 9 | 775 | 10 | 58 | 0.474 | 0.134 | 0.209 | |
| −0.2390 | 6 | 597 | 4 | 32 | 0.600 | 0.158 | 0.250 | |
| −0.3943 | 15 | 28 | 34 | 0 | 0.306 | 1.000 | 0.469 | |
| −0.3461 | 21 | 204 | 25 | 4 | 0.457 | 0.840 | 0.592 | |
| −0.2532 | 12 | 372 | 3 | 38 | 0.800 | 0.240 | 0.369 | |
| −0.2617 | 13 | 197 | 9 | 13 | 0.591 | 0.500 | 0.542 | |
True positives (TP), true negatives (TN), false positives (FP), false negatives (FN), F1 measure (F1, the harmonic mean of precision and recall).