| Literature DB >> 30984504 |
Aidin Abedi1, Lidwine B Mokkink2,3, Shayan Abdollah Zadegan4, Permsak Paholpak5, Koji Tamai6, Jeffrey C Wang1, Zorica Buser1.
Abstract
STUDYEntities:
Keywords: classification; injury severity score; reliability; reproducibility of results; spinal fractures; spinal injuries; spine; thoracolumbar; trauma; validity
Year: 2018 PMID: 30984504 PMCID: PMC6448204 DOI: 10.1177/2192568218806847
Source DB: PubMed Journal: Global Spine J ISSN: 2192-5682
Figure 1.PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram[33] of the screening and selection process.
Characteristics of the Included Studies.
| First Author (Year) | Sampling Method | Cases | Observers | Imaging Modality | |||
|---|---|---|---|---|---|---|---|
| n | Injury Distributiona | No. of Injury Levels | n | Characteristics | |||
| Vaccaro (2013)[ | Random | n = 40 | Type A: 54% | NR | n = 9 | Spine surgeons | NR |
| Kepler (2016)[ | Purposive | n = 25 | Type A: 53% | NR | n = 100 | Surgeons from Africa, Asia, Europe, North America, and South America | High-quality CT |
| Azimi (2015)[ | Random | n = 56 | Type A: 41.9% | 74 | n = 2 | Spine surgeons | Plain X-ray, CT, and MRI |
| Barcelos (2016)[ | Consecutive | n = 43 | Type A: 32.5% | NR | n = 3 | Spine surgeons | CT (axial, sagittal, and coronal) |
| Kaul (2016)[ | Consecutive | n = 50 | Type A: 39.45% | NR | n = 11 | Surgeons from 4 countries. | Plain X-ray, CT, and MRI |
| Sadiqi (2015)[ | Purposive | n = 25 | Subtype A0: 4% | NR | n = 100 | International group of spine surgeons naïve to the
classification. | High-quality CT |
| Schroeder (2015) [ | Purposive | n = 6 | Subtype A3: n = 3 | NR | n = 100 | International group of spine surgeons naïve to the classification | High-quality CT |
| Urrutia (2015)[ | Purposive | n = 70 | Type A: 49.8% | NR | n = 6 | Spine surgeons (fellowship-trained): n = 3 | Plain X-ray (anteroposterior and lateral), multislice 64 channel CT (axial view and sagittal reconstruction) |
| Yacoub (2015)[ | NR | n = 54 | NR | NR | n = 2 | Spine surgeon: n = 1 | Multislice 64-channel CT with reconstruction |
Abbreviations: NR, not reported; CT, computed tomography; MRI, magnetic resonance imaging.
a Based on AOSpine Thoracolumbar Injury Classification System.[3]
Summary of the Results of Studies on Reliability and Measurement Error of the AOSpine Thoracolumbar Injury Classification System.
| First Author (Year) | Findings | ||||
|---|---|---|---|---|---|
| Intraobserver | Interobserver | ||||
| Time Interval | Reliability (Kappa Values) | Measurement Error ( | Reliability (Kappa Values) | Measurement Error ( | |
| Vaccaro (2013)[ | 1 month | Overall: Mean = 0.77 (range: 0.6-0.97) | Overall: 0.64 | Overall: 35% | |
| Kepler (2016)[ | 1 month | Overall: mean = 0.68 (range: 0.22-1) | Overall: 0.56 | Overall: 0% | |
| Azimi (2015)[ | 5 weeks | Type A: 0.84 (95% CI: 0.82-0.91) | Type A: 0.88 (95% CI: 0.80-0.94) | ||
| Barcelos (2016)[ | |||||
| First assessment | Overall without subtypes: 0.526 | ||||
| Second assessment | Overall without subtypes: 0.645 | ||||
| Kaul (2016)[ | 6 weeks | Overall: 0.61 (SE = 0.13) | Overall: 0.45 (SE = 0.01) | Overall: 32% | |
| Sadiqi (2015)[ | 1 month | Overall: | Agreement with predefined gold standard: | ||
| Schroeder (2015)[ | Agreement with predefined gold standard: | ||||
| Urrutia (2015) [ | 6 weeks | Overall: | Overall: 75.71% | Overall: 0.55 (95% CI: 0.52-0.57) | Overall: 30% |
| Yacoub (2015)[ | 8 weeks | Type A: 0.75 | Overall: | Subtypes: 0 (A2) to 0.85 (C) | Overall: 55.5% |
Abbreviations: CI, confidence interval; SE, standard error.
Summary of Evidence on Intraobserver Reliability of the AOSpine Thoracolumbar Injury Classification System.
| First Author (Year) | Study Quality | Quality of Findingsa | ||||||
|---|---|---|---|---|---|---|---|---|
| Morphologic Classification | Neurologic Injury | Modifiers | ||||||
| Overall | Overall Without Subtypes | Type A | Type B | Type C | ||||
| Vaccaro (2013)[ | Fair | + | + | + | − | 0 | 0 | 0 |
| Kepler (2016)[ | Fair | − | + | − | − | 0 | 0 | 0 |
| Azimi (2015)[ | Fair | 0 | 0 | + | + | + | 0 | 0 |
| Kaul (2016)[ | Fair | − | − | 0 | 0 | 0 | + | 0 |
| Sadiqi (2015)[ | Fair | − | + | 0 | 0 | 0 | 0 | 0 |
| Urrutia (2015)[ | Fair | + | + | 0 | 0 | 0 | 0 | 0 |
| Yacoub (2015)[ | Fair | 0 | 0 | + | + | + | 0 | 0 |
| Overall quality of findingsa | Conflicting | + | Conflicting | Conflicting | + | + | 0 | |
| Overall quality of evidenceb | Low | Moderate | Low | Low | Moderate | Low | Unknown | |
a +, positive rating; ?, indeterminate rating; −, negative rating; 0, not reported.[18]
b According to the criteria by Prinsen et al[19] (Appendix D).
Summary of Evidence on Interobserver Reliability of the AOSpine Thoracolumbar Injury Classification System.
| First Author (Year) | Study Quality | Quality of Findingsa | ||||||
|---|---|---|---|---|---|---|---|---|
| Morphologic Classification | Neurologic Classification | Modifiers | ||||||
| Overall | Overall Without Subtypes | Type A | Type B | Type C | ||||
| Vaccaro (2013)[ | Fair | − | + | + | − | + | 0 | 0 |
| Kepler (2016)[ | Fair | − | + | + | − | + | 0 | 0 |
| Azimi (2015)[ | Fair | 0 | 0 | + | + | + | 0 | 0 |
| Barcelos (2016)[ | Fair | 0 | − | −/+b | − | − | 0 | 0 |
| Kaul (2016)[ | Fair | − | − | − | − | + | + | 0 |
| Urrutia (2015)[ | Fair | − | − | − | − | − | 0 | 0 |
| Yacoub (2015)[ | Fair | 0 | 0 | 0 | 0 | 0 | 0 | 0 |
| Overall quality of findingsa | − | Conflicting | Conflicting | − | Conflicting | + | 0 | |
| Overall quality of evidencec | Moderate | Low | Low | Moderate | Low | Low | Unknown | |
a +, positive rating; ?, indeterminate rating; −, negative rating; 0, not reported.[18]
b Two assessments had conflicting findings.
c According to the criteria by Prinsen et al[19] (Appendix D).
Adapted COSMIN Checklist for Evaluation of the Methodological Quality of Studies on Reliability and Measurement Error of Ordinal Classification Systems.
| Revised COSMIN Checklist—Reliability and Measurement Error | ||||||
|---|---|---|---|---|---|---|
| Excellent | Good | Fair | Poor | Not Applicable | ||
|
| ||||||
| 1 | Were at least 2 measurements available? | At least 2 measurements | Only 1 measurement | |||
| 2 | Were the administrations independent? | Independent measurements | Assumable that the measurements were independent | Doubtful whether the measurements were independent | measurements NOT independent | |
| 3 | Was the time interval stated? | Time interval stated | Time interval NOT stated | * | ||
| 4 | Were patients stable in the interim period on the construct to be measured? | Patients were stable (evidence provided) | Assumable that patients were stable | Unclear if patients were stable | Patients were NOT stable | * |
| 5 | Were observers stable in the interim period? | Observers were stable (evidence provided) | Assumable that observers were stable | Unclear if observers were stable | Observers were NOT stable, eg, received additional training | * |
| 6 | Was the time interval appropriate? | Time interval appropriate | Doubtful whether time interval was appropriate | Time interval NOT appropriate | * | |
| 7 | Were the test conditions similar for both measurements? For example, type of administration, environment, instructions | Test conditions were similar (evidence provided) | Assumable that test conditions were similar | Unclear if test conditions were similar | Test conditions were NOT similar | |
| 8 | Were there any important flaws in the design or methods of the study? | No other important methodological flaws in the design or execution of the study | Other minor methodological flaws in the design or execution of the study, e.g. lack of blinding regarding the clinical information | Other important methodological flaws in the design or execution of the study | ||
|
| ||||||
| 9 | Reliability studies: Was kappa calculated? | Kappa calculated | Kappa not calculated | |||
| 10 | Reliability studies: Was a weighted kappa calculated? | Weighted Kappa calculated | Unweighted kappa calculated | |||
| 11 | Reliability studies: Was the weighting scheme described? For example, linear, quadratic | Weighting scheme described | Weighting scheme NOT described | * | ||
| 12 | Measurement error studies: Was percentage agreement calculated? | Percentage agreement calculated | Percentage agreement not calculated | |||
Adapted from Terwee et al[17] under a Creative Commons Attribution–Noncommercial (http://creativecommons.org/licenses/by-nc/4.0/).
Criteria for Evaluation of the Quality of Results.
| Measurement Property | Ratinga | Criteria |
|---|---|---|
| Content validity (including face validity) | + | All items refer to relevant aspects of the construct to be measured AND are relevant for the target population AND are relevant for the context of use AND together comprehensively reflect the construct to be measured. |
| ? | Not all information for “+” reported. | |
| − | Criteria for “+” not met. | |
| Reliability | + | ICC or weighted kappa ≥0.70. |
| ? | ICC or weighted kappa not reported. | |
| − | Criteria for “+” not met. | |
| Measurement error | + | SDC or LoA < MIC. |
| ? | MIC not defined. | |
| − | Criteria for “+” not met. | |
| Construct validity | + | At least 75% of the results are in accordance with the hypotheses. |
| ? | No correlations with instrument(s) measuring related construct(s) AND no differences between relevant groups reported. | |
| − | Criteria for “+” not met. |
Adapted from Prinsen et al[19] (as modified from Terwee et al[18]) under a Creative Commons Attribution 4.0 (http://creativecommons.org/licenses/by/4.0/).
Abbreviations: ICC, intraclass correlation coefficient; SDC, smallest detectable change; LoA, limits of agreement; MIC, minimal important change.
a +, positive rating; ?, indeterminate rating; −, negative rating.
Criteria for Evaluation of the Quality of Evidence.
| Quality Rating | Criteria |
|---|---|
| High | Consistent findings in multiple studies of at least good quality OR one study of excellent quality AND a total sample size of ≥100 patients |
| Moderate | Conflicting findings in multiple studies of at least good quality OR consistent findings in multiple studies of at least fair quality OR one study of good quality AND a total sample size of ≥50 patients |
| Low | Conflicting findings in multiple studies of at least fair quality OR one study of fair quality AND a total sample size of ≥30 patients |
| Very low | Only studies of poor quality OR a total sample size of <30 patients |
| Unknown | No studies |
Reused from Prinsen et al. [19] under a Creative Commons Attribution 4.0 (http://creativecommons.org/licenses/by/4.0/).