Literature DB >> 32871264

Comparison of the accuracy of telehealth examination versus clinical examination in the detection of shoulder pathology.

Kendall E Bradley1, Chad Cook2, Emily K Reinke3, Emily N Vinson4, Richard C Mather3, Jonathan Riboh3, Tally Lassiter3, Jocelyn R Wittstein3.   

Abstract

HYPOTHESIS/
BACKGROUND: In 2017, the American Orthopaedic Association advocated for the increased use of telehealth as an assessment and treatment platform, and demand has significantly increased during the coronavirus disease 2019 pandemic. Diagnostic effectiveness (also called overall diagnostic accuracy) and reliability of a telehealth clinical examination vs. a traditional shoulder clinical examination (SCE) has not been established. Our objective is to compare the diagnostic effectiveness of a telehealth shoulder examination against an SCE for rotator cuff tear (RCT), using magnetic resonance imaging (MRI) as a reference standard; secondary objectives included assessing agreement between test platforms and validity of individualized tests. We hypothesize that tests provided in a telehealth platform would not have inferior diagnostic effectiveness to an SCE.
METHODS: The study is a case-based, case-control design. Two clinicians selected movement, strength, and special tests for the SCE that are associated with the diagnosis of RCT and identified similar tests to replicate for a simulated telehealth-based examination (STE). Consecutive patients with no prior shoulder surgery or advanced imaging underwent both the SCE and STE in the same visit using 2 separate assessors. We randomized the order of the SCE or STE. A blinded reader assessed an MRI, to be used as a reference standard. We calculated diagnostic effectiveness, which provides values from 0% to 100% as well as agreement statistics (Kappa) between tests by an assessment platform, and sensitivity, specificity, and likelihood ratios for individual tests in both SCE and STE. We compared the diagnostic effectiveness (overall) of the SCE and STE with a Mann-Whitney U test.
RESULTS: We included 62 consecutive patients with shoulder pain, aged 40 or older; 50 (81%) received an MRI as a reference standard. The diagnostic effectiveness of stand-alone tests was poor regardless of the group, with the exception of a few tests with high specificity. None had greater than 70% accuracy. There was no significant difference between the overall diagnostic effectiveness of the STE and SCE (P = .98). Overall agreement between the STE tests and the SCE tests ranged from poor to moderate (Kappa, 0.07-0.87).
CONCLUSION: This study identified initial feasibility and noninferiority of the physician-guided, patient-performed STE when compared with an SCE in the detection of RCTs. Although these results are promising, larger studies are needed for further validation of an STE assessment platform.
Copyright © 2020 Journal of Shoulder and Elbow Surgery Board of Trustees. Published by Elsevier Inc. All rights reserved.

Entities:  

Keywords:  Shoulder; clinical assessment; diagnostic accuracy; imaging; magnetic resonance; rotator cuff tear; telehealth

Year:  2020        PMID: 32871264      PMCID: PMC7455801          DOI: 10.1016/j.jse.2020.08.016

Source DB:  PubMed          Journal:  J Shoulder Elbow Surg        ISSN: 1058-2746            Impact factor:   3.019


Shoulder pain is a common cause of disability in the adult population, with rotator cuff tendon tears increasing in frequency each decade after the age of 40 years and affecting approximately 25% of adults over the age of 50. , Differentiating shoulder pain requires a careful assessment of patient report, physical examination, and imaging. Meta-analyses have demonstrated that the stand-alone diagnostic utility of movement, physical testing, and special tests of the shoulder are poor. , Imaging fares better and improves the ability to identify full-thickness rotator cuff tears (RCTs) primarily in addition to other shoulder pathologies. These challenges suggest that a shoulder specialist is required to distinguish between the prevalent shoulder pathologies, variabilities in test findings, and imaging findings. , Unfortunately, there is a lack of access to subspecialized orthopedic care, including shoulder specialists, in rural settings as compared with urban settings. , , , , Access issues will worsen with a projected shortfall of 20,000-30,000 surgical specialists by 2030. To address this shortfall, some shoulder surgeons have integrated telehealth as an alternative to conventional clinical care. Telehealth evaluations are cost-effective and provide access to specialized care in a variety of orthopedic conditions in the United States and abroad. , , , , , , In Finland, teleconsultations allowed general practitioners to examine and diagnose 25% of patients who otherwise would have required referral to an outside provider. The United States Army began using telehealth medicine for orthopedic surgery in July 2007 for soldiers deployed overseas. The use of telehealth as an assessment and treatment platform has been advocated by the American Orthopaedic Association. Public demand for orthopedic telehealth services has surged with the arrival of the coronavirus disease 2019 (COVID-19) pandemic in early 2020. To date, most telehealth-based studies are survey oriented or involve physician and patient perceptions of care. There are no studies that compare the diagnostic effectiveness (also known as overall diagnostic accuracy) of a telehealth examination platform with a standard clinical examination (SCE) of the shoulder. One study examined the accuracy of self-administered hip examination for FAI and actually showed that the accuracy of telehealth-based assessment was slightly higher. Our objective is to compare the diagnostic effectiveness of a simulated telehealth shoulder examination (STE) against an SCE for RCT, using magnetic resonance imaging (MRI) as a reference standard. Secondary objectives included assessing agreement between test platforms and diagnostic validity of individualized tests (eg, sensitivity, specificity, likelihood ratios). We hypothesize that the STE would be noninferior to an SCE in accurately diagnosing RCTs.

Materials and methods

Study design

The study is a case-based, case-control design. We used the Standards for Reporting Diagnostic Accuracy Studies (STARD) reporting standards to guide this study. All care in outpatient, orthopedic specialists’ practices was provided by 3 orthopedic surgeons (J.R.W., T.L., J.R.). Patients provided electronic informed consent and then underwent both the SCE and STE during the same visit but by 2 different providers to avoid confirmation bias. We randomized the order in which this examination was performed.

Eligibility and exclusion criteria

We recruited consecutive patients, 40 years of age or older, presenting with shoulder pain and seen in the Duke Sports Medicine Clinic when research staff were available for the study. We excluded patients from the study if the shoulder being evaluated had prior shoulder arthroplasty, instability, prior imaging revealing a rotator cuff injury, history of fracture/dislocation prior advanced imaging, or contraindications to advanced imaging. We did not compensate subjects. We provided MRIs for all patients, paid for through research funds, but not change other patient management for the standard of care.

Index testing

This study involved 2 sets of index tests for 2 assessment platforms (SCE and STE). We selected the SCE tests if they were commonly used tests and measures from the literature and practice, and if they achieved reasonable diagnostic accuracy in the summated study. We designed tests to identify all types of RCTs (ie, supraspinatus, infraspinatus, subscapularis) and selected tests for either having high sensitivity or specificity, or when available, both. The senior author, an orthopedic surgeon with 10 years of experience, and a physical therapist who is PhD and has specialized in diagnostic accuracy research for 15 years created the STE tests. The goal of both clinicians was to identify tests that reflected a clinical examination; each SCE test had an analogous “sister” STE test they created to reflect its purpose in clinical practice. We included tests that detected other shoulder pathologies in order to provide the standard of care to patients for a shoulder examination. The goal was also to create tests that were transferable to any telehealth setting with a video feed. A description of these testing procedures is given in Table I .
Table I

Shoulder examination

Examination maneuverDescriptionTelehealth modificationPositive test
Rotator cuff—supraspinatus/infraspinatus
Drop armPatient actively lowers the arm from the abducted position to the sideNo modificationInability to control the arm while lowering it from flexion
Shoulder shrug24Inability to abduct the arm to 90° without elevation of the scapulaNo modificationScapula elevates
ER lag sign18Elbow passive flexed to 90° with shoulder near maximum external rotation in abduction. Wrist is releasedArm supported by a table; patient passively externally rotates the affected shoulder to the maximal ER, and then releases the supportShoulder internally rotates/forearm falls toward the table
Active elevation deficitActive and passive forward flexion recorded, with notation of deficitPassive forward flexion performed by gripping the wand and pushing the affected shoulder into maximal passive flexion using the contralateral armDeficit of active flexion relative to passive flexion
ER weaknessSubjective rating of MMTIsometric ER testing with resistance by the contralateral hand at the dorsal wristMMT <4+Patient report of subjective weakness
Abduction weakness26Isometric abduction testing against the wall
ER pain with strength testingPatient asked if test painfulNo modification+ Pain
Abduction pain with strength testingNo modification
Rotator cuff—subscapularis
Lift-off sign10Patient attempts to lift the hand away from his or her backNo modificationPatient unable to lift the hand off back
Belly press9Patient presses the hands to the abdomen while the elbows push forwardNo modificationElbows drop backward with wrist flexion
IR weakness26Subjective rating of MMTIsometric IR testing with resistance by the contralateral fist at the palmar aspect of the handMMT <4+Patient report of subjective weakness
IR pain with strength testingPatient asked if test painfulNo modification+ Pain
Impingement
Hawkins Kennedy14Arm flexed 90° while passively internally rotating the shoulderPatient flexes the shoulder, supports the elbow, and passively internally rotates the shoulder with the contralateral armPain
Neer’s sign38Arm flexed causing impingement of the greater tuberosity against the acromionPatient forward flexes the shoulder with the hand pronatedPain at the anterior edge of the acromion
Painful arc28Shoulder actively abducted from 60° to 180°No modificationPain with the elevation of the shoulder
Adhesive capsulitis
PROM flexionExaminer passively maximally flexes the arm while the patient standing and controls movement of the scapulaPatient holds the stick vertically and the opposite arm assists forward flexion0-max degrees (continuous)Positive test was a 25% limitation of the affected side to the nonaffected
ER affected to contralateral limitationWith elbows at side and elbow flexed to 90°, the examiner maximally externally rotates and repeats on the contralateral sidePatient holds the stick horizontally and the opposite arm assists to externally rotate0-max degrees (continuous)Positive test was a 10° reduction of the affected side to contralateral
IR limitationPatient turns around and attempts to reach toward the scapulaNo modificationBy max level (to lateral thigh, buttock, LS region, waist, T12, interscapular region). Positive test was inability to lift to T12

ER, external rotation; MMT, manual motor testing; IR, internal rotation; PROM, passive range of motion.

Shoulder examination ER, external rotation; MMT, manual motor testing; IR, internal rotation; PROM, passive range of motion.

Clinical testing

We randomized patients to undergo the STE or SCE first. Randomization was assigned using randomly assorted blocks of sizes 4 and 6 and stratified by the attending surgeon. All patients underwent both examinations. For the SCE, 1 of 3 fellowship-trained sports medicine orthopedic surgeons (J.R.W., T.L., J.R.) performed a set of shoulder examination procedures in the predetermined order. They performed the shoulder examination maneuvers in the same order as the analogous STE maneuvers. To avoid bias, the senior surgeon obtained a detailed history and reviewed radiographs only after all examination procedures were complete.

Simulated telehealth testing

For the STE, time permitting, we invited the patient to view a tutorial video of the STE. A research coordinator used a portable electronic device (Apple iPad; Apple, Cupertino, CA, USA) equipped with a video camera to image the patient, and a senior orthopedic resident (PGY4-6) served as the telehealth examiner performing the physician-guided, patient-performed telehealth examination from a separate room at a desktop interface (Cisco Webex DX80 or Cisco Webex DX70; Cisco, San Jose, CA, USA). We used a senior in order to eliminate the bias of the attending surgeon might have if they performed both examinations. A total of 9 senior residents and fellows participated in the telehealth version of the examination (45% by 4 fellows, 32% by 2 PGY5 residents, and 23% by 3 PGY4 residents). These were the residents and fellows on service with the attending. Both devices had video capability such that the patient and examiner could see, hear, and observe each other. For the STE, the senior author developed a script to ask standardized questions regarding the quality of the shoulder pain in order to minimize differences between trainees. The resident then directed the patient through a series of self-examination maneuvers, using the script provided. The examinations were meant to mimic traditional clinic testing.

Data management

We collected and managed all study data, with the exception of the MRI images, using REDCap electronic data capture tools hosted at Duke University. REDCap (Research Electronic Data Capture; , Nashville, TN, ) is a secure web-based software platform designed to support data capture for research studies, providing: (1) an interface for validated data capture, (2) audit trails for tracking data manipulation and export procedures, (3) export procedures for data downloads to statistical packages, and (4) procedures for data integration and interoperability with external sources. ,

Reference standard

We compared the STE of shoulder pain using patient self-examination with the SCE using MRI as the gold standard to determine the accuracy of detecting RCTs. Fifty (81%) patients had shoulder imaging obtained using MRI with a dedicated shoulder coil. All 50 patients underwent a nonarthrographic shoulder MRI examination performed on a 3.0-Tesla MR scanner (Trio TIM; Siemens Healthcare, Erlangen, Germany) using a phased array 8-channel shoulder coil (Invivo). The imaging protocol consisted of the following sequences: axial, oblique sagittal, and oblique coronal fat-suppressed fast spin-echo T2-weighted sequences (slice thickness, 3.0 mm; FOV, 16 cm; TR/TE, 3000/65); axial fat-suppressed fast spin-echo intermediate-weighted sequence (slice thickness, 3.0 mm; FOV, 16 cm; TR/TE, 3000/23); and oblique sagittal T1-weighted sequence (slice thickness, 3.0 mm; FOV, 16 cm; TR/TE, 688/11). A single musculoskeletal radiologist who was blinded to the clinical findings reviewed the research MRIs prospectively and recorded the presence or absence of partial and complete RCTs as well as tear locations. MRI was considered positive if the tear was complete.

Sample size estimate

When determining the sample size of a noninferiority trial, one aim is to show that a new testing platform is not unacceptably worse than an older one (SCE). To do so, we need to select a noninferiority margin, calculate the confidence window around the difference between the treatments, and then determine the acceptability of difference one is willing to accept. We estimated the sample size with the intention to treat analyses, and these forms of analyses typically lead to noninferiority between groups. We assumed a 20% difference in overall diagnostic effectiveness between groups as unacceptable, and at 95% power, our projected use of a Mann-Whitney U test for comparison of differences, and an error probability of .05, this led to a projected sample size of 60. Because of COVID-19 and the subsequent closure of the MRI facility, 50 of these patients completed MRIs before the publication of this paper, whereas the remaining 12 subjects were indefinitely unable to access the research scanner. Complete data for all 62 were available in 100% of physical examination test items.

Data analysis and statistical considerations

We performed the analysis using SPSS (V26.0; IBM, Armonk, NY, USA) and a publicly available online software calculator from the University of Illinois, Chicago (http://araw.mede.uic.edu/cgi-bin/testcalc.pl). We first tabulated summary values for age, gender, and diagnoses based on a radiology read. We calculated agreement between tests using Cohen’s Kappa. Cohen’s Kappa calculates the chance-corrected agreement between 2 or more raters. Kappa values may range from 0 (perfect lack of agreement) to +1.0 (perfect agreement). It is possible for the statistic to be negative, which suggests that the agreement is worse than random. Although arbitrary, Landis and Koch provided cutoff values for interpretation as follows: <0, no agreement; 0-0.20, slight; 0.21-0.40, fair; 0.41-0.60, moderate; 0.61-0.80, substantial; and 0.81-1, almost perfect agreement. We calculated diagnostic accuracy measures of sensitivity (SN), specificity (SP), positive likelihood ratio (LR+), and negative likelihood ratio (LR−) for each component of the STE and SCE. SN is the ability of the test to identify a positive finding when the targeted diagnosis is present. SP is the ability of the test to identify a negative finding when the targeted is negative. LR+ indicates a shift in probability supporting the existence of a disorder if the test is found to be positive. Values greater than 1 indicate an increased ability to diagnose a condition with a positive finding. LR− indicates a shift in probability supporting the absence of a disorder if the test is found to be negative. Values less than 1 and closer to 0 are used to rule out a diagnosis. Jaeschke et al provided arbitrary cutoffs for likelihood variables, suggesting that LR+ of >10.0 and LR− of <0.10 have large increases or decreases in the likelihood of the disease. LR+ of >5 to 10 and LR− of <0.20 to 0.10 have moderate increases and decreases in the likelihood of the disease. For this study, we did not attempt to differentiate location of tear during determination of accuracy of each test (ie, subscapularis vs. supraspinatus), although location can be noted. For our primary objective, we analyzed overall diagnostic effectiveness ((TP + TN)/(TP + TN + FP + FN) × 100). , We calculated this for both SCE and STE tests and summated to a grand mean. For each individual test, values can range from 0% (completely inaccurate) to 100% (accurate in ruling in and out). We compared summated mean diagnostic effectiveness using a Mann-Whitney U nonparametric test. We defined a statistically significant finding as P < .05.

Results

From August 2019 to March 2020, 96 consecutive patients who met the inclusion criteria were considered for enrollment (Fig. 1 ). Of the 96 patients, 34 declined to participate in the study. Of the remaining 62 patients, 2 were unable to tolerate an enclosed MRI or had a contraindication to MRI, as discovered by the MRI technician. As previously stated, we included 62 patients for the analysis of agreement between tests (secondary objective) and 50 (81%) for diagnostic effectiveness analysis (primary objective).
Figure 1

Patient enrollment flowchart.

Patient enrollment flowchart. The mean age was 57.9 years (±11.2) and 31 (51.7%) of these patients were women. The 50 patients who received an MRI exhibited a similar demographic distribution: 52% women, with the mean age of 58.2 years. The final diagnosis, per official radiology read, indicated that 22% of patients had a full-thickness supraspinatus tear, whereas 62% of patients had partial tearing of one of their rotator cuff tendons. All of the full-thickness supraspinatus tears were accompanied by other complete or partial tears, most commonly of the infraspinatus tendon. There were no isolated full-thickness tears of subscapularis nor infraspinatus (Fig. 2 ).
Figure 2

Location of rotator cuff tear on MRI as read by the radiologist. SS, supraspinatus; IS, infraspinatus; SSc, subscapularis.

Location of rotator cuff tear on MRI as read by the radiologist. SS, supraspinatus; IS, infraspinatus; SSc, subscapularis. Table II shows the agreement between the clinical and telehealth testing and includes only those tests that have been represented as rotator cuff–related tests or movements in previous studies. Night pain reported by patients had almost perfect agreement between telehealth examination and clinical testing. The painful arc test, shoulder shrug with active abduction, and active internal rotation limitation all had a moderate amount of agreement. External rotation (ER) lag sign, drop arm test, Neer’s sign, abduction weakness with strength testing, active to passive flexion limitation, and passive ER affected to contralateral limitation all had fair agreement.
Table II

Agreement between clinical testing and telehealth testing

TestsKappa statisticStrength of agreementP value
ER lag sign0.32Fair.007
Painful arc test0.42Moderate<.001
Shoulder shrug0.59Moderate<.001
Drop arm test0.36Fair.005
Belly press test0.17Slight.167
Lift-off sign0.03Slight.738
Hawkins Kennedy test0.07Slight.601
Neer’s sign0.22Fair.085
Night pain0.87Almost perfect<.001
ER pain with strength testing0.15Slight.243
IR pain with strength testing0.19Slight.107
Abduction pain with strength testing0.17Slight.148
ER weakness with strength testing0.04Slight.758
IR weakness with strength testing0.14Slight.268
Abduction weakness with strength testing0.32Fair.007
IR limitation0.51Moderate<.001
Active to passive flexion limitation0.35Fair.005
ER affected to contralateral limitation0.27Fair.027

ER, external rotation; IR, internal rotation.

Agreement between clinical testing and telehealth testing ER, external rotation; IR, internal rotation. The STE and SCE both had large LR+ for the drop arm test and active to passive flexion limitation (Table III ). For the SCE, the ER lag sign also had a high LR+, whereas strength testing with ER had a moderate likelihood ratio, LR+ = 4.92. The remaining tests did not provide moderate to large changes in post-test probability but did exhibit differences depending on the platform. The belly press test and IR weakness test had LR+ around 2.0 for both SCE (2.11 and 2.46, respectively) and STE (2.28 and 2.46, respectively). Whereas the SCE had an increased LR+ for abduction weakness (1.93), the STE did not (1.17) (Table III).
Table III

Diagnostic accuracy of each clinical and telehealth test and measure

TestSensitivitySpecificityPositive likelihood ratioNegative likelihood ratio
Clinical testing (SCE)
 ER lag sign2.70100Infinite0.97
 Painful arc test75.723.1.981.05
 Shoulder shrug35.169.21.140.94
 Drop arm test16.2100Infinite0.84
 Belly press test16.292.32.110.91
 Lift-off sign5.484.60.351.12
 Hawkins Kennedy test62.246.21.150.82
 Neer’s sign67.623.10.881.41
 Night pain73.015.40.861.76
 ER pain with strength testing48.653.81.050.95
 IR pain with strength testing32.430.80.472.20
 Abduction pain with strength testing64.915.40.772.28
 ER weakness with strength testing37.892.34.920.67
 IR weakness with strength testing18.992.32.460.88
 Abduction weakness with strength testing59.569.21.930.59
 IR limitation54.169.21.760.66
 Active to passive flexion limitation13.5100Infinite0.86
 ER affected to contralateral limitation37.846.20.71.35
Telehealth testing (STE)
 ER lag sign8.192.31.051.00
 Painful arc test75.723.10.981.05
 Shoulder shrug37.876.91.640.81
 Drop arm test8.3100Infinite0.92
 Belly press test35.184.62.280.77
 Lift-off sign45.976.91.990.70
 Hawkins Kennedy test72.215.40.851.81
 Neer’s sign75.738.51.230.63
 Night pain75.723.10.981.05
 ER pain with strength testing45.961.51.190.88
 IR pain with strength testing27.084.61.760.86
 Abduction pain with strength testing48.638.50.791.34
 ER weakness with strength testing21.653.80.471.46
 IR weakness with strength testing18.992.32.460.88
 Abduction weakness with strength testing27.076.91.170.95
 IR limitation18.976.90.821.05
 Active to passive flexion limitation8.1100Infinite0.92
 ER affected to contralateral limitation51.461.51.340.79

SCE, shoulder clinical examination; ER, external rotation; IR, internal rotation; STE, simulated telehealth-based examination.

Diagnostic accuracy of each clinical and telehealth test and measure SCE, shoulder clinical examination; ER, external rotation; IR, internal rotation; STE, simulated telehealth-based examination. When comparing the diagnostic effectiveness (overall accuracy) for examination maneuvers, the overall accuracy was similar; neither examination was very accurate (Table IV ). The mean diagnostic effectiveness for SCE was 45.53% compared with STE, which was 45.72 % (P = .98). The SCE and STE both had accuracies of around 60% for the painful arc test and night pain. The STE performed poorly for strength testing, with all values being less than 50%. The SCE performed slightly better for abduction weakness, with a value of 62%. The STE performed better for the lift-off sign (54% vs. 26%) and Neer’s sign (66% vs 56%), although these tests still had low accuracy.
Table IV

Comparative analysis of diagnostic effectiveness (overall accuracy) with 50 patients for rotator cuff tears

The testsClinical values of diagnostic accuracy (% correct)Telehealth values of diagnostic accuracy (% correct)
ER lag sign2830
Painful arc test6262
Shoulder shrug4448
Drop arm test3833
Belly press test3446
Lift-off sign2654
Hawkins Kennedy test5856
Neer’s sign5666
Night pain5862
ER pain with strength testing5050
IR pain with strength testing3242
Abduction pain with strength testing4846
ER weakness with strength testing5230
IR weakness with strength testing3838
Abduction weakness with strength testing6240
IR limitation5834
Active to passive flexion limitation3632
ER affected to contralateral limitation4054
Mean diagnostic accuracy45.53%45.72%

ER, external rotation; IR, internal rotation.

P value = .961 (no significant difference).

Comparative analysis of diagnostic effectiveness (overall accuracy) with 50 patients for rotator cuff tears ER, external rotation; IR, internal rotation. P value = .961 (no significant difference).

Discussion

This study endeavored to compare the diagnostic effectiveness (overall accuracy) of an STE with that of an SCE. In this case-based study, we identified commonly used clinical tests and created tests that would complement the clinical tests but could be used in a simulated telehealth setting. Further, 50 of the 62 subjects received an MRI as a reference standard, in which the rater was blinded to the clinical findings of the patient. Our goal was to determine if the overall diagnostic effectiveness, which reflects whether true positives and true negatives were better in one group or another. We think our findings are promising and timely, especially during a global pandemic, when virtual appointments become increasingly important. Several areas we feel are worth discussing.

Agreement

There was fair to good agreement between the SCE and STE. Night pain, as expected, had almost perfect agreement. This test was a question to the patient, and it is interesting that the answer changed in the minutes between the SCE and STE for a few patients. Some patients gave a conditional response, such as “I have pain at night, but it doesn’t keep me awake,” or their answer changed while they were talking. Intuitively, tests that required minimal intervention by an examiner also had higher agreement. For example, the painful arc test, shoulder shrug with active abduction, and active internal rotation limitation all had a moderate amount of agreement. Tests that required explanation or subjective grading, such as the ER lag sign, drop arm test, abduction weakness with strength testing, active to passive flexion limitation, and ER affected to contralateral limitation still exhibited fair agreement.

Diagnostic effectiveness (overall accuracy)

We found that neither the SCE nor STE was accurate to identify an RCT. The tests were usually either sensitive or specific, and in some cases, the tests were neither. Although the accuracy between the SCE and STE was low, these findings are consistent with past meta-analytic literature. , Particularly, this is well represented in studies with smaller sample sizes that did not differentiate tendon tear type (we did not) or which included a case-based designed as we did in our study. It is worth noting that emerging data suggest that differentiating tendon type may also lack accuracy. , Combining tests or examination findings may improve accuracy for the identification of shoulder pathology, although this was not the purpose of our study, nor did we have the power to combine tests to look at conditions. ,

Perception

Most studies evaluating telehealth have focused on the surgeon’s subjective perception of the quality of the examination. We did not ask the providers their perception of the SCE and STE in this study. In Norway, surgeons were asked to evaluate how they felt the quality of the appointment was compared with an in-office consultation. A total of 98% of these surgeons felt that the consultations were good or very good, whereas the visits took about the same amount of time. Both visits took about equal amount of time. There have been some studies that have tried to look at a more objective comparison, such as examination with smartphone photography compared with a manual goniometer to measure elbow range of motion. Future studies with this model should also evaluate the difference in time of visit as well as both patient and clinician satisfaction.

Limitations

This study had a relatively small sample size, and although the results are promising, larger studies will be needed for validation. Our study was also limited by only being able to complete 50 of 62 (81%) MRIs before restrictions on research being placed due to COVID-19. We calculated our power analysis using a 20% difference in overall diagnostic effectiveness. Although this may seem high, diagnostic tests are typically either sensitive or specific, with the majority of tests ranging between 50% and 70% accuracy; thus a 20% difference in plausible. That said, we are comfortable in reporting that there are no differences among groups. If we used the current data, the same statistical analyses, and differences between groups, our post hoc power analysis suggests that we would need over 60,000 subjects in each group to show a statistically significant difference. Although we describe the locations of the specific tendon tears, we did not analyze overall accuracy for full-thickness tears by individual tendon or other shoulder pathologies. Emerging data question the utility of this with clinical tests. , Out of concern for asymptomatic cuff tears, we have intentionally included only full-thickness tears and not partial-thickness tears in our analysis of accuracy for detecting cuff tears. We felt that new shoulder pain in patients over the age of 40 with full-thickness RCTs would likely be symptomatic, possibly with other pathologies as well. Future research into this data set will address alternative diagnoses. , , , , We performed pooled accuracy, rather than dividing specific tests by tear type, as it is well known that clinical tests do not effectively distinguish tears of the rotator cuff across all tendon types. Indeed, cross-over positive findings are very common with some diagnostic clinical tests being used for multiple conditions. For example, a drop arm test will be positive for a supraspinatus tendon injury, but it is also positive for infraspinatus problems as well as impingement. Finally, different trainees performed telehealth examinations, which may have led to some bias on how patients were “coached.” Subsequent larger studies may benefit from using a single telehealth examiner for consistency of administration and for this examination to be performed by a second senior surgeon instead of a trainee. The age of the trainee may have led to increased ease of use of telehealth technology, as compared with an attending surgeon. We attempted to minimize the effect of the telehealth examiner variation by providing an instructional video to the patients before participation and by using a script for the administration of the examination. The senior author also reviewed the examination with the trainee before their administration of the examination. We felt that it was important for the senior author to perform all in-person examinations as part of routine clinical care; however, randomization of the senior author to perform either a telehealth examination or an in-person examination may decrease bias in the future.

Future directions

We felt that, although we were unable to complete all 62 MRI studies, the utility of this study revealing noninferiority of a telehealth examination was important in the wake of the current global pandemic. Future studies should expand the sample size, consider analyzing tests specifically for each RCT lesion, consider accuracy for other shoulder pathologies, and determine cost-effectiveness and management consequences examination using the STE. In addition, assessing patient satisfaction is important as telehealth continues to increase in use. It will also be important to note the impacts of misdiagnosis and malpractice as the telehealth use increases. We also feel that the data gleaned from this study will be useful in future studies of clinical decision-making based on telehealth examination findings and encounters.

Conclusion

This study demonstrated the noninferiority of an STE for rotator cuff pathology. This may increase the geographic footprint of health care networks and give providers an opportunity to evaluate patients in the midst of a pandemic. Future studies are underway to test the accuracy of the STE for different shoulder pathologies, assess clinical decision-making based on the STE, evaluate patient satisfaction, and calculate the cost-effectiveness of this tool.

Disclaimer

Duke Institute for Health Innovations generously provided a grant of $59,700 (DIHI grant no. 2019040104). This provided for purchasing of the electronic equipment and research support for this study. The authors, their immediate families, and any research foundations with which they are affiliated have not received any financial payments or other benefits from any commercial entity related to the subject of this article.
  4 in total

1.  Good Comes From Evil: COVID-19 and the Advent of Telemedicine in Orthopedics.

Authors:  Joseph D Lamplot; Samuel A Taylor
Journal:  HSS J       Date:  2021-02-21

2.  What is the content of virtually delivered pain management programmes for people with persistent musculoskeletal pain? A systematic review.

Authors:  Gregory Booth; Deborah Williams; Hasina Patel; Anthony W Gilbert
Journal:  Br J Pain       Date:  2021-06-13

3.  How Are Orthopaedic Sports Medicine Physicians Triaging Cases and Using Telehealth in Response to COVID-19? A Survey of AOSSM Membership.

Authors:  Brian C Lau; Jocelyn R Wittstein; Oke A Anakwenze
Journal:  Orthop J Sports Med       Date:  2021-03-03

4.  Multicenter Evaluation of Telehealth Utilization in Hip and Knee Arthroplasty Before and for One Year During the COVID-19 Pandemic.

Authors:  Stefano Bini; Yu-Fen Chiu; Michael Ast; Chad Krueger; Joseph Maratt; Ilya Bendich
Journal:  Arthroplast Today       Date:  2021-10-02
  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.