Literature DB >> 31320932

The feasibility, reliability and concurrent validity of the MSReactor computerized cognitive screening tool in multiple sclerosis.

Daniel Merlo1, David Darby2, Tomas Kalincik3, Helmut Butzkueven4, Anneke van der Walt4.   

Abstract

BACKGROUND: Multiple sclerosis (MS) cognitive tests are resource intensive and limited by practice effects that prevent frequent retesting. Brief, reliable and valid monitoring tools are urgently needed to detect subtle, subclinical cognitive changes in people with MS. Cognitive monitoring over time could contribute to a new definition of disease progression, supplementing routine clinical monitoring.
METHODS: MSReactor is a web-based battery that measures psychomotor (processing) speed, visual attention and working memory, using simple reaction time tasks. Clinic-based tasks were completed at baseline and 6 monthly with home testing 1-3 monthly. Acceptability, quality of life, depression and anxiety surveys were completed. We studied its correlation with the Symbol Digit Modalities Test, practice effects, test-retest reliability and the discriminative ability of MSReactor.
RESULTS: A total of 450 people with MS were recruited over 18 months, with 81% opting to complete home-based testing. Most participants (96%) would be happy (or neutral) to repeat the tasks again and just four reported the tasks made them 'very anxious'. Persistence of home testing was high and practice effects stabilized within three tests. MSReactor tasks correlated with Symbol Digit Modalities Test scores and participants with MS performed slower than healthy controls.
CONCLUSION: MSReactor is a scalable and reliable cognitive screening tool that can be used in the clinic and remotely. MSReactor task performance correlated with another highly validated cognitive test, was sensitive to MS and baseline predictors of cognitive performance were identified.

Entities:  

Keywords:  attention; cognition; multiple sclerosis; neuropsychology; processing speed; working memory

Year:  2019        PMID: 31320932      PMCID: PMC6628542          DOI: 10.1177/1756286419859183

Source DB:  PubMed          Journal:  Ther Adv Neurol Disord        ISSN: 1756-2856            Impact factor:   6.570


Introduction

Cognitive impairment affects 40–65% of people with multiple sclerosis (MS), leading to lower rates of employment, social isolation and affected activities of daily living.[1] Cognitive impairment occurs throughout the MS disease course,[2] most commonly impacting information-processing speed, attention, working memory and executive function.[3] In its early stages, cognitive change is, however, difficult to detect, both by clinicians[4] and by standard neuropsychological tests[5] because individuals with cognitive decline will remain within the normal range of standard tests at this time. Complex cognitive batteries and even simpler, adapted tests such as the Brief International Cognitive Assessment for MS[6] require dedicated resources to administer and score, making it impractical to use in under-resourced outpatient clinics. The Symbol Digit Modalities Test (SDMT) is recommended for use as a brief and valid cognitive screening measure where time is limited.[7] Despite the availability of alternate versions of this test, learning effects still occur and this limits their use in situations where frequent and repeated cognitive screening is required, for example, when monitoring for early signs of a treatment response.[8] Other commonly used cognitive screening tools also lack sensitivity to preclinical cognitive change in MS.[9] This is important, as intervention with disease modifying treatments have the greatest impact on physical disability accumulation if used early in the disease course.[10] The same beneficial effects potentially apply to cognition,[11,12] but conclusive evidence regarding long-term effects of current therapies on cognition is lacking.[13] The ability to perform regular cognitive monitoring in the outpatient clinic is currently an unmet need in MS[14] and requires the development of a screening test that can be repeated frequently, with minimal learning effects. An ideal screening test needs to be brief, interesting and self-administered, in addition to being valid, reliable and sensitive to subtle cognitive changes. Computerized screening tests have the potential to address many of these issues. Computerized cognitive batteries have gained traction in other fields of neurology[15] and efficiently screen broad cognitive functions such as information-processing speed, attention and working memory.[16] Where early computerized cognitive tests aimed to replicate existing ‘pen and paper’ tests, recent studies have investigated the basic speed of a response, a measure of information-processing speed. This is a key foundational cognitive domain that can be responsible for impairments in higher cognitive abilities, including working memory and executive function.[17] Computerized cognitive batteries are highly useable,[18] stable and reliable across a range of ages in healthy and impaired populations,[19] can be self-administered and have a relative lack of practice effects due to the ability to generate many alternate versions. In our previous work investigating the use of a computerized battery in MS, the detection (Simple Reaction Time, SRT), identification (Choice Reaction Time, ChRT) and One-Back (OBK) tasks of the CogState brief battery were able to discriminate between 70 MS and 37 healthy controls, with the detection and identification tasks more sensitive to cognitive change over 12 months than the Paced Auditory Serial Addition Test (PASAT).[20] ‘MSReactor’, adapted from tests made available by uBrain (http://ubrain.com.br), is a web-based battery to monitor cognitive abilities in three commonly affected cognitive domains. In this study, we explored the usability, test–retest reliability and practice effects of the MSReactor battery. In addition, we determined the correlation with the SDMT score and compared performance on the cognitive tasks between MS patients and healthy controls (HCs).

Materials and methods

Participants and recruitment

Adult MS participants were recruited between March 2016 and September 2017 from two tertiary MS clinics in Melbourne, Australia. Inclusion criteria included: (a) diagnosis of relapsing–remitting or secondary-progressive MS; (b) no upper limb, visual, or cognitive deficits that preclude performance on a touch-screen device in the clinic; and (c) willingness to use their own computer or tablet device with internet access for home-based testing. HC participants were recruited via community notices, self-enrolled and completed testing via the testing website. The study was approved by the relevant Ethics Committees and all participants provided written informed consent.

Study design

A prospective convenience sample of MS participants were enrolled during their outpatient visit and provided with a unique password to access the testing website. Clinic-based testing was completed at baseline and each subsequent clinic visit (approximately 6 monthly). Optional home-based testing was offered to all participants and performed 1–3 monthly. HC participants completed home-based testing only. All participants completed at least one (maximum of two) brief practice test prior to their baseline test and were encouraged to perform a practice test prior to the home-based test. Immediately following completion of the tasks, electronic surveys assessing acceptability, quality of life (QoL), anxiety and depression were presented. Total clinic-based testing time was 12–15 min. Surveys were omitted from home tests, resulting in a testing time of about 5 min. Persistence was encouraged by two automated email reminders (sent 1 week apart) if no scored test (clinic or home) had been recorded for 3 months.

Computerized cognitive battery (MSReactor)

MSReactor is accessible via any modern internet browser. The battery consisted of three tasks using a set of universal, very simple stimuli presented in a visual game-like interface, including a psychomotor (processing) speed (SRT) test, a visual attention (ChRT) test and a working memory (OBK) test where participants reacted to soccer balls or custom playing cards appearing on the screen. Participants were required to become familiar with the ‘yes’ and ‘no’ buttons and each task displayed a textual instruction screen. For the SRT task, participants pressed the ‘Yes’ button when they detected a yellow ball appear on the screen. For the ChRT task, participants indicated ‘yes’ if the ball was red and ‘no’ if the ball was not red. For the OBK task, participants responded ‘yes’ if the face-up card was identical to the immediately previous card and ‘no’ if the card was different to the previous card. The cards presented in the OBK task consisted of combinations of four colours, four shapes and eight numbers, allowing for 128 unique possibilities in stimuli. All tasks had a prestimulus interval of 1000 ms, a 100–5000 ms stimulus presentation followed by a randomly variable poststimulus interval of between 0 and 1000 ms. These measures ensured alternate forms of the tasks were generated. On completion of the tasks, results were uploaded to a central database, automatically analysed, collated with prior results for the same participant and made available for review by the participants treating physician.

Acceptability, quality of life, depression and anxiety surveys

Participants completed an acceptability questionnaire to assess the enjoyability, level of anxiety, engagement, duration and repeatability of the tasks [Supplementary File (a)]. Depression was assessed using the Patient Health Questionnaire (PHQ-9);[21] anxiety using the Penn State Worry Questionnaire (PSWQ);[22] and QoL assessed using the Multiple Sclerosis Quality-of-Life score.[23]

Concurrent validity and discriminative ability

A convenience subset of MS participants without relapse or steroid treatment completed the pen-and-paper version of the SDMT in addition to the MSReactor in the same testing session. To determine the ability of the MSReactor tasks to discriminate between MS patients and controls without MS, the baseline task performance of this subset of participants was compared with the baseline task performance of HC participants and controlled for education attainment.

Data analysis

Descriptive data are presented as mean and standard deviation (SD), median and interquartile range (IQR) where appropriate and frequency data as proportions. Acceptability was recorded on Likert scales, ranging from a negative response (0) to a positive response (10) and recoded to 5-point ordinal dummy variables for analysis. For each task, the speed of performance was the average reaction time (ms) for the first 30 correct responses. Individual performance speeds were log-transformed and mean reaction times calculated. Accuracy was defined as the proportion of correct responses made for each task, normalized with an arcsine square-root transformation. The probability of discontinuing home testing was assessed using a Cox proportional hazards model, with covariates of age and quartiles of baseline task performance. Correlation between baseline task performance and QoL, depression and anxiety were assessed using a Spearman rank coefficient. To assess baseline associations between task performance and disease and demographic factors, multivariable linear regression was performed with task performance as the dependent variable and age, Extended Disability Status Scale (EDSS) and disease duration as independent variables. The effect of time between repeat testing and the number of completed tests on practice effects was assessed in separate linear mixed-effects models and then together using a multivariate analysis with task performance as the dependent variable. Test–retest reliability was assessed by calculating the concordance correlation coefficient (CCC) between each consecutive pair of tests. To visualize the mean distribution of reaction time over the first 10 repeat tests, a curve was interpolated through each timepoint using nonparametric bootstrap for 10,000 resamples and bias-adjusted confidence intervals calculated from the bootstrapped distributions. The mean first derivative, or slope of a line tangent to the interpolated curve, was calculated for each timepoint and bias-adjusted confidence intervals calculated. One-sample t test was used to compare the first derivative at each timepoint (n = 10,000) to a hypothesized first derivative mean of zero (mu = 0). Performance at the second clinic test (approximately 6 months from baseline) was compared with the preceding home test using a linear mixed-effects model. Devices used to perform home tests were summarized. A general linear model was used to compare baseline performance between MS and controls, with all models controlled for years of education. Raw correlations between MSReactor and SDMT scores were calculated using Pearson’s correlation coefficient. Disattenuated correlation coefficients between the latent test scores were then calculated by adjusting for reliability of the MSReactor (following stabilization of learning effect) and previously published reliability data for the SDMT[24] for the equivalent testing epoch.

Results

Participant characteristics

Characteristics of the 450 MS participants who completed baseline clinic tests are shown in Table 1. Of these, 364 (81%) opted to complete additional home testing, with most of these participants (80%) completing a home test within 3 months of baseline. Most participants completing home testing used the Windows operating system (42%), followed by iOS (38%), Macintosh operating system (13%) and ‘Other’ platform (7%). Seventeen participants (3.8%) withdrew from the study. A subset of 30 MS participants completed the MSReactor tasks and SDMT in the same testing session and the baseline task performance of this subset was compared with the baseline performance of HC participants (n = 30).
Table 1.

MS participant characteristics.

ParticipantsWithdrawers
n (%)n (%)
Total 45017
 RRMS435 (97)17 (100)
 SPMS15 (3%)0
Female 338 (75)12 (70.5)
Age (SD)43.1 years (11.09)44.7 years (9)
EDSS; median (IQR)2 (1–3.5)2 (1–4)
Disease duration (SD)13.52 years (8.14)14.07 (7.56)
Opted to complete home testing 364 (81%)
 Repeated testing within 3 months of enrollment289 (80%)
Withdrawn from study 17 (3.8%)

EDSS, Extended Disability Status Scale; IQR, interquartile range; MS, multiple sclerosis; SD, standard deviation; SPMS, secondary progressive multiple sclerosis; RRMS, relapsing–remitting multiple sclerosis.

MS participant characteristics. EDSS, Extended Disability Status Scale; IQR, interquartile range; MS, multiple sclerosis; SD, standard deviation; SPMS, secondary progressive multiple sclerosis; RRMS, relapsing–remitting multiple sclerosis.

Home-testing persistence

Home-based testing was discontinued by 40 participants (11%) who reverted to clinic-only testing. In multivariate survival analysis, lower quartile (or slower reaction time) performance on all tasks [SRT: hazard ratio (HR) 1.48; 95% confidence interval (CI) 1.10–1.99; ChRT: HR 1.44; 95% CI 1.08–1.93; and OBK: HR 1.35; 95% CI 1.01–1.80] was significantly associated with greater rates of home-testing discontinuation [Figure 1(a–c)]. In addition, older participants were more likely to persist with home testing.
Figure 1.

Probability of home-testing persistence based on quartiles of baseline task performance.

(a) Home-testing persistence based on Simple Reaction Time task performance; (b) home-testing persistence based on Choice Reaction Time task performance; and (c) home-testing persistence based on One-Back task performance.

Probability of home-testing persistence based on quartiles of baseline task performance. (a) Home-testing persistence based on Simple Reaction Time task performance; (b) home-testing persistence based on Choice Reaction Time task performance; and (c) home-testing persistence based on One-Back task performance.

Acceptability

Acceptability surveys were completed by 438 (97.3%) participants at baseline. Participant-rated acceptability of the cognitive tasks was high and is summarized in Table 2.
Table 2.

Baseline acceptability of the MSReactor tasks.

Not anxious at allNot anxiousNeutralSlightly anxiousVery anxiousTotal
Did the test make you anxious? 227 (52%)63 (14.5%)120 (27%)24 (5.5%)4 (1%)438
Very much A little bit Neutral Not really Not at all Total
Did you enjoy the test? 79 (18%)126 (28.8%)222 (51%)10 (2%)1 (0.2%)438
Very interesting A little bit interesting Neutral Not that interesting Very boring Total
Did you find the test interesting? 22 (5%)39 (9%)317 (72%)48 (11%)12 (3%)438
Very happy Happy Neutral Unhappy Very unhappy Total
Would you be happy to repeat the test? 197 (45%)111 (25%)116 (26%)7 (2%)7 (2%)438
Too short Slightly too short About right Slightly too long Too long Total
What did you think about the duration of the test? 3 (0.5%)15 (3.5%)409 (93%)7 (2%)4 (1%)438
Baseline acceptability of the MSReactor tasks.

Quality of life, depression and anxiety

Most participants completed baseline QoL (95.5%), depression (94.9%) and anxiety surveys (94.9%). QoL scores correlated weakly with reaction time on the SRT (r = −0.26, p < 0.001), ChRT (r = −0.29, p < 0.001) and OBK (r = −0.26, p < 0.001). PHQ-9 scores correlated weakly with reaction time on the SRT (r = 0.24, p < 0.001), ChRT (r = 0.26, p < 0.001) and OBK (r = 0.26, p < 0.001). PSWQ scores did not significantly correlate with performance on any of the speed measures (p > 0.05).

Cognitive performance and baseline predictors

Baseline task performance was independently associated with EDSS and age, but not disease duration (Table 3). For the SRT, ChRT and OBK tasks, each one step increase in EDSS resulted in slowing of the transformed reaction times by between 0.015 and 0.02 log milliseconds, translating to a prolonging of between 13 ms and 25 ms in reaction time per step of increase in EDSS. For each year increase in age, reaction times slowed between 0.001 and 0.002 log milliseconds (or 1 ms and 3.2 ms). Sex was associated with faster reaction times on the OBK task only, with males performing 0.029 log milliseconds (or approximately 44 ms) faster than females.
Table 3.

Multivariable linear regression estimates of the association between baseline patient characteristics and the performance on the MSReactor tasks.

MSReactor taskIndependent variable ß 95% confidence intervalp value
Simple Reaction Time Intercept2.4963151
EDSS0.0180.013–0.024<0.0001[*]
Age0.00140.0004–0.00240.006[*]
Sex (male)−0.138−0.14 to 0.010.51
Disease duration0.0005−0.0008 to 0.0020.46
Choice Reaction Time Intercept2.6872189
EDSS0.0170.012–0.022<0.0001[*]
Age0.0010.0002–0.0020.018[*]
Sex (male)−0.01−0.03 to 0.0050.16
Disease duration0.0003−0.0008 to 0.0010.54
One Back Intercept2.8354256
EDSS0.0160.01–0.02<0.0001[*]
Age0.0020.0008–0.003<0.001[*]
Sex (male)−0.029−0.05 to −0.0090.005[*]
Disease duration0.0003−0.0009 to 0.00160.61

Denotes statistical significance.

EDSS, Extended Disability Status Scale.

Multivariable linear regression estimates of the association between baseline patient characteristics and the performance on the MSReactor tasks. Denotes statistical significance. EDSS, Extended Disability Status Scale.

Learning effects and test–retest reliability

To assess learning effects and test–retest reliability, task performance was examined in MS participants (n = 328) who had performed up to 10 successful testing sessions. In this home-testing cohort, the median time interval between tests was 82 days between the first and second test, reducing to 31 days between the second and third test, 29 between the third and fourth test and then stabilizing around 27 days between subsequent tests. In the nonparametric bootstrap fitted data, mean reaction time performance on all tests improved after baseline as evidenced by the slope of the curve being significantly different to the hypothesized mean of zero at baseline (p < 0.001). The slope of the fitted curve stabilized rapidly and no more learning effect was evident from the second test for the SRT task and from the third test for the ChRT and OBK tasks, respectively (Figure 2; one-sample t test shown in Appendix 1). The reliability of the tasks improved over time following stabilization of learning effect and the CCC for test 4–5 was 0.77, 0.71 and 0.83; and for tests 8–9 was 0.83, 0.81 and 0.86 for the SRT, ChRT and OBK, respectively (Figure 3; all CCCs shown in Appendix 2). Mean reaction time performance on all tasks at the second clinic testing session was not significantly different from the preceding home test (p > 0.05).
Figure 2.

Fitted curves and first derivatives for each of Simple Reaction Time, Choice Reaction Time and One-Back memory task reaction time.

Cubic splines were fitted to the distribution of the first 10 tests for each task using nonparametric bootstrap and bias-corrected confidence intervals calculated (a, c, e). The mean first derivative was calculated for each timepoint and bias-adjusted confidence intervals calculated for each timepoint (b, d, f).

*Indicates timepoints where H0 is rejected (p < 0.05) in one-sample t test (μ = 0).

Appendix 1.

One-sample t test of first derivative of bootstrap fitted curves (n = 10,000).

TaskTest numberFirst derivative of fitted curve n SDp value
SRT1−57.091000011092.69E−07[*]
2−7.78100005680.17
33.48100004870.47
4−6.50100006260.30
57.61100006380.23
6−10.02100007150.16
70.88100007510.91
88.92100007860.26
9−2.61100008270.75
10−7.711000016110.63
TaskTest numberFirst derivative of fitted curve n SDp value
ChRT1−74.231000018807.92E−05[*]
2−20.46100007540.007[*]
310.86100009050.23
4−7.35100008930.41
5−0.471000010090.96
6−7.15100009000.43
7−2.841000011500.80
814.441000016370.38
91.591000012170.90
10−15.001000031590.63
TaskTest numberFirst derivative of fitted curve n SDp value
OBK1−91.431000031580.004[*]
2−49.001000011984.34E−05[*]
3−7.321000013930.60
4−7.221000012990.58
5−12.441000013980.37
6−18.971000015360.22
7−4.361000014910.77
89.291000021990.67
90.301000018120.99
10−7.331000039110.85

Denotes statistical significance.

ChRT, Choice Reaction Time; OBK, One-Back memory task; SRT, Simple Reaction Time; SD, standard deviation.

Figure 3.

Test–retest reliability.

The concordance correlation coefficient (CCC) was calculated for performance between consecutive pairs of tests for the Simple Reaction Time, Choice Reaction Time and One-Back memory tasks. The CCC improves over time from between test 1 and 2 (a, b, c); to tests 4 and 5 (d, e, f) and tests 8 and 9 (g, h, i).

Appendix 2.

Concordance correlation coefficient (CCC) between subsequent tests for SRT, ChRT and OBK tasks.

Between testTaskCCCLLCIULCISE
1 and 2 SRT0.490.390.570.047
2 and 3 SRT0.610.520.690.044
3 and 4 SRT0.730.650.780.04
4 and 5 SRT0.770.690.830.04
5 and 6 SRT0.730.620.810.05
6 and 7 SRT0.780.660.850.05
7 and 8 SRT0.750.610.840.06
8 and 9 SRT0.830.720.910.05
9 and 10 SRT0.720.510.850.08
Between testTaskCCCLLCIULCISE
1 and 2 ChRT0.560.470.640.04
2 and 3 ChRT0.680.60.750.04
3 and 4 ChRT0.750.630.780.04
4 and 5 ChRT0.710.60.790.05
5 and 6 ChRT0.780.680.850.04
6 and 7 ChRT0.820.730.880.04
7 and 8 ChRT0.780.640.860.05
8 and 9 ChRT0.810.670.890.055
9 and 10 ChRT0.730.510.860.09
Between testTaskCCCLLCIULCISE
1 and 2 OBK0.680.620.740.03
2 and 3 OBK0.770.70.810.03
3 and 4 OBK0.840.780.880.02
4 and 5 OBK0.830.770.880.03
5 and 6 OBK0.860.80.90.025
6 and 7 OBK0.820.740.880.04
7 and 8 OBK0.850.760.90.04
8 and 9 OBK0.860.760.920.04
9 and 10 OBK0.90.80.950.034

ChRT, Choice Reaction Time; LLCI, lower-level confidence interval; OBK, One-Back memory task; ULCI, upper-level confidence interval; SE, standard error; SRT, Simple Reaction Time.

Fitted curves and first derivatives for each of Simple Reaction Time, Choice Reaction Time and One-Back memory task reaction time. Cubic splines were fitted to the distribution of the first 10 tests for each task using nonparametric bootstrap and bias-corrected confidence intervals calculated (a, c, e). The mean first derivative was calculated for each timepoint and bias-adjusted confidence intervals calculated for each timepoint (b, d, f). *Indicates timepoints where H0 is rejected (p < 0.05) in one-sample t test (μ = 0). Test–retest reliability. The concordance correlation coefficient (CCC) was calculated for performance between consecutive pairs of tests for the Simple Reaction Time, Choice Reaction Time and One-Back memory tasks. The CCC improves over time from between test 1 and 2 (a, b, c); to tests 4 and 5 (d, e, f) and tests 8 and 9 (g, h, i). SDMT scores correlated moderately with SRT performance (Pearson’s r = −0.51, p = 0.004), ChRT performance (r = −0.59, p < 0.001) and OBK performance (r = −0.43, p = 0.015; Figure 4). Disattenuated correlation coefficients were rdis = −0.68, rdis = −0.73 and rdis = −0.50, respectively.
Figure 4.

Correlations between SDMT and MSReactor tasks.

Pearson product-moment correlation coefficient was calculated for a subset of participants (n = 30) who completed the MSReactor battery and the Symbol Digit Modalities Test in the same testing session. Pearson’s r for the Simple Reaction Time (a), Choice Reaction Time (b) and One-Back (c) tasks are shown.

SDMT, Symbol Digit Modalities Test.

Correlations between SDMT and MSReactor tasks. Pearson product-moment correlation coefficient was calculated for a subset of participants (n = 30) who completed the MSReactor battery and the Symbol Digit Modalities Test in the same testing session. Pearson’s r for the Simple Reaction Time (a), Choice Reaction Time (b) and One-Back (c) tasks are shown. SDMT, Symbol Digit Modalities Test. MS (n = 30) and HC participants (n = 30) were well balanced with regards to age [MS mean 41.5 years (SD 11.13) and HC mean 38 years (SD 14.25)], sex [77% (23/30) female and 72% (13/18) female] and years of education [MS mean 15 years (SD 2.72) and HC mean 16.4 years (SD 2.53)] respectively. The mean baseline difference between MS and HC participants for the SRT, ChRT, and OBK tasks was −59.5 ms (95% CI 28–94 ms, p < 0.001), −89 ms (95% CI 23–162 ms, p = 0.01) and −127 ms (95% CI 21–249 ms, p = 0.02), respectively, independent of years of education.

Discussion

To our knowledge, this study is the first to investigate the feasibility of implementing a web-based computerized cognitive screening tool in both the clinic-based and home-based setting for MS. We studied the usability (acceptability, efficiency, stability[18]) and concurrent validity of a computerized cognitive screening platform, MSReactor. Assessing the usability of the battery is an important first step in defining its utility in the clinic setting. Any test that uses an individual’s previous test scores to detect subtle change in cognition needs to be administered regularly. Factors that maintain a patient’s motivation for testing are therefore critical and the task needs to be brief, nonanxiety provoking and reasonably interesting to perform. Participant response to MSReactor tasks were favourable, with most being happy to repeat the testing and the majority indicating that they thought the duration of the tasks was ‘about right’. Only a small fraction of participants found that the tasks made them feel anxious, in contrast to prior studies with tests such as the PASAT, which is frequently reported as aversive and stressful.[25] Implementation of MSReactor is uncomplicated and allows rapid recruitment of large groups of participants. In this study, it allowed 450 participants to be enrolled by a single, nonexpert member of the research team over 18 months at just two clinic sessions per week. The brief testing time of 5–15 min and self-administration of the battery means most participants were able to complete the testing, on their own, while waiting for their clinical consultation with no extra time required. This ease of use and lack of requiring a technical support person[26] is a major practical advantage that makes MSReactor suitable for use in a busy tertiary MS clinic. The majority of participants chose to enrol and also persisted with home testing over time. Benefits of home testing include testing in a familiar or remote environment and allowing frequent testing. This can increase fidelity of serial assessments and should enable earlier detection of change. Home-testing performance over time was equivalent to repeat outpatient clinic testing. The ability to complete testing on a range of everyday electronic screen devices reduced the barrier to remote testing and did not affect the overall performance measures. On the other hand, disadvantages of home testing could include testing in a variable environment, technical support challenges and the possibility of tester substitution. Although compliance for home testing was high during the follow-up period, 40 participants (11%) chose to revert to clinic-only testing. Interestingly, younger participants were less likely to persist with home testing than older participants, a difference possibly attributable to age-related lifestyle and social differences. Poorer baseline performance on MSReactor tasks was also associated with lower home-testing persistence and possibly reflects lack of motivation, frustration, or apathy.[27] Identification of patients who are noncompliant with remote testing could prompt more detailed cognitive evaluation, in addition to offering tailored support to improve testing persistence, including increased email reminders or mobile-phone-optimized platforms. Practice effects can be evident in cognitive measurement tools where regular use leads to improvements in test scores in the absence of neurological change. Although practice effects were not eliminated completely with the MSReactor computerized battery, the learning curve is steep and task performance stabilized within two to three retests, with subsequent high test–retest reliability demonstrated. Task performance correlated only weakly with depression and QoL scores, but not with anxiety. The ability to perform regular testing to identify and quantify the practice effects using a computerized battery is an advantage to standard tools where limited number of alternate versions restrict retest frequency. In a recent study of a computerized version of the SDMT, the Processing Speed Test (PST), Rao and colleagues found significant practice effects in both MS patients and HCs when administered across two sessions (2–3 h apart); however, the persistence of these practice effects in subsequent testing was not explored.[28] Like the PST, the MSReactor tasks demonstrated excellent test–retest reliability following the second administration of the tasks, coinciding with a shorter intertest interval. MSReactor task performance and SDMT scores were moderately correlated. The SDMT is a commonly used, valid and reliable tool that correlates with lesion burden and brain atrophy,[29,30] yet despite these advantages, the SDMT remains impractical to administer in a busy outpatient clinic. Self-administered computerized cognitive batteries such as MSReactor and the PST may be able to address this limitation. The CogState brief battery, a computerized battery employing a similar testing paradigm to MSReactor, was shown to be construct valid, with the strongest associations between the identification task (processing speed) and the SDMT.[16] Although the MSReactor cognitive tasks described here do not interrogate just a single neuropsychological construct [psychomotor (processing) speed, visual attention], the good concurrent correlations with the SDMT provide preliminary evidence of measuring comparative neuropsychological functions. Further work is planned to comprehensively validate the MSReactor battery. The MSReactor tasks were able to discriminate between MS participants and those without MS. Performance on any cognitive task can be influenced by demographics such as educational attainment, age and sex; thus, any meaningful interpretation of cognitive impairment from a test battery must be derived from standardized scores based on normative values. Although the ultimate aim of a screening tool such as MSReactor is to monitor for cognitive change within an individual, where demographics do not change, collection of normative data from people without MS remains a focus of current work. This study had some limitations. Participation in the study was limited to (predominantly) participants with relapsing–remitting MS (RRMS). We are now broadening the population to include clinically isolated syndrome (CIS), as cognitive impairment is present in up to 30% of patients with CIS. As early intervention with disease-modifying therapies has the greatest impact on disability trajectories, we predict that detection of cognitive change in periods of pretreatment observation or during early therapy in CIS and early RRMS is most likely to improve long-term outcome. MSReactor is an innovative and self-administered web-based cognitive battery which is highly scalable, well accepted and reliable, suggesting it should be evaluated further as a cognitive screening tool in MS. It is important to note that computerized cognitive batteries are not intended to replace neuropsychological testing but to act as sensitive screening tools that can prompt further clinical testing.[31] Having a brief self-administered monitoring tool could also provide the treating team and the patient with an earlier indication of subtle changes or cognitive relapses. If confirmed using neuropsychological testing, this could lead to early intervention with education on coping strategies and positive efforts to maintain employability. The results from this study forms the basis of future research to define cognitive trajectories across the MS disease course and impact of treatment change on these trajectories. Click here for additional data file. Supplemental material, Supplementary_file_A_-_The_feasibility,_reliability_and_validity_of_the_MSReactor_computerised_screening_tool_in_multiple_sclerosis for The feasibility, reliability and concurrent validity of the MSReactor computerized cognitive screening tool in multiple sclerosis by Daniel Merlo, David Darby, Tomas Kalincik, Helmut Butzkueven and Anneke van der Walt in Therapeutic Advances in Neurological Disorders
  31 in total

1.  Validity of the CogState brief battery: relationship to standardized tests and sensitivity to cognitive impairment in mild traumatic brain injury, schizophrenia, and AIDS dementia complex.

Authors:  Paul Maruff; Elizabeth Thomas; Lucette Cysique; Bruce Brew; Alex Collie; Peter Snyder; Robert H Pietrzak
Journal:  Arch Clin Neuropsychol       Date:  2009-03-25       Impact factor: 2.813

Review 2.  A comprehensive review of the Paced Auditory Serial Addition Test (PASAT).

Authors:  Tom N Tombaugh
Journal:  Arch Clin Neuropsychol       Date:  2005-11-14       Impact factor: 2.813

3.  Neocortical atrophy, third ventricular width, and cognitive dysfunction in multiple sclerosis.

Authors:  Ralph H B Benedict; Jared M Bruce; Michael G Dwyer; Nadir Abdelrahman; Sarah Hussein; Bianca Weinstock-Guttman; Neeta Garg; Frederick Munschauer; Robert Zivadinov
Journal:  Arch Neurol       Date:  2006-09

4.  Validity of the minimal assessment of cognitive function in multiple sclerosis (MACFIMS).

Authors:  Ralph H B Benedict; Diane Cookfair; Rebecca Gavett; Mark Gunther; Frederick Munschauer; Neeta Garg; Bianca Weinstock-Guttman
Journal:  J Int Neuropsychol Soc       Date:  2006-07       Impact factor: 2.892

5.  Thalamic atrophy and cognition in multiple sclerosis.

Authors:  M K Houtchens; R H B Benedict; R Killiany; J Sharma; Z Jaisani; B Singh; B Weinstock-Guttman; C R G Guttmann; R Bakshi
Journal:  Neurology       Date:  2007-09-18       Impact factor: 9.910

6.  Repeated assessment of neuropsychological deficits in multiple sclerosis using the Symbol Digit Modalities Test and the MS Neuropsychological Screening Questionnaire.

Authors:  R H B Benedict; J A Duquin; S Jurgensen; R A Rudick; J Feitcher; F E Munschauer; M A Panzara; B Weinstock-Guttman
Journal:  Mult Scler       Date:  2008-06-23       Impact factor: 6.312

7.  The relationship between cognitive deficits and everyday functional activities in multiple sclerosis.

Authors:  Jessica H Kalmar; Elizabeth A Gaudino; Nancy B Moore; June Halper; John Deluca
Journal:  Neuropsychology       Date:  2008-07       Impact factor: 3.295

8.  Validation of the Multiple Sclerosis International Quality of Life questionnaire.

Authors:  Mc Simeoni; P Auquier; O Fernandez; P Flachenecker; S Stecchi; Cs Constantinescu; E Idiman; A Boyko; Ag Beiske; T Vollmer; N Triantafyllou; P O'Connor; Y Barak; L Biermann; E Cristiano; S Atweh; Dl Patrick; S Robitail; N Ammoury; A Beresniak; J Pelletier
Journal:  Mult Scler       Date:  2007-10-17       Impact factor: 6.312

9.  Long-term effect of early treatment with interferon beta-1b after a first clinical event suggestive of multiple sclerosis: 5-year active treatment extension of the phase 3 BENEFIT trial.

Authors:  Ludwig Kappos; Mark S Freedman; Chris H Polman; Gilles Edan; Hans-Peter Hartung; David H Miller; Xavier Montalbán; Frederik Barkhof; Ernst-Wilhelm Radü; Carola Metzig; Lars Bauer; Vivian Lanius; Rupert Sandbrink; Christoph Pohl
Journal:  Lancet Neurol       Date:  2009-09-10       Impact factor: 44.182

Review 10.  Cognitive impairment in multiple sclerosis.

Authors:  Nancy D Chiaravalloti; John DeLuca
Journal:  Lancet Neurol       Date:  2008-12       Impact factor: 44.182

View more
  3 in total

1.  Cognitive effects of onabotulinumtoxinA in chronic migraine.

Authors:  Susan Ho; David Darby; Natasha Bear
Journal:  BMJ Neurol Open       Date:  2020-01-09

Review 2.  Mobile health (mHealth) usage, barriers, and technological considerations in persons with multiple sclerosis: a literature review.

Authors:  Elizabeth S Gromisch; Aaron P Turner; Jodie K Haselkorn; Albert C Lo; Thomas Agresta
Journal:  JAMIA Open       Date:  2020-12-15

Review 3.  Practice effects in performance outcome measures in patients living with neurologic disorders - A systematic review.

Authors:  Sven P Holm; Arnaud M Wolfer; Grégoire H S Pointeau; Florian Lipsmeier; Michael Lindemann
Journal:  Heliyon       Date:  2022-08-17
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.