Literature DB >> 33392382

Validation of SATURN, a free, electronic, self-administered cognitive screening test.

David Bissig1, Jeffrey Kaye2, Deniz Erten-Lyons3.   

Abstract

BACKGROUND: Cognitive screening is limited by clinician time and variability in administration and scoring. We therefore developed Self-Administered Tasks Uncovering Risk of Neurodegeneration (SATURN), a free, public-domain, self-administered, and automatically scored cognitive screening test, and validated it on inexpensive (<$100) computer tablets.
METHODS: SATURN is a 30-point test including orientation, word recall, and math items adapted from the Saint Louis University Mental Status test, modified versions of the Stroop and Trails tasks, and other assessments of visuospatial function and memory. English-speaking neurology clinic patients and their partners 50 to 89 years of age were given SATURN, the Montreal Cognitive Assessment (MoCA), and a brief survey about test preferences. For patients recruited from dementia clinics (n = 23), clinical status was quantified with the Clinical Dementia Rating (CDR) scale. Care partners (n = 37) were assigned CDR = 0.
RESULTS: SATURN and MoCA scores were highly correlated (P < .00001; r = 0.90). CDR sum-of-boxes scores were well-correlated with both tests (P < .00001) (r = -0.83 and -0.86, respectively). Statistically, neither test was superior. Most participants (83%) reported that SATURN was easy to use, and most either preferred SATURN over the MoCA (47%) or had no preference (32%). DISCUSSION: Performance on SATURN-a fully self-administered and freely available (https://doi.org/10.5061/dryad.02v6wwpzr) cognitive screening test-is well-correlated with MoCA and CDR scores.
© 2020 The Authors. Alzheimer's & Dementia: Translational Research & Clinical Interventions published by Wiley Periodicals, Inc. on behalf of Alzheimer's Association.

Entities:  

Keywords:  Alzheimer's disease; cognitive screening; computer‐based test; dementia screening; psychometrics; self‐administered cognitive test

Year:  2020        PMID: 33392382      PMCID: PMC7771179          DOI: 10.1002/trc2.12116

Source DB:  PubMed          Journal:  Alzheimers Dement (N Y)        ISSN: 2352-8737


BACKGROUND

Dementia is prevalent and costly , but underdiagnosed: <20% of dementia cases are detected while in the “mild” stage. , Although there are dozens of cognitive screening instruments, most lack some essential features when compared to successful detection tools for other conditions (like sphygmomanometry for hypertension). These include low cost, high accuracy, relative freedom from language barriers, negligible time investment from the clinician, and potential for remote use, either via telemedicine or with assistance from non‐clinicians. Legacy tools like the Mini‐Mental State Examination (MMSE) and the Montreal Cognitive Assessment (MoCA) are illustrative. These once low‐cost tests have now been commercialized. , For the MoCA, commercialization via mandatory training was felt necessary to improve high variability in its administration and scoring: In one study of non‐clinicians, 5% of tests had administration errors and 32% had scoring errors, justifying the authors’ advice to retrain testers every 2 to 3 months. Furthermore, poor hearing and low vision are common in the older adults being screened. Reliance on both sensory modalities complicates clinical use of the MoCA and MMSE, and an attempt to adapt the Saint Louis University Mental Status (SLUMS) test to an electronic format was limited by hearing impairment. Although several legacy tests have been translated into other languages, some fluency is needed to administer the test. Finally, the 10 minutes spent administering one test can occupy an entire primary care visit, and discourages use in hospital settings where cognitive screening has prognostic value. , , Moving cognitive screening to mobile computers may mitigate the preceding issues, provided that the computer guides participants through self‐administered tasks and automatically scores results. Prior efforts—reviewed elsewhere , , and summarized in Table 1—fall short of clinical screening needs by requiring a trained test administrator, , lacking validation in dementia, , or being commercialized and restricted. , , , Responding to this gap in the literature, we developed and validated Self‐Administered Tasks Uncovering Risk of Neurodegeneration (SATURN)—a free, public domain, and automatically scored cognitive screening test.
TABLE 1

Summary of computerized self‐administered cognitive screening tests

Cognitive Domains AssessedStudy Diagnoses
Test NameMemoryOrientationVisuospatialExecutiveCalculationSustained Attention b Expressive LanguageTime (min) c Accommodates Hearing Loss d Participants (n) Age ≥ 50 yrMCIDementiaFree to Use e Reference
ACAD+    + 20++32    33
BCSD++ +  +15++101++  34
CAMCI+  + + 30+263+  35
CANS‐MCI+ ++  +>30 310   36
5097++ 37
CANTAB PAL+       +58++ 21
ClockMe  ++   <2+20    38
40   39
CNS Vital Signs+  + + 30++347++ 40
Cogstate Brief Battery+    + 20++1273   41
765++ 42
C‐TOC+ +++ +45+76++  25
CUPDE+++ +    30    10
CUPDE2+++ + ? ?21    43
DETECT+  + + 10+405++ 44
Dtmt  ++    +81+   45
eCTT  ++    ++21    43
eSAGE+++++ +18++66++ 20
GrayMatters+  +   20+251 + 46
IVR+    ++10 61   47
MicroCog/ACS+ ++++ >30 102 + 48
Revere+        153+  49

Abbreviations: ACAD, Automatic Cognitive Assessment Delivery; ACS, Assessment of Cognitive Skills; BCSD, Brief Computerized Self‐screen for Dementia; CADi, Cognitive Assessment for Dementia iPad version; CADi2, Cognitive Assessment for Dementia iPad version revised; CAMCI, Computerized Assessment of Mild Cognitive Impairment; CANS‐MCI, Computer‐Administered Neuropsychological Screen for Mild Cognitive Impairment; CANTAB PAL, Cambridge Neuropsychological Test Automated Battery Paired Associate Learning; C‐TOC, Cognitive Testing on Computer; CUPDE, Cambridge University Pen to Digital Equivalence; CUPDE2, Cambridge University Pen to Digital Equivalence revised; DETECT, Display Enhanced Testing for Cognitive impairment and Traumatic brain injury; dTMT, digital version of the Trails Making Task parts A and B; eCTT, electronic version of the Color Trails Test; eSAGE, electronic version of the Self‐Administered Gerocognitive Examination; IVR, Interactive Voice Response; MCS, Mobile Cognitive Screening; SATURN, Self‐Administered Tasks Uncovering Risk of Neurodegeneration; TPDAS, Touch Panel‐type Dementia Assessment Scale; TPST, Touch‐Panel Computer Assisted Screening Tool.

We omit studies that only tested those age <50 years, studies without an English‐language publication (eg, the CogVal‐Senior listed in a recent review ), and studies where a skilled operator participated in testing, as when training a participant (eg, task demonstration), , reinforcing instructions, or similar. , For brevity, only one or two publications are selected for each test. The total number of participants, and inclusion (or not) of patients with dementia and mild cognitive impairment (MCI), was extracted from those specific publications. The cumulative literature experience with some tests, like MicroCog, is more extensive than listed here.

Includes tasks that better challenge and report on participant attention than the “Simple Attention” items in SATURN. Examples include backward digit span and working memory tasks. ,

Where a range of times was provided, we tabled the larger value to reflect expectations for cognitively impaired participants, who are generally slower than age‐matched controls. The tabled value for SATURN is above the median time for participants with a clinical dementia rating scale global score of 1.0 (Table S1). Except for Revere (for which time is difficult to report due to a 20 minutes delay built into their test procedure), values are left blank if insufficient information was provided to judge test duration.

++ indicates that all testing was done without an audio device (speakers or headphones). + indicates that the study used some audio, but trivial protocol changes might make it audio‐free (eg, written instructions accompanied by a recording of those instructions read aloud, for redundancy). Outside of those categories, some tests with auditory stimuli nevertheless carry many audio‐free items, and truncated versions may be useful in the hearing impaired.

++ indicates public domain. – indicates that the test has been commercialized, including cases using automatic speech recognition technology , and cases where we are unsure if the company is still active.

Summary of computerized self‐administered cognitive screening tests Abbreviations: ACAD, Automatic Cognitive Assessment Delivery; ACS, Assessment of Cognitive Skills; BCSD, Brief Computerized Self‐screen for Dementia; CADi, Cognitive Assessment for Dementia iPad version; CADi2, Cognitive Assessment for Dementia iPad version revised; CAMCI, Computerized Assessment of Mild Cognitive Impairment; CANS‐MCI, Computer‐Administered Neuropsychological Screen for Mild Cognitive Impairment; CANTAB PAL, Cambridge Neuropsychological Test Automated Battery Paired Associate Learning; C‐TOC, Cognitive Testing on Computer; CUPDE, Cambridge University Pen to Digital Equivalence; CUPDE2, Cambridge University Pen to Digital Equivalence revised; DETECT, Display Enhanced Testing for Cognitive impairment and Traumatic brain injury; dTMT, digital version of the Trails Making Task parts A and B; eCTT, electronic version of the Color Trails Test; eSAGE, electronic version of the Self‐Administered Gerocognitive Examination; IVR, Interactive Voice Response; MCS, Mobile Cognitive Screening; SATURN, Self‐Administered Tasks Uncovering Risk of Neurodegeneration; TPDAS, Touch Panel‐type Dementia Assessment Scale; TPST, Touch‐Panel Computer Assisted Screening Tool. We omit studies that only tested those age <50 years, studies without an English‐language publication (eg, the CogVal‐Senior listed in a recent review ), and studies where a skilled operator participated in testing, as when training a participant (eg, task demonstration), , reinforcing instructions, or similar. , For brevity, only one or two publications are selected for each test. The total number of participants, and inclusion (or not) of patients with dementia and mild cognitive impairment (MCI), was extracted from those specific publications. The cumulative literature experience with some tests, like MicroCog, is more extensive than listed here. Includes tasks that better challenge and report on participant attention than the “Simple Attention” items in SATURN. Examples include backward digit span and working memory tasks. , Where a range of times was provided, we tabled the larger value to reflect expectations for cognitively impaired participants, who are generally slower than age‐matched controls. The tabled value for SATURN is above the median time for participants with a clinical dementia rating scale global score of 1.0 (Table S1). Except for Revere (for which time is difficult to report due to a 20 minutes delay built into their test procedure), values are left blank if insufficient information was provided to judge test duration. ++ indicates that all testing was done without an audio device (speakers or headphones). + indicates that the study used some audio, but trivial protocol changes might make it audio‐free (eg, written instructions accompanied by a recording of those instructions read aloud, for redundancy). Outside of those categories, some tests with auditory stimuli nevertheless carry many audio‐free items, and truncated versions may be useful in the hearing impaired. ++ indicates public domain. – indicates that the test has been commercialized, including cases using automatic speech recognition technology , and cases where we are unsure if the company is still active.

METHODS

Participants

This study was approved by the Oregon Health and Science University (OHSU) Institutional Review Board. We recruited patients and study partners from OHSU neurology clinics. Clinicians alerted study personnel when they had eligible (English‐speaking adults 50 to 89 years of age) and interested patients, generating a convenience sample of 42 dyads. We prioritized recruitment from the dementia and movement disorders clinics, seeking participants with cognitive or motor impairment that might complicate use of a computer tablet. After obtaining informed consent, both dyad members participated in all study procedures, which we completed at the end of a scheduled clinic visit. We logged the primary diagnosis from each clinic visit, and when the patient was seen in our dementia clinic, he or she was assigned Clinical Dementia Rating scale global (CDRglobal) and sum‐of‐boxes (CDRSOB) scores. As part of a typical clinic visit, information about the study partner's activities of daily living is provided by descriptions of how he or she assists the patient. No study partners reported any significant cognitive or functional impairment. Subsequent review of partner MoCA scores supported the expectation that this group was cognitively intact. They were therefore assigned CDR scores of 0.

HIGHLIGHTS

We validate SATURN (Self‐Administered Tasks Uncovering Risk of Neurodegeneration). SATURN is a free, self‐administered, automatically scored cognitive screening test. SATURN is highly correlated with the Montreal Cognitive Assessment (MoCA). SATURN and MoCA are equally associated with Clinical Dementia Rating (CDR) score. All SATURN materials are freely available (https://doi.org/10.5061/dryad.02v6wwpzr).

RESEARCH IN CONTEXT

Systematic review: The authors reviewed the literature with standard sources (e.g., PubMed). There is a need for public domain, self‐administered, and automatically scored cognitive screening tests, but we find no such examples that have been validated to screen for persons with dementia. Pertinent citations, including recent review articles on cognitive screening tests, are provided in the article. Interpretation: We develop and validate a cognitive screening test that satisfies the unmet need defined by the literature. It performs favorably when compared to the Montreal Cognitive Assessment and the Clinical Dementia Rating scale. Future directions: Because our test is in the public domain, it can be adapted or modified without restriction, which will ease deployment to clinical settings that would benefit from high‐quality cognitive screening. Additional validation in diverse samples (socioeconomic, ethnic/racial, linguistic) is warranted, and draft translations to Korean, Vietnamese, and simplified Chinese are provided in a data repository.

Study procedures

The patient and study partner were tested with SATURN and the MoCA (either version 7.1 or 7.2, which use different stimuli). For each dyad, we randomized which MoCA version was administered to the patient, and randomized the order of testing (MoCA before or after SATURN). Afterwards, participants completed a brief survey asking age, sex, years of education, and self‐identified race/ethnicity, and two questions comparing the MoCA to SATURN: “Was the tablet easy to use?” (“yes” or “no”), and “Which did you prefer, the paper‐and‐pencil test, or the tablet?” (“paper‐and‐pencil,” “tablet,” or “I felt the same about both”).

SATURN development and hardware

All code was written for VisionEgg in Python 2.5.4. All stimuli were either novel, or newly adapted from public domain cognitive tests. One author (DB) produced the code and stimuli as part of his Veterans Affairs Advanced Fellowship. Therefore, SATURN is fully in the public domain. Testing was performed on low‐cost (retail <$100) Ematic EWT935DK tablets running Windows 10. Stimuli appeared in the landscape orientation on the 19.7 × 11.5 cm (1024 × 600 pixel) display, with all text written in capitalized bold Arial font (each letter ≥4.5 mm tall, equivalent to ≥18 pt printed font). Through its development, SATURN was sequentially tested in three versions. As in prior work, we expected that some tasks would have limited utility when adapted for a computer tablet. We therefore included extra tasks in the first version of SATURN, with plans to remove or replace low‐utility tasks in later versions. Our experience with each version is detailed in the Results section. Scoring rules were devised for each version before testing began. As in prior work, each version was tested on n>9 participants to gauge usability. Expecting literature‐typical correlation coefficients (r > 0.80), , this provides adequate power (α = 0.05; 1‐β > 0.80) to detect a correlation between SATURN and MoCA at each development stage. Because almost all tasks in the final version were present in the earlier versions, we also aggregated data by retroactively applying the final version's scoring rules and time limit to all three versions.

Tasks included in SATURN's final version

Like the SLUMS and MoCA, SATURN includes high‐yield tests of orientation and delayed recall of intentionally encoded words. A test of incidental memory was included to ensure that any memory deficits are not merely due to poor encoding effort. To capture non‐amnestic cognitive impairment, SATURN also contains brief tests of calculation, executive function, and visuospatial function. Although SATURN does not test verbal fluency, language is assessed by estimating reading speed (time spent viewing each instruction screen divided by the number of words per screen). For all versions of SATURN, the initial tasks were the same: First, participants must read and act on the prompt “CLOSE YOUR EYES,” which was written in the smallest font used throughout, and meant to verify that vision and literacy were adequate for testing. To this end, the experimenter would offer generic encouragement and re‐prompt the participant multiple times as needed. Second, participants needed to eventually complete the three “simple attention” tasks described below. Even if a participant's effort and number of errors on those tasks exceeded expectations, it served as a second check that vision and literacy were adequate for testing. Per the study protocol, any participants unable to eventually complete these tasks were deemed “unscorable” and excluded from analyses (save for post hoc intent‐to‐screen analyses). For all tasks, a valid input is needed to proceed (a participant cannot skip a task, and is prompted to make a selection if this is attempted) but unless otherwise noted, the input can be incorrect. The final version of SATURN is composed of these assessments, in order of their appearance (for a video of normal operation, see https://doi.org/10.5061/dryad.02v6wwpzr): Simple attention is tested when the participant must (1) select the one word (of eight) that starts with “J”, (2) pick the two nouns (of four) that are fruit, and (3) copy the number 1239 with an on‐screen number pad. For each of these tasks—which also acclimate the participant to SATURN, and provide materials for later testing of incidental memory—any incorrect selection leads to a prompt to change one's selection. The participant cannot proceed without selecting the correct choice. For each of these three tasks, +2 points are awarded if the correct selection is made on the first try. Participants are then asked to remember five words—presented one at a time, for two seconds each—adapted from the free recall task on the SLUMS. Participants’ incidental memory is tested next by asking (1) which (of six) commands they read at the start of SATURN, (2) which word (of eight) they selected earlier, and (3) to re‐enter the prior four‐digit number. For each of these three tasks, +1 point is awarded for a correct answer. Participants’ orientation for the (1) month, (2) year, (3) day of the week, and (4) state are tested, earning +1 point for a correct answer in each of these four tasks. Except for the year, which is entered with an on‐screen number pad, orientation items are selected from a list of options. Next, memory is tested by showing a list of 100 nouns and asking the participant to select which were the five previously studied words, earning +1 point for each correct selection. Calculation is tested using updated versions of the SLUMS questions, where +1 point is awarded for entering the total spent on a $60 tricycle and $7 of apples, and +2 points are awarded for entering how much out of $100 remains after that purchase. Next, visuospatial function is tested with an adaptation of Pintner (1919)’s picture completion task : Participants are shown a large drawing (eg, a circle centered on the top corner of a pentagon) and asked to select which two of six smaller drawings compose the larger one (eg, the circle, and the pentagon). Four such large drawings are shown sequentially, and for each, +1 point is earned for selecting the correct pair of smaller drawings. Next, executive function is probed with 12 incongruent color‐word Stroop items. To reinforce task instructions, patients are prompted to input the correct answer after any error. Each error subtracts 1 point from a possible +3 points earned for perfect Stroop performance. Finally, participants earn +1 point each for error‐free completion of a mini‐Trails A (1 through 5) and mini‐Trails B (1 through D). The maximum total score is 30.

Statistics

Statistics were performed in R (https://www.r-project.org, v.3.6.1). We evaluated the association between MoCA and SATURN scores with linear regression. We compared both MoCA and SATURN scores to clinical status as rated by CDRglobal (one‐way analysis of variance [ANOVA]) and CDRSOB (linear regression), then tested whether MoCA or SATURN were superior predictors of CDRSOB (Davidson‐MacKinnon J‐test). We similarly tested for relationships between SATURN's reading speed estimate, CDR scores, and total scores on the SATURN and MoCA. Using R's pROC library, we compared receiver‐operating characteristic (ROC) curves for the sensitivity and specificity of MoCA and SATURN scores for cognitive impairment. Additional univariate comparisons (as with demographics) used linear regression, t‐tests, or Fisher exact test. Two‐tailed P < .05 was considered significant. Task‐specific findings are detailed in the Appendix.

RESULTS

We refined SATURN until its third and final version. Here we present this development process, followed by analyses of aggregated data using the final version's scoring rules.

SATURN development, Version 1

Of seven dyads (n = 14), one study partner fell outside of our pre‐defined age range, and two patients were “unscorable” (see Section 3.5). Of the remaining participants (n = 11), the initial scoring plan for SATURN yielded results strongly correlated with the MoCA (P < .00001; r = 0.95). A few problems were noted. (1) Even though the six study partners appeared cognitively normal (mean ± SD MoCA scores of 27.2 ± 1.5) only one achieved better than 33% correct on an unscored spatial memory task. Due to floor effects, it was removed from subsequent versions. (2) We took note of small aesthetic choices that would improve usability, and implemented these in subsequent phases. For instance, some elements of the Stroop task instructions were mistaken for input buttons. For all subsequent versions, we therefore simplified instructions and added three unscored “warm up” trials using non‐color words (eg, “BOOK” written in red) to verify task comprehension. (3) This version of SATURN shared the clock drawing task with the MoCA, but those scores were not well‐correlated (P = .052, r = 0.60): nine (82%; including all six study partners) drew a normal clock on the MoCA, compared to three (27%) on the tablet. In subsequent versions of SATURN, the clock drawing was replaced by the four‐item adaptation of Pintner (1919)’s picture completion task. For later analyses (and presentation in Figure 1) we re‐calculated Version 1 scores to approximate the final version: Lacking the picture completion task, we awarded 4 points for the clock (with 2 points for hand position), but otherwise matched the tasks and scoring rules of the final version (including a retroactively applied time limit (see Section 3.2)). Scoring changes had no effect on the relationship between total SATURN and MoCA scores (P = .00004; r = 0.93 vs the original r = 0.95).
FIGURE 1

Relationship between scores on the Montreal Cognitive Assessment (MoCA) and Self‐Administered Tasks Uncovering Risk of Neurodegeneration (SATURN), and their association with clinical status, quantified by the Clinical Dementia Rating (CDR) scale. (A) SATURN and MoCA scores are strongly correlated. Scorable participant data from all three versions of SATURN (n = 75) are shown as translucent circles, so that overlapping points appear darker. The ×s denote data from “unscorable” participants and are staggered slightly where they would otherwise overlap. (These conventions are carried through the plots in B, although, by design, the beeswarm plots for CDRglobal have no overlapping points.) The best‐fit line for the correlation among scorable participants is pictured (solid line; MoCA = 0.786 × SATURN + 5.663; r = 0.90; P < .00001) along with its 95% confidence interval (dashed lines). The 50% prediction interval (not shown) is roughly 1.5 points above and below the regression line throughout. Thus at least half of those with a SATURN score of 28 will score between a 26 and 29 on the MoCA. (B) The relationship between CDRglobal scores and SATURN scores (top) and MoCA scores (bottom) are detailed by beeswarm plots overlaid on box‐and‐whisker plots. Both SATURN and MoCA were strongly associated with CDRglobal score (analysis of variance [ANOVA]; F[3,56] = 46.8 and F[3,56] = 75.8 respectively, both P < .00001). Harmonizing this plot with Table 1, we note that those 10 participants with a CDRglobal = 0.5 had clinical diagnoses of either mild cognitive impairment (n = 3) or mild dementia (n = 7) per American Academy of Neurology guidelines. To the right of the figure, those with nonzero CDRglobal scores are re‐plotted according to their specific CDRSOB scores. Since CDRSOB = 0 for all those with CDRglobal = 0.0, their data are not re‐plotted into the CDRSOB plot. The relationship between CDRSOB and both SATURN and MoCA was robust (linear regression; r = −0.83 and r = −0.86 respectively; P < .00001). (C) Receiver‐operating characteristic (ROC) curves detail the ability of both MoCA (gray) and SATURN (black) to detect cognitive impairment. On the left, cognitive impairment is defined as CDRglobal >0. Based on these data, it is optimal to label one cognitively impaired if one's score is <24 on SATURN (sensitivity 82%, specificity 92%) or <26 on the MoCA (sensitivity 91%, specificity 82%). Area under the curve (AUC) was similar for each test (P > .19; 0.95 for MoCA [95% CI 0.89 to 1.0], versus 0.90 for SATURN [95% CI 0.82 to 0.95]). For illustrative purposes, points are overlaid on the ROC curves at cutoffs of <22, <24, <26, and <28 for both tests. On the right, cognitive impairment is defined as CDRglobal >0.5. Based on these data, it is optimal to label one cognitively impaired if one's score is <21 on SATURN (sensitivity 92%, specificity 88%) or <23 on the MoCA (sensitivity 92%, specificity 88%). AUC was similar for each test (P = .8; 0.94 for MoCA [95% CI 0.85 to 1.0], versus 0.95 for SATURN [95% CI 0.87 to 1.0]). Points are overlaid on the ROC curves cutoffs of <19, <21, <23, and <25 for both tests

Relationship between scores on the Montreal Cognitive Assessment (MoCA) and Self‐Administered Tasks Uncovering Risk of Neurodegeneration (SATURN), and their association with clinical status, quantified by the Clinical Dementia Rating (CDR) scale. (A) SATURN and MoCA scores are strongly correlated. Scorable participant data from all three versions of SATURN (n = 75) are shown as translucent circles, so that overlapping points appear darker. The ×s denote data from “unscorable” participants and are staggered slightly where they would otherwise overlap. (These conventions are carried through the plots in B, although, by design, the beeswarm plots for CDRglobal have no overlapping points.) The best‐fit line for the correlation among scorable participants is pictured (solid line; MoCA = 0.786 × SATURN + 5.663; r = 0.90; P < .00001) along with its 95% confidence interval (dashed lines). The 50% prediction interval (not shown) is roughly 1.5 points above and below the regression line throughout. Thus at least half of those with a SATURN score of 28 will score between a 26 and 29 on the MoCA. (B) The relationship between CDRglobal scores and SATURN scores (top) and MoCA scores (bottom) are detailed by beeswarm plots overlaid on box‐and‐whisker plots. Both SATURN and MoCA were strongly associated with CDRglobal score (analysis of variance [ANOVA]; F[3,56] = 46.8 and F[3,56] = 75.8 respectively, both P < .00001). Harmonizing this plot with Table 1, we note that those 10 participants with a CDRglobal = 0.5 had clinical diagnoses of either mild cognitive impairment (n = 3) or mild dementia (n = 7) per American Academy of Neurology guidelines. To the right of the figure, those with nonzero CDRglobal scores are re‐plotted according to their specific CDRSOB scores. Since CDRSOB = 0 for all those with CDRglobal = 0.0, their data are not re‐plotted into the CDRSOB plot. The relationship between CDRSOB and both SATURN and MoCA was robust (linear regression; r = −0.83 and r = −0.86 respectively; P < .00001). (C) Receiver‐operating characteristic (ROC) curves detail the ability of both MoCA (gray) and SATURN (black) to detect cognitive impairment. On the left, cognitive impairment is defined as CDRglobal >0. Based on these data, it is optimal to label one cognitively impaired if one's score is <24 on SATURN (sensitivity 82%, specificity 92%) or <26 on the MoCA (sensitivity 91%, specificity 82%). Area under the curve (AUC) was similar for each test (P > .19; 0.95 for MoCA [95% CI 0.89 to 1.0], versus 0.90 for SATURN [95% CI 0.82 to 0.95]). For illustrative purposes, points are overlaid on the ROC curves at cutoffs of <22, <24, <26, and <28 for both tests. On the right, cognitive impairment is defined as CDRglobal >0.5. Based on these data, it is optimal to label one cognitively impaired if one's score is <21 on SATURN (sensitivity 92%, specificity 88%) or <23 on the MoCA (sensitivity 92%, specificity 88%). AUC was similar for each test (P = .8; 0.94 for MoCA [95% CI 0.85 to 1.0], versus 0.95 for SATURN [95% CI 0.87 to 1.0]). Points are overlaid on the ROC curves cutoffs of <19, <21, <23, and <25 for both tests

SATURN development, Version 2

Of 13 dyads, two patients were “unscorable.” Of the remaining participants (n = 24), the initial scoring plan for SATURN yielded scores strongly correlated with those from the MoCA (P < .00001; r = 0.84). Before moving on to the final development, we addressed the following issues. (1) We had been piloting a full version of the Trials B task in Versions 1 and 2. For only this task, we allowed incomplete responses, which were more common in those with cognitive impairment (67%) than in their study partners (16%) (P = .007). We therefore opted to remove the full Trails B task from the final version of SATURN. (2) Spatial tasks borrowed from the SLUMS—drawing an “x” in a triangle and picking the largest of three shapes—showed ceiling effects, and were removed from the final version of SATURN. (3) Especially in patients with cognitive impairment, the time spent on SATURN was felt to be excessive (for Version 2, mean ± SD of 17.9 ± 8.4 min, vs 10.8 ± 4.4 in study partners; P = .041). We therefore built a time limit into SATURN: The program ends when one finishes whichever tasks were started within 15 minutes. Zero points are awarded for never‐started tasks, and the total is still scored out of 30 points. Recalculating Version 2 scores with the final task set and time limit (as presented in Figure 1) had no effect on the correlation between SATURN and MoCA scores (P < .00001; r = 0.85 vs the original r = 0.84).

SATURN, Version 3

Of 22 dyads, four study partners fell outside of our pre‐defined age range. All remaining participants (n = 40) were scorable. SATURN and MoCA scores were well correlated (P < .00001; r = 0.93). CDRglobal score and CDRSOB were strongly associated with MoCA (respectively, F[3,25] = 38.4, and r = −0.89) and SATURN scores (F[3,25] = 21.0; r = −0.92)(for all, P < .00001).

Scorable participants’ results aggregated from all SATURN versions

Summary demographics, diagnoses, survey results, MoCA scores, and SATURN scores from the aforementioned n = 75 scorable participants are provided in Table 2.
TABLE 2

Demographics, mean performance, and survey results of “scorable” participants

AgeSexEducationTest preference (%)
Participant sourcen(years)(%women)(years)MoCASATURN%Report SATURN “Easy to Use”MoCANo Pref.SATURN
Study partner3765.5 ± 9.25916.3 ± 2.527.0 ± 1.827.1 ± 2.69783062
Dementia clinic2471.5 ± 10.34615.5 ± 3.019.1 ± 6.117.5 ± 6.463423821
 Not Demented a 474.3 ± 6.15015.8 ± 1.624.8 ± 3.824.3 ± 4.17525750
 Alzheimer's b 1368.5 ± 11.34616.4 ± 2.618.5 ± 5.616.4 ± 6.054543115
Other Dementia c 775.6 ± 9.22913.7 ± 3.816.9 ± 6.615.7 ± 6.271292943
Other Clinic1465.9 ± 7.63615.5 ± 3.624.8 ± 3.323.6 ± 3.679212950
 Multiple Sclerosis262.5 ± 7.810015.0 ± 1.428.0 ± 1.426.5 ± 2.11005050
 Essential Tremor166100142725100100
 Parkinson1166.5 ± 8.21815.7 ± 4.024.0 ± 3.323.0 ± 3.773183645

Shaded rows break down patients from each clinic into more specific diagnoses. Other than group sizes (n), data are presented as percentages or as mean ± standard deviation. The MoCA and SATURN are scored out of 30 points.

Includes three patients with mild cognitive impairment, and a cognitively normal carrier of an apolipoprotein E (APOE) ε4 allele

The clinical diagnosis was Alzheimer's disease for all but one, for whom cerebrospinal fluid implicated Alzheimer's disease as the cause of corticobasal syndrome

Includes one case of Lewy body dementia, three cases predominantly or exclusively caused by cerebrovascular disease, and another three of mixed or uncertain cause. To simplify presentation, this row includes a patient with known dementia who was seen in general neurology clinic for peripheral neuropathy on the day of testing (and therefore not assigned CDR scores).

Demographics, mean performance, and survey results of “scorable” participants Shaded rows break down patients from each clinic into more specific diagnoses. Other than group sizes (n), data are presented as percentages or as mean ± standard deviation. The MoCA and SATURN are scored out of 30 points. Includes three patients with mild cognitive impairment, and a cognitively normal carrier of an apolipoprotein E (APOE) ε4 allele The clinical diagnosis was Alzheimer's disease for all but one, for whom cerebrospinal fluid implicated Alzheimer's disease as the cause of corticobasal syndrome Includes one case of Lewy body dementia, three cases predominantly or exclusively caused by cerebrovascular disease, and another three of mixed or uncertain cause. To simplify presentation, this row includes a patient with known dementia who was seen in general neurology clinic for peripheral neuropathy on the day of testing (and therefore not assigned CDR scores).

Demographics

Compared to dementia clinic patients, study partners were similar in terms of sex (P = .4), and years of education (P = .4), but were slightly younger (P = .039). Race and ethnicity were not analyzed, as 93% of our participants self‐identified as non‐Hispanic Caucasian.

Survey results

Overall, 47% of participants preferred SATURN over the MoCA, and another 32% had no preference (Table 2). Compared to dementia clinic patients, study partners were more likely to favor SATURN over MoCA, and report that SATURN was easy to use (P < .0005).

Test scores in relation to demographics

MoCA scores among study partners were unrelated to age and sex (P = .2) but were lower in less‐educated partners (P = .027). SATURN scores were unrelated to sex (P = .9) and years of education (P = .4), but were lower in older partners (P = .019)

Test scores in relation to one‐another and CDR

MoCA and SATURN scores were well correlated (P < .00001; r = 0.90). This remained true when the comparison was restricted to dementia clinic patients (n = 24 detailed in Table 2; P < .00001; r = 0.84), patients from other clinics (n = 14; P < .00001; r = 0.95; including those with Parkinson disease (n = 11; P = .00001; r = 0.94)), and study partners (n = 37; P = .00009; r = 0.60). CDRglobal score and CDRSOB were strongly associated with MoCA (respectively, F[3,56] = 75.8, and r = ‐0.86) and SATURN (F[3,56] = 46.8; r = −0.83)(all P < .00001). Some variance in the relationship between MoCA and CDRSOB was further explained by SATURN score (P = .048) and vice versa (P = .00021) (J‐tests), which indicates that SATURN and MoCA tests are complementary, without clear evidence that one is superior to the other. Multivariate analyses comparing CDRSOB to either MoCA or SATURN and including demographic variables (any combination of age, education, and sex) as covariates revealed negligible impact on the relationship between CDRSOB and test scores, and showed no relationship between demographics and CDRSOB. As detailed in Figure 1 and its legend, for those assigned CDR scores (n = 60), MoCA and SATURN had similar ROC curves for their ability to distinguish cognitively normal participants (CDRglobal = 0) from those with any impairment (CDRglobal > 0), and for their ability to distinguish those without dementia (CDRglobal ≤ 0.5) from those with dementia (CDRglobal > 0.5).

Reading speed

We compared participants’ CDR scores (n = 60) to reading speed estimates (in seconds per word). Those with cognitive impairment were slower to advance through instruction screens (r = 0.67 for correlation with CDRSOB; F[3,56] = 24.68 for ANOVA comparison with CDRglobal; both P < .00001)(see Table A.1). Multivariate analyses demonstrated that reading speed estimates complemented the 30‐point test scores, explaining some additional variance in the relationship between CDRSOB and SATURN (P = .0062) and between CDRSOB and MoCA scores (P = .014) (J‐tests). (See Appendix for additional secondary analyses of reading speed and diagnostic categories.)

“Unscorable” participants

Across all SATURN versions, three patients were unable to read the prompt to close their eyes (all with CDRglobal ≥ 2.0, and MoCA ≤ 2). We initiated the program for all, since none had an ophthalmologic or educational history that would limit reading, but none earned points. The fourth unscorable patient (CDRglobal = 2.0, and MoCA = 14) earned points for the first simple attention task, but erred when asked to select two fruit words from a list of four, and then seemed to confuse selection versus de‐selection of responses, preventing her further advance. A post hoc intention‐to‐screen analyses including these four participants (age 70.8 ± 7.0 years, 15.8 ± 2.8 years education, half men, all non‐Hispanic Caucasian) strengthened results: MoCA and SATURN scores remained well‐correlated (P < .00001; r = 0.93). CDRglobal score remained strongly associated with MoCA (P < .00001; F[4,59] = 90.1) and SATURN (P < .00001; F[4,59] = 63.7), as did CDRSOB (r = ‐0.92 for MoCA, r = −0.89 for SATURN; both P < .00001) for which both SATURN and MoCA were complementary (J‐tests both P < .04). When “unscorable” patients were included in the aforementioned ROC analyses, area under the curve, sensitivity, and specificity were stable to improved.

DISCUSSION

We developed and validated an electronic cognitive screening task with several desirable features. SATURN is fully self‐administered and automatically scored, sparing clinician time and removing opportunities for error. Compared to prior efforts, the scope of cognitive domains tested by SATURN is favorable, and its existence in the public domain is unique (Table 1). It is strongly correlated with the previously validated MoCA, and we found no evidence that SATURN was inferior to MoCA at sorting patients by overall clinical status, as characterized with the CDR. Given the high prevalence of hearing impairment in older adults, and based on prior work, we strategically avoided auditory stimuli in SATURN. In turn, we expect SATURN will be of limited use in those with poor vision or low baseline literacy. In regular clinical practice, when a patient is unable to read the initial stimulus, an alternative and appropriately normed test should be selected. In familiar patients with good vision and high baseline literacy—as with our four “unscorable” participants—difficulty with the first stimulus may instead reflect significant cognitive impairment, and equally should trigger additional testing. The exclusive use of visually presented and automatically scored tasks may make it easier to mitigate other communication barriers: We believe that SATURN will be relatively simple to translate to other languages, and to facilitate development, we have included Korean, Vietnamese, and simplified Chinese drafts in a repository (https://doi.org/10.5061/dryad.02v6wwpzr). Language‐specific versions of some tasks included in SATURN (eg, Stroop and Trails ) have already been validated, which may further shorten development times. One weakness of this study is the lack of demographic diversity in our sample, which was disproportionately well‐educated, Caucasian, and recruited from a regional referral center. Future population‐based testing in wider age ranges, wider ranges of educational achievement, and greater inclusion of racial/ethnic minorities is needed to reassure against biases. Although lacking demographic diversity, our participants were clinically diverse. We initially wondered whether patients with movement disorders or multiple sclerosis would find SATURN harder to use than the MoCA. Reassuringly, test scores remained well correlated in these patients, who tended to prefer SATURN over the MoCA. The total number of participants in this validation study is typical of prior efforts, , and it was not powered to probe for specific cognitive patterns associated with each clinical group. Like most computerized tests, it is therefore not yet known if SATURN can go beyond a screening instrument, and distinguish between different types of dementia. Protocols exploring this possibility could employ additional test batteries with good discriminative power, may leverage performance times recorded by SATURN (Table S1), and will benefit from greater enrollment than in the present study. Overall, SATURN provides high‐quality cognitive screening without clinician input, and is suitable for a wide variety of health care settings, including community efforts relying on non‐clinicians. Although the present data were collected in‐person with inexpensive tablets, versions of SATURN appropriate for fully remote use—as through an internet browser—are foreseeable. Integration with electronic health records would be a worthwhile: For instance, the cognitive screening portion of a Medicare Annual Wellness Visit could be completed for most patients while sitting in the waiting room. It is doubtful that a single group or company could adapt testing to every useful setting in perpetuity. We therefore make all SATURN materials freely available for download (https://doi.org/10.5061/dryad.02v6wwpzr), and encourage readers to use, share, and adapt SATURN without restriction.

CONFICTS OF INTEREST

Deniz Erten‐Lyons, within the past year, served on the Independent Adjudication Committee for the HARMONY study by Acadia Pharmaceuticals. David Bissig and Jeffrey Kaye have no conflicts of interest. Supporting Information Click here for additional data file. Supporting Information Click here for additional data file.
  51 in total

1.  Getting closer? Differences remain in neuropsychological assessments converted to mobile devices.

Authors:  Áine Maguire; Jennifer Martin; Hannes Jarke; Kai Ruggeri
Journal:  Psychol Serv       Date:  2018-11-08

2.  Normative study of the Stroop Color and Word Test in an educationally diverse elderly population.

Authors:  Eun Hyun Seo; Dong Young Lee; Il Han Choo; Shin Gyeom Kim; Ki Woong Kim; Jong Choul Youn; Jin Hyeong Jhoo; Jong Inn Woo
Journal:  Int J Geriatr Psychiatry       Date:  2008-10       Impact factor: 3.485

3.  Importance of functional measures in predicting mortality among older hospitalized patients.

Authors:  S K Inouye; P N Peduzzi; J T Robison; J S Hughes; R I Horwitz; J Concato
Journal:  JAMA       Date:  1998-04-15       Impact factor: 56.272

4.  The worldwide economic impact of dementia 2010.

Authors:  Anders Wimo; Linus Jönsson; John Bond; Martin Prince; Bengt Winblad
Journal:  Alzheimers Dement       Date:  2013-01       Impact factor: 21.566

5.  C-TOC (Cognitive Testing on Computer): investigating the usability and validity of a novel self-administered cognitive assessment tool in aging and early dementia.

Authors:  Claudia Jacova; Joanna McGrenere; Hyunsoo S Lee; William W Wang; Sarah Le Huray; Emily F Corenblith; Matthew Brehmer; Charlotte Tang; Sherri Hayden; B Lynn Beattie; Ging-Yuek R Hsiung
Journal:  Alzheimer Dis Assoc Disord       Date:  2015 Jul-Sep       Impact factor: 2.703

6.  The FAB: a Frontal Assessment Battery at bedside.

Authors:  B Dubois; A Slachevsky; I Litvan; B Pillon
Journal:  Neurology       Date:  2000-12-12       Impact factor: 9.910

7.  Multicomponent analysis of a digital Trail Making Test.

Authors:  Robert P Fellows; Jessamyn Dahmen; Diane Cook; Maureen Schmitter-Edgecombe
Journal:  Clin Neuropsychol       Date:  2016-10-03       Impact factor: 3.535

8.  Revision of the Cognitive Assessment for Dementia, iPad version (CADi2).

Authors:  Keiichi Onoda; Shuhei Yamaguchi
Journal:  PLoS One       Date:  2014-10-13       Impact factor: 3.240

Review 9.  Mobile Technology for Cognitive Assessment of Older Adults: A Scoping Review.

Authors:  Bon Mi Koo; Lisa M Vizer
Journal:  Innov Aging       Date:  2019-01-01

Review 10.  Brief cognitive screening instruments for early detection of Alzheimer's disease: a systematic review.

Authors:  Ellen Elisa De Roeck; Peter Paul De Deyn; Eva Dierckx; Sebastiaan Engelborghs
Journal:  Alzheimers Res Ther       Date:  2019-02-28       Impact factor: 6.982

View more
  3 in total

1.  Digital Phenotyping in Clinical Neurology.

Authors:  Anoopum S Gupta
Journal:  Semin Neurol       Date:  2022-01-11       Impact factor: 3.212

2.  Self-administered Web-Based Tests of Executive Functioning and Perceptual Speed: Measurement Development Study With a Large Probability-Based Survey Panel.

Authors:  Ying Liu; Stefan Schneider; Bart Orriens; Erik Meijer; Jill E Darling; Tania Gutsche; Margaret Gatz
Journal:  J Med Internet Res       Date:  2022-05-09       Impact factor: 7.076

Review 3.  Digital Cognitive Biomarker for Mild Cognitive Impairments and Dementia: A Systematic Review.

Authors:  Zihan Ding; Tsz-Lok Lee; Agnes S Chan
Journal:  J Clin Med       Date:  2022-07-19       Impact factor: 4.964

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.