Literature DB >> 28690635

The Development of the Dementia Screening Battery-100: Instrument Presentation, Reliability, and Construct Validity.

Tarek Bellaj1, Sonia Ben Jemaa2, Maher Khelifa1, Mona Ben Djebara3, Riadh Gouider3, Didier Le Gall4.   

Abstract

BACKGROUND/AIM: The screening of dementia in non-Westerners has so far relied on translations and adaptations of reputed instruments. Other efforts focused on developing culturally appropriate tests or tests in touch with new developments in the field. This study presents the rationale behind the construction of a new dementia screening test: the Dementia Screening Battery-100 (DSB-100).
METHODS: The DSB-100 was administered to 46 demented individuals and 159 healthy matched controls. All demented participants met the DSM-IV criteria for dementia. The healthy controls showed no cognitive impairment and were independent in activities of daily living. The DSB-100 was administered as part of a larger neuropsychological assessment to collect additional indices on the severity of patients' dementia, depression, and frontal dysfunctions. The same information was used for comparisons with DSB-100 scores.
RESULTS: Multiple regression analysis suggested that age and education, but not the variable sex, are essential in predicting cognitive performance. Construct validation yielded 4 factors, namely attention-visuospatial factors, memory, language, and executive functions. The results showed that the DSB-100 has a high interrater reliability and an acceptable overall internal homogeneity.
CONCLUSION: These results validate the DSB-100 and suggest its appropriateness for dementia screening in Tunisian elderly and possibly elderly people from other cultures with modifications to some subscales.

Entities:  

Keywords:  Developing countries; Elderly population; Neurodegenerative diseases; Neuropsychological assessment; Non-Western cultures; Screening test; Tunisia; Validation

Year:  2017        PMID: 28690635      PMCID: PMC5498961          DOI: 10.1159/000477437

Source DB:  PubMed          Journal:  Dement Geriatr Cogn Dis Extra        ISSN: 1664-5464


Introduction

Among the neurological syndromes, dementia is one of the most complex due to diagnostic difficulties, especially in its early stages. Dementia is accompanied by cognitive deficits affecting different functions. Examining these functions through neuropsychological testing is necessary to document deficits directly that define dementia itself. With the emergence of the diagnostic entity “mild cognitive impairment” (MCI) with single or multiple domain impairment, this assessment becomes all the more useful in the absence of other efficient markers to distinguish these deficits or their progression [1]. For this purpose, it is important to distinguish between formal neuropsychological assessment, which is quite long and rich, and screening tools, which quickly yield the different levels of impairment. Other than a short assessment time, screening tools must be simple to score, easily acceptable to patients, and should not require special or heavy equipment. They also need to be objective, quantitative, balanced between sensitivity and specificity, and have to satisfy classical psychometric qualities such as validity and reliability. Most importantly, they must be normalized and culturally appropriate for the population they assess. Hence, there is a real need for new screening tools that take into consideration and deal with all existing methodological limitations. It is especially important to address the many conceptual issues related to the universality of cognitive systems [2], the relationship between culture and cognition [3], the importance of demographic specificities of target populations [4], and cultural bias in psychological assessments [5]. Therefore, when we had to choose between translating, adapting, or developing a new instrument to assess dementia in Tunisia, we opted for the last to address 2 main concerns in existing measures in addition to the well-documented cultural and methodological issues [6, 7]. The first relates to methodological, conceptual, and ethical issues with available procedures and tests. The second relates to keeping up with rapidly changing knowledge in the field of dementia and its neuropsychological assessment. As far as the first issue is concerned, typically any translation or adaptation includes serious problems of equivalence. The final product of any translation or adaptation inevitably includes equivalence bias and rarely equates with the original theoretical construct, administration method, and item content [8]. For instance, even when we are attentive to the procedures of translation and back-translation, the equivalence of tasks is always far from guaranteed. But avoiding verbal tests as a way to deal with bias is also not a solution, given that a large part of human cognitive activities are mediated by language and because non-verbal tests do not necessarily mean “culture-free tests” [9]. For translated and adapted tests, choosing items and pictures from local normalized databases – similar to Battig et al.'s category word frequency data [10] and Snodgrass and Vanderwart's norms for name agreement, familiarity, and visual complexity for pictures [11] – is not always appropriate, since scores that are “numerically identical can have a psychologically different meaning across cultures” [12]. This is especially true for adaptation procedures where inferences drawn from the adapted version scores can prove problematic as far as the psychological constructs are concerned [13]. This establishes the need to re-examine the construct validity of any adapted test, even if the items are based on standardized local databases. Regarding the second concern, we noted a growing dissatisfaction with the increasing knowledge gap between the rapidly changing knowledge in the field of dementia and the existing dementia screening tools. Indeed, the most widely used instruments for the detection of dementia suffer from the following major limitations that our proposed instrument seeks to address: Most existing dementia screening instruments assess orientation in time and space, a complex function not altered in the early stages of dementia. This affects the discriminatory power of the measures, as they allot many points to such non-sensitive tasks for the early detection of dementia. This is the case for the Mini-Mental State Examination (MMSE) [14], the Montreal Cognitive Assessment (MoCA) [15], the Mattis Dementia Rating Scale (Mattis) [16], the Short Portable Mental Status Questionnaire [17], the modified MMSE (3MS) [18], the Addenbrooke's Cognitive Examination Revised [19], the Neurobehavioral Cognitive Status Examination (NCSE) [20], and the CERAD Neuropsychological Instrument [21]. Several dementia screening instruments assess memory inappropriately using infra-span tasks (less than 6–7 units), even though long-term episodic memory deficit is a major criterion in the diagnosis of dementia according to the DSM-IV [22]. This is the case for the MMSE, 3MS, Mini-Cog, [23], MoCA, and for the Rowland Universal Dementia Assessment Scale (RUDAS) [24]. For example, registration and recall in the MMSE can be considered weak measures of memory, as each of the tasks used is far less than the span. In addition, registration is a measure of short-term memory that is preserved even in moderate dementia [25]. Recall is underscored in many of these tests. In the MMSE, it represents only 3 points (10% of the total score), this is quantitatively a low weighting for such an important consideration compared to the presumed importance of memory impairment in the early stages of dementia. A good number of existing dementia screening instruments use reading (MMSE, Mattis, Addenbrooke's Cognitive Examination Revised), writing (MMSE, 3MS, and Mattis), and/or drawing tasks (MMSE, 3MS, Mini-Cog, MoCA, Mattis, RUDAS, NCSE, and CERAD Neuropsychological Instrument) that are irrelevant to illiterate subjects who represent the majority of the elderly population especially in Africa, Asia, and Latin America [see UNESCO Institute for Statistics, 26]. For instance, how do you interpret the poor graphic reproduction of a person who has never use a pen in their life? To include reading, writing, and drawing task scores in the assessment of dementia may produce false positives by inappropriately inflating the total scores. Strangely enough, dementia screening instruments neglect the assessment of executive functions (MMSE, Mini-Cog, RUDAS, Short Portable Mental Status Questionnaire, and the NCSE) that are essential for the diagnosis of frontotemporal dementia and subcortical dementia among others. These tests lag behind the current developments in our knowledge base of dementia. The aforementioned limitations of the existing dementia screening tests argue for the development of a new screening tool to address these issues. We consider the instrument we are proposing, the Dementia Screening Battery-100 (DSB-100), as more effective in the detection of dementia in Tunisian elderly since it addresses and corrects all of the above-mentioned issues, and its items and administration format are derived from the local cultural background. Indeed, cultural appropriateness provides more reliability to the inferences drawn from the test scores and the interpretations of differences in the patterns of the scores. In this study, we present the DSB-100 and the rationale behind its development. This instrument is an analytic battery, composed of 10 subtests, developed and derived from established neuropsychological procedures, targeting memory, executive functions, praxia, language, and attentional and visuospatial deficits. The developed instrument attempts to screen dementia and type cognitive dysfunctions among the elderly Tunisian population who is mostly illiterate. Additionally, other goals of the current study were also to verify the impact of sociodemographic factors (sex, age, and education) on the DSB-100 scores and to examine its construct validity and its reliability.

Method

The DSB-100

The DSB-100 is a test battery consisting of 10 subtests, which assess memory, executive functions, praxia, language, and attention and visuospatial functions. Its administration requires, on average, 15–20 min. Of note, the DSB-100 uses tasks that are classically known to be closely associated with attentional processes namely forward and backward span, even though the latter also engages working memory processes [27]. For these tasks, we opted to use names for forward and backward span. I.e., we present series of names, saying that we are reading male or female names as if they represent brothers or sisters from the same family. The task is to repeat the names immediately from the oldest to the youngest (Forward Names Span) – respecting the order of presentation – or from the youngest to the oldest (Backward Names Span). The use of names instead of digits removes the resistance found in illiterate individuals. Indeed, illiterates often complain that they cannot complete the forward or backward span task because they are not well-grounded in maths, even though the task does not call for it. The names span task appears more ecological, as it is similar to real world activities in which we say or repeat siblings' names in ascending or descending order. We assessed language using 2 subtests: the comprehension of simple to complex orders and the naming of different parts of the body. These activities are important to assess receptive and expressive language processing. For the comprehension task, we selected, from the Token Test [28], 10 items of progressive difficulty. The most simple orders start with a verb, an adjective, and a noun. On the next higher level, we add “size” to the instructions. Then, the subjects must take into account the form and the color of 2 items. Next, we ask for the size, the color, and the form of 2 items. For the most complex level of the task, instructions include various suffixes, conjunctions, and adverbs. For the DSB-100 naming subtest, we know that all cultures and languages have specific names for parts of the human body [29]. In the naming task, we ask the subjects to name 10 parts of the body pointed at or touched by the tester (ears, teeth, nails, neck, shoulder, forehead, eyebrow, chin, elbow, and the index finger). Naming these body parts provides subjects with a range of difficulties – going from the eyes, nose, etc. to the eyebrows, nail, etc. [30]. On top, this task does not require any additional material (e.g., images and objects). We assessed the subjects' visuospatial capacities through a modified version of the Line-Object Orientation Judgment Test [31]. The test consists of 10 items with different levels of difficulty. It is a motor-free measure of the ability to estimate angular relationships between line segments by visually matching angled line pairs. The stimuli appear on the upper half of the page, and at the bottom of the page, there is an array of 11 lines, arranged in a semi-circle and pointing to a picture of an object. For each item, the participants are randomly provided with a set of two lines, and the task is to identify the object each line leads to. We examined memory by presenting the participants an A4 page containing pictures of 10 objects, selected from the Snodgrass and Vanderwart database [11]. These pictures represent a mix of semantic living (animals, fruits, and vegetables) and non-living categories (furniture, clothing, school furniture, toys, tools, kitchen utensils, and transportation). The items are neither prototypic nor rare. They are ordered from 6 to 20 in each corresponding category frequency table. The items are also chosen considering naming agreement, visual complexity, and familiarity. After presenting the picture, we invite the individuals to name and memorize each picture. This procedure ensures a deeper semantic encoding strategy and activates the explicit memory system. After an interfering activity (digit span), the subjects are requested to recall the presented items. Then, a recognition task is proposed. For this task, a card with 40 pictures is presented, and the individuals have to decide which pictures have been shown in the learning phase. This card contains the 10 target items and 30 distractors (10 not related to the target items either semantically or formally; 10 semantically related, i.e., they are prototypes of the categories; and 10 formally related, i.e. they have almost the same form). Comparing performance in recall and in recognition is very useful for the understanding of memory-processing defects in dementia. The analysis of error types in recognition is also valuable. We assessed executive functions through 2 tasks: the cognitive estimation task and the category verbal fluency task. The first is based on a well-known theoretically and experimentally validated test developed by Shallice and Evans [32]. The second is classically used in neuropsychological assessment as part of executive function evaluation [26]. For this purpose, we prefer a letter fluency task but used a semantic fluency test, since the latter is less sensitive to educational level and school experience [33]. We examined constructive apraxia, which is the type of apraxia most sensitive to brain lesions [34] using a short version of the “Stick construction task” developed by Butters and Barton [35], in which subjects are invited to put sticks together to reproduce 2-dimensional geometric constructions. With its 10 subtests, the DSB-100 attempts to screen dementia and to type cognitive dysfunctions in individuals including those with little or no formal education. The goal of the current study was to present the DSB-100, the rationale behind its construction, and to examine its construct validity and reliability after verifying and correcting for the impact of sociodemographic factors namely sex, age, and education on its scores.

Participants

The study was conducted in Tunisia, North Africa. Two groups of participants took part in this study: a group diagnosed as demented and a group of healthy controls. The first group consisted of 42 demented patients; 27 female and 15 male. The mean age of this group was 69.38 years (SD = 7.75), and the mean number of years of education was 5.93 (SD = 6.84). All met the DSM-IV criteria for dementia [22]. The diagnosis of dementia was taken by a panel discussion including neurologists and neuropsychologists according to gold standards based on a complete medical and neurological exam, a neuropsychological testing of major cognitive and conative functions, an objective assessment of daily life activities using an adapted version of the Instrumental Activities of Daily Living [36], biological exams (ApoE, hemocysteine, vitamin B12, folate, etc.) and a CT scan, MRI, or SPECT imaging. The demented participants were outpatients recruited from the Memory Center of the Department of Neurology at Razi Hospital in Tunis, Tunisia. Among the patients, 26 were diagnosed as having mild dementia (CDR = 1) and 16 as moderate (CDR = 2). Patients having very mild/uncertain dementia (CDR = 0.5) or very severe dementia (CDR >3) were excluded from the analysis of the current study, in addition to all patients having other chronic neurological conditions such as epilepsy, multiple sclerosis, patients living in nursing homes, or patients having significant hearing, sight, or motor deficits, and those with any combination of the above that may interfere with neuropsychological testing. Healthy matched controls were 159 subjects; 81 female and 78 male. The mean age was 67.80 years (SD = 7.25), and the mean number of years of education was 5.47 (SD = 6.64). All subjects were examined by a physician for general health. All showed no cognitive impairment (CDR = 0) and were independent in activities of daily living. Individuals having sensory, motor, or intellectual deficiencies or a personal history of a psychiatric disorder, addiction, or a central nervous system disorder were excluded from the study. The controls were relatives of the patients and normal healthy individuals living in normal non-institutional residence. Individuals who did not live anymore in their family setting or who lived permanently in elderly centers were also excluded. As illustrated in Table 1, no significant differences were observed between the groups in terms of sex (χ2(1) = 2.37, p = 0.123), age (t(199) = 1.23, p = 0.218), and education (t(199) = 0.46, p = 0.645).
Table 1

Demographic and general neuropsychological data of the dementia patients and healthy controls

Dementia patients (n = 42)Healthy controls (n = 159)p value
Sex
 Female, n (%)27 (64.29%)81 (50.94%)0.123
 Male, n (%)15 (35.71%)78 (49.06%)
Age, years
 Mean ± SD69.38±7.7567.80±7.250.218
 Min-max55–8955–90
Education, years
 Mean ± SD5.93±6.845.47±6.640.645
 Min–max0–210–21

A-MMSE23.65±3.5728.14±1.530.001
A-FAB10.59±2.5515.86±1.480.001
A-GDS12.65±6.986.91±3.920.001

χ2 was used for sex comparison. The Student t test was used for age and education, A-MMSE, A-FAB, and A-GDS comparisons. A- indicates the Arabic version of the respective tests; MMSE, Mini-Mental State Examination; FAB, Frontal Assessment Battery; GDS, Geriatric Depression Scale.

Procedure

All participants underwent a structured interview to obtain demographic data, a detailed personal and family medical history, and information on alcohol or drug consumption. For patients with dementia, all required information was collected from relatives if needed. Cognitive tests were administered individually by clinical neuropsychologists after obtaining informed consent. The DSB-100 was administered as part of a larger neuropsychological assessment including the A-MMSE [37], the A-GDS [38], and the A-FAB [39] to have additional indices on the severity of patients' dementia, depression, and potential frontal dysfunctions. The same information was also used for comparisons with DSB-100 scores.

Statistical Analysis

To examine the effects of demographic variables on performance, we used a multiple linear regression analysis where the DSB-100 total score was the dependent variable and sex, age, and education were considered as predictors (independent variables). The Student t test was used to compare the groups on demographic variables or overall difference in their DSB-100 scores. A MANOVA was also conducted to compare the group differences in the DSB-100 subtests followed by omnibus univariate tests. η2 was used for measuring effect size, and the significance level was set at p < 0.05. To validate the DSB-100, we opted for the construct validation rather than other forms in the absence of any other available concurrent measure. Construct validity was done using principal component factor analysis (PCFA) with varimax rotation. Factor extraction was determined using a minimum eigenvalue of 1. To determine reliability, 2 types of coefficients were used: internal consistency using Cronbach's alpha coefficient and test-retest reliability using the Pearson Product Moment Correlation Coefficient, where 27 patients and 34 healthy controls were retested by the same clinician 4–6 weeks after the initial assessment.

Results

The Kolmogorov-Smirnov test showed normal DSB-100 distribution for the healthy controls (Dmax = 0.075, p > 0.20) and for the demented patients (Dmax = 0.105, p > 0.20). Table 2 presents means and standard deviations for each group and each subtest. For both groups, the highest scores were seen in the Comprehension, Naming, and Recognition subtests. The poorest scores were seen in the Backward Names Span followed by the Forward Names Span, and Line-Object Orientation Judgment subtests.
Table 2

Univariate analysis of the DSB-100 subtest performance in the dementia patients and healthy controls

SubtestDementia patients (n = 42)Healthy controls (n = 159)p valueη2
Forward Names Span3.95 (1.27)5.47 (1.12)0.0010.19
Backward Names Span2.59 (0.99)4.18 (1.24)0.0010.18
Recall4.76 (0.87)7.08 (1.69)0.0010.28
Recognition6.24 (1.16)8.92 (1.25)0.0010.42
Verbal Fluency4.90 (1.61)8.72 (1.39)0.0010.57
Comprehension7.57 (1.71)9.63 (0.79)0.0010.34
Naming7.24 (1.28)9.56 (0.80)0.0010.39
Cognitive Estimation5.29 (1.67)8.39 (1.53)0.0010.38
Line-Object Orientation Judgment4.26 (1.38)5.93 (1.75)0.0010.11
Constructive Praxia5.64 (1.41)8.84 (1.50)0.0010.39

Total DSB-10052.45 (5.79)76.74 (7.54)0.0010.74

All values are represented as means (SD).

For the total DSB-100 data, the overall multivariate test (Wilk's λ) was significant (F(10, 190) = 62.49, p < 0.001, η2 = 0.76). Using the adjusted A-GDS score as a covariate in the multivariate analysis slightly attenuated the effect size, but it did not change the pattern of the results (F(10, 190) = 44.48, p < 0.001, η2 = 0.74). For the MANOVA, omnibus univariate tests were significant (p < 0.001) for all subtests. Effect sizes ranged from 0.11 for the Line-Object Orientation Judgment subtest to 0.57 for the Verbal Fluency subtest. Table 2 presents the means and standard deviations by group along with significance tests and effect sizes for each subtest.

Effects of Demographic Variables on the DSB-100

The effect of demographic variables (sex, age, and education) on the DSB-100 scores were analyzed by step-wise multiple linear regression analysis. The results indicated that, while the DSB-100 score was significantly sensitive to age and education (p < 0.001), it was not affected by sex (p = 0.262). The original scores of the DSB-100 should be adjusted for age and education: final score = 82.05 + [(number of years of education × 0.778) + (age × −0.160)]. Regression analysis retains “age” and “education” as variables with significant importance, while “sex” was excluded as insignificant. The results in Table 3 show the value of R2 and suggest that the percentage of variance explained by the 2 sociodemographic variables age and education was 64%. The DSB-100 performance depends heavily on age and education. Thus, the scores obtained by the DSB-100 must be adjusted according to the relative weight of the significant demographic variables, namely age and education. For this purpose, we ran a calculation using the multiple linear regression of the quantitative impact of the variables age and education on the total DSB-100 score. Table 4 provides the suggested appropriate corrections to adjust the observed total scores by adding the necessary correction (positive or negative) based on age and education level of the participants.
Table 3

Relationship between the total score of the DSB-100 and the sociodemographic variables: value of standardized β coefficients and coefficients of determination R2

PredictorsCoefficients
Model
non-standardized βstandardized βStudent t testp valueF (ANOVA)p valueR2
Constant82.05145.590.00010.64

Sex0.0050.960.924
Education1.090.77816.410.0001
Age−0.166−0.160−3.370.001
Table 4

Values of corrections to the initial scores to obtain adjusted total scores

Age, years
55–5758–6364–6970–7475–8081–86≥87
Education, years
0–11234567
2–30123456
4–5−1012345
6–7−2−101234
8–9−3−2−10123
10–11−4−3−2−1012
12–13−5−4−3−2−101
14–15−6−5−4−3−2−10
16–17−7−6−5−4−3−2−1
18–19−8−7−6−5−4−3−2
20–21−9−8−7−6−5−4−3
Hereafter, we provide 3 examples for the use of the correction table (Table 4). For instance, for an illiterate 67-year-old man with a total observed score of 60 on the DSB-100, 3 points must be added to his initial score to arrive at the adjusted score where the effects of age and education are taken into account. His adjusted score is 63. Another example, for a 55-year-old woman who had had only 2 years of education and who obtained 87 points on the DSB-100, we keep the score as it is because the correction value from Table 4 is 0. A last example, for a woman aged 60, with a Master of Science (equivalent to 17 years of education) and a score of 75 on the DSB-100, we need to deduct 6 points from the total score, so her adjusted score is 69. When we applied a multiple regression analysis to each DSB-100 subtest, the results revealed that the effect of education was highly significant for all subtests, while the effect of age was merely significant for the Constructive Praxia (p < 0.001) and Recognition (p < 0.05) subtests (Table 5). An effect of sex was observed only for the Forward Digit Span subtest scores (p < 0.05).
Table 5

Standardized scores for the effects of sex, age, and education on the DSB-100 subtest scores

Subtests and total DSB-100SexAgeEducation
Forward Digit Span0.151*−0.1150.574***
Backward Digit Span0.118−0.0660.588***
Recall−0.043−0.0240.582***
Recognition−0.034−0.186*0.357***
Verbal Fluency−0.140−0.0320.244**
Comprehension−0.110−0.1200.215**
Naming−0.080−0.1380.254***
Cognitive Estimation0.042−0.0360.185*
Line-Object Orientation Judgment0.042−0.0590.870***
Constructive Praxia−0.026−0.263***0.382***

p < 0.05

p < 0.01

p < 0.001.

Validity of the DSB-100

To validate the DSB-100, we opted for the construct validation, which is considered the most appropriate for adapted or newly developed instruments. Construct validity was done using PCFA on the healthy control group's adjusted scores. The performed exploratory PCFA using eigenvalues >1 yielded a 4-factor solution. Subjected to an orthogonal rotation (varimax), the 4 extracted factors accounted for 67.05% of the total variance with 29.96, 15.92, 11.09, and 10.08%, respectively. The factors can be labeled as: (a) “Attentional and Visuospatial”, composed of efficiency measures on the Line-Object Orientation Judgment, Forward Names Span, Backward Names Span, and Constructive Praxia subtests, (b) “Memory”, defined by high loadings for the Recall and Recognition subtests, (c) “Language”, based on the scores of the Comprehension and Naming subtests, and (d) “Executive Functions”, consisting of high loadings on the Cognitive Estimation and Verbal Fluency subtests (Table 6).
Table 6

Varimax with a Kaiser normalization-rotated component matrix for the DSB-100 items in the healthy control group

SubtestsFactors
Attentional and VisuospatialMemoryLanguageExecutive Functions
Line-Object Orientation Judgment0.840.08−0.060.05
Forward Names Span0.830.10−0.010.02
Backward Names Span0.820.210.180.13
Constructive Praxia0.470.360.060.24
Recall0.110.820.03−0.01
Recognition0.200.76−0.020.02
Comprehension0.050.100.840.09
Naming0.01−0.060.820.06
Cognitive Estimation0.020.15−0.040.84
Verbal Fluency0.21−0.140.280.72

Explained variance, %29.9615.9211.0910.08

Reliability of the DSB-100

As illustrated in Table 7, the internal consistency of the DSB-100, obtained by Cronbach's alpha for the healthy controls, was acceptable (α = 0.76). This indicates that all items measure the same construct without redundancy.
Table 7

Item characteristics of the DSB-100

ItemsCronbach's alpha if item is deleted
Forward Names Span0.72
Backward Names Span0.72
Recall0.71
Recognition0.73
Verbal Fluency0.76
Comprehension0.75
Naming0.75
Cognitive Estimation0.76
Line-Object Orientation Judgment0.69
Constructive Praxia0.73

Total DSB-100 score0.76
We calculated the test-retest reliability by re-testing 27 patients and 34 controls, by the same clinician, 4–6 weeks after the initial assessment. The test-retest correlation coefficients for the total DSB-100 score calculated for each sample were high for patients (r = 0.81, p < 0.001) and for healthy matched controls (r = 0.87, p < 0.001). This indicates that performance on the DSB-100 is stable over time.

Discussion

The first objective of this study was to introduce the DSB-100 and to present the rationale behind its construction. This new dementia screening battery was designed to address shortcomings found in existing dementia screening instruments (such as the MMSE, Mattis, MoCA, or 3MS) by including more accurate assessments of long-term memory, an ecological assessment of short-term and working memory, a contextualized assessment of visuospatial functions, an assessment of executive functions, and the inclusion of less culturally biased procedures and instrument contents. The second objective was to assess the effect of sociodemographic variables, namely sex, age, and education, on the battery scores and to examine the instrument's psychometric properties. Concerning the rationale behind the construction of the DSB-100, 2 main reasons motivated its conception. The first sought to address fundamental issues when using existing dementia screening tools in non-Western cultures. The second sought to bridge the gap between the rapidly changing knowledge in the field of dementia and the validity of dementia screening tools. To this effect, the proposed instrument sought to correct biases of cultural nature in available dementia screening instruments and to address methodological issues arising from the interplay of both the fundamental and cultural factors combined. These problems affect the assessment of dementia and called for the construction of a new screening instrument to overcome the many limitations of widely used dementia screening tools. There are 2 main shortcomings of fundamental nature with the available dementia screening tools: they miss some major functions, and they do not detect early stages of dementia. Currently, there is a clear incongruity between the psychological functions tapped in existing screening tools and the evolving diagnostic criteria of dementia. For instance, executive functions are a major criterion for dementia, ever since the publication of the DSM IV [22], and are currently critically important diagnostic criteria of dementia. However, they are not measured in widely used dementia screening instruments such as the MMSE and 3MS. False negatives can easily occur for patients suffering from frontotemporal dementia, corticobasal dementia, vascular dementia, and subcortical dementia. In addition, existing screening tools do not dissociate between recall and recognition in long-term memory tasks, even though it is important to distinguish between cortical and subcortical dementias [40]. Available screening instruments offer just gross measures of visuoconstructive functions and do not dissociate between visuospatial and visuoconstructive deficits – that can be helpful in differentiating between Lewy Body dementia and Alzheimer disease [41]. Screening tools should be able to detect dementia as early as possible to enable early intervention, rehabilitation, and disease management. Nevertheless, most of the existing screening tools (e.g., the MMSE, MoCA, and Mattis) seem to serve rather as a severity measure than an early detection tool. For instance, the MMSE devotes one third of its total score to temporal and spatial orientation, alterations happen here at later stages of dementia, while it fails to detect executive and working memory dysfunctions, which occur in the early stages of many types of dementia [42, 43, 44]. Existing dementia screening instruments are notorious for overlooking important cultural dimensions and for including biases that affect the proper assessment of dementia and, thus, its detection [45]. The development of a culturally sensitive and appropriate instrument has the potential to reduce assessment biases and interpretation errors. Studies in cultural neuroscience offer proof of cultural differences in brain activation patterns. Such differences were observed in both lower-level processes [46, 47, 48, 49, 50] as well as higher-order ones [51, 52, 53, 54]. Findings argue for a brain architecture that is not universal (for a review, see Caramazza [55]) despite the fact that the brain works according to identical laws [56]. The aforementioned arguments represent findings from studies on brain-injured illiterate patients [30, 57], brain-injured bilinguals [57], and evidence from experimental neuroscience with normal monolinguals and bilinguals [58]. All testify that differences observed in performance in cognitive style do not vary only according to biological factors, the nature and extent of brain lesion, laterality, sex, and age. These differences also depend on cultural factors that arrange cerebral organization itself. Among these factors, we noted the effect of the level of education, bilingualism, and the importance of the type of the script used to write on brain organization. In addition to the importance of the impact of cultural factors on cognitive functioning, demographic specificities are well-established determinants of cognitive performance. Knowing specific demographic facts is essential to conceptualize and to build any health program for dementia. African countries, for example, are characterized by a high level of fertility, a high level of multiculturalism, and a high level of illiteracy [59]. By 2050, the worldwide population over 60 years will reach two billion people, three quarters of which will be from African countries. These countries are the hallmark of multicultural, multilingual, and multiethnic societies and have the highest rates of illiteracy in the world [60]. For instance, in Tunisia, more than 85% of all individuals older than 60 years are illiterate [61], and the country is predominantly multicultural. Throughout history, the indigenous Berber culture of Tunisia has been influenced by multiple invading civilizations including the Carthaginians, Romans, Vandals, Bezants, Africans, Arabs, Turkish, Spanish, and French, but the current main cultural influences remain Arabic and Islamic [62]. Besides, the sociolinguistic context is quite complex, since Phoenician, Latin, Arabic, Hebrew, Italian, Spanish, Turkish, and French have been added to the Berber basis [62]. In the Western world, illiteracy is not generally a problem among the elderly [63]. However, in non-Western and low-to-middle-income countries such as Tunisia, illiteracy continues to be characteristic of elderly people especially women [26]. Existing dementia screening tools assess writing, reading, calculating, and drawing. When the assessed elderly is “culturally different” and illiterate like it is the case for many people in African, Asian and South-American countries, existing dementia screening tools do not only contain cultural biases but are inherently inappropriate as screening tools. To continuously use them would systematically introduce methodological issues and fundamental diagnostic errors resulting in high rates of false positives [64]. Adapting the DSB-100 to the Tunisian context produced a culturally appropriate measure that has a great potential for application in similar contexts (such as elderly from developing countries of non-Western cultures and from countries with low literacy levels; the DSB-100 may also be the instrument of choice with immigrants from developing countries of non-Western cultures residing in Western countries). The DSB-100 has also the potential to be easily adapted to many other cultures (Western and non-Western) with straightforward modifications to some of its subscales. It simply requires the inclusion of verbal and pictorial stimuli that are equivalent to the DSB-100 items, based on local standardized databases similar to Battig and Montague's verbal categorical database [10] and Snodgrass and Vanderwart's pictorial database [11]. In addition to the importance of the impact of cultural factors no cognitive functioning, demographic specificities are well-established determinants of cognitive performance [64]. The study findings outlined the sensitivity of the DSB-100 to some sociodemographic variables. It revealed that besides all the efforts to select and develop tasks that are applicable to illiterate and low-educated individuals, the influence of education remains significant. Task-taking behavior may inherently contribute to lowering test performance in illiterate individuals, as they tend to perform poorly on tests due to a discernible lack of training and clear deficiencies in test taking [65]. Illiteracy in itself affects cognitive capacities (for a review, see Julayanont and Ruthirago [66]), and this may explain why the educational effect was significant in all tasks, even if the task acceptability and instruction comprehension were fine. The adjustment of obtained data may, in part, solve the problem. Barona et al.'s adjustment technique [67] was useful to attenuate general education and age effects, even though it may be criticized, since it presumes a linearity of the effect of education and age on cognitive performance [68]. The study was conducted in Tunisia, where Tunisian, an Arabic dialect, is spoken. Since Tunisia is a developing country, we selected 55 years of age as our lower bound – as was the case for Chinese [69] or Brazilian populations [70]. The DSB-100 displayed good psychometric properties. The total score's reliability concerning its stability was high in the test-retest method and quite satisfactory regarding its internal homogeneity. For the internal consistency of the DSB-100 subtests, Cronbach's alpha was acceptable in the healthy matched controls for all of the subtests Factor analysis was a useful quantitative method to examine the dimensionality of the DSB-100. In fact, data reduction yielded 4 factors, namely “Attentional and Visuospatial”, “Memory”, “Language”, and “Executive Functions”, providing a starting point for clinicians in the diagnosis of dementia types. Evidence from the construct validity proved that the clinical utility of the DSB-100 was comparable to most widely used tests like the Mattis [16]. Thus, the DSB-100 is an instrument for the detection and the assessment of cognitive changes associated with dementia. It was constructed for a non-Western elderly population with low literacy/illiterate individuals living in a developing country. However, as noted previously, the instrument is easily adaptable to many other cultures (Western and non-Western) with straightforward modifications to some of its subscales. This is especially the case as tasks composing this instrument were selected from various well-known neuropsychological tests. This study showed that the DSB-100 is a suitable, reliable, and valid tool for the screening of dementia. It yields information about 4 cognitive domains: attention and visuospatial functions, memory, language, and executive functions. It can also be useful to orient clinicians toward the type of dementia condition via the analysis of the cognitive deficit profile and to plan intervention treatment strategies. Possessing normative data and deciding about cutoff scores are essential to take diagnostic decisions and to apply the instrument to follow-up patients' change under cognitive rehabilitation and/or appropriate medication. Other aspects of the psychometric properties of the DSB-100 need to be examined through further studies, especially concurrent validity and clinical validity respecting different stages of severity of cognitive decline in different types of dementia.

Disclosure Statement

The authors declare no conflict of interest.
  42 in total

1.  Culture, control, and perception of relationships in the environment.

Authors:  L J Ji; K Peng; R E Nisbett
Journal:  J Pers Soc Psychol       Date:  2000-05

2.  Cancellation task in very low educated people.

Authors:  Sonia Maria Dozzi Brucki; Ricardo Nitrini
Journal:  Arch Clin Neuropsychol       Date:  2008-02-21       Impact factor: 2.813

Review 3.  Distinguishing Alzheimer's disease from other major forms of dementia.

Authors:  Stella Karantzoulis; James E Galvin
Journal:  Expert Rev Neurother       Date:  2011-11       Impact factor: 4.618

4.  Neuropsychological assessment in illiterates. II. Language and praxic abilities.

Authors:  M Rosselli; A Ardila; P Rosas
Journal:  Brain Cogn       Date:  1990-03       Impact factor: 2.310

5.  Effect of parietal lobe damage on the performance of reversible operations in space.

Authors:  N Butters; M Barton
Journal:  Neuropsychologia       Date:  1970-04       Impact factor: 3.139

6.  Differential recognition of tachistoscopically presented English and Hebrew words in right and left visual fields.

Authors:  M I Barton; H Goodglass; A Shai
Journal:  Percept Mot Skills       Date:  1965-10

7.  A standardized set of 260 pictures: norms for name agreement, image agreement, familiarity, and visual complexity.

Authors:  J G Snodgrass; M Vanderwart
Journal:  J Exp Psychol Hum Learn       Date:  1980-03

8.  Visuospatial judgment. A clinical test.

Authors:  A L Benton; N R Varney; K D Hamsher
Journal:  Arch Neurol       Date:  1978-06

9.  Neuropsychological assessment in illiterates: visuospatial and memory abilities.

Authors:  A Ardila; M Rosselli; P Rosas
Journal:  Brain Cogn       Date:  1989-11       Impact factor: 2.310

Review 10.  Effects of development and enculturation on number representation in the brain.

Authors:  Daniel Ansari
Journal:  Nat Rev Neurosci       Date:  2008-04       Impact factor: 34.870

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.