Literature DB >> 36092691

Development and design of a diagnostic report to support communication in dementia: Co-creation with patients and care partners.

Aniek M van Gils1,2, Leonie N C Visser1,2,3,4, Heleen M A Hendriksen1,2, Jean Georges5, Wiesje M van der Flier1,2,6, Hanneke F M Rhodius-Meester1,2,7.   

Abstract

Introduction: Clear communication of diagnostic test results and dementia diagnosis is challenging yet important to empower patients and care partners. A personalized diagnostic report could support the communication of dementia diagnostics and aid patients' understanding of diagnosis. In this study, we aimed to design a diagnostic report in co-creation with patients and care partners.
Methods: We used a mixed-methods approach, combining surveys with focus groups in iteration. Phase 1 consisted of an international survey assessing needs among patients (n = 50) and care partners (n = 46), and phase 2 consisted of focus group meetings (n = 3) to co-create the content and to hands-on co-design the layout of the diagnostic report with patients (n = 7) and care partners (n = 7). Phase 3 validated results from phase 2 in a survey among patients (n = 28) and care partners (n = 12), and phase 4 comprised final feedback by dementia (care) experts (n = 5). Descriptive statistics were used to report quantitative results and directed content analysis was used to analyze qualitative data.
Results: Most patients (39/50, 78%) and care partners (38/46, 83%) positively valued a diagnostic report to summarize test results. The report should be brief, straightforward, and comprise results of the diagnostic tests, including brain imaging and information on future expectations. Despite a clear preference for visual display of test results, several visualization options were deemed best and were equally comprehended. Discussion: In this study, we developed a prototype of a personalized patient report through an iterative design process and learned that co-creation is highly valuable to meet the specific needs of end-users.
© 2022 The Authors. Alzheimer's & Dementia: Diagnosis, Assessment & Disease Monitoring published by Wiley Periodicals, LLC on behalf of Alzheimer's Association.

Entities:  

Keywords:  brain imaging; communication; dementia; diagnosis; diagnostic testing; neuropsychology; prevention; prognosis; progression

Year:  2022        PMID: 36092691      PMCID: PMC9446898          DOI: 10.1002/dad2.12333

Source DB:  PubMed          Journal:  Alzheimers Dement (Amst)        ISSN: 2352-8729


BACKGROUND

The diagnostic workup of dementia requires an elaborate set of investigations, including neuropsychological testing and brain imaging, sometimes supplemented with biomarkers. , , , , , With increasing use of these tests, a new challenge for the clinician arises: how to adequately discuss these results with the patient? Should they explain every test result or summarize the conclusion? How to stimulate understanding while preventing information overload? Patients and care partners need to remember and understand the given information to cope with their condition and its implications. A lack of understanding about dementia might lead to uncertainty about the future and suboptimal disease management. However, patients and care partners frequently feel that test results and their meaning are inadequately explained. , This might result from patients’ recall of information, as patients generally remember only half of the provided information directly after consultation and even less in the long term. Furthermore, the more information the clinician supplies, the more the patient will forget —which is problematic considering the increasing number of diagnostic tests. Several communication strategies could be used to present information in a comprehensible and orderly manner to stimulate understanding and recall in cognitively impaired patients. Health communication research has shown that tailored information is more effective for recall than standard information. In other medical fields like oncology, written summaries have promoted patient understanding and recall. , Effective explanation of test results could also be facilitated by visual aids, for example, adding visualizations to written and spoken language increases patients’ comprehension and recall, , especially for those with low literacy. , Overall, an individualized diagnostic report containing tailored information and visualized test results could be a delicate way of presenting a patient's health data in a comprehensible manner, supporting clinician‐patient communication about test results, and stimulating information recall and understanding by patients and their care partners. , In this study, we aim to design a diagnostic report in an iterative design process, in co‐creation with patients and care partners, following the concept of design thinking. By creating the diagnostic report in iteration with the end‐users, we aim to meet their specific needs and preferences.

RESEARCH IN CONTEXT

Systematic Review: Patients and care partners need to remember and understand all given information on diagnostic test results and dementia diagnosis to cope with their condition and its implications. The literature shows that several communication strategies could be used to present information in a comprehensible manner. A tailored diagnostic report including diagnostic test results could aid communication and stimulate understanding and recall of all given information. Interpretation: Through several iterative steps involving patients and care partners, this study resulted in the development of a prototype of a tailored diagnostic report including information on diagnosis, visualization of diagnostic test results, brain health, practical issues, and information sources. Future Directions: This study showed that one size does not fit all regarding the diagnostic report. Preferably, diagnostic reports should be compiled on the spot, attuned to patients’ and care partners’ informational needs, preferences, phase of the disease, and health literacy.

METHODS

Study design

We used an iterative, mixed‐methods study design based on the design thinking framework consisting of four phases (Figure 1). Table 1 shows an overview of the study participants per phase of the study. The study was reviewed by the board of the Medical Ethics Committee of the VU University Medical Center, Amsterdam UMC. All participants gave (digital) informed consent before participation.
FIGURE 1

Study flowchart of developing the diagnostic report

TABLE 1

Participant characteristics

1. Needs survey2. Focus groups3. Validation survey
PhasePatientsCare partnersPatientsCare partnersPatientsCare partners
N 50 a 467728 b 12
Age, years 73 ± 8 c 65 ± 12 c 68 ± 670 ± 475 ± 7 c 69 ± 6 c
Female (%) 17 (41%) c 25 (60%) c 4 (57%)4 (57%)5 (18%)3 (25%)
Education, years 11.8 ± 4.112.2 ± 4.412.3 ± 2.612.3 ± 2.612.3 ± 4.411.9 ± 4.1
Relation to patient, n (%)
SpouseN/A33 (72%)N/A6 (86%)N/A12 (100%)
(Grand)childN/A12 (26%)N/A0N/A0
Sister/brotherN/A1 (2%)N/A0N/A0
OtherN/A0N/A1 (14%)N/A0

Diagnosis of patient, n (%) d

SCD21 (42%)2 (4%)5 (71%)2 (29%)10 (36%)0
MCI16 (32%)8 (17%)2 (29%)2 (29%)11 (39%)1 (8%)
Dementia13 (26%)36 (78%)0 (0%)3 (43%)7 (25%)11 (92%)

Note: Data represent mean ± SD or n (%). The groups of participants in phase 2 and phase 3 are subgroups of the participants from phase 1.

Abbreviations: MCI, mild cognitive impairment; N/A, not applicable; SD, standard deviation; SCD, subjective cognitive decline.

Of whom n = 20 (44%) participated together with their care partner.

Of whom n = 10 (36%) participated together with their care partner.

Pairwise comparisons indicate a group difference (P < .05).

Self‐reported data.

Study flowchart of developing the diagnostic report Participant characteristics Diagnosis of patient, Note: Data represent mean ± SD or n (%). The groups of participants in phase 2 and phase 3 are subgroups of the participants from phase 1. Abbreviations: MCI, mild cognitive impairment; N/A, not applicable; SD, standard deviation; SCD, subjective cognitive decline. Of whom n = 20 (44%) participated together with their care partner. Of whom n = 10 (36%) participated together with their care partner. Pairwise comparisons indicate a group difference (P < .05). Self‐reported data.

Phase 1—Needs assessment

Online survey

We performed a needs assessment using an online survey sent out between July and November 2020. The survey was part of a more extensive survey, of which the methods are described in detail in our previous study. We added specific topics relevant to the present study to explore the views of patients and their care partners regarding a diagnostic report. To explore topics they wanted to include in a report, we used a predefined list of 10 topics , (Table 2), which they could complement with their topics.
TABLE 2

Preferences regarding pre‐selected list of ten topics during phase 1—the needs survey

Topics (n, %)Total (n = 76)Patients (n = 42)Care partners (n = 34) P‐value
Short explanation of the diagnosis68 (90%)39 (93%)29 (85%).244
The test results63 (83%)37 (88%)26 (77%).151
Future – What can I expect over the course of the symptoms63 (83%)35 (83%)28 (82%).574
Tips for brain health62 (82%)36 (86%)26 (77%).230
Diagnosis60 (79%)31 (74%)29 (85%).174
The best way (for the person I care for) to cope with (their) symptoms58 (76%)28 (67%)30 (88%).025
The best way to cope with the person I care forN/AN/A30 (88%)N/A
Pictures of brain imaging45 (59%)25 (60%)20 (59%).568
Where to find more information38 (50%)19 (45%)19 (45%).245
Test results compared with other people of the same age29 (38%)19 (45%)10 (29%).120

Abbreviations: N/A, not applicable.

Participants

An international, memory clinic population from the Amsterdam Dementia Cohort (ADC), , the Amsterdam Ageing Cohort (AAC), Alzheimer Europe and Alzheimer's Society UK, consisting of patients (n = 50) with subjective memory decline (SCD, n = 21, 42%), mild cognitive impairment (MCI, n = 16, 32%), and dementia (n = 13, 26%); and care partners (n = 46) of patients with SCD (n = 2, 4%), MCI (n = 8, 17%), or dementia (n = 36, 78% dementia) participated (Table 1).

Analysis

Quantitative survey results were analyzed using IBM SPSS Statistics version 26. P‐values <.05 were considered statistically significant. Frequencies and percentages of topics in the topic list were calculated, and similarities between patients and care partners were studied.

Phase 2—Design

Subsequently, we used focus group meetings (1) to further examine the most relevant topics to be summarized, using the topic list from phase 1 as input; and (2) to design the diagnostic report.

Focus group meetings

In November and October 2020, we conducted three focus groups, each consisting of four to six participants with a duration of 2 hours. We recruited participants from all Dutch survey respondents. Patients were eligible if they had an Mini‐Mental Status Examination (MMSE) score ≥18. All focus groups were conducted in the Alzheimer Center Amsterdam, Amsterdam UMC. Due to the coronavirus disease 2019 (COVID‐19) situation, the focus groups had a hybrid format, with some participants (n = 3) participating through a video connection and others being present live. All groups included a mix of patients and their care partners, patients alone, or care partners alone, and were heterogeneous in gender, age, and disease stage. All participants were relatively highly educated (Table 1).

Procedures

The focus group procedure was based on the nominal group technique (NTG) and consisted of three parts. All focus groups were audio‐recorded. First, we provided participants with topics from phase 1 and asked them to rate these on a 5‐point Likert scale (1, unimportant; 2, not very important; 3, neutral; 4 important; and 5, very important). Afterward, we asked each participant to explain which topics they had rated as important (Likert 4 or 5) and why. Results were discussed, and participants were asked if they would change their initial ratings. Then we asked participants to select five topics they considered most important. Second, we supplied the participants with six visual options to display test results (Figure 2). The visualizations were based on original designs from Visualizing Health, a collaborative project between the University of Michigan and the Robert Wood Johnson Foundation made publicly available via Creative Commons license and derived from a literature search. , , We included visualizations with no text, a small amount of text, and only text, and we asked participants to rank the visualizations. Afterward, the rankings were discussed.
FIGURE 2

Visualizations to depict test results

Visualizations to depict test results Third, participants were instructed to design their preferred diagnostic report. All topics from part 1 and all visualizations from part 2 were available during this hands‐on design exercise. Participants were encouraged to write down additional information from their perspectives. One author (A.G.) transcribed the audiotapes using intelligent verbatim transcription. A second author (H.H.) checked the transcripts. The transcripts were analyzed using directed content analysis using MAXQDA software. One author (A.G.) generated a start list of codes. Two authors (A.G., H.H.) independently coded all audio records and the designed diagnostic reports. Any text that could not be categorized based on the start list was given a new code. Afterward, A.G. and H.H. discussed the codes until consensus was reached. To select the most important topics to be summarized and the best visualization(s) for test results, and to build the initial prototype, quantitative and qualitative analyses were combined.

Phase 3—Validation survey

According to phase 2, there was no consensus on the visualization of test results. To further investigate the ideal visualization and validate the layout of the prototype, we conducted a second survey among patients and care partners in which Dutch participants from the initial survey were invited. Patients who participated in the focus groups were excluded. In total, n = 28 patients and n = 12 care partners participated (Table 1). We provided the participants with a fictive case of a person with dementia and the diagnostic report with summarized diagnostic test results. To test the comprehensibility of the report, we asked questions about the content of the report (e.g., “what is the diagnosis of this patient?”). To test the comprehensibility of the individual visualizations, we supplied two versions of each visualization—one correctly illustrated the test result mentioned in the text and one that was not in correspondence to the written test result—and asked the participants which visual corresponded to the patients’ results as summarized in the report. In addition, we asked which type of visualization they would prefer to summarize test results.

Phase 4—Feedback round

Phase 4 emerged during the study, spurred by spontaneous written feedback from one of the care partners participating in the survey and additionally from a formal caregiver to whom a participating care partner supplied the prototype. Accordingly, we decided to present the prototype to other dementia (care) experts (two dementia care nurses and one elderly care physician) to gather more feedback, after which the final prototype was compiled.

RESULTS

As part of the more extensive project, we performed a survey to assess the views of clinicians, patients, and care partners toward computer tools for support during the diagnostic process (Clinical Decision Support Systems [CDSS]). If a CDSS could be used to explain test results in “layperson” language, 78% of patients and 83% of the care partners would want their doctor to use such a tool. Ninety‐two percent (71/77) of those patients and care partners would like to receive a hand‐out of this explanation on paper. Patients and care partners would use this hand‐out to re‐read it at home (72/96, 75%) or to inform family and friends about their situation (47/96, 49%). These percentages did not differ between patients and care partners (P‐values >.20), but patients were more inclined to use this summary to compare future test results (33/50, 66%) than care partners (21/46, 46%, P = .036). Table 2 shows our list of pre‐selected topics and the proportion of patients and care partners selecting topics to be included in a diagnostic report. Nine of 11 pre‐selected topics were chosen by over half of the patients and care partners. Apart from the topic, the best way (for the person I care for) to cope with (their) symptoms (patients 67%, care partners 88%, P = .025), no differences were found in answers between patients and care partners. Suggestions for additional topics were: contact details of the memory clinic, information on facilities, services, and support groups. Preferences regarding pre‐selected list of ten topics during phase 1—the needs survey Abbreviations: N/A, not applicable.

Focus groups—topics

Given the high percentages of favored topics in phase 1, selecting the most relevant topics was challenging. Based on phase 1 and the first focus group results, we divided the topic “test results” into (1) Lumbar puncture, and (2) Cognitive test results and added the topic “Explanation of brain imaging,” resulting in a list of 13 topics to present to the second and third focus group participants. After the first part of the focus group, all participants were asked to compose their top‐five topics of the 13‐item topic list. Next, we calculated how often topics appeared in each top‐five and defined the most important topics to be included in the diagnostic report: diagnosis (100%) including a short explanation (85%), test results (cognitive and lumbar puncture) (62%), future expectations (69%), explanation of brain imaging (54%), and how to deal with complaints (54%).

Focus groups—layout

Visualizations

According to the rankings, the top‐three of visualizations of test results consisted of (1) Paragraph (Figure 2D; 57%). (2) Icon (Figure 2C; 43%) and Bar chart with comparative score (Figure 2E; 43%), and (3) Smileys (Figure 2A; 36%). Thematic content analyses revealed that visualization preferences differed between patients and care partners and within patients and care partners groups. One of the participants (patient with MCI, 60 years of age) said,“[..] what phase you are in, and is it for you or your partner? That is quite a difference. Suppose you take the paragraph of text, and you are already a bit further [in the disease process]; it is quite difficult to read, and it is the clearest to show an image.” Furthermore, the participants had different opinions on whether to show their test scores compared to the test scores of other people of the same age. Most participants were positive and thought it could help them better understand their situation. On the other hand, some were more negative and did not see the point in comparing their scores to others. In addition, the visualization with the smileys was judged twofold. On the one hand, negative: “Not the smiley's, I think they do not fit this miserable [situation][…]” (partner of SCD patient, 67 years of age), and on the other hand, positive: “Well I think of course if you use this it is just very clear with all those emoticons. Everybody understands.” (MCI patient, 60 years of age).

Layout

Several important themes regarding the layout of the diagnostic report were identified. The themes were divided into two main categories: (1) Content and (2) Display. Regarding “content,” the report should contain information on where reliable information can be found and how to get formal help. It should entail (practical) tips, for example, how to get in touch with other patients and care partners, and tips for brain health, such as eating healthy and exercising regularly. Furthermore, participants wanted to include information about the future, such as follow‐up procedures and future expectations. Regarding “display,” the report must be brief and concise, that is, include the main points. One of the participants (patient with SCD, 76 years of age) said: “I have been writing [during the consultation], […] and then something goes through your head I have to ask, and a little later you forget it. So that is very nice if you know that you get the main points home.” Furthermore, the report should contain clear language, comprehensible for a layman. One of the participants (care partner, 71 years of age) mentioned: “I can look for that letter [from the general practitioner], […]; however, when I am going to read it, I do not understand it […].”

Focus groups—hands‐on designing the diagnostic report

At the final stage of phase 2, a total of 13 report pages were made by the participants (Figure 3). One report page was constructed by a patient‐care partner couple. Analyses of the hands‐on designing of the layout of the diagnostic report showed that diagnosis was included in 11 of 13 report pages and explanation of diagnosis in 10 of 13. Then, 10 of 13 contained future expectations. Relating this topic, two participants wrote down wanting to know about both expected symptoms and progression rates. Brain imaging was placed on 10 of 13 pages. Both cognitive test results and how to deal with complaints (of loved ones) were included on 9 of 13 pages. A slight majority of 7 of 13 pages contained the topic tips, and lumbar puncture result was placed on 6 of 9 pages. According to phases 1 and 2, a prototype was built.
FIGURE 3

Overview of the hands‐on design session during the focus group meeting.

Overview of the hands‐on design session during the focus group meeting.

Phase 3—Validation

The focus group results indicated a preference but no consensus on the ideal visualization of test results. However, we could exclude two visualizations (Figure 2B,F). To keep the report comprehensible for all end‐users, we also decided to exclude the visual paragraph (Figure 2D). We presented the three remaining visualizations (Figure 2A,C,E) to a larger group in the validation survey. Most participants preferred either the icon (Figure 2C, n = 33, 38%) or the bar chart (Figure 2E, n = 37, 43%). When asked to correctly match one of two visualization versions to the written test result, we found that most participants matched the correct visualization version to the results summarized on the report page. The highest scores were found for the smiley (Figure 2A; 81% answered correctly) and the icon bar (Figure 2C; 79% answered correctly) compared to the bar chart (Figure 2E; 66% answered correctly). After combining the results from the focus groups and the validation survey, the icon (Figure 2C) was chosen as the visualization to display test results for the prototype. After the validation survey, we received written feedback from one of the participants (care partner) and a formal caregiver involved through this care partner. They pointed out that the textual information was not clear enough, contained too many words, and was too academic. They advised us to rephrase some sentences to make the report more comprehensible. According to this feedback, we adapted the prototype and decided to present it to three other dementia (care) experts. They provided us with verbal feedback, and together we made changes to linguistics and layout. Afterward, the final prototype was constructed.

Prototype

The final prototype of the diagnostic report consisted of two pages, shown in Figure 4. The first page includes information on diagnosis and displays the results from diagnostic tests. The second page entails information on the course of the disease, tips for brain health, information on practical issues, and where to find more information.
FIGURE 4

Prototype of the diagnostic report

Prototype of the diagnostic report

DISCUSSION

This study reports the development of a diagnostic report for dementia diagnosis in co‐creation with patients and care partners. Overall, patients and care partners were enthusiastic and eager to join, and they provided constructive input. In the first phase of the study, in which we conducted a needs assessment, we found that patients and care partners positively value a diagnostic report to summarize and communicate test results. In the second and third phases, consensus was reached on the topics to be outlined, and a choice was made on the visualization to explain test results. A prototype was made, and in phase 4, we asked experts for their opinion. Subsequently, we designed the diagnostic report according to the results from all phases, in close collaboration with patients and care partners. Our study intends a first step toward developing a personalized diagnostic report that contributes to a personalized medicine approach. By portraying the needs of patients and care partners through the design thinking process, we developed an example of what a diagnostic report should entail. The diagnostic report developed in the current study can support the harmonization of clinical practice and is aimed at supporting both clinicians and patients. Clinicians might use the report as a guide to explain test results and diagnoses. For patients, the report can improve their understanding of their test results and diagnosis. They can take the report home on paper as a reference page and facilitate communication about test results with relatives. We strived to develop an acceptable and comprehensible diagnostic report that eventually contributes to effective implementation. However, the feasibility of generating and structurally using this report in clinical practice remains to be tested. In a subsequent study, we collaborate in a public‐private partnership to study the implementation of this report in various memory clinics. In that study, the report page will be generated from a CDSS that combines digital cognitive testing, magnetic resonance imaging (MRI), and lumbar puncture results in a simple machine learning method. We will study how clinicians use the report page in a real‐life setting, what barriers and facilitators can be identified for successful implementation, and what requirements must be met to generate a report for all memory clinic patients. In addition, we will further examine outcomes and experiences reported by patients and care partners. In designing the diagnostic report, we aimed to balance the opinions of patients and care partners and what is considered to be advantageous based on research and expertise. For instance, as existing literature suggests that adding visual information to a written report enhances recall, we aimed to display test results visually. , It was difficult to get a consensus on one ideal visualization, as patients and care partners differed in their preferences. Because almost all visual options were rated with high scores, the final choice to use the icon (Figure 2C) as visualization for the diagnostic report in our study was made on the high scoring combined with our expert‐based opinion. Preferably, in the future, diagnostic reports should not be static but should leave room for adaptation to individual patients. Different visualizations could be generated (e.g., a bar chart and an icon array; see, e.g., www.ADappt.health ), and the clinician and patient could decide together which visualization appeals most to the patient and should appear in the printed report page, or the clinician could switch between both. In this study and others, , , we found that patients want to know about their expected symptoms and progression rates. Furthermore, they indicated feeling left in ignorance after receiving the diagnosis. With the ongoing developments in biomarker‐based prognosis for patients with early stages of Alzheimer's disease, there might be a future in which a personalized prediction of progression is possible and could be added to the diagnostic report. , Before we get there, supplying information on coping strategies, practical tips, and tips stimulating brain health can make patients and care partners feel less alone after receiving the diagnosis and enable adequate coping. The strength of our study is the careful and iterative design, based on the proven design thinking methodology, and carried out in co‐creation with a heterogeneous group of patients and care partners. As a result, we optimally understood and depicted their needs and preferences regarding a diagnostic report. The limitation of our study was that the study participants were relatively highly educated. Therefore, we do not yet know whether the diagnostic report is also suited for patients with less education. However, we strove for comprehensibility, even for lower educated or illiterate patients, by using easily understandable visualizations to depict test results.

CONCLUSION

We co‐created a diagnostic report to communicate test results in dementia diagnostics to meet the urgent patient need for more information on their diagnosis. By developing the diagnostic report with patients and care partners, we learned that not one size fits all. It is most important that all essential information is delivered in line with their preferences in a comprehensible manner. Next, we will perform a pilot study in clinical practice to study the usability and feasibility of this prototype of the diagnostic report, thereby strengthening the results of this study.

CONFLICTS OF INTEREST

HRM performs contract research for Combinostics; all funding is paid to her institution. WF performs contract research for Biogen. Research programs of WF have been funded by ZonMW, NWO, EU‐FP7, EU‐JPND, Alzheimer Nederland, CardioVascular Onderzoek Nederland, Health‐Holland, Topsector Life Sciences & Health, stichting Dioraphte, Gieskes‐Strijbis fonds, stichting Equilibrio, Pasman stichting, stichting Alzheimer & Neuropsychiatrie Foundation, Philips, Biogen MA Inc., Novartis‐NL, Life‐MI, AVID, Roche BV, Fujifilm, Combinostics. WF holds the Pasman chair. WF is recipient of ABOARD, which is a public–private partnership receiving funding from ZonMW (#73305095007) and Health‐Holland, Topsector Life Sciences & Health (PPP‐allowance; #LSHM20106). WF has performed contract research for Biogen MA Inc. and Boehringer Ingelheim. WF has been an invited speaker at Boehringer Ingelheim, Biogen MA Inc., Danone, Eisai, WebMD Neurology (Medscape), Springer Healthcare. WF is consultant to Oxford Health Policy Forum CIC, Roche, and Biogen MA Inc. WF participated in advisory boards of Biogen MA Inc. and Roche; all funding is paid to her institution. WF was associate editor of Alzheimer, Research & Therapy in 2020/2021. WF is associate editor at Brain. JG is Executive Director of Alzheimer Europe which receives funding from EU research and health programmes as well as private foundations and companies. All other coauthors report no conflicts of interest.
  35 in total

1.  Three approaches to qualitative content analysis.

Authors:  Hsiu-Fang Hsieh; Sarah E Shannon
Journal:  Qual Health Res       Date:  2005-11

2.  Classification of primary progressive aphasia and its variants.

Authors:  M L Gorno-Tempini; A E Hillis; S Weintraub; A Kertesz; M Mendez; S F Cappa; J M Ogar; J D Rohrer; S Black; B F Boeve; F Manes; N F Dronkers; R Vandenberghe; K Rascovsky; K Patterson; B L Miller; D S Knopman; J R Hodges; M M Mesulam; M Grossman
Journal:  Neurology       Date:  2011-02-16       Impact factor: 9.910

3.  'How do they want to know?' Doctors' perspectives on making and communicating a diagnosis of dementia.

Authors:  Cate Bailey; Jemima Dooley; Rose McCabe
Journal:  Dementia (London)       Date:  2018-04-15

4.  Letters to patients: improving communication in cancer care.

Authors:  D Damian; M H Tattersall
Journal:  Lancet       Date:  1991-10-12       Impact factor: 79.321

5.  Mortality Risk and Its Association with Geriatric Domain Deficits in Older Outpatients: The Amsterdam Ageing Cohort.

Authors:  Hanneke Frederica Maria Rhodius-Meester; Sara A J van de Schraaf; Mike J L Peters; Emma E F Kleipool; Marijke C Trappenburg; Majon Muller
Journal:  Gerontology       Date:  2021-01-13       Impact factor: 5.140

6.  Prospective randomised controlled trial of written supplement to verbal communication of results to patients at the time of flexible cystoscopy.

Authors:  Frank D'Arcy; Chew Lin Yip; Kiran Manya; Paul McGivern; Rustom P Manecksha; Damien Bolton; Shomik Sengupta
Journal:  World J Urol       Date:  2018-02-14       Impact factor: 4.226

7.  Vascular dementia: diagnostic criteria for research studies. Report of the NINDS-AIREN International Workshop.

Authors:  G C Román; T K Tatemichi; T Erkinjuntti; J L Cummings; J C Masdeu; J H Garcia; L Amaducci; J M Orgogozo; A Brun; A Hofman
Journal:  Neurology       Date:  1993-02       Impact factor: 9.910

8.  The experience of providing care in the early stages of dementia: an interpretative phenomenological analysis.

Authors:  Catherine Quinn; Linda Clare; Alison Pearce; Michael van Dijkhuizen
Journal:  Aging Ment Health       Date:  2008-11       Impact factor: 3.658

9.  Biomarker-based prognosis for people with mild cognitive impairment (ABIDE): a modelling study.

Authors:  Ingrid S van Maurik; Stephanie J Vos; Isabelle Bos; Femke H Bouwman; Charlotte E Teunissen; Philip Scheltens; Frederik Barkhof; Lutz Frolich; Johannes Kornhuber; Jens Wiltfang; Wolfgang Maier; Oliver Peters; Eckart Rüther; Flavio Nobili; Giovanni B Frisoni; Luiza Spiru; Yvonne Freund-Levi; Asa K Wallin; Harald Hampel; Hilkka Soininen; Magda Tsolaki; Frans Verhey; Iwona Kłoszewska; Patrizia Mecocci; Bruno Vellas; Simon Lovestone; Samantha Galluzzi; Sanna-Kaisa Herukka; Isabel Santana; Ines Baldeiras; Alexandre de Mendonça; Dina Silva; Gael Chetelat; Stephanie Egret; Sebastian Palmqvist; Oskar Hansson; Pieter Jelle Visser; Johannes Berkhof; Wiesje M van der Flier
Journal:  Lancet Neurol       Date:  2019-09-13       Impact factor: 44.182

Review 10.  Communicating mild cognitive impairment diagnoses with and without amyloid imaging.

Authors:  Joshua D Grill; Liana G Apostolova; Szofia Bullain; Jeffrey M Burns; Chelsea G Cox; Malcolm Dick; Dean Hartley; Claudia Kawas; Sarah Kremen; Jennifer Lingler; Oscar L Lopez; Mark Mapstone; Aimee Pierce; Gil Rabinovici; J Scott Roberts; Seyed Ahmad Sajjadi; Edmond Teng; Jason Karlawish
Journal:  Alzheimers Res Ther       Date:  2017-05-04       Impact factor: 6.982

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.