| Literature DB >> 35468084 |
Roel Boumans1, Yana van de Sande1, Serge Thill2, Tibor Bosse1.
Abstract
BACKGROUND: Older adults often have increasing memory problems (amnesia), and approximately 50 million people worldwide have dementia. This syndrome gradually affects a patient over a period of 10-20 years. Intelligent virtual agents may support people with amnesia.Entities:
Keywords: Alzheimer; amnesia; dementia; intelligent virtual agent; mobile phone; systematic review
Year: 2022 PMID: 35468084 PMCID: PMC9086881 DOI: 10.2196/32473
Source DB: PubMed Journal: JMIR Aging ISSN: 2561-7605
Figure 1A virtual agent under development at the Behavioural Science Institute.
Figure 2Study flow.
Selected studies.
| Authors and title | Study design | Experiment population | Intervention | Human-agent interaction | Results | Agent image | |||
| Ali R et al [ | Country: United States; study design: pilot randomized controlled trial; N=18; intervention and control group; study aim: to assess feasibility and acceptability of the VAa | N=18, 1 lost to follow-up; aged >60 years; having self-reported mild communication difficulties with social skills that could be attributable to memory problems | In the intervention group, the VA had weekly sessions with each participant, including three 2- to 3-minute open dialogues with the participant on a selection of general topics (eg, weather, pets, retirement, life goals, growing older, and spirituality). The control group was provided with videos. | The VA was a web-based application using computer system including camera and microphone from the participant’s home to record participant utterances and expressions. Open dialogue included ASRb and TTSc. The ASR used a hierarchical tree to classify participant utterances and determine responses. | The system usability score was 69.5 on a 0-100 scale, where 68 was considered as “good” usability. Participants randomized to the VA demonstrated significantly fewer impairments in nonverbal communication at follow-up compared with the control group, whereas the results were nonsignificant for participants with verbal impairment. |
| |||
| Razavi et al [ | Country: United States; N=8. Evaluative study aimed at older adults where each participant had a 10- to 20-minute conversation with the VA. The system was designed for geriatric patients and with input from gerontologists. Study aim: examination of conversation quality. | N=8; older adults. No details given. | Have a short conversation on several topics, categorized as easy, medium, and hard. They covered 30 themes among which were hobbies, weather, cooking, life goals, and spirituality. Average duration was a few minutes with 3-5 turns; no details given. | A 2-way interaction; VA asks question or reacts on participant response [ | Participants were asked to score 4 variables on a 5-point scale from strongly disagree (1) to strongly agree (5). Ease of use was scored as 4.3, learnability as 3.9, confidence in using as 4.3, and user-friendliness as 4.6. |
| |||
| Wargnier et al [ | Country: France; feasibility study; N=14; no control group. Study aim was usability assessment of the LOUISE system. | N=14; aged >65 years; diagnosed with cognitive impairment | VA asked participant to perform 4 tasks: drink water, take a pill, measure blood pressure, and select meal. Participants could choose between 2 VA embodiments: “Louise” (left image) and “Charlotte” (right image). | The VA questions were provided verbally. Participants could verbally answer with “yes” or “no.” Microsoft Speech ASR was used. | All but one participant could interact with LOUISE. Of the 14 participants, 11 completed the 4 scenarios, but from these 11 situations, only 1 was conducted “in WoZ mode,” and 1 showed sensor failures. Thus, ultimately data from 9 participants were available. Participants often forgot they could only say “yes” or “no.” ASR error rate was 20%. |
| |||
| Tokunaga et al [ | Country: Japan; feasibility study; N=11; no controls. | N=11; 9 women; older inhabitants of the daycare center | VA conducted greeting, confirmation of basic personal information, quiz, and playing music. Measurement was done using a 10-question usability score between 1 and 4. Design of the avatar was similar to Tokunaga et al [ | The VA makes statements and asks questions. The paper does not provide details on the user input types. No description of NLPf function was given. | Average usability score was 3.58. Participants found playing music as especially useful. |
| |||
| Tokunaga et al [ | Country: Japan; exploratory study; N=17. Study aim was to confirm that the patients could interact with the agent service using some interactions (eg, voice or touch). | N=17; older adult patients, aged 46-84 years, 12 women. Mean MMSEg 22.9, meaning some cognitive impairment. | A nondetailed scenario in which participant had to perform certain tasks upon verbal instruction of the VA. The participant could respond by voice or by touch button. | VA asks questions, and the participant reacts. No details were provided. NLP characteristics were unclear. | Participants did not always hear the agent because of hearing impairments or microphone quality. Touch button operation was difficult for the older adults not accustomed to smartphones or tablets. VA did hear the patient only after a second utterance. Patients were sometimes surprised and did not know what to do if the system did not react as expected. |
| |||
| Tsiourti et al [ | Countries: Switzerland, Portugal, and the Netherlands; N=20; design: longitudinal evaluation study; goals: to examine empirically interaction with ECA at home and explore ECA acceptance, perceived usability, and usefulness. | A total of 24 older adults living at home with average age of 77.9 years in 2 countries: the Netherlands (N=11) and Switzerland (N=13). In Netherlands, the number of dropouts was 4. | In the Netherlands, researchers visited participants for joint sessions with the VA 2-3 times a week. System user options were tried in no specific order or method but included reminders and memory programs. Detailed use scenarios on Switzerland were unclear, participants seemed to use the system autonomously. | The users interacted with the companion using a multimodal interface including automatic speech recognition and a graphical touch-based user interface menu (messages and agenda). ASR used Kinect for Windows SDKh to perform speech recognition for predefined speech commands that the users had to remember. | Empirical findings were problems with speech recognitions, remembering of user interaction options by participants, and the nonintuitiveness of the user interface. Acceptance was “well received.” Usability was 62.2 for participants from Switzerland and 52 for those from the Netherlands on an unspecified scale. Usefulness in participants from the Netherlands was not reported and in participants from Switzerland as 2.3-2.5 on a scale of 0 to 5. |
| |||
| Jegundo et al [ | Country: Portugal; observational study; N=46. | Target group was older adults needing formal care. Convenience sample of older adults in daycare centers; 34 women, 12 men; mean age 63.6 (SD 20.5) years. | All interactions were by touch; only interaction on “News” was by voice. Sessions were all witnessed by researchers. | VA responded verbally by “News” commands (“Open news” or “Read news”). All other interactions were through touch buttons (play game or show agenda). | CaMeLii presents a good degree of usefulness, satisfaction, and ease of use. CaMeLi was a barrier to 11 participants and a facilitator for 35 participants. |
| |||
| Oliveira et al [ | Country: Portugal; design: quiz game with VA as host; n=21; no control group; study aim: feasibility | A total of 2 groups: “Young” (aged 24-28 years; n=4) and “Elder” (aged 59-88 years; n=17); divided into 2 subgroups: “tested at home” (n=9) and “tested at the senior university” (n=8). Target group comprised people with dementia, but cognitive status of participating older adults was not described. | Participants had to answer quiz questions by a web-based quiz host. Answers could be open answers. | ASR and TTS were developed at the university laboratory. A 2-way spoken question and answer was used. VA used was a passive cartoon. | Participants enjoyed the game and were stimulated to interact with each other. |
| |||
aVA: virtual agent.
bASR: automated speech recognition.
cTTS: text-to-speech.
dLOUISE: Lovely User Interface for Servicing Elders.
eECA: embodied conversational agent.
fNLP: natural language processing.
gMMSE: Mini Mental State Examination.
hSDK: software development kit.
iCaMeLi: Care Me for Life.
Scores of methodological quality assessment of the included studies.
|
| Ali et al [ | Razavi et al [ | Wargnier et al [ | Tokunaga et al [ | Tokunaga et al [ | Tsiourti et al [ | Jegundo et al [ | Oliveira et al [ |
| Objectives | 2 | 1.5 | 1 | 1 | 1 | 1 | 1.5 | 1.5 |
| Study design | 1.5 | 1 | 1.5 | 1.5 | 1.5 | 2 | 2 | 1.5 |
| Method | 2 | 1.5 | 1 | 0.5 | 0.5 | 1 | 1.5 | 1.5 |
| Participants | 2 | 0.5 | 1.5 | 1 | 1.5 | 1 | 1.5 | 1.5 |
| Random allocation | 1 | N/Aa | N/A | N/A | N/A | N/A | N/A | N/A |
| Blinding investigators | 2 | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| Blinding participants | 0 | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| Outcomes | 1.5 | 2 | 0.5 | 2 | 1 | 1.5 | 1.5 | 0.5 |
| Sample size | 1 | 1.5 | 1.5 | 1 | 1.5 | 1.5 | 1.5 | 1.5 |
| Analytic methods | 1 | N/A | N/A | N/A | 1.5 | N/A | 1.5 | 2 |
| Variance estimates | 2 | 2 | 1 | 1 | N/A | N/A | 2 | 1 |
| Confounding controls | 1.5 | N/A | N/A | N/A | N/A | N/A | N/A | N/A |
| Results reporting | 1.5 | 2 | 1.5 | 2 | 2 | 1 | 1.5 | 1 |
| Conclusions | 1.5 | 1.5 | 1 | 1 | 1 | 1.5 | 2 | 1.5 |
| Total | 20.5 | 13.5 | 10.5 | 11 | 11.5 | 10.5 | 16.5 | 13.5 |
| Maximum score | 28 | 18 | 18 | 18 | 18 | 16 | 20 | 20 |
| Summary score | 0.73 | 0.75 | 0.58 | 0.61 | 0.64 | 0.66 | 0.83 | 0.68 |
aN/A: not applicable.
Comparison between usability scores.
| Reference | Usability |
| Ali et al [ | Mean 4.17 (SD 0.68) on a scale of 1 to 6, where 1=awful, 2=poor, 3=okay, 4=good, 5=excellent, and 6=best imaginable |
| Razavi et al [ | Mean 4.33 (SD 0.67) on a scale of 1 to 5 (“strongly agree” to “strongly disagree”) |
| Wargnier et al [ | Pleasantness: mean 3.38 (SD 0.43); ease of following instructions: mean 3.38 (SD 0.47) on a scale of 1 to 4 |
| Tokunaga et al [ | Experimental questionnaire: 1=lowest, 4=highest; mean 3.58 (SD not given) |
| Tokunaga et al [ | No quantitative usability data |
| Tsiourti et al [ | Usability: mean 62.2 for Switzerland and mean 52 for the Netherlands on an unspecified scale (SD not provided) |
| Jegundo et al [ | USEa questionnaire from Lund [ |
| Oliveira et al [ | No quantitative usability data |
aUSE: Usefulness, Satisfaction, and Ease of Use.