Literature DB >> 35692231

A simulated experiment to explore robotic dialogue strategies for people with dementia.

Fengpei Yuan1, Amir Sadovnik2, Ran Zhang3, Devin Casenhiser4, Eun Jin Paek4, Xiaopeng Zhao1.   

Abstract

Introduction: Persons with dementia (PwDs) often show symptoms of repetitive questioning, which brings great burdens on caregivers. Conversational robots hold promise of helping cope with PwDs' repetitive behavior. This paper develops an adaptive conversation strategy to answer PwDs' repetitive questions, follow up with new questions to distract PwDs from repetitive behavior, and stimulate their conversation and cognition.
Methods: We propose a general reinforcement learning model to interact with PwDs with repetitive questioning. Q-learning is exploited to learn adaptive conversation strategy (from the perspectives of rate and difficulty level of follow-up questions) for four simulated PwDs. A demonstration is presented using a humanoid robot.
Results: The designed Q-learning model performs better than random action selection model. The RL-based conversation strategy is adaptive to PwDs with different cognitive capabilities and engagement levels. In the demonstration, the robot can answer a user's repetitive questions and further come up with a follow-up question to engage the user in continuous conversations. Conclusions: The designed Q-learning model demonstrates noteworthy effectiveness in adaptive action selection. This may provide some insights towards developing conversational social robots to cope with repetitive questioning by PwDs and increase their quality of life.
© The Author(s) 2022.

Entities:  

Keywords:  Alzheimer’s dementia; Conversational robot; adaptive robot behavior; human-robot interaction; reinforcement learning

Year:  2022        PMID: 35692231      PMCID: PMC9174559          DOI: 10.1177/20556683221105768

Source DB:  PubMed          Journal:  J Rehabil Assist Technol Eng        ISSN: 2055-6683


Introduction

According to the World Alzheimer Report 2018, there were 50 million people living with Alzheimer’s Disease and Alzheimer’s Disease Related Dementias (AD/ADRD) in 2018 around the world, with one new case of dementia every 3 s. Alzheimer’s disease is the most common form of dementia, contributing to 60 − 70% of cases. Due to memory impairment, persons with AD/ADRD (PwDs) often show behavior of repetitive questioning, which can be very frustrating, tedious and exhausting to their caregivers.[3,4] Worse yet, caregivers often do not receive sufficient training on how to communicate with people with dementia and do not have the time to learn the best communication methods to create a strong relationship.[5,6] Especially in senior homes, many PwDs do not receive meaningful daily interactions. Social robots, an emerging common assistive technology to support dementia care,[8-10] could play the role of conversational companion engaging in conversations with PwDs. For example, a very recent late-breaking report proposed proactive robotic listeners to encourage more responses from PwDs. The robot will ask a follow-up question related to the most recent topic or start a topic introduction if a PwD shows silence during conversation. In the situation when a PwD shows repetitive questioning, a conversational robot can be employed to answer those repetitive questions, with the advantages of high repeatability and no complaints and no fatigue. Meaningful daily communication with a robot, as indicated in previous studies, may help reduce symptoms of AD/ADRD and improve quality of life and independence of people with AD/ADRD.[5,6,13-15] It has been found that having a robot to talk to can decrease the feelings of loneliness.[16-18] On the other hand, adaptive robot interactions are necessary to provide a comfortable and effective interactions with target users, i.e. PwDs in our study. An adaptive dialogue system would facilitate meaningful, effective communication and a more trusting relationship between the PwD and robot.[6,13] For example, Rudzicz et al. confirms that the entire communication system will be more effective if the individual’s mental state is taken into account. The adaptive (or autonomous) behaviors in robots have been investigated using the technique of Wizard of Oz (WoZ), where the robot is usually controlled by a human operator.[21,22] However, WoZ has been demonstrated to not be a sustainable technique in long term. WoZ may be sufficient for narrow task domains and very specific user interactions, but it is limited in terms of flexibility and adaptivity to different individual users. Particularly regarding the population of people with AD/ADRD, each individual may have a different personality, preference and cognitive abilities, and show time-varying behaviors, emotions (e.g. behavioral and psychological symptoms of dementia, BPSD), and personality in both short and long term, as well as time-decreasing cognitive capabilities. The limitations of WoZ can be compensated by using adaptive algorithms such as reinforcement learning (RL), which enables robots to learn from the interaction with the environment (e.g. users) and makes it possible to adapt and optimize robotic policies to different individual users. RL allows to integrate developer-defined rewards that better mimic the goal of conversational robot and to model the long-term influence of a generated response in an ongoing dialogue with PwDs. It has been applied in some studies for dialogue management. For example, Cuayáhuitl used deep reinforcement learning to perform action selection from raw text for the context of restaurant. The state space was defined as word-based features, and the action space included 35 dialogue actions in response to users’ intentions. In another study of using RL to learn a conversation strategy for autonomous robotic dialogue system for PwDs, the authors designed state space as the robot’s internal motivation (closely associated with user’s motivation) and previously selected action. Their action space was represented by three types of robot’s action, including short response (simple agreement/encouragement), long response (question) and topic change. The results showed the robot was capable of maintaining conversations with seniors for at least 20 min. Regarding the specific context of repetitive questioning in PwDs, the conversational robot would be expected to not only answer those repetitive questions for PwDs, but also further communicate with PwDs to distract their attention from the repetitive behaviors and stimulate their cognitive activities. Asking appropriate questions proactively by the robot can be a good approach to start conversations with PwDs. However, the question difficulty level (e.g. closed-ended vs open-ended questions) must be taken into careful consideration during communication with PwDs. Inspired by previous relevant research,[32,33] asking questions with optimal difficulty level with respect to PwD’s cognitive capability may help engage PwD and keep them interested in interacting with the robot, and maximally stimulate their cognitive activities. Our long-term goal is to build a cost-effective and adaptive conversational robot, as a complement of the caregivers, to cope with repetitive questioning in PwDs by answering their questions and following up with new questions to distract them from the repetitive behaviors. Inspired by previous studies,[32,33] asking questions with optimal difficulty level geared to a PwD’s cognitive capability may keep them engaged and interested in interacting with the robot, and consequently, stimulate their cognitive activities to the most extent. In this paper, we propose a general RL framework for this problem (i.e. repetitive questioning), to explore an adaptive conversation strategy for a robot by tuning the frequency and difficulty level of asking a follow-up question when communicating with PwDs. To the best of our knowledge, this work is the first to learn adaptive conversation strategy for social robots to specifically cope with PwDs’ repetitive questioning. The strategy is adaptive to PwDs with different cognitive capabilities and engagement levels. We expect this general framework would serve as a useful benchmark on the design of the adaptive conversation strategy for the robots when additional sensing modalities are available. The remainder of the paper is organized as follows. The Section Method introduces the methodology, including the definition of Markov Decision Process, the modelling of the persons with AD/ADRD, and experiments. The simulation results and a demonstration are presented in the Section Results and discussed in the section of Discussion. A discussion of limitations in this study as well as the future work are listed in Section Limitation and future work. The last section concludes the paper.

Method

The technique of reinforcement learning is used to learn from the PwD-robot conversation (Figure 1) and investigate the optimal policy for a robot. To offset the stress of PwD’s repetitively questioning on a caregiver, the robot should be able to always answer the questions asked by the PwD, propose some follow-up questions to distract PwD from the repetitive behaviors, and also to stimulate their cognitive activities. However, if a follow-up question is too challenging, difficult or complicated to the PwD, the PwD may not respond to it at all. Therefore, the robot should also be able to identify if the follow-up question is too difficult for the PwD, and adapt the difficulty level to users with different cognitive capabilities. Therefore, the optimal policy maps the PwD-robot interaction/dialogue to the robot’s follow-up rate and difficulty level of follow-up question, so that there will be frequent conversation between the PwD and the robot, together with the brain activities in people with AD/ADRD being maximally stimulated.
Figure 1.

A schematic framework demonstrates the adaptive strategy of PwD-robot dialogue.

A schematic framework demonstrates the adaptive strategy of PwD-robot dialogue.

Definition of reinforcement learning

According to the aforementioned goals for PwD-robot dialogue, we define the key elements of Markov decision process (MDP) model as follows : where the variable QuestionDifficulty can be 1, 2, or 3, separately corresponding to easy, moderately difficult, and difficult follow-up questions. The expression Q, NF, RR, IR, and NR represent the possible states in the MDP model. The symbol → indicates the state transition. Although meaning conversation (i.e. relevant responses from PwD to robot’s questions) is the most suggested, irrelevant responses are still meaningful to PwDs. Thus, a slightly positive reward (i.e. +0.5) was assigned to Q → IR. • State space: A state is defined according to a user’s situation during a conversational context. There are five potential states: the user asking a question, denoted by Q or Question, the user’s question being simply answered by robot without following-up, denoted by NF, the user providing relevant response to the robot’s follow-up question, denoted by RR, the user providing irrelevant response to the robot’s follow-up question, denoted by IR, and the user providing no response to robot’s follow-up question, denoted by NR. Noticeably, underlying the states RR, IR, and NR, there are always processes of the user’s question being answered by the robot and further the robot asking a follow-up question. • Action: An action includes two elements: the follow-up rate (i.e. the probability of the robot asking a question after answering PwD’s question) for the robot and the difficulty level of a follow-up question. For simplification, currently the optional follow-up rate is 0.1, 0.4, 0.7 or 1.0. And there are three question difficulty levels: easy, moderately difficult, and difficult. Table 1 shows examples of follow-up questions with different difficulty levels.
Table 1.

Examples of follow-up questions by a robot.

Question difficultyExample 
Easy“Would you like some tea?”     
Moderate“What would you like to drink?”     
Difficult“What do you think about this tea?”
• Reward: According to our goals for a social robot in the situation of repetitive questioning, i.e. answering PwDs’ repetitive questions, asking follow-up questions to distract PwDs from repetitive questioning and to promote daily conversation, and avoiding too difficult follow-up questions towards PwDs, we build our immediate reward function as following, Examples of follow-up questions by a robot. The PwD-robot dialogue is a complicated interactive task, which can be continuing (i.e. the left diagram in Figure 2) or episodic tasks. In this paper, we start our simulated exploration study with the most simple modelling of situation, one-step episodic task, as illustrated in the right diagram of Figure 2. Each episodic task starts by a PwD asking a repetitive question, that is, the state s = Q, and ends with a terminal state, i.e. s = NF, RR, IR, or NR.
Figure 2.

The Markov decision process diagram in one episode of PwD-robot dialogue interaction.

The Markov decision process diagram in one episode of PwD-robot dialogue interaction.

Simulated person with Alzheimer’s disease and Alzheimer’s disease related dementias

Because the real-world samples of PwD-robot conversational interaction is costly in terms of time and labor and might be related to ethical issues, we start from simulated PwD-robot dialogue using RL modelling, which hopefully provides us insight into real-world PwD-robot dialogue in next step. A simulated individual with AD/ADRD during the PwD-robot conversation is characterized by their response to the robot’s follow-up question. The response can be categorized into three types of response: relevant, irrelevant, and no response to the follow-up question. Therefore, an individual is characterized by the relevant response rate P and irrelevant response rate P . Notice here the rate of no response P is dependent on P and P . The sum of these three variables is always 1. We assume that an individual’s relevant and irrelevant response rates are influenced by the individual cognitive capability and their engagement, as well as the question difficulty level. Engagement of a person with AD/ADRD is defined as the act of being occupied or involved with an external stimulus, i.e. the conversational robot in our case. The engagement of an individual with AD/ADRD can be influenced by robot attributes and the individual’s attributes (e.g. personality and preference). There are three basic rules to create models of PwDs. Firstly, with the same engagement level, a person with a lower cognitive capability will show a lower relevant response probability. Also, we expect a person with lower cognitive capability will have a higher irrelevant response probability except the most severe PwD who will have extremely low response (both relevant and irrelevant) to all questions. Secondly, we assume that an individual with a lower engagement level will have a lower response rate (i.e. the sum of P and P ) and thus a greater no response rate, P . The specific effect on P and P is dependent on individuals. Therefore, the effects on individual response rate P and P are random in our simulation. Thirdly, given a more difficult follow-up question, the same user with the same engagement is expected to show a lower relevant response rate P and a higher no response rate P . Naturally, it takes people with AD/ADRD more cognitive workload to answer a more difficult follow-up question. For example, PwDs only need to answer “yes” or “no” to the easy question in Table 1. Comparatively, to answer the difficult question in this table, PwDs need to think more to understand the question and express their opinions. During our simulation, there are four basic simulated users, each with three different levels of engagement, i.e. high, medium, and low. The parameters of the simulated User 1 − 4 with different engagement levels are listed in Tables 2–5. User 1, 2, 3 and 4, correspond to older adults without cognitive impairment, with mild cognitive impairment, moderate dementia, and severe dementia, respectively.
Table 2.

Parameters of simulated User 1 without cognitive impairment.

EngagementQuestion difficulty P Rresp P IRresp P Nresp
HighEasy100
Moderate100
Difficult100
MediumEasy0.9500.05
Moderate0.9200.08
Difficult0.9000.10
LowEasy0.9000.10
Moderate0.8800.12
Difficult0.8500.15
Table 5.

Parameters of simulated User 4 with severe dementia.

EngagementQuestion difficulty P Rresp P IRresp P Nresp
HighEasy0.040.080.88
Moderate0.020.080.90
Difficult0.010.040.95
MediumEasy0.020.050.93
Moderate0.010.040.95
Difficult0.0050.020.975
LowEasy0.010.040.95
Moderate00.020.98
Difficult00.010.99
Parameters of simulated User 1 without cognitive impairment. Parameters of simulated User 2 with mild cognitive impairment. Parameters of simulated User 3 with moderate dementia. Parameters of simulated User 4 with severe dementia.

Experiments

On the basis of the aforementioned definition for the three key elements (state, action, and reward function) of MDP, we train our reinforcement learning model for our four simulated users with three different engagement levels. We apply the technique of Off-policy Q-learning to solve the MDP problem, with the ϵ-greedy policy (ϵ = 0.1), constant learning rate α = 0.03, and discount factor, γ = 0.95. An episode starts by a PwD asking a repetitive question and ends with a terminal state (NF, RR, IR, or NR), as illustrated on the right of Figure 2. We evaluate the performance of our model with the metrics proposed by, including the average return per epoch, the starting state value V (Question) in each epoch, and the sum of Q-value updates during each epoch. Here, due to our definition of one-step episodic task, we only need to consider the starting state value in the starting state, Question (i.e. the PwD asking repetitive question). The start state value V (Question) indicates the expected return. Each epoch is composed of 150 episodes. Furthermore, we analyze the performance of Q-learning by comparing the learning results using randomized action selection.

Results

Based on the evaluation metrics, we observe that in all cases (i.e. User 1 − 4 with different engagement level in Tables 2–5) our RL model converges within 30 epochs. Therefore, we only show the learning processes of the model during epoch 1 − 100 in Figure 3. In this figure, the first, second, and third column correspond to the learning process (i.e. average return) of all users with high, medium, and low engagement, respectively. The solid and dashed curves with the same color in each sub-figure represent the learning results for the same user using Q-learning and random action selection, respectively. A user with a higher engagement level is associated with greater average return, compared to the same user with lower engagement level. With the same engagement level, the converged average return in User 1 and 2 is obviously greater than that in User 3 and 4. The optimal policies learned by the RL agent for different users with different engagement levels are listed in Table 6.
Figure 3.

The learning results of average return for User 1 − 4 with high (first column), medium (second column), and low (third column) engagement. In each sub-figure, the solid and dashed curves represent the learning results by Q-learning and random action selection model, respectively.

Table 6.

Optimal policy suggested by the Q-learning.

EngagementUser 1User 2User 3User 4
High[1.0, D][1.0, D][1.0, D] or [1.0, M][0.1, E]
Medium[1.0, D][1.0, D][1.0, M][0.1, E]
Low[1.0, D][1.0, D] or [1.0, M][1.0, E][0.1, E]

Note User 1, 2, 3 and 4 represent a person without cognitive impairment, with mild cognitive impairment, moderate dementia and severe dementia. D = difficult follow-up questions; M = moderately difficult follow-up questions; E = easy follow-up questions.

The learning results of average return for User 1 − 4 with high (first column), medium (second column), and low (third column) engagement. In each sub-figure, the solid and dashed curves represent the learning results by Q-learning and random action selection model, respectively. Optimal policy suggested by the Q-learning. Note User 1, 2, 3 and 4 represent a person without cognitive impairment, with mild cognitive impairment, moderate dementia and severe dementia. D = difficult follow-up questions; M = moderately difficult follow-up questions; E = easy follow-up questions.

Demonstration of the dialogue policy

For purpose of demonstration, we implement the optimal dialogue policies obtained in Table 6 using a humanoid robot, Pepper, and show the interaction between Pepper and a participant to simulate scenarios of repetitive questioning. Here, a researcher plays the role of a PwD with repetitive questioning behaviors. Pepper is 1.2-m tall and has 17 joints to support expressive and appealing gestures and other body movements. Thanks to its attractive appearance and capability of multimodal interaction (e.g. verbal communication, body movement, and eye contact), the Pepper robot is perceived to be acceptable, appealing, and engaging to human users in daily activities.[38-40] Pepper is equipped with a tablet in its chest, which can be used to display dialogues and pictures. In addition, the robot has a range of sensors, which can be used to evaluate the human user’s response relevance, engagement level, and emotions. We use Pepper’s software development kit (SDK), QiSDK, to implement the conversational policy (as listed in Table 6) so that Pepper is able to respond to the repetitive questions as well as to come up with follow-up questions to distract users from their repetitive questioning behaviors. The robot’s body movement, gestures and eye contact are used to support the interaction and engage the human user in the interaction. For example, during the conversation, the robot Pepper will always make eye contact with the human user. All the verbal conversation between the user and the robot is displayed in Pepper’s tablet. A list of repetitive questions are incorporated into the robot using the Application Programming Interface (API), QiChatbot and topic, through which we can define how the robot will verbally respond to a human user’s verbal input (e.g. repetitive questions) and come up with a follow-up question for the user. Figure 4 shows a screenshot of the verbal interaction between the user and the robot Pepper. The robot is coded with the policy of [1.0, M] in Table 6. During the scenario, when the robot Pepper detects a repetitive question (e.g. “What time is the party?”) from the user, the robot first answers the question with the pre-defined answer (e.g. “It is at 7p.m.”) and then comes up with a follow-up question for the user.
Figure 4.

Verbal interaction between a human user (text in green) and the social robot Pepper (text in pink) in a scenario of repetitive questioning.

Verbal interaction between a human user (text in green) and the social robot Pepper (text in pink) in a scenario of repetitive questioning.

Discussion

From Figure 3, we can see that the implemented RL agent is able to learn the best policy within 30 epochs. The solid average return curves for all users learned by Q-learning are greater than dashed curves learned by random action selection, which indicates that Q-learning here is helpful for action selection. The learning processes here converge with spikes, which is due to the simulation of stochastic response rate (Tables 2–5). In Figure 3, as the learning curves converge, the average return for an individual with higher cognitive capability (e.g. User 1 and 2) are greater than an individual with lower cognitive capability (e.g. User 1 and 2). This makes sense considering that an individual without or with mild cognitive impairment is more likely to answer a question. Comparing the optimal policies learned by our RL agent for four types of user with the same engagement level, for example, the row of medium engagement, the agent is able to learn the best policy for PwDs with different cognitive capabilities. More specifically, the optimal policy (i.e. follow-up rate and question difficulty) for a user with higher cognitive capability (e.g. User 1 and 2) is always asking the difficult follow-up question. However, towards users with lower cognitive capability, User 3 and 4, the RL agent separately adapts the policy to always following up with moderate-level difficult question, and following up very rarely also with easy questions. Comparing the learning results for the same user but with different engagement level, for example, the green curve representing User 3 in the three sub-figures in Figure 3, the average return obtained per epoch decreases as the user’s engagement decreases, which makes sense because an individual with lower engagement is expected to less likely join activities and conversations. Moreover, the column of User 3 in Table 6 shows that the question difficulty needs to decrease as an adaptation to individual’s decreasing engagement level. This indicates that, although merely observing relevance of user’s response, the RL agent is able to detect the change of latent variable (i.e. user’s engagement level) and adaptively adjust the question difficulty level accordingly.

Limitation and future work

There are some limitations in the study. First, a state in our MDP only considers the relevance of user’s response to a robot’s question. Although the current MDP seems to adapt the policy when the latent variables (e.g. user’s cognitive capability and engagement level) changes, the inclusion of more variables associated with users, e.g. PwD’s cognitive and affective states, may facilitate better learning performance. Previous studies has showed that a PwD’s engagement level can be read using sensing technologies such as camera, heart-rate sensor, and passive brain-computer interfaces (BCIs, e.g. electroencephalography). In the future, we will take into consideration one or some PwD’s cognitive and affective states as additional dimension(s) of state space, by integrating sensing technologies in a cost-effective way (e.g. in terms of computational complexity). Second, in the case of demonstration, we manually coded the optimal policy suggested by RL as well as the list of follow-up questions in the robot, which is an impediment to the development of adaptive conversational robot from long term. In the future, we will integrate the robot Pepper and the RL model into one framework, where the input will be raw data of user’s states and the output will be Pepper automatically performing an action suggested by RL. We will also work on the list of follow-up questions, which could be appropriate to users’ environment and social context. Third, it is simplified that an individual’s cognitive capability and engagement level is consistent during the whole PwD-robot dialogue, which is usually not true in real world considering the intra-individual variability and disease progression in people with AD/ADRD. In future, we will conduct more research on this direction. We will collect real-world repetitive questions from PwDs who show time-varying cognitive capability and time-varying engagement in either the short and long term and adapt our RL model accordingly. Then we will investigate the performance of the RL model for those PwDs in both the short and long term in real world.

Conclusions

In this paper, we have developed a general RL framework, to learn the adaptive conversation strategy for a robot to cope with the problem of repetitive questioning by persons with dementia. The model allows the robot’s dialogue system to adjust an appropriate rate of asking a follow-up question and the question difficulty level, given the robot only perceives PwDs’ spoken conversation (e.g. repeating a question and relevance of response). We have also presented a demonstration of the optimal dialogue policy, where a physically embodied social robot interacts with a human user with repetitive questioning behaviors. During the scenario, the robot is able to respond to the user’s repetitive questions with appropriate answers and further come up with a follow-up question. This study may allow a conversational social robot to help caregivers with PwDs’ repetitive questioning behaviors. Moreover, the design of follow-up question may also distract PwDs from repetitive behaviors and stimulate PwDs’ brain activities through the conversation with different difficulty level. In the future, the MDP model will be improved by leveraging the multimodal sensing technologies. And more work is needed to train the RL model with users who has time-decreasing cognitive capabilities and time-varying engagement level. Also, future study is needed to integrate the RL agent with the robot Pepper and investigate the performance with PwDs in real world.
Table 3.

Parameters of simulated User 2 with mild cognitive impairment.

EngagementQuestion difficulty P Rresp P IRresp P Nresp
HighEasy0.90.10
Moderate0.860.140
Difficult0.820.180
MediumEasy0.830.110.06
Moderate0.750.150.10
Difficult0.680.200.12
LowEasy0.750.140.11
Moderate0.650.160.19
Difficult0.500.180.32
Table 4.

Parameters of simulated User 3 with moderate dementia.

EngagementQuestion difficulty P Rresp P IRresp P Nresp
HighEasy0.700.200.10
Moderate0.630.220.15
Difficult0.500.230.27
MediumEasy0.600.210.19
Moderate0.500.250.25
Difficult0.300.200.50
LowEasy0.350.150.50
Moderate0.200.130.67
Difficult0.080.100.82
  14 in total

1.  Robots to assist daily activities: views of older adults with Alzheimer's disease and their caregivers.

Authors:  Rosalie H Wang; Aishwarya Sudhama; Momotaz Begum; Rajibul Huq; Alex Mihailidis
Journal:  Int Psychogeriatr       Date:  2016-09-23       Impact factor: 3.878

2.  Engagement in persons with dementia: the concept and its measurement.

Authors:  Jiska Cohen-Mansfield; Maha Dakheel-Ali; Marcia S Marx
Journal:  Am J Geriatr Psychiatry       Date:  2009-04       Impact factor: 4.105

3.  Can We Talk through a Robot As if Face-to-Face? Long-Term Fieldwork Using Teleoperated Robot for Seniors with Alzheimer's Disease.

Authors:  Kaiko Kuwamura; Shuichi Nishio; Shinichi Sato
Journal:  Front Psychol       Date:  2016-07-19

4.  Exploration of verbal repetition in people with dementia using an online symptom-tracking tool.

Authors:  Emily Reeve; Pierre Molin; Amaris Hui; Kenneth Rockwood
Journal:  Int Psychogeriatr       Date:  2017-03-09       Impact factor: 3.878

5.  The associations between behavioral-psychological symptoms of dementia (BPSD) and coping strategy, burden of care and personality style among low-income caregivers of patients with dementia.

Authors:  Aishah Diyana Baharudin; Normah Che Din; Ponnusamy Subramaniam; Rosdinom Razali
Journal:  BMC Public Health       Date:  2019-06-13       Impact factor: 3.295

6.  Reducing Patient Loneliness With Artificial Agents: Design Insights From Evolutionary Neuropsychiatry.

Authors:  Kate Loveys; Gregory Fricchione; Kavitha Kolappa; Mark Sagar; Elizabeth Broadbent
Journal:  J Med Internet Res       Date:  2019-07-08       Impact factor: 5.428

7.  Effects of a Humanoid Socially Assistive Robot Versus Tablet Training on Psychosocial and Physical Outcomes of Persons With Dementia: Protocol for a Mixed Methods Study.

Authors:  Sandra Schüssler; Julia Zuschnegg; Lucas Paletta; Maria Fellner; Gerald Lodron; Josef Steiner; Sandra Pansy-Resch; Lara Lammer; Dimitrios Prodromou; Sebastian Brunsch; Magdalena Holter; Lorenzo Carnevale; Silvia Russegger
Journal:  JMIR Res Protoc       Date:  2020-02-03

Review 8.  Passive Brain-Computer Interfaces for Enhanced Human-Robot Interaction.

Authors:  Maryam Alimardani; Kazuo Hiraki
Journal:  Front Robot AI       Date:  2020-10-02

9.  Repetitive Questioning Exasperates Caregivers.

Authors:  R C Hamdy; J V Lewis; R Copeland; A Depelteau; A Kinser; T Kendall-Wilson; K Whalen
Journal:  Gerontol Geriatr Med       Date:  2018-01-09
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.