Literature DB >> 34254385

Clients and professionals elicit long-term care preferences by using 'What matters to me': A process evaluation in the Netherlands.

Catharina M van Leersum1, Albine Moser1,2, Ben van Steenkiste1, Judith R L M Wolf3, Trudy van der Weijden1.   

Abstract

BACKGROUND: 'What matters to me' is a five-category preference elicitation tool to assist clients and professionals in choosing long-term care. This study aimed to evaluate the use of and experiences with this tool.
METHODS: A mixed-method process evaluation was applied. Participants were 71 clients or relatives, and 12 professionals. They were all involved in decision-making on long-term care. Data collection comprised online user activity logs (N = 71), questionnaires (N = 38) and interviews (N = 20). Descriptive statistics was used for quantitative data, and a thematic analysis for qualitative data.
RESULTS: Sixty-nine per cent of participants completed one or more categories in an average time of 6.9 (±0.03) minutes. The tool was rated 6.63 (±0.88) of 7 in the Post-Study System Usability Questionnaire (PSSUQ). Ninety-five per cent experienced the tool as useful in practice. Suggestions for improvement included a separate version for relatives and a non-digital version. Although professionals thought the potentially extended consultation time could be problematic, all participants would recommend the tool to others.
CONCLUSION: 'What matters to me' seems useful to assist clients and professionals with preference elicitation in long-term care. Evaluation of the impact on consultations between clients and professionals by using 'What matters to me' is needed.
© 2021 The Authors. Health and Social Care in the Community published by John Wiley & Sons Ltd.

Entities:  

Keywords:  decision support technique; long-term care; patient preferences; preference elicitation; process evaluation

Mesh:

Year:  2021        PMID: 34254385      PMCID: PMC9291068          DOI: 10.1111/hsc.13509

Source DB:  PubMed          Journal:  Health Soc Care Community        ISSN: 0966-0410


What is known about this topic

Preference elicitation is beneficial in the decision‐making in long‐term care. Preference elicitation tools can make the decision‐making easier for clients and professionals. A process evaluation is a supportive step during implementation of innovations by showing the use and the experiences with a tool.

What this paper adds

A process evaluation showing the first use and experiences with the tool ‘What matters to me’. ‘What matters to me’ is supportive in real decision‐making practice for long‐term care. Clients and professionals would recommend the tool to others because they experience support in clarifying preferences for life and care and in discussing these during consultations.

BACKGROUND

The growing demand for long‐term care causes changes in the setting and delivery of healthcare (Christensen et al., 2009; Ward & Schiller, 2010). In the Netherlands, the Dutch health policy aims to enable people to stay in their home rather than in institutional care. The increasing number of older people and the prevalence of chronic disease have caused this growing demand for long‐term care, and the Dutch health policy caused the growing demand for care at home. Besides the economic burden of long‐term care, this means an increase in the workload of healthcare professionals (Reiners et al., 2019). When professionals discuss options of care with their client, their intention is to take the client's situation into account in a collaborative manner (Elwyn et al., 2017). Discussing and considering a client's preferences, and choosing suitable long‐term care, whether someone is able to stay in their home or whether someone would rather be in institutional care. Shared decision‐making (SDM) is an approach whereby clients who are faced with a decision receive information from their professionals on which to base their decision (Elwyn et al., 2012). SDM starts with the ‘team talk’. The client is invited to cooperate in the decision‐making process, options are explained, support is given and information about goals is elicited. The second stage – the ‘option talk’ – covers possible options and their harms and benefits. Lastly comes the ‘decision talk’ at which time preference‐based decisions are made (Elwyn et al., 2017). SDM is important in order to be able to understand the client's situation and to decide on the best course of action (Wieringa et al., 2019). To provide personalised care, the SDM dialogue should include preferences in order to discover those things that matter most to clients while taking the options into account (Jesus & Silva, 2016). These things play an important role in the decision‐making process in which the views and preferences of professionals and clients are elicited and discussed (Spatz et al., 2017). Clients in need of long‐term care might experience challenges in articulating preferences. Personal preferences are care or health related as well as daily life related, such as family, activities or finances. Without the assistance of preference elicitation, decision‐making could be difficult for these clients (Emmanuel & Emmanuel, 1992; van Leersum et al., 2019). Attempts to address the demand for long‐term care benefit from the application of digital technology (Krick et al., 2019; Ostherr et al., 2016). Digital technology is regarded as improving client centeredness, self‐management, engagement in consultations, quality of care and as lowering the costs of healthcare (Arcury et al., 2020; Barello et al., 2015; Granja et al., 2018). It can make clients feel secure and in control, while simultaneously appreciating the immediate professional responses and peer support during decision‐making (Karisalmi et al., 2019). To assist preference elicitation, digital technology can improve SDM by addressing needs and preferences (Ostherr et al., 2016). This also supports professionals in understanding their clients and building a relationship (Stans et al., 2019). Currently, digital technologies to assist preference elicitation are investigated and under development (Granja et al., 2018). Most of the technologies ensuring a preference‐guided decision are single‐disease oriented (Gray et al., 2014; Jayadevappa et al., 2015; Kaiser et al. 2015). These tools are designed to evaluate goals, preferences, capabilities, values or wishes in health and life. The model of positive health, for example, aims to visualise someone's state of health by means of physical and mental functioning, spirituality, quality of life, social participation and daily functioning (Huber et al., 2016). Another tool, the Outcome Prioritization Tool, supports decision‐making for treatment. The tool facilitates a talk and prioritisation of preferences for a specific outcome. This results in enhanced engagement of the client and a deepening of the relationship (van Summeren et al., 2017). The digital preference elicitation tool ‘What matters to me’ is designed to elicit preferences for long‐term care. Clients in long‐term care face the challenge that they need support in knowing and articulating personal preferences considering their life and care during the decision‐making process (van Leersum et al., 2019). The tool aims to assist in the construction and discussion of preferences during client–professional consultations in four long‐term care sectors in the Netherlands: nursing and care of older people, mental healthcare, care of people with disabilities and social care that refers to support services to citizens with needs related to, for example, psychosocial problems, debts, housing, and addiction, provided under the Dutch Social Support Act in 2015 (Wmo 2015). All institutionalised care as well as home or ambulant care. Examples of preferences could include the desire to have a pet, to take meals with other people or the importance of a ‘click’ with a caregiver (van Leersum et al., 2019, 2020). Preference elicitation may indeed assist professionals in acquiring insight into client preferences and in discussing the clients’ view on the long‐term care options. The problem is that these open conversations on preferences and values are not or rarely spontaneously performed in daily practice. The aim of tools like ‘What matters most to me’ is to stimulate and facilitate such conversations, not to replace these conversations (Table 1). The tool ‘What matters to me’ was designed to assist clients and their professionals with preference elicitation during the decision‐making process for long‐term care. Either in need or searching for home care or institutional care, or in cases where advanced care planning is at stake. In all cases, clients have personal preferences for care and their life, which the tool ‘What matters to me’ might assist to articulate.
TABLE 1

‘What matters to me’ (van Leersum, Moser, et al., 2020; van Leersum, van Steenkiste, et al., 2020)

The tool ‘What matters to me’ is designed to assist with preference elicitation for long‐term health and social care settings: nursing and care of older people, mental healthcare, care of people with disabilities and social care. The design was based on existing tools and a qualitative study into preferences of clients in long‐term care settings. The preference elicitation tool ‘What matters to me’ comprises five categories:

‘Health’ is defined as client's care needs and their preferences on how to receive care. This domain helps clients identify to what extent they are self‐reliant in providing their care, and in what areas they need assistance; it also includes preferences for care professionals.

‘Family and friends’ is defined as the importance of relatives and all kind of social contacts, addressing all people important in someone's life.

‘Living conditions’ is defined as client's preferences regarding housing, different options and environment as well as social interaction in their living environment.

‘Daily life’ is defined as client's preferences for all kinds of activities ranging from work to sport, and culture to religion.

‘Finances’ is defined as financial resources and considers the role money or debts plays in life, and the preferences for assistance with finances or administration.

‘What matter to me’ is web based consisting of four essential pages (Additional file 1 shows screenshots of the web‐based tool).

On the homepage, the user can read information about the purpose and the use of the tool. The homepage also has the function of portal towards the preference elicitation section, by clicking the start‐button.

After clicking the start button, the ‘category page’ opens. The user can choose one or more categories to his or her preference, and is invited to click on a category symbol to answer the propositions belonging to this category.

When choosing a category, different proposition pages will follow. Each category has a different amount of propositions. Each proposition belonging to the category of choice appears on separate pages. Users can click on the answers that match their opinion and continue with the subsequent proposition. After the last proposition of a category, the user will be automatically returned back to the category page.

When a user finished answering the propositions of one or more categories, he or she can click on the button ‘overview’. The overview lists all the answers a user has given. The answers are shown per category and users are able to change their answers. This overview can be printed or send to the user by email.

‘What matters to me’ (van Leersum, Moser, et al., 2020; van Leersum, van Steenkiste, et al., 2020) The tool ‘What matters to me’ is designed to assist with preference elicitation for long‐term health and social care settings: nursing and care of older people, mental healthcare, care of people with disabilities and social care. The design was based on existing tools and a qualitative study into preferences of clients in long‐term care settings. The preference elicitation tool ‘What matters to me’ comprises five categories: ‘Health’ is defined as client's care needs and their preferences on how to receive care. This domain helps clients identify to what extent they are self‐reliant in providing their care, and in what areas they need assistance; it also includes preferences for care professionals. ‘Family and friends’ is defined as the importance of relatives and all kind of social contacts, addressing all people important in someone's life. ‘Living conditions’ is defined as client's preferences regarding housing, different options and environment as well as social interaction in their living environment. ‘Daily life’ is defined as client's preferences for all kinds of activities ranging from work to sport, and culture to religion. ‘Finances’ is defined as financial resources and considers the role money or debts plays in life, and the preferences for assistance with finances or administration. ‘What matter to me’ is web based consisting of four essential pages (Additional file 1 shows screenshots of the web‐based tool). On the homepage, the user can read information about the purpose and the use of the tool. The homepage also has the function of portal towards the preference elicitation section, by clicking the start‐button. After clicking the start button, the ‘category page’ opens. The user can choose one or more categories to his or her preference, and is invited to click on a category symbol to answer the propositions belonging to this category. When choosing a category, different proposition pages will follow. Each category has a different amount of propositions. Each proposition belonging to the category of choice appears on separate pages. Users can click on the answers that match their opinion and continue with the subsequent proposition. After the last proposition of a category, the user will be automatically returned back to the category page. When a user finished answering the propositions of one or more categories, he or she can click on the button ‘overview’. The overview lists all the answers a user has given. The answers are shown per category and users are able to change their answers. This overview can be printed or send to the user by email. The implementation of new interventions remains a challenge for healthcare, as implementation is often difficult due to lack of knowledge about healthcare environments (Christie et al., 2018; Krick et al., 2019). Furthermore, there are barriers to implementation, such as costs, time and difficulty in fitting it into the current workflow (Granja et al., 2018). To increase the knowledge about the use of tools in real practice, evaluation is needed to assist implementation (Barello et al., 2015; Enam et al., 2018; Escoffrey et al., 2004; Ross et al., 2016). The aim of this study is to perform a process evaluation to evaluate the use of, and experiences with, the preference elicitation tool ‘What matters to me’. This process evaluation was guided by two research questions. To what extent was ‘What matters to me’ used in long‐term care settings? How was using ‘What matters to me’ experienced? These questions were studied from the perspectives of both professional and client.

METHODS

Design

A process evaluation was chosen to describe the actual use of the tool (Research question 1) and to gain insight into the experiences with the tool (Research question 2). This process evaluation aims to plan strategies for successful implementation in practice (Enam et al., 2018; Escoffrey et al., 2004). The process evaluation is based on six process components, that is, context, fidelity, dose delivered, dose received, reach and recruitment (Saunders et al., 2005). These components generate insight into mechanisms of a program and investigate its impact and outcome. The component ‘context’ refers to the environment in which ‘What matters to me’ was evaluated. ‘Fidelity’ is the extent to which it was implemented as originally proposed in the blueprint. ‘Dose delivered’ covers the materials provided or delivered by the professionals, and ‘What matters to me’. ‘Dose received’ and ‘satisfaction’ question the extent to which participants interacted and used the tool as recommended, and their satisfaction with its use. ‘Reach’ or ‘participation rate’ is the proportion of the intended stakeholders who actually participated in the process. Lastly, ‘recruitment’ covers approaching and including users. The components context, fidelity, dose delivered and received and reach were used to answer research question 1. The components fidelity, dose delivered, satisfaction, reach, and recruitment were used to answer Research question 2. Table 2 will show the components of the process evaluation with the quantitative and qualitative data sources.
TABLE 2

Summary of the methodology. The first and second columns contain the six components of a process evaluation (Saunders et al., 2005) and the categories within these components. For each component, the quantitative and qualitative data sources are shown

ComponentsCategoriesQuantitative data sourceQualitative data source
ContextEnvironment of useQuestionnairesInterviews
FidelityQualityQuestionnairesInterviews
Actual use and implementationsLog dataInterviews
Dose deliveredProvided materialsQuestionnairesInterviews
Dose received and satisfactionUse as recommendedLog data and questionnaires
SatisfactionInterviews
ReachParticipationLog data and questionnairesInterviews
RecruitmentWay of inclusionPhone calls and email conversations with profs
Summary of the methodology. The first and second columns contain the six components of a process evaluation (Saunders et al., 2005) and the categories within these components. For each component, the quantitative and qualitative data sources are shown

Setting

Long‐term care in the Netherlands comprises four sectors: nursing and care of older people, mental healthcare, care of people with disabilities and social care. Twelve long‐term care organisations spread over three provinces of the Netherlands participated. These organisations provide home care or residential care. Three organisations operated in the field of nursing and care of older people, three in the care of people with disabilities, one in mental health care, three in social care and two provided services in all care sectors. The professionals were caregivers, independent care coordinators and professionals responsible for intake. They shared information with clients during the decision‐making process, and discussed options clients might have.

Participants and recruitment

The participants in this study were professionals and clients. Professionals were included if they were actively involved and consulted by clients during their search for long‐term care. These professionals will guide their client throughout the decision‐making process. They could work within a care facility or work as independent advisers belonging to the municipality. The clients were actual clients, and their relatives or informal caregivers. Inclusion criteria were active searching for long‐term care and consulting a professional for support in decision‐making. Exclusion criteria were the inability to use ‘What matters to me’, and a command of the Dutch language below level A2. There were four strategies followed during the recruitment. First, the researchers contacted 18 long‐term care organisations divided over the four long‐term care sectors. Information about ‘What matters to me’ and the study was given on visits to these organisations. Twelve organisations were included through purposeful sampling. Second, the professionals recruited clients by means of an information letter, a card with a link to ‘What matters to me’ and a questionnaire. The professionals were contacted by the researchers every 2 weeks by email or telephone to follow the progression of the recruitment. All professionals together recruited 24 clients for the study by means of an information letter, a card with a link to ‘What matters to me’ and a questionnaire. Third, clients were also able to participate independent of professional referral. Two clients were recruited via a pop‐up built into the open access tool ‘What matters to me’. These participants received the information letter and questionnaire by mail. In total, 26 clients completed the questionnaire. At the end of the questionnaire, clients could indicate if they were willing to participate in a follow‐up interview. As a result, 20 clients agreed to an interview by phone (convenience sampling). And fourth, there was self‐recruitment of users. The tool was visited and used by 102 unique visitors of which 71 completed one or more categories of the tool.

Ethics

Approval was obtained from the Ethics Committee of Zuyderland Zuyd (dossier number 17‐N‐79). The participants gave written informed consent and were informed about the goal and process of the study, and their right to withdraw at any time. Data were anonymised and confidentiality was maintained.

Intervention

The study intervention comprised of four steps: (1) the participating professionals received 1 hr on the spot training on ‘What matters to me’ (Table 1 and File S1) and information about research procedures. During the training, the professionals were advised to give the link to their clients to prepare for a follow‐up consultation. However, some professionals chose to discuss and use the tool during the consultation. The professionals were free to choose the most suitable procedure for their clients. (2) During a consultation with their professional, clients received a link to ‘What matters to me’ and oral or written information about the study. When clients were unable to use the tool by themselves, the relatives were invited to use the tool and complete the propositions considering the preferences of the client. (3) The clients used ‘What matters to me’ to prepare for the follow‐up consultation. (4) The clients returned to their professional for the follow‐up consultation to discuss the preferences generated by the tool.

Data collection

Data were collected between April and August 2018. Quantitative data were collected by 71 user activity logs of the tracking system of the website. These logs show the number of hits, the route a participant took through the website, which categories were used and the time spent on particular pages. Second, 26 client questionnaires considered use, usability and user‐friendliness. The questionnaire included the Post System Study usability Questionnaire (PSSUQ), and additional questions on experiences and the time they spent on it (Table 3). The PSSUQ is an instrument specifically designed to measure the usefulness, information quality, and interface quality of a system (Lewis, 1993). The questionnaire evaluates usability characteristics such as perceived time taken to complete the work, ease of learning, quality of documentation and information, functional adequacy and speed of acquisition (Fruhling & Lee, 2005; Lewis, 2002). Participants were asked to answer statements using 7‐point Likert scales, ranging from 1 ‘strongly disagree’ to 7 ‘strongly agree’, scores were used to calculate the score over all statements (Lewis, 2002; Phongpheaw & Jiamsanguangwong, 2016).
TABLE 3

Post‐Study System Usability Questionnaire Items (Lewis, 1993)

QuestionScoring
1I am satisfied with how easy it is to use this system.1–2–3–4–5–6–7
2It was simple to use this system.1–2–3–4–5–6–7
3I could effectively complete the tasks and scenarios using this system.1–2–3–4–5–6–7
4I was able to complete the tasks and scenarios quickly using this system.1–2–3–4–5–6–7
5I was able to efficiently complete the tasks and scenarios using this system.1–2–3–4–5–6–7
6I felt comfortable using this system.1–2–3–4–5–6–7
7It was easy to learn to use this system.1–2–3–4–5–6–7
8I believe I could become understand quickly how to use this system.1–2–3–4–5–6–7
9The system gave error messages that clearly told me how to fix problems.1–2–3–4–5–6–7
10Whenever I made a mistake using the system, I could recover easily and quickly.1–2–3–4–5–6–7
11The information (such as online help, on‐screen messages, and other documentation) provided with this system was clear.1–2–3–4–5–6–7
12It was easy to find the information I needed.1–2–3–4–5–6–7
13The information provided for the system was easy to understand.1–2–3–4–5–6–7
14The information was effective in helping me complete the tasks and scenarios.1–2–3–4–5–6–7
15The organisation of information on the system screens was clear.1–2–3–4–5–6–7
16The interface of this system was pleasant.1–2–3–4–5–6–7
17I liked using the interface of this system.1–2–3–4–5–6–7
18This system has all the functions and capabilities I expect it to have.1–2–3–4–5–6–7
19Overall, I am satisfied with this system1–2–3–4–5–6–7
Post‐Study System Usability Questionnaire Items (Lewis, 1993) Third, 12 professionals answered a questionnaire about their experiences by rating it on a 1–10 scale, their involvement with the clients during consultations, the time they spent on using the tool and their experiences with ‘What matters to me’ during consultations. Qualitative data were collected by semi‐structured phone interviews with 20 clients or relatives. The interview guide was developed and reviewed by two client representatives. The interviews lasted approximately 15 min, audiotapes were made and field notes taken. Another data source was 59 emails and 47 phone calls with the professionals. Field notes contain failures and successes in asking clients to participate and use ‘What matters to me’.

Data analysis

To analyse the quantitative data, SPSS was used for descriptive statistics. To analyse the qualitative data, interviews were transcribed verbatim. Content analysis with a deductive approach (Elo & Kyngas, 2008) was performed on all transcripts guided by the six components of the process evaluation model of Saunders et al., (2005). During the organising phase of the analysis, a matrix was developed comprising the six components. Categories were created within each of the components of the analysis matrix (File S2). Two researchers performed the analysis. The analysis and findings were discussed during weekly meetings with the research team. During this phase, the researchers compared the components with the categories. The findings of the analysed data based on the six components are described in the results section. Data management was performed using the NVivo version 11 software package.

Validity

Validity was established by means of triangulation and face validity (Giddings & Grant, 2009). Data triangulation was established by including four different quantitative and qualitative data sources: log data files, questionnaires from clients, questionnaires from professionals and interviews. Methodological triangulation was reached by including clients twice with a short time interval in the quantitative and qualitative data collection. Two researchers read, analysed and compared findings multiple times. The project team was informed about the findings at weekly meetings at which the scientific and organisational aspects were discussed to reach investigator triangulation. Face validity was reached by acknowledging the use of the six components to apply a process evaluation. Member checking was performed by sharing transcripts with participants, and by an invitational conference. All participants and those who have an interest were invited to the conference where the findings were presented and discussed. Five participants were present during the conference. Their feedback given during the discussions of the conference was used for further development of the tool and to develop plans for implementation.

RESULTS

The result section begins with a description of the participants involved in this study. Thereafter, the results are structured by the two research questions: research question 1, use, and research question 2, experience. The components context, fidelity, dose delivered and received and reach are used to describe the use of ‘What matters to me’, and the components fidelity, dose delivered, satisfaction, reach and recruitment are used to describe the experiences with the tool. A division is made between those data obtained from the professionals and from the clients. At the end of the results section, Table 4 shows the six components of the process evaluation with a summary of the findings.
TABLE 4

Summary of findings, quantitative and qualitative results are presented for each component (Saunders et al., 2005)

ComponentsQuantitative resultsQualitative results
ContextThe users are involved in care of older adults, care for people with disabilities, mental healthcare and social support.All clients were in need of care in the near future.
Fidelity

71 users filled‐in the tool and completed one or more categories. All categories were used.

The average time spent was 6.9 (±0.03) minutes.

The tool was rated 6.63 (±0.88) of 7 by clients and 7.73 (±0.75, range 7 to 9) = of 10 by professionals.

Recommendations for implementations were given, including a separate version for relatives and a non‐digital version.
Dose deliveredAll clients knew about the tool and received the link via a professional or relative.The professionals handed out information about the tool to some of their clients. The additional questionnaire was less often given to their clients.
Dose received and satisfactionThe tool was filled in as recommended, but the overview was not discussed by all professionals during consultations with their clients.

The tool was helpful to prepare for consultations, set the agenda and build a relationship.

All clients and professionals would recommend the tool to others.

Reach

26 of the 50 clients were asked by professionals to fill in the additional questionnaire.

71 of the 102 visitors of the tool filled‐in one or more categories.

The reach was expected to increase by the inclusion of more organisations and professionals, the use of (social) media or other channels.
RecruitmentProfessionals hesitated to ask clients to participate based on assumptions of pressure for the client. Due to work pressure, it was difficult to recruit professionals.
Summary of findings, quantitative and qualitative results are presented for each component (Saunders et al., 2005) 71 users filled‐in the tool and completed one or more categories. All categories were used. The average time spent was 6.9 (±0.03) minutes. The tool was rated 6.63 (±0.88) of 7 by clients and 7.73 (±0.75, range 7 to 9) = of 10 by professionals. The tool was helpful to prepare for consultations, set the agenda and build a relationship. All clients and professionals would recommend the tool to others. 26 of the 50 clients were asked by professionals to fill in the additional questionnaire. 71 of the 102 visitors of the tool filled‐in one or more categories.

Participants

Twelve professionals filled in the questionnaire at the end of the study. Seventeen per cent were men, and the average age of the professionals was 46 (±14.6 years) ranging from 19 to 63 years. All had Dutch nationality and a higher educational level. Seven professionals worked in the care of people with disabilities, one in mental health care, two in social care and two provided care in more than one sector. The professionals recruited 24 clients and 2 clients were reached via the pop‐up built in ‘What matters to me’. These 26 clients used the tool and completed the additional questionnaire as part of this study. Of the 26 clients, 54% were men, and the average age of the clients was 45 (±24.9 years), ranging from 18 to 92 years. Forty‐two per cent had a low educational level. Twenty‐three participants had Dutch nationality, the others were of Turkish or Indian origin. Five clients received care in the nursing and care of older people sector, nine clients in the care of people with disabilities sector, three clients in the mental healthcare sector, seven clients in the social care sector and two clients received care in more than one sector. Fifteen clients used ‘What matters to me’ independently, three relatives, four professionals and four clients completed it together with a professional.

Use of ‘what matters to me’

To answer the first research question: To what extent was ‘What matters to me’ used in long‐term care settings? The process evaluation component context was described, and the components fidelity, dose delivered, dose received and reach were explored. The context of the professionals and clients is described in the previous section. Considering the components dose delivered and reach, the 12 professionals asked 50 clients to use ‘What matters to me’. Six professionals asked all their clients who fulfilled the inclusion criteria to take part in the study and use ‘What matters to me’. Four professionals asked a small number of clients, and two did not ask their clients. Some professionals gave information and asked their clients to use the tool without mentioning the additional questionnaire. Twenty‐one users knew about ‘What matters to me’ via their professionals, and five already heard about it through informal routes, a friend or partner. Considering the fidelity and dose received, the average estimate professionals made of the time they spent together with their clients ranged from <10 min to 30 min. The total number of spontaneous and referred visitors of the tool was 102, of which 71 clicked and filled‐in on one or more categories of the tool. The log data showed that 33 of these 71 visitors filled in all five categories. None of the categories was filled in less frequently than the others. The option to write free‐text comments on the answers was used by 27% of the participants. The average actual time the participants spent was 6.9 (±0.03) minutes, ranging from 2 to more than 30 min (Table 5). The generated overview with preferences was printed or emailed by 74% of the clients. Six of 12 professionals discussed the overview with their client. However, the interviews revealed that more clients would have liked to discuss the overview with their professional:
TABLE 5

Actual use of ‘What matters to me’, the categories and time spent obtained from the log data of the process evaluation

VariableNumber of participants (N = 71)
Number of categories filled in
120
211
36
41
533
Categories filled in
Health45
Living conditions42
Family and friends45
Daily life52
Finances46
Addition of extra comments to answers18
Time spent
0–10 min54
11–20 min12
21–30 min4
>30 min1
‘Coincidentally, the next day I had a consultation and wanted to show the overview to my coach, but my coach was not interested and did not even look at it’. [Client 3] Actual use of ‘What matters to me’, the categories and time spent obtained from the log data of the process evaluation

Experiences with ‘what matters to me’

The process evaluation components fidelity, dose delivered, satisfaction, reach and recruitment were explored to answer the second research question, How was using ‘What matters to me’ experienced? In the user context, all clients were in need of care in the near future, but their need was not classified as urgent. For example, the father of an autistic son acknowledged he was ageing and was uncertain about the future care for his son: ‘My son is autistic, he is 46 now and has just received a long‐term care indication. I am 72 and have just had an intestinal operation. This has made me worry about the future. My wife died last January, and of course I worry about what will happen to my son’. [Client 1] When questioned about the dose delivered and dose received, all clients said that they were told about ‘What matters to me’ by a professional or relative, and had received the link to the tool and relevant use information from them: ‘Someone came to discuss housing and living conditions for my future with us. It is not immediately necessary at this time, but I would like to be covered for later on. Then she told me about a tool that could help us and asked if I was willing to use it. I said ‘I would really like to use it’. [Client 5] Considering the component recruitment in the process evaluation, some professionals hesitated to ask clients to participate not only due to work pressure of the professional but also based on their assumption that the client would not be willing to participate. Other professionals were hesitant at first, but the clients were motivated to use and discuss their preferences: ‘Together with colleagues, I handed out some flyers with the link to “What matters to me”. Although our population (mental disabilities) does not seem to be very motivated to take part in research, we often discussed the ‘What matters to me’ questions during consultations. We looked at the questionnaires a few times, and this clearly added value to the conversation’. [Professional 5] Clients thought about possibilities to extend the reach by naming professionals and organisations that could help in commercial advertising of ‘What matters to me’ to their clients. They also suggested using the media, social media and diverse awareness‐raising channels, and posting reports or recommendations to others: ‘A sort of “like” button that can be shared in all kinds of ways, a share button and somewhere to leave a comment. I think these things could help people when they generate a search, because when I googled “What matters to me” you were not really near the top’. [Client 7] Considering the fidelity, participants who did not fill in all categories were questioned for their reasoning. All replied that the categories of which they did not answer the propositions were considered not useful for their situation. The professionals graded the quality of the tool at a mean of 7.73 (±0.75, range 7 to 9) of 10. Recommendations for improvements were given in the comment field, including a map showing organisations where a client could receive care, a non‐digital version and a separate section for informal caregivers. The clients evaluated ‘What matters to me’ by answering the PSSUQ, and gave it a mean score of 6.24 (±0.76) of seven. During the interviews, clients graded the user‐friendliness at an 8 or 9 (of 10). Relatives who used ‘What matters to me’ wished to label their role as the client's proxy: ‘The thing I missed from “What matters to me” was that there is nowhere to indicate that you are filling in the answers on behalf of someone else, not for yourself’. [Relative 2] All clients and professionals would recommend ‘What matters to me’ to others. Satisfaction was further explored by asking how clients experienced the use of the tool. All clients who used the tool wanted to talk about the overview with a professional. The clients and professionals who used the overview at the follow‐up consultation were positive about its usefulness. It was helpful to prepare for consultations, set the agenda and to build a relationship with someone in a short time since the issues that mattered in someone's life were directly on the table: ‘“What matters to me” is useful to get to know someone a bit better. Thus, asking the person to fill in the questions will save a lot of time, and I will acquire more knowledge in a shorter time’. [Professional 7] ‘I really liked the idea of having the overview with me. It could help me to remember things during the consultation. We had to drive there, then find the right place, and then getting there on time, very stressful situation. Having this overview gave me a good feeling. We now have something in reserve, and I have discussed all the important things’. [Client 4]

DISCUSSION

The aim of the process evaluation was to evaluate the use of (research question 1) and experiences with (research question 2) the tool ‘What matters to me’. Considering the six components for a process evaluation, starting with the context component, the participants were involved in care of older people, care for people with disabilities, mental healthcare and social support. Sixty‐nine per cent of the visitors of the tool completed one or more categories, the average time spent was 6.9 (±0.03) minutes and the tool was rated at 6.63 (±0.88) of 7 in the PSSUQ. Almost all participants, both clients and professionals, experienced ‘What matters to me’ as useful in practice, and would recommend it to others. Suggestions to improve the tool included a separate version specifically for relatives, and a non‐digital version. Considering time span, the attention span of eHealth users is short and will shorten even further if the time requirement is too long (van der Heijden, 2000). When using eHealth, the ideal attention span is 6 min (Geri et al., 2017). Therefore, the short time spent by the users of the tool will benefit by its implementation. Distractions and difficulties in understanding reduce the attention span (Wagner et al., 2014); therefore, the interface should be attractive with few distractions and simple to understand. Fidelity showed that the tool is attractive and easy to understand, which will be beneficial for the attention span and the willingness to use. Although satisfaction was high, the professionals stated that time was often a barrier to use a tool. A tool should meet expectations of professionals and fit in with their workflow; time is an important aspect in care settings (Granja et al., 2018; Reiners et al., 2019). Our results showed that the time professionals had spent on using the tool varied depending on how they used it. Professionals adapted their workflow by providing information about ‘What matters to me’. During a follow‐up consultation, some professionals discussed the overview with preferences. The time these professionals spent with their client ranged from 5 to 15 min, but those professionals who filled in the tool together with their clients spent approximately 30 min. Evaluating the workflow could enhance a long‐term implementation and support the healthcare setting in which ‘What matters to me’ would be used (Enam et al., 2018). The study showed ways to implement ‘What matters to me’ in the workflow of professionals. Its implementation introduced new tasks and responsibilities for the professional (Karisalmi et al. 2019). ‘What matters to me’ could become part of SDM for clients during the decision‐making process about long‐term care, but more guidance in using eHealth in SDM seems necessary (Wieringa et al., 2019). Concerning when it should be used in consultations, ‘What matters to me’ could become part of the team talk, at which exploring important preferences has great potential to discuss the range of options a client has (Elwyn et al., 2017). The context component showed that clients from diverse long‐term care sectors were included, which is a strength of this study. However, no professionals working in the nursing and care of older people sector participated. The clients in need of nursing and care consulted professionals employed in multiple care sectors. Professionals had difficulty recruiting clients. The component recruitment showed that the reasons for this were the assumed extra burden for clients to fill in a questionnaire, and the time constraints of the professional. Besides a reach following in 26 questionnaires, an additional 71 online user activity logs were obtained. Activity logs enabled data collection without any extra effort by the participants to fill in a questionnaire (Sieverink et al., 2017). A limitation is that these data do not give the complete view of the participant. A difficulty with the components dose delivered, reach and recruitment is the possible confusions by the terminology. It shows overlap with terminology of a methodology. As defined by Saunders (2005), ‘dose delivered’ covers the materials provided or delivered to a user. ‘Reach’ or ‘participation rate’ is the proportion of the intended users who actually participated, and ‘recruitment’ covers findings considering approaching and including of users. In this paper, the methodology focussed on how to gain information about the six components, and the findings considering these components were presented in the results section. By discussing the components beforehand, the researchers had a clear understanding of the differences which helped in the possible confusions caused by similarities in terminology. This process evaluation raises questions on how to effectively implement ‘What matters to me’ in long‐term care settings. The findings provided insight into identifying potential problems related to implementation (van der Krieke et al., 2013). No participants experienced challenges with the tool as such; clients did not report anything that might hamper possible implementation. However, professionals identified challenges to use in practice. They were unsure which clients they should ask to use it, mainly due to urgency of situations and competences of clients. To foster its use, professionals and clients should be given information on how to access and use it (Karisalmi et al., 2019). This could become part of the training in advance. The training could include scenarios to show applicability of the tool in different settings with a variety of clients. During the implementation of ‘What matters to me’, professionals should play a key role in providing access, as even the most active clients might not be aware of the tool.

CONCLUSION

In this study, clients and professionals used ‘What matters to me’ in real decision‐making practice for long‐term care. The process evaluation demonstrated that the tool seems useful in practice, and may assist with preference elicitation. Although a follow‐up study should evaluate the perceived impact on SDM and consultations for long‐term care, this study shows that the tool is experienced beneficial to clarify preferences of clients in long‐term care. Also, this clarification is shown helpful for the clients and professionals to discuss these preferences for life and care. In health and social care practice, the tool seems to fit as part of consultations considering long‐term care needs and preferences.

CONFLICT OF INTERESTS

All authors declare they have no competing interests.

AUTHOR CONTRIBUTION

All authors designed the study protocol. AM, BvS and CvL contributed to the data collection. CvL performed the analysis of the data, and CvL, AM and BvS contributed to the interpretation of the data. CvL undertook the drafting of the first manuscript. TW was the project leader, with AM co‐leading and the applicant. All authors contributed to drafting and critically reviewing the manuscript, and all approved the final version. All agreed to be accountable for all aspects of the work. File S1 Click here for additional data file. File S2 Click here for additional data file.
  32 in total

1.  The qualitative content analysis process.

Authors:  Satu Elo; Helvi Kyngäs
Journal:  J Adv Nurs       Date:  2008-04       Impact factor: 3.187

2.  Encouraging the Use of eHealth Services: A Survey of Patients' Experiences.

Authors:  Nina Karisalmi; Johanna Kaipio; Sari Kujala
Journal:  Stud Health Technol Inform       Date:  2019

3.  Prime Time for Shared Decision Making.

Authors:  Erica S Spatz; Harlan M Krumholz; Benjamin W Moulton
Journal:  JAMA       Date:  2017-04-04       Impact factor: 56.272

4.  Older Adult Internet Use and eHealth Literacy.

Authors:  Thomas A Arcury; Joanne C Sandberg; Kathryn P Melius; Sara A Quandt; Xiaoyan Leng; Celine Latulipe; David P Miller; D Alden Smith; Alain G Bertoni
Journal:  J Appl Gerontol       Date:  2018-10-24

Review 5.  Ageing populations: the challenges ahead.

Authors:  Kaare Christensen; Gabriele Doblhammer; Roland Rau; James W Vaupel
Journal:  Lancet       Date:  2009-10-03       Impact factor: 79.321

6.  Outcome prioritisation tool for medication review in older patients with multimorbidity: a pilot study in general practice.

Authors:  Jojanneke Jgt van Summeren; Jan Schuling; Flora M Haaijer-Ruskamp; Petra Denig
Journal:  Br J Gen Pract       Date:  2017-03-27       Impact factor: 5.386

7.  Opening the Black Box of Electronic Health: Collecting, Analyzing, and Interpreting Log Data.

Authors:  Floor Sieverink; Saskia Kelders; Mannes Poel; Lisette van Gemert-Pijnen
Journal:  JMIR Res Protoc       Date:  2017-08-07

8.  Evidence-Based Evaluation of eHealth Interventions: Systematic Literature Review.

Authors:  Amia Enam; Johanna Torres-Bonilla; Henrik Eriksson
Journal:  J Med Internet Res       Date:  2018-11-23       Impact factor: 5.428

9.  Getting to grips with the process of decision-making in long-term care. Descriptive cases illustrate the chaotic reality of the construction of preferences.

Authors:  Catharina M van Leersum; Albine Moser; Ben van Steenkiste; Judith R L M Wolf; Trudy van der Weijden
Journal:  PLoS One       Date:  2019-05-24       Impact factor: 3.240

10.  Factors Determining the Success and Failure of eHealth Interventions: Systematic Review of the Literature.

Authors:  Conceição Granja; Wouter Janssen; Monika Alise Johansen
Journal:  J Med Internet Res       Date:  2018-05-01       Impact factor: 5.428

View more
  1 in total

1.  Clients and professionals elicit long-term care preferences by using 'What matters to me': A process evaluation in the Netherlands.

Authors:  Catharina M van Leersum; Albine Moser; Ben van Steenkiste; Judith R L M Wolf; Trudy van der Weijden
Journal:  Health Soc Care Community       Date:  2021-07-12
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.