Literature DB >> 28720558

My Team of Care Study: A Pilot Randomized Controlled Trial of a Web-Based Communication Tool for Collaborative Care in Patients With Advanced Cancer.

Teja Voruganti1, Eva Grunfeld1,2, Trevor Jamieson3,4,5, Allison M Kurahashi6, Bhadra Lokuge6, Monika K Krzyzanowska1,7, Muhammad Mamdani1,5,8,9,10, Rahim Moineddin2,10, Amna Husain2,6.   

Abstract

BACKGROUND: The management of patients with complex care needs requires the expertise of health care providers from multiple settings and specialties. As such, there is a need for cross-setting, cross-disciplinary solutions that address deficits in communication and continuity of care. We have developed a Web-based tool for clinical collaboration, called Loop, which assembles the patient and care team in a virtual space for the purpose of facilitating communication around care management.
OBJECTIVE: The objectives of this pilot study were to evaluate the feasibility of integrating a tool like Loop into current care practices and to capture preliminary measures of the effect of Loop on continuity of care, quality of care, symptom distress, and health care utilization.
METHODS: We conducted an open-label pilot cluster randomized controlled trial allocating patients with advanced cancer (defined as stage III or IV disease) with ≥3 months prognosis, their participating health care team and caregivers to receive either the Loop intervention or usual care. Outcome data were collected from patients on a monthly basis for 3 months. Trial feasibility was measured with rate of uptake, as well as recruitment and system usage. The Picker Continuity of Care subscale, Palliative care Outcomes Scale, Edmonton Symptom Assessment Scale, and Ambulatory and Home Care Record were patient self-reported measures of continuity of care, quality of care, symptom distress, and health services utilization, respectively. We conducted a content analysis of messages posted on Loop to understand how the system was used.
RESULTS: Nineteen physicians (oncologists or palliative care physicians) were randomized to the intervention or control arms. One hundred twenty-seven of their patients with advanced cancer were approached and 48 patients enrolled. Of 24 patients in the intervention arm, 20 (83.3%) registered onto Loop. In the intervention and control arms, 12 and 11 patients completed three months of follow-up, respectively. A mean of 1.2 (range: 0 to 4) additional healthcare providers with an average total of 3 healthcare providers participated per team. An unadjusted between-arm increase of +11.4 was observed on the Picker scale in favor of the intervention arm. Other measures showed negligible changes. Loop was primarily used for medical care management, symptom reporting, and appointment coordination.
CONCLUSIONS: The results of this study show that implementation of Loop was feasible. It provides useful information for planning future studies further examining effectiveness and team collaboration. Numerically higher scores were observed for the Loop arm relative to the control arm with respect to continuity of care. Future work is required to understand the incentives and barriers to participation so that the implementation of tools like Loop can be optimized. TRIAL REGISTRATION: ClinicalTrials.gov NCT02372994; https://clinicaltrials.gov/ct2/show/NCT02372994 (Archived by WebCite at http://www.webcitation.org/6r00L4Skb). ©Teja Voruganti, Eva Grunfeld, Trevor Jamieson, Allison M Kurahashi, Bhadra Lokuge, Monika K Krzyzanowska, Muhammad Mamdani, Rahim Moineddin, Amna Husain. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 18.07.2017.

Entities:  

Keywords:  MeSH: Internet; adult; chronic disease; communication; continuity of patient care; interdisciplinary communication; neoplasms; outcome assessment (health care); patient care team; professional-patient relations

Mesh:

Year:  2017        PMID: 28720558      PMCID: PMC5539387          DOI: 10.2196/jmir.7421

Source DB:  PubMed          Journal:  J Med Internet Res        ISSN: 1438-8871            Impact factor:   5.428


Introduction

With advances in medical care enabling people to live longer, patients with chronic diseases and their families have increasingly complex care needs requiring the expertise of many health care providers from multiple settings and more frequent use of the health care system [1-3]. Important contextual information is not consistently exchanged between health care providers, and coordinated delivery of patient care as a team is lacking [4-7]. As such, there is a need for solutions that are cross-organizational, cross-setting, and that improve continuity of care, defined as the extent to which delivery of care by different providers is coherent, connected, and timely [8]. Organizations such as the Institute of Medicine and the Agency for Healthcare Research and Quality have called for solutions that build on the growing momentum of health information technology to address the deficits in continuity of care and coordinated delivery of care [9-13]. With over 80% of the populations of Canada and the United States having access to the Internet and mobile phones [14], Web and mobile-based communication are ideally positioned to improve the sharing of knowledge, expertise, and decision making between providers (ie, collaboration) [15], to involve patients and, by extension, improve continuity of care [16,17]. Solutions have generally been limited to one-to-one secure messaging or email, possibly as additions to information systems such as patient health records [18]. Reviews on the impact of tools for patient-physician communication have shown promising evidence of improvement on such outcomes as patient self-efficacy, satisfaction with care, and on clinical/psychosocial outcomes [19,20]. However, few tools exist with the express intent of facilitating secure team-based communication, which can enable sharing of information between different providers, across health events and settings, and promote collaborative care [21]. Previous studies examining tools that enable patients to communicate with their health care team have been observational in design and focused their examination on implementation in the pediatric [22], general primary care [23], elderly [24], and cerebral palsy [25] populations. These studies did not consider such outcomes as continuity of care, which is particularly pertinent to the patient population with complex care needs. In this study, we evaluated a Web-based tool for clinical collaboration, called Loop. The purpose of Loop is to assemble care teams that include patients and caregivers in order to facilitate communication and collaboration [26]. We conducted a pilot randomized controlled trial in a population of patients with advanced cancer, as prototypical of a population with complex care needs [27,28]. Our objective was to evaluate the feasibility of integrating a tool like Loop into current care processes and to capture preliminary measures of the effect of Loop on continuity of care, quality of care, symptom distress, and health care utilization.

Methods

Trial Design

We conducted a 15-month multicentered, nonblinded, pragmatic pilot cluster-randomized controlled trial (cRCT), called the My Team of Care study, allocating participants to receive access to Loop as the intervention arm or to usual care as the control arm. The unit of randomization was at the level of the physician, and the unit of analysis was at the level of the individual patient. See Multimedia Appendix 1 for the CONSORT EHEALTH checklist [29].

Setting and Participants

The study took place at the Temmy Latner Center for Palliative Care at Mount Sinai Hospital and Princess Margaret Cancer Center in Toronto, Ontario, Canada, from January 2015 to April 2016. The Temmy Latner Center is the largest home-based palliative care program in Canada, consisting of 23 palliative care physicians. The Princess Margaret Cancer Center is a University of Toronto‒affiliated research hospital with 46 medical oncologists and 42 radiation oncologists. At both study sites, patients generally access their physician through visits and telephone messages; some health care providers are contactable via email. Participants consisted of eligible patients plus their principal cancer physician (oncologist or palliative care physician), and if interested, their family caregiver (informal, unpaid). For the intervention arm, additional health care providers as identified by the patient were also invited to participate as members of the circle of care to use Loop. Eligible patients were aged 18 or older; had stage IV cancer or stage III cancer with poor prognosis as determined by their oncologist (a survival prognosis of ˃3 months but ˂2 years); Eastern Cooperative Oncology Group (ECOG) performance status score of 0, 1, or 2, as assessed by their oncologist or palliative care physician at time of enrollment; English literacy and language competency to provide informed consent and complete questionnaires; and patients or caregivers had access to a computer and Internet. Exclusion criteria were currently receiving or a candidate for hormone therapy for breast or prostate cancer (given the impact on prognosis); impaired mental status assessed with the Bedside Confusion Scale [30] (score of ≥2 suggesting cognitive impairment); or participation in another study precluding participation in this study.

Intervention

Participants randomly allocated to the intervention arm received access to Loop. Loop is a secure online communication tool for team-based clinical collaboration that enables patients and caregivers to communicate asynchronously with multiple members of the health care team involved in providing their direct care (ie, not individuals hired for the purpose of research), as well as for health care providers including physicians, nurses, and allied health professionals to communicate with each other. The development of Loop followed a user-centered design approach [31] with substantial end-user and stakeholder involvement (including caregivers, health care providers from several specialties, and patients with different conditions). As described by Kurahashi et al [26], this process included initial needs assessments, ethnographic observational studies, and affinity diagramming leading to the development of a prototype. This was followed by simulation activities, usability testing in laboratory and real-world settings (ie, home, clinics, hospitals, offices), and piloting with patients and clinical teams. Screenshot of the Loop interface on desktop computer. Screenshot of the Loop interface on mobile phone. Loop was developed with an intuitive interface to allow for use ideally without prior training. A patient profile and space are created that can be viewed by the patient, health care providers, and caregivers on a computer or mobile phone after logging in with an email address and password (Figures 1 and 2). Each patient’s Loop is a secure space partitioned from other Loops and can be joined only if involved in the patient’s care and authenticated by a study administrator. Health care providers may be a part of multiple patients’ Loops, but patients cannot access the Loops of other patients. On the main page, individuals can write and post text-based messages. All messages posted by any member of a patient Loop can be read and responded to by members of that given Loop. All entries remain on the patient space, allowing for previous posts to be viewed. The messages are threaded in conversations and can be searched using various filters. In addition to posting messages, users may label posts with user-defined “tags” and an “Attention To” feature that specifies individuals to be alerted to a post by a generic email. No updates to the system were made during the trial.
Figure 1

Screenshot of the Loop interface on desktop computer.

Figure 2

Screenshot of the Loop interface on mobile phone.

Recruitment and Study Procedures

In order to ensure that there would be at least one health care provider on each team, medical/radiation oncologists or palliative care physicians were recruited first (“initiating physician”), randomized, and patients from their practices were then approached in clinic prior to appointments or over the phone. Physicians at the two study sites were notified of the study through announcements at educational rounds, and physicians who expressed interest in participating were followed up with directly. All participants provided consent and were asked to complete a baseline demographics form and an internally developed 11-item survey on participants’ access, comfort, and usage of computers and Internet [32]. Initiating physicians and patients in the intervention arm were then invited to register on Loop. Registration involved completion of a form requesting participant name, address, and, in the case of health care providers, professional license number. Once registered on the tool, an intervention patient was considered active in the study. Through the tool, patients could invite family caregivers and additional health care providers to join their Loop. Study administrators contacted, explained the study, obtained consent from all additional members of a patient Loop prior to registration, and posted an introductory message welcoming participants. Study administrators were part of patient Loops only for the purpose of providing assistance during registration and with using the tool, as requested. When a patient was no longer part of the study, Loops remained open for 2 weeks to allow for message exchange records to be exported and saved. Use of the tool and the type of communication that could occur on the tool were not prescribed. The intervention protocol did not specify intent to replace existing care practices or methods of communication; the intervention was additive. As Loop was not meant to be used for urgent communication, this was reinforced during the consent process and in the Loop terms of use. Recruitment of initiating physicians and their patients was conducted similarly in the control arm.

Usage of Loop

Usage of the intervention was evaluated from message exchange transcripts and audit data from Loop. Data included time to registration on Loop from consent date, number of participants who registered on Loop, number of messages exchanged, number of times additional features (Attention To, Tagging) were used, and number of views and posts by participants. We conducted a content analysis of Loop messages to summarize the content of messages on patient Loops. Two coders independently reviewed messages exchanged in each patient Loop and assigned categories thematically that emerged from the data (see Multimedia Appendix 2 for coding framework and definitions) [33]. Categories were assigned to messages and any responses or follow-up posts. Categories were assigned only once per Loop and not quantified. If multiple categories were perceived in a single post, then each was included once as a category identified in that particular Loop. Messages posted by administrators to welcome team members were excluded.

Outcomes

The primary feasibility outcomes were participant recruitment rate and implementation fidelity defined as the proportion of participants who were randomized, completed the baseline demographics and computer usage questionnaire, and if in the intervention arm, registered on Loop, with ≥70% completion indicating feasibility success. This threshold was selected as an adequate threshold to justify further study and has been suggested previously in the literature [34]. Secondary outcomes were measured using standardized instruments to assess the impact of Loop. Mean difference over the course of the study (from baseline to months 1, 2, and 3) for each instrument was calculated. Patient-reported continuity and coordination of care was measured with the 8-item Picker Ambulatory Cancer Care Survey (Picker) Continuity and Coordination subscale questionnaire. The Picker scale is scored by summing absolute positive responses, divided by the total number of responses (scores range from 0-100, with higher scores being better), and a minimal clinically important difference of 10 points has been previously found to be significant [35,36]. The Palliative care Outcomes Scale (POS) was used to assess patient-reported quality of care and well-being. The POS is a 12-item self-administered questionnaire (total scores range from 0-40, with higher scores being worse); a difference of one point on each item is considered clinically meaningful [37-39]. The Edmonton Symptom Assessment Scale (ESAS) is a 9-item, patient-reported questionnaire of symptom intensity (pain, tiredness, nausea, depression, anxiety, drowsiness, appetite, well-being, shortness of breath) with each item rated from 0 (worst) to 10 (best). Individual symptoms are summed for the Total Symptom Distress score (ranging from 0-90, with higher scores indicating worse symptom distress) [40]. Health care utilization was measured as number of visits to the emergency department and number of hospitalizations and was self-reported at each monthly assessment using the Ambulatory and Home Care Record [41]. Data were collected monthly for 3 months from baseline (four time points). Questionnaires were distributed electronically using online surveys emailed to study participants via Research Electronic Data Capture version 6.16.7 [42], a data management system hosted at the Applied Health Research Center of St. Michael’s Hospital, Toronto, Ontario. Patients who did not respond within 1 week were followed up via a reminder email or telephone call and were considered lost to follow-up if not reachable after four contacts. We piloted outcome assessments and survey administration prior to the study. We also collected qualitative interview data at assessments, which will be reported separately.

Randomization and Blinding

This study was designed as a cRCT with initiating physicians recruited first and randomized in order to minimize contamination between study arms. Participating patients were allocated to the study arm to which their initiating physician had been randomized. Randomization was done by a statistical team independent of the study using a computer-generated randomization sequence consisting of permuted blocks of varying size, and assigned initiating physicians in a 1:1 ratio to the intervention and control arms. It was not possible to blind patients completely to study arm, but control patients provided consent without being informed of the existence of another arm. This was done to minimize bias of control patients basing their decision to participate on study arm assignment. Control patients were informed that they were taking part in a study on patient-physician communication to improve health care delivery and care management. Investigators and initiating physicians were aware of study arm assignment.

Sample Size and Statistical Analysis

A formal sample size calculation was not computed, as this was a pilot study with primary feasibility outcomes. We set a target sample size of 20-25 patients per study arm, which has been previously justified as sufficient for pilot evaluations [34,43] The primary analysis was intention-to-treat with available cases. We did not make adjustments for missing data but secondarily report comparison of data for complete cases (participants who completed outcome assessments at all time points). Descriptive statistics were used to describe each study arm. Analysis compared mean change scores and unadjusted differences in mean change scores on the preliminary effectiveness outcomes between study arms. Statistical tests of difference were not conducted since the study was not powered to undertake hypothesis testing. Baseline patient and family caregiver characteristics by treatment arm. aAge-adjusted Charlson Comorbidity Index is a measure of comorbidity based on risk of mortality. The score is weighted by age, increasing for each decade over age 40 [44]. bECOG scale is scored from 1-5 with 1 being well and 4 indicating complete disability. A value of 5 indicates death. cMean summed scores are presented for POS with a maximum score of 40. Higher scores indicate worse quality of care. dThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care. eMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress.

Ethics

All participants provided written, informed consent to participate. Research Ethics Boards of the University Health Network, Mount Sinai Hospital, University of Toronto, and the Community Care Access Centers of Toronto, Ontario, approved the study.

Results

We recruited 10 palliative care physicians and 9 medical oncologists and sequentially randomized 10 to the intervention arm and 9 to the control arm. We assessed 127 patients for eligibility of whom 94 were eligible. We recruited 24 patients each to the intervention and control arms. Figure 3 shows the randomization of clusters (initiating physicians), reasons patients declined to participate in the study, and patient follow-up. In each arm, the baseline questionnaire was completed by 21 patients. In the intervention arm, 18 family caregivers participated and in the control arm, 8 family caregivers participated. There were two instances of initiating physicians from the intervention arm serving as additional health care providers on other intervention patient Loops.
Figure 3

Participant Flow Diagram.

Between arms, there was minimal difference between patients on demographic characteristics, with some modest discrepancies resulting from small sample size (Table 1). There was differential distribution of patients’ primary cancer diagnoses by arm reflecting differences in clinical subspecialty of the participating physicians: lymphoma (6 in intervention vs 0 in control), breast (1 in intervention vs 10 in control) and lung (3 in intervention vs 6 in control). There was minimal difference at baseline in comorbidity and performance status as measured with the ECOG score. Participants were comfortable with using computers and less so Internet-enabled devices (tablets and mobile phones), as described in Multimedia Appendix 2. Initiating physicians in both arms showed similar demographic and practice characteristics (Table 2). All were from academic settings, and most had an alternative payment plan fee structure.
Table 1

Baseline patient and family caregiver characteristics by treatment arm.

CharacteristicsIntervention arm (n=21)Control arm (n=21)
Patients
Age in years, mean (SD)60 (12.8)59.5 (13.8)
Female sex, n (%)13 (61.9)16 (76.2)
Primary cancer site, n (%)
Breast1 (4.8)10 (47.6)
Colorectal2 (9.5)1 (4.8)
Lung3 (14.3)6 (28.6)
Prostate2 (9.5)0
Ovarian01 (4.8)
Thyroid2 (9.5)0
Lymphoma6 (28.6)0
Melanoma01 (4.8)
Brain1 (4.8)0
Other4 (19.0)2 (9.5)
Annual household income in CDN$, n (%)
$0-$21,9992 (9.5)4 (19.1)
$22,000-$49,9992 (9.5)2 (9.5)
$50,000-$89,9997 (33.3)4 (19.1)
>$90,0004 (19.1)5 (20.8)
Prefer not to disclose6 (28.6)6 (28.6)
Primary language, n (%)
English20 (95.2)20 (95.2)
Other1 (4.8)1 (4.8)
Age-adjusted Charlson Comorbidity Index, mean (SD)a5.2 (2.5)5.8 (1.9)
Caregiver, n (%)
Yes4 (19.1)6 (28.6)
No17 (81.0)15 (71.4)
Highest education attained, n (%)
Primary school
High school4 (19.1)6 (28.6)
College/University8 (38.1)8 (38.1)
Professional/Graduate degree9 (42.9)7 (33.3)
ECOG score, median (interquartile range)b1.5 (1-2)1 (1-2)
Outcome measures (n=39)
POS, mean (SD)c9.3 (6.8)9.8 (5.4)
Picker Continuity and Coordination subscale, mean (SD)d47.9 (28.5)62.5 (25.3)
ESAS (Total Symptom Distress Score), mean (SD)e21.2 (17.1)23.4 (12.9)
Family caregivers of consented patient participants(n=18)(n=8)
Age in years, mean (SD)57 (15.9)54 (14.6)
Female sex, n (%)9 (60.0)6 (33.3)
Missing, n (%)3 (16.7)
Relationship to patient, n (%)
Spouse7 (38.9)4 (22.2)
Immediate family5 (27.8)8 (44.4)
Other3 (16.7)
Missing data3 (16.7)

aAge-adjusted Charlson Comorbidity Index is a measure of comorbidity based on risk of mortality. The score is weighted by age, increasing for each decade over age 40 [44].

bECOG scale is scored from 1-5 with 1 being well and 4 indicating complete disability. A value of 5 indicates death.

cMean summed scores are presented for POS with a maximum score of 40. Higher scores indicate worse quality of care.

dThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care.

eMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress.

Table 2

Baseline health care provider demographics.

CharacteristicsIntervention arm (n=10)Control arm (n=9)
Initiating physicians
Age in years, mean (SD)44 (7.9)43 (6.1)
Female sex, n (%)5 (50.0)3 (33.3)
Years in health care, mean (SD)16 (8.8)15 (6.5)
Initiating physician profession, n (%)
Medical oncologist4 (40.0)2 (22.2)
Radiation oncologist1 (10.0)2 (22.2)
Palliative care physician5 (50.0)5 (55.6)
Primary practice setting, n (%)
Hospital-based6 (60.0)4 (44.4)
Home-based care4 (40.0)5 (55.6)
Other
Type of practice, n (%)
Community setting
Academic setting10 (100)9 (100)
Practice fee structure, n (%)
Fee-for-service
Alternate payment plan8 (80.0)7 (77.8)
Salaried1 (10.0)1 (11.1)
Other1 (10.0)1 (11.1)
Provides after-hours care, n (%)
Telehealth
Phone support2 (20.0)4 (44.4)
Phone support with visit when needed6 (60.0)5 (55.6)
Other
None2 (20.0)
Additional health care providers identified, N51
Additional health care providersa (n=22)
Profession, n (%)
Family physician1 (4.5)
Nurse4 (18.2)
Case manager1 (4.5)
Palliative care physician4 (18.2)
Medical oncologist5 (22.7)
Naturopath1 (4.5)
Oncology nurse1 (4.5)
Otolaryngologist1 (4.5)
Personal support worker1 (4.5)
Psychiatrist1 (4.5)
Pharmacist1 (4.5)
Physiotherapist1 (4.5)
Additional health care providers who consented and registered on Loop, n (%)16 (72.7)
Additional health care providers identified per patient, mean (range)2.4 (1-5)

aRecruited as part of the intervention arm and who provided consent.

Regarding team assembly in the intervention arm, an average of 3 health care providers, including the initiating health care provider, participated per patient Loop. Patients suggested between 1 to 5 additional health care team members to participate on Loop; 43% (22/51) consented to participate in the study (mean 1.2 per patient [range 0-4]), and 65% (13/20) of patient Loops had an additional health care provider register on the tool (Table 2). In the intervention arm, 83% (20/24) of patients who consented, registered on Loop (Table 3). In terms of health care provider load, the mean number of patient Loops per initiating physician was 1.6 (range 0-7). Registration on Loop required that the baseline questionnaire be completed beforehand. The time from consent to registration on Loop varied considerably, with the mean time to registration being 39 days. Some patients experienced disease worsening between consent and registration, and one patient delayed taking the step to register for 156 days due to personal circumstances. Over the study period, the majority (85%, 17/20) of Loops had message exchanges, with 45% (13/20) having more than six messages exchanged. During the study, there were 358 logins by all participants: 43 on the mobile version and 315 on the desktop version. Patients viewed their Loops more often relative to their number of posts (a difference of 14.3) compared to initiating physicians viewing and posting to their patients’ Loops (a difference of 3.9).
Table 3

Usage of Loop (intervention arm participants, n=24).

Loop usageMean (range) or n (%)
Loop compositiona
Patients who registered on Loop (regardless of baseline questionnaire completion), n (%)20 (83)
Initiating physicians (intervention arm, n=10) who, n (%):
Registered on Loop9 (90)
Used the tool (posted at least 1 message or viewed a patient Loop)7 (70)
Health care providers (including initiating physician) per patient Loop, mean (range)3 (0-5)
Additional health care providers suggested by each patient, mean (range)2.4 (1-5)
Additional health care providers per patient Loop, mean (range)1.25 (0-4)
Patient Loops health care provider is a part of, mean (range)1.6 (0-7)
Family caregivers per patient Loop, mean (range)0.5 (0-1)
Frequency of use of the tool, nLoops, n
Messages exchanged per Loop by registered participants (n=20)
03
1-25
3-53
6-106
>103
Views of a patient’s own Loop by the patient or caregiver (n=20)
00
1-23
3-54
6-105
>108
Posts to a patient’s own Loop by the patient or caregiver (n=20)
06
1-25
3-53
6-104
>102
Views of a patient Loop by an initiating physician (n=9)
02
1-21
3-52
6-104
>100
Posts to all their patient Loops by an initiating physician (n=9)
03
1-25
3-51
6-100
>100
Use of additional features
Time from consent to registration on Loop (days), mean (range)39 (2-156)
Times an issue was tagged, mean (range)1 (1)
Times Attention To feature was used by a patient or caregiver, mean (range)3 (0-14)
Times Attention To feature was used by a health care provider per Loop, mean (range)0.6 (0-3)

aA “Loop” is an aggregation of a patient and/or caregiver and at least the initiating physician allocated to the intervention arm, and registered on the intervention tool.

Content analysis of messages revealed that of Loops with messages exchanged, messages regarding medical care management, reporting of symptoms, and appointment coordination predominated (these categories were identified in 50%, 45%, and 45% of the 20 patient Loops, respectively), while only 10% of the Loops had messages that were prescription-related queries (see Figure 4). No urgent messages were exchanged during the study.
Figure 4

Categories of messages on Loops with messages exchanged.

Baseline health care provider demographics. aRecruited as part of the intervention arm and who provided consent. Participant Flow Diagram. Usage of Loop (intervention arm participants, n=24). aA “Loop” is an aggregation of a patient and/or caregiver and at least the initiating physician allocated to the intervention arm, and registered on the intervention tool. Categories of messages on Loops with messages exchanged. For the primary outcomes, the mean number of patients recruited per initiating physician was similar between study arms (in intervention arm 2.4 [range 0-7] in intervention vs control arm 2.7 [range 0-7]) (Table 4). With respect to implementation fidelity, 88% (21/24) of control patients who consented completed the baseline questionnaire and 75% (18/24) of intervention patients who consented completed the baseline questionnaire, along with registering on the tool. Regarding patient retention, in the intervention arm, 3 patients withdrew due to declining health, 1 patient withdrew because they were no longer interested, and 1 patient died. In the control arm, 1 patient withdrew due to declining health, and 5 patients died. Instrument and item completion were approximately proportional to patient retention. Of patients completing the baseline outcome assessment, 52% (11/21) in the intervention arm and 52% (11/21) in the control arm completed Month 3 outcome assessments.
Table 4

Feasibility outcomes by treatment arm.

Intervention arm (n=24)Control arm (n=24)
Patients from oncology practices, n1813
Patients from palliative care practices, n611
Initiating physicians, n109
Consenting initiating physicians approached who provided at least one patient, n97
Patients who completed baseline and, if in the intervention arm, registered on Loop, n (%)18 (75)21 (87.5)
Patients recruited per initiating physician, mean (SD)2.4 (2.2)2.7 (2.6)
Patients with a family caregiver who participated in study, n188
Teams with an additional health care provider, n13
Results described are based on available cases. Mean change scores and unadjusted difference in change scores between study arms for preliminary effectiveness outcomes are presented in Table 5. At Month 3, there was an increase in Picker scale scores (in intervention arm +10.2 [SD 31.5] vs control arm -1.1 [SD 30.3]), a negligible change in POS (in intervention arm +0.8 [SD 4.4] vs control arm +0.5 [SD 5.4]), and an increase in ESAS Total Symptom Distress score (in intervention arm +2.3 [SD 10.7] vs control arm +3.4 [SD 8.7]). The number of patients with emergency room visits self-reported at baseline was 3 in the intervention arm and 1 in the control arm. At the third month, no visits were reported in the intervention arm and 3 were reported in the control arm. Similar numbers were observed for number of patients with hospitalizations. On complete case analysis, the Picker scale showed a between-arm difference of +18.5 [47.4] in favor of the intervention arm (Table 6).
Table 5

Preliminary measures of effectiveness by treatment arm, available case analysis.

Score at each month in the intervention armScore at each month in the control armMean observed change from baseline (SD) in the intervention armMean observed change from baseline (SD) in the control armUnadjusted difference between change scores (SD)
Picker Continuity and Coordination subscale, mean (SD)a
1 month58.7 (23.0)64.4 (23.3)-3.4 (29.6)-6.7 (11.0)3.3 (31.6)
2 months66.7 (27.4)69.8 (24.7)9.1 (25.1)4.2 (22.8)4.9 (33.9)
3 months63.5 (25.8)60.2 (24.9)10.2 (31.5)-1.1 (30.3)11.4 (43.8)
POS, mean (SD)b
1 month7.5 (5.1)8.3 (5.3)-1.3 (4.0)-1.0 (3.4)-0.3 (5.2)
2 months7.2 (4.5)9.2 (6.0)0.0 (3.2)-0.6 (4.0)0.6 (5.1)
3 months8.2 (4.8)10.3 (6.5)0.8 (4.4)0.5 (5.4)0.4 (7.0)
ESAS (Total Symptom Distress Score), mean (SD)c
1 month14.6 (11.8)24.7 (15.2)-3.0 (9.0)1.4 (12.2)-4.4 (15.2)
2 months15.2 (12.1)21.1 (11.7)-1.6 (9.4)1.1 (8.0)-2.7 (12.4)
3 months19.2 (9.3)23.3 (17.0)2.3 (10.7)3.4 (8.7)-1.1 (13.8)
Patients with an emergency room visit in previous 4 weeks, n
Baseline31
1 month12
2 months02
3 months03
Patients with a hospitalization in previous 4 weeks, n
Baseline30
1 month11
2 months00
3 months03

aThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care.

bMean summed scores are presented for the POS with a maximum score of 40. Higher scores indicate worse quality of care.

cMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress.

Table 6

Complete case analysis preliminary measures of effectiveness by treatment arm.

Score at each month in the intervention armScore at each month in the control armMean observed change from baseline (SD) in the intervention armMean observed change from baseline (SD) in the control armUnadjusted difference between change scores (SD)
Picker Continuity and Coordination subscale, mean (SD)a
Baseline57.8 (27.5)76.8 (21.0)
1 month51.6 (20.5)73.2 (28.3)-6.3 (34.7)-3.6 (11.9)-2.7 (36.7)
2 months62.5 (25.0)76.8 (24.4)4.7 (26.7)0.0 (21.7)4.7 (34.4)
3 months65.6 (60.0)66.1 (29.5)7.8 (36.6)-10.7 (30.1)18.5 (47.4)
POS, mean (SD)b
Baseline8.0 (6.0)8.0 (6.2)
1 month6.5 (3.7)6.0 (5.9)-1.5 (4.4)-2.0 (3.0)0.5 (5.3)
2 months8.3 (3.5)7.1 (5.1)0.3 (3.8)-0.9 (2.0)1.1 (4.3)
3 months8.3 (3.1)8.4 (7.6)0.3 (4.7)0.4 (6.0)-0.2 (7.6)
ESAS (Total Symptom Distress Score), mean (SD)c
Baseline16.8 (10.3)16.8 (12.8)
1 month11.7 (8.0)19.4 (13.2)-5.1 (9.7)2.5 (7.4)-7.7 (12.2)
2 months14.3 (8.3)20.3 (13.6)-2.5 (11.0)3.5 (7.4)-6.8 (13.3)
3 months18.9 (8.4)21.3 (21.4)2.1 (11.6)4.4 (10.3)-2.4 (15.5)
Patients with an emergency room visit in previous 4 weeks, n
Baseline21
1 month11
2 months00
3 months02
Patients with a hospitalization in previous 4 weeks, n
Baseline20
1 month10
2 months00
3 months03

aThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care.

bMean summed scores are presented for the POS with a maximum score of 40. Higher scores indicate worse quality of care.

cMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress.

Feasibility outcomes by treatment arm. Preliminary measures of effectiveness by treatment arm, available case analysis. aThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care. bMean summed scores are presented for the POS with a maximum score of 40. Higher scores indicate worse quality of care. cMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress. Complete case analysis preliminary measures of effectiveness by treatment arm. aThe Picker Continuity and Coordination subscale is a proportion of total number of positive responses to total number of responses. Higher scores indicate the higher perceived continuity of care. bMean summed scores are presented for the POS with a maximum score of 40. Higher scores indicate worse quality of care. cMean summed scores are presented for the ESAS with a maximum score of 90. Higher scores indicate higher symptom distress.

Discussion

Principal Findings

In this pilot cRCT evaluating an online communication tool for clinical collaboration, trial feasibility conditions and implementation goals were met. The study was not powered to observe changes in outcomes between study arms, but we did observe an increase in continuity of care scores in the intervention arm at last follow-up, which was maintained on complete case analysis. Regarding the assembly of teams, though each patient identified at least one additional health care provider, only 65% of patient Loops had an additional health care provider register on the tool. Loop was primarily used for medical care management, symptom-related discussions, and appointment coordination.

Interpretation

As a population with complex care needs [45], the advanced cancer population served as an exemplar patient population in which to evaluate Loop but also proved challenging from a participation standpoint. Although the proportion of eligible patients who consented in this study was slightly higher than two previous studies conducted at the same institution with the same population (38% here vs 10%) [36,46], a number of patients withdrew due to ill health or died over the course of the study. This was expected given the uncertainty in prognosis in this population. Instrument completion rates reflected patient retention rates, indicating that questionnaire administration was feasible despite the nature of the patient population. Loop was designed to connect patients and caregivers to their team of health care providers in a virtual space where communication might be facilitated outside of appointments and across care settings [26]. While we did not assess differences in measures of effect for statistical significance, preliminary Picker scale results appear to support potential for this tool to improve continuity of care in future studies of adequate statistical power. Contextualized with the content of messages, the findings of this study may suggest that there were important needs that could be dealt with between appointments by using the tool, contributing to increased perceptions of continuity of care. The care of patients with complex needs requires a redefining of the relationship between health care providers and patients to a team-based model of care that engages the patient [47]. These patients often have interdependent issues and thus require collaborative approaches to care (negotiated decision making between individuals in a synergistic manner around shared goals) over coordination between providers (alignment of functioning among independent individuals to address common needs) [48]. Here, the greater number of posts by patients over health care providers and the patient-driven content of messages (eg, Updates, Appointment Coordination) are suggestive that in this study, coordination tasks were addressed to some extent but collaboration did not occur. Given these results, we recognize that Loop in isolation did not produce collaboration, and further consideration into building relationships among these teams is required [49]. We further found that assembling the team was difficult in this study, with few health care providers from outside academic practices, who were identified by the patient, agreeing to join. Other studies have found that barriers to health care provider participation and uptake of studies of eHealth tools include lack of provider compensation and perceived worry about the burden of patient overuse [50]. Although this increased burden has not been observed thus far [51], better strategies to improve integration into clinical workflow need to be examined, especially for physicians with large patient rosters. Implementation of incentive schemes, akin to what has been done in the province of Ontario, Canada, for electronic consultations [52] may also improve uptake of eHealth tools, like Loop, into practice. In this study, Loop was intentionally provided to teams without training. We observed that participants were able to understand and use the core functionality of Loop, that is, to post and read messages. We further observed that patients viewed their Loop more often than they posted compared to health care providers, who posted nearly as often as they viewed a Loop. This could be interpreted as showing that patients were more proactive tool users, while health care providers are more likely to wait for notifications before logging in.

Comparison With Previous Studies

While many tools for patient-physician communication exist (frequently as secure email or part of patient portals) [19,53], few have considered the potential value of team-based communication, which is crucial for complex care scenarios or situations requiring ongoing care. At least four studies have evaluated variations of tools connecting patients and caregivers with multiple health care providers. Gulmans et al [25] found that patient groups who used their tool more often tended to have a larger care network (number of professionals registered per patient). Furthermore, Ralston et al [23], in evaluating secure messaging as part of a portal, found that messaging increased proportionally with patient morbidity, which reinforces the suggestion that messaging is of more value in complex care. While our study was too small to examine such associations, these findings support the increased value that eHealth communication tools may have as complexity of team and illness increase. In a study by Hsiao et al [22], as has been noted elsewhere [54,55], participants felt that text-based communication may diminish the therapeutic relationship gained from in-person visits or unstructured voice-based contact (such as telephone). This suggests that such forms of communication should supplement, but not replace, appointments or calls. The ZWIP tool, by Robben et al [24], allows for patient-provider and between-provider communication. Evaluation in frail elderly patients found that use by both patients and providers depended on provider use of the tool. Health care providers considered implementation strategies (such as training to use the tool) “very necessary” to make the most use of ZWIP. This finding may reflect the need for guided implementation to facilitate integration into clinical workflow and to improve the use of Loop.

Limitations

The results of this study should be interpreted within the context of the study’s strengths and limitations. As a pilot study, we aimed for a sample size that was adequate to determine feasibility of implementation. However, this sample size limited the ability to test the effectiveness of the intervention. All health care providers described themselves as very comfortable with computers and worked in fully computerized practices; however, this may not be true of every medical practice, limiting study generalizability. Similarly, the complexity of clinical cases and nature of physician practice may be different elsewhere. While patients with advanced cancer are a prototypical population of patients with complex care needs and have involvement of multiple providers, similar results may not be reflected in other populations. The cRCT design, involving recruitment and randomization of initiating physicians (clusters) sequentially, and their patients prospectively, may have led to selection bias because of differential recruitment rates by provider and differences in their clinical subspecialty. We also observed that more patients died in the control arm than in the intervention arm possibly indicating unmeasured confounding. As use of the tool was voluntary, there is also a risk of confounding by indication, with patients who have more issues needing to use the tool more often, or functionally limited patients using the tool less often.

Future Research

The results of this study will inform the next phase of research, which aims to (1) understand the conditions that affect tool adoption and assembly of teams, (2) understand the relationship between use and outcomes such as continuity and quality of care, and (3) examine the contexts and target populations where the benefits of a tool like Loop may be best realized and where the effort to assemble the care team is justified. We anticipate that by making adjustments to the implementation approach [56] through use of site champions, consideration of strategies to foster team collaboration as a co-intervention with Loop, and considering an initiating physician (the “index provider”) from those additional provider disciplines that were less represented, we may address issues with team assembly and optimize collaboration on Loop. We also expect that with a longer duration of follow-up, as could be done in other patient populations, participants’ comfort with using Loop may improve, and the content of communication may become more oriented towards care planning and decision making, over coordination.

Conclusion

In this study, we found that it was feasible to implement Loop in clinical practice and that the tool may have the potential to improve continuity of care. We observed that Loop messages reflected message categories of medical care management, symptom reporting, and appointment coordination, among others. Usage of the tool suggests that some coordination tasks were improved but further strategies to build collaboration among team members may be needed. As an ongoing goal of eHealth development, the integration of the dynamic components of care (communication and collaboration) with the static repositories of medical records would enable a more seamless provision of health care. However challenging this may be in the current environment of multiple electronic health records across organizations, studying collaborative tools like Loop advances this goal.
  44 in total

Review 1.  The impact of travel on cancer patients' experiences of treatment: a literature review.

Authors:  S Payne; N Jarrett; D Jeffs
Journal:  Eur J Cancer Care (Engl)       Date:  2000-12       Impact factor: 2.520

2.  Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support.

Authors:  Paul A Harris; Robert Taylor; Robert Thielke; Jonathon Payne; Nathaniel Gonzalez; Jose G Conde
Journal:  J Biomed Inform       Date:  2008-09-30       Impact factor: 6.317

3.  Secure web messaging in a pediatric chronic care clinic: a slow takeoff of "kids' airmail".

Authors:  Allen L Hsiao; Alia Bazzy-Asaad; Concettina Tolomeo; Diana Edmonds; Beverly Belton; Andrea L Benin
Journal:  Pediatrics       Date:  2011-01-10       Impact factor: 7.124

4.  An internet-based self-management program with telephone support for adolescents with arthritis: a pilot randomized controlled trial.

Authors:  Jennifer N Stinson; Patrick J McGrath; Ellen D Hodnett; Brian M Feldman; Ciaran M Duffy; Adam M Huber; Lori B Tucker; C Ross Hetherington; Shirley M L Tse; Lynn R Spiegel; Sarah Campillo; Navreet K Gill; Meghan E White
Journal:  J Rheumatol       Date:  2010-07-01       Impact factor: 4.666

5.  Family caregiving skill: development of the concept.

Authors:  K L Schumacher; B J Stewart; P G Archbold; M J Dodd; S L Dibble
Journal:  Res Nurs Health       Date:  2000-06       Impact factor: 2.228

6.  Processes for effective communication in primary care.

Authors:  Saul J Weiner; Beth Barnet; Tina L Cheng; Timothy P Daaleman
Journal:  Ann Intern Med       Date:  2005-04-19       Impact factor: 25.391

Review 7.  The impact of patient-physician web messaging on healthcare service provision.

Authors:  David McGeady; Jaakko Kujala; Karita Ilvonen
Journal:  Int J Med Inform       Date:  2006-12-22       Impact factor: 4.046

8.  Patients' supportive care needs beyond the end of cancer treatment: a prospective, longitudinal survey.

Authors:  Jo Armes; Maggie Crowe; Lynne Colbourne; Helen Morgan; Trevor Murrells; Catherine Oakley; Nigel Palmer; Emma Ream; Annie Young; Alison Richardson
Journal:  J Clin Oncol       Date:  2009-11-02       Impact factor: 44.544

Review 9.  The effect of patient portals on quality outcomes and its implications to meaningful use: a systematic review.

Authors:  Clemens Scott Kruse; Katy Bolton; Greg Freriks
Journal:  J Med Internet Res       Date:  2015-02-10       Impact factor: 5.428

10.  A tutorial on pilot studies: the what, why and how.

Authors:  Lehana Thabane; Jinhui Ma; Rong Chu; Ji Cheng; Afisi Ismaila; Lorena P Rios; Reid Robson; Marroon Thabane; Lora Giangregorio; Charles H Goldsmith
Journal:  BMC Med Res Methodol       Date:  2010-01-06       Impact factor: 4.615

View more
  11 in total

1.  Disruption or innovation? A qualitative descriptive study on the use of electronic patient-physician communication in patients with advanced cancer.

Authors:  Teja Voruganti; Amna Husain; Eva Grunfeld; Fiona Webster
Journal:  Support Care Cancer       Date:  2018-03-04       Impact factor: 3.603

Review 2.  How technology can improve communication and health outcomes in patients with advanced cancer: an integrative review.

Authors:  Natasha Ansari; Christina M Wilson; Mallorie B Heneghan; Kathie Supiano; Kathi Mooney
Journal:  Support Care Cancer       Date:  2022-04-12       Impact factor: 3.603

Review 3.  Interventions for interpersonal communication about end of life care between health practitioners and affected people.

Authors:  Rebecca E Ryan; Michael Connolly; Natalie K Bradford; Simon Henderson; Anthony Herbert; Lina Schonfeld; Jeanine Young; Josephine I Bothroyd; Amanda Henderson
Journal:  Cochrane Database Syst Rev       Date:  2022-07-08

Review 4.  Routine provision of feedback from patient-reported outcome measurements to healthcare providers and patients in clinical practice.

Authors:  Chris Gibbons; Ian Porter; Daniela C Gonçalves-Bradley; Stanimir Stoilov; Ignacio Ricci-Cabello; Elena Tsangaris; Jaheeda Gangannagaripalli; Antoinette Davey; Elizabeth J Gibbons; Anna Kotzeva; Jonathan Evans; Philip J van der Wees; Evangelos Kontopantelis; Joanne Greenhalgh; Peter Bower; Jordi Alonso; Jose M Valderas
Journal:  Cochrane Database Syst Rev       Date:  2021-10-12

5.  Use of Mobile Messaging System for Self-Management of Chemotherapy Symptoms in Patients with Advanced Cancer.

Authors:  Connie Jaenicke; Dan Greenwood; Kathleen Nelson; Mark Klein; Bev Foss; Lara Carson; Kim Smith; Evan Mariash
Journal:  Fed Pract       Date:  2019-08

6.  A Clinical Communication Tool (Loop) for Team-Based Care in Pediatric and Adult Care Settings: Hybrid Mixed Methods Implementation Study.

Authors:  Amna Husain; Eyal Cohen; Raluca Dubrowski; Trevor Jamieson; Allison Miyoshi Kurahashi; Bhadra Lokuge; Adam Rapoport; Stephanie Saunders; Elaine Stasiulis; Jennifer Stinson; Saranjah Subramaniam; Pete Wegier; Melanie Barwick
Journal:  J Med Internet Res       Date:  2021-03-03       Impact factor: 5.428

7.  The Perceived Ease of Use and Usefulness of Loop: Evaluation and Content Analysis of a Web-Based Clinical Collaboration System.

Authors:  Allison M Kurahashi; Jennifer N Stinson; Margaret van Wyk; Stephanie Luca; Trevor Jamieson; Peter Weinstein; Joseph A Cafazzo; Bhadra Lokuge; Eyal Cohen; Adam Rapoport; Amna Husain
Journal:  JMIR Hum Factors       Date:  2018-01-09

8.  Technology-Based Interventions for Cancer Caregivers: Concept Analysis.

Authors:  Zhaohui Su; Xiaoshan Li; Dean McDonnell; Andrea A Fernandez; Bertha E Flores; Jing Wang
Journal:  JMIR Cancer       Date:  2021-11-16

Review 9.  The Extent of Engagement With Telehealth Approaches by Patients With Advanced Cancer: Systematic Review.

Authors:  William Goodman; Anne-Marie Bagnall; Laura Ashley; Desiree Azizoddin; Felix Muehlensiepen; David Blum; Michael I Bennett; Matthew Allsop
Journal:  JMIR Cancer       Date:  2022-02-17

Review 10.  Digital Health Technologies Enabling Partnerships in Chronic Care Management: Scoping Review.

Authors:  Carolina Wannheden; Matilda Åberg-Wennerholm; Marie Dahlberg; Åsa Revenäs; Sara Tolf; Elena Eftimovska; Mats Brommels
Journal:  J Med Internet Res       Date:  2022-08-01       Impact factor: 7.076

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.