Literature DB >> 33887443

Considerations for assessing physical function and physical activity in clinical trials during the COVID-19 pandemic.

Seol Ju E Moon1, Annette DeVito Dabbs2, Andrea L Hergenroeder3, Melissa L Vendetti2, Kristen B Jones2, Bryan M Willey2, Matthew R Morrell4, Christopher C Imes2.   

Abstract

The coronavirus disease-2019 (COVID-19) pandemic has changed the conduct of clinical trials. For studies with physical function and physical activity outcomes that require in-person participation, thoughtful approaches in transitioning to the remote research environment are critical. Here, we share our experiences in transitioning from in-person to remote assessments of physical function and activity during the pandemic and highlight key considerations for success. Details on the development of the remote assessment protocol, integration of a two-way video platform, and implementation of remote assessments are addressed. In particular, procedural challenges and considerations in transitioning and conducting remote assessments will be discussed in terms of efforts to maintain participant safety, maximize study efficiency, and sustain trial integrity. Plans for triangulation and analysis are also discussed. Although the role of telehealth platforms and research activities in remote settings are still growing, our experiences suggest that adopting remote assessment strategies are useful and convenient in assessing study outcomes during, and possibly even beyond, the current pandemic. Trial register and number: ClinicalTrials.gov [NCT03728257].
Copyright © 2021 Elsevier Inc. All rights reserved.

Entities:  

Keywords:  COVID-19; Clinical trials; Physical activity; Physical function; Remote assessments

Year:  2021        PMID: 33887443      PMCID: PMC8055496          DOI: 10.1016/j.cct.2021.106407

Source DB:  PubMed          Journal:  Contemp Clin Trials        ISSN: 1551-7144            Impact factor:   2.226


Introduction

To date, coronavirus disease-2019 (COVID-19), has resulted in over 31 million cases and 560,000 deaths in the U.S. alone [1]. It has led to socio-economic losses across all sectors [2]. It is no exaggeration that COVID-19 has had detrimental effects on global health and its ripple effects have disrupted every aspect of life as we know it. These changes have impacted the conduct of highly choreographed clinical trials, which require timely assessments to maintain scientific validity in evaluating the safety and efficacy of new interventions [3]. COVID-19 restrictions and infection mitigation strategies have greatly restricted participation in face-to-face clinical trial research activities. However, when possible, it is critical to support continuing clinical trials using creative and thoughtful approaches to facilitate continued evaluation of intervention impact for improving health and preventing disability. Rather than halting clinical trials, investigators are considering ways to conduct clinical trials remotely when possible, while maintaining safety, following good clinical practice standards, and preserving the integrity of the trial. The first priority when considering any change in a study protocol is to protect the rights, safety and well-being of trial participants [4]. This is particularly important during a pandemic for trials like ours that include lung transplant recipients (LTR) whose immunosuppression contributes to increased vulnerability to emerging infectious agents [5]. Ironically, our ongoing randomized controlled trial, Lung Transplant GO (LTGO): Improving Self-Management of Exercise After Lung Transplantation, was designed to evaluate a behavioral exercise intervention that provides individualized exercise training for LTR in their homes via a telehealth platform. While the intervention is delivered remotely via telehealth, many of our primary outcome measures were designed for face-to-face assessment. Based on the guidance of our university, state-wide health officials, and Human Research Protection Office, our study was able to continue remotely with appropriately modified protocols. While our original protocol already included strategies to recruit participants, collect self-report data, and deliver our telehealth intervention remotely, there were other factors to consider before converting to remote assessments for our primary study outcomes: physical function and physical activity. The transition of physical function and activity outcomes to remote assessment were guided by recent recommendations for conducting remote assessments in clinical trials during COVID-19 [6] and preserving trial integrity during the pandemic [7,8]. These considerations included the availability of remote counterparts to face-to-face measures with established reliability and validity; the feasibility and efficacy of performing the assessment remotely; or the need to consider alternative approaches for remote assessment such as the collection of self-reported data that could be used as surrogate measures. We also weighed potential disadvantages of performing assessments of physical function and activity remotely, such as safety concerns like falling, and availability of equipment, against the potential advantages of maintaining health and safety during pandemic. Therefore, the purposes of this paper are to share our experience adapting outcome measures to the remote environment to allow the LTGO trial to proceed during the COVID-19 pandemic, and to inform other researchers who have similar outcomes so that remote measurements are valid, safe, and trial integrity is sustained.

LTGO trial

LTGO, a multi-component telerehabilitation exercise intervention, is currently being evaluated in an ongoing randomized, controlled trial (National Clinical Trial: 03728257). LTR are randomized to an enhanced usual care group or the LTGO intervention, a behavioral exercise intervention that provides individualized exercise training and behavioral coaching for LTR in their home via a telehealth platform. The LTGO intervention is delivered in two phases. Phase 1: Intensive home-based exercise training and behavioral coaching sessions via a videoconferencing platform; and Phase 2: Transition to exercise self-management with a behavioral contract followed by 3-monthly telephone sessions to provide behavioral coaching and exercise reinforcement. The primary study aims are to evaluate efficacy of LTGO in improving physical function and physical activity from baseline to 3 and 6 months. Eligibility criteria are shown in Table 1 .
Table 1

Inclusion and exclusion criteria for study population.

Inclusion criteriaExclusion criteria

Age: ≥18 years

Greater than 1-year post-transplant hospital discharge

Discharged home after lung transplant surgery

Chronic conditions that may severely limit participation in exercise training (e.g., cardiac, musculoskeletal or cognitive impairments)

>4 weeks post-surgery

No home internet or smart device with Bluetooth capabilities

Physician report of difficulty walking 0.25 mile or climbing 10 steps without resting

Medical issue precluding participation

Medical monitor approves patient eligibility for participation

Unwilling to be screened or approached about research

Inclusion and exclusion criteria for study population. Age: ≥18 years Greater than 1-year post-transplant hospital discharge Discharged home after lung transplant surgery Chronic conditions that may severely limit participation in exercise training (e.g., cardiac, musculoskeletal or cognitive impairments) >4 weeks post-surgery No home internet or smart device with Bluetooth capabilities Physician report of difficulty walking 0.25 mile or climbing 10 steps without resting Medical issue precluding participation Medical monitor approves patient eligibility for participation Unwilling to be screened or approached about research

Outcome measures: pre and during COVID-19

The measures of physical function and physical activity were examined to determine which outcomes could be assessed remotely, and if not suitable for remote assessment, what other surrogate measures were available as shown in Table 2 .
Table 2

Summary of measures.

Pre-COVID-19 measures
During COVID-19 measure
Outcome variablesMeasureRemote measure
Physical functionWalking ability6 Minute Walk Test assessed in person: distance walked in feet/ 6 min assessed (longer distance indicates better walking ability)30-Second Sit-to-stand Test (30s-STS) assessed remotely via Zoom
BalanceBerg Balance Scale assessed in person: 14-items related to balance using a 5-point ordinal scale from 0 (lowest level) to 4 (highest level); Total scores range from 0 to 56 (higher scores indicate better balance)Berg Balance Scale assessed remotely via Zoom
Lower body strength30-Second Sit-to-stand Test assessed in person: the number of sit-stand repetitions performed in 30 s (more repetitions indicate better lower body strength)30s-STS assessed remotely via Zoom
Maximal exercise capacityCardiopulmonary Exercise Test performed in person in specialized laboratory: watts and blood pressure at isoworkloadUnable to perform during COVID-19 remotely due to potential aerosolized spread
Quality of lifeSt. George Respiratory Questionnaire (SGRQ) administered in person or remotely: 50-item to assess overall health, daily life, and perceived wellbeing (higher scores indicate more limitations and poorer quality of life.)SGRQ administered remotely via phone
Physical activitySteps per dayFitbit: total steps/dayFitbit: total steps/day
Time spent in light, moderate, & vigorous activitiesActigraph GT3X: mins/day per level of activityActigraph GT3X: mins/day per level of activity
International Physical Activity Questionnaire Short Form administered remotely via phone
Summary of measures.

Physical function

Walking ability

The 6 Minute Walk Test (6MWT) is a standardized, well-validated measure of functional capacity [9]. Testing is conducted according to American Thoracic Society Guidelines and participants are asked to walk on a ~ 37.56-m track as far as possible in 6 min while an assessor observes and keeps time. Conducting the 6MWT remotely would require long, unobstructed hallways beyond the space of most places of residence. A PubMed search for “remote 6MWT” revealed several promising options for using a smartphone to conduct the 6MWT [10,11]. However, the apps and algorithms were not available for download by the public or by the research team, preventing their use for remote 6MWTs for our study. Furthermore, the 4-m gait speed (4MGS) test is strongly correlated with the 6WMT (r = 0.85) test [12]. However, the test involves a 1-m acceleration zone, a 4-m testing zone, and 1-m deceleration zone (6 m = 19.7 ft) [13]. Incorporating mobility assessments that required the procuring of a certain distance raised concerns of feasibility and accuracy in measurement given the uncontrolled remote setting (e.g., lack of space in participant's home, difficulty in determining an accurate distance visually via video conferencing). The 30-s sit-to-stand test (30-s STS) measures lower body strength and has been recommended as an alternative when conditions do not permit the 6MWT [14]. Multiple studies in different samples found that the 30-s STS, was moderately correlated with the 6MWT, including correlations of r = 0.528 to 0.660 (p < .001) in a sample of healthy young adults [15]; adults with chronic obstructive pulmonary disease [16]; and pulmonary hypertension [17]. Since the 30-s STS test was already included in our protocol for assessment of lower body strength, it was selected as a reliable remote surrogate measure for walking ability.

Balance

The Berg Balance Scale (BBS) [18,19] tests the participant's ability to handle tasks that require balance (e.g., sitting to standing, placing alternate foot on stool). We modified the protocol to perform the test remotely via two-way video; methods are further described in the remote Physical Function Assessment (PFA) protocol below.

Lower body strength

The 30-s STS measures lower body strength [20]. A trained assessor instructs the participant to: 1) sit in the middle of a chair (17-in. height, with a straight back without armrests); 2) place hands on the opposite shoulder crossed at the wrists; 3) keep feet flat on the floor and back straight. The participant is asked to rise to a full standing position, sit down on the chair and repeat this move for up to 30 s. The protocol was modified to perform the 30-s STS remotely via two-way video; methods are further described in the remote PFA protocol below.

Maximal exercise capacity

A Cardio Pulmonary Exercise Test (CPET) is performed in a laboratory setting to determine physical function (maximum exercise watts for age, height, weight, and gender) based on isoworkload measurements of cardiopulmonary metabolic responses at the point of participant's volitional exhaustion [21]. Due to COVID-19 restrictions, the clinical laboratory suspended non-essential pulmonary testing forcing us to exclude the direct measure of maximal exercise capacity until laboratory testing can be resumed.

Quality of life

The St. George Respiratory Questionnaire (SGRQ) [22], a 2 part, 50-item, self-report questionnaire administered in person or by phone to assess overall health, daily life, and perceived wellbeing. Administration of the SGRQ already includes remote data collection options, so no modification was required.

Physical activity

Steps per day (SPD)

All participants are provided a Fitbit Charge 3 (Fitbit Inc., San Francisco, CA). The number of days that the participant wears the Fitbit is used to document adherence to self-monitoring of walking goals. SPD are already monitored remotely, so no modification was required.

Time spent in light, moderate, and vigorous activities

All participants are provided an Actigraph GT3X (ActiGraph, LLC., Pensacola, FL) and accelerometer to measure physical activity for a 7-day period at baseline, 3 and 6 months. Participants are instructed to: 1) wear the Actigraph on their waist during waking hours (except during imaging studies such as bronchoscopy, bathing or showering), 2) keep a paper activity diary, and 3) return the device and diary by mail in a pre-paid envelope. The Actigraph provides tri-axial vector data in activity units, metabolic equivalent tasks (METs), or kilocalories. Validity has been established in people with chronic lung disease [23]. Physical activity is reported as average minutes spent per day in light (<3 METs) and moderate/vigorous (≥3 METs) activity at each time point. Due to challenges of not having a centralized work location, limited face-to face participant contact, and increased turnaround time for receiving devices from participants during COVID-19, Actigraph data were not always available. Therefore, the International Physical Activity Questionnaire Short Form (IPAQ-S) [24] was added as a surrogate measure, administered over the telephone, of time spent in certain activities during the pandemic. The IPAQ-S is a 7-item, self-reported physical activity assessment completed either by telephone or mailed. The data can be expressed as a categorical score (low, moderate, or high) or a continuous score of metabolic equivalent-minutes per week. The psychometric evaluations report good reliability, with a test-retest Spearman's coefficient of 0.64 for the telephone version and 0.79 for the self-administered version, which we adapted in this study. Spearman's coefficient between accelerometer derived activity and self-reported activity of r = 0.42 to 0.52 for total physical activity and 0.77 to 0.86 for greater than 150 min of physical activity/week provided evidence of its validty [24].

Development of the remote PFA protocol

Video platform

We selected Zoom as the two-way, remote assessment platform to expand the reach, scalability, and performance of physical assessments (30-s STS and BBS) during COVID-19 and/or situations where participants are limited in leaving their homes or where access to care is restricted such as in rural areas [25]. Zoom is Health Insurance Portability and Accountability Act of 1996 compliant through applying mandatory account settings to customers' accounts. Zoom protects and encrypts all audio, video, and screen sharing data, and does not have access to identifiable personal health information. Other notable Zoom security features are shown in Table 3 . Quality control monitoring for reliability and reproducibility was performed before conducting assessments remotely via videoconferencing in participants' homes. Practice PFA sessions were conducted between assessors with other research staff with no knowledge of the assessment as participants. Training focused on providing consistent participant instructions and determining scores based on standard criteria for consistency and reproducibility. For quality assurance, each assessor was evaluated by a gold-standard reviewer prior to taking the role as an assessor.
Table 3

Considerations for video conference settings using Zoom.

Points to consider in setting video conferences
Meeting invite/link/settings

Use Health Insurance Portability and Accountability Act of 1996 compliant account (i.e., require encryption for 3rd party endpoints; recording disclaimer)

Disable participants to join meeting before host

Enable waiting room feature (allows host to admit participants individually)

Enable participants to join meeting online (using browser) to bypass the application download process

Provide contact of videoconferencing software company to troubleshoot problems to participants

Chats

Enable all meeting participants to send and receive messages

Disable participants from saving chats or messages to minimize breach of confidentiality and privacy

Recording

Enable recording disclaimer to inform all participants that the session is being recorded

Security

For 2-person meetings, disable peer to peer connection for increased security

Create meeting invites with unique meeting identification and passcode (for security purposes)

Only the host should be able to record meetings to minimize breach of confidentiality

Disable saving in cloud (save recordings on local computer)

Enable only host and co-hosts to have meeting controls

Confidentiality and privacy

Disable display of profile pictures

Considerations for video conference settings using Zoom. Use Health Insurance Portability and Accountability Act of 1996 compliant account (i.e., require encryption for 3rd party endpoints; recording disclaimer) Disable participants to join meeting before host Enable waiting room feature (allows host to admit participants individually) Enable participants to join meeting online (using browser) to bypass the application download process Provide contact of videoconferencing software company to troubleshoot problems to participants Enable all meeting participants to send and receive messages Disable participants from saving chats or messages to minimize breach of confidentiality and privacy Enable recording disclaimer to inform all participants that the session is being recorded For 2-person meetings, disable peer to peer connection for increased security Create meeting invites with unique meeting identification and passcode (for security purposes) Only the host should be able to record meetings to minimize breach of confidentiality Disable saving in cloud (save recordings on local computer) Enable only host and co-hosts to have meeting controls Disable display of profile pictures

Equipment and space needs

To perform remote assessments, participants need a laptop or tablet with a functioning video camera and speakers, a chair with arms, an armless chair, a ruler, tape, a cup, and a step-stool. Adequate space is also required to conduct the assessments; ideally near a wall for safety that allows for roughly 10 ft between the participant and the laptop/tablet so the assessors are able to view the participant clearly from head to toe.In order to ensure that a participant has the necessary equipment and space, a pre-assessment protocol was developed (described below). However, as the remote assessments occurred, scripts and procedures were revised to address unanticipated problems/issues. For example, some participants did not have a ruler in the home, or the ruler markings were difficult for the assessor to see on the video camera during the reach test. To address this problem, we created a paper ruler with bold markings that could be visible to the remote assessor from a distance and included it in the mailed assessment packet. Similarly, instead of displaying the Borg Rating of Perceived Exertion (RPE) scale via screen-sharing, we included this scale in the mailed assessment packet.

Participant instructions

For in-person PFA the assessor sets-up all of the equipment, however for remote assessments, participants are provided additional instructions before and during the assessment to set-up equipment themselves. Once set-up is complete, the assessor demonstrates each test and observes the participant for proper procedure and adequacy of the screen view. The assessor frequently needs to instruct the participant to adjust the distance and angle of the camera so that their entire body (head to toe) is in view to allow the assessor to score the assessment or ensure safety. When the view of the participant is not clear to the assessor despite equipment adjustments, the participants are asked to describe their surroundings such as the distance from the wall.

Safety

In addition to the changes required to make remote testing feasible, the remote PFA protocol adopted additional precautions to ensure participant safety. The original in-person safety protocol outlined circumstances (e.g., signs and symptoms) that warranted stopping the PFA, activating a rapid response team, and notifying the principal investigator (PI) and/or key personnel. In order to make the existing safety protocol applicable for remote PFA, changes included adding instructions to dial 911 in case of an emergency and the national hotline and crisis line in the event of suicidal ideation. Additional safety precautions for the remote setting included assessing whether participants planned to exercise alone, and if so, their comfort level in doing so. This allows the assessor to know who to inform in the event of an emergency. Furthermore, prior to initiating the 30-s STS, the participant's ability to safely stand up from chair without help is assessed. If the participant is unable to stand independently, the test is stopped for safety purposes and the safety protocol is initiated by notifying the PI for further risk assessment and instructions. For the BBS, additional verbal instructions and demonstration of most test-items are included. For tests that require freestanding, the participant is instructed to stand near, but not touching a wall, for safety purposes, in case of loss of balance. If space or room layout prevents standing near a wall, free-standing is permitted if the participant verbalizes comfort and safety. In addition, instructions for suspending the test or initiating the safety protocol are included, such as the participant scores too low on a task or safety is a concern. To prevent possible fatigue during testing, we offer frequent breaks and monitor participant's condition during testing.

Addition of an observer

When assessors began to train for remote PFA protocol in preparation for remote assessment, we realized the assessor was juggling multiple responsibilities including correct equipment set-up in participant's homes, providing instructions, assessing the participant's understanding of instructions, conducting the test, assessing the participant's safety throughout testing, evaluating the performance, and scoring the test. This extensive list of activities often overwhelmed the assessor and introduced testing errors, thus, the observer role was created. The observer's primary responsibility is to assess the testing environment and ensure the assessor has an appropriate view of the participant to score the test. During each PFA session, the observer notes potential issues with: 1) camera positioning; 2) assessment equipment; 3) lighting; 4) video technology; and/or 5) participant safety. The observer serves to assess the quality of the testing and to optimize testing conditions by providing the participant with additional instructions if warranted.

Implementation of remote physical assessments

The steps to prepare and implement remote PFAs are shown in Fig. 1 . First, after a participant enrolls in the study, a research staff member contacts the participant for scheduling and pre-assessment checks. Logistics of remote setup and preparation instructions are discussed by phone and email. Once scheduled, the participant is introduced to the video software (Zoom), provided with the assessment meeting link and passcode, and informed that a box will be mailed including all study materials for the scheduled visit. One or two days prior to the assessment, a reminder phone call and pre-assessment video check is performed with the participant. Any issues with video conferencing software are addressed, including internet connectivity, audio and video quality. The participant is guided on how to use the items in the mailed box, including Fitbit application download and set-up, and a testing location is chosen in the home based on the space needs. On the day of testing, the assessor and the observer sign-in to Zoom. The participant joins the video conference by clicking the Zoom link and entering the unique passcode and prompted to provide consent to record the meeting. The assessment session is recorded and starts with a brief introduction and health questions to determine if any new health or safety concerns have arisen. Once the participant is cleared to continue, the assessor begins testing of the 30-s STS and BBS. For each testing item, a verbal instruction is provided followed by demonstration. The participant performs the test and the assessor times with a stopwatch and scores the performance accordingly. The Borg RPE rating is obtained before and after the 30-s STS subtests. Throughout the assessment, the assessor and participant make adjustments to the camera and equipment and visibility is optimized. When audio or video disconnection occurs during a test item, the participant is asked to repeat the item. If the problem persists, the item is skipped. Both the assessor and observer score the participant's performance and the observer makes additional notes for each test item about concerns that might impact the quality of the tests. After the zoom session the assessor and observer debrief about any test items or areas in which scoring may have been impacted due to limited visualization, lack of equipment, or any other reasons. If there is a disagreement about any limitations that may impact scoring, the assessor and the observer may ask another research team member to review the recorded session to reach consensus. The day after the remote PFA session the participant receives a reminder phone call to wear study devices, record wear-time for Actigraph, to sync their Fitbit data daily, and to assist with any issues with device setup and use. A similar call is made seven days after the remote PFA with additional instructions to return the Actigraph device and diary.
Fig. 1

Steps of the remote physical function assessment implementation process.

Steps of the remote physical function assessment implementation process.

Challenges and considerations for future

Transitioning from in-person to remote PFA testing required us to recognize challenges, barriers, and considerations (Table 4 ). Given the nature of remote testing, one of our biggest challenges was overcoming technical issues. The most frequently experienced problems were starting and maintaining connectivity during video conferencing and loss of audiovisual features in the course of assessment. Although pre-assessments were helpful in preparing for the assessment, technical issues still occurred on the day of assessment and had to be resolved in real-time (e.g. calling participant's phone when audio failed, reconnecting to video conference link, using a mobile hotspot or contacting the internet provider to resolve weak or unstable connectivity). If an issue was not resolved promptly, the assessment was rescheduled for a later date.
Table 4

Key considerations in planning remote physical assessments.

Considerations
Researcher's openness

Be creative, flexible and open-minded when it comes to adapting to participants' needs in the remote testing environment

Surrogate measures

Explore literature for evidence of possible surrogate measures to measure study outcomes (recently published examples include the rapid reviews by Holland and colleagues [26] and Houchen-Wolloff and colleagues [12])

Quality assurance

Practice and perform repeated quality assurance reviews before performing remote testing with participants to ensure that testing procedures are reproducible and consistent between participants (e.g., some participants might have more space for testing than other participants)

Participant engagement

Maximize participant engagement by sending frequent appointment reminders, emphasizing study benefits, and using contact and scheduling strategies [27]

Use preferred digital communication channels for efficient communication (e.g., email, phone call, text)

Develop pre-assessment protocols to assist participants with any preparatory tasks prior to assessment an (e.g., review video conferencing software)

Provide follow up calls to assist with device set-up and use

Participant safety

Consider safety protocols that are effective and appropriate in remote versus face-to-face settings (e.g., calling a suicide hotline or 911 for medical emergencies such as suicidal ideation, falls, dyspnea, rather than local resources); consider risk, patient safety and consult IRB for protocol modifications prior to testing if needed

Staffing

Consider the adequacy of staff to complete remote testing (e.g., establishing the role of an observer for remote assessments)

Budget

Consider budget implications to conduct assessments remotely: although remote assessments may reduce costs in areas such as transportation, additional costs should be anticipated for mailing or paying for digital services

Time

Account for the extra time needed for scheduling, remoting testing, and conducting phone surveys for assessors and participants

Account for the increased turnaround time for sending and receiving back mailed devices (e.g., extend device activation time period)

Technical issues

Consider ways to resolve technical issues and potential participant burdens [28]: Use mobile phones (regular/video call) if audio or visual issues occur with video conferencing software; contact internet provider for unstable internet connectivity)

Communication

Maintain current participants' contact information (primary and secondary) for efficient communication

Establish good internal and external communication channel between research staff and participants (e.g., appoint contact person for specific subject/task, use preferred channels and devices)

Key considerations in planning remote physical assessments. Be creative, flexible and open-minded when it comes to adapting to participants' needs in the remote testing environment Explore literature for evidence of possible surrogate measures to measure study outcomes (recently published examples include the rapid reviews by Holland and colleagues [26] and Houchen-Wolloff and colleagues [12]) Practice and perform repeated quality assurance reviews before performing remote testing with participants to ensure that testing procedures are reproducible and consistent between participants (e.g., some participants might have more space for testing than other participants) Maximize participant engagement by sending frequent appointment reminders, emphasizing study benefits, and using contact and scheduling strategies [27] Use preferred digital communication channels for efficient communication (e.g., email, phone call, text) Develop pre-assessment protocols to assist participants with any preparatory tasks prior to assessment an (e.g., review video conferencing software) Provide follow up calls to assist with device set-up and use Consider safety protocols that are effective and appropriate in remote versus face-to-face settings (e.g., calling a suicide hotline or 911 for medical emergencies such as suicidal ideation, falls, dyspnea, rather than local resources); consider risk, patient safety and consult IRB for protocol modifications prior to testing if needed Consider the adequacy of staff to complete remote testing (e.g., establishing the role of an observer for remote assessments) Consider budget implications to conduct assessments remotely: although remote assessments may reduce costs in areas such as transportation, additional costs should be anticipated for mailing or paying for digital services Account for the extra time needed for scheduling, remoting testing, and conducting phone surveys for assessors and participants Account for the increased turnaround time for sending and receiving back mailed devices (e.g., extend device activation time period) Consider ways to resolve technical issues and potential participant burdens [28]: Use mobile phones (regular/video call) if audio or visual issues occur with video conferencing software; contact internet provider for unstable internet connectivity) Maintain current participants' contact information (primary and secondary) for efficient communication Establish good internal and external communication channel between research staff and participants (e.g., appoint contact person for specific subject/task, use preferred channels and devices) Other challenges in preparing and conducting remote PFA were related to participant engagement, participant privacy and safety, and minimizing assessor subjectivity and bias. The PFA protocols and procedures were adapted to overcome such challenges by incorporating various strategies as described in 2, 2.2. Specifically for participant engagement, pre-assessment protocol was developed to address any preparatory difficulties before the assessment, and follow-up calls were made after each visit to address outstanding questions and assist with device set-up and use. For effective and efficient communication with participants it is essential to have accurate and updated participant contact information. To ensure successful delivery and return of study materials, make sure that participants know the value of the items to be mailed and use mailing tracking. In addition, communication among all the members of the research team is critical. All members must feel comfortable discussing their concerns and possible solutions to issues that arise. Overall, it is important to establish, test and iterate PFA protocols. As in any study, data loss may still occur, such as inability to score the complete BBS when participants do not have the necessary equipment (e.g., a chair without armrests) or monitoring devices fail to collect data (e.g., technical problems or failure to wear the devices). Additionally, lack of a remote surrogate measure for the CPET, precludes assessments of the impact of the intervention on change in maximal exercise capacity between baseline and three months. Nevertheless, assessing PFA remotely offers numerous benefits, such as convenience and reduction in patient burden, since the need to travel to the testing center is eliminated. Key observations related to study adherence are also noted during the implementation of PFA (Table 5 ). Transplant patients often reside hours from the program site, so converting from in-person to remote PFA also allows more individuals to participate since distance from the study center is no longer an exclusion criterion. Lastly, remote PFA is more consistent with the telehealth philosophy and increases the generalizability of the study's findings.
Table 5

Anecdotal observations related to study adherence.

Observations
Retention rates remained strong during the transition from in-person to virtual

Study visits are easier to schedule and reschedule remotely

Removal of arrangements and activities that required physical presence for a study visit such as parking and walking over to assessment sites may have influenced the high retention rate

Consent process is easier because patients are not required to report in-person
Data collection is maintained with minimal issues and easier to administer questionnaires remotely

Missing data are noted as some test items require equipment and supplies tat participants may not have at home (e.g., chair with and without armrests, wall space, ruler, etc.)

Anecdotal observations related to study adherence. Study visits are easier to schedule and reschedule remotely Removal of arrangements and activities that required physical presence for a study visit such as parking and walking over to assessment sites may have influenced the high retention rate Missing data are noted as some test items require equipment and supplies tat participants may not have at home (e.g., chair with and without armrests, wall space, ruler, etc.)

Plans for triangulation and analysis

The primary outcomes for the study did not change with the move to remote assessments. Therefore, the original data analysis plan for testing hypotheses of between-group differences in physical function and physical activity between baseline, 3 and 6 months is maintained. However, the shift to remote assessments raises issues for how to analyze outcomes when data are collected using different in-person, remote and surrogate measures. Below is a description of the two study outcomes that were affected by the change to remote assessment and how we plan to transform the disparate data sources for use in the analysis. Prior to COVID-19, the 6MWT was the only outcome measure for walking ability. Since moving to remote assessments, conducting the 30-s STS during a two-way video session is used as a ‘surrogate’ measure for 6MWT. Scores for 6MWT and 30-s STS are treated as continuous variables and will be described using measures of central tendency, including the mode, median, and mean, and measures of distribution and spread, including range, quartiles, absolute deviation, variance and standard deviation in their original metric. However, to combine raw values from the continuous level 6MWT and 30-s STS tests, we will convert them to the standard scores, more commonly referred to as Z-scores [29], which enables comparison of scores that are from different distributions or measures. The result will be a range of standardized scores to compare differences in walking ability between the LTGO and enhanced usual care groups regardless of the original metric. Similarly, prior to COVID-19, the data recorded using an accelerometer (Actigraph) for 7-days was the only physical activity outcome measure available to quantify the time spent in light and moderate/vigorous activities as continuous scores. To overcome some of the challenges of relying on Actigraphs, the IPAQ-S [24] was added as a surrogate measure. The IPAQ-S data are reported as a continuous score of metabolic equivalent-minutes per week including median values and interquartile ranges for walking, moderate-intensity activities. As with the measures of walking ability, to summarize data from each measure of time spent in levels of physical activity in a meaningful way, descriptive statistics will be reported. Data from the Actigraph and the IPAQ-S will be converted to Z-scores to examine differences in time spent in certain activity levels per day or week between the two groups. Since some participants may have Actigraph and/or IPAQ-S data collected, we will also be able to compare reliability and validity between the objective and self-report measures. Furthermore, as suggested by other clinical trialists [7,30], when all pandemic restrictions are lifted and trials have the option to resume in-person activities, consideration will be given to how best to measure certain variables (such as self-reported measures of physical activity like the IPAQ-S, objective measures such as accelerometers, or in some cases, in combination) to measure outcomes reliably and possibly reduce burden on participants. To this end, we will conduct post-hoc analyses to determine whether any differences in hypothesized effects were explained by combining measurement methods (in person or remote).

Conclusion

In response to the ongoing COVID-19 pandemic, the research team swiftly transitioned from in-person to remote physical function assessments, in compliance with national and local guidelines. During this process, we identified a wide range of challenges which required creative and flexible solutions, while ensuring participant safety, good practice standards, and clinical trial integrity. Although remote assessments may not fully represent the experience of in-person assessments, it is an integral component of research, particularly during public health emergencies. Remote assessment strategies have the potential to make research more efficient and convenient for both participants and researchers. Therefore, aforementioned considerations associated with the planning and implementation of remote assessment should be carefully evaluated in conducting research activities.

Funding

This work was supported by the [R01NR01719601A1].

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  2 in total

1.  Developing a virtual assessment protocol for the AMPLIFI Randomized Controlled Trial due to COVID-19: From assessing participants' preference to preparing the team.

Authors:  Maria Pisu; Iman Omairi; Teri Hoenemeyer; Karina I Halilova; Yu-Mei M Schoenberger; Laura Q Rogers; Kelly M Kenzik; Robert A Oster; Nataliya V Ivankova; Dori Pekmezi; Kevin Fontaine; Wendy Demark-Wahnefried; Michelle Y Martin
Journal:  Contemp Clin Trials       Date:  2021-10-29       Impact factor: 2.226

2.  Test/Retest Reliability and Validity of Remote vs. In-Person Anthropometric and Physical Performance Assessments in Cancer Survivors and Supportive Partners.

Authors:  Teri W Hoenemeyer; William W Cole; Robert A Oster; Dorothy W Pekmezi; Andrea Pye; Wendy Demark-Wahnefried
Journal:  Cancers (Basel)       Date:  2022-02-21       Impact factor: 6.575

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.