Literature DB >> 36245571

Physician perceptions of surveillance: Wearables, Apps, and Chatbots for COVID-19.

Alexandra R Linares1, Katrina A Bramstedt2, Mohan M Chilukuri3, P Murali Doraiswamy1.   

Abstract

Background and Purpose: To characterize the global physician community's opinions on the use of digital tools for COVID-19 public health surveillance and self-surveillance. Materials and
Methods: Cross-sectional, random, stratified survey done on Sermo, a physician networking platform, between September 9 and 15, 2020. We aimed to sample 1000 physicians divided among the USA, EU, and rest of the world. The survey questioned physicians on the risk-benefit ratio of digital tools, as well as matters of data privacy and trust. Statistical Analysis Used: Descriptive statistics examined physicians' characteristics and opinions by age group, gender, frontline status, and geographic region. ANOVA, t-test, and Chi-square tests with P < 0.05 were viewed as qualitatively different. As this was an exploratory study, we did not adjust for small cell sizes or multiplicity. We used JMP Pro 15 (SAS), as well as Protobi.
Results: The survey was completed by 1004 physicians with a mean (standard deviation) age of 49.14 (12) years. Enthusiasm was highest for self-monitoring smartwatches (66%) and contact tracing apps (66%) and slightly lower (48-56%) for other tools. Trust was highest for health providers (68%) and lowest for technology companies (30%). Most respondents (69.8%) felt that loosening privacy standards to fight the pandemic would lead to misuse of privacy in the future.
Conclusion: The survey provides foundational insights into how physicians think of surveillance. Copyright:
© 2020 DIGITAL MEDICINE.

Entities:  

Keywords:  Apps; Privacy; Surveillance; Trust; Wearables

Year:  2022        PMID: 36245571      PMCID: PMC9549767          DOI: 10.4103/digm.digm_28_21

Source DB:  PubMed          Journal:  Digit Med        ISSN: 2226-8561


INTRODUCTION

Public health surveillance is the systematic collection and analysis of health-related data to prevent or control disease, followed by its application for public health action.[1] The global scale of the COVID-19 pandemic has accelerated the use of non-traditional, technology-based, public health, and self-surveillance mechanisms to control the spread of severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2).[234567891011121314151617181920212223] Examples of such tools include contact-tracing apps, analyses of global positioning systems and social media data for population movement tracking, fever-sensing infrared thermal detection systems, symptom self-screeners (e.g. chatbots), and smartwatch applications to detect physiological signs of infection.[4567810] Digital technologies can rapidly collect, store, analyze, and share numerically encoded information, making them potentially highly useful in a pandemic such as COVID-19. Blue Dot, a Canadian digital health company, reportedly identified the emergence of COVID-19 through the aggregation of big data from sources such as social media and air travel, before even the WHO issued an alert.[14] However, these digital surveillance tools are experimental, and their accuracy across different settings is not fully established.[151617181920] For example, studies have shown that the accuracy of facial recognition technologies differs by race, gender, and age.[22] These tools also come with a number of potential legal and ethical risks,[1822232425262728] such as privacy concerns, discrimination, and over-reach of the data mission that “highlight the long-standing tensions between individual and collective rights.”[18] Notably, the morbidity and mortality of the COVID-19 pandemic have heightened worldwide anxiety to an extent that digital public health surveillance has become ubiquitous (e.g. national requirements for downloading contact tracing apps; thermal scanning by employers and private businesses; personal location data collection via QR codes; texting of COVID-19 assay results to patients). COVID-19 is not the world’s first pandemic, nor will it be the last. Thus, it is vital to understand the views of physicians, as they are involved in many facets of health data and its application to COVID-19 care. The aim of this report is to characterize the views of physicians regarding the benefits and risks of surveillance technologies

MATERIALS AND METHODS

Ethics

This study was deemed as exempt research by Duke University Medical Center’s Institutional Review Board

Study sample

To characterize the opinions of physicians on this topic, we analyzed data from a cross-sectional, random, stratified survey of physicians registered with Sermo, a secure digital platform for medical crowdsourcing and anonymous surveys. The Sermo platform is exclusive to verified and licensed physicians and has over 800,000 registered physicians, of all specialties, worldwide Following informed consent, the English-language survey sampled physicians between September 9 and September 15, 2020 (before initiation of SARS-CoV-2 vaccination), with a target sample size of 1000 doctors equally divided between the US, EU, and rest of the world (RoW). The survey results were de-identified to create anonymized data for analysis

Survey instrument

Five questions in the survey [Figure 1 and Supplemental Table 1] asked physicians their opinions on the benefits and risks/harms of using smartwatch sensor alerts, contact tracing apps, thermal cameras for mass fever screening, chatbots, and social media tracking for public health or self-surveillance. These questions focused on the risk-benefit ratio, and the answer options were “Yes” “No” and “Uncertain” For the questions on sensors (wearables and thermal cameras), we provided accuracy estimates derived from published studies. We specified that informational chatbots do not require regulatory approval in most countries for the question on chatbots, as physicians may not be aware of this. The survey then asked about the level of trust in different organizations (technology companies, government, employer, medical providers, educational universities/nonprofit bodies, or no one) to protect private surveillance data. The answer choices for this question were “Very much” “Somewhat” “Neutral” “Not really” or “Not at all” Respondents were then asked about the impact of current surveillance on future privacy standards. A final question asked physicians to provide brief qualitative comments to elaborate on their views. The results of two survey questions are reported elsewhere.[2930]
Figure 1

Schematic illustration of surveillance tools and issues queried in the survey. The survey examined the risk-benefit ratio of two self-screening (purple) and three public health (blue) surveillance digital tools. It also addressed issues around trust and misuse (red)

Schematic illustration of surveillance tools and issues queried in the survey. The survey examined the risk-benefit ratio of two self-screening (purple) and three public health (blue) surveillance digital tools. It also addressed issues around trust and misuse (red)

Data and statistical methods

Descriptive statistics examined physicians’ characteristics and opinions by age group, gender, frontline status, and geographic region. To test the effect of age, subjects were grouped as “younger” or “older” by age 49 years. To test the effect of frontline status, physicians more directly involved in COVID-19 care were grouped as frontline (e.g. internal medicine, ICU, ED), whereas the rest were categorized as non-frontline (although we recognize that all physicians may interact with or consult on COVID-19 patients). For geographic analyses, we pooled doctors into three groups based on the location of practice (US, European Union, RoW), while recognizing these subgroups are not homogeneous. The five categories relating to trust were combined into three categories as trusted (“Somewhat” or “Very much” responses), not trusted (“Not really” or “Not at all”), and “Neutral”. Gender analyses were restricted to those who categorized themselves as male or female. ANOVA, t-test, and Chi-square tests with P < 0.05 were viewed as qualitatively different. As this was an exploratory study, we did not adjust for small cell sizes or multiplicity. We used JMP Pro 15 (SAS), as well as Protobi

RESULTS

Sample characteristics

The final respondent sample consisted of 1004 physicians representing 40 countries in North and South America, Europe, and Asia-Pacific [Supplemental Table 2]. The average age of the sample was 49.1 ± 12.3 years and 49% of respondent physicians were characterized as frontline. Of the sample, 40% were male, 20.6% were female, and 39% opted out of indicating their gender

Utility of surveillance tools

Response rates for the support of various digital surveillance tools are shown in Figure 2 and Table 1. Smart watches were supported by 65.5% of respondents (χ2 = 468.58, P < 0.0001), contact tracing apps were supported by 66.4% (χ2 = 496.85, P < 0.0001), fever cameras were supported by 58.9% (χ2 = 306.34, P < 0.0001), symptom screener chatbots by 47.6% (χ2 = 92.18, P < 0.0001), and social media by 50.9% (χ2 = 145.68, P < 0.0001)
Figure 2

Physician perceptions of digital surveillance tools relevant to COVID-19. Graph illustrates percent of respondents who were supportive (green bars), uncertain (blue bars) or unsupportive of the use of surveillance tools. Please see text for statistical differences

Table 1

Physician perceptions of surveillance tools

Physician perceptions of surveillance tools Physician perceptions of digital surveillance tools relevant to COVID-19. Graph illustrates percent of respondents who were supportive (green bars), uncertain (blue bars) or unsupportive of the use of surveillance tools. Please see text for statistical differences

Age differences

Younger physicians (69.3%) were more likely to support the use of smart watch sensors compared with older physicians (61.3%) (χ2 = 7.06, P = 0.03). Younger physicians (69.7%) were marginally more likely to support the use of contact tracing apps compared to older physicians (62.8%) (P = 0.02). Younger physicians (63.6%) were more likely to support the use of mass fever screenings compared with older physicians (53.6%) (χ2 = 11.62, P = 0.003), whereas older physicians were more likely to be uncertain. Younger physicians (56.1%) were more likely to support the use of social media for population movement tracking versus older physicians (45.2%) (χ2 = 14.34, P = 0.0007), and older physicians tended to be more uncertain. Responses did not significantly differ by physician age for the utility of symptom screener chatbots

Gender differences

Male physicians (60%) were slightly more likely to support the use of fever cameras than female physicians (51%) (P = 0.012). Responses did not differ by gender for the other surveillance tools

Frontline status differences

Responses did not differ by frontline status

Which entity do you trust the most with your personal surveillance data?

Physicians picked “medical providers” as the most trusted entity to protect the privacy of COVID-19 surveillance data, with about 68% of respondents reporting that they trusted their medical provider Figure 3. The second most trusted group was “educational/non-profit bodies” with a combined 52% of respondents reporting “somewhat” and “very much” levels of trust. Conversely, the most distrusted group was “technology companies” with only 30% of respondents reporting “somewhat” or “very much” and 46% reporting “not really” or “not at all”. Following technology companies, respondents reported low levels of trust for the “government” with only 36% responding “somewhat” or “very much” Older physicians were more likely to be distrustful of technology companies (48.9%), the government (44.5%), and educational universities/non-profit bodies (26.9%) compared with younger physicians (42.4%, 33.9%, 16.9%, respectively) (P = 0.038, 0.001, <0.001). US physicians (54.1%) were more likely to be distrustful of technology companies, compared with both EU (41.0%) and RoW (41.3%) physicians (P < 0.001). US physicians (51.8%) were also more distrustful of the government, compared with RoW (38.6%) and EU (27.5%) physicians (P < 0.001)
Figure 3

Physicians perceptions of trust in various entities to protect their personal data. Colors show the percentages reported for the 5 trust level categories. The percentages reported inside the bars combine “Somewhat”/”Very much” and “Not really”/”Not at all” categories

Physicians perceptions of trust in various entities to protect their personal data. Colors show the percentages reported for the 5 trust level categories. The percentages reported inside the bars combine “Somewhat”/”Very much” and “Not really”/”Not at all” categories

Effect of current surveillance on future misuse of privacy

The majority of respondents (69.8%) believed that potentially loosening privacy standards to fight the pandemic would lead to misuse of privacy in future (χ2 = 601.50, P < 0.0001) [Figure 4. Frontline physicians (73.2%) were more likely to voice concern compared with nonfrontline physicians (66.5%) (χ2 = 7.65 P = 0.022). More male physicians (71.9%) believed that a loosening of privacy standards would lead to misuse, compared with female physicians (61.8%) (χ2 = 9.58 P = 0.048)
Figure 4

Physician perceptions of the risk for future misuse of data. Red illustrates the percent who agreed that loosening privacy laws would result in misuse. Green represents those who disagreed. Blue represents those who were uncertain. (*P < 0.05)

Physician perceptions of the risk for future misuse of data. Red illustrates the percent who agreed that loosening privacy laws would result in misuse. Green represents those who disagreed. Blue represents those who were uncertain. (*P < 0.05)

Selected qualitative comments by physicians about digital surveillance

Respondents also had an option to provide qualitative comments on digital surveillance. Supportive comments [Supplemental Table 3] included statements such as “the Future is here,” “must be made mandatory,” “anything that prevents deaths is fine, I don’t worry about privacy,” and “During the 1940–1941 bombing of London called The Blitz, I believe there were zero residents of London who said ‘I have a constitutional right to leave my lights on a night if I feel like it.’” Concerns about efficacy [Supplemental Table 4] included statements like “way too early,” “bad for patient and physician,” and “data should be analyzed in clinical trials.” Concerns about harms [Supplemental Table 5] included statements such as “Pandora’s box,” “creepy, extreme slippery slope,” “any great idea can have unforeseen consequences” and “I fear the behavior of people not technologies.”

DISCUSSION

Data is currency. Technology companies know this, governments know this, and so does the public. Like with any currency, data can be accumulated, bought and sold, or even be stolen. Hence, its storage needs to be secure. Health data are a form of personal information that generally people want to keep private, and many regulations have been implemented to safeguard personal data privacy rights.[3132] During a pandemic, public health interests allow for broader powers, governments, and health systems to collect, use, store, and share personal information. However, as our survey shows, this creates concern among even the physicians who are part of this process (and concurrently attempting to prevent and treat the implicated illness)

Key findings

Overall, support varied from 48% to 66% for the various surveillance tools. Two-thirds of physicians voiced support for the use of smartwatches in self-monitoring. This appears consistent with recent studies documenting the promise of consumer smart watch-based physiological signals (e.g. heart rate, sleep, activity, skin temperature) for discriminating COVID-19 test positive cases from negative cases, as well as for detecting pre-symptomatic COVID-19 infection.[45] Further, a smartwatch-linked platform, Aura, recently received an EU CE mark for this purpose based on its sensitivity of 94% and ability to detect an infection signal on average 2.64 days after inoculation.[6] The minority of respondents who oppose smart watch-based infection detection technology were likely concerned about the potential for noisy data leading to misdiagnosis and unnecessary testing.[19] Two-thirds of physicians also voiced support for contact tracing apps, even those that collected personal data. Many countries have implemented contact tracing apps, and physicians are well versed in traditional contact tracing principles for infection control, both of which likely increased physician confidence in their utility. However, since our survey, some studies have questioned the effectiveness (e.g. sensitivity of only 7% in one study) and ethics of digital contact tracing.[202122] This suggests that the optimism of respondents in our survey may have been premature Physician support was slightly lower (59%) for “fever cameras” but still optimistic, consistent with the utility for mass screening offered by their high negative predictive value,[8] However, the positive predictive value (<20% in one study) of these systems remains low,[89] suggesting the need for further optimization to reduce false positives Support for the use of social media tracking (51%) and chatbots (48%) was also slightly lower. Social media tracking is a promising tool that offers real-time data for public health officials to monitor citizen movement or social interactions during lockdowns.[10111213] However, questions remain about lack of consent, accuracy, and misuse potential. Chatbots, especially those designed using WHO or CDC guidelines, very likely helped large numbers of users (over 200 million messages by some estimates) quickly get reliable information. However, to our knowledge, there are no published accuracy or outcomes data on the utility of chatbots for pandemic self-screening. Hence, there is a need for further research into the effectiveness and potential for the spread of misinformation.[2021] Respondents in our survey also voiced concerns over privacy risks and over-reach of the data mission. Respondents had low trust in technology companies (30%) and governments (36%) to safeguard surveillance data. Trust was highest with medical providers (68%), followed by non-profit organizations. The higher level of trust among EU physicians may be due to the stricter data privacy laws in the EU versus the US.[31] These concerns are legitimate since some technology platforms rely on selling user data to advertisers,[27] and studies have found apps and chatbots share information with a variety of third parties.[20272829] The risks also go beyond privacy breaches.[27] Historically, surveillance has worsened stigmatization and discrimination against racial or religious minorities who were often falsely blamed for disease outbreaks.[27] Further, some governments have reportedly used the pandemic to rank citizens by health status or analyze personal telecommunications traffic.[27] Hence, surveillance done wrong may “invite mission creep into adjacent fields, such as automated policing and content control.”[24] “No turning back” is a famous quote used in many settings, and our research makes it pertinent to digital health as well. The fast portability of health data, along with the complexity of legal regulations and voluminous “Terms of Use” documents that are rarely read by users,[30] create a reality of data that has the potential to quickly bounce to all corners of the world. In addition, there is the very real presence of hackers.[31] Therefore, some of the data receivers have motives that have nothing to do with “public interest” Accordingly, the fears and lack of trust we observed are likely well-founded and highlight the need for risk mitigation to harness the full promise of public health surveillance during a pandemic

Strengths and limitations

This is the first global survey, to our knowledge, to investigate the opinions of physicians about the utility, trust, and risks of commonly used public health digital surveillance tools. Our survey data is from a relatively large and diverse sample of verified practicing physicians. Potential limitations include cross-sectional design, the limited number of respondents from developing countries, inability to control for all possible confounding variables (e.g. personal medical history, socio-political beliefs, local data privacy regulations, knowledge about digital tools), and inability to deduce causality. Further, physician perceptions may change over time if infection risk and prevalence decrease, due to vaccination and herd immunity. Our findings should be interpreted within this context. Nevertheless, they provide a useful baseline for future surveys

Interpretation and implications

Physicians were optimistic but not equally supportive of all surveillance tools suggesting the need for further research on effectiveness. There was also variation in physician opinions by age group. This may in part reflect differences in physician knowledge about emerging technologies and/or risk-benefit analyses, which would benefit from further education. The low level of trust in technology companies to protect personal data suggests that independent entities (governed by stricter privacy laws) should be the gatekeepers of such data. Current regulations fall short of addressing the risks posed by these new technological developments. It has been said that “data moves at the speed of trust.” During public health emergencies, any data collection through such newer tools should be both time-limited and scope-limited, with decisions made in a transparent way before the launch of surveillance activity.[2227] In parallel, we may need to strengthen other data privacy rules to ensure any temporary loosening during public health emergencies does not result in future misuse in normal times. We hope that insights from surveys such as this may spur public health agencies and technology innovators to work together to develop the evidence base and balance individual versus societal versus commercial needs.[24] As aptly noted by one of the survey respondents, “we can learn from films like Spiderman and The Dark Knight – with great power comes great responsibility.”

Financial support and sponsorship

Sermo provided non-financial technical platform support for the study

Conflicts of interest

PMD has received research grants and/or advisory/ board fees from health and technology companies. PMD owns shares in companies and is a co-inventor on patents. MMC reported personal fees outside the submitted work. KB has received consulting fees outside of the submitted work. ARL has no conflicts of interest to report
  18 in total

1.  Public health surveillance and knowing about health in the context of growing sources of health data.

Authors:  Lisa M Lee; Stephen B Thacker
Journal:  Am J Prev Med       Date:  2011-12       Impact factor: 5.043

2.  Healthcare Data Are Remarkably Vulnerable to Hacking: Connected Healthcare Delivery Increases the Risks.

Authors:  Ross Koppel; Craig Kuziemsky
Journal:  Stud Health Technol Inform       Date:  2019

3.  Digital health during COVID-19: lessons from operationalising new models of care in ophthalmology.

Authors:  Dinesh V Gunasekeran; Yih-Chung Tham; Daniel S W Ting; Gavin S W Tan; Tien Y Wong
Journal:  Lancet Digit Health       Date:  2021-02

4.  Presence and mobility of the population during the first wave of Covid-19 outbreak and lockdown in Italy.

Authors:  Paolo Beria; Vardhman Lunkar
Journal:  Sustain Cities Soc       Date:  2020-11-22       Impact factor: 7.587

5.  Utilization of Self-Diagnosis Health Chatbots in Real-World Settings: Case Study.

Authors:  Xiangmin Fan; Daren Chao; Zhan Zhang; Dakuo Wang; Feng Tian; Xiaohua Li
Journal:  J Med Internet Res       Date:  2021-01-06       Impact factor: 5.428

6.  Analyzing the Human Rights Impact of Increased Digital Public Health Surveillance during the COVID-19 Crisis.

Authors:  Sharifah Sekalala; Stéphanie Dagron; Lisa Forman; Benjamin Mason Meier
Journal:  Health Hum Rights       Date:  2020-12

7.  Physician Perceptions of Catching COVID-19: Insights from a Global Survey.

Authors:  P Murali Doraiswamy; Mohan M Chilukuri; Dan Ariely; Alexandra R Linares
Journal:  J Gen Intern Med       Date:  2021-03-29       Impact factor: 5.128

8.  Early Perceptions of COVID-19 Contact Tracing Apps in German-Speaking Countries: Comparative Mixed Methods Study.

Authors:  Stuart McLennan; Alena Buyx; Bettina Maria Zimmermann; Amelia Fiske; Barbara Prainsack; Nora Hangel
Journal:  J Med Internet Res       Date:  2021-02-08       Impact factor: 5.428

9.  Pre-symptomatic detection of COVID-19 from smartwatch data.

Authors:  Tejaswini Mishra; Meng Wang; Ahmed A Metwally; Gireesh K Bogu; Andrew W Brooks; Amir Bahmani; Arash Alavi; Alessandra Celli; Emily Higgs; Orit Dagan-Rosenfeld; Bethany Fay; Susan Kirkpatrick; Ryan Kellogg; Michelle Gibson; Tao Wang; Erika M Hunting; Petra Mamic; Ariel B Ganz; Benjamin Rolnik; Xiao Li; Michael P Snyder
Journal:  Nat Biomed Eng       Date:  2020-11-18       Impact factor: 29.234

10.  Integrating emerging technologies into COVID-19 contact tracing: Opportunities, challenges and pitfalls.

Authors:  Elliot Mbunge
Journal:  Diabetes Metab Syndr       Date:  2020-08-26
View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.