| Literature DB >> 33647989 |
Ignacio Perez-Pozuelo1,2, Dimitris Spathis3, Jordan Gifford-Moore4, Jessica Morley5,6, Josh Cowls2,5.
Abstract
In this perspective we want to highlight the rise of what we call "digital phenotyping" or inferring insights about peopleãs health and behavior from their digital devices and data, and the challenges this introduces. Indeed, the collection, processing, and storage of data comes with significant ethical, security and data governance considerations. The COVID-19 pandemic has laid bare the importance of scientific data and modeling, both to understand the nature and spread of the disease, and to develop treatment. But digital devices have also played a (controversial) role, with track and trace systems and increasingly "vaccine passports" being rolled out to help societies open back up. These systems epitomize a wider and longer-standing trend towards seeing almost any form of personal data as potentially health data, especially with the rise of consumer health trackers and other gadgets. Here, we offer an overview of the risks this introduces, drawing on the earlier revolution in genomic sequencing, and propose guidelines to help protect privacy whilst utilizing personal data to help get society back up to speed.Entities:
Mesh:
Year: 2021 PMID: 33647989 PMCID: PMC8363798 DOI: 10.1093/jamia/ocab012
Source DB: PubMed Journal: J Am Med Inform Assoc ISSN: 1067-5027 Impact factor: 4.497
Figure 1.Differences between genotyping data and digital phenotyping data. Digital phenotyping can never be said to be complete, because new data are generated continuously to reflect changing patterns of user behavior. Although sophisticated data analysis often requires considerable infrastructure and expertise, the cost of processing and analyzing each additional data point is usually negligible.
Building on developments in genetics to establish a path for digital phenotyping
| Lessons From Genetics | Digital Phenotyping | |
|---|---|---|
|
|
Open, tiered, and managed data access Innovative consent models |
Federated learning and zero-knowledge proofs could allow secure data sharing even between competing organizations Differential privacy (adding noise to individual datapoints) can be applied, especially where location/GPS data is recorded |
|
|
Culture of research ethics and oversight Clear classification of consent to participation in secondary data use and explanation of the implications |
Transparent presentation of terms, employing modern UX practices Innovation in dynamic consent and user interfaces for research |
|
|
Developing standards for the reliability of results in clinical settings Leveraging data-sharing practices for genome-wide association studies |
FACT AI and models that actively adjust for demographic parity Provision of reliable outputs and contextualizing clinically relevant information |
|
|
Multistakeholder input into applicable guidelines and regulatory frameworks Scope for public–private collaboration and research exemptions included in applicable laws |
Enabling regulatory oversight through allocation of resources and cross-domain expertise to existing consumer protection regulators Developing practical industry codes, guidance, and oversight mechanisms (eg, an enforceable GDPR code) |
|
|
Direct-to-consumer models and collaboration among clinicians, commercial providers, and researchers Anonymization of individual-level data Biobanks that balance the values and rights of participants while ensuring their long-term sustainability. |
Diverse potential harms from data collection and processing due to heterogeneous data Heightening the differences and disparities of historically marginalized groups Transparency regarding the validity of inferences and associated behavioral interventions Commercially led approach to traditionally public research |
AI: artificial intelligence; FACT AI: Fairness, Accountability and Transparency in AI; GDPR: General Data Protection Regulation; UX: user experience.