| Literature DB >> 29893864 |
Cao Xiao1, Edward Choi2, Jimeng Sun2.
Abstract
Objective: To conduct a systematic review of deep learning models for electronic health record (EHR) data, and illustrate various deep learning architectures for analyzing different data sources and their target applications. We also highlight ongoing research and identify open challenges in building deep learning models of EHRs. Design/method: We searched PubMed and Google Scholar for papers on deep learning studies using EHR data published between January 1, 2010, and January 31, 2018. We summarize them according to these axes: types of analytics tasks, types of deep learning model architectures, special challenges arising from health data and tasks and their potential solutions, as well as evaluation strategies.Entities:
Year: 2018 PMID: 29893864 PMCID: PMC6188527 DOI: 10.1093/jamia/ocy068
Source DB: PubMed Journal: J Am Med Inform Assoc ISSN: 1067-5027 Impact factor: 4.497
Figure 1.Illustration of literature search and selection procedure.
Figure 2.Transform longitudinal EHR data into input vectors (top left), which could support different analytics tasks described in the survey (top right). The underlying deep learning models are visually described at the bottom (a): Feedforward neural networks use multiple layers of fully connected neural networks and non-linear activations (eg., sigmoid or rectified linear unit). (b): Recurrent neural networks can process variable-length input sequence using its recurrent connection. (c): Restricted Boltzmann Machines are bipartite neural networks that consist of binary stochastic nodes. They can capture the latent representation of the input data by learning their generative probability. (d): Generative adversarial networks can generate realistic synthetic samples by training the generator and the discriminator in an adversarial game. (e): Convolutional neural networks capture local features of the input data, and stack those features up via a sequence of convolution to derive global features. (f): Word2vec exploits the co-occurrence information of discrete concepts (eg., words in text, codes in EHR data) to derive concept representations. (g): Denoising autoencoders (AE) try to reconstruct original input from its corrupted version, thus learning robust representations of the input data.
Distributions of models over analytic tasks
| Disease Detection or Classification | Sequential Prediction of Clinical Events | Concept Embedding | Data Augmentation | EHR Privacy | |
|---|---|---|---|---|---|
| [ | [ | [ | [ | [ | |
| [ | [ | [ | NA | NA | |
| [ | NA | [ | [ | NA | |
| [ | [ | [ | NA | [ | |
| NA | [ | NA | [ | [ |