| Literature DB >> 31450991 |
Nobuyuki Kagiyama1, Sirish Shrestha1, Peter D Farjo1, Partho P Sengupta1.
Abstract
Entities:
Keywords: artificial intelligence; deep learning; machine learning; risk model; risk prediction; statistics; telemedicine
Mesh:
Year: 2019 PMID: 31450991 PMCID: PMC6755846 DOI: 10.1161/JAHA.119.012788
Source DB: PubMed Journal: J Am Heart Assoc ISSN: 2047-9980 Impact factor: 5.501
Figure 1Supervised, unsupervised, and reinforcement learning. Machine learning tasks are categorized into supervised, unsupervised, or reinforcement learning. Supervised learning is used for prediction (classification or regression), whereas unsupervised learning aims to reveal hidden patterns in data. Reinforcement learning is another way of learning where an algorithm learns the best action based on its consequences, and is well suited for game theory and control theory. However, it has not had a significant role in clinical research because it requires simulating many “wrong actions” to learn. *Dimensionality reduction can also be performed in a supervised manner.
Figure 2Pipelines of medical research using traditional statistics and AI. Traditional medical research formulates a hypothesis first, and tests it using statistical analysis. Medical research using AI can be hypothesis‐free and data‐driven. Compared with traditional statistics, AI can deal with various types of data, including unstructured data such as images, signals, and EHR. In contrast to traditional medical research that focuses on validation of hypotheses and understanding causality and mechanisms, the main goal of research using AI is to predict new data and identify a hidden pattern in the data. AI indicates artificial intelligence; EHR, electronic health record.
Representative Machine Learning Algorithms
| Algorithm | Description | Use |
|---|---|---|
| Logistic regression | An algorithm that estimates probability of dichotomized outcome from multiple covariates using logistic function. | Classification |
| Decision tree | A flow chart–like algorithm that divides data into branches by considering information gain. The final branches represent output of the algorithm (class or value). | Classification/regression |
| (simple) Neural network | An algorithm inspired by human brain architecture. Layers consisting of nodes are connected to one another with edges weighted as per training results. | Classification/regression |
| K nearest neighbor | A simple algorithm that classifies observations by comparing k examples that exist in the nearest locations (=examples with the most similar features). | Classification/regression |
| Support vector machine | Support vector machine draws a boundary line that maximizes margins from each class. New observations are classified using this line. | Classification/regression |
| K means | A clustering method that makes k clusters in which each observation belongs to the cluster that has its mean in the nearest locations from the observation. | Clustering |
| Hierarchical clustering | A type of cluster analysis that builds a dendrogram with a hierarchy of clusters. Pairs of clusters are merged to form clusters as they move up the hierarchy (agglomerative approach). | Clustering |
| Principal component analysis | An algorithm that converts high dimensional data into lower dimensional data with keeping important information as much as possible by orthogonal transformation | Dimensionality reduction |
Figure 3Structure of deep learning. Deep learning consists of input layers, hidden layers, and output layers. Through multiple hidden layers, raw input is gradually converted into more abstract and essential features that represent the original data. In image recognition, the input layer indicates raw pixels of the image, then first layers identify simple features of the image such as edges and lines. Succeeding layers identify somewhat more complex features such as ears, eyes, and tails. Finally, last layers recognize features of cats and dogs. As such, deep learning extracts key features from raw unstructured data and returns outputs as classification or regression.
Examples of ML in Cardiovascular Research
| Data Structure | Year | First Author | Journal/Conference | Task | Model | Summary |
|---|---|---|---|---|---|---|
| Structured data | ||||||
| 2016 | Motowani | Eur Heart J | Classification: Prognostic prediction | Ensemble | Using 69 clinical and CT parameters of 10 030 CAD patients, a ML model predicted mortality better than traditional statistics | |
| 2018 | Kakadiaris | JAHA | Classification: Prognostic prediction | SVM | Using 9 parameters that consist of ACC/AHA risk calculator, a ML model showed better prediction than original ACC/AHA risk score. | |
| 2016 | Narula | JACC | Classification: Diagnosis of HCM | Ensemble (SVM, RF, and ANN) | Using clinical and echocardiographic parameters, ML algorithms discriminated HCM from ATH with 87% sensitivity and 82% specificity. | |
| 2019 | Lancaster | JACC CV Imaging | Clustering | Hierarchical clustering | Using echocardiographic parameters that guidelines recommend for assessment of LVDD, hierarchical clustering identified clusters that discriminate patient prognosis better than guidelines‐based classification | |
| 2019 | Casaclang‐Verzosa | JACC CV Imaging | Clustering with dimensionality reduction | Topological data analysis | Topological data analysis was able to visualize patient‐patient similarity network that is created from 4 parameters. Relative location of patients in the network were associated with disease phenotypes and prognosis. | |
| Unstructured data | ||||||
| Echocardiographic images | 2018 | Zhang | Circulation | Classification: Automatic interpretation of echocardiography | CNN | Using 14 035 echocardiograms, CNN enabled automatic classification of views, identification of chambers, measurements of cardiac volumes, and discrimination of diseases from healthy controls (see text for details) |
| MRI images | 2019 | Zhang | Radiology | Classification: Prediction of MI from non‐enhanced MRI | LSTM+Optical flow | In 212 patients and 87 controls, algorithms were able to detect chronic MI (validated by LGE) with 90% sensitivity and 99% specificity using nonenhanced cine MRI. |
| CT images | 2016 | Shandmi | Med Image Anal | Classification: Coronary artery calcium in a voxel | CNN | Using 3D CTA of 250 patients, after localization of volume of interest using 3 CNNs, 2 CNNs were used to classify voxels to calcium or noncalcium. Agatston score calculated based on the voxel classification showed excellent agreement with reference standard (accuracy 83%). |
| ECG signals | 2019 | Hannun | Nat Med | Classification: Arrhythmia detection | DNN | Using 91 232 single‐lead ECG, trained algorithm showed better prediction of 12 types of heart rhythm than cardiologist (F‐measure 0.84 vs 0.78). |
| Heart sound signals | 2016 | Potes | 2016 CinC | Classification: Normal and abnormal heart sound | AdaBoost+CNN | Combination of AdaBoost and CNN showed 94.2% sensitivity and 77.8% specificity for identifying abnormal heart sound in PhysioNet/CinC data set. |
| EHR | 2019 | Mallya | arXiv | Classification: Prognostic prediction | LSTM | Using >23 000 patients time‐series data, LSTM algorithm successfully predicted the onset of heart failure 15 mo in advance (AUC 0.91) |
| EHR: medical letters (text) | 2019 | Diller | Eur Heart J | Classification: Diagnosis, symptoms and prognosis | CNN+LSTM | Using natural language processing, diagnosis (accuracy 91%) and symptoms (90.6%) were extracted from medical letters. Also, prognostic prediction using the same data was useful (HR 34.0) |
ACC indicates American College of Cardiology; AHA, American Heart Association; ANN, artificial neural network; ATH, athlete; AUC, area under the receiver‐operating‐characteristic curves; CAD, coronary artery disease; CNN, convolutional neural network; CT, computed tomography; CTA, computed tomography angiography; DNN, deep neural network; HCM, hypertrophic cardiomyopathy; HR, hazard ratio; LGE, late gadolinium enhancement; LSTM, long short time memory; LVDD, left ventricular diastolic dysfunction; MI, myocardial infarction; ML, machine learning; MRI, magnetic resonance imaging; RF, random forest; SVM, support vector machine.
Figure 4Topological data analysis in patients with AS. Topological data analysis enables integration of multidimensional complex data and visualization of hidden patterns in the data. Each node represents 1 or more patients with similar echocardiographic parameters of AS, and nodes including similar patients are connected by edges. Each panel is colored by 1 parameter written on the top right and color of the nodes represents the mean value of the parameter in the nodes. Although the network was created only from the parameters of aortic stenosis, preserved and reduced LV function (systolic and diastolic) were segregated in different regions. AS indicates aortic stenosis; AV, aortic valve; LV, left ventricle. Reprinted from Casaclang‐Verzosa et al19 with permission. Copyright ©2019, Elsevier.
Figure 5Underfitting, optimal fitting, and overfitting. The upper row shows regression models created in sample data (=training data), and new data from the same population were added in the bottom row. A simple linear model on the left panel was underfitted to the data, with low variance (ie, fluctuations in predicted value) but high bias (ie, difference between predicted and true value). In contrast, a complex model on the right panel was overfitted with low bias but high variance, because it also modeled random noise in the sample data. As a model becomes more complex, goodness‐of‐fit increases and bias decreases. However, overfitted models do not capture real association in data and cannot work well for new data.
Figure 6Development and evaluation of machine learning model. Since machine learning aims to predict new data in supervised learning, the test set is always preserved during when the machine learning model is built in order to guarantee generalizability. Ordinarily, the remaining data are further split into the training set, which is used to build models (calculate weights), and the validation set, which is used to validate the generated models and to tune hyperparameters. This training‐validation process is performed using a cross‐validation or holdout method. Finally, performance of the created model is evaluated using a test set that is not used in the model‐building process. MAE indicates mean absolute error; RMSE, root mean square error.