| Literature DB >> 35354911 |
Joseph Caffarini1, Klevest Gjini2, Brinda Sevak2, Roger Waleffe3, Mariel Kalkach-Aparicio2, Melanie Boly2,4, Aaron F Struck5,6.
Abstract
In this study, we designed two deep neural networks to encode 16 features for early seizure detection in intracranial EEG and compared them and their frequency responses to 16 widely used engineered metrics to interpret their properties: epileptogenicity index (EI), phase locked high gamma (PLHG), time and frequency domain Cho Gaines distance (TDCG, FDCG), relative band powers, and log absolute band powers (from alpha, beta, theta, delta, low gamma, and high gamma bands). The deep learning models were pretrained for seizure identification on the time and frequency domains of 1 s, single-channel clips of 127 seizures (from 25 different subjects) using "leave-one-out" (LOO) cross validation. Each neural network extracted unique feature spaces that were interpreted using spectral power modulations before being used to train a Random Forest Classifier (RFC) for seizure identification. The Gini Importance of each feature was calculated from the pretrained RFC, enabling the most significant features (MSFs) for each task to be identified. The MSFs were extracted to train another RFC for UPenn and Mayo Clinic's Seizure Detection Kaggle Challenge. They obtained an AUC score of 0.93, demonstrating a transferable method to identify and interpret biomarkers for seizure detection.Entities:
Mesh:
Substances:
Year: 2022 PMID: 35354911 PMCID: PMC8967852 DOI: 10.1038/s41598-022-09429-w
Source DB: PubMed Journal: Sci Rep ISSN: 2045-2322 Impact factor: 4.379
Figure 1Neural network architectures, leave one out (LOO) pretraining, and evaluation. (a) Network schematic for each neural network. (b) LOO test set performance of each neural network on clinical (broadband) iEEG signals. (c) LOO test set performance on narrowband iEEG signals. (d) Visualization of model LOO test set predictions on iEEG data.
Classifier subnetwork parameters.
| Classifier subnetwork | |
|---|---|
| Layers | Intermediate tensor shape |
| Input, | (N, 16) |
| (N, 14) | |
| (N, 12) | |
| (N, 10) | |
| (N, 8) | |
| (N, 6) | |
| (N, 4) | |
| (N, 2) | |
| (N, 1) | |
This classifier subnetwork is used in all neural networks with the same parameters. Dropout was adjusted to prevent overfitting for each. N is the batch number of the input tensor.
Trainable layers are given in bold.
Time domain black box (TDBB) encoder network parameters.
| TDBB encoder | |
|---|---|
| Layers | Intermediate tensor shape |
| Input | (N, 1, 400) |
| (N, 2, 400) | |
| (N, 4, 400) | |
| 1D max pooling | (N, 4, 198) |
| (N, 5, 198) | |
| (N, 7, 198) | |
| 1D max pooling | (N, 7, 97) |
| (N, 9, 97) | |
| (N, 10, 97) | |
| (N, 11, 97) | |
| 1D max pooling | (N, 11, 46) |
| (N, 12, 46) | |
| (N, 13, 46) | |
| (N, 14, 46) | |
| 1D max pooling | (N, 14, 21) |
| (N, 15, 21) | |
| 1D max pooling | (N, 15, 8) |
| (N, 16, 1) | |
The convolutional layers were designed to act as a series of discrete filters with trainable parameters. N is the batch number of the input tensor.
Trainable layers are given in bold.
Frequency domain black box (FDBB) encoder network parameters.
| FDBB encoder | |
|---|---|
| Layers | Intermediate tensor shape |
| Input, | (N, 200) |
| (N, 100) | |
| (N, 50) | |
| (N, 16) | |
The fully connected subnetwork was designed to learn how to derive relationships between the power densities of each frequency band. N is the batch number of the input tensor.
Trainable layers are given in bold.
Figure 2Process for identifying and interpreting significant features during pretraining using random forest. (a) Process using Random Forest Classifiers (RFCs) to identify the most significant features (MSFs). (b) Relative Gini Importance of all features. (c) Frequency response of encoded feature spaces. (d) Correlation between feature spaces.
Figure 3Kaggle test scores using most and least significant feature (MSF and LSF) ensembles. (a) Kaggle contest training process on a hypothetical canine data point. (b) Kaggle Autograder Scores on the test set using feature ensembles with varying importance.