| Literature DB >> 34283328 |
Marcos Fabietti1, Mufti Mahmud2,3,4, Ahmad Lotfi1, M Shamim Kaiser5, Alberto Averna6, David J Guggenmos7, Randolph J Nudo7, Michela Chiappalone8, Jianhui Chen9,10.
Abstract
Neuronal signals generally represent activation of the neuronal networks and give insights into brain functionalities. They are considered as fingerprints of actions and their processing across different structures of the brain. These recordings generate a large volume of data that are susceptible to noise and artifacts. Therefore, the review of these data to ensure high quality by automatically detecting and removing the artifacts is imperative. Toward this aim, this work proposes a custom-developed automatic artifact removal toolbox named, SANTIA (SigMate Advanced: a Novel Tool for Identification of Artifacts in Neuronal Signals). Developed in Matlab, SANTIA is an open-source toolbox that applies neural network-based machine learning techniques to label and train models to detect artifacts from the invasive neuronal signals known as local field potentials.Entities:
Keywords: Artifacts; Local field potential; Machine learning; Neural networks; Neuronal signals
Year: 2021 PMID: 34283328 PMCID: PMC8292498 DOI: 10.1186/s40708-021-00135-3
Source DB: PubMed Journal: Brain Inform ISSN: 2198-4026
Open-source toolboxes and noise detection and removal functionalities
| Artifact detection | Digital filtering | Data visual. | Spectral analysis | Stim. art. removal | File oper. | Multiple formats | |
|---|---|---|---|---|---|---|---|
| Brainstorm [ | X | X | |||||
| BSMART [ | X | X | X | X | |||
| Chronux [ | X | X | X | X | X | ||
| Elephant [ | X | X | X | X | |||
| Fieldtrip [ | X | X | X | ||||
| Klusters, NeuroScope, NDManager [ | X | X | |||||
| Neo [ | X | X | X | X | |||
| NeuroChaT [ | X | X | X | X | |||
| Spycode [ | X | X | X | ||||
| SANTIA |
Data visualization, stimulation artifact removal and file operations (i.e., file splitting, concatenation, column rearranging)
Advancements of SANTIA over SigMate
| Toolbox | SAD | UNoC | SE | Up | DF | DV | SA | SAR | FO | MF |
|---|---|---|---|---|---|---|---|---|---|---|
| SigMate [ | X | X | X | X | ||||||
| SANTIA |
SAD state-of-the-art artifact detection, UNoC unlimited number of channels, SE supported environment, Up updates, DF digital filtering, DV data visualization, SA spectral analysis, SAR stimulation artifact removal, FO file operations, MF multiple formats
Fig. 1Distribution of formats of local field potential signals in open datasets, extracted from [29]
Fig. 2Recording of extracellular neuronal signals from behaving rodents using linear implantable neural probe (shown in gray). Representative local field potential signals with and without movement artifacts are shown from two datasets. The blue traces denote signals without artifacts and the red traces show examples of movement artifacts present in the signals
Fig. 3Architectures of different neural network models: multi-layer perceptron (A), long short-term memory (B), and one-dimension convolutional neural network (C). Each circle represents a neuron, multiple rectangles a layer’s depth, and the arrows how the information is propagated throughout each network
Performance comparison, extracted from [72]
| Network | Accuracy | Parameters | Computational time (s) |
|---|---|---|---|
| 1D-CNN [ | 95.1 | 561218 | 2.27 ± 0.13 |
| MLP [ | 93.2 | 1532 | 2.57 ± 0.06 |
| LSTM [ | 87.1 | 4418 | 3.47 ± 0.04 |
Fig. 4Screenshots of the SANTIA toolbox graphical user interface: Data Labeling (A), Neural Network Training (B), and Classify New Unlabeled Data (C)
Fig. 5Functional block diagram of the Toolbox.Arrows in black correspond to the “Data Labeling” module , in red to the “Neural Network Training” module, in dark blue to the “Classify New Unlabeled Data” module, and the purple arrows indicate the progress output
Fig. 6Workflow of the SANTIA toolbox, where the “Data Labeling” modules are colored yellow, the “Neural Network Training” modules in green, and “Classify New Unlabeled Data” modules in blue
Guide to determine best channels and epochs to use of baseline walk and rest recordings in medial prefrontal cortex (mPFC) and the mediodorsal (MD) thalamus, as mentioned in the file named “Coherence Phase Plot Guide”
| Rat | mPFC chan1 | mPFC chan2 | MD chan1 | MD chan2 | Walk epoch | Rest epoch |
|---|---|---|---|---|---|---|
| KF9 | 5 | 6 | 3 | 7 | 960–1160 | 3780–3820 |
| KF10 | 3 | 4 | 3 | 8 | 670–860 | 1260–1390 |
| KF14 | 2 | 6 | 5 | 7 | 740–940 | 3350–3550 |
| KF15 | 3 | 4 | 5 | 7 | 450–640 | 1600–1700 |
| KF25 | 2 | 6 | 2 | 5 | 1480–1680 | 1700–1800 |
| KF26 | 1 | 6 | 1 | 6 | 1180–1380 | 1050–1150 |
| KF27 | 2 | 4 | 5 | 8 | 480–680 | 2160–2250 |
The first column is the rat identification, column 2 and 3 the selected two best channels of the mPFC recordings, and 4 and 5 of the MD recordings. Finally, column 6 shows the range of artifact-free epochs during walking and column 7 during resting, respectively [74]
Fig. 7Training plots for models trained with the first dataset
First dataset’s results for different architectures and sequence length: training loss, validation accuracy, testing accuracy, and testing AUROC
| Network | Sequence length (ms) | Training loss | Val. Acc. | Test Acc. | Test AUROC |
|---|---|---|---|---|---|
| MLP | 50 | 0.20 | 0.92 | 0.92 | 0.98 |
| 100 | 0.41 | 0.82 | 0.81 | 0.90 | |
| 150 | 0.39 | 0.83 | 0.83 | 0.90 | |
| 200 | 0.24 | 0.91 | 0.91 | 0.97 | |
| 1D-CNN | |||||
| 100 | 0.39 | 0.84 | 0.84 | 0.89 | |
| 150 | 0.37 | 0.83 | 0.83 | 0.91 | |
| 200 | 0.36 | 0.83 | 0.83 | 0.91 | |
| LSTM | 50 | 0.16 | 0.93 | 0.94 | 0.99 |
| 100 | 0.26 | 0.90 | 0.91 | 0.97 | |
| 150 | 0.25 | 0.89 | 0.90 | 0.97 | |
| 200 | 0.25 | 0.91 | 0.90 | 0.97 |
Values pertaining to model’s best performance are highlighted in bold
Fig. 8Training plots for models trained with the second dataset
Second dataset’s results for different architectures and sequence length: training loss, validation accuracy, testing accuracy, and testing AUROC
| Network | Sequence length (ms) | Training loss | Val. Acc. | Test Acc. | Test AUROC |
|---|---|---|---|---|---|
| MLP | 50 | 0.24 | 0.78 | 0.78 | 0.857 |
| 100 | 0.27 | 0.89 | 0.86 | 0.94 | |
| 150 | 0.15 | 0.94 | 0.95 | 0.99 | |
| 200 | 0.16 | 0.94 | 0.96 | 0.98 | |
| 1D-CNN | 50 | 0.18 | 0.92 | 0.91 | 0.97 |
| 100 | 0.15 | 0.94 | 0.96 | 0.97 | |
| 200 | 0.08 | 0.98 | 0.97 | 0.99 | |
| LSTM | 50 | 0.25 | 0.86 | 0.86 | 0.94 |
| 100 | 0.26 | 0.89 | 0.89 | 0.96 | |
| 150 | 0.02 | 0.97 | 0.97 | 0.99 | |
| 200 | 0.07 | 0.96 | 0.96 | 0.99 |
Values pertaining to model’s best performance are highlighted in bold
Fig. 9Screenshots of the toolbox’s threshold selection outputs: threshold selection table (A), a window of a non-artifactual signal (B), and a window of an artifactual signal (C)
Fig. 10Screenshots of the toolbox’s network training outputs: multi-layer perceptron training process (A) and one dimensional convolutional neural network training process (B)
Fig. 11Screenshots of the toolbox’s network test set results outputs: confusion matrix (A), AUROC curve (B), threshold selection window with default (C), and custom values (D)
Fig. 12Screenshots of the toolbox’s saved files: labeled data (A), trained network and results (B), and new data labels (C)