| Literature DB >> 28701911 |
Evangelos Stromatias1, Miguel Soto1, Teresa Serrano-Gotarredona1, Bernabé Linares-Barranco1.
Abstract
This paper introduces a novel methodology for training an event-driven classifier within a Spiking Neural Network (SNN) System capable of yielding good classification results when using both synthetic input data and real data captured from Dynamic Vision Sensor (DVS) chips. The proposed supervised method uses the spiking activity provided by an arbitrary topology of prior SNN layers to build histograms and train the classifier in the frame domain using the stochastic gradient descent algorithm. In addition, this approach can cope with leaky integrate-and-fire neuron models within the SNN, a desirable feature for real-world SNN applications, where neural activation must fade away after some time in the absence of inputs. Consequently, this way of building histograms captures the dynamics of spikes immediately before the classifier. We tested our method on the MNIST data set using different synthetic encodings and real DVS sensory data sets such as N-MNIST, MNIST-DVS, and Poker-DVS using the same network topology and feature maps. We demonstrate the effectiveness of our approach by achieving the highest classification accuracy reported on the N-MNIST (97.77%) and Poker-DVS (100%) real DVS data sets to date with a spiking convolutional network. Moreover, by using the proposed method we were able to retrain the output layer of a previously reported spiking neural network and increase its performance by 2%, suggesting that the proposed classifier can be used as the output layer in works where features are extracted using unsupervised spike-based learning methods. In addition, we also analyze SNN performance figures such as total event activity and network latencies, which are relevant for eventual hardware implementations. In summary, the paper aggregates unsupervised-trained SNNs with a supervised-trained SNN classifier, combining and applying them to heterogeneous sets of benchmarks, both synthetic and from real DVS chips.Entities:
Keywords: DVS sensors; convolutional neural networks; event driven processing; fully connected neural networks; neuromorphic; spiking neural networks; supervised learning
Year: 2017 PMID: 28701911 PMCID: PMC5487436 DOI: 10.3389/fnins.2017.00350
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Figure 12-D histograms and raster plots for different encoding schemes and neuromorphic data sets. (A) Poisson 28 × 28 input size sample. (B) Latency 28 × 28 input size sample. (C) MNIST-DVS 128 × 128 input size sample. (D) N-MNIST 34 × 34 input size sample. (E) Fast-Poker DVS 32 × 32 input size sample. (F) Slow-Poker DVS 128 × 128 input size sample.
Figure 2The topology of the event-driven Convnet used for this work. Where n is the size of the input layer, k is the size of the kernel.
Figure 3The pre-computed 7 × 7 gabor kernels used for the first convolutional layer.
The parameters used to generate the 18 2D Gabor kernels.
| θ | [ 0, 20, 40, 60, 80, 100, 120, 140, 160]° |
| σ | 4 |
| λ | 8 |
| ψ | [0.0, 1.7] |
| γ | 0.5 |
Figure 4The topology of the fully connected network as trained by O'Connor et al. (2013).
Figure 5The evolution and spike generation sequence for a spiking neuron with linear leakage and refractory period.
Parameters used for the C1 Convolution Layers and the FC Classifiers.
| Synthetic MNISTs | 28 × 28 | 18 | 22 × 22 | 7 × 7 | 1.2 × 108 | −1.2 × 108 | 1.2 × 104 | 1.2 × 104 | 10 | 107 | 1.2 × 104 |
| N-MNIST | 34 × 34 | 18 | 28 × 28 | 7 × 7 | 2 × 107 | −231 | 106 | 106 | 10 | 107 | 106 |
| MNIST-DVS | 128 × 128 | 18 | 108 × 108 | 21 × 21 | 8 × 108 | −231 | 1.5 × 104 | −1.5 × 104 | 10 | 107 | 1.2 × 104 |
| Slow-Poker-DVS | 128 × 128 | 18 | 114 × 114 | 15 × 15 | 1.5 × 107 | −1.5 × 107 | 1.5 × 104 | 1.5 × 104 | 4 | 107 | 1.5 × 105 |
| Fast-Poker-DVS | 32 × 32 | 18 | 26 × 26 | 7 × 7 | 1.5 × 107 | −231 | 1.5 × 104 | 1.5 × 104 | 4 | 107 | 1.5 × 105 |
Summary of performance results.
| Latency | Full | 98.45 | 98.42 | −0.03 | 151.12 | 957.37 | 10.89 | 98.41 | 98.39 | −0.02 | 98.37 | −0.04 |
| Poisson | Full | 98.15 | 98.20 | −0.05 | 1,000 | 16,070.28 | 11.38 | 98.3 | 98.25 | −0.05 | 98.32 | −0.02 |
| N-MNIST | Full | 97.77 | 97.23 | −0.54 | 4,203.61 | 29,9425.57 | 2,901.10 | 97.76 | 97.08 | −0.68 | 97.09 | −0.67 |
| MNIST-DVS | Full | 97.3 | 97.25 | −0.05 | 73,520.96 | 139,707.66 | 58,918.25 | 97.95 | 97.9 | −0.05 | 97.95 | 0.00 |
| Slow-Poker-DVS | 100 ms | 99.77 | 99.70 | −0.07 | 1,418.94 | 154,483.18 | 774.27 | 98.95 | 98.59 | −0.36 | 98.88 | −0.07 |
| 2 ms | 94.44 | 100 | +5.66 | 231662.29 | 100.67 | 92.78 | 100 | −7.22 | 100 | +7.22 | ||
| Fast-Poker-DVS | 5 ms | 95.00 | 94.65 | −0.35 | 2539.77 | 224598.88 | 124 | 95 | 93.13 | −1.87 | 93.89 | −1.11 |
| 10 ms | 100 | 83.97 | −6.03 | 222681.92 | 126.21 | 100 | 85.50 | −14.5 | 85.50 | −14.5 | ||
Figure 6Classification accuracies of the SNNs for each data set together with their corresponding confidence intervals.
Figure 7The classification accuracy of the SNNs for each data set as (A) a function of average percentage of input events per symbol presentation, and (B) average absolute number of input events per symbol presentation.
Figure 8Classification accuracy of the SNN using (A) the LOOCV method over the 131 samples Fast-Poker-DVS data set and (B) the LOOCV method over the 40 samples Fast-Poker-DVS data set.
Summary of fine-tuning an already trained SNN.
| Original Netowrk | 95.2 (O'Connor et al., | 95.26 |
| Fine-tuned (this work) | 97.24 | 97.25 |
Figure 9The mean and standard deviation of the classification latency of the SNNs for each data set.
Comparison of classification accuracies (CA) of SNNs on the MNIST data set.
| Spiking RBM (Neftci et al., | Poisson | Unsupervised | Event-Based CD | 91.9 |
| FC (2 layer network) (Querlioz et al., | Poisson | Unsupervised | STDP | 93.5 |
| FC (4 layer network) (O'Connor et al., | Poisson | Unsupervised | CD | 94.1 |
| FC (2 layer network) (Diehl and Cook, | Poisson | Unsupervised | STDP | 95.0 |
| Synaptic Sampling Machine (3 layer network) (Neftci et al., | Poisson | Unsupervised | Event-Based CD | 95.6 |
| FC (4 layer network) (this work(O'Connor et al., | Poisson | Supervised | Stochastic GD | 97.25 |
| FC (4 layer network) (O'Connor and Welling, | – | Supervised | Fractional SGD | 97.8 |
| FC (4 layer network) (Hunsberger and Eliasmith, | Not reported | Supervised | Backprop soft LIF neurons | 98.37 |
| FC (4 layer network) (Diehl et al., | Poisson | Supervised | Stochastic GD | 98.64 |
| CNN (Kheradpisheh et al., | Latency | Unsupervised | STDP | 98.4 |
| CNN (Diehl et al., | Poisson | Supervised | Stochastic GD | 99.14 |
| Sparsely Connected Network (×64) (Esser et al., | Poisson | Supervised | Backprop | 99.42 |
| CNN (Rueckauer et al., | Poisson | Supervised | Stochastic GD | 99.44 |
| CNN (this work) | Latency | Supervised | Stochastic GD | 98.42 |
| CNN (this work) | Poisson | Supervised | Stochastic GD | 98.20 |
Comparison of classification accuracies (CA) of SNNs on the N-MNIST data set.
| CNN (Orchard et al., | None | Unsupervised | HFirst | 71.15 |
| FC (2 layer network) (Cohen et al., | None | Supervised | OPIUM (van Schaik and Tapson, | 92.87 |
| CNN (Neil and Liu, | Centering | Supervised | – | 95.72 |
| FC (3 layer network) (Lee et al., | None | Supervised | Backpropagation | 98.66 |
| CNN (this work) | None | Supervised | SGD | 97.77 |
Comparison of classification accuracies (CA) of SNNs on the 40 cards Fast-Poker-DVS data set.
| CNN (Pérez-Carrasco et al., | Supervised | Backprop | 90.1 − 91.6 |
| CNN (Orchard et al., | Unsupervised | HFirst | 97.5 ± 3.5 |
| CNN (Lagorce et al., | Supervised | HOTS | 100 |
| CNN (this work) | Supervised | Stochastic GD | 100 |