| Literature DB >> 29375284 |
Bodo Rueckauer1, Iulia-Alexandra Lungu1, Yuhuang Hu1, Michael Pfeiffer1,2, Shih-Chii Liu1.
Abstract
Spiking neural networks (SNNs) can potentially offer an efficient way of doing inference because the neurons in the networks are sparsely activated and computations are event-driven. Previous work showed that simple continuous-valued deep Convolutional Neural Networks (CNNs) can be converted into accurate spiking equivalents. These networks did not include certain common operations such as max-pooling, softmax, batch-normalization and Inception-modules. This paper presents spiking equivalents of these operations therefore allowing conversion of nearly arbitrary CNN architectures. We show conversion of popular CNN architectures, including VGG-16 and Inception-v3, into SNNs that produce the best results reported to date on MNIST, CIFAR-10 and the challenging ImageNet dataset. SNNs can trade off classification error rate against the number of available operations whereas deep continuous-valued neural networks require a fixed number of operations to achieve their classification error rate. From the examples of LeNet for MNIST and BinaryNet for CIFAR-10, we show that with an increase in error rate of a few percentage points, the SNNs can achieve more than 2x reductions in operations compared to the original CNNs. This highlights the potential of SNNs in particular when deployed on power-efficient neuromorphic spiking neuron chips, for use in embedded applications.Entities:
Keywords: artificial neural network; deep learning; deep networks; object classification; spiking network conversion; spiking neural network
Year: 2017 PMID: 29375284 PMCID: PMC5770641 DOI: 10.3389/fnins.2017.00682
Source DB: PubMed Journal: Front Neurosci ISSN: 1662-453X Impact factor: 4.677
Figure 1Distribution of all non-zero activations in the first convolution layer of a CNN, for 16666 CIFAR10 samples, and plotted in log-scale. The dashed line in both plots indicates the 99.9th percentile of all ReLU activations across the dataset, corresponding to a normalization scale λ = 6.83. This is more than three times less than the overall maximum of λ = 23.16. (B) Distribution of maximum ReLU activations for the same 16666 CIFAR10 samples. For most samples their maximum activation is far from λ. (A) ANN activations; (B) Maximum ANN activation.
Classification error rate on MNIST, CIFAR-10 and ImageNet for our converted spiking models, compared to the original ANNs, and compared to spiking networks from other groups.
| MNIST [ours] | 0.56 | 8 k | 1.2 M | |
| MNIST [Zambrano and Bohte, | 0.86 | 0.86 | 27 k | 6.6 M |
| CIFAR-10 [ours, BinaryNet sign] | 11.03 | 11.75 | 0.5 M | 164 M |
| CIFAR-10 [ours, BinaryNet Heav] | 11.58 | 12.55 | 0.5 M | 164 M |
| CIFAR-10 [ours, BinaryConnect, binarized at infer.] | 16.81 | 16.65 | 0.5 M | 164 M |
| CIFAR-10 [ours, BinaryConnect, full prec. at infer.] | 8.09 | 0.5 M | 164 M | |
| CIFAR-10 [ours] | 11.13 | 11.18 | 0.1 M | 23 M |
| CIFAR-10 [Esser et al., | NA | 12.50 | 8 M | NA |
| CIFAR-10 [Esser et al., | NA | 17.50 | 1 M | NA |
| CIFAR-10 [Hunsberger and Eliasmith, | 14.03 | 16.46 | 50 k | NA |
| CIFAR-10 [Cao et al., | 20.88 | 22.57 | 35 k | 7.4 M |
| ImageNet [ours, VGG-16] | 36.11 (15.14) | 50.39 (18.37) | 15 M | 3.5 B |
| ImageNet [ours, Inception-V3] | 23.88 (7.01) | 11.7 M | 0.5 B | |
| ImageNet [Hunsberger and Eliasmith, | NA | 48.20 (23.80) | 0.5 M | NA |
The reported error rate is top-1, with top-5 in brackets for ImageNet.
Cropped to 24x24.
Cropped to 24x24.
On a subset of 2570 samples, using single-scale images of size 224x224.
On a subset of 1382 samples, using single-scale images of size 299x299.
On a subset of 3072 samples. The values in bold highlight the best SNN result for a particular data set.
Figure 2Influence of novel mechanisms for ANN-to-SNN conversion on the SNN error rate for CIFAR-10.
Figure 3Accuracy-latency trade-off. Robust parameter normalization (red) enables our spiking network to correctly classify CIFAR-10 samples much faster than using our previous max-normalization (green). Not normalizing leads to classification at chance level (blue).
Figure 4Classification error rate vs number of operations for the BinaryNet ANN and SNN implementation on the complete CIFAR-10 dataset.
Figure 5Classification error rate vs number of operations for the LeNet ANN and SNN implementation on the MNIST dataset.