Literature DB >> 31417340

RAPA-ConvNets: Modified Convolutional Networks for Accelerated Training on Architectures With Analog Arrays.

Malte J Rasch1, Tayfun Gokmen1, Mattia Rigotti1, Wilfried Haensch1.   

Abstract

Analog arrays are a promising emerging hardware technology with the potential to drastically speed up deep learning. Their main advantage is that they employ analog circuitry to compute matrix-vector products in constant time, irrespective of the size of the matrix. However, ConvNets map very unfavorably onto analog arrays when done in a straight-forward manner, because kernel matrices are typically small and the constant time operation needs to be sequentially iterated a large number of times. Here, we propose to parallelize the training by replicating the kernel matrix of a convolution layer on distinct analog arrays, and randomly divide parts of the compute among them. With this modification, analog arrays execute ConvNets with a large acceleration factor that is proportional to the number of kernel matrices used per layer (here tested 16-1024). Despite having more free parameters, we show analytically and in numerical experiments that this new convolution architecture is self-regularizing and implicitly learns similar filters across arrays. We also report superior performance on a number of datasets and increased robustness to adversarial attacks. Our investigation suggests to revise the notion that emerging hardware architectures that feature analog arrays for fast matrix-vector multiplication are not suitable for ConvNets.

Entities:  

Keywords:  analog computing; convolutional networks; emerging technologies; hardware acceleration of deep learning; machine learning; resistive cross-point devices

Year:  2019        PMID: 31417340      PMCID: PMC6682637          DOI: 10.3389/fnins.2019.00753

Source DB:  PubMed          Journal:  Front Neurosci        ISSN: 1662-453X            Impact factor:   4.677


  6 in total

1.  Equivalent-accuracy accelerated neural-network training using analogue memory.

Authors:  Stefano Ambrogio; Pritish Narayanan; Hsinyu Tsai; Robert M Shelby; Irem Boybat; Carmelo di Nolfo; Severin Sidler; Massimo Giordano; Martina Bodini; Nathan C P Farinha; Benjamin Killeen; Christina Cheng; Yassine Jaoudi; Geoffrey W Burr
Journal:  Nature       Date:  2018-06-06       Impact factor: 49.962

2.  Memristive devices for computing.

Authors:  J Joshua Yang; Dmitri B Strukov; Duncan R Stewart
Journal:  Nat Nanotechnol       Date:  2013-01       Impact factor: 39.213

3.  Energy-Efficient Neuromorphic Classifiers.

Authors:  Daniel Martí; Mattia Rigotti; Mingoo Seok; Stefano Fusi
Journal:  Neural Comput       Date:  2016-08-24       Impact factor: 2.026

4.  Training Deep Convolutional Neural Networks with Resistive Cross-Point Devices.

Authors:  Tayfun Gokmen; Murat Onen; Wilfried Haensch
Journal:  Front Neurosci       Date:  2017-10-10       Impact factor: 4.677

5.  Training LSTM Networks With Resistive Cross-Point Devices.

Authors:  Tayfun Gokmen; Malte J Rasch; Wilfried Haensch
Journal:  Front Neurosci       Date:  2018-10-24       Impact factor: 4.677

6.  Acceleration of Deep Neural Network Training with Resistive Cross-Point Devices: Design Considerations.

Authors:  Tayfun Gokmen; Yurii Vlasov
Journal:  Front Neurosci       Date:  2016-07-21       Impact factor: 4.677

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.