Literature DB >> 26431018

Subdominant Dense Clusters Allow for Simple Learning and High Computational Performance in Neural Networks with Discrete Synapses.

Carlo Baldassi1,2, Alessandro Ingrosso1,2, Carlo Lucibello1,2, Luca Saglietti1,2, Riccardo Zecchina1,2,3.   

Abstract

We show that discrete synaptic weights can be efficiently used for learning in large scale neural systems, and lead to unanticipated computational performance. We focus on the representative case of learning random patterns with binary synapses in single layer networks. The standard statistical analysis shows that this problem is exponentially dominated by isolated solutions that are extremely hard to find algorithmically. Here, we introduce a novel method that allows us to find analytical evidence for the existence of subdominant and extremely dense regions of solutions. Numerical experiments confirm these findings. We also show that the dense regions are surprisingly accessible by simple learning protocols, and that these synaptic configurations are robust to perturbations and generalize better than typical solutions. These outcomes extend to synapses with multiple states and to deeper neural architectures. The large deviation measure also suggests how to design novel algorithmic schemes for optimization based on local entropy maximization.

Year:  2015        PMID: 26431018     DOI: 10.1103/PhysRevLett.115.128101

Source DB:  PubMed          Journal:  Phys Rev Lett        ISSN: 0031-9007            Impact factor:   9.161


  7 in total

1.  Unreasonable effectiveness of learning neural networks: From accessible states and robust ensembles to basic algorithmic schemes.

Authors:  Carlo Baldassi; Christian Borgs; Jennifer T Chayes; Alessandro Ingrosso; Carlo Lucibello; Luca Saglietti; Riccardo Zecchina
Journal:  Proc Natl Acad Sci U S A       Date:  2016-11-15       Impact factor: 11.205

2.  PAC Bayesian Performance Guarantees for Deep (Stochastic) Networks in Medical Imaging.

Authors:  Anthony Sicilia; Xingchen Zhao; Anastasia Sosnovskikh; Seong Jae Hwang
Journal:  Med Image Comput Comput Assist Interv       Date:  2021-09-21

3.  Shaping the learning landscape in neural networks around wide flat minima.

Authors:  Carlo Baldassi; Fabrizio Pittorino; Riccardo Zecchina
Journal:  Proc Natl Acad Sci U S A       Date:  2019-12-23       Impact factor: 11.205

4.  Variational Characterizations of Local Entropy and Heat Regularization in Deep Learning.

Authors:  Nicolas García Trillos; Zachary Kaplan; Daniel Sanz-Alonso
Journal:  Entropy (Basel)       Date:  2019-05-20       Impact factor: 2.524

5.  Efficiency of quantum vs. classical annealing in nonconvex learning problems.

Authors:  Carlo Baldassi; Riccardo Zecchina
Journal:  Proc Natl Acad Sci U S A       Date:  2018-01-30       Impact factor: 11.205

6.  Optimization of neural networks via finite-value quantum fluctuations.

Authors:  Masayuki Ohzeki; Shuntaro Okada; Masayoshi Terabe; Shinichiro Taguchi
Journal:  Sci Rep       Date:  2018-07-02       Impact factor: 4.379

7.  Generalization properties of neural network approximations to frustrated magnet ground states.

Authors:  Tom Westerhout; Nikita Astrakhantsev; Konstantin S Tikhonov; Mikhail I Katsnelson; Andrey A Bagrov
Journal:  Nat Commun       Date:  2020-03-27       Impact factor: 14.919

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.