Literature DB >> 32454374

Sparsity through evolutionary pruning prevents neuronal networks from overfitting.

Richard C Gerum1, André Erpenbeck2, Patrick Krauss3, Achim Schilling4.   

Abstract

Modern Machine learning techniques take advantage of the exponentially rising calculation power in new generation processor units. Thus, the number of parameters which are trained to solve complex tasks was highly increased over the last decades. However, still the networks fail - in contrast to our brain - to develop general intelligence in the sense of being able to solve several complex tasks with only one network architecture. This could be the case because the brain is not a randomly initialized neural network, which has to be trained from scratch by simply investing a lot of calculation power, but has from birth some fixed hierarchical structure. To make progress in decoding the structural basis of biological neural networks we here chose a bottom-up approach, where we evolutionarily trained small neural networks in performing a maze task. This simple maze task requires dynamic decision making with delayed rewards. We were able to show that during the evolutionary optimization random severance of connections leads to better generalization performance of the networks compared to fully connected networks. We conclude that sparsity is a central property of neural networks and should be considered for modern Machine learning approaches.
Copyright © 2020 The Author(s). Published by Elsevier Ltd.. All rights reserved.

Keywords:  Artificial neural networks; Biological plausibility; Evolution; Evolutionary algorithm; Maze task; Overfitting

Year:  2020        PMID: 32454374     DOI: 10.1016/j.neunet.2020.05.007

Source DB:  PubMed          Journal:  Neural Netw        ISSN: 0893-6080


  4 in total

1.  Dynamics and Information Import in Recurrent Neural Networks.

Authors:  Claus Metzner; Patrick Krauss
Journal:  Front Comput Neurosci       Date:  2022-04-27       Impact factor: 3.387

2.  Intrinsic Noise Improves Speech Recognition in a Computational Model of the Auditory Pathway.

Authors:  Achim Schilling; Richard Gerum; Claus Metzner; Andreas Maier; Patrick Krauss
Journal:  Front Neurosci       Date:  2022-06-08       Impact factor: 5.152

3.  The impact of sparsity in low-rank recurrent neural networks.

Authors:  Elizabeth Herbert; Srdjan Ostojic
Journal:  PLoS Comput Biol       Date:  2022-08-09       Impact factor: 4.779

4.  Neural network based successor representations to form cognitive maps of space and language.

Authors:  Paul Stoewer; Christian Schlieker; Achim Schilling; Claus Metzner; Andreas Maier; Patrick Krauss
Journal:  Sci Rep       Date:  2022-07-04       Impact factor: 4.996

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.