Literature DB >> 12662875

Discovering Neural Nets with Low Kolmogorov Complexity and High Generalization Capability.

Jurgen Schmidhuber1.   

Abstract

Many neural net learning algorithms aim at finding "simple" nets to explain training data. The expectation is that the "simpler" the networks, the better the generalization on test data (--> Occam's razor). Previous implementations, however, use measures for "simplicity" that lack the power, universality and elegance of those based on Kolmogorov complexity and Solomonoff's algorithmic probability. Likewise, most previous approaches (especially those of the "Bayesian" kind) suffer from the problem of choosing appropriate priors. This paper addresses both issues. It first reviews some basic concepts of algorithmic complexity theory relevant to machine learing, and how the Solomonoff-Levin distribution (or universal prior) deals with the prior problem. The universal prior leads to a probabilistic method for finding "algorithmically simple" problem solutions with high generalization capability. The method is based on Levin complexity (a time-bounded generalization of Kolmogorov complexity) and inspired by Levin's optimal universal search algorithm. For a given problem, solution candidates are computed by efficient "self-sizing" programs that influence their own runtime and storage size. The probabilistic search algorithm finds the "good" programs (the ones quickly computing algorithmically probable solutions fitting the training data). Simulations focus on the task of discovering "algorithmically simple" neural networks with low Kolmogorov complexity and high generalization capability. It is demonstrated that the method, at least with certain toy problems where it is computationally feasible, can lead to generalization results unmatchable by previous neural network algorithms. Much remains to be done, however, to make large scale applications and "incremental learning" feasible. Copyright 1997 Elsevier Science Ltd.

Entities:  

Year:  1997        PMID: 12662875     DOI: 10.1016/s0893-6080(96)00127-x

Source DB:  PubMed          Journal:  Neural Netw        ISSN: 0893-6080


  3 in total

1.  Divide et impera: subgoaling reduces the complexity of probabilistic inference and problem solving.

Authors:  Domenico Maisto; Francesco Donnarumma; Giovanni Pezzulo
Journal:  J R Soc Interface       Date:  2015-03-06       Impact factor: 4.118

2.  Problem Solving as Probabilistic Inference with Subgoaling: Explaining Human Successes and Pitfalls in the Tower of Hanoi.

Authors:  Francesco Donnarumma; Domenico Maisto; Giovanni Pezzulo
Journal:  PLoS Comput Biol       Date:  2016-04-13       Impact factor: 4.475

3.  Cryostructuring of Polymeric Systems : Application of Deep Neural Networks for the Classification of Structural Features Peculiar to Macroporous Poly(vinyl alcohol) Cryogels Prepared without and with the Additives of Chaotropes or Kosmotropes.

Authors:  Ilya I Kurochkin; Ilya N Kurochkin; Olga Yu Kolosova; Vladimir I Lozinsky
Journal:  Molecules       Date:  2020-09-29       Impact factor: 4.411

  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.