Literature DB >> 33343224

Learning feature spaces for regression with genetic programming.

William La Cava1, Jason H Moore1.   

Abstract

Genetic programming has found recent success as a tool for learning sets of features for regression and classification. Multidimensional genetic programming is a useful variant of genetic programming for this task because it represents candidate solutions as sets of programs. These sets of programs expose additional information that can be exploited for building block identification. In this work, we discuss this architecture and others in terms of their propensity for allowing heuristic search to utilize information during the evolutionary process. We investigate methods for biasing the components of programs that are promoted in order to guide search towards useful and complementary feature spaces. We study two main approaches: 1) the introduction of new objectives and 2) the use of specialized semantic variation operators. We find that a semantic crossover operator based on stagewise regression leads to significant improvements on a set of regression problems. The inclusion of semantic crossover produces state-of-the-art results in a large benchmark study of open-source regression problems in comparison to several state-of-the-art machine learning approaches and other genetic programming frameworks. Finally, we look at the collinearity and complexity of the data representations produced by different methods, in order to assess whether relevant, concise, and independent factors of variation can be produced in application.

Entities:  

Keywords:  feature construction; regression; representation learning; variation

Year:  2020        PMID: 33343224      PMCID: PMC7748157          DOI: 10.1007/s10710-020-09383-4

Source DB:  PubMed          Journal:  Genet Program Evolvable Mach        ISSN: 1389-2576            Impact factor:   2.522


  8 in total

1.  Evolving neural networks through augmenting topologies.

Authors:  Kenneth O Stanley; Risto Miikkulainen
Journal:  Evol Comput       Date:  2002       Impact factor: 3.277

2.  Why Deep Learning Works: A Manifold Disentanglement Perspective.

Authors:  Pratik Prabhanjan Brahma; Dapeng Wu; Yiyuan She
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2015-12-07       Impact factor: 10.451

3.  A hypercube-based encoding for evolving large-scale neural networks.

Authors:  Kenneth O Stanley; David B D'Ambrosio; Jason Gauci
Journal:  Artif Life       Date:  2009       Impact factor: 0.667

4.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

Review 5.  Representation learning: a review and new perspectives.

Authors:  Yoshua Bengio; Aaron Courville; Pascal Vincent
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2013-08       Impact factor: 6.226

6.  Diagnosis of multiple cancer types by shrunken centroids of gene expression.

Authors:  Robert Tibshirani; Trevor Hastie; Balasubramanian Narasimhan; Gilbert Chu
Journal:  Proc Natl Acad Sci U S A       Date:  2002-05-14       Impact factor: 11.205

7.  A Probabilistic and Multi-Objective Analysis of Lexicase Selection and ε-Lexicase Selection.

Authors:  William La Cava; Thomas Helmuth; Lee Spector; Jason H Moore
Journal:  Evol Comput       Date:  2018-05-10       Impact factor: 4.766

8.  PMLB: a large benchmark suite for machine learning evaluation and comparison.

Authors:  Randal S Olson; William La Cava; Patryk Orzechowski; Ryan J Urbanowicz; Jason H Moore
Journal:  BioData Min       Date:  2017-12-11       Impact factor: 2.522

  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.