Literature DB >> 33478002

Information-Theoretic Generalization Bounds for Meta-Learning and Applications.

Sharu Theresa Jose1, Osvaldo Simeone1.   

Abstract

Meta-learning, or "learning to learn", refers to techniques that infer an inductive bias from data corresponding to multiple related tasks with the goal of improving the sample efficiency for new, previously unobserved, tasks. A key performance measure for meta-learning is the meta-generalization gap, that is, the difference between the average loss measured on the meta-training data and on a new, randomly selected task. This paper presents novel information-theoretic upper bounds on the meta-generalization gap. Two broad classes of meta-learning algorithms are considered that use either separate within-task training and test sets, like model agnostic meta-learning (MAML), or joint within-task training and test sets, like reptile. Extending the existing work for conventional learning, an upper bound on the meta-generalization gap is derived for the former class that depends on the mutual information (MI) between the output of the meta-learning algorithm and its input meta-training data. For the latter, the derived bound includes an additional MI between the output of the per-task learning procedure and corresponding data set to capture within-task uncertainty. Tighter bounds are then developed for the two classes via novel individual task MI (ITMI) bounds. Applications of the derived bounds are finally discussed, including a broad class of noisy iterative algorithms for meta-learning.

Entities:  

Keywords:  generalization bounds; meta-learning; mutual information; noisy iterative algorithms

Year:  2021        PMID: 33478002      PMCID: PMC7835863          DOI: 10.3390/e23010126

Source DB:  PubMed          Journal:  Entropy (Basel)        ISSN: 1099-4300            Impact factor:   2.524


  4 in total

1.  Algorithmic stability and sanity-check bounds for leave-one-out cross-validation.

Authors:  M Kearns; D Ron
Journal:  Neural Comput       Date:  1999-08-15       Impact factor: 2.026

2.  General conditions for predictivity in learning theory.

Authors:  Tomaso Poggio; Ryan Rifkin; Sayan Mukherjee; Partha Niyogi
Journal:  Nature       Date:  2004-03-25       Impact factor: 49.962

3.  A general framework for updating belief distributions.

Authors:  P G Bissiri; C C Holmes; S G Walker
Journal:  J R Stat Soc Series B Stat Methodol       Date:  2016-02-23       Impact factor: 4.488

4.  Towards a Unified Theory of Learning and Information.

Authors:  Ibrahim Alabdulmohsin
Journal:  Entropy (Basel)       Date:  2020-04-13       Impact factor: 2.524

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.