Literature DB >> 30951462

Learning Representations for Neural Network-Based Classification Using the Information Bottleneck Principle.

Rana Ali Amjad, Bernhard C Geiger.   

Abstract

In this theory paper, we investigate training deep neural networks (DNNs) for classification via minimizing the information bottleneck (IB) functional. We show that the resulting optimization problem suffers from two severe issues: First, for deterministic DNNs, either the IB functional is infinite for almost all values of network parameters, making the optimization problem ill-posed, or it is piecewise constant, hence not admitting gradient-based optimization methods. Second, the invariance of the IB functional under bijections prevents it from capturing properties of the learned representation that are desirable for classification, such as robustness and simplicity. We argue that these issues are partly resolved for stochastic DNNs, DNNs that include a (hard or soft) decision rule, or by replacing the IB functional with related, but more well-behaved cost functions. We conclude that recent successes reported about training DNNs using the IB framework must be attributed to such solutions. As a side effect, our results indicate limitations of the IB framework for the analysis of DNNs. We also note that rather than trying to repair the inherent problems in the IB functional, a better approach may be to design regularizers on latent representation enforcing the desired properties directly.

Year:  2019        PMID: 30951462     DOI: 10.1109/TPAMI.2019.2909031

Source DB:  PubMed          Journal:  IEEE Trans Pattern Anal Mach Intell        ISSN: 0098-5589            Impact factor:   6.226


  9 in total

1.  The Convex Information Bottleneck Lagrangian.

Authors:  Borja Rodríguez Gálvez; Ragnar Thobaben; Mikael Skoglund
Journal:  Entropy (Basel)       Date:  2020-01-14       Impact factor: 2.524

2.  An Information Theoretic Approach to Reveal the Formation of Shared Representations.

Authors:  Akihiro Eguchi; Takato Horii; Takayuki Nagai; Ryota Kanai; Masafumi Oizumi
Journal:  Front Comput Neurosci       Date:  2020-01-29       Impact factor: 2.380

3.  On the Information Bottleneck Problems: Models, Connections, Applications and Information Theoretic Views.

Authors:  Abdellatif Zaidi; Iñaki Estella-Aguerri; Shlomo Shamai Shitz
Journal:  Entropy (Basel)       Date:  2020-01-27       Impact factor: 2.524

4.  Pareto-Optimal Data Compression for Binary Classification Tasks.

Authors:  Max Tegmark; Tailin Wu
Journal:  Entropy (Basel)       Date:  2019-12-19       Impact factor: 2.524

5.  Convergence Behavior of DNNs with Mutual-Information-Based Regularization.

Authors:  Hlynur Jónsson; Giovanni Cherubini; Evangelos Eleftheriou
Journal:  Entropy (Basel)       Date:  2020-06-30       Impact factor: 2.524

6.  Variational Information Bottleneck for Semi-Supervised Classification.

Authors:  Slava Voloshynovskiy; Olga Taran; Mouad Kondah; Taras Holotyak; Danilo Rezende
Journal:  Entropy (Basel)       Date:  2020-08-27       Impact factor: 2.524

7.  The Conditional Entropy Bottleneck.

Authors:  Ian Fischer
Journal:  Entropy (Basel)       Date:  2020-09-08       Impact factor: 2.524

8.  Examining the Causal Structures of Deep Neural Networks Using Information Theory.

Authors:  Scythia Marrow; Eric J Michaud; Erik Hoel
Journal:  Entropy (Basel)       Date:  2020-12-18       Impact factor: 2.524

9.  Heuristic Attention Representation Learning for Self-Supervised Pretraining.

Authors:  Van Nhiem Tran; Shen-Hsuan Liu; Yung-Hui Li; Jia-Ching Wang
Journal:  Sensors (Basel)       Date:  2022-07-10       Impact factor: 3.847

  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.