Literature DB >> 33733123

Interpretability With Accurate Small Models.

Abhishek Ghose1, Balaraman Ravindran2.   

Abstract

Models often need to be constrained to a certain size for them to be considered interpretable. For example, a decision tree of depth 5 is much easier to understand than one of depth 50. Limiting model size, however, often reduces accuracy. We suggest a practical technique that minimizes this trade-off between interpretability and classification accuracy. This enables an arbitrary learning algorithm to produce highly accurate small-sized models. Our technique identifies the training data distribution to learn from that leads to the highest accuracy for a model of a given size. We represent the training distribution as a combination of sampling schemes. Each scheme is defined by a parameterized probability mass function applied to the segmentation produced by a decision tree. An Infinite Mixture Model with Beta components is used to represent a combination of such schemes. The mixture model parameters are learned using Bayesian Optimization. Under simplistic assumptions, we would need to optimize for O(d) variables for a distribution over a d-dimensional input space, which is cumbersome for most real-world data. However, we show that our technique significantly reduces this number to a fixed set of eight variables at the cost of relatively cheap preprocessing. The proposed technique is flexible: it is model-agnostic, i.e., it may be applied to the learning algorithm for any model family, and it admits a general notion of model size. We demonstrate its effectiveness using multiple real-world datasets to construct decision trees, linear probability models and gradient boosted models with different sizes. We observe significant improvements in the F1-score in most instances, exceeding an improvement of 100% in some cases.
Copyright © 2020 Ghose and Ravindran.

Entities:  

Keywords:  Bayesian optimization; ML; density estimation; infinite mixture models; interpretable machine learning

Year:  2020        PMID: 33733123      PMCID: PMC7861231          DOI: 10.3389/frai.2020.00003

Source DB:  PubMed          Journal:  Front Artif Intell        ISSN: 2624-8212


  9 in total

Review 1.  Completely derandomized self-adaptation in evolution strategies.

Authors:  N Hansen; A Ostermeier
Journal:  Evol Comput       Date:  2001       Impact factor: 3.277

2.  Rotation forest: A new classifier ensemble method.

Authors:  Juan J Rodríguez; Ludmila I Kuncheva; Carlos J Alonso
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2006-10       Impact factor: 6.226

3.  A comparison of methods for multiclass support vector machines.

Authors:  Chih-Wei Hsu; Chih-Jen Lin
Journal:  IEEE Trans Neural Netw       Date:  2002

4.  Optimization by simulated annealing.

Authors:  S Kirkpatrick; C D Gelatt; M P Vecchi
Journal:  Science       Date:  1983-05-13       Impact factor: 47.728

5.  Learning interactions via hierarchical group-lasso regularization.

Authors:  Michael Lim; Trevor Hastie
Journal:  J Comput Graph Stat       Date:  2015-09-16       Impact factor: 2.302

6.  Searching for exotic particles in high-energy physics with deep learning.

Authors:  P Baldi; P Sadowski; D Whiteson
Journal:  Nat Commun       Date:  2014-07-02       Impact factor: 14.919

7.  Interpretable Decision Sets: A Joint Framework for Description and Prediction.

Authors:  Himabindu Lakkaraju; Stephen H Bach; Leskovec Jure
Journal:  KDD       Date:  2016-08

8.  Distributed Newton Methods for Deep Neural Networks.

Authors:  Chien-Chih Wang; Kent Loong Tan; Chun-Ting Chen; Yu-Hsiang Lin; S Sathiya Keerthi; Dhruv Mahajan; S Sundararajan; Chih-Jen Lin
Journal:  Neural Comput       Date:  2018-04-13       Impact factor: 2.026

9.  Detection of non-coding RNAs on the basis of predicted secondary structure formation free energy change.

Authors:  Andrew V Uzilov; Joshua M Keegan; David H Mathews
Journal:  BMC Bioinformatics       Date:  2006-03-27       Impact factor: 3.169

  9 in total
  1 in total

1.  Post-Analysis of Predictive Modeling with an Epidemiological Example.

Authors:  Christina Brester; Ari Voutilainen; Tomi-Pekka Tuomainen; Jussi Kauhanen; Mikko Kolehmainen
Journal:  Healthcare (Basel)       Date:  2021-06-24
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.