Literature DB >> 34094697

Optimizing Nondecomposable Data Dependent Regularizers via Lagrangian Reparameterization Offers Significant Performance and Efficiency Gains.

Sathya N Ravi1, Abhay Venkatesh2, Glenn M Fung3, Vikas Singh2.   

Abstract

Data dependent regularization is known to benefit a wide variety of problems in machine learning. Often, these regularizers cannot be easily decomposed into a sum over a finite number of terms, e.g., a sum over individual example-wise terms. The F β measure, Area under the ROC curve (AUCROC) and Precision at a fixed recall (P@R) are some prominent examples that are used in many applications. We find that for most medium to large sized datasets, scalability issues severely limit our ability in leveraging the benefits of such regularizers. Importantly, the key technical impediment despite some recent progress is that, such objectives remain difficult to optimize via backpropapagation procedures. While an efficient general-purpose strategy for this problem still remains elusive, in this paper, we show that for many data-dependent nondecomposable regularizers that are relevant in applications, sizable gains in efficiency are possible with minimal code-level changes; in other words, no specialized tools or numerical schemes are needed. Our procedure involves a reparameterization followed by a partial dualization - this leads to a formulation that has provably cheap projection operators. We present a detailed analysis of runtime and convergence properties of our algorithm. On the experimental side, we show that a direct use of our scheme significantly improves the state of the art IOU measures reported for MSCOCO Stuff segmentation dataset.

Entities:  

Year:  2020        PMID: 34094697      PMCID: PMC8174799          DOI: 10.1609/aaai.v34i04.5999

Source DB:  PubMed          Journal:  Proc Conf AAAI Artif Intell        ISSN: 2159-5399


  5 in total

1.  Beyond the t-test: statistical equivalence testing.

Authors:  Giselle B Limentani; Moira C Ringo; Feng Ye; Mandy L Berquist; Ellen O McSorley
Journal:  Anal Chem       Date:  2005-06-01       Impact factor: 6.986

2.  Optimal Thresholding of Classifiers to Maximize F1 Measure.

Authors:  Zachary C Lipton; Charles Elkan; Balakrishnan Naryanaswamy
Journal:  Mach Learn Knowl Discov Databases       Date:  2014

3.  SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.

Authors:  Vijay Badrinarayanan; Alex Kendall; Roberto Cipolla
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2017-01-02       Impact factor: 6.226

4.  Object Detection With Deep Learning: A Review.

Authors:  Zhong-Qiu Zhao; Peng Zheng; Shou-Tao Xu; Xindong Wu
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2019-01-28       Impact factor: 10.451

5.  When can Multi-Site Datasets be Pooled for Regression? Hypothesis Tests, 2-consistency and Neuroscience Applications.

Authors:  Hao Henry Zhou; Yilin Zhang; Vamsi K Ithapu; Sterling C Johnson; Grace Wahba; Vikas Singh
Journal:  Proc Mach Learn Res       Date:  2017-08
  5 in total
  2 in total

1.  Differentiable Optimization of Generalized Nondecomposable Functions using Linear Programs.

Authors:  Zihang Meng; Lopamudra Mukherjee; Yichao Wu; Vikas Singh; Sathya N Ravi
Journal:  Adv Neural Inf Process Syst       Date:  2021-12

2.  Physarum Powered Differentiable Linear Programming Layers and Applications.

Authors:  Zihang Meng; Sathya N Ravi; Vikas Singh
Journal:  Proc Conf AAAI Artif Intell       Date:  2021-05-18
  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.