Literature DB >> 33816904

Beyond the topics: how deep learning can improve the discriminability of probabilistic topic modelling.

Noura Al Moubayed1, Stephen McGough2, Bashar Awwad Shiekh Hasan3.   

Abstract

The article presents a discriminative approach to complement the unsupervised probabilistic nature of topic modelling. The framework transforms the probabilities of the topics per document into class-dependent deep learning models that extract highly discriminatory features suitable for classification. The framework is then used for sentiment analysis with minimum feature engineering. The approach transforms the sentiment analysis problem from the word/document domain to the topics domain making it more robust to noise and incorporating complex contextual information that are not represented otherwise. A stacked denoising autoencoder (SDA) is then used to model the complex relationship among the topics per sentiment with minimum assumptions. To achieve this, a distinct topic model and SDA per sentiment polarity is built with an additional decision layer for classification. The framework is tested on a comprehensive collection of benchmark datasets that vary in sample size, class bias and classification task. A significant improvement to the state of the art is achieved without the need for a sentiment lexica or over-engineered features. A further analysis is carried out to explain the observed improvement in accuracy.
© 2020 Al Moubayed et al.

Entities:  

Keywords:  Sentiment analysis; Stacked denoising autoencoders; Text classification; Topic modelling

Year:  2020        PMID: 33816904      PMCID: PMC7924555          DOI: 10.7717/peerj-cs.252

Source DB:  PubMed          Journal:  PeerJ Comput Sci        ISSN: 2376-5992


  8 in total

1.  Nonlinear autoassociation is not equivalent to PCA.

Authors:  N Japkowicz; M A Gluck
Journal:  Neural Comput       Date:  2000-03       Impact factor: 2.026

2.  Linear recursive distributed representations.

Authors:  Thomas Voegtlin; Peter F Dominey
Journal:  Neural Netw       Date:  2005-09

3.  A fast learning algorithm for deep belief nets.

Authors:  Geoffrey E Hinton; Simon Osindero; Yee-Whye Teh
Journal:  Neural Comput       Date:  2006-07       Impact factor: 2.026

4.  Reducing the dimensionality of data with neural networks.

Authors:  G E Hinton; R R Salakhutdinov
Journal:  Science       Date:  2006-07-28       Impact factor: 47.728

Review 5.  Representation learning: a review and new perspectives.

Authors:  Yoshua Bengio; Aaron Courville; Pascal Vincent
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2013-08       Impact factor: 6.226

6.  Auto-association by multilayer perceptrons and singular value decomposition.

Authors:  H Bourlard; Y Kamp
Journal:  Biol Cybern       Date:  1988       Impact factor: 2.086

7.  Probabilistic Topic Models: A focus on graphical model design and applications to document and image analysis.

Authors:  David Blei; Lawrence Carin; David Dunson
Journal:  IEEE Signal Process Mag       Date:  2010-11-01       Impact factor: 12.551

8.  How inherently noisy is human sensory processing?

Authors:  Peter Neri
Journal:  Psychon Bull Rev       Date:  2010-12
  8 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.