Literature DB >> 32584774

Automated Social Text Annotation With Joint Multilabel Attention Networks.

Hang Dong, Wei Wang, Kaizhu Huang, Frans Coenen.   

Abstract

Automated social text annotation is the task of suggesting a set of tags for shared documents on social media platforms. The automated annotation process can reduce users' cognitive overhead in tagging and improve tag management for better search, browsing, and recommendation of documents. It can be formulated as a multilabel classification problem. We propose a novel deep learning-based method for this problem and design an attention-based neural network with semantic-based regularization, which can mimic users' reading and annotation behavior to formulate better document representation, leveraging the semantic relations among labels. The network separately models the title and the content of each document and injects an explicit, title-guided attention mechanism into each sentence. To exploit the correlation among labels, we propose two semantic-based loss regularizers, i.e., similarity and subsumption, which enforce the output of the network to conform to label semantics. The model with the semantic-based loss regularizers is referred to as the joint multilabel attention network (JMAN). We conducted a comprehensive evaluation study and compared JMAN to the state-of-the-art baseline models, using four large, real-world social media data sets. In terms of F1 , JMAN significantly outperformed bidirectional gated recurrent unit (Bi-GRU) relatively by around 12.8%-78.6% and the hierarchical attention network (HAN) by around 3.9%-23.8%. The JMAN model demonstrates advantages in convergence and training speed. Further improvement of performance was observed against latent Dirichlet allocation (LDA) and support vector machine (SVM). When applying the semantic-based loss regularizers, the performance of HAN and Bi-GRU in terms of F1 was also boosted. It is also found that dynamic update of the label semantic matrices (JMANd) has the potential to further improve the performance of JMAN but at the cost of substantial memory and warrants further study.

Entities:  

Year:  2021        PMID: 32584774     DOI: 10.1109/TNNLS.2020.3002798

Source DB:  PubMed          Journal:  IEEE Trans Neural Netw Learn Syst        ISSN: 2162-237X            Impact factor:   10.451


  2 in total

1.  Multi-label classification for biomedical literature: an overview of the BioCreative VII LitCovid Track for COVID-19 literature topic annotations.

Authors:  Qingyu Chen; Alexis Allot; Robert Leaman; Rezarta Islamaj; Jingcheng Du; Li Fang; Kai Wang; Shuo Xu; Yuefu Zhang; Parsa Bagherzadeh; Sabine Bergler; Aakash Bhatnagar; Nidhir Bhavsar; Yung-Chun Chang; Sheng-Jie Lin; Wentai Tang; Hongtong Zhang; Ilija Tavchioski; Senja Pollak; Shubo Tian; Jinfeng Zhang; Yulia Otmakhova; Antonio Jimeno Yepes; Hang Dong; Honghan Wu; Richard Dufour; Yanis Labrak; Niladri Chatterjee; Kushagri Tandon; Fréjus A A Laleye; Loïc Rakotoson; Emmanuele Chersoni; Jinghang Gu; Annemarie Friedrich; Subhash Chandra Pujari; Mariia Chizhikova; Naveen Sivadasan; Saipradeep Vg; Zhiyong Lu
Journal:  Database (Oxford)       Date:  2022-08-31       Impact factor: 4.462

2.  JLAN: medical code prediction via joint learning attention networks and denoising mechanism.

Authors:  Xingwang Li; Yijia Zhang; Faiz Ul Islam; Deshi Dong; Hao Wei; Mingyu Lu
Journal:  BMC Bioinformatics       Date:  2021-12-13       Impact factor: 3.169

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.