| Literature DB >> 35627429 |
Khishigsuren Davagdorj1, Ling Wang2, Meijing Li3, Van-Huy Pham4, Keun Ho Ryu4,5, Nipon Theera-Umpon5,6.
Abstract
The increasing expansion of biomedical documents has increased the number of natural language textual resources related to the current applications. Meanwhile, there has been a great interest in extracting useful information from meaningful coherent groupings of textual content documents in the last decade. However, it is challenging to discover informative representations and define relevant articles from the rapidly growing biomedical literature due to the unsupervised nature of document clustering. Moreover, empirical investigations demonstrated that traditional text clustering methods produce unsatisfactory results in terms of non-contextualized vector space representations because that neglect the semantic relationship between biomedical texts. Recently, pre-trained language models have emerged as successful in a wide range of natural language processing applications. In this paper, we propose the Gaussian Mixture Model-based efficient clustering framework that incorporates substantially pre-trained (Bidirectional Encoder Representations from Transformers for Biomedical Text Mining) BioBERT domain-specific language representations to enhance the clustering accuracy. Our proposed framework consists of main three phases. First, classic text pre-processing techniques are used biomedical document data, which crawled from the PubMed repository. Second, representative vectors are extracted from a pre-trained BioBERT language model for biomedical text mining. Third, we employ the Gaussian Mixture Model as a clustering algorithm, which allows us to assign labels for each biomedical document. In order to prove the efficiency of our proposed model, we conducted a comprehensive experimental analysis utilizing several clustering algorithms while combining diverse embedding techniques. Consequently, the experimental results show that the proposed model outperforms the benchmark models by reaching performance measures of Fowlkes mallows score, silhouette coefficient, adjusted rand index, Davies-Bouldin score of 0.7817, 0.3765, 0.4478, 1.6849, respectively. We expect the outcomes of this study will assist domain specialists in comprehending thematically cohesive documents in the healthcare field.Entities:
Keywords: document clustering; natural language processing; pre-trained language representation model
Mesh:
Year: 2022 PMID: 35627429 PMCID: PMC9141535 DOI: 10.3390/ijerph19105893
Source DB: PubMed Journal: Int J Environ Res Public Health ISSN: 1660-4601 Impact factor: 4.614
Figure 1Contextualized bidirectional encoder representations from transformers-based clustering framework for biomedical documents.
List of pre-trained corpora for BioBERT.
| Corpus | Number of Words | Domain |
|---|---|---|
| BooksCorpus | 2.5 | General |
| English Wikipedia | 0.8 | General |
| PubMed abstracts | 4.5 | Biomedical |
| PMC full-text articles | 13.5 | Biomedical |
Figure 2General architecture of BioBERT model.
Figure 3The architecture of the Word2Vec models: CBOW and skip gram.
Figure 4The architecture of principal component analysis.
Figure 5The architecture of the auto-encoder model.
The vector representation models and their characteristics.
| Name | Unit | Level |
|---|---|---|
| BioBERT | Contextual string embedding | Sentences |
| Word2Vec | Words | Local context |
| TF–IDF | Words | Corpus |
| GloVe | Words | Corpus |
| BioWordVec | Contextual sub-word | Local context |
Evaluation results of Gaussian-mixture-model-based clustering for biomedical documents.
| Representations | Fowlkes–Mallows Score | Silhouette Coefficient | Adjusted Rand Index | Davies–Bouldin Score |
|---|---|---|---|---|
| BioBERT | 0.7817 | 0.3765 | 0.4478 | 1.6849 |
| Word2Vec | 0.5919 | 0.3162 | 0.1143 | 3.1435 |
| GloVe | 0.6994 | 0.2175 | 0.3375 | 2.2419 |
| TF–IDF with PCA | 0.5308 | 0.0969 | 0.0863 | 4.5658 |
| TF–IDF with AE | 0.5659 | 0.0493 | 0.0751 | 3.7854 |
| BioWordVec | 0.7621 | 0.3854 | 0.4095 | 1.7309 |
Evaluation results of k-means clustering for biomedical documents.
| Representations | Fowlkes–Mallows Score | Silhouette Coefficient | Adjusted Rand Index | Davies–Bouldin Score |
|---|---|---|---|---|
| BioBERT | 0.7712 | 0.3041 | 0.4369 | 1.8507 |
| Word2Vec | 0.5794 | 0.2395 | 0.1025 | 2.7911 |
| GloVe | 0.5929 | 0.2658 | 0.2904 | 2.8612 |
| TF–IDF with PCA | 0.4672 | 0.0623 | 0.0719 | 3.8127 |
| TF–IDF with AE | 0.5531 | 0.0867 | 0.2758 | 3.4395 |
| BioWordVec | 0.7283 | 0.2624 | 0.4294 | 1.9204 |
Evaluation results of expectation–maximization clustering for biomedical documents.
| Representations | Fowlkes–Mallows Score | Silhouette Coefficient | Adjusted Rand Index | Davies–Bouldin Score |
|---|---|---|---|---|
| BioBERT | 0.6798 | 0.2762 | 0.3258 | 2.2121 |
| Word2Vec | 0.5339 | 0.2112 | 0.0875 | 4.8135 |
| Glove | 0.5573 | 0.2355 | 0.2216 | 3.8381 |
| TF–IDF with PCA | 0.4545 | 0.0752 | 0.0626 | 5.7447 |
| TF–IDF with AE | 0.5102 | 0.1118 | 0.0684 | 4.8371 |
| BioWordVec | 0.6356 | 0.2995 | 0.3055 | 1.9482 |
Figure 6Comparison charts of the biomedical document clustering models based on different evaluation metrics: (a) Fowlkes–Mallows score, (b) silhouette coefficient, (c) adjusted Rand index, (d) Davies–Bouldin score.
Figure 7Word cloud representations for the biomedical documents clusters.