Literature DB >> 35239700

Hierarchical lifelong topic modeling using rules extracted from network communities.

Muhammad Taimoor Khan1, Nouman Azam1, Shehzad Khalid2, Furqan Aziz3.   

Abstract

Topic models extract latent concepts from texts in the form of topics. Lifelong topic models extend topic models by learning topics continuously based on accumulated knowledge from the past which is updated continuously as new information becomes available. Hierarchical topic modeling extends topic modeling by extracting topics and organizing them into a hierarchical structure. In this study, we combine the two and introduce hierarchical lifelong topic models. Hierarchical lifelong topic models not only allow to examine the topics at different levels of granularity but also allows to continuously adjust the granularity of the topics as more information becomes available. A fundamental issue in hierarchical lifelong topic modeling is the extraction of rules that are used to preserve the hierarchical structural information among the rules and will continuously update based on new information. To address this issue, we introduce a network communities based rule mining approach for hierarchical lifelong topic models (NHLTM). The proposed approach extracts hierarchical structural information among the rules by representing textual documents as graphs and analyzing the underlying communities in the graph. Experimental results indicate improvement of the hierarchical topic structures in terms of topic coherence that increases from general to specific topics.

Entities:  

Year:  2022        PMID: 35239700      PMCID: PMC8893656          DOI: 10.1371/journal.pone.0264481

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

Topic models process a collection of documents to extract hidden thematic structures called topics [1, 2]. Each topic is an ordered lists of contextually correlated words representing a concept. Topic modeling has many extensions based on the learning paradigm such as supervised, semi-supervised, transfer learning, hybrid and knowledge-based models. These extensions are designed to meet the specific demands and needs found in different application areas [3-5]. In this paper, we examine a new type of topic model called hierarchical lifelong topic models. Lifelong topic models (LTM) is an important extension to topic models where topics are continuously learned and refined based on previously accumulated knowledge and processing of new information [3, 6]. Unlike other extensions, it does not require external guidance to analyze a dataset or to mine rules from it. More specifically, it exploits the huge volume of data processed in the past and the overlap in their concepts to identify global patterns [7]. Such an approach is highly desired in an unsupervised analysis of continuously arriving data from diverse sources [8]. Hierarchical topic modeling is another exciting extension which organizes the topics in a hierarchical structure thereby allowing the topics to be viewed at different levels of granularity. The topics at higher levels are abstract and get more specific down the hierarchy [9]. The incorporation of hierarchical topics in lifelong topic models will allow for examining the topics at different levels of granularity and adjusting the granularity as new information becomes available. This may be useful in applications such as summarization of news stories at different levels based on continuous streaming of news items [10]. A fundamental issue for incorporating hierarchical topics in lifelong topic models is the efficient and effective extraction of rules that are used to construct topics. Existing approaches are not suitable for this task. Although, the rule extraction approaches in LTM models facilitate the production of compact and coherent topics. They are, however, not well suited to arrange the topics in a hierarchy [8, 11–13]. In particular, they provide little or no information regarding the structural relationships that may occur among the topics [14, 15]. On the other hand, the rule extraction approaches in hierarchical topic models do not benefit from the past experience and rely on some level of supervision or manual guidance. In addition to this, whenever new information is available, the same support is required. To overcome these issues, we present an approach called network communities based rule mining for hierarchical lifelong topic models or NHLTM. The proposed NHLTM approach extract rules that provides the hierarchical structure among the rules and update the hierarchical relationships among the rules whenever more information becomes available. In particular, rules are extracted by constructing a graph based on the data in textual documents and then communities of related words are detected within the graph using spectral clustering. Experimental results on datasets of Chen 2014 electronics having 50 categories, Chen 2014 non-electronic having 50 categories and Reuters 21,578 having 90 categories suggest topic hierarchies with improved topics interpretability [16, 17]. The hierarchical structure of the topics is rationalized by gradual improvement of topic coherence across the topics at different levels. The proposed approach shows an improvement in topic coherence as we move down the hierarchy. The empirical log-likelihood of the held-out test indicates stable behavior of the model for predicting the unseen data at varying levels and number of topics in the hierarchy.

Topic models, lifelong and hierarchical topic models

Topic models

A topic consist of contextually related words that collectively represents a concept [18, 19]. An inference technique is typically used that iteratively determines the distribution of words across different topics based on words relationships in documents [20]. Words that co-occur across many documents have high chance of being selected under the same topic [21]. Topic models are used in different applications to explore, categorize, summarize and analyze textual data [22, 23]. Topic models have been extended in variety of ways and lead to models such as supervised models, hybrid models, transfer learning models, semi-supervised models, knowledge-based models and lifelong learning models [19, 24, 25]. In supervised models, the documents in training data are tagged with manually provided set of topic labels [26, 27]. Semi-supervised topic models incorporate expert guidance in associating certain words to topics while the other words are grouped around them in their respective topics [7]. Hybrid topic models are trained on certain features using a small labeled data and is provided for unsupervised analysis on remaining data. It is used in case of limited availability of labeled training data. The supervised and hybrid topic models help to produce interpretable topics, however, its use is only limited to the datasets for which such labeled training data exists. Transfer learning based topic models are trained on one dataset and applied on another having limited or noisy documents [28, 29]. Knowledge-based topic models are manually provided with set of rules that constraint the model to observe them while extracting topics [11, 30, 31]. The inference technique i.e., Gibbs sampling is modified in knowledge-based topic models where it has to incorporate the rules impact as well, as external guidance, in deciding the associations among words for a topic. Lifelong learning approach with topic modeling can be considered as the knowledge-based topic modeling where the rule mining mechanism is automated [32]. Hierarchical topic modeling uses some level of supervision or external guidance to combine the topics of a dataset into hierarchical structure [33]. The high level topics that are closer to root are general topics while the topics away from root are specialized topics. The literature of lifelong topic modeling and hierarchical topic modeling is presented below and their shortcomings are highlighted.

Lifelong topic modeling

Lifelong topic models continue to mine new rules from the dataset of each task, while it benefits from the relevant rules that were mined in the previous tasks [32, 34]. Considering the diversity of the datasets processed in the past, only the rules that are relevant to the current task can be of help [3, 6, 15, 35]. For this purpose, the background or context of a rule is compared with the vocabulary of the current dataset. It may be considered as resolving polysemy at the rule level. Although the quality of automatically mined rules may not be as good as that of a human expert, however, it is compensated by the quantity of rules and the freedom that it can be applied to data of any nature [12, 36, 37]. Generally speaking, the lifelong learning cycle contains four main components [38]. The first component deals with mining rules from a task completed. The rules are first evaluated based on some evaluation criteria reflecting their quality. Important rules are mined based on suitable thresholds on the evaluation criteria. The rules may represent semantic or contextual correlations among words [31, 39]. The approaches used for evaluating the quality of rules in literature are Normalized point-wise mutual information, frequent itemset mining, multi-support frequent itemset mining etc [8, 12, 13]. This components executes after performing each task of extracting topics from a dataset. The second component deals with representing the rules in a format where they can be efficiently and effectively retrieved when required [31]. In literature, the rules are represented as word pairs with positive correlation and strength of correlation, called must-link rules. For example, the must-link rule mustLink(word1, word2, +, 0.67) indicates a positive correlation between the words word1 and word2, indicated by the + sign. This rule adds bias into the gibbs sampling based inference to increase the probability of word2 for topics where word1 has higher probability and vice versa with an impact of 0.67 [32]. Similarly, the rules with word pairs having negative correlation with a strength of negative correlation is called cannot-link rules. For example, the cannot-link rule cannotLink(word1, word3, −, 0.86) indicates a negative correlation between the words word1 and word3, shown by the − sign. The rule adds a bias into the gibbs sampling based inference to decrease the probability of word3 for topics where word1 has high probability and vice versa with an impact of 0.86 [8, 12, 34]. Thus, in the modified gibbs sampling the intuition of arranging words into topics comes from the current dataset and relevant rules of the past [34]. The third component deals with the retention and maintenance of rules to preserve them for future [32]. The rules are maintained with different levels of abstraction, providing a tradeoff between storage space and efficiency [8, 12, 34, 36]. Lifelong learning can also be considered as a longer chain of transfer learning with unknown target datasets, multiple source datasets and that can be processed in any sequence. Therefore, it demands perseverance of rules for longer duration. For example, the rules learnt from 2 task may be utilized by the 10 or 50 task. On the other hand, the rules mined automatically are expected to have wrong, irrelevant, duplicate and even contradictory rules. Therefore, the retention and maintenance module undergo a refining process to resolve these issues and keep only consistent and good quality rules. The fourth component deals with the transfer of rules relevant to the current task and benefiting from it by incorporating the bias from the rules [11]. The gibbs sampling inference technique is extended with the Generalized polya urn (GPU) model [12]. The GPU model and its other variants are responsible for introducing automatic external guidance for different types of rules into the inference of topic models [30, 32, 34, 36, 40]. In recent times, word embeddings have improved various natural language processing approaches and is successfully used with topic models [41, 42]. Large text corpora have issues of long tailed vocabularies that result in losing the context. Using topic models on word embeddings helps better grouping of words into topics as topic models also attempt to group words based on their context [43]. Topic embeddings are also used to represent topics as dense vectors and help evaluate the correlation among topics [41]. Embeddings are not introduced in the current LTM approaches and may prove to be very effective in preserving context of words within topics and topics within a dataset.

Hierarchical topic modeling

Hierarchies are crucial in displaying the various components of a system in a tree-like structure. It has generic concepts at higher levels, while specific concepts are at lower levels [33]. Topics that are structured in a hierarchy can be analyzed at different levels of granularity and compared to one another [44]. There are some studies on hierarchical topic modeling with traditional topic modeling approaches [9, 28, 33]. We refer the approaches that are not using lifelong learning approach as the traditional approaches. They perform hierarchical topic modeling for a dataset in isolation. The hLDA model is the earliest attempt in organizing topics into a hierarchical structure [14, 45]. It make use of the nested chinese restaurant process (nCRP) to set a prior on possible hierarchical trees using the topic prior β. The gibbs sampling technique is also modified to predict new topic for the sampled word in the given document while it also predict the level of topics associated with the same document. The SSHLDA model organize the topics of a dataset into a hierarchical structure using labeled training data [9]. Transfer learning based topic models are used to transfer the hierarchical structure of one dataset into another related dataset [28]. It may be noted that hierarchical document topic modeling is also a closely related area where the topic hierarchies are extracted by representing the textual documents in the form of a graph. There are many useful approaches based on hierarchical document topic modeling including [46-48]. The essential assumption in these approaches is that the explicit linkage among documents in order to arrange them into a hierarchical structure [49, 50]. Although this assumption is useful in many cases, however, in our study, we do not make any such assumptions. We only rely on words and their co-occurrence in order to represent the given textual data in the form of a graph. The words co-occurrence is generally vague due to large heavy-tailed vocabularies but useful patterns can be identified when combined over a number of datasets. The notable attempts for extracting topic hierarchies in literature either require external support or hyper-parameters for the hierarchical structure which limits their application to certain scenarios only. Moreover, they perform hierarchical topic modeling in isolation and would require the same support each time. On the other hand, lifelong topic models benefit from the topic modeling performed on the previous datasets to improve the coherence of topics for the relevant future datasets. But the existing lifelong topic modeling approaches make use of statistical measures to mine rules. These approaches mine, maintain and transfer the rules in isolation, which makes them inconclusive towards generating topic hierarchies [8, 12, 13]. Therefore the existing lifelong topic modeling approaches are not well suited for finding the associations among topics [51, 52]. In particular, they lack in conveying structural information which is a key constituent for extracting topic hierarchies [52, 53]. In the next section, we demonstrate this limitation of the existing rule mining approaches and indicate towards a possible solution.

A limitation in existing rule mining approaches for topic modeling

In this section, we highlight a limitation in rule mining approaches that are used in existing lifelong topic modeling approaches. Consider Table 1 which represents a sample dataset. The rows of the table correspond to documents and the columns correspond to words. A “-” in a cell means that the corresponding word appears in the respective document. For instance, RAM is present in document D1, D2, D5, D6, D7, D8 and D9. We are interested in co-occurrences of words across the documents. Table 2 is constructed for this purpose based on data in Table 1. The rows and columns of Table 2 correspond to the words. A particular entry in the table reflects the co-occurrence between a pair of words. For instance, a value of 6 in the second column corresponding to the first row means that the words RAM and Cache has co-occurred in 6 documents. This can be easily seen in Table 1, i.e., RAM and Cache both occurred together in Documents D1, D2, D5, D6, D8 and D9.
Table 1

Sample dataset with 10 documents containing 9 words.

RAM Cache Intel Core Price Grapahics Video keyboard Cost
D 1 ----
D 2 -------
D 3 ----
D 4 ------
D 5 ------
D 6 ----
D 7 --------
D 8 ----
D 9 -----
D 10 ---

A dash (-) indicates the presence of a word in a document.

Table 2

Term-term matrix based on the data in Table 1.

RAM Cache Intel Core Price Graphics Video Keyboard Cost
RAM -64434433
Cache -4423333
Intel -633333
Core -33333
Price -1125
Graphics -531
Video -31
Keyboard -3
Cost -
A dash (-) indicates the presence of a word in a document. The typical lifelong topic modeling approaches consider rules in the form of word pairs. Among the rules, only those are selected whose evaluation is at or above a certain threshold value. Lets assume that the words co-occurrence is used as an evaluation measure with a threshold being set to 5. This will result in selection of four rules, namely, r1(RAM, Cache), r2(Intel, Core), r3(Price, Cost) and r4(Graphics, Video). Since these rules have a co-occurrence frequency of 5 or more. We may note that the rule are evaluated in isolation, irrespective of their associations with other rules. More specifically, it ignores the relationships among different rules. For example, the rules r1 and r2 have words having very strong association such as RAM and Intel having a co-occurrence value of 4. This relationship is however not being considered in the existing approaches. The issue of mining rules in isolation can be intuitively addressed by considering the association among the rules in the form of co-occurrence of the words belonging to different rules. We define the associations based on co-occurrence information between words belonging to different rules as, where w and w′ are distinct words belonging to different rules r and r, respectively. The association between two rules is the summation of all word co-occurrences across the two rules. For instance, consider the rules r1 and r2, the association between these two rules according to Eq (1) is, The association among rules are calculated for all the other rules as well and are given in Table 3. These associations can be used to find which rules are more closely related to each other as compared to others. Considering all the rules as leaf nodes of a tree, the rules r1 and r2 have the highest association of 4 according to Table 3. They are linked at the next higher level to form a more general rule. We call this the rule r1,2, which can also be considered as the parent rule for the two rules r1 and r2. Now at the next higher level, there are three rules i.e., r1,2, r3 and r4. Following the same, the association among these rules is evaluated where the rules r1,2 and r4 have the next highest association with an association score of 3.25. The association score of rules r1,2 and r3 is 3 and that of rules r3 and r4 is only 1. Thus, the rule r1,2,4 is the next level parent rule that has the rules r1,2 and r4 as its descendants. Finally, the two available rules i.e., r1,2,4 and r3 having association score of 2.3 are linked to form a single rule of r1,2,3,4. This is one way of finding the structural association among rules.
Table 3

Values of association among the rules based on Eq 1.

r 1 r 2 r 3 r 4
r 1 433.5
r 2 33
r 3 1
r 4
Exploiting the structural association among rules, a hierarchical structure may be generated for the topics of a given dataset. Different approaches are available for finding the association among rules that could be translated into improving the arrangement of topics within a dataset and that of words within a topic. For instance, consider topics T1, T2, … T. The topic T1 is containing the words of the rule r1 and T2 is containing the words of the rule r2. Since the rules r1 and r2 have the closest association, therefore, topics T1 and T2 are merged at the higher level to represent a single topic T1,2. The structure between the topics can be successively generated by combining the topics containing rules with highest associations. This structure will lead to hierarchy with the dataset as root node and the specific topics as leaf nodes while the intermediate nodes representing general topics at different levels. The model has incomplete knowledge, and therefore some topics may not be represented by rules. Such topics link to the root of the hierarchy as outliers. The number of such topics depends on the number of relevant rules in the knowledge base. They are expected to decrease as the number of rules and their diversity increases with experience. It may be noted that the existing studies generally consider rules in the form of word pairs only. We consider rules in the form of word pairs and in addition to that we also consider rules that may contain any number of words. A critical issue in the above intuitive solution is how to determine and model the association of rules based on sound theoretical notions. The exploration of this issue will open up new ways for exploiting the relationships among the rules. Graph based community detection is widely used for exploratory data analysis in many areas [54]. Representing the data as graph and performing the graph Laplacian enhances the clustering properties in the data. Such properties are generally not found in conventional clustering algorithms [55, 56]. For any type of data represented as a graph, community detection techniques can help in detecting the communities and their hierarchical structures from the data [57, 58]. These approaches have been used successfully for research problems in the fields of psychology, sociology, biology, statistics and computer science [57, 59, 60]. Community detection has been used to identify clusters of similar articles using a citation network [61]. In this paper, we explore the use of community detection technique to mine structurally associated rules.

Network communities based hierarchical lifelong topic modeling (NHLTM) approach

In this section, we present our proposed NHLTM approach. The lifelong learning aspect allows the model to preserve rules of diverse nature over experience. The proposed community detection approach for mining rules facilitate in improving coherence of topics and arrange them into a hierarchical structure.

Architecture of NHLTM approach

The Architecture that realizes the NHLTM approach is presented in Fig 1. A dataset is provided to this architecture as input. Hierarchically structured topics are produced from the given dataset as output. The rules mined from the datasets processed in the past contribute towards organizing words into topics and topics into a hierarchical structure. Our contribution in this research is to introduce a community detection based approach named NHLTM to mine structurally associated rules. The architecture initiates the rule mining module whenever a new dataset is made available. The other modules i.e., rule representation and rule transfer are also updated based on new datasets which helps in efficient and effective utilization of the newly introduced community based rules. The knowledge base retains all the rules for long term future use while the maintenance module ensures the refinement of rules based on new data.
Fig 1

Architecture of NHLTM approach.

Community detection based rule mining

In order to perform community detection, the following four step process is generally used [56, 62]. The first step is to represent the given data as a graph. The next step is to recursively split communities into smaller and smaller communities until the stopping criteria is met. Generally, the communities at leaf nodes have strong correlation and represent contextually correlated words. There may be few exceptions with leaf node communities representing noise words in the dataset. The third step is to evaluate communities and discard those that do not have enough compactness among words. The last step is to only retain those communities that have an acceptable level of compactness as correlation among its words. Our proposed NHLTM approach uses the aforementioned steps of community detection for mining rules. The details of NHLTM approach is given in Algorithm 1. It takes the current dataset D and a graph G corresponding to all previously processed datasets D1 to D as an input. The output of the algorithm are the updated topic hierarchies and the updated graph G. The algorithm first extract the relevant knowledge rules from the graph G of previously processed datasets. The while loop in the algorithm corresponds to the mechanism of transferring the impact of rules into LTM which relates to the final step of the NHLTM approach. In line 10, the structural similarity among rules are evaluated based on similarity in the assigned hexadecimal codes to organize them into a hierarchical structure. Lines 11—14 are related to the lifelong learning components of the NHLTM approach. We now elaborate the community detection process for rule mining employed in the algorithm in some detail. Algorithm 1 Proposed NHLTM Approach Input: Dataset D, G of previously processed datasets from D1 to D Output: Topic Hierarchy for D, updated Graph G 1: Get relevant rules from prior knowledge rules 2: LTM random initialization 3: while Words are not settled in their respective topics (Inference) do 4:  Predict the most appropriate topic t for the sampled word w according to the current state 5:  Update the word-topic and topic-document distributions accordingly 6:  if w in rules then 7:   Increase the probability of all words w′ in topic t that are in rule r ∈ rules with the word w 8:  end if 9: end while 10: Extract Topic Hierarchy of Dataset D, using the structural association codes of rules 11: Add words as nodes and their co-occurrence as edges from D to Graph G of D1 to D 12: Extract communities using spectral clustering 13: Filter communities with weak intra-community associations 14: Represented the communities as rules 15: Generate new rules i.e., and add to the knowledge base 16: return Topic Hierarchies, Graph G Step 1: Representing the data as graph The algorithm 1 constructs a graph G from the textual data by considering words as nodes and their co-occurrence as edges. In particular, the textual data corresponding to datasets D1 to D are used for this purpose. It is an undirected graph as the edges represent the words co-occurrence that can be interpreted both ways. The edges are weighted highlighting the co-occurrence frequency of the two words forming an edge. Considering the power law distribution in the graph where few words have very high occurrence frequency as compared to others and therefore, are more frequent in stable co-occurrence associations. Such graphs can be categorized as scale-free networks that require the weaker filtering the weaker associations. Network based representation has many useful applications in text analytics including keywords extraction, document categorization and knowledge graphs [63]. In this paper, the network of words across all datasets represent a hierarchical knowledge graph that assists topic modeling improving topic coherence and generating hierarchical structure of topics within a dataset. The weaker nodes and edges are filtered using minimum node degree and minimum edge weight thresholds respectively. The advantage of dropping weak edges having low weights and weak nodes having low degrees is to reduce the computational cost. Since the graph is expected to grow in size, only reliable rules are extracted. The intuition for clipping weak branches is that they are not going to make strong rules due to their weak associations. They usually represent noise or rare vocabulary that is not frequently used by other users [56, 64]. Step 2: Identifying and detecting word communities Communities can be extracted using different community detection algorithms [54, 55]. For extracting communities, we use the Fiedler eigenvector. It represents the nodes with positive and negative values, indicating a possible split as the nodes of positive values may form one sub-community while the others may form another sub-community. It is obtained by representing the graph G as a normalized Laplacian matrix and decomposing it into eigenvectors [65]. The conductance cut is next used to evaluate if it is feasible to proceed with splitting the graph into communities and sub-communities recursively. The evaluation is based on the cost associated with inter-community associations. It is worth mentioning that there are many community detection algorithms that can be used for extracting communities however, we used spectral clustering for this purpose. The edges within a community help in identifying useful and important rules while the edges between communities help in identifying the hierarchical association among the rules. This approach continuously updates the communities based on new information which results in updated hierarchical information among the rules. Spectral clustering performs better for dense graphs and has reasonable results in various text processing tasks. The graph of multiple datasets with pruning weak nodes and their edges ensures that the graph is dense, thus making spectral clustering more favorable. Spectral clustering algorithms are based on computing eigenvalues that allow considering different levels of information and have dimensionality reduction feature which is a typical issue in textual processing tasks. All this makes spectral clustering a more favorable approach for our study. Step 3: Evaluating compactness of communities The communities extracted in step 2 are next evaluated based on the associations among the nodes contained in the community. The measure of weighted connectivity index or WGCI is used for this purpose [55]. It measures the intra-community associations in relation to the inter-community relations. Nodes with higher intra-community associations have higher WGCI score as the intra-community edge weights increases proportional to the degree of nodes. In contrast, the nodes with high inter-community associations have low Weighted graph connectivity index WGCI as the intra-community edges will have low weights and their respective nodes have high degrees. The WGCI of a community comm can be calculated as, Where e is the edge between the words w and w′ belonging to the community comm. In addition to WGCI, we also consider the measure of Von Neumann entropy for evaluating communities [66] as, where N is the total number of nodes in the community and represents their eigenvalues of the normalized adjacency matrix corresponding to the second lowest value. By putting suitable thresholds on the two measures, we selected communities that have considerable associations among its nodes. The threshold values are experimentally chosen to keep only those communities that have higher intra-community associations [67]. Step 4: Representing communities as rules The communities selected are abstracted and represented as rules. The objective is to facilitate efficient storage, retrieval and transfer of rules into a task [68]. The words in a community are retained as the rule words which also helps in setting up the context for the rules. The structural information of a community is retained in the rule by preserving the order of splits that lead to the generation of each community. The impact of a rule is calculated using Lins Approximation [55]. The value of Lins Approximation is aggregated across all edges of the community to give a measure of entropy for the community. Step 5: Transferring the impact of rules into Gibbs sampling This step deals with transferring the impact of relevant rules into the Gibbs sampling inference of topic distribution for the current task. The default inference technique continuously update the document-topic and topic-word distributions. It starts from a random state and converges on the suitable distribution in an iterative way [18]. The distributions in a particular iteration refer to the state of the model, where new state is determined based on the previous state following markov chain [69]. The probability of a suitable topic for the sample word is [70], where a sampled word w from document d has its probability updated for topic t as θ times ϕ. The θ and ϕ are given by, and respectively. In the Eq (6), the document-topic distribution i.e., θ is updated using as the probability of topic t in document d based on the current state of the model. It is computed by excluding the sampled occurrence of the word, indicated by −i. α is the smoothing factor. The denominator normalizes the value using the sum of probabilities of all the other topics from k = 1, …, T for the same document d. Similarly in Eq (7), the topic-word distribution i.e., ϕ is updated using which is the probability of word w in topic t based on the current state of the model. It is again computed by excluding the sampled occurrence of the word w, indicated by −i. Moreover, β is used as a smoothing factor. The denominator normalizes the value using the sum of probabilities of all the words from u = 1, …, V for the same topic t. In order to select relevant rules for a task, the overlap of a rule is measured with the vocabulary of the task and are compared against a threshold. It is given by [12] as, where ξ is the threshold for overlap between rule words and the vocabulary of the current task. It helps to resolve the context and address the word sense disambiguation for a rule. The bias from the rules effects the topic-word distribution i.e., ϕ only. When the probability of a word w is increased for a topic t and w is part of a rule w ∈ r, then the other words of the same rule w′ ∈ r also have their probabilities increased for the same topic t, using equation given by [12] as, where the probabilities of all other words w′ ∈ r are updated for topic t by the ν factor of the rule when w ∈ r has its probability updated for t. However, in case w ∉ r, the default Eq (7) will be used to update ϕ.

Experimental results and discussion

In this section, we present detailed experimental results of the proposed NHLTM model for a sequence of tasks. The Chen 2014 dataset with 50 electronic product categories, the Chen 2014 dataset with 50 non-electronic product categories and Reuters R21578 with 90 categories are considered for this purpose [16, 17]. These datasets have documents in multiple categories assumed to be available in a sequence and are used for lifelong topic modeling in literature. In case of the first two datasets, i.e., Chen 2014 electronic and Chen 2014 non-electronic, there are more than 5000 in each category. Each document is a short product review that is collected from Amazon.com. The third dataset has skewed document representation with categories having documents in thousands, hundreds, and below that. Each category is considered a dataset for extracting topic hierarchies and is processed in a sequence. Therefore, it fits more into the context of mining rules continuously as an automatic process. Lifelong topic modeling cannot be applied to a single dataset as it requires multiple tasks to improve with experience. Table 4 shows the details of the graphs for each of the three datasets. Only 10% of the past data is considered for mining rules to reduce computational cost. The datasets are preprocessed to retain only meaningful words as verbs, adverbs, adjectives, and nouns. The words as nodes, their edges and corresponding weights are mentioned before and after graph filtering. Through aggressive filtering, the weak nodes and edges are removed as they are least likely to be part of compact communities. It results in increasing the average node degree and edge weight. The filtering mechanism refines the graph for efficient analysis. It drops the nodes and edges that have a low degree and edge weight, respectively. Thus we have a smaller but more densely connected network after the filtering phase.
Table 4

Graph properties before and after pruning weak nodes and edges from the dataset.

DatasetPropertyBefore FilteringAfter Filtering
Electronic Chen 2014Word Nodes5,5741,228
Co-occurrence Edges6,55,2613,46,687
Per Edge Weight3.434.84
Non-Electronic Chen 2014Word nodes7,3912,572
Co-occurrence Edges5,83,0204,24,066
Per Edge Weight2.192.49
Reuters R21578Word Nodes5,1932470
Co-occurrence Edges9,74,7936,73,193
Per Edge Weight1.621.82
In the second step, the graph is processed to identify and extract possible communities. The Chen 2014 electronic dataset has a total 148 communities extracted that are interrelated to each other through structural associations. Structures are preserved through iteratively splitting the communities into sub-communities. The communities that are separated in later iterations are considered closer as compared to the communities that are separated in the former iterations. The structural position of each community in the graph of all prior data is derived through subsequent splits, and is presented as hexadecimal codes as IDs in Table 5. The communities with a smaller difference of hexadecimal codes indicate a nearer common parent and a closer relationship and vice versa. These values can be integers or binary but we have presented them as hexadecimal values for compactness. Fig 2 shows the first five iterations of recursively splitting communities into sub-communities. The communities with grey background will continue with the process. The others with white background have reached their stopping criteria of either too few nodes or too expensive cut to continue with.
Table 5

Sample communities with their codes, top words and neighboring communities.

ID (Hexa Codes)Top WordsNeighboring Community ID
9Adaptor, cam, filter, stand, indicatorC
CException, weird, superior, break9
56Field, section, flat8A
8AStorage, airport, convenient56
21EEnvironment, rest, world, answer23D
23DFolder, label, credit21E
439Profile, background, scene, rock43B
43BTray, blueray, manual, gripe439
10A9Alarm, longer, beep43B
2146Police, ticket, state, ka2151
2151Dead, transmitter, awful2146
8528Hard, top, lap, scroll, pad2151
10A63Outstanding, playback, release, model10A70
10A70Gamer, thrive bigger, hub10A63
10B3EToshiba, lifetime, replacement, warranty10A70
21091Foot, pace, distance, run10B3E
79A0DHeavy, metal, cover, bottom21091
10C0C3Traffic, highway, city, road79A0D
214C93Bluetooth, reception, works, feedback214C90
214C90Bright, display, control214C93
Fig 2

Hierarchical breakdown of graph into communities till first 5 levels, nodes per community and their respective graph cut weight.

In the third step, the communities extracted are evaluated for their quality. Communities with weakly associated nodes cannot effectively communicate their impact and are therefore filtered out. Only 84 communities were left behind after the WGCI and entropy based filtering. The threshold values for WGCI and entropy were set to 1 and 0.2, respectively. These values are experimentally evaluated for the given dataset. Dropping the thresholds add to computational cost without showing reasonable improvement in results. Table 5 shows the structural information and the top few words for 20 sample rules. Table 6 shows other properties, i.e., size, entropy, Lins approximation, WGCI and average degree node across community for the corresponding rules in Table 5. These evaluation scores are used to retain only good quality communities with higher compactness within its nodes while filter others.
Table 6

Evaluation scores of the communities shown in Table 5.

Comm Sr.No.SizeEntropyLins Approx. WGCI Avg. Node Degree
172.7871.87.07356.893
2133.8163.2490.23612.153
351.1081.2490.0202
451.1301.2490.0404.6
561.4401.4990.0403.83
651.0551.2490.0302.4
782.0871.9990.0614.125
851.0851.2490.0546.6
961.4251.4990.867.83
1051.1351.2490.33619.2
1161.4491.4990.0363.33
12281.7451.7490.08336.85
1361.4881.4990.092310.83
1451.1091.2490.04143.8
1551.0711.2490.322748.6
1671.6981.7491.538103.285
1751.1291.2490.34852.8
1871.8171.7490.82952.85
1951.1411.2490.0626
2041.7461.4480.08575.4
In the fourth step, the communities are transformed and represented as rules. The hub node in the community is represented as rule head. The Lins Approximation value of each community is represented as the impact of its corresponding rule. It refers to the bias with which the inference technique is manipulated in favor of the rule by using generalized polya urn model. Step five is related to transferring the impact of rules into the new task. It can be observed that most of the rules have their sizes ranging from 4 to 7 nodes as the natural arrangement of words in the given data. It may vary for other datasets and therefore, the learning mechanism is kept flexible to adapt to it. The quality of the topics extracted is evaluated using topic coherence. It presents the association or correlation among the words of a topic [71]. Thus, the words of a topic that have high contextual correlation results in a higher topic coherence score for the topic. It is the dominant topic evaluation technique that has the highest relevance to human judgment. The topic coherence for a topic k can be calculated as, where T represents the k topic in the given corpus while w and w′ are two distinct words belonging to the topic T. It is the double sum of log of conditional probabilities across all the words in the topic. The top 30 words with highest probability in a topic are used to calculate topic coherence as in literature [71]. The topic coherence of a dataset is determined by averaging the topic coherence over all the topics in a given dataset. The learning curve of lifelong topic modeling is evaluated based on improvement in the quality of topics as topic coherence. Fig 3 shows the comparison of NHLTM with some of the existing approaches including lifelong topic modeling (LTM) [31], automatic must link cannot link (AMC) [34], automatic must link cannot link with must links only (AMC-M) [34], online learning based automatic must link cannot link (OAMC) [12] and the basic latent dirichlet allocation approach. All approaches are provided with the same experience in the form of 10% of the data to mine rules from. They are then presented with a new category as current data to extract interpretable topics by minimizing the impact of noise in the new categories with the help of the mined rules. The top 30 words in a topic are used to calculate the topic coherence of a topic. The NHLTM approach has resulted in consistently improving the topic coherence across all the datasets in comparison to others.
Fig 3

Comparison of NHLTM with existing approaches.

A hierarchy of topics is extracted for each dataset that shows the concepts discussed within the dataset at different granularity levels. The hierarchies are compared in terms of their topic coherence at different levels in the hierarchy as the number of topics increases. The empirical log-likelihood indicates generality of the models as its ability to estimate the unseen data. Topic coherence of the proposed NHLTM approach shows higher improvement down the hierarchy from level one to level six, in comparison to the HLDA approach. The topics at higher levels represent generic concepts and therefore attain lower topic coherence. The lower levels have specific concepts with compact representation. The given results indicate the compactness of topics produced by NHLTM at all levels of the hierarchy. The value of empirical log-likelihood should preferably increase or stabilize as the model makes better predictions of the unseen data. The empirical log-likelihood values produced by NHLTM model are maintained lower than HLDA and are also depicted in Fig 4.
Fig 4

Comparison of NHLTM and HLDA with increasing number of topics and levels of hierarchy for the Alarm clock dataset in large-scale data.

The topics of the Alarm Clock dataset are more scattered having more branches at the lower level in Fig 5. A possible reason is that the users discussing topics of Alarm Clock referred to the same topics using a variety of expressions. Discussing more topics within a document leads a closer relation among those topics. The topics Noise wake and Design flaw have very weak association with all the other topics and are therefore, directly linked to the root at the highest level. The root represents the domain of the dataset, which is Alarm Clock in this case. The topics Small bedside and Easy cheap are closely related. The other topics that are closely related, represent the operational topics as Light feature, Simple stuff and Digital number. The topics Radio instructions and Display control represent the soft features or controls of the product. These topics are linked to the previous group of topics as operations and controls are somewhat closely related. The topics Battery star and Plastic piece doesn’t have a very clear sense but loosely represent the exterior of the product. The next few topics also have weaker association. The non-electronic products like Cars, Boats and Food would have more diverse vocabulary as compared to electronic products and therefore, their topic hierarchies could be more easily comprehendible.
Fig 5

Topic hierarchy of Alarm Clock dataset.

Conclusion

Lifelong topic models enable the conventional topic model to learn topics continuously from the knowledge accumulated from the past which is updated regularly based on new information. On the other hand, hierarchical topic modeling extends topic modeling for grouping into a hierarchical structure. This study combines the two and proposes hierarchical lifelong topic models which allows the examination of topics at different levels of granularity and continuously adjust the granularity of the topics as more information is made available. The existing hierarchical topic models lack the lifelong learning mechanism and therefore, cannot generate topic hierarchies from the word associations only. On the other hand, the existing LTM approaches do not attempt to generate topic hierarchies. NHLTM has introduced spectral clustering and preserve the communities as rules along with their associations that contribute towards extracting topic hierarchies. A key issue in hierarchical lifelong topic modeling is to extract rules that will not only preserve the hierarchical information among the rules but will also regularly update them as information evolves. We introduced a network community based hierarchical lifelong topic models or NHLTM to address this issue. Experimental results on the datasets of Chen 2014 electronic, Chen 2014 non-electronic and R21578 Reuters indicate improvement of the hierarchical topic structures for maintaining high topic coherence that increases from general topics to specific topics. The proposed approach has opened up doors for further research. More specifically, one may extend the NHLTM approach to include topics with multiple parents and to consider polysemy at the rules level. Moreover, the effect of using embeddings in the rule mining component and its impact on the quality of extracted hierarchy may provide further useful insights. 12 Nov 2020 PONE-D-20-29449 Hierarchical Lifelong Topic Modeling Using Rules Extracted From Network Communities PLOS ONE Dear Dr. Khan, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Dec 27 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Diego Raphael Amancio Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. Thank you for stating the following in the Acknowledgments Section of your manuscript: "This work is partially supported by faculty research support fund of National University of Computer and Emerging Sciences (NUCES), Pakistan." 2.1. We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: [xxx] 2.2. Please provide an amended statement that declares *all* the funding or sources of support (whether external or internal to your organization) received during this study, as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now.  Please also include the statement “There was no additional external funding received for this study.” in your updated Funding Statement. Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 3.  We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For more information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. In your revised cover letter, please address the following prompts: a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially sensitive information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. We will update your Data Availability statement on your behalf to reflect the information you provide. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: No ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This paper presents a hierarchical lifelong topic model called NHLTM. This new model combines lifelong learning, with network community detection and topic modelling. Overall I found that the manuscript was well organized and the contributions were clear. There were several non-native grammar errors. I probably did not catch them all, but I did upload my notes to this review. Please look at my ink-notes for detailed comments. My broad remarks are as follows: 1. Reporting raw likelihood values is not normal and not meaningful. Please do not report -856, etc. as results. Just say that it was improved. 2. Section 1 and section 2 are otherwise well organized with only minor typographical and grammatical errors. 3. In section 4 you talk about treating the text corpora as a graph. There exists the hierarchical document topic model (HDTM) which does something similar here that you could look at. 4. The input of Alg 1, does it require preprocessing of the whole data each time? The input looks like it does. 5. In general, Section 4 is too verbose. You do not need to describe each line of Alg 1. Alg 2, Alg 3, and Alg 4 are very straightforward and can just be described in prose. Likewise Eq 5 and Eq 6 are well known functions. Just say normalized Laplacian and Fiedler Vector. You don't need to derive it. Any reader will know what those things are. Eq 7 is called "conductance cut" and is well known too - there is no need to describe these things. I think that Section 4 could be reduced to 3-4 pages total. 6. In section 5, you only perform experiments on a single dataset. This is not enough for conclusive study. There exist enormous datasets for study. You need to perform a much more thorough analysis with many datasets. 7. In Table 5, what is the hexadecimal data supposed to show? Why not just represent things as ints? What is Comm Sr. No. column? 8. Fig 2 us supposed to have some nodes with a grey background, but my printout does not have anything in grey. 9. In Section 5, where do you test the "lifelong" portion of NHLTM? 10. Whats the purpose of Table 6? And on page 23 generally, raw values are just not informative. 11. Fig 3 has the y-axis flipped. It goes from largest on the bottom to smallest on the top. This is quite abnormal. 12. Table 7 should be a figure. 13. Conclusions are not strong. What did we learn in this paper? What are the limitations of NHLTM? What is the future work. Expand this section please. Overall I think the technical contribution of this paper is ok, but the experiments are severely lacking. Significant additional work is needed before this paper can be accepted. Reviewer #2: While interesting, I believe this paper needs major revisions for improved clarity and a more succinct narrative. - What do you mean by rules? You should present a clear example and definition very early. In one case, it appears that a rule consists of a correlation of two words, but a subsequent equation seems to contradict this. - Am I right in thinking you used raw frequency? In "classic" approaches to topic modeling, raw frequency is rarely used. It's more common to see something like positive pointwise mutual information (PPMI). Were alternatives to raw frequency explored? If so, why wasn't this reported? - Moreover, typically, term term matrices have a zero diagonal. - I recommend that you avoid using "Extract communities using Algorithm 3" when you can instead reference your favored technique by name. - Your tables and figures all warrant real captions so that they can be digested on their own. - Word embeddings have revolutionized natural language processing. Recent approaches to topic modeling that have appeared in venues such as TACL and various ACL conferences have all involved word embeddings to some degree. Why is no discussion or comparison made here? While this might require a second evaluation on another dataset, I think it's necessary. - Speaking of recent approaches, how do neural topic models (ex. VAE-based) compare to your approach? For example, a comparison to Gupta et al's Neural Topic Modeling with Continual Lifelong Learning (ICML 2020) would be nice. - How does the approach work with different community detection algorithms? Why did you select this particular one without any comparison? - How are you evaluating the lifelong aspect of your algorithm? - AMC-M, AMC and OAMC - Are the abbreviated forms of these acronyms introduced anywhere? - Table 5: this should be reformatted to make it more obvious which communities are "close". - Please make your code available for reproducibility and future comparisons. Doing so is now standard practice. Ideally, I would like to see how you compared the different models. - Though not a serious problem, there are numerous typographical errors (repeated words, inconsistent formatting, agreement errors, missing determiners, etc.). ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. Submitted filename: 0658_001.pdf Click here for additional data file. 1 Jan 2021 We are thankful to the reviewers for taking time to review the paper and liking it. We tried to incorporate all the concerns of the reviewers and we hope that the revisions in the paper will meet the expectations of the reviewer. Please find detailed responses and the resulting changes in the manuscript in the ResponseNHLTMv4 document attached. Submitted filename: Response NHLTMv4.doc Click here for additional data file. 8 Jun 2021 PONE-D-20-29449R1 Hierarchical Lifelong Topic Modeling Using Rules Extracted From Network Communities PLOS ONE Dear Dr. Khan, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. ============================== The reviewer pointed out that major issues have been addressed. However, language should be checked in order to improve readability. ============================== Please submit your revised manuscript by Jul 23 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript: A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see:  http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols . Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at  https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols . We look forward to receiving your revised manuscript. Kind regards, Diego Raphael Amancio Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The author has addressed all of my comments. The new parts of the paper need a thorough proofread - there are minor grammatical issues, but nothing major. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 1 Jul 2021 Comments by Reviewer 1 The author has addressed all of my comments. The new parts of the paper need a thorough proofread - there are minor grammatical issues, but nothing major. Response: We are thankful to the reviewer for accepting the modifications we have made to the manuscript in response to the comments provided. We have fixed the grammatical issues as highlighted by the reviewer and have proof read the document multiple times. Below are some of the updates that we made with regards to the grammar in this version of the manuscript. Some examples of correct usage of the article. ‘….50 electronic product categories, Chen 2014 dataset …’ changed to ‘….50 electronic product categories, the Chen 2014 dataset …’ ‘Filtering out edges with smaller weights and nodes …’ changed to ‘The filtering out edges with smaller weights and nodes …’ ‘A higher value indicates better correlation among the words of a topic.’ changed to ‘A higher value indicates a better correlation among the words of a topic. ’ ‘All approaches are provided same experience in the … ’ changed to ‘All approaches are provided the same experience in the …’ ‘Top 30 words in a topic are used to calculate the topic coherence of a topic.’ changed to ‘The top 30 words in a topic are used to calculate the topic coherence of a topic.’ ‘The communities with smaller difference … ‘ changed to ‘The communities with a smaller difference …’ ‘ … in comparison to HLDA approach.’ Changed to ‘ … in comparison to the HLDA approach. ‘ … The given results indicate compactness …’ changed to ‘ … The given results indicate the compactness …’ Some Examples of adding comma at the right place for better readability ‘In this paper we examine a new type of topic model …’ changed to ‘In this paper, we examine a new type of topic model …’ ‘… meaningful words as verbs, adverbs, adjectives and nouns.’ changed to ‘… meaningful words as verbs, adverbs, adjectives, and nouns.’ ‘… case of the first two datasets i.e., Chen 2014 Electronic …’ changed to ‘…case of the first two datasets, i.e., Chen 2014 Electronic …’ ‘… having documents in thousands, hundreds and below that.’ changed to ‘… having documents in thousands, hundreds, and below that.’ ‘…is derived through subsequent splits and …’ changed to ‘…is derived through subsequent splits, and …’ ‘Table 6 shows other properties i.e., size, entropy …’ changed to ‘Table 6 shows other properties, i.e., size, entropy …’ ‘ … generic concepts and therefore, attain …’ changed to ‘ … generic concepts and therefore attain …’ Some Examples of preposition correction ‘… has the highest relevance with human judgment.’ changed to ‘… has the highest relevance to human judgment.’ ‘… impact of noise in the new categories with the help the mined rules.’ changed to ‘…impact of noise in the new categories with the help of the mined rules’ ‘ … All approaches are provided same experience in the …’ changed to ‘ … All approaches are provided with same experience in the …’ Some examples of correct singular/plural usage ‘… are directly linked to the root of the hierarchy as an outlier’ changed to ‘…topics link to the root of the hierarchy as outliers’ ‘…extracting topic hierarchies and is processed …’ changed to ‘…extracting topic hierarchies and are processed …’ ‘These type of classes are frequently encountered in …’ changed to ‘These types of classes are frequently encountered in …’ Some examples of correct form of verb ‘Lifelong topic models enables the conventional …’ changed to ‘Lifelong topic models enable the conventional …’ ‘This study combines the two and proposed a hierarchical …’ changed to ‘This study combines the two and proposes a hierarchical … ‘… word embeddings has improved various natural … ’ changed to ‘…word embeddings have improved various natural …’ ‘… topic models on word embeddings help better grouping… ’ changed to ‘… topic models on word embeddings helps better grouping…’ ‘… the communities based on new information which result in updated … ’ changed to ‘…the communities based on new information which results in updated …’ ‘… computing eigenvalues that allows considering …’ changed to ‘…computing eigenvalues that allow considering …’ ‘ … of information and has dimensionally reduction …’ changed to ‘ … of information and have dimensionally reduction …’ Some examples of improving the structure of sentence for easy to follow understanding. ‘Moreover, such support is required each time when new information is processed in the form of new dataset.’ changed to ‘Moreover, each time new information is available as a dataset, the same support is required.’ ‘Hierarchies are highly important in displaying the various components of a system in a tree like structure that has generic concepts at higher levels while specific concepts at lower levels [33].’ changed to ‘Hierarchies are crucial in displaying the various components of a system in a tree-like structure. It has generic concepts at higher levels, while specific concepts are at lower levels [33].’ ‘Filtering out edges with smaller weights and nodes with fewer edges, the graph is refined for efficient analysis. Thus we have a smaller but more densely connected network after the filtering phase.’ changed to ‘The filtering mechanism refines the graph for efficient analysis. It drops the nodes and edges that have a low degree and edge weight, respectively. Thus we have a smaller but more densely connected network after the filtering phase.’ ‘In this section, we present detailed experimental results of the proposed NHLTM model for extracting topic hierarchies for a sequence of tasks.’ changed to ‘In this section, we present detailed experimental results of the proposed NHLTM model for a sequence of tasks.’ ‘Aggressive filtering is applied to remove weak nodes and their edges as they cannot be part of compact communities which also results in the average edge weight.’ changed to ‘Through aggressive filtering, the weak nodes and edges are removed as they are least likely to be part of compact communities. It results in increasing the average node degree and edge weight.’ At the end, we would like to pay thanks again to the editor in chief, associate editor and the reviewers. Indeed, their constructive comments helped to improve the overall quality of the manuscript. Submitted filename: Response to Reviewers.doc Click here for additional data file. 30 Dec 2021
PONE-D-20-29449R2
Hierarchical Lifelong Topic Modeling Using Rules Extracted From Network Communities
PLOS ONE Dear Dr. Khan, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Feb 13 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Diego Raphael Amancio Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #3: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #3: (No Response) ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Grammatical changes appear to be satisfactory. I have no further comments to make regarding the revision of this paper. Reviewer #3: This paper proposes an interesting approach for topic modeling. More specifically, the authors put together concepts of NLP and network science to develop a topic modeling approach. Since the author already answered the other reviewers, I have only some points concerning the text. - Since the method involves techniques from different areas, I believe it would be better to explain in more detail some basic concepts. For example, how to calculate and interpret "topic coherence" and "Von Neumann entropy". Note that these concepts are simple, but researchers only from one of the areas involved in this work would not know these concepts. - On page 12, the authors describe that a co-occurrence graph is used. However, it would help the reader to describe more characteristics of this graph. More specifically, if the edges are directed and/or weighted and how it is determined. - Topic modeling based on communities of complex networks have been studied in: Silva FN, Amancio DR, Bardosova M, Costa LD, Oliveira Jr ON. Using network science and text analytics to produce surveys in a scientific topic. Journal of Informetrics. 2016 May 1;10(2):487-502. Furthermore, a multi-scale extension of this method was proposed in: Ceribeli C, de Arruda HF, da Fontoura Costa L. How coupled are capillary electrophoresis and mass spectrometry?. Scientometrics. 2021 May;126(5):3841-51. Considering that both papers are significantly related to the paper, I think these references could be considered. - In step 5 of section 4.2, it is written: "The α as the smoothing factor." I think that it should be: "α is the smoothing factor." Furthermore, in "Moreover, the β is used as a smoothing factor", I think there is no "the". ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
15 Jan 2022 Comments by Reviewer 1 Grammatical changes appear to be satisfactory. I have no further comments to make regarding the revision of this paper. Response: We are thankful to the reviewer for his valuable time and feedback. The suggestions were very helpful and contributed in improving the overall quality of our manuscript. Comments by Reviewer 3 This paper proposes an interesting approach for topic modeling. More specifically, the authors put together concepts of NLP and network science to develop a topic modeling approach. Since the author already answered the other reviewers, I have only some points concerning the text. Response: We are thankful to the reviewer for taking time to review the paper and liking it. We tried to incorporate all the concerns of the reviewer and we hope that the revisions in the paper will meet the expectations of the reviewer. A detailed comment by comment response is provided in a separate document named "response to reviewers". We are very thankful to the reviewers and editor for contributing to improve the quality of this manuscript. Submitted filename: Round3 Response to Reviewers.doc Click here for additional data file. 14 Feb 2022 Hierarchical Lifelong Topic Modeling Using Rules Extracted From Network Communities PONE-D-20-29449R3 Dear Dr. Khan, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Jerry Chun-Wei Lin Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Thank you for addressing my reviews. The current version of the paper does not pose any obvious flaws according to the PLOS ONE publication model. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No 23 Feb 2022 PONE-D-20-29449R3 Hierarchical Lifelong Topic Modeling Using Rules Extracted From Network Communities Dear Dr. Khan: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Prof. Jerry Chun-Wei Lin Academic Editor PLOS ONE
  7 in total

1.  How Hierarchical Topics Evolve in Large Text Corpora.

Authors:  Weiwei Cui; Shixia Liu; Zhuofeng Wu; Hao Wei
Journal:  IEEE Trans Vis Comput Graph       Date:  2014-12       Impact factor: 4.579

2.  Incorporating Domain Knowledge into Topic Modeling via Dirichlet Forest Priors.

Authors:  David Andrzejewski; Xiaojin Zhu; Mark Craven
Journal:  Proc Int Conf Mach Learn       Date:  2009

3.  Heat diffusion: thermodynamic depth complexity of networks.

Authors:  Francisco Escolano; Edwin R Hancock; Miguel A Lozano
Journal:  Phys Rev E Stat Nonlin Soft Matter Phys       Date:  2012-03-14

4.  Discovering health topics in social media using topic models.

Authors:  Michael J Paul; Mark Dredze
Journal:  PLoS One       Date:  2014-08-01       Impact factor: 3.240

5.  Redundancy-aware topic modeling for patient record notes.

Authors:  Raphael Cohen; Iddo Aviram; Michael Elhadad; Noémie Elhadad
Journal:  PLoS One       Date:  2014-02-13       Impact factor: 3.240

6.  DEFENDER: Detecting and Forecasting Epidemics Using Novel Data-Analytics for Enhanced Response.

Authors:  Nicholas Thapen; Donal Simmie; Chris Hankin; Joseph Gillard
Journal:  PLoS One       Date:  2016-05-18       Impact factor: 3.240

7.  Online Knowledge-Based Model for Big Data Topic Extraction.

Authors:  Muhammad Taimoor Khan; Mehr Durrani; Shehzad Khalid; Furqan Aziz
Journal:  Comput Intell Neurosci       Date:  2016-04-19
  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.