Literature DB >> 36157791

Heterogeneous deep graph convolutional network with citation relational BERT for COVID-19 inline citation recommendation.

Tao Dai1, Jie Zhao2, Dehong Li2, Shun Tian1, Xiangmo Zhao3, Shirui Pan4.   

Abstract

The outbreak of COVID-19 brings almost the biggest explosions of scientific literature ever. Facing such volume literature, it is hard for researches to find desired citation when carrying out COVID-19 related research, especially for junior researchers. This paper presents a novel neural network based method, called citation relational BERT with heterogeneous deep graph convolutional network (CRB-HDGCN), for COVID-19 inline citation recommendation task. The CRB-HDGCN contains two main stages. The first stage is to enhance the representation learning of BERT model for COVID-19 inline citation recommendation task through CRB. To achieve the above goal, an augmented citation sentence corpus, which replaces the citation placeholder with the title of the cited papers, is used to lightly retrain BERT model. In addition, we extract three types of sentence pair according citation relation, and establish sentence prediction tasks to further fine-tune the BERT model. The second stage is to learn effective dense vector of nodes among COVID-19 bibliographic graph through HDGCN. The HDGCN contains four layers which are essentially all sub neural networks. The first layer is initial embedding layer which generates initial input vectors with fixed size through CRB and a multilayer perceptron. The second layer is a heterogeneous graph convolutional layer. In this layer, we expand traditional homogeneous graph convolutional network into heterogeneous by subtly adding heterogeneous nodes and relations. The third layer is a deep attention layer. This layer uses trainable project vectors to reweight the node importance simultaneously according to both node types and convolution layers, which further promotes the performance of learnt node vectors. The last decoder layer recovers the graph structure and let the whole network trainable. The recommendation is finally achieved by integrating the high performance heterogeneous vectors learnt from CRB-HDGCN with the query vectors. We conduct experiments on the CORD-19 and LitCovid datasets. The results show that compared with the second best method CO-Search, CRB-HDGCN improves MAP, MRR, P@100 and R@100 with 21.8%, 22.7%, 37.6% and 21.2% on CORD-19, and 29.1%, 25.9%, 15.3% and 11.3% on LitCovid, respectively.
© 2022 Elsevier Ltd. All rights reserved.

Entities:  

Keywords:  COVID-19 citation recommendation; Citation enhanced BERT; Deep graph convolutional network; Heterogeneous graph; Text representation learning

Year:  2022        PMID: 36157791      PMCID: PMC9482209          DOI: 10.1016/j.eswa.2022.118841

Source DB:  PubMed          Journal:  Expert Syst Appl        ISSN: 0957-4174            Impact factor:   8.665


Introduction

The COVID-19 pandemic broken out in early 2020 has brought a worldwide public health crisis, and results in an exponentially growth of COVID-19 research papers. According to previous studies, the volume of COVID-19 literature is doubling increasing every two weeks (Brainard, 2020). Facing such large amount of papers, it is hard for researchers to acquire desired cited COVID-19 scholarly papers which meet their requirements. Therefore, it is necessary to provide an efficient an effective COVID-19 citation recommendation service for researchers, especially for who are new to COVID-19 research field. Many approaches have been proposed to retrieve COVID-19 papers (Fister et al., 2020, MacAvaney et al., 2020, Wise et al., 2020), and many of them applied pretrained language models (PLMs) such as BERT (Kenton & Toutanova, 2019) which has been widely used in many other tasks. However, there are still two defects for the above methods. Firstly, the used word vectors in most previous studies lack deep understanding of COVID-19 papers. In COVID-19 literature, there are a lot of frequently used specific words which rarely appear in general fields. Only using common papers to train word vectors will result in deviation in the understanding of COVID-19 information (Andre et al., 2020, MacAvaney et al., 2020, Mass and Roitman, 2020), which further decrease the performance of COVID-19 inline citation recommendation. Secondly, most previous methods solely obtain paper vectors based on their content, and neglect valuable citation relationship information between COVID-19 papers. Citation link is a special data that widely existing in academic literature, which contains latent semantic co-occurrences between papers. Semantic co-occurrences relations has been proved as a vital factor when training PLMs (Gan et al., 2022). For example, a citing sentence contains “coronavirus cross immunity” usually citing papers that contains “antibody-depend enhancement from coronvarius-like”. On the surface, there is only one common key word “coronavirus” between them, but the two sentences actually express nearly similar meaning and are co-occurred with the form of citation relations. Therefore, how to use these citation relationships to further enhance the semantical understanding of PLMs for COVID-19 papers is an important issue for COVID-19 citation recommendation task. Besides the above defects, existing COVID-19 paper retrieval and recommendation methods are difficult to effectively utilize heterogeneous information. Due to the burstiness and severity, the COVID-19 pandemic has evolved into a cross-border crisis and attracted multiple areas of research, including social management, medicine, biology and computer science (Aristovnik, Ravšelj, & Umek, 2020). Comparing with common scientific literature data, the information among COVID-19 papers, such as text content, authors and venues, have more character of heterogeneous. For paper and citation recommendation task, the heterogeneous character will bring more obstacles for finding desired citations (Zhu et al., 2021). Therefore, how to fully mine the association between heterogeneous nodes, and distinguish the importance of them to further refine optimal node representation, is a key issue for improving the performance of COVID-19 citation recommendation. To address the aforementioned problems, we propose a novel deep learning based method, called citation relational BERT with heterogeneous deep graph convolutional network (CRB-HDGCN), for COVID-19 inline citation recommendation task. The proposed method first uses a citation augmented corpus to lightly retrain and fine-tune BERT model. Then the method uses a heterogeneous deep graph convolutional network to obtain reliable vector representations for different nodes. Specifically, we introduce a deep attention layer for convolution network, which is able to distinguish node importance by considering both the convolutional layers and node types simultaneously. Experimental results on CORD-19 and LitCovid datasets show that our CRB-HDGCN model is more effective than other baseline methods. The main contributions of this paper are summarized as follows. (1) We propose a novel neural network based citation recommendation method, called CRB-HDGCN, for COVID-19 related papers. CRB-HDGCN is able to fully utilize citation relations and heterogeneous links to obtain reliable vectors for papers, authors and venues related to COVID-19. (2) We propose a new solution to retrain and fine-tune PLMs by utilizing citation relations among scientific papers, which is benefit PLMs to better understand citing activities. (3) We expand homogeneous graph convolutional network into heterogeneous deep graph convolutional network. The new network can fully utilize the heterogeneous relations to learn reliable node vectors, and is able to distinguish node importance by considering both the depth of network and the node types simultaneously. (4) Thorough experimental studies on CORD-19 and LitCovid datasets are carried out to validate the effectiveness of the proposed method. The rest of the paper are organized as follows. Section 2 reviews related work. The retrain and fine-tune manner of citation relational BERT is introduced in Section 3. Section 4 presents heterogeneous deep graph convolutional network in detail. Section 5 presents the experimental results and analysis. The paper is concluded in Section 6.

Related work

COVID-19 paper retrieval

As far as we know, there is no research work on COVID-19 citation recommendation. The most closely research work is COVID-19 paper retrieval. Therefore, we briefly overview COVID-19 paper retrieval methods in this subsection. The related previous works of COVID-19 paper retrieval are summarized in Table 1. Since the outbreak of COVID-19 pandemic, scientists worldwide have been carried many works to deal with this global public health crisis, and the research literature about COVID-19 is increasing with an extraordinary rate. In order to further accelerate the research of COVID-19, Allen AI Institute and other top research groups publicly released the COVID-19 open research dataset (CORD-19) (Wang et al., 2020) in March 2020, which covers all current coronavirus research literature and continuously updated weekly. Meanwhile, National Library of Medicinethis published a lightweight COVID-19 paper dataset LitCovid (Chen, Allot, & Lu, 2021), which is a collection of recently published PubMed articles which are directly related to the 2019 coronavirus. From that on, a variety of COVID-19 paper retrieval approaches have been proposed in the literature. The research team of Amazon released AWS CORD-19 Search,1 a literature search engine for the field of COVID-19, by using named entity recognition, knowledge graph and topic model technologies. The research team of Google released COVID-19 Research Explorer2 which uses biomedical field data to train the BERT model (Kenton & Toutanova, 2019), and combined keywords and deep paper vectors to achieve COVID-19 paper search. However, according to the research of Soni and Roberts (2021), the accuracy of these two commercial COVID-19 search engines are not satisfied. Zhang, Gupta, Nogueira, Cho, and Lin (2020) proposed a coronavirus literature retrieval method Neural Covidex based on BioBERT (Lee et al., 2020). Neural Covidex obtains deep vector representation for the query and candidate COVID-19 papers, and establishes paragraph level index to increase the search granularity. Rohatgi et al. (2020) expanded the CORD-19 dataset, and developed a literature search engine COVIDSeer by using SciBERT (Beltagy, Lo, & Cohan, 2019). Wise et al. (2020) combined SciBERT and knowledge graph embedding to obtain paper vector to achieve COVID-19 paper retrieval. Fister et al. (2020) analyzed papers in COVID-19 field by using word frequency information, and then used association rules based on information cartography to obtain COVID-19 papers. Andre et al. (2020) linearly combined TF-IDF vector, BM25 vector and Sentence-BERT (Reimers & Gurevych, 2019) vector to calculate the similarity between query and candidate papers to generate final retrieval results. Ostendorff, Ruas, Blume, Gipp, and Rehm (2020) used chapter title in cited literature as an aspect to measure the similarity between cited literature. They also compared the performance of different pre-training models when applied to COVID-19 paper retrieval. MacAvaney et al. (2020) used BM25 vectors to obtain initial search list, and then used SciBERT to reorder the list so as to obtain the final COVID-19 retrieve list. Ahamed and Samad (2020) applied a betweenness centrality measurement algorithm in social networks filed, and used keywords related to drugs, diseases, pathogens, hosts of pathogens, and biomolecules to construct a keywords graph for COVID-19 field, which can be used for fine-grained COVID-19 paper retrieval. Nguyen, Rybinski, Karimi, and Xing (2020) proposed a hybrid COVID-19 literature retrieval method. The method first combines the BM25 similarity and cosine similarity between the query and the candidate literature. Then it uses divergence from randomness similarity to reorder the retrieval list, and finally applied BioBERT vectors to sort the retrieval list again to obtain the final results. Xiong et al. (2020) mitigated the domain discrepancy and label scarcity problems by the domain-adaptive pretraining and few-shot learning technologies, and integrated dense retrieval to alleviate vocabulary mismatch obstacle of sparse retrieval. Mass and Roitman (2020) proposed a COVID-19 literature retrieval method based on weakly supervised learning. The method first fine-tunes BERT and GPT (Radford et al., 2019) models with title-abstract relationship, and then integrates multiple similarity to generate output the final paper retrieval results according to a comprehensive score. Tran et al. (2021) proposed a COVID-19 paper retrieval system from extracted entities and relations. They use BERT to encode each relation, and score them using pointwise mutual information. Uday, Pavani, Lakshmi, and Chivukula (2022) combined vectors from bag of words, Word2Vec, BERT and TF-IDF to to find relevant COVID-19 documents. The main defect of existing COVID-19 paper retrieval methods is the retraining process of PLMs totally neglect citation information, which will result in lacking deep understand for citing activities. Moreover, they all neglect valuable heterogeneous information such as author and venue, which are vital for inline citation recommendation task. Our method solves the first defect by expanding original citing sentence with cited titles, which subtly infuses citation relations into PLMs without breaking its originally training procedure. We also use several sentence relations extracted from citation relations to further enhance the PLMs comprehension ability for citing activities. The second defect is solved by introducing HDGCN which is easy to integrate heterogeneous information when learning the high level representations for COVID-19 corpus, which can essentially improve the performance of COVID-19 inline recommendation.
Table 1

Related works summarization for COVID-19 paper retrieval.

Title DateUsing techniques
COVID-19 academic datasetsCORD-19: The COVID-19 open research dataset.
2020

LitCovid: An open database of COVID-19 literature.2020

Commercial searching enginesAmazon AWS CORD-19 search.
2020
Knowledge Graph; Topic model
Google COVID-19 research explorer.2020BERT; Hybrid Neural model

COVID-19 paper retrieve methodsRapidly deploying a neural search engine for the COVID-19 open research dataset.
2020
BioBERT; BM25; Paragraph-level index
Covidseer: Extending the CORD-19 dataset.
2020
TF-IDF; SciBERT
COVID-19 knowledge graph: accelerating information retrieval and discovery for scientific literature.
2020
Knowledge graph; SciBERT
Discovering associations in COVID-19 related research papers.
2020
Association rule; Information cartography
Co-search: COVID-19 information retrieval with semantic search, question answering, and abstractive summarization.
2020
TF-IDF; BM25; Sentence-BERT
Aspect-based document similarity for research papers.
2020
PLMs
SLEDGE-Z: A zero-shot baseline for COVID-19 literature search.
2020
BM25; SciBERT
Information mining for COVID-19 research from a large volume of scientific literature.
2020
Betweenness centrality algorithm
Searching scientific literature for answers on COVID-19 questions.
2020
BM25; BioBERT
CMT in TREC-COVID round 2: mitigating the generalization gaps from web to special domain search.
2020
Domain-adaptive pretraining; SciBERT
Ad-hoc document retrieval using weak supervision with BERT and GPT2.
2020
BERT; GPT2
Covrelex: A covid-19 retrieval system with relation extraction.
2021
BERT; named entity
COVID-19 literature mining and retrieval using text mining approaches.2022BOW; TF-IDF; Word2Vec;BERT
Related works summarization for COVID-19 paper retrieval. Related works summarization for citation recommendation.

Inline citation recommendation

In the academic landscape, the rapid rise of big scholarly data (BSD) brings about new issues and challenges with respect to data management and analysis (Xia, Wang, Bekele, & Liu, 2017). Exploring BSD can provide great benefits for various stakeholders, including community detection (Mercorio, Mezzanzanica, Moscato, Picariello, & Sperlí, 2019), academic social network analysis (Kong, Shi, Yu, Liu, & Xia, 2019), scientific impact evaluation (Kong et al., 2020), etc. Citation recommendation is one of key application for BSD which aims to provide researchers a list of references satisfying their citing expectations. According to different usages, citation recommendation can be divided into global recommendation and inline recommendation. The global citation recommendation (Ayala-Gomez et al., 2018, Dai et al., 2021, Zhu et al., 2021) returns the reference that are relevant to the entire query paper. In contrast, inline citation recommendation, also called context-aware citation recommendation, only provides references for a citation placeholder according to its local context. The work of this paper belongs to context-aware citation recommendation. He, Pei, Kifer, Mitra, and Giles (2010) first introduced the concept of context-aware citation recommendation in 2010. From then on, many context-aware citation recommendation methods have been proposed in the literature (Dai et al., 2021, Duma et al., 2016, Ebesu and Fang, 2017, Färber et al., 2018, Gu et al., 2022, He et al., 2011, Huang et al., 2015, Huang et al., 2014, Jiang et al., 2018, Khadka and Knoth, 2018, Peng et al., 2016, Totti et al., 2016, Yin and Li, 2017, Zhang and Ma, 2022). Recently, neural networks have been widely used in context-aware citation recommendation field due to its promising performance. Huang et al. (2015) presented the first neural network based context-aware citation recommendation method. The method is formed as a multi-layer neural network to predict a cited scientific paper by given a query citation context. Ebesu and Fang (2017) proposed a context-aware citation recommendation method by using neural encoder–decoder architecture. The encoder part of the model is based on max time delay neural network (TDNN) which encode a query citation context into a hidden vector. The decoder is based on recurrent neural network (RNN) to recover the hidden vector into the title of cited paper. The method also considers author information when generating hidden vectors, which is proved to be effective for recommendation. Yin and Li (2017) proposed a personalized convolutional neural networks (PCNN) to recommend citations when given a citation context. PCNN also considers author information in the input layer. Khadka and Knoth (2018) tried to avoid topic drifting problem in citation recommendation by combining citation context with citation position. Dai, Zhu, Wang, and Carley (2019) propose a context-aware citation recommendation method ASL by combining stacked denoising autoencoders (SDAE) and Bi-LSTM. For a scientific paper, they use attention mechanism to extend vanilla SDAE architecture into attentive SDAE (ASDAE). They also extended Bi-LSTM with a local attention layer to learn the hidden vectors for citation context, which can not only capture key information for effective embedding, but also benefit for extracting suitable citation context. Gu et al. (2022) used a hierarchical attention network to excavate the text embeddings of cited papers, and use these embedding to prefetch a limited number of relevant documents. The used hierarchical attention network is coupled with a SciBERT reranker fine-tuned on contex-aware citation recommendation task. Zhang and Ma (2022) proposed a embedding-based neural network called dual attention model for citation recommendation task. The model comprises two attention mechanisms: self-attention and additive attention. The self-attention aims to capture the relatedness between the contextual words and structural context, and additive attention aims to learn importance of them. The two attention mechanisms is able to interpret “relatedness” and “importance” through the learned attention weights. For readers to quick check the precious works of citation recommendation, we list them in Table 2.
Table 2

Related works summarization for citation recommendation.

Title DateUsing techniques
Global citation recommendationGlobal citation recommendation using knowledge graphs.
2018
Knowledge graph; Learning to rank
Recommending scientific paper via heterogeneous knowledge embedding based attentive recurrent neural networks.
2021
Graph learning; Bi-LSTM
Gated relational stacked denoising autoencoder with localized author embedding for global citation recommendation.2021SDAE; attention mechanism

Inline citation recommendationContext-aware citation recommendation.
2010
Non-parametric probabilistic model
A neural probabilistic model for context based citation recommendation.
2015
Multi-layer neural network
Neural citation network for context-aware citation recommendation.
2017
Time delay neural network; RNN
Personalized citation recommendation via convolutional neural networks.
2017
CNN
Using citation-context to reduce topic drifting on pure citation-based recommendation.
2018
Topic model; Word2Vec
Attentive stacked denoising autoencoder with Bi-LSTM for personalized context-aware citation recommendation.
2019
SDAE; Bi-LSTM; attention mechanism
Local citation recommendation with hierarchical-attention text encoder and SciBERT-based reranking.
2022
Sci-BERT; attention mechanism
Dual attention model for citation recommendation with analyses on explainability of attention mechanisms and qualitative experiments.2022Neural model; Dual attention mechanism
Due to the reason of generality, the existing inline citation recommendation methods are all focus on common scientific literature, which lead to poor vector performance for COVID-19 domain specific inline citation recommendation task. Our method solves the above problem by using citation relational BERT that retrained and fine-tuned on citation relations among COVID-19 corpus, which improves the domain representation of vectors while fitting inline citation recommendation task. Moreover, HDGCN in our method expands the learning ability of conventional GCN. It can be easily expanded with various heterogeneous information. Moreover, the deep attention layer in HDGCN makes it can be applied into large graph when carrying out multiple deep graph convolution, which is also benefit for common citation recommendation task.

Citation relational BERT on COVID-19 papers

Contexuralised representations of pretrained language models (PLMs) like ELMo (Peters, Neumann, Iyyer, Gardner, & Zettlemoyer, 2018), BERT (Kenton & Toutanova, 2019), and GPT (Radford et al., 2019) have accomplished promising results in many natural language processing (NLP) tasks. Especially, Hugging Face provided a bert-small-cord19 model3 which uses a subset of CORD-19 dataset to fine-tune vanilla BERT. Chen, Du, Allot, and Lu (2022) presented LitMC-BERT4 which trains BioBERT (Lee et al., 2020) on LitCovid dataset for multi-label classification task. Nevertheless, these PLMs still neglect vital relations between COVID-19 papers, which may results in lacking understanding citing activities in papers. In this section, we use citation relations among COVID-19 papers to lightly retrain and fine-tune BERT to let PLM is able to generate more reliable word vectors that more suitable for citation recommendation task. We call the new BERT model as citation relational BERT (CRB). For CORD-19 dataset, the retraining and fine-tuning process is carried on bert-small-cord19 model. For LitCovid dataset, this process is carried on the back bone of LitMC-BERT. The brief retraining and fine-tuning process is demonstrated in Fig. 1.
Fig. 1

The brief retraining and fine-tuning process of CRB.

The brief retraining and fine-tuning process of CRB.

Retraining language model with augmented citing sentences

Existing BERT based COVID-19 paper retrieval and recommendation methods all simply delete citation information in papers. Inspired by the work of Huang et al. (2012), we believe the appearance of references has its specific meaning which may conceal particular latent information for citation recommendation task. The latent semantic information can be revealed in two aspects. The first one is the particular citing places. When an author want to cite a paper, written a citing sentence is the pre-order action. According to the research of Qazvinian and Radev (2010), citing sentences usually have their own particular syntax compared with non-citing sentences, and different pattern of citing sentences may existed in different citing places. The second one is the deep semantic of citing sentences. Subjected to the length of papers, authors usually briefly introduce the work of cited papers, and present the reference for readers to further understand related works. This means a citing sentence contains much more deep semantic information that beyond the citing sentence itself. Therefore, both from the syntax and semantic level, simply neglecting citations in text body will loss these latent information, and result in deviation in understanding citation activities. To further explore the latent information in citations, we construct a special citation augmented corpus with COVID-19 related papers. The construction manner is demonstrated in Table 3. As can be seen, we regard the citation placeholder in the original citing sentence as a special fragment, and replace the citation placeholder with the title of the cited papers to form an augmented citation sentence. As can be seen, for one hand, the added title is among the citing sentence, which does not break the structure context of other non-citing sentences. On the other hand, the added title provides a further supplement for the citing sentence, and enriches its meaning. The added title reorganizes the written logical of citing sentence, which makes the original paper presents a new language structure.
Table 3

An example construction of citation augmented corpus.

Original citing sentenceAugmented citing sentence
These structures defined a novel family of enzymes, now called sedolisins or serine-carboxyl peptidases, that is characterized by the utilization of a fully conserved catalytic triad (Ser, Glu, Asp) and by the presence of an Asp in the oxyanion hole [8].These structures defined a novel family of enzymes, now called sedolisins or serine-carboxyl peptidases, that is characterized by the utilization of a fully conserved catalytic triad (Ser, Glu, Asp) and by the presence of an Asp in the oxyanion hole , Aorsin, a novel serine proteinase with trypsin-like specificity at acidic pH.
We select 1,861,412 and 235,421 citing sentences from CORD-19 and LitCovid datasets respectively, and use them to form augmented corpus. Then we use all these augmented sentences to retrain BERT by applying the Masked Language Model (MLM) objective. The model is retrained for 100 epochs in batches of 64 samples by employing the Adam with learning rate 5e-5. An example construction of citation augmented corpus.

Fine-tuning language model with additional link information

To enable BERT further effectively understand the citing relations between COVID-19 papers, we further use sentence relation prediction task to lightly fine-tune it. We first extract three types of sentences pair, including: (1) the title of citing and cited papers; (2) the title of papers that cite same paper; (3) the title of papers that are cited by same paper. After obtaining the above three sentence pairs, we use retained BERT in previous subsection to predict whether there is correlation between them to execute fine-tune process. For example, suppose we have sentence and which are the titles of citing and cited papers, we send them into retained BERT and obtain all word vectors of them. Then we sum all word vectors and calculate mean vectors to get sentence vectors and . After that, we calculate the relevance score of and by , and use the following cross entropy loss to execute fine-tune process. where is the overall number of sentence pairs, is 1 if and is a selected sentence pair. The main novelty of CRB is we infuse the learning of citation relations by a masked tokens prediction task, which can be easily carried on by partial original training process of BERT model. Such a practice is not like common approaches that heavily retrain BERT model with full COVID-19 corpus following whole original training process (Kenton and Toutanova, 2019, Müller et al., 2020), which totally ignore the citation relations. It is also not like other solutions that consider relation prediction as a downstream task to fine-tune BERT model (Andre et al., 2020, Mass and Roitman, 2020), which cannot acquire fine-grained semantic co-occurrences among citation relations.

Heterogeneous deep graph convolutional network for COVID-19 paper recommendation

In previous section, we have obtained CRB by excavating latent citation information to obtain more robust word vectors. To achieve promising citation recommendation for COVID-19 papers, we still need to obtain reliable paper vectors. Moreover, other related items, such as authors and venues with their correlations, are also need to be utilized. In this section, we present heterogeneous deep graph convolutional network to achieve the above goal. The detailed architecture of proposed network are shown in Fig. 2, which contains initial embedding layer, heterogeneous graph convolutional layer, deep attention layer and decoder layer.
Fig. 2

The detailed architecture of heterogeneous deep graph convolutional network. There are essentially four layers in the network. The initial embedding layer firstly transfers nodes in COVID-19 paper dataset into vectors with fixed size by a multilayer perceptron. Secondly, the heterogeneous graph convolutional layer embeds the initial vectors into heterogeneous matrices by utilizing multiple relations among them. Thirdly, for each th vector matrices , the deep attention layer utilizes three learnable project vectors and to extract importance regulating factors and , respectively. The final vectors are then generated by combing all reweighted vectors in all convolutional layers. Finally, the decoder layer recovers the network structure from obtained vectors to ensure the trainability of the whole network.

The main novelties of our HDGCN are reflected in two aspects. The first one is we extend homogeneous GCN (Jeong, Jang, Park, & Choi, 2020) into heterogeneous by subtly expanding the normalized Laplacian with various nodes and relations. The second one is the deep attention layer in HDGCN is able to promote the importance of long distance heterogeneous nodes when stacking plenty convolution layer, thus relieves over-smoothing issue (Chen et al., 2020, Li et al., 2018, Liu et al., 2020, Xu et al., 2018) of deep GCN in heterogeneous condition. The detailed architecture of heterogeneous deep graph convolutional network. There are essentially four layers in the network. The initial embedding layer firstly transfers nodes in COVID-19 paper dataset into vectors with fixed size by a multilayer perceptron. Secondly, the heterogeneous graph convolutional layer embeds the initial vectors into heterogeneous matrices by utilizing multiple relations among them. Thirdly, for each th vector matrices , the deep attention layer utilizes three learnable project vectors and to extract importance regulating factors and , respectively. The final vectors are then generated by combing all reweighted vectors in all convolutional layers. Finally, the decoder layer recovers the network structure from obtained vectors to ensure the trainability of the whole network.

Initial embedding layer

The initial embedding layer is aim to embed raw words into vectors. For th COVID-19 paper , we send its all sentences into CRB to obtain all word vectors. Then we calculate the mean vectors from obtained word vectors to get paper vector . Then we use a multilayer perceptron (MLP) to embed it into a lower dimension vector where is a fully connected linear function. The layer dimension in MLP is set to . After getting initial vectors for all papers, the next step is to obtain initial vectors of authors and venues. for th author and th venue , we obtain their initial vectors and as follows. where avg() is vector mean function. represents all papers published by author , and represents all papers included in venue .

Heterogeneous graph convolutional layer

In previous subsection, we have obtained initial vector representations COVID-19 related papers, authors and venues. These vectors only contain information based on text content, and neglect vital link relationship between them. In this subsection, we will utilize link relationships to build the connection between these vectors, and use heterogeneous graph convolutional layer to further boost the performance of learnt vectors. A demonstration graph which illustrates the involved relationships and nodes in bibliographical heterogeneous graph. We first represent COVID-19 paper dataset as a heterogeneous graph , where denotes node vectors of papers, authors and venues. The value of is obtained from the initial embedding layer in previous subsection. According to different types of nodes, edge set contains 6 types of relations, including 3 types of homogeneous relations and heterogeneous relations: paper citation relations , co-author relations , venue relations , author writing relations , paper publication relations , and author publication relations . Besides , other relations have explicit meaning and are easily to be established. For example, if author published a paper in venue , then the value of is 1; otherwise 0. The is calculated according to the cosine similarity between venue vectors. We select the top 30% nodes as semantic neighbors for each venue. Then we set the value to 1 if a venue is a semantic neighbor of an other venue. A demonstration graph is presented in Fig. 3 which illustrates the involved relationships and nodes in heterogeneous bibliographical graph .
Fig. 3

A demonstration graph which illustrates the involved relationships and nodes in bibliographical heterogeneous graph.

Based on , we construct an augmented matrix as follows. The next step is to send node vectors into heterogeneous graph convolutional network. The network contains multiply layers. The th layer is executed as follows. where is the rectifier activation function. is a learnable convolution kernel. , and . represents the node representation in th layer. , and are the overall number of papers, authors and venues in dataset, respectively. Particular, is formed by the initial node vectors. The exactly construction of and are as follows. where , and are the paper, author and venue vectors in th layers, respectively.

Deep attention layer

Through the layered by layered manner, graph convolutional network is able to incorporate heterogeneous correlations between nodes to improve the performance of node vectors. However, there is as serious problem for a deep GCN: each layer more considers immediate neighbors, and the performance decreases when the whole network goes deeper. Several recent studies attribute this performance deterioration to the over-smoothing issue, which states that repeated propagation makes node representations of different classes indistinguishable (Chen et al., 2020, Li et al., 2018, Xu et al., 2018). Liu et al. (2020) propose a solution to address the above problem. Nevertheless, their solution only considered homogeneous graphs. In a scholar network, different types of nodes obviously plays different roles when the author want to cite a paper. For example, one researcher cited a paper because it was published in an influential venue, while another researcher cited the paper because it was the work of a coauthor. This enlightens us that we need to develop a mechanism to distinguish node importance by considering both the depth of network and the type of nodes simultaneously. We use the deep attention layer in this subsection to achieve the above goal. The core of deep attention layer is three trainable project vectors , which are used to learn the importance of papers, authors and venues in different covolutional layers. As shown in Fig. 2, we stack output vectors in all layers according to different type, thereby obtaining three tensors , and for papers, authors and venues, respectively. Then we use the following process to generate importance regulating factors for all layers. where represents sigmoid function. is a learnable mapping matrix. , , are generated importance regulating factors for paper vectors, author vectors and venue vectors in all layers. Next, we use the obtained importance regulating factors to reweight all output vectors to obtain final vector representation . where represents Hadamard product, and denotes obtaining the sum mean vector.

Decoder layer

In order to ensure the effectiveness of obtained vectors, we put into a decoder layer to recover the link information of heterogeneous graph with probability , which equals to maximum the following likelihood function. where is the overall number of nodes. and represent any two node vectors in . is the entry value of the augmented matrix related to th node and th node.

COVID-19 paper recommendation

When finishing training, we can use our method to recommend COVID-19 citations. Given a citation context with query author and query venue , we first send them into CRB, and then use initial embedding layer in Section 4.1 to obtain query vectors , where is the mean vector of , is the mean vector of the papers published by the provided author and is the mean vector of the papers contained in the provided venue. Then we calculate the score for each candidate paper as follows. where , and are the paper vector, author vector and venue of candidate paper . The final recommendation list is generated according to with the higher score.

A running example for HDGCN

To let readers understand the process of our HDGCN more clearly, we present a running example of a HDGCN with 2 convolutional layers. As shown in the left part of Fig. 4, the input heterogeneous graph is consist of three types of nodes. We use round, diamond and pentagram to represent paper, author and venue, respectively. According to the degree of nodes, the input graph can be divided into three circles. We only observe the vector changing of nodes in the inner circle.
Fig. 4

A running example of a HDGCN with 2 convolutional layers.

The initial vectors are shown in the top part of Fig. 4. As can be proved in Li et al. (2018), a single layer convolution means a propagation process which makes nodes gathering neighborhood information, and more layers of convolution will involve neighbors that have longer distance. Therefore, as can be seen in the middle part of Fig. 4, when carrying out 1 layer heterogeneous graph convolution, the propagation mechanism makes the nodes in the inner circle gathering their neighborhood information. The colorless nodes denote the neighbors that have not been gathered, while purple nodes is the neighbors that have been gathered. Because we only carried out convolution for one time now, only neighbors that have 1 degree are involved. For the sake of simplicity, let us assume that a node is more important only when it has more information gathering ability. In this time, we can see that the author node (inner red diamond) contributes more information gathering for it has more neighbors with 1 degree. Therefore, our attention layer preserves the vector of author and decrease the importance of the other two nodes. When we continue carrying out 2 layer convolution, the inner nodes will continue gather the nodes with 2 degree. As can be seen in the bottom of Fig. 4, our deep attention layer preserves the vector of venue node (inner green pentagram), for it has more neighbors with 2 degree. It is worth noting that the color of preserved venue vector is a little lighter, for our deep layer considers venue has less impact for paper retrieval task. The final vectors are then generated according to the reweighed vectors in all convolution layers. In this running example, we can see that our HDGCN is able to reweight node vectors by not only the information of local or global neighbors, but also the contribution of node types, thus generate more reliable vector representations. A running example of a HDGCN with 2 convolutional layers.

Experiments and evaluations

Datasets and preprocessing

In order to evaluate the performance of our proposed method, we use two publicly available datasets: CORD-19 and LitCovid. The briefly introduction of used datasets are as follows. CORD-19 dataset5 : the COVID-19 Open Research Dataset (CORD-19) (Wang et al., 2020) is established by the U.S. White House, along with the U.S. National Institutes of Health, the Allen Institute for AI, the Chan-Zuckerberg Initiative, Microsoft Research, and Georgetown University. It is one of the earliest scientific paper datasets related to COVID-19 and similar coronaviruses such as SARS, MERS and other earlier coronaviruses since 2002. The sources of these papers including the WHO, and bioRxiv, medRxiv, etc. For evaluation purpose, we only keep the paper that cited over 5 times and related citing paper. We finally obtain a subset that contains 327,619 papers. LitCovid dataset6 : this lightweight dataset is published by National Library of Medicine (Chen et al., 2021). It is a collection of recently published PubMed articles which are directly related to the 2019 novel coronavirus. The dataset contains upwards of 240,000 articles and growing every day, making it a comprehensive resource for keeping researchers up to date with the current COVID-19 crisis. For our recommendation purpose, we select as subset that contains 103,175 articles. For each papers, we pre-process it by removing stop words and stemming using Porter stemmer. To reduce the impact of short words, we remove the words which consist of less than two characters and appear less than ten times in the dataset. We consider the citing sentence with additional 2 sentences around the citation placeholder as a query, and the corresponding cited papers as ground truth retrieve result. For evaluating the performance of citation recommendation, we select 37,289 and 6,613 citation context in CORD-19 and LitCovid as test set, respectively. Table 4 summarizes the statistics of used two datasets.
Table 4

Statistics of used datasets.

PapersAuthorsVenues
CORD-19train327,619284,58115,924
test37,28935,4177,423

LitCovidtrain103,17590,6926,305
test6,6135,9342,081
Statistics of used datasets.

Evaluation metrics

To thoroughly evaluate the recommendation of our method, we choose the following four types of metric. Precision and recall: These two types of metric are the most widely used metrics in information retrieval area. For test samples, we evaluate precision and recall according to the top results using citing score. Precision@N (P@N) measures the ratio of correctly retrieved citations to N. Recall@N (R@N) measures the ratio of correctly retrieved citations to all ground truths. These two metrics are defined as follows. where is a query citation context. is the testing set. represents the top recommended results for . denotes all ground truth citations of . Mean Reciprocal Rank (MRR): This metric is calculated by the reciprocal value of position for the first matched results. MRR is defined as follows. where is the size of testing set. is the first matched recommend result for query citation context . It should be noted that because a query citation context may contains multiple references, we treat them equally and calculate the MRR score for each ground truth citation. Mean Average Precision (MAP): Unlike MRR that limits a query only have one matched result, MAP considers the situation of multiple matched results. For a query , we calculate average precision(AveP) as follows. where is the overall number of ground truths of query . is the position of th matched ground truth in . We calculate AveP value for all testing samples, and then average them to obtain the final MAP value.

Comparing with other approaches

We use the following compared methods to evaluate the performance of our method for COVID-19 citation recommendation. TF-IDF: We extract TF-IDF vectors for all papers and query citation context, and retrieve the papers that have high cosine similarity with the query. To promote the calculation, we only preserve top 8000 words according to word frequency. BM25 (Robertson & Walker, 1994): BM25 is a well-known ranking method for measuring the relevance of matching documents to a query based on the text. We calculate the text similarity between the papers and query citation context by using both TF and IDF for BM25. SciBERT (Beltagy et al., 2019): We send all papers and query into SciBERT, and obtain the mean vector according to their output word vectors. Then same with TF-IDF method, we consider the papers with higher cosine similarity with the query as the recommend results. Association rules (AR) (Fister et al., 2020): By calculating term frequency and inverse term frequency (TF/ITF) and constructing term graph, AR uses information cartography to extract associate rules to explore knowledge among COVID-19 papers. We identify the terms that occur in the best metro map for queries and candidate papers, and return the results that have higher term matching. BERT-DVGCN (Jeong et al., 2020): This method is originally for general citation recommendation task. It uses vanilla BERT to encode query sentences into context embeddings, and uses GCN to embeds paper vectors which is obtained by Doc2Vec (Le & Mikolov, 2014) into graph embedding according to citation relationships. Then Bert-DVGCN concatenates them and uses a softmax output layer to generate recommendation results. CO-Search (Esteva et al., 2021): This method linearly combines TF-IDF vector, Siamese-BERT vector and BM25 vector of queries and answers to generate recommendation results. We use same queries as our method, and consider cited papers as answers for CO-Search. As we can see from Table 5, Table 6, our CRB-HDGCN is superior to other methods in terms of all metrics. For example, compared with the second best method CO-Search, CRB-HDGCN improves MAP, MRR, P@100 and R@100 with 21.8%, 22.7%, 37.6% and 21.2% on CORD-19, and 29.1%, 25.9%, 15.3% and 11.3% on LitCovid. This can be ascribed to our method uses domain related citation relations to let PLM interprets latent citation information between the COVID-19 papers more clearly. Moreover, we excavate a fine-grained paper graph by using multiple relations, and use deep graph neural network to differentiate the importance of nodes according to both type and range, which is able to obtained reliable paper vectors and consequently further improve the performance of citation recommendation. We can find that due to only utilize word frequency information, both TF-IDF and BM25 are obviously worse than all other methods. SciBERT works much better than TF-IDF and BM25, which indicates that the latent information excavated by PLMs is much more useful than word frequency in finding relevant papers. By utilizing link relations, BERT-DVGCN further improves the performance, which demonstrated that link information plays an important role in citation recommendation task. However, both SciBERT and BERT-DVGCN are only trained on common papers, which will results in understanding deviation for domain specific corpus such as COVID-19. AR uses association rules to deeply mine the domain knowledge among COVID-19 papers, and its performance is greatly improved compared with SciBERT and BERT-DVGCN. By employing high performance pre-trained word vector in COVID-19 field, CO-Search further improves performance compared with AR which only utilize word occurrence information. However, the training of word vectors in CO-Search fails to effectively mine citation information among COVID-19 papers, and does not utilize vital author and venue information. Our CRB-HDGCN mitigates above problems by combining citation enhanced PLM with deep heterogeneous graph learning.
Table 5

Performance comparison between different methods on CORD-19.

MethodsMAPMRRP@25P@50P@75P@100R@25R@50R@75R@100
TF-IDF0.05410.06040.01910.01480.00620.00130.07190.10050.12510.1694
BM250.05620.06890.02240.01610.00740.00170.09430.11390.13540.1825
SciBERT0.07160.08450.03140.01930.01240.00480.11520.13820.15840.2136
BERT-DVGCN0.08260.09120.04080.02140.01470.00850.12820.14260.16710.2216
AR0.11070.11690.05940.03770.02720.01420.14420.16810.18690.2454
CO-Search0.12270.12890.06970.04930.04230.01970.15690.17270.20360.2563
CRB-HDGCN0.14950.15820.08480.06080.05720.02710.19530.22790.24830.3107
Table 6

Performance comparison between different methods on LitCovid.

MethodsMAPMRRP@25P@50P@75P@100R@25R@50R@75R@100
TF-IDF0.07050.07530.03340.02370.01020.00340.11530.14060.16900.1907
BM250.07230.08310.03650.02530.01790.00650.12680.15120.17840.2198
SciBERT0.09230.10050.05640.03010.02400.01850.17060.18800.20370.2610
BERT-DVGCN0.10810.11380.06260.04590.03070.02040.17920.20650.22040.2709
AR0.14920.16730.10510.07610.05290.03060.20540.22750.23730.3096
CO-Search0.15940.17170.11650.08520.06210.03780.21070.23880.24630.3159
CRB-HDGCN0.20570.21610.12950.09520.07840.04310.24090.27700.29510.3517
Performance comparison between different methods on CORD-19. Performance comparison between different methods on LitCovid.

Comparing with variants of our method

Our proposed CRB-HDGCN is consists of multiple sub-networks. In order to evaluate the effectiveness of these sub-networks, we conduct experiment with partial CRB-HDGCN on selected datasets. The compared variants of our method are summarized as follows. CRB: This variant only uses citation relational BERT part to obtain mean vectors for papers, authors and venues. BERT-HDGCN-PA: This variant uses vanilla BERT to obtain initial vectors, and uses HDGCN with paper and author nodes. BERT-HDGCN-PJ: This variant is similar with BERT-HDGCN-PA except that it uses paper and venue nodes in HDGCN. BERT-HDGCN: This variant extends BERT-HDGCN-PJ by adding author nodes. CRB-HDGCN-PA: This variant replaces BERT-HDGCN-PA with CRB. CRB-HDGCN-PJ: This variant replaces author nodes in CRB-HDGCN-PA by venue nodes. The experimental results are shown in Table 7, Table 8. We can see that CRB performs worse than any other variants because it only utilizes content information. Nevertheless, it can be seen that CRB still outperforms than SciBERT and BERT-DVGCN (in Table 5, Table 6). It is should be noticed that CRB does not using the whole COVID-19 literature dataset to heavily train BERT, but only chooses reference sentence relations to slightly retrain and fine-tune it. This reveals that it is necessary to use citation information to further retrain and fine-tune PLMs for citation recommendation task. As can be seen when introducing heterogeneous deep graph convolution, both BERT-HDGCN-PA and BERT-HDGCN-PJ perform better than CRB and BERT-DVGCN (in Table 5, Table 6). Particularly, we can see that BERT-HDGCN-PA is better than BERT-HDGCN-PJ, which shows that author information is more important than venue information. This can be ascribe into author usually has customized citing activity, for an author is tend to cite papers that he/she read or cite before, which result in the citing activities have high time continuity. Nevertheless, when using all types of nodes, BERT-HDGCN is able to automatically differentiate the importance of each node type and to generate more robust node vector representation, and consequently reaches better performance. We can also see that the performance of variants that using CRB is consistently outperform than that only using vanilla BERT.
Table 7

Performance comparison between different variants on CORD-19.

MethodsMAPMRRP@25P@50P@75P@100R@25R@50R@75R@100
CRB0.08420.09290.04150.02370.01510.00930.12930.14420.16890.2236
BERT-HDGCN-PA0.12240.13350.06530.03740.03490.01850.17020.19830.22630.2752
BERT-HDGCN-PJ0.11350.12670.06060.03440.03160.01690.16830.19010.21850.2703
BERT-HDGCN0.12650.13880.07170.04520.03920.02150.17620.20430.23080.2841
CRB-HDGCN-PA0.14520.15410.08030.05590.05430.02460.19140.22180.24110.2983
CRB-HDGCN-PJ0.13850.14790.07620.05120.05150.02260.18920.21760.23850.2925
CRB-HDGCN0.14950.15820.08480.06080.05720.02710.19530.22790.24830.3107
Table 8

Performance comparison between different variants on LitCovid.

MethodsMAPMRRP@25P@50P@75P@100R@25R@50R@75R@100
CRB0.14030.14920.08720.05930.03620.02610.18130.20960.22830.2856
BERT-HDGCN-PA0.17820.18520.11050.07620.06140.03350.22030.23730.25520.3169
BERT-HDGCN-PJ0.16460.17910.10250.07070.05850.03080.21020.22850.24970.3102
BERT-HDGCN0.18920.19150.11750.08120.06930.06410.22610.24960.26340.3253
CRB-HDGCN-PA0.19930.20930.12260.08850.07240.03820.23360.26980.28670.3428
CRB-HDGCN-PJ0.19250.20370.11760.08240.07000.03600.23010.26350.28390.3380
CRB-HDGCN0.20570.21610.12950.09520.07840.04310.24090.27700.29510.3517
Performance comparison between different variants on CORD-19. Performance comparison between different variants on LitCovid.

Efficiency evaluation

In this subsection, we conduct experiment on CORD-19 and LitCovid to evaluate the efficiency of our method. The compared methods are BERT-DVGCN and CO-Search. We first use 10% of samples in dataset to carry out training and testing process, and then incrementally increase the volume of samples to check the difference of running time and R@100 between compared methods. Because citation relational BERT part in our method can be considered as a pre-processing step for getting better PLMs model, we separately list the retraining time and fine-tune time of CRB in Table 9, and only present rest time of our method for the fairness of comparison.
Table 9

Retraining and fine-tune time of CRB.

DatasetRetrainingFine-tune
CORD-1910,401 s6,729 s
LitCovid6,158 s4,613 s
The comparison of efficiency is demonstrated in Table 10, Table 11. We can see that our CRB-HDGCN uses competitive time compared with BERT-DVGCN, but generates much better performance. There are two reasons for this result. The first reason is the convolutional part of CRB-HDGCN is similar with BERT-DVGCN, except adding more heterogeneous nodes which can be simultaneously executed by convolution process. The second reason is the added deep attention layer is essentially a shallow network compared with stacked deep convolution, which also brings little burden for training and testing. The training time of CO-Search is less than our CRB-HDGCN, for it only applied paragraphs and citations tuples to training Siamese-BERT. However, in testing stage, CO-Search executes a much more time consuming clusting and re-ranking process, which results in it spends much more time than CRB-HDGCN.
Table 10

Efficiency evaluation on CORD-19.

Methods10% samples
50% samples
100% samples
train timetest timeR@100train timetest timeR@100train timetest timeR@100
BERT-DVGCN14,052 s582 s0.093773,269 s2,353 s0.1547143,839 s4,523 s0.2216
CO-Search11,503 s904 s0.103756,731 s3,462 s0.1835116,493 s6,620 s0.2563
CRB-HDGCN14,937 s597 s0.135274,820 s2,495 s0.2235149,522 s4,782 s0.3107
Table 11

Efficiency evaluation on LitCovid.

Methods10% samples
50% samples
100% samples
train timetest timeR@100train timetest timeR@100train timetest timeR@100
BERT-DVGCN4,641 s117 s0.116321,659 s450 s0.170347,628 s953 s0.2709
CO-Search3,128 s182 s0.132816,510 s738 s0.209442,627 s1,773 s0.3159
CRB-HDGCN5,002 s128 s0.156123,569 s497 s0.255449,953 s1,084 s0.3517
Retraining and fine-tune time of CRB. Efficiency evaluation on CORD-19. Efficiency evaluation on LitCovid.

Parameter tuning

There are several manually set hyperparameters in our CRB-HDGCN, which will affect the performance of recommendation. In this subsection, we conduct experiment to investigate the optimal value of them. These essential hyperparameters are : , and . is the overall number of convolution layers. According previous study on Liu et al. (2020), we set the range of from 1 to 25. is the size of hidden vector dimension in heterogeneous deep graph convolutional neural network. Because determines the learning ability of HDGCN, we set the value of from 50 to 500. is the dimension of trainable project vectors in deep attention layer. Same as , the range of is set from 50 to 500. We only demonstrate the tuning results on R@100, for other metrics generate similar results. We first empirically fix , and evaluate the effect of with different . As can be seen in Fig. 5(a) and (b), when the layer of HDGCN is small, the performance of recommendation is not satisfied. As the number of layers increases, the performance gradually improves. When the number of layers reaches 15, the performance reaches to optimum. After that, when the number of layers increases, the performance drops slightly. We believe the reason behind this is that more graph convolution layers will let more long distant neighbors participate the learning process of node representation. In the scientific literature network, more long distant nodes will bring more research drift, which meantime bring more noise, and consequently resulting in performance degradation.It also can be seen that when dimension of is small (), the recommendation performances with different layers are relatively close. As increases, the performance gradually increases. When the value of reaches about 300, the performance gradually stabilize. Since a larger will lead to larger size of parameter scale and will increase training difficulty and time, the optimal value of is set to 300.
Fig. 5

The impact of hyperparameters with different convolution layers : (a) on CORD-19; (b) on LitCovid; (c) on CORD-19; (d) on LitCovid.

We then fix , and study the impact of with different . The results are demonstrated in Fig. 5(c) and (d). It can be seen that when the convolutional layer is small (), the change of has little impact on recommendation performance. This is due to limited long distance nodes are involved for node representation learning in this situation. Therefore, the deep attention layer in our model is hard to take full advantage. Nevertheless, as the number of layers increases, we can be seen that the performance curve rises steeply at the beginning increasing of . This is because in this situation, more long distance nodes are involved to participate in node representation learning, and the deep attention layer is able to effectively distinguish the importance of these nodes and greatly improving the recommendation performance. The performance reaches the best when . After that, the performance declines slightly with the increase of . This is because a large will bring more overfitting and decrease the performance. According to the above experimental results, the optimal value of is set to 200. The impact of hyperparameters with different convolution layers : (a) on CORD-19; (b) on LitCovid; (c) on CORD-19; (d) on LitCovid.

Vector visualization and case study

To vividly demonstrate the effectiveness of our method, we show the vector visualization of CRB in our method, and present some recommendation cases. We choose 6 hot keywords in COVID-19 area, including: “coronavirus”, “bacteria”, “antibodies”, “quarantine”, “vaccine”, “respiratory” and “dysgeusia”. Given a sentence, BERT will tokenize it into a word list by using tokenizer. Because the wordpiece vocabulary of BERT contains limited words, some COVID-19 related words will be tokenized into several sub-token. For example, “coronavirus”, “quarantine” and “dysgeusia” will be tokenized into [’corona’, ’##virus’], [’qu’, ’##aran’, ’##tine’] and [’d’, ’##ys’,’##ge’,’##usia’], respectively. In order to obtain reliable vectors of these words, we choose 1000 sentences for each words that contain it, and send these sentences to obtain output vector. The final word vector is calculated by the sum mean of all related sub-token in all chose sentences. After that, we compute the cosine similarity between these 6 keywords. We select different BERT model to generate results, including vanilla BERT, SciBERT, COVID-Twitter-BERT (Müller et al., 2020) and CRB. The visualization results are shown in Fig. 6. The diagonal is covered with white color for the cosine similarity with itself is equal to 1. We can see that vanilla BERT generate pretty low correlations between these words, which is because vanilla BERT is trained on common corpus and lacks deep understanding COVID-19 related content. SciBERT seems generates similar result. But SciBERT can still reveal correlation between some words. For example the cosine similarity between “bacteria” and “quarantine” generated by SciBERT is higher than vanilla BERT. This is because SciBERT is trained through scientific papers, which is helpful to discover more deep relation between text content. It can be seen that COVID-Twitter-BERT can excavate more correlations. For example, the cosine similarity between “coronavirus” and “quarantine” of COVID-Twitter-BERT are much higher than vanilla BERT and SciBERT. The reason is COVID-Twitter-BERT is trained on COVID-19 related twitters, which is benefit for COVID-Twitter-BERT to find co-occurrences of these two words. However, only using twitter data to train BERT may results in understanding drifting problem, which is a serious problem for COVID-19 scientific citation recommendation task. We can see that our CRB is able to reveal more relations between these words, which shows using citation link and text is help for PLM comprehending COVID-19 related content.
Fig. 6

The visualization of cosine similarity between words. The to represent “coronavirus”, “bacteria”, “antibodies”, “quarantine”, “vaccine”, “respiratory” and “dysgeusia”, respectively. The sub-graphs are the results by choosing different BERT model, including: (a) vanilla BERT, (b) SciBERT, (c) COVID-Twitter-BERT and (d) CRB.

To further show the effectiveness our method, we present some recommendation results in this subsection. We choose a query citation context from CORD-19 dataset. The paper ID is “PMC7096991” (Wack, Terczyńska-Dyla, & Hartmann, 2015) and the citing sentence is “This event creates docking sites for STAT1 and STAT2 signaling molecules, which leads to their recruitment and subsequent phosphorylation”. We also list the titles of three ground truth papers. The compared methods are using 5 and 15 layers of heterogeneous deep graph convolution, which are denoted as CRB-HDGCN-5 and CRB-HDGCN-15. As can be seen in Table 12, CRB-HDGCN-15 is able to correctly recommend all ground truths, while CRB-HDGCN-5 miss one ground truth. We can see that the correctly recommended results by CRB-HDGCN-5 are all published on same venue “Nature immunology”, which is also the publication venue of query paper. It is also interesting to further check the wrong results. We can see that CRB-HDGCN-5 still generate papers that published on the venue that very close to “Nature immunology”, while CRB-HDGCN-15 generate a much far wrong result. We checked the wrong result of CRB-HDGCN-15 and find it is the cited paper of ground truth 3, which means this wrong result may still have high possibility that cited by the query. The above findings indicate that CRB-HDGCN-5 can only find results that have more closer distance in paper graph. In contrast, CRB-HDGCN-15 is able to improve the importance of nodes that are far from the query, and consequently generate more effectiveness recommendation results.
Table 12

An example of 4 recommended citations by CRB-HDGCN-5 and CRB-HDGCN-15 for paper “PMC7096991”.

Ground truths:CRB-HDGCN-5CRB-HDGCN-15
Query: This event creates docking sites for STAT1 and STAT2 signaling molecules, which leads to their recruitment and subsequent phosphorylation. (venue: Nature Immunology)[O] result 1: IFN-λ s mediate antiviral protection through a distinct class II cytokine receptor complex. (venue: Nature Immunology)[O] result 1: IFN-λ s mediate antiviral protection through a distinct class II cytokine receptor complex. (venue: Nature Immunology)
ground truth 1: IFN-λ s mediate antiviral protection through a distinct class II cytokine receptor complex. (venue: Nature Immunology)[X] result 2: Diverse intracellular pathogens activate type III interferon expression from peroxisomes. (venue: Nature Immunology)[O] result 2: IL-28, IL-29 and their class II cytokine receptor IL-28R. (venue: Nature Immunology)
ground truth 2: IL-28, IL-29 and their class II cytokine receptor IL-28R. (venue: Nature Immunology)[O] result 3:IL-28, IL-29 and their class II cytokine receptor IL-28R. (venue: Nature Immunology)[X] result 3: Cloning of a new type II cytokine receptor activating signal transducer and activator of transcription (STAT) 1, STAT2 and STAT3. (venue: Biochemical Journal)
ground truth 3: Role of the interleukin (IL)-28 receptor tyrosine residues for antiviral and antiproliferative activity of IL-29/interferon-λ 1: similarities with type I interferon signaling. (venue: Journal of Biological Chemistry)[X] result 4: Mechanisms of type-I-and type-II-interferon-mediated signalling. (venue: Nature Reviews Immunology)[O] result 4: Role of the interleukin (IL)-28 receptor tyrosine residues for antiviral and antiproliferative activity of IL-29/interferon-λ 1: similarities with type I interferon signaling. (venue: Journal of Biological Chemistry)
An example of 4 recommended citations by CRB-HDGCN-5 and CRB-HDGCN-15 for paper “PMC7096991”. The visualization of cosine similarity between words. The to represent “coronavirus”, “bacteria”, “antibodies”, “quarantine”, “vaccine”, “respiratory” and “dysgeusia”, respectively. The sub-graphs are the results by choosing different BERT model, including: (a) vanilla BERT, (b) SciBERT, (c) COVID-Twitter-BERT and (d) CRB.

Conclusion

This paper presented an efficient and effective COVID-19 citation recommendation method CRB-HDGCN. The method first augments citing sentences among COVID-19 papers with the title of cited paper, and uses these augmented citing sentences to lightly retrain BERT model. Then we choose three types of sentence pairs between citing and cited paper to further fine-tune BERT, in order to let PLM better comprehend the citing activities among COVID-19 paper. Secondly, we develop a heterogeneous deep graph convolutional network (HDGCN) to learn vector representations for papers, authors and venues in COVID-19 literature dataset. The network uses an initial embedding layer to transfer the vector of all types of node into fixed length, and uses a heterogeneous graph convolutional layer to obtain hidden vectors for all nodes. Moreover, we add a deep attention layer to let the network is able to automatically differentiate the importance of nodes both according to types and distance, which is benefit for getting reliable vectors. The experimental results carried on CORD-19 and LitCovid datasets validated the effectiveness of our method. In the future, we will try to add more domain knowledge to help PLMs better understanding COVID-19 papers. We will also try to use more available fine-grained correlations among COVID-19 papers to further help graph neural network to get more robust vector representations.

CRediT authorship contribution statement

Tao Dai: Conceptualization, Methodology, Writing – original draft, Implementation of algorithm, Writing – review & editing. Jie Zhao: Conceptualization, Methodology, Writing – reviewing, Editing, Supervision. Dehong Li: Data curation, Visualization, Carrying out experiments. Shun Tian: Implementation of algorithm, Carrying out experiments. Xiangmo Zhao: Supervision. Shirui Pan: Validation, Investigation, Editing.

Declaration of Competing Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
  7 in total

1.  New tools aim to tame pandemic paper tsunami.

Authors:  Jeffrey Brainard
Journal:  Science       Date:  2020-05-29       Impact factor: 47.728

2.  LitMC-BERT: Transformer-Based Multi-Label Classification of Biomedical Literature With An Application on COVID-19 Literature Curation.

Authors:  Qingyu Chen; Jingcheng Du; Alexis Allot; Zhiyong Lu
Journal:  IEEE/ACM Trans Comput Biol Bioinform       Date:  2022-10-10       Impact factor: 3.702

Review 3.  Guarding the frontiers: the biology of type III interferons.

Authors:  Andreas Wack; Ewa Terczyńska-Dyla; Rune Hartmann
Journal:  Nat Immunol       Date:  2015-08       Impact factor: 25.606

4.  LitCovid: an open database of COVID-19 literature.

Authors:  Qingyu Chen; Alexis Allot; Zhiyong Lu
Journal:  Nucleic Acids Res       Date:  2020-11-09       Impact factor: 16.971

5.  COVID-19 information retrieval with deep-learning based semantic search, question answering, and abstractive summarization.

Authors:  Andre Esteva; Anuprit Kale; Romain Paulus; Kazuma Hashimoto; Wenpeng Yin; Dragomir Radev; Richard Socher
Journal:  NPJ Digit Med       Date:  2021-04-12

6.  An evaluation of two commercial deep learning-based information retrieval systems for COVID-19 literature.

Authors:  Sarvesh Soni; Kirk Roberts
Journal:  J Am Med Inform Assoc       Date:  2021-01-15       Impact factor: 4.497

7.  BioBERT: a pre-trained biomedical language representation model for biomedical text mining.

Authors:  Jinhyuk Lee; Wonjin Yoon; Sungdong Kim; Donghyeon Kim; Sunkyu Kim; Chan Ho So; Jaewoo Kang
Journal:  Bioinformatics       Date:  2020-02-15       Impact factor: 6.937

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.