Literature DB >> 31737809

Representing Multiword Chemical Terms through Phrase-Level Preprocessing and Word Embedding.

Liyuan Huang1, Chen Ling1.   

Abstract

In recent years, data-driven methods and artificial intelligence have been widely used in chemoinformatic and material informatics domains, for which the success is critically determined by the availability of training data with good quality and large quantity. A potential approach to break this bottleneck is by leveraging the chemical literature such as papers and patents as alternative data resources to high throughput experiments and simulation. Compared to other domains where natural language processing techniques have established successes, the chemical literature contains a large portion of phrases of multiple words that create additional challenges for accurate identification and representation. Here, we introduce a chemistry domain suitable approach to identify multiword chemical terms and train word representations at the phrase level. Through a series of special-designed experiments, we demonstrate that our multiword identifying and representing method effectively and accurately identifies multiword chemical terms from 119, 166 chemical patents and is more robust and precise to preserve the semantic meaning of chemical phrases compared to the conventional approach, which represents constituent single words first and combine them afterward. Because the accurate representation of chemical terms is the first and essential step to provide learning features for downstream natural language processing tasks, our results pave the road to utilize the large volume of chemical literature in future data-driven studies.
Copyright © 2019 American Chemical Society.

Entities:  

Year:  2019        PMID: 31737809      PMCID: PMC6854573          DOI: 10.1021/acsomega.9b02060

Source DB:  PubMed          Journal:  ACS Omega        ISSN: 2470-1343


Introduction

Artificial intelligence (AI) has demonstrated plenty of remarkable progresses in the recent years, from game playing strategies AlphaGo[1] to accurately classify skin cancer[2] and from image classification[3] to neural machine translation system.[4] Material science has also been greatly propelled by these data-driven-based AI technologies.[5] For example, machine learning and deep learning techniques were used for predicting target properties of chemicals,[6−14] in chemical discovery pipelines[15−17] and in reaction prediction.[18,19] Among all of these applications, the foundations of AI models and algorithms are to learn underlaying statistical relationships and patterns in the training data.[20] In material science, AI models can be trained on data generated through “high-throughput” simulation and/or experiments. For example, several publicly accessible repositories are available, such as Materials Project,[21] Open Quantum Materials Database (OQMD),[22] and AFLOW software framework,[23] to supply materials data from high-throughput density functional theory calculation in a large quantity and well-controlled quality. However, due to the computational cost, they are typically limited to a few fundamental properties such as formation energy and band gap values. Materials data can also be obtained by consolidating different handbooks. This way usually generates a set of a few hundred examples.[24] An alternative to acquire material data is to extract information from existing material literature such as patents and publications by text mining. For example, Kim et al. mined the synthesis parameters of oxide materials from 640 000 journal articles.[25] Based on the extracted information, they succeeded in predicting the important parameters required to inorganic materials leveraging machine learning techniques.[26−30] Elton et al. extracted information on the properties and functionalities of energetic materials based on a total of 3136 patents.[31] More recently, Tshitoyan et al. conducted an unsupervised word embedding based on a total of 1.5 million abstracts from materials science, physics, and chemistry publications to capture complex materials science concepts and successfully applied the learned knowledge to recommend new thermoelectric materials.[32] Before widely implemented in material science, text mining had been successfully and frequently employed in other domains, such as linguistic, biomedical, and marketing areas.[33−36] Compared to the general literature used by these studies, chemical literature is unique in terms of the abundance of chemistry-specific linguistic lexicons. Townsend et al. showed that at least 14 kinds of chemistry-specific lexicons were necessary to sufficiently describe chemical concepts in organic chemistry.[37] It becomes more concerning because the majority of these terms are composed of multiple words. For example, most chemical names are composed of more than one single word, such as “lithium chloride” and “lithium hydroxide monohydrate”. Reactions such as “double replacement” and “oxygen evolution reaction” and apparatus such as “graduated cylinders” are also examples of this type. The frequent appearance of these multiword phrases also creates great challenges to appropriately and accurately embed the semantic meanings in vectors, which serve as key features in a variety of downstream tasks, such as part-of-speech tagging,[38] sentiment analysis,[39] and text summarization.[40] Currently, the most frequently used technique for the phrase identification is named entity recognition, which leverages linguistic grammar-based techniques or statistical models,[41] while for the word representation, the conventional method includes two steps: first, represent each single constituent word independently and then combine them up to a multiword phrase representation afterward. For example, to represent the phrase “lithium chloride,” one needs to represent the individual component of “lithium” and “chloride” and takes the sum for the representation of the target. Here, we propose a new means for the multiword phrase representation of chemical terms. Our Multiword Identifying and Representing (MIR) method starts with recognizing the multiword phrases in the chemical literature by an unsupervised data-driven model and then represent the identified phrases as new words in the vocabulary at the phrase level. Using the same example of representing “lithium chloride”, the MIR method treats this phrase as a new, independent word different from “lithium” and “chloride”, then represents these three terms separately. Through a series of designed experiments, the results indicated that the current phrase identification model sufficiently achieves the goal of obtaining a better representation of words. In addition, we showed that the MIR method has less information loss compared to the conventional approach in representing the multiword phrases. Our results demonstrated that the MIR method has the capability to accurately represent chemical terms and pave the road to improve the effectiveness and performance of downstream machine learning algorithm-incorporated tasks, where the representations of chemical terms will be used as learning features.

Methodology

Overview

The overall workflow of our approach and conventional approach is presented in Figure . Besides demonstrating the MIR method, we constructed a pipeline as the baseline model by employing the conventional method of representing multiword chemical terms. Both the conventional method and MIR method started with the same data source and tokenization. In the conventional method, word embedding is performed right after tokenization, and the representation of a phrase is obtained through a post-vector addition. In the MIR method, a new step is incorporated to identify multiword phrases and add the detected terms to the vocabulary. The word embedding is performed afterward at the phrase level. Finally, a series of experiments are designed to evaluate the representing performances of the conventional method and MIR method.
Figure 1

Workflows of the convectional approach and the MIR approach to represent multiword phrases in the chemical literature.

Workflows of the convectional approach and the MIR approach to represent multiword phrases in the chemical literature.

Data Source

The training data is composed of patents downloaded from the United States Patent and Trademark Office (USPTO). To focus on the chemistry domain, we kept patents that contain keywords of “lithium” and “synthesis”. A total of 119 166 patents were preserved as the training corpora to the current study. For comparison, the patents extracted with a cooperative patent classification code under the “chemistry” section had a total 356 612 patents. Thus, the training corpora used in this work served as good representatives of the chemistry domain.

Tokenization

The raw patent files were processed using the following procedures. First, each patent was segmented into sentences using the function sent_tokenize() from the NLTK library.[42] After that, words were tokenized to keep necessary punctuation characters. The word tokens were then converted into lower case, and stop words were removed according to the stop words list from NLTK. We purposely kept those elemental names coincide with the stop words such as “I” as iodine, “At” as astatine, and “O” as oxygen. Finally, numeric values were removed from the tokens. We also removed ambiguous tokens such as “a.11.b.2.i” using several regular expression filters. The tokenization should preserve tokens that contain valid punctuation characters such as the chemical identifiers, IUPAC name, InChi key, and SMILES. Instead of simply chopping sentence on whitespace and removing all punctuation characters from word tokens, we used several punctuations as token delimiters, including the whitespace symbol, the forward slash symbol, the equal symbol, and the number symbol. After that, a subset of punctuation characters frequently appeared in chemical terms were put into a whitelist to allow valid tokens to pass through. In some special cases when punctuation characters were confusing with sentence boundaries such as “degree.” and “no.”, we trimmed them out of the tokens.

Multiword Phrase Detection

There are multiple systems to identify multiword chemical phrases from chemistry literature, which can be categorized as dictionary-based, rule-based, supervised machine learning-based, unsupervised machine learning-based, and hybrid systems.[41,43] Dictionary-based and rule-based systems require manually preparation based on public resources. Supervised machine learning-based systems require feature-based representation of sample data under human preparation and supervision. We chose an unsupervised statistical model for its capability to leverage statistical information and linguistic features and do the training without human annotation or supervision. In the unsupervised method, many measurements have been previously created to identify phrases in the text, such as mutual information (MI), point mutual information (PMI), and the normalized version of them.[44] These methods, however, have a common drawback of heavy computing requirement. We adopted a data-driven phrase identification function proposed by Mikolov et al.[45] Compared to other models, this data-driven method is able to identify not only chemical name phrases (e.g., compound names) but also some other kinds of phrases such as operation names, brand names, equipment names, etc. Meanwhile, it has an acceptable phrase identification rate, which brings additional advantages in terms of computational efficiency. The multiword identification function started from tokenized and trimmed single words (unigrams) in the sentence context obtained from the previous tokenization step. A scoring function from a third-party unsupervised semantic modeling python library Gensim[46] was implemented to identify reasonable phrases (bigrams) formed by two co-occur unigrams. The bigram scoring function is defined aswhere unigram_a_count and unigram_b_count are numbers of occurrences of the first and second component words. bigram_count is the number of co-occurrences of the first and the second component words that are composing a phrase. len_vocab is the size of the vocabulary. min_count is the minimum collocation count threshold of a bigram among all patents. If S passes the predefined threshold value of t, the bigram is determined as a valid vocabulary phrase and added to the vocabulary. Phrases consisting of more than two words such as three-word phrases (trigrams) and four-word phrases (quadrigrams) are aggregated from one unigram one bigram and two bigrams, respectively. In principle, repeating this process could identify multiple-word phrases composed of n single words (n-grams, n is up to infinity), although it would be a memory-intensive task. As suggested by Mikolov et al.,[45] phrases composed of two to four words are a typical selected range for the training section. We, therefore, limited the detection to phrases containing no more than four words.

Word Embedding

After establishing the training corpus either directly from tokenization as the conventional approach does or after adding identified phrases to the vocabulary as the MIR approach does, we trained the word embedding model to map every word w from the vocabulary V to a numerical vector representation as for all w in V, where D is the dimensionality of representation . We used the Word2Vec model for the word embedding.[45,47] The original paper of Word2Vec model reported two architectures of training, the continuous bag-of-word (CBOW) and skip-gram (SG). Compared to other word embedding techniques, the CBOW architecture is more robust and better performed in relatedness and analogy tasks.[48] Continuous bag-of-words (CBOW) structures were employed for all of the representation works in entire study. The CBOW model is trained by maximizing the probability of the center word being wc (“target” word to predict) if all of the surrounding words wc–, ..., wc–1, wc+1, ..., wc+ (training words) are given in terms of a softmax function, where m is the distance between center word wc and the farthest context wordwhere score (wc|wc–,...,wc–1,wc+1,...,wc+) is the compatibility of word wc with surrounding words. A dot product is often used here as the score function. The objective function of this model is to minimize its negative log-likelihood on the training datasetBecause of the large size of training vocabulary V, the sum operation in the denominator of the softmax function becomes very expensive. Two replacing techniques are introduced to improve the efficiency: hierarchical softmax and negative sampling.[45] In addition, we used the subsampling technique to balance between frequently appearing and rarely appearing words. The implementation of word embedding used the Python library Gensim.

Post-Vector Addition

In the MIR approach, the word embedding directly outputs the representations for multiword phrases p for downstream tasks. However, in the conventional approach, the word embedding only outputs the representation for a single word. To represent a multiword phrase, a common way is to use the arithmetic aggregation of individual representations[45,49]where c is a vector summed from n component word vectors and is the representation for each component word.

Evaluation

To assess the performance of the representation of c from the conventional approach and the representation of p from the MIR approach, we created a total of six datasets for a series of evaluations. The first four datasets contain names of chemical compounds extracted from PubChem, a database of chemical molecules and their activities against biological assays maintained by the National Center for Biotechnology Information.[50] Since we filtered the patent sources by only keeping those containing keywords of “lithium” and “synthesis”, the first dataset (D1) was designed to contain the names of 50 lithium organic and inorganic compounds. The second (D2), third (D3), and fourth(D4) dataset contained the names of inorganic compounds formed by two word, three word, and four word, respectively, with those overlapping with D1 removed. To evaluate the performance of representing lithium-related and non-lithium-related chemical phrase terms, the fifth (D5) and sixth (D6) dataset were manually crafted by pulling phrases from the Wikipedia pages about “lithium battery” and “unit operations & separation process”. These crafted datasets were continuously used throughout our study, and the full lists of these datasets are provided in the Supporting Information.

Results

Phrase Detection

First, we evaluated the capability of the MIR approach to correctly identify the chemical phrases. As shown in the Methodology section, a joint word is positively identified as a phrase if the score from eq exceeds a manually defined threshold t. In practice, it is recommended to use a smaller threshold for phrases with longer sizes.[45] However, to the best of our knowledge, there is no established rule on setting the threshold value to maximize the identification rate while keeping reasonable computational cost. We, therefore, used different t values in the current study. For each t value, we calculated the rate of successfully detecting the phrases in the evaluation datasets, as reported in Figure . For the min_count parameter, we used 30 in our study to keep a sufficient amount of bigram phrase candidates while maintaining the computational cost at an affordable level.
Figure 2

Rate to effectively identify phrases in the evaluation datasets. The scalers of x-axis were normalized comparing to a total of 1 313 918 single vocabulary words count, which is the largest size for a single word vocabulary in this study. The numbers between x-axis and the lowest curve represent the different testing threshold configurations of the models.

Rate to effectively identify phrases in the evaluation datasets. The scalers of x-axis were normalized comparing to a total of 1 313 918 single vocabulary words count, which is the largest size for a single word vocabulary in this study. The numbers between x-axis and the lowest curve represent the different testing threshold configurations of the models. From the results, we verified that the MIR approach detects most of the bigrams in the evaluation datasets. All of the models with t ≤ 1.25 achieve an identification rate higher than 50% on D1 and D2, and the highest identification rate 86% is scored on D5. For trigrams (D3) and quadrigrams (D4), all of the models achieve lower detection rates no matter how their threshold values were set, because the trigram and quadrigram chemical compound phrases appear less frequently than unigram or bigram compound names, hence making the denominator of the scoring function much larger than the enumerator. Nonetheless, the detection rate for these two datasets is still higher than 30%, indicating a decent success to find these high order phrases. An important observation from Figure is that the detection rate increased with the number of phrases in the vocabulary. This trend is not surprising, since the more phrases we detected, the more chance we find the correct one. In an extreme scenario, assuming that we retain every multiword phrase as long as it appears in the source text, we should achieve the perfect detection rate while also obtaining the largest possible vocabulary. A large vocabulary means more computational demanding for the word embedding as well as other downstream tasks. Thus, it is important to balance the tradeoff between phrase detection rate and the total amount of phrases by appropriately tuning the threshold of t. From Figure , we see that after a rapid increase for larger t values, the detection rate was nearly plateaued starting from t = 0.75 to 0.25. We used t = 0.75 for the rest of our studies. Additionally, the main results of our study were also verified using t = 0.25.

Chemical Name and Formula Relatedness Scoring

At this point, a new means to represent phrases from the chemistry-domain had been developed. We trained word representations through this new means. Better representations of phrases further contribute to the whole semantic vector space, which not only contains multiword phrases but also single-word vocabulary. We hypothesized that the connections between multiword phrases and single words are expected to become closer and more obvious in the MIR approach. Therefore, a lot of applications can benefit from this new characteristic of the vector space. To verify our hypothesis, a series of experiments was designed to evaluate the representations generated through conventional approach and MIR approach. The existing schemes to evaluate vector representations trained on the general domain of textual resources such as Wikipedia and Google News can be split into intrinsic and extrinsic evaluation in general.[48] In intrinsic evaluation, word embeddings are evaluated by measuring performance among themselves in relatedness, analogy, categorization, and selectional preference tasks. In extrinsic evaluation, vector representations are used as input features to downstream tasks and their performance is evaluated based on these tasks. In the current study, we performed the intrinsic evaluation of the possible enhancements of the vector representations brought by the MIR method. First, we consider the task of measuring the similarity between the name of a compound and its corresponding formula. This task is designed to examine whether the representations embed the precise semantic relationships or distances among chemical terms. For example, a two-word phrase of “lithium chloride” contains the same chemical meaning as a single word of “LiCl”. A good representation should, thus, represent the vector for “lithium chloride” similar to that of “LiCl”. This relatedness can be quantitively assessed using the cosine similarity function Datasets used in this experiment were obtained from D1, D2, and D3. We chose the ones with both chemical name and corresponding formula included in the vocabularies of two approaches. As shown in Figure , the relatedness score for the c representation varies between 0.4 and 0.5. By using the MIR approach, it is increased to around 0.6, suggesting that the p representation captures the semantic relationship between the chemical name and formula more accurately than c. The highest similarity score was obtained for D1 at 0.65.
Figure 3

Relatedness scores on compound names and formulas as measured by cosine similarity.

Relatedness scores on compound names and formulas as measured by cosine similarity.

Chemical Name and Formula Inferring

While the previous experiment directly examined the similarities between a given pair of chemical name and formulas, we designed another experiment to examine if the corresponding formulas can be correctly inferred by searching the geometric neighbors of chemical names. For example, providing the name of “lithium chloride”, the task is to look up the neighbor words ranging from 0 to vocabulary size V to yield the correct answer “LiCl”. In principle, a good representation should rank semantically similar words to a target word higher, while rank less similar words lower, thus enabling the discovery of the correct formula in the vicinity of the corresponding chemical name. We considered two ways to define the range of neighborhood. In the first run, we looked for the correct chemical formula within the first 20, 50, and 100 neighbors of a chemical name in two vector spaces. It, therefore, evaluated whether the formula is correctly located in a given number of neighbors of the target chemical name without considering the influence of local word density. For a dense population around a given word, the target word may remain in the vicinity but fall well outside of the detection range. Thus, in the second test, we looked for the formula within a given radius of similarity to the given chemical name. Results for this experiment are presented in Figure . We see that the c representation performed poorly on this inferring task, especially for the compound names in D1. On the contrary, for the p representation. the inferring task achieved a much better success rate with large portions of examples from D1 and D2 being correctly inferred. These two experiments, thus, concluded that semantic relationships between two similar/identical terms are better captured and presented by p than c.
Figure 4

Inferring results on compound names and formulas. (a) Searching formulas within a given number of nearest neighbors. (b) Searching formulas within a given range of cosine similarities.

Inferring results on compound names and formulas. (a) Searching formulas within a given number of nearest neighbors. (b) Searching formulas within a given range of cosine similarities.

Synonyms, Acronyms, and Abbreviations Finding

Other ways to verify our hypothesis are the synonym finding and acronym/abbreviation finding tasks. In previous sections, we demonstrated how MIR contributes to finding the formula of a given chemical name. Here, we show that the same procedure can be applied here to find synonyms and acronyms/abbreviations based on the representations generated through MIR. For synonym finding, we used p to generate the nearest neighbors for 30 keywords (vocabulary-included) randomly selected from D1, D5, and D6 and checked if the neighbors included the correct synonyms (see full results in the Supporting Information). Table shows several examples from this experiment.
Table 1

Partial Synonyms Candidates Found through p Evaluation Results

target words/phrasestop 10 nearest neighbors (from high to low ranking)
portable_electronicselectric_vehicles; consumer_electronics; portable_electronic_devices; power_tools; electric_cars; uninterruptible_power_supplies; electrical_vehicles; laptop_computers; cell_phones; portable_devices
lithium_platingsei_layer; capacity_loss; dendrite_growth; self-discharge; battery_discharge; lithium_anode; overcharge; electrolyte_decomposition; charge_discharge_cycles; deep_discharge
heat_exchangeheat_exchanger; heat_exchangers; heat_transfer; indirect_heat_exchange; boiler; cooling_medium; steam_generator; heat_removal; compressor; cooler
lithium_aluminium_hydridelithium_aluminum_hydride; lithium_borohydride; borane-tetrahydrofuran_complex; lithium_aluminiumhydride; borane-thf_complex; sodium_borohydride; diisobutylaluminium_hydride; borane_tetrahydrofuran_complex; lialh; lialh4
lithium-air_batteryair_battery; lithium_battery; lithium_ion_battery; metal-air_battery; electrochemical_device; lithium-ion_battery; lithium-sulfur_battery; lithium_metal_battery; lithium_ion; li-ion_battery
It is apparent that synonyms were found for all of the words evaluated. Furthermore, not only the synonyms were found but also some related concepts or representatives were found such as “power tools,” “laptop computer,” and “cell phone” were found for word “portable electronics.” A core component of “lithium plating” was found, which is “sei layer.” These findings indicate that it is feasible to find synonyms through our MIR method only based on the similarities. For the acronym/abbreviation finding, we first extracted 40 vocabulary-included acronyms/abbreviations of chemical terms from a reference book of chemistry visualization.[51] Out of these testing terms, we found that within the range of 250 nearest neighbors, the correct acronyms/abbreviations were detected for 21 terms, as listed in Table . This decent success rate indicates that correct acronyms/abbreviations for a given term can be identified solely utilizing the similarity between the representations. It is, therefore, anticipated that better success rate can be achieved by combining the MIR representation with other established methods such as monolingual syntax-based methods or statistical information-based methods for synonym finding[52−54] and string matching or natural language processing for acronyms/abbreviations finding.[55]
Table 2

Correct Acronyms/Abbreviations Found through p Evaluation Results

definition wordsfound acronyms/abbreviationsranking of acronymscosine similarity
atomic_force_microscopyafm20.8087
charge-coupled_deviceccd20.7761
cetyltrimethylammonium_bromidectab460.6714
fourier_transform_infrared_spectroscopyftir40.7988
green_fluorescent_proteingfp80.8007
graphical_user_interfacegui70.7328
high-performance_liquid_chromatographyhplc190.7203
infraredir560.6272
low-density_lipoproteinldl350.7178
magnetic_resonance_imagingmri10.8624
near_infrarednir100.7870
nuclear_magnetic_resonancenmr1580.6536
protein_data_bankpdb20.7800
red_fluorescent_proteinrfp790.6382
scanning_electron_microscopysem250.7674
second_harmonic_generationshg10.7791
secondary_ion_mass_spectrometrysims10.8549
scanning_probe_microscopyspm2340.4587
transmission_electron_microscopetem100.8261
terahertzthz350.6882
ultravioletuv40.8107

Chemical Terms Clustering

We further evaluated if the representation of p preserved the semantic relation between different phrases. For this purpose, we designed the experiment to examine the capability of recovering a clustering of labeled keywords into separate groups. After the phrase identification procedure, the model with threshold value t set to 0.75 identified 25, 617, 41, 4, 30, 26 phrases out of D1, D2, D3, D4, D5, and D6, respectively. To balance between separate groups, we sampled 24 identified phrases each from D1, D5, and D6. For the easiness of implementation and the capability of handling different kinds of similarities or distances, Hierarchical agglomerative clustering (HAC) was selected and performed to create a hierarchy of these keywords clusters according to the Euclidean distance metric (ward linkage), which was implemented through python Scikit-learn library.[56] Euclidean similarity measures the magnitude (frequency) over the direction (semantic) of word vectors, and it can maximize the intercluster distance comparing to cosine similarity.[48,57] Agglomerative (bottom-up) algorithm treats each document as a singleton cluster at the beginning, then successively combines nearby pair of document clusters into one cluster. Eventually, all clusters will be merged into one big cluster, which contains all of the documents. To visualize the clustering results, we plotted part of the dataset phrases in the principle component space, as shown in Figure . We find that the p representation impressively recovered the correct clustering. Only the phrase of “lithium salt” and “charging speed” were clustered to the groups of compound names and operation, respectively. On the other side, the c representation had much more phrases that were mistakenly grouped and a closer examination suggested that in the c representation, the phrases from the compound name group and “lithium battery”-related group are difficult to distinguish from each other.
Figure 5

Visualization of clustering results on part of the samples. Only first two principal components of the representations were used to build the graph. Color graded represents clusters predicted by the hierarchical agglomerative algorithm. Black arrow and empty symbols represent wrongly clustered samples compared to the ground truth. (a) Partial of clustering results based on c representation. (b) Partial of clustering based on p representation. (c) Names of the examined phrases. The samples from D1, D5, and D6 datasets are represented as uppercase letters, lowercase letters, and numbers, respectively.

Visualization of clustering results on part of the samples. Only first two principal components of the representations were used to build the graph. Color graded represents clusters predicted by the hierarchical agglomerative algorithm. Black arrow and empty symbols represent wrongly clustered samples compared to the ground truth. (a) Partial of clustering results based on c representation. (b) Partial of clustering based on p representation. (c) Names of the examined phrases. The samples from D1, D5, and D6 datasets are represented as uppercase letters, lowercase letters, and numbers, respectively. Quantitatively, the clustering results were evaluated by external evaluation metrics, including the mutual information-based score (MI),[58] adjusted rand index (ARI),[59] homogeneity, completeness, V-measure[60] and Fowlkes-mallows index (FMI).[61] As the scoring metrics in Table indicate, the clustering performance based on representation p greatly surpasses the clustering performance based on representation c. The external clustering measurement scores in other domains indicated that these metrics are normally ranging from 0 to 0.8, and the best clustered result can approach to 0.9.[62−64] Although it is not fair to compare our result with results from these works, which were measured based on completely different and integral datasets, our clustering result can still suggest that p represent the semantic meanings among groups of phrases more accurately than conventional ones.
Table 3

External Clustering Evaluation Results of Hierarchical Agglomerative Clustering

 MIARIhomogeneitycompletenessV-measureFMI
Wc0.59480.56750.60550.67020.63620.7230
Wp0.89100.91700.89380.89530.89460.9439

Discussion

The evaluation results prove the capability of the MIR approach on effectively and accurately identifying and representing several types of multiword chemical phrases. Comparing to the conventional approach, less information loss is achieved in the MIR approach. We analyzed why p performed better than c representation on the evaluation tasks. In c, the representation of a phrase is constructed through individual components, each of which is trained on the single-word level from its own context. For example, to obtain the representation of “lithium battery”, we need to represent “lithium” by learning its co-occurrence with “chloride” when mentioned as a chemical name in the text resources, with “silvery-white” when describing the physical occurrence and with “battery” when describing in a specific application. Apparently, the information from the first two context serves as unnecessary noise in deciphering the actual meaning of “lithium” in “lithium battery”. Same situation happens when representing “battery”. On the contrary, in the p representation, the phrase of “lithium battery” is treated as an independent word that differs with “lithium”, “lithium chloride”, and others, thus minimizes the influences from noisy background contexts. Another information loss in the c occurs in summing the representations of individual words with equal weight. For example, we should anticipate that the phrase of “lithium battery”, which describes a specific type of battery, to be more similar to “battery” than “lithium”. In other words, the representation of “battery” should have a higher weight than “lithium” in “lithium battery”. On the contrary, equally weighting “lithium” and “battery” for the representation of “lithium battery” suggests that “lithium battery” is almost equally similar to “battery” and “lithium” (Figure a). By treating “lithium battery” as an independent phrase and training the embedding at the phrase level, the preservation of semantic meaning is improved. As a result, the representation of “lithium battery” is closer to “battery” than to “lithium” in the p representation (Figure b).
Figure 6

Illustration of relationships between c, p, and their constituent word vectors in two vector spaces. Cosine similarities are measured from real models. (a) In c representation, “lithium battery” is almost equally similar to “battery” and “lithium”. (b) In p representation, “lithium battery” is more similar to “battery” than to “lithium”.

Illustration of relationships between c, p, and their constituent word vectors in two vector spaces. Cosine similarities are measured from real models. (a) In c representation, “lithium battery” is almost equally similar to “battery” and “lithium”. (b) In p representation, “lithium battery” is more similar to “battery” than to “lithium”. As we have shown in Section 3.1, a critical factor in our MIR method is the occurrences from eq that determines the identification of a phrase in the text resource. One restriction on the occurrences is the threshold of the minimum appearance of a phrase (min_count). For a phrase with a number of appearances less than the min_count, eq yields a negative score and hence directly rejects to add it to the vocabulary. However, there are certain circumstances that chemical terms only appear a few times in the text resources but should be considered as a phrase, especially for some compound names such as “lithium magnesium sodium silicate”. This limitation of phrase detection function can be alleviated by introducing chemical databases and third-party tools for looking up. We anticipate the dictionary-based, rule-based, or hybrid methods in cooperation with the current proposed MIR approach could result in a more accurate detection of chemical phrases without significantly sacrificing computing resources. There are still several limitations of our work. One of them is the noise phrase generated accompanying identifying real chemical multiword phrases. Although improvements to word representations achieved by moving the phrases identification procedure forward had already been verified under the current model settings, it is reasonable to infer that the performance will become better if more noise phrases were removed from the current context. As mentioned before, some additional approaches, such as dictionary-lookup and rule-based parsing, can be incorporated with MIR in the future to help create a “clean” context to get better word embeddings. Another limitation of our current study is the lack of extrinsic evaluations. One purpose of word representation is to create a numerical vector space for the downstream machine learning tasks. From this point of view, the evaluation of the performance of representation should also be thoroughly investigated in these downstream jobs. While the current study has extensively evaluated the sematic relatedness in the MIR approach, it is of great interest to see how the MIR representation performs as input in future downstream machine learning models.

Conclusions

To summary, the current work introduces the multiword identifying and representing method, a fine-tuned workflow to detect multiword chemical terms in the chemical literature. The MIR starts with identifying multiword phrases, adding it to the vocabulary, and performing the word embedding at the phrase level. Through a series of special-designed experiments, we demonstrated that the MIR approach effectively and accurately identifies multiword chemical terms from chemical patents. Compared to the conventional approach representing constituent single words first and combining them afterward, the MIR method is more robust and accurate to preserve the semantic meaning of multiword chemical phrases. The accurate representation of chemical terms is the first and essential step to provide learning features for downstream natural language processing tasks. The MIR method, thus, paves the road to utilize a large volume of chemical literature in future data-driven studies.
  16 in total

1.  A simple algorithm for identifying abbreviation definitions in biomedical text.

Authors:  Ariel S Schwartz; Marti A Hearst
Journal:  Pac Symp Biocomput       Date:  2003

2.  Machine-Learning-Augmented Chemisorption Model for CO2 Electroreduction Catalyst Screening.

Authors:  Xianfeng Ma; Zheng Li; Luke E K Achenie; Hongliang Xin
Journal:  J Phys Chem Lett       Date:  2015-08-27       Impact factor: 6.475

3.  Unsupervised word embeddings capture latent knowledge from materials science literature.

Authors:  Vahe Tshitoyan; John Dagdelen; Leigh Weston; Alexander Dunn; Ziqin Rong; Olga Kononova; Kristin A Persson; Gerbrand Ceder; Anubhav Jain
Journal:  Nature       Date:  2019-07-03       Impact factor: 49.962

4.  "Found in Translation": predicting outcomes of complex organic chemistry reactions using neural sequence-to-sequence models.

Authors:  Philippe Schwaller; Théophile Gaudin; Dávid Lányi; Costas Bekas; Teodoro Laino
Journal:  Chem Sci       Date:  2018-06-22       Impact factor: 9.825

5.  Mastering the game of Go without human knowledge.

Authors:  David Silver; Julian Schrittwieser; Karen Simonyan; Ioannis Antonoglou; Aja Huang; Arthur Guez; Thomas Hubert; Lucas Baker; Matthew Lai; Adrian Bolton; Yutian Chen; Timothy Lillicrap; Fan Hui; Laurent Sifre; George van den Driessche; Thore Graepel; Demis Hassabis
Journal:  Nature       Date:  2017-10-18       Impact factor: 49.962

6.  Dermatologist-level classification of skin cancer with deep neural networks.

Authors:  Andre Esteva; Brett Kuprel; Roberto A Novoa; Justin Ko; Susan M Swetter; Helen M Blau; Sebastian Thrun
Journal:  Nature       Date:  2017-01-25       Impact factor: 49.962

7.  ElemNet: Deep Learning the Chemistry of Materials From Only Elemental Composition.

Authors:  Dipendra Jha; Logan Ward; Arindam Paul; Wei-Keng Liao; Alok Choudhary; Chris Wolverton; Ankit Agrawal
Journal:  Sci Rep       Date:  2018-12-04       Impact factor: 4.379

8.  A Machine Learning Approach to Zeolite Synthesis Enabled by Automatic Literature Data Extraction.

Authors:  Zach Jensen; Edward Kim; Soonhyoung Kwon; Terry Z H Gani; Yuriy Román-Leshkov; Manuel Moliner; Avelino Corma; Elsa Olivetti
Journal:  ACS Cent Sci       Date:  2019-04-19       Impact factor: 14.553

9.  Machine-learned and codified synthesis parameters of oxide materials.

Authors:  Edward Kim; Kevin Huang; Alex Tomala; Sara Matthews; Emma Strubell; Adam Saunders; Andrew McCallum; Elsa Olivetti
Journal:  Sci Data       Date:  2017-09-12       Impact factor: 6.444

10.  PubChem 2019 update: improved access to chemical data.

Authors:  Sunghwan Kim; Jie Chen; Tiejun Cheng; Asta Gindulyte; Jia He; Siqian He; Qingliang Li; Benjamin A Shoemaker; Paul A Thiessen; Bo Yu; Leonid Zaslavsky; Jian Zhang; Evan E Bolton
Journal:  Nucleic Acids Res       Date:  2019-01-08       Impact factor: 16.971

View more
  1 in total

Review 1.  Opportunities and challenges of text mining in aterials research.

Authors:  Olga Kononova; Tanjin He; Haoyan Huo; Amalie Trewartha; Elsa A Olivetti; Gerbrand Ceder
Journal:  iScience       Date:  2021-02-06
  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.