| Literature DB >> 33072184 |
Yuzhuo Wang1, Chengzhi Zhang1.
Abstract
In the era of big data, the advancement, improvement, and application of algorithms in academic research have played an important role in promoting the development of different disciplines. Academic papers in various disciplines, especially computer science, contain a large number of algorithms. Identifying the algorithms from the full-text content of papers can determine popular or classical algorithms in a specific field and help scholars gain a comprehensive understanding of the algorithms and even the field. To this end, this article takes the field of natural language processing (NLP) as an example and identifies algorithms from academic papers in the field. A dictionary of algorithms is constructed by manually annotating the contents of papers, and sentences containing algorithms in the dictionary are extracted through dictionary-based matching. The number of articles mentioning an algorithm is used as an indicator to analyze the influence of that algorithm. Our results reveal the algorithm with the highest influence in NLP papers and show that classification algorithms represent the largest proportion among the high-impact algorithms. In addition, the evolution of the influence of algorithms reflects the changes in research tasks and topics in the field, and the changes in the influence of different algorithms show different trends. As a preliminary exploration, this paper conducts an analysis of the impact of algorithms mentioned in the academic text, and the results can be used as training data for the automatic extraction of large-scale algorithms in the future. The methodology in this paper is domain-independent and can be applied to other domains.Entities:
Keywords: Algorithm entity; Full-text content; Influence of algorithms
Year: 2020 PMID: 33072184 PMCID: PMC7548120 DOI: 10.1016/j.joi.2020.101091
Source DB: PubMed Journal: J Informetr ISSN: 1751-1577 Impact factor: 5.107
Fig. 1Framework of our work.
Fig. 2The number of papers accepted by ACL conference each year.
Example of algorithms extracted from articles.
| ID | Title | Alg. 1 | Alg. 2 | Alg. 3 | Alg. 4 | Alg. 5 | Alg. 6 |
|---|---|---|---|---|---|---|---|
| P01-1049 | Building Semantic Perceptron Net for Topic Spotting | Naïve Bayes | KNN | SVM | NNet | LSF | BP algorithm |
Examples about different names of algorithms.
| Standard Name | Abbreviations and Aliases |
|---|---|
| Support Vector Machine | svm, svms, support vector machines, support-vector machines, support-vector machine |
| Conditional Random Field | crf, crfs, conditional random fields |
| Maximum Entropy | me, maxent, max-ent, maximum-entropy |
| Naive Bayes | nb, naïve bayesian, naivebayes |
| Expectation Maximization | em, em-algorithm, expectation-maximization, expectation and maximization, expectation-maximisation, expectation and maximisation, expectationmaximization |
| K Nearest Neighbor | knn, k-nn, k-nearest neighbor, k-nearest-neighbor, k-nearest neighbors, k-nearest-neighbors, k-nearest neighbour, k-nearest-neighbour, k-nearest neighbours, k-nearest-neighbours, k nearest-neighbor, k nearest neighbors, k nearest-neighbors, k nearest neighbour, k nearest-neighbour, k nearest neighbours, k nearest-neighbours |
| Context Free Grammar | cfg, cfgs, context free grammars, context-free grammar, context-free grammars, contextfree grammar, contextfree grammars |
Example of sentence matched by the algorithm.
| ID | Algorithm | Full name | Matched Sentence |
|---|---|---|---|
| P01-1049 | SVM | Support vector machine | A large number of techniques have been proposed to tackle the problem, including regression model, nearest neighbor classification, Bayesian probabilistic model, decision tree, inductive rule learning, neural network, on-line learning, and, SVM. |
The top-10 most influential algorithms in ACL conference papers.
| Rank | Algorithm | Rank | Algorithm |
|---|---|---|---|
| 1 | Support vector machine | 2 | Context-free grammar |
| 3 | BLEU algorithm | 4 | Maximum entropy |
| 5 | Word2vec | 6 | Expectation maximization |
| 7 | Hidden markov model | 8 | Dynamic programming algorithm |
| 9 | Decision-tree | 10 | Cosine similarity |
The type of high-impact algorithms.
| Rank | Type | Number of algorithms | Average influence |
|---|---|---|---|
| 1 | Classification algorithm | 15 | 0.0448 |
| 2 | Probabilistic graphical models | 7 | 0.0410 |
| 3 | Grammar | 14 | 0.0400 |
| 4 | Optimization algorithm | 16 | 0.0382 |
| 5 | Neural networks | 10 | 0.0364 |
| 6 | Ensemble learning algorithm | 2 | 0.0354 |
| 7 | Metric algorithm | 11 | 0.0342 |
| 8 | Unique algorithms in NLP domain | 4 | 0.0327 |
| 9 | Regression algorithm | 2 | 0.0264 |
| 10 | Search algorithm | 3 | 0.0240 |
| 11 | Dimension reduction algorithm | 1 | 0.0233 |
| 12 | Other | 9 | 0.0233 |
| 13 | Link analysis algorithms | 2 | 0.0222 |
| 14 | Clustering algorithm | 4 | 0.0146 |
Top-10 algorithms with higher influence in each year.
| Year | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | 10 |
|---|---|---|---|---|---|---|---|---|---|---|
| 1979 | Augmented transition network | Context free grammars | Hobbs algorithm | Kernel methods | Case frame | Local maxima | Discrimination network | * | * | * |
| 1980 | Augmented transition network | Hobbs algorithm | Context free grammars | Semantic grammars | Case frame | Fuzzy matching | Higher order unification | Structure grammars | Cocke younger kasami | Plan recognition |
| 1981 | Augmented transition network | Structure grammars | Lexical functional grammars | Left corner | Hobbs algorithm | Context free grammars | Decision tree | Semantic grammars | Case frame | Cocke younger kasami |
| 1982 | Structure grammars | Augmented transition network | Context free grammars | Semantic grammars | Plan recognition | Classification and regression tree | Dynamic programming | Hyperlink induced topic search | Generalized phrase structure grammars | Lexical functional grammars |
| 1983 | Context free grammars | Augmented transition network | Lexical functional grammars | Structure grammars | Tree adjoining grammars | Definite clause grammars | Semantic grammars | Earley algorithm | Unification grammars | Hobbs algorithm |
| 1984 | Augmented transition network | Context free grammars | Lexical functional grammars | Structure grammars | Case frame | Definite clause grammars | Unification grammars | Semantic grammars | Dependency grammars | Functional unification grammars |
| 1985 | Context free grammars | Lexical functional grammars | Structure grammars | Definite clause grammars | Earley algorithm | Head grammars | Plan recognition | Left corner | Head drive phrase structure grammar | Bag of words |
| 1986 | Context free grammars | Tree adjoining grammars | Generalized phrase structure grammars | Lexical functional grammars | Augmented transition network | Kimmo generation | Definite clause grammars | Head grammars | Earley algorithm | Plan recognition |
| 1987 | Context free grammars | Categorial grammars | Earley algorithm | Augmented transition network | Definite clause grammars | Tree adjoining grammars | Lexical functional grammars | Head grammars | Functional unification grammars | Constraint propagation |
| 1988 | Context free grammars | Categorial grammars | Augmented transition network | Tree adjoining grammars | Head drive phrase structure grammar | Unification grammars | Categorial unification grammar | Linear indexed grammar | Earley algorithm | Definite clause grammars |
| 1989 | Unification grammars | Context free grammars | Categorial grammars | Head drive phrase structure grammar | Unification categorial grammars | Augmented transition network | Tree adjoining grammars | Structure grammars | Lexical functional grammars | Centering algorithm |
| 1990 | Context free grammars | Tree adjoining grammars | Combinatory categorial grammars | Unification grammars | Categorial grammars | Head drive phrase structure grammar | Functional unification grammars | Cocke younger kasami | Lexical functional grammars | Categorial unification grammar |
| 1991 | Head drive phrase structure grammar | Context free grammars | Unification grammars | Tree adjoining grammars | Categorial grammars | Dynamic programming | Plan recognition | Earley algorithm | Centering algorithm | Hidden markov model |
| 1992 | Tree adjoining grammars | Context free grammars | Unification grammars | Combinatory categorial grammars | Left corner | Categorial grammars | Plan recognition | Functional unification grammars | Structure grammars | Lexical functional grammars |
| 1993 | Context free grammars | Unification grammars | Tree adjoining grammars | Head drive phrase structure grammar | Expectation maximization | Centering algorithm | Left corner | Categorial grammars | Cocke younger kasami | Hidden markov model |
| 1994 | Context free grammars | Head drive phrase structure grammar | Left corner | Dynamic programming | Unification grammars | Expectation maximization | Hidden markov model | Decision tree | Maximal likelihood estimation | Head grammars |
| 1995 | Tree adjoining grammars | Head drive phrase structure grammar | Decision tree | Context free grammars | Expectation maximization | Combinatory categorial grammars | Categorial grammars | Yarowsky | Earley algorithm | Centering algorithm |
| 1996 | Context free grammars | Tree adjoining grammars | Dynamic programming | Head drive phrase structure grammar | Maximal likelihood estimation | Kaplan and kay | Categorial grammars | Yarowsky | Decision tree | Earley algorithm |
| 1997 | Decision tree | Expectation maximization | Maximal likelihood estimation | Context free grammars | Dynamic programming | Head drive phrase structure grammar | Tree adjoining grammars | Hidden markov model | Definite clause grammars | Binary trees |
| 1998 | Context free grammars | Head drive phrase structure grammar | Dynamic programming | Tree adjoining grammars | Neural networks | Decision tree | Dependency grammars | Hidden markov model | Yarowsky | N best |
| 1999 | Context free grammars | Probabilistic context free grammar | Dynamic programming | Decision tree | Maximum entropy | Head drive phrase structure grammar | Tree adjoining grammars | N best | C4.5 | Left corner |
| 2000 | Decision tree | Maximum entropy | Hidden markov model | Maximal likelihood estimation | Expectation maximization | Context free grammars | Head drive phrase structure grammar | Tree adjoining grammars | Dependency grammars | Edit distance |
| 2001 | Context free grammars | Maximum entropy | Dynamic programming | Hidden markov model | Head drive phrase structure grammar | Tree adjoining grammars | Decision tree | Dependency grammars | Probabilistic context free grammar | Categorial grammars |
| 2002 | Maximum entropy | Decision tree | Context free grammars | Expectation maximization | Probabilistic context free grammar | Yarowsky | Dynamic programming | Hidden markov model | N best | Naive Bayes |
| 2003 | Maximum entropy | Decision tree | Context free grammars | Dynamic programming | Hidden markov model | Head drive phrase structure grammar | Support vector machine | Expectation maximization | N best | Maximal likelihood estimation |
| 2004 | Maximum entropy | Expectation maximization | Probabilistic context free grammar | Support vector machine | Decision tree | Hidden markov model | Context free grammars | Dynamic programming | Naive Bayes | Maximal likelihood estimation |
| 2005 | Maximum entropy | Support vector machine | Dynamic programming | Hidden markov model | Naive Bayes | Bootstrapping | N best | Bleu | Expectation maximization | Decision tree |
| 2006 | Support vector machine | Maximum entropy | Dynamic programming | Hidden markov model | Expectation maximization | Bleu | Decision tree | Bootstrapping | N best | Context free grammars |
| 2007 | Support vector machine | Maximum entropy | Bleu | Log linear | Hidden markov model | Expectation maximization | Dynamic programming | Bootstrapping | Perceptron | N best |
| 2008 | Support vector machine | Maximum entropy | Bleu | Hidden markov model | Expectation maximization | Log linear | Dynamic programming | Decision tree | N best | Probabilistic context free grammar |
| 2009 | Support vector machine | Bleu | Maximum entropy | Dynamic programming | Log linear | Expectation maximization | Conditional random fields | Minimum error rate training | Hidden markov model | Bootstrapping |
| 2010 | Support vector machine | Bleu | Expectation maximization | Maximum entropy | Hidden markov model | Dynamic programming | Probabilistic context free grammar | Log linear | Bootstrapping | N best |
| 2011 | Support vector machine | Bleu | Maximum entropy | Expectation maximization | Dynamic programming | Log linear | Cosine similarity | Hidden markov model | Bootstrapping | Minimum error rate training |
| 2012 | Support vector machine | Bleu | Expectation maximization | Conditional random fields | Dynamic programming | Log linear | Maximum entropy | Minimum error rate training | Gibbs sampling | N best |
| 2013 | Support vector machine | Bleu | Graph based | Conditional random fields | Maximum entropy | Cosine similarity | Latent dirichlet allocation | Bag of words | Expectation maximization | Dynamic programming |
| 2014 | Support vector machine | Cosine similarity | Bleu | Expectation maximization | Log linear | Maximum entropy | Perceptron | Conditional random fields | Neural networks | Logistic regression |
| 2015 | Neural networks | Support vector machine | Gradient descent | Stochastic gradient descent | Word2vec | Logistic regression | Bag of words | Cosine similarity | Skip gram | Maximum-entropy |
Note: “*”means that there are less than 10 algorithms in this year.
Fig. 3The number of papers accepted by ACL conference each year.
Fig. 4Ranking of representative algorithms in the top-10 each year.
Fig. 5Algorithms with rapidly growing influence.
Fig. 6Algorithms with steadily growing influence.
Fig. 7Algorithms with steadily declining influence.
Rising span of different algorithms.
| Algorithms | The year of first appearance | The year of highest impact | Time span | |
|---|---|---|---|---|
| Algorithms with rapidly growing influence | Neural networks | 1984 | 2015 | 31 |
| Back propagation | 1991 | 2015 | 24 | |
| Word embedding | 2013 | 2015 | 2 | |
| Word2vec | 2013 | 2015 | 2 | |
| Adagrad | 2012 | 2015 | 3 | |
| Adadelta | 2015 | 2015 | 0 | |
| LSTM | 2013 | 2015 | 2 | |
| Skip-gram | 2009 | 2015 | 6 | |
| Gradient descent | 1994 | 2015 | 21 | |
| Algorithms with steadily growing influence | Support-vector machine | 2000 | 2015 | 15 |
| Cosine similarity | 1998 | 2014 | 16 | |
| Conditional random fields | 1998 | 2012 | 14 | |
| Bag of words | 1985 | 2015 | 30 | |
| Beam-search | 1990 | 2015 | 25 | |
| Logistic regression | 1997 | 2015 | 18 | |
| Singular-value decomposition | 1993 | 2015 | 22 | |
| Latent dirichlet allocation | 2006 | 2013 | 7 | |
| K-means | 1993 | 2015 | 22 | |
| Algorithms with steadily declining influence | Augmented transition network | 1979 | 1979 | 0 |
| Categorial grammars | 1985 | 1987 | 2 | |
| Unification grammars | 1983 | 1989 | 6 | |
| Structure grammars | 1980 | 1982 | 2 | |
| Earley algorithm | 1981 | 1987 | 6 | |
| Context free grammars | 1979 | 1983 | 4 | |
| Hidden markov model | 1991 | 2000 | 9 | |
| Maximum entropy | 1993 | 2005 | 12 | |
| Decision tree | 1983 | 2002 | 19 | |
| Naïve bayes | 1998 | 2005 | 7 |
The influence of top-10 date mining algorithms in NLP domain.
| No. | Algorithm | # papers (Ratio) | No. | Algorithm | # papers (Ratio) |
|---|---|---|---|---|---|
| 1 | SVM | 774 (44.33 %) | 6 | KNN | 67 (3.84 %) |
| 2 | EM | 403 (23.0 %) | 7 | C4.5 | 57 (3.26 %) |
| 3 | Naive Bayes | 190 (10.88 %) | 8 | AdaBoost | 21 (1.20 %) |
| 4 | K-Means | 115 (6.59 %) | 9 | Apriori | 14 (0.80 %) |
| 5 | PageRank | 92 (5.27 %) | 10 | CART | 13 (0.74 %) |