Literature DB >> 32191738

Question classification based on Bloom's taxonomy cognitive domain using modified TF-IDF and word2vec.

Manal Mohammed1,2, Nazlia Omar1.   

Abstract

The assessment of examination questions is crucial in educational institutes since examination is one of the most common methods to evaluate students' achievement in specific course. Therefore, there is a crucial need to construct a balanced and high-quality exam, which satisfies different cognitive levels. Thus, many lecturers rely on Bloom's taxonomy cognitive domain, which is a popular framework developed for the purpose of assessing students' intellectual abilities and skills. Several works have been proposed to automatically handle the classification of questions in accordance with Bloom's taxonomy. Most of these works classify questions according to specific domain. As a result, there is a lack of technique of classifying questions that belong to the multi-domain areas. The aim of this paper is to present a classification model to classify exam questions based on Bloom's taxonomy that belong to several areas. This study proposes a method for classifying questions automatically, by extracting two features, TFPOS-IDF and word2vec. The purpose of the first feature was to calculate the term frequency-inverse document frequency based on part of speech, in order to assign a suitable weight for essential words in the question. The second feature, pre-trained word2vec, was used to boost the classification process. Then, the combination of these features was fed into three different classifiers; K-Nearest Neighbour, Logistic Regression, and Support Vector Machine, in order to classify the questions. The experiments used two datasets. The first dataset contained 141 questions, while the other dataset contained 600 questions. The classification result for the first dataset achieved an average of 71.1%, 82.3% and 83.7% weighted F1-measure respectively. The classification result for the second dataset achieved an average of 85.4%, 89.4% and 89.7% weighted F1-measure respectively. The finding from this study showed that the proposed method is significant in classifying questions from multiple domains based on Bloom's taxonomy.

Entities:  

Year:  2020        PMID: 32191738      PMCID: PMC7081997          DOI: 10.1371/journal.pone.0230442

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


1. Introduction

Examination assessment plays a fundamental role in evaluating how students’ proficiency in the course content [1]. Writing high quality and balanced exam questions that satisfy different levels of cognition is not an easy task [2]. Constructing the examination in a comprehensive manner must also take into consideration the difficulty levels, that match the standard objectives and outcomes of the course in a standard way such as Bloom’s taxonomy [1,3]. Bloom’s taxonomy involves three domains, i.e. cognitive, affective, and psychomotor domains. The cognitive domain, covers different thinking skills starting from the most straightforward to the most complex one [4,5] as shown in Fig 1. The key to writing learning outcomes by Bloom’s taxonomy is verbs [4].
Fig 1

Bloom’s taxonomy cognitive domain levels.

Recently, researchers [5-9] have shown an increased interest in automating evaluating examination based on Bloom’s taxonomy cognitive domain. Various techniques have been used along with different features such as lexical features, and syntactic features, while few of them used semantic features. On the other hand, most previous works handled classifying questions from a specific domain, where there is a lack of techniques on classifying questions based on Bloom’s taxonomy over the multi-domain area [10,11]. Therefore, this study aims to build a question classification model based on Bloom’s taxonomy cognitive domain. Hence, considerable room for further improvement still remains, particularly in open domain area. In general, questions classification is unlike documents classification, since the questions are written in short forms. Documents classification aids users to get and extract useful information easily due to the extensive available information. Whereas, short text suffers from the lack of gained information and sparsity [12,13]. Thus, it is not suitable to use the pure statistical method in order to perform question classification such as N-gram and TF-IDF, since these techniques need a massive amount of data in order to get high accuracy [5]. In addition, removing stop-words during the text classification pre-processing stage of the document is a common step to reduce insignificant words. Nevertheless, some of these stop-words such as what, when, where, and how are valuable in the process of question classification [14]. Another main issue in classifying questions based on Bloom’s cognitive is assigning a suitable weight for keywords that determine the level of the question, especially for the words that might appear in more than one taxonomy, such as the word ‘define’ that belongs to the Knowledge and Comprehension levels. To handle this issue, [15,16] proposed assigning weight to the words from the perspective of experts. Using this method may lead to inconsistencies due to the variety of background knowledge of each expert. This resulted in the poor performance of the classification process. Furthermore, there are still other methods to classifying questions based on Bloom’s cognitive level which might not have been investigated yet such as using word embedding, e.g., word2vec which have shown good results in sentiment analysis and question classification in the question answering system [17-19]. Thus, it might be useful to gain benefit from the use of word embedding with a combination of the modified feature TF-IDF, to enhance the classification process.

2. Related work

The issue of classifying exam questions based on Bloom’s taxonomy has received a considerable critical attention in recent years. In order to handle this task, researchers use different techniques and features. For example, [15] used a pre-defined list of keywords on their online test system, in which the verbs on the questions checked against the keyword list. Their dataset consists of 228 questions written in English. This method can only classify questions that belong to the knowledge level. A rule-based approach with natural language processing techniques was used by [16] to return significant words in a question, then assign weights to these words which are defined previously by experts. Their dataset consists of 100 computer programming questions written in English where the dataset is split into 70% training set and 30% test set. Similarly, [2] used a rule-based approach with N-gram, to enhance the classification result and produce 86% average F1. The dataset that is used in this work consists of 135 questions in computer programming subject written in English language which is split into 100 questions as a training set and 35 questions as a testing set. Likewise, [8] used rules to classify questions from three subjects; Object-oriented programming, software engineering and algorithm and data structure. The dataset contains 53 questions written in English. This work [8] used WordNet with cosine similarity to define question pattern. The accuracy of the outcome from the classification process reached 71%. However, other researchers [5,6,20-24] used machine learning techniques to classify Bloom’s taxonomy into cognitive levels. A neural network is used by [22] with different feature sets: whole feature, document frequency, and category frequency-document frequency to classify questions by training the model with a scaled conjugate gradient learning algorithm. The result shows that document frequency reduction is more effective with regard to the classification and coverage time compared to other feature reductions. Their dataset consists of 274 questions, divided into 192 training set and 82 testing set. TF-IDF feature is extracted extensively in many works, [23] performed pre-processing for the dataset which contains 272 questions collected from different websites. Then TF-IDF was calculated and fed into Linear SVM, which produced satisfactory results in term of accuracy and precision, but not in F-measure and recall. This work was extended in [20] by increasing the size of the dataset to 600 questions and testing the effect of classification with and without stop words using the same TF-IDF feature. As a result, removing stop words does not make a big impact on the result. Similarly, [6] used the same dataset prepared by [20], but with enhanced TF-IDF which is multiplied with impact factor. Then the classification process handled by three classifiers NB, KNN, and SVM where SVM performance superior other classifiers. In addition, an ensemble technique is used by [5,24] to classify questions into Bloom’s taxonomy. [5] applied voting algorithm on a combination of three classifiers i.e. SVM, KNN and NB with chi-square, mutual information & odd ratio on questions from programming subject. [24] used ensemble to handle keywords overlapping issue by combining four classifiers namely rule-based, SVM, KNN and NB using majority voting algorithm and WordNet similarity. The dataset that is used in this work contains 100 programming questions which divided into 60 questions training set and 40 questions test set. The results show that the ensemble returns the best performance compared to each individual classifier by achieving 95% F-measure. One of the most popular weighting methods is TF-IDF, which stands for Term Frequency- Inverse Document Frequency. It assigns scores to the importance of a word in a document, based on the lexical and morphological properties of the text. It is extensively used in many studies [25-29], and in question classification [3,6,20,23,30-32] However, some researchers used TF-IDF as it is, while others proposed some enhancement to the way TF-IDF is calculated in order to improve it. This is because classical TF-IDF does not capture some useful information such as the impact of word distribution between various classes [6]. A Massive Open Online Courses (MOOC) search engine proposed by [26], used a modified TF-IDF, that assigned the important words in the query to specific weights according to their parts of speech. This study changed the way of calculating the term ‘frequency’ by assigning a higher weight for the most important word, which is a noun or verb. Then in the second priority, the adjective and adverb are assigned to the weight that is less than a noun and verb but higher than other parts of speech which are assigned to the lowest weight. This method produces a significant result. Other researchers [33-35] also use the syntactic feature of the parts of speech to assign a suitable weight to the terms. Another example of improving the way of calculating the TF-IDF is proposed by [36] who introduced an impact factor that is multiplied by the traditional TF-IDF. The purpose of the impact factor equation is to calculate the weights of the class distribution. This enhancement performs better than the classical method. Similarly, [6] used the impact factor to enhance the result of TF-IDF to classify question based on Bloom’s taxonomy. A comprehensive review in question classification by [37] mentioned that most of the works in question classification used semantic and syntactic features rather than pure statistical features such as bag-of-word and n-gram. Moreover, [37] stated that semantic features have been significantly used in question classification in question answering system and information retrieval and produced considerable accuracy, while there is a lack of extracting semantic features in educational environments. On the other hand, many Natural Language Processing (NLP) studies with deep learning models have included learning word vector representation. The word vectors are represented in a dense form known as word embedding, in which the words that are semantically and syntactically related are close to each other in the embedding space [13,38-40]. Word embedding has been used efficiently in many NLP tasks [41]. Word2vec is one of the common word embeddings, where vectors representation is learnt via neural-network language model. It was developed based on Google by [38]. It required extensive training text as input and established vocabulary, then the model learned word vectors representation in such a way that the words that shared the same context are close to each other in the vector space. These word vectors can be used as features in various NLP tasks. As [17] mentioned, the pre-trained vectors can be considered as universal feature extractors. Many NLP tasks had benefited by word2vec, e.g., sentiment analysis, machine translation, and paraphrase detection [38]. [17] proposed a model that classified questions in question answering system, and sentiment analysis using Convolutional Neural Network (CNN) that was trained on the top of pre-trained word2vec, which produced excellent results. Similarly, [18] used CNN for Arabic sentiment analysis with Arabic pre-trained word2vec, which produced a significant result and outperformed previously existed methods. The work [42] proposed a method that classified texts via SVM classifier, with the use of a semantic feature based on word2vec weighted by TF-IDF. In this study, several experiments compared if TF-IDF was better than combining it with word2vec as well as whether with or without stop words. Three steps were performed to produce the vector representation of word2vec with TF-IDF. The first step was summing the word2vec vectors to produce a single vector. The second step was the summation of multiplication of word2vec and TF-IDF. The third step was to concatenate the outcome of first and second step. For example, if the size of the vector of the first step was 200, and the vector size of the second step was 2000, then after appending them together the final representation of the vector would be 2200 which represented a document. The result showed that the best performance was achieved by word2vec without stop words being combined with TF-IDF without stop words. Another example of using a combination of word2vec and TF-IDF was [36]. This work used three classifiers; SVM, KNN and Radial Basis Function (RBF) network, in order to classify Chinese documents. They performed enhancement to the TF-IDF by multiplying with impact factor, the improved feature called ITF-IDF. Then the ITF-IDF was combined with word2vec by applying the following equation: where t is a term in document d, and J is the number of words in document d. For example, if the size of word2vec vectors is 200, then the size of the vector representation of the document can be obtained by applying Eq (1) is 200. The result of this study showed that the combination of both features outperformed the using of each feature individually with all classifiers. The work of [32] classified questions based on a Persian question answering system. This study proposed three feature extraction methods; one of them was word2vec with weighted TF-IDF. The TF-IDF was raised to the power of TF-IDF factor and then multiplied by word2vec. The study stated that tuning the value of TF-IDF factor helped enhanced the accuracy significantly. SVM and Neural Network used as classification algorithms, in which they produced results that showed the effectiveness of the vector representation.

3. Method

The process of the proposed model is shown in Fig 2, and it involves five stages. The details of these stages will be discussed in detail in this section.
Fig 2

Proposed model.

3.1 Question collection

Since the aim of this study was to build a model to classify questions based on Bloom’s Taxonomy from different domains, two open domain datasets had been used to ensure the stability of the proposed method. The first dataset were collected from several websites, books, and previous work [2]; the total number of questions in this dataset was 141 open-ended questions. The second dataset was introduced by Yahya et al. (2012), and it consisted of 600 open-ended questions. Both datasets were labeled and annotated into six classes. Table 1 shows the distribution of questions at each level in both datasets. Table 2 shows a sample of questions from both datasets.
Table 1

The number of questions in each dataset.

Cognitive LevelCollected DatasetYahya et al. (2012)
Knowledge Level26100
Comprehension Level23100
Application Level15100
Analysis Level23100
Synthesis Level30100
Evaluation Level24100
Total141600
Table 2

Sample of questions from each dataset.

Collected datasetYahya et al. (2012) dataset
LevelQuestionAreaQuestionArea
KNOWLEDGEMemorize and recall the periodic tableChemistryLabel the parts of the microscope shown on the rightBiology
Name three 19th-century women English authors.LiteratureList reserved words in C programming.Computer programming
COMPREHENSIONExplain how the heart is like a pump.BiologyRetell the story in your words.Literature
Draw a diagram explaining how air pressure affects the weather.PhysicsDetermine the next number in a sequenceMath
APPLICATIONDesign or sketch a marketing strategy for your product using a known strategy as a model.Social marketingShow how E-CRM can be used to improve marketing positioning as explained in the articleSocial marketing
Write a C++ statement to declare a variable of type music Type name MyTune.Computer programmingSketch a prediction of the field lines for the arrangement of electrodes shown in Fig 2Physics
ANALYSISWhat is the relationship between probability and statistical analysis?MathBy comparing the map of the tectonic plates to the earthquake map, what inferences can you make?Geology
Analyze a work of art in terms of form, color and texture.ArtBreak down the components of a standard film camera and explain how they interact to make the machine work.Physics
SYNTHESISWrite a JAVA program to show the Overloading conceptComputer programmingExplain how the biological concept of symbiotic relationships could be used to help solve socially created problems like water pollution, overflowing garbage landfills, or homelessness.Ecology
Use your imagination to create a picture about the story. Then, add one new thing that was not in the story.LiteratureDevelop a SQA Plan for a software development project which is defined in the attached document.Software engineering
EVALUATIONYour advice has been sought to settle the following dispute in Company X. Referring to appropriate legal principles, write a short report advising the company on the best course of action to adopt.LawConclude and support which economic system leads to a higher standard of living.Economy
"Don’t use public instance variables" is defensive programming techniques. discuss why it is good advice?Computer programmingExamine the stated positions of both major political candidates with regard to a particular issue and state good reasons (based on principles discussed in class) for why one candidates position is more likely to be effective than the other’s.Political science

3.2 Pre-processing

Performing pre-processing for unstructured data is highly recommended in any framework which uses machine learning [43] to reduce the unnecessary, duplicated, irrelevant, and noisy data. Therefore, pre-processing is implemented before the feature extraction phase and the classification process. However, in the pre-processing stage, the questions had to pass through several steps: normalization, tokenization, part-of-speech tagging, and stemming in order to refine and prepare them for the next stage. The first step was the normalization process in which the unwanted data such as punctuation marks, numbers, and non-English characters will be eliminated. Then, all the words in the question will be converted into lowercase. In addition, stop-words are discarded from the question. In this step, not all stop-words will be eliminated from the question. The reason behind keeping some of the stop-words is that some questions contain important words, which can be considered as stop words in the default stop-words list, while these words have a significant impact in determining the level of the question. For example, the question ‘Define transferable skill’ belongs to Knowledge level, while the question ‘In your own words, how would you define transferable skills’ belongs to the Comprehension level. In case of eliminating the stop-words from the latter example, the question will be ‘words, define transferable skills’ where the following words will be deleted ‘your, own, how, you’, which means the question will look like Knowledge level question; these words can be effectively used in determining the class of question during classification process. After performing normalization for the question, the tokenization will take place by splitting the question into individual words based on the whitespaces, known as tokens. Then the Stanford Part-of-Speech (POS) tagger (version 3.9.1) [44] tags the question words. The word will be labelled with a syntactic class such as noun (NN), verb (VB), adjective (JJ), and adverb (RB). Finally, the last step in pre-processing is stemming, the process of deleting the prefix and suffix in order to return the word into its original root. For example, waits, waiting, and waited after stemming will be wait. However, in order to perform stemming one of the most popular stemmers is used, which is the Porter stemmer in NLTK toolkit [45]. Unfortunately, the stemmer sometimes suffers from a problem of failure to return the correct root of some words e.g., recalls, recalling and recalled after stemming will be recal instead of recall. Although, in this case, it does not matter since similar terms are stored at the same index, which means during search and retrieval, the desirable word will be retrieved correctly. Moreover, POS tagging and the stemming will be used only with the TFPOS-IDF feature, and not word2vec as it will be discussed later.

3.3 Feature extraction

After pre-processing have been being applied, now the questions are ready for features extraction phase. Feature extraction is a process of transforming the raw input data into a meaningful set of features [46], in such a way that it can be understandable to the machine learning classifiers. Two features will be extracted. The aim of the first feature is to develop a weighting method to set the priority of words based on modified TF-IDF with Part-of-Speech (POS). Whereas the second feature is used to enhance the classification process, by extracting the semantic feature based on pre-trained word2vec. These two features will be combined to produce the vectors that will be used in the classification process.

3.3.1 Modified TF-IDF (term weighting method TFPOS-IDF)

The Term Frequency (TF) is known as a local term weight, whereas the Inverse Document Frequency (IDF) known as a global term weight, is calculated using the following formulas: where c(t,d) indicates the occurrence of term t appears in document d, and the denominator indicates the total number of terms in document d, D is the total number of documents in the dataset, and d is the number of the documents a term t appeared in. Many researchers [26–28, 36] tries to improve the performance of TF-IDF by proposing some modification to the original equation of TF-IDF. In the case of Bloom’s taxonomy verb plays an important role to determine the level of the question. Relying on the pre-defined list does not always guarantee a good performance, especially for words that appear in more than one level. Therefore, the proposed feature Term Frequency (TF)-Inverse Document Frequency (IDF) based on Part-Of-Speech (POS) is introduced known as TFPOS-IDF. The idea beyond this approach is inspired from [26]. In order to determine the suitable rank for word weight according to Bloom’s taxonomy, several cases have been examined. The experiments have been run for fifteen times using both dataset and three classifiers. Table 3 demonstrates the average of weighted F1-measure using KNN, LR, and SVM with the collected dataset. Table 4 displays the average of weighted F1-measure using KNN, LR, and SVM with the Yahya et al.'s (2012) dataset.
Table 3

Weighted F1-measure of different weight cases with collected dataset using KNN, LR and SVM.

Casesw1w2w3KNNLRSVM
CASE 1VBOTHERWISE-0.6340.7330.738
CASE 2VBNNOTHERWISE0.6560.7510.759
CASE 3VBNN–JJOTHERWISE0.6590.7560.764
CASE 4VBNN-RBOTHERWISE0.6470.7390.749
CASE 5VBNN-JJ-RBOTHERWISE0.6330.7430.751
CASE 6VB-WHOTHERWISE-0.5680.7030.718
CASE 7VB- WHNNOTHERWISE0.6210.6980.652
CASE 8VB- WHNN-JJOTHERWISE0.6140.7010.721
CASE 9VB- WHNN-RBOTHERWISE0.6250.6980.718
CASE 10VB- WHNN-JJ-RBOTHERWISE0.6050.7010.721
CASE 11VB-WH-MDOTHERWISE-0.5740.6870.715
CASE 12VB-WH-MDNNOTHERWISE0.6100.6870.713
CASE 13VB-WH-MDNN-JJOTHERWISE0.6060.6940.717
CASE 14VB-WH-MDNN-RBOTHERWISE0.6100.6870.713
CASE 15VB-WH-MDNN-JJ-RBOTHERWISE0.6190.6940.717
Table 4

Weighted F1-measure of different weight cases with Yahya et al. (2012) dataset using KNN, LR and SVM.

Casesw1w2w3KNNLRSVM
CASE 1VBOTHERWISE-0.8000.8280.847
CASE 2VBNNOTHERWISE0.8240.8480.861
CASE 3VBNN–JJOTHERWISE0.8340.8560.866
CASE 4VBNN-RBOTHERWISE0.8170.8340.841
CASE 5VBNN-JJ-RBOTHERWISE0.8260.8370.855
CASE 6VB-WHOTHERWISE-0.7490.8040.790
CASE 7VB- WHNNOTHERWISE0.7660.8120.796
CASE 8VB- WHNN-JJOTHERWISE0.7800.8150.800
CASE 9VB- WHNN-RBOTHERWISE0.7700.8140.797
CASE 10VB- WHNN-JJ-RBOTHERWISE0.7770.8170.804
CASE 11VB-WH-MDOTHERWISE-0.7470.8050.799
CASE 12VB-WH-MDNNOTHERWISE0.7620.8150.803
CASE 13VB-WH-MDNN-JJOTHERWISE0.7700.8160.808
CASE 14VB-WH-MDNN-RBOTHERWISE0.7630.8170.806
CASE 15VB-WH-MDNN-JJ-RBOTHERWISE0.7640.8160.808
where MD, WH, VB, NN, JJ, and RB stand for the model, wh-question word, verb, noun, adjective, and adverb, respectively. From the results shown in Table 3, the best result achieved with the collected dataset when verb assigned to w1, noun and adjective assigned to w2, and other POS assigned to w3. Similar experiments had been implemented using Yahya et al.'s (2012) dataset. Table 4 shows the average of weighted F1-measure using KNN, LR, and SVM with Yahya et al.'s (2012) dataset. The findings stated from Tables 3 and 4, confirm that verbs are the most important words to determine the level of the question according to Bloom’s taxonomy. In addition, assigning a higher weight for wh-question words, or model along with verbs does not help to improve the performance of classification since these words appear in all levels. Furthermore, nouns often carry significant information in question and document classification [10]. That is obviously clear by comparing the results of Case 1 versus Case 2, Case 6 versus Case 7, and Case 11 versus Case 12. Giving noun some priority over other POS improves the performance of the classification. Lastly, it can be observed that the best result is obtained in Case 3. Which means adjective also carries a significant information in question classification based on Bloom’s taxonomy, since it improves the classification outcome. Thus, in the case of question classification based on Bloom’s taxonomy cognitive domain, the highest priority will be given to verb by assigning w1. Moreover, from the experiments, the best result with both datasets and both classifiers are obtained when the noun and adjective are assigned to w2. Therefore, noun and adjective will be in the second rank, in which their weight will be less than the verbs but more significant than other POS. Lastly, other POS such as preposition, and pronoun will be assigned to the lowest weights. The method for assigning a weight for the term t is illustrated by Eq 5 where w1>w2>w3> 0, assume that the weights values for w1,w2, and w3 are w1 = 5, w2 = 3 and w3 = 1, as in [26] study. Now the new equation that represents TFPOS, i.e. the term frequency based on POS is: finally, the TFPOS-IDF(t,d) is calculated for term t in document d by the following equation: The output of this phase is the sparse matrix, i.e. high-dimensional vectors. After calculating TFPOS-IDF the feature vectors will be normalized to prevent the numerical complexities during the calculation. L2 norm is used for normalization, which is the most popular norm. The representation of the L2 norm is demonstrated by the following equation: Now the feature vector representation is already normalized in which the sum of square values of a row is equal to one. For example, the values of each term in question “Recall the main components of the flowchart” after pre-processing and applying TF-IDF and TFPOS-IDF is shown in Table 5. The higher value indicates that it is a more relevant and important term in this document. For example, the TF-IDF’s value for recall is 0.774, but for the modified feature TFPOS-IDF is 0.878 since this word is a verb, it has a higher value, which denotes that recall appears more in this document. Whereas other words main, components and flowchart have a higher value in TF-IDF comparing to the modified feature TFPOS-IDF which reduces the values of these words, since it is not as important as verb.
Table 5

Example of weighting method using classical TF-IDF and modified TFPOS-IDF.

Tagged termsrecall / VBmain /JJcomponents /NNSflowchart/NN
Stemmed termsrecalmaincomponflowchart
TF-IDF0.7740.5810.1940.161
TFPOS-IDF0.8780.4400.1460.122

3.3.2 Word2vec

The first proposed feature TFPOS-IDF focuses on giving verbs higher priority over other words. Focuses on the verbs might not be enough. Therefore, it is worthwhile to try extracting another feature, which is the semantic feature word embedding; to the best of our knowledge has not been applied yet in the question classification based on Bloom’s taxonomy. Since the context of the question might carry an important information, this may help in determining the level of question. This study used pre-trained word2vec, which trains approximately one hundred billion tokens using Google News dataset. The word vector for each word consists of floating-point values, and the meanings of these feature values are hard to explain, except it denotes that the words with similar meaning have a similar vector representation. For example, the top five words that have a similar vector to the word ‘Examine’ are ‘Scrutinize, Investigate, Determine, Assess, Evaluate’. Fig 3 shows an example of converting question into vector representation. The tokenized question that was produced after the pre-processing step, will be converted into the matrix of floating-point values with the size of 300 dimensions. Gensim library is used to retrieve the word vectors from the word2vec. Lastly, if the word does not exist on the word2vec, then vectors will be initialized by zeros.
Fig 3

Example of converting question into a word vector.

As mentioned previously, the stemming process is performed only in the TFPOS-IDF step. The question word in this step is not stemmed. The reason behind keeping the word as it is without stemming, is that the pre-trained word2vec, i.e. (Google News) is not pre-processed before the process of learning and establishing the word vectors, and it builds the relationships between these vectors. For example, there is no word vector for the stemmed wordrecal’. But on the other hand, there are word vectors for ‘recall’, ‘recalling’, and ‘recalls’ which are not totally equal but approximately have similar vectors. The strength of these features lies in producing similar word vectors for the words that have similar meanings, in low dimensional vectors representation. After retrieving the words vector for each question in the corpus, it is a time to combine them with the improved feature TFPOS-IDF.

3.3.3 Combination of word2vec and TFPOS-IDF (W2VTFPOS-IDF)

The aim of this phase is to produce a question with high-quality feature vectors representation by combining the two proposed features, discussed before in the previous two sections. [32] used word2vec with TF-IDF to represent feature vector for question classification in the question answering system, by multiplying the term value of TF-IDF with the word vector of that term. Then after that, all the terms vector of the question is summed together to produce a single vector. This study uses the same approach as demonstrated in Eq (9). where the term t belongs to the document d. The output of this phase is a dense vector, i.e. low-dimensional vector. The size of the vector is 300 dimensions since it is based on the pre-trained word2vec with the dimensions of 300. The advantage of representing question vector in the dense representation (low dimensional) over sparse representation (high dimensional) lies in decreasing the complexity of the classification process [36]. Fig 4 demonstrate an example of producing question vectors by combining word2vec with TFPOS-IDF.
Fig 4

Example of combining word2vec with TFPOS-IDF.

Now the questions vectors W2V-TFPOSIDF are ready to be fed into the classification algorithms, which will be implemented in the next section.

3.4 Classification

Three of the most common supervised machine learning classification algorithms are used, which are the and K-Nearest Neighbour (KNN), Logistic Regression (LR), and Support Vector Machine (SVM).

3.4.1 K-Nearest Neighbours (KNN)

K-Nearest Neighbours (KNN) is one of the simplest algorithms in machine learning. KNN performs well with low-dimensional vectors as shown in[36,47]. This algorithm is dominated by three crucial components [48]; (K) which is the number of nearest neighbours, the labelled training set, and how to estimate the distance among the points. The way K-NN behave is by looking to the K nearest neighbours among the labelled training questions in which the similarity between a test question and K training questions is calculated to predict the class of the test question. The class or the category of the test question is assigned based on the highest score (majority class) of the K training questions. For example, for the Question X, if the algorithm found five nearest neighbours to X in which one belongs to the Knowledge class, one belongs to the Analysis class, two belongs to the Application class, and one belongs to the Evaluation class, then the question X is assigned to the Application class.

3.4.2 Logistic Regression (LR)

Logistic Regression classifier is based on logistic function, also known as the sigmoid function, which has been used in several text classification applications [49,50] and shows significant results. The idea behind it is to find the relationship among features and specific output. Logistics regression algorithm applied one versus all approach for the multiclass classification task. For example, to predict the question class, six binary classification problem are considered i.e. whether the class is knowledge, comprehension and so on with all classes. Following this, the Maximum Likelihood Estimation is used to assign the predicted class and it is implemented by the following equation [51]:

3.4.3 Support Vector Machine (SVM)

Support Vector Machine (SVM) was proposed by [52] at AT&T Bell Laboratories. It is widely used in text classification problems [53]. The idea behind SVM is to find an appropriate hyperplane that best split-up the two sets of data from each other, by maximizing the margin between the hyperplane and the set of data points nearest to it. In-text classification SVM with linear kernel perform well [54]. Therefore, this study will use it. Given N data points from questions in the training set , where x is the i-th input features (i.e. pattern), and y is the i-th output class label. The SVM linear kernel equation can be written as the following [55]: where n is the number of support vectors, xi is the pattern of i-th question, yi is the class label of the i-th question, and θ(x,x) is linear kernel function, which can be written as:

3.5 Evaluation metrics

To measure the effectiveness of the proposed model, the weighted recall, precision and F1-mueasre will be calculated. In order to define these metrics, the terms True Positive (TP), False Positive (FP) and False Negative (FN) that must be introduced. TP is the occurrence of instances a classifier correctly classified to the suitable class. FP is the occurrence of instances incorrectly classified. FN is the occurrence of instances that have not been classified. The value of recall, precision, and F1-measure varies from 0 to 1. The closer value to zero of recall, precision, and F1-measure implies the poor performance. Whereas, the closer value to 1 denotes a good performance. Recall metric measures the perfection of classifiers by computing this formula Precision metric measures the fineness of classifiers by calculating this equation Finally, the F-measures or also known as F-score, is calculated by When the value of β = 1 that represents F1-measure, where the equation will be: Since the training data and test data are randomly selected, and usually imbalanced, the weighted recall, precision, and F1-measure will be calculated. Scikit-learn defined the term weighted “Calculate metrics for each label, and find their average, weighted by support (the number of true instances for each label). This alters ‘macro’ to account for label imbalance; it can result in an F-score that is not between precision and recall.”. The equations for calculating the weighted precision, weighted recall and weighted F1-measure are:

4. Experiment setting

In order to evaluate the effectiveness of the proposed model, several experiments had been conducted, with three features and three classifiers. The first feature was the traditional statistical feature TF-IDF. The second feature was the modified feature TFPOS-IDF. The purpose of mentioning the traditional feature TF-IDF was to compare it with the improved feature TFPOS-IDF, in order to check whether the improvement is significant or not. The third feature W2V-TFPOSIDF is word2vec, which was weighted by the improved feature TFPOS-IDF. All these features were fed into three classifiers KNN, LR, and SVM with a linear kernel, which were implemented using Scikit-learn library in Python with the default settings. The default value of K in the KNN classifier is five, and the default value of the penalty is l2 in LR. The default setting of the parameter C in SVM is 1. However, the experiments were run 15 times; during each run, the training and testing set were randomly picked up. The dataset was divided into 80% training set and 20% test set, for the first dataset, the size of the training set was 112 and 29 for the testing set. The size of the training set in the second dataset is 480, and the testing set is 120.

4.1 Result of KNN

This section discusses the experiments result that has been obtained from the collected dataset and Yahya et al. (2012) dataset using the KNN classifier in terms of weighted recall, precision, and F1-measure for each Bloom class. The experiments measured the performance of KNN with three features TF-IDF, TFPOS-IDF, and W2V-TFPOSIDF. It was observed from the result demonstrated in Table 6 that the modified feature TFPOS-IDF performs better than the traditional feature TF-IDF from all aspects. In addition, the proposed features W2V-TFPOSIDF outperform the improved feature by 5.2% in terms of F1-measure. On the other hand, the score of recall, precision, and F1-measure for the application class with all features are low compared to other classes. This is due to the small number of instances which belongs to the application level in the collected dataset. Therefore, the learning algorithm KNN misclassified the questions that belong to the application level.
Table 6

Results of using KNN with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.7810.8000.7770.7940.8870.8220.7820.9470.832
Comprehension0.8890.7780.8190.9200.7830.8370.8570.6170.691
Application0.1340.3160.1730.1340.3410.1830.3070.5800.375
Analysis0.7420.4640.5530.7410.5260.6010.7070.7980.729
Synthesis0.6910.6600.6570.8010.6940.7290.8810.7770.808
Evaluation0.4640.7420.5540.5590.7490.6270.6550.8250.694
Weighted Average0.6320.6510.6110.6830.6870.6590.7240.7730.711
Based on the results demonstrated in Table 7, it can be observed that the classifier learnt more about Bloom’s cognitive levels using the modified feature TFPOS-IDF, since the performance has increased from all aspects. The classification results of using the proposed feature W2V-TFPOSIDF outperforms the modified feature TFPOS-IDF. The result with this dataset is better than the obtained result from the collected dataset, since the size of Yahya et al. (2012) dataset is larger than the collected dataset. Moreover, the number of the questions in Yahya et al. (2012) dataset is distributed evenly over all classes, unlike the collected dataset.
Table 7

Results of using KNN with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the Yahya et al. (2012) dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.9460.7350.8240.9870.7980.8800.9330.8660.895
Comprehension0.7960.8140.8000.8120.8650.8340.8800.9190.896
Application0.6600.8260.7340.8200.8700.8460.8690.8470.856
Analysis0.8750.6870.7670.9030.8370.8650.9200.8150.864
Synthesis0.6860.7950.7300.7250.7880.7500.7880.8120.794
Evaluation0.6340.8620.7230.7550.9170.8200.7290.9470.817
Weighted Average0.7660.7860.7630.8350.8460.8340.8550.8660.854

4.2 Result of LR

This section discusses the results of the experiment that have been obtained with the LR classifier in terms of the weighted recall, precision, and F1-measure for each Bloom class. Table 8 demonstrates the results of the collected dataset. Table 9 shows the results of Yahya et al. (2012) dataset.
Table 8

Results of LR with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.8330.8980.8510.8330.9240.8640.8850.9600.910
Comprehension0.8730.7980.8140.9370.8970.9070.8570.8860.860
Application0.0770.3080.1210.1540.5960.2410.3270.7260.419
Analysis0.7590.5920.6450.8100.6460.6950.9480.8440.882
Synthesis0.9100.7010.7820.9700.7470.8350.9300.8250.861
Evaluation0.6550.8230.7090.7500.9170.8110.8930.8830.879
Weighted Average0.7220.7120.6880.7790.8020.7560.8340.8600.823
Table 9

Results of LR TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for Yahya et al. (2012) dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.9070.8070.8500.9970.8260.9010.9870.9170.950
Comprehension0.8030.9130.8520.8060.9300.8590.8580.9700.908
Application0.7280.9050.8040.7970.8980.8430.8580.8760.865
Analysis0.9070.7330.8070.9270.8640.8930.9440.9390.940
Synthesis0.7420.8630.7930.7770.8560.8100.8450.8350.834
Evaluation0.8080.7650.7790.8400.8310.8290.8680.8680.865
Weighted Average0.8140.8330.8140.8570.8680.8560.8930.9010.894
Tables 8 and 9 show the performance of the LR classifier with both datasets in which the W2V-TFPOSIDF feature produces a better result than TFPOS-IDF and TF-IDF.

4.3 Result of SVM

This section discusses the results of the experiment that have been obtained with the SVM classifier in the terms of the weighted recall, precision, and F1-measure for each Bloom class using both dataset; collected dataset, and Yahya et al. (2012) dataset. As shown in Table 10 the SVM classifier with the modified feature TFPOS-IDF outperforms the traditional feature TF-IDF in terms of all the evaluation metrics recall, precision and F1-measure, as well as W2V-TFPOSIDF which outperform the two other features.
Table 10

Results of SVM with TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for the collected dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.8340.9900.8950.8471.0000.9080.8850.9780.920
Comprehension0.9520.8270.8760.9360.8730.8970.9040.8900.892
Application0.1530.5380.2290.1730.5960.2610.4620.6290.507
Analysis0.8620.5350.6450.8960.5400.6630.9660.8650.905
Synthesis0.9200.7800.8370.9700.8100.8770.8810.8500.855
Evaluation0.6080.8170.6790.7010.9390.7790.8700.8590.853
Weighted Average0.7500.7700.7240.7860.8170.7640.8440.8560.837
The detailed results of using SVM classifier with Yahya et al. (2012) dataset have been shown in Table 11, where TFPOS-IDF outperforms TF-IDF, and W2V-TFPOSIDF Superior both features.
Table 11

Results of SVM TF-IDF, TFPOS-IDF, W2V-TFPOSIDF for Yahya et al. (2012) dataset.

Cognitive LevelTF-IDFTFPOS-IDFW2V-TFPOSIDF
RecallPrecisionF1-measureRecallPrecisionF1-measureRecallPrecisionF1-measure
Knowledge0.9800.8410.9020.9870.8930.9350.9930.9290.961
Comprehension0.7990.9250.8540.8160.9320.8680.8710.9450.905
Application0.7660.8530.8030.8390.8530.8450.8800.8530.864
Analysis0.9140.7880.8450.9380.9100.9210.9610.9290.943
Synthesis0.7140.8680.7780.7910.8310.8050.8410.8500.841
Evaluation0.7970.7560.7710.8250.8230.8190.8290.9150.866
Weighted Average0.8280.8390.8260.8670.8740.8660.8970.9020.897
Based on the results from Tables (6–11), it can be concluded that LR and SVM approximately produce similar results, and achieve better performance compared to KNN in classifying questions from different domains in accordance with Bloom’s taxonomy, using both datasets.

5. Statistical test

The statistical test, the t-test can be used to decide whether two samples of data are significantly different from each other. In order to check the significance of the first modified feature TFPOS-IDF and the second proposed feature W2V-TFPOSIDF, the weighted F1-measure values of each run were used. This section discusses the results of the significant test in two phases, using both the datasets and the three classifiers KNN, LR and SVM. In the first phase, the first modified feature TFPOS-IDF was checked against the traditional feature TF-IDF, to check if the proposed improvement was significant or otherwise. The second phase, the TFPOS-IDF was checked against the W2V-TFPOSIDF to check the significance of adding TFPOS-IDF to the word2vec. A two-sample t-test of the null hypothesis was used. The null hypothesis meant that the value of the hypothesized mean difference was equal to zero, which assumed that there was no difference between the two samples. The value of alpha P-value was equal to 0.05. The idea behind t-test was to check if the obtained value of alpha was less than the assumed one (i.e. P-value = 0.05), if so then the null hypothesis would be rejected, which means there was a significant difference between the two samples. Otherwise, if the P-value was greater than or equal to the assumed one, then the null hypothesis was accepted, which meant that there was no significant difference between the two samples. The alpha values of the t-test are shown in Table 12.
Table 12

Alpha values of t-test.

ClassifierDatasetTF-IDF vs. TFPOS-IDFTFPOS-IDF vs. W2V-TFPOSIDF
KNNCollected Dataset0.0020560.041744
Yahya et al. (2012) Dataset9.95E-060.049986
LRCollected Dataset0.0011630.006075
Yahya et al. (2012) Dataset1.03E-056.63E-06
SVMCollected Dataset0.0010890.00065
Yahya et al. (2012) Dataset6.34E-055.61E-05
As previously mentioned, if the P-value of the t-test is less than 0.05, then there is a significant difference between the two samples. Therefore, from the results shown in Table 11, it is clear that in all cases, the modified feature TFPOS-IDF is statistically significant compared to the classical one TF-IDF. In addition, the proposed feature W2V-TFPOSIDF is statistically significant compared to the modified feature TFPOS-IDF.

6. Conclusion

In conclusion, this paper proposed a method to classify open-domain questions based on Bloom’s taxonomy cognitive domain. In order to check the robustness of the proposed model, two datasets with three classifiers have been examined. In addition, the statistical test shows that there is a significant improvement between TF-IDF versus TFPOS-IDF and TFPOS-IDF versus W2V-TFPOSIDF. This method can be useful for educators and lecturers to analyze the exam question to fulfill the requirement for the different levels of education such as Bachelor’s degree or Master’s level. For example, the exam paper at a Master’s level usually contains the higher thinking order questions more than the lower thinking order. In addition, it can be effectively used throughout different kinds of applications such as automatic test generation systems, intelligent tutoring systems and even more in serious game such as in [56]. Nevertheless, further study could improve, enhance or extend this work by expanding the dataset through adding different kinds of questions. The dataset used in this study only consists of open-ended questions, in which there are neither multiple choices nor true or false questions. Most of the previous studies are conducted with an English dataset where there is a lack of studies in other languages. Since there is no public benchmark dataset that contains real exams questions labeled based on Bloom’s taxonomy, providing one would be beneficial to relevant researchers in this area.

Collected dataset.

(DOCX) Click here for additional data file.

Yahya et al. (2012) dataset.

(DOCX) Click here for additional data file.
  2 in total

1.  Malay sentiment analysis based on combined classification approaches and Senti-lexicon algorithm.

Authors:  Ahmed Al-Saffar; Suryanti Awang; Hai Tao; Nazlia Omar; Wafaa Al-Saiagh; Mohammed Al-Bared
Journal:  PLoS One       Date:  2018-04-23       Impact factor: 3.240

2.  Word2vec convolutional neural networks for classification of news articles and tweets.

Authors:  Beakcheol Jang; Inhwan Kim; Jong Wook Kim
Journal:  PLoS One       Date:  2019-08-22       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.