| Literature DB >> 35965951 |
S Poomagal1, B Malar1, E M Ranganayaki1, K Deepika1, G Dheepak1.
Abstract
Classifying product reviews is one of the tasks in Natural Language Processing by which the sentiment of the reviewer towards a product can be identified. This identification is useful for the growth of the business by increasing the number of satisfied customers through product quality improvement. Bigram models are more popular in performing this classification since it considers the occurrence of two words consecutively in the reviews. In the existing works on bigram models, semantically similar words to the words present in bigrams are not considered. As the reviewers use different words with the same meaning to express their feeling, we proposed improved bigram models in which semantically similar words to the words in bigrams are also used for classifying the reviews. In the proposed models, sentiment polarity thesaurus is constructed by including sentiment words and their synonyms. The combinations of constructed thesaurus, Synset and Word2Vec are used for extracting synonyms for the words in the reviews. Performance of the proposed models is compared with the traditional bigram model and state-of-the-art methods. It is observed from the results that our models are able to achieve better performance than traditional model and recent methods.Entities:
Keywords: Bigram; Classification; Natural Language Processing; Synset; Unigram; Word2Vec
Year: 2022 PMID: 35965951 PMCID: PMC9362630 DOI: 10.1007/s42979-022-01305-8
Source DB: PubMed Journal: SN Comput Sci ISSN: 2661-8907
Scenarios
| M1 | Traditional bigram model |
| Methods which use single resource | |
| M2 | Similar words extraction from only Synset |
| M3 | Similar words extraction from only Word2Vec |
| M4 | Similar words extraction from thesaurus only for sentiment words |
| M5 | Similar words extraction from Synset only for sentiment words |
| M6 | Similar words extraction from Word2Vec only for sentiment words |
| Methods which use various combinations of resources | |
| M7 | Similar words extraction from thesaurus and Synset for sentiment words |
| M8 | Similar words extraction from thesaurus and Word2Vec for sentiment words |
| M9 | Similar words extraction from Synset and Word2Vec for sentiment words |
| M10 | Similar words extraction from thesaurus, Synset and Word2Vec for sentiment words |
| M11 | Similar words extraction from thesaurus for sentiment words and from Synset for non-sentiment words |
| M12 | Similar words extraction from thesaurus for sentiment words and from Word2Vec for non-sentiment words |
| M13 | Similar words extraction from thesaurus for sentiment words and Synset and Word2Vec for non-sentiment words |
Actual and predicted class
| Predicted | Actual | |
|---|---|---|
| Positive class | Negative class | |
| Positive class | TP | FP |
| Negative class | FN | TN |
Sample sentiment words in thesaurus
| Words | Synonyms |
|---|---|
| Beautiful | 'alluring', 'appealing', 'charming', 'cute', 'dazzling', 'delicate', 'delightful', 'elegant', 'exquisite', 'fascinating', 'fine', 'good-looking', 'gorgeous', 'graceful', 'grand', 'handsome', 'lovely', 'magnificent', 'marvelous', 'pleasing', 'pretty', 'splendid', 'stunning', 'superb', 'wonderful', 'admirable', 'angelic', 'beauteous', 'bewitching', 'classy', 'comely', 'divine', 'enticing', 'excellent', 'fair', 'foxy', 'ideal', 'nice', 'pulchritudinous', 'radiant', 'ravishing', 'refined', 'resplendent', 'shapely', 'sightly', 'statuesque', 'sublime', 'symmetrical', 'taking', 'well-formed' |
| Better | 'exceptional', 'improved', 'superior', 'choice', 'exceeding', 'fitter', 'preferred', 'sophisticated', 'surpassing', 'bigger', 'finer', 'greater', 'larger', 'more desirable', 'more suitable', 'more valuable', 'preferable', 'prominent' |
| Disagree | 'clash', 'contradict', 'differ', 'dissent', 'diverge', 'conflict', 'counter', 'depart', 'deviate', 'discord', 'disharmonize', 'vary', 'war', 'be dissimilar' |
| Fail | 'decline', 'fall', 'abozt', 'backslide', 'blunder', 'deteriorate', 'fizzle', 'flop', 'flounder', 'fold', 'founder', 'miscarry', 'miss', 'slip', 'be defeated', 'be found lacking', 'be ruined', 'come to nothing', 'fall short', 'go astray', 'go down swinging', 'go up in smoke', 'hit bottom', 'lose control', 'lose status', 'miss the boat', 'run aground' |
Semantically similar retrieved using Synset and Word2Vec
| Words | Synset | Word2Vec |
|---|---|---|
| Picture | 'picture', 'image', 'impression', 'scene', 'movie', 'film', 'pic', 'video', 'photograph', 'photo', 'fancy', 'see', 'figure', 'show' | 'pictures', 'photograph', 'photo', 'photos', 'images', 'image' |
| Guess | 'guess', 'guessing', 'think', 'suppose', 'imagine', 'pretend' | 'suppose', 'think', 'yeah', 'maybe', 'probably', 'anyway', 'know', 'hey' |
| Place | 'place', 'position', 'shoes', 'home', 'post', 'office', 'situation', 'space', 'put', 'set', 'lay', 'rate', 'range', 'order', 'grade', 'locate', 'site', 'point', 'send' | 'places', 'placed', 'finish' |
| Thing | 'thing', 'matter' | 'things', 'something', 'stuff', 'really', 'think', ‘aspect’, 'reason', 'kind' |
| Sleep | 'sleep', 'slumber', 'sopor', 'nap', 'rest', 'eternal_rest', 'eternal_sleep', 'quietus', 'kip' | 'sleeping', 'restful_sleep', 'restorative_sleep', 'slept', 'nap', 'wakings', 'naps', 'fitful_sleep', 'Sleep', 'doze’ |
| Happy | 'happy', 'felicitous', 'glad', 'well-chosen' | 'glad', 'pleased', 'ecstatic', 'overjoyed', 'thrilled', 'satisfied', 'delighted’, 'disappointed', 'excited' |
| Sad | 'sad', 'deplorable', 'distressing', 'lamentable', 'pitiful', 'sorry' | 'saddening', 'Sad’, 'saddened', 'heartbreaking', 'disheartening', 'saddens_me', 'distressing', `reminders_bobbing' |
| Good | 'good’, ‘goodness', 'commodity', 'trade_good’, ‘full’, 'estimable', 'honorable', 'respectable', 'beneficial', 'just', 'upright', 'adept', 'expert', 'practiced', 'proficient', 'skillful', 'skilful', 'dear', 'near', 'dependable', 'safe', 'secure', 'right', 'ripe', 'well', 'effective', 'in_effect', 'in_force', 'serious', 'sound', 'salutary', 'honest', 'undecomposed', 'unspoiled', 'unspoilt’, ‘thoroughly', 'soundly' | 'great', 'terrific', 'decent', 'nice', 'excellent', 'fantastic', 'better', 'solid', 'lousy' |
| Bad | 'bad', 'badness', 'big', 'tough’, ‘spoiled', 'spoilt’, ‘regretful', 'sorry’, ‘uncollectible', 'risky’, ‘high-risk', 'speculative', 'unfit', 'unsound’, ‘forged', 'defective', 'badly' | 'good', 'terrible', 'horrible', 'Bad', 'lousy', 'crummy', 'horrid', 'awful', 'dreadful', 'horrendous' |
Sample bigram combinations using similar words
| Existing bigram | Sample new bigrams |
|---|---|
| ('phone', 'good') | ('phone', 'beneficial') ('phone', 'sound') ('phone', 'effective') ('telephone', 'beneficial') ('telephone', 'sound') ('telephone', 'effective') ('headphone', 'beneficial') ('headphone', 'sound') ('headphone', 'effective') ('earphone', 'beneficial') ('earphone', 'sound') ('earphone', 'effective') ('telephone', 'good') ('headphone', 'sound') ('earphone', 'sound') |
| ('good', 'price') | ('beneficial', 'cost') ('beneficial', 'price') ('sound', 'cost') ('sound', 'price') ('effective', 'cost') ('effective', 'price') |
| ('price', 'paid,') | ('cost', 'paid,') ('cost', 'give,') ('cost', 'pay,') ('cost', 'devote,') ('cost', 'bear,') ('price', 'give,') ('price', 'pay,') ('price', 'devote,') ('price', 'bear,') |
| ('battery', 'lasts') | ('battery', 'survive') ('barrage', 'survive') ('barrage', 'lasts') ('battery', 'live') ('barrage', 'live') |
| ('camera', 'decent') | ('camera', 'nice') ('camera', 'adequate') ('camera', 'enough') ('camera', 'properly') ('camera', 'right') |
Fig. 1Number of words for which similar words returned by Thesaurus, Synset and Word2Vec
Number of words for which similar words retrieved
| Dataset | Total number of words | Therasus + Synset | Therasus + Word2Vec |
|---|---|---|---|
| Apparel | 5924 | 1306 | 1485 |
| Books | 14,576 | 2993 | 4666 |
| DVD | 17,451 | 3180 | 5136 |
| Electronics | 11,628 | 2070 | 3146 |
| Health | 8274 | 1574 | 2548 |
| Kitchen | 8199 | 1703 | 2525 |
| Music | 12,067 | 2185 | 3745 |
| Sports | 9818 | 1834 | 3003 |
| Toys | 9416 | 1692 | 2752 |
| Video | 13,970 | 2593 | 4456 |
Accuracy of M1–M6
| Dataset | Accuracy | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 70.5 | 75 | 74 | 71 | 72.5 | 73 | M2 (75%) |
| Books | 70 | 74.5 | 75.5 | 76 | 75.5 | 73 | M4 (76%) |
| DVD | 66 | 74 | 68.5 | 70.5 | 74.5 | 67 | M5 (74.5%) |
| Electronics | 75 | 79 | 74 | 76 | 78.5 | 73 | M2 (79%) |
| Health | 74 | 79 | 77 | 77 | 77.5 | 75.5 | M2 (79%) |
| Kitchen | 72 | 77.5 | 76 | 73.5 | 74 | 71.5 | M2 (77.5%) |
| Music | 65.5 | 72 | 70 | 68.5 | 71.5 | 67 | M2 (72%) |
| Sports | 68 | 73.5 | 74 | 73 | 72.5 | 67 | M3 (74%) |
| Toys | 68 | 67 | 71.5 | 70.5 | 69 | 70.5 | M3 (71.5%) |
| Video | 70.5 | 78.5 | 73.5 | 74 | 75.5 | 71 | M2 (78.5%) |
Words and sample meaningless synonyms returned by Word2Vec
| Word | Useless words returned by Word2Vec |
| Guess | “hey”, “know” |
| Thing | “really”, “think”, “kind” |
| See | “expect”, “imagine” |
| Came | “got”, “gave”, “ran” |
| Supposed | “going”, “trying”, ”wanted”, “not”, “want” |
Accuracy of M7–M13
| Dataset | Accuracy | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 69 | 69 | 68 | 69 | 79 | 77 | 70.5 | M11 (79%) |
| Books | 78.5 | 78 | 79.5 | 80 | 76 | 74 | 74 | M10 (80%) |
| DVD | 75 | 77.5 | 76.5 | 76 | 74.5 | 70 | 70 | M8 (77.5%) |
| Electronics | 75 | 75.5 | 72.5 | 72 | 79 | 76 | 67.5 | M11 (79%) |
| Health | 76.5 | 79 | 77 | 76 | 80 | 76.5 | 70.5 | M11 (80%) |
| Kitchen | 74 | 72.5 | 77 | 73.5 | 80.5 | 78.5 | 67.5 | M11 (80.5%) |
| Music | 65.5 | 66 | 68 | 64.5 | 73.5 | 71 | 69 | M11 (73.5%) |
| Sports | 69.5 | 69.5 | 71 | 69.5 | 74 | 72 | 71.5 | M11 (74%) |
| Toys | 74 | 73 | 72 | 72.5 | 69.5 | 73.5 | 69 | M7 (74%) |
| Video | 73.5 | 74.5 | 75.5 | 74.5 | 78 | 73.5 | 73 | M11 (78%) |
Analysis of Tables 7 and 8
| Dataset | Winner from Table | Winner from Table | Overall winner |
|---|---|---|---|
| Apparel | M2 (75%) | M11 (79%) | M11 |
| Books | M4 (76%) | M10 (80%) | M10 |
| DVD | M5 (74.5%) | M8 (77.5%) | M8 |
| Electronics | M2 (79%) | M11 (79%) | M2 and M11 |
| Health | M2 (79%) | M11 (80%) | M11 |
| Kitchen | M2 (77.5%) | M11 (80.5%) | M11 |
| Music | M2 (72%) | M11 (73.5%) | M11 |
| Sports | M3 (74%) | M11 (74%) | M3 and M11 |
| Toys | M3 (71.5%) | M7 (74%) | M7 |
| Video | M2 (78.5%) | M11 (78%) | M2 |
Precision of M1–M6
| Dataset | Precision | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 0.695238 | 0.727273 | 0.718182 | 0.669355 | 0.68595 | 0.7053571 | M2 (0.727273) |
| Books | 0.75 | 0.752577 | 0.831169 | 0.788889 | 0.78022 | 0.75 | M3 (0.831169) |
| DVD | 0.728571 | 0.772727 | 0.717647 | 0.766234 | 0.775281 | 0.765625 | M5 (0.775281) |
| Electronics | 0.797619 | 0.795918 | 0.76087 | 0.765306 | 0.776699 | 0.7613636 | M1 (0.797619) |
| Health | 0.815789 | 0.808511 | 0.793478 | 0.787234 | 0.772277 | 0.8 | M1 (0.815789) |
| Kitchen | 0.72 | 0.752294 | 0.745283 | 0.715596 | 0.706897 | 0.7087379 | M2 (0.752294) |
| Music | 0.653465 | 0.683333 | 0.675439 | 0.67619 | 0.674797 | 0.6770833 | M2 (0.683333) |
| Sports | 0.676471 | 0.697479 | 0.714286 | 0.694915 | 0.669173 | 0.6416667 | M3 (0.714286) |
| Toys | 0.691489 | 0.67 | 0.708738 | 0.699029 | 0.669643 | 0.7070707 | M3 (0.708738) |
| Video | 0.71134 | 0.761468 | 0.72381 | 0.726415 | 0.721739 | 0.71875 | M2 (0.761468) |
Precision of M7–M13
| Dataset | Precision | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 0.6376812 | 0.6376812 | 0.6343284 | 0.6357143 | 0.768518 | 0.79347826 | 0.6752137 | M12 (0.79347826) |
| Books | 0.728 | 0.7372881 | 0.7398374 | 0.734375 | 0.795454 | 0.84285714 | 0.7926829 | M12 (0.84285714) |
| DVD | 0.7083333 | 0.7477477 | 0.7226891 | 0.7063492 | 0.810127 | 0.8125 | 0.7272727 | M12 (0.8125) |
| Electronics | 0.6984127 | 0.7107438 | 0.6771654 | 0.6641791 | 0.822222 | 0.825 | 0.7058824 | M12 (0.825) |
| Health | 0.7190083 | 0.7457627 | 0.725 | 0.7063492 | 0.840909 | 0.8630137 | 0.7303371 | M12 (0.8630137) |
| Kitchen | 0.673913 | 0.6642336 | 0.7045455 | 0.6666667 | 0.808080 | 0.82022472 | 0.6767677 | M12 (0.82022472) |
| Music | 0.602649 | 0.6111111 | 0.6267606 | 0.5947712 | 0.707965 | 0.73333333 | 0.6727273 | M12 (0.73333333) |
| Sports | 0.6444444 | 0.6444444 | 0.6544118 | 0.6402878 | 0.706897 | 0.75581395 | 0.6806723 | M12 (0.75581395) |
| Toys | 0.6904762 | 0.6916667 | 0.6692308 | 0.6717557 | 0.696970 | 0.79012346 | 0.6862745 | M12 (0.79012346) |
| Video | 0.6850394 | 0.699187 | 0.704 | 0.6899225 | 0.769231 | 0.79012346 | 0.7053571 | M12 (0.79012346) |
Recall of M1–M6
| Dataset | Recall | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 0.73 | 0.8 | 0.79 | 0.83 | 0.83 | 0.79 | M4 and M5 (0.83) |
| Books | 0.6 | 0.73 | 0.64 | 0.71 | 0.71 | 0.6 | M2 (0.73) |
| DVD | 0.51 | 0.68 | 0.61 | 0.59 | 0.69 | 0.49 | M5 (0.69) |
| Electronics | 0.67 | 0.78 | 0.7 | 0.75 | 0.8 | 0.67 | M5 (0.8) |
| Health | 0.62 | 0.76 | 0.73 | 0.74 | 0.78 | 0.68 | M5 (0.78) |
| Kitchen | 0.72 | 0.82 | 0.79 | 0.78 | 0.82 | 0.73 | M2 and M5 (0.82) |
| Music | 0.66 | 0.82 | 0.77 | 0.71 | 0.83 | 0.65 | M5 (0.83) |
| Sports | 0.69 | 0.83 | 0.8 | 0.82 | 0.89 | 0.77 | M5 (0.89) |
| Toys | 0.65 | 0.67 | 0.73 | 0.72 | 0.75 | 0.7 | M5 (0.75) |
| Video | 0.69 | 0.83 | 0.76 | 0.77 | 0.83 | 0.69 | M2 and M5 (0.83) |
Recall of M7–M13
| Dataset | Recall | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 0.88 | 0.88 | 0.85 | 0.89 | 0.83 | 0.73 | 0.79 | M10 (0.89) |
| Books | 0.91 | 0.87 | 0.91 | 0.94 | 0.7 | 0.59 | 0.65 | M10 (0.94) |
| DVD | 0.85 | 0.83 | 0.86 | 0.89 | 0.64 | 0.52 | 0.64 | M10 (0.89) |
| Electronics | 0.88 | 0.86 | 0.86 | 0.89 | 0.74 | 0.66 | 0.6 | M10 (0.89) |
| Health | 0.87 | 0.88 | 0.87 | 0.89 | 0.74 | 0.63 | 0.65 | M10 (0.89) |
| Kitchen | 0.93 | 0.91 | 0.93 | 0.94 | 0.8 | 0.73 | 0.67 | M10 (0.94) |
| Music | 0.91 | 0.88 | 0.89 | 0.91 | 0.8 | 0.66 | 0.74 | M7 and M10 (0.91) |
| Sports | 0.87 | 0.87 | 0.89 | 0.89 | 0.82 | 0.65 | 0.81 | M9 and M10 (0.89) |
| Toys | 0.87 | 0.83 | 0.87 | 0.88 | 0.69 | 0.64 | 0.7 | M10 (0.88) |
| Video | 0.87 | 0.86 | 0.88 | 0.89 | 0.8 | 0.64 | 0.79 | M10 (0.89) |
F-measure of M1–M6
| Dataset | F-measure | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 0.712195122 | 0.761904762 | 0.752380952 | 0.741071429 | 0.751131222 | 0.74528302 | M2 (0.761904762) |
| Books | 0.666666667 | 0.741116751 | 0.723163842 | 0.747368421 | 0.743455497 | 0.66666667 | M4 (0.747368421) |
| DVD | 0.6 | 0.723404255 | 0.659459459 | 0.666666667 | 0.73015873 | 0.59756098 | M5 (0.73015873) |
| Electronics | 0.72826087 | 0.787878788 | 0.729166667 | 0.757575758 | 0.78817734 | 0.71276596 | M5 (0.78817734) |
| Health | 0.704545455 | 0.783505155 | 0.760416667 | 0.762886598 | 0.776119403 | 0.73513514 | M2 (0.783505155) |
| Kitchen | 0.72 | 0.784688995 | 0.766990291 | 0.746411483 | 0.759259259 | 0.71921182 | M2 (0.784688995) |
| Music | 0.656716418 | 0.745454545 | 0.719626168 | 0.692682927 | 0.744394619 | 0.66326531 | M2 (0.745454545) |
| Sports | 0.683168317 | 0.757990868 | 0.754716981 | 0.752293578 | 0.763948498 | 0.7 | M2 (0.757990868) |
| Toys | 0.670103093 | 0.67 | 0.719211823 | 0.709359606 | 0.70754717 | 0.70351759 | M3 (0.719211823) |
| Video | 0.700507614 | 0.794258373 | 0.741463415 | 0.747572816 | 0.772093023 | 0.70408163 | M2 (0.794258373) |
F-measure of M7–M13
| Dataset | F-measure | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 0.7394958 | 0.7394958 | 0.72649573 | 0.74166667 | 0.798076923 | 0.760416667 | 0.7281106 | M11 (0.798076923) |
| Books | 0.80888889 | 0.79816514 | 0.8161435 | 0.8245614 | 0.744680851 | 0.694117647 | 0.71428571 | M10 (0.8245614) |
| DVD | 0.77272727 | 0.78672986 | 0.78538813 | 0.78761062 | 0.715083799 | 0.634146341 | 0.68085106 | M10 (0.78761062) |
| Electronics | 0.77876106 | 0.77828054 | 0.75770925 | 0.76068376 | 0.778947368 | 0.733333333 | 0.64864865 | M11 (0.778947368) |
| Health | 0.78733032 | 0.80733945 | 0.79090909 | 0.78761062 | 0.787234043 | 0.728323699 | 0.68783069 | M8 (0.80733945) |
| Kitchen | 0.78151261 | 0.76793249 | 0.80172414 | 0.78008299 | 0.804020101 | 0.772486772 | 0.67336683 | M11 (0.804020101) |
| Music | 0.7250996 | 0.72131148 | 0.73553719 | 0.71936759 | 0.751173709 | 0.694736842 | 0.7047619 | M11 (0.751173709) |
| Sports | 0.74042553 | 0.74042553 | 0.75423729 | 0.74476987 | 0.759259259 | 0.698924731 | 0.73972603 | M11 (0.759259259) |
| Toys | 0.7699115 | 0.75454545 | 0.75652174 | 0.76190476 | 0.693467337 | 0.70718232 | 0.69306931 | M7 (0.7699115) |
| Video | 0.76651982 | 0.77130045 | 0.78222222 | 0.77729258 | 0.784313725 | 0.70718232 | 0.74528302 | M11 (0.784313725) |
Analysis of Tables 15 and 16
| Dataset | Winner from Table | Winner from Table | Overall winner |
|---|---|---|---|
| Apparel | M2 (0.761904762) | M11 (0.798076923) | M11 |
| Books | M4 (0.747368421) | M10 (0.8245614) | M10 |
| DVD | M5 (0.73015873) | M10 (0.78761062) | M10 |
| Electronics | M5 (0.78817734) | M11 (0.778947368) | M5 |
| Health | M2 (0.783505155) | M8 (0.80733945) | M8 |
| Kitchen | M2 (0.784688995) | M11 (0.804020101) | M11 |
| Music | M2 (0.745454545) | M11 (0.751173709) | M11 |
| Sports | M2 (0.757990868) | M11 (0.759259259) | M11 |
| Toys | M3 (0.719211823) | M7 (0.7699115) | M7 |
| Video | M2 (0.794258373) | M11 (0.784313725) | M2 |
Completeness of M1–M6
| Dataset | Completeness | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 0.365 | 0.4 | 0.395 | 0.415 | 0.415 | 0.395 | M4 and M5 (0.415) |
| Books | 0.3 | 0.365 | 0.32 | 0.355 | 0.355 | 0.3 | M2 (0.365) |
| DVD | 0.255 | 0.34 | 0.305 | 0.295 | 0.345 | 0.245 | M5 (0.345) |
| Electronics | 0.335 | 0.39 | 0.35 | 0.375 | 0.4 | 0.335 | M5 (0.4) |
| Health | 0.31 | 0.38 | 0.365 | 0.37 | 0.39 | 0.34 | M5 (0.39) |
| Kitchen | 0.36 | 0.41 | 0.395 | 0.39 | 0.41 | 0.365 | M2 and M5 (0.41) |
| Music | 0.33 | 0.41 | 0.385 | 0.355 | 0.415 | 0.325 | M5 (0.415) |
| Sports | 0.345 | 0.415 | 0.4 | 0.41 | 0.445 | 0.385 | M5 (0.445) |
| Toys | 0.325 | 0.335 | 0.365 | 0.36 | 0.375 | 0.35 | M5 (0.375) |
| Video | 0.345 | 0.415 | 0.38 | 0.385 | 0.415 | 0.345 | M2 and M5 (0.415) |
Completeness of M7–M13
| Dataset | Completeness | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 0.44 | 0.44 | 0.425 | 0.445 | 0.415 | 0.365 | 0.395 | M10 (0.445) |
| Books | 0.455 | 0.435 | 0.455 | 0.47 | 0.35 | 0.295 | 0.325 | M10 (0.47) |
| DVD | 0.425 | 0.415 | 0.43 | 0.445 | 0.32 | 0.26 | 0.32 | M10 (0.445) |
| Electronics | 0.44 | 0.43 | 0.43 | 0.445 | 0.32 | 0.33 | 0.3 | M10 (0.445) |
| Health | 0.435 | 0.44 | 0.435 | 0.445 | 0.37 | 0.315 | 0.325 | M10 (0.445) |
| Kitchen | 0.465 | 0.455 | 0.465 | 0.47 | 0.37 | 0.365 | 0.335 | M10 (0.47) |
| Music | 0.455 | 0.44 | 0.445 | 0.455 | 0.4 | 0.33 | 0.37 | M7 and M10 (0.455) |
| Sports | 0.435 | 0.435 | 0.445 | 0.445 | 0.41 | 0.325 | 0.405 | M9 and M10 (0.445) |
| Toys | 0.435 | 0.415 | 0.435 | 0.44 | 0.345 | 0.32 | 0.35 | M10 (0.44) |
| Video | 0.435 | 0.43 | 0.44 | 0.445 | 0.4 | 0.32 | 0.395 | M10 (0.445) |
Analysis of Tables 18 and 19
| Dataset | Winner from Table | Winner from Table | Overall winner |
|---|---|---|---|
| Apparel | M4 and M5 (0.415) | M10 (0.445) | M10 |
| Books | M2 (0.365) | M10 (0.47) | M10 |
| DVD | M5 (0.345) | M10 (0.445) | M10 |
| Electronics | M5 (0.4) | M10 (0.445) | M10 |
| Health | M5 (0.39) | M10 (0.445) | M10 |
| Kitchen | M2 and M5 (0.41) | M10 (0.47) | M10 |
| Music | M5 (0.415) | M7 and M10 (0.455) | M7 and M10 |
| Sports | M5 (0.445) | M9 and M10 (0.445) | M5, M9 and M10 |
| Toys | M5 (0.375) | M10 (0.44) | M10 |
| Video | M2 and M5 (0.415) | M10 (0.445) | M10 |
False alarm rate of M1–M6
| Dataset | False alarm rate | Winner | |||||
|---|---|---|---|---|---|---|---|
| M1 | M2 | M3 | M4 | M5 | M6 | ||
| Apparel | 0.27 | 0.2 | 0.21 | 0.17 | 0.17 | 0.21 | M4 and M5 (0.17) |
| Books | 0.4 | 0.27 | 0.36 | 0.29 | 0.29 | 0.4 | M2 (0.27) |
| DVD | 0.49 | 0.32 | 0.39 | 0.41 | 0.31 | 0.51 | M5 (0.31) |
| Electronics | 0.33 | 0.22 | 0.3 | 0.25 | 0.2 | 0.33 | M5 (0.2) |
| Health | 0.38 | 0.24 | 0.27 | 0.26 | 0.22 | 0.32 | M5 (0.22) |
| Kitchen | 0.28 | 0.18 | 0.21 | 0.22 | 0.18 | 0.27 | M2 and M5 (0.18) |
| Music | 0.34 | 0.18 | 0.23 | 0.29 | 0.17 | 0.35 | M5 (0.17) |
| Sports | 0.31 | 0.17 | 0.2 | 0.18 | 0.11 | 0.23 | M5 (0.11) |
| Toys | 0.35 | 0.33 | 0.27 | 0.28 | 0.25 | 0.3 | M5 (0.25) |
| Video | 0.31 | 0.17 | 0.24 | 0.23 | 0.17 | 0.31 | M2 and M5 (0.17) |
False alarm rate of M7–M13
| Dataset | False alarm rate | Winner | ||||||
|---|---|---|---|---|---|---|---|---|
| M7 | M8 | M9 | M10 | M11 | M12 | M13 | ||
| Apparel | 0.12 | 0.12 | 0.15 | 0.11 | 0.17 | 0.27 | 0.21 | M10 (0.11) |
| Books | 0.09 | 0.13 | 0.09 | 0.06 | 0.3 | 0.41 | 0.35 | M10 (0.06) |
| DVD | 0.15 | 0.17 | 0.14 | 0.11 | 0.36 | 0.48 | 0.36 | M10 (0.11) |
| Electronics | 0.12 | 0.14 | 0.14 | 0.11 | 0.26 | 0.34 | 0.4 | M10 (0.11) |
| Health | 0.13 | 0.12 | 0.13 | 0.11 | 0.26 | 0.37 | 0.35 | M10 (0.11) |
| Kitchen | 0.07 | 0.09 | 0.07 | 0.06 | 0.2 | 0.27 | 0.33 | M10 (0.06) |
| Music | 0.09 | 0.12 | 0.11 | 0.09 | 0.2 | 0.34 | 0.26 | M7 and M10 (0.09) |
| Sports | 0.13 | 0.13 | 0.11 | 0.11 | 0.18 | 0.35 | 0.19 | M9 and M10 (0.11) |
| Toys | 0.13 | 0.17 | 0.13 | 0.12 | 0.31 | 0.36 | 0.3 | M10 (0.12) |
| Video | 0.13 | 0.14 | 0.12 | 0.11 | 0.2 | 0.36 | 0.21 | M10 (0.11) |
Analysis of Tables 21 and 22
| Dataset | Winner from Table | Winner from Table | Overall winner |
|---|---|---|---|
| Apparel | M4 and M5 (0.17) | M10 (0.11) | M10 |
| Books | M2 (0.27) | M10 (0.06) | M10 |
| DVD | M5 (0.31) | M10 (0.11) | M10 |
| Electronics | M5 (0.2) | M10 (0.11) | M10 |
| Health | M5 (0.22) | M10 (0.11) | M10 |
| Kitchen | M2 and M5 (0.18) | M10 (0.06) | M10 |
| Music | M5 (0.17) | M7 and M10 (0.09) | M7 and M10 |
| Sports | M5 (0.11) | M9 and M10 (0.11) | M5, M9 and M10 |
| Toys | M5 (0.25) | M10 (0.12) | M10 |
| Video | M2 and M5 (0.17) | M10 (0.11) | M10 |
Fig. 2Comparison of proposed models with NB using accuracy
Fig. 3Comparison of proposed models with NB using precision
Fig. 4Comparison of proposed models with NB using recall
Fig. 5Comparison of proposed models with NB using F-measure
Fig. 6Comparison of proposed models with NB using completeness
Fig. 7Comparison of proposed models with NB using false alarm rate
Fig. 8Accuracy of 5 runs or CNN for 10 datasets
Fig. 9Comparison of proposed models with CNN using accuracy
Fig. 10Comparison of proposed models with CNN using precision
Fig. 11Comparison of proposed models with CNN using recall
Fig. 12Comparison of proposed models with CNN using F-measure
Fig. 13Comparison of proposed models with CNN using completeness
Fig. 14Comparison of proposed models with CNN using false alarm rate