| Literature DB >> 24255824 |
Kanako Komiya1, Yuji Abe, Hajime Morita, Yoshiyuki Kotani.
Abstract
Question Answering (QA) is a task of answering natural language questions with adequate sentences. This paper proposes two methods to improve the performance of the QA system using a Q&A site corpus. The first method is for the relevant document retrieval module. We proposed modification of measure of mutual information for the query expansion; we calculate it between two words in each question and a word in its answer in the Q&A site corpus not to choose the words that are not suitable. The second method is for the candidate answer evaluation module. We proposed to evaluate candidate answers using the two measures together, i.e., the Web relevance score and the translation probability. The experiments were carried out using a Japanese Q&A site corpus. They revealed that the first proposed method was significantly better than the original method when their accuracies and MRR (Mean Reciprocal Rank) were compared and the second method was significantly better than the original methods when their MRR were compared.Entities:
Year: 2013 PMID: 24255824 PMCID: PMC3825096 DOI: 10.1186/2193-1801-2-396
Source DB: PubMed Journal: Springerplus ISSN: 2193-1801
Figure 1Outlines of Web document retrieval and candidate answer evaluation with and without query expansion. This figure shows the outlines of the Web document retrieval and the candidate answer evaluation with and without query expansion. D1∪D2 and D3 each represents the candidate answers with and without query expansion. The number of the documents of D1∪D2 was equalized to that of D3 for the fair comparison.
Results of experiments of query expansion
| Without query | Original method | Proposed method | |
|---|---|---|---|
| expansion | |||
| Accuracy | 0.420 | 0.400 | 0.450 |
| MRR | 0.262 | 0.233 | 0.273 |
This table summarizes the top-5 accuracies and MRR of the systems for the experiments of the query expansion. Original method in the table represents the method proposed by Berger et al. (2000), where the words to be added are chosen based on mutual information between a word from a question and another word in its answer. This table indicates that the system with the proposed method outperformed the two systems: the system without query expansion and the system with the method proposed by Berger et al. (2000).
Figure 2Performance of proposed method. This figure shows the top-5 accuracies and MRR of the experiments of the candidate answer evaluation when the value of γ changed from 0 to 1. The top-5 accuracy was maximized to 0.59 when γ = 0.93 and the MRR was maximized to 0.461 when γ =0.98.
Results of experiments of answer candidate evaluation
| Top-5 accuracy | MRR | |
|---|---|---|
| Only Web relevanve ( | 0.55 | 0.423 |
| Only translation probability ( | 0.49 | 0.318 |
| Proposed method ( | 0.59 | 0.395 |
| Proposed method ( | 0.57 | 0.461 |
This table summarizes the top-5 accuracies and MRR of the systems for the experiments of the candidate answer evaluation. As for MRR, the proposed method was significantly better than the original methods according to a Wilcoxon signed-rank test.
Examples of translation probability
| Index | Given word | 1st | 2nd | 3rd | 4th | 5th |
|---|---|---|---|---|---|---|
| (1) |
|
|
|
|
|
|
| Medical care | Medical care | Hospital | Fare | Admission | Operation | |
| (2) |
|
|
|
|
|
|
| Lawsuit | Lawsuit | Judgment | Sue over | Advocate | Right | |
| (3) |
|
|
|
|
|
|
| Salt water | Salt water | Water | Tastes salty | Method | Shell | |
| (4) |
|
|
|
|
|
|
| Landform | Landform | Yokohama | Times | Typhoon | Geography | |
| (5) |
|
|
|
|
|
|
| Channel | Channel | Island | World | Takeshima | Tohoku | |
| (6) |
|
|
|
|
|
|
| Arrestment | Arrestment | PASSIVE | Get caught | Do | PAST | |
| (7) |
|
| ? (.031) |
|
|
|
| Cell | Cell | ? | PREDICATION | Why | Human | |
| (8) |
|
|
|
|
|
|
| Information | Information | Of | PREDICATION | Do | AGENT | |
| (9) |
|
|
|
|
| ?(.038) |
| Technology | Technology | Of | TOPIC MARKER | PREDICATION | ? | |
| (10) |
|
|
|
|
| ?(.021) |
| President | President | America | Bush | Of | ? | |
| (11) |
|
|
|
|
|
|
| Prime minister | Prime minister | Koizumi | Minister | Mr. | Prime minister | |
| (12) |
|
| ?(.170) |
|
|
|
| Shepherd’s-purse | Seven herbs | ? | As for | Of | QUESTION | |
| (13) |
| ?(.064) |
|
|
|
|
| Because | ? | Of | PREDICATION | TOPIC MARKER | QUESTION | |
| (14) |
|
|
| ? (.056) |
|
|
| Because | Of | PREDICATION | ? | QUESTION | AGENT | |
| (15) |
|
|
| ? (.055) |
|
|
| For | Of | PREDICATION | ? | QUESTION | For | |
| (16) |
|
|
|
|
|
|
| Reason | Reason | Why | AGENT | Of | QUESTION |
This table has examples of the top-5 words that maximize P(q|a), which is the translation probability from a word a in an answer to a word q in a question when a is given. The English words and the numbers in brackets are the English translations and the translation probabilities, respectively. The functions of Japanese words are shown when the English words are written in upper case.