| Literature DB >> 35360634 |
Abstract
In order to improve text reading ability, a human-computer interaction method based on artificial intelligence (AI) human-computer interaction is proposed. Firstly, the design of the AI human-computer interaction model is constructed, which includes the Stanford Question Answering Dataset (SQuAD) and the designed baseline model. There are three components: the coding layer is based on a cyclic neural network (recurrent neural network [RNN] encoder layer), which aims to encode the problem and text into a hidden state; the interaction layer is used to integrate problems and text representation; the output layer connects two independent soft Max layers after a fully connected layer, one is used to obtain the starting position of the answer in the text and the other is used to obtain the ending position. In the interaction layer of the model, this manuscript uses hierarchical attention and aggregation mechanism to improve text coding. The traditional model interaction layer has a simple structure, which leads to weak relevance between text and problems, and poor understanding ability of the model. Finally, the self-attention model is used to further enhance the feature representation of text. The experimental results show that the improved model in this manuscript is compared with the public AI human-computer interaction reading comprehension model. According to the data in the table, the accuracy of the model in this manuscript is better than that of the baseline model, in which the exact match (EM) value is increased by 1.4% and the F1 value is increased by 2.7%. However, compared with improvement point 2, the EM and F1 values of the model have decreased by 0.7%. It shows that the output layer has a certain impact on the effect of the model, and the improvement and optimization of the output layer can also improve the performance of the model. It is proved that AI human-computer interaction can effectively improve text reading ability.Entities:
Keywords: AI human-computer interaction; SQuAD dataset; attention mechanism; neural network; reading comprehension
Year: 2022 PMID: 35360634 PMCID: PMC8963353 DOI: 10.3389/fpsyg.2022.853066
Source DB: PubMed Journal: Front Psychol ISSN: 1664-1078
FIGURE 1Artificial intelligence (AI)-human machine interaction.
FIGURE 2Schematic diagram of baseline model architecture.
FIGURE 3Schematic diagram of the convolution operation.
Baseline model evaluation results.
| Method | EM | F1 |
| Original paper effect | 66.8% | 80.6% |
| This paper realizes the effect | 72.8% | 81.5% |
Comparative experimental results based on Bidirectional Encoder Representation from Transformers (BERT).
| Method | EM | F1 |
| Original effect | 69.3% | 80.6% |
| Reproduction effect | 72.4% | 83.5% |
| Improved results based on BERT | 72.6% | 85.7% |
Comparison of experimental results based on word vector and context coding.
| Method | EM | F1 |
| Original effect | 72.6% | 80.7% |
| Reproducing results | 71.6% | 81.9% |
| Coding results based on word vector and context | 73.6% | 83.7% |
FIGURE 4Text length distribution.
FIGURE 5Problem length distribution.
FIGURE 6Answer length distribution.
Baseline model parameters.
| Super parameter | Numerical value |
| Random inactivation | 0.15 |
| Text length N | 300 |
| Problem length M | 30 |
| Word vector dimension d | 100 |
| GRU hidden state dimension H (RNN coding layer) | 150 |
| GRU hidden state dimension H (interaction layer) | 150 |
| Sliding window length w | 15 |
Super parameters of the improved model.
| Super parameter | Numerical value |
| Convolution layer number K | 6 |
| Number of convolution layer filters l | 50 |
| Convolution layer filter width k | 2, 3, 4, 5, 6, 7 |
Effect verification of output layer based on sliding window prediction method.
| Prediction mode | F1 | EM |
| Not based on sliding window | 68.7 | 58.0 |
| Based on sliding window | 70.7 | 59.7 |
Comparison of exact match (EM) and F1 scores of baseline model and improved model.
| Coding layer | F1 | EM |
| RNN | 70.7 | 59.7 |
| CNN | 66.9 | 54.9 |
FIGURE 7Comparison of exact match (EM) scores of two models on different length answers.
Comparison of experimental results based on attention and aggregation mechanism.
| Method | EM | F1 |
| Original effect | 72% | 79.9% |
| Reproduction effect | 70.9% | 80.6% |
| Results based on word vector and context coding | 71.9% | 83.7% |
| Results based on attention and aggregation mechanism | 76.7% | 89.8% |