| Literature DB >> 34306049 |
Abstract
Deep learning is the latest trend of machine learning and artificial intelligence research. As a new field with rapid development over the past decade, it has attracted more and more researchers' attention. Convolutional Neural Network (CNN) model is one of the most important classical structures in deep learning models, and its performance has been gradually improved in deep learning tasks in recent years. Convolutional neural networks have been widely used in image classification, target detection, semantic segmentation, and natural language processing because they can automatically learn the feature representation of sample data. Firstly, this paper analyzes the model structure of a typical convolutional neural network model to increase the network depth and width in order to improve its performance, analyzes the network structure that further improves the model performance by using the attention mechanism, and then summarizes and analyzes the current special model structure. In order to further improve the text language processing effect, a convolutional neural network model, Hybrid convolutional neural network (CNN), and Long Short-Term Memory (LSTM) based on the fusion of text features and language knowledge are proposed. The text features and language knowledge are integrated into the language processing model, and the accuracy of the text language processing model is improved by parameter optimization. Experimental results on data sets show that the accuracy of the proposed model reaches 93.0%, which is better than the reference model in the literature.Entities:
Mesh:
Year: 2021 PMID: 34306049 PMCID: PMC8279871 DOI: 10.1155/2021/2578422
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The recognition rates of models in the literature on classification tasks.
Figure 2The structure of the convolutional neural network.
Figure 3The network structure of RNN.
Figure 4The LSTM network structure.
Comparative analysis of four models.
| Advantages and disadvantages of model mechanisms | Structures | The model's structure | Application suggestions | Stacked structure model |
|---|---|---|---|---|
| There is only one backbone network | Network structure is simple | Network is hard to train | Network-in-network structure model | Multiple network branches at multiple scales |
|
| ||||
| It has the ability of reaggregating, except the number of parameters | It is applied to tasks with small data sets | Try to avoid applying to tight resources | The equipment is missing | Residual structural model |
|
| ||||
| Is a short circuit mechanism network | It solves the problem of deep network modeling | Phenomenon of random depth | Enhance the autonomous feature extraction ability | The amount of computation |
|
| ||||
| Used to build deep networks | Attention mechanism model | Channel attention and spatial attention | The performance of the original model | Generalization ability |
Figure 5Dares Net adopts the residual unit structure.
Figure 6The Hybrid CNN and LSTM framework proposed in this paper.
The experimental results of feature fusion of the classification layer fusion model Hybrid CNN and LSTM.
| Characteristics of the combination | Negative (%) | Positive (%) | Macro average (%) | ||||||
|---|---|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| |
| 1 | 90.5 | 86.8 | 88.6 | 86.9 | 87.1 | 85.5 | 90.7 | 89 | 88.7 |
| 2 | 88.9 | 88.8 | 88.8 | 88.8 | 89.6 | 85.1 | 89.6 | 87.2 | 87.1 |
| 3 | 89.6 | 89.3 | 89.3 | 89.3 | 87.1 | 89.4 | 87.1 | 89.4 | 82.1 |
| 4 | 87.1 | 89.4 | 87.1 | 89.4 | 87.1 | 89.4 | 88.2 | 89.5 | 83.4 |
| 5 | 92.2 | 93.4 | 92.8 | 92.9 | 91.6 | 92.2 | 92.5 | 92.5 | 92.5 |
| 6 | 92.8 | 93.7 | 93.2 | 93.3 | 92.3 | 92.8 | 93 | 93 | 93 |
| 1.WV | 2.POSV | 3.SWV | 4.WV + POSV | 5.WV + SWV | 6.WV + POSV + SWV | ||||
Figure 7The fitting curve of Hybrid CNN and LSTM.
Figure 8Comparison of display models and classification algorithm accuracy between Hybrid CNN and LSTM and BOW + CNN.