| Literature DB >> 36188679 |
Lei Zhang1,2.
Abstract
Credit evaluation is a difficult problem in the process of financing and loan for small and medium-sized enterprises. Due to the high dimension and nonlinearity of enterprise behavior data, traditional logistic regression (LR), random forest (RF), and other methods, when the feature space is very large, it is easy to show low accuracy and lack of robustness. However, recurrent neural network (RNN) will have a serious gradient disappearance problem under long sequence training. This paper proposes a compound neural network model based on the attention mechanism to meet the needs of enterprise credit evaluation. The convolutional neural network (CNN) and the long short-term memory (LSTM) network were used to establish the model, using soft attention, the gradient propagates back to other parts of the model through the attention mechanism module. In the multimodel comparison experiment and three different enterprise data experiments, the CNN-LSTM-ATT model proposed in this paper is superior to the traditional models LR, RF, CNN, LSTM, and CNN-LSTM in most cases. The experimental results under multimodel comparison reflect the higher accuracy of the model, and the group test reflects the higher robustness of the model.Entities:
Mesh:
Year: 2022 PMID: 36188679 PMCID: PMC9522511 DOI: 10.1155/2022/6826573
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The unit structure of LSTM.
Figure 2The structure of CNN-LSTM-ATT.
The experimental environment configuration.
| Experimental environment | Parameter |
|---|---|
| Operating system | Windows 10 × 64 |
| Processor | CPU intel ( |
| Video card | GTX 1650 4G |
| Memory | 8G |
| Hard disk | 500G |
| Development language |
|
The indexes of calculating ROC.
| Indexes | Meaning of indexes |
|---|---|
| TP | No default and correct prediction |
| TN | No default and wrong prediction |
| FP | No default and correct prediction |
| FN | No default and wrong prediction |
The parameter setting of the CNN-LSTM-ATT model.
| Parameter | Meaning of parameter | Value of parameter |
|---|---|---|
| time_series | The step length | 20 |
| behavior_num | Number of behavioral indexes | 14 |
| iB-LSTM_units | Number of neurons for LSTM | 64 |
| Kernel_size1 | The size of the convolution kernel for convolution layer 1 | 1 × 3 |
| Kernel_size1 | The size of the convolution kernel for convolution layer 2 | 3 × 3 |
| Pool | Pooling method | Max |
| Stride | Length of pooling or convolutional | 1 |
| Dropout | Rate of LSTM dropout | 0.3 |
| batch_size | The size of batch | 128 |
| Epoch | Number of iterations | 20 |
| Optimizer | The optimizer | SGD |
| Learning rate | The learning rate | 0.0001 |
Figure 3Experimental results of super-parameter setting. (a) Optimizer. (b) False positive rate.
Figure 4ROC curves of four models under different learning rates.
Figure 5ROC curve of four models under different dropout values.
The AUC of different models under three groups data.
| Model | Group-0 | Group-1 | Group-2 | Mean |
|---|---|---|---|---|
|
|
|
|
|
|
| CNN-LSTM | 0.89 | 0.86 | 0.94 | 0.89 |
| CNN | 0.82 | 0.88 | 0.72 | 0.80 |
| LSTM | 0.83 | 0.75 | 0.82 | 0.80 |
Figure 6ROC curve of four models under three groups data.