| Literature DB >> 35214201 |
Soyoung Yoo1, Junghyun Kim1,2, Guang J Choi2,3.
Abstract
In recent research on the formulation prediction of oral dissolving drugs, deep learning models with significantly improved performance compared to machine learning models were proposed. However, the performance degradation due to limitations of an imbalanced dataset with a small size and inefficient neural network structure has still not been resolved. Therefore, we propose new deep learning-based prediction models that maximize the prediction performance for disintegration time of oral fast disintegrating films (OFDF) and cumulative dissolution profiles of sustained-release matrix tablets (SRMT). In the case of OFDF, we use principal component analysis (PCA) to reduce the dimensionality of the dataset, thereby improving the prediction performance and reducing the training time. In the case of SRMT, the Wasserstein generative adversarial network (WGAN), a neural network-based generative model, is used to overcome the limitation of performance improvement due to the lack of experimental data. To the best of our knowledge, this is the first work that utilizes WGAN for pharmaceutical formulation prediction. Experimental results show that the proposed methods have superior performance than the existing methods for all the performance metrics considered.Entities:
Keywords: Wasserstein GAN; deep learning; imbalanced data; pharmaceutical formulation; principal component analysis; small data
Year: 2022 PMID: 35214201 PMCID: PMC8880629 DOI: 10.3390/pharmaceutics14020467
Source DB: PubMed Journal: Pharmaceutics ISSN: 1999-4923 Impact factor: 6.321
Figure 1Architecture of two proposed models. (a) OFDF; (b) SRMT.
Figure 2Scree plot of OFDF dataset.
Figure 3Correlation matrices of OFDF datasets before and after applying PCA.
The number of trainable parameters of the existing OFDF model [11].
| Layer | Shape | Trainable Parameters |
|---|---|---|
| Input | 24 | 0 |
| Hidden layer 1 | 50 | 1250 |
| Hidden layer 2 | 50 | 2550 |
| Hidden layer 3 | 50 | 2550 |
| Hidden layer 4 | 50 | 2550 |
| Hidden layer 5 | 50 | 2550 |
| Hidden layer 6 | 50 | 2550 |
| Hidden layer 7 | 50 | 2550 |
| Hidden layer 8 | 50 | 2550 |
| Hidden layer 9 | 50 | 2550 |
| Hidden layer 10 | 50 | 2550 |
| Output | 1 | 51 |
|
|
|
The number of trainable parameters of the proposed OFDF model without PCA.
| Layer | Shape | Trainable Parameters |
|---|---|---|
| Input | 24 | 0 |
| Hidden layer 1 | 50 | 1250 |
| Hidden layer 2 | 25 | 1275 |
| Hidden layer 3 | 16 | 416 |
| Output | 1 | 17 |
|
|
|
The number of trainable parameters of the proposed OFDF model with PCA.
| Layer | Shape | Trainable Parameters |
|---|---|---|
| Input | 24 | 0 |
| Hidden layer 1 | 50 | 300 |
| Hidden layer 2 | 25 | 1275 |
| Hidden layer 3 | 16 | 416 |
| Output | 1 | 17 |
|
|
|
A comparison of the total number of trainable parameters, total number of updates, and training time between the existing OFDF model [11], the proposed OFDF model without PCA, and the proposed OFDF model with PCA.
| Model | Total Number of Trainable Parameters | Total Number of Updates | Training Time (min:s) | |
|---|---|---|---|---|
| Ref. [ | 24,251 | 81,900 | 01:47 | |
| Proposed | Without PCA | 2958 | 2400 | 00:05 |
| With PCA | 2008 | 2400 | 00:04 | |
Performance comparison between the existing OFDF model [11], the proposed OFDF model without PCA, and the proposed OFDF model with PCA.
| Model | Training Data | Validation Data | Test Data | |||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy (%) | RMSE | MAE | Accuracy (%) | RMSE | MAE | Accuracy (%) | RMSE | MAE | ||
| Ref. [ | 97.8 | 0.0382 | 0.0277 | 80 | 0.0765 | 0.0629 | 75 | 0.1009 | 0.0720 | |
| Proposed | Without | 100 | 0.0256 | 0.0192 | 85 | 0.0755 | 0.0598 | 85 | 0.0813 | 0.0536 |
| With | 100 | 0.0338 | 0.0243 | 95 | 0.0568 | 0.0427 | 95 | 0.0843 | 0.0578 | |
The number of samples for each class of the original training data for SRMT and the new training data with the oversampling techniques.
| Class | Original Training Data | ROS | SMOTE | ADASYN |
|---|---|---|---|---|
| 1 | 5 | 20 | 20 | 21 |
| 2 | 4 | 20 | 20 | 19 |
| 3 | 7 | 20 | 20 | 21 |
| 4 | 8 | 20 | 20 | 21 |
| 5 | 9 | 20 | 20 | 21 |
| 6 | 13 | 20 | 20 | 22 |
| 7 | 20 | 20 | 20 | 20 |
| 8 | 15 | 20 | 20 | 21 |
| 9 | 9 | 20 | 20 | 18 |
| 10 | 15 | 20 | 20 | 19 |
|
|
|
|
|
|
Two cases of whether cumulative characteristic is satisfied in the cumulative dissolution profile for SRMT.
| Satisfaction | Dissatisfaction | ||||||
|---|---|---|---|---|---|---|---|
|
|
|
|
|
|
|
|
|
| 0.070 | 0.234 | 0.824 | 0.855 | 0.244 | 0.513 |
|
|
| 0.302 | 0.575 | 0.785 | 0.910 | 0.365 | 0.316 |
|
|
| 0.100 | 0.386 | 0.934 | 0.942 | 0.358 | 0.469 |
|
|
| 0.214 | 0.305 | 0.409 | 0.441 |
|
|
|
|
| 0.131 | 0.328 | 0.455 | 0.497 |
|
|
|
|
Figure 4Distribution of data samples generated by (a) GAN and (b) WGAN.
The number of trainable parameters of the existing SRMT model [11].
| Layer | Shape | Trainable Parameters |
|---|---|---|
| Input | 21 | 0 |
| Hidden layer 1 | 30 | 660 |
| Hidden layer 2 | 30 | 930 |
| Hidden layer 3 | 30 | 930 |
| Hidden layer 4 | 30 | 930 |
| Hidden layer 5 | 30 | 930 |
| Hidden layer 6 | 30 | 930 |
| Hidden layer 7 | 30 | 930 |
| Hidden layer 8 | 30 | 930 |
| Hidden layer 9 | 30 | 930 |
| Output | 4 | 124 |
|
|
|
The number of trainable parameters of the proposed SRMT model [11].
| Layer | Shape | Trainable Parameters |
|---|---|---|
| Input | 21 | 0 |
| Hidden layer 1 | 150 | 3300 |
| Hidden layer 2 | 130 | 19,630 |
| Hidden layer 3 | 100 | 13,100 |
| Hidden layer 4 | 50 | 5050 |
| Hidden layer 5 | 30 | 1530 |
| Output | 4 | 124 |
|
|
|
A comparison of the total number of trainable parameters, total number of updates, and training time between the existing SRMT model [11], the proposed SRMT model.
| Model | Total Number of Trainable Parameters | Total Number of Updates | Training Time (min:s) |
|---|---|---|---|
| Ref. [ | 8224 | 273,000 | 02:47 |
| Proposed SRMT Model | 42,734 | 37,500 | 01:04 |
Performance comparison between the existing SRMT model [11] and the proposed SRMT model with data augmentation techniques.
| Model | Training Data | Validation Data | Test Data | |||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Accuracy (%) | RMSE | MAE | Accuracy (%) | RMSE | MAE | Accuracy (%) | RMSE | MAE | ||
| Ref. [ | 99.05 | 0.0384 | 0.0283 | 65 | 0.1625 | 0.1053 | 60 | 0.2019 | 0.1264 | |
| Proposed | ROS | 99 | 0.0176 | 0.0079 | 65 | 0.1347 | 0.9090 | 65 | 0.1116 | 0.0833 |
| SMOTE | 100 | 0.0082 | 0.0045 | 65 | 0.1469 | 0.0990 | 55 | 0.1282 | 0.0993 | |
| ADASYN | 100 | 0.0073 | 0.0038 | 65 | 0.1438 | 0.0965 | 60 | 0.1292 | 0.0992 | |
| WGAN | 100 | 0.0134 | 0.0101 | 75 | 0.1142 | 0.0762 | 70 | 0.1086 | 0.0831 | |