| Literature DB >> 35458972 |
Adarsh Vulli1, Parvathaneni Naga Srinivasu2, Madipally Sai Krishna Sashank1, Jana Shafi3, Jaeyoung Choi4, Muhammad Fazal Ijaz5.
Abstract
Lymph node metastasis in breast cancer may be accurately predicted using a DenseNet-169 model. However, the current system for identifying metastases in a lymph node is manual and tedious. A pathologist well-versed with the process of detection and characterization of lymph nodes goes through hours investigating histological slides. Furthermore, because of the massive size of most whole-slide images (WSI), it is wise to divide a slide into batches of small image patches and apply methods independently on each patch. The present work introduces a novel method for the automated diagnosis and detection of metastases from whole slide images using the Fast AI framework and the 1-cycle policy. Additionally, it compares this new approach to previous methods. The proposed model has surpassed other state-of-art methods with more than 97.4% accuracy. In addition, a mobile application is developed for prompt and quick response. It collects user information and models to diagnose metastases present in the early stages of cancer. These results indicate that the suggested model may assist general practitioners in accurately analyzing breast cancer situations, hence preventing future complications and mortality. With digital image processing, histopathologic interpretation and diagnostic accuracy have improved considerably.Entities:
Keywords: 1-cycle policy; DenseNet-169; FastAI; cancer; computational histopathology; diagnostic odds ratio; lymph nodes; whole-slide images
Mesh:
Year: 2022 PMID: 35458972 PMCID: PMC9025766 DOI: 10.3390/s22082988
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.847
A detailed description of the various models in image processing.
| Approach | Objective | Challenges of the Approach |
|---|---|---|
| Genetic Algorithm (GA) [ | A genetic algorithm selects the beginning population at random through a probabilistic approach. It performs crossover and mutation processes concurrently until the necessary portions are reached. | The algorithm fails in producing the best output and is more time-consuming. |
| Fully Convolutional Residual Network (FCRN) [ | FCRN technique employs encoder and decoder layers for image classification that use low-high level features. The feature processing is exceptionally important for the appropriate classification. | A completely Conventional Layer handles overfitting well, yet the model is complex in design and implementation. Adding batch normalization might also make the model less efficient. |
| Decision Tree (DT) [ | Handling discrete data necessitates the usage of models based on decision trees which is a rule-based technique for predictions. It is effective in dealing with non-linear factors. | The Decision Tree model is unreliable if the input data is changed even by a small proportion, and at times DT models will lead to overfitting while training. |
| Bayesian Learning (BL) [ | The Bayesian Learning technique effectively manages continuous and discrete data by avoiding the incorrect binary and multi-class classification characteristics. | The Bayesian Classifier is often an improper probabilistic model since it is unsuited for unsupervised learning applications. |
| Deep Neural Networks [ | Deep Neural Networks may process structured and unstructured data. Models are capable of working with unlabelled data and delivering the expected results. | DNN model is a black-box decision model, and models are complex and need tremendous development efforts. |
| K-Nearest Neighbourhood [ | KNN based models work on unlabelled data and classify data into different categories using feature selection and similarity matching. These models use the distance between two instances to identify their correlation. | The trained model’s accuracy is closely related to the quality of the data used to train it. In addition, the time needed to make a forecast may be much longer if the sample size is bigger. |
| Support Vector Machine [ | Support Vector Machine is a data processing system that uses as little computing and memory as possible. | It is difficult to determine the feature-based parameters using the Support Vector Machine method, which is inefficient for noisy data. |
| Artificial Neural Networks [ | Linear relationships between dependent and independent parameters may be easily recognized using Artificial Neural Networks, storing data across the network nodes. | Using Artificial Neural Network models is a good way to deal with a lack of knowledge of the issue. There is a good chance that the ANN will miss the spatial elements of the picture. The gradient’s diminishment and explosion are also major concerns. |
Figure 1Image denoting the block diagram of the proposed model.
Dataset descript associated with pcam.
| Description | Specification |
|---|---|
| Format | TIF |
| Input Size | 96 × 96 |
| Number of Channels | 3 |
| Bits per Channel | 8 |
| Data Type | Unsigned Char |
| Image Compression Approach | Jpeg |
Figure 2Random sampling of the dataset.
Figure 3Cropped histopathological scan image.
Figure 4Random augmentation of cropped scan images.
Figure 5The architecture of DenseNet-169 used to implement the proposed method.
DenseNet-169 layered architecture.
| Layer | Kernel Size | Parameters | Tensor Size |
|---|---|---|---|
| Convolution | 7 × 7 (Conv) | Stride = 2, ReLu | 112 × 112 |
| Pooling | 3 × 3 (MaxPool) | Stride = 2 | 56 × 56 |
| Dense-1 Layer | 1 × 1 × 6 (Conv) | Dropout = 0.2 | 56 × 56 |
| Transition-1 Layer | 1 × 1 (Conv) | Stride = 2 | 56 × 56 |
| Dense-2 block | 1 × 1 × 12 (Conv) | Dropout = 0.2 | 28 × 28 |
| Transition-2 Layer | 1 × 1 (Conv) | Stride = 2 | 28 × 28 |
| Dense-3 Layer | 1 × 1 × 32 (Conv) | Dropout = 0.2 | 14 × 14 |
| Transition-3 Layer | 1 × 1 (Conv) | Stride = 2 | 14 ×14 |
| Dense-4 Layer | 1 × 1 × 32 (Conv) | Dropout = 0.2 | 7 × 7 |
| Classification Layer | 1 × 1 (Global AvgPool) | 1 × 1 |
Figure 6Graph representing the learning rate associated with weight decay.
Figure 7Graphs representing the learning rate for 1-Cycle policy.
Figure 8Graphs of learning rate and momentum over iterations before fine-tuning DenseNet-169.
Figure 9Graphs of learning rate and momentum over iterations after fine-tuning DenseNet-169.
Figure 10Loss associated with batches processed before fine-tuning the model.
Figure 11Loss associated with batches processed after fine-tuning the model.
The hyperparameter values are associated with various models.
| Training | Testing | |||
|---|---|---|---|---|
| Loss | Accuracy | Loss | Accuracy | |
| CNN [ | 0.124 | 92.25 | 0.565 | 81.93 |
| CNN + Augmentation [ | 0.164 | 93.82 | 0.621 | 82.13 |
| VGG-16 [ | 0.008 | 99.75 | 0.290 | 79.00 |
| ConcatNet [ | 0.108 | 95.90 | 0.435 | 86.23 |
| DenseNet-169 | 0.152 | 94.61 | 0.411 | 95.57 |
| Fine-tuned DenseNet-169 | 0.123 | 95.45 | 0.125 | 97.45 |
Details of Implementation Environment.
| Environment Details | Specifications |
|---|---|
| Operating System | Microsoft Windows 11 |
| Processor | Intel(R) Core (TM) i7-8750H |
| Architecture | 64-Bit |
| Memory Allotted | 3 GB |
| GPU | Nvidia (TM) 1050 Ti |
| Language | Python |
| Framework | FastAI, PyTorch, DL |
| Libraries Used | Pandas, Numpy, cv2, Matplotlib, Scikit-learn, os |
Figure 12(a) Confusion matrix for DenseNet-169 (b) Confusion matrix for fine-tuned DenseNet-169.
Figure 13Probabilities scores associated with samples in the testing phase.
Comparison of DenseNet-169 model with state-of-art models.
| Accuracy | Sensitivity | Specificity | F1-Score | Precision | |
|---|---|---|---|---|---|
| Logistic regression [ | 87.0 | 86.4 | 87.6 | 0.87 | - |
| NN [ | 82.8 | 74.4 | 91.0 | 0.81 | - |
| NN feature subset [ | 91.3 | 85.7 | 96.8 | 0.91 | - |
| Random Forest [ | 93.0 | 92.6 | 93.3 | 0.93 | - |
| SVM [ | 88.3 | 85.9 | 90.6 | 0.88 | - |
| CNN [ | 76.4 | 74.6 | 80.4 | - | - |
| CNN + Augmentation [ | 78.8 | 80.2 | 81.4 | - | - |
| VGG-16 [ | 76.5 | 75.3 | 82.6 | - | - |
| ConcatNet [ | 84.1 | 82.0 | 87.8 | - | - |
| Multimodal Deep Neural Networks [ | 79.4 | 80.0 | - | - | 0.875 |
| SVM [ | 77.5 | 87.8 | - | - | 0.811 |
| RF [ | 77.0 | 90.2 | - | - | 0.787 |
| RF [ | 80.1 | 91.0 | - | - | - |
| LR [ | 75.4 | 96.3 | - | - | 0.563 |
| Inception V3 [ | 80.5 | 82.0 | 79.0 | 0.81 | - |
| Inception-RestNet V2 [ | 82.0 | 80.0 | 85.0 | 0.82 | - |
| ResNet-101 [ | 78.0 | 78.0 | 79.0 | 0.78 | - |
| DenseNet-169 | 95.5 | 93.1 | 97.1 | 0.94 | 0.971 |
| Fine-tuned DenseNet-169 | 96.7 | 95.2 | 97.8 | 0.96 | 0.978 |
Figure 14(a) Receiver operator characteristic (ROC) curve before fine-tuning (b) ROC curve after fine-tuning.
Figure 15Results after applying the test time augmentation (TTA).
Figure 16Block diagram of practical implication model.
Figure 17Mobile framework for app integration with fine-tuned DenseNet-169.
Figure 18Image of user-interface of the future perspective model.