| Literature DB >> 35205491 |
Máximo Eduardo Sánchez-Gutiérrez1, Pedro Pablo González-Pérez2.
Abstract
Medical data includes clinical trials and clinical data such as patient-generated health data, laboratory results, medical imaging, and different signals coming from continuous health monitoring. Some commonly used data analysis techniques are text mining, big data analytics, and data mining. These techniques can be used for classification, clustering, and machine learning tasks. Machine learning could be described as an automatic learning process derived from concepts and knowledge without deliberate system coding. However, finding a suitable machine learning architecture for a specific task is still an open problem. In this work, we propose a machine learning model for the multi-class classification of medical data. This model is comprised of two components-a restricted Boltzmann machine and a classifier system. It uses a discriminant pruning method to select the most salient neurons in the hidden layer of the neural network, which implicitly leads to a selection of features for the input patterns that feed the classifier system. This study aims to investigate whether information-entropy measures may provide evidence for guiding discriminative pruning in a neural network for medical data processing, particularly cancer research, by using three cancer databases: Breast Cancer, Cervical Cancer, and Primary Tumour. Our proposal aimed to investigate the post-training neuronal pruning methodology using dissimilarity measures inspired by the information-entropy theory; the results obtained after pruning the neural network were favourable. Specifically, for the Breast Cancer dataset, the reported results indicate a 10.68% error rate, while our error rates range from 10% to 15%; for the Cervical Cancer dataset, the reported best error rate is 31%, while our proposal error rates are in the range of 4% to 6%; lastly, for the Primary Tumour dataset, the reported error rate is 20.35%, and our best error rate is 31%.Entities:
Keywords: discriminant pruning; feature selection; information-entropy measures; machine learning; medical data and signals; restricted Boltzmann machine
Year: 2022 PMID: 35205491 PMCID: PMC8870840 DOI: 10.3390/e24020196
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Definitions.
| Term | Definition |
|---|---|
| Artificial neural | Artificial neural networks are information |
| Artificial neuron | An artificial neuron is a mathematical function that |
| Visible neuron/unit | The visible units or neurons make up the input layer |
| Hidden neuron/unit | The hidden units or neurons make up the |
| Network topology | The network topology refers to the number of layers, |
| Size of the network | The size of an artificial neural network is related to |
| Pruning method | The technique used to reduce the size of a neural |
| Restricted Boltzmann | A RBM is a neural network that models a non- |
| Deep neural network | A deep neural network uses a hierarchy of |
| Deep belief network (DBN) | A DBN can be seen as a deep neural network where |
| k Nearest Neighbor (k-NN) | The k-NN algorithm can be defined as memory- |
Figure 1Generic multi–layer perceptron model. The input layer corresponds to the dimension of the input vectors. The hidden layer contains (# input features + # classes)/2 neurons. Finally, the number of neurons in the output layer corresponds to the same number of classes.
Figure 2Operations done by a neuron.
Figure 3Sigmoid function.
Figure 4Discriminative pruning of hidden neurons. The process includes a classifier that takes the outputs of the most discriminative neurons from an RBM.
Cancer datasets.
| Cancer Dataset | # of Original Features | # of Instances | # of Classes | Class Distribution |
|---|---|---|---|---|
| Breast Cancer | 9 | 286 | 2 | [201, 85] |
| Cervical Cancer | 35 | 858 | 2 | [803, 55] |
| Primary Tumour | 17 | 339 | 21 | [84, 20, 9, 14, 39, |
Figure 5Cross-validation schema. In the k-fold cross-validation, the training set is split into k smaller sets. A model is trained using of the folds as training data; the resulting model is validated on the test set to compute a performance measure. The performance measure reported by k-fold cross-validation is then the average of the values computed in the loop.
Figure 6Pruning performance on the Breast Cancer dataset. Experiments were carried out with 5 (a), 9 (b), 18 (c), 36 (d), 54 (e), 72 (f), 144 (g) and 288 (h) initial hidden units.
Figure 7Pruning performance on the Cervical Cancer dataset. Experiments were carried out with 14 (a), 28 (b), 56 (c), 112 (d), 168 (e), 224 (f), 448 (g) and 896 (h) initial hidden units.
Figure 8Pruning performance on the Primary Tumour dataset. Experiments were carried out with 9 (a), 17 (b), 34 (c), 68 (d), 102 (e), 136 (f), 272 (g) and 544 (h) initial hidden units.
Architecture of the restricted Boltzmann machines.
| Cancer Dataset | # of Features | # Test Hidden Layer Architectures |
|---|---|---|
| Breast Cancer | 9 | [5, 9, 18, 36, 54, 72, 144, 288] |
| Cervical Cancer | 28 | [14, 28, 56, 112, 168, 224, 448, 896] |
| Primary Tumour | 17 | [9, 17, 34, 68, 102, 136, 272, 544] |
Reported best error rate vs. best error found with the proposed pruning method for the cancer datasets under analysis.
| Cancer Dataset | Reported Error Rate | Error Rate Found with the Proposed Pruning Method |
|---|---|---|
| Breast Cancer | (ANN) 10.68% [ | 10% to 15% |
| Cervical Cancer | (Linear Regression) 31% [ | 4% to 6% |
| Primary Tumour | (ANN) 20.35% [ | 31% |