| Literature DB >> 33805937 |
Filip Pałka1, Wojciech Książek1, Paweł Pławiak1,2, Michał Romaszewski2, Kamil Książek2,3.
Abstract
This study is focused on applying genetic algorithms (GAs) to model and band selection in hyperspectral image classification. We use a forensic-inspired data set of seven hyperspectral images with blood and five visually similar substances to test GA-optimised classifiers in two scenarios: when the training and test data come from the same image and when they come from different images, which is a more challenging task due to significant spectral differences. In our experiments, we compare GA with a classic model optimisation through a grid search. Our results show that GA-based model optimisation can reduce the number of bands and create an accurate classifier that outperforms the GS-based reference models, provided that, during model optimisation, it has access to examples similar to test data. We illustrate this with experiments highlighting the importance of a validation set.Entities:
Keywords: SVM; blood; genetic algorithm; hyperspectral classification; machine learning
Year: 2021 PMID: 33805937 PMCID: PMC8037346 DOI: 10.3390/s21072293
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Visualisation of the dataset used in experiments. Upper panels present classes as a coloured ground truth on RGB images created from hyperspectral cubes. Middle panels present mean class spectra. Bottom panels present the PCA projection of data for the first two principal components. Images come from [14].
Figure 2Visualisation of the impact of preprocessing and feature extraction on example spectra of the “blood” class from the dataset. Spectra in plot (a,b) were normalised by dividing each pixel by its median. Spectra in plot (c) were transformed by computing their first order derivatives.
The structure of a chromosome corresponding to optimized parameters of the -SVM classifier along with selected hyperspectral bands. RBF: radial basis function.
| Parameter | Range of Values |
|---|---|
|
| {RBF, polynomial, sigmoid} |
|
|
|
|
|
|
|
|
|
|
|
|
| band |
|
Kernel function; parameter of the polynomial kernel; parameter of the RBF kernel; parameter of the polynomial and sigmoid kernel.
Figure 3Visualisation of a one-point crossover between two individuals.
Parameters of the genetic algorithm (GA) used in experiments.
| Parameter | Value |
|---|---|
| Size of the population | 200 |
| Number of epochs | 100 |
| Fitness function | Accuracy |
| Selection algorithm | Tournament selection, size 3 |
| Crossover method | Uniform crossover |
| Mutation method | One-point mutation |
| Probability of crossover | 0.8 |
| Probability of mutation | 0.8 |
| Elitist strategy | 1 individual |
Own implementation.
Grid-search (GS) parameters used in experiments.
| Classifier | Parameter | Values |
|---|---|---|
| SVM |
| {RBF, polynomial, sigmoid} |
|
|
| |
|
|
| |
|
|
| |
|
|
| |
| LSVM | loss | {hinge, squared} |
|
|
| |
|
| {RBF, polynomial, sigmoid} | |
|
|
| |
|
|
| |
|
|
| |
|
|
| |
| KNN | Dist. metric | {Euclidean, Manhattan, Chebyshev} |
| Weights | {Uniform, distance} | |
| No. neighbors |
| |
| MLP | No. hidden layers |
|
| Number of neurons on consecutive layers |
| |
| Dropout |
| |
| Learning rate |
| |
| Batch size |
| |
| Number of iterations |
| |
| Weights initialisation | Glorot method [ |
Kernel function; parameter of the polynomial kernel; parameter of the RBF kernel; parameter of the polynomial and sigmoid kernel; linear Support Vector Machine (SVM), implemented in liblinear library. KNN: K-nearest neighbour; MLP: Multilayer Perceptron.
Figure 4The overview scheme of experiments.
Figure 5Visualisation of the model optimisation stage in the hyperspectral inductive classification (HIC) scenario, using 10-fold cross-validation on a selected training set from “Frame” images.
Figure 6Visualisation of the model optimisation stage in the hyperspectral inductive classification with a validation set (HICVS) experiment with a small training set and 10-fold cross-validation.
Results of the HTC scenario for classification with GA and reference classifiers trained with a grid search (GS). The highest result in each day is denoted with a bold font. SVC: SVM with the regularisation parameter C.
| Model Optimisation | Classifier | Accuracy/Day | |||
|---|---|---|---|---|---|
| 1 | 7 | 21 | All | ||
| GS | SVC |
|
|
|
|
| LSVC |
|
|
|
| |
|
|
|
|
| ||
| KNN |
|
|
|
| |
| MLP |
|
|
|
| |
| GA |
|
|
|
| |
SVM with a linear kernel; results for combined data from all days.
Results of the HIC scenario for classification with GA and reference classifiers trained with a grid search (GS). The highest result in each day is denoted with bold font.
| Model Optimisation | Classifier | Accuracy/Day | |||
|---|---|---|---|---|---|
| 1 | 7 | 21 | All | ||
| GS | SVC |
|
|
|
|
| LSVC |
|
|
|
| |
|
|
|
|
| ||
| KNN |
|
|
|
| |
| MLP |
|
|
|
| |
| GA |
|
|
|
| |
SVM with a linear kernel; results for combined data from all days.
Results of the HICVS scenario for classification with GA and reference classifiers trained with a grid search (GS). The highest result in each day is denoted with bold font.
| Model Optimisation | Classifier | Accuracy/Day | |||
|---|---|---|---|---|---|
| 1 | 7 | 21 | All | ||
| GS | SVC |
|
|
|
|
| LSVC |
|
|
|
| |
|
|
|
|
| ||
| KNN |
|
|
|
| |
| MLP |
|
|
|
| |
| GA |
|
|
|
| |
Denotes SVM with a linear kernel; Results for combined data from all days.