| Literature DB >> 35009784 |
Qasem Abu Al-Haija1, Ahmad Al-Badawi2.
Abstract
Network Intrusion Detection Systems (NIDSs) are indispensable defensive tools against various cyberattacks. Lightweight, multipurpose, and anomaly-based detection NIDSs employ several methods to build profiles for normal and malicious behaviors. In this paper, we design, implement, and evaluate the performance of machine-learning-based NIDS in IoT networks. Specifically, we study six supervised learning methods that belong to three different classes: (1) ensemble methods, (2) neural network methods, and (3) kernel methods. To evaluate the developed NIDSs, we use the distilled-Kitsune-2018 and NSL-KDD datasets, both consisting of a contemporary real-world IoT network traffic subjected to different network attacks. Standard performance evaluation metrics from the machine-learning literature are used to evaluate the identification accuracy, error rates, and inference speed. Our empirical analysis indicates that ensemble methods provide better accuracy and lower error rates compared with neural network and kernel methods. On the other hand, neural network methods provide the highest inference speed which proves their suitability for high-bandwidth networks. We also provide a comparison with state-of-the-art solutions and show that our best results are better than any prior art by 1~20%.Entities:
Keywords: Internet of Things; cybersecurity; ensemble learning; intrusion classification; intrusion detection; network layer
Mesh:
Year: 2021 PMID: 35009784 PMCID: PMC8749547 DOI: 10.3390/s22010241
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1NIDS typical deployment in computer networks.
Machine-learning algorithms selection criteria.
| Method | Assumptions | Explainability | Execution Time | Data Engineering | |
|---|---|---|---|---|---|
| Train | Test | ||||
| EBT, ERT | No assumptions about predictors or response variable | Intuitively explainable as rule-based knowledge system | Slow | Fast | Minimal effort |
| ESK | No assumptions | Intuitively explainable via similarity measures | Slow | Very slow | Minimal effort |
| SNN, BNN | No assumptions | Black box | Depends on network architecture | Fast | Medium effort |
| LRK | Linearity between predictors and response variable | Relatively explainable | Fast | Slow | Essential |
Comparison among some state-of-the-art NIDS solutions. ANN stands for Artificial Neural Networks, NB for Naïve Bayes, RF for Random Forests, SOM for Self-Organizing Maps, LDA for Linear Discriminant Analysis, QDA for Quadratic Discriminant Analysis, PCA for Principal Component Analysis, RT for Regression Trees, LR for Logistic Regression, U2R for User to Root, R2L for Remote to Local.
| Ref. | Methods | Datasets | Attacks |
|---|---|---|---|
| [ | ANN, SVM, NB, RF, SOM | NSL-KDD, KDD Cup 1999, CIC DOS, ADFA-LD12, UNSWNB15, WSN-DS | DDoS, flooding, U2R, Jamming |
| [ | Auto-Encoder, RF, NB, LDA, QDA | CICIDS2017 | DDoS, Heartbleed, SQL Injection, Botnet. |
| [ | ANN, SVM | NSL-KDD | DDoS, R2L, U2R |
| [ | Ensemble Learning (Extra Trees) | UNSW-NB15, BoT-IoT, | DDoS, Botnet, |
| [ | Statistical Analysis | Kitsune, ISCX, IoT | Botnet, DDoS, MITM |
| [ | XGBoost, PCA | ToN-IoT and | DDoS, Botnet, |
| [ | Ensemble-based voting classifier | Ton-IoT | DDoS, Botnet, |
| [ | Shallow CNN | NSL-KDD | Normal, DoS, Probe, R2L, U2R |
| [ | XGBoost | NSL-KDD | SynFlood, UDP Flood, Smurf, and others |
| [ | LR, LDA, RT, RF, and NB | TON-IoT | DDoS, Password, |
| [ | RF, KNN, NB | Simulated dataset | Satori, Reaper, Amnesia, Masuta, Mirai, others |
| [ | Fuzzy C-means clustering and fuzzy interpolation | Kitsune | Botnet, MitM, DoS |
| [ | Generative adversarial networks (GAN) | Kitsune, CICIDS | Artificially generated attacks |
| [ | Extreme Value Analysis | Kitsune | Botnet, MitM, DoS |
Simulation Environment Specifications (Hardware and Software).
| Item | Descriptions |
|---|---|
| Operation System | Windows 11, Edition 21H2, 64-bit operating system, x64-based processor |
| Processing Component | 11th Gen Intel(R) Core(TM) i7-11800H @ 2.30 GHz⋯30 GHz |
| Computing Component | NVIDIA GeForce RTX 3050 Ti Laptop GPU@ 4 GBye |
| Memory Component | 16.0 GB, DDR4 1.2v @ Memory Speed: 2933 MHz (PC4-23400) |
| Storage Component | 500 GB Kingston NV1 M.2 (2280) PCIe NVMe Gen 3.0 (×4) SSD |
| Development Platform | MATLAB 2021b + Parallel Computing + Machine Learning Packages. |
Figure 2Workflow Diagram for attack-aware IoT network traffic routing via ML techniques.
Summary of Dataset Distribution: distilled-Kitsune-2018 and NSL-KDD.
| Samples Distribution for Distilled-Kitsune-2018 | ||||
|---|---|---|---|---|
| Attack | No. Training Packets | No. Normal Test Packets | No. Malicious Test Packets | |
| OS Scan | 6000 | 13,500 | 1499 | |
| Fuzzing | 1200 | 9000 | 999 | |
| Video Inj. | 4000 | 9000 | 999 | |
| ARP | 6000 | 13,500 | 1499 | |
| Wiretap | 4000 | 9000 | 999 | |
| SSDP F. | 6000 | 13,500 | 1499 | |
| SYN DoS | 1200 | 9000 | 999 | |
| SSL R. | 6000 | 13,500 | 1499 | |
| Mirai | 6000 | 9000 | 999 | |
|
| ||||
|
|
|
|
| |
| Training | 67,343 | 45,927 | 11,656 | 995 |
| Testing | 9711 | 7458 | 2754 | 2421 |
| Total | 77,054 | 53,385 | 14,410 | 3416 |
Summary of System Development Parameters.
| ML Model | Models Parameters |
|---|---|
| Ensemble Boosted Trees (EBT) | Ensemble method: AdaBoost, Learner type: Decision tree, Maximum number of splits: 20, Number of learners: 30, Learning rate: 0.1, 5-Fold Cross Validation. |
| Ensemble Subspace kNN (ESK) | Ensemble method: Subspace, Learner type: Nearest Neighbors, number of learners: 30, Subspace Dimension: 58, 5-Fold Cross Validation |
| Ensemble RUS_Boosted Trees (ERT) | Ensemble method: RUSBoost, Learner type: Decision tree, Maximum number of splits: 20, Number of learners: 30, Learning rate: 0.1, 5-Fold Cross Validation |
| Shallow Neural Network (SNN) | Number of fully connected layers = one hidden layer with size = 30, Activation: Sigmoid, Iteration limit: 1000, Standardize data: Yes, Regularization strength (Lambda): 0, |
| Bilayered Neural Network (BNN) | Number of fully connected layers: 2, First layer size: 10 |
| Logistic Regression Kernel (LRK) | Learner: Logistic Regression, Number of expansion dimensions: Auto, Regularization strength (Lambda): Auto, Kernel scale: Auto, Multiclass method: One-vs-One, Iteration limit: 1000, 5-Fold Cross Validation |
Figure 3Confusion matrix with other performance evaluation measures.
Summary of system evaluation results comparing performance of Kitsune and NSL-KDD datasets.
| ML Models | CA% | MCR% | TC (#) | CS (Obs/Sec) | ||||
|---|---|---|---|---|---|---|---|---|
| Kitsune | NSL-KDD | Kitsune | NSL-KDD | Kitsune | NSL-KDD | Kitsune | NSL-KDD | |
| EBT | 99.8 | 99.1 | 0.2 | 0.9 | 249 | 1332 | 90,000 | 84,000 |
| ESK | 99.4 | 98.4 | 0.6 | 1.6 | 780 | 2346 | 41 | 84 |
| ERT | 98.1 | 97.2 | 1.9 | 2.8 | 2495 | 4211 | 90,000 | 90,000 |
| BNN | 97.5 | 96.1 | 2.5 | 3.9 | 3250 | 5702 | 290,000 | 420,000 |
| SNN | 96.7 | 94.6 | 3.3 | 5.4 | 4290 | 8014 | 240,000 | 390,000 |
| LRK | 94.5 | 93.7 | 5.5 | 6.3 | 7215 | 9198 | 440 | 1900 |
Figure 4Timing complexity of both datasets using the six above mentioned ML models.
Figure 5Confusion matrix for Ensemble Boosted Trees (EBT) classifier.
Figure 6Matrix of PPV vs. FDR for each individual class using EBT classifier.
Figure 7Matrix of TPR vs. FNR for each individual class using EBT classifier.
Summary of System Evaluation Results for EBT.
| CA% | PR% | RC% | MCR% | FDR% | FNR% | TC (#) | CS (Obs/Sec) |
|---|---|---|---|---|---|---|---|
| 99.8 | 99.7 | 98.1 | 0.2 | 0.9 | 1.7 | 249 | 90,000 |
Comparison with other existing ML-based IoT-IDS systems.
| Research | Year | ML Model | No. Classes | ACC% | PPV% | TPR% |
|---|---|---|---|---|---|---|
| Sarhan et al. [ | 2021 | XRT Classifier | 2–10 | 98.05 | 84.61 | - |
| Ashraf et al. [ | 2021 | STL Classifier | 3 | 99.20 | - | - |
| Kumar et al. [ | 2021 | XGB Classifier | 10 | 97.81 | 87.55 | 85.43 |
| Khan et al. [ | 2021 | HYB Classifier | 7 | 76.00 | 75.00 | 75.00 |
| Al-Haija et al. [ | 2020 | S-CNN Classifier | 5 | 98.20 | 98.27 | 98.20 |
| Jinxin et al. [ | 2020 | XGB Classifier | 5 | 97.0 | - | - |
| Alsaedi et al. [ | 2020 | CART Classifier | 9 | 77.00 | 77.00 | 77.00 |
| Al-Haija et al. [ | 2020 | S-CNN Classifier | 2 | 99.30 | 99.33 | 99.18 |
| Kumar et al. [ | 2019 | kNN Classifier | 3 | 94.44 | 92.00 | 100.0 |
| Proposed model | 2021 | EDT Classifier | 10 | 99.80 | 99.69 | 98.10 |