| Literature DB >> 30939764 |
Zainib Noshad1, Nadeem Javaid2, Tanzila Saba3, Zahid Wadud4, Muhammad Qaiser Saleem5, Mohammad Eid Alzahrani6, Osama E Sheta7.
Abstract
Wireless Sensor Networks (WSNs) are vulnerable to faults because of their deployment in unpredictable and hazardous environments. This makes WSN prone to failures such as software, hardware, and communication failures. Due to the sensor's limited resources and diverse deployment fields, fault detection in WSNs has become a daunting task. To solve this problem, Support Vector Machine (SVM), Convolutional Neural Network (CNN), Stochastic Gradient Descent (SGD), Multilayer Perceptron (MLP), Random Forest (RF), and Probabilistic Neural Network (PNN) classifiers are used for classification of gain, offset, spike, data loss, out of bounds, and stuck-at faults at the sensor level. Out of six faults, two of them are induced in the datasets, i.e., spike and data loss faults. The results are compared on the basis of their Detection Accuracy (DA), True Positive Rate (TPR), Matthews Correlation Coefficients (MCC), and F1-score. In this paper, a comparative analysis is performed among the classifiers mentioned previously on real-world datasets. Simulations show that the RF algorithm secures a better fault detection rate than the rest of the classifiers.Entities:
Keywords: WSNs; convolutional neural network; fault detection; machine learning; random forest; support vector machine
Year: 2019 PMID: 30939764 PMCID: PMC6480196 DOI: 10.3390/s19071568
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
List of acronyms.
| AN | Absolute Norm |
| ANOVA | Analysis Of Variance |
| BBA | Basic Belief Assignment |
| CCAD-SW | Collective Contextual Anomaly Detection using Sliding Window |
| CNN | Convolutional Neural Network |
| CMOS | Complementary Metal Oxide Semiconductor |
| DNN | Deep Neural Network |
| DA | Detection Accuracy |
| ECOC-SVM | Error-Correcting Output Coding-Support Vector Machine |
| EN | Euclidean Norm |
| ED | Euclidean Distance |
| EAD | Ensemble Anomaly Detection |
| FFNN | FeedFoward Neural Network |
| GA | Genetic Algorithm |
| GLDT | Gradient Lifting Decision Tree |
| GLR | Generalized Likelihood Ratio |
| KDE | Kernel Density Estimation |
| LNSM | Log-Normal Shadowing Model |
| MMSE | Minimum Mean Squared Error |
| MSE | Mean Squared Error |
| MLP | Multilayer Perceptron |
| MCC | Matthews Correlation Coefficients |
| OCSVM | One-Class Support Vector Machine |
| PSO-ANN | Particle Swarm Optimization Artificial Neural Network |
| Probability Distributed Function | |
| PNN | Probabilistic Neural Network |
| PDR | Packet Drop Ratio |
| RF | Random Forest |
| SVM | Support Vector Machine |
| SGD | Stochastic Gradient Descent |
| SVR | Support Vector Regression |
| SHM | Structural Health Monitoring |
| SCADA | Supervisory Control And Data Acquisition |
| TPR | True Positive Rate |
| TA-SSVM | Trend Analysis least Squares Support Vector Machine |
| WSNs | Wireless Sensor Networks |
| XGBoost | Extreme Gradient Boost |
Figure 1Types of classifiers.
Summary of related work.
| Techniques | Contributions | Limitations | Future Work |
|---|---|---|---|
| SVM and Statistical Learning Theory [ | Fault detection in WSNs | Upcoming faults are not identified, predicted, and quantified; it does not consider single-hop indoor datasets | Forecasting of upcoming faults |
| OCSVM and SGD [ | Anomaly detection in WSNs | No proposed scheme for anomaly identification mechanism | Multi-class, one-class classification problems for efficient anomaly detection |
| LNSM and PSO-ANN [ | Measured the RSSI in real environments, determined the distance between sensor and anchor nodes in WSNs, and improved the accuracy of estimated distance | For outdoor environments, the effect of anchor node density on localization accuracy was not calculated | Algorithms can be evaluated through more metrics such as MAPE and MAE |
| SVR and Neighbor Coordination [ | Fault detection for WSNs | The proposed algorithm is specific to meteorological elements, and there is no comparison with any other classifier | Comparisons with more than two classifiers for validation |
| CNN [ | Image classification and object detection | No mechanism for fault identification | Fault identification classifiers can be incorporated with CNN |
| RF and XGBoost [ | To detect fault in wind turbines | The number of features is limited to 10 | More comprehensive training data in multiple wind turbine working conditions, including all the unconsidered faults |
| SVM and TA [ | Improvised sensor fault diagnosis | Single type of ECOC coding matrix in both feature extraction and fault classification | Parameter optimization processes |
| RF and SVR [ | Anomaly detection in energy consumption sensors | No comparison with other hybrid classifiers that are trained on different datasets and features | More robust voting techniques should be explored such as weighted voting |
| Stochastic systems [ | Distributed soft fault detection in WSNs | Limitation of the network-induced delay and event-triggered mechanism | Non-linear systems with complex and limited communication |
| PNN [ | Heterogeneous fault diagnosis | No variation in the classifier | Further, this methodology can be applied on body area sensor networks, vehicular ad-hoc networks, and UWSN |
| SVM and Statistical Time-Domain Features [ | Sensor fault classification | The scheme is implemented on only five faults | More faults should be considered |
| MMSE, Multiple Hypothesis Test, and GLR Test [ | Wireless sensor fault detection, identification, and quantification | A large number of time-synchronized samples are required for MMSE identification | Future work in energy efficiency of WSNs |
| DNN and Conventional RF [ | Fault detection in a direct drive wind turbine | No evaluation through comparisons | Comparison of DNN with other neural network algorithms |
| GA [ | Fault detection in cluster head and cluster members | Limited use of the optimization technique | For better evaluation, multiple optimization techniques can be used to validate the scheme |
| LSVM [ | A data-driven framework for reliable link classification of WSNs | Not suitable for unsupervised data | The technique can be enhanced for anomaly identification in link failure |
| SVM, GA, and Tukey Test [ | To adjust the transmission rate in WSNs | No scheme for identifying and detecting anomalies in the transmission or congestion rate | Multi-classification to detect faulty nodes |
Figure 2System model of fault detection.
MCC of SVM, RFC, SGD, MLP, CNN, and PNN.
| Technique | Matthews Correlation Coefficient (MCC) | Rank |
|---|---|---|
| SVM | 0.65 | 2 |
| RF | 0.73 | 1 |
| SGD | −0.32 | 6 |
| MLP | −0.46 | 3 |
| CNN | −0.43 | 4 |
| PNN | −0.39 | 5 |
Figure 3SVM detection accuracy of fault types.
Figure 4MLP detection accuracy of fault types.
Figure 5SGD detection accuracy of fault types.
Figure 6CNN detection accuracy of fault types.
Figure 7RF detection accuracy of fault types.
Figure 8PNN detection accuracy of fault types.
Figure 9DA comparison of SVM, RF, SGD, MLP, CNN, and PNN detection of four types of fault.
Figure 10MCC score comparison of SVM, RF, SGD, MLP, CNN, and PNN classifiers.
Figure 11TPR of SVM, RF, SGD, MLP, CNN, and PNN.
Figure 12F1-score of SVM, RF, SGD, MLP, CNN, and PNN.