Literature DB >> 33364482

Alcoholic EEG signal classification with Correlation Dimension based distance metrics approach and Modified Adaboost classification.

Sunil Kumar Prabhakar1, Harikumar Rajaguru2.   

Abstract

The basic function of the brain is severely affected by alcoholism. For the easy depiction and assessment of the mental condition of a human brain, Electroencephalography (EEG) signals are highly useful as it can record and measure the electrical activities of the brain much to the satisfaction of doctors and researchers. Utilizing the standard conventional techniques is quite hectic to derive the useful information as these signals are highly non-linear and non-stationary in nature. While recording the EEG signals, the activities of the neurons are recorded from various scalp regions which has varied characteristics and has a very low magnitude. Therefore, human interpretation of such signals is very difficult and consumes a lot of time. Hence, with the advent of Computer Aided Diagnosis (CAD) Techniques, identifying the normal versus alcoholic EEG signals has been of great utility in the medical field. In this work, we perform the initial clustering of the alcoholic EEG signals by means of using Correlation Dimension (CD) for easy feature extraction and then the suitable features are selected in it by means of employing various distance metrics like correlation distance, city block distance, cosine distance and chebyshev distance. Proceeding in such a methodology aids and assures that a good discrimination could be achieved between normal and alcoholic EEG signals using non-linear features. Finally, classification is then carried out with the suitable classifiers chosen such as Adaboost.RT classifier, the proposed Modified Adaboost.RT classifier by means of introducing Ridge and Lasso based soft thresholding technique, Random Forest with bootstrap resampling technique, Artificial Neural Networks (ANN) such as Radial Basis Functions (RBF) and Multi-Layer Perceptron (MLP), Support Vector Machine (SVM) with Linear, Polynomial and RBF Kernel, Naïve Bayesian Classifier (NBC), K-means classifier, and K Nearest Neighbor (KNN) Classifier and the results are analyzed. Results report a comparatively high classification accuracy of about 98.99% when correlation distance metrics are utilized with CD and the proposed Modified Adaboost.RT classifier using Ridge based soft thresholding technique.
© 2020 The Author(s).

Entities:  

Keywords:  Alcoholism; Classification; Computer science; Correlation distance; Distance metrics; EEG

Year:  2020        PMID: 33364482      PMCID: PMC7750377          DOI: 10.1016/j.heliyon.2020.e05689

Source DB:  PubMed          Journal:  Heliyon        ISSN: 2405-8440


Introduction

One of the most common kinds of mental abuse is by means of both acute and chronic alcoholism [1]. Depending on the specific pattern of drinking and volume of alcohol consumed, the seriousness of it can be easily assessed. Alcoholism is a state where patients are obsessed with too much drinking alcohol, knowing well that alcohol can lead to a lot of impairments in the human brain and body and yet they cannot resist their desire to drink alcohol [2]. Too much of alcohol consumption leads to serious behavioral and cognitive problems in the human body by especially affecting the peripheral and central nervous systems. The patient develops an incapability to remember new things after chronic alcoholic consumption thereby taking a toll on the memory aspect too [3]. A major reason of worldwide mortality rates every year is due to alcoholism and its related disorders. Alcohol-related cancers are also on the significant rise affecting vital organs like stomach, liver, kidneys etc [4]. Alcoholism is also partially responsible for causing stomach ulcers, liver cirrhosis, pancreatic and gall bladder problems too. Alcoholism even contributes to deaths caused by accident, homicide, suicide, depression etc [5]. The social and personal relationships at colleges, offices, social gatherings, meetings etc is fully spoilt by alcoholism. Factors due to environment, genes and psychology contribute a lot to alcoholism. Because of such problems, alcoholism has to be definitely addressed in prominence so that early detection of it through simple and non-invasive techniques can save precious human lives [6]. Using the standard techniques, it is difficult to assess the cases related to alcoholism. Diagnosis of alcoholic patients in clinics is done by assessing the responses of alcoholic patients by measuring factors such as quantity of drinking, feeling of guilt, urge to drink incessantly, responses to criticism etc. Due to the societal stigmatization and fear, patients fail to reveal the exact information and so the rate of positive screening of alcoholism in hospitals is quite less. Therefore, with the advent of EEG, the measurement of the brain signals of the alcoholic patients can be easily done in a non-invasive manner [7]. For evaluating the various mental disorders like screening and diagnosis, EEG signals are predominantly utilized. By nature, EEG signals are highly dynamic and non-linear in nature and so it is utilized in various tasks such as classification of pilot mental states [8], drowsiness level detection [9], epilepsy detection [10], autism disorder detection [11], Alzheimer's disease detection [12], stroke analysis [13], consciousness and unconsciousness analysis [14], motor imagery classification [15], sleep related disorders [16], schizophrenia related disorders [17] etc. A more precise work for alcoholic EEG signal classification was done by Acharya et al in [18], where the alcoholic EEG signals was split from normal EEG signals by using non-linear features and Higher Order Spectra (HOS) features. Depending on the maximal weight matching concept, the functional connectivity in alcoholic EEG signals was done and evaluated by Zhu et al [19]. The alcoholic EEG signals were identified by their respective rhythms in [20]. The alcoholism disorder was explored by an EEG-based Machine Learning technique that utilizes resting state EEG features to classify the alcoholic patients by Mumtaz et al [21]. A wavelet filter bank technique which is orthogonal in nature and having three bands was utilized with Least Squares Support Vector Machine (LS-SVM) for identifying the alcoholic EEG signals by Sharma et al [22]. A medical expert who is trained to find out the fine variation in EEG signals between an alcoholic subject and normal subject for the diagnosis of alcoholism is quite difficult because the inspection of these subtle variations is difficult to analyze with the naked eye [23]. Moreover, to interpret the EEG frequency spectrum in a clinical manner, it involves the analysis of EEG frequency spectrum which is hectic to do. Owing to its strenuous and time-consuming nature, CAD is incorporated which involves both signal processing and soft computing techniques as a base to diagnose the disease using EEG signals [24]. A variety of signal processing techniques utilized for feature extraction are reported in [25]. Some of the commonly used soft computing techniques for analyzing the EEG signals are reported in [26]. To trace the deviation from the normal ones, certain linear and non-linear techniques [27] are utilized such as Hurst exponent, Lyapunov exponent, Fractal dimension etc. In this paper, for the alcoholic EEG signals, initially the clusters are computed with the help of CD so that the features are extracted completely. Then the appropriate features are selected using distance metrics utilized here such as correlation distance, city block distance, cosine distance and chebyshev distance. Finally, it is classified with twelve suitable post classifiers for analyzing the alcoholic risk levels in EEG signals. The organization of the paper is as follows; In section 2, the materials and methods are discussed followed by the classification methodology utilized in section 3. Section 4 discusses the results and discussion and the paper is concluded in section 5 followed by suitable reference materials.

Materials and methods

Under this topic, the information about the dataset utilized, clustering through CD methodology for feature extraction and feature selection through distance metrics is explained.

Dataset utilized

The alcoholic EEG data utilized in this work was obtained from University of California, Irvine Knowledge Discovery database (UCI KDD), commonly known as UCI KDD archive [28]. To analyze the correlation between the EEG signals and alcoholic EEG signals, this dataset was acquired. Based on the standard 10/20 International montage, this dataset has the EEG recordings of totally 122 normal and alcoholic EEG patients. Based on the standard nomenclature of the American EEG committee, the electrode positions were placed in a standard manner. The impedance of the electrode was very much less than 5. Every subject has undergone 120 trials for various stimuli. Event- related potentials (ERP) were present in the EEG signals obtained by the 64 channel electrodes. The sampling rate of the system was about 256 Hz and a resolution range of 12 bits. Undesirable artifacts caused due to muscle movement, random eye movement and body movement were eliminated with a help of a simple preprocessing Independent Component Analysis (ICA) technique. With an artifact-free EEG signal, a reasonably accurate classification of alcohol levels is quite possible. But with artifacts, this problem enhances thereby leading to cases of false detection which in turn spoils the efficacy of the classification system and so it is very vital to perform pre-processing techniques initially. Then for both the normal and alcoholic EEG datasets, recordings with the appropriate data files were considered and stored in separate file formats each having a length of 2560 samples. The simplified version of the block diagram of the work to provide an easy understanding is given in Figure 1. The EEG signals are initially preprocessed and then clustered through CD methodology so that the features are extracted and finally efficient features are selected using various distance metrics. Then it is classified using suitable post classifiers for analyzing the alcohol levels from EEG signals.
Figure 1

A simplified block diagram of the work for easy understanding.

A simplified block diagram of the work for easy understanding.

Clustering through Correlation Dimension method for easy feature extraction

For physiological signals, the CD has always been a useful parameter and complements spectral analysis to a great extent [29]. Application of CD to analyze the EEG signals obtained from Parkinson's disease, Alzheimer's disease, glaucoma, and schizophrenia are widely reported in literature [30]. The CD acts as a good diagnostic parameter for the neural changes occurring in various neurological disorders. To explain the energy levels of various neural tasks also, this CD has been used always. For a set of random points, CDis nothing but a simple measure of the dimensionality of the space occupied. To determine the fractal object dimension, the CD is of great use. Other techniques of measuring dimension are using techniques such as box-counting dimension, Hausdroff dimension, information dimension etc. Being straightforward and relatively easy to calculate, CD is widely used as it is less noisy with limited number of available points. For a particular set of points in a - dimensional space, the feature representation is expressed as: Then the correlation integral is computed aswhere denotes the total number of pairs of points. As the total number of points leads to infinity, the distance in between them leads to zero and therefore the correlation integral will assume the form as follows A log-log graph of the correlation integral versus gives an estimate of the CD, if the points are pretty large and evenly distributed. For high dimensional objects, there are several ways for points to be much closer to each other. Therefore, the number of pairs closer to each other will be rising more for high dimension. Thus, in order to differentiate random and chaotic behavior, this technique is widely used. Therefore, with the help of this CD technique, clustering is done as follows. There are 2560 samples per channel and here there are 64 such channels available thereby making the total number of EEG samples as 163840. CD is utilized to reduce the 2560 samples per channel into 256 Correlated Dimension values. This accounts for ten times reduction in the size of the sample per channel. Hence, 16384 CD features are representing the EEG signal of an alcoholic patient. In order to identify the preservation of the non-linearity in the CD initially, we have to test the CD values with histogram plot. Histogram is a simple representation of repeated patterns based on their frequency of occurrence. It is observed from the analysis of CD values of EEG signals that non-linear dynamics is responsible for the presence of non- linearity in the histogram. The histogram for a patient is depicted in Figure 2(a). Table 1 shows the average value of statistical parameters like mean, variance, skewness, and kurtosis of CD features for alcoholic EEG signals of a subject. These CD values closely resemble the CD values reported by Acharya et al [18]. Figure 2(b) illustrates the Histogram of Chi Square Probability Density Function (PDF) for CD with ten degree of freedom for alcoholic EEG Signal of a patient. It is observed from Figure 2(b) that the histogram of Chi Square PDF is highly non-linear. Therefore, to explore further presence of non-linearity in the CD values, we subject the CD through Hilbert transform process and the histogram was plotted as shown in Figure 3.
Figure 2

Histogram plot (a) CD for alcoholic EEG Signals for a patient.(b) Chi-Square PDF for C D with ten degree of Freedom for. alcoholic EEG Signal of a patient.

Table 1

Average value of Statistical Parameters of CD Features for Alcoholic EEG Signal of a Subject.

Sl. NoStatistical ParameterNumerical Value
1Mean4.091731
2Variance10.73556
3Skewness0.849836
4Kurtosis0.477126
Figure 3

Histogram of Hilbert transform of normalized CD values for alcoholic EEG signal.

Histogram plot (a) CD for alcoholic EEG Signals for a patient.(b) Chi-Square PDF for C D with ten degree of Freedom for. alcoholic EEG Signal of a patient. Average value of Statistical Parameters of CD Features for Alcoholic EEG Signal of a Subject. Histogram of Hilbert transform of normalized CD values for alcoholic EEG signal. As shown in Figures 2(a) and 3, the histogram pattern looks alike and hence, the presence of nonlinearity in the CD values are predominantly evident. Figure 4(a), (b) depicts the Cumulative Distribution Function (CDF) plot for CD values for mean and skewness parameters. The presence of peak and valley points in Figure 4(a), (b) specifies the further evidence of non-linearity in the CD features.
Figure 4

Cumulative Distribution Function (CDF) Plot for (a) CD Mean values b) CD Skewness values.

Cumulative Distribution Function (CDF) Plot for (a) CD Mean values b) CD Skewness values. Figure 5(a) exhibits the histogram of Chi Square PDF with ten degree of freedom for statistical parameters of CD values. The peaked distribution of skewness and kurtosis in Figure 5(a) indicates the non-linearity in the CD values. The mean and variance exhibit the highly flatten condition of the histogram which shows the abnormal trend in the CD values. Figure 5(b) shows the Histogram of CD with four selected Distance features. Figure 5(b) illustrate that the reduced distance feature selection process will ease the burden of the classifiers.
Figure 5

(a) Histogram of Chi Square PDF with ten degree of freedom for statistical parameters of CD (b) Histogram of CD with Distance features.

(a) Histogram of Chi Square PDF with ten degree of freedom for statistical parameters of CD (b) Histogram of CD with Distance features.

Feature selection through distance metrics

Once the clustering is done through CD methodology, then the efficient features are selected from the clusters through distance metrics concept. The distance metrics analyzed here for feature selection are Correlation Distance, City Block distance, Cosine distance and Chebyshev distance measures. Correlation Distance The correlation distance is a simple measure of dependence between 2 paired random vectors of equal dimension [31]. If the random vectors are independent, then the correlation distance coefficient is considered zero. Between two random vectors or variables, the linear and non-linear association is easily measured by Correlation Distance. Pearson's correlation can easily detect the linear relation between two random variables whereas Correlation distance can detect the linear and non-linear relation between two random variables. Let be a statistical sample considered from a pair of real valued random variables. The correlation distance of two random variables is obtained by finding the ratio of their covariance distance by the product of their standard deviation distance. The correlation distance is mathematically expressed as The correlation distance has three important properties such as: and , this is different when compared to Pearson's correlation as it can be negative , if and only if andare independent denotes that the linear subspaces designated by and samples are almost equal. City Block Distance It comes under taxicab geometry. It is a type in which the general metric of Euclidean geometry is replaced by a new metric. In such a case, the distance between the two points is the sum of the absolute differences of their respective coordinates. It is also called as Manhattan, rectilinear and taxicab metric [32]. It is defined as the sum of the lengths of the projections of a particular line segment between the points onto the coordinate axis. It is represented as follows: Chebyshev Distance The Chebyshev distance between 2 vector points and, with standard coordinates and respectively is expressed as It is also called as chessboard distance as in the game of chess, the minimum number of moves required by a king to go from one square to another equals to the Chebyshev distance between the centers of the square [33]. Cosine Distance Cosine similarity is a method utilized to measure how similar the entities are irrespective of their size [34] and is expressed as Therefore, Cosine Distance = 1 – Cosine Similarity. The scatter plot depicts how the features are closely correlated to each other. As from Figure 6(a), it is identified that a scatter plot of correlation and City block distance features are closely correlated one. From this plot the target value is selected as the maximal regression point which is identified as 0.45. Selection of target value for the classifiers will be easy in the plot due to the center cluster of the plot is settled at 0.45. Therefore, the target for the classifiers is set at 0.45.
Figure 6

Scatter Plot (a) between Correlation and City block Distance Features of CD Values (b) between Cosine and Chebyshev Distance Features of CD Values.

Scatter Plot (a) between Correlation and City block Distance Features of CD Values (b) between Cosine and Chebyshev Distance Features of CD Values.

Classification methodology

The obtained best features through distance metrics are then classified with the help of the standard Adaboost.RT algorithm, modified version of Adaboost.RT algorithm, Random Forest classifier, K means Classifier, ANN like RBF and MLP, SVM with Linear, Polynomial and RBF kernels, NBC and KNN classifier.

The Standard Adaboost.RT algorithm

The Adaboost.RT algorithm was proposed by Solomatine & Sherstha [35]. For problems dealing with regression, it is a good boost algorithm where R represents regression and T represents threshold respectively. The Adaboost.RT algorithm in its original form is explained in Algorithm 1. Standard Adaboost.RT algorithm. The following are fed as input: a specific sequence of ‘q’ samples where output weak learning procedural step The number of iterations specified as integer For approximating and differentiating the correct from incorrect predictions, a threshold is used where The following are initialized: Iteration Distribution for all Error rate The learning steps are as follows while iterating Step 1: The weak learning is called for and it is provided with approximate distribution Step 2: The regression model is then built up as Step 3: For each training example, the Absolute Relative Error (ARE) are calculated as Step 4: The error rate of is calculated as Step 5: The , where the power coefficient is denoted as Step 6: The distribution is updated as follows: If , then , else , where denotes a normalization factor chosen such that is its respective distribution. Step 7: Finally, set The final output hypothesis is as follows: The regression problem is being predicted as the binary classification problem by the Adaboost.RT algorithm. Based on the Absolute Relative Error (ARE), this algorithm can predict whether the samples are correct or incorrect. If the ARE of a sample value is not large than the threshold , then it is termed as correct predictor otherwise it is turned as incorrect predictor. In this classification problem, it gives an idea about correct classification and misclassification. To those weak learners, very large weights are being arranged by the algorithm in order to reach a high prediction rate. The final hypothesis is the simple combination of the weak learner outputs using the respective computed weights. Due to the requirement of manual selection of threshold , Adaboost.RT has a slight problem and sometimes considered unstable. Sometimes it may cause overfitting and produces a low convergence efficiency.

Modified Adaboost.RT algorithm based on Ridge, Lasso and Soft Thresholding

The proposed modification is that of the Ridge, Lasso and Soft Thresholding based Adaboost.RT algorithm. Based on the intrinsic property of the input data samples, the initial threshold values can be easily determined by Algorithm 2. When the data suffers from multicollinearity, ridge regression technique is widely used [36]. Here the Least squares estimates are unbiased, but it has a very high variance value. Hence to the regression estimates, adding degree of bias reduces the standard errors. Therefore, the prediction errors are decomposed into two sub-components due to the bias and variance. Through the shrinkage parameter, the multicollinearity problem is solved by Ridge regression. The equation is simple and expressed as , where it indicates the combination of the loss and penalty component. The first term denotes the least square term and the second term denotes the summation of the coefficient . In order to reduce the parameter to a very low variance, this term is added. The operation of Lasso regression is quite similar to ridge regression and helps to penalize the absolute size of the regression co-efficients [37]. In order to improve the accuracy and to mitigate the variability, the model is used and is expressed as . Instead of squares, absolute values are used in the penalty function and that is a major difference in the lasso regression when compared to ridge regression. Therefore, it leads to penalizing certain parameter estimates to turn to negligible condition. If the penalty applied is larger, then the estimates tend to shrink towards zero. Modified Adaboost.RT algorithm based on Ridge, Lasso and Soft Thresholding The following should be considered as input: sequence of samples , where output , weak learner and total number of iterations denoted as . The following are initialized: iteration index , distribution for all , the weight vector:for all , error rate The following are iterated while The weak learner is called , and it is provided with appropriate distribution represented as , where is a normalization factor chosen and represents the distribution. The regression model is built as and then ridge regression and lasso regression are applied to it. To introduce the soft thresholding concept now, the signal received is initially expressed as , where denotes the unknown signal which has to be detected and represents the Gaussian distribution noise factor. The soft thresholding is implemented as: Error is calculated as: The error rate is calculated as , where , denotes the expected value and is denoted as robust threshold value, where denotes the Standard deviation. Here the relative factor is told as .If , then prioritize and terminate the loop. Initialize The contribution of is calculated to the final result: . The weight vectors are updated as: If , then ; Else i) Finally set The is normalized so that. The final hypotheses output is determined as: For the approximate error distribution, the standard deviation is utilized as a criterion for every iteration of this proposed algorithm. If the data points are close to the mean values, then it indicates that there is a low Standard Deviation and on the other half, if the data points are far from the mean values, then it indicates that there is a high Standard Deviation. For a particular set of predictions, Standard deviation serves as a standard for uncertainty. From the predictions to the original values, if the average distance obtained is small, then the regression model utilized here need not be revised. The prediction accuracy set is generally low if the respective sample points falls outside the range of values. In this proposed method, the approximate error of the weak learner, , for an input dataset could be indicated as one standard distribution with parameters , where denotes the expected value and denotes the standard deviation and denotes the relative factor ranging from 0 to 1. The threshold value for is assessed by the scaled standard deviation. The generation of small prediction errors is done by the trained weak learners so that. For the regression errors, the obtained means are fluctuating around zero and is within a small range. The individual samples can fully determine the standard deviation. For most cases is very large than most outputs and is located within the ranges , thereby contributing to the instability of the boosting process. The relative factor is implemented to standard deviation to standardize the stability of the threshold value. If the samples are present within the threshold range , it is considered as accepted samples and if it falls outside the threshold ranges, then it is called rejected samples.

Random forest classifier with bootstrap resampling technique

The feature values obtained through distance metrics are utilized as input to a classifier with the sole intention of classifying the alcoholic signals. The RF classifiers mainly depend on the classification results of various classification trees [38]. Then a random vector is assigned to each tree and these assigned Random vectors has the same kind of distribution and are not dependent on one another. Therefore, to perform classification, both the training data and the random vector assigned provide the necessary support to tree to perform the classification. The validation of the classification performance is done using 10-fold cross validation technique. Then the classification performance parameters such as sensitivity, specificity and accuracy are used to evaluate the method. The RF classification algorithm is utilized in two phases. Initially, using the bootstrap resampling technique the extraction of the subsamples from the original samples are done. Secondly, the classification of the decision trees is done and then a simple vote is implemented with the largest vote of classification obtained as the ultimate result of the prediction and this procedure is shown in Algorithm 3. For a training dataset , where are features, represents a class response feature, denotes the total number of training samples, denotes the total number of features and a RF model told in Algorithm 3. Assumeis the predictor of tree given input. The prediction output of random forest with tree is given as follows Since every tree is obtained from a bagged sample set, it is grown with only two third of the sample in the training set and it is called as in-bag samples. To estimate the prediction error, out-of-bag (OOB) samples are utilized which are only about one third of the samples left out. The out of bag predicted values is expressed as , where , where represents the in bay sample index and represent the out of bay sample index. represents the size of the OOB sub dataset. The prediction error of OOB is expressed as where denotes the error function and represents the sample size of OOB. Random Forest Classification with Bootstrap Resampling Technique Input: the training dataset : the number of trees, : subspace size Output: A random forest RF For to do A bagged subset of samples is drawn to from While (failure of stopping criteria) do Random selection of m sub features are done For to do Reduction in the node input is computed The feature which mitigates the inputs to the utmost level is chosen and then the node is divided into 2 children nodes. The trees are combined to form a random forest.

ANN based classifier

To process neurobiological signals extracted from EEG, ANN's are widely used [39]. In this paper, two kinds of ANNs are analyzed to determine the most convenient configuration. RBF network with 64 neurons in the hidden layer, 32 input and 1 output linear neurons. MLP, where 32 input linear neurons and one output neuron with logistic activation function is present. It has a single hidden layer comprising 64 neurons with hyperbolic tangent as an activation function. Recognizing the non-linear data requires more hidden layers with non-linear activation functions and that is provided in MLP The training MSE and testing MSE for both the RBF architecture and MLP architecture are shown in Table 2 and Table 3 respectively. In this work, only 32-64-1 architecture has been utilized for both RBF and MLP as it produced a very less training and testing MSE.
Table 2

MSE analysis for RBF architecture.

Sl.NoRBF ArchitectureTraining MSETesting MSE
132-1-10.000193210.002506079
232-2-10.0005456237.84E-06
332-4-10.0000018.56625E-06
432-8-19.39067E-057.05094E-06
532-16-11.156E-052.84625E-06
632-32-11.96E-064.84E-06
732-64-11.44E-064.69844E-07
832-128-11.76241E-051.96E-06
932-148-12.90631E-054.41E-06
Table 3

MSE analysis for MLP architecture.

Sl.NoMLP ArchitectureTraining MSETesting MSE
132-1-10.0001639290.000001
232-2-18.836E-057.84E-06
332-4-11.369E-054E-06
432-8-10.0000366.4E-06
532-16-14.80644E-055.24E-06
632-32-11.31567E-051.156E-05
732-64-18.1E-071E-08
832-128-14.84E-064.84E-06
932-148-11E-060
MSE analysis for RBF architecture. MSE analysis for MLP architecture.

Support Vector Machine

Here SVM classifier with linear kernel, polynomial kernel order 1 and RBF kernels are utilized [40]. A hyperplane is searched as a decision surface in SVM which tends to split the two classes to the maximum extent by a very high margin. The hyperplane which is separated will be inclined to the shortest line in a perpendicular aspect thereby separating the convex hulls of each class. Assuming, is the separating hyperplane, where is its normal. For separable data, the wisely chosen optimum boundary with maximum margin criterion is obtained by mitigating, provided , for all . The support vectors subset of the training data are the solutions obtained for optimum boundary level and is attained in dual form easily. In order to get the decision boundary in an optimum manner, Quadratic programming techniques are utilized to solve this optimization problem where the random vectors are linearly combined [38]. With the aid of polynomial and RBF kernels, the non-linear boundary problems can be easily obtained. As far as the hyper parameter selection is done for SVM-RBF Classifier, the gamma value of the RBF kernel is selected such as starting from 0.2, 0.4, 0.6 and up to 2.6. While experimenting it is observed that for Gamma value of 2.0, lower MSE of 4.00E-08 is attained at the number of iterations of 250. Furthermore, in this case an increase in the number of iteration is depicted as reduced MSE value. For all other gamma values, the SVM-RBF classifiers is either plugged with local minima or has a flattened MSE effect. Therefore, for SVM-RBF Classifier the Gamma value is selected as 2.0. For SVM-Polynomial method, the order of the polynomial is controlled by using grid search method. For SVM-linear method, random search procedure is used to select the hyper parameter.

Naïve Bayesian Classifier

One of the most famous to predict the specific probability of a class membership is Naïve Bayesian classifier [41]. In bio signal processing, a high accuracy with a very low computation time is yielded by Naïve Bayesian process. The Naïve Bayesian depends on an assumption that the particular influence of an observation in a specific class is independent of the values present or obtained from other observation. Therefore, a class conditional independence is always assumed because of this criterion. Let is a particular sample that consists of ‘q’ points. Assume ‘P’ is a hypothesis for each set of that belongs to a particular class. In Bayes rule, is considered as an evidence and finds to match each point of is considered as an evidence and finds to match each point of to the probability class containing the highest posteriori.

K means classifier

It is used to solve a clustering problem and it comes under the category of unsupervised classification technique. For a given population, this technique separates it into a number of clusters on the position of K centroids for every cluster [42]. By mitigating the Euclidean distance between the cluster centroid and the observation point, it can be easily achieved. Every observation which pertains to the given population is related to the nearest centroid. To obtain the initial clustering level, this step is repeated at every iteration and therefore the calculation of the new K centroids is done.

K-nearest neighbor classifier

As it is quite robust with huge noisy datasets, it is one of the most commonly used classification techniques. It is also adaptive in nature and is known for the prediction of unknown data. Based on the frequency of a class from its nearest neighbours present in a particular feature space, the classification task is performed in such an easy manner. To define the distance in the KNN algorithm, various distance metrics are utilized. Depending on the training session, the nature of distance utilized here in this work is Euclidean [43]. The similarities with K-nearest neighbours are calculated and the class which corresponds to the maximum number of votes is assigned as the output class of the instance. As far as the hyper parameter selection is done for KNN Classifier, the value is set as 4 initially and for various values such as K = 4,6,8, the procedure is carried out to trace the optimum number of iterations for a low MSE value and after experimentation K = 4 gave a lowest MSE.

Results and discussion

In this section, the performance metrics analysis, and the summary of the previous works with the obtained results are compared and analyzed.

Performance metrics analysis

By utilizing the clustering technique through means of CD method and then utilizing distance metrics as suitable feature extraction technique, it is finally classified with the help of post classifiers. For the risk level classification of alcohol clustering and distance metrics feature extraction from EEG signals, based on the parameters like Sensitivity, Specificity, Performance Index and Accuracy, the average results are computed. The Mean Square Error (MSE), Good Detection Rate (GDR) and Error Rate analysis for the classifiers is also computed. In this work, fold cross validation is used. Initially the dataset comprising the features is split into ‘k’ equal size points. groups are then utilized to check the performance in every step. The validation is repeated for number of times and here . Therefore, 90% of the data was utilized for training and 10% for testing. The computation of the average if all performance metrics at the end of 10-fold process is done. The mathematical formulae for the Performance Index (PI), Sensitivity, Specificity and Accuracy are expressed as follows:where Perfect Classification – PC, Missed Classification – MC and the False Alarm – FA. The Sensitivity, Specificity and Accuracy measures are mathematically defined by the following GDR: It is one of the most significant criteria of a detector, representing the ability of a detector in successful detection and specified as The MSE is represented as followswhereis the observed value at time, is the target value at model; = 1 to 64, and is the total number of observations per patient in our case, it is 122. This research uses all distance features of EEG data both for training and testing classifiers. The training was progressed regressively and the MSE Values of classifiers were decreased to minimum. At most all the classifiers are trained with zero training error of MSE. The consolidated result analysis of correlation distance metric with the classifiers is shown in Table 4. The consolidated result analysis of city block distance metric with the classifiers is shown in Table 5. Table 6 explains the results of the cosine distance metric with the classifiers and Table 7 explains the results of the chebyshev distance metric with the classifiers. Table 8 explains the consolidated MSE and GDR result analysis with all the distance metrics and the classifiers.
Table 4

Consolidated Result Analysis of Correlation Distance metric with classifiers.

ClassifiersPI(%)Sensitivity (%)Specificity (%)Accuracy (%)
Adaboost.RT89.9890.6310095.315
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding98.9397.9810098.99
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding96.3129296.454110098.22705
Random Forest95.6595.8310097.915
RBF97.8797.9210098.96
MLP81.4784.3810092.19
Linear SVM84.31586.4610093.23
Polynomial SVM82.9385.4210092.71
RBF SVM86.3788.022510094.01125
NBC85.787.510093.75
K-means51.16367.18810083.594
KNN84.31586.4610093.23
Table 5

Consolidated Result Analysis of City Block Distance metric with classifiers.

ClassifiersPI(%)Sensitivity (%)Specificity (%)Accuracy (%)
Adaboost.RT83.3145886.7777810093.38889
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding97.8797.9210098.96
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding97.4197.4810098.74
Random Forest85.787.510093.75
RBF90.7891.1510095.575
MLP94.4994.7910097.395
Linear SVM77.92510080.22590.1125
Polynomial SVM84.31510086.4693.23
RBF SVM89.9890.6310095.315
NBC78.46510082.29591.1475
K-means66.667510087.5
KNN88.3810089.5994.795
Table 6

Consolidated Result Analysis of Cosine Distance metric with classifiers.

ClassifiersPI(%)Sensitivity (%)Specificity (%)Accuracy (%)
Adaboost.RT76.9281.2510090.625
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding94.4994.7910097.395
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding97.8310097.8898.94
Random Forest80.0183.3410091.67
RBF93.61593.7510096.875
MLP82.56585.1610092.58
Linear SVM88.3889.5910094.795
Polynomial SVM93.3393.7510096.875
RBF SVM96.7696.87510098.4375
NBC63.7673.4310086.71
K-means4062.510081.25
KNN77.92580.22510090.1125
Table 7

Consolidated Result Analysis of Chebyshev Distance metric with classifiers.

ClassifiersPI(%)Sensitivity (%)Specificity (%)Accuracy (%)
Adaboost.RT93.3310093.7596.875
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding92.45592.7110096.355
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding97.5768198.4080699.2777898.84292
Random Forest89.9890.6310095.315
RBF90.7891.1510095.575
MLP91.7991.9310095.965
Linear SVM85.710087.593.75
Polynomial SVM81.4710084.3892.19
RBF SVM92.01792.1910096.095
NBC63.7610073.4386.71
K-means51.16367.18810083.594
KNN75.8678.1510089.075
Table 8

Consolidated MSE and GDR Result Analysis with Distance metric and Classifiers.

Consolidated AnalysisCorrelation Distance MetricCity Block Distance MetricChebyshev Distance MetricCosine Distance Metric
ClassifiersMSEGDRMSEGDRMSEGDRMSEGDR
Adaboost.RT1.44E-0690.638.267E-0686.777784.9E-0793.333331.296E-0581.25
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding1E-0898.024E-0897.928.1E-0792.713.6E-0794.79474
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding1.873E-0796.460862.5E-0997.481.239E-0797.67054E-0897.83482
Random Forest2.5E-0795.839584E-0687.51.44E-0690.639E-0683.34
RBF4E-0897.921.21E-0691.151.21E-0691.156.4E-0793.75
MLP7.84E-0684.383.6E-0794.798.1E-0791.935.76E-0685.16
Linear SVM4.84E-0686.461.44E-0575.327274E-0685.714292.25E-0689.59
Polynomial SVM6.76E-0685.424.84E-0684.339587.29E-0681.48856.4E-0793.75
RBF SVM2.89E-0688.02251.44E-0690.636.4E-0792.199E-0896.875
NBC4E-0687.51.09E-0578.485932.3E-0563.829792.21E-0573.45
K-means4.1E-0567.182632.03E-05754.1E-0567.1888.65E-0562.5
KNN5.76E-0685.942.25E-0688.38041.68E-0578.121.37E-0580.21
Consolidated Result Analysis of Correlation Distance metric with classifiers. Consolidated Result Analysis of City Block Distance metric with classifiers. Consolidated Result Analysis of Cosine Distance metric with classifiers. Consolidated Result Analysis of Chebyshev Distance metric with classifiers. Consolidated MSE and GDR Result Analysis with Distance metric and Classifiers.

Comparison of previous works

The result analysis with the standard previous works with respect to automated detection of alcohol risk levels have been compared and analyzed with our work in Table 9.
Table 9

Comparison of our works with previous works.

AuthorsFeatures ObtainedClassification techniqueClassification Accuracy (%)
Patidar et al [44]Correlation EntropyLS -SVM97.02
Faust et al [45]Higher Order Spectra (HOS) FeaturesFuzzy Sugeno Classifier (FSC)92.40
Acharya et al [18]Entropy, HOS and LLESVM with different kernels91.7
Kannathal et al [46]CD and entropyDistinct ranges90
Proposed WorkCD with the Correlation Distance MetricsAdaboost.RT95.315
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding98.99
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding98.227
Random Forest97.915
RBF98.96
MLP92.19
Linear SVM93.23
Polynomial SVM92.71
RBF SVM94.011
NBC93.75
K-means83.594
KNN93.23
Proposed WorkCD with the City Block Distance MetricsAdaboost.RT93.388
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding98.96
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding98.74
Random Forest93.75
RBF95.575
MLP97.395
Linear SVM90.112
Polynomial SVM93.23
RBF SVM95.315
NBC91.147
K-means87.5
KNN94.795
Proposed WorkCD with the Cosine Distance MetricsAdaboost.RT90.625
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding97.395
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding98.94
Random Forest91.67
RBF96.875
MLP92.58
Linear SVM94.795
Polynomial SVM96.875
RBF SVM98.437
NBC86.71
K-means81.25
KNN90.1125
Proposed WorkCD with the Chebyshev Distance MetricsAdaboost.RT96.875
Ridge Based Modified Adaboost.RT dependent on Soft Thresholding96.355
Lasso Based Modified Adaboost.RT dependent on Soft Thresholding98.842
Random Forest95.315
RBF95.575
MLP95.965
Linear SVM93.75
Polynomial SVM92.19
RBF SVM96.095
NBC86.71
K-means83.594
KNN89.075
Comparison of our works with previous works. The main contribution in this work is the usage of CD for clustering and feature extraction and then using distance metrics for feature selection before classifying it using different post classifiers to determine the most efficient classification performance of alcoholism identification. In the last decade using a lot of linear and non-linear methods, many attempts have been utilized to discriminate alcoholic signals from normal signals. As depicted in Table 9, Patidar utilized a technique based on Tunable Q-wavelet transform (TQWT) and correntropy to identify the important differences in the alcoholic EEG signals and classify it with LS-SVM and reported a classification accuracy of 97.02% [44]. Faust et al utilized HOS features with FSC and obtained a classification accuracy of 92.40% [45]. Acharya et.al utilized approximate entropy, adaptive entropy, and HOS scheme to classify using LS-SVM and reported a classification accuracy of about 91.7% [18]. Kannathal et al. utilized CD and entropy to get the required features for alcohol detection from EEG signals and these features were helpful in the measurement of correlation and self-similarity properties and a classification accuracy of 90% was obtained [46]. In this study, about twelve different classifiers were utilized for the classification of alcoholism from EEG signals and the results show that when CD is utilized with correlation distance metrics and classified with Ridge based Modified Adaboost.RT with soft thresholding a good classification accuracy of 98.99% is obtained and when it is classified with Lasso based Modified Adaboost.RT with soft thresholding a good classification accuracy of 98.22% is obtained. When classifying with K means classifier, a low classification accuracy of 83.59% is obtained and while classifying with Random forest and RBF a classification accuracy of about 97.91% and 98.96% were reported respectively. Similarly, when CD is utilized with city block distance metrics and classified with Ridge based Modified Adaboost.RT with soft thresholding a high classification accuracy of 98.96% is obtained and when it is classified with Lasso based Modified Adaboost.RT with soft thresholding technique a high classification accuracy of 98.74% is obtained. With K-means classifier, it had a low classification accuracy of 87.5% and with MLP and RBF it reported an average classification accuracy of 97.39% and 95.57% respectively. Considering the analysis of CD with cosine distance metrics and classified with Ridge based Modified Adaboost.RT with soft thresholding technique a high classification accuracy of 97.39% is obtained and when it is classified with Lasso based Modified Adaboost.RT with soft thresholding technique a high classification accuracy of 98.94% is obtained. K-means and NBC produced a pretty less classification accuracy of about 81.25% and 86.71% respectively. With RBF and with SVM based on RBF kernel, it produced a classification accuracy of 96.87% and 98.43% respectively. When the CD is dealt with Chebyshev distance metrics for analysis and then when it is classified with Ridge based Modified Adaboost.RT with soft thresholding technique a high classification accuracy of 96.35% is obtained and when it is classified with Lasso based Modified Adaboost.RT with soft thresholding technique a good classification accuracy of 98.84% is obtained. With NBC and KNN, a less classification accuracy of about 86.71% and 89.07% is obtained and a comparatively good classification accuracy of 95.57%, 95.96% and 96.09% is obtained when utilizing RBF, MLP and SVM with RBF kernel. The primary advantages are the efficiency and versatility of the system as it supports both robustness and reliability as 10-fold cross validation approach is proved to be more versatile yielding good results. Installation of this technique is quite easier and there is zero inter-observer variability. This proposed framework seems to be good in the area of alcoholic EEG signal classification, as prior to feature extraction basic pre-processing steps were taken and so the computation is fast and simple.

Conclusion and future work

This paper incorporates the uses of signal processing and pattern recognition techniques for alcoholic risk level detection using non-linear methods. In this work, CD was utilized for clustering and then using suitable distance metrics the features were extracted. Later the classification was done using two proposed classifiers and ten existing classifiers and the performance metrics was analyzed. It is concluded that using a methodology of clustering with CD and then feature extraction with distance metrics and classifying it with appropriate classifiers yields best results and follows a very systematic methodology. In this work, overall a very good classification accuracy of 98.99% was obtained when CD with correlation distance metrics for feature extraction was classified with Ridge based Modified Adaboost.RT classifier incorporating soft thresholding technique. A very good classification accuracy of 98.96% was obtained when CD with city block distance metrics for feature extraction was classified with Ridge based Modified Adaboost.RT classifier incorporating soft thresholding technique. A very good classification accuracy of 98.94% was obtained when CD with cosine distance block distance metrics for feature extraction was classified with Lasso based Modified Adaboost.RT classifier incorporating soft thresholding technique. A very good classification accuracy of 98.84% was obtained when CD with chebyshev distance metrics for feature extraction was classified with Lasso based Modified Adaboost.RT classifier incorporating soft thresholding technique. Therefore, the proposed methodology performs well with all the kinds of distance metrics and the classifiers utilized here yielding good classification accuracy. Future works aim to study and perform the analysis with other different type of features for the efficient alcohol risk level classification.

Declarations

Author contribution statement

S. K. Prabhakar, H. Rajaguru: Conceived and designed the experiments; Performed the experiments; Analyzed and interpreted the data; Contributed reagents, materials, analysis tools or data; Wrote the paper.

Funding statement

This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors.

Data availability statement

Data will be made available on request.

Declaration of interests statement

The authors declare no conflict of interest.

Additional information

No additional information is available for this paper.
  2 in total

1.  A Holistic Strategy for Classification of Sleep Stages with EEG.

Authors:  Sunil Kumar Prabhakar; Harikumar Rajaguru; Semin Ryu; In Cheol Jeong; Dong-Ok Won
Journal:  Sensors (Basel)       Date:  2022-05-07       Impact factor: 3.847

2.  Performance Analysis of Ovarian Cancer Detection and Classification for Microarray Gene Data.

Authors:  M Kalaiyarasi; Harikumar Rajaguru
Journal:  Biomed Res Int       Date:  2022-07-15       Impact factor: 3.246

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.