Literature DB >> 35989223

DKPNet41: Directed knight pattern network-based cough sound classification model for automatic disease diagnosis.

Mutlu Kuluozturk1, Mehmet Ali Kobat2, Prabal Datta Barua3, Sengul Dogan4, Turker Tuncer5, Ru-San Tan6, Edward J Ciaccio7, U Rajendra Acharya8.   

Abstract

PROBLEM: Cough-based disease detection is a hot research topic for machine learning, and much research has been published on the automatic detection of Covid-19. However, these studies are useful for the diagnosis of different diseases. AIM: In this work, we collected a new and large (n=642 subjects) cough sound dataset comprising four diagnostic categories: 'Covid-19', 'heart failure', 'acute asthma', and 'healthy', and used it to train, validate, and test a novel model designed for automatic detection.
METHOD: The model consists of four main components: novel feature generation based on a specifically directed knight pattern (DKP), signal decomposition using four pooling methods, feature selection using iterative neighborhood analysis (INCA), and classification using the k-nearest neighbor (kNN) classifier with ten-fold cross-validation. Multilevel multiple pooling decomposition combined with DKP yielded 41 feature vectors (40 extracted plus one original cough sound). From these, the ten best feature vectors were selected. Based on each vector's misclassification rate, redundant feature vectors were eliminated and then merged. The merged vector's most informative features automatically selected using INCA were input to a standard kNN classifier.
RESULTS: The model, called DKPNet41, attained a high accuracy of 99.39% for cough sound-based multiclass classification of the four categories.
CONCLUSIONS: The results obtained in the study showed that the DKPNet41 model automatically and efficiently classifies cough sounds for disease diagnosis.
Copyright © 2022. Published by Elsevier Ltd.

Entities:  

Keywords:  Covid-19; DKPNet41; Directed knight pattern; acute asthma; cough sound; heart failure; multiple pooling

Year:  2022        PMID: 35989223      PMCID: PMC9356574          DOI: 10.1016/j.medengphy.2022.103870

Source DB:  PubMed          Journal:  Med Eng Phys        ISSN: 1350-4533            Impact factor:   2.356


Introduction

The Covid-19 pandemic has affected people's lives worldwide. Despite the availability of vaccination, many countries still experience a resurgence in Covid-19 infection rates, especially with variant viral strains [1, 2]. While the definitive diagnosis of Covid-19 infection requires confirmatory viral assays, upstream clinical triage can identify patients at risk, so that isolation can be used to avert further spread. As respiratory symptoms and signs in Covid-19 may resemble that of other conditions such as heart failure, methods that can discern the Covid-19 presentation from other ailments would be clinically useful. Artificial intelligence methods have found burgeoning applications in medicine [3], especially for the computer-aided diagnosis of diverse diseases [4], [5], [6]. Unsurprisingly, research interest in unmanned or automatic computer-aided diagnosis of Covid-19 has grown apace in the last year. Biomedical signal readouts of X-ray computed tomography and cough sound types can be fed into various machine learning or deep learning models to automatically detect Covid-19 [7]. Once validated, intelligent medical assistant systems can be prioritized for deployment in medical centers and hospitals to facilitate disease screening [8]. In this work, we focused on disease detection using cough sound signals, which are widely available and inexpensive to acquire. We collected a large dataset of cough sounds from healthy subjects as well as patients with Covid-19 infection, heart failure, and acute asthma. Acute asthma is a common disease that may present with altered cough sounds from vocal cord dysfunction [9]. Left heart failure may cause fluid to accumulate in the lungs, leading to respiratory symptoms. Nocturnal cough is a clinical sign of heart failure [10]. Artificial intelligence-based disorder detection/diagnosis has become a very popular area of research, with increasing numbers of articles on automatic disease diagnosis being published [11, 12]. Unsupervised deep learning models are often used to train, validate, and test biomedical data. Their high levels of attained classification accuracy render them suitable for many biomedical applications [13, 14]. However, automated disease diagnostic systems incorporating machine learning based on handcrafted features can also be effective [15], [16], [17]. While deep networks can achieve high accuracy with large data sets [18], handcrafted models exact lower computational costs [16]. As such, either method can be deployed in intelligent medical systems to optimize accuracy or computational efficiency [19], [20], [21]. Various automated computer-aided diagnostic methods for the detection of asthma, Covid-19, and heart failure have been published using diverse medical imaging and biomedical signals as input [22, 23]. Using the RapidMiner application [24], Yunus et al. [25] designed a heart failure classification system based on 11 data attributes (age, sex, smoking, anemia, platelets, diabetes, ejection fraction, high blood pressure, serum sodium, serum creatine, time). They tested the model on 299 samples from the UCI machine learning repository [26] and achieved classification accuracy rates of 86.95% and 94.31% with the k-nearest neighbor (kNN) and random forest implementations, respectively. Al-khassaweneh and Abdelrahman [27] used a method based on the Wigner distribution and wavelet transform to diagnose asthma, and attained a 100% accuracy rate, albeit with small study subject numbers (12 asthmatics and 12 healthy subjects). From cough sounds acquired with a system comprising a thermos camera, airflow sensor, and microphone and then uploaded to the Internet, Belkacem et al. [28] were able to diagnose Covid-19 remotely via a mobile connection. The technique incorporated a computationally demanding model based on convolution, recurrent neural networks, and mel-frequency cepstral coefficients. Brown et al. [29] reported a precision rate and area under the curve of 80% and 82%, respectively, using a support vector machine classifier. Islam et al. [30] studied an artificial neural network model for asthma diagnosis based on multichannel analyses of lung sounds collected from 30 asthmatics and 30 healthy subjects. Among the four channels analyzed, Channel 2 yielded the highest accuracy of 82.3% with an artificial neural network classifier, while the accuracy rates for Channels 1, 3, and 4 were less salutary (66.70%, 70.0%, and 63.3%, respectively, with support vector machine classifiers). With a larger dataset comprising 728 asthmatics and 522 healthy medical personnel, Badnjević et al. [31] reported a high accuracy rate of 98.85% for the detection of asthmatic and healthy subjects using a model that was also based on an artificial neural network. In the study, the training was implemented as 80% and the validation was set at 20%. Hassan et al. [32] applied a model based on a recurrent neural network and long short-term memory to analyze breath, cough, and voice signals from 20 Covid-19 and 60 healthy subjects. Implementing a 70:30 split ratio, they attained a high accuracy rate of 97% for Covid-19 detection using cough sounds. Studies in the literature generally focus on diagnosing a single disease [25]. Therefore, no automatic machine learning approach can diagnose more than one disease [31]. A major research gap in the literature is that cough-based disease diagnosis models demonstrate a low level of accuracy [29, 30]. The few studies reaching high accuracy values use small-sized datasets [27]. Another research gap in the literature is that the proposed methods have a high computational complexity [28, 32]. To fill these gaps, a new and unique dataset with a higher subject count was collected in this study. The collected dataset includes four diagnostic types. These types are: ‘Covid-19’, ‘heart failure’, ‘acute asthma’, and ‘healthy’ categories. The categories can be detected based on cough sounds. In addition, a new paradigm with low computational complexity is proposed. In this context, we developed a machine learning model based on a specifically directed graph called the directed knight pattern (DKP) that outputs 41 feature vectors (DKPNet41) for downstream feature vector and feature selection as well as classification. A machine learning method should not only generate features but must also be able to select the salient feature vectors and features. We employed a histogram-based function using DKP as the main feature generator, in which two kernels and chess moves were used to generate 1,536 features per run. As the pattern itself cannot discern high-level features, an appropriate pooling method for decomposition is needed for multilevel feature generation. However, pooling methods are not satisfactory for routing according to the capsule network [33] as only peak values are routed with maximum pooling. To ameliorate, we deployed four pooling methods: average, minimum, maximum, and conditional pooling, to generate forty sub-bands. Together with the original cough sound signal, 41 feature vectors were obtained from which the DKPNet41 selected the ten best feature vectors to create the final vector of length 1536  ×  10 = 15360. We next used the iterative neighborhood component (INCA) feature selector [34] to sieve out the most informative 726 features from the extracted 15,360 features in the final vector, which were then input to the kNN classifier [35]. This new model was devised with high accuracy and low computational complexity in mind. The model was trained and validated on a new large multiclass cough sound dataset that was prospectively acquired. The research gaps evident in the literature, such as a small number of subjects, low accuracy, high computational complexity, and lack of satisfactory diagnosis when there are multiple disease categories, were rectified with this study. In this context, the main contributions of this research are: Artificial intelligence-enabled cough sound classification an urgently needed area of research. The new multiclass cough sound dataset was prospectively acquired and made publicly accessible with hyperlink http://web.firat.edu.tr/turkertuncer/acute_asthma_cough.rar. A novel DKP-based model was presented that combined chess moves and two binary feature extraction kernels to generate features. The handcrafted DKPNet41 is a computationally lightweight cognitive learning model comprising multiple pooling decomposition, multilevel feature extraction, iterative feature selector, and classification steps. Multilevel multiple pooling decomposition combined with directed knight pattern yielded 41 feature vectors (40 extracted plus one original cough sound). From these, the best ten feature vectors were selected. Based on each vector's misclassification rate, redundant feature vectors were eliminated, and then merged. The merged vector's most informative features automatically selected using INCA were input to a standard kNN classifier.

Dataset

From December 2020 to April 2021, 842 subjects who were diagnosed with Covid-19, heart failure, or acute asthma, or who were assessed as healthy (see Table 1 ), were recruited from the pulmonology and cardiology clinics of Firat University Hospital, Elazığ, Turkey. Smartphone microphones were used to collect cough sounds, which were saved in ogg, mpeg, mp4, or m4a formats at sampling rates of 44.1 or 48 kHz, depending on the type of mobile phone. We recorded 120, 2, 1, and 719 participants’ cough sounds in ogg, mp4, mpeg, and m4a format, respectively, to show the robust classification ability of the proposed model. Furthermore, we have not used any preprocessing algorithm, only cough sounds have been segmented. In addition, few speech signals were deleted manually. These cough sounds were collected from definitively diagnosed patients. For diagnosis, doctors checked PCR test results of the Covid19 patients, and thereafter we collected cough sounds from these patients. Each cough sound signal was subdivided into smaller segments (the length of each segment is one second) for a total of 2944 samples and 842 subjects used in the analysis (see Figure 1 for examples). The collected dataset is publicly accessible and downloadable at the hyperlink http://web.firat.edu.tr/turkertuncer/acute_asthma_cough.rar.
Table 1

Details of cough sound dataset.

NoClassNumber of subjectsNumber of samples
1Acute asthma110787
2Healthy247696
3Covid-19241907
4Heart failure244554
Total8422,944
Fig. 1

Spectrogram images of the four cough sample of our collected dataset per the class.

Details of cough sound dataset. Spectrogram images of the four cough sample of our collected dataset per the class. To denote the distinctive of these cough sounds per the classes, we extracted spectrogram images of cough sound samples, and the obtained images are shown in Figure 1. As can be seen from spectrogram images (see Figure 1), these cough has variable attributes and Figure 1 clearly demonstrated seperable attributes of the collected cough sounds per the category.

Method

Feature engineering using unsupervised deep learning is computationally demanding. In this study, we designed a novel computationally lightweight machine learning model that employed a specific DKP-based pattern -where the direction of each edge defined the generated bits- as the primary local feature generation function to develop handcrafted binary features from two directed kernels, i.e., signum and ternary. The DKP is a local textural feature generator that generates 1,536 features per run, but it cannot perform high-level feature extraction, for which a decomposer is generally utilized. Although pooling decomposers have been employed to create layers with deep learning models such as convolutional neural networks [36] and multi-layer perceptron mixers [37], pooling decomposers are not necessarily the optimal routers for handcraft modeling. Instead, we deployed four pooling methods in a ten-level architecture to generate 4 × 10 sub-bands. Together with the original cough sound signal, 41 feature vectors were obtained per cough sound signal segment. From these 41 feature vectors, the best ten feature vectors were selected based on the individual vectors’ accuracy rates to eliminate redundancy. Then and then merged to form a final vector of size 10 × 1,536 = 15,360. From the merged vector, INCA automatically selected the most informative 726 features, which were then input to a standard kNN classifier (see Figure 2 ). The detailed steps of the DKPNet41 model are explained in the following sections.
Fig. 2

Schematic of the directed knight pattern-based model.

Schematic of the directed knight pattern-based model. As given in Figure 2, the model acquires cough sound as input. In the study, segmented cough sounds are used. The developed model produces features from the main cough and decomposition sounds. In this phase, DKP is employed as a unique feature extraction approach. In the decomposition process, sub-bands are obtained using the average pooling method, and features are produced from each sub-band using the DKP method. The decomposition process is repeated ten times. Each iteration produces four feature vectors via the pooling process. In the next phase, the model selects and combines the best feature vectors (projection from 41 feature vectors to 10 feature vectors). Finally, the most significant features are selected from the feature vector, and these features are classified as categories ‘acute asthma’, ‘Covid-19’, ‘heart failure’, and ‘healthy’. The details of the proposed method are given in the subsections.

Feature extraction

Read cough sound. Create sub-bands using average, maximum, minimum, and conditional pooling functions. These functions use two-sized non-overlapping windows to decompose cough sounds. In algorithm 1 , the maximum, minimum and average pooling methods were used with maximum, minimum, and average statistical moments, respectively. Conditional pooling used the three statistical moments according to the condition defined in Equation 1.
Algorithm 1

Pseudocode for sub-band creation using multilevel multiple pooling.

Input: Cough sound (cs)Output: 40 sub-band (sb)
00: Read cs from the collected dataset.01: cnt = 1;02: for k=1 to 10 do03: sbcnt = minp(cs); // Minimum pooling applying.04: sbcnt + 1 = maxp(cs); // Maximum pooling applying.05: sbcnt + 2 = avep(cs); // Average pooling applying.06: sbcnt + 3 = conp(cs); // Conditional pooling applying.07: cs = sbcnt + 2; // Decompose cough sound using average pooling and update08: cnt = cnt + 4;09: end for k
Pseudocode for sub-band creation using multilevel multiple pooling. Where the conp(.) function represents conditional pooling; x, the input vector; L, the length of x vector; and max (.), min (.),  ave(.), the maximum, minimum, and average statistical moments, respectively. Next, DKP is deployed for local feature extraction using the 40 sub-bands generated in Algorithm 1, plus the original cough sound.where f is the tth generated feature vector; and DKP(.), the proposed DKP feature generator function. The latter is explained in detail below.where b and x are nine-sized overlapping blocks and a one-dimensional signal, respectively. Extract 1,536 features from each sub-band and cough sound using DKP. Step 2.1: Deploy overlapping window division. Here, the one-dimensional signal input is divided into nine overlapping blocks. Step 2.2: Generate a 3 × 3 sized matrix by deploying vector-to-matrix transformation. In Equation 5, m defines a 3 × 3 sized matrix, which is used to apply the DKP. Step 2.3: Apply the DKP using the created 3 × 3 sized matrix (see Figure 3 ).
Fig. 3

Graphical depiction of directed chess moves (knight patterns) and directions used to generate bits.

Graphical depiction of directed chess moves (knight patterns) and directions used to generate bits. In Figure 3, the DKP consists of two directed graph patterns denoted as black and red arrows, with edges (enumerated) that can be used to generate the binary features. The black arrows trace the starting and destination points of sequential knight moves on a 3 × 3 matrix computerized “chessboard", while the red arrows point to the destination points after each corresponding knight move, with reference to the center of the “chessboard”. Directed upper and lower ternary functions and directed signum functions are employed to generate bits. Using the directions in the analytical plane (see Figure 3), bits are calculated as follows.andwhere lth bit of zth bit group using zth directed kernel (∂). Here, Equation 6 and Equation 7 define feature bit generation using the first and second graphs depicted using black and red arrows, respectively, in Figure 3. In both graph patterns, six-bit groups, each with a length of eight, are generated. Here, the main feature generators are directed kernels defined in equations (8) to (11). where ∂1(.,.), ∂2(.,.), and ∂3(.,.) define the directed signum, directed upper ternary and directed lower ternary binary feature generators, respectively. a, b are input values; dir., the direction, which is calculated using the edge (see Figure 3); and d, threshold value, which is calculated using the standard deviation (std(.)) of the one-dimensional input signal (s). Step 2.4. Generate bits using both DKP patterns and directed kernels. Step 2.5: Calculate map signals by deploying binary-to-decimal conversion. From Equation 12, the DKP generates six maps (mp) signals for feature extraction. Step 2.6: Extract histograms of the generated six map signals. In Equation 13, the initial histogram values are assigned null values. Equation (14) defines the histogram extraction mathematically. Step 2.7: Merge the extracted six histograms to obtain a feature vector of 256  ×  6 = 1, 536. Here, mf represents the merged features. Calculate the accuracy rates of each of the extracted 41 feature vectors in Step 2, and deploy the kNN classifier with five-fold cross-validation (see Figure 4 ).
Fig. 4

Accuracy rate calculation of the generated 41 feature vector.

Accuracy rate calculation of the generated 41 feature vector. Select top feature vectors based on individually calculated accuracy rates in Step 3. The proposed DKPNet41 is a parametric method. This work selects the top ten feature vectors to obtain optimum performance. Concatenate the top ten feature vectors to obtain the final feature vector (X) of length 15,360.

Feature selection using INCA

Select top informative features by deploying INCA [34]. INCA is an iterative and improved version of neighborhood components analysis. The latter is a weight-based selector without the ability to automatically select the most suitable number of features, unlike INCA, which is a parametric selector. These parameters are the misclassification rate generator, the initial value of iteration, and the end value of iteration. The parameters used in this work are listed in Table 2 .
Table 2

Parameters of the INCA selector in the model.

ParameterValue
Misclassification rate generatorkNN (k is 1; distance, Manhattan; voting, none; and standardize, true)
Initial value40
End value1000
Parameters of the INCA selector in the model. By using these parameters, the INCA selector evaluates 1000 − 40 + 1 = 961 feature vectors to select the most appropriate features, i.e., 726 features from the 15,360 features in the final merged vector from Step 5.

Classification

Classify the chosen 726 features by deploying standard kNN (parameters are defined similar to Table 2) with ten-fold cross-validation.

Performance analysis

To evaluate the performance of the DKPNet41 model, we evaluated standard performance evaluation metrics based on the mathematical formulae given below. where tp denotes true positives; tn, true negatives; fp, false positives; fn, false negatives; Acc, accuracy; Pre, precision; Rec, recall; and F1, F1-score (harmonic average of precision and recall). Individual performance metrics for all and individual diagnostic categories, respectively, were calculated and then listed in Table 3 .
Table 3

Calculated overall and individual performance metrics of the DKPNet41 model using ten-fold cross-validation.

Performance metricCategoryResults (%)
AccuracyOverall99.39
RecallAcute asthma100
Healthy100
Covid-1998.02
Heart failure100
Overall unweighted average recall99.51
PrecisionAcute asthma99.24
Healthy98.31
Covid-19100
Heart failure100
Overall99.39
F1-scoreAcute asthma99.62
Healthy99.15
Covid-1999
Heart failure100
Overall99.45
Calculated overall and individual performance metrics of the DKPNet41 model using ten-fold cross-validation. As provided in Table 3, the developed model reached an average accuracy of 99.39% with the ten-fold-cross-validation strategy. In addition, the model achieved a 100% classification success for acute asthma, healthy, and heart failure categories. The confusion matrix of these results is shown in Figure 5 .
Fig. 5

Confusion matrix of the DKPNet41 model using ten-fold cross-validation.

Confusion matrix of the DKPNet41 model using ten-fold cross-validation. In summary, the DKPNet41 model attained excellent overall multiclass classification accuracy, unweighted average recall, and an F1-score of 99.39%, 99.51%, and 99.45%, respectively, with ten-fold cross-validation. Moreover, fold-wise accuracies were calculated, and the fold-wise results were consistently excellent (see Figure 6 ). Fold-wise results were calculated to demonstrate the consistency of the proposed method. As can be seen from Figure 6, the proposed method indicates a very high classification success for all folds.
Fig. 6

Fold-wise classification accuracies.

Fold-wise classification accuracies. As an alternative to cross-validation, we performed hold-out validation at a range of split ratios: 90:10, 80:20, 70:30, 60:40, and 50:50. Similar high accuracy rates, unweighted average recall, and F1-scores at varying split ratios (Table 4 ) were obtained, which confirmed the robustness of the DKPNet14 model.
Table 4

Performance metrics at varying split ratios using hold-out validation.

Split ratioAccuracy (%)Unweighted average recall (%)F1 score (%)
90:10100100100
80:2099.8399.8699.85
70:3098.5398.6998.63
60:4098.0597.9698.09
50:5097.0897.0297.08
Performance metrics at varying split ratios using hold-out validation. The results given in Table 4 show that the proposed method outperforms the hold-out validation strategy. In addition, the high success rate applies to all split ratios. The results are given in Figure 6, and Table 4 thus show the superiority of the proposed method. Finally, we performed a complexity to assess the model's time burden using big O notation (Table 5 ). This analysis is provided in Table 5.
Table 5

Complexity analysis of the DKPNet41 model.

StepComplexity
Multilevel multiple pooling decompositionO(klogn)
DKP-based feature generationO(nlogn)
Top ten feature vector selectionO(knd)
Feature vector mergingO(d)
INCAO(d2 + hnd)
Classification with kNNO(nd)
TotalO(klogn + nlogn + d(d + hn + n))
n is the length of the vector; k, the number of feature vectors; d, the dimension; and h, the number of iterations
Complexity analysis of the DKPNet41 model. The most complex step of the DKPNet41 model is feature selection with INCA, which possesses a relatively low computational complexity. However, when the total complexity is examined, it is apparent that the proposed method has low computational complexity. Furthermore, the obtained accuracy values (fold-wise and split ratio(s) results) and computational complexity demonstrate that the proposed method is of practical use.

Discussion

We have developed a novel computationally efficient handcrafted DKPNet41 model for the automated diagnosis of cough sounds and confirmed its accurate performance by training and validating the model on a new large multiclass cough sound dataset that was prospectively acquired from healthy subjects and patients with Covid-19, heart failure and acute asthma. As far as we know, the method proposed in this study is the first of its kind and produces results that can be used in the preliminary diagnosis of more than one disease. The proposed method can automatically classify cough sounds collected through a simple microphone with a high accuracy rate. In addition, this method has low computational complexity. For performance, multilevel pooling-based decompositions and DPK feature generation function were applied to cough sounds to generate 40 sub-bands with 1,536 features each. Forty-one feature vectors (40 sub-bands plus original cough sound) were obtained. The model automatically selected the top ten feature vectors with the highest accuracy rates (i.e., least miscalculation) using kNN with 5-fold cross-validation (Figure 7 ), which were then merged to form a final feature vector of size 15,360. INCA was deployed to select from the final merged feature vector the most informative 726 features, which were then fed into the final kNN classifier. kNN was deployed at three steps in the DKPNet41 model: selection of ten feature vectors selection, selection of the most informative features from the merged final feature vector using INCA, and final classification. Regarding the latter, we tested and compared the results of various classifiers on our model. These classifiers can be broadly separated into five categories: decision tree (DT), discriminant (D), naïve Bayes (NB), support vector machine (SVM), and kNN. In this work, we have used Fine DT, Medium DT, Coarse DT, Linear D, Gaussian NB, Kernel NB, Linear SVM, Quadratic SVM, Cubic SVM, Fine Gaussian SVM, Medium Gaussian SVM, Coarse Gaussian SVM, Fine kNN, Medium kNN, Coarse kNN, Cosine kNN, Cubic kNN and Weighted kNN classifiers enumerated from 1 to 18, respectively. These 18 classifiers were implemented using the MATLAB classification learner tool with default settings. Figure 8 shows the calculated accuracy rates after testing. The classifier with the best results was Fine kNN (13th classifier), which attained a 98.23% accuracy using default settings.
Fig. 7

Individual accuracy rates of all 41 generated feature vectors. The best and worst calculated classification accuracy rates were with the first or original cough sound signal (98.51%) and the 40th feature vector (93.65%), respectively.

Fig. 8

Accuracy rates of the 18 tested classifiers

Individual accuracy rates of all 41 generated feature vectors. The best and worst calculated classification accuracy rates were with the first or original cough sound signal (98.51%) and the 40th feature vector (93.65%), respectively. Accuracy rates of the 18 tested classifiers To refine the Fine kNN, Bayesian optimization [38] was implemented with 30 iterations to tune its parameters. At initialization, k was set equal to 1; the distance metric was Manhattan, and the voting value was null. By doing so, the accuracy rate climbed from 98.23% to 98.51%. The top ten feature vectors were concatenated to improve accuracy, and INCA was applied to the merged feature vector. INCA evaluated 961 feature vectors. The misclassification rates of these feature vectors (see Figure 9 ) was lowest at a feature vector length of 726, and the classification accuracy peaked at 99.39% at this reduced feature vector length.
Fig. 9

Misclassification rates of the 961 feature vectors during INCA feature selection.

Misclassification rates of the 961 feature vectors during INCA feature selection. The 726 features were selected from the ten best feature vectors. The top ten feature vectors are: 1st (raw signal), 3rd, 7th, 6th, 9th, 4th, 8th, 13th, 12th, and 5th feature vectors, and these feature vectors contain 1536 features. Moreover, the classification rates of these features range from 97.89% to 98.51%. The selected number of the features per used feature vector is depicted in Figure 10 .
Fig. 10

The distribution of the selected features per generated feature vector.

The distribution of the selected features per generated feature vector. As can be seen from Figure 10, we incorporated the chosen ten feature vectors to create a final feature vector. The performance of the DKPNet41 model compares favorably with other state-of-the-art cough sound classification methods, attaining superior accuracy rates for multiclass classification. Our study results as compared with other results in the literature, are summarized in Table 6 .
Table 6

Comparison of DKPNet41 model with published cough sound classification methods.

StudyMethodClassifierSubject/ SamplesSplit ratioResults (%)
Amrulloh et al. [39]Neural networkArtificial neural network9 asthma9 pneumoniaLeave one out validationSen: 89.00Spe: 100.0Kappa: 89.00
Hee et al. [40]Mel-frequency cepstral coefficients, constantQ cepstral coefficientsGaussian mixture model–universal background model89 asthma89 healthy70:30Sen: 82.81Spe: 84.76
Yadav et al. [41]Mel-frequency cepstralcoefficientsSupport vector machine47 asthma48 healthy5-fold cross validationAcc: 77.80
Mouawad et al. [42]Recurrence quantification analysis, Mel frequency cepstral coefficientsXGBoost20 Covid-191468 healthy20-fold cross validationAcc: 97.00AUC: 84.00
Knocikova et al. [43]Wavelet transformLinear discriminant analysis26 healthy17 asthma22 chronicobstructive pulmonary diseaseLeave one subject out validationAcc: 90.00
Loey and Mirjalili [44]Convolutional neural networksConvolutional neural networks114 Covid-19102 healthy70:15:15Acc: 94.90Sen: 94.44Spe: 95.37
Islam et al. [45]Mel frequency cepstral coefficientsDeep neural network50 Covid-1950 healthy5-fold cross validationAcc: 97.50
Chowdhury et al. [46]Multi-criteria decision making, Mel frequency cepstral coefficientsExtra-Trees622 Covid-191610 healthy10-fold cross validationAUC: 78.00Pre: 61.00Rec: 67.00
Bagad et al. [47]Convolutional neural networksSoftMax376 Covid-19663 healthy5-fold cross validationAUC: 68.00
Pahar et al. [48]Convolutional neural network, long short term memoryConvolutional neural network1. 92 Covid-191079 healthy2. 8 Covid-1913 healthyLeave-p-out crossvalidation1. Acc: 95.30Sen: 93.00Spe: 98.00AUC: 97.592. Acc: 92.91Sen: 91.00Spe: 96.00AUC: 93.75
Pal and Sankarasubbu [7]Deep neural networks, Mel frequency cepstral coefficientsDeep neural network150 Covid-19, bronchitis, healthy, asthmaNAF1: 90.60Pre: 90.40Sen: 90.10Spe: 90.30Acc: 90.80 for cough data
Our studyDirected knight patternkNN110 acute asthma247 healthy241 Covid-19244 heart failure1. 90:102. 80:203. 70:304. 60:405. 50:506. 10-fold cross-validationAcc:1. 100.02. 99.833. 98.534. 98.055. 97.086. 99.39
Comparison of DKPNet41 model with published cough sound classification methods. To date, there is no automated cough sound classification study in the literature in which ‘acute asthma’, ‘Covid-19’, ‘heart failure’ and ‘healthy’ classes are used together. Therefore, we present a novel automated cough sound classification model using a four-class system. Many studies use Covid-19 versus Control classes or Asthma versus healthy classes in the literature. These works solve a binary classification problem by analyzing a cough sound dataset from a two-class perspective. Others, such as Pal and Sankarasubbu [7] used a dataset with four classes (Table 6). The classes are ‘Covid-19’, ‘bronchitis’, ‘healthy’ and ‘asthma’. They used a deep neural network to classify the sounds. Pal and Sankarasubbu's [7] method attained a 90.80% accuracy rate for these four classes. By comparison, our DKPNet41 attained a 99.39% accuracy for four-class classification. Additionally, Pal and Sankarasubbu's [6] method's computational complexity is higher than ours. Another multiclass classification method was proposed by Knocikova et al. [43]. ‘Asthma’, ‘chronic obstructive pulmonary disease’ and ‘healthy’ classes were categorized in this study. Knocikova et al. [43] achieved a 90% classification accuracy in the study. In this context, the highlights of our DKPNet41 model are listed below. The benefits of this are: Cough sounds can be easily collected at a low cost using a simple smartphone microphone. We prospectively collected and made available a large multiclass cough sound dataset that can be downloaded as well by other investigators for new model development. It can be noted from Table 6 that this is the first paper to classify normal, asthma, Covid-19, and heart failure using cough sounds (4-class system). DKPNet41 achieved the highest classification performance using six different validation techniques. Hence, the proposed DKPNet41 model is robust and accurate. We studied cough sounds from patients with Covid-19, heart failure, and acute asthma patients. The latter is a common disorder that may not manifest abnormality on lung imaging, which precludes the automated diagnosis of asthma using an image-based classification model. The DKPNet41 model uses a novel graph pattern to extract handcrafted local features and is then able to cognitively select the best feature vectors and features automatically. The model possesses an excellent multiclass classification accuracy of 99.39% for cough sound-based diagnosis. The model is computationally lightweight with low complexity and time burden (Table 5). The Limitations of this method are as follows: A larger dataset can be collected from more patients to test the DKPNet41 model. More disorders can be detected using cough sounds.

Conclusions

We have developed a novel handcrafted DKPNet41 learning model and trained and validated it on a new cough sound dataset. The model extracted features using a new local feature generator based on DKP, combined with two directed kernels to generate comprehensive features. Forty sub-bands were generated using four pooling decomposers. Together with the original cough sound signal, 41 feature vectors underwent further automatic feature vector and feature selections. From accuracy measures employed to analyze the 41 feature vectors, the top ten feature vectors were selected, and then the most informative features from these ten feature vectors were chosen using the INCA selector. INCA selected 726 features per run fed to a standard kNN classifier. The model attained a 99.39% accuracy with ten-fold cross-validation. Hold-out validation analyses at 90:10, 80:20, 70:30, 60:40 and 50:50 split ratios demonstrated consistently high corresponding accuracy rates of 100%, 99.83%, 98.53%, 98.05% and 97.08%, respectively, which support the robustness of the model. The limitation of this work is that we have used a relatively small dataset to develop our proposed DKPNet41 system. In the future, we will validate our automated system with larger cough datasets. New intelligent chip diagnostic models and applications can be developed for artificial intelligence-enabled cough sound diagnosis. Our work demonstrates that cough sound-based intelligent disease diagnostics is feasible and can be implemented in pulmonology and cardiology clinics. Newer handcrafted learning models/networks similar to DKPNet41 can be developed that will be more computationally efficient than the available deep networks. In the near future, a new generation handcrafted feature-based deep networks/models can be devised. It is among our future plans to test and revise the developed model using more data and disease types. In addition, it is aimed to ensure its practical usability by integrating it into an embedded card.

Authors’ contributions

Conceptualization:Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua, Sengul Dogan, Turker Tuncer, Ru-San Tan, Edward J. Ciaccio, U Rajendra Acharya; Methodology: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua, Sengul Dogan, Turker Tuncer; Software: Turker Tuncer; Validation: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua; Formal analysis: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua; Investigation: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua, Sengul Dogan, Turker Tuncer, Ru-San Tan, Edward J. Ciaccio, U Rajendra Acharya; Resources: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua, Sengul Dogan, Turker Tuncer; Data Curation: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua; Writing - Original Draft: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua, Sengul Dogan, Turker Tuncer; Writing - Review & Editing: Prabal Datta Barua, Ru-San Tan, Edward J. Ciaccio, U Rajendra Acharya; Visualization: Mutlu Kuluozturk, Mehmet Ali Kobat,Prabal Datta Barua; Supervision: Ru-San Tan, Edward J. Ciaccio, U Rajendra Acharya; Project administration: U Rajendra Acharya.

Funding

The authors state that this work has not received any funding.

Ethical approval

This research has been approved on ethical grounds by Non-Invasive Ethics Committee, Firat University on 08 April 2021 (2021/05-10).

Conflicts of Interest

The authors of this manuscript declare no conflicts of interest.
  22 in total

1.  Multichannel lung sound analysis for asthma detection.

Authors:  Md Ariful Islam; Irin Bandyopadhyaya; Parthasarathi Bhattacharyya; Goutam Saha
Journal:  Comput Methods Programs Biomed       Date:  2018-03-09       Impact factor: 5.428

2.  Unrecognized heart failure in elderly patients with stable chronic obstructive pulmonary disease.

Authors:  Frans H Rutten; Maarten-Jan M Cramer; Diederick E Grobbee; Alfred P E Sachs; Johannes H Kirkels; Jan-Willem J Lammers; Arno W Hoes
Journal:  Eur Heart J       Date:  2005-04-28       Impact factor: 29.983

3.  Wavelet analysis of voluntary cough sound in patients with respiratory diseases.

Authors:  J Knocikova; J Korpas; M Vrabec; M Javorka
Journal:  J Physiol Pharmacol       Date:  2008-12       Impact factor: 3.011

4.  Arrhythmia detection using deep convolutional neural network with long duration ECG signals.

Authors:  Özal Yıldırım; Paweł Pławiak; Ru-San Tan; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2018-09-15       Impact factor: 4.589

5.  Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks.

Authors:  Ali Narin; Ceren Kaya; Ziynet Pamuk
Journal:  Pattern Anal Appl       Date:  2021-05-09       Impact factor: 2.580

6.  Automated detection of COVID-19 cases using deep neural networks with X-ray images.

Authors:  Tulin Ozturk; Muhammed Talo; Eylul Azra Yildirim; Ulas Baran Baloglu; Ozal Yildirim; U Rajendra Acharya
Journal:  Comput Biol Med       Date:  2020-04-28       Impact factor: 4.589

7.  The COVID-19 epidemic.

Authors:  Thirumalaisamy P Velavan; Christian G Meyer
Journal:  Trop Med Int Health       Date:  2020-02-16       Impact factor: 2.622

8.  Robust Detection of COVID-19 in Cough Sounds: Using Recurrence Dynamics and Variable Markov Model.

Authors:  Pauline Mouawad; Tammuz Dubnov; Shlomo Dubnov
Journal:  SN Comput Sci       Date:  2021-01-12

9.  COVID-19 cough sound symptoms classification from scalogram image representation using deep learning models.

Authors:  Mohamed Loey; Seyedali Mirjalili
Journal:  Comput Biol Med       Date:  2021-11-10       Impact factor: 4.589

10.  COVID-19 cough classification using machine learning and global smartphone recordings.

Authors:  Madhurananda Pahar; Marisa Klopper; Robin Warren; Thomas Niesler
Journal:  Comput Biol Med       Date:  2021-06-17       Impact factor: 4.589

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.