| Literature DB >> 29558434 |
Duona Zhang1, Wenrui Ding2, Baochang Zhang3, Chunyu Xie4, Hongguang Li5, Chunhui Liu6, Jungong Han7.
Abstract
Deep learning has recently attracted much attention due to its excellent performance in processing audio, image, and video data. However, few studies are devoted to the field of automatic modulation classification (AMC). It is one of the most well-known research topics in communication signal recognition and remains challenging for traditional methods due to complex disturbance from other sources. This paper proposes a heterogeneous deep model fusion (HDMF) method to solve the problem in a unified framework. The contributions include the following: (1) a convolutional neural network (CNN) and long short-term memory (LSTM) are combined by two different ways without prior knowledge involved; (2) a large database, including eleven types of single-carrier modulation signals with various noises as well as a fading channel, is collected with various signal-to-noise ratios (SNRs) based on a real geographical environment; and (3) experimental results demonstrate that HDMF is very capable of coping with the AMC problem, and achieves much better performance when compared with the independent network.Entities:
Keywords: automatic modulation classification; classifier fusion; convolutional neural network; deep learning; long short-term memory
Year: 2018 PMID: 29558434 PMCID: PMC5876703 DOI: 10.3390/s18030924
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1Illustration of the traditional and classifier methods in this study for automatic modulation classification (AMC). The traditional methods usually separate feature extraction and the classification process. Meanwhile, they usually employ handcrafted features, which might contribute to limitations in representing the samples. By contrast, we deploy deep learning to solve the AMC problem, due to its high capacity for feature representation. In addition, deep learning is generally performed in the end-to-end framework, which performs the feature extraction and classification in the same process. Our deep methods achieve a much lower computational complexity during testing compared with the training process. The upshot is that AMC is implemented more efficiently with a heterogeneous deep model fusion (HDMF) method.
Figure 2Long short-term memory (LSTM) memory cell structure.
Figure 3Fusion model structure of heterogeneous deep model fusion (HDMF) in parallel and series modes. We note that two HDMF models are used separately to solve the AMC problem.
Figure 4The geographic simulation environment. (a) Short-distance perspective of the real geographical environment; (b) Long-distance perspective of the real geographical environment.
Dataset descriptions.
| Content | Detailed description |
|---|---|
| Modulation mode | Eleven types of single-carrier modulation modes (MASK, MFSK, MPSK, MQAM) |
| Carrier frequency | 20 MHz to 2 GHz |
| Noise | 0 dB to 20 dB |
| Attenuation | A fading channel based on a real geographical environment |
| Sample value | 22,000 samples (11,000 training samples and 11,000 test samples) |
Figure 5Classification accuracy of convolutional neural network (CNN) and LSTM models. (a) Classification accuracy of CNN when the number of convolution kernels is from 8 to 64; (b) Classification accuracy of CNN when the size of convolution kernels is from 10 to 40; (c) Classification accuracy of CNN when the number of convolution layers is from 1 to 4; (d) Classification accuracy of Bi-LSTM when the number of memory cells is from 16 to 128; (e) Classification accuracy of Bi-LSTM when the number of hidden layers is from 1 to 3.
Training parameters and computational complexity of CNNs.
| Kernels | Parameters (M) | Training Time (s) | Testing Time (s) | |
|---|---|---|---|---|
| CNN1 (with size 20) | 8 | 1.537 | 72 | 0.4 |
| 16 | 3.073 | 96 | 0.6 | |
| 32 | 6.146 | 118 | 1.1 | |
| CNN2 (with size 20) | 8-8 | 1.539 | 96 | 1.0 |
| 16-16 | 3.079 | 144 | 1.5 | |
| 32-32 | 6.166 | 250.5 | 2.85 | |
| CNN3 (with size 20) | 8-8-8 | 1.540 | 148 | 1.55 |
| 16-16-16 | 3.084 | 196 | 2.16 | |
| 32-32-32 | 6.187 | 420 | 4.3 | |
| CNN4 (with size 20) | 8-8-8-8 | 1.541 | 165 | 2.3 |
| 16-16-16-16 | 3.089 | 296.5 | 3.3 | |
| 32-32-32-32 | 6.207 | 507.5 | 5.9 |
Classification accuracy of different methods without noise.
| Methods | Wavelet/SVM | CNN | Bi-LSTM | Parallel Fusion | Serial Fusion |
|---|---|---|---|---|---|
| Accuracy | 92.8% | 91.2% | 92.5% | 93.1% | 98.9% |
Classification accuracy of different methods with signal-to-noise ratio (SNR) from 0 to 20dB.
| SNR Methods | 20 dB | 16 dB | 12 dB | 8 dB | 4 dB | 0 dB |
|---|---|---|---|---|---|---|
| Wavelet/SVM | 85.2% | 84.1% | 83.2% | 81.6% | 79.0% | 77.5% |
| CNN | 86.1% | 84.0% | 82.1% | 78.1% | 73.6% | 62.1% |
| Bi-LSTM | 87.2% | 84.9% | 82.7% | 77.5% | 72.5% | 66.0% |
| Parallel fusion | 89.1% | 85.2% | 84.6% | 80.0% | 75.4% | 67.9% |
| Serial fusion | 98.2% | 95.6% | 94.3% | 91.5% | 86.2% | 78.5% |
Figure 6Comparison of classification accuracy between the deep learning models and the traditional method. (a) Classification accuracy of different methods without noise; (b) Classification accuracy of different methods with SNR from 0 dB to 20 dB.
Figure 7Probability matrix of series fusion model. (a) Probability matrix of series fusion model for 20 dB SNR; (b) Probability matrix of series fusion model for 10 dB SNR; (c) Probability matrix of series fusion model for 0 dB SNR.