| Literature DB >> 32276392 |
Liyan Luo1,2, Hongming Qin1,2, Xiyu Song1,2, Mei Wang1,2,3, Hongbing Qiu1,2, Zou Zhou1,2.
Abstract
Nowadays, urban noise emerges as a distinct threat to people's physiological and psychological health. Previous works mainly focus on the measurement and mapping of the noise by using Wireless Acoustic Sensor Networks (WASNs) and further propose some methods that can effectively reduce the noise pollution in urban environments. In addition, the research on the combination of environmental noise measurement and acoustic events recognition are rapidly progressing. In a real-life application, there still exists the challenges on the hardware design with enough computational capacity, the reduction of data amount with a reasonable method, the acoustic recognition with CNNs, and the deployment for the long-term outdoor monitoring. In this paper, we develop a novel system that utilizes the WASNs to monitor the urban noise and recognize acoustic events with a high performance. Specifically, the proposed system mainly includes the following three stages: (1) We used multiple sensor nodes that are equipped with various hardware devices and performed with assorted signal processing methods to capture noise levels and audio data; (2) the Convolutional Neural Networks (CNNs) take such captured data as inputs and classify them into different labels such as car horn, shout, crash, explosion; (3) we design a monitoring platform to visualize noise maps, acoustic event information, and noise statistics. Most importantly, we consider how to design effective sensor nodes in terms of cost, data transmission, and outdoor deployment. Experimental results demonstrate that the proposed system can measure the urban noise and recognize acoustic events with a high performance in real-life scenarios.Entities:
Keywords: CNNs; WASNs; acoustic events recognition; noise measurement; real-life scenarios
Mesh:
Year: 2020 PMID: 32276392 PMCID: PMC7180790 DOI: 10.3390/s20072093
Source DB: PubMed Journal: Sensors (Basel) ISSN: 1424-8220 Impact factor: 3.576
Figure 1The overview of the proposed system.
The cost of each sensor node.
| Description | Material Cost |
|---|---|
| Acoustic sensor | $12.8 |
| Embedded processing board | $34.3 |
| Wireless transmission module | $28.5 |
| Solar panel | $10.9 |
Figure 2The block diagram of the proposed system.
Data information in MySQL (My Structured Query Language).
| Num | Time | Coords (latitude, longitude) | IMEI | dB |
|---|---|---|---|---|
| 158 | 20190702143712 | 25.289270, 110.343036 | 0005 | 56 |
| 159 | 20190702143714 | 25.288436, 110.343405 | 0003 | 49 |
| 160 | 20190702143715 | 25.289270, 110.342036 | 0004 | 50 |
Figure 3The process of feature extraction.
Figure 4The architecture of deep convolutional neural networks (CNNs).
Figure 5The structure of a sensor node. (a) External appearance of the waterproof box; (b) interior view of the waterproof box; (c) the solar panel with a battery and power controller inside; (d) the connection of different parts inside the box; (e) a whole sensor node.
Figure 6Data obtained from the sound level meter and the sensor nodes in the field test. (a) In the school road (low Signal to Noise Ratio (SNR)); (b) in the school gate (high SNR).
Figure 7The spectrogram of the sound. (a) The spectrogram of the sound we recorded by the sensor in a box; (b) the spectrum of the sound we recorded by the sensor towards the outside of the box.
Figure 8The validity of endpoint detection used in the system.
Figure 9The influence of compression with the adaptive differential pulse code modulation (ADPCM). (a) The waveform of car horn before compression; (b) the waveform of car horn after compression; (c) the spectrograms of car horn before compression; (d) the spectrograms of car horn after compression.
The number of data received on the server.
| Node Number | The Number of Data Received on the Server | Missing Rate |
|---|---|---|
| 1 | 2997 | 0.10% |
| 2 | 2995 | 0.16% |
| 3 | 2998 | 0.07% |
| 4 | 2997 | 0.10% |
The accuracy of three kinds of sounds with four classification methods.
| Methods | Gunshots | Screams | Car Horns |
|---|---|---|---|
| RF | 83% | 78.5% | 77% |
| KNN | 79% | 56.3% | 86.2% |
| SVM | 80.7% | 70.7% | 73.1% |
| CNN | 97.3% | 88.5% | 94% |
Time-consuming of feature extraction and recognition.
| Samples | Samples | Time-Consuming | Time-Consuming | Time-Consuming |
|---|---|---|---|---|
| 10,000 | 139,120,15 | 182,100,1 | 2340 | 18,233,41 |
Figure 10Example of the place where the sensor nodes are being deployed.
The number of samples in our dataset.
| Species | Car Horn | Background |
|---|---|---|
| Train | 1160 | 1052 |
| Test | 523 | 474 |
| C-V | 59 | 53 |
| Total | 1742 | 1579 |
Confusion matrix of the CNN model before compression.
| Car Horn | Background | |
|---|---|---|
|
| 491 | 32 |
|
| 10 | 464 |
Confusion matrix of the CNN model after compression.
| Car Horn | Background | |
|---|---|---|
|
| 480 | 43 |
|
| 9 | 465 |
Recognition result with audio steaming.
| Metric | Result |
|---|---|
| TP | 71 |
| FP | 16 |
| FN | 3 |
| Recall | 95.9% |
| Precision | 81.6% |
| F1-score | 88.2% |
Figure 11The monitoring platform.
Figure 12The presentation of acoustic event information.
Figure 13The function of data query.