| Literature DB >> 35486577 |
Xing Cheng Han1,2, Chenxi Ren1,2, Liming Wang1,2, Yunjiao Bai3.
Abstract
To improve the recognition accuracy of underwater acoustic targets by artificial neural network, this study presents a new recognition method that integrates a one-dimensional convolutional neural network and a long short-term memory network. This new network framework is constructed and applied to underwater acoustic target recognition for the first time. Ship acoustic data are used as input to evaluate the network performance. A visual analysis of the recognition results is performed. The results show that this method can realize the recognition and classification of underwater acoustic targets. Compared with a single neural network, the relevant indices, such as the recognition accuracy of the joint network are considerably higher. This provides a new direction for the application of deep learning in the field of underwater acoustic target recognition.Entities:
Mesh:
Year: 2022 PMID: 35486577 PMCID: PMC9053803 DOI: 10.1371/journal.pone.0266425
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Fig 11D-Convolutional neural network model.
Fig 2LSTM block model diagram.
Dataset classification.
| Category | Ship types |
|---|---|
|
| Fishing boats; Trawlers; Mussel boats; Tugboats; Dredgers |
|
| Motorboats; Pilot boats; Sailboats |
|
| Passenger ferries |
|
| Ocean liner; Ro-Ro vessels |
|
| Background noise recordings |
Fig 3Data processing flow chart.
Fig 4t-SNE visualization result of the Mel spectrum, MFCCs feature and fusion feature.
(a) Mel spectrum; (b) MFCCs; (c) fusion feature.
Actual size of the dataset used.
| Category | Acoustic signal serial number | The number of data | Total |
|---|---|---|---|
|
| 13,15,28, 46–49,66,73–76,80, 93–96 | 1040 | 4900 |
|
| 26,27,29,30,33,50–52,56,57,68, 70,72,77,79 | 790 | |
|
| 6,10,40,42,43,52–54,59–65,67 | 1340 | |
|
| 18–20,22,24,25,58,69,71,78 | 1135 | |
|
| 81–92 | 595 |
Fig 5Joint network model.
Network parameter table.
| Layer | Output Shape | Param |
|---|---|---|
|
| 191×64 | 256 |
|
| 63×64 | 0 |
|
| 62×128 | 24704 |
|
| 20×128 | 0 |
|
| 20×128 | 0 |
|
| 32×1 | 20608 |
|
| 32×1 | 0 |
|
| 5×1 | 165 |
Network training parameter.
| Parameters | Parameter Settings |
|---|---|
|
| Categorical_crossentropy |
|
| Adam |
|
| Accuracy |
|
| 64 |
|
| 100 |
|
| ReLU |
|
| Sigmoid |
Fig 6Variation of accuracy.
Fig 7Variation of loss.
Comparison of three kinds of network recognition results.
| Network | Accuracy of training set | Accuracy of test set |
|---|---|---|
|
| 82.27% | 76.10% |
|
| 85.98% | 84.18% |
|
| 96.73% | 92.14% |
Fig 8Confusion matrices for three networks.
(a) LTSM; (b) 1D-CNN; (c) Joint Network.
Various types of recognition.
| Network | Accuracy of test set | ||||
|---|---|---|---|---|---|
| A | B | C | D | E | |
|
| 79.00% | 72.66% | 69.34% | 88.09% | 71.43% |
|
| 77.00% | 70.50% | 91.63% | 91.48% | 98.32% |
|
| 94.50% | 76.26% | 91.99% | 96.60% | 98.32% |
Fig 9Precision about different categories.
Fig 11F1 Score about different categories.
Fig 12Comparison results of the recognition accuracy of the three networks.
(a) Class A; (b) Class B; (c) Class C; (d) Class D; (e) Class E; (f) Overall recognition accuracy.