| Literature DB >> 35917311 |
Liquan Zhao1, Leilei Wang1, Yanfei Jia2, Ying Cui3.
Abstract
To improve accuracy of the MobileNet network, a new lightweight deep neural network is designed based on the MobileNetV2 network. Firstly, it modifies the network depth of MobileNetV2 to balance the image resolution, network width and depth to keep the gradient stable, which reduces the generation of gradient vanishing or gradient exploding. Secondly, it proposes an improved Bottleneck module by introducing channel attention mechanism. It assigns different weights for different channels according to the degree of relevance between the object features and channels. Therefore, the network can extract more effective features from a complex background. In the end, a new usage strategy of the improved Bottleneck is proposed. It uses the improved Bottleneck module in the second, fourth and fifth stages of MobileNetV2, and uses the original Bottleneck module in other states. Compared with MobileNetV2, MobileNetV3, ShuffleNetV2, GhostNet and HBONetmethods, the proposed method has the highest classification accuracy on the ImageNet-1K dataset, CIFAR-10 and CIFAR-100. Compared with YOLOV4-Lite methods based on these lightweight network networks, YOLOV4-Lite based on our proposed network also has the highest detection accuracy on the PASCAL VOC07+12 dataset.Entities:
Mesh:
Year: 2022 PMID: 35917311 PMCID: PMC9345353 DOI: 10.1371/journal.pone.0271225
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.752
Results on IMAGENET-100.
|
| TOP-1(%) | Parameters | FLOPs |
|---|---|---|---|
| 77.86 | 3.50M | 314.19M | |
| 78.52(+0.66) | 3.18M | 298.51M | |
| 78.21(+0.35) | 2.86M | 282.83M |
Results on CIFAR-100.
|
| TOP-1(%) | Parameters | FLOPs |
|---|---|---|---|
| 77.79 | 3.50M | 314.19M | |
| 78.31(+0.52) | 3.18M | 298.51M | |
| 77.51(-0.28) | 2.86M | 282.83M |
Results on CIFAR-10.
|
| TOP-1(%) | Parameters | FLOPs |
|---|---|---|---|
| 92.34 | 3.50M | 314.19M | |
| 92.89(+0.55) | 3.18M | 298.51M | |
| 91.89(-0.36) | 2.86M | 282.83M |
Fig 1The original Bottleneck module and the improved Bottleneck module.
A: original Bottleneck module. B:depthwise convolution. C: improved Bottleneck module.
Fig 2The detail network structure of improved Bottleneck module with the attention mechanism.
The network structure of improved MobileNetV2.
|
| In_C | Operator | n | t | out_c | s |
|---|---|---|---|---|---|---|
| 224×224 | 3 | conv2d | 1 | − | 32 | 2 |
| 112×112 | 32 | Bottleneck | 1 | 1 | 16 | 1 |
| 112×112 | 16 | Bottleneck | 2 | 6 | 24 | 2 |
| 56×56 | 24 |
| 3 | 6 | 32 | 2 |
| 28×28 | 32 | Bottleneck | 4 | 6 | 64 | 2 |
| 14×14 | 64 |
|
| 6 | 96 | 1 |
| 14×14 | 96 |
| 2 | 6 | 160 | 2 |
| 7×7 | 160 |
| 1 | 6 | 320 | 1 |
| 7×7 | 320 | Conv2d1×1 | 1 | − | 1280 | 1 |
| 7×7 | 1280 | Avgpool 7×7 | 1 | − | − | − |
| 1×1 | 1280 | Conv2d1×1 | 1 | − | k | − |
The network structure of improved MobileNetV2.
|
| In_C | Operator | n | t | out_c | s |
|---|---|---|---|---|---|---|
| 224×224 | 3 | conv2d | 1 | − | 32 | 2 |
| 112×112 | 32 | Bottleneck | 1 | 1 | 16 | 1 |
| 112×112 | 16 | Bottleneck | 2 | 6 | 24 | 2 |
| 56×56 | 24 | Bottleneck | 3 | 6 | 32 | 2 |
| 28×28 | 32 | Bottleneck | 4 | 6 | 64 | 2 |
| 14×14 | 64 | Bottleneck | 3 | 6 | 96 | 1 |
| 14×14 | 96 | Bottleneck | 3 | 6 | 160 | 2 |
| 7×7 | 160 | Bottleneck | 1 | 6 | 320 | 1 |
| 7×7 | 320 | Conv2d1×1 | 1 | − | 1280 | 1 |
| 7×7 | 1280 | Avgpool 7×7 | 1 | − | − | − |
| 1×1 | 1280 | Conv2d1×1 | 1 | − | k | − |
Results for different position of attention module in bottleneck.
|
| TOP-1(%) | Parameters | FLOPs |
|---|---|---|---|
| 77.86 | 3.50M | 314.19M | |
| 78.52(+0.66) | 3.18M | 298.51M | |
| MobileNetV2(5)_E | 79.34(+1.48) | 3.28M | 312.25M |
| MobileNetV2(5)_M | 80.01(+2.15) | 3.67M | 323.70M |
Results for different position of improved bottleneck in network.
| Stage 1 | Stage 2 | Stage 3 | Stage 4 | Stage 5 | Top-1 | |
|---|---|---|---|---|---|---|
| MO | Original Bottleneck | Original Bottleneck | Original Bottleneck | Original Bottleneck | Original Bottleneck | 78.52 |
| M1 | Improved Bottleneck | Improved Bottleneck | Improved Bottleneck | Improved Bottleneck | Improved Bottleneck | 79.73 |
| M2 | Original Bottleneck | Improved Bottleneck | Original Bottleneck | Improved Bottleneck | Improved Bottleneck | 80.01 |
TOP-1 accuracies on on ImageNet-1K dataset.
| Lightweight network | TOP-1(%) | Parameters | FLOPs |
|---|---|---|---|
| ShuffleNetV2 | 71.66 | 3.51 M | 304.51 M |
| GhostNet | 72.22 | 5.18 M | 148.79 M |
| HBONet | 73.53 | 4.56 M | 316.07 M |
| MobileNetV3 | 73.27 | 5.48 M | 229.65 M |
| MobileNetV2(baseline) | 72.07 | 3.50 M | 314.19 M |
| Our proposed network | 75.04 | 3.67 M | 323.70 M |
TOP-1 accuracies on CIFAR100 and CIFAR10 datasets.
| Lightweight network | CIFAR-10 | CIFAR-100 |
|---|---|---|
| ShuffleNetV2 | 91.49 | 77.16 |
| GhostNet | 93.38 | 78.21 |
| HBONet | 92.96 | 77.94 |
| MobileNetV3 | 91.82 | 77.39 |
| MobileNetV2(baseline) | 92.34 | 77.79 |
| Our proposed network | 94.71 | 79.97 |
Fig 3Object detection based on different lightweight networks.
Results for different lightweight networks on object detection.
| Lightweight network | mAP(%) | average detection time/image(ms) |
|---|---|---|
| YOLOV4-Lite+ShuffleNetV2 | 61.94 | 44.0 |
| YOLOV4-Lite +GhostNet | 64.64 | 65.8 |
| YOLOV4 +HBONet | 63.19 | 50.3 |
| YOLOV4-Lite + MobileNetV3 | 64.86 | 46.2 |
| YOLOV4-Lite + MobileNetV2 | 64.39 | 47.4 |
| YOLOV4-Lite + our proposed network | 67.28 | 49.5 |