| Literature DB >> 34764617 |
Liang Li1, Weidong Jin1,2, Yingkun Huang1.
Abstract
This paper presents a novel discriminative Few-shot learning architecture based on batch compact loss. Currently, Convolutional Neural Network (CNN) has achieved reasonably good performance in image recognition. Most existing CNN methods facilitate classifiers to learn discriminating patterns to identify existing categories trained with large samples. However, learning to recognize novel categories from a few examples is a challenging task. To address this, we propose the Residual Compact Network to train a deep neural network to learn hierarchical nonlinear transformations to project image pairs into the same latent feature space, under which the distance of each positive pair is reduced. To better use the commonality of class-level features for category recognition, we develop a batch compact loss to form robust feature representations relevant to a category. The proposed methods are evaluated on several datasets. Experimental evaluations show that our proposed method achieves acceptable results in Few-shot learning.Entities:
Keywords: Batch compact loss; Few-shot contrastive learning; Insulator identification; Partial residual embedding module
Year: 2021 PMID: 34764617 PMCID: PMC8412402 DOI: 10.1007/s10489-021-02769-6
Source DB: PubMed Journal: Appl Intell (Dordr) ISSN: 0924-669X Impact factor: 5.019
Fig. 1Residual Compact Network
Fig. 2Partial residual embedding module
Specifications of the partial residual embedding module
| Layers | Layer name | type | Depth | Stride | Padding |
|---|---|---|---|---|---|
| 1 | ConvBlock1 | 3 × 3Conv+BN+ReLU | 64 | 1 | 1 |
| 2 | ConvBlock2 | 3 × 3Conv+BN | 64 | 1 | 1 |
| 3 × 3Conv+BN | 64 | 1 | 1 | ||
| 3 | ConvBlock3 | 3 × 3Conv+BN | 128 | 2 | 1 |
| 3 × 3Conv+BN | 128 | 1 | 1 | ||
| ShortBlock1 | 1 × 1Conv+BN | 128 | 2 | 0 | |
| 4 | ConvBlock4 | 3 × 3Conv+BN | 256 | 2 | 1 |
| 3 × 3Conv+BN | 256 | 1 | 1 | ||
| ShortBlock2 | 1 × 1Conv+BN | 256 | 2 | 0 | |
| 5 | Feature Vectors | Pooling |
Fig. 3Batch compact loss
Fig. 4Comparison of different loss functions
The first stage training algorithm
The second stage training algorithm
Fig. 5States of different types of insulators. a Normal. b umbrella petticoats damaged. c foreign body. d polluted. e broken
Fig. 6Accuracy of different residual modules
Specifications of the partial residual embedding module
| Name | Framework | Blocks |
|---|---|---|
| S1 | Conv64|Conv64|Conv64|Conv64 | 4 |
| S2 | 2 ×Conv64|2 ×Conv64|2 ×Conv64|2 ×Conv64 | 4 |
| S3 | Conv64|Conv128|Conv256|Conv512 | 4 |
| S4 | 2 ×Conv64|2 ×Conv128|2 ×Conv256|2 ×Conv512 | 4 |
| S5 | Conv64|Conv64|Conv64 | 3 |
| S6 | 2 ×Conv64|2 ×Conv64|2 ×Conv64 | 3 |
| S7 | Conv64|Conv128|Conv256 | 3 |
| S8 | 2 ×Conv64|2 ×Conv128|2 ×Conv256 | 3 |
The accuracy on Omniglot dataset
| Model | FT | 5-way Acc | 20-way Acc | ||
|---|---|---|---|---|---|
| 1-shot | 5-shot | 1-shot | 5-shot | ||
| MANN [ | N | 82.8% | 94.9% | - | - |
| MATCHING NETS [ | N | 98.1% | 98.9% | 93.8% | 98.5% |
| MATCHING NETS [ | Y | 97.9% | 98.7% | 93.5% | 98.7% |
| CNAPS [ | Y | 97.4 ± 0.3% | 99.4 ± 0.1% | 95.3 ± 0.2% | 98.4 ± 0.1% |
| SIAMESE NETS WITH MEMORY [ | N | 98.4% | 99.6% | 95.0% | 98.6% |
| NEURAL STATISTICIAN [ | N | 98.1% | 99.5% | 93.2% | 98.1% |
| META NETS [ | N | 99.0% | - | 97.0% | - |
| PROTOTYPICAL NETS [ | N | 98.8% | 99.7% | 96.0% | 98.9% |
| MAML [ | Y | 98.7 ± 0.4% | 95.8 ± 0.3% | 98.9 ± 0.2% | |
| RELATION NET [ | N | 99.6 ± 0.2% | 99.8 ± 0.1% | 97.6 ± 0.2% | 99.1 ± 0.1% |
| RCNet | N | 99.6 ± 0.2% | 99.7 ± 0.1% | 97.7 ± 0.2% | 99.1 ± 0.1% |
| RCNet with batch compact loss | N | 99.8 ± 0.1% | |||
The accuracy on miniImagenet dataset
| Model | FT | 5-way Acc | |
|---|---|---|---|
| 1-shot | 5-shot | ||
| MATCHING NETS [ | N | 43.56 ± 0.84% | 55.31 ± 0.73% |
| META-LEARN LSTM [ | N | 43.44 ± 0.77% | 60.60 ± 0.71% |
| MAML [ | Y | 48.70 ± 1.84% | 63.11 ± 0.92% |
| META NETS [ | N | 49.21 ± 0.96% | - |
| PROTOTYPICAL NETS [ | N | 49.42 ± 0.78% | 68.20 ± 0.66% |
| RELATION NET [ | N | 50.44 ± 0.82% | 65.32 ± 0.70% |
| CovaMNet [ | N | 51.19 ± 0.76% | 67.65 ± 0.63% |
| RCNet | N | 51.33 ± 0.86% | 67.69 ± 0.79% |
| RCNet with batch compact loss | N | ||
The accuracy on tieredImagenet dataset
| Model | FT | 5-way Acc | |
|---|---|---|---|
| 1-shot | 5-shot | ||
| MAML [ | Y | 51.67 ± 1.81% | 70.30 ± 1.75% |
| SSL [ | N | 52.39 ± 0.44% | 69.88 ± 0.22% |
| PROTOTYPICAL NETS [ | N | 53.31 ± 0.89% | 72.69 ± 0.74% |
| CovaMNet [ | N | 54.07 ± 0.91% | 70.34 ± 0.75% |
| RELATION NET [ | N | 54.48 ± 0.93% | 71.31 ± 0.78% |
| RCNet | N | 56.92 ± 0.97% | 73.41 ± 0.80% |
| RCNet with batch compact loss | N | ||
The accuracy on OCS insulator sub-dataset
| Model | FT | 5-way Acc | |
|---|---|---|---|
| 1-shot | 5-shot | ||
| MAML [ | N | 81.44 ± 1.24% | 88.93 ± 0.68% |
| MATCHING NETS [ | N | 82.12 ± 0.62% | 86.79 ± 0.23% |
| META-LEARN LSTM [ | N | 84.16 ± 0.19% | 87.80 ± 0.13% |
| CovaMNet [ | N | 84.93 ± 0.91% | 87.81 ± 0.33% |
| PROTOTYPICAL NETS [ | N | 83.77 ± 0.14% | 86.69 ± 0.11% |
| RELATION NET [ | N | 84.31 ± 0.43% | 89.11 ± 0.28% |
| RCNet | N | 85.21 ± 0.65% | 89.59 ± 0.36% |
| RCNet with batch compact loss | N | ||