| Literature DB >> 35769671 |
Zhou Tao1,2, Chang XiaoYu1, Lu HuiLing3, Ye XinYu1, Liu YunCan1, Zheng XiaoMin4.
Abstract
Deep learning has become a research hotspot in multimedia, especially in the field of image processing. Pooling operation is an important operation in deep learning. Pooling operation can reduce the feature dimension, the number of parameters, the complexity of computation, and the complexity of time. With the development of deep learning models, pooling operation has made great progress. The main contributions of this paper on pooling operation are as follows: firstly, the steps of the pooling operation are summarized as the pooling domain, pooling kernel, step size, activation value, and response value. Secondly, the expression form of pooling operation is standardized. From the perspective of "invariable" to "variable," this paper analyzes the pooling domain and pooling kernel in the pooling operation. Pooling operation can be classified into four categories: invariable of pooling domain, variable of pooling domain, variable of pooling kernel, and the pooling of invariable "+" variable. Finally, the four types of pooling operation are summarized and discussed with their advantages and disadvantages. There is great significance to the research of pooling operations and the iterative updating of deep learning models.Entities:
Mesh:
Year: 2022 PMID: 35769671 PMCID: PMC9236794 DOI: 10.1155/2022/4067581
Source DB: PubMed Journal: Biomed Res Int Impact factor: 3.246
Figure 1The process of pooling operation.
Figure 2“Invariable” to “variable” of pooling operation.
Figure 3Summary of pooling methods.
Figure 4Invariable of pooling domain.
Summary of basic pooling operation.
| Pooling method | Characteristic | Sketch map |
|---|---|---|
| Max pooling | The max pooling operation is simple, but the features with strong influencing factors are ignored and lost information |
|
| Maximum two-mean pooling | Maximum two-mean pooling operation improves the disadvantage of max pooling to a certain extent, which is ignoring features with larger influencing factors. But there is some problem of loss information |
|
|
|
|
|
| Average pooling | Average pooling operation takes global information into account, no information is lost, and overfitting is reduced. However, the features tend to be smooth, and the information of prominent features cannot be extracted |
|
| Median pooling | Median pooling operation can learn the characteristics of edge and texture structure and has a strong antinoise ability |
|
| Sum pooling | Sum pooling operation considers global information. There is no information loss, but it is vulnerable to extreme information | — |
| Overlapping pooling [ | Overlapping pooling operation reduces the characteristic dimension of the output and reduces overfitting, but there are information that will be lost |
|
Summary of improved pooling operation.
| Pooling method | Characteristic | Sketch map |
|---|---|---|
| Matrix2-norm pooling [ | Matrix 2-norm pooling operation takes the energy of the image as the information transmitted to the next layer network to make geometric distortion of the image is highly invariable |
|
| Moment pooling [ | The randomness of moment pooling operation makes each choice different, to prevent over inhibition |
|
| Covariance pooling [ | Covariance pooling operation can capture more information on the feature map |
|
| Dynamic adaptive pooling [ | Dynamic adaptive pooling operation can adaptively adjust the weight of pooling and extract more accurate features |
|
| Nested invariance pooling [ | Nested invariance pooling operation can be extended to any arbitrary transformation set, allowing any arbitrary output feature dimension to be specified |
|
| Spectral pooling [ | Spectral pooling operation has higher efficiency and lower cost | — |
Figure 5Variable of pooling domain.
Figure 6Schematic diagram of strip pooling [21].
Summary of variable of pooling domain.
| Pooling method | Characteristic | Sketch map |
|---|---|---|
| Region of interest pooling [ | ROI pooling operation converts feature maps within a region of interest of any size into feature maps of fixed size |
|
| Fractional max pooling [ | Fractional max pooling operation allows the pooling domain to be noninteger values and reduces overfitting |
|
| Strip pooling [ | Strip pooling operation is easy to establish remote dependencies between discrete distributed regions. It can capture local details and can be easily embedded into any building block |
|
| Chunk-max pooling [ | Chunk-max pooling operation retains the relative order information of multiple local max eigenvalues but does not retain the absolute position information |
|
| Multiscale orderless pooling [ | Multiscale orderless pooling operation improves the invariance of neural network activations without reducing its resolution |
|
Figure 7Variable of pooling kernel.
Figure 8Rank-based pooling [28].
Summary of variable of the pooling kernel.
| Pooling method | Characteristic | Sketch map | |
|---|---|---|---|
| Generalized max pooling [ | Generalized max pooling operation balances the influence of frequent pixels and rare data and improves the ability to extract fine-grained data |
| |
| Parameter pooling [ | Parameter pooling operation converts the correlation operation into an interpretable pooling operation, which retains information and reduces errors |
| |
| LP pooling [ | LP pooling operation balances the effects of max pooling and average pooling. |
| |
| Spatial attention pooling [ | Spatial attention pooling operation mitigates the effects of distracting factors and focuses on meaningful parts of the image |
| |
| Rank-based pooling [ | Rank-based average pooling | Alleviate the problem of loss of information in max pooling and loss of discriminative information in average pooling |
|
| Rank-based weighted pooling | Assigning a weight value to each pixel in the pooling domain can improve performance |
| |
| Rank-based stochastic pooling | Alleviates the problem that random pooling is limited to nonnegative values and reduces overfitting |
| |
| Stochastic pooling [ | Stochastic pooling operation is simple and has a strong generalization ability |
| |
| Spatial pyramid pooling [ | Spatial pyramid pooling operation can handle images of different scales, is very flexible to use, and can effectively prevent overfitting. However, there are also practical constraints to consider |
| |
Figure 9The pooling of invariable “+” variable.
Figure 10Multiscale pooling.
Summary of multichannel pooling.
| Pooling method | Characteristic | Sketch map |
|---|---|---|
| Intermediate value pooling [ | It takes into account the average pooling and the max pooling, so that it has smaller model error and higher stability |
|
| Mixed pooling [ | It solves the problem of which to choose between max pooling and average pooling, but only using one of these pooling methods still has the disadvantage of max pooling or average pooling |
|
| Spatial pyramid and global average hybrid pooling [ | It can extract local and global information, respectively, which learns more information |
|
| Multiscale pooling | It is more flexible and can be used multiple times at the beginning, middle, or end of the network |
|
| Scalable overlapping slide pooling | It considers the saliency of features at different scales and the relationship between adjacent feature elements, so that coarse-grained, medium-grained, and fine-grained multiple features can be extracted |
|