Pooling Revisited: Your Receptive Field is Suboptimal
The size and shape of the receptive field determine how the network aggregates local information and affect the overall performance of a model considerably. Many components in a neural network, such as kernel sizes and strides for convolution and pooling operations, influence the configuration of a...
Saved in:
Main Authors | , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
30.05.2022
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The size and shape of the receptive field determine how the network
aggregates local information and affect the overall performance of a model
considerably. Many components in a neural network, such as kernel sizes and
strides for convolution and pooling operations, influence the configuration of
a receptive field. However, they still rely on hyperparameters, and the
receptive fields of existing models result in suboptimal shapes and sizes.
Hence, we propose a simple yet effective Dynamically Optimized Pooling
operation, referred to as DynOPool, which optimizes the scale factors of
feature maps end-to-end by learning the desirable size and shape of its
receptive field in each layer. Any kind of resizing modules in a deep neural
network can be replaced by the operations with DynOPool at a minimal cost.
Also, DynOPool controls the complexity of a model by introducing an additional
loss term that constrains computational cost. Our experiments show that the
models equipped with the proposed learnable resizing module outperform the
baseline networks on multiple datasets in image classification and semantic
segmentation. |
---|---|
DOI: | 10.48550/arxiv.2205.15254 |