CB-FPN: object detection feature pyramid network based on context information and bidirectional efficient fusion

Feature pyramid network (FPN) is a typical structure in object detection. It can improve the accuracy of detection results by fusing feature information at different resolutions and enhancing the expression ability of different levels of features. Among them, the mismatch between the resolution of f...

Full description

Saved in:
Bibliographic Details
Published inPattern analysis and applications : PAA Vol. 26; no. 3; pp. 1441 - 1452
Main Authors Liu, Zhibo, Cheng, Jian
Format Journal Article
LanguageEnglish
Published London Springer London 01.08.2023
Springer Nature B.V
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Feature pyramid network (FPN) is a typical structure in object detection. It can improve the accuracy of detection results by fusing feature information at different resolutions and enhancing the expression ability of different levels of features. Among them, the mismatch between the resolution of feature information and the receptive field and the limited way of feature fusion hinder the full exchange of feature information. To solve the above problems, this paper designs a new structure called an object detection feature pyramid network based on context information and an efficient bidirectional fusion (CB-FPN): (1) Before feature fusion, this study designs a context enhancement module with cross stage partial network (CSPNet) module (CEM-CSP). By using carefully designed dilated convolutions on high-level features, rich context information and receptive fields are obtained to match appropriate feature information. (2) In feature fusion, this study designed a bidirectional efficient feature pyramid network (BE-FPN) module to fuse features efficiently. After adding these two modified architectures to Faster R-CNN with ResNet-50, the average precision (AP) improves from 37.5 to 39.2 on COCO val-2017 data set. In addition, extensive experiments show the effectiveness of our methods on one-stage, two-stage, and anchor-free models.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:1433-7541
1433-755X
DOI:10.1007/s10044-023-01173-9