FEANet: Feature-Enhanced Attention Network for RGB-Thermal Real-time Semantic Segmentation

The RGB-Thermal (RGB-T) information for semantic segmentation has been extensively explored in recent years. However, most existing RGB-T semantic segmentation usually compromises spatial resolution to achieve real-time inference speed, which leads to poor performance. To better extract detail spati...

Full description

Saved in:
Bibliographic Details
Published in2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) pp. 4467 - 4473
Main Authors Deng, Fuqin, Feng, Hua, Liang, Mingjian, Wang, Hongmin, Yang, Yong, Gao, Yuan, Chen, Junfeng, Hu, Junjie, Guo, Xiyue, Lam, Tin Lun
Format Conference Proceeding
LanguageEnglish
Published IEEE 27.09.2021
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The RGB-Thermal (RGB-T) information for semantic segmentation has been extensively explored in recent years. However, most existing RGB-T semantic segmentation usually compromises spatial resolution to achieve real-time inference speed, which leads to poor performance. To better extract detail spatial information, we propose a two-stage Feature-Enhanced Attention Network (FEANet) for the RGB-T semantic segmentation task. Specifically, we introduce a Feature-Enhanced Attention Module (FEAM) to excavate and enhance multi-level features from both the channel and spatial views. Benefited from the proposed FEAM module, our FEANet can preserve the spatial information and shift more attention to high-resolution features from the fused RGB-T images. Extensive experiments on the urban scene dataset demonstrate that our FEANet outperforms other state-of-the-art (SOTA) RGB-T methods in terms of objective metrics and subjective visual comparison (+2.6% in global mAcc and +0.8% in global mIoU). For the 480 × 640 RGB-T test images, our FEANet can run with a real-time speed on an NVIDIA GeForce RTX 2080 Ti card.
ISSN:2153-0866
DOI:10.1109/IROS51168.2021.9636084