Weakly supervised classification model for zero-shot semantic segmentation

As one of the most fundamental tasks in computer vision, semantic segmentation assigns per pixel prediction of object categories. Training a robust model for semantic segmentation is challenging since pixel-level annotations are expensive to obtain. To alleviate the burden of annotations, the author...

Full description

Saved in:
Bibliographic Details
Published inElectronics letters Vol. 56; no. 23; pp. 1247 - 1250
Main Authors Shen, Fengli, Wang, Zong-Hui, Lu, Zhe-Ming
Format Journal Article
LanguageEnglish
Published The Institution of Engineering and Technology 12.11.2020
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:As one of the most fundamental tasks in computer vision, semantic segmentation assigns per pixel prediction of object categories. Training a robust model for semantic segmentation is challenging since pixel-level annotations are expensive to obtain. To alleviate the burden of annotations, the authors propose a weakly-supervised framework for zero-shot semantic segmentation, which can segment images having target classes without any pixel-level labelled instances. Under the assumption that the accessibility to image-level annotations of target classes does not violate the principle of zero pixel-level label in zero-shot semantic segmentation, we utilised image-level annotations to improve the proposed model's ability to extract pixel-level features. Furthermore, unlike existing zero-shot semantic segmentation methods, which use semantic embeddings as class embeddings to transfer knowledge from source classes to target classes, we use image-level features as their class embeddings to transfer knowledge since the distribution of pixel-level features is more similar to the distribution of image-level features rather than the distribution of semantic embeddings. Experimental results on the PASCAL-VOC data set under different data splits demonstrate that the proposed model achieves promising results.
ISSN:0013-5194
1350-911X
1350-911X
DOI:10.1049/el.2020.2270