IPE Transformer for Depth Completion with Input-Aware Positional Embeddings

In contrast to traditional transformer blocks using a set of pre-defined parameters as positional embeddings, we propose the input-aware positional embedding (IPE) which is dynamically generated according to the input feature. We implement this idea by designing the IPE transformer, which enjoys str...

Full description

Saved in:
Bibliographic Details
Published inPattern Recognition and Computer Vision Vol. 13022; pp. 263 - 275
Main Authors Li, Bocen, Li, Guozhen, Wang, Haiting, Wang, Lijun, Gong, Zhenfei, Zhang, Xiaohua, Lu, Huchuan
Format Book Chapter
LanguageEnglish
Published Switzerland Springer International Publishing AG 2021
Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:In contrast to traditional transformer blocks using a set of pre-defined parameters as positional embeddings, we propose the input-aware positional embedding (IPE) which is dynamically generated according to the input feature. We implement this idea by designing the IPE transformer, which enjoys stronger generalization powers across arbitrary input sizes. To verify its effectiveness, we integrate the newly-designed transformer into NLSPN and GuideNet, two remarkable depth completion networks. The experimental result on a large scale outdoor depth completion dataset shows that the proposed transformer can effectively model long-range dependency with a manageable memory overhead.
ISBN:3030880125
9783030880125
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-030-88013-2_22