vHeat: Building Vision Models upon Heat Conduction

A fundamental problem in learning robust and expressive visual representations lies in efficiently estimating the spatial relationships of visual semantics throughout the entire image. In this study, we propose vHeat, a novel vision backbone model that simultaneously achieves both high computational...

Full description

Saved in:
Bibliographic Details
Main Authors Wang, Zhaozhi, Liu, Yue, Liu, Yunfan, Yu, Hongtian, Wang, Yaowei, Ye, Qixiang, Tian, Yunjie
Format Journal Article
LanguageEnglish
Published 26.05.2024
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:A fundamental problem in learning robust and expressive visual representations lies in efficiently estimating the spatial relationships of visual semantics throughout the entire image. In this study, we propose vHeat, a novel vision backbone model that simultaneously achieves both high computational efficiency and global receptive field. The essential idea, inspired by the physical principle of heat conduction, is to conceptualize image patches as heat sources and model the calculation of their correlations as the diffusion of thermal energy. This mechanism is incorporated into deep models through the newly proposed module, the Heat Conduction Operator (HCO), which is physically plausible and can be efficiently implemented using DCT and IDCT operations with a complexity of $\mathcal{O}(N^{1.5})$. Extensive experiments demonstrate that vHeat surpasses Vision Transformers (ViTs) across various vision tasks, while also providing higher inference speeds, reduced FLOPs, and lower GPU memory usage for high-resolution images. The code will be released at https://github.com/MzeroMiko/vHeat.
DOI:10.48550/arxiv.2405.16555