CR-UNIT: Unsupervised Image-to-Image Translation with Content Reconstruction

The goal of unsupervised image-to-image translation (UNIT) is to translate images between two different domains without any paired data. Recent research has shown great progress in the UNIT subject, but it still faces challenges when translating between artificial and natural objects. To overcome th...

Full description

Saved in:
Bibliographic Details
Published in2023 IEEE International Conference on Image Processing (ICIP) pp. 1165 - 1169
Main Authors Shi, Xiao-Wen, Wang, Yuan-Gen
Format Conference Proceeding
LanguageEnglish
Published IEEE 08.10.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The goal of unsupervised image-to-image translation (UNIT) is to translate images between two different domains without any paired data. Recent research has shown great progress in the UNIT subject, but it still faces challenges when translating between artificial and natural objects. To overcome this challenge, we propose a novel unsupervised image-to-image translation with content reconstruction (CR-UNIT), a two-stage and from-coarse-to-fine training framework. Specifically, CR-UNIT builds a corresponding relationship among the content features of different domains on a coarse granularity in the first stage. In the second stage, a new content reconstruction module is constructed to extract the fine-grained content and style features, obtaining more detailed semantic correspondence and better fusion of the content and style features. Furthermore, we design a content reconstruction loss to facilitate the training of our model. Extensive experimental results demonstrate the superiority of the proposed CR-UNIT over the existing methods. Especially for the translation task between the artificial and natural objects, our CR-UNIT achieves outstanding effect in terms of perceptive quality and objective metric. The code is available at https://github.com/GZHU-DVL/CR-UNIT.
DOI:10.1109/ICIP49359.2023.10222213