LoopNet for fine-grained fashion attributes editing

Generative Adversarial Networks (GANs) have revolutionized the field of image synthesis by transforming randomly sampled latent codes into high-fidelity synthesized images. However, current methods fall short in manipulating a wide range of fashion attributes due to semantic ambiguity and lack of di...

Full description

Saved in:
Bibliographic Details
Published inExpert systems with applications Vol. 259; p. 125182
Main Authors Zou, Xingxing, Zhu, Shumin, Wong, Wai Keung
Format Journal Article
LanguageEnglish
Published Elsevier Ltd 01.01.2025
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Generative Adversarial Networks (GANs) have revolutionized the field of image synthesis by transforming randomly sampled latent codes into high-fidelity synthesized images. However, current methods fall short in manipulating a wide range of fashion attributes due to semantic ambiguity and lack of disentanglement. This work focuses on fashion attribute editing by proposing an encoder-based GAN inversion method, namely LoopNet. To enable high-fidelity image inversion and fine-grained attribute editing, it refines edited images through two encoder–decoder stages, utilizing predefined directions from principal component analysis on latent codes and canny loss for detail enhancement. Experiments show LoopNet’s effectiveness in attribute disentanglement and manipulation, outperforming seven state-of-the-art image inversion methods. •A novel image inversion framework, LoopNet, designed for fashion attribute editing.•The use of semantic attribute directions to enhance model fidelity and editability.•The introduction of a Canny loss function to refine the details of inverted images.•Comprehensive experimental shows LoopNet’s ability in fashion attributes editing.
ISSN:0957-4174
DOI:10.1016/j.eswa.2024.125182