Learning Spatially-Adaptive Style-Modulation Networks for Single Image Synthesis

Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the...

Full description

Saved in:
Bibliographic Details
Published in2023 IEEE International Conference on Image Processing (ICIP) pp. 1455 - 1459
Main Authors Shen, Jianghao, Wu, Tianfu
Format Conference Proceeding
LanguageEnglish
Published IEEE 08.10.2023
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the training image. However, SinGAN cannot learn high level semantics of the image, and thus their synthesized samples tend to have unrealistic spatial layouts. To address this issue, this paper proposes a spatially adaptive style-modulation (SASM) module that learns to preserve realistic spatial configuration of images. Specifically, it extracts style vector (in the form of channel-wise attention) and latent spatial mask (in the form of spatial attention) from a coarse level feature separately. The style vector and spatial mask are then aggregated to modulate features of deeper layers. The disentangled modulation of spatial and style attributes enables the model to preserve the spatial structure of the image without overfitting. Experimental results show that the proposed module learns to generate samples with better fidelity than prior works.
AbstractList Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the training image. However, SinGAN cannot learn high level semantics of the image, and thus their synthesized samples tend to have unrealistic spatial layouts. To address this issue, this paper proposes a spatially adaptive style-modulation (SASM) module that learns to preserve realistic spatial configuration of images. Specifically, it extracts style vector (in the form of channel-wise attention) and latent spatial mask (in the form of spatial attention) from a coarse level feature separately. The style vector and spatial mask are then aggregated to modulate features of deeper layers. The disentangled modulation of spatial and style attributes enables the model to preserve the spatial structure of the image without overfitting. Experimental results show that the proposed module learns to generate samples with better fidelity than prior works.
Author Wu, Tianfu
Shen, Jianghao
Author_xml – sequence: 1
  givenname: Jianghao
  surname: Shen
  fullname: Shen, Jianghao
  email: jshen27@ncsu.edu
  organization: North Carolina State University,Department of Electrical and Computer Engineering,USA
– sequence: 2
  givenname: Tianfu
  surname: Wu
  fullname: Wu, Tianfu
  email: tianfu_wu@ncsu.edu
  organization: North Carolina State University,Department of Electrical and Computer Engineering,USA
BookMark eNo1j8tOwzAURI0EC1r4AyT8Ayn29aPxsop4RApQKbCunPqmWLhOlBhQ_r6RgNUs5syRZkHOYxeRkFvOVpwzc1cW5VYaocwKGIgVZwAgc3FGFnwNOTe5UPqSbCu0Q_TxQOveJm9DmLKNs33y30jrNAXMnjv3Feaui_QF0083fI607QZaz6uAtDzaw4xOMX3g6McrctHaMOL1Xy7J-8P9W_GUVa-PZbGpMg9MpsxIRMZcY3KnlYWmsWKvc-kMNE6BgL0C3joLRiOCE6oBw6wxfJ0brrVkYklufr0eEXf94I92mHb_L8UJfVhNJw
ContentType Conference Proceeding
DBID 6IE
6IH
CBEJK
RIE
RIO
DOI 10.1109/ICIP49359.2023.10222483
DatabaseName IEEE Electronic Library (IEL) Conference Proceedings
IEEE Proceedings Order Plan (POP) 1998-present by volume
IEEE Xplore All Conference Proceedings
IEEE Xplore: IEL
IEEE Proceedings Order Plans (POP) 1998-present
DatabaseTitleList
Database_xml – sequence: 1
  dbid: RIE
  name: IEEE Electronic Library (IEL)
  url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/
  sourceTypes: Publisher
DeliveryMethod fulltext_linktorsrc
EISBN 1728198356
9781728198354
EndPage 1459
ExternalDocumentID 10222483
Genre orig-research
GroupedDBID 6IE
6IH
CBEJK
RIE
RIO
ID FETCH-LOGICAL-i204t-94ee00db98d65a2bba3c684d92bd5232c521fda296ee2d35b290a991789166403
IEDL.DBID RIE
IngestDate Wed Jan 10 09:27:48 EST 2024
IsPeerReviewed false
IsScholarly true
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-i204t-94ee00db98d65a2bba3c684d92bd5232c521fda296ee2d35b290a991789166403
PageCount 5
ParticipantIDs ieee_primary_10222483
PublicationCentury 2000
PublicationDate 2023-Oct.-8
PublicationDateYYYYMMDD 2023-10-08
PublicationDate_xml – month: 10
  year: 2023
  text: 2023-Oct.-8
  day: 08
PublicationDecade 2020
PublicationTitle 2023 IEEE International Conference on Image Processing (ICIP)
PublicationTitleAbbrev ICIP
PublicationYear 2023
Publisher IEEE
Publisher_xml – name: IEEE
Score 2.2419465
Snippet Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications,...
SourceID ieee
SourceType Publisher
StartPage 1455
SubjectTerms Aggregates
generative learning
Image synthesis
Layout
Modulation
Semantics
single image
spatially-adaptive
Three-dimensional displays
Training
Title Learning Spatially-Adaptive Style-Modulation Networks for Single Image Synthesis
URI https://ieeexplore.ieee.org/document/10222483
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8MwGA66kycVJ36Tg9d0bZp-5CjDsQkbgznYbSR938hwdsN1h_nrTdJOURC8hZLQJinvR_I870PIvbV_TvQoYSbhkgnMIqYMAMtMlkUSlBLc8Z2Ho7Q_FU-zZNaQ1T0XBhE9-AwD1_R3-bAqtu6orOOzE5HHh-TQZm41WavBbEWh7Ay6g7FwTNPAaYIH-94_dFO82-gdk9H-hTVa5DXYVjooPn7VYvz3F52Q9jdDj46_fM8pOcDyjIybYqkv1AkN2x9ruWMPoNbOotFJtVsiG66g0euioxoAvqE2bKUTO2qJdPBmzQud7EobFm4WmzaZ9h6fu33WKCawBQ9FxaRADEPQMoc0UVxrFRdpLkByDTbj5IV11gYUlykihzjRXIbKRohZbqPEVITxOWmVqxIvCLVTNRDprDDC6aUXSkMh_aWfibSOxCVpu-WYr-uiGPP9Slz98fyaHLldqeFzN6RVvW_x1vrzSt_5ffwEu3ah3A
linkProvider IEEE
linkToHtml http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8MwGA46D3pSceK3OXhN16TpR44yHKtuZbANdhv5qgy3brjuMH-9SdopCoK3EhqaJuV93jTP8z4APJj4Z02PQpSHhCGqY4x4rhSK8zjGTHFOidU797OoO6bPk3BSi9WdFkZr7chn2rOX7ixfLeXG_iprud0JTYJ9cGCAP8SVXKtmbWGftdJ2OqBWa-pZV3Bvd_8P5xQHHJ1jkO0eWfFF3rxNKTz58asa47_HdAKa3xo9OPhCn1Owp4szMKjLpb5CazVsPq35Fj0qvrIxDQ7L7Vyj_lLVjl0wqyjga2gSVzg0veYapgsTYOBwW5jEcD1bN8G48zRqd1HtmYBmxKclYlRr31eCJSoKORGCBzJKqGJEKLPnJNLAda44YZHWRAWhIMznJkeME5MnRtQPzkGjWBb6AkDzqrnCIpY5tY7pkgslmTv2y7EQmF6Cpp2O6aoqizHdzcTVH-334LA76vemvTR7uQZHdoUqMt0NaJTvG31r0L0Ud25NPwFc36Ul
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2023+IEEE+International+Conference+on+Image+Processing+%28ICIP%29&rft.atitle=Learning+Spatially-Adaptive+Style-Modulation+Networks+for+Single+Image+Synthesis&rft.au=Shen%2C+Jianghao&rft.au=Wu%2C+Tianfu&rft.date=2023-10-08&rft.pub=IEEE&rft.spage=1455&rft.epage=1459&rft_id=info:doi/10.1109%2FICIP49359.2023.10222483&rft.externalDocID=10222483