Learning Spatially-Adaptive Style-Modulation Networks for Single Image Synthesis
Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the...
Saved in:
Published in | 2023 IEEE International Conference on Image Processing (ICIP) pp. 1455 - 1459 |
---|---|
Main Authors | , |
Format | Conference Proceeding |
Language | English |
Published |
IEEE
08.10.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the training image. However, SinGAN cannot learn high level semantics of the image, and thus their synthesized samples tend to have unrealistic spatial layouts. To address this issue, this paper proposes a spatially adaptive style-modulation (SASM) module that learns to preserve realistic spatial configuration of images. Specifically, it extracts style vector (in the form of channel-wise attention) and latent spatial mask (in the form of spatial attention) from a coarse level feature separately. The style vector and spatial mask are then aggregated to modulate features of deeper layers. The disentangled modulation of spatial and style attributes enables the model to preserve the spatial structure of the image without overfitting. Experimental results show that the proposed module learns to generate samples with better fidelity than prior works. |
---|---|
AbstractList | Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications, collecting large dataset is not feasible. Existing work like SinGAN is able to synthesize novel images that resemble the patch distribution of the training image. However, SinGAN cannot learn high level semantics of the image, and thus their synthesized samples tend to have unrealistic spatial layouts. To address this issue, this paper proposes a spatially adaptive style-modulation (SASM) module that learns to preserve realistic spatial configuration of images. Specifically, it extracts style vector (in the form of channel-wise attention) and latent spatial mask (in the form of spatial attention) from a coarse level feature separately. The style vector and spatial mask are then aggregated to modulate features of deeper layers. The disentangled modulation of spatial and style attributes enables the model to preserve the spatial structure of the image without overfitting. Experimental results show that the proposed module learns to generate samples with better fidelity than prior works. |
Author | Wu, Tianfu Shen, Jianghao |
Author_xml | – sequence: 1 givenname: Jianghao surname: Shen fullname: Shen, Jianghao email: jshen27@ncsu.edu organization: North Carolina State University,Department of Electrical and Computer Engineering,USA – sequence: 2 givenname: Tianfu surname: Wu fullname: Wu, Tianfu email: tianfu_wu@ncsu.edu organization: North Carolina State University,Department of Electrical and Computer Engineering,USA |
BookMark | eNo1j8tOwzAURI0EC1r4AyT8Ayn29aPxsop4RApQKbCunPqmWLhOlBhQ_r6RgNUs5syRZkHOYxeRkFvOVpwzc1cW5VYaocwKGIgVZwAgc3FGFnwNOTe5UPqSbCu0Q_TxQOveJm9DmLKNs33y30jrNAXMnjv3Feaui_QF0083fI607QZaz6uAtDzaw4xOMX3g6McrctHaMOL1Xy7J-8P9W_GUVa-PZbGpMg9MpsxIRMZcY3KnlYWmsWKvc-kMNE6BgL0C3joLRiOCE6oBw6wxfJ0brrVkYklufr0eEXf94I92mHb_L8UJfVhNJw |
ContentType | Conference Proceeding |
DBID | 6IE 6IH CBEJK RIE RIO |
DOI | 10.1109/ICIP49359.2023.10222483 |
DatabaseName | IEEE Electronic Library (IEL) Conference Proceedings IEEE Proceedings Order Plan (POP) 1998-present by volume IEEE Xplore All Conference Proceedings IEEE Xplore: IEL IEEE Proceedings Order Plans (POP) 1998-present |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: RIE name: IEEE Electronic Library (IEL) url: https://proxy.k.utb.cz/login?url=https://ieeexplore.ieee.org/ sourceTypes: Publisher |
DeliveryMethod | fulltext_linktorsrc |
EISBN | 1728198356 9781728198354 |
EndPage | 1459 |
ExternalDocumentID | 10222483 |
Genre | orig-research |
GroupedDBID | 6IE 6IH CBEJK RIE RIO |
ID | FETCH-LOGICAL-i204t-94ee00db98d65a2bba3c684d92bd5232c521fda296ee2d35b290a991789166403 |
IEDL.DBID | RIE |
IngestDate | Wed Jan 10 09:27:48 EST 2024 |
IsPeerReviewed | false |
IsScholarly | true |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-i204t-94ee00db98d65a2bba3c684d92bd5232c521fda296ee2d35b290a991789166403 |
PageCount | 5 |
ParticipantIDs | ieee_primary_10222483 |
PublicationCentury | 2000 |
PublicationDate | 2023-Oct.-8 |
PublicationDateYYYYMMDD | 2023-10-08 |
PublicationDate_xml | – month: 10 year: 2023 text: 2023-Oct.-8 day: 08 |
PublicationDecade | 2020 |
PublicationTitle | 2023 IEEE International Conference on Image Processing (ICIP) |
PublicationTitleAbbrev | ICIP |
PublicationYear | 2023 |
Publisher | IEEE |
Publisher_xml | – name: IEEE |
Score | 2.2419465 |
Snippet | Recently there has been a growing interest in learning generative models from a single image. This task is important as in many real world applications,... |
SourceID | ieee |
SourceType | Publisher |
StartPage | 1455 |
SubjectTerms | Aggregates generative learning Image synthesis Layout Modulation Semantics single image spatially-adaptive Three-dimensional displays Training |
Title | Learning Spatially-Adaptive Style-Modulation Networks for Single Image Synthesis |
URI | https://ieeexplore.ieee.org/document/10222483 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8MwGA66kycVJ36Tg9d0bZp-5CjDsQkbgznYbSR938hwdsN1h_nrTdJOURC8hZLQJinvR_I870PIvbV_TvQoYSbhkgnMIqYMAMtMlkUSlBLc8Z2Ho7Q_FU-zZNaQ1T0XBhE9-AwD1_R3-bAqtu6orOOzE5HHh-TQZm41WavBbEWh7Ay6g7FwTNPAaYIH-94_dFO82-gdk9H-hTVa5DXYVjooPn7VYvz3F52Q9jdDj46_fM8pOcDyjIybYqkv1AkN2x9ruWMPoNbOotFJtVsiG66g0euioxoAvqE2bKUTO2qJdPBmzQud7EobFm4WmzaZ9h6fu33WKCawBQ9FxaRADEPQMoc0UVxrFRdpLkByDTbj5IV11gYUlykihzjRXIbKRohZbqPEVITxOWmVqxIvCLVTNRDprDDC6aUXSkMh_aWfibSOxCVpu-WYr-uiGPP9Slz98fyaHLldqeFzN6RVvW_x1vrzSt_5ffwEu3ah3A |
linkProvider | IEEE |
linkToHtml | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwjV1NS8MwGA46D3pSceK3OXhN16TpR44yHKtuZbANdhv5qgy3brjuMH-9SdopCoK3EhqaJuV93jTP8z4APJj4Z02PQpSHhCGqY4x4rhSK8zjGTHFOidU797OoO6bPk3BSi9WdFkZr7chn2rOX7ixfLeXG_iprud0JTYJ9cGCAP8SVXKtmbWGftdJ2OqBWa-pZV3Bvd_8P5xQHHJ1jkO0eWfFF3rxNKTz58asa47_HdAKa3xo9OPhCn1Owp4szMKjLpb5CazVsPq35Fj0qvrIxDQ7L7Vyj_lLVjl0wqyjga2gSVzg0veYapgsTYOBwW5jEcD1bN8G48zRqd1HtmYBmxKclYlRr31eCJSoKORGCBzJKqGJEKLPnJNLAda44YZHWRAWhIMznJkeME5MnRtQPzkGjWBb6AkDzqrnCIpY5tY7pkgslmTv2y7EQmF6Cpp2O6aoqizHdzcTVH-334LA76vemvTR7uQZHdoUqMt0NaJTvG31r0L0Ud25NPwFc36Ul |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2023+IEEE+International+Conference+on+Image+Processing+%28ICIP%29&rft.atitle=Learning+Spatially-Adaptive+Style-Modulation+Networks+for+Single+Image+Synthesis&rft.au=Shen%2C+Jianghao&rft.au=Wu%2C+Tianfu&rft.date=2023-10-08&rft.pub=IEEE&rft.spage=1455&rft.epage=1459&rft_id=info:doi/10.1109%2FICIP49359.2023.10222483&rft.externalDocID=10222483 |