Learning Energy-Based Prior Model with Diffusion-Amortized MCMC
Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in the field of generative modeling due to its flexibility in the formulation and strong modeling power of the latent space. However, the common practice of learning latent space EBMs with non-co...
Saved in:
Main Authors | , , , , , , |
---|---|
Format | Journal Article |
Language | English |
Published |
04.10.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Abstract | Latent space Energy-Based Models (EBMs), also known as energy-based priors,
have drawn growing interests in the field of generative modeling due to its
flexibility in the formulation and strong modeling power of the latent space.
However, the common practice of learning latent space EBMs with non-convergent
short-run MCMC for prior and posterior sampling is hindering the model from
further progress; the degenerate MCMC sampling quality in practice often leads
to degraded generation quality and instability in training, especially with
highly multi-modal and/or high-dimensional target distributions. To remedy this
sampling issue, in this paper we introduce a simple but effective
diffusion-based amortization method for long-run MCMC sampling and develop a
novel learning algorithm for the latent space EBM based on it. We provide
theoretical evidence that the learned amortization of MCMC is a valid long-run
MCMC sampler. Experiments on several image modeling benchmark datasets
demonstrate the superior performance of our method compared with strong
counterparts |
---|---|
AbstractList | Latent space Energy-Based Models (EBMs), also known as energy-based priors,
have drawn growing interests in the field of generative modeling due to its
flexibility in the formulation and strong modeling power of the latent space.
However, the common practice of learning latent space EBMs with non-convergent
short-run MCMC for prior and posterior sampling is hindering the model from
further progress; the degenerate MCMC sampling quality in practice often leads
to degraded generation quality and instability in training, especially with
highly multi-modal and/or high-dimensional target distributions. To remedy this
sampling issue, in this paper we introduce a simple but effective
diffusion-based amortization method for long-run MCMC sampling and develop a
novel learning algorithm for the latent space EBM based on it. We provide
theoretical evidence that the learned amortization of MCMC is a valid long-run
MCMC sampler. Experiments on several image modeling benchmark datasets
demonstrate the superior performance of our method compared with strong
counterparts |
Author | Zhu, Yaxuan Gao, Ruiqi Zhu, Song-Chun Yu, Peiyu Ma, Xiaojian Wu, Ying Nian Xie, Sirui |
Author_xml | – sequence: 1 givenname: Peiyu surname: Yu fullname: Yu, Peiyu – sequence: 2 givenname: Yaxuan surname: Zhu fullname: Zhu, Yaxuan – sequence: 3 givenname: Sirui surname: Xie fullname: Xie, Sirui – sequence: 4 givenname: Xiaojian surname: Ma fullname: Ma, Xiaojian – sequence: 5 givenname: Ruiqi surname: Gao fullname: Gao, Ruiqi – sequence: 6 givenname: Song-Chun surname: Zhu fullname: Zhu, Song-Chun – sequence: 7 givenname: Ying Nian surname: Wu fullname: Wu, Ying Nian |
BackLink | https://doi.org/10.48550/arXiv.2310.03218$$DView paper in arXiv |
BookMark | eNotz8tOAjEYBeAudIHIA7CiL1Bs_05vK4MjoslMMJH9pDPTYhNoTQcv-PQiujrJyclJvit0EVN0CE0ZnRdaCHpj81f4mAM_FZQD0yN0WzmbY4hbvIwub4_kzg6ux885pIzr1Lsd_gyHV3wfvH8fQopksU_5EL5Po7qsy2t06e1ucJP_HKOXh-WmfCTVevVULipipdJEKMFaKYVUXSt024IxigEvRA_gBXVQgKaKaUa90Z2VjBYMDNVOGrCm42M0-3s9A5q3HPY2H5tfSHOG8B8PZUIn |
ContentType | Journal Article |
Copyright | http://creativecommons.org/licenses/by-sa/4.0 |
Copyright_xml | – notice: http://creativecommons.org/licenses/by-sa/4.0 |
DBID | AKY EPD GOX |
DOI | 10.48550/arxiv.2310.03218 |
DatabaseName | arXiv Computer Science arXiv Statistics arXiv.org |
DatabaseTitleList | |
Database_xml | – sequence: 1 dbid: GOX name: arXiv.org url: http://arxiv.org/find sourceTypes: Open Access Repository |
DeliveryMethod | fulltext_linktorsrc |
ExternalDocumentID | 2310_03218 |
GroupedDBID | AKY EPD GOX |
ID | FETCH-LOGICAL-a678-5751b66567cb58bb299712345d22f50e2428071810f98ca610412908e692a9c3 |
IEDL.DBID | GOX |
IngestDate | Mon Jan 08 05:45:24 EST 2024 |
IsDoiOpenAccess | true |
IsOpenAccess | true |
IsPeerReviewed | false |
IsScholarly | false |
Language | English |
LinkModel | DirectLink |
MergedId | FETCHMERGED-LOGICAL-a678-5751b66567cb58bb299712345d22f50e2428071810f98ca610412908e692a9c3 |
OpenAccessLink | https://arxiv.org/abs/2310.03218 |
ParticipantIDs | arxiv_primary_2310_03218 |
PublicationCentury | 2000 |
PublicationDate | 2023-10-04 |
PublicationDateYYYYMMDD | 2023-10-04 |
PublicationDate_xml | – month: 10 year: 2023 text: 2023-10-04 day: 04 |
PublicationDecade | 2020 |
PublicationYear | 2023 |
Score | 1.900823 |
SecondaryResourceType | preprint |
Snippet | Latent space Energy-Based Models (EBMs), also known as energy-based priors,
have drawn growing interests in the field of generative modeling due to its... |
SourceID | arxiv |
SourceType | Open Access Repository |
SubjectTerms | Computer Science - Artificial Intelligence Computer Science - Learning Statistics - Machine Learning |
Title | Learning Energy-Based Prior Model with Diffusion-Amortized MCMC |
URI | https://arxiv.org/abs/2310.03218 |
hasFullText | 1 |
inHoldings | 1 |
isFullTextHit | |
isPrint | |
link | http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1LS8QwEB529-RFFJX1SQ5eo2mTNslJ1rrrIlQFFXpbkjaVgi_qroi_3kla0YvXZAhkAvN9M5OZATiWQpWSMUONlQkVUemockyjl8JNzHltpfAFzvl1On8QV0VSDID81MKY9rP56PoD2_dTTz5OGEcYGsIwjv2XrcuboktOhlZcvfyvHHLMsPQHJGYbsN6zOzLpnmMTBu5lC876HqaPZBoK7eg5IkdFbtvmtSV-FtkT8dFQctHU9coHr-jkGUlx84VCeZZn23A3m95nc9rPLaAGTT_1mQybIk-SpU2UtWjwJeKDSKo4rhPmEBQVAruKWK1VaZC_CB8MUi7VsdEl34ERev5uDCTSMqlEJfG0FMHGaGlkba2nYNxYzXZhHC67eOs6Uyy8HhZBD3v_b-3Dmp-ZHn6kiQMYLduVO0RkXdqjoN5vAvB1XQ |
link.rule.ids | 228,230,783,888 |
linkProvider | Cornell University |
openUrl | ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Learning+Energy-Based+Prior+Model+with+Diffusion-Amortized+MCMC&rft.au=Yu%2C+Peiyu&rft.au=Zhu%2C+Yaxuan&rft.au=Xie%2C+Sirui&rft.au=Ma%2C+Xiaojian&rft.date=2023-10-04&rft_id=info:doi/10.48550%2Farxiv.2310.03218&rft.externalDocID=2310_03218 |