Learning Energy-Based Prior Model with Diffusion-Amortized MCMC

Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in the field of generative modeling due to its flexibility in the formulation and strong modeling power of the latent space. However, the common practice of learning latent space EBMs with non-co...

Full description

Saved in:
Bibliographic Details
Main Authors Yu, Peiyu, Zhu, Yaxuan, Xie, Sirui, Ma, Xiaojian, Gao, Ruiqi, Zhu, Song-Chun, Wu, Ying Nian
Format Journal Article
LanguageEnglish
Published 04.10.2023
Subjects
Online AccessGet full text

Cover

Loading…
Abstract Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in the field of generative modeling due to its flexibility in the formulation and strong modeling power of the latent space. However, the common practice of learning latent space EBMs with non-convergent short-run MCMC for prior and posterior sampling is hindering the model from further progress; the degenerate MCMC sampling quality in practice often leads to degraded generation quality and instability in training, especially with highly multi-modal and/or high-dimensional target distributions. To remedy this sampling issue, in this paper we introduce a simple but effective diffusion-based amortization method for long-run MCMC sampling and develop a novel learning algorithm for the latent space EBM based on it. We provide theoretical evidence that the learned amortization of MCMC is a valid long-run MCMC sampler. Experiments on several image modeling benchmark datasets demonstrate the superior performance of our method compared with strong counterparts
AbstractList Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in the field of generative modeling due to its flexibility in the formulation and strong modeling power of the latent space. However, the common practice of learning latent space EBMs with non-convergent short-run MCMC for prior and posterior sampling is hindering the model from further progress; the degenerate MCMC sampling quality in practice often leads to degraded generation quality and instability in training, especially with highly multi-modal and/or high-dimensional target distributions. To remedy this sampling issue, in this paper we introduce a simple but effective diffusion-based amortization method for long-run MCMC sampling and develop a novel learning algorithm for the latent space EBM based on it. We provide theoretical evidence that the learned amortization of MCMC is a valid long-run MCMC sampler. Experiments on several image modeling benchmark datasets demonstrate the superior performance of our method compared with strong counterparts
Author Zhu, Yaxuan
Gao, Ruiqi
Zhu, Song-Chun
Yu, Peiyu
Ma, Xiaojian
Wu, Ying Nian
Xie, Sirui
Author_xml – sequence: 1
  givenname: Peiyu
  surname: Yu
  fullname: Yu, Peiyu
– sequence: 2
  givenname: Yaxuan
  surname: Zhu
  fullname: Zhu, Yaxuan
– sequence: 3
  givenname: Sirui
  surname: Xie
  fullname: Xie, Sirui
– sequence: 4
  givenname: Xiaojian
  surname: Ma
  fullname: Ma, Xiaojian
– sequence: 5
  givenname: Ruiqi
  surname: Gao
  fullname: Gao, Ruiqi
– sequence: 6
  givenname: Song-Chun
  surname: Zhu
  fullname: Zhu, Song-Chun
– sequence: 7
  givenname: Ying Nian
  surname: Wu
  fullname: Wu, Ying Nian
BackLink https://doi.org/10.48550/arXiv.2310.03218$$DView paper in arXiv
BookMark eNotz8tOAjEYBeAudIHIA7CiL1Bs_05vK4MjoslMMJH9pDPTYhNoTQcv-PQiujrJyclJvit0EVN0CE0ZnRdaCHpj81f4mAM_FZQD0yN0WzmbY4hbvIwub4_kzg6ux885pIzr1Lsd_gyHV3wfvH8fQopksU_5EL5Po7qsy2t06e1ucJP_HKOXh-WmfCTVevVULipipdJEKMFaKYVUXSt024IxigEvRA_gBXVQgKaKaUa90Z2VjBYMDNVOGrCm42M0-3s9A5q3HPY2H5tfSHOG8B8PZUIn
ContentType Journal Article
Copyright http://creativecommons.org/licenses/by-sa/4.0
Copyright_xml – notice: http://creativecommons.org/licenses/by-sa/4.0
DBID AKY
EPD
GOX
DOI 10.48550/arxiv.2310.03218
DatabaseName arXiv Computer Science
arXiv Statistics
arXiv.org
DatabaseTitleList
Database_xml – sequence: 1
  dbid: GOX
  name: arXiv.org
  url: http://arxiv.org/find
  sourceTypes: Open Access Repository
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID 2310_03218
GroupedDBID AKY
EPD
GOX
ID FETCH-LOGICAL-a678-5751b66567cb58bb299712345d22f50e2428071810f98ca610412908e692a9c3
IEDL.DBID GOX
IngestDate Mon Jan 08 05:45:24 EST 2024
IsDoiOpenAccess true
IsOpenAccess true
IsPeerReviewed false
IsScholarly false
Language English
LinkModel DirectLink
MergedId FETCHMERGED-LOGICAL-a678-5751b66567cb58bb299712345d22f50e2428071810f98ca610412908e692a9c3
OpenAccessLink https://arxiv.org/abs/2310.03218
ParticipantIDs arxiv_primary_2310_03218
PublicationCentury 2000
PublicationDate 2023-10-04
PublicationDateYYYYMMDD 2023-10-04
PublicationDate_xml – month: 10
  year: 2023
  text: 2023-10-04
  day: 04
PublicationDecade 2020
PublicationYear 2023
Score 1.900823
SecondaryResourceType preprint
Snippet Latent space Energy-Based Models (EBMs), also known as energy-based priors, have drawn growing interests in the field of generative modeling due to its...
SourceID arxiv
SourceType Open Access Repository
SubjectTerms Computer Science - Artificial Intelligence
Computer Science - Learning
Statistics - Machine Learning
Title Learning Energy-Based Prior Model with Diffusion-Amortized MCMC
URI https://arxiv.org/abs/2310.03218
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwdV1LS8QwEB529-RFFJX1SQ5eo2mTNslJ1rrrIlQFFXpbkjaVgi_qroi_3kla0YvXZAhkAvN9M5OZATiWQpWSMUONlQkVUemockyjl8JNzHltpfAFzvl1On8QV0VSDID81MKY9rP56PoD2_dTTz5OGEcYGsIwjv2XrcuboktOhlZcvfyvHHLMsPQHJGYbsN6zOzLpnmMTBu5lC876HqaPZBoK7eg5IkdFbtvmtSV-FtkT8dFQctHU9coHr-jkGUlx84VCeZZn23A3m95nc9rPLaAGTT_1mQybIk-SpU2UtWjwJeKDSKo4rhPmEBQVAruKWK1VaZC_CB8MUi7VsdEl34ERev5uDCTSMqlEJfG0FMHGaGlkba2nYNxYzXZhHC67eOs6Uyy8HhZBD3v_b-3Dmp-ZHn6kiQMYLduVO0RkXdqjoN5vAvB1XQ
link.rule.ids 228,230,783,888
linkProvider Cornell University
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Ajournal&rft.genre=article&rft.atitle=Learning+Energy-Based+Prior+Model+with+Diffusion-Amortized+MCMC&rft.au=Yu%2C+Peiyu&rft.au=Zhu%2C+Yaxuan&rft.au=Xie%2C+Sirui&rft.au=Ma%2C+Xiaojian&rft.date=2023-10-04&rft_id=info:doi/10.48550%2Farxiv.2310.03218&rft.externalDocID=2310_03218