On the Gradient Formula for learning Generative Models with Regularized Optimal Transport Costs

The use of optimal transport costs for learning generative models has become popular with Wasserstein Generative Adversarial Networks (WGANs). Training a WGAN requires the computation of the differentiation of the optimal transport cost with respect to the parameters of the generative model. In this...

Full description

Saved in:
Bibliographic Details
Published inTransactions on Machine Learning Research Journal
Main Authors Houdard, Antoine, Leclaire, Arthur, Papadakis, Nicolas, Rabin, Julien
Format Journal Article
LanguageEnglish
Published [Amherst Massachusetts]: OpenReview.net, 2022 14.07.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:The use of optimal transport costs for learning generative models has become popular with Wasserstein Generative Adversarial Networks (WGANs). Training a WGAN requires the computation of the differentiation of the optimal transport cost with respect to the parameters of the generative model. In this work, we provide sufficient conditions for the existence of a gradient formula in two different frameworks: the case of semi-discrete optimal transport (i.e. with a discrete target distribution) and the case of regularized optimal transport (i.e. with an entropic penalty). Both cases are based on the dual formulation of the transport cost, and the gradient formula involves a solution of the dual problem. The learning problem is addressed with an alternate algorithm, whose behavior is examined for the problem of MNIST digits generation. In particular, we analyze the impact of entropic regularization both on visual results and convergence speed.
ISSN:2835-8856