On the Gradient Formula for learning Generative Models with Regularized Optimal Transport Costs
The use of optimal transport costs for learning generative models has become popular with Wasserstein Generative Adversarial Networks (WGANs). Training a WGAN requires the computation of the differentiation of the optimal transport cost with respect to the parameters of the generative model. In this...
Saved in:
Published in | Transactions on Machine Learning Research Journal |
---|---|
Main Authors | , , , |
Format | Journal Article |
Language | English |
Published |
[Amherst Massachusetts]: OpenReview.net, 2022
14.07.2023
|
Subjects | |
Online Access | Get full text |
Cover
Loading…
Summary: | The use of optimal transport costs for learning generative models has become popular with Wasserstein Generative Adversarial Networks (WGANs). Training a WGAN requires the computation of the differentiation of the optimal transport cost with respect to the parameters of the generative model. In this work, we provide sufficient conditions for the existence of a gradient formula in two different frameworks: the case of semi-discrete optimal transport (i.e. with a discrete target distribution) and the case of regularized optimal transport (i.e. with an entropic penalty). Both cases are based on the dual formulation of the transport cost, and the gradient formula involves a solution of the dual problem. The learning problem is addressed with an alternate algorithm, whose behavior is examined for the problem of MNIST digits generation. In particular, we analyze the impact of entropic regularization both on visual results and convergence speed. |
---|---|
ISSN: | 2835-8856 |