资源论文Generalization and Equilibrium in Generative Adversarial Nets (GANs)

Generalization and Equilibrium in Generative Adversarial Nets (GANs)

2020-03-09 | |  74 |   31 |   0

Abstract

It is shown that training of generative adversarial network (GAN) may not have good generalization properties; e.g., training may appear successful but the trained distribution may be far from target distribution in standard metrics. However, generalization does occur for a weaker metric called neural net distance. It is also shown that an approximate pure equilibrium exists in the discriminator/generator game for a natural training objective (Wasserstein) when generator capacity and training set sizes are moderate. This existence of equilibrium inspires MIX + GAN protocol, which can be combined with any existing GAN training, and empirically shown to improve some of them.

上一篇:Gradient Coding: Avoiding Stragglers in Distributed Learning

下一篇:Distributed Batch Gaussian Process Optimization

用户评价
全部评价

热门资源

  • Learning to Predi...

    Much of model-based reinforcement learning invo...

  • Stratified Strate...

    In this paper we introduce Stratified Strategy ...

  • The Variational S...

    Unlike traditional images which do not offer in...

  • A Mathematical Mo...

    Direct democracy, where each voter casts one vo...

  • Rating-Boosted La...

    The performance of a recommendation system reli...