论文标题
生成对抗网络的有效动态
Effective Dynamics of Generative Adversarial Networks
论文作者
论文摘要
生成对抗网络(GAN)是一类机器学习模型,它们使用对抗性训练来生成与培训样本相同(可能非常复杂)统计的新样本。一种主要的训练失败形式,称为模式崩溃,涉及发电机未能重现目标概率分布中模式的全部多样性。在这里,我们提出了一个有效的GAN训练模型,该模型通过用输出空间中的粒子收集来代替发电机神经网络来捕获学习动力。颗粒由通用内核对某些宽神经网络和高维输入有效。我们简化模型的通用性使我们能够研究发生模式崩溃的条件。实际上,改变发生器有效核的实验揭示了模式塌陷过渡,其形状可以通过频率原理与鉴别器的类型相关。此外,我们发现中间强度的梯度正规化器可以通过发电机动力学的严重阻尼来最佳地产生收敛。因此,我们有效的GAN模型为理解和改善对抗性训练提供了可解释的物理框架。
Generative adversarial networks (GANs) are a class of machine-learning models that use adversarial training to generate new samples with the same (potentially very complex) statistics as the training samples. One major form of training failure, known as mode collapse, involves the generator failing to reproduce the full diversity of modes in the target probability distribution. Here, we present an effective model of GAN training, which captures the learning dynamics by replacing the generator neural network with a collection of particles in the output space; particles are coupled by a universal kernel valid for certain wide neural networks and high-dimensional inputs. The generality of our simplified model allows us to study the conditions under which mode collapse occurs. Indeed, experiments which vary the effective kernel of the generator reveal a mode collapse transition, the shape of which can be related to the type of discriminator through the frequency principle. Further, we find that gradient regularizers of intermediate strengths can optimally yield convergence through critical damping of the generator dynamics. Our effective GAN model thus provides an interpretable physical framework for understanding and improving adversarial training.