论文标题
具有量化梯度的生成对抗网络的分布式培训算法
A Distributed Training Algorithm of Generative Adversarial Networks with Quantized Gradients
论文作者
论文摘要
以分布式方式培训生成的对抗网络(GAN)是一项有前途的技术,因为它有助于对GAN进行培训,以在现实世界中有效地对大量数据进行培训。但是,已知GAN难以通过SGD型方法(可能无法融合)训练,并且分布式SGD型方法也可能遭受大量的通信成本。在本文中,我们提出了一种{分布式gans训练算法,具有量化梯度,称为dqgan,},这是第一种具有量化gan的梯度的分布式训练方法。新方法基于一种称为乐观镜下降(OMD)算法的特定单机器算法的gans,并且适用于满足一般$δ$δ$ appapproximate压缩器的任何梯度压缩方法。我们设计的错误反馈操作用于补偿由压缩引起的偏差,并确保新方法的收敛性。从理论上讲,我们建立了{dqgan}算法到一阶固定点的非反应收敛性,这表明所提出的算法可以在参数服务器模型中实现线性加速。从经验上讲,我们的实验表明,我们的{dqgan}算法可以降低通信成本并节省训练时间,并在合成数据集和真实数据集上略有性能下降。
Training generative adversarial networks (GAN) in a distributed fashion is a promising technology since it is contributed to training GAN on a massive of data efficiently in real-world applications. However, GAN is known to be difficult to train by SGD-type methods (may fail to converge) and the distributed SGD-type methods may also suffer from massive amount of communication cost. In this paper, we propose a {distributed GANs training algorithm with quantized gradient, dubbed DQGAN,} which is the first distributed training method with quantized gradient for GANs. The new method trains GANs based on a specific single machine algorithm called Optimistic Mirror Descent (OMD) algorithm, and is applicable to any gradient compression method that satisfies a general $δ$-approximate compressor. The error-feedback operation we designed is used to compensate for the bias caused by the compression, and moreover, ensure the convergence of the new method. Theoretically, we establish the non-asymptotic convergence of {DQGAN} algorithm to first-order stationary point, which shows that the proposed algorithm can achieve a linear speedup in the parameter server model. Empirically, our experiments show that our {DQGAN} algorithm can reduce the communication cost and save the training time with slight performance degradation on both synthetic and real datasets.