河北大学学报(自然科学版) ›› 2021, Vol. 41 ›› Issue (6): 734-744.DOI: 10.3969/j.issn.1000-1565.2021.06.014

• • 上一篇    

多生成器生成对抗网络

申瑞彩,翟俊海,侯璎真   

  • 发布日期:2021-12-08
  • 通讯作者: 翟俊海(1964—)
  • 作者简介:申端彩(1993—),女,河北邯郸人,河北大学在读硕士研究生,主要从事深度学习研究.E-mail:1943303808@qq.com
  • 基金资助:
    河北省科技计划重点研发计划项目(19210310D);河北省自然科学基金资助项目(F2021201020)

Multi-generator generative adversarial networks

SHEN Ruicai, ZHAI Junhai, HOU Yingzhen   

  1. Hebei Key Laboratory of Machine Learning and Computational Intelligence, College of Mathematics and Information Science, Hebei University, Baoding 071002, China
  • Published:2021-12-08

摘要: 生成对抗网络(Generative adversarial networks,GAN)广泛应用于各种领域,尤其在图像生成方面.该模型由生成网络与判别网络2部分组成,在无监督的训练方式下,2个网络相互竞争相互提高.然而,GAN在训练时经常出现模式崩溃问题,进而导致模型收敛较慢,生成样本多样性较差.为解决这一问题,在深度卷积神经网络的基础上提出了一种多生成器生成对抗网络模型.该模型包含多个生成网络,每个生成网络均使用残差网络进行搭建,同时在生成网络间引入协作机制,以加快模型获取信息并减少参数量,最后将各生成网络的特征图进行融合得到最终图像输入到判别网络中.GAN在训练过程中还会出现梯度消失、训练不稳定问题.为避免出现这些问题,将Wasserstein距离和梯度惩罚引入模型的损失函数.通过在多个数据集上与多种相关方法进行实验比较,结果表明提出的模型在缓解模式崩溃问题、加快模型收敛速度以及减少参数量上均明显优于其他几种方法.

关键词: 生成对抗网络, 残差网络, 集成学习, 模式崩溃, Wasserstein距离

Abstract: Generative adversarial networks(GAN)are widely used in various fields, especially in image generation. The model consists of two parts, a generative network and a discriminative network, and the two networks compete with each other to improve each other in an unsupervised training method. However, GAN often suffers from pattern collapse during training, which leads to slow convergence of the model and poor diversity of generated samples. To solve this problem, a multi-generator generative adversarial network model based on deep convolutional neural networks is proposed. The model consists of multiple generative networks, each of which is built using residual networks, and a collaboration mechanism is introduced among the generative networks to speed up the information acquisition and reduce the number of parameters, and finally the feature maps of each generative network are fused to obtain the final image and input to the discriminative network. GAN also has the problems of gradient disappearance and training instability. To address the two problems, Wasserstein distance and gradient penalty are introduced into the loss function of the mode. Through experimental comparison with several related methods on multiple datasets, the results show that the proposed model significantly outperforms several- DOI:10.3969/j.issn.1000-1565.2021.06.014多生成器生成对抗网络申瑞彩,翟俊海,侯璎真(河北大学 数学与信息科学学院,河北省机器学习与计算智能重点实验室,河北 保定 071002)摘 要:生成对抗网络(Generative adversarial networks,GAN)广泛应用于各种领域,尤其在图像生成方面.该模型由生成网络与判别网络2部分组成,在无监督的训练方式下,2个网络相互竞争相互提高.然而,GAN在训练时经常出现模式崩溃问题,进而导致模型收敛较慢,生成样本多样性较差.为解决这一问题,在深度卷积神经网络的基础上提出了一种多生成器生成对抗网络模型.该模型包含多个生成网络,每个生成网络均使用残差网络进行搭建,同时在生成网络间引入协作机制,以加快模型获取信息并减少参数量,最后将各生成网络的特征图进行融合得到最终图像输入到判别网络中.GAN在训练过程中还会出现梯度消失、训练不稳定问题.为避免出现这些问题,将Wasserstein距离和梯度惩罚引入模型的损失函数.通过在多个数据集上与多种相关方法进行实验比较,结果表明提出的模型在缓解模式崩溃问题、加快模型收敛速度以及减少参数量上均明显优于其他几种方法.关键词:生成对抗网络;残差网络;集成学习;模式崩溃;Wasserstein距离 中图分类号:TP181 文献标志码:A 文章编号:1000-1565(2021)06-0734-11Multi-generator generative adversarial networksSHEN Ruicai, ZHAI Junhai, HOU Yingzhen(Hebei Key Laboratory of Machine Learning and Computational Intelligence, College of Mathematics and Information Science, Hebei University, Baoding 071002,China)Abstract: Generative adversarial networks(GAN)are widely used in various fields, especially in image generation. The model consists of two parts, a generative network and a discriminative network, and the two networks compete with each other to improve each other in an unsupervised training method. However, GAN often suffers from pattern collapse during training, which leads to slow convergence of the model and poor diversity of generated samples. To solve this problem, a multi-generator generative adversarial network model based on deep convolutional neural networks is proposed. The model consists of multiple generative networks, each of which is built using residual networks, and a collaboration mechanism is introduced among the generative networks to speed up the information acquisition and reduce the number of parameters, and finally the feature maps of each generative network are fused to obtain the final image and input to the discriminative network. GAN also has the problems of gradient disappearance and training instability. To address the two problems, Wasserstein distance and gradient penalty are introduced into the loss function of the mode. Through experimental comparison with several related methods on multiple datasets, the results show that the proposed model significantly outperforms several- 收稿日期:2021-05-26 基金项目:河北省科技计划重点研发计划项目(19210310D); 河北省自然科学基金资助项目(F2021201020) 第一作者:申端彩(1993—),女,河北邯郸人,河北大学在读硕士研究生,主要从事深度学习研究.E-mail:1943303808@qq.com 通信作者:翟俊海(1964—),男, 河北易县人, 河北大学教授, 博士生导师, 主要从事云计算与大数据处理和深度学习方向研究.E-mail:mczjh@hbu.cn第6期申瑞彩等:多生成器生成对抗网络other methods in alleviating the pattern collapse problem, speeding up model convergence, and reducing the number of parameters.

Key words: generative adversarial networks(GAN), residual networks, ensemble learning, pattern collapse, Wasserstein distance

中图分类号: