TY - GEN
T1 - Adversarial knowledge distillation for a compact generator
AU - Tsunashima, Hideki
AU - Kataoka, Hirokatsu
AU - Yamato, Junji
AU - Chen, Qiu
AU - Morishima, Shigeo
N1 - Funding Information:
This research is supported by the JST ACCEL (JPMJAC1602), JST-Mirai Program (JPMJMI19B2), JSPS KAKENHI JP17H06101, JP19H01129 and JP19H04137. Computational resource of AI Bridging Cloud Infrastructure (ABCI) provided by National Institute of Advanced Industrial Science and Technology (AIST) was used. We also want to thank Naoto Inoue, Seito Kasai, Yuchi Ishikawa, Naofumi Akimoto, Kensho Hara, and Yoshihiro Fukuhara for their helpful comments during research discussions.
Funding Information:
This research is supported by the JST ACCEL (JPM-JAC1602), JST-Mirai Program (JPMJMI19B2), JSPS KAK-ENHI JP17H06101, JP19H01129 and JP19H04137. Computational resource of AI Bridging Cloud Infrastructure (ABCI) provided by National Institute of Advanced Industrial Science and Technology (AIST) was used. We also want to thank Naoto Inoue, Seito Kasai, Yuchi Ishikawa, Naofumi Akimoto, Kensho Hara, and Yoshihiro Fukuhara for their helpful comments during research discussions.
Publisher Copyright:
© 2020 IEEE
PY - 2020
Y1 - 2020
N2 - In this paper, we propose memory-efficient Generative Adversarial Nets (GANs) in line with knowledge distillation. Most existing GANs have a shortcoming in terms of the number of model parameters and low processing speed. Here, to tackle the problem, we propose Adversarial Knowledge Distillation for Generative models (AKDG) for highly efficient GANs, in terms of unconditional generation. Using AKDG, model size and processing speed are substantively reduced. Through an adversarial training exercise with a distillation discriminator, a student generator successfully mimics a teacher generator in fewer model layers and fewer parameters and at a higher processing speed. Moreover, our AKDG is network architecture-agnostic. A Comparison of AKDG-applied models to vanilla models suggests that it achieves closer scores to a teacher generator and more efficient performance than a baseline method with respect to Inception Score (IS) and Frechet Inception Distance (FID). In CIFAR-10 experiments, improving IS/FID 1.17pt/55.19pt and in LSUN bedroom experiments, improving FID 71.1pt in comparison to the conventional distillation method for GANs.
AB - In this paper, we propose memory-efficient Generative Adversarial Nets (GANs) in line with knowledge distillation. Most existing GANs have a shortcoming in terms of the number of model parameters and low processing speed. Here, to tackle the problem, we propose Adversarial Knowledge Distillation for Generative models (AKDG) for highly efficient GANs, in terms of unconditional generation. Using AKDG, model size and processing speed are substantively reduced. Through an adversarial training exercise with a distillation discriminator, a student generator successfully mimics a teacher generator in fewer model layers and fewer parameters and at a higher processing speed. Moreover, our AKDG is network architecture-agnostic. A Comparison of AKDG-applied models to vanilla models suggests that it achieves closer scores to a teacher generator and more efficient performance than a baseline method with respect to Inception Score (IS) and Frechet Inception Distance (FID). In CIFAR-10 experiments, improving IS/FID 1.17pt/55.19pt and in LSUN bedroom experiments, improving FID 71.1pt in comparison to the conventional distillation method for GANs.
UR - http://www.scopus.com/inward/record.url?scp=85110446870&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85110446870&partnerID=8YFLogxK
U2 - 10.1109/ICPR48806.2021.9413150
DO - 10.1109/ICPR48806.2021.9413150
M3 - Conference contribution
AN - SCOPUS:85110446870
T3 - Proceedings - International Conference on Pattern Recognition
SP - 10636
EP - 10643
BT - Proceedings of ICPR 2020 - 25th International Conference on Pattern Recognition
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 25th International Conference on Pattern Recognition, ICPR 2020
Y2 - 10 January 2021 through 15 January 2021
ER -