zoukankan      html  css  js  c++  java
  • GAN Compression: Efficient Architectures for Interactive Conditional GANs

    GAN Compression: Efficient Architectures for Interactive Conditional GANs

    Background

    Conditional Generated Adversarial Network (CGAN) provides controlled image synthesis for many computer vision and graphics applications. However, the computational strength of CGAN is 1-2 orders of magnitude higher than that of CNN used for recognition.

    Contributions

    • The author transfers the knowledge from the teacher generator to student generator and find it beneficial to create pseudo pairs using the teacher model’s output for unpaired training. This transforms the unpaired learning to a paired learning.
    • The author uses neural architecture search (NAS) to automatically find an efficient network with significantly fewer computation costs and parameters.

     Method

             

    As shown in equation 1, the author unifies the unpaired and paired learning in the model

    compression setting. Since CGANs usually output a deterministic image, rather than a probabilistic distribution. Therefore, it is difficult to distill the dark knowledge from the teacher’s output pixels. The author use equation 3 L2 loss as the distill loss for the regression in different stages. 

    To reducing the number of channels, the author automatically first selects the channel width in the generators using channel pruning [23, 22, 44, 81, 47] to remove the redundancy. For each convolution layers, the number of channels is multiples of 8. A straight-forward approach is to traverse all the possible channel configuration, train it to find which is the best. However, to this approach, if we have 8 layers and which layers have 4 channel candidates {8,16,32,64}, so we need to train the eighth power of four models, which is very time-consuming. Hence, the author trains a “once-for-all” network [7] that supports different channel numbers. Each sub-network with different numbers of channels are equally trained and can operate independently. Sub-networks share the weights with the “once-for-all” network. In this case, they can only need to train once, but they can evaluate all the possible channel configurations without further training,

    Experiments

    With the fast GAN compression proposed by authors, we can find this method can largely reduce the training time.

    As shown in Table 1, by using the best performing subnetwork from the “once-for-all” network, GAN Compression not only reduce the model size but also remain the performance.

    In order to test how the GAN compression perform in practical devices, they measure the inference speed of our compressed models on several devices with different computational powers. As shown in Table 2, the proposed method performs well on CPU devices.

    Strength

    Overall, this paper proposed a compression framework for reducing the computational cost and model size of generators in CGANs and use knowledge distillation and neural architecture search to alleviate the training instability and increase the model efficiency. Extensive experiments have shown that this method can compress several CGAN models while preserving the visual quality.

    Weakness

    To be honest, the proposed method is interesting and can be used in a lot of image to image tasks, like image super-resolution. The extensive experiments also demonstrate the effectiveness of the proposed method. I think the only weakness is the description of the part of NAS and Decouple Training and Search is so vague that I take a long time to capture the author’s ideas.

  • 相关阅读:
    Android权限大全(链接地址整理)
    Android6.0运行时权限(基于RxPermission开源库)
    Android6.0机型上调用系统相机拍照返回的resultCode值始终等于0的问题
    使用AccessibilityService模拟点击事件失败的分析
    Android混淆打包配置总结
    okhttputils开源库的混淆配置(Eclipse)
    Android布局中的空格以及占一个汉字宽度的空格,实现不同汉字字数对齐
    java开发者大会:总结
    JAVA开发者大会:拍拍贷MQ系统原理与应用
    消息总线真的能保证幂等?
  • 原文地址:https://www.cnblogs.com/echoboy/p/14918705.html
Copyright © 2011-2022 走看看