zoukankan      html  css  js  c++  java
  • GAN Compression: Efficient Architectures for Interactive Conditional GANs

    GAN Compression: Efficient Architectures for Interactive Conditional GANs

    Background

    Conditional Generated Adversarial Network (CGAN) provides controlled image synthesis for many computer vision and graphics applications. However, the computational strength of CGAN is 1-2 orders of magnitude higher than that of CNN used for recognition.

    Contributions

    • The author transfers the knowledge from the teacher generator to student generator and find it beneficial to create pseudo pairs using the teacher model’s output for unpaired training. This transforms the unpaired learning to a paired learning.
    • The author uses neural architecture search (NAS) to automatically find an efficient network with significantly fewer computation costs and parameters.

     Method

             

    As shown in equation 1, the author unifies the unpaired and paired learning in the model

    compression setting. Since CGANs usually output a deterministic image, rather than a probabilistic distribution. Therefore, it is difficult to distill the dark knowledge from the teacher’s output pixels. The author use equation 3 L2 loss as the distill loss for the regression in different stages. 

    To reducing the number of channels, the author automatically first selects the channel width in the generators using channel pruning [23, 22, 44, 81, 47] to remove the redundancy. For each convolution layers, the number of channels is multiples of 8. A straight-forward approach is to traverse all the possible channel configuration, train it to find which is the best. However, to this approach, if we have 8 layers and which layers have 4 channel candidates {8,16,32,64}, so we need to train the eighth power of four models, which is very time-consuming. Hence, the author trains a “once-for-all” network [7] that supports different channel numbers. Each sub-network with different numbers of channels are equally trained and can operate independently. Sub-networks share the weights with the “once-for-all” network. In this case, they can only need to train once, but they can evaluate all the possible channel configurations without further training,

    Experiments

    With the fast GAN compression proposed by authors, we can find this method can largely reduce the training time.

    As shown in Table 1, by using the best performing subnetwork from the “once-for-all” network, GAN Compression not only reduce the model size but also remain the performance.

    In order to test how the GAN compression perform in practical devices, they measure the inference speed of our compressed models on several devices with different computational powers. As shown in Table 2, the proposed method performs well on CPU devices.

    Strength

    Overall, this paper proposed a compression framework for reducing the computational cost and model size of generators in CGANs and use knowledge distillation and neural architecture search to alleviate the training instability and increase the model efficiency. Extensive experiments have shown that this method can compress several CGAN models while preserving the visual quality.

    Weakness

    To be honest, the proposed method is interesting and can be used in a lot of image to image tasks, like image super-resolution. The extensive experiments also demonstrate the effectiveness of the proposed method. I think the only weakness is the description of the part of NAS and Decouple Training and Search is so vague that I take a long time to capture the author’s ideas.

  • 相关阅读:
    Spring MVC Controller中GET方式传过来的中文参数会乱码的问题
    解决Scala异常处理java.lang.OutOfMemoryError: Java heap space error
    sbt的安装设置
    Oracle VM VirtualBox 设置本机与虚拟机共享文件
    Hadoop fs命令详解
    Oracle 11g CPU负载很高紧急处理【负载100%】
    Oracle单表备份三种方案
    Nginx1.9.xx以后版支持4层tcp转发
    用SQL命令查看Mysql数据库大小
    Jenkins: ERROR: Exception when publishing, exception message [java.net.SocketException: Broken pipe (Write failed)]
  • 原文地址:https://www.cnblogs.com/echoboy/p/14918705.html
Copyright © 2011-2022 走看看