zoukankan      html  css  js  c++  java
  • Activation Functions

    Sigmoid

    Sigmoids saturate and kill gradients.

    Sigmoid outputs are not zero-centered.

    Exponential function is a little computational expensive.

    Tanh

    Kill gradients when saturated.

    It's zero-centered! : )

    ReLU

    Does not saturate. ( in positive region)

    Very computational efficient.

    Converges much faster than sigmoid/tanh in practice. (6 times)

    Seems more biologically plausible than sigmoid.

    BUT!

    Not zero-centered.

    No gradient when x<0.

    Take care of learning rate when using ReLU.

    Leakly ReLU

    Does not saturate.

    Very computational efficient.

    Converges much faster than sigmoid/tanh in practice. (6 times)

    will not "die"

     

    Parametric ReLU

    Exponential Linear Unit

  • 相关阅读:
    Friends ZOJ
    2^x mod n = 1 HDU
    Paint the Grid Reloaded ZOJ
    Treap 模板
    bzoj进度条
    。。。
    bzoj
    。。。
    bzoj
    题解continue
  • 原文地址:https://www.cnblogs.com/hizhaolei/p/10623472.html
Copyright © 2011-2022 走看看