zoukankan      html  css  js  c++  java
  • 04 Regularization

    Regularization

    for Linear Regression and Logistic Regression

    Define

    1. under-fitting 欠拟合(high bias)
    2. over-fitting 过拟合 (high variance):have too many features, fail to generalize(泛化) to new examples.

    Addressing over-fitting

    1. Reduce number of features.
      • Manually select which features to keep.
      • Model selection algorithm.
    2. Regularization
      • Keep all the features. but reduce magnitude/values of parameters ( heta_j).
      • Works well when we have a lot of features, each of whitch contributes a bit to predicting (y).

    Regularized Cost Function

    • [min_ heta dfrac{1}{2m} sum_{i=1}^m (h_ heta(x^{(i)}) - y^{(i)})^2 + lambda sum_{j=1}^n heta_j^2]

    Regularized Linear Regression

    1. Gradient Descent
      [
      egin{align*} & ext{Repeat} lbrace ewline & heta_0 := heta_0 - alpha frac{1}{m} sum_{i=1}^m (h_ heta(x^{(i)}) - y^{(i)})x_0^{(i)} ewline & heta_j := heta_j - alpha left[ left( frac{1}{m} sum_{i=1}^m (h_ heta(x^{(i)}) - y^{(i)})x_j^{(i)} ight) + frac{lambda}{m} heta_j ight] & j in lbrace 1,2...n brace ewline & brace end{align*}
      ]

      • 等价于
        [
        heta_j := heta_j(1 - alphafrac{lambda}{m}) - alphafrac{1}{m} sum_{i=1}^m(h_ heta(x^{(i)}) - y^{(i)})x_j^{(i)}
        ]
    2. Normal Equation
      [
      egin{align*}& heta = left( X^TX + lambda cdot L ight)^{-1} X^Ty ewline& ext{where} L = egin{bmatrix} 0 & & & & ewline & 1 & & & ewline & & 1 & & ewline & & & ddots & ewline & & & & 1 ewlineend{bmatrix}end{align*}
      ]

    • 对于不可逆的((X^TX)), ((X^TX + lambda.L)) 会可逆

    Regularized Logistic Regression

    1. Cost Function

    [
    J( heta) = -frac{1}{m} sum_{i=1}^m[y^{(i)}log(h_ heta(x^{(i)})) + (1 - y^{(i)})log(1 - h_ heta(x^{(i)}))] + frac{lambda}{2m}sum_{j=1}^n heta_j^2
    ]

    1. Gradient descent
      [
      egin{align*} & ext{Repeat} lbrace ewline & heta_0 := heta_0 - alpha frac{1}{m} sum_{i=1}^m (h_ heta(x^{(i)}) - y^{(i)})x_0^{(i)} ewline & heta_j := heta_j - alpha left[ left( frac{1}{m} sum_{i=1}^m (h_ heta(x^{(i)}) - y^{(i)})x_j^{(i)} ight) + frac{lambda}{m} heta_j ight] & j in lbrace 1,2...n brace ewline & brace end{align*}
      ]
  • 相关阅读:
    Android开发-API指南-<uses-configuration>
    Android开发-API指南-<supports-screens>
    Android开发-API指南-<supports-gl-texture>
    Android开发-API指南-<service>
    Android开发-API指南-<receiver>
    Android开发-API指南-<provider>
    Android开发-API指南-<permission-tree>
    maven scope 以及依赖传递
    转发和重定向
    Apache
  • 原文地址:https://www.cnblogs.com/QQ-1615160629/p/04-Regularization.html
Copyright © 2011-2022 走看看