Regularizing your neural network 神经网络正则化


Logistic regression regularization

先用简单的逻辑回归正则化作为例子,因为神经网络的参数W是2维的。

  1. 无正则

    J(w,b)=1mi=1mL(y^(i)y(i))

  2. L2 正则

    J(w,b)=1mi=1mL(y^(i)y(i))+λ2m||w||22

    ||w||22=j=1nxwj2=wTw

  3. L1 正则
    J(w,b)=1mi=1mL(y^(i)y(i))+λm||w||1

||w||1=j=1nx|w|j

Neural network regularization

  1. Frobenius正则(类似L2正则)
    J(w[1],b[1],,w[l],b[l])=1mi=1mL(y^(i),y(i))+12ml=1L||w[l]||F2

    ||w[l]||F2=i=1n[l]j=1n[l1](wij[l])2

相较于无正则化的反向传播,正则化的反向传播在更新W时,会对其进行权重衰减(weight decay),并下降。

dw[l]=(from backpropagation)+λmw[l]

w[l]:=w[l]αdw[l]=w[l]αλmw[l]α(from backpropagation)=(1αλm)w[l]α(from backpropagation)

Deep learning II - I Practical aspects of deep learning - Regularizing your neural network 神经网络范数正则化
Deep learning II - I Practical aspects of deep learning - Regularizing your neural network 神经网络范数正则化

相关文章: