5.1 Cost Function

假设训练样本为:{(x1),y(1)),(x(2),y(2)),...(x(m),y(m))}

L  = total no.of layers in network

sL= no,of units(not counting bias unit) in layer L

K = number of output units/classes

吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)如图所示的神经网络,L = 4,s1 = 3,s2 = 5,s3 = 5, s4 = 4

逻辑回归的代价函数:

            吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)

神经网络的代价函数:

   吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)

 

 5.2 反向传播算法 Backpropagation

 关于反向传播算法的一篇通俗的解释http://blog.csdn.net/shijing_0214/article/details/51923547

吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)

 

 吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)

 5.3 Training a neural network

 吴恩达机器学习第5周Neural Networks(Cost Function and Backpropagation)

隐藏层的单元数一般一样,隐藏层一般越多越好,但计算量会较大。

Training a Neural Network

  1. Randomly initialize the weights
  2. Implement forward propagation to get )
  3. Implement the cost function
  4. Implement backpropagation to compute partial derivatives
  5. Use gradient checking to confirm that your backpropagation works. Then disable gradient checking.
  6. Use gradient descent or a built-in optimization function to minimize the cost function with the weights in theta.

 

相关文章:

  • 2022-12-23
  • 2021-04-27
  • 2021-09-17
  • 2021-09-11
  • 2021-08-25
  • 2021-05-28
  • 2021-05-23
猜你喜欢
  • 2022-12-23
  • 2021-04-01
  • 2021-10-22
  • 2021-09-16
  • 2021-07-28
  • 2021-07-08
  • 2021-11-15
相关资源
相似解决方案