梯度下降算法:
Repeat
{
θj=θjαθjJ(θ0,θ1...θn)\theta_j=\theta_j-\alpha\frac{\partial}{\partial\theta_j}J(\theta_0,\theta_1...\theta_n)
}simultaneously update for every j=0,1…n)

θj=θjα1mi=1m(hθ(x(i))y(i))xj(i)\theta_j=\theta_j-\alpha\frac{1}{m}\sum_{i=1}^{m} (h_{\theta}(x^{(i)})-y^{(i)})x_j^{(i)}
Feature Scaling以及Mean normalizaition

α\alpha太大:slow convergence
α\alpha太小:J(θ\theta) mat not decrease on every iteration,may not converge
尝试不同的α\alpha,绘制J(θ\theta)随迭代次数变化的曲线

polynominal regression(多项式回归)

Normal equation(正规方程)

θjJ(θ)=0\frac{\partial}{\partial\theta_j}J(\theta)=0 for every j

机器学习第二课
Gradient Descent 和 Normal Equation各自的优缺点
机器学习第二课

相关文章:

  • 2021-07-13
  • 2021-08-11
  • 2021-07-25
  • 2021-11-30
  • 2021-04-17
  • 2021-11-16
  • 2021-12-31
  • 2021-12-19
猜你喜欢
  • 2021-09-12
  • 2021-11-03
  • 2021-11-28
  • 2021-10-16
  • 2021-04-23
  • 2021-05-27
相关资源
相似解决方案