1、Model representation
- Our Training Set [训练集]:
- We will start with this ‘’Housing price prediction‘’ example first of fitting linear functions, and we will build on this to eventually have more complex models
2、Cost function
- 代价函数(平方误差函数):It figures out how to fit the best possible straight line to our data
- So how to choose θi's ?
- and just try:
- The parameters we choose determine the accuracy of the straight line we get relative to our training set
- But there is modeling error 建模误差
Our goal is to select the model parameters that minimize the sum of squares of modeling errors
That is to minimize the cost function!
summary:
2-1、Cost function introduction I
- We look up some plots to understand the cost function
2-2、Cost function introduction II
- Let's take a look at the three-dimensional space diagram of the cost function(also called a convex function 凸函数)
- And here is an example of a contour figure:
- The contour figure is a more convenient way to visualize the cost function
3、Gradient descent
- It turns out gradient descent(梯度下降) is a more general algorithm and is used not only in linear regression. I will introduce how to use gradient descent for minimizing some arbitrary function J
- The formula of the batch gradient descent algorithm :
4、Gradient descent intuition
Derivative term purpose :get closer to the minimum
Learning rate α :
- But what if my parameter θ1 is already at a local minimum?
- Gradient descent can converge to a local minimum, even with the learning rate α fixed