CPSC 340 Gradient Descent

Posted by Luna's Blog on February 14, 2022

Gradient Descent

  • starts with a guess w0
  • use the gradient f(w0) to generate a better w1
  • the limit of wt goes to has f(wt)=0

It converges to a global optimum if f is convex. More details Gradient Descent

Gradient Descent for Least Squares

f(w)=12||Xwy||2

f(w)=XT(Xwy)

wt+1=wtαtXT(Xwy)

Cost of each iteration is O(nd), t iterations.

How to Know a Convex Function