Difference between revisions of "Gradient Descent"

From Cohen Courses
Jump to navigationJump to search
(Created page with 'Gradient Descent is a method to find local minimum or maximum of an optimization of a function, in which we steps are taken in direction and proportional to negative or positive …')
 
 
Line 1: Line 1:
 
Gradient Descent is a method to find local minimum or maximum of an optimization of a function, in which we steps are taken in direction and proportional to negative or positive of gradient of the function we are optimizing.  
 
Gradient Descent is a method to find local minimum or maximum of an optimization of a function, in which we steps are taken in direction and proportional to negative or positive of gradient of the function we are optimizing.  
 +
 
[http://en.wikipedia.org/wiki/Gradient_descent Wiki Link]
 
[http://en.wikipedia.org/wiki/Gradient_descent Wiki Link]
 +
 +
== Relevant Papers ==
 +
 +
{{#ask: [[UsesMethod::Gradient Descent]]
 +
| ?AddressesProblem
 +
| ?UsesDataset
 +
}}

Latest revision as of 15:52, 31 March 2011

Gradient Descent is a method to find local minimum or maximum of an optimization of a function, in which we steps are taken in direction and proportional to negative or positive of gradient of the function we are optimizing.

Wiki Link

Relevant Papers