Open main menu
Home
Random
Recent changes
Special pages
Community portal
Preferences
About Wikipedia
Disclaimers
Incubator escapee wiki
Search
User menu
Talk
Dark mode
Contributions
Create account
Log in
Editing
Gradient descent
(section)
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== Theoretical properties == The properties of gradient descent depend on the properties of the objective function and the variant of gradient descent used (for example, if a [[line search]] step is used). The assumptions made affect the convergence rate, and other properties, that can be proven for gradient descent.<ref name=":1">{{cite arXiv|last=Bubeck |first=Sébastien |title=Convex Optimization: Algorithms and Complexity |date=2015 |class=math.OC |eprint=1405.4980 }}</ref> For example, if the objective is assumed to be [[Strongly convex function|strongly convex]] and [[Lipschitz continuity|lipschitz smooth]], then gradient descent converges linearly with a fixed step size.<ref name="auto"/> Looser assumptions lead to either weaker convergence guarantees or require a more sophisticated step size selection.<ref name=":1" />
Edit summary
(Briefly describe your changes)
By publishing changes, you agree to the
Terms of Use
, and you irrevocably agree to release your contribution under the
CC BY-SA 4.0 License
and the
GFDL
. You agree that a hyperlink or URL is sufficient attribution under the Creative Commons license.
Cancel
Editing help
(opens in new window)