Tho Le
A Data Scientist. Looking for knowledge!
Navigation
Home
About Me
My publications
My Projects
XML Feed
Optimizers
14 Mar 2025
»
ml
,
dl
,
optimizers
Adam
RMSProp
SGD (Stochastic Gradient Descent)
Momentum SGD
Adagrad (Adaptive Gradient)
Adadelta
AdamW (Adam with Weight Decay)
Share this on →
Tweet
Related Posts
Deep Learning Explained
(Categories:
ai
,
dl
,
explain
)
Forecast Resources
(Categories:
techniques
,
ml
,
forecast
)
Some notes on para/hyper-para of a few ML models
(Categories:
ml
,
parameters
)
ML Design Notes
(Categories:
system
,
ml
,
design
)
Deep Learning - Hands-On
(Categories:
ml
,
dl
,
hands-on
)
ML Resources
(Categories:
techniques
,
ml
,
resources
)
« Some notes on para/hyper-para of a few ML models
LLM Tools »