@deepbean
  @deepbean
DeepBean | Optimization for Deep Learning (Momentum, RMSprop, AdaGrad, Adam) @deepbean | Uploaded 1 year ago | Updated 2 days ago
Here we cover six optimization schemes for deep neural networks: stochastic gradient descent (SGD), SGD with momentum, SGD with Nesterov momentum, RMSprop, AdaGrad and Adam.

Chapters
---------------
Introduction 00:00
Brief refresher 00:27
Stochastic gradient descent (SGD) 03:16
SGD with momentum 05:01
SGD with Nesterov momentum 07:02
AdaGrad 09:46
RMSprop 12:20
Adam 13:23
SGD vs Adam 15:03
Optimization for Deep Learning (Momentum, RMSprop, AdaGrad, Adam)Transformers: The Model Behind ChatGPTTime Dilation and Length Contraction | Special RelativityBackpropagation: How Neural Networks LearnEinsteins Twin Paradox - The REAL SolutionWhat is Spacetime? | Special RelativityHow YOLO Object Detection WorksThe Physics of Nuclear WeaponsEinsteins Ladder Paradox; Simply ExplainedThe Geiger-Marsden Experiments | Nuclear Physics

Optimization for Deep Learning (Momentum, RMSprop, AdaGrad, Adam) @deepbean