@3blue1brown
  @3blue1brown
3Blue1Brown | Gradient descent, how neural networks learn | Chapter 2, Deep learning @3blue1brown | Uploaded 6 years ago | Updated 17 minutes ago
Enjoy these videos? Consider sharing one or two.
Help fund future projects: patreon.com/3blue1brown
Special thanks to these supporters: http://3b1b.co/nn2-thanks
Written/interactive form of this series: 3blue1brown.com/topics/neural-networks

This video was supported by Amplify Partners.
For any early-stage ML startup founders, Amplify Partners would love to hear from you via 3blue1brown@amplifypartners.com

To learn more, I highly recommend the book by Michael Nielsen
http://neuralnetworksanddeeplearning.com
The book walks through the code behind the example in these videos, which you can find here:
github.com/mnielsen/neural-networks-and-deep-learning

MNIST database:
http://yann.lecun.com/exdb/mnist

Also check out Chris Olah's blog:
http://colah.github.io
His post on Neural networks and topology is particular beautiful, but honestly all of the stuff there is great.

And if you like that, you'll *love* the publications at distill:
https://distill.pub/

For more videos, Welch Labs also has some great series on machine learning:
youtu.be/i8D90DkCLhI
youtu.be/bxe2T-V8XRs

"But I've already voraciously consumed Nielsen's, Olah's and Welch's works", I hear you say. Well well, look at you then. That being the case, I might recommend that you continue on with the book "Deep Learning" by Goodfellow, Bengio, and Courville.

Thanks to Lisha Li (@lishali88) for her contributions at the end, and for letting me pick her brain so much about the material. Here are the articles she referenced at the end:
arxiv.org/abs/1611.03530
arxiv.org/abs/1706.05394
arxiv.org/abs/1412.0233

Music by Vincent Rubinetti:
vincerubinetti.bandcamp.com/album/the-music-of-3blue1brown

Thanks to these viewers for their contributions to translations
Hebrew: Omer Tuchfeld
Italian: @teobucci

-------------------
Video timeline
0:00 - Introduction
0:30 - Recap
1:49 - Using training data
3:01 - Cost functions
6:55 - Gradient descent
11:18 - More on gradient vectors
12:19 - Gradient descent recap
13:01 - Analyzing the network
16:37 - Learning more
17:38 - Lisha Li interview
19:58 - Closing thoughts
------------------

3blue1brown is a channel about animating math, in all senses of the word animate. And you know the drill with YouTube, if you want to stay posted on new videos, subscribe, and click the bell to receive notifications (if you're into that).

If you are new to this channel and want to see more, a good place to start is this playlist: http://3b1b.co/recommended

Various social media stuffs:
Website: 3blue1brown.com
Twitter: twitter.com/3Blue1Brown
Patreon: patreon.com/3blue1brown
Facebook: facebook.com/3blue1brown
Reddit: reddit.com/r/3Blue1Brown
Gradient descent, how neural networks learn | Chapter 2, Deep learningConvolutions | Why X+Y in probability is a beautiful messWhat is backpropagation really doing? | Chapter 3, Deep learningThe determinant | Chapter 6, Essence of linear algebraSnells law proof using springsPositioned as the hardest question on a Putnam exam  (#6, 1992)Tattoos on MathThe surface area of a sphereBut why would light slow down? | Optics puzzles 3Exponential growth and epidemicsBut what is a convolution?Beyond the Mandelbrot set, an intro to holomorphic dynamics

Gradient descent, how neural networks learn | Chapter 2, Deep learning @3blue1brown

SHARE TO X SHARE TO REDDIT SHARE TO FACEBOOK WALLPAPER