There is a connection between gradient descent based optimizers and the dynamics of damped harmonic oscillators. What does that mean? We now have a better theory for optimization algorithms.
In this episode I explain how all this works.
All the formulas I mention in the episode can be found in the post The physics of optimization algorithms
Enjoy the show.
Attacking machine learning for fun and profit (with the authors of SecML Ep. 80)
[RB] How to scale AI in your organisation (Ep. 79)
Replicating GPT-2, the most dangerous NLP model (with Aaron Gokaslan) (Ep. 78)
Training neural networks faster without GPU [RB] (Ep. 77)
How to generate very large images with GANs (Ep. 76)
[RB] Complex video analysis made easy with Videoflow (Ep. 75)
[RB] Validate neural networks without data with Dr. Charles Martin (Ep. 74)
How to cluster tabular data with Markov Clustering (Ep. 73)
Waterfall or Agile? The best methodology for AI and machine learning (Ep. 72)
Training neural networks faster without GPU (Ep. 71)
Validate neural networks without data with Dr. Charles Martin (Ep. 70)
Complex video analysis made easy with Videoflow (Ep. 69)
Episode 68: AI and the future of banking with Chris Skinner [RB]
Episode 67: Classic Computer Science Problems in Python
Episode 66: More intelligent machines with self-supervised learning
Episode 65: AI knows biology. Or does it?
Episode 64: Get the best shot at NLP sentiment analysis
Episode 63: Financial time series and machine learning
Episode 62: AI and the future of banking with Chris Skinner
Episode 61: The 4 best use cases of entropy in machine learning
Create your
podcast in
minutes
It is Free
Insight Story: Tech Trends Unpacked
Zero-Shot
Fast Forward by Tomorrow Unlocked: Tech past, tech future
The Unbelivable Truth - Series 1 - 26 including specials and pilot
Lex Fridman Podcast