gradient ascent maximize and descent minimize function
Iteratively subtract the derivative of Loss Function wrt weight from the weight with the learning rate and then Cost Function decreases
Repeat until the weight convergence (Local extremum point)
Gradient Descent Notion
adventures/dual-numbers-and-gradient-descent.ipynb at master · samuelbelko/adventures
Simple math oriented projects using Julia and JupyterLab. - adventures/dual-numbers-and-gradient-descent.ipynb at master · samuelbelko/adventures
Online gradient descent written in SQL • Max Halford
Edit: this post generated a few insightful comments on Hacker News. I’ve also put the code in a notebook for ease of use. Introduction Modern MLOps is complex because it involves too many components. You need a message bus, a stream processing engine, an API, a model store, a feature store, a monitoring service, etc. Sadly, containerisation software and the unbundling trend have encouraged an appetite for complexity. I believe MLOps shouldn’t be this complex.