Update weights & manage learning rate
An algorithm that adjusts the parameters of a model in order to minimize the difference between the predicted output and the actual output of the training data.
When applying a fixed learning rate, the model may oscillate or fail to converge. With constant gradients like the sign function in L1 norm, why adaptive learning rates are applied.
Model Optimizers
Model Optimizer Notion

Seonglae Cho


![[논문 리뷰] AdamW에 대해 알아보자! Decoupled weight decay regularization 논문 리뷰(1)](https://www.notion.so/image/https%3A%2F%2Fhiddenbeginner.github.io%2Fstatic%2Fimg%2FHDBG.png?table=block&id=f5f2830c-e206-4c69-bd32-6ae2f9d0083f&cache=v2)