Facebook AI's MADGRAD optimizer improves neural network training
Facebook AI's MADGRAD optimizer improves neural network training
Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization
arXiv paper abstract https://arxiv.org/abs/2101.11075
arXiv PDF paper https://arxiv.org/pdf/2101.11075.pdf
We're introducing an optimizer for deep learning, MADGRAD. This method matches or exceeds the performance of the Adam optimizer across a varied set of realistic large-scale deep learning training problems.
Read the Docs documentation https://madgrad.readthedocs.io/en/latest
Please like and share this post if you enjoyed it using the buttons at the bottom!
Stay up to date. Subscribe to my posts https://morrislee1234.wixsite.com/website/contact
Web site with my other posts by category https://morrislee1234.wixsite.com/website
Comments