Faruk Ahmad’s Post

View profile for Faruk Ahmad, graphic

Senior AI Engineer @ Deloitte | AWS Certified Solution Architect Associate | Google Certified TensorFlow developer | Data Science Enthusiast | Opinions are my own.

Exciting development in optimization! 👏 Researchers from the University of Tokyo have introduced ADOPT, a new adaptive gradient method that addresses the convergence issues of Adam without the need for specific Hyperparameter tuning. ADOPT achieves an optimal convergence rate and shows superior performance across multiple tasks, including image classification and large language models. [The paper has been accepted in NeurIPS 2024] For anyone working with adaptive optimizers, this is a must-read! Check out the paper for detailed insights and theoretical analysis. Arxiv Link: https://lnkd.in/g4sZvDzd GitHub Implementation: https://lnkd.in/ga2NUTfj #AI #MachineLearning #DeepLearning #Optimization #Research

GitHub - iShohei220/adopt: Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"

GitHub - iShohei220/adopt: Official Implementation of "ADOPT: Modified Adam Can Converge with Any β2 with the Optimal Rate"

github.com

To view or add a comment, sign in

Explore topics