4.7 Review

Recent Advances in Stochastic Gradient Descent in Deep Learning

期刊

MATHEMATICS
卷 11, 期 3, 页码 -

出版社

MDPI
DOI: 10.3390/math11030682

关键词

stochastic gradient descent; machine learning; deep learning

向作者/读者索取更多资源

In the age of artificial intelligence, finding the best approach to handle massive data is a challenging task. Stochastic gradient descent (SGD) stands out among machine learning models as it is simple yet highly effective. This study examines various contemporary deep learning applications, including natural language processing (NLP), visual data processing, and voice and audio processing. The study also presents different versions of SGD and its variant available in the PyTorch optimizer, such as SGD, Adagrad, adadelta, RMSprop, Adam, AdamW, etc. Additionally, theoretical conditions for the applicability of these methods are proposed, highlighting the existing gap between theoretical convergence and practical implementation.
In the age of artificial intelligence, the best approach to handling huge amounts of data is a tremendously motivating and hard problem. Among machine learning models, stochastic gradient descent (SGD) is not only simple but also very effective. This study provides a detailed analysis of contemporary state-of-the-art deep learning applications, such as natural language processing (NLP), visual data processing, and voice and audio processing. Following that, this study introduces several versions of SGD and its variant, which are already in the PyTorch optimizer, including SGD, Adagrad, adadelta, RMSprop, Adam, AdamW, and so on. Finally, we propose theoretical conditions under which these methods are applicable and discover that there is still a gap between theoretical conditions under which the algorithms converge and practical applications, and how to bridge this gap is a question for the future.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据