4.6 Article

OPTIMAL STOCHASTIC APPROXIMATION ALGORITHMS FOR STRONGLY CONVEX STOCHASTIC COMPOSITE OPTIMIZATION, II: SHRINKING PROCEDURES AND OPTIMAL ALGORITHMS

期刊

SIAM JOURNAL ON OPTIMIZATION
卷 23, 期 4, 页码 2061-2089

出版社

SIAM PUBLICATIONS
DOI: 10.1137/110848876

关键词

stochastic approximation; convex optimization; strong convexity; complexity; optimal method; large deviation

资金

  1. NSF [CMMI-1000347]
  2. ONR [N00014-13-1-0036]
  3. NSF CAREER Award [CMMI-1254446]

向作者/读者索取更多资源

In this paper we study new stochastic approximation (SA) type algorithms, namely, the accelerated SA (AC-SA), for solving strongly convex stochastic composite optimization (SCO) problems. Specifically, by introducing a domain shrinking procedure, we significantly improve the large-deviation results associated with the convergence rate of a nearly optimal AC-SA algorithm presented by Ghadimi and Lan in [SIAM J. Optim., 22 (2012), pp 1469-1492]. Moreover, we introduce a multistage AC-SA algorithm, which possesses an optimal rate of convergence for solving strongly convex SCO problems in terms of the dependence on not only the target accuracy, but also a number of problem parameters and the selection of initial points. To the best of our knowledge, this is the first time that such an optimal method has been presented in the literature. From our computational results, these AC-SA algorithms can substantially outperform the classical SA and some other SA type algorithms for solving certain classes of strongly convex SCO problems.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据