4.6 Article

STOCHASTIC FIRST- AND ZEROTH-ORDER METHODS FOR NONCONVEX STOCHASTIC PROGRAMMING

期刊

SIAM JOURNAL ON OPTIMIZATION
卷 23, 期 4, 页码 2341-2368

出版社

SIAM PUBLICATIONS
DOI: 10.1137/120880811

关键词

stochastic approximation; nonconvex optimization; stochastic programming; simulation-based optimization

资金

  1. NSF [CMMI-1000347, DMS-1319050]
  2. ONR [N00014-13-1-0036]
  3. NSF CAREER Award [CMMI-1254446]

向作者/读者索取更多资源

In this paper, we introduce a new stochastic approximation type algorithm, namely, the randomized stochastic gradient (RSG) method, for solving an important class of nonlinear (possibly nonconvex) stochastic programming problems. We establish the complexity of this method for computing an approximate stationary point of a nonlinear programming problem. We also show that this method possesses a nearly optimal rate of convergence if the problem is convex. We discuss a variant of the algorithm which consists of applying a postoptimization phase to evaluate a short list of solutions generated by several independent runs of the RSG method, and we show that such modification allows us to improve significantly the large-deviation properties of the algorithm. These methods are then specialized for solving a class of simulation-based optimization problems in which only stochastic zeroth-order information is available.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据