4.2 Article

Fast Bayesian hyperparameter optimization on large datasets

期刊

ELECTRONIC JOURNAL OF STATISTICS
卷 11, 期 2, 页码 4945-4968

出版社

INST MATHEMATICAL STATISTICS
DOI: 10.1214/17-EJS1335SI

关键词

-

资金

  1. European Research Council (ERC) under the European Union's Horizon 2020 research and innovation programme [716721]
  2. European Commission [H2020-ICT-645403-ROBDREAM]
  3. German Research Foundation (DFG) under Priority Programme Autonomous Learning [SPP 1527, BR 3815/8-1, HU 1900/3-1]

向作者/读者索取更多资源

Bayesian optimization has become a successful tool for optimizing the hyperparameters of machine learning algorithms, such as support vector machines or deep neural networks. Despite its success, for large datasets, training and validating a single configuration often takes hours, days, or even weeks, which limits the achievable performance. To accelerate hyperparameter optimization, we propose a generative model for the validation error as a function of training set size, which is learned during the optimization process and allows exploration of preliminary configurations on small subsets, by extrapolating to the full dataset. We construct a Bayesian optimization procedure, dubbed Fabolas, which models loss and training time as a function of dataset size and automatically trades off high information gain about the global optimum against computational cost. Experiments optimizing support vector machines and deep neural networks show that Fabolas often finds high-quality solutions 10 to 100 times faster than other state-of-the-art Bayesian optimization methods or the recently proposed bandit strategy Hyperband.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.2
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据