Preliminary Evaluation of Hyperopt Algorithms on HPOLib

ICML 2014 AutoML Workshop, 2014

James Bergstra, Brent Komer, Chris Eliasmith, David Warde-Farley

Abstract

Model selection, also known as hyperparameter tuning, can be viewed as a blackbox optimization problem. Recently the HPOlib benchmarking suite was advanced to facilitate algorithm comparison between hyperparameter optimization algorithms. We compare seven optimization algorithms implemented in the Hyperopt optimization package, including a new annealing-type algorithm and a new family of Gaussian Process-based SMBO methods, on four screening problems from HPOLib. We find that methods based on Gaussian Processes (GPs) are the most call-efficient. Vanilla GP-based methods using stationary RBF kernels and maximum likelihood kernel parameter estimation provide a near-perfect ability to optimize the benchmarks. Despite being slower than more heuristic baselines, a Theano-based GP-SMBO implementation requires at most a few seconds to produce a candidate evaluation point. We compare this vanilla approach to Hybrid Monte-Carlo integration of the kernel lengthscales and fail to find compelling advantages of this more expensive procedure.

Full text links

 PDF

 External link

Supplementary information

 Poster

Conference Proceedings

Booktitle
ICML 2014 AutoML Workshop
Pages
7

Cite

Plain text

BibTeX