Publication | Closed Access
Random search for hyper-parameter optimization
7.9K
Citations
22
References
2012
Year
Unknown Venue
Grid search and manual search are the most widely used strategies for hyper-parameter optimization. This paper shows empirically and theoretically that randomly chosen trials are more efficient for hyper-parameter optimization than trials on a grid. Empirical evidence comes from a comparison with a large previous study that used grid search and manual search to configure neural networks and deep belief networks. Compared with neural networks configured by a pure grid search, we find that random search over the same domain is able to find models that are as good or better within a small fraction of the computation time. Granting random search the same computational budget, random search finds better models by effectively searching a larger, less promising configuration space. Compared with deep belief networks configured by a thoughtful combination of manual search and grid search, purely random search over the same 32-dimensional configuration space found statistically equal performance on four of seven data sets, and superior performance on one of seven. A Gaussian process analysis of the function from hyper-parameters to validation set performance reveals that for most data sets only a few of the hyper-parameters really matter, but that different hyper-parameters are important on different data sets. This phenomenon makes
| Year | Citations | |
|---|---|---|
1998 | 56.5K | |
1983 | 44K | |
2011 | 41.1K | |
1965 | 28.5K | |
1994 | 18.7K | |
2006 | 16.2K | |
2010 | 12.6K | |
2000 | 7.5K | |
2008 | 7.2K | |
2003 | 2.5K |
Page 1
Page 1