Details zu Publikationen

Hyperparameter Optimization: Foundations, Algorithms, Best Practices and Open Challenges

verfasst von
Bernd Bischl, Martin Binder, Michel Lang, Tobias Pielok, Jakob Richter, Stefan Coors, Janek Thomas, Theresa Ullmann, Marc Becker, Anne-Laure Boulesteix, Difan Deng, Marius Lindauer
Abstract

Most machine learning algorithms are configured by one or several hyperparameters that must be carefully chosen and often considerably impact performance. To avoid a time consuming and unreproducible manual trial-and-error process to find well-performing hyperparameter configurations, various automatic hyperparameter optimization (HPO) methods, e.g., based on resampling error estimation for supervised machine learning, can be employed. After introducing HPO from a general perspective, this paper reviews important HPO methods such as grid or random search, evolutionary algorithms, Bayesian optimization, Hyperband and racing. It gives practical recommendations regarding important choices to be made when conducting HPO, including the HPO algorithms themselves, performance evaluation, how to combine HPO with ML pipelines, runtime improvements, and parallelization. This work is accompanied by an appendix that contains information on specific software packages in R and Python, as well as information and recommended hyperparameter search spaces for specific learning algorithms. We also provide notebooks that demonstrate concepts from this work as supplementary files.

Organisationseinheit(en)
Institut für Künstliche Intelligenz
Externe Organisation(en)
Ludwig-Maximilians-Universität München (LMU)
Technische Universität Dortmund
Typ
Artikel
Journal
Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery
Band
13
Anzahl der Seiten
70
ISSN
1942-4787
Publikationsdatum
10.03.2023
Publikationsstatus
Veröffentlicht
Peer-reviewed
Ja
ASJC Scopus Sachgebiete
Informatik (insg.)
Elektronische Version(en)
http://10.48550/arXiv.2107.05847 (Zugang: Offen)
https://doi.org/10.1002/widm.1484 (Zugang: Offen)