Foundations and Trends® in Machine Learning > Vol 18 > Issue 6

Hyperparameter Optimization in Machine Learning

By Luca Franceschi, Amazon Web Services, USA | Michele Donini, Helsing, Germany | Valerio Perrone, Amazon Web Services, USA | Aaron Klein, Leipzig University, Germany | Cédric Archambeau, Helsing, Germany | Matthias Seeger, Amazon Web Services, USA | Massimiliano Pontil, Istituto Italiano di Tecnologia, Italy and University College London, UK | Paolo Frasconi, Università di Firenze, Italy, paolo.frasconi@pm.me

 
Suggested Citation
Luca Franceschi, Michele Donini, Valerio Perrone, Aaron Klein, Cédric Archambeau, Matthias Seeger, Massimiliano Pontil and Paolo Frasconi (2025), "Hyperparameter Optimization in Machine Learning", Foundations and Trends® in Machine Learning: Vol. 18: No. 6, pp 1054-1201. http://dx.doi.org/10.1561/2200000088

Publication Date: 02 Oct 2025
© 2025 L. Franceschi et al.
 
Subjects
Model choice,  Bayesian learning,  Classification and prediction,  Deep learning,  Evaluation
 

Free Preview:

Download extract

Share

Download article
In this article:
1. Introduction
2. Hyperparameter Optimization
3. Elementary Algorithms
4. Model-based Methods
5. Multi-fidelity Methods
6. Population-based Methods
7. Gradient-based Optimization
8. Further Topics
9. Hyperparameter Optimization Systems
10. Research Directions
Acknowledgements
References

Abstract

Hyperparameters are configuration variables controlling the behavior of machine learning algorithms. They are ubiquitous in machine learning and artificial intelligence and the choice of their values determines the effectiveness of systems based on these technologies. Manual hyperparameter search is often time-consuming and becomes infeasible when the number of hyperparameters is large. Automating the search is an important step towards advancing, streamlining, and systematizing machine learning, freeing researchers and practitioners alike from the burden of finding a good set of hyperparameters by trial and error. In this survey, we present a unified treatment of hyperparameter optimization, providing the reader with examples, insights into the state-of-the-art, and numerous links to further reading. We cover the main families of techniques to automate hyperparameter search, often referred to as hyperparameter optimization or tuning, including random and quasi-random search, bandit-, model-, population-, and gradient-based approaches. We further discuss extensions, including online, constrained, and multi-objective formulations, touch upon connections with other fields, such as meta-learning and neural architecture search, and conclude with open questions and future research directions.

DOI:10.1561/2200000088
ISBN: 978-1-63828-604-2
162 pp. $99.00
Buy book (pb)
 
ISBN: 978-1-63828-605-9
162 pp. $160.00
Buy E-book (.pdf)
Table of contents:
1. Introduction
2. Hyperparameter Optimization
3. Elementary Algorithms
4. Model-based Methods
5. Multi-fidelity Methods
6. Population-based Methods
7. Gradient-based Optimization
8. Further Topics
9. Hyperparameter Optimization Systems
10. Research Directions
Acknowledgements
References

Hyperparameter Optimization in Machine Learning

Hyperparameters are configuration variables controlling the behavior of machine learning algorithms. They are ubiquitous in machine learning and artificial intelligence and the choice of their values determines the effectiveness of systems based on these technologies. Manual hyperparameter search is often time-consuming and becomes infeasible when the number of hyperparameters is large. Automating the search is an important step towards advancing, streamlining, and systematizing machine learning, freeing researchers and practitioners alike from the burden of finding a good set of hyperparameters by trial and error.

In this monograph, a unified treatment of hyperparameter optimization is presented, providing the reader with examples, insights into the state-of-the-art, and numerous links to further reading. The main families of techniques to automate hyperparameter search are covered, often referred to as hyperparameter optimization or tuning, including random and quasi-random search, bandit-, model-, population-, and gradient-based approaches. Also discussed are extensions, including online, constrained, and multi-objective formulations, and connections with other fields are identified, such as meta-learning and neural architecture search. The monograph concludes with open questions and future research directions.

 
MAL-088