HyperQ-Opt: Q-learning for Hyperparameter Optimization
Hyperparameter optimization (HPO) is critical for enhancing the performance of machine learning models, yet it often involves a computationally intensive search across a large parameter space. Traditional approaches such as Grid Search and Random Search suffer from inefficiency and limited scalabili...
Gespeichert in:
1. Verfasser: | |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Hyperparameter optimization (HPO) is critical for enhancing the performance
of machine learning models, yet it often involves a computationally intensive
search across a large parameter space. Traditional approaches such as Grid
Search and Random Search suffer from inefficiency and limited scalability,
while surrogate models like Sequential Model-based Bayesian Optimization (SMBO)
rely heavily on heuristic predictions that can lead to suboptimal results. This
paper presents a novel perspective on HPO by formulating it as a sequential
decision-making problem and leveraging Q-learning, a reinforcement learning
technique, to optimize hyperparameters. The study explores the works of H.S.
Jomaa et al. and Qi et al., which model HPO as a Markov Decision Process (MDP)
and utilize Q-learning to iteratively refine hyperparameter settings. The
approaches are evaluated for their ability to find optimal or near-optimal
configurations within a limited number of trials, demonstrating the potential
of reinforcement learning to outperform conventional methods. Additionally,
this paper identifies research gaps in existing formulations, including the
limitations of discrete search spaces and reliance on heuristic policies, and
suggests avenues for future exploration. By shifting the paradigm toward
policy-based optimization, this work contributes to advancing HPO methods for
scalable and efficient machine learning applications. |
---|---|
DOI: | 10.48550/arxiv.2412.17765 |