论文标题
MFES-HB:具有多余性质量测量的有效超频带
MFES-HB: Efficient Hyperband with Multi-Fidelity Quality Measurements
论文作者
论文摘要
超参数优化(HPO)是自动机器学习(AUTOML)的基本问题。但是,由于模型的昂贵评估成本(例如,在大型数据集中训练深度学习模型或培训模型),香草贝叶斯优化(BO)通常在计算上是不可行的。为了减轻此问题,HyperBand(HB)通过提前终止那些表现不佳的配置来使用早期停止机制来加快配置评估。这导致了两种质量测量:(1)用于早期停滞的配置的许多低保真测量值,以及(2)几乎没有进行评估而没有早期停止的配置的高保真测量。最先进的HB风格方法BOHB旨在结合BO和HB的好处。 BOHB不是在HB中随机取样配置,而是基于BO替代模型的样品配置,该配置仅以高保真测量构建。但是,高保真测量的稀缺性极大地阻碍了BO指导构造搜索的效率。在本文中,我们提出了MFES-HB,这是一种有效的超级带方法,能够同时利用高保真性和低保真测量来加速HPO任务的收敛性。设计MFES-HB并非微不足道,因为低保真测量可能是有偏见但有益的,可以指导配置搜索。因此,我们建议基于专家框架的广义产品建立一个多忠诚集合替代(MFE),该框架可以有效地整合有用的多余性测量信息。关于现实世界自动化任务的实证研究表明,MFES-HB可以在最先进的方法-BOHB上实现3.3-8.9倍的速度。
Hyperparameter optimization (HPO) is a fundamental problem in automatic machine learning (AutoML). However, due to the expensive evaluation cost of models (e.g., training deep learning models or training models on large datasets), vanilla Bayesian optimization (BO) is typically computationally infeasible. To alleviate this issue, Hyperband (HB) utilizes the early stopping mechanism to speed up configuration evaluations by terminating those badly-performing configurations in advance. This leads to two kinds of quality measurements: (1) many low-fidelity measurements for configurations that get early-stopped, and (2) few high-fidelity measurements for configurations that are evaluated without being early stopped. The state-of-the-art HB-style method, BOHB, aims to combine the benefits of both BO and HB. Instead of sampling configurations randomly in HB, BOHB samples configurations based on a BO surrogate model, which is constructed with the high-fidelity measurements only. However, the scarcity of high-fidelity measurements greatly hampers the efficiency of BO to guide the configuration search. In this paper, we present MFES-HB, an efficient Hyperband method that is capable of utilizing both the high-fidelity and low-fidelity measurements to accelerate the convergence of HPO tasks. Designing MFES-HB is not trivial as the low-fidelity measurements can be biased yet informative to guide the configuration search. Thus we propose to build a Multi- Fidelity Ensemble Surrogate (MFES) based on the generalized Product of Experts framework, which can integrate useful information from multi-fidelity measurements effectively. The empirical studies on the real-world AutoML tasks demonstrate that MFES-HB can achieve 3.3-8.9x speedups over the state-of-the-art approach - BOHB.