贝叶斯优化算法全面解析-图文

合集下载
  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

Bayesian Optimization CSC2541 - Topics in Machine Learning

Scalable and Flexible Models of Uncertainty

University of Toronto - Fall 2017

Overview

1.Bayesian Optimization of Machine Learning Algorithms

2.Gaussian Process Optimization in the Bandit Setting

3.Exploiting Structure for Bayesian Optimization

Bayesian Optimization of Machine Learning Algorithms J. Snoek, A. Krause, H. Larochelle, and R.P. Adams (2012)

Practical Bayesian Optimization of Machine Learning Algorithms

J. Snoek et al. (2015)

Scalable Bayesian Optimization Using Deep Neural Nets

Presentation by: Franco Lin, Tahmid Mehdi, Jason Li

Motivation

Performance of Machine Learning algorithms are usually dependent on the choice of hyperparameters

Picking the optimal hyperparameter values are hard

-Ex. grid search, random search, etc.

-Instead could we use a model to select which hyperparameters will be good next?

Bayes Opt. of Machine Learning Algorithms

-Bayesian Optimization uses all of the information from previous evaluations and performs some computation to determine the next point to try

-If our model takes days to train, it would be beneficial to have a well structured way of selecting the next combination of hyperparameters to try

-Bayesian Optimization is much better than a person finding a good combination of hyperparameters

Bayesian Optimization

Intuition:

We want to find the peak of our true function (eg. accuracy as a function of hyperparameters)

To find this peak, we will fit a Gaussian Process to our observed points and pick our next best point where we believe the maximum will be.

This next point is determined by an acquisition function - that trades of exploration and exploitation

Bayesian Optimization Tutorial

Bayesian Optimization Tutorial

that maximizes acquisition function Find the next best point x

n

Bayesian Optimization Tutorial

Evaluate ƒ at the new observation x

and update posterior

n

Update acquisition function from new posterior and find the next best point

Acquisition Function Intuition

-We will use the acquisition function

Probability of Improvement (PI) as

an example.

-We want to find the point with the

largest area above our best value

-This corresponds to the maximum

of our acquisition function

Acquisition Functions

-Guides the optimization by determining which point to observe next and is easier to optimize to find the next sample point

Probability of Improvement (PI)

Expected Improvement (EI)

GP-Upper/Lower Confidence Bound (GP-UCB/LCB)

相关文档
最新文档