Bayesian Optimisation (AutoML)
Bayesian Optimisation (AutoML)
We will first explore the theory of Bayesian Optimization with details in the following areas:
Bayesian Optimization
Gaussian Processes
Tree-Based Approaches
Gaussian Processes
Figure 2) GP for
a single iteration
Tree-Based Approaches
Tree Parzen Estimators (TPE) [2]
These model and rather than .
Where a split is created at the specified threshold and the 2 new conditionals are
modeled by tree Parzen estimators.
Multi-Task Bayesian Optimization (MTBO) by K. Swersky et al. emerges as one technique that
introduces dataset size as an additional task to optimize [4]
Fast Bayesian Optimization on Large DataSets (Fabolas) is a method that uses a unique
acquisition function that models the dataset size and function evaluation time on small data
subsets [6]
Applications of BO
Bayesian optimization is ideal for globally optimizing
black-box functions that are expensive to evaluate. It
has been successfully been applied to:
There is large number of tools available for languages used for data science, such as Python,
MATLAB, R, and Java.
We list a few tools and frameworks implementing the mentioned algorithms for Python, a very popular
language for machine learning and Bayesian optimization research implementations. Tools also exist
for Java, such as Auto-WEKA, and for R, such as H2O. However, we stick to Python for its prevalence
in the research community
Tools, platforms and frameworks
Below are some Python packages implementing various algorithms
A sub-field of deep reinforcement learning aims to use RL to train deep neural networks to
perform complex tasks such as playing videogames
Choosing which neural network architecture, which network hyperparameters, and which RL
algorithm to use is a tough problem for humans to tackle
Deep RL itself is an approach that finds many applications in robotics, game-playing agents, and
adversarial systems.
Improving model selection and hyperparameter optimization for deep RL can create more
efficient agents, and accelerate research into certain application areas
Deep RL Tutorials and Videogame Agents
Deep RL is state-of-the-art for playing many adversarial games:
(labelled with figures alongside)
d)
a) Spinning pendulum
b) Atari games such as Breakout, Pong
c) Pole on a cart
d) DOOM (!)
a) b) c)
a
)
Experimental Details
Hardware:
16GB RAM
Software:
If you had to count our dependencies, you may run out of natural numbers to do it with.
Also, sorry Mac OS, and Windows users, this one is strictly Linux (blame dependencies)
Experimental Details Cont...
For each of the algorithms, the models to choose from were: (Below, images refers to
preprocessed images)
ReLU, Sigmoid, Linear, Tanh, Leaky ReLU activation functions for each layer, and
Choices of reinforcement learning algorithms are Q-learning, SARSA, A3C, CEM, and TD-
lambda. Models were allowed to play a total of 10 000 episodes for training and 10 test episodes
for evaluation.
This budget constraint makes good models a necessity
Demo - Show Me the Money
This is the part where we show off all the things
Conclusions
We have seen how Bayesian Optimization (BO) can be used as a technique for
automated machine learning.
We have learnt some of the theory and seen a few of the state-of-the-art techniques used
to combat dimensionality issues and large datasets.
We have seen BO applied to Deep Reinforcement Learning to play some simple and
some more complex videogames at a better-than human level of performance, while
removing the need for model configuration from the experimenter.
References
[1] Shahriari, B., Swersky, K., Wang, Z., Adams, R. P., & de Freitas, N. (2016).
Taking the human out of the loop: A review of bayesian optimization.
Proceedings of the IEEE, 104(1), 148-175
[4] Wang, Z., Zoghi, M., Hutter, F., Matheson, D., & De Freitas, N. (2013, August).
Bayesian Optimization in High Dimensions via Random Embeddings.
In IJCAI (pp. 1778-1784)
[6] Klein, A., Falkner, S., Bartels, S., Hennig, P., & Hutter, F. (2016).
Fast bayesian optimization of machine learning hyperparameters on large datasets.
arXiv preprint arXiv:1605.07079