Abstract

Limbo (LIbrary for Model-Based Optimization) is an open-source C++11 library for Gaussian Processes and data-efficient optimization (e.g., Bayesian optimization, see (Shahriari et al. 2016)) that is designed to be both highly flexible and very fast. It can be used as a state-of-the-art optimization library or to experiment with novel algorithms with plugin components. Limbo is currently mostly used for data-efficient policy search in robot learning (Lizotte et al. 2007) and online adaptation because computation time matters when using the low-power embedded computers of robots. For example, Limbo was the key library to develop a new algorithm that allows a legged robot to learn a new gait after a mechanical damage in about 10-15 trials (2 minutes) (Cully et al. 2015), and a 4-DOF manipulator to learn neural networks policies for goal reaching in about 5 trials (Chatzilygeroudis et al. 2017). The implementation of Limbo follows a policy-based design (Alexandrescu 2001) that leverages C++ templates: this allows it to be highly flexible without the cost induced by classic object-oriented designs (Driesen and Holzle 1996) (cost of virtual functions). The regression benchmarks 1 show that the query time of Limbo's Gaussian processes is several orders of magnitude better than the one of GPy (a state-of-the-art Python library for Gaussian processes 2) for a similar accuracy (the learning time highly depends on the optimization algorithm chosen to optimize the hyper-parameters). The black-box optimization benchmarks 3 demonstrate that Limbo is about 2 times faster than BayesOpt (a C++ library for data-efficient optimization, (Martinez-Cantin 2014)) for a similar accuracy and data-efficiency. In practice, changing one of the components of the algorithms in Limbo (e.g., changing the acquisition function) usually requires changing only a template definition in the source code. This design allows users to rapidly experiment and test new ideas while keeping the software as fast as specialized code. Limbo takes advantage of multi-core architectures to parallelize the internal optimization processes (optimization of the acquisition function, optimization of the hyper-parameters of a Gaussian process) and it vectorizes many of the linear algebra operations (via the Eigen 3 library 4 and optional bindings to Intel's MKL).

Highlights

  • A state-of-the-art optimization library or to experiment with novel algorithms with “plugin” components

  • The regression benchmarks1 show that the query time of Limbo’s Gaussian processes is several orders of magnitude better than the one of GPy for a similar accuracy

  • Limbo takes advantage of multi-core architectures to parallelize the internal optimization processes and it vectorizes many of the linear algebra operations

Read more

Summary

Introduction

A state-of-the-art optimization library or to experiment with novel algorithms with “plugin” components. Limbo is currently mostly used for data-efficient policy search in robot learning (Lizotte et al 2007) and online adaptation because computation time matters when using the low-power embedded computers of robots. Limbo was the key library to develop a new algorithm that allows a legged robot to learn a new gait after a mechanical damage in about 10-15 trials (2 minutes) (Cully et al 2015), and a

Results
Conclusion
Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call