Abstract
Ensuring safety in optimization is challenging if the underlying functional forms of either the constraints or the objective function are unknown. The challenge can be addressed by using Gaussian processes to provide confidence intervals used to find solutions that can be considered safe. To iteratively find a trade-off between finding the solution and ensuring safety, the SafeOpt algorithm builds on algorithms using only the upper bounds (UCB-type algorithms) by performing an exhaustive search on the entire search space to find a safe iterate. That approach can quickly become computationally expensive. We reformulate the exhaustive search as a series of optimization problems to find the next recommended points. We show that the proposed reformulation allows using a wide range of available optimization solvers, such as derivative-free methods. We show that by exploiting the properties of the solver, we enable the introduction of new stopping criteria into safe learning methods and increase flexibility in trading off solver accuracy and computational time. The results from a non-convex optimization problem and an application for controller tuning confirm the flexibility and the performance of the proposed reformulation. Show more
Permanent link
https://doi.org/10.3929/ethz-b-000615589Publication status
publishedEvent
Subject
Machine learning in modelling; prediction; control and automation; Learning for control; Bayesian methods; Safe learning; Derivative-free optimizationOrganisational unit
02650 - Institut für Automatik / Automatic Control Laboratory03751 - Lygeros, John / Lygeros, John
Funding
180545 - NCCR Automation (phase I) (SNF)
787845 - Optimal control at large (EC)
Related publications and datasets
Is supplemented by: https://doi.org/10.3929/ethz-b-000616522
Is new version of: https://doi.org/10.48550/arXiv.2211.14104
Notes
Preprint available: https://arxiv.org/abs/2211.14104More
Show all metadata
ETH Bibliography
yes
Altmetrics