Skip to Main content Skip to Navigation
Journal articles

Robust Reinforcement Learning with Bayesian Optimisation and Quadrature

Abstract : Bayesian optimisation has been successfully applied to a variety of reinforcement learning problems. However, the traditional approach for learning optimal policies in simulators does not utilise the opportunity to improve learning by adjusting certain environment variables: state features that are unobservable and randomly determined by the environment in a physical setting but are controllable in a simulator. This article considers the problem of finding a robust policy while taking into account the impact of environment variables. We present alternating optimisation and quadrature (ALOQ), which uses Bayesian optimisation and Bayesian quadrature to address such settings. We also present transferable ALOQ (TALOQ), for settings where simulator inaccuracies lead to difficulty in transferring the learnt policy to the physical system. We show that our algorithms are robust to the presence of significant rare events, which may not be observable under random sampling but play a substantial role in determining the optimal policy. Experimental results across different domains show that our algorithms learn robust policies efficiently.
Complete list of metadatas

Cited literature [60 references]  Display  Hide  Download

https://hal.inria.fr/hal-02943567
Contributor : Jean-Baptiste Mouret <>
Submitted on : Saturday, September 19, 2020 - 9:25:19 PM
Last modification on : Tuesday, September 22, 2020 - 3:57:10 AM

File

18-216.pdf
Files produced by the author(s)

Identifiers

  • HAL Id : hal-02943567, version 1

Citation

Supratik Paul, Konstantinos Chatzilygeroudis, Kamil Ciosek, Jean-Baptiste Mouret, Michael Osborne, et al.. Robust Reinforcement Learning with Bayesian Optimisation and Quadrature. Journal of Machine Learning Research, Microtome Publishing, 2020, 21, pp.1 - 31. ⟨hal-02943567⟩

Share

Metrics

Record views

17

Files downloads

55