Applying reinforcement learning to robotic systems poses a number of challenging problems. A key requirement is the ability to handle continuous state and action spaces while remaining within a limited time and resource budget. Additionally, for safe operation, the system must make robust decisions under hard constraints. To address these challenges, we propose a model based approach that combines Gaussian Process regression and Receding Horizon Control. Using sparse spectrum Gaussian Processes, we extend previous work by updating the dynamics model incrementally from a stream of sensory data. This results in an agent that can learn and plan in real-time under non-linear constraints. We test our approach on a cart pole swing-up environment and demonstrate the benefits of online learning on an autonomous racing task. The environment’s dynamics are learned from limited training data and can be reused in new task instances without retraining.
Reference:
Van Niekerk, B., Damianou, A. and Rosman, B.S. 2017. Online constrained model-based reinforcement learning. Uncertainty in Artificial Intelligence, 11-15 August 2017, Sydney, Australia
Van Niekerk, B., Damianou, A., & Rosman, B. S. (2017). Online constrained model-based reinforcement learning. Association for Uncertainty in Artificial Intelligence (AUAI). http://hdl.handle.net/10204/9617
Van Niekerk, B, A Damianou, and Benjamin S Rosman. "Online constrained model-based reinforcement learning." (2017): http://hdl.handle.net/10204/9617
Van Niekerk B, Damianou A, Rosman BS, Online constrained model-based reinforcement learning; Association for Uncertainty in Artificial Intelligence (AUAI); 2017. http://hdl.handle.net/10204/9617 .