Safe Reinforcement Learning (Safe RL) typically assumes a fixed safety bound or cost threshold, constraining the agent’s behavior during training and evaluation. However, in many real-world applications (e.g., robotics, autonomous driving, or healthcare), safety requirements are not static. Regulatory limits, environmental constraints, or critical tolerances may shift over time, sometimes becoming stricter, other times relaxing.
This project will introduce a continual learning setting for Safe RL, where the safety bound (cost threshold) changes periodically or in unknown patterns across tasks. The agent must continually adapt to stricter or looser safety requirements while retaining competence across all previously encountered bounds. This presents a dual challenge: avoiding catastrophic forgetting (as in continual learning) and satisfying evolving safety constraints (as in Safe RL).
Tristan Tomilin, Meng Fang, Yudi Zhang, and Mykola Pechenizkiy. COOM: A Game Benchmark for Continual Reinforcement Learning. NeurIPS 2023
Tristan Tomilin, Meng Fang, and Mykola Pechenizkiy. HASARD: A Benchmark for Vision-Based Safe Reinforcement Learning in Embodied Agents. ICLR 2025.
Javier Garcıa and Fernando Fernández. A comprehensive survey on safe reinforcement learning. Journal of Machine Learning Research 16.1 (2015): 1437-1480.
Khetarpal, Khimya, et al. Towards continual reinforcement learning: A review and perspectives. Journal of Artificial Intelligence Research 75 (2022): 1401-1476.
Tristan Tomilin
Thiago Simão