To address this problem, we propose a two new methods to reduce the interventions of our safety shield, action replacement and projection. At each RL step, we check if the selected action is likely to cause a safety intervention and alter the action if that is the case. Both methods cannot fully prevent all safety interventions as this would result in a very restrictive robot behavior. However, our experiments show that the action projection method reduces the number of interventions to almost zero in two different robotic domains. The safety reduction does not come with any loss of reward performance.
Our real-world evaluations on a six degree-of-freedom manipulator showed strong improvements in the behavior of the safe RL agent when using our proposed projection method, as you can see in the video on top.
For more a detailed description of our proposed methods, see our pre-print:
- IROS 2023: Reducing Safety Interventions in Provably Safe Reinforcement Learning [IEEExplore, Arxiv]