Poster
Enhancing Robustness in Deep Reinforcement Learning: A Lyapunov Exponent Approach
Rory Young · Nicolas Pugeault
Deep Reinforcement Learning agents achieve state-of-the-art performance in a wide range of simulated control tasks. However, successful applications to real-world problems remain limited. One essential limitation of deep RL is the lack of stability and robustness to noise and adversarial attacks of the learnt policies. In this paper, we investigate the robustness of deep RL policies to a single small state perturbation in deterministic continuous control tasks. We demonstrate that RL policies can be deterministically chaotic as small perturbations to the system state have a large impact on subsequent state and reward trajectories. This unstable non-linear behaviour has two consequences: First, inaccuracies in sensor readings, or adversarial attacks, can cause significant performance degradation; Second, even policies that show robust performance in terms of rewards, may have unpredictable behaviour in practice. These two facets of chaos in RL policies drastically restrict the application of deep RL to real-world problems. To address this issue, we propose an improvement on the successful Dreamer V3 architecture, implementing a Maximal Lyapunov Exponent regularisation. This new approach reduces the chaotic state dynamics, rendering the learnt policies more resilient to sensor noise and adversarial attacks and thereby improving their suitability for real-world applications.
Live content is unavailable. Log in and register to view live content