|
Regret-Based Defense in Adversarial Reinforcement Learning Speaker (s):  BELAIRE Roman Lok-Ming PhD Candidate School of Computing and Information Systems Singapore Management University | Date: Time:
Venue:
| | 25 April 2024, Thursday 2:00pm - 2:30pm
Meeting room 4.4, Level 4 School of Computing and Information Systems 1, Singapore Management University, 80 Stamford Road, Singapore 178902
Please register by 24 April 2024.

|
|
About the Talk Deep Reinforcement Learning (DRL) policies have been shown to be vulnerable to small adversarial noise in observations. Such adversarial noise can have disastrous consequences in safety-critical environments. For instance, a self-driving car receiving adversarially perturbed sensory observations about nearby signs (e.g., a stop sign physically altered to be perceived as a speed limit sign) or objects (e.g., cars altered to be recognized as trees) can be fatal. Existing approaches for making RL algorithms robust to an observation-perturbing adversary have focused on reactive approaches that iteratively improve against adversarial examples generated at each iteration. While such approaches have been shown to provide improvements over regular RL methods, they are reactive and can fare significantly worse if certain categories of adversarial examples are not generated during training. To that end, we pursue a more proactive approach that relies on directly optimizing a well-studied robustness measure, regret instead of expected value. We provide a principled approach that minimizes maximum regret over a "neighborhood" of observations to the received "observation". Our regret criterion can be used to modify existing value- and policy-based Deep RL methods. We demonstrate that our approaches provide a significant improvement in performance across a wide variety of benchmarks against leading approaches for robust Deep RL.
This is a Pre-Conference talk for 23rd International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2024). About the Speaker Roman BELAIRE is a Ph.D. candidate at SCIS, under the supervision of Prof. Pradeep VARAKANTHAM. Roman's research focuses on reinforcement learning methods for policies that are robust to adversarial perturbation.
|