Please click here if you are unable to view this page.

Regret-Based Defense in Adversarial Reinforcement Learning

Speaker (s): 

 

BELAIRE Roman Lok-Ming
PhD Candidate
School of Computing and Information Systems
Singapore Management University

 

 

 

 

Date:


Time:


Venue:

 

 

 

25 April 2024, Thursday


2:00pm - 2:30pm


Meeting room 4.4, Level 4
School of Computing and Information Systems 1, 
Singapore Management University, 
80 Stamford Road, Singapore 178902

Please register by 24 April 2024.
 

 

About the Talk

Deep Reinforcement Learning (DRL) policies have been shown to be vulnerable to small adversarial noise in observations. Such adversarial noise can have disastrous consequences in safety-critical environments. For instance, a self-driving car receiving adversarially perturbed sensory observations about nearby signs (e.g., a stop sign physically altered to be perceived as a speed limit sign) or objects (e.g., cars altered to be recognized as trees) can be fatal. Existing approaches for making RL algorithms robust to an observation-perturbing adversary have focused on reactive approaches that iteratively improve against adversarial examples generated at each iteration. While such approaches have been shown to provide improvements over regular RL methods, they are reactive and can fare significantly worse if certain categories of adversarial examples are not generated during training. To that end, we pursue a more proactive approach that relies on directly optimizing a well-studied robustness measure, regret instead of expected value. We provide a principled approach that minimizes maximum regret over a "neighborhood" of observations to the received "observation". Our regret criterion can be used to modify existing value- and policy-based Deep RL methods. We demonstrate that our approaches provide a significant improvement in performance across a wide variety of benchmarks against leading approaches for robust Deep RL. 

This is a Pre-Conference talk for 23rd International Conference on Autonomous Agents and Multiagent Systems (AAMAS 2024).
 

About the Speaker

Roman BELAIRE is a Ph.D. candidate at SCIS, under the supervision of Prof. Pradeep VARAKANTHAM. Roman's research focuses on reinforcement learning methods for policies that are robust to adversarial perturbation.