Skip to content

PyTorch implementation of **Towards Robust Policy: Enhancing Offline Reinforcement Learning with Adversarial Attacks and Defenses**

Notifications You must be signed in to change notification settings

thanhkaist/robust_offline_rl

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

4 Commits
Β 
Β 

Repository files navigation

Robust Offline RL: Adversarial Attacks and Defenses (PyTorch Implementation)

PyTorch implementation of
Towards Robust Policy: Enhancing Offline Reinforcement Learning with Adversarial Attacks and Defenses
[Paper on arXiv]


πŸ“˜ Overview

Offline reinforcement learning (RL) enables training policies from pre-collected datasets without costly or risky online exploration. However, this paradigm is vulnerable to observation perturbations and intentional adversarial attacks, which can degrade policy robustness and real-world performance.

This project proposes a robust offline RL framework that:

  • πŸ“Œ Applies adversarial attacks on both the actor and critic during training by perturbing observations
  • πŸ›‘οΈ Incorporates adversarial defenses as regularization strategies to improve policy robustness
  • πŸ§ͺ Evaluates the framework using the D4RL benchmark

πŸš€ Features

  • βš”οΈ 4 types of adversarial attacks (actor-side, critic-side, joint, observation-space)
  • 🧠 2 adversarial defenses to mitigate the effects of attacks
  • πŸ”¬ Evaluation on D4RL tasks using standard offline RL baselines (e.g., CQL, TD3+BC)
  • πŸ”§ Plug-and-play support for custom attack/defense modules

πŸ§ͺ Installation

git clone https://github.com/thanhkaist/robust_offline_rl.git
cd robust_offline_rl
pip install -r requirements.txt

πŸ“Š Results

Extensive experiments on the D4RL benchmark show:

  • Offline RL policies are vulnerable to small adversarial perturbations
  • Proposed defenses significantly improve policy robustness across tasks
  • Attacks on the critic lead to greater performance drops than on the actor alone

πŸ“š Citation

If you find this work useful, please cite:

@inproceedings{nguyen2024towards,
  title={Towards robust policy: Enhancing offline reinforcement learning with adversarial attacks and defenses},
  author={Nguyen, Thanh and Luu, Tung M and Ton, Tri and Yoo, Chang D},
  booktitle={International Conference on Pattern Recognition and Artificial Intelligence},
  pages={310--324},
  year={2024},
  organization={Springer}
}

About

PyTorch implementation of **Towards Robust Policy: Enhancing Offline Reinforcement Learning with Adversarial Attacks and Defenses**

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published