A unified framework of Adversarial Attacks for Robust Reinforcement Learning.
- π Content of the Repository
- πΎ Installation
- π¦ Content of the Library
- π‘οΈ Adversarial Attacks
- π οΈ Usage
Use python 3.10
git clone git@github.com:IRT-SystemX/adaro-rl.git
pip install -e adaro-rladaro_rl/
βββ attacks/
βββ pipelines/
βββ wrappers/
βββ zoo/
βββ __init__.pysrc/adaro_rl/attacks/contains the implementation of many adversarial attacks in a unified framework.src/adaro_rl/pipelines/contains the implementation of scripts to train, test, attack, and adversarially train agents.src/adaro_rl/wrappers/contains the implementations of wrappers of gymnasium environments to apply attacks from different perspectives (the agent or the adversary) and with different supports for perturbation (observations or environment).src/adaro_rl/zoo/contains the implementation of tools and functionality to run the scripts.
RandomUniformAttackorRUA: generates perturbation with random uniform sampling.RandomNormalAttackorRNA: generates perturbation with random normal sampling.RandomSignedAttackorRSA: generates perturbation with random sampling between -1 and 1.
FGM_DforFastGradientMethod_DiscreteAction: generates perturbation with fast gradient method to attack discrete action policies.FGM_CforFastGradientMethod_ContinuousAction: generates perturbation with fast gradient method to attack continuous action policies.FGM_VforFastGradientMethod_V_Critic: generates perturbation with fast gradient method to attack value critics.FGM_QCforFastGradientMethod_Q_Critic: generates perturbation with fast gradient method to attack the q critic of a q actor-critic agent.FGM_QACforFastGradientMethod_Q_ActorCritic: generates perturbation with fast gradient method to attack both models of a q actor-critic agent.
FGSM_Dfor the sign version ofFGM_D.FGSM_Cfor the sign version ofFGM_C.FGSM_Vfor the sign version ofFGM_V.FGSM_QCfor the sign version ofFGM_QC.FGSM_QACfor the sign version ofFGM_QAC.
The common parameters for all attacks are the following:
epsdefines the amount of perturbation applied with an attack.normdefines the norm of the delta. Options: 0, 1, 2, ... inf.max_epsdefines the maximum amount allowed on each dimension of the perturbation. It can be a number, the same for each dimension. It can also be an array, defining a specific amount of perturbation for each dimension. By default, max_eps is set as delta.is_proportional_maskis an array of booleans that defines for each dimension, if the perturbation has to be applied proportionally to the value of the support on which the perturbation is applied. By default, is_prop_perturb_mask is None.
Other parameters for gradient attacks are the following:
targetdefines the target followed by the attack. Options: 'untargeted', 'targeted', 'min', 'max' or 'target_fct'. Default is 'untargeted'.
The first way to use the ADARO-RL library is to directly apply the attack within your infrastructure with your training and testing scripts.
from adaro_rl.attacks import make_attack
attack = make_attack(attack_name, **attack_kwargs)
perturbation = attack.generate_perturbation(observation)
adv_obs = observation + perturbation- Here, use one of the
attack_namelisted inadaro_rl.attacks.attack_names. - Add the
attack_kwargsneeded for the attack_name specified. - Use an
observationof your environment|dataset.
The second way to use the ADARO-RL library is to use the application available in adaro_rl in command line. This enables to train, test, attack, and adversarial train agents, it is needed to choose a configuration of an existing use case provided in the adaro_rl.zoo, add your configurations in the zoo, or provide a new zoo module to use.
Here are some basic command line examples :
adaro_rl train --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo'
adaro_rl test --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo'
adaro_rl online_attack --config 'HalfCheetah-v5' --zoo 'adaro_rl.zoo' --attack-name 'FGM_D' --eps 0.01 --norm "2"The different pipelines that can be executed by the application are the following:
traintestonline_attackadversarial_train
Use any given script with the option --help or check the source codes available in adaro-rl/pipelines to see how to use them.
You can also use the scripts directly in Python. Here are some examples equivalent to the ones presented above in command lines:
import adaro_rl.zoo as zoo
config = zoo.configs['HalfCheetah-v5']
adaro_rl.train(config=config)
adaro_rl.test(config=config)
adaro_rl.online_attack(config=config, attack_name='FGM_D', eps=0.1, norm=2)Example scripts and jupyter notebooks are available in examples folder of the repository. You will find there :
- Attacks examples and
examples/attacks.py: a jupyter notebook and a python script showing the usages of the attacks available. - Pipelines examples and
examples/pipelines.sh: a jupyter notebook and a bash script showing the usages of the high level pipelines available.
docs/contains the documentation the library
To build the HTML documentation:
cd docs
make htmlThe output will be in:
docs/build/html/index.html
Open this file in a browser to view the docs locally.