You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Safe Reinforcement Learning with Stability Guarantees
This code accompanies the paper [1] and implements the code for estimating the region of attraction for a policy and optimizing the policy subject to stability constraints. For the old numpy-based code to estimate the region of attraction in [2] see the lyapunov-learning repository. The code for learning Lyapunov functions from [3] can be found in the examples folder.
This library is tested based on both python 2.7 and 3.5, together with the following dependencies, since pip>=19 does not support --process-dependency-links (see below)
pip install pip==18.1
pip install numpy==1.14.5
Based on this, you can install the library by cloning the repository and installing it with
pip install . --process-dependency-links
To run the tests with the bash script in scripts/test_code.sh, you need to install additional dependencies with
pip install ".[test]" --process-dependency-links
The --process-dependency-links flag is needed to install gpflow==0.4.0, which is not on pypi. You can skip it if that particular version of the library is already installed.
You can the find example jupyter notebooks and the experiments in the paper in the examples folder.
About
Safe reinforcement learning with stability guarantees