You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Modular Reinforcement Learning (RL) library (implemented in PyTorch, JAX, and NVIDIA Warp) with support for Gymnasium/Gym, NVIDIA Isaac Lab, Brax and other environments
skrl is an open-source modular library for Reinforcement Learning written in Python (implemented in PyTorch, JAX and NVIDIA Warp) and designed with a focus on modularity, readability, simplicity, and transparency of algorithm implementation. In addition to supporting the OpenAI Gym, Farama Gymnasium and PettingZoo, Google DeepMind and Brax, ManiSkill, among other environment interfaces, it allows loading and configuring NVIDIA Isaac Lab environments, enabling agents' simultaneous training by scopes (subsets of environments among all available environments), which may or may not share resources, in the same run.
Please, visit the documentation for usage details and examples
Note: This project is under active continuous development. Please make sure you always have the latest version. Visit the develop branch or its documentation to access the latest updates to be released.
Citing this library
To cite this library in publications, please use the following reference:
@article{serrano2023skrl,
author = {Antonio Serrano-Muñoz and Dimitrios Chrysostomou and Simon Bøgh and Nestor Arana-Arexolaleiba},
title = {skrl: Modular and Flexible Library for Reinforcement Learning},
journal = {Journal of Machine Learning Research},
year = {2023},
volume = {24},
number = {254},
pages = {1--9},
url = {https://jmlr.org/papers/v24/23-0112.html}
}
About
Modular Reinforcement Learning (RL) library (implemented in PyTorch, JAX, and NVIDIA Warp) with support for Gymnasium/Gym, NVIDIA Isaac Lab, Brax and other environments