Files
Gymnasium/docs/environments/third_party_environments.md

234 lines
14 KiB
Markdown
Raw Normal View History

```{eval-rst}
:tocdepth: 2
```
2022-09-13 20:27:34 +01:00
2023-03-19 18:51:09 -04:00
# Third-Party Environments
2022-09-13 20:27:34 +01:00
2023-03-19 18:51:09 -04:00
The Farama Foundation maintains a number of other [projects](https://farama.org/projects), most of which use Gymnasium. Topics include:
multi-agent RL ([PettingZoo](https://pettingzoo.farama.org/)),
offline-RL ([Minari](https://minari.farama.org/)),
gridworlds ([Minigrid](https://minigrid.farama.org/)),
robotics ([Gymnasium-Robotics](https://robotics.farama.org/)),
multi-objective RL ([MO-Gymnasium](https://mo-gymnasium.farama.org/))
many-agent RL ([MAgent2](https://magent2.farama.org/)),
3D navigation ([Miniworld](https://miniworld.farama.org/)), and many more.
2022-09-13 20:27:34 +01:00
2023-03-19 18:51:09 -04:00
*This page contains environments which are not maintained by Farama Foundation and, as such, cannot be guaranteed to function as intended.*
*If you'd like to contribute an environment, please reach out on [Discord](https://discord.gg/nHg2JRN489).*
### [highway-env: Autonomous driving and tactical decision-making tasks](https://github.com/eleurent/highway-env)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/eleurent/highway-env)]()
An environment for behavioral planning in autonomous driving, with an emphasis on high-level perception and decision rather than low-level sensing and control.
### [sumo-rl: Reinforcement Learning using SUMO traffic simulator](https://github.com/LucasAlegre/sumo-rl)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.3-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/LucasAlegre/sumo-rl)]()
Gymnasium wrapper for various environments in the SUMO traffic simulator. Supports both single and multiagent settings (using [pettingzoo](https://pettingzoo.farama.org/)).
### [panda-gym: Robotics environments using the PyBullet physics engine](https://github.com/qgallouedec/panda-gym/)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.3-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/qgallouedec/panda-gym)]()
PyBullet based simulations of a robotic arm moving objects.
### [tmrl: TrackMania 2020 through RL](https://github.com/trackmania-rl/tmrl/)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/trackmania-rl/tmrl)]()
tmrl is a distributed framework for training Deep Reinforcement Learning AIs in real-time applications. It is demonstrated on the TrackMania 2020 video game.
2023-03-25 04:41:28 -04:00
### [gym-jiminy: Training Robots in Jiminy](https://github.com/duburcqa/jiminy)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.28.0-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/duburcqa/jiminy)]()
gym-jiminy presents an extension of the initial Gym for robotics using [Jiminy](https://github.com/duburcqa/jiminy), an extremely fast and light-weight simulator for poly-articulated systems using Pinocchio for physics evaluation and Meshcat for web-based 3D rendering.
2023-03-19 18:51:09 -04:00
### [Safety-Gymnasium: Ensuring safety in real-world RL scenarios](https://github.com/PKU-MARL/safety-gymnasium)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.3-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/PKU-MARL/safety-gymnasium)]()
Highly scalable and customizable Safe Reinforcement Learning library.
2022-09-13 20:27:34 +01:00
### [stable-retro: Classic retro games, a maintained version of OpenAI Retro](https://github.com/MatPoliquin/stable-retro)
2023-03-19 18:51:09 -04:00
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/MatPoliquin/stable-retro)]()
2023-03-25 04:41:28 -04:00
Supported fork of [gym-retro](https://openai.com/research/gym-retro): turn classic video games into Gymnasium environments.
2022-09-13 20:27:34 +01:00
### [gym-cellular-automata: Cellular Automata environments](https://github.com/elbecerrasoto/gym-cellular-automata)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.28.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/elbecerrasoto/gym-cellular-automata)]()
2023-03-19 18:51:09 -04:00
Environments where the agent interacts with _Cellular Automata_ by changing its cell states.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [gym-saturation: Environments used to prove theorems](https://github.com/inpefess/gym-saturation)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/inpefess/gym-saturation)]()
An environment for guiding automated theorem provers based on saturation algorithms (e.g. [Vampire](https://github.com/vprover/vampire)).
2023-03-19 18:51:09 -04:00
### [matrix-mdp: Easily create discrete MDPs](https://github.com/Paul-543NA/matrix-mdp-gym)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.26.2-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/Paul-543NA/matrix-mdp-gym)]()
An environment to easily implement discrete MDPs as gym environments. Turn a set of matrices (`P_0(s)`, `P(s'| s, a)` and `R(s', s, a)`) into a gym environment that represents the discrete MDP ruled by these dynamics.
### [flappy-bird-gymnasium: A Flappy Bird environment for Gymnasium](https://github.com/markub3327/flappy-bird-gymnasium)
[![Gymnasium version dependency](https://img.shields.io/badge/Gymnasium-v0.27.1-blue)]()
[![GitHub stars](https://img.shields.io/github/stars/markub3327/flappy-bird-gymnasium)]()
A simple environment for single-agent reinforcement learning algorithms on a clone of [Flappy Bird](https://en.wikipedia.org/wiki/Flappy_Bird), the hugely popular arcade-style mobile game. Both state and pixel observation environments are available.
## Third-Party Environments using Gym
2023-03-19 18:51:09 -04:00
There are a large number of third-party environments using various versions of [Gym](https://github.com/openai/gym).
Many of these can be adapted to work with gymnasium (see [Compatibility with Gym](https://gymnasium.farama.org/content/gym_compatibility/)), but are not guaranteed to be fully functional.
## Video Game environments
### [gym-derk: GPU accelerated MOBA environment](https://gym.derkgame.com/)
2023-03-25 04:41:28 -04:00
A 3v3 MOBA environment where you train creatures to fight each other.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [SlimeVolleyGym: A simple environment for Slime Volleyball game](https://github.com/hardmaru/slimevolleygym)
2023-03-25 04:41:28 -04:00
A simple environment for benchmarking single and multi-agent reinforcement learning algorithms on a clone of Slime Volleyball game.
2022-09-13 20:27:34 +01:00
### [Unity ML Agents: Environments for Unity game engine](https://github.com/Unity-Technologies/ml-agents)
2023-03-25 04:41:28 -04:00
Gym (and PettingZoo) wrappers for arbitrary and premade environments with the Unity game engine.
2022-09-13 20:27:34 +01:00
### [PGE: Parallel Game Engine](https://github.com/222464/PGE)
2023-03-25 04:41:28 -04:00
Uses The [Open 3D Engine](https://www.o3de.org/) for AI simulations and can interoperate with the Gym. Uses [PyBullet](https://github.com/bulletphysics/bullet3) physics.
2022-09-13 20:27:34 +01:00
## Robotics environments
2022-09-13 20:27:34 +01:00
### [gym-pybullet-drones: Environments for quadcopter control](https://github.com/JacopoPan/gym-pybullet-drones)
2022-09-13 20:27:34 +01:00
A simple environment using [PyBullet](https://github.com/bulletphysics/bullet3) to simulate the dynamics of a [Bitcraze Crazyflie 2.x](https://www.bitcraze.io/documentation/hardware/crazyflie_2_1/crazyflie_2_1-datasheet.pdf) nanoquadrotor.
### [MarsExplorer: Environments for controlling robot on Mars](https://github.com/dimikout3/MarsExplorer)
Mars Explorer is a Gym compatible environment designed and developed as an initial endeavor to bridge the gap between powerful Deep Reinforcement Learning methodologies and the problem of exploration/coverage of an unknown terrain.
2022-09-13 20:27:34 +01:00
### [robo-gym: Real-world and simulation robotics](https://github.com/jr-robotics/robo-gym)
Robo-gym provides a collection of reinforcement learning environments involving robotic tasks applicable in both simulation and real-world robotics.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [Offworld-gym: Control real robots remotely for free](https://github.com/offworld-projects/offworld-gym)
2023-03-25 04:41:28 -04:00
Gym environments that let you control real robots in a laboratory via the internet.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [safe-control-gym: Evaluate safety of RL algorithms](https://github.com/utiasDSL/safe-control-gym)
2023-03-25 04:41:28 -04:00
Evaluate safety, robustness and generalization via PyBullet based CartPole and Quadrotor environments—with [CasADi](https://web.casadi.org) (symbolic) *a priori* dynamics and constraints.
2022-09-13 20:27:34 +01:00
### [gym-softrobot: Soft-robotics environments](https://github.com/skim0119/gym-softrobot/)
A large-scale benchmark for co-optimizing the design and control of soft robots.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [iGibson: Photorealistic and interactive robotics environments](https://github.com/StanfordVL/iGibson/)
2023-03-25 04:41:28 -04:00
A simulation environment with high-quality realistic scenes, with interactive physics using [PyBullet](https://github.com/bulletphysics/bullet3).
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [DexterousHands: Dual dexterous hand manipulation tasks](https://github.com/PKU-MARL/DexterousHands/)
2023-03-25 04:41:28 -04:00
This is a library that provides dual dexterous hand manipulation tasks through Isaac Gym.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [OmniIsaacGymEnvs: Gym environments for NVIDIA Omniverse Isaac ](https://github.com/NVIDIA-Omniverse/OmniIsaacGymEnvs/)
2023-03-25 04:41:28 -04:00
Reinforcement Learning Environments for [Omniverse Isaac simulator](https://docs.omniverse.nvidia.com/app_isaacsim/app_isaacsim/overview.html).
2022-09-13 20:27:34 +01:00
## Autonomous Driving environments
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [gym-duckietown: Lane-following for self-driving cars](https://github.com/duckietown/gym-duckietown)
2022-11-03 01:35:43 -04:00
A lane-following simulator built for the [Duckietown](http://duckietown.org/) project (small-scale self-driving car course).
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [gym-electric-motor: Gym environments for electric motor simulations](https://github.com/upb-lea/gym-electric-motor)
2023-03-25 04:41:28 -04:00
An environment for simulating a wide variety of electric drives taking into account different types of electric motors and converters.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [CommonRoad-RL: Motion planning for traffic scenarios ](https://commonroad.in.tum.de/tools/commonroad-rl)
2022-11-03 01:35:43 -04:00
A Gym for solving motion planning problems for various traffic scenarios compatible with [CommonRoad benchmarks](https://commonroad.in.tum.de/scenarios), which provides configurable rewards, action spaces, and observation spaces.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [racing_dreamer: Latent imagination in autonomous racing](https://github.com/CPS-TUWien/racing_dreamer/)
2023-03-25 04:41:28 -04:00
Train a model-based RL agent in simulation and, without finetuning, transfer it to small-scale race cars.
2022-09-13 20:27:34 +01:00
### [l2r: Multimodal control environment where agents learn how to race](https://github.com/learn-to-race/l2r/)
2022-11-03 01:35:43 -04:00
An open-source reinforcement learning environment for autonomous racing.
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
### [racecar_gym: Miniature racecar env using PyBullet](https://github.com/axelbr/racecar_gym/)
2023-03-25 04:41:28 -04:00
A gym environment for a miniature racecar using the [PyBullet](https://github.com/bulletphysics/bullet3) physics engine.
2022-09-13 20:27:34 +01:00
## Other environments
2022-09-13 20:27:34 +01:00
### [CompilerGym: Optimise compiler tasks](https://github.com/facebookresearch/CompilerGym)
2022-11-03 01:35:43 -04:00
Reinforcement learning environments for compiler optimization tasks, such as LLVM phase ordering, GCC flag tuning, and CUDA loop nest code generation.
2022-09-13 20:27:34 +01:00
### [CARL: context adaptive RL](https://github.com/automl/CARL)
Configurable reinforcement learning environments for testing generalization, e.g. CartPole with variable pole lengths or Brax robots with different ground frictions.
2022-09-13 20:27:34 +01:00
### [gym-sokoban: 2D Transportation Puzzles](https://github.com/mpSchrader/gym-sokoban)
2023-03-25 04:41:28 -04:00
The environment consists of transportation puzzles in which the player's goal is to push all boxes to the warehouse's storage locations.
2022-09-13 20:27:34 +01:00
### [DACBench: Benchmark Library for Dynamic Algorithm configuration](https://github.com/automl/DACBench)
2023-03-25 04:41:28 -04:00
Environments for hyperparameter configuration using RL. Includes cheap surrogate benchmarks as well as real-world algorithms.
2022-09-13 20:27:34 +01:00
### [NLPGym: A toolkit to develop RL agents to solve NLP tasks](https://github.com/rajcscw/nlp-gym)
2023-03-25 04:41:28 -04:00
[NLPGym](https://arxiv.org/pdf/2011.08272v1.pdf) provides interactive environments for standard NLP tasks such as sequence tagging, question answering, and sequence classification.
2022-09-13 20:27:34 +01:00
### [ShinRL: Environments for evaluating RL algorithms](https://github.com/omron-sinicx/ShinRL/)
2022-11-03 01:35:43 -04:00
ShinRL: A Library for Evaluating RL Algorithms from Theoretical and Practical Perspectives (Deep RL Workshop 2021)
2022-09-13 20:27:34 +01:00
### [gymnax: Hardware Accelerated RL Environments](https://github.com/RobertTLange/gymnax/)
RL Environments in JAX which allows for highly vectorised environments with support for a number of environments, Gym, MinAtari, bsuite and more.
2022-09-13 20:27:34 +01:00
### [gym-anytrading: Financial trading environments for FOREX and STOCKS](https://github.com/AminHP/gym-anytrading)
AnyTrading is a collection of Gym environments for reinforcement learning-based trading algorithms with a great focus on simplicity, flexibility, and comprehensiveness.
2022-09-13 20:27:34 +01:00
### [gym-mtsim: Financial trading for MetaTrader 5 platform](https://github.com/AminHP/gym-mtsim)
2023-03-25 04:41:28 -04:00
MtSim is a simulator for the [MetaTrader 5](https://www.metatrader5.com/) trading platform for reinforcement learning-based trading algorithms.
2022-09-13 20:27:34 +01:00
### [openmodelica-microgrid-gym: Environments for controlling power electronic converters in microgrids](https://github.com/upb-lea/openmodelica-microgrid-gym)
The OpenModelica Microgrid Gym (OMG) package is a software toolbox for the simulation and control optimization of microgrids based on energy conversion by power electronic converters.
2022-09-13 20:27:34 +01:00
### [mobile-env: Environments for coordination of wireless mobile networks](https://github.com/stefanbschneider/mobile-env/)
An open, minimalist Gym environment for autonomous coordination in wireless mobile networks.
2022-09-13 20:27:34 +01:00
### [GymFC: A flight control tuning and training framework](https://github.com/wil3/gymfc/)
2022-09-13 20:27:34 +01:00
2023-03-25 04:41:28 -04:00
GymFC is a modular framework for synthesizing neuro-flight controllers. Has been used to generate policies for the world's first open-source neural network flight control firmware [Neuroflight](https://github.com/wil3/neuroflight).