gym
Here are 478 public repositories matching this topic...
For a lot of exercises, there are different variations (wide grip, narrow grip, reverse, etc. etc.). It would be nice if this could be modelled and presented to the user.
We would basically only need a new many-to-many table, linking all the exercises together, most of the work here would be to go through the DB and actually group them together. Perhaps some text similarity script could help
-
Updated
Nov 28, 2020 - Python
-
Updated
Nov 21, 2020 - Python
-
Updated
Apr 15, 2020 - Python
-
Updated
Jul 25, 2020 - Python
Per this comment in #12
-
Updated
Aug 5, 2019 - Python
-
Updated
Jul 14, 2019 - Python
-
Updated
Oct 1, 2020 - Python
-
Updated
May 31, 2020 - Python
There seem to be some vulnerabilities in our code that might fail easily. I suggest adding more unit tests for the following:
- Custom agents (there's only VPG and PPO on CartPole-v0 as of now. We should preferably add more to cover discrete-offpolicy, continuous-offpolicy and continuous-onpolicy)
- Evaluation for the Bandits and Classical agents
- Testing of convergence of agents as proposed i
-
Updated
Nov 29, 2020 - Jupyter Notebook
-
Updated
Nov 30, 2020 - Python
Problem Description
Procgen Environments (https://github.com/openai/procgen) are new environments to test out the generalization ability of agents. It would be nice to include some of the games into the Open RL Benchmark (http://benchmark.cleanrl.dev/)
This is a good first issue for contributors. I think contributors can simply modify the network model slightly (https://github.com/vwxyzjn/c
-
Updated
May 8, 2020 - JavaScript
-
Updated
Sep 13, 2020 - Jupyter Notebook
-
Updated
Jul 4, 2019 - Python
-
Updated
Apr 16, 2017 - Python
-
Updated
Nov 30, 2020 - Python
-
Updated
Oct 7, 2020 - Python
-
Updated
Nov 26, 2020 - Python
-
Updated
Jul 20, 2018 - Python
-
Updated
Nov 6, 2020 - Python
-
Updated
Feb 10, 2019 - Python
-
Updated
Dec 4, 2019 - Python
Improve this page
Add a description, image, and links to the gym topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the gym topic, visit your repo's landing page and select "manage topics."
The following applies to DDPG and TD3, and possibly other models. The following libraries were installed in a virtual environment:
numpy==1.16.4
stable-baselines==2.10.0
gym==0.14.0
tensorflow==1.14.0
Episode rewards do not seem to be updated in
model.learn()beforecallback.on_step(). Depending on whichcallback.localsvariable is used, this means that: