baselines
Here are 39 public repositories matching this topic...
I'll post it as a question as I am not quite sure that it is a bug. I have been experimenting for a while with the library in a custom environment for a school project and I am really interested in the reproducibility of the result. I have read the disclaimer in the documentation that reads that reproducible results are not guaranteed across multiple platforms or different versions of Pytorch. Ho
-
Updated
Jun 24, 2021
I noticed it is quite tricky at the moment to generate a benchmark with an unbalanced number of examples for each step.
It would be nice to have an option in ni_scenario, nc_scenario and similar to set the number of examples for each step.
-
Updated
Apr 5, 2021 - Python
-
Updated
Apr 6, 2021 - Python
-
Updated
Mar 10, 2020 - Python
-
Updated
Dec 26, 2018 - Python
-
Updated
Dec 14, 2018
-
Updated
Jul 1, 2020 - C++
-
Updated
Aug 19, 2021 - Python
-
Updated
Jun 3, 2019 - Python
-
Updated
Apr 23, 2018 - Python
-
Updated
Jul 15, 2020 - Python
-
Updated
Jul 17, 2019 - HTML
-
Updated
Jan 30, 2021 - Python
-
Updated
Aug 22, 2021 - Python
-
Updated
May 11, 2021
-
Updated
Jun 11, 2021 - TypeScript
-
Updated
Jul 22, 2020
-
Updated
Mar 1, 2020 - Python
-
Updated
Jul 14, 2020 - Python
-
Updated
May 20, 2021 - Jupyter Notebook
-
Updated
Nov 10, 2019 - Python
-
Updated
Jan 10, 2019 - C++
-
Updated
Aug 23, 2021 - Jupyter Notebook
-
Updated
Mar 29, 2021 - Java
-
Updated
Jun 8, 2021 - Python
-
Updated
Sep 29, 2020 - Python
Improve this page
Add a description, image, and links to the baselines topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the baselines topic, visit your repo's landing page and select "manage topics."
The following applies to DDPG and TD3, and possibly other models. The following libraries were installed in a virtual environment:
numpy==1.16.4
stable-baselines==2.10.0
gym==0.14.0
tensorflow==1.14.0
Episode rewards do not seem to be updated in
model.learn()beforecallback.on_step(). Depending on whichcallback.localsvariable is used, this means that: