Skip to content
#

reinforcement-learning

Here are 7,958 public repositories matching this topic...

gjoliver
gjoliver commented Apr 13, 2022

Description

There are multiple user requests of using GraphNN data (node and edge lists) as sample batches into a custom RLlib model.

https://discuss.ray.io/t/rllib-variable-length-observation-spaces-without-padding/726
https://discuss.ray.io/t/working-with-graph-neural-networks-varying-state-space/5730/2

The recommended method today is to use Repeated observation space and VariableVal

good first issue enhancement P2 rllib-models
annotated_deep_learning_paper_implementations

🧑‍🏫 50! Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, ...), gans(cyclegan, stylegan2, ...), 🎮 reinforcement learning (ppo, dqn), capsnet, distillation, ... 🧠

  • Updated Apr 15, 2022
  • Jupyter Notebook
stable-baselines
calerc
calerc commented Nov 23, 2020

The following applies to DDPG and TD3, and possibly other models. The following libraries were installed in a virtual environment:

numpy==1.16.4
stable-baselines==2.10.0
gym==0.14.0
tensorflow==1.14.0

Episode rewards do not seem to be updated in model.learn() before callback.on_step(). Depending on which callback.locals variable is used, this means that:

  • episode rewards may n
good first issue question

Improve this page

Add a description, image, and links to the reinforcement-learning topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the reinforcement-learning topic, visit your repo's landing page and select "manage topics."

Learn more