question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

[question] Model Based RL with Stable Baselines?

See original GitHub issue

I would like to do some model based RL with stable baselines. From what I’ve read, it seems that to do this, I just have to train a model f to predict the next state, given the current state and action, i.e. s{t+1} <- f(s{t}, a{t}). I want to do this so that the RL can plan several steps into the future.

Anyways, I need access to the obs and actions from the learn method in PPO2. Basically, after each rollout, I want to use the data collected to train f in a supervised manner. I am trying to use a callback to do this. Is this possible?

Thanks!

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:11

github_iconTop GitHub Comments

1reaction
araffincommented, Jul 6, 2020

Thanks again for commenting. In case anyone wants to use the new patch, you can use !pip install git+https://github.com/hill-a/stable-baselines.git

this is actually already in the documentation 😉 https://stable-baselines.readthedocs.io/en/master/guide/install.html#bleeding-edge-version

1reaction
PartiallyTypedcommented, Jul 6, 2020

It’s not on pypi, the patch #787 was completed recently (^^’).

Read more comments on GitHub >

github_iconTop Results From Across the Web

RL Algorithms — Stable Baselines 2.10.3a0 documentation
This table displays the rl algorithms that are implemented in the stable baselines project, along with some useful characteristics: support for recurrent ...
Read more >
Reinforcement Learning with Stable Baselines 3 - YouTube
Welcome to a tutorial series covering how to do reinforcement learning with the Stable Baselines 3 (SB3) package.
Read more >
Reinforcement Learning with Stable Baselines 3 (P.3)
How to incorporate custom environments with stable baselines 3Text- based tutorial and sample code: ...
Read more >
Reinforcement Learning with Stable Baselines 3 (P.4)
Helping our reinforcement learning algorithm to learn better by tweaking the environment rewards.Text- based tutorial and sample code: ...
Read more >
RL Tutorial on Stable Baselines - Antonin Raffin
Model Free: No explicit representation of the environment · Model Based: Rely on a the model of the environment. known, given to the...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found