question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

[rllib] MADDPG no exploration options applied?

See original GitHub issue

Is no exploration (parameter noise, ou exploration) possible/applied to chosen action values as in DDPG in the current version of contrib/MADDPG?

Also, would saving the trained model and restoring it be possible, to subsequently use compute_action methods to query the trained MADDPG model?

Issue Analytics

  • State:closed
  • Created 4 years ago
  • Comments:12 (6 by maintainers)

github_iconTop GitHub Comments

3reactions
jkterry1commented, Feb 12, 2020

Hey guys, my schedule just freed up. I’ll fix the action space issue and add exploration myself in the next few days.

On Wed, Feb 12, 2020 at 2:27 PM Justin Terry notifications@github.com wrote:

The specifics of the bug in terms of action spaces is still unclear (dealing with it is on my list), but I don’t think it should keep it from learning at all?

— You are receiving this because you are subscribed to this thread.

Reply to this email directly, view it on GitHub https://github.com/ray-project/ray/issues/7069?email_source=notifications&email_token=AEUF33EQXNNVSUIDGODPBK3RCRESVA5CNFSM4KQVYXE2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOELSCDAQ#issuecomment-585376130, or unsubscribe https://github.com/notifications/unsubscribe-auth/AEUF33B3YVUHNSCYCSC52IDRCRESVANCNFSM4KQVYXEQ .

– Thank you for your time, Justin Terry

0reactions
stale[bot]commented, Nov 27, 2020

Hi again! The issue will be closed because there has been no more activity in the 14 days since the last message.

Please feel free to reopen or open a new issue if you’d still like it to be addressed.

Again, you can always ask for help on our discussion forum or Ray’s public slack channel.

Thanks again for opening the issue!

Read more comments on GitHub >

github_iconTop Results From Across the Web

Algorithms — Ray 2.2.0 - the Ray documentation
The challenge in applying existing Q-learning algorithms to offline RL lies in the overestimation of the Q-function, as well as, the lack of...
Read more >
ray.rllib.algorithms.maddpg.maddpg — Ray 2.2.0
"""Contributed port of MADDPG from OpenAI baselines. ... Apply a state preprocessor with spec given by the "model" config option (like other RL...
Read more >
Algorithms — Ray 1.13.0
Contextual bandit algorithms typically consist of an action-value model (Q model) and an exploration strategy (epsilon-greedy, UCB, Thompson Sampling etc.).
Read more >
RLlib Algorithms — Ray 0.7.5 documentation
"vtrace": False, # == These two options only apply if vtrace: False == # If true, use the Generalized Advantage Estimator (GAE) #...
Read more >
RLlib Algorithms — Ray 0.8.7 documentation
No. No. LinUCB, LinTS. torch. Yes +parametric. No. Yes. MADDPG ... "vtrace": False, # == These two options only apply if vtrace: False...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found