Hyper-parameters for successful DQN Agent
See original GitHub issueHi @marload,
Great repository you have here 😄! I am running your DQN script and I am trying to solve CartPole
with it (consistently get >200
score).
I ran the script with the default parameters, but the agent is having trouble learning a successful policy. All I get is fluctuating scores between 10 and 100 for the first 800 episodes I trained it on. There was one episode with >200 but it was early in the training and having in mind that eps
would have been very high at this point I think this must have been due to chance.
So my question is - if you have trained a successful agent with this algorithm can you provide me with “working” parameters? Or maybe DQN is just unstable in nature and I should run the script a couple of more times and hope for something better?
I have not reviewed the code thoroughly, because I wanted to see it working first, but at first glance, it looks clean and simple.
Anyway, thanks for posting it on Reddit, not sure why it was deleted. I hope I can learn a thing or two from it since I am working on something similar at the moment. 😄
Have a great day!
Issue Analytics
- State:
- Created 4 years ago
- Reactions:1
- Comments:5 (3 by maintainers)
Thx 😃
Sorry for replying late! I made the DQN and DRQN work normally. In addition, the code style has been changed to a better architecture. Thank you for waiting!