Inverse Reinforcement Learning
See original GitHub issueHi,
Firstly, thanks for putting together such an awesome project!
I’ve been playing around with the singleagent.py
problems recently and was wondering if there is any way to incorporate demonstration learning / inverse reinforcement learning into these tasks. Instead of hard-coding the reward functions in _computeReward
, I wanted to try learning a reward function from expert demonstrations. (See: Let’s do IRL). I’d appreciate if you could give me some pointers on how to modify the _computeReward
and the learning process to include an IRL_reward
.
Cheers
Issue Analytics
- State:
- Created 2 years ago
- Comments:10 (5 by maintainers)
Top Results From Across the Web
Inverse Reinforcement Learning - People @EECS
Inverse Reinforcement Learning ... Apprenticeship learning/Imitation learning through ... E.g., support vector machine, neural network, decision tree,.
Read more >Learning from humans: what is inverse reinforcement learning?
Inverse reinforcement learning (IRL), as described by Andrew Ng and Stuart Russell in 2000, flips the problem and instead attempts to extract ...
Read more >[1806.06877] A Survey of Inverse Reinforcement Learning
Abstract: Inverse reinforcement learning (IRL) is the problem of inferring the reward function of an agent, given its policy or observed ...
Read more >What is Inverse Reinforcement Learning? - Analytics Steps
Inverse Reinforcement Learning (IRL), as described by Andrew Ng and Stuart Russell in 2000, reverses the problem and instead tries to extract ...
Read more >Algorithms for Inverse Reinforcement Learning - Stanford AI Lab
The inverse reinforcement learning (IRL) problem can be characterized informally as follows (Russell, 1998):. Given 1) measurements of an agent's behavior ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I fixed it by adding a scaling factor for the thrust and the torques after realising that their scales and clamps are completely different to what
DynAviary
expects.I see, I think it can make sense. If you run a few tests and eventually want to create a PR for your modified
DSLPIDControlDyn
(I’d recommend to subclass fromDSLPIDControl
for as much as possible), contributions are always welcome!