-
If you're unfamiliar with eligibility traces, they basically unify temporal-difference learning with monte carlo methods -- essentially you hold a buffer in memory of an agent's experience and perform…
-
![auth](https://user-images.githubusercontent.com/8428372/46090268-5279f600-c1eb-11e8-99ef-1494b9f77653.png)
![body](https://user-images.githubusercontent.com/8428372/46090269-5279f600-c1eb-11e8-95a9…
-
@berndbischl @markdumke I think we do not have that much time to wait until a perfect design, I would suggest
- I help Markus to finish the first version (at least we wrote those class designs tog…
-
hello everyone, i develop a custom RL-based scheduler for TSCH, i inspired from orchestra,alice to differentiate between EB packets and RPL packets, but the link selector i use in my case provokes que…
-
RL has been added to original ConvNetJS.
Will you be adding that too?
Any plans for LSTM?
Thanks
-
I have been working through the Agents vignette (https://cran.rstudio.com/web/packages/reinforcelearn/vignettes/agents.html) and I found a few errors.
1. In the "Value Functions" section, half way …
-
Not sure if you are interested but I have written a tutorial for building a basic agent:
https://medium.com/@skjb/building-a-basic-pysc2-agent-b109cde1477c
https://medium.com/@skjb/building-a-smar…
-
Given this code that's similar to https://github.com/thomasdeneux/param_qt/blob/master/test_ParamQt.py#L98:
```
class PosPar(pm.Parameterized):
shape = GObjectSelector('star',
…
-
Hi,
I'm trying to integrate a model-based RL algorithm to the d3rlpy library but couldn't find any documentation to know where & how to begin. Would really appreciate if you could please point me w…
-
Hi,it is me again (hhh
When I run the example like this:
```
import d3rlpy
# prepare dataset
dataset, env = d3rlpy.datasets.get_d4rl('hopper-medium-v0')
# prepare algorithm
cql = d3rlpy.alg…