-
### Describe the bug
Here: https://isaac-orbit.github.io/orbit/source/setup/sample.html#reinforcement-learning
In this example:
```bash
./orbit.sh -p source/standalone/workflows/rsl_rl/play.py…
-
As noted in SasView/sasview#912 the raspberry model needs to be doing volume calculation outside Iq function to get polydispersity working properly.
Migrated from http://trac.sasview.org/ticket/845
`…
-
- [ ] [DeepSeek-V2: A Strong, Economical, and Efficient MoE LLM of 236B total parameters](https://github.com/deepseek-ai/DeepSeek-V2)
# DeepSeek-V2: A Strong, Economical, and Efficient MoE LLM of 2…
-
Hello,
is it possible to return the differential of the step reward function (with respect to the action) at least for the simplest envs like pendulum, cartple ?
Best, Jacek
dzako updated
7 months ago
-
Hi, i want to plan a robot which is holding and object on the end effector. Thus, I only want to take into account the collision of the object but still plan for the end effector position of the robot…
-
Link: [OpenReview](https://openreview.net/forum?id=S1xCPJHtDB)
Code: http://bit.ly/2wjgn1a
Published on 26 Sep 2019 ICLR 2020
Comment from OpenReview:
> Decision: Accept (Spotlight)
> Comm…
-
Please add some comments to the definition of the reinforcement learning models to understand, why they were designed like this.
Checklist
- [X] Modify `rl_gym_environments.py` ✓ https://github.co…
-
I've looked into the available documentation and examples, but haven't been able to figure out if it is possible to use the ML.NET in its current state for (non-deep) reinforcement learning. If it is …
-
### 🚀 Feature
Can a recent BBF algorithm be added to the library- https://github.com/google-research/google-research/tree/master/bigger_better_faster?
### Motivation
It is model free single agent R…
-
### Is your proposal related to a problem?
We would like to have the ability to ask "follow-up questions" based on the answers to previous questions in the booking request. For example, we could h…