-
Dear Hongzi,
I was trying to figure out the matching between the RL agent's state s(t) in the code and the input info in the paper.
**Input:** After the download of each chunk t, Pensieve’s lea…
-
- `@testing-library/react` version:
"@testing-library/jest-dom": "^5.17.0",
"@testing-library/react": "^14.0.0",
"@testing-library/user-event": "^14.0.0",
- Testing Framework and v…
-
## cli command
```bash
CUDA_VISIBLE_DEVICES=5 xtuner train internlm_chat_7b_qlora_colorist_e5.py
```
## content of `internlm_chat_7b_qlora_colorist_e5.py`
```python
# Copyright (c) OpenMMLab. …
-
Traceback (most recent call last):
File "explorer.py", line 54, in
Explorer(**args).show()
File "explorer.py", line 14, in __init__
self.env.reset()
File "/home/zy/reproduce/visual…
-
### Operating system
Ubuntu 18.04
### Bug description
When running the imitation learning baseline, I am sometimes getting a floating point exception. Unfortunately, It's not deterministic an…
-
**Question**:
According to the appendix section in the paper, you used supervised learning to train graph neural networks for the sanity check.
I presume the target (label) is the critical path wh…
-
The Cartpole example crashes when trying to run the learning agent (`epochs = 5` in the notebook).
```
ic| msg: 'Learning phase started. (1/5)'
[/usr/local/lib/python3.10/dist-packages/scipy/stat…
-
This is similar to #656. But I am making another issue since that issue is still not resolved. Also, @sguada mentioned in #702 that PPO agent can take 1-D action spaces.
I have the following actio…
-
Hey guys,
I appreciated reading your paper. However I just wanted to see your eval results for myself and running the script results in the follwoing error when I execute ```python evaluate_hotpot_q…
-
In the course of playing around with `fishing-v1`, I've come across a few peculiar things that are worthy of exploration or at least some public disclosure.
Action Space Size: In training an agent,…