-
Here I have put together a simple example that just checks a SHA256 preimage in the inner snark, but using primary inputs. The proof is supposed to be checked in an outer snark. It fails during proof …
-
### ❓ Question
I would think 0.99 is the most widely used default value, so I suggest using that for PPO as well.
### Checklist
- [X] I have checked that there is no similar [issue](https://github.…
-
-
When I run python3 pytorch_rl/main.py --no-vis --env-name Duckietown-small_loop-v0 --algo a2c --lr 0.0002 --max-grad-norm 0.5 --num-steps 20,there is an erro that tell me does't has the directory pyto…
-
Consider the following lines from `multibody_plant.cc`:
https://github.com/RobotLocomotion/drake/blob/59de1fd0bb25b0628b598d4a89cddde99b3e6f53/multibody/plant/multibody_plant.cc#L208-L209
`M_PPo_P…
-
Hi,
I am trying to use Arena in my research project. I have several general questions:
1) The [baseline tutorial videos ](https://sites.google.com/view/arena-unity/home/tutorials-baselines?authu…
-
Sorry to bother.
When I'm trying to export the model using `tf.saved_model` to deploy it to a production environment (tf serving in docker), there are some problems.
The exported graph format r…
-
Hi, thanks for the great for on the highway env. However, I am using the racetrack-env and encountered a few issues:
i) AttributeError: 'numpy.random._generator.Generator' object has no attribute '…
-
Hi,
Thanks for the awesome work and open source code!
The Hugging face link for the MM-GAG data is unavailable at the moment, i.e. https://huggingface.co/datasets/MVRL/MMGAG. Looking forward to …
-
### Description
The code section provided in the documentation does not run:
```python
from ray.rllib.algorithms.ppo import PPOConfig
from ray.rllib.algorithms.callbacks import MemoryTrackingC…