Closed haje01 closed 1 year ago
cc @michaelzhiluo can you open source your examples?
Experiencing a similar problem with an updated AsyncPPOTFPolicy. My policy behaves as expected with 'num_gpus: 1' but throws the same error at me with 'num_gpus: 2'.
I have the same problem with AsyncPPOTFPolicy. My policy behaves as expected with 'num_gpus: 1' but throws the same error at me with 'num_gpus: 2'.
Is there any known way to fix this?
I have the same problem with AsyncPPOTFPolicy. My policy behaves as expected with 'num_gpus: 1' but throws the same error at me with 'num_gpus: 2'.
Is there any known way to fix this?
I have not worked on that issue, sorry. Does the execution plan automatically create a multi GPU Learner thread if you hand it more than one GPU resource?
Hi, I'm a bot from the Ray team :)
To help human contributors to focus on more relevant issues, I will automatically add the stale label to issues that have had no activity for more than 4 months.
If there is no further activity in the 14 days, the issue will be closed!
You can always ask for help on our discussion forum or Ray's public slack channel.
System information
Hi. I am trying to find a distributed(and hopefully faster) training setting for continuous action tasks, like Humanoid or Halfcheetah. Since I couldn't find one in tuned examples, I tried it myself. My first idea was changing existing tuned APPO example(
halfcheetah-appo.yaml
) into Multi-GPU setting:But the training raised following error:
What's the problem? By the way, I wish we could have more tuned examples on distributed continuous action tasks.
Thank you.