Open kekulai-fredchang opened 1 month ago
Ray cluster launcher can not recover cached-stopped-nodes when nodes fail (e.g. from overcapacity) on the azure cloud end.
cache_stopped_nodes: False
running
state, but there is a provisioning
error.cache_stopped_nodes: False
in which the state of the VM is guaranteed, do I see reliable ray operation using these problematic A100, H100 workers.cache_stopped_nodes: False
, it persistently retries to communicated with the broken worker.
What happened + What you expected to happen
When requesting resources for a ray cluster, when the actor times out (could be error in the minimal code example) it seems to leave the ray workers in a state where they don't respond to ssh.
1) ray up ray.yml 2) ray dashboard ray.yml
3) seq 6 | parallel -n0 ray job submit --entrypoint-num-gpus 1 --entrypoint-num-cpus 24 --working-dir . -- nvcc --version // this successfully loads 6 gpu workers 4) time ray job submit --runtime-env ./ray_runtime_env.yml --address http://localhost:8265 -- python test.py // this is the problematic command where the worker will timeout and return command line 5) seq 6 | parallel -n0 ray job submit --entrypoint-num-gpus 1 --entrypoint-num-cpus 24 --working-dir . -- nvcc --version // when running the following command that ran fine at step 3), instead this instance the ray head node will be stuck not able to communicate with the workers.
when looking at the monitor.out file I see that the ssh to the workers is not working. however when you attach to the head and ssh from there I can ssh into the workers. (also workers are running fine in azure console
Versions / Dependencies
ray 3.0.0.dev
Reproduction script
ray.yml.txt ray-env2.yml.txt test2.py.txt ray_runtime_env2.yml.txt
Issue Severity
High: It blocks me from completing my task.