Closed gordonwatts closed 6 months ago
@fengpinghu - do we know how many simultaneous request S3 should handle? Perhaps we have to limit it somehow?
Tried, but SX isn't transforming right now. Being followed up in #servicex.
Will need backoff to get this to work. So, we'll leave this as unfixed and try out #44
After fixing #20 , we were able to run with more DASK workers. But we can still trigger this error:
This is testing with workers already setup (not dynamically scaling). It occurs with:
I think what is happening is 200 workers hit S3 at exactly the same time and that causes its
slow down
message. With dynamic scaling, the nodes slowly come up, and so the S3 load is spread out a little bit.