Closed waiyip-aquabyte closed 3 years ago
Hi @waiyip-aquabyte
I don't see the issue reproducing in me setup with my aws account
It's sound's like a AWS network issue or a boto3 connection issue.
Are you still seeing it reproducing? Maybe can you see in cloudWatch if s3 network load is high?
About S3Path it's almost the same running:
>>> with p.open('w') as fp:
... json.dump([None]*5000,fp)
AND
>>> p.write_text(json.dumps([None]*5000))
I don't see way it can behave differently.
@waiyip-aquabyte are you still seeing the issue?
@waiyip-aquabyte are you still seeing the issue? Or can we close this issue?
Sorry, I need find a chance to upgrade and verify. The problem was reproducible.
We already start to talk about this issue in a deferent issue #55
@impredicative you talked about your use case
When you are using S3Path and smart_open
.
Maybe we can use smart_open
instead of optimising our current code.
What do you think about smart_open
project?
I think this is the same bug I ran into. I noticed it continuously uploading data and never finishing.
Using smart_open
might be a good idea since it is much more widely used and tested. It does magic things though like decompression of certain file types based on extension which would need to be disabled.
For writing small files, s3path
seems fine as is. For a single-step write, boto3
works. For a streaming write smart_open
is best.
In all cases I rely on s3path
to manipulate paths and generate URIs.
When it is stuck, I've seen a lot of log messages like this.
This looks like an infinite loop. But after a long time, it completed successfully.