Open brucehoff opened 5 years ago
It would be useful to have a page with broad/general guidance when collaborating on large file transfer. If someone at Sage wants to move 100TB to (or from) Synapse, what do they do? How do they find out if their collaborator has enough bandwidth to transfer the data? How do they figure out what the best tool is to use? How long will it take and how much would it cost (if anything)? What steps are done by the Sage scientist and what steps are done by the collaborator?
Good suggestion @brucehoff. @ychae wrote a nice doc of one way we've done this:
https://github.com/Sage-Bionetworks/AD-DCC/blob/master/large_data_uploads.md
Also, to partially answer the question (and ask if it's a tool you've used elsewhere) I've used this site to test AWS connection before (linked for S3, but they have lots of other tests available):
http://cloudharmony.com/speedtest-for-aws:s3-us-east-1-and-aws:s3-us-east-2
File transfer is a core feature of Synapse and there's a lot of engineering under the hood to make it efficient. It may be worth adding a page that discusses some of the features of file transfer: