fatiando / rockhound

NOTICE: This library is no longer being developed. Use Ensaio instead (https://www.fatiando.org/ensaio). -- Download geophysical models/datasets and load them in Python
BSD 3-Clause "New" or "Revised" License
34 stars 15 forks source link

Move Fatiando sample data to Rockhound #84

Closed santisoler closed 2 years ago

santisoler commented 4 years ago

Add functions to fetch sample datasets from Harmonica and Verde. Add South Africa gravity and British aeromagnetic anomalies datasets. Add gallery examples for new datasets. Set dask, rasterio and netcdf4 as optional dependencies.

After last Fatiando Community Call we decided to move all sample data from Harmonica and Verde to Rockhound.

Reminders:

Reminders for Maintainers:

santisoler commented 4 years ago

@leouieda I think this is ready. I just need to upload the gallery files, but I'll do it just before merging so the review doesn't have too many files.

Remember that the CIs are just running style checks and availability tests. Please run the full tests and compile the gallery before approving the review.

Just a quick note. I'm experiencing the same problem as Verde (https://github.com/fatiando/verde/issues/269): Cartopy is not correctly plotting lands and/or oceans. I tried to pin cartopy to 0.17.* but got a Connection timed out error when cartopy was trying to download http://naciscdn.org/naturalearth/110m/physical/ne_110m_land.zip . Probably the website is currently down and I should try again later. If it end up working just fine, I will pin it and push the new environment.yml and requirements-dev.txt.

leouieda commented 4 years ago

Trying to build the gallery locally but running into the same issue with the Cartopy download. Will try again later on.

leouieda commented 4 years ago

I managed to test the gallery locally finally. I'm making some edits and will push them soon. I also checked that we can download individual files from a Zenodo archive.

santisoler commented 4 years ago

Thanks @leouieda!

I also checked that we can download individual files from a Zenodo archive.

Great! We should upload all sample data files to Zenodo and then change the URLs in the registry file. @mtb-za would you like to tackle it? I think we might need to add you to the Fatiando group in Zenodo to do so, but after that you'll be ready to start uploading the files.

mtb-za commented 4 years ago

I will carve out some time for it this week, yeah.

mtb-za commented 4 years ago

What sort of license can we use? Is everything fine with CC-0 (because it is public domain), or do we need something else?

For now, I am going to call this a version 1.0.0 upload, containing everything in https://github.com/fatiando/verde/tree/master/data https://github.com/fatiando/harmonica/tree/master/data including a combined readme. Am I missing a data file somewhere? (Boule, Pooch and Rockhound do not have any.)

I am also not too sure what to do for authors, since none of us are really the authors of any of these datasets. Can we credit the agencies as authors directly?

Pinging @leouieda @santisoler.

santisoler commented 4 years ago

What sort of license can we use? Is everything fine with CC-0 (because it is public domain), or do we need something else?

This is tricky, because not every dataset has the same license. For example, the British aeromagnetic data is released under the Open Goverment License which requires acknowledgement, so we cannot publish it under public domain. Don't know if we can omit a global license for the whole upload and specify the license of each dataset on a README.

For now, I am going to call this a version 1.0.0 upload, containing everything in https://github.com/fatiando/verde/tree/master/data https://github.com/fatiando/harmonica/tree/master/data including a combined readme. Am I missing a data file somewhere? (Boule, Pooch and Rockhound do not have any.)

Yes, we only need to upload datasets from Verde and Harmonica. Although, don't add the Rio magnetic data: it doesn't have a clear license and we will deprecate it on Verde 2.0.0. Users will still be able to download it until that release (for keeping backward compatibility), but they are encouraged to use another dataset.

I am also not too sure what to do for authors, since none of us are really the authors of any of these datasets. Can we credit the agencies as authors directly?

Maybe we can put something as Fatiando Developer Team? What do you think @leouieda ?

mtb-za commented 4 years ago

Don't know if we can omit a global license for the whole upload and specify the license of each dataset on a README.

It looks like we can do multiple uploads with different licenses. We should therefore try and group them, if any of them have the same license. Not sure if I can add the next lot to the same DOI though, I will have to experiment. We might just have to bite the bullet and have a new upload for each file. This is the info under the text-box for entering the license detail:

Selected license applies to all of your files displayed on the top of the form. If you want to upload some of your files under different licenses, please do so in separate uploads. If you cannot find the license you're looking for, include a relevant LICENSE file in your record and choose one of the Other licenses available (Other (Open), Other (Attribution), etc.). The supported licenses in the list are harvested from opendefinition.org and spdx.org . If you think that a license is missing from the list, please contact us.

Although, don't add the Rio magnetic data: it doesn't have a clear license and we will deprecate it on Verde 2.0.0. Users will still be able to download it until that release (for keeping backward compatibility), but they are encouraged to use another dataset.

I will leave the Rio data out of the upload. Do we have an alternative magnetic dataset to upload instead?

santisoler commented 4 years ago

Yes, we have the British one.