Problem
One of the major goal of the platform is to streamline data search, access, exploration and analysis for geospatial workflows. Considering a dataset is available (via the catalog), design mechanisms to import it in the platform for further analysis. The import process by itself needs to be as transparent as possible from the users perspective with minimal interaction.
Considering the size, data can either can made available in the Jupyter environment directly or can be imported indirectly in the processing environment (HPC).
Identify representative examples of a large cloud optimized dataset that cannot be staged locally & a smaller dataset that can be staged locally in Jupyter.
Demonstrate how these two datasets can be discovered in the catalog and launched in the Jupyter environment.
Problem One of the major goal of the platform is to streamline data search, access, exploration and analysis for geospatial workflows. Considering a dataset is available (via the catalog), design mechanisms to import it in the platform for further analysis. The import process by itself needs to be as transparent as possible from the users perspective with minimal interaction.
Considering the size, data can either can made available in the Jupyter environment directly or can be imported indirectly in the processing environment (HPC).
Potential Solution GeoEDF Connectors
Pull Request(s) ToDo ...
Data Catalog Components![iguide-workshop-data-catalog-diagram](https://github.com/I-GUIDE/CI_Platform/assets/11427468/dd2b19a5-5e74-4ffc-a43e-6d55f217be53)