dtcc-platform / dtcc

DTCC Platform
MIT License
0 stars 4 forks source link

Workflow: DTE/4PEDs Viewer #55

Open anderslogg opened 1 year ago

anderslogg commented 1 year ago

Workflow for supporting DTE (energy viewer). See workflows/workflow_dte.py in DTCC module.

Questions to be answered:

Overview: Explain what the overall idea is

Input data: List input data, file formats etc in detail. How to obtain the data? Also attach example data to this issue.

Output data: List input data, file formats etc in detail. How is the output data used?

How is this done today: Explain the workflow as it is today with existing tools.

Comments: Any comments? How would this work in an ideal world?

andreasrudena commented 1 year ago

Overview:

The scenario data of Gothenburg buildings are currently built into tiles containing (too) much information about the different indicators and result from the simulations. We need to off-load the tiles with a geospatial database accessible through an API so that the client can fetch scenario data dynamically.

Input data:

Data is based on 1. GeoJSON (from Shapefile) with buildings of Gothenburg and 2..n. Scenario data that maps to the buildings with id. The data has two version depending on if is the current state of the city or the future state of the city, e.g:

.. etc (we have only 2 and 3 above from Claudio, but for example 4. Solar potential is on discussion

(note that 1 is protected by license from Lantmäteriet and cannot be hosted publicly)

On top of this there are 6 aggregation boundary levels represented by GeoJSON:

The data should be aggregated by those boundaries so that boundary features are queryable.

Output data:

The buildings should be queried from a REST API (preferably in a standardised way)

(The tiles are not taken into consideration at this stage of discussion)

How this is done today?

  1. Data is loaded and mapped to merge the attributes from scenario data to the building polygons of the features
  2. The attributes a aggregated into statistical values and put on aggregated boundaries
  3. Colors and some other values are precalculated for different indicator levels
  4. Tiles are generated to file
  5. Tiles are hosted on a tile server

Ideal world?

  1. Create different processes to be run in data pipeline workflows
  2. Allow for reusable indicators (and project specific indicators)
  3. Generate tiles and host them from a tile server API
  4. Load data into one or several databases and give access to the data through standard REST API
  5. Let the user modify parts of the data depending on need (might need to rerun pipelines)

Comments

andreasrudena commented 11 months ago

@vbassn A new dataset of scenario data arrived recently. Better check directly with Claudio to get access directly to the source.