SolidLabResearch / Challenges

24 stars 0 forks source link

Manage workflows to derive/provide data to solid pods #50

Open ajuvercr opened 2 years ago

ajuvercr commented 2 years ago

Data workflow can be configured and executed, providing new or derived data in solid pods.

Actors/actresses

Preconditions

Steps

Acceptance criteria

The workflow ingests the strava data, saving it as linked data. The workflow derives and stores the total running distance and average speed of the ingested data.

Postconditions

Note: 3rd party data doesn't have to be actual 3rd party. Data from the end-users solid pod can also be seen as 3rd party data.

RubenVerborgh commented 2 years ago

Can we apply this to a concrete use case, and have a specific demo, such that the challenge has a clear end?

pheyvaer commented 2 years ago

@ajuvercr Did you already have a chance to look into making the requested changes?

ajuvercr commented 2 years ago

A demo could be a pipeline that ingests gpx data and creates an eventsource from this data (as time-based versioned LDES).

Concretely the user posts a gpx file to a server running the pipeline and the solid server. The pipeline ingests this data and creates with RML linked data which is then transformed into a time-based versioned LDES on the users solid pod.

If this can be done the challenge is done because such a configuration file had to be generated and is able to run. Sounds good?

pheyvaer commented 2 years ago

@RubenVerborgh What do you think? I see that this issue is marked as a scenario and not a challenge though.

RubenVerborgh commented 2 years ago

I'd need a real highly concrete use case. Like: we will ingest a person's location history, or we will ingest 1000-measurement/s sensor data.

pheyvaer commented 1 year ago

@ajuvercr Can you implement the necessary changes?

ajuvercr commented 1 year ago

Added a concrete use case that calculates basic statistics of a strava run, similar to @woutslabbinck work