-
As a developer, when running our Platform ETL process, I would like for the pipeline to have an option where computation for pre-existing results is skipped.
## Background
_Platform ETL_ process i…
-
Hi,
Great product!
I see an issue though - I want to specify mode-specific config eg S3_ETL_BUCKET. If I don't specify this key in the default config, when creating my pipeline with:
dataduc…
-
Notes 8/30:
* pipeline utils functions moved into each capsule (i.e. dont need special codebase to translate data between capsules)
* capsules are being added to github for each processing step (ex: a…
-
I want to use the Custom to FHIR pipeline but this class does not seem to exist in any of the branches:
com.google.cloud.healthcare.etl.runner.customtofhir.CustomToFhirStreamingRunner
Can you mak…
-
-
### Search before asking
- [X] I had searched in the [feature](https://github.com/apache/seatunnel/issues?q=is%3Aissue+label%3A%22Feature%22) and found no similar feature requirement.
### Descripti…
-
### Dagster version
1.4.17
### What's the issue?
LeaseLost error when writing outputs to Azure Blob Storage using Dagster Azure ADLS2 IO Manager.
During the execution of a Dagster pipeline, spec…
-
I am trying to use this minimal config described here: http://dataduct.readthedocs.org/en/latest/config.html
But it appears that it's insufficient.
dataduct wouldn't start without a logging section …
-
Extend the Flask API to read from a .jsonl file of scraped Commands data and insert it into the Postgres DB.
**Is your feature request related to a problem? Please describe.**
As part of our data …
-
Currently, Flow can only read XML, but we should start looking into saving XML files as well.
There are two libraries that could help us achieve that:
- https://github.com/veewee/xml
- https://git…