Closed kirantodekar123 closed 1 year ago
[2022-08-16, 18:49:11 UTC] {taskinstance.py:1262} INFO - Executing <Task(PythonOperator): datalake_ingest> on 2022-08-15 18:33:18.794266+00:00
[2022-08-16, 18:49:11 UTC] {standard_task_runner.py:52} INFO - Started process 129 to run task
[2022-08-16, 18:49:11 UTC] {standard_task_runner.py:76} INFO - Running: ['airflow', 'tasks', 'run', 'datahub_datalake_ingest', 'datalake_ingest', 'scheduled__2022-08-15T18:33:18.794266+00:00', '--job-id', '4', '--raw', '--subdir', 'DAGS_FOLDER/datahub/datahub_datalake_metadata_dag.py', '--cfg-path', '/tmp/tmp48rx7iq3', '--error-file', '/tmp/tmpf8k0n0r3']
[2022-08-16, 18:49:11 UTC] {standard_task_runner.py:77} INFO - Job 4: Subtask datalake_ingest
[2022-08-16, 18:49:11 UTC] {logging_mixin.py:109} INFO - Running <TaskInstance: datahub_datalake_ingest.datalake_ingest scheduled2022-08-15T18:33:18.794266+00:00 [running]> on host 9e5ae13b54b4
[2022-08-16, 18:49:11 UTC] {taskinstance.py:1429} INFO - Exporting the following env vars:
AIRFLOW_CTX_DAG_EMAIL=data-engineering@digitalturbine.com
AIRFLOW_CTX_DAG_OWNER=data-engineering
AIRFLOW_CTX_DAG_ID=datahub_datalake_ingest
AIRFLOW_CTX_TASK_ID=datalake_ingest
AIRFLOW_CTX_EXECUTION_DATE=2022-08-15T18:33:18.794266+00:00
AIRFLOW_CTX_DAG_RUN_ID=scheduled2022-08-15T18:33:18.794266+00:00
[2022-08-16, 18:49:11 UTC] {datahub_datalake_metadata_dag.py:39} INFO - /airflow/dags/datahub/recipes/dev/datahub/app_media.yaml,
Resolved close this
Describe the bug I am ingesting the data from s3 delta lake as source to own server as sink . After ingesting the data I am able to get only path but no data table and other things are coming . I am running those config file through apache airflow DAG .
config file is below
source: type: s3 config: path_specs:
sink: type: "datahub-rest" config: server: "https://datahub-sbx-gms.amer-dev.xxxxx.com:8080/" (edited)
Expected behavior data table should come at server side .
Additional context Errors Message