The Python file to be executed. Cloud file URIs (such as dbfs:/, s3:/, adls:/, gcs:/) and workspace paths are supported. For python files stored in the Databricks workspace, the path must be absolute and begin with /. For files stored in a remote repository, the path must be relative. This field is required.
Create a similar abstraction to databricks.labs.ucx.mixins.wspath.WorkspacePath to load files from DBFS. Focus on open(...) method. We already have the most of DBFS APIs mapped out in databricks.sdk.mixins.files.DbfsExt and accessible through workspace client.
Is there an existing issue for this?
Problem statement
https://databricks-sdk-py.readthedocs.io/en/latest/dbdataclasses/jobs.html#databricks.sdk.service.jobs.SparkPythonTask
Proposed Solution
Create a similar abstraction to
databricks.labs.ucx.mixins.wspath.WorkspacePath
to load files from DBFS. Focus onopen(...)
method. We already have the most of DBFS APIs mapped out indatabricks.sdk.mixins.files.DbfsExt
and accessible through workspace client.Additional Context
No response