Closed dejan closed 1 year ago
@dejan a quick update on this. The Delta folks at Databricks are looking at supporting dynamic partition overwrite. It's prioritized in their roadmap, I'll post back here once it's released.
Thanks @bilalaslamseattle !
Any updates on this?
@creativedutchmen this capability it is in Delta Lake 2.0. We now have to implement it in dbt-databricks. It's on our radar. @superdupershant @ueshin and @allisonwang-db FYI.
+1 on this. Adding one data point that this will be a blocker for us for adapting dbt
Great, thanks! This will greatly reduce the runtime of some of our heaviest models :)
I submitted a PR to include it in dbt-spark
if you want to look at it: https://github.com/dbt-labs/dbt-spark/pull/430
This issue has been marked as Stale because it has been open for 180 days with no activity. If you would like the issue to remain open, please remove the stale label or comment on the issue.
Still relevant for me
@bilalaslamseattle please share some updates on this.
This issue has been address by #310
I don't think it was a good decision to close this because #310 has a major flaw which was not resolved (only documented) https://github.com/databricks/dbt-databricks/issues/334
I haven't confirmed this but by looking at the documentation insert_overwrite apparently now supports dynamic partition overwrite (it no longer errors for Delta), however it is stated that it only works for All-purpose clusters which is also a major drawback as that's not cost-efficient.
Can someone please confirm the status quo and plans on properly providing such a basic and common case such as partition overwrite?
however it is stated that it only works for All-purpose clusters which is also a major drawback as that's not cost-efficient. @andrefurlan-db can you confirm if this is correct? Seems odd.
@bilalaslamseattle / @andrefurlan-db was there an update on this one or a thread somewhere else? When I run partition overwrite against SQL warehouses, I get:
Error running query: [_LEGACY_ERROR_TEMP_DBR_0222] org.apache.spark.sql.catalyst.ExtendedAnalysisException: Configuration spark.sql.sources.partitionOverwriteMode is not available.
Looks like its attempting to set some custom spark config which I don't think is allowed on SQL warehouses.
Are there any plans to implement it? Or any updates?
Describe the feature
This was reported in https://github.com/dbt-labs/dbt-spark/issues/155 but I think you might be more interested in resolving the issue
Delta already supports partitions overwrite but it seems that dbt adapter implementation is not making use of it.
Describe alternatives you've considered
I could not find a way to atomically overwrite a partition.
Who will this benefit?
Everyone using dbt and Delta.