Closed shreyas-goenka closed 6 hours ago
If integration tests don't run automatically, an authorized user can run them manually by following the instructions below:
Trigger: go/deco-tests-run/cli
Inputs:
382a4efd6efb98adbc7532a3286d81521c899843
Checks will be approved automatically on success.
Test Details: go/deco-tests/11974490258
Changes
This PR adds a warning validating that the configuration for a single node cluster is valid for interactive, job, job-task, and pipeline clusters.
Note: We skip the validation if a cluster policy is configured because the policy is likely to configure
spark_conf
/custom_tags
itself.Note: Terrform originally only had validation for interactive, job, and job-task clusters. This PR adding the validation for pipeline clusters as well is new.
This PR follows the same logic as we used to have in Terraform. The validation was removed from Terraform because we had no way to demote the error to a warning: https://github.com/databricks/terraform-provider-databricks/pull/4222
Background
Single-node clusters require
spark_conf
andcustom_tags
to be correctly set in the cluster definition for them to function optimally. The cluster will be created even if incorrectly configured, but its performance will not be great.For example, if both
spark_conf
andcustom_tags
are not set andnum_workers
is 0, then only the driver process will be launched on the cluster compute instance thus leading to sub-optimal utilization of available compute resources and no parallelization across worker processes when processing a spark query.Issue
This PR addresses some issues reported in https://github.com/databricks/cli/issues/1546
Tests
Unit tests and manually.
Example output of the warning: