mirror of https://github.com/databricks/cli.git
Enable Spark JAR task test (#1658)
## Changes Enable Spark JAR task test ## Tests ``` Updating deployment state... Deleting files... Destroy complete! --- PASS: TestAccSparkJarTaskDeployAndRunOnVolumes (194.13s) PASS coverage: 51.9% of statements in ./... ok github.com/databricks/cli/internal/bundle 194.586s coverage: 51.9% of statements in ./... ```
This commit is contained in:
parent
f3ffded3bf
commit
9d1fbbb39c
|
@ -24,6 +24,10 @@
|
|||
"artifact_path": {
|
||||
"type": "string",
|
||||
"description": "Path to the remote base path for artifacts"
|
||||
},
|
||||
"instance_pool_id": {
|
||||
"type": "string",
|
||||
"description": "Instance pool id for job cluster"
|
||||
}
|
||||
}
|
||||
}
|
||||
|
|
|
@ -22,6 +22,7 @@ resources:
|
|||
num_workers: 1
|
||||
spark_version: "{{.spark_version}}"
|
||||
node_type_id: "{{.node_type_id}}"
|
||||
instance_pool_id: "{{.instance_pool_id}}"
|
||||
spark_jar_task:
|
||||
main_class_name: PrintArgs
|
||||
libraries:
|
||||
|
|
|
@ -6,15 +6,14 @@ import (
|
|||
|
||||
"github.com/databricks/cli/internal"
|
||||
"github.com/databricks/cli/internal/acc"
|
||||
"github.com/databricks/cli/libs/env"
|
||||
"github.com/google/uuid"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func runSparkJarTest(t *testing.T, sparkVersion string) {
|
||||
t.Skip("Temporarily skipping the test until auth / permission issues for UC volumes are resolved.")
|
||||
|
||||
env := internal.GetEnvOrSkipTest(t, "CLOUD_ENV")
|
||||
t.Log(env)
|
||||
cloudEnv := internal.GetEnvOrSkipTest(t, "CLOUD_ENV")
|
||||
t.Log(cloudEnv)
|
||||
|
||||
if os.Getenv("TEST_METASTORE_ID") == "" {
|
||||
t.Skip("Skipping tests that require a UC Volume when metastore id is not set.")
|
||||
|
@ -24,14 +23,16 @@ func runSparkJarTest(t *testing.T, sparkVersion string) {
|
|||
w := wt.W
|
||||
volumePath := internal.TemporaryUcVolume(t, w)
|
||||
|
||||
nodeTypeId := internal.GetNodeTypeId(env)
|
||||
nodeTypeId := internal.GetNodeTypeId(cloudEnv)
|
||||
tmpDir := t.TempDir()
|
||||
instancePoolId := env.Get(ctx, "TEST_INSTANCE_POOL_ID")
|
||||
bundleRoot, err := initTestTemplateWithBundleRoot(t, ctx, "spark_jar_task", map[string]any{
|
||||
"node_type_id": nodeTypeId,
|
||||
"unique_id": uuid.New().String(),
|
||||
"spark_version": sparkVersion,
|
||||
"root": tmpDir,
|
||||
"artifact_path": volumePath,
|
||||
"node_type_id": nodeTypeId,
|
||||
"unique_id": uuid.New().String(),
|
||||
"spark_version": sparkVersion,
|
||||
"root": tmpDir,
|
||||
"artifact_path": volumePath,
|
||||
"instance_pool_id": instancePoolId,
|
||||
}, tmpDir)
|
||||
require.NoError(t, err)
|
||||
|
||||
|
|
Loading…
Reference in New Issue