mirror of https://github.com/databricks/cli.git
32 lines
1004 B
Go
32 lines
1004 B
Go
package config_tests
|
|
|
|
import (
|
|
"testing"
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
"github.com/stretchr/testify/require"
|
|
)
|
|
|
|
func TestJobWithSparkConf(t *testing.T) {
|
|
b := loadTarget(t, "./job_with_spark_conf", "default")
|
|
assert.Len(t, b.Config.Resources.Jobs, 1)
|
|
|
|
job := b.Config.Resources.Jobs["job_with_spark_conf"]
|
|
assert.Len(t, job.JobClusters, 1)
|
|
assert.Equal(t, "test_cluster", job.JobClusters[0].JobClusterKey)
|
|
|
|
// This test exists because of https://github.com/databricks/cli/issues/992.
|
|
// It is solved for bundles as of https://github.com/databricks/cli/pull/1098.
|
|
require.Len(t, job.JobClusters, 1)
|
|
cluster := job.JobClusters[0]
|
|
assert.Equal(t, "14.2.x-scala2.12", cluster.NewCluster.SparkVersion)
|
|
assert.Equal(t, "i3.xlarge", cluster.NewCluster.NodeTypeId)
|
|
assert.Equal(t, 2, cluster.NewCluster.NumWorkers)
|
|
assert.Equal(t, map[string]string{
|
|
"spark.string": "string",
|
|
"spark.int": "1",
|
|
"spark.bool": "true",
|
|
"spark.float": "1.2",
|
|
}, cluster.NewCluster.SparkConf)
|
|
}
|