package bundle import ( "fmt" "os" "path/filepath" "testing" "github.com/databricks/cli/internal" "github.com/databricks/cli/internal/acc" "github.com/databricks/databricks-sdk-go" "github.com/databricks/databricks-sdk-go/service/jobs" "github.com/google/uuid" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" ) func TestAccBindJobToExistingJob(t *testing.T) { env := internal.GetEnvOrSkipTest(t, "CLOUD_ENV") t.Log(env) ctx, wt := acc.WorkspaceTest(t) gt := &generateJobTest{T: t, w: wt.W} nodeTypeId := internal.GetNodeTypeId(env) uniqueId := uuid.New().String() bundleRoot, err := initTestTemplate(t, ctx, "basic", map[string]any{ "unique_id": uniqueId, "spark_version": "13.3.x-scala2.12", "node_type_id": nodeTypeId, }) require.NoError(t, err) jobId := gt.createTestJob(ctx) t.Cleanup(func() { gt.destroyJob(ctx, jobId) require.NoError(t, err) }) t.Setenv("BUNDLE_ROOT", bundleRoot) c := internal.NewCobraTestRunner(t, "bundle", "deployment", "bind", "foo", fmt.Sprint(jobId), "--auto-approve") _, _, err = c.Run() require.NoError(t, err) // Remove .databricks directory to simulate a fresh deployment err = os.RemoveAll(filepath.Join(bundleRoot, ".databricks")) require.NoError(t, err) err = deployBundle(t, ctx, bundleRoot) require.NoError(t, err) w, err := databricks.NewWorkspaceClient() require.NoError(t, err) // Check that job is bound and updated with config from bundle job, err := w.Jobs.Get(ctx, jobs.GetJobRequest{ JobId: jobId, }) require.NoError(t, err) require.Equal(t, job.Settings.Name, fmt.Sprintf("test-job-basic-%s", uniqueId)) require.Contains(t, job.Settings.Tasks[0].SparkPythonTask.PythonFile, "hello_world.py") c = internal.NewCobraTestRunner(t, "bundle", "deployment", "unbind", "foo") _, _, err = c.Run() require.NoError(t, err) // Remove .databricks directory to simulate a fresh deployment err = os.RemoveAll(filepath.Join(bundleRoot, ".databricks")) require.NoError(t, err) err = destroyBundle(t, ctx, bundleRoot) require.NoError(t, err) // Check that job is unbound and exists after bundle is destroyed job, err = w.Jobs.Get(ctx, jobs.GetJobRequest{ JobId: jobId, }) require.NoError(t, err) require.Equal(t, job.Settings.Name, fmt.Sprintf("test-job-basic-%s", uniqueId)) require.Contains(t, job.Settings.Tasks[0].SparkPythonTask.PythonFile, "hello_world.py") } func TestAccAbortBind(t *testing.T) { env := internal.GetEnvOrSkipTest(t, "CLOUD_ENV") t.Log(env) ctx, wt := acc.WorkspaceTest(t) gt := &generateJobTest{T: t, w: wt.W} nodeTypeId := internal.GetNodeTypeId(env) uniqueId := uuid.New().String() bundleRoot, err := initTestTemplate(t, ctx, "basic", map[string]any{ "unique_id": uniqueId, "spark_version": "13.3.x-scala2.12", "node_type_id": nodeTypeId, }) require.NoError(t, err) jobId := gt.createTestJob(ctx) t.Cleanup(func() { gt.destroyJob(ctx, jobId) destroyBundle(t, ctx, bundleRoot) }) // Bind should fail because prompting is not possible. t.Setenv("BUNDLE_ROOT", bundleRoot) t.Setenv("TERM", "dumb") c := internal.NewCobraTestRunner(t, "bundle", "deployment", "bind", "foo", fmt.Sprint(jobId)) // Expect error suggesting to use --auto-approve _, _, err = c.Run() assert.ErrorContains(t, err, "failed to bind the resource") assert.ErrorContains(t, err, "This bind operation requires user confirmation, but the current console does not support prompting. Please specify --auto-approve if you would like to skip prompts and proceed") err = deployBundle(t, ctx, bundleRoot) require.NoError(t, err) w, err := databricks.NewWorkspaceClient() require.NoError(t, err) // Check that job is not bound and not updated with config from bundle job, err := w.Jobs.Get(ctx, jobs.GetJobRequest{ JobId: jobId, }) require.NoError(t, err) require.NotEqual(t, job.Settings.Name, fmt.Sprintf("test-job-basic-%s", uniqueId)) require.Contains(t, job.Settings.Tasks[0].NotebookTask.NotebookPath, "test") } func TestAccGenerateAndBind(t *testing.T) { env := internal.GetEnvOrSkipTest(t, "CLOUD_ENV") t.Log(env) ctx, wt := acc.WorkspaceTest(t) gt := &generateJobTest{T: t, w: wt.W} uniqueId := uuid.New().String() bundleRoot, err := initTestTemplate(t, ctx, "with_includes", map[string]any{ "unique_id": uniqueId, }) require.NoError(t, err) w, err := databricks.NewWorkspaceClient() require.NoError(t, err) jobId := gt.createTestJob(ctx) t.Cleanup(func() { _, err = w.Jobs.Get(ctx, jobs.GetJobRequest{ JobId: jobId, }) if err == nil { gt.destroyJob(ctx, jobId) } }) t.Setenv("BUNDLE_ROOT", bundleRoot) c := internal.NewCobraTestRunnerWithContext(t, ctx, "bundle", "generate", "job", "--key", "test_job_key", "--existing-job-id", fmt.Sprint(jobId), "--config-dir", filepath.Join(bundleRoot, "resources"), "--source-dir", filepath.Join(bundleRoot, "src")) _, _, err = c.Run() require.NoError(t, err) _, err = os.Stat(filepath.Join(bundleRoot, "src", "test.py")) require.NoError(t, err) matches, err := filepath.Glob(filepath.Join(bundleRoot, "resources", "test_job_key.yml")) require.NoError(t, err) require.Len(t, matches, 1) c = internal.NewCobraTestRunner(t, "bundle", "deployment", "bind", "test_job_key", fmt.Sprint(jobId), "--auto-approve") _, _, err = c.Run() require.NoError(t, err) err = deployBundle(t, ctx, bundleRoot) require.NoError(t, err) err = destroyBundle(t, ctx, bundleRoot) require.NoError(t, err) // Check that job is bound and does not extsts after bundle is destroyed _, err = w.Jobs.Get(ctx, jobs.GetJobRequest{ JobId: jobId, }) require.ErrorContains(t, err, "does not exist.") }