package bundle import ( "context" "encoding/json" "io" "path" "strconv" "testing" "github.com/databricks/cli/bundle/config" "github.com/databricks/cli/bundle/metadata" "github.com/databricks/cli/internal" "github.com/databricks/cli/internal/acc" "github.com/databricks/cli/libs/env" "github.com/databricks/cli/libs/filer" "github.com/google/uuid" "github.com/stretchr/testify/assert" "github.com/stretchr/testify/require" ) func TestAccJobsMetadataFile(t *testing.T) { ctx, wt := acc.WorkspaceTest(t) w := wt.W nodeTypeId := internal.GetNodeTypeId(env.Get(ctx, "CLOUD_ENV")) uniqueId := uuid.New().String() bundleRoot, err := initTestTemplate(t, ctx, "job_metadata", map[string]any{ "unique_id": uniqueId, "node_type_id": nodeTypeId, "spark_version": "13.2.x-snapshot-scala2.12", }) require.NoError(t, err) // deploy bundle err = deployBundle(t, ctx, bundleRoot) require.NoError(t, err) // Cleanup the deployed bundle t.Cleanup(func() { err = destroyBundle(t, ctx, bundleRoot) require.NoError(t, err) }) // assert job 1 is created jobName := "test-job-metadata-1-" + uniqueId job1, err := w.Jobs.GetBySettingsName(context.Background(), jobName) require.NoError(t, err) assert.Equal(t, job1.Settings.Name, jobName) // assert job 2 is created jobName = "test-job-metadata-2-" + uniqueId job2, err := w.Jobs.GetBySettingsName(context.Background(), jobName) require.NoError(t, err) assert.Equal(t, job2.Settings.Name, jobName) // Compute root path for the bundle deployment root := getBundleRemoteRootPath(w, t, uniqueId) f, err := filer.NewWorkspaceFilesClient(w, root) require.NoError(t, err) // Read metadata object from the workspace r, err := f.Read(context.Background(), "state/metadata.json") require.NoError(t, err) b, err := io.ReadAll(r) require.NoError(t, err) actualMetadata := metadata.Metadata{} err = json.Unmarshal(b, &actualMetadata) require.NoError(t, err) // expected value for the metadata expectedMetadata := metadata.Metadata{ Version: metadata.Version, Config: metadata.Config{ Bundle: metadata.Bundle{ Git: config.Git{ BundleRootPath: ".", }, }, Workspace: metadata.Workspace{ FilePath: path.Join(root, "files"), }, Resources: metadata.Resources{ Jobs: map[string]*metadata.Job{ "foo": { ID: strconv.FormatInt(job1.JobId, 10), RelativePath: "databricks.yml", }, "bar": { ID: strconv.FormatInt(job2.JobId, 10), RelativePath: "a/b/resources.yml", }, }, }, }, } // Assert metadata matches what we expected. assert.Equal(t, expectedMetadata, actualMetadata) }