2024-02-16 20:54:38 +00:00
|
|
|
package tfdyn
|
|
|
|
|
|
|
|
import (
|
|
|
|
"context"
|
|
|
|
"fmt"
|
2024-09-26 13:22:22 +00:00
|
|
|
"sort"
|
2024-02-16 20:54:38 +00:00
|
|
|
|
|
|
|
"github.com/databricks/cli/bundle/internal/tf/schema"
|
|
|
|
"github.com/databricks/cli/libs/dyn"
|
|
|
|
"github.com/databricks/cli/libs/dyn/convert"
|
|
|
|
"github.com/databricks/cli/libs/log"
|
|
|
|
"github.com/databricks/databricks-sdk-go/service/jobs"
|
|
|
|
)
|
|
|
|
|
|
|
|
func convertJobResource(ctx context.Context, vin dyn.Value) (dyn.Value, error) {
|
|
|
|
// Normalize the input value to the underlying job schema.
|
|
|
|
// This removes superfluous keys and adapts the input to the expected schema.
|
|
|
|
vin, diags := convert.Normalize(jobs.JobSettings{}, vin)
|
|
|
|
for _, diag := range diags {
|
|
|
|
log.Debugf(ctx, "job normalization diagnostic: %s", diag.Summary)
|
|
|
|
}
|
|
|
|
|
2024-09-26 13:22:22 +00:00
|
|
|
// Sort the tasks of each job in the bundle by task key. Sorting
|
|
|
|
// the task keys ensures that the diff computed by terraform is correct and avoids
|
|
|
|
// recreates. For more details see the NOTE at
|
|
|
|
// https://registry.terraform.io/providers/databricks/databricks/latest/docs/resources/job#example-usage
|
|
|
|
// and https://github.com/databricks/terraform-provider-databricks/issues/4011
|
|
|
|
// and https://github.com/databricks/cli/pull/1776
|
|
|
|
vout := vin
|
|
|
|
var err error
|
|
|
|
tasks, ok := vin.Get("tasks").AsSequence()
|
|
|
|
if ok {
|
|
|
|
sort.Slice(tasks, func(i, j int) bool {
|
|
|
|
// We sort the tasks by their task key. Tasks without task keys are ordered
|
|
|
|
// before tasks with task keys. We do not error for those tasks
|
|
|
|
// since presence of a task_key is validated for in the Jobs backend.
|
|
|
|
tk1, ok := tasks[i].Get("task_key").AsString()
|
|
|
|
if !ok {
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
tk2, ok := tasks[j].Get("task_key").AsString()
|
|
|
|
if !ok {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
return tk1 < tk2
|
|
|
|
})
|
|
|
|
vout, err = dyn.Set(vin, "tasks", dyn.V(tasks))
|
|
|
|
if err != nil {
|
|
|
|
return dyn.InvalidValue, err
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2024-02-16 20:54:38 +00:00
|
|
|
// Modify top-level keys.
|
2024-09-26 13:22:22 +00:00
|
|
|
vout, err = renameKeys(vout, map[string]string{
|
2024-02-16 20:54:38 +00:00
|
|
|
"tasks": "task",
|
|
|
|
"job_clusters": "job_cluster",
|
|
|
|
"parameters": "parameter",
|
2024-04-22 11:44:34 +00:00
|
|
|
"environments": "environment",
|
2024-02-16 20:54:38 +00:00
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return dyn.InvalidValue, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Modify keys in the "git_source" block
|
2024-03-07 13:56:50 +00:00
|
|
|
vout, err = dyn.Map(vout, "git_source", func(_ dyn.Path, v dyn.Value) (dyn.Value, error) {
|
2024-02-16 20:54:38 +00:00
|
|
|
return renameKeys(v, map[string]string{
|
|
|
|
"git_branch": "branch",
|
|
|
|
"git_commit": "commit",
|
|
|
|
"git_provider": "provider",
|
|
|
|
"git_tag": "tag",
|
|
|
|
"git_url": "url",
|
|
|
|
})
|
|
|
|
})
|
|
|
|
if err != nil {
|
|
|
|
return dyn.InvalidValue, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Modify keys in the "task" blocks
|
2024-03-07 13:56:50 +00:00
|
|
|
vout, err = dyn.Map(vout, "task", dyn.Foreach(func(_ dyn.Path, v dyn.Value) (dyn.Value, error) {
|
2024-04-05 15:52:39 +00:00
|
|
|
// Modify "library" blocks for for_each_task
|
|
|
|
vout, err = dyn.Map(v, "for_each_task.task", func(_ dyn.Path, v dyn.Value) (dyn.Value, error) {
|
|
|
|
return renameKeys(v, map[string]string{
|
|
|
|
"libraries": "library",
|
|
|
|
})
|
|
|
|
})
|
|
|
|
|
|
|
|
if err != nil {
|
|
|
|
return dyn.InvalidValue, err
|
|
|
|
}
|
|
|
|
|
|
|
|
return renameKeys(vout, map[string]string{
|
2024-02-16 20:54:38 +00:00
|
|
|
"libraries": "library",
|
|
|
|
})
|
|
|
|
}))
|
|
|
|
if err != nil {
|
|
|
|
return dyn.InvalidValue, err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Normalize the output value to the target schema.
|
|
|
|
vout, diags = convert.Normalize(schema.ResourceJob{}, vout)
|
|
|
|
for _, diag := range diags {
|
|
|
|
log.Debugf(ctx, "job normalization diagnostic: %s", diag.Summary)
|
|
|
|
}
|
|
|
|
|
|
|
|
return vout, err
|
|
|
|
}
|
|
|
|
|
|
|
|
type jobConverter struct{}
|
|
|
|
|
|
|
|
func (jobConverter) Convert(ctx context.Context, key string, vin dyn.Value, out *schema.Resources) error {
|
|
|
|
vout, err := convertJobResource(ctx, vin)
|
|
|
|
if err != nil {
|
|
|
|
return err
|
|
|
|
}
|
|
|
|
|
|
|
|
// Add the converted resource to the output.
|
|
|
|
out.Job[key] = vout.AsAny()
|
|
|
|
|
|
|
|
// Configure permissions for this resource.
|
|
|
|
if permissions := convertPermissionsResource(ctx, vin); permissions != nil {
|
|
|
|
permissions.JobId = fmt.Sprintf("${databricks_job.%s.id}", key)
|
|
|
|
out.Permissions["job_"+key] = permissions
|
|
|
|
}
|
|
|
|
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
|
|
|
|
func init() {
|
|
|
|
registerConverter("jobs", jobConverter{})
|
|
|
|
}
|