mirror of https://github.com/databricks/cli.git
Compare commits
5 Commits
8f59695d0d
...
5194889605
Author | SHA1 | Date |
---|---|---|
|
5194889605 | |
|
aed0e0a92f | |
|
43c4b581d4 | |
|
d3b30f73ec | |
|
40505b867a |
|
@ -170,32 +170,29 @@ func convertLinksToAbsoluteUrl(s string) string {
|
||||||
base := "https://docs.databricks.com"
|
base := "https://docs.databricks.com"
|
||||||
referencePage := "/dev-tools/bundles/reference.html"
|
referencePage := "/dev-tools/bundles/reference.html"
|
||||||
|
|
||||||
// Regular expression to match Markdown-style links
|
// Regular expression to match Markdown-style links like [_](link)
|
||||||
re := regexp.MustCompile(`\[_\]\(([^)]+)\)`)
|
re := regexp.MustCompile(`\[_\]\(([^)]+)\)`)
|
||||||
result := re.ReplaceAllStringFunc(s, func(match string) string {
|
result := re.ReplaceAllStringFunc(s, func(match string) string {
|
||||||
// Extract the URL inside parentheses
|
|
||||||
matches := re.FindStringSubmatch(match)
|
matches := re.FindStringSubmatch(match)
|
||||||
if len(matches) < 2 {
|
if len(matches) < 2 {
|
||||||
return match // Return original if no match found
|
return match
|
||||||
}
|
}
|
||||||
link := matches[1]
|
link := matches[1]
|
||||||
|
|
||||||
var text, absoluteURL string
|
var text, absoluteURL string
|
||||||
|
|
||||||
if strings.HasPrefix(link, "#") {
|
if strings.HasPrefix(link, "#") {
|
||||||
text = strings.TrimPrefix(link, "#")
|
text = strings.TrimPrefix(link, "#")
|
||||||
absoluteURL = fmt.Sprintf("%s%s%s", base, referencePage, link)
|
absoluteURL = fmt.Sprintf("%s%s%s", base, referencePage, link)
|
||||||
} else if strings.HasPrefix(link, "/") {
|
|
||||||
// Handle relative paths like /dev-tools/bundles/resources.html#dashboard
|
// Handle relative paths like /dev-tools/bundles/resources.html#dashboard
|
||||||
|
} else if strings.HasPrefix(link, "/") {
|
||||||
|
absoluteURL = strings.ReplaceAll(fmt.Sprintf("%s%s", base, link), ".md", ".html")
|
||||||
if strings.Contains(link, "#") {
|
if strings.Contains(link, "#") {
|
||||||
parts := strings.Split(link, "#")
|
parts := strings.Split(link, "#")
|
||||||
text = parts[1]
|
text = parts[1]
|
||||||
absoluteURL = fmt.Sprintf("%s%s", base, link)
|
|
||||||
} else {
|
} else {
|
||||||
text = "link"
|
text = "link"
|
||||||
absoluteURL = fmt.Sprintf("%s%s", base, link)
|
|
||||||
}
|
}
|
||||||
absoluteURL = strings.ReplaceAll(absoluteURL, ".md", ".html")
|
|
||||||
} else {
|
} else {
|
||||||
return match
|
return match
|
||||||
}
|
}
|
||||||
|
|
|
@ -19,7 +19,9 @@ github.com/databricks/cli/bundle/config.Artifact:
|
||||||
The location where the built artifact will be saved.
|
The location where the built artifact will be saved.
|
||||||
"type":
|
"type":
|
||||||
"description": |-
|
"description": |-
|
||||||
The type of the artifact. Valid values are wheel or jar.
|
The type of the artifact.
|
||||||
|
"markdown_description": |-
|
||||||
|
The type of the artifact. Valid values are `wheel` or `jar`
|
||||||
github.com/databricks/cli/bundle/config.ArtifactFile:
|
github.com/databricks/cli/bundle/config.ArtifactFile:
|
||||||
"source":
|
"source":
|
||||||
"description": |-
|
"description": |-
|
||||||
|
@ -191,7 +193,9 @@ github.com/databricks/cli/bundle/config.Root:
|
||||||
Defines attributes for experimental features.
|
Defines attributes for experimental features.
|
||||||
"include":
|
"include":
|
||||||
"description": |-
|
"description": |-
|
||||||
PLACEHOLDER
|
Specifies a list of path globs that contain configuration files to include within the bundle.
|
||||||
|
"markdown_description": |-
|
||||||
|
Specifies a list of path globs that contain configuration files to include within the bundle. See [_](/dev-tools/bundles/settings.md#include)
|
||||||
"permissions":
|
"permissions":
|
||||||
"description": |-
|
"description": |-
|
||||||
Defines the permissions to apply to experiments, jobs, pipelines, and models defined in the bundle
|
Defines the permissions to apply to experiments, jobs, pipelines, and models defined in the bundle
|
||||||
|
@ -204,9 +208,9 @@ github.com/databricks/cli/bundle/config.Root:
|
||||||
Defines bundle deployment presets. See [_](/dev-tools/bundles/deployment-modes.md#presets).
|
Defines bundle deployment presets. See [_](/dev-tools/bundles/deployment-modes.md#presets).
|
||||||
"resources":
|
"resources":
|
||||||
"description": |-
|
"description": |-
|
||||||
PLACEHOLDER
|
Specifies information about the Databricks resources used by the bundle
|
||||||
"markdown_description": |-
|
"markdown_description": |-
|
||||||
See [_](/dev-tools/bundles/resources.md).
|
Specifies information about the Databricks resources used by the bundle. See [_](/dev-tools/bundles/resources.md).
|
||||||
"run_as":
|
"run_as":
|
||||||
"description": |-
|
"description": |-
|
||||||
The identity to use to run the bundle.
|
The identity to use to run the bundle.
|
||||||
|
@ -259,9 +263,9 @@ github.com/databricks/cli/bundle/config.Target:
|
||||||
The Git version control settings for the target. See [_](#git).
|
The Git version control settings for the target. See [_](#git).
|
||||||
"mode":
|
"mode":
|
||||||
"description": |-
|
"description": |-
|
||||||
The deployment mode for the target. Valid values are development or production.
|
The deployment mode for the target.
|
||||||
"markdown_description": |-
|
"markdown_description": |-
|
||||||
The deployment mode for the target. Valid values are development or production. See [_](/dev-tools/bundles/deployment-modes.md).
|
The deployment mode for the target. Valid values are `development` or `production`. See [_](/dev-tools/bundles/deployment-modes.md).
|
||||||
"permissions":
|
"permissions":
|
||||||
"description": |-
|
"description": |-
|
||||||
The permissions for deploying and running the bundle in the target.
|
The permissions for deploying and running the bundle in the target.
|
||||||
|
@ -294,7 +298,7 @@ github.com/databricks/cli/bundle/config.Target:
|
||||||
The custom variable definitions for the target. See [_](/dev-tools/bundles/settings.md#variables) and [_](/dev-tools/bundles/variables.md).
|
The custom variable definitions for the target. See [_](/dev-tools/bundles/settings.md#variables) and [_](/dev-tools/bundles/variables.md).
|
||||||
"workspace":
|
"workspace":
|
||||||
"description": |-
|
"description": |-
|
||||||
The Databricks workspace for the target. _.
|
The Databricks workspace for the target.
|
||||||
"markdown_description": |-
|
"markdown_description": |-
|
||||||
The Databricks workspace for the target. [_](#workspace)
|
The Databricks workspace for the target. [_](#workspace)
|
||||||
github.com/databricks/cli/bundle/config.Workspace:
|
github.com/databricks/cli/bundle/config.Workspace:
|
||||||
|
@ -409,7 +413,7 @@ github.com/databricks/cli/bundle/config/variable.TargetVariable:
|
||||||
PLACEHOLDER
|
PLACEHOLDER
|
||||||
"description":
|
"description":
|
||||||
"description": |-
|
"description": |-
|
||||||
The description of the variable
|
The description of the variable.
|
||||||
"lookup":
|
"lookup":
|
||||||
"description": |-
|
"description": |-
|
||||||
The name of the alert, cluster_policy, cluster, dashboard, instance_pool, job, metastore, pipeline, query, service_principal, or warehouse object for which to retrieve an ID.
|
The name of the alert, cluster_policy, cluster, dashboard, instance_pool, job, metastore, pipeline, query, service_principal, or warehouse object for which to retrieve an ID.
|
||||||
|
@ -418,7 +422,7 @@ github.com/databricks/cli/bundle/config/variable.TargetVariable:
|
||||||
The type of the variable.
|
The type of the variable.
|
||||||
"markdown_description":
|
"markdown_description":
|
||||||
"description": |-
|
"description": |-
|
||||||
The type of the variable. Valid values are `complex`.
|
The type of the variable.
|
||||||
github.com/databricks/cli/bundle/config/variable.Variable:
|
github.com/databricks/cli/bundle/config/variable.Variable:
|
||||||
"default":
|
"default":
|
||||||
"description": |-
|
"description": |-
|
||||||
|
@ -429,9 +433,11 @@ github.com/databricks/cli/bundle/config/variable.Variable:
|
||||||
"lookup":
|
"lookup":
|
||||||
"description": |-
|
"description": |-
|
||||||
The name of the alert, cluster_policy, cluster, dashboard, instance_pool, job, metastore, pipeline, query, service_principal, or warehouse object for which to retrieve an ID.
|
The name of the alert, cluster_policy, cluster, dashboard, instance_pool, job, metastore, pipeline, query, service_principal, or warehouse object for which to retrieve an ID.
|
||||||
|
"markdown_description": |-
|
||||||
|
The name of the `alert`, `cluster_policy`, `cluster`, `dashboard`, `instance_pool`, `job`, `metastore`, `pipeline`, `query`, `service_principal`, or `warehouse` object for which to retrieve an ID."
|
||||||
"type":
|
"type":
|
||||||
"description": |-
|
"description": |-
|
||||||
The type of the variable. Valid values are complex.
|
The type of the variable.
|
||||||
github.com/databricks/databricks-sdk-go/service/serving.Ai21LabsConfig:
|
github.com/databricks/databricks-sdk-go/service/serving.Ai21LabsConfig:
|
||||||
"ai21labs_api_key":
|
"ai21labs_api_key":
|
||||||
"description": |-
|
"description": |-
|
||||||
|
|
|
@ -44,7 +44,8 @@ func addInterpolationPatterns(typ reflect.Type, s jsonschema.Schema) jsonschema.
|
||||||
case jsonschema.ArrayType, jsonschema.ObjectType:
|
case jsonschema.ArrayType, jsonschema.ObjectType:
|
||||||
// arrays and objects can have complex variable values specified.
|
// arrays and objects can have complex variable values specified.
|
||||||
return jsonschema.Schema{
|
return jsonschema.Schema{
|
||||||
AnyOf: []jsonschema.Schema{
|
// OneOf is used because we don't expect more than 1 match and schema-based auto-complete works better with OneOf
|
||||||
|
OneOf: []jsonschema.Schema{
|
||||||
s,
|
s,
|
||||||
{
|
{
|
||||||
Type: jsonschema.StringType,
|
Type: jsonschema.StringType,
|
||||||
|
@ -55,7 +56,7 @@ func addInterpolationPatterns(typ reflect.Type, s jsonschema.Schema) jsonschema.
|
||||||
// primitives can have variable values, or references like ${bundle.xyz}
|
// primitives can have variable values, or references like ${bundle.xyz}
|
||||||
// or ${workspace.xyz}
|
// or ${workspace.xyz}
|
||||||
return jsonschema.Schema{
|
return jsonschema.Schema{
|
||||||
AnyOf: []jsonschema.Schema{
|
OneOf: []jsonschema.Schema{
|
||||||
s,
|
s,
|
||||||
{Type: jsonschema.StringType, Pattern: interpolationPattern("resources")},
|
{Type: jsonschema.StringType, Pattern: interpolationPattern("resources")},
|
||||||
{Type: jsonschema.StringType, Pattern: interpolationPattern("bundle")},
|
{Type: jsonschema.StringType, Pattern: interpolationPattern("bundle")},
|
||||||
|
|
|
@ -0,0 +1,5 @@
|
||||||
|
targets:
|
||||||
|
production:
|
||||||
|
variables:
|
||||||
|
myvar:
|
||||||
|
default: true
|
|
@ -41,32 +41,32 @@ func TestJsonSchema(t *testing.T) {
|
||||||
resourceJob := walk(s.Definitions, "github.com", "databricks", "cli", "bundle", "config", "resources.Job")
|
resourceJob := walk(s.Definitions, "github.com", "databricks", "cli", "bundle", "config", "resources.Job")
|
||||||
fields := []string{"name", "continuous", "tasks", "trigger"}
|
fields := []string{"name", "continuous", "tasks", "trigger"}
|
||||||
for _, field := range fields {
|
for _, field := range fields {
|
||||||
assert.NotEmpty(t, resourceJob.AnyOf[0].Properties[field].Description)
|
assert.NotEmpty(t, resourceJob.OneOf[0].Properties[field].Description)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Assert descriptions were also loaded for a job task definition.
|
// Assert descriptions were also loaded for a job task definition.
|
||||||
jobTask := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "jobs.Task")
|
jobTask := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "jobs.Task")
|
||||||
fields = []string{"notebook_task", "spark_jar_task", "spark_python_task", "spark_submit_task", "description", "depends_on", "environment_key", "for_each_task", "existing_cluster_id"}
|
fields = []string{"notebook_task", "spark_jar_task", "spark_python_task", "spark_submit_task", "description", "depends_on", "environment_key", "for_each_task", "existing_cluster_id"}
|
||||||
for _, field := range fields {
|
for _, field := range fields {
|
||||||
assert.NotEmpty(t, jobTask.AnyOf[0].Properties[field].Description)
|
assert.NotEmpty(t, jobTask.OneOf[0].Properties[field].Description)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Assert descriptions are loaded for pipelines
|
// Assert descriptions are loaded for pipelines
|
||||||
pipeline := walk(s.Definitions, "github.com", "databricks", "cli", "bundle", "config", "resources.Pipeline")
|
pipeline := walk(s.Definitions, "github.com", "databricks", "cli", "bundle", "config", "resources.Pipeline")
|
||||||
fields = []string{"name", "catalog", "clusters", "channel", "continuous", "development"}
|
fields = []string{"name", "catalog", "clusters", "channel", "continuous", "development"}
|
||||||
for _, field := range fields {
|
for _, field := range fields {
|
||||||
assert.NotEmpty(t, pipeline.AnyOf[0].Properties[field].Description)
|
assert.NotEmpty(t, pipeline.OneOf[0].Properties[field].Description)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Assert enum values are loaded
|
// Assert enum values are loaded
|
||||||
schedule := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "pipelines.RestartWindow")
|
schedule := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "pipelines.RestartWindow")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "MONDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "MONDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "TUESDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "TUESDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "WEDNESDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "WEDNESDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "THURSDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "THURSDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "FRIDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "FRIDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "SATURDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "SATURDAY")
|
||||||
assert.Contains(t, schedule.AnyOf[0].Properties["days_of_week"].Enum, "SUNDAY")
|
assert.Contains(t, schedule.OneOf[0].Properties["days_of_week"].Enum, "SUNDAY")
|
||||||
|
|
||||||
providers := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "jobs.GitProvider")
|
providers := walk(s.Definitions, "github.com", "databricks", "databricks-sdk-go", "service", "jobs.GitProvider")
|
||||||
assert.Contains(t, providers.Enum, "gitHub")
|
assert.Contains(t, providers.Enum, "gitHub")
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
|
@ -70,6 +70,9 @@ type Schema struct {
|
||||||
// Schema that must match any of the schemas in the array
|
// Schema that must match any of the schemas in the array
|
||||||
AnyOf []Schema `json:"anyOf,omitempty"`
|
AnyOf []Schema `json:"anyOf,omitempty"`
|
||||||
|
|
||||||
|
// Schema that must match one of the schemas in the array
|
||||||
|
OneOf []Schema `json:"oneOf,omitempty"`
|
||||||
|
|
||||||
// Title of the object, rendered as inline documentation in the IDE.
|
// Title of the object, rendered as inline documentation in the IDE.
|
||||||
// https://json-schema.org/understanding-json-schema/reference/annotations
|
// https://json-schema.org/understanding-json-schema/reference/annotations
|
||||||
Title string `json:"title,omitempty"`
|
Title string `json:"title,omitempty"`
|
||||||
|
|
Loading…
Reference in New Issue