mirror of https://github.com/databricks/cli.git
Add top level docs for bundle json schema (#313)
Co-authored-by: Pieter Noordhuis <pieter.noordhuis@databricks.com> Co-authored-by: PaulCornellDB <paul.cornell@databricks.com>
This commit is contained in:
parent
bd11da88eb
commit
417839021b
|
@ -16,8 +16,9 @@ These descriptions are rendered in the inline documentation in an IDE
|
||||||
1. You can autogenerate empty descriptions for the new fields by running
|
1. You can autogenerate empty descriptions for the new fields by running
|
||||||
`bricks bundle schema --only-docs > ~/bricks/bundle/schema/docs/bundle_descriptions.json`
|
`bricks bundle schema --only-docs > ~/bricks/bundle/schema/docs/bundle_descriptions.json`
|
||||||
2. Manually edit bundle_descriptions.json to add your descriptions
|
2. Manually edit bundle_descriptions.json to add your descriptions
|
||||||
3. Again run `bricks bundle schema --only-docs > ~/bricks/bundle/schema/docs/bundle_descriptions.json` to copy over any applicable descriptions to `environments`
|
3. Build again to embed the new `bundle_descriptions.json` into the binary (`go build`)
|
||||||
4. push to repo
|
4. Again run `bricks bundle schema --only-docs > ~/bricks/bundle/schema/docs/bundle_descriptions.json` to copy over any applicable descriptions to `environments`
|
||||||
|
5. push to repo
|
||||||
|
|
||||||
|
|
||||||
### SOP: Update descriptions in resources from a newer openapi spec
|
### SOP: Update descriptions in resources from a newer openapi spec
|
||||||
|
|
|
@ -2,7 +2,7 @@
|
||||||
"description": "Root of the bundle config",
|
"description": "Root of the bundle config",
|
||||||
"properties": {
|
"properties": {
|
||||||
"artifacts": {
|
"artifacts": {
|
||||||
"description": "Artifacts contains a description of all code artifacts in this bundle.",
|
"description": "A description of all code artifacts in this bundle.",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -18,10 +18,10 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"bundle": {
|
"bundle": {
|
||||||
"description": "Bundle contains details about this bundle, such as its name,\nversion of the spec (TODO), default cluster, default warehouse, etc.\n",
|
"description": "The details for this bundle.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"name": {
|
"name": {
|
||||||
"description": ""
|
"description": "The name of the bundle."
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
@ -31,7 +31,7 @@
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
"artifacts": {
|
"artifacts": {
|
||||||
"description": "Artifacts contains a description of all code artifacts in this bundle.",
|
"description": "A description of all code artifacts in this bundle.",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -47,10 +47,10 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"bundle": {
|
"bundle": {
|
||||||
"description": "Bundle contains details about this bundle, such as its name,\nversion of the spec (TODO), default cluster, default warehouse, etc.\n",
|
"description": "The details for this bundle.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"name": {
|
"name": {
|
||||||
"description": ""
|
"description": "The name of the bundle."
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
|
@ -58,10 +58,10 @@
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"resources": {
|
"resources": {
|
||||||
"description": "Specification of databricks resources to instantiate",
|
"description": "Collection of Databricks resources to deploy.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"experiments": {
|
"experiments": {
|
||||||
"description": "",
|
"description": "List of MLflow experiments",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -1049,7 +1049,7 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"models": {
|
"models": {
|
||||||
"description": "",
|
"description": "List of MLflow models",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -1394,6 +1394,14 @@
|
||||||
"items": {
|
"items": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
"file": {
|
||||||
|
"description": "",
|
||||||
|
"properties": {
|
||||||
|
"path": {
|
||||||
|
"description": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
"jar": {
|
"jar": {
|
||||||
"description": "URI of the jar to be installed. Currently only DBFS and S3 URIs are supported.\nFor example: `{ \"jar\": \"dbfs:/mnt/databricks/library.jar\" }` or\n`{ \"jar\": \"s3://my-bucket/library.jar\" }`.\nIf S3 is used, please make sure the cluster has read access on the library. You may need to\nlaunch the cluster with an IAM role to access the S3 URI."
|
"description": "URI of the jar to be installed. Currently only DBFS and S3 URIs are supported.\nFor example: `{ \"jar\": \"dbfs:/mnt/databricks/library.jar\" }` or\n`{ \"jar\": \"s3://my-bucket/library.jar\" }`.\nIf S3 is used, please make sure the cluster has read access on the library. You may need to\nlaunch the cluster with an IAM role to access the S3 URI."
|
||||||
},
|
},
|
||||||
|
@ -1485,27 +1493,19 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"workspace": {
|
"workspace": {
|
||||||
"description": "",
|
"description": "Configures which workspace to connect to and locations for files, state, and similar locations within the workspace file tree.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"artifact_path": {
|
"artifact_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize build artifacts to. This defaults to `${workspace.root}/artifacts`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"azure_client_id": {
|
"azure_client_id": {
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"azure_environment": {
|
"azure_environment": {
|
||||||
"description": ""
|
"description": "Azure environment, one of (Public, UsGov, China, Germany)."
|
||||||
},
|
},
|
||||||
"azure_login_app_id": {
|
"azure_login_app_id": {
|
||||||
"description": ""
|
"description": "Azure Login Application ID."
|
||||||
},
|
},
|
||||||
"azure_tenant_id": {
|
"azure_tenant_id": {
|
||||||
"description": ""
|
"description": ""
|
||||||
|
@ -1514,41 +1514,25 @@
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"azure_workspace_resource_id": {
|
"azure_workspace_resource_id": {
|
||||||
"description": ""
|
"description": "Azure Resource Manager ID for Azure Databricks workspace."
|
||||||
},
|
},
|
||||||
"file_path": {
|
"file_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize local files artifacts to. This defaults to `${workspace.root}/files`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"google_service_account": {
|
"google_service_account": {
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"host": {
|
"host": {
|
||||||
"description": ""
|
"description": "Host url of the workspace."
|
||||||
},
|
},
|
||||||
"profile": {
|
"profile": {
|
||||||
"description": ""
|
"description": "Connection profile to use. By default profiles are specified in ~/.databrickscfg."
|
||||||
},
|
},
|
||||||
"root": {
|
"root_path": {
|
||||||
"description": ""
|
"description": "The base location for synchronizing files, artifacts and state. Defaults to `/Users/jane@doe.com/.bundle/${bundle.name}/${bundle.environment}`"
|
||||||
},
|
},
|
||||||
"state_path": {
|
"state_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize bundle state to. This defaults to `${workspace.root}/state`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1556,16 +1540,16 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"include": {
|
"include": {
|
||||||
"description": "",
|
"description": "A list of patterns of file names to load and merge into the this configuration. It defaults to loading `*.yml` and `*/*.yml`.",
|
||||||
"items": {
|
"items": {
|
||||||
"description": ""
|
"description": ""
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"resources": {
|
"resources": {
|
||||||
"description": "Specification of databricks resources to instantiate",
|
"description": "Collection of Databricks resources to deploy.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"experiments": {
|
"experiments": {
|
||||||
"description": "",
|
"description": "List of MLflow experiments",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -2553,7 +2537,7 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"models": {
|
"models": {
|
||||||
"description": "",
|
"description": "List of MLflow models",
|
||||||
"additionalproperties": {
|
"additionalproperties": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
@ -2898,6 +2882,14 @@
|
||||||
"items": {
|
"items": {
|
||||||
"description": "",
|
"description": "",
|
||||||
"properties": {
|
"properties": {
|
||||||
|
"file": {
|
||||||
|
"description": "",
|
||||||
|
"properties": {
|
||||||
|
"path": {
|
||||||
|
"description": ""
|
||||||
|
}
|
||||||
|
}
|
||||||
|
},
|
||||||
"jar": {
|
"jar": {
|
||||||
"description": "URI of the jar to be installed. Currently only DBFS and S3 URIs are supported.\nFor example: `{ \"jar\": \"dbfs:/mnt/databricks/library.jar\" }` or\n`{ \"jar\": \"s3://my-bucket/library.jar\" }`.\nIf S3 is used, please make sure the cluster has read access on the library. You may need to\nlaunch the cluster with an IAM role to access the S3 URI."
|
"description": "URI of the jar to be installed. Currently only DBFS and S3 URIs are supported.\nFor example: `{ \"jar\": \"dbfs:/mnt/databricks/library.jar\" }` or\n`{ \"jar\": \"s3://my-bucket/library.jar\" }`.\nIf S3 is used, please make sure the cluster has read access on the library. You may need to\nlaunch the cluster with an IAM role to access the S3 URI."
|
||||||
},
|
},
|
||||||
|
@ -2989,27 +2981,19 @@
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"workspace": {
|
"workspace": {
|
||||||
"description": "",
|
"description": "Configures which workspace to connect to and locations for files, state, and similar locations within the workspace file tree.",
|
||||||
"properties": {
|
"properties": {
|
||||||
"artifact_path": {
|
"artifact_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize build artifacts to. This defaults to `${workspace.root}/artifacts`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"azure_client_id": {
|
"azure_client_id": {
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"azure_environment": {
|
"azure_environment": {
|
||||||
"description": ""
|
"description": "Azure environment, one of (Public, UsGov, China, Germany)."
|
||||||
},
|
},
|
||||||
"azure_login_app_id": {
|
"azure_login_app_id": {
|
||||||
"description": ""
|
"description": "Azure Login Application ID."
|
||||||
},
|
},
|
||||||
"azure_tenant_id": {
|
"azure_tenant_id": {
|
||||||
"description": ""
|
"description": ""
|
||||||
|
@ -3018,43 +3002,27 @@
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"azure_workspace_resource_id": {
|
"azure_workspace_resource_id": {
|
||||||
"description": ""
|
"description": "Azure Resource Manager ID for Azure Databricks workspace."
|
||||||
},
|
},
|
||||||
"file_path": {
|
"file_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize local files artifacts to. This defaults to `${workspace.root}/files`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
},
|
},
|
||||||
"google_service_account": {
|
"google_service_account": {
|
||||||
"description": ""
|
"description": ""
|
||||||
},
|
},
|
||||||
"host": {
|
"host": {
|
||||||
"description": ""
|
"description": "Host url of the workspace."
|
||||||
},
|
},
|
||||||
"profile": {
|
"profile": {
|
||||||
"description": ""
|
"description": "Connection profile to use. By default profiles are specified in ~/.databrickscfg."
|
||||||
},
|
},
|
||||||
"root": {
|
"root_path": {
|
||||||
"description": ""
|
"description": "The base location for synchronizing files, artifacts and state. Defaults to `/Users/jane@doe.com/.bundle/${bundle.name}/${bundle.environment}`"
|
||||||
},
|
},
|
||||||
"state_path": {
|
"state_path": {
|
||||||
"description": "",
|
"description": "The remote path to synchronize bundle state to. This defaults to `${workspace.root}/state`"
|
||||||
"properties": {
|
|
||||||
"dbfs": {
|
|
||||||
"description": ""
|
|
||||||
},
|
|
||||||
"workspace": {
|
|
||||||
"description": ""
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
Loading…
Reference in New Issue