mirror of https://github.com/databricks/cli.git
Compare commits
7 Commits
6a5bfda10e
...
b9b1858f6d
Author | SHA1 | Date |
---|---|---|
|
b9b1858f6d | |
|
24ac8d8d59 | |
|
5d392acbef | |
|
272ce61302 | |
|
878fa80322 | |
|
8d849fe868 | |
|
e93f1ed8e8 |
|
@ -1,15 +1,19 @@
|
|||
{
|
||||
"formatter": "go run golang.org/x/tools/cmd/goimports@latest -w $FILENAMES && go fmt ./...",
|
||||
"formatter": "go run golang.org/x/tools/cmd/goimports@latest -w cmd && go fmt ./...",
|
||||
"services": {
|
||||
".codegen/service.go.tmpl": "cmd/{{if .IsAccounts}}account{{else}}workspace{{end}}/{{(.TrimPrefix \"account\").KebabName}}/{{(.TrimPrefix \"account\").KebabName}}.go"
|
||||
".codegen/service.go.tmpl": "cmd/{{if .IsAccounts}}account{{else}}workspace{{end}}/{{(.TrimPrefix \"account\").KebabName}}/{{(.TrimPrefix \"account\").KebabName}}.go",
|
||||
".codegen/cmd_script.tmpl": "acceptance/help/cmd/{{if .IsAccounts}}account{{else}}workspace{{end}}/{{(.TrimPrefix \"account\").KebabName}}/{{(.TrimPrefix \"account\").KebabName}}/script"
|
||||
},
|
||||
"batch": {
|
||||
".codegen/cmds-workspace.go.tmpl": "cmd/workspace/cmd.go",
|
||||
".codegen/cmds-account.go.tmpl": "cmd/account/cmd.go"
|
||||
},
|
||||
"toolchain": {
|
||||
"required": ["go"],
|
||||
"required": [
|
||||
"go"
|
||||
],
|
||||
"post_generate": [
|
||||
"go test ./acceptance -v -update -run 'TestAccept/help/cmd' ",
|
||||
"go test -timeout 240s -run TestConsistentDatabricksSdkVersion github.com/databricks/cli/internal/build",
|
||||
"make schema",
|
||||
"echo 'bundle/internal/tf/schema/\\*.go linguist-generated=true' >> ./.gitattributes",
|
||||
|
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
{{- $isAccount := .IsAccounts }}
|
||||
{{- $cmdGrp := (.TrimPrefix "account").KebabName }}
|
||||
{{- range .Methods}}
|
||||
trace $CLI {{if $isAccount }}account {{end}}{{$cmdGrp}} {{.KebabName}} --help
|
||||
{{- end}}
|
|
@ -1,3 +1,139 @@
|
|||
acceptance/help/cmd/account/access-control/access-control/script linguist-generated=true
|
||||
acceptance/help/cmd/account/billable-usage/billable-usage/script linguist-generated=true
|
||||
acceptance/help/cmd/account/budgets/budgets/script linguist-generated=true
|
||||
acceptance/help/cmd/account/credentials/credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/account/csp-enablement-account/csp-enablement-account/script linguist-generated=true
|
||||
acceptance/help/cmd/account/custom-app-integration/custom-app-integration/script linguist-generated=true
|
||||
acceptance/help/cmd/account/disable-legacy-features/disable-legacy-features/script linguist-generated=true
|
||||
acceptance/help/cmd/account/encryption-keys/encryption-keys/script linguist-generated=true
|
||||
acceptance/help/cmd/account/esm-enablement-account/esm-enablement-account/script linguist-generated=true
|
||||
acceptance/help/cmd/account/federation-policy/federation-policy/script linguist-generated=true
|
||||
acceptance/help/cmd/account/groups/groups/script linguist-generated=true
|
||||
acceptance/help/cmd/account/ip-access-lists/ip-access-lists/script linguist-generated=true
|
||||
acceptance/help/cmd/account/log-delivery/log-delivery/script linguist-generated=true
|
||||
acceptance/help/cmd/account/metastore-assignments/metastore-assignments/script linguist-generated=true
|
||||
acceptance/help/cmd/account/metastores/metastores/script linguist-generated=true
|
||||
acceptance/help/cmd/account/network-connectivity/network-connectivity/script linguist-generated=true
|
||||
acceptance/help/cmd/account/networks/networks/script linguist-generated=true
|
||||
acceptance/help/cmd/account/o-auth-published-apps/o-auth-published-apps/script linguist-generated=true
|
||||
acceptance/help/cmd/account/personal-compute/personal-compute/script linguist-generated=true
|
||||
acceptance/help/cmd/account/private-access/private-access/script linguist-generated=true
|
||||
acceptance/help/cmd/account/published-app-integration/published-app-integration/script linguist-generated=true
|
||||
acceptance/help/cmd/account/service-principal-federation-policy/service-principal-federation-policy/script linguist-generated=true
|
||||
acceptance/help/cmd/account/service-principal-secrets/service-principal-secrets/script linguist-generated=true
|
||||
acceptance/help/cmd/account/service-principals/service-principals/script linguist-generated=true
|
||||
acceptance/help/cmd/account/settings/settings/script linguist-generated=true
|
||||
acceptance/help/cmd/account/storage-credentials/storage-credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/account/storage/storage/script linguist-generated=true
|
||||
acceptance/help/cmd/account/usage-dashboards/usage-dashboards/script linguist-generated=true
|
||||
acceptance/help/cmd/account/users/users/script linguist-generated=true
|
||||
acceptance/help/cmd/account/vpc-endpoints/vpc-endpoints/script linguist-generated=true
|
||||
acceptance/help/cmd/account/workspace-assignment/workspace-assignment/script linguist-generated=true
|
||||
acceptance/help/cmd/account/workspaces/workspaces/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/access-control-proxy/access-control-proxy/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/access-control/access-control/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/aibi-dashboard-embedding-access-policy/aibi-dashboard-embedding-access-policy/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/aibi-dashboard-embedding-approved-domains/aibi-dashboard-embedding-approved-domains/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/alerts-legacy/alerts-legacy/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/alerts/alerts/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/apps/apps/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/artifact-allowlists/artifact-allowlists/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/automatic-cluster-update/automatic-cluster-update/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/catalogs/catalogs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/clean-room-assets/clean-room-assets/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/clean-room-task-runs/clean-room-task-runs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/clean-rooms/clean-rooms/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/cluster-policies/cluster-policies/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/clusters/clusters/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/command-execution/command-execution/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/compliance-security-profile/compliance-security-profile/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/connections/connections/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/consumer-fulfillments/consumer-fulfillments/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/consumer-installations/consumer-installations/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/consumer-listings/consumer-listings/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/consumer-personalization-requests/consumer-personalization-requests/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/consumer-providers/consumer-providers/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/credentials-manager/credentials-manager/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/credentials/credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/current-user/current-user/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/dashboard-widgets/dashboard-widgets/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/dashboards/dashboards/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/data-sources/data-sources/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/dbfs/dbfs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/dbsql-permissions/dbsql-permissions/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/default-namespace/default-namespace/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/disable-legacy-access/disable-legacy-access/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/disable-legacy-dbfs/disable-legacy-dbfs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/enhanced-security-monitoring/enhanced-security-monitoring/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/experiments/experiments/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/external-locations/external-locations/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/files/files/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/functions/functions/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/genie/genie/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/git-credentials/git-credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/global-init-scripts/global-init-scripts/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/grants/grants/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/groups/groups/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/instance-pools/instance-pools/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/instance-profiles/instance-profiles/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/ip-access-lists/ip-access-lists/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/jobs/jobs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/lakeview/lakeview/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/libraries/libraries/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/metastores/metastores/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/model-registry/model-registry/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/model-versions/model-versions/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/notification-destinations/notification-destinations/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/online-tables/online-tables/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/permission-migration/permission-migration/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/permissions/permissions/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/pipelines/pipelines/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/policy-compliance-for-clusters/policy-compliance-for-clusters/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/policy-compliance-for-jobs/policy-compliance-for-jobs/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/policy-families/policy-families/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-exchange-filters/provider-exchange-filters/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-exchanges/provider-exchanges/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-files/provider-files/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-listings/provider-listings/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-personalization-requests/provider-personalization-requests/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-provider-analytics-dashboards/provider-provider-analytics-dashboards/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/provider-providers/provider-providers/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/providers/providers/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/quality-monitors/quality-monitors/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/queries-legacy/queries-legacy/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/queries/queries/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/query-history/query-history/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/query-visualizations-legacy/query-visualizations-legacy/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/query-visualizations/query-visualizations/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/recipient-activation/recipient-activation/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/recipients/recipients/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/registered-models/registered-models/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/repos/repos/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/resource-quotas/resource-quotas/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/restrict-workspace-admins/restrict-workspace-admins/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/schemas/schemas/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/secrets/secrets/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/service-principals/service-principals/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/serving-endpoints-data-plane/serving-endpoints-data-plane/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/serving-endpoints/serving-endpoints/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/settings/settings/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/shares/shares/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/statement-execution/statement-execution/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/storage-credentials/storage-credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/system-schemas/system-schemas/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/table-constraints/table-constraints/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/tables/tables/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/temporary-table-credentials/temporary-table-credentials/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/token-management/token-management/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/tokens/tokens/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/users/users/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/vector-search-endpoints/vector-search-endpoints/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/vector-search-indexes/vector-search-indexes/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/volumes/volumes/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/warehouses/warehouses/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/workspace-bindings/workspace-bindings/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/workspace-conf/workspace-conf/script linguist-generated=true
|
||||
acceptance/help/cmd/workspace/workspace/workspace/script linguist-generated=true
|
||||
cmd/account/access-control/access-control.go linguist-generated=true
|
||||
cmd/account/billable-usage/billable-usage.go linguist-generated=true
|
||||
cmd/account/budget-policy/budget-policy.go linguist-generated=true
|
||||
|
|
4
NOTICE
4
NOTICE
|
@ -114,3 +114,7 @@ dario.cat/mergo
|
|||
Copyright (c) 2013 Dario Castañé. All rights reserved.
|
||||
Copyright (c) 2012 The Go Authors. All rights reserved.
|
||||
https://github.com/darccio/mergo/blob/master/LICENSE
|
||||
|
||||
https://github.com/gorilla/mux
|
||||
Copyright (c) 2023 The Gorilla Authors. All rights reserved.
|
||||
https://github.com/gorilla/mux/blob/main/LICENSE
|
||||
|
|
|
@ -11,6 +11,7 @@ import (
|
|||
"os"
|
||||
"os/exec"
|
||||
"path/filepath"
|
||||
"regexp"
|
||||
"runtime"
|
||||
"slices"
|
||||
"sort"
|
||||
|
@ -26,6 +27,7 @@ import (
|
|||
"github.com/databricks/cli/libs/testdiff"
|
||||
"github.com/databricks/cli/libs/testserver"
|
||||
"github.com/databricks/databricks-sdk-go"
|
||||
"github.com/databricks/databricks-sdk-go/service/iam"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
|
@ -72,7 +74,8 @@ func TestInprocessMode(t *testing.T) {
|
|||
if InprocessMode {
|
||||
t.Skip("Already tested by TestAccept")
|
||||
}
|
||||
require.Equal(t, 1, testAccept(t, true, "selftest"))
|
||||
require.Equal(t, 1, testAccept(t, true, "selftest/basic"))
|
||||
require.Equal(t, 1, testAccept(t, true, "selftest/server"))
|
||||
}
|
||||
|
||||
func testAccept(t *testing.T, InprocessMode bool, singleTest string) int {
|
||||
|
@ -118,14 +121,12 @@ func testAccept(t *testing.T, InprocessMode bool, singleTest string) int {
|
|||
uvCache := getUVDefaultCacheDir(t)
|
||||
t.Setenv("UV_CACHE_DIR", uvCache)
|
||||
|
||||
ctx := context.Background()
|
||||
cloudEnv := os.Getenv("CLOUD_ENV")
|
||||
|
||||
if cloudEnv == "" {
|
||||
defaultServer := testserver.New(t)
|
||||
AddHandlers(defaultServer)
|
||||
// Redirect API access to local server:
|
||||
t.Setenv("DATABRICKS_HOST", defaultServer.URL)
|
||||
t.Setenv("DATABRICKS_DEFAULT_HOST", defaultServer.URL)
|
||||
|
||||
homeDir := t.TempDir()
|
||||
// Do not read user's ~/.databrickscfg
|
||||
|
@ -148,27 +149,12 @@ func testAccept(t *testing.T, InprocessMode bool, singleTest string) int {
|
|||
// do it last so that full paths match first:
|
||||
repls.SetPath(buildDir, "[BUILD_DIR]")
|
||||
|
||||
var config databricks.Config
|
||||
if cloudEnv == "" {
|
||||
// use fake token for local tests
|
||||
config = databricks.Config{Token: "dbapi1234"}
|
||||
} else {
|
||||
// non-local tests rely on environment variables
|
||||
config = databricks.Config{}
|
||||
}
|
||||
workspaceClient, err := databricks.NewWorkspaceClient(&config)
|
||||
require.NoError(t, err)
|
||||
|
||||
user, err := workspaceClient.CurrentUser.Me(ctx)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, user)
|
||||
testdiff.PrepareReplacementsUser(t, &repls, *user)
|
||||
testdiff.PrepareReplacementsWorkspaceClient(t, &repls, workspaceClient)
|
||||
testdiff.PrepareReplacementsUUID(t, &repls)
|
||||
testdiff.PrepareReplacementsDevVersion(t, &repls)
|
||||
testdiff.PrepareReplacementSdkVersion(t, &repls)
|
||||
testdiff.PrepareReplacementsGoVersion(t, &repls)
|
||||
|
||||
repls.Repls = append(repls.Repls, testdiff.Replacement{Old: regexp.MustCompile("dbapi[0-9a-f]+"), New: "[DATABRICKS_TOKEN]"})
|
||||
|
||||
testDirs := getTests(t)
|
||||
require.NotEmpty(t, testDirs)
|
||||
|
||||
|
@ -180,8 +166,7 @@ func testAccept(t *testing.T, InprocessMode bool, singleTest string) int {
|
|||
}
|
||||
|
||||
for _, dir := range testDirs {
|
||||
testName := strings.ReplaceAll(dir, "\\", "/")
|
||||
t.Run(testName, func(t *testing.T) {
|
||||
t.Run(dir, func(t *testing.T) {
|
||||
if !InprocessMode {
|
||||
t.Parallel()
|
||||
}
|
||||
|
@ -203,7 +188,8 @@ func getTests(t *testing.T) []string {
|
|||
name := filepath.Base(path)
|
||||
if name == EntryPointScript {
|
||||
// Presence of 'script' marks a test case in this directory
|
||||
testDirs = append(testDirs, filepath.Dir(path))
|
||||
testName := filepath.ToSlash(filepath.Dir(path))
|
||||
testDirs = append(testDirs, testName)
|
||||
}
|
||||
return nil
|
||||
})
|
||||
|
@ -239,7 +225,6 @@ func runTest(t *testing.T, dir, coverDir string, repls testdiff.ReplacementsCont
|
|||
}
|
||||
|
||||
repls.SetPathWithParents(tmpDir, "[TMPDIR]")
|
||||
repls.Repls = append(repls.Repls, config.Repls...)
|
||||
|
||||
scriptContents := readMergedScriptContents(t, dir)
|
||||
testutil.WriteFile(t, filepath.Join(tmpDir, EntryPointScript), scriptContents)
|
||||
|
@ -253,38 +238,79 @@ func runTest(t *testing.T, dir, coverDir string, repls testdiff.ReplacementsCont
|
|||
cmd := exec.Command(args[0], args[1:]...)
|
||||
cmd.Env = os.Environ()
|
||||
|
||||
var workspaceClient *databricks.WorkspaceClient
|
||||
var user iam.User
|
||||
|
||||
// Start a new server with a custom configuration if the acceptance test
|
||||
// specifies a custom server stubs.
|
||||
var server *testserver.Server
|
||||
|
||||
// Start a new server for this test if either:
|
||||
// 1. A custom server spec is defined in the test configuration.
|
||||
// 2. The test is configured to record requests and assert on them. We need
|
||||
// a duplicate of the default server to record requests because the default
|
||||
// server otherwise is a shared resource.
|
||||
if len(config.Server) > 0 || config.RecordRequests {
|
||||
server = testserver.New(t)
|
||||
server.RecordRequests = config.RecordRequests
|
||||
server.IncludeRequestHeaders = config.IncludeRequestHeaders
|
||||
if cloudEnv == "" {
|
||||
// Start a new server for this test if either:
|
||||
// 1. A custom server spec is defined in the test configuration.
|
||||
// 2. The test is configured to record requests and assert on them. We need
|
||||
// a duplicate of the default server to record requests because the default
|
||||
// server otherwise is a shared resource.
|
||||
|
||||
// If no custom server stubs are defined, add the default handlers.
|
||||
if len(config.Server) == 0 {
|
||||
databricksLocalHost := os.Getenv("DATABRICKS_DEFAULT_HOST")
|
||||
|
||||
if len(config.Server) > 0 || config.RecordRequests {
|
||||
server = testserver.New(t)
|
||||
server.RecordRequests = config.RecordRequests
|
||||
server.IncludeRequestHeaders = config.IncludeRequestHeaders
|
||||
|
||||
for _, stub := range config.Server {
|
||||
require.NotEmpty(t, stub.Pattern)
|
||||
items := strings.Split(stub.Pattern, " ")
|
||||
require.Len(t, items, 2)
|
||||
server.Handle(items[0], items[1], func(fakeWorkspace *testserver.FakeWorkspace, req *http.Request) (any, int) {
|
||||
statusCode := http.StatusOK
|
||||
if stub.Response.StatusCode != 0 {
|
||||
statusCode = stub.Response.StatusCode
|
||||
}
|
||||
return stub.Response.Body, statusCode
|
||||
})
|
||||
}
|
||||
|
||||
// The earliest handlers take precedence, add default handlers last
|
||||
AddHandlers(server)
|
||||
databricksLocalHost = server.URL
|
||||
}
|
||||
|
||||
for _, stub := range config.Server {
|
||||
require.NotEmpty(t, stub.Pattern)
|
||||
server.Handle(stub.Pattern, func(fakeWorkspace *testserver.FakeWorkspace, req *http.Request) (any, int) {
|
||||
statusCode := http.StatusOK
|
||||
if stub.Response.StatusCode != 0 {
|
||||
statusCode = stub.Response.StatusCode
|
||||
}
|
||||
return stub.Response.Body, statusCode
|
||||
})
|
||||
// Each local test should use a new token that will result into a new fake workspace,
|
||||
// so that test don't interfere with each other.
|
||||
tokenSuffix := strings.ReplaceAll(uuid.NewString(), "-", "")
|
||||
config := databricks.Config{
|
||||
Host: databricksLocalHost,
|
||||
Token: "dbapi" + tokenSuffix,
|
||||
}
|
||||
cmd.Env = append(cmd.Env, "DATABRICKS_HOST="+server.URL)
|
||||
workspaceClient, err = databricks.NewWorkspaceClient(&config)
|
||||
require.NoError(t, err)
|
||||
|
||||
cmd.Env = append(cmd.Env, "DATABRICKS_HOST="+config.Host)
|
||||
cmd.Env = append(cmd.Env, "DATABRICKS_TOKEN="+config.Token)
|
||||
|
||||
// For the purposes of replacements, use testUser.
|
||||
// Note, users might have overriden /api/2.0/preview/scim/v2/Me but that should not affect the replacement:
|
||||
user = testUser
|
||||
} else {
|
||||
// Use whatever authentication mechanism is configured by the test runner.
|
||||
workspaceClient, err = databricks.NewWorkspaceClient(&databricks.Config{})
|
||||
require.NoError(t, err)
|
||||
pUser, err := workspaceClient.CurrentUser.Me(context.Background())
|
||||
require.NoError(t, err, "Failed to get current user")
|
||||
user = *pUser
|
||||
}
|
||||
|
||||
testdiff.PrepareReplacementsUser(t, &repls, user)
|
||||
testdiff.PrepareReplacementsWorkspaceClient(t, &repls, workspaceClient)
|
||||
|
||||
// Must be added PrepareReplacementsUser, otherwise conflicts with [USERNAME]
|
||||
testdiff.PrepareReplacementsUUID(t, &repls)
|
||||
|
||||
// User replacements come last:
|
||||
repls.Repls = append(repls.Repls, config.Repls...)
|
||||
|
||||
if coverDir != "" {
|
||||
// Creating individual coverage directory for each test, because writing to the same one
|
||||
// results in sporadic failures like this one (only if tests are running in parallel):
|
||||
|
@ -295,15 +321,6 @@ func runTest(t *testing.T, dir, coverDir string, repls testdiff.ReplacementsCont
|
|||
cmd.Env = append(cmd.Env, "GOCOVERDIR="+coverDir)
|
||||
}
|
||||
|
||||
// Each local test should use a new token that will result into a new fake workspace,
|
||||
// so that test don't interfere with each other.
|
||||
if cloudEnv == "" {
|
||||
tokenSuffix := strings.ReplaceAll(uuid.NewString(), "-", "")
|
||||
token := "dbapi" + tokenSuffix
|
||||
cmd.Env = append(cmd.Env, "DATABRICKS_TOKEN="+token)
|
||||
repls.Set(token, "[DATABRICKS_TOKEN]")
|
||||
}
|
||||
|
||||
// Write combined output to a file
|
||||
out, err := os.Create(filepath.Join(tmpDir, "output.txt"))
|
||||
require.NoError(t, err)
|
||||
|
@ -320,7 +337,7 @@ func runTest(t *testing.T, dir, coverDir string, repls testdiff.ReplacementsCont
|
|||
|
||||
for _, req := range server.Requests {
|
||||
reqJson, err := json.MarshalIndent(req, "", " ")
|
||||
require.NoError(t, err)
|
||||
require.NoErrorf(t, err, "Failed to indent: %#v", req)
|
||||
|
||||
reqJsonWithRepls := repls.Replace(string(reqJson))
|
||||
_, err = f.WriteString(reqJsonWithRepls + "\n")
|
||||
|
|
|
@ -13,13 +13,13 @@
|
|||
|
||||
=== Inside the bundle, profile flag not matching bundle host. Badness: should use profile from flag instead and not fail
|
||||
>>> errcode [CLI] current-user me -p profile_name
|
||||
Error: cannot resolve bundle auth configuration: config host mismatch: profile uses host https://non-existing-subdomain.databricks.com, but CLI configured to use [DATABRICKS_URL]
|
||||
Error: cannot resolve bundle auth configuration: config host mismatch: profile uses host https://non-existing-subdomain.databricks.com, but CLI configured to use [DATABRICKS_TARGET]
|
||||
|
||||
Exit code: 1
|
||||
|
||||
=== Inside the bundle, target and not matching profile
|
||||
>>> errcode [CLI] current-user me -t dev -p profile_name
|
||||
Error: cannot resolve bundle auth configuration: config host mismatch: profile uses host https://non-existing-subdomain.databricks.com, but CLI configured to use [DATABRICKS_URL]
|
||||
Error: cannot resolve bundle auth configuration: config host mismatch: profile uses host https://non-existing-subdomain.databricks.com, but CLI configured to use [DATABRICKS_TARGET]
|
||||
|
||||
Exit code: 1
|
||||
|
||||
|
|
|
@ -5,4 +5,8 @@ Badness = "When -p flag is used inside the bundle folder for any CLI commands, C
|
|||
# This is a workaround to replace DATABRICKS_URL with DATABRICKS_HOST
|
||||
[[Repls]]
|
||||
Old='DATABRICKS_HOST'
|
||||
New='DATABRICKS_URL'
|
||||
New='DATABRICKS_TARGET'
|
||||
|
||||
[[Repls]]
|
||||
Old='DATABRICKS_URL'
|
||||
New='DATABRICKS_TARGET'
|
||||
|
|
|
@ -0,0 +1,12 @@
|
|||
{
|
||||
"headers": {
|
||||
"Authorization": [
|
||||
"Basic [ENCODED_AUTH]"
|
||||
],
|
||||
"User-Agent": [
|
||||
"cli/[DEV_VERSION] databricks-sdk-go/[SDK_VERSION] go/[GO_VERSION] os/[OS] cmd/current-user_me cmd-exec-id/[UUID] auth/basic"
|
||||
]
|
||||
},
|
||||
"method": "GET",
|
||||
"path": "/api/2.0/preview/scim/v2/Me"
|
||||
}
|
|
@ -0,0 +1,4 @@
|
|||
{
|
||||
"id":"[USERID]",
|
||||
"userName":"[USERNAME]"
|
||||
}
|
|
@ -0,0 +1,8 @@
|
|||
# Unset the token which is configured by default
|
||||
# in acceptance tests
|
||||
export DATABRICKS_TOKEN=""
|
||||
|
||||
export DATABRICKS_USERNAME=username
|
||||
export DATABRICKS_PASSWORD=password
|
||||
|
||||
$CLI current-user me
|
|
@ -0,0 +1,4 @@
|
|||
# "username:password" in base64 is dXNlcm5hbWU6cGFzc3dvcmQ=, expect to see this in Authorization header
|
||||
[[Repls]]
|
||||
Old = "dXNlcm5hbWU6cGFzc3dvcmQ="
|
||||
New = "[ENCODED_AUTH]"
|
|
@ -0,0 +1,34 @@
|
|||
{
|
||||
"headers": {
|
||||
"User-Agent": [
|
||||
"cli/[DEV_VERSION] databricks-sdk-go/[SDK_VERSION] go/[GO_VERSION] os/[OS]"
|
||||
]
|
||||
},
|
||||
"method": "GET",
|
||||
"path": "/oidc/.well-known/oauth-authorization-server"
|
||||
}
|
||||
{
|
||||
"headers": {
|
||||
"Authorization": [
|
||||
"Basic [ENCODED_AUTH]"
|
||||
],
|
||||
"User-Agent": [
|
||||
"cli/[DEV_VERSION] databricks-sdk-go/[SDK_VERSION] go/[GO_VERSION] os/[OS]"
|
||||
]
|
||||
},
|
||||
"method": "POST",
|
||||
"path": "/oidc/v1/token",
|
||||
"raw_body": "grant_type=client_credentials\u0026scope=all-apis"
|
||||
}
|
||||
{
|
||||
"headers": {
|
||||
"Authorization": [
|
||||
"Bearer oauth-token"
|
||||
],
|
||||
"User-Agent": [
|
||||
"cli/[DEV_VERSION] databricks-sdk-go/[SDK_VERSION] go/[GO_VERSION] os/[OS] cmd/current-user_me cmd-exec-id/[UUID] auth/oauth-m2m"
|
||||
]
|
||||
},
|
||||
"method": "GET",
|
||||
"path": "/api/2.0/preview/scim/v2/Me"
|
||||
}
|
|
@ -0,0 +1,4 @@
|
|||
{
|
||||
"id":"[USERID]",
|
||||
"userName":"[USERNAME]"
|
||||
}
|
|
@ -0,0 +1,8 @@
|
|||
# Unset the token which is configured by default
|
||||
# in acceptance tests
|
||||
export DATABRICKS_TOKEN=""
|
||||
|
||||
export DATABRICKS_CLIENT_ID=client_id
|
||||
export DATABRICKS_CLIENT_SECRET=client_secret
|
||||
|
||||
$CLI current-user me
|
|
@ -0,0 +1,5 @@
|
|||
# "client_id:client_secret" in base64 is Y2xpZW50X2lkOmNsaWVudF9zZWNyZXQ=, expect to
|
||||
# see this in Authorization header
|
||||
[[Repls]]
|
||||
Old = "Y2xpZW50X2lkOmNsaWVudF9zZWNyZXQ="
|
||||
New = "[ENCODED_AUTH]"
|
|
@ -0,0 +1,12 @@
|
|||
{
|
||||
"headers": {
|
||||
"Authorization": [
|
||||
"Bearer dapi1234"
|
||||
],
|
||||
"User-Agent": [
|
||||
"cli/[DEV_VERSION] databricks-sdk-go/[SDK_VERSION] go/[GO_VERSION] os/[OS] cmd/current-user_me cmd-exec-id/[UUID] auth/pat"
|
||||
]
|
||||
},
|
||||
"method": "GET",
|
||||
"path": "/api/2.0/preview/scim/v2/Me"
|
||||
}
|
|
@ -0,0 +1,4 @@
|
|||
{
|
||||
"id":"[USERID]",
|
||||
"userName":"[USERNAME]"
|
||||
}
|
|
@ -0,0 +1,3 @@
|
|||
export DATABRICKS_TOKEN=dapi1234
|
||||
|
||||
$CLI current-user me
|
|
@ -0,0 +1,20 @@
|
|||
LocalOnly = true
|
||||
|
||||
RecordRequests = true
|
||||
IncludeRequestHeaders = ["Authorization", "User-Agent"]
|
||||
|
||||
[[Repls]]
|
||||
Old = '(linux|darwin|windows)'
|
||||
New = '[OS]'
|
||||
|
||||
[[Repls]]
|
||||
Old = " upstream/[A-Za-z0-9.-]+"
|
||||
New = ""
|
||||
|
||||
[[Repls]]
|
||||
Old = " upstream-version/[A-Za-z0-9.-]+"
|
||||
New = ""
|
||||
|
||||
[[Repls]]
|
||||
Old = " cicd/[A-Za-z0-9.-]+"
|
||||
New = ""
|
|
@ -14,11 +14,7 @@ import (
|
|||
|
||||
func StartCmdServer(t *testing.T) *testserver.Server {
|
||||
server := testserver.New(t)
|
||||
|
||||
// {$} is a wildcard that only matches the end of the URL. We explicitly use
|
||||
// /{$} to disambiguate it from the generic handler for '/' which is used to
|
||||
// identify unhandled API endpoints in the test server.
|
||||
server.Handle("/{$}", func(w *testserver.FakeWorkspace, r *http.Request) (any, int) {
|
||||
server.Handle("GET", "/", func(_ *testserver.FakeWorkspace, r *http.Request) (any, int) {
|
||||
q := r.URL.Query()
|
||||
args := strings.Split(q.Get("args"), " ")
|
||||
|
||||
|
|
|
@ -0,0 +1,72 @@
|
|||
|
||||
>>> $CLI account access-control get-assignable-roles-for-resource --help
|
||||
Get assignable roles for a resource.
|
||||
|
||||
Gets all the roles that can be granted on an account level resource. A role is
|
||||
grantable if the rule set on the resource can contain an access rule of the
|
||||
role.
|
||||
|
||||
Arguments:
|
||||
RESOURCE: The resource name for which assignable roles will be listed.
|
||||
|
||||
Usage:
|
||||
databricks account access-control get-assignable-roles-for-resource RESOURCE [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-assignable-roles-for-resource
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account access-control get-rule-set --help
|
||||
Get a rule set.
|
||||
|
||||
Get a rule set by its name. A rule set is always attached to a resource and
|
||||
contains a list of access rules on the said resource. Currently only a default
|
||||
rule set for each resource is supported.
|
||||
|
||||
Arguments:
|
||||
NAME: The ruleset name associated with the request.
|
||||
ETAG: Etag used for versioning. The response is at least as fresh as the eTag
|
||||
provided. Etag is used for optimistic concurrency control as a way to help
|
||||
prevent simultaneous updates of a rule set from overwriting each other. It
|
||||
is strongly suggested that systems make use of the etag in the read ->
|
||||
modify -> write pattern to perform rule set updates in order to avoid race
|
||||
conditions that is get an etag from a GET rule set request, and pass it
|
||||
with the PUT update request to identify the rule set version you are
|
||||
updating.
|
||||
|
||||
Usage:
|
||||
databricks account access-control get-rule-set NAME ETAG [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-rule-set
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account access-control update-rule-set --help
|
||||
Update a rule set.
|
||||
|
||||
Replace the rules of a rule set. First, use get to read the current version of
|
||||
the rule set before modifying it. This pattern helps prevent conflicts between
|
||||
concurrent updates.
|
||||
|
||||
Usage:
|
||||
databricks account access-control update-rule-set [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update-rule-set
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account access-control get-assignable-roles-for-resource --help
|
||||
trace $CLI account access-control get-rule-set --help
|
||||
trace $CLI account access-control update-rule-set --help
|
|
@ -0,0 +1,33 @@
|
|||
|
||||
>>> $CLI account billable-usage download --help
|
||||
Return billable usage logs.
|
||||
|
||||
Returns billable usage logs in CSV format for the specified account and date
|
||||
range. For the data schema, see [CSV file schema]. Note that this method might
|
||||
take multiple minutes to complete.
|
||||
|
||||
**Warning**: Depending on the queried date range, the number of workspaces in
|
||||
the account, the size of the response and the internet speed of the caller,
|
||||
this API may hit a timeout after a few minutes. If you experience this, try to
|
||||
mitigate by calling the API with narrower date ranges.
|
||||
|
||||
[CSV file schema]: https://docs.databricks.com/administration-guide/account-settings/usage-analysis.html#schema
|
||||
|
||||
Arguments:
|
||||
START_MONTH: Format: YYYY-MM. First month to return billable usage logs for. This
|
||||
field is required.
|
||||
END_MONTH: Format: YYYY-MM. Last month to return billable usage logs for. This
|
||||
field is required.
|
||||
|
||||
Usage:
|
||||
databricks account billable-usage download START_MONTH END_MONTH [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for download
|
||||
--personal-data Specify whether to include personally identifiable information in the billable usage logs, for example the email addresses of cluster creators.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,2 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account billable-usage download --help
|
|
@ -0,0 +1,101 @@
|
|||
|
||||
>>> $CLI account budgets create --help
|
||||
Create new budget.
|
||||
|
||||
Create a new budget configuration for an account. For full details, see
|
||||
https://docs.databricks.com/en/admin/account-settings/budgets.html.
|
||||
|
||||
Usage:
|
||||
databricks account budgets create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account budgets delete --help
|
||||
Delete budget.
|
||||
|
||||
Deletes a budget configuration for an account. Both account and budget
|
||||
configuration are specified by ID. This cannot be undone.
|
||||
|
||||
Arguments:
|
||||
BUDGET_ID: The Databricks budget configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account budgets delete BUDGET_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account budgets get --help
|
||||
Get budget.
|
||||
|
||||
Gets a budget configuration for an account. Both account and budget
|
||||
configuration are specified by ID.
|
||||
|
||||
Arguments:
|
||||
BUDGET_ID: The budget configuration ID
|
||||
|
||||
Usage:
|
||||
databricks account budgets get BUDGET_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account budgets list --help
|
||||
Get all budgets.
|
||||
|
||||
Gets all budgets associated with this account.
|
||||
|
||||
Usage:
|
||||
databricks account budgets list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-token string A page token received from a previous get all budget configurations call.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account budgets update --help
|
||||
Modify budget.
|
||||
|
||||
Updates a budget configuration for an account. Both account and budget
|
||||
configuration are specified by ID.
|
||||
|
||||
Arguments:
|
||||
BUDGET_ID: The Databricks budget configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account budgets update BUDGET_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account budgets create --help
|
||||
trace $CLI account budgets delete --help
|
||||
trace $CLI account budgets get --help
|
||||
trace $CLI account budgets list --help
|
||||
trace $CLI account budgets update --help
|
|
@ -0,0 +1,92 @@
|
|||
|
||||
>>> $CLI account credentials create --help
|
||||
Create credential configuration.
|
||||
|
||||
Creates a Databricks credential configuration that represents cloud
|
||||
cross-account credentials for a specified account. Databricks uses this to set
|
||||
up network infrastructure properly to host Databricks clusters. For your AWS
|
||||
IAM role, you need to trust the External ID (the Databricks Account API
|
||||
account ID) in the returned credential object, and configure the required
|
||||
access policy.
|
||||
|
||||
Save the response's credentials_id field, which is the ID for your new
|
||||
credential configuration object.
|
||||
|
||||
For information about how to create a new workspace with this API, see [Create
|
||||
a new workspace using the Account API]
|
||||
|
||||
[Create a new workspace using the Account API]: http://docs.databricks.com/administration-guide/account-api/new-workspace.html
|
||||
|
||||
Usage:
|
||||
databricks account credentials create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account credentials delete --help
|
||||
Delete credential configuration.
|
||||
|
||||
Deletes a Databricks credential configuration object for an account, both
|
||||
specified by ID. You cannot delete a credential that is associated with any
|
||||
workspace.
|
||||
|
||||
Arguments:
|
||||
CREDENTIALS_ID: Databricks Account API credential configuration ID
|
||||
|
||||
Usage:
|
||||
databricks account credentials delete CREDENTIALS_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account credentials get --help
|
||||
Get credential configuration.
|
||||
|
||||
Gets a Databricks credential configuration object for an account, both
|
||||
specified by ID.
|
||||
|
||||
Arguments:
|
||||
CREDENTIALS_ID: Databricks Account API credential configuration ID
|
||||
|
||||
Usage:
|
||||
databricks account credentials get CREDENTIALS_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account credentials list --help
|
||||
Get all credential configurations.
|
||||
|
||||
Gets all Databricks credential configurations associated with an account
|
||||
specified by ID.
|
||||
|
||||
Usage:
|
||||
databricks account credentials list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account credentials create --help
|
||||
trace $CLI account credentials delete --help
|
||||
trace $CLI account credentials get --help
|
||||
trace $CLI account credentials list --help
|
|
@ -0,0 +1,110 @@
|
|||
|
||||
>>> $CLI account csp-enablement-account get --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account csp-enablement-account update --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
3
acceptance/help/cmd/account/csp-enablement-account/csp-enablement-account/script
generated
Executable file
3
acceptance/help/cmd/account/csp-enablement-account/csp-enablement-account/script
generated
Executable file
|
@ -0,0 +1,3 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account csp-enablement-account get --help
|
||||
trace $CLI account csp-enablement-account update --help
|
|
@ -0,0 +1,101 @@
|
|||
|
||||
>>> $CLI account custom-app-integration create --help
|
||||
Create Custom OAuth App Integration.
|
||||
|
||||
Create Custom OAuth App Integration.
|
||||
|
||||
You can retrieve the custom OAuth app integration via
|
||||
:method:CustomAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account custom-app-integration create [flags]
|
||||
|
||||
Flags:
|
||||
--confidential This field indicates whether an OAuth client secret is required to authenticate this client.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--name string Name of the custom OAuth app.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account custom-app-integration delete --help
|
||||
Delete Custom OAuth App Integration.
|
||||
|
||||
Delete an existing Custom OAuth App Integration. You can retrieve the custom
|
||||
OAuth app integration via :method:CustomAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account custom-app-integration delete INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account custom-app-integration get --help
|
||||
Get OAuth Custom App Integration.
|
||||
|
||||
Gets the Custom OAuth App Integration for the given integration id.
|
||||
|
||||
Arguments:
|
||||
INTEGRATION_ID: The OAuth app integration ID.
|
||||
|
||||
Usage:
|
||||
databricks account custom-app-integration get INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account custom-app-integration list --help
|
||||
Get custom oauth app integrations.
|
||||
|
||||
Get the list of custom OAuth app integrations for the specified Databricks
|
||||
account
|
||||
|
||||
Usage:
|
||||
databricks account custom-app-integration list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--include-creator-username
|
||||
--page-size int
|
||||
--page-token string
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account custom-app-integration update --help
|
||||
Updates Custom OAuth App Integration.
|
||||
|
||||
Updates an existing custom OAuth App Integration. You can retrieve the custom
|
||||
OAuth app integration via :method:CustomAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account custom-app-integration update INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
6
acceptance/help/cmd/account/custom-app-integration/custom-app-integration/script
generated
Executable file
6
acceptance/help/cmd/account/custom-app-integration/custom-app-integration/script
generated
Executable file
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account custom-app-integration create --help
|
||||
trace $CLI account custom-app-integration delete --help
|
||||
trace $CLI account custom-app-integration get --help
|
||||
trace $CLI account custom-app-integration list --help
|
||||
trace $CLI account custom-app-integration update --help
|
|
@ -0,0 +1,165 @@
|
|||
|
||||
>>> $CLI account disable-legacy-features delete --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account disable-legacy-features get --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account disable-legacy-features update --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
4
acceptance/help/cmd/account/disable-legacy-features/disable-legacy-features/script
generated
Executable file
4
acceptance/help/cmd/account/disable-legacy-features/disable-legacy-features/script
generated
Executable file
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account disable-legacy-features delete --help
|
||||
trace $CLI account disable-legacy-features get --help
|
||||
trace $CLI account disable-legacy-features update --help
|
|
@ -0,0 +1,118 @@
|
|||
|
||||
>>> $CLI account encryption-keys create --help
|
||||
Create encryption key configuration.
|
||||
|
||||
Creates a customer-managed key configuration object for an account, specified
|
||||
by ID. This operation uploads a reference to a customer-managed key to
|
||||
Databricks. If the key is assigned as a workspace's customer-managed key for
|
||||
managed services, Databricks uses the key to encrypt the workspaces notebooks
|
||||
and secrets in the control plane, in addition to Databricks SQL queries and
|
||||
query history. If it is specified as a workspace's customer-managed key for
|
||||
workspace storage, the key encrypts the workspace's root S3 bucket (which
|
||||
contains the workspace's root DBFS and system data) and, optionally, cluster
|
||||
EBS volume data.
|
||||
|
||||
**Important**: Customer-managed keys are supported only for some deployment
|
||||
types, subscription types, and AWS regions that currently support creation of
|
||||
Databricks workspaces.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform or on a select custom plan that allows multiple workspaces per
|
||||
account.
|
||||
|
||||
Usage:
|
||||
databricks account encryption-keys create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account encryption-keys delete --help
|
||||
Delete encryption key configuration.
|
||||
|
||||
Deletes a customer-managed key configuration object for an account. You cannot
|
||||
delete a configuration that is associated with a running workspace.
|
||||
|
||||
Arguments:
|
||||
CUSTOMER_MANAGED_KEY_ID: Databricks encryption key configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account encryption-keys delete CUSTOMER_MANAGED_KEY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account encryption-keys get --help
|
||||
Get encryption key configuration.
|
||||
|
||||
Gets a customer-managed key configuration object for an account, specified by
|
||||
ID. This operation uploads a reference to a customer-managed key to
|
||||
Databricks. If assigned as a workspace's customer-managed key for managed
|
||||
services, Databricks uses the key to encrypt the workspaces notebooks and
|
||||
secrets in the control plane, in addition to Databricks SQL queries and query
|
||||
history. If it is specified as a workspace's customer-managed key for storage,
|
||||
the key encrypts the workspace's root S3 bucket (which contains the
|
||||
workspace's root DBFS and system data) and, optionally, cluster EBS volume
|
||||
data.
|
||||
|
||||
**Important**: Customer-managed keys are supported only for some deployment
|
||||
types, subscription types, and AWS regions.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform.",
|
||||
|
||||
Arguments:
|
||||
CUSTOMER_MANAGED_KEY_ID: Databricks encryption key configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account encryption-keys get CUSTOMER_MANAGED_KEY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account encryption-keys list --help
|
||||
Get all encryption key configurations.
|
||||
|
||||
Gets all customer-managed key configuration objects for an account. If the key
|
||||
is specified as a workspace's managed services customer-managed key,
|
||||
Databricks uses the key to encrypt the workspace's notebooks and secrets in
|
||||
the control plane, in addition to Databricks SQL queries and query history. If
|
||||
the key is specified as a workspace's storage customer-managed key, the key is
|
||||
used to encrypt the workspace's root S3 bucket and optionally can encrypt
|
||||
cluster EBS volumes data in the data plane.
|
||||
|
||||
**Important**: Customer-managed keys are supported only for some deployment
|
||||
types, subscription types, and AWS regions.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform.
|
||||
|
||||
Usage:
|
||||
databricks account encryption-keys list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account encryption-keys create --help
|
||||
trace $CLI account encryption-keys delete --help
|
||||
trace $CLI account encryption-keys get --help
|
||||
trace $CLI account encryption-keys list --help
|
|
@ -0,0 +1,110 @@
|
|||
|
||||
>>> $CLI account esm-enablement-account get --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account esm-enablement-account update --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
3
acceptance/help/cmd/account/esm-enablement-account/esm-enablement-account/script
generated
Executable file
3
acceptance/help/cmd/account/esm-enablement-account/esm-enablement-account/script
generated
Executable file
|
@ -0,0 +1,3 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account esm-enablement-account get --help
|
||||
trace $CLI account esm-enablement-account update --help
|
|
@ -0,0 +1,94 @@
|
|||
|
||||
>>> $CLI account federation-policy create --help
|
||||
Create account federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account federation-policy create [flags]
|
||||
|
||||
Flags:
|
||||
--description string Description of the federation policy.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--name string Resource name for the federation policy.
|
||||
--policy-id string The identifier for the federation policy.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account federation-policy delete --help
|
||||
Delete account federation policy.
|
||||
|
||||
Arguments:
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account federation-policy delete POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account federation-policy get --help
|
||||
Get account federation policy.
|
||||
|
||||
Arguments:
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account federation-policy get POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account federation-policy list --help
|
||||
List account federation policies.
|
||||
|
||||
Usage:
|
||||
databricks account federation-policy list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int
|
||||
--page-token string
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account federation-policy update --help
|
||||
Update account federation policy.
|
||||
|
||||
Arguments:
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account federation-policy update POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
--description string Description of the federation policy.
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--name string Resource name for the federation policy.
|
||||
--update-mask string The field mask specifies which fields of the policy to update.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account federation-policy create --help
|
||||
trace $CLI account federation-policy delete --help
|
||||
trace $CLI account federation-policy get --help
|
||||
trace $CLI account federation-policy list --help
|
||||
trace $CLI account federation-policy update --help
|
|
@ -0,0 +1,131 @@
|
|||
|
||||
>>> $CLI account groups create --help
|
||||
Create a new group.
|
||||
|
||||
Creates a group in the Databricks account with a unique name, using the
|
||||
supplied group details.
|
||||
|
||||
Usage:
|
||||
databricks account groups create [flags]
|
||||
|
||||
Flags:
|
||||
--display-name string String that represents a human-readable group name.
|
||||
--external-id string
|
||||
-h, --help help for create
|
||||
--id string Databricks group ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account groups delete --help
|
||||
Delete a group.
|
||||
|
||||
Deletes a group from the Databricks account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a group in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account groups delete ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account groups get --help
|
||||
Get group details.
|
||||
|
||||
Gets the information for a specific group in the Databricks account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a group in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account groups get ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account groups list --help
|
||||
List group details.
|
||||
|
||||
Gets all details of the groups associated with the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account groups list [flags]
|
||||
|
||||
Flags:
|
||||
--attributes string Comma-separated list of attributes to return in response.
|
||||
--count int Desired number of results per page.
|
||||
--excluded-attributes string Comma-separated list of attributes to exclude in response.
|
||||
--filter string Query by which the results have to be filtered.
|
||||
-h, --help help for list
|
||||
--sort-by string Attribute to sort the results.
|
||||
--sort-order ListSortOrder The order to sort the results. Supported values: [ascending, descending]
|
||||
--start-index int Specifies the index of the first result.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account groups patch --help
|
||||
Update group details.
|
||||
|
||||
Partially updates the details of a group.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a group in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account groups patch ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for patch
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account groups update --help
|
||||
Replace a group.
|
||||
|
||||
Updates the details of a group by replacing the entire group entity.
|
||||
|
||||
Arguments:
|
||||
ID: Databricks group ID
|
||||
|
||||
Usage:
|
||||
databricks account groups update ID [flags]
|
||||
|
||||
Flags:
|
||||
--display-name string String that represents a human-readable group name.
|
||||
--external-id string
|
||||
-h, --help help for update
|
||||
--id string Databricks group ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,7 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account groups create --help
|
||||
trace $CLI account groups delete --help
|
||||
trace $CLI account groups get --help
|
||||
trace $CLI account groups list --help
|
||||
trace $CLI account groups patch --help
|
||||
trace $CLI account groups update --help
|
|
@ -0,0 +1,172 @@
|
|||
|
||||
>>> $CLI account ip-access-lists create --help
|
||||
Create access list.
|
||||
|
||||
Creates an IP access list for the account.
|
||||
|
||||
A list can be an allow list or a block list. See the top of this file for a
|
||||
description of how the server treats allow lists and block lists at runtime.
|
||||
|
||||
When creating or updating an IP access list:
|
||||
|
||||
* For all allow lists and block lists combined, the API supports a maximum of
|
||||
1000 IP/CIDR values, where one CIDR counts as a single value. Attempts to
|
||||
exceed that number return error 400 with error_code value QUOTA_EXCEEDED.
|
||||
* If the new list would block the calling user's current IP, error 400 is
|
||||
returned with error_code value INVALID_STATE.
|
||||
|
||||
It can take a few minutes for the changes to take effect.
|
||||
|
||||
Arguments:
|
||||
LABEL: Label for the IP access list. This **cannot** be empty.
|
||||
LIST_TYPE: Type of IP access list. Valid values are as follows and are
|
||||
case-sensitive:
|
||||
|
||||
* ALLOW: An allow list. Include this IP or range. * BLOCK: A block
|
||||
list. Exclude this IP or range. IP addresses in the block list are
|
||||
excluded even if they are included in an allow list.
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists create LABEL LIST_TYPE [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account ip-access-lists delete --help
|
||||
Delete access list.
|
||||
|
||||
Deletes an IP access list, specified by its list ID.
|
||||
|
||||
Arguments:
|
||||
IP_ACCESS_LIST_ID: The ID for the corresponding IP access list
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists delete IP_ACCESS_LIST_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account ip-access-lists get --help
|
||||
Get IP access list.
|
||||
|
||||
Gets an IP access list, specified by its list ID.
|
||||
|
||||
Arguments:
|
||||
IP_ACCESS_LIST_ID: The ID for the corresponding IP access list
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists get IP_ACCESS_LIST_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account ip-access-lists list --help
|
||||
Get access lists.
|
||||
|
||||
Gets all IP access lists for the specified account.
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account ip-access-lists replace --help
|
||||
Replace access list.
|
||||
|
||||
Replaces an IP access list, specified by its ID.
|
||||
|
||||
A list can include allow lists and block lists. See the top of this file for a
|
||||
description of how the server treats allow lists and block lists at run time.
|
||||
When replacing an IP access list: * For all allow lists and block lists
|
||||
combined, the API supports a maximum of 1000 IP/CIDR values, where one CIDR
|
||||
counts as a single value. Attempts to exceed that number return error 400 with
|
||||
error_code value QUOTA_EXCEEDED. * If the resulting list would block the
|
||||
calling user's current IP, error 400 is returned with error_code value
|
||||
INVALID_STATE. It can take a few minutes for the changes to take effect.
|
||||
|
||||
Arguments:
|
||||
IP_ACCESS_LIST_ID: The ID for the corresponding IP access list
|
||||
LABEL: Label for the IP access list. This **cannot** be empty.
|
||||
LIST_TYPE: Type of IP access list. Valid values are as follows and are
|
||||
case-sensitive:
|
||||
|
||||
* ALLOW: An allow list. Include this IP or range. * BLOCK: A block
|
||||
list. Exclude this IP or range. IP addresses in the block list are
|
||||
excluded even if they are included in an allow list.
|
||||
ENABLED: Specifies whether this IP access list is enabled.
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists replace IP_ACCESS_LIST_ID LABEL LIST_TYPE ENABLED [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for replace
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account ip-access-lists update --help
|
||||
Update access list.
|
||||
|
||||
Updates an existing IP access list, specified by its ID.
|
||||
|
||||
A list can include allow lists and block lists. See the top of this file for a
|
||||
description of how the server treats allow lists and block lists at run time.
|
||||
|
||||
When updating an IP access list:
|
||||
|
||||
* For all allow lists and block lists combined, the API supports a maximum of
|
||||
1000 IP/CIDR values, where one CIDR counts as a single value. Attempts to
|
||||
exceed that number return error 400 with error_code value QUOTA_EXCEEDED.
|
||||
* If the updated list would block the calling user's current IP, error 400 is
|
||||
returned with error_code value INVALID_STATE.
|
||||
|
||||
It can take a few minutes for the changes to take effect.
|
||||
|
||||
Arguments:
|
||||
IP_ACCESS_LIST_ID: The ID for the corresponding IP access list
|
||||
|
||||
Usage:
|
||||
databricks account ip-access-lists update IP_ACCESS_LIST_ID [flags]
|
||||
|
||||
Flags:
|
||||
--enabled Specifies whether this IP access list is enabled.
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--label string Label for the IP access list.
|
||||
--list-type ListType Type of IP access list. Supported values: [ALLOW, BLOCK]
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,7 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account ip-access-lists create --help
|
||||
trace $CLI account ip-access-lists delete --help
|
||||
trace $CLI account ip-access-lists get --help
|
||||
trace $CLI account ip-access-lists list --help
|
||||
trace $CLI account ip-access-lists replace --help
|
||||
trace $CLI account ip-access-lists update --help
|
|
@ -0,0 +1,114 @@
|
|||
|
||||
>>> $CLI account log-delivery create --help
|
||||
Create a new log delivery configuration.
|
||||
|
||||
Creates a new Databricks log delivery configuration to enable delivery of the
|
||||
specified type of logs to your storage location. This requires that you
|
||||
already created a [credential object](:method:Credentials/Create) (which
|
||||
encapsulates a cross-account service IAM role) and a [storage configuration
|
||||
object](:method:Storage/Create) (which encapsulates an S3 bucket).
|
||||
|
||||
For full details, including the required IAM role policies and bucket
|
||||
policies, see [Deliver and access billable usage logs] or [Configure audit
|
||||
logging].
|
||||
|
||||
**Note**: There is a limit on the number of log delivery configurations
|
||||
available per account (each limit applies separately to each log type
|
||||
including billable usage and audit logs). You can create a maximum of two
|
||||
enabled account-level delivery configurations (configurations without a
|
||||
workspace filter) per type. Additionally, you can create two enabled
|
||||
workspace-level delivery configurations per workspace for each log type, which
|
||||
means that the same workspace ID can occur in the workspace filter for no more
|
||||
than two delivery configurations per log type.
|
||||
|
||||
You cannot delete a log delivery configuration, but you can disable it (see
|
||||
[Enable or disable log delivery
|
||||
configuration](:method:LogDelivery/PatchStatus)).
|
||||
|
||||
[Configure audit logging]: https://docs.databricks.com/administration-guide/account-settings/audit-logs.html
|
||||
[Deliver and access billable usage logs]: https://docs.databricks.com/administration-guide/account-settings/billable-usage-delivery.html
|
||||
|
||||
Usage:
|
||||
databricks account log-delivery create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account log-delivery get --help
|
||||
Get log delivery configuration.
|
||||
|
||||
Gets a Databricks log delivery configuration object for an account, both
|
||||
specified by ID.
|
||||
|
||||
Arguments:
|
||||
LOG_DELIVERY_CONFIGURATION_ID: Databricks log delivery configuration ID
|
||||
|
||||
Usage:
|
||||
databricks account log-delivery get LOG_DELIVERY_CONFIGURATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account log-delivery list --help
|
||||
Get all log delivery configurations.
|
||||
|
||||
Gets all Databricks log delivery configurations associated with an account
|
||||
specified by ID.
|
||||
|
||||
Usage:
|
||||
databricks account log-delivery list [flags]
|
||||
|
||||
Flags:
|
||||
--credentials-id string Filter by credential configuration ID.
|
||||
-h, --help help for list
|
||||
--status LogDeliveryConfigStatus Filter by status ENABLED or DISABLED. Supported values: [DISABLED, ENABLED]
|
||||
--storage-configuration-id string Filter by storage configuration ID.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account log-delivery patch-status --help
|
||||
Enable or disable log delivery configuration.
|
||||
|
||||
Enables or disables a log delivery configuration. Deletion of delivery
|
||||
configurations is not supported, so disable log delivery configurations that
|
||||
are no longer needed. Note that you can't re-enable a delivery configuration
|
||||
if this would violate the delivery configuration limits described under
|
||||
[Create log delivery](:method:LogDelivery/Create).
|
||||
|
||||
Arguments:
|
||||
LOG_DELIVERY_CONFIGURATION_ID: Databricks log delivery configuration ID
|
||||
STATUS: Status of log delivery configuration. Set to ENABLED (enabled) or
|
||||
DISABLED (disabled). Defaults to ENABLED. You can [enable or disable
|
||||
the configuration](#operation/patch-log-delivery-config-status) later.
|
||||
Deletion of a configuration is not supported, so disable a log delivery
|
||||
configuration that is no longer needed.
|
||||
|
||||
Usage:
|
||||
databricks account log-delivery patch-status LOG_DELIVERY_CONFIGURATION_ID STATUS [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for patch-status
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account log-delivery create --help
|
||||
trace $CLI account log-delivery get --help
|
||||
trace $CLI account log-delivery list --help
|
||||
trace $CLI account log-delivery patch-status --help
|
|
@ -0,0 +1,111 @@
|
|||
|
||||
>>> $CLI account metastore-assignments create --help
|
||||
Assigns a workspace to a metastore.
|
||||
|
||||
Creates an assignment to a metastore for a workspace
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastore-assignments create WORKSPACE_ID METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastore-assignments delete --help
|
||||
Delete a metastore assignment.
|
||||
|
||||
Deletes a metastore assignment to a workspace, leaving the workspace with no
|
||||
metastore.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastore-assignments delete WORKSPACE_ID METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastore-assignments get --help
|
||||
Gets the metastore assignment for a workspace.
|
||||
|
||||
Gets the metastore assignment, if any, for the workspace specified by ID. If
|
||||
the workspace is assigned a metastore, the mappig will be returned. If no
|
||||
metastore is assigned to the workspace, the assignment will not be found and a
|
||||
404 returned.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
|
||||
Usage:
|
||||
databricks account metastore-assignments get WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastore-assignments list --help
|
||||
Get all workspaces assigned to a metastore.
|
||||
|
||||
Gets a list of all Databricks workspace IDs that have been assigned to given
|
||||
metastore.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastore-assignments list METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastore-assignments update --help
|
||||
Updates a metastore assignment to a workspaces.
|
||||
|
||||
Updates an assignment to a metastore for a workspace. Currently, only the
|
||||
default catalog may be updated.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastore-assignments update WORKSPACE_ID METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
6
acceptance/help/cmd/account/metastore-assignments/metastore-assignments/script
generated
Executable file
6
acceptance/help/cmd/account/metastore-assignments/metastore-assignments/script
generated
Executable file
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account metastore-assignments create --help
|
||||
trace $CLI account metastore-assignments delete --help
|
||||
trace $CLI account metastore-assignments get --help
|
||||
trace $CLI account metastore-assignments list --help
|
||||
trace $CLI account metastore-assignments update --help
|
|
@ -0,0 +1,97 @@
|
|||
|
||||
>>> $CLI account metastores create --help
|
||||
Create metastore.
|
||||
|
||||
Creates a Unity Catalog metastore.
|
||||
|
||||
Usage:
|
||||
databricks account metastores create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastores delete --help
|
||||
Delete a metastore.
|
||||
|
||||
Deletes a Unity Catalog metastore for an account, both specified by ID.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastores delete METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
--force Force deletion even if the metastore is not empty.
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastores get --help
|
||||
Get a metastore.
|
||||
|
||||
Gets a Unity Catalog metastore from an account, both specified by ID.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastores get METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastores list --help
|
||||
Get all metastores associated with an account.
|
||||
|
||||
Gets all Unity Catalog metastores associated with an account specified by ID.
|
||||
|
||||
Usage:
|
||||
databricks account metastores list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account metastores update --help
|
||||
Update a metastore.
|
||||
|
||||
Updates an existing Unity Catalog metastore.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account metastores update METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account metastores create --help
|
||||
trace $CLI account metastores delete --help
|
||||
trace $CLI account metastores get --help
|
||||
trace $CLI account metastores list --help
|
||||
trace $CLI account metastores update --help
|
|
@ -0,0 +1,184 @@
|
|||
|
||||
>>> $CLI account network-connectivity create-network-connectivity-configuration --help
|
||||
Create a network connectivity configuration.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the network connectivity configuration. The name can contain
|
||||
alphanumeric characters, hyphens, and underscores. The length must be
|
||||
between 3 and 30 characters. The name must match the regular expression
|
||||
^[0-9a-zA-Z-_]{3,30}$.
|
||||
REGION: The region for the network connectivity configuration. Only workspaces in
|
||||
the same region can be attached to the network connectivity configuration.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity create-network-connectivity-configuration NAME REGION [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create-network-connectivity-configuration
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity create-private-endpoint-rule --help
|
||||
Create a private endpoint rule.
|
||||
|
||||
Create a private endpoint rule for the specified network connectivity config
|
||||
object. Once the object is created, Databricks asynchronously provisions a new
|
||||
Azure private endpoint to your specified Azure resource.
|
||||
|
||||
**IMPORTANT**: You must use Azure portal or other Azure tools to approve the
|
||||
private endpoint to complete the connection. To get the information of the
|
||||
private endpoint created, make a GET request on the new private endpoint
|
||||
rule. See [serverless private link].
|
||||
|
||||
[serverless private link]: https://learn.microsoft.com/azure/databricks/security/network/serverless-network-security/serverless-private-link
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
RESOURCE_ID: The Azure resource ID of the target resource.
|
||||
GROUP_ID: The sub-resource type (group ID) of the target resource. Note that to
|
||||
connect to workspace root storage (root DBFS), you need two endpoints, one
|
||||
for blob and one for dfs.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity create-private-endpoint-rule NETWORK_CONNECTIVITY_CONFIG_ID RESOURCE_ID GROUP_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create-private-endpoint-rule
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity delete-network-connectivity-configuration --help
|
||||
Delete a network connectivity configuration.
|
||||
|
||||
Deletes a network connectivity configuration.
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity delete-network-connectivity-configuration NETWORK_CONNECTIVITY_CONFIG_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete-network-connectivity-configuration
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity delete-private-endpoint-rule --help
|
||||
Delete a private endpoint rule.
|
||||
|
||||
Initiates deleting a private endpoint rule. If the connection state is PENDING
|
||||
or EXPIRED, the private endpoint is immediately deleted. Otherwise, the
|
||||
private endpoint is deactivated and will be deleted after seven days of
|
||||
deactivation. When a private endpoint is deactivated, the deactivated field
|
||||
is set to true and the private endpoint is not available to your serverless
|
||||
compute resources.
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
PRIVATE_ENDPOINT_RULE_ID: Your private endpoint rule ID.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity delete-private-endpoint-rule NETWORK_CONNECTIVITY_CONFIG_ID PRIVATE_ENDPOINT_RULE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete-private-endpoint-rule
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity get-network-connectivity-configuration --help
|
||||
Get a network connectivity configuration.
|
||||
|
||||
Gets a network connectivity configuration.
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity get-network-connectivity-configuration NETWORK_CONNECTIVITY_CONFIG_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-network-connectivity-configuration
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity get-private-endpoint-rule --help
|
||||
Get a private endpoint rule.
|
||||
|
||||
Gets the private endpoint rule.
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
PRIVATE_ENDPOINT_RULE_ID: Your private endpoint rule ID.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity get-private-endpoint-rule NETWORK_CONNECTIVITY_CONFIG_ID PRIVATE_ENDPOINT_RULE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-private-endpoint-rule
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity list-network-connectivity-configurations --help
|
||||
List network connectivity configurations.
|
||||
|
||||
Gets an array of network connectivity configurations.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity list-network-connectivity-configurations [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list-network-connectivity-configurations
|
||||
--page-token string Pagination token to go to next page based on previous query.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account network-connectivity list-private-endpoint-rules --help
|
||||
List private endpoint rules.
|
||||
|
||||
Gets an array of private endpoint rules.
|
||||
|
||||
Arguments:
|
||||
NETWORK_CONNECTIVITY_CONFIG_ID: Your Network Connectvity Configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account network-connectivity list-private-endpoint-rules NETWORK_CONNECTIVITY_CONFIG_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list-private-endpoint-rules
|
||||
--page-token string Pagination token to go to next page based on previous query.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
9
acceptance/help/cmd/account/network-connectivity/network-connectivity/script
generated
Executable file
9
acceptance/help/cmd/account/network-connectivity/network-connectivity/script
generated
Executable file
|
@ -0,0 +1,9 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account network-connectivity create-network-connectivity-configuration --help
|
||||
trace $CLI account network-connectivity create-private-endpoint-rule --help
|
||||
trace $CLI account network-connectivity delete-network-connectivity-configuration --help
|
||||
trace $CLI account network-connectivity delete-private-endpoint-rule --help
|
||||
trace $CLI account network-connectivity get-network-connectivity-configuration --help
|
||||
trace $CLI account network-connectivity get-private-endpoint-rule --help
|
||||
trace $CLI account network-connectivity list-network-connectivity-configurations --help
|
||||
trace $CLI account network-connectivity list-private-endpoint-rules --help
|
|
@ -0,0 +1,91 @@
|
|||
|
||||
>>> $CLI account networks create --help
|
||||
Create network configuration.
|
||||
|
||||
Creates a Databricks network configuration that represents an VPC and its
|
||||
resources. The VPC will be used for new Databricks clusters. This requires a
|
||||
pre-existing VPC and subnets.
|
||||
|
||||
Arguments:
|
||||
NETWORK_NAME: The human-readable name of the network configuration.
|
||||
|
||||
Usage:
|
||||
databricks account networks create NETWORK_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--vpc-id string The ID of the VPC associated with this network.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account networks delete --help
|
||||
Delete a network configuration.
|
||||
|
||||
Deletes a Databricks network configuration, which represents a cloud VPC and
|
||||
its resources. You cannot delete a network that is associated with a
|
||||
workspace.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform.
|
||||
|
||||
Arguments:
|
||||
NETWORK_ID: Databricks Account API network configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account networks delete NETWORK_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account networks get --help
|
||||
Get a network configuration.
|
||||
|
||||
Gets a Databricks network configuration, which represents a cloud VPC and its
|
||||
resources.
|
||||
|
||||
Arguments:
|
||||
NETWORK_ID: Databricks Account API network configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account networks get NETWORK_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account networks list --help
|
||||
Get all network configurations.
|
||||
|
||||
Gets a list of all Databricks network configurations for an account, specified
|
||||
by ID.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform.
|
||||
|
||||
Usage:
|
||||
databricks account networks list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account networks create --help
|
||||
trace $CLI account networks delete --help
|
||||
trace $CLI account networks get --help
|
||||
trace $CLI account networks list --help
|
|
@ -0,0 +1,19 @@
|
|||
|
||||
>>> $CLI account o-auth-published-apps list --help
|
||||
Get all the published OAuth apps.
|
||||
|
||||
Get all the available published OAuth apps in Databricks.
|
||||
|
||||
Usage:
|
||||
databricks account o-auth-published-apps list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int The max number of OAuth published apps to return in one page.
|
||||
--page-token string A token that can be used to get the next page of results.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
2
acceptance/help/cmd/account/o-auth-published-apps/o-auth-published-apps/script
generated
Executable file
2
acceptance/help/cmd/account/o-auth-published-apps/o-auth-published-apps/script
generated
Executable file
|
@ -0,0 +1,2 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account o-auth-published-apps list --help
|
|
@ -0,0 +1,165 @@
|
|||
|
||||
>>> $CLI account personal-compute delete --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account personal-compute get --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
||||
|
||||
>>> $CLI account personal-compute update --help
|
||||
Databricks Account Commands
|
||||
|
||||
Usage:
|
||||
databricks account [command]
|
||||
|
||||
Identity and Access Management
|
||||
access-control These APIs manage access rules on resources in an account.
|
||||
groups Groups simplify identity management, making it easier to assign access to Databricks account, data, and other securable objects.
|
||||
service-principals Identities for use with jobs, automated tools, and systems such as scripts, apps, and CI/CD platforms.
|
||||
users User identities recognized by Databricks and represented by email addresses.
|
||||
workspace-assignment The Workspace Permission Assignment API allows you to manage workspace permissions for principals in your account.
|
||||
|
||||
Unity Catalog
|
||||
metastore-assignments These APIs manage metastore assignments to a workspace.
|
||||
metastores These APIs manage Unity Catalog metastores for an account.
|
||||
storage-credentials These APIs manage storage credentials for a particular metastore.
|
||||
|
||||
Settings
|
||||
ip-access-lists The Accounts IP Access List API enables account admins to configure IP access lists for access to the account console.
|
||||
network-connectivity These APIs provide configurations for the network connectivity of your workspaces for serverless compute resources.
|
||||
settings Accounts Settings API allows users to manage settings at the account level.
|
||||
|
||||
Provisioning
|
||||
credentials These APIs manage credential configurations for this workspace.
|
||||
encryption-keys These APIs manage encryption key configurations for this workspace (optional).
|
||||
networks These APIs manage network configurations for customer-managed VPCs (optional).
|
||||
private-access These APIs manage private access settings for this account.
|
||||
storage These APIs manage storage configurations for this workspace.
|
||||
vpc-endpoints These APIs manage VPC endpoint configurations for this account.
|
||||
workspaces These APIs manage workspaces for this account.
|
||||
|
||||
Billing
|
||||
billable-usage This API allows you to download billable usage logs for the specified account and date range.
|
||||
budgets These APIs manage budget configurations for this account.
|
||||
log-delivery These APIs manage log delivery configurations for this account.
|
||||
usage-dashboards These APIs manage usage dashboards for this account.
|
||||
|
||||
OAuth
|
||||
custom-app-integration These APIs enable administrators to manage custom OAuth app integrations, which is required for adding/using Custom OAuth App Integration like Tableau Cloud for Databricks in AWS cloud.
|
||||
o-auth-published-apps These APIs enable administrators to view all the available published OAuth applications in Databricks.
|
||||
published-app-integration These APIs enable administrators to manage published OAuth app integrations, which is required for adding/using Published OAuth App Integration like Tableau Desktop for Databricks in AWS cloud.
|
||||
service-principal-secrets These APIs enable administrators to manage service principal secrets.
|
||||
|
||||
Flags:
|
||||
-h, --help help for account
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
Use "databricks account [command] --help" for more information about a command.
|
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account personal-compute delete --help
|
||||
trace $CLI account personal-compute get --help
|
||||
trace $CLI account personal-compute update --help
|
|
@ -0,0 +1,157 @@
|
|||
|
||||
>>> $CLI account private-access create --help
|
||||
Create private access settings.
|
||||
|
||||
Creates a private access settings object, which specifies how your workspace
|
||||
is accessed over [AWS PrivateLink]. To use AWS PrivateLink, a workspace must
|
||||
have a private access settings object referenced by ID in the workspace's
|
||||
private_access_settings_id property.
|
||||
|
||||
You can share one private access settings with multiple workspaces in a single
|
||||
account. However, private access settings are specific to AWS regions, so only
|
||||
workspaces in the same AWS region can use a given private access settings
|
||||
object.
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Arguments:
|
||||
PRIVATE_ACCESS_SETTINGS_NAME: The human-readable name of the private access settings object.
|
||||
REGION: The cloud region for workspaces associated with this private access
|
||||
settings object.
|
||||
|
||||
Usage:
|
||||
databricks account private-access create PRIVATE_ACCESS_SETTINGS_NAME REGION [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--private-access-level PrivateAccessLevel The private access level controls which VPC endpoints can connect to the UI or API of any workspace that attaches this private access settings object. Supported values: [ACCOUNT, ENDPOINT]
|
||||
--public-access-enabled Determines if the workspace can be accessed over public internet.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account private-access delete --help
|
||||
Delete a private access settings object.
|
||||
|
||||
Deletes a private access settings object, which determines how your workspace
|
||||
is accessed over [AWS PrivateLink].
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].",
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Arguments:
|
||||
PRIVATE_ACCESS_SETTINGS_ID: Databricks Account API private access settings ID.
|
||||
|
||||
Usage:
|
||||
databricks account private-access delete PRIVATE_ACCESS_SETTINGS_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account private-access get --help
|
||||
Get a private access settings object.
|
||||
|
||||
Gets a private access settings object, which specifies how your workspace is
|
||||
accessed over [AWS PrivateLink].
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].",
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Arguments:
|
||||
PRIVATE_ACCESS_SETTINGS_ID: Databricks Account API private access settings ID.
|
||||
|
||||
Usage:
|
||||
databricks account private-access get PRIVATE_ACCESS_SETTINGS_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account private-access list --help
|
||||
Get all private access settings objects.
|
||||
|
||||
Gets a list of all private access settings objects for an account, specified
|
||||
by ID.
|
||||
|
||||
Usage:
|
||||
databricks account private-access list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account private-access replace --help
|
||||
Replace private access settings.
|
||||
|
||||
Updates an existing private access settings object, which specifies how your
|
||||
workspace is accessed over [AWS PrivateLink]. To use AWS PrivateLink, a
|
||||
workspace must have a private access settings object referenced by ID in the
|
||||
workspace's private_access_settings_id property.
|
||||
|
||||
This operation completely overwrites your existing private access settings
|
||||
object attached to your workspaces. All workspaces attached to the private
|
||||
access settings are affected by any change. If public_access_enabled,
|
||||
private_access_level, or allowed_vpc_endpoint_ids are updated, effects of
|
||||
these changes might take several minutes to propagate to the workspace API.
|
||||
|
||||
You can share one private access settings object with multiple workspaces in a
|
||||
single account. However, private access settings are specific to AWS regions,
|
||||
so only workspaces in the same AWS region can use a given private access
|
||||
settings object.
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Arguments:
|
||||
PRIVATE_ACCESS_SETTINGS_ID: Databricks Account API private access settings ID.
|
||||
PRIVATE_ACCESS_SETTINGS_NAME: The human-readable name of the private access settings object.
|
||||
REGION: The cloud region for workspaces associated with this private access
|
||||
settings object.
|
||||
|
||||
Usage:
|
||||
databricks account private-access replace PRIVATE_ACCESS_SETTINGS_ID PRIVATE_ACCESS_SETTINGS_NAME REGION [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for replace
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--private-access-level PrivateAccessLevel The private access level controls which VPC endpoints can connect to the UI or API of any workspace that attaches this private access settings object. Supported values: [ACCOUNT, ENDPOINT]
|
||||
--public-access-enabled Determines if the workspace can be accessed over public internet.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account private-access create --help
|
||||
trace $CLI account private-access delete --help
|
||||
trace $CLI account private-access get --help
|
||||
trace $CLI account private-access list --help
|
||||
trace $CLI account private-access replace --help
|
|
@ -0,0 +1,96 @@
|
|||
|
||||
>>> $CLI account published-app-integration create --help
|
||||
Create Published OAuth App Integration.
|
||||
|
||||
Create Published OAuth App Integration.
|
||||
|
||||
You can retrieve the published OAuth app integration via
|
||||
:method:PublishedAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account published-app-integration create [flags]
|
||||
|
||||
Flags:
|
||||
--app-id string App id of the OAuth published app integration.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account published-app-integration delete --help
|
||||
Delete Published OAuth App Integration.
|
||||
|
||||
Delete an existing Published OAuth App Integration. You can retrieve the
|
||||
published OAuth app integration via :method:PublishedAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account published-app-integration delete INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account published-app-integration get --help
|
||||
Get OAuth Published App Integration.
|
||||
|
||||
Gets the Published OAuth App Integration for the given integration id.
|
||||
|
||||
Usage:
|
||||
databricks account published-app-integration get INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account published-app-integration list --help
|
||||
Get published oauth app integrations.
|
||||
|
||||
Get the list of published OAuth app integrations for the specified Databricks
|
||||
account
|
||||
|
||||
Usage:
|
||||
databricks account published-app-integration list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int
|
||||
--page-token string
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account published-app-integration update --help
|
||||
Updates Published OAuth App Integration.
|
||||
|
||||
Updates an existing published OAuth App Integration. You can retrieve the
|
||||
published OAuth app integration via :method:PublishedAppIntegration/get.
|
||||
|
||||
Usage:
|
||||
databricks account published-app-integration update INTEGRATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
6
acceptance/help/cmd/account/published-app-integration/published-app-integration/script
generated
Executable file
6
acceptance/help/cmd/account/published-app-integration/published-app-integration/script
generated
Executable file
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account published-app-integration create --help
|
||||
trace $CLI account published-app-integration delete --help
|
||||
trace $CLI account published-app-integration get --help
|
||||
trace $CLI account published-app-integration list --help
|
||||
trace $CLI account published-app-integration update --help
|
|
@ -0,0 +1,103 @@
|
|||
|
||||
>>> $CLI account service-principal-federation-policy create --help
|
||||
Create service principal federation policy.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal id for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-federation-policy create SERVICE_PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
--description string Description of the federation policy.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--name string Resource name for the federation policy.
|
||||
--policy-id string The identifier for the federation policy.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-federation-policy delete --help
|
||||
Delete service principal federation policy.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal id for the federation policy.
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-federation-policy delete SERVICE_PRINCIPAL_ID POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-federation-policy get --help
|
||||
Get service principal federation policy.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal id for the federation policy.
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-federation-policy get SERVICE_PRINCIPAL_ID POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-federation-policy list --help
|
||||
List service principal federation policies.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal id for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-federation-policy list SERVICE_PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int
|
||||
--page-token string
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-federation-policy update --help
|
||||
Update service principal federation policy.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal id for the federation policy.
|
||||
POLICY_ID: The identifier for the federation policy.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-federation-policy update SERVICE_PRINCIPAL_ID POLICY_ID [flags]
|
||||
|
||||
Flags:
|
||||
--description string Description of the federation policy.
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--name string Resource name for the federation policy.
|
||||
--update-mask string The field mask specifies which fields of the policy to update.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
6
acceptance/help/cmd/account/service-principal-federation-policy/service-principal-federation-policy/script
generated
Executable file
6
acceptance/help/cmd/account/service-principal-federation-policy/service-principal-federation-policy/script
generated
Executable file
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account service-principal-federation-policy create --help
|
||||
trace $CLI account service-principal-federation-policy delete --help
|
||||
trace $CLI account service-principal-federation-policy get --help
|
||||
trace $CLI account service-principal-federation-policy list --help
|
||||
trace $CLI account service-principal-federation-policy update --help
|
|
@ -0,0 +1,64 @@
|
|||
|
||||
>>> $CLI account service-principal-secrets create --help
|
||||
Create service principal secret.
|
||||
|
||||
Create a secret for the given service principal.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal ID.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-secrets create SERVICE_PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-secrets delete --help
|
||||
Delete service principal secret.
|
||||
|
||||
Delete a secret from the given service principal.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal ID.
|
||||
SECRET_ID: The secret ID.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-secrets delete SERVICE_PRINCIPAL_ID SECRET_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principal-secrets list --help
|
||||
List service principal secrets.
|
||||
|
||||
List all secrets associated with the given service principal. This operation
|
||||
only returns information about the secrets themselves and does not include the
|
||||
secret values.
|
||||
|
||||
Arguments:
|
||||
SERVICE_PRINCIPAL_ID: The service principal ID.
|
||||
|
||||
Usage:
|
||||
databricks account service-principal-secrets list SERVICE_PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-token string An opaque page token which was the next_page_token in the response of the previous request to list the secrets for this service principal.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
4
acceptance/help/cmd/account/service-principal-secrets/service-principal-secrets/script
generated
Executable file
4
acceptance/help/cmd/account/service-principal-secrets/service-principal-secrets/script
generated
Executable file
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account service-principal-secrets create --help
|
||||
trace $CLI account service-principal-secrets delete --help
|
||||
trace $CLI account service-principal-secrets list --help
|
|
@ -0,0 +1,138 @@
|
|||
|
||||
>>> $CLI account service-principals create --help
|
||||
Create a service principal.
|
||||
|
||||
Creates a new service principal in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals create [flags]
|
||||
|
||||
Flags:
|
||||
--active If this user is active.
|
||||
--application-id string UUID relating to the service principal.
|
||||
--display-name string String that represents a concatenation of given and family names.
|
||||
--external-id string
|
||||
-h, --help help for create
|
||||
--id string Databricks service principal ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principals delete --help
|
||||
Delete a service principal.
|
||||
|
||||
Delete a single service principal in the Databricks account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a service principal in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals delete ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principals get --help
|
||||
Get service principal details.
|
||||
|
||||
Gets the details for a single service principal define in the Databricks
|
||||
account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a service principal in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals get ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principals list --help
|
||||
List service principals.
|
||||
|
||||
Gets the set of service principals associated with a Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals list [flags]
|
||||
|
||||
Flags:
|
||||
--attributes string Comma-separated list of attributes to return in response.
|
||||
--count int Desired number of results per page.
|
||||
--excluded-attributes string Comma-separated list of attributes to exclude in response.
|
||||
--filter string Query by which the results have to be filtered.
|
||||
-h, --help help for list
|
||||
--sort-by string Attribute to sort the results.
|
||||
--sort-order ListSortOrder The order to sort the results. Supported values: [ascending, descending]
|
||||
--start-index int Specifies the index of the first result.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principals patch --help
|
||||
Update service principal details.
|
||||
|
||||
Partially updates the details of a single service principal in the Databricks
|
||||
account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a service principal in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals patch ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for patch
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account service-principals update --help
|
||||
Replace service principal.
|
||||
|
||||
Updates the details of a single service principal.
|
||||
|
||||
This action replaces the existing service principal with the same name.
|
||||
|
||||
Arguments:
|
||||
ID: Databricks service principal ID.
|
||||
|
||||
Usage:
|
||||
databricks account service-principals update ID [flags]
|
||||
|
||||
Flags:
|
||||
--active If this user is active.
|
||||
--application-id string UUID relating to the service principal.
|
||||
--display-name string String that represents a concatenation of given and family names.
|
||||
--external-id string
|
||||
-h, --help help for update
|
||||
--id string Databricks service principal ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
7
acceptance/help/cmd/account/service-principals/service-principals/script
generated
Executable file
7
acceptance/help/cmd/account/service-principals/service-principals/script
generated
Executable file
|
@ -0,0 +1,7 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account service-principals create --help
|
||||
trace $CLI account service-principals delete --help
|
||||
trace $CLI account service-principals get --help
|
||||
trace $CLI account service-principals list --help
|
||||
trace $CLI account service-principals patch --help
|
||||
trace $CLI account service-principals update --help
|
|
@ -0,0 +1,3 @@
|
|||
script: line 65: syntax error near unexpected token `)'
|
||||
|
||||
Exit code: 2
|
|
@ -0,0 +1 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
|
@ -0,0 +1,118 @@
|
|||
|
||||
>>> $CLI account storage-credentials create --help
|
||||
Create a storage credential.
|
||||
|
||||
Creates a new storage credential. The request object is specific to the cloud:
|
||||
|
||||
* **AwsIamRole** for AWS credentials * **AzureServicePrincipal** for Azure
|
||||
credentials * **GcpServiceAcountKey** for GCP credentials.
|
||||
|
||||
The caller must be a metastore admin and have the
|
||||
**CREATE_STORAGE_CREDENTIAL** privilege on the metastore.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account storage-credentials create METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage-credentials delete --help
|
||||
Delete a storage credential.
|
||||
|
||||
Deletes a storage credential from the metastore. The caller must be an owner
|
||||
of the storage credential.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
STORAGE_CREDENTIAL_NAME: Name of the storage credential.
|
||||
|
||||
Usage:
|
||||
databricks account storage-credentials delete METASTORE_ID STORAGE_CREDENTIAL_NAME [flags]
|
||||
|
||||
Flags:
|
||||
--force Force deletion even if the Storage Credential is not empty.
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage-credentials get --help
|
||||
Gets the named storage credential.
|
||||
|
||||
Gets a storage credential from the metastore. The caller must be a metastore
|
||||
admin, the owner of the storage credential, or have a level of privilege on
|
||||
the storage credential.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
STORAGE_CREDENTIAL_NAME: Name of the storage credential.
|
||||
|
||||
Usage:
|
||||
databricks account storage-credentials get METASTORE_ID STORAGE_CREDENTIAL_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage-credentials list --help
|
||||
Get all storage credentials assigned to a metastore.
|
||||
|
||||
Gets a list of all storage credentials that have been assigned to given
|
||||
metastore.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
|
||||
Usage:
|
||||
databricks account storage-credentials list METASTORE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage-credentials update --help
|
||||
Updates a storage credential.
|
||||
|
||||
Updates a storage credential on the metastore. The caller must be the owner of
|
||||
the storage credential. If the caller is a metastore admin, only the __owner__
|
||||
credential can be changed.
|
||||
|
||||
Arguments:
|
||||
METASTORE_ID: Unity Catalog metastore ID
|
||||
STORAGE_CREDENTIAL_NAME: Name of the storage credential.
|
||||
|
||||
Usage:
|
||||
databricks account storage-credentials update METASTORE_ID STORAGE_CREDENTIAL_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
6
acceptance/help/cmd/account/storage-credentials/storage-credentials/script
generated
Executable file
6
acceptance/help/cmd/account/storage-credentials/storage-credentials/script
generated
Executable file
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account storage-credentials create --help
|
||||
trace $CLI account storage-credentials delete --help
|
||||
trace $CLI account storage-credentials get --help
|
||||
trace $CLI account storage-credentials list --help
|
||||
trace $CLI account storage-credentials update --help
|
|
@ -0,0 +1,86 @@
|
|||
|
||||
>>> $CLI account storage create --help
|
||||
Create new storage configuration.
|
||||
|
||||
Creates new storage configuration for an account, specified by ID. Uploads a
|
||||
storage configuration object that represents the root AWS S3 bucket in your
|
||||
account. Databricks stores related workspace assets including DBFS, cluster
|
||||
logs, and job results. For the AWS S3 bucket, you need to configure the
|
||||
required bucket policy.
|
||||
|
||||
For information about how to create a new workspace with this API, see [Create
|
||||
a new workspace using the Account API]
|
||||
|
||||
[Create a new workspace using the Account API]: http://docs.databricks.com/administration-guide/account-api/new-workspace.html
|
||||
|
||||
Usage:
|
||||
databricks account storage create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage delete --help
|
||||
Delete storage configuration.
|
||||
|
||||
Deletes a Databricks storage configuration. You cannot delete a storage
|
||||
configuration that is associated with any workspace.
|
||||
|
||||
Arguments:
|
||||
STORAGE_CONFIGURATION_ID: Databricks Account API storage configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account storage delete STORAGE_CONFIGURATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage get --help
|
||||
Get storage configuration.
|
||||
|
||||
Gets a Databricks storage configuration for an account, both specified by ID.
|
||||
|
||||
Arguments:
|
||||
STORAGE_CONFIGURATION_ID: Databricks Account API storage configuration ID.
|
||||
|
||||
Usage:
|
||||
databricks account storage get STORAGE_CONFIGURATION_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account storage list --help
|
||||
Get all storage configurations.
|
||||
|
||||
Gets a list of all Databricks storage configurations for your account,
|
||||
specified by ID.
|
||||
|
||||
Usage:
|
||||
databricks account storage list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account storage create --help
|
||||
trace $CLI account storage delete --help
|
||||
trace $CLI account storage get --help
|
||||
trace $CLI account storage list --help
|
|
@ -0,0 +1,40 @@
|
|||
|
||||
>>> $CLI account usage-dashboards create --help
|
||||
Create new usage dashboard.
|
||||
|
||||
Create a usage dashboard specified by workspaceId, accountId, and dashboard
|
||||
type.
|
||||
|
||||
Usage:
|
||||
databricks account usage-dashboards create [flags]
|
||||
|
||||
Flags:
|
||||
--dashboard-type UsageDashboardType Workspace level usage dashboard shows usage data for the specified workspace ID. Supported values: [USAGE_DASHBOARD_TYPE_GLOBAL, USAGE_DASHBOARD_TYPE_WORKSPACE]
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--workspace-id int The workspace ID of the workspace in which the usage dashboard is created.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account usage-dashboards get --help
|
||||
Get usage dashboard.
|
||||
|
||||
Get a usage dashboard specified by workspaceId, accountId, and dashboard type.
|
||||
|
||||
Usage:
|
||||
databricks account usage-dashboards get [flags]
|
||||
|
||||
Flags:
|
||||
--dashboard-type UsageDashboardType Workspace level usage dashboard shows usage data for the specified workspace ID. Supported values: [USAGE_DASHBOARD_TYPE_GLOBAL, USAGE_DASHBOARD_TYPE_WORKSPACE]
|
||||
-h, --help help for get
|
||||
--workspace-id int The workspace ID of the workspace in which the usage dashboard is created.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,3 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account usage-dashboards create --help
|
||||
trace $CLI account usage-dashboards get --help
|
|
@ -0,0 +1,145 @@
|
|||
|
||||
>>> $CLI account users create --help
|
||||
Create a new user.
|
||||
|
||||
Creates a new user in the Databricks account. This new user will also be added
|
||||
to the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account users create [flags]
|
||||
|
||||
Flags:
|
||||
--active If this user is active.
|
||||
--display-name string String that represents a concatenation of given and family names.
|
||||
--external-id string External ID is not currently supported.
|
||||
-h, --help help for create
|
||||
--id string Databricks user ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--user-name string Email address of the Databricks user.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account users delete --help
|
||||
Delete a user.
|
||||
|
||||
Deletes a user. Deleting a user from a Databricks account also removes objects
|
||||
associated with the user.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a user in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account users delete ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account users get --help
|
||||
Get user details.
|
||||
|
||||
Gets information for a specific user in Databricks account.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a user in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account users get ID [flags]
|
||||
|
||||
Flags:
|
||||
--attributes string Comma-separated list of attributes to return in response.
|
||||
--count int Desired number of results per page.
|
||||
--excluded-attributes string Comma-separated list of attributes to exclude in response.
|
||||
--filter string Query by which the results have to be filtered.
|
||||
-h, --help help for get
|
||||
--sort-by string Attribute to sort the results.
|
||||
--sort-order GetSortOrder The order to sort the results. Supported values: [ascending, descending]
|
||||
--start-index int Specifies the index of the first result.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account users list --help
|
||||
List users.
|
||||
|
||||
Gets details for all the users associated with a Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account users list [flags]
|
||||
|
||||
Flags:
|
||||
--attributes string Comma-separated list of attributes to return in response.
|
||||
--count int Desired number of results per page.
|
||||
--excluded-attributes string Comma-separated list of attributes to exclude in response.
|
||||
--filter string Query by which the results have to be filtered.
|
||||
-h, --help help for list
|
||||
--sort-by string Attribute to sort the results.
|
||||
--sort-order ListSortOrder The order to sort the results. Supported values: [ascending, descending]
|
||||
--start-index int Specifies the index of the first result.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account users patch --help
|
||||
Update user details.
|
||||
|
||||
Partially updates a user resource by applying the supplied operations on
|
||||
specific user attributes.
|
||||
|
||||
Arguments:
|
||||
ID: Unique ID for a user in the Databricks account.
|
||||
|
||||
Usage:
|
||||
databricks account users patch ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for patch
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account users update --help
|
||||
Replace a user.
|
||||
|
||||
Replaces a user's information with the data supplied in request.
|
||||
|
||||
Arguments:
|
||||
ID: Databricks user ID. This is automatically set by Databricks. Any value
|
||||
provided by the client will be ignored.
|
||||
|
||||
Usage:
|
||||
databricks account users update ID [flags]
|
||||
|
||||
Flags:
|
||||
--active If this user is active.
|
||||
--display-name string String that represents a concatenation of given and family names.
|
||||
--external-id string External ID is not currently supported.
|
||||
-h, --help help for update
|
||||
--id string Databricks user ID.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--user-name string Email address of the Databricks user.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,7 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account users create --help
|
||||
trace $CLI account users delete --help
|
||||
trace $CLI account users get --help
|
||||
trace $CLI account users list --help
|
||||
trace $CLI account users patch --help
|
||||
trace $CLI account users update --help
|
|
@ -0,0 +1,109 @@
|
|||
|
||||
>>> $CLI account vpc-endpoints create --help
|
||||
Create VPC endpoint configuration.
|
||||
|
||||
Creates a VPC endpoint configuration, which represents a [VPC endpoint] object
|
||||
in AWS used to communicate privately with Databricks over [AWS PrivateLink].
|
||||
|
||||
After you create the VPC endpoint configuration, the Databricks [endpoint
|
||||
service] automatically accepts the VPC endpoint.
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
[VPC endpoint]: https://docs.aws.amazon.com/vpc/latest/privatelink/vpc-endpoints.html
|
||||
[endpoint service]: https://docs.aws.amazon.com/vpc/latest/privatelink/privatelink-share-your-services.html
|
||||
|
||||
Arguments:
|
||||
VPC_ENDPOINT_NAME: The human-readable name of the storage configuration.
|
||||
|
||||
Usage:
|
||||
databricks account vpc-endpoints create VPC_ENDPOINT_NAME [flags]
|
||||
|
||||
Flags:
|
||||
--aws-vpc-endpoint-id string The ID of the VPC endpoint object in AWS.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--region string The AWS region in which this VPC endpoint object exists.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account vpc-endpoints delete --help
|
||||
Delete VPC endpoint configuration.
|
||||
|
||||
Deletes a VPC endpoint configuration, which represents an [AWS VPC endpoint]
|
||||
that can communicate privately with Databricks over [AWS PrivateLink].
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[AWS VPC endpoint]: https://docs.aws.amazon.com/vpc/latest/privatelink/concepts.html
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Arguments:
|
||||
VPC_ENDPOINT_ID: Databricks VPC endpoint ID.
|
||||
|
||||
Usage:
|
||||
databricks account vpc-endpoints delete VPC_ENDPOINT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account vpc-endpoints get --help
|
||||
Get a VPC endpoint configuration.
|
||||
|
||||
Gets a VPC endpoint configuration, which represents a [VPC endpoint] object in
|
||||
AWS used to communicate privately with Databricks over [AWS PrivateLink].
|
||||
|
||||
[AWS PrivateLink]: https://aws.amazon.com/privatelink
|
||||
[VPC endpoint]: https://docs.aws.amazon.com/vpc/latest/privatelink/concepts.html
|
||||
|
||||
Arguments:
|
||||
VPC_ENDPOINT_ID: Databricks VPC endpoint ID.
|
||||
|
||||
Usage:
|
||||
databricks account vpc-endpoints get VPC_ENDPOINT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account vpc-endpoints list --help
|
||||
Get all VPC endpoint configurations.
|
||||
|
||||
Gets a list of all VPC endpoints for an account, specified by ID.
|
||||
|
||||
Before configuring PrivateLink, read the [Databricks article about
|
||||
PrivateLink].
|
||||
|
||||
[Databricks article about PrivateLink]: https://docs.databricks.com/administration-guide/cloud-configurations/aws/privatelink.html
|
||||
|
||||
Usage:
|
||||
databricks account vpc-endpoints list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account vpc-endpoints create --help
|
||||
trace $CLI account vpc-endpoints delete --help
|
||||
trace $CLI account vpc-endpoints get --help
|
||||
trace $CLI account vpc-endpoints list --help
|
|
@ -0,0 +1,86 @@
|
|||
|
||||
>>> $CLI account workspace-assignment delete --help
|
||||
Delete permissions assignment.
|
||||
|
||||
Deletes the workspace permissions assignment in a given account and workspace
|
||||
for the specified principal.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: The workspace ID for the account.
|
||||
PRINCIPAL_ID: The ID of the user, service principal, or group.
|
||||
|
||||
Usage:
|
||||
databricks account workspace-assignment delete WORKSPACE_ID PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspace-assignment get --help
|
||||
List workspace permissions.
|
||||
|
||||
Get an array of workspace permissions for the specified account and workspace.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: The workspace ID.
|
||||
|
||||
Usage:
|
||||
databricks account workspace-assignment get WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspace-assignment list --help
|
||||
Get permission assignments.
|
||||
|
||||
Get the permission assignments for the specified Databricks account and
|
||||
Databricks workspace.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: The workspace ID for the account.
|
||||
|
||||
Usage:
|
||||
databricks account workspace-assignment list WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspace-assignment update --help
|
||||
Create or update permissions assignment.
|
||||
|
||||
Creates or updates the workspace permissions assignment in a given account and
|
||||
workspace for the specified principal.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: The workspace ID.
|
||||
PRINCIPAL_ID: The ID of the user, service principal, or group.
|
||||
|
||||
Usage:
|
||||
databricks account workspace-assignment update WORKSPACE_ID PRINCIPAL_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
5
acceptance/help/cmd/account/workspace-assignment/workspace-assignment/script
generated
Executable file
5
acceptance/help/cmd/account/workspace-assignment/workspace-assignment/script
generated
Executable file
|
@ -0,0 +1,5 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account workspace-assignment delete --help
|
||||
trace $CLI account workspace-assignment get --help
|
||||
trace $CLI account workspace-assignment list --help
|
||||
trace $CLI account workspace-assignment update --help
|
|
@ -0,0 +1,279 @@
|
|||
|
||||
>>> $CLI account workspaces create --help
|
||||
Create a new workspace.
|
||||
|
||||
Creates a new workspace.
|
||||
|
||||
**Important**: This operation is asynchronous. A response with HTTP status
|
||||
code 200 means the request has been accepted and is in progress, but does not
|
||||
mean that the workspace deployed successfully and is running. The initial
|
||||
workspace status is typically PROVISIONING. Use the workspace ID
|
||||
(workspace_id) field in the response to identify the new workspace and make
|
||||
repeated GET requests with the workspace ID and check its status. The
|
||||
workspace becomes available when the status changes to RUNNING.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_NAME: The workspace's human-readable name.
|
||||
|
||||
Usage:
|
||||
databricks account workspaces create WORKSPACE_NAME [flags]
|
||||
|
||||
Flags:
|
||||
--aws-region string The AWS region of the workspace's data plane.
|
||||
--cloud string The cloud provider which the workspace uses.
|
||||
--credentials-id string ID of the workspace's credential configuration object.
|
||||
--deployment-name string The deployment name defines part of the subdomain for the workspace.
|
||||
-h, --help help for create
|
||||
--is-no-public-ip-enabled Whether no public IP is enabled for the workspace.
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--location string The Google Cloud region of the workspace data plane in your Google account.
|
||||
--managed-services-customer-managed-key-id string The ID of the workspace's managed services encryption key configuration object.
|
||||
--network-id string
|
||||
--no-wait do not wait to reach RUNNING state
|
||||
--pricing-tier PricingTier The pricing tier of the workspace. Supported values: [
|
||||
COMMUNITY_EDITION,
|
||||
DEDICATED,
|
||||
ENTERPRISE,
|
||||
PREMIUM,
|
||||
STANDARD,
|
||||
UNKNOWN,
|
||||
]
|
||||
--private-access-settings-id string ID of the workspace's private access settings object.
|
||||
--storage-configuration-id string The ID of the workspace's storage configuration object.
|
||||
--storage-customer-managed-key-id string The ID of the workspace's storage encryption key configuration object.
|
||||
--timeout duration maximum amount of time to reach RUNNING state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspaces delete --help
|
||||
Delete a workspace.
|
||||
|
||||
Terminates and deletes a Databricks workspace. From an API perspective,
|
||||
deletion is immediate. However, it might take a few minutes for all workspaces
|
||||
resources to be deleted, depending on the size and number of workspace
|
||||
resources.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform or on a select custom plan that allows multiple workspaces per
|
||||
account.
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
|
||||
Usage:
|
||||
databricks account workspaces delete WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspaces get --help
|
||||
Get a workspace.
|
||||
|
||||
Gets information including status for a Databricks workspace, specified by ID.
|
||||
In the response, the workspace_status field indicates the current status.
|
||||
After initial workspace creation (which is asynchronous), make repeated GET
|
||||
requests with the workspace ID and check its status. The workspace becomes
|
||||
available when the status changes to RUNNING.
|
||||
|
||||
For information about how to create a new workspace with this API **including
|
||||
error handling**, see [Create a new workspace using the Account API].
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform or on a select custom plan that allows multiple workspaces per
|
||||
account.
|
||||
|
||||
[Create a new workspace using the Account API]: http://docs.databricks.com/administration-guide/account-api/new-workspace.html
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
|
||||
Usage:
|
||||
databricks account workspaces get WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspaces list --help
|
||||
Get all workspaces.
|
||||
|
||||
Gets a list of all workspaces associated with an account, specified by ID.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform or on a select custom plan that allows multiple workspaces per
|
||||
account.
|
||||
|
||||
Usage:
|
||||
databricks account workspaces list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI account workspaces update --help
|
||||
Update workspace configuration.
|
||||
|
||||
Updates a workspace configuration for either a running workspace or a failed
|
||||
workspace. The elements that can be updated varies between these two use
|
||||
cases.
|
||||
|
||||
### Update a failed workspace You can update a Databricks workspace
|
||||
configuration for failed workspace deployment for some fields, but not all
|
||||
fields. For a failed workspace, this request supports updates to the following
|
||||
fields only: - Credential configuration ID - Storage configuration ID -
|
||||
Network configuration ID. Used only to add or change a network configuration
|
||||
for a customer-managed VPC. For a failed workspace only, you can convert a
|
||||
workspace with Databricks-managed VPC to use a customer-managed VPC by adding
|
||||
this ID. You cannot downgrade a workspace with a customer-managed VPC to be a
|
||||
Databricks-managed VPC. You can update the network configuration for a failed
|
||||
or running workspace to add PrivateLink support, though you must also add a
|
||||
private access settings object. - Key configuration ID for managed services
|
||||
(control plane storage, such as notebook source and Databricks SQL queries).
|
||||
Used only if you use customer-managed keys for managed services. - Key
|
||||
configuration ID for workspace storage (root S3 bucket and, optionally, EBS
|
||||
volumes). Used only if you use customer-managed keys for workspace storage.
|
||||
**Important**: If the workspace was ever in the running state, even if briefly
|
||||
before becoming a failed workspace, you cannot add a new key configuration ID
|
||||
for workspace storage. - Private access settings ID to add PrivateLink
|
||||
support. You can add or update the private access settings ID to upgrade a
|
||||
workspace to add support for front-end, back-end, or both types of
|
||||
connectivity. You cannot remove (downgrade) any existing front-end or back-end
|
||||
PrivateLink support on a workspace. - Custom tags. Given you provide an empty
|
||||
custom tags, the update would not be applied. - Network connectivity
|
||||
configuration ID to add serverless stable IP support. You can add or update
|
||||
the network connectivity configuration ID to ensure the workspace uses the
|
||||
same set of stable IP CIDR blocks to access your resources. You cannot remove
|
||||
a network connectivity configuration from the workspace once attached, you can
|
||||
only switch to another one.
|
||||
|
||||
After calling the PATCH operation to update the workspace configuration,
|
||||
make repeated GET requests with the workspace ID and check the workspace
|
||||
status. The workspace is successful if the status changes to RUNNING.
|
||||
|
||||
For information about how to create a new workspace with this API **including
|
||||
error handling**, see [Create a new workspace using the Account API].
|
||||
|
||||
### Update a running workspace You can update a Databricks workspace
|
||||
configuration for running workspaces for some fields, but not all fields. For
|
||||
a running workspace, this request supports updating the following fields only:
|
||||
- Credential configuration ID - Network configuration ID. Used only if you
|
||||
already use a customer-managed VPC. You cannot convert a running workspace
|
||||
from a Databricks-managed VPC to a customer-managed VPC. You can use a network
|
||||
configuration update in this API for a failed or running workspace to add
|
||||
support for PrivateLink, although you also need to add a private access
|
||||
settings object. - Key configuration ID for managed services (control plane
|
||||
storage, such as notebook source and Databricks SQL queries). Databricks does
|
||||
not directly encrypt the data with the customer-managed key (CMK). Databricks
|
||||
uses both the CMK and the Databricks managed key (DMK) that is unique to your
|
||||
workspace to encrypt the Data Encryption Key (DEK). Databricks uses the DEK to
|
||||
encrypt your workspace's managed services persisted data. If the workspace
|
||||
does not already have a CMK for managed services, adding this ID enables
|
||||
managed services encryption for new or updated data. Existing managed services
|
||||
data that existed before adding the key remains not encrypted with the DEK
|
||||
until it is modified. If the workspace already has customer-managed keys for
|
||||
managed services, this request rotates (changes) the CMK keys and the DEK is
|
||||
re-encrypted with the DMK and the new CMK. - Key configuration ID for
|
||||
workspace storage (root S3 bucket and, optionally, EBS volumes). You can set
|
||||
this only if the workspace does not already have a customer-managed key
|
||||
configuration for workspace storage. - Private access settings ID to add
|
||||
PrivateLink support. You can add or update the private access settings ID to
|
||||
upgrade a workspace to add support for front-end, back-end, or both types of
|
||||
connectivity. You cannot remove (downgrade) any existing front-end or back-end
|
||||
PrivateLink support on a workspace. - Custom tags. Given you provide an empty
|
||||
custom tags, the update would not be applied. - Network connectivity
|
||||
configuration ID to add serverless stable IP support. You can add or update
|
||||
the network connectivity configuration ID to ensure the workspace uses the
|
||||
same set of stable IP CIDR blocks to access your resources. You cannot remove
|
||||
a network connectivity configuration from the workspace once attached, you can
|
||||
only switch to another one.
|
||||
|
||||
**Important**: To update a running workspace, your workspace must have no
|
||||
running compute resources that run in your workspace's VPC in the Classic data
|
||||
plane. For example, stop all all-purpose clusters, job clusters, pools with
|
||||
running clusters, and Classic SQL warehouses. If you do not terminate all
|
||||
cluster instances in the workspace before calling this API, the request will
|
||||
fail.
|
||||
|
||||
### Wait until changes take effect. After calling the PATCH operation to
|
||||
update the workspace configuration, make repeated GET requests with the
|
||||
workspace ID and check the workspace status and the status of the fields. *
|
||||
For workspaces with a Databricks-managed VPC, the workspace status becomes
|
||||
PROVISIONING temporarily (typically under 20 minutes). If the workspace
|
||||
update is successful, the workspace status changes to RUNNING. Note that you
|
||||
can also check the workspace status in the [Account Console]. However, you
|
||||
cannot use or create clusters for another 20 minutes after that status change.
|
||||
This results in a total of up to 40 minutes in which you cannot create
|
||||
clusters. If you create or use clusters before this time interval elapses,
|
||||
clusters do not launch successfully, fail, or could cause other unexpected
|
||||
behavior. * For workspaces with a customer-managed VPC, the workspace status
|
||||
stays at status RUNNING and the VPC change happens immediately. A change to
|
||||
the storage customer-managed key configuration ID might take a few minutes to
|
||||
update, so continue to check the workspace until you observe that it has been
|
||||
updated. If the update fails, the workspace might revert silently to its
|
||||
original configuration. After the workspace has been updated, you cannot use
|
||||
or create clusters for another 20 minutes. If you create or use clusters
|
||||
before this time interval elapses, clusters do not launch successfully, fail,
|
||||
or could cause other unexpected behavior.
|
||||
|
||||
If you update the _storage_ customer-managed key configurations, it takes 20
|
||||
minutes for the changes to fully take effect. During the 20 minute wait, it is
|
||||
important that you stop all REST API calls to the DBFS API. If you are
|
||||
modifying _only the managed services key configuration_, you can omit the 20
|
||||
minute wait.
|
||||
|
||||
**Important**: Customer-managed keys and customer-managed VPCs are supported
|
||||
by only some deployment types and subscription types. If you have questions
|
||||
about availability, contact your Databricks representative.
|
||||
|
||||
This operation is available only if your account is on the E2 version of the
|
||||
platform or on a select custom plan that allows multiple workspaces per
|
||||
account.
|
||||
|
||||
[Account Console]: https://docs.databricks.com/administration-guide/account-settings-e2/account-console-e2.html
|
||||
[Create a new workspace using the Account API]: http://docs.databricks.com/administration-guide/account-api/new-workspace.html
|
||||
|
||||
Arguments:
|
||||
WORKSPACE_ID: Workspace ID.
|
||||
|
||||
Usage:
|
||||
databricks account workspaces update WORKSPACE_ID [flags]
|
||||
|
||||
Flags:
|
||||
--aws-region string The AWS region of the workspace's data plane (for example, us-west-2).
|
||||
--credentials-id string ID of the workspace's credential configuration object.
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--managed-services-customer-managed-key-id string The ID of the workspace's managed services encryption key configuration object.
|
||||
--network-connectivity-config-id string
|
||||
--network-id string The ID of the workspace's network configuration object.
|
||||
--no-wait do not wait to reach RUNNING state
|
||||
--private-access-settings-id string The ID of the workspace's private access settings configuration object.
|
||||
--storage-configuration-id string The ID of the workspace's storage configuration object.
|
||||
--storage-customer-managed-key-id string The ID of the key configuration object for workspace storage.
|
||||
--timeout duration maximum amount of time to reach RUNNING state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI account workspaces create --help
|
||||
trace $CLI account workspaces delete --help
|
||||
trace $CLI account workspaces get --help
|
||||
trace $CLI account workspaces list --help
|
||||
trace $CLI account workspaces update --help
|
|
@ -0,0 +1,5 @@
|
|||
|
||||
>>> $CLI access-control-proxy get-assignable-roles-for-resource --help
|
||||
Error: unknown command "access-control-proxy" for "databricks"
|
||||
|
||||
Exit code: 1
|
4
acceptance/help/cmd/workspace/access-control-proxy/access-control-proxy/script
generated
Executable file
4
acceptance/help/cmd/workspace/access-control-proxy/access-control-proxy/script
generated
Executable file
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI access-control-proxy get-assignable-roles-for-resource --help
|
||||
trace $CLI access-control-proxy get-rule-set --help
|
||||
trace $CLI access-control-proxy update-rule-set --help
|
|
@ -0,0 +1,16 @@
|
|||
|
||||
>>> $CLI access-control check-policy --help
|
||||
Check access policy to a resource.
|
||||
|
||||
Usage:
|
||||
databricks access-control check-policy [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for check-policy
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,2 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI access-control check-policy --help
|
|
@ -0,0 +1,5 @@
|
|||
|
||||
>>> $CLI aibi-dashboard-embedding-access-policy delete --help
|
||||
Error: unknown command "aibi-dashboard-embedding-access-policy" for "databricks"
|
||||
|
||||
Exit code: 1
|
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI aibi-dashboard-embedding-access-policy delete --help
|
||||
trace $CLI aibi-dashboard-embedding-access-policy get --help
|
||||
trace $CLI aibi-dashboard-embedding-access-policy update --help
|
|
@ -0,0 +1,5 @@
|
|||
|
||||
>>> $CLI aibi-dashboard-embedding-approved-domains delete --help
|
||||
Error: unknown command "aibi-dashboard-embedding-approved-domains" for "databricks"
|
||||
|
||||
Exit code: 1
|
|
@ -0,0 +1,4 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI aibi-dashboard-embedding-approved-domains delete --help
|
||||
trace $CLI aibi-dashboard-embedding-approved-domains get --help
|
||||
trace $CLI aibi-dashboard-embedding-approved-domains update --help
|
|
@ -0,0 +1,119 @@
|
|||
|
||||
>>> $CLI alerts-legacy create --help
|
||||
Create an alert.
|
||||
|
||||
Creates an alert. An alert is a Databricks SQL object that periodically runs a
|
||||
query, evaluates a condition of its result, and notifies users or notification
|
||||
destinations if the condition was met.
|
||||
|
||||
**Note**: A new version of the Databricks SQL API is now available. Please use
|
||||
:method:alerts/create instead. [Learn more]
|
||||
|
||||
[Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html
|
||||
|
||||
Usage:
|
||||
databricks alerts-legacy create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--parent string The identifier of the workspace folder containing the object.
|
||||
--rearm int Number of seconds after being triggered before the alert rearms itself and can be triggered again.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts-legacy delete --help
|
||||
Delete an alert.
|
||||
|
||||
Deletes an alert. Deleted alerts are no longer accessible and cannot be
|
||||
restored. **Note**: Unlike queries and dashboards, alerts cannot be moved to
|
||||
the trash.
|
||||
|
||||
**Note**: A new version of the Databricks SQL API is now available. Please use
|
||||
:method:alerts/delete instead. [Learn more]
|
||||
|
||||
[Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html
|
||||
|
||||
Usage:
|
||||
databricks alerts-legacy delete ALERT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts-legacy get --help
|
||||
Get an alert.
|
||||
|
||||
Gets an alert.
|
||||
|
||||
**Note**: A new version of the Databricks SQL API is now available. Please use
|
||||
:method:alerts/get instead. [Learn more]
|
||||
|
||||
[Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html
|
||||
|
||||
Usage:
|
||||
databricks alerts-legacy get ALERT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts-legacy list --help
|
||||
Get alerts.
|
||||
|
||||
Gets a list of alerts.
|
||||
|
||||
**Note**: A new version of the Databricks SQL API is now available. Please use
|
||||
:method:alerts/list instead. [Learn more]
|
||||
|
||||
[Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html
|
||||
|
||||
Usage:
|
||||
databricks alerts-legacy list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts-legacy update --help
|
||||
Update an alert.
|
||||
|
||||
Updates an alert.
|
||||
|
||||
**Note**: A new version of the Databricks SQL API is now available. Please use
|
||||
:method:alerts/update instead. [Learn more]
|
||||
|
||||
[Learn more]: https://docs.databricks.com/en/sql/dbsql-api-latest.html
|
||||
|
||||
Usage:
|
||||
databricks alerts-legacy update ALERT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--rearm int Number of seconds after being triggered before the alert rearms itself and can be triggered again.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI alerts-legacy create --help
|
||||
trace $CLI alerts-legacy delete --help
|
||||
trace $CLI alerts-legacy get --help
|
||||
trace $CLI alerts-legacy list --help
|
||||
trace $CLI alerts-legacy update --help
|
|
@ -0,0 +1,100 @@
|
|||
|
||||
>>> $CLI alerts create --help
|
||||
Create an alert.
|
||||
|
||||
Creates an alert.
|
||||
|
||||
Usage:
|
||||
databricks alerts create [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts delete --help
|
||||
Delete an alert.
|
||||
|
||||
Moves an alert to the trash. Trashed alerts immediately disappear from
|
||||
searches and list views, and can no longer trigger. You can restore a trashed
|
||||
alert through the UI. A trashed alert is permanently deleted after 30 days.
|
||||
|
||||
Usage:
|
||||
databricks alerts delete ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts get --help
|
||||
Get an alert.
|
||||
|
||||
Gets an alert.
|
||||
|
||||
Usage:
|
||||
databricks alerts get ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts list --help
|
||||
List alerts.
|
||||
|
||||
Gets a list of alerts accessible to the user, ordered by creation time.
|
||||
**Warning:** Calling this API concurrently 10 or more times could result in
|
||||
throttling, service degradation, or a temporary ban.
|
||||
|
||||
Usage:
|
||||
databricks alerts list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int
|
||||
--page-token string
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI alerts update --help
|
||||
Update an alert.
|
||||
|
||||
Updates an alert.
|
||||
|
||||
Arguments:
|
||||
ID:
|
||||
UPDATE_MASK: Field mask is required to be passed into the PATCH request. Field mask
|
||||
specifies which fields of the setting payload will be updated. The field
|
||||
mask needs to be supplied as single string. To specify multiple fields in
|
||||
the field mask, use comma as the separator (no space).
|
||||
|
||||
Usage:
|
||||
databricks alerts update ID UPDATE_MASK [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,6 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI alerts create --help
|
||||
trace $CLI alerts delete --help
|
||||
trace $CLI alerts get --help
|
||||
trace $CLI alerts list --help
|
||||
trace $CLI alerts update --help
|
|
@ -0,0 +1,308 @@
|
|||
|
||||
>>> $CLI apps create --help
|
||||
Create an app.
|
||||
|
||||
Creates a new app.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app. The name must contain only lowercase alphanumeric
|
||||
characters and hyphens. It must be unique within the workspace.
|
||||
|
||||
Usage:
|
||||
databricks apps create NAME [flags]
|
||||
|
||||
Flags:
|
||||
--description string The description of the app.
|
||||
-h, --help help for create
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--no-compute If true, the app will not be started after creation.
|
||||
--no-wait do not wait to reach ACTIVE state
|
||||
--timeout duration maximum amount of time to reach ACTIVE state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps delete --help
|
||||
Delete an app.
|
||||
|
||||
Deletes an app.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps delete NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for delete
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps deploy --help
|
||||
Create an app deployment.
|
||||
|
||||
Creates an app deployment for the app with the supplied name.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps deploy APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
--deployment-id string The unique id of the deployment.
|
||||
-h, --help help for deploy
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
--mode AppDeploymentMode The mode of which the deployment will manage the source code. Supported values: [AUTO_SYNC, SNAPSHOT]
|
||||
--no-wait do not wait to reach SUCCEEDED state
|
||||
--source-code-path string The workspace file system path of the source code used to create the app deployment.
|
||||
--timeout duration maximum amount of time to reach SUCCEEDED state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps get --help
|
||||
Get an app.
|
||||
|
||||
Retrieves information for the app with the supplied name.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps get NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps get-deployment --help
|
||||
Get an app deployment.
|
||||
|
||||
Retrieves information for the app deployment with the supplied name and
|
||||
deployment id.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The name of the app.
|
||||
DEPLOYMENT_ID: The unique id of the deployment.
|
||||
|
||||
Usage:
|
||||
databricks apps get-deployment APP_NAME DEPLOYMENT_ID [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-deployment
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps get-permission-levels --help
|
||||
Get app permission levels.
|
||||
|
||||
Gets the permission levels that a user can have on an object.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The app for which to get or manage permissions.
|
||||
|
||||
Usage:
|
||||
databricks apps get-permission-levels APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-permission-levels
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps get-permissions --help
|
||||
Get app permissions.
|
||||
|
||||
Gets the permissions of an app. Apps can inherit permissions from their root
|
||||
object.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The app for which to get or manage permissions.
|
||||
|
||||
Usage:
|
||||
databricks apps get-permissions APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get-permissions
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps list --help
|
||||
List apps.
|
||||
|
||||
Lists all apps in the workspace.
|
||||
|
||||
Usage:
|
||||
databricks apps list [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list
|
||||
--page-size int Upper bound for items returned.
|
||||
--page-token string Pagination token to go to the next page of apps.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps list-deployments --help
|
||||
List app deployments.
|
||||
|
||||
Lists all app deployments for the app with the supplied name.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps list-deployments APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for list-deployments
|
||||
--page-size int Upper bound for items returned.
|
||||
--page-token string Pagination token to go to the next page of apps.
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps set-permissions --help
|
||||
Set app permissions.
|
||||
|
||||
Sets permissions on an object, replacing existing permissions if they exist.
|
||||
Deletes all direct permissions if none are specified. Objects can inherit
|
||||
permissions from their root object.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The app for which to get or manage permissions.
|
||||
|
||||
Usage:
|
||||
databricks apps set-permissions APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for set-permissions
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps start --help
|
||||
Start an app.
|
||||
|
||||
Start the last active deployment of the app in the workspace.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps start NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for start
|
||||
--no-wait do not wait to reach ACTIVE state
|
||||
--timeout duration maximum amount of time to reach ACTIVE state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps stop --help
|
||||
Stop an app.
|
||||
|
||||
Stops the active deployment of the app in the workspace.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app.
|
||||
|
||||
Usage:
|
||||
databricks apps stop NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for stop
|
||||
--no-wait do not wait to reach STOPPED state
|
||||
--timeout duration maximum amount of time to reach STOPPED state (default 20m0s)
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps update --help
|
||||
Update an app.
|
||||
|
||||
Updates the app with the supplied name.
|
||||
|
||||
Arguments:
|
||||
NAME: The name of the app. The name must contain only lowercase alphanumeric
|
||||
characters and hyphens. It must be unique within the workspace.
|
||||
|
||||
Usage:
|
||||
databricks apps update NAME [flags]
|
||||
|
||||
Flags:
|
||||
--description string The description of the app.
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI apps update-permissions --help
|
||||
Update app permissions.
|
||||
|
||||
Updates the permissions on an app. Apps can inherit permissions from their
|
||||
root object.
|
||||
|
||||
Arguments:
|
||||
APP_NAME: The app for which to get or manage permissions.
|
||||
|
||||
Usage:
|
||||
databricks apps update-permissions APP_NAME [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update-permissions
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
|
@ -0,0 +1,15 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI apps create --help
|
||||
trace $CLI apps delete --help
|
||||
trace $CLI apps deploy --help
|
||||
trace $CLI apps get --help
|
||||
trace $CLI apps get-deployment --help
|
||||
trace $CLI apps get-permission-levels --help
|
||||
trace $CLI apps get-permissions --help
|
||||
trace $CLI apps list --help
|
||||
trace $CLI apps list-deployments --help
|
||||
trace $CLI apps set-permissions --help
|
||||
trace $CLI apps start --help
|
||||
trace $CLI apps stop --help
|
||||
trace $CLI apps update --help
|
||||
trace $CLI apps update-permissions --help
|
|
@ -0,0 +1,44 @@
|
|||
|
||||
>>> $CLI artifact-allowlists get --help
|
||||
Get an artifact allowlist.
|
||||
|
||||
Get the artifact allowlist of a certain artifact type. The caller must be a
|
||||
metastore admin or have the **MANAGE ALLOWLIST** privilege on the metastore.
|
||||
|
||||
Arguments:
|
||||
ARTIFACT_TYPE: The artifact type of the allowlist.
|
||||
|
||||
Usage:
|
||||
databricks artifact-allowlists get ARTIFACT_TYPE [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for get
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
||||
|
||||
>>> $CLI artifact-allowlists update --help
|
||||
Set an artifact allowlist.
|
||||
|
||||
Set the artifact allowlist of a certain artifact type. The whole artifact
|
||||
allowlist is replaced with the new allowlist. The caller must be a metastore
|
||||
admin or have the **MANAGE ALLOWLIST** privilege on the metastore.
|
||||
|
||||
Arguments:
|
||||
ARTIFACT_TYPE: The artifact type of the allowlist.
|
||||
|
||||
Usage:
|
||||
databricks artifact-allowlists update ARTIFACT_TYPE [flags]
|
||||
|
||||
Flags:
|
||||
-h, --help help for update
|
||||
--json JSON either inline JSON string or @path/to/file.json with request body (default JSON (0 bytes))
|
||||
|
||||
Global Flags:
|
||||
--debug enable debug logging
|
||||
-o, --output type output type: text or json (default text)
|
||||
-p, --profile string ~/.databrickscfg profile
|
||||
-t, --target string bundle target to use (if applicable)
|
3
acceptance/help/cmd/workspace/artifact-allowlists/artifact-allowlists/script
generated
Executable file
3
acceptance/help/cmd/workspace/artifact-allowlists/artifact-allowlists/script
generated
Executable file
|
@ -0,0 +1,3 @@
|
|||
# Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
||||
trace $CLI artifact-allowlists get --help
|
||||
trace $CLI artifact-allowlists update --help
|
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in New Issue