mirror of https://github.com/databricks/cli.git
Compare commits
No commits in common. "58f1ce9498eacce21114eac01e03d25dc989d946" and "037ded96789814458e89bd188a4a150eb97480e5" have entirely different histories.
58f1ce9498
...
037ded9678
|
@ -1 +1 @@
|
||||||
cf9c61453990df0f9453670f2fe68e1b128647a2
|
0c86ea6dbd9a730c24ff0d4e509603e476955ac5
|
|
@ -54,7 +54,6 @@ cmd/workspace/dashboards/dashboards.go linguist-generated=true
|
||||||
cmd/workspace/data-sources/data-sources.go linguist-generated=true
|
cmd/workspace/data-sources/data-sources.go linguist-generated=true
|
||||||
cmd/workspace/default-namespace/default-namespace.go linguist-generated=true
|
cmd/workspace/default-namespace/default-namespace.go linguist-generated=true
|
||||||
cmd/workspace/disable-legacy-access/disable-legacy-access.go linguist-generated=true
|
cmd/workspace/disable-legacy-access/disable-legacy-access.go linguist-generated=true
|
||||||
cmd/workspace/disable-legacy-dbfs/disable-legacy-dbfs.go linguist-generated=true
|
|
||||||
cmd/workspace/enhanced-security-monitoring/enhanced-security-monitoring.go linguist-generated=true
|
cmd/workspace/enhanced-security-monitoring/enhanced-security-monitoring.go linguist-generated=true
|
||||||
cmd/workspace/experiments/experiments.go linguist-generated=true
|
cmd/workspace/experiments/experiments.go linguist-generated=true
|
||||||
cmd/workspace/external-locations/external-locations.go linguist-generated=true
|
cmd/workspace/external-locations/external-locations.go linguist-generated=true
|
||||||
|
|
|
@ -1,3 +1,3 @@
|
||||||
package schema
|
package schema
|
||||||
|
|
||||||
const ProviderVersion = "1.54.0"
|
const ProviderVersion = "1.53.0"
|
||||||
|
|
|
@ -1,15 +0,0 @@
|
||||||
// Generated from Databricks Terraform provider schema. DO NOT EDIT.
|
|
||||||
|
|
||||||
package schema
|
|
||||||
|
|
||||||
type DataSourceNotificationDestinationsNotificationDestinations struct {
|
|
||||||
DestinationType string `json:"destination_type,omitempty"`
|
|
||||||
DisplayName string `json:"display_name,omitempty"`
|
|
||||||
Id string `json:"id,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type DataSourceNotificationDestinations struct {
|
|
||||||
DisplayNameContains string `json:"display_name_contains,omitempty"`
|
|
||||||
Type string `json:"type,omitempty"`
|
|
||||||
NotificationDestinations []DataSourceNotificationDestinationsNotificationDestinations `json:"notification_destinations,omitempty"`
|
|
||||||
}
|
|
|
@ -1,32 +0,0 @@
|
||||||
// Generated from Databricks Terraform provider schema. DO NOT EDIT.
|
|
||||||
|
|
||||||
package schema
|
|
||||||
|
|
||||||
type DataSourceRegisteredModelModelInfoAliases struct {
|
|
||||||
AliasName string `json:"alias_name,omitempty"`
|
|
||||||
VersionNum int `json:"version_num,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type DataSourceRegisteredModelModelInfo struct {
|
|
||||||
BrowseOnly bool `json:"browse_only,omitempty"`
|
|
||||||
CatalogName string `json:"catalog_name,omitempty"`
|
|
||||||
Comment string `json:"comment,omitempty"`
|
|
||||||
CreatedAt int `json:"created_at,omitempty"`
|
|
||||||
CreatedBy string `json:"created_by,omitempty"`
|
|
||||||
FullName string `json:"full_name,omitempty"`
|
|
||||||
MetastoreId string `json:"metastore_id,omitempty"`
|
|
||||||
Name string `json:"name,omitempty"`
|
|
||||||
Owner string `json:"owner,omitempty"`
|
|
||||||
SchemaName string `json:"schema_name,omitempty"`
|
|
||||||
StorageLocation string `json:"storage_location,omitempty"`
|
|
||||||
UpdatedAt int `json:"updated_at,omitempty"`
|
|
||||||
UpdatedBy string `json:"updated_by,omitempty"`
|
|
||||||
Aliases []DataSourceRegisteredModelModelInfoAliases `json:"aliases,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type DataSourceRegisteredModel struct {
|
|
||||||
FullName string `json:"full_name"`
|
|
||||||
IncludeAliases bool `json:"include_aliases,omitempty"`
|
|
||||||
IncludeBrowse bool `json:"include_browse,omitempty"`
|
|
||||||
ModelInfo []DataSourceRegisteredModelModelInfo `json:"model_info,omitempty"`
|
|
||||||
}
|
|
|
@ -36,9 +36,7 @@ type DataSources struct {
|
||||||
NodeType map[string]any `json:"databricks_node_type,omitempty"`
|
NodeType map[string]any `json:"databricks_node_type,omitempty"`
|
||||||
Notebook map[string]any `json:"databricks_notebook,omitempty"`
|
Notebook map[string]any `json:"databricks_notebook,omitempty"`
|
||||||
NotebookPaths map[string]any `json:"databricks_notebook_paths,omitempty"`
|
NotebookPaths map[string]any `json:"databricks_notebook_paths,omitempty"`
|
||||||
NotificationDestinations map[string]any `json:"databricks_notification_destinations,omitempty"`
|
|
||||||
Pipelines map[string]any `json:"databricks_pipelines,omitempty"`
|
Pipelines map[string]any `json:"databricks_pipelines,omitempty"`
|
||||||
RegisteredModel map[string]any `json:"databricks_registered_model,omitempty"`
|
|
||||||
Schema map[string]any `json:"databricks_schema,omitempty"`
|
Schema map[string]any `json:"databricks_schema,omitempty"`
|
||||||
Schemas map[string]any `json:"databricks_schemas,omitempty"`
|
Schemas map[string]any `json:"databricks_schemas,omitempty"`
|
||||||
ServicePrincipal map[string]any `json:"databricks_service_principal,omitempty"`
|
ServicePrincipal map[string]any `json:"databricks_service_principal,omitempty"`
|
||||||
|
@ -94,9 +92,7 @@ func NewDataSources() *DataSources {
|
||||||
NodeType: make(map[string]any),
|
NodeType: make(map[string]any),
|
||||||
Notebook: make(map[string]any),
|
Notebook: make(map[string]any),
|
||||||
NotebookPaths: make(map[string]any),
|
NotebookPaths: make(map[string]any),
|
||||||
NotificationDestinations: make(map[string]any),
|
|
||||||
Pipelines: make(map[string]any),
|
Pipelines: make(map[string]any),
|
||||||
RegisteredModel: make(map[string]any),
|
|
||||||
Schema: make(map[string]any),
|
Schema: make(map[string]any),
|
||||||
Schemas: make(map[string]any),
|
Schemas: make(map[string]any),
|
||||||
ServicePrincipal: make(map[string]any),
|
ServicePrincipal: make(map[string]any),
|
||||||
|
|
|
@ -1448,7 +1448,6 @@ type ResourceJobWebhookNotifications struct {
|
||||||
|
|
||||||
type ResourceJob struct {
|
type ResourceJob struct {
|
||||||
AlwaysRunning bool `json:"always_running,omitempty"`
|
AlwaysRunning bool `json:"always_running,omitempty"`
|
||||||
BudgetPolicyId string `json:"budget_policy_id,omitempty"`
|
|
||||||
ControlRunState bool `json:"control_run_state,omitempty"`
|
ControlRunState bool `json:"control_run_state,omitempty"`
|
||||||
Description string `json:"description,omitempty"`
|
Description string `json:"description,omitempty"`
|
||||||
EditMode string `json:"edit_mode,omitempty"`
|
EditMode string `json:"edit_mode,omitempty"`
|
||||||
|
|
|
@ -19,10 +19,9 @@ type ResourceOnlineTableSpec struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourceOnlineTable struct {
|
type ResourceOnlineTable struct {
|
||||||
Id string `json:"id,omitempty"`
|
Id string `json:"id,omitempty"`
|
||||||
Name string `json:"name"`
|
Name string `json:"name"`
|
||||||
Status []any `json:"status,omitempty"`
|
Status []any `json:"status,omitempty"`
|
||||||
TableServingUrl string `json:"table_serving_url,omitempty"`
|
TableServingUrl string `json:"table_serving_url,omitempty"`
|
||||||
UnityCatalogProvisioningState string `json:"unity_catalog_provisioning_state,omitempty"`
|
Spec *ResourceOnlineTableSpec `json:"spec,omitempty"`
|
||||||
Spec *ResourceOnlineTableSpec `json:"spec,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -142,26 +142,10 @@ type ResourcePipelineGatewayDefinition struct {
|
||||||
GatewayStorageSchema string `json:"gateway_storage_schema,omitempty"`
|
GatewayStorageSchema string `json:"gateway_storage_schema,omitempty"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjectsReportTableConfiguration struct {
|
|
||||||
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
|
||||||
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
|
||||||
ScdType string `json:"scd_type,omitempty"`
|
|
||||||
SequenceBy []string `json:"sequence_by,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjectsReport struct {
|
|
||||||
DestinationCatalog string `json:"destination_catalog,omitempty"`
|
|
||||||
DestinationSchema string `json:"destination_schema,omitempty"`
|
|
||||||
DestinationTable string `json:"destination_table,omitempty"`
|
|
||||||
SourceUrl string `json:"source_url,omitempty"`
|
|
||||||
TableConfiguration *ResourcePipelineIngestionDefinitionObjectsReportTableConfiguration `json:"table_configuration,omitempty"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjectsSchemaTableConfiguration struct {
|
type ResourcePipelineIngestionDefinitionObjectsSchemaTableConfiguration struct {
|
||||||
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
||||||
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
||||||
ScdType string `json:"scd_type,omitempty"`
|
ScdType string `json:"scd_type,omitempty"`
|
||||||
SequenceBy []string `json:"sequence_by,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjectsSchema struct {
|
type ResourcePipelineIngestionDefinitionObjectsSchema struct {
|
||||||
|
@ -176,7 +160,6 @@ type ResourcePipelineIngestionDefinitionObjectsTableTableConfiguration struct {
|
||||||
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
||||||
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
||||||
ScdType string `json:"scd_type,omitempty"`
|
ScdType string `json:"scd_type,omitempty"`
|
||||||
SequenceBy []string `json:"sequence_by,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjectsTable struct {
|
type ResourcePipelineIngestionDefinitionObjectsTable struct {
|
||||||
|
@ -190,7 +173,6 @@ type ResourcePipelineIngestionDefinitionObjectsTable struct {
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinitionObjects struct {
|
type ResourcePipelineIngestionDefinitionObjects struct {
|
||||||
Report *ResourcePipelineIngestionDefinitionObjectsReport `json:"report,omitempty"`
|
|
||||||
Schema *ResourcePipelineIngestionDefinitionObjectsSchema `json:"schema,omitempty"`
|
Schema *ResourcePipelineIngestionDefinitionObjectsSchema `json:"schema,omitempty"`
|
||||||
Table *ResourcePipelineIngestionDefinitionObjectsTable `json:"table,omitempty"`
|
Table *ResourcePipelineIngestionDefinitionObjectsTable `json:"table,omitempty"`
|
||||||
}
|
}
|
||||||
|
@ -199,7 +181,6 @@ type ResourcePipelineIngestionDefinitionTableConfiguration struct {
|
||||||
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
PrimaryKeys []string `json:"primary_keys,omitempty"`
|
||||||
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
SalesforceIncludeFormulaFields bool `json:"salesforce_include_formula_fields,omitempty"`
|
||||||
ScdType string `json:"scd_type,omitempty"`
|
ScdType string `json:"scd_type,omitempty"`
|
||||||
SequenceBy []string `json:"sequence_by,omitempty"`
|
|
||||||
}
|
}
|
||||||
|
|
||||||
type ResourcePipelineIngestionDefinition struct {
|
type ResourcePipelineIngestionDefinition struct {
|
||||||
|
|
|
@ -21,7 +21,7 @@ type Root struct {
|
||||||
|
|
||||||
const ProviderHost = "registry.terraform.io"
|
const ProviderHost = "registry.terraform.io"
|
||||||
const ProviderSource = "databricks/databricks"
|
const ProviderSource = "databricks/databricks"
|
||||||
const ProviderVersion = "1.54.0"
|
const ProviderVersion = "1.53.0"
|
||||||
|
|
||||||
func NewRoot() *Root {
|
func NewRoot() *Root {
|
||||||
return &Root{
|
return &Root{
|
||||||
|
|
|
@ -320,11 +320,9 @@ func (r *jobRunner) Cancel(ctx context.Context) error {
|
||||||
func (r *jobRunner) Restart(ctx context.Context, opts *Options) (output.RunOutput, error) {
|
func (r *jobRunner) Restart(ctx context.Context, opts *Options) (output.RunOutput, error) {
|
||||||
// We don't need to cancel existing runs if the job is continuous and unpaused.
|
// We don't need to cancel existing runs if the job is continuous and unpaused.
|
||||||
// the /jobs/run-now API will automatically cancel any existing runs before starting a new one.
|
// the /jobs/run-now API will automatically cancel any existing runs before starting a new one.
|
||||||
//
|
|
||||||
// /jobs/run-now will not cancel existing runs if the job is continuous and paused.
|
|
||||||
// New job runs will be queued instead and will wait for existing runs to finish.
|
|
||||||
// In this case, we need to cancel the existing runs before starting a new one.
|
|
||||||
continuous := r.job.JobSettings.Continuous
|
continuous := r.job.JobSettings.Continuous
|
||||||
|
// TODO: Is this correct wrt pause status?
|
||||||
|
// TODO: test all permutations, even pipelines.
|
||||||
if continuous != nil && continuous.PauseStatus == jobs.PauseStatusUnpaused {
|
if continuous != nil && continuous.PauseStatus == jobs.PauseStatusUnpaused {
|
||||||
return r.Run(ctx, opts)
|
return r.Run(ctx, opts)
|
||||||
}
|
}
|
||||||
|
|
|
@ -131,78 +131,69 @@ func TestJobRunnerCancelWithNoActiveRuns(t *testing.T) {
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestJobRunnerRestart(t *testing.T) {
|
func TestJobRunnerRestart(t *testing.T) {
|
||||||
for _, jobSettings := range []*jobs.JobSettings{
|
job := &resources.Job{
|
||||||
{},
|
ID: "123",
|
||||||
{
|
JobSettings: &jobs.JobSettings{},
|
||||||
Continuous: &jobs.Continuous{
|
}
|
||||||
PauseStatus: jobs.PauseStatusPaused,
|
b := &bundle.Bundle{
|
||||||
},
|
Config: config.Root{
|
||||||
},
|
Resources: config.Resources{
|
||||||
} {
|
Jobs: map[string]*resources.Job{
|
||||||
job := &resources.Job{
|
"test_job": job,
|
||||||
ID: "123",
|
|
||||||
JobSettings: jobSettings,
|
|
||||||
}
|
|
||||||
b := &bundle.Bundle{
|
|
||||||
Config: config.Root{
|
|
||||||
Resources: config.Resources{
|
|
||||||
Jobs: map[string]*resources.Job{
|
|
||||||
"test_job": job,
|
|
||||||
},
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
},
|
||||||
|
|
||||||
runner := jobRunner{key: "test", bundle: b, job: job}
|
|
||||||
|
|
||||||
m := mocks.NewMockWorkspaceClient(t)
|
|
||||||
b.SetWorkpaceClient(m.WorkspaceClient)
|
|
||||||
ctx := context.Background()
|
|
||||||
ctx = cmdio.InContext(ctx, cmdio.NewIO(flags.OutputText, &bytes.Buffer{}, &bytes.Buffer{}, &bytes.Buffer{}, "", ""))
|
|
||||||
ctx = cmdio.NewContext(ctx, cmdio.NewLogger(flags.ModeAppend))
|
|
||||||
|
|
||||||
jobApi := m.GetMockJobsAPI()
|
|
||||||
jobApi.EXPECT().ListRunsAll(mock.Anything, jobs.ListRunsRequest{
|
|
||||||
ActiveOnly: true,
|
|
||||||
JobId: 123,
|
|
||||||
}).Return([]jobs.BaseRun{
|
|
||||||
{RunId: 1},
|
|
||||||
{RunId: 2},
|
|
||||||
}, nil)
|
|
||||||
|
|
||||||
// Mock the runner cancelling existing job runs.
|
|
||||||
mockWait := &jobs.WaitGetRunJobTerminatedOrSkipped[struct{}]{
|
|
||||||
Poll: func(time time.Duration, f func(j *jobs.Run)) (*jobs.Run, error) {
|
|
||||||
return nil, nil
|
|
||||||
},
|
|
||||||
}
|
|
||||||
jobApi.EXPECT().CancelRun(mock.Anything, jobs.CancelRun{
|
|
||||||
RunId: 1,
|
|
||||||
}).Return(mockWait, nil)
|
|
||||||
jobApi.EXPECT().CancelRun(mock.Anything, jobs.CancelRun{
|
|
||||||
RunId: 2,
|
|
||||||
}).Return(mockWait, nil)
|
|
||||||
|
|
||||||
// Mock the runner triggering a job run
|
|
||||||
mockWaitForRun := &jobs.WaitGetRunJobTerminatedOrSkipped[jobs.RunNowResponse]{
|
|
||||||
Poll: func(d time.Duration, f func(*jobs.Run)) (*jobs.Run, error) {
|
|
||||||
return &jobs.Run{
|
|
||||||
State: &jobs.RunState{
|
|
||||||
ResultState: jobs.RunResultStateSuccess,
|
|
||||||
},
|
|
||||||
}, nil
|
|
||||||
},
|
|
||||||
}
|
|
||||||
jobApi.EXPECT().RunNow(mock.Anything, jobs.RunNow{
|
|
||||||
JobId: 123,
|
|
||||||
}).Return(mockWaitForRun, nil)
|
|
||||||
|
|
||||||
// Mock the runner getting the job output
|
|
||||||
jobApi.EXPECT().GetRun(mock.Anything, jobs.GetRunRequest{}).Return(&jobs.Run{}, nil)
|
|
||||||
|
|
||||||
_, err := runner.Restart(ctx, &Options{})
|
|
||||||
require.NoError(t, err)
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
runner := jobRunner{key: "test", bundle: b, job: job}
|
||||||
|
|
||||||
|
m := mocks.NewMockWorkspaceClient(t)
|
||||||
|
b.SetWorkpaceClient(m.WorkspaceClient)
|
||||||
|
ctx := context.Background()
|
||||||
|
ctx = cmdio.InContext(ctx, cmdio.NewIO(flags.OutputText, &bytes.Buffer{}, &bytes.Buffer{}, &bytes.Buffer{}, "", "..."))
|
||||||
|
ctx = cmdio.NewContext(ctx, cmdio.NewLogger(flags.ModeAppend))
|
||||||
|
|
||||||
|
jobApi := m.GetMockJobsAPI()
|
||||||
|
jobApi.EXPECT().ListRunsAll(mock.Anything, jobs.ListRunsRequest{
|
||||||
|
ActiveOnly: true,
|
||||||
|
JobId: 123,
|
||||||
|
}).Return([]jobs.BaseRun{
|
||||||
|
{RunId: 1},
|
||||||
|
{RunId: 2},
|
||||||
|
}, nil)
|
||||||
|
|
||||||
|
// Mock the runner cancelling existing job runs.
|
||||||
|
mockWait := &jobs.WaitGetRunJobTerminatedOrSkipped[struct{}]{
|
||||||
|
Poll: func(time time.Duration, f func(j *jobs.Run)) (*jobs.Run, error) {
|
||||||
|
return nil, nil
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jobApi.EXPECT().CancelRun(mock.Anything, jobs.CancelRun{
|
||||||
|
RunId: 1,
|
||||||
|
}).Return(mockWait, nil)
|
||||||
|
jobApi.EXPECT().CancelRun(mock.Anything, jobs.CancelRun{
|
||||||
|
RunId: 2,
|
||||||
|
}).Return(mockWait, nil)
|
||||||
|
|
||||||
|
// Mock the runner triggering a job run
|
||||||
|
mockWaitForRun := &jobs.WaitGetRunJobTerminatedOrSkipped[jobs.RunNowResponse]{
|
||||||
|
Poll: func(d time.Duration, f func(*jobs.Run)) (*jobs.Run, error) {
|
||||||
|
return &jobs.Run{
|
||||||
|
State: &jobs.RunState{
|
||||||
|
ResultState: jobs.RunResultStateSuccess,
|
||||||
|
},
|
||||||
|
}, nil
|
||||||
|
},
|
||||||
|
}
|
||||||
|
jobApi.EXPECT().RunNow(mock.Anything, jobs.RunNow{
|
||||||
|
JobId: 123,
|
||||||
|
}).Return(mockWaitForRun, nil)
|
||||||
|
|
||||||
|
// Mock the runner getting the job output
|
||||||
|
jobApi.EXPECT().GetRun(mock.Anything, jobs.GetRunRequest{}).Return(&jobs.Run{}, nil)
|
||||||
|
|
||||||
|
_, err := runner.Restart(ctx, &Options{})
|
||||||
|
require.NoError(t, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestJobRunnerRestartForContinuousUnpausedJobs(t *testing.T) {
|
func TestJobRunnerRestartForContinuousUnpausedJobs(t *testing.T) {
|
||||||
|
|
|
@ -115,5 +115,6 @@ func TestPipelineRunnerRestart(t *testing.T) {
|
||||||
}, nil)
|
}, nil)
|
||||||
|
|
||||||
_, err := runner.Restart(ctx, &Options{})
|
_, err := runner.Restart(ctx, &Options{})
|
||||||
|
// TODO: assert on the output.
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
}
|
}
|
||||||
|
|
|
@ -209,10 +209,6 @@
|
||||||
{
|
{
|
||||||
"type": "object",
|
"type": "object",
|
||||||
"properties": {
|
"properties": {
|
||||||
"budget_policy_id": {
|
|
||||||
"description": "The id of the user specified budget policy to use for this job.\nIf not specified, a default budget policy may be applied when creating or modifying the job.\nSee `effective_budget_policy_id` for the budget policy used by this workload.",
|
|
||||||
"$ref": "#/$defs/string"
|
|
||||||
},
|
|
||||||
"continuous": {
|
"continuous": {
|
||||||
"description": "An optional continuous property for this job. The continuous property will ensure that there is always one run executing. Only one of `schedule` and `continuous` can be used.",
|
"description": "An optional continuous property for this job. The continuous property will ensure that there is always one run executing. Only one of `schedule` and `continuous` can be used.",
|
||||||
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/jobs.Continuous"
|
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/jobs.Continuous"
|
||||||
|
@ -3905,10 +3901,6 @@
|
||||||
{
|
{
|
||||||
"type": "object",
|
"type": "object",
|
||||||
"properties": {
|
"properties": {
|
||||||
"report": {
|
|
||||||
"description": "Select tables from a specific source report.",
|
|
||||||
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/pipelines.ReportSpec"
|
|
||||||
},
|
|
||||||
"schema": {
|
"schema": {
|
||||||
"description": "Select tables from a specific source schema.",
|
"description": "Select tables from a specific source schema.",
|
||||||
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/pipelines.SchemaSpec"
|
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/pipelines.SchemaSpec"
|
||||||
|
@ -4241,40 +4233,6 @@
|
||||||
}
|
}
|
||||||
]
|
]
|
||||||
},
|
},
|
||||||
"pipelines.ReportSpec": {
|
|
||||||
"anyOf": [
|
|
||||||
{
|
|
||||||
"type": "object",
|
|
||||||
"properties": {
|
|
||||||
"destination_catalog": {
|
|
||||||
"description": "Required. Destination catalog to store table.",
|
|
||||||
"$ref": "#/$defs/string"
|
|
||||||
},
|
|
||||||
"destination_schema": {
|
|
||||||
"description": "Required. Destination schema to store table.",
|
|
||||||
"$ref": "#/$defs/string"
|
|
||||||
},
|
|
||||||
"destination_table": {
|
|
||||||
"description": "Required. Destination table name. The pipeline fails if a table with that name already exists.",
|
|
||||||
"$ref": "#/$defs/string"
|
|
||||||
},
|
|
||||||
"source_url": {
|
|
||||||
"description": "Required. Report URL in the source system.",
|
|
||||||
"$ref": "#/$defs/string"
|
|
||||||
},
|
|
||||||
"table_configuration": {
|
|
||||||
"description": "Configuration settings to control the ingestion of tables. These settings override the table_configuration defined in the IngestionPipelineDefinition object.",
|
|
||||||
"$ref": "#/$defs/github.com/databricks/databricks-sdk-go/service/pipelines.TableSpecificConfig"
|
|
||||||
}
|
|
||||||
},
|
|
||||||
"additionalProperties": false
|
|
||||||
},
|
|
||||||
{
|
|
||||||
"type": "string",
|
|
||||||
"pattern": "\\$\\{(var(\\.[a-zA-Z]+([-_]?[a-zA-Z0-9]+)*(\\[[0-9]+\\])*)+)\\}"
|
|
||||||
}
|
|
||||||
]
|
|
||||||
},
|
|
||||||
"pipelines.SchemaSpec": {
|
"pipelines.SchemaSpec": {
|
||||||
"anyOf": [
|
"anyOf": [
|
||||||
{
|
{
|
||||||
|
@ -4323,7 +4281,7 @@
|
||||||
"$ref": "#/$defs/string"
|
"$ref": "#/$defs/string"
|
||||||
},
|
},
|
||||||
"destination_table": {
|
"destination_table": {
|
||||||
"description": "Optional. Destination table name. The pipeline fails if a table with that name already exists. If not set, the source table name is used.",
|
"description": "Optional. Destination table name. The pipeline fails If a table with that name already exists. If not set, the source table name is used.",
|
||||||
"$ref": "#/$defs/string"
|
"$ref": "#/$defs/string"
|
||||||
},
|
},
|
||||||
"source_catalog": {
|
"source_catalog": {
|
||||||
|
@ -4371,10 +4329,6 @@
|
||||||
"SCD_TYPE_1",
|
"SCD_TYPE_1",
|
||||||
"SCD_TYPE_2"
|
"SCD_TYPE_2"
|
||||||
]
|
]
|
||||||
},
|
|
||||||
"sequence_by": {
|
|
||||||
"description": "The column names specifying the logical order of events in the source data. Delta Live Tables uses this sequencing to handle change events that arrive out of order.",
|
|
||||||
"$ref": "#/$defs/slice/string"
|
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"additionalProperties": false
|
"additionalProperties": false
|
||||||
|
|
|
@ -101,6 +101,7 @@ task or a Python wheel task, the second example applies.
|
||||||
}
|
}
|
||||||
|
|
||||||
runOptions.NoWait = noWait
|
runOptions.NoWait = noWait
|
||||||
|
|
||||||
var output output.RunOutput
|
var output output.RunOutput
|
||||||
if restart {
|
if restart {
|
||||||
output, err = runner.Restart(ctx, &runOptions)
|
output, err = runner.Restart(ctx, &runOptions)
|
||||||
|
|
|
@ -28,6 +28,9 @@ func New() *cobra.Command {
|
||||||
Annotations: map[string]string{
|
Annotations: map[string]string{
|
||||||
"package": "apps",
|
"package": "apps",
|
||||||
},
|
},
|
||||||
|
|
||||||
|
// This service is being previewed; hide from help output.
|
||||||
|
Hidden: true,
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add methods
|
// Add methods
|
||||||
|
|
|
@ -1,220 +0,0 @@
|
||||||
// Code generated from OpenAPI specs by Databricks SDK Generator. DO NOT EDIT.
|
|
||||||
|
|
||||||
package disable_legacy_dbfs
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
|
|
||||||
"github.com/databricks/cli/cmd/root"
|
|
||||||
"github.com/databricks/cli/libs/cmdio"
|
|
||||||
"github.com/databricks/cli/libs/flags"
|
|
||||||
"github.com/databricks/databricks-sdk-go/service/settings"
|
|
||||||
"github.com/spf13/cobra"
|
|
||||||
)
|
|
||||||
|
|
||||||
// Slice with functions to override default command behavior.
|
|
||||||
// Functions can be added from the `init()` function in manually curated files in this directory.
|
|
||||||
var cmdOverrides []func(*cobra.Command)
|
|
||||||
|
|
||||||
func New() *cobra.Command {
|
|
||||||
cmd := &cobra.Command{
|
|
||||||
Use: "disable-legacy-dbfs",
|
|
||||||
Short: `When this setting is on, access to DBFS root and DBFS mounts is disallowed (as well as creation of new mounts).`,
|
|
||||||
Long: `When this setting is on, access to DBFS root and DBFS mounts is disallowed (as
|
|
||||||
well as creation of new mounts). When the setting is off, all DBFS
|
|
||||||
functionality is enabled`,
|
|
||||||
|
|
||||||
// This service is being previewed; hide from help output.
|
|
||||||
Hidden: true,
|
|
||||||
}
|
|
||||||
|
|
||||||
// Add methods
|
|
||||||
cmd.AddCommand(newDelete())
|
|
||||||
cmd.AddCommand(newGet())
|
|
||||||
cmd.AddCommand(newUpdate())
|
|
||||||
|
|
||||||
// Apply optional overrides to this command.
|
|
||||||
for _, fn := range cmdOverrides {
|
|
||||||
fn(cmd)
|
|
||||||
}
|
|
||||||
|
|
||||||
return cmd
|
|
||||||
}
|
|
||||||
|
|
||||||
// start delete command
|
|
||||||
|
|
||||||
// Slice with functions to override default command behavior.
|
|
||||||
// Functions can be added from the `init()` function in manually curated files in this directory.
|
|
||||||
var deleteOverrides []func(
|
|
||||||
*cobra.Command,
|
|
||||||
*settings.DeleteDisableLegacyDbfsRequest,
|
|
||||||
)
|
|
||||||
|
|
||||||
func newDelete() *cobra.Command {
|
|
||||||
cmd := &cobra.Command{}
|
|
||||||
|
|
||||||
var deleteReq settings.DeleteDisableLegacyDbfsRequest
|
|
||||||
|
|
||||||
// TODO: short flags
|
|
||||||
|
|
||||||
cmd.Flags().StringVar(&deleteReq.Etag, "etag", deleteReq.Etag, `etag used for versioning.`)
|
|
||||||
|
|
||||||
cmd.Use = "delete"
|
|
||||||
cmd.Short = `Delete the disable legacy DBFS setting.`
|
|
||||||
cmd.Long = `Delete the disable legacy DBFS setting.
|
|
||||||
|
|
||||||
Deletes the disable legacy DBFS setting for a workspace, reverting back to the
|
|
||||||
default.`
|
|
||||||
|
|
||||||
cmd.Annotations = make(map[string]string)
|
|
||||||
|
|
||||||
cmd.Args = func(cmd *cobra.Command, args []string) error {
|
|
||||||
check := root.ExactArgs(0)
|
|
||||||
return check(cmd, args)
|
|
||||||
}
|
|
||||||
|
|
||||||
cmd.PreRunE = root.MustWorkspaceClient
|
|
||||||
cmd.RunE = func(cmd *cobra.Command, args []string) (err error) {
|
|
||||||
ctx := cmd.Context()
|
|
||||||
w := root.WorkspaceClient(ctx)
|
|
||||||
|
|
||||||
response, err := w.Settings.DisableLegacyDbfs().Delete(ctx, deleteReq)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return cmdio.Render(ctx, response)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Disable completions since they are not applicable.
|
|
||||||
// Can be overridden by manual implementation in `override.go`.
|
|
||||||
cmd.ValidArgsFunction = cobra.NoFileCompletions
|
|
||||||
|
|
||||||
// Apply optional overrides to this command.
|
|
||||||
for _, fn := range deleteOverrides {
|
|
||||||
fn(cmd, &deleteReq)
|
|
||||||
}
|
|
||||||
|
|
||||||
return cmd
|
|
||||||
}
|
|
||||||
|
|
||||||
// start get command
|
|
||||||
|
|
||||||
// Slice with functions to override default command behavior.
|
|
||||||
// Functions can be added from the `init()` function in manually curated files in this directory.
|
|
||||||
var getOverrides []func(
|
|
||||||
*cobra.Command,
|
|
||||||
*settings.GetDisableLegacyDbfsRequest,
|
|
||||||
)
|
|
||||||
|
|
||||||
func newGet() *cobra.Command {
|
|
||||||
cmd := &cobra.Command{}
|
|
||||||
|
|
||||||
var getReq settings.GetDisableLegacyDbfsRequest
|
|
||||||
|
|
||||||
// TODO: short flags
|
|
||||||
|
|
||||||
cmd.Flags().StringVar(&getReq.Etag, "etag", getReq.Etag, `etag used for versioning.`)
|
|
||||||
|
|
||||||
cmd.Use = "get"
|
|
||||||
cmd.Short = `Get the disable legacy DBFS setting.`
|
|
||||||
cmd.Long = `Get the disable legacy DBFS setting.
|
|
||||||
|
|
||||||
Gets the disable legacy DBFS setting.`
|
|
||||||
|
|
||||||
cmd.Annotations = make(map[string]string)
|
|
||||||
|
|
||||||
cmd.Args = func(cmd *cobra.Command, args []string) error {
|
|
||||||
check := root.ExactArgs(0)
|
|
||||||
return check(cmd, args)
|
|
||||||
}
|
|
||||||
|
|
||||||
cmd.PreRunE = root.MustWorkspaceClient
|
|
||||||
cmd.RunE = func(cmd *cobra.Command, args []string) (err error) {
|
|
||||||
ctx := cmd.Context()
|
|
||||||
w := root.WorkspaceClient(ctx)
|
|
||||||
|
|
||||||
response, err := w.Settings.DisableLegacyDbfs().Get(ctx, getReq)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return cmdio.Render(ctx, response)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Disable completions since they are not applicable.
|
|
||||||
// Can be overridden by manual implementation in `override.go`.
|
|
||||||
cmd.ValidArgsFunction = cobra.NoFileCompletions
|
|
||||||
|
|
||||||
// Apply optional overrides to this command.
|
|
||||||
for _, fn := range getOverrides {
|
|
||||||
fn(cmd, &getReq)
|
|
||||||
}
|
|
||||||
|
|
||||||
return cmd
|
|
||||||
}
|
|
||||||
|
|
||||||
// start update command
|
|
||||||
|
|
||||||
// Slice with functions to override default command behavior.
|
|
||||||
// Functions can be added from the `init()` function in manually curated files in this directory.
|
|
||||||
var updateOverrides []func(
|
|
||||||
*cobra.Command,
|
|
||||||
*settings.UpdateDisableLegacyDbfsRequest,
|
|
||||||
)
|
|
||||||
|
|
||||||
func newUpdate() *cobra.Command {
|
|
||||||
cmd := &cobra.Command{}
|
|
||||||
|
|
||||||
var updateReq settings.UpdateDisableLegacyDbfsRequest
|
|
||||||
var updateJson flags.JsonFlag
|
|
||||||
|
|
||||||
// TODO: short flags
|
|
||||||
cmd.Flags().Var(&updateJson, "json", `either inline JSON string or @path/to/file.json with request body`)
|
|
||||||
|
|
||||||
cmd.Use = "update"
|
|
||||||
cmd.Short = `Update the disable legacy DBFS setting.`
|
|
||||||
cmd.Long = `Update the disable legacy DBFS setting.
|
|
||||||
|
|
||||||
Updates the disable legacy DBFS setting for the workspace.`
|
|
||||||
|
|
||||||
cmd.Annotations = make(map[string]string)
|
|
||||||
|
|
||||||
cmd.PreRunE = root.MustWorkspaceClient
|
|
||||||
cmd.RunE = func(cmd *cobra.Command, args []string) (err error) {
|
|
||||||
ctx := cmd.Context()
|
|
||||||
w := root.WorkspaceClient(ctx)
|
|
||||||
|
|
||||||
if cmd.Flags().Changed("json") {
|
|
||||||
diags := updateJson.Unmarshal(&updateReq)
|
|
||||||
if diags.HasError() {
|
|
||||||
return diags.Error()
|
|
||||||
}
|
|
||||||
if len(diags) > 0 {
|
|
||||||
err := cmdio.RenderDiagnosticsToErrorOut(ctx, diags)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
return fmt.Errorf("please provide command input in JSON format by specifying the --json flag")
|
|
||||||
}
|
|
||||||
|
|
||||||
response, err := w.Settings.DisableLegacyDbfs().Update(ctx, updateReq)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
return cmdio.Render(ctx, response)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Disable completions since they are not applicable.
|
|
||||||
// Can be overridden by manual implementation in `override.go`.
|
|
||||||
cmd.ValidArgsFunction = cobra.NoFileCompletions
|
|
||||||
|
|
||||||
// Apply optional overrides to this command.
|
|
||||||
for _, fn := range updateOverrides {
|
|
||||||
fn(cmd, &updateReq)
|
|
||||||
}
|
|
||||||
|
|
||||||
return cmd
|
|
||||||
}
|
|
||||||
|
|
||||||
// end service DisableLegacyDbfs
|
|
|
@ -1557,7 +1557,6 @@ func newSubmit() *cobra.Command {
|
||||||
cmd.Flags().Var(&submitJson, "json", `either inline JSON string or @path/to/file.json with request body`)
|
cmd.Flags().Var(&submitJson, "json", `either inline JSON string or @path/to/file.json with request body`)
|
||||||
|
|
||||||
// TODO: array: access_control_list
|
// TODO: array: access_control_list
|
||||||
cmd.Flags().StringVar(&submitReq.BudgetPolicyId, "budget-policy-id", submitReq.BudgetPolicyId, `The user specified id of the budget policy to use for this one-time run.`)
|
|
||||||
// TODO: complex arg: email_notifications
|
// TODO: complex arg: email_notifications
|
||||||
// TODO: array: environments
|
// TODO: array: environments
|
||||||
// TODO: complex arg: git_source
|
// TODO: complex arg: git_source
|
||||||
|
|
|
@ -9,7 +9,6 @@ import (
|
||||||
compliance_security_profile "github.com/databricks/cli/cmd/workspace/compliance-security-profile"
|
compliance_security_profile "github.com/databricks/cli/cmd/workspace/compliance-security-profile"
|
||||||
default_namespace "github.com/databricks/cli/cmd/workspace/default-namespace"
|
default_namespace "github.com/databricks/cli/cmd/workspace/default-namespace"
|
||||||
disable_legacy_access "github.com/databricks/cli/cmd/workspace/disable-legacy-access"
|
disable_legacy_access "github.com/databricks/cli/cmd/workspace/disable-legacy-access"
|
||||||
disable_legacy_dbfs "github.com/databricks/cli/cmd/workspace/disable-legacy-dbfs"
|
|
||||||
enhanced_security_monitoring "github.com/databricks/cli/cmd/workspace/enhanced-security-monitoring"
|
enhanced_security_monitoring "github.com/databricks/cli/cmd/workspace/enhanced-security-monitoring"
|
||||||
restrict_workspace_admins "github.com/databricks/cli/cmd/workspace/restrict-workspace-admins"
|
restrict_workspace_admins "github.com/databricks/cli/cmd/workspace/restrict-workspace-admins"
|
||||||
)
|
)
|
||||||
|
@ -34,7 +33,6 @@ func New() *cobra.Command {
|
||||||
cmd.AddCommand(compliance_security_profile.New())
|
cmd.AddCommand(compliance_security_profile.New())
|
||||||
cmd.AddCommand(default_namespace.New())
|
cmd.AddCommand(default_namespace.New())
|
||||||
cmd.AddCommand(disable_legacy_access.New())
|
cmd.AddCommand(disable_legacy_access.New())
|
||||||
cmd.AddCommand(disable_legacy_dbfs.New())
|
|
||||||
cmd.AddCommand(enhanced_security_monitoring.New())
|
cmd.AddCommand(enhanced_security_monitoring.New())
|
||||||
cmd.AddCommand(restrict_workspace_admins.New())
|
cmd.AddCommand(restrict_workspace_admins.New())
|
||||||
|
|
||||||
|
|
2
go.mod
2
go.mod
|
@ -7,7 +7,7 @@ toolchain go1.22.7
|
||||||
require (
|
require (
|
||||||
github.com/Masterminds/semver/v3 v3.3.0 // MIT
|
github.com/Masterminds/semver/v3 v3.3.0 // MIT
|
||||||
github.com/briandowns/spinner v1.23.1 // Apache 2.0
|
github.com/briandowns/spinner v1.23.1 // Apache 2.0
|
||||||
github.com/databricks/databricks-sdk-go v0.49.0 // Apache 2.0
|
github.com/databricks/databricks-sdk-go v0.48.0 // Apache 2.0
|
||||||
github.com/fatih/color v1.17.0 // MIT
|
github.com/fatih/color v1.17.0 // MIT
|
||||||
github.com/ghodss/yaml v1.0.0 // MIT + NOTICE
|
github.com/ghodss/yaml v1.0.0 // MIT + NOTICE
|
||||||
github.com/google/uuid v1.6.0 // BSD-3-Clause
|
github.com/google/uuid v1.6.0 // BSD-3-Clause
|
||||||
|
|
|
@ -32,8 +32,8 @@ github.com/cncf/udpa/go v0.0.0-20191209042840-269d4d468f6f/go.mod h1:M8M6+tZqaGX
|
||||||
github.com/cpuguy83/go-md2man/v2 v2.0.4/go.mod h1:tgQtvFlXSQOSOSIRvRPT7W67SCa46tRHOmNcaadrF8o=
|
github.com/cpuguy83/go-md2man/v2 v2.0.4/go.mod h1:tgQtvFlXSQOSOSIRvRPT7W67SCa46tRHOmNcaadrF8o=
|
||||||
github.com/cyphar/filepath-securejoin v0.2.4 h1:Ugdm7cg7i6ZK6x3xDF1oEu1nfkyfH53EtKeQYTC3kyg=
|
github.com/cyphar/filepath-securejoin v0.2.4 h1:Ugdm7cg7i6ZK6x3xDF1oEu1nfkyfH53EtKeQYTC3kyg=
|
||||||
github.com/cyphar/filepath-securejoin v0.2.4/go.mod h1:aPGpWjXOXUn2NCNjFvBE6aRxGGx79pTxQpKOJNYHHl4=
|
github.com/cyphar/filepath-securejoin v0.2.4/go.mod h1:aPGpWjXOXUn2NCNjFvBE6aRxGGx79pTxQpKOJNYHHl4=
|
||||||
github.com/databricks/databricks-sdk-go v0.49.0 h1:VBTeZZMLIuBSM4kxOCfUcW9z4FUQZY2QeNRD5qm9FUQ=
|
github.com/databricks/databricks-sdk-go v0.48.0 h1:46KtsnRo+FGhC3izUXbpL0PXBNomvsdignYDhJZlm9s=
|
||||||
github.com/databricks/databricks-sdk-go v0.49.0/go.mod h1:ds+zbv5mlQG7nFEU5ojLtgN/u0/9YzZmKQES/CfedzU=
|
github.com/databricks/databricks-sdk-go v0.48.0/go.mod h1:ds+zbv5mlQG7nFEU5ojLtgN/u0/9YzZmKQES/CfedzU=
|
||||||
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
github.com/davecgh/go-spew v1.1.0/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
github.com/davecgh/go-spew v1.1.1 h1:vj9j/u1bqnvCEfJOwUhtlOARqs3+rkHYY13jYWTU97c=
|
||||||
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
github.com/davecgh/go-spew v1.1.1/go.mod h1:J7Y8YcW2NihsgmVo/mv3lAwl/skON4iLHjSsI+c5H38=
|
||||||
|
|
Loading…
Reference in New Issue