From 52ca599cd5c87a4131d96863a90cef69da51e095 Mon Sep 17 00:00:00 2001 From: Pieter Noordhuis Date: Tue, 23 Jul 2024 18:15:02 +0200 Subject: [PATCH] Upgrade TF provider to 1.49.0 (#1617) ## Changes This includes a fix for model serving endpoints. See https://github.com/databricks/terraform-provider-databricks/pull/3690. ## Tests n/a --- bundle/internal/tf/codegen/schema/version.go | 2 +- .../internal/tf/schema/data_source_cluster.go | 232 +++++++++++++++++- .../internal/tf/schema/data_source_schema.go | 36 +++ .../internal/tf/schema/data_source_volume.go | 38 +++ bundle/internal/tf/schema/data_sources.go | 4 + .../internal/tf/schema/resource_dashboard.go | 21 ++ .../tf/schema/resource_permissions.go | 1 + .../tf/schema/resource_workspace_binding.go | 12 + bundle/internal/tf/schema/resources.go | 4 + bundle/internal/tf/schema/root.go | 2 +- 10 files changed, 339 insertions(+), 13 deletions(-) create mode 100644 bundle/internal/tf/schema/data_source_schema.go create mode 100644 bundle/internal/tf/schema/data_source_volume.go create mode 100644 bundle/internal/tf/schema/resource_dashboard.go create mode 100644 bundle/internal/tf/schema/resource_workspace_binding.go diff --git a/bundle/internal/tf/codegen/schema/version.go b/bundle/internal/tf/codegen/schema/version.go index 63a4b1b7..aecb2736 100644 --- a/bundle/internal/tf/codegen/schema/version.go +++ b/bundle/internal/tf/codegen/schema/version.go @@ -1,3 +1,3 @@ package schema -const ProviderVersion = "1.48.3" +const ProviderVersion = "1.49.0" diff --git a/bundle/internal/tf/schema/data_source_cluster.go b/bundle/internal/tf/schema/data_source_cluster.go index fff66dc9..94d67bbf 100644 --- a/bundle/internal/tf/schema/data_source_cluster.go +++ b/bundle/internal/tf/schema/data_source_cluster.go @@ -10,7 +10,9 @@ type DataSourceClusterClusterInfoAutoscale struct { type DataSourceClusterClusterInfoAwsAttributes struct { Availability string `json:"availability,omitempty"` EbsVolumeCount int `json:"ebs_volume_count,omitempty"` + EbsVolumeIops int `json:"ebs_volume_iops,omitempty"` EbsVolumeSize int `json:"ebs_volume_size,omitempty"` + EbsVolumeThroughput int `json:"ebs_volume_throughput,omitempty"` EbsVolumeType string `json:"ebs_volume_type,omitempty"` FirstOnDemand int `json:"first_on_demand,omitempty"` InstanceProfileArn string `json:"instance_profile_arn,omitempty"` @@ -18,10 +20,16 @@ type DataSourceClusterClusterInfoAwsAttributes struct { ZoneId string `json:"zone_id,omitempty"` } +type DataSourceClusterClusterInfoAzureAttributesLogAnalyticsInfo struct { + LogAnalyticsPrimaryKey string `json:"log_analytics_primary_key,omitempty"` + LogAnalyticsWorkspaceId string `json:"log_analytics_workspace_id,omitempty"` +} + type DataSourceClusterClusterInfoAzureAttributes struct { - Availability string `json:"availability,omitempty"` - FirstOnDemand int `json:"first_on_demand,omitempty"` - SpotBidMaxPrice int `json:"spot_bid_max_price,omitempty"` + Availability string `json:"availability,omitempty"` + FirstOnDemand int `json:"first_on_demand,omitempty"` + SpotBidMaxPrice int `json:"spot_bid_max_price,omitempty"` + LogAnalyticsInfo *DataSourceClusterClusterInfoAzureAttributesLogAnalyticsInfo `json:"log_analytics_info,omitempty"` } type DataSourceClusterClusterInfoClusterLogConfDbfs struct { @@ -49,12 +57,12 @@ type DataSourceClusterClusterInfoClusterLogStatus struct { } type DataSourceClusterClusterInfoDockerImageBasicAuth struct { - Password string `json:"password"` - Username string `json:"username"` + Password string `json:"password,omitempty"` + Username string `json:"username,omitempty"` } type DataSourceClusterClusterInfoDockerImage struct { - Url string `json:"url"` + Url string `json:"url,omitempty"` BasicAuth *DataSourceClusterClusterInfoDockerImageBasicAuth `json:"basic_auth,omitempty"` } @@ -139,12 +147,212 @@ type DataSourceClusterClusterInfoInitScripts struct { Workspace *DataSourceClusterClusterInfoInitScriptsWorkspace `json:"workspace,omitempty"` } +type DataSourceClusterClusterInfoSpecAutoscale struct { + MaxWorkers int `json:"max_workers,omitempty"` + MinWorkers int `json:"min_workers,omitempty"` +} + +type DataSourceClusterClusterInfoSpecAwsAttributes struct { + Availability string `json:"availability,omitempty"` + EbsVolumeCount int `json:"ebs_volume_count,omitempty"` + EbsVolumeIops int `json:"ebs_volume_iops,omitempty"` + EbsVolumeSize int `json:"ebs_volume_size,omitempty"` + EbsVolumeThroughput int `json:"ebs_volume_throughput,omitempty"` + EbsVolumeType string `json:"ebs_volume_type,omitempty"` + FirstOnDemand int `json:"first_on_demand,omitempty"` + InstanceProfileArn string `json:"instance_profile_arn,omitempty"` + SpotBidPricePercent int `json:"spot_bid_price_percent,omitempty"` + ZoneId string `json:"zone_id,omitempty"` +} + +type DataSourceClusterClusterInfoSpecAzureAttributesLogAnalyticsInfo struct { + LogAnalyticsPrimaryKey string `json:"log_analytics_primary_key,omitempty"` + LogAnalyticsWorkspaceId string `json:"log_analytics_workspace_id,omitempty"` +} + +type DataSourceClusterClusterInfoSpecAzureAttributes struct { + Availability string `json:"availability,omitempty"` + FirstOnDemand int `json:"first_on_demand,omitempty"` + SpotBidMaxPrice int `json:"spot_bid_max_price,omitempty"` + LogAnalyticsInfo *DataSourceClusterClusterInfoSpecAzureAttributesLogAnalyticsInfo `json:"log_analytics_info,omitempty"` +} + +type DataSourceClusterClusterInfoSpecClusterLogConfDbfs struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecClusterLogConfS3 struct { + CannedAcl string `json:"canned_acl,omitempty"` + Destination string `json:"destination"` + EnableEncryption bool `json:"enable_encryption,omitempty"` + EncryptionType string `json:"encryption_type,omitempty"` + Endpoint string `json:"endpoint,omitempty"` + KmsKey string `json:"kms_key,omitempty"` + Region string `json:"region,omitempty"` +} + +type DataSourceClusterClusterInfoSpecClusterLogConf struct { + Dbfs *DataSourceClusterClusterInfoSpecClusterLogConfDbfs `json:"dbfs,omitempty"` + S3 *DataSourceClusterClusterInfoSpecClusterLogConfS3 `json:"s3,omitempty"` +} + +type DataSourceClusterClusterInfoSpecClusterMountInfoNetworkFilesystemInfo struct { + MountOptions string `json:"mount_options,omitempty"` + ServerAddress string `json:"server_address"` +} + +type DataSourceClusterClusterInfoSpecClusterMountInfo struct { + LocalMountDirPath string `json:"local_mount_dir_path"` + RemoteMountDirPath string `json:"remote_mount_dir_path,omitempty"` + NetworkFilesystemInfo *DataSourceClusterClusterInfoSpecClusterMountInfoNetworkFilesystemInfo `json:"network_filesystem_info,omitempty"` +} + +type DataSourceClusterClusterInfoSpecDockerImageBasicAuth struct { + Password string `json:"password"` + Username string `json:"username"` +} + +type DataSourceClusterClusterInfoSpecDockerImage struct { + Url string `json:"url"` + BasicAuth *DataSourceClusterClusterInfoSpecDockerImageBasicAuth `json:"basic_auth,omitempty"` +} + +type DataSourceClusterClusterInfoSpecGcpAttributes struct { + Availability string `json:"availability,omitempty"` + BootDiskSize int `json:"boot_disk_size,omitempty"` + GoogleServiceAccount string `json:"google_service_account,omitempty"` + LocalSsdCount int `json:"local_ssd_count,omitempty"` + UsePreemptibleExecutors bool `json:"use_preemptible_executors,omitempty"` + ZoneId string `json:"zone_id,omitempty"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsAbfss struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsDbfs struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsFile struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsGcs struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsS3 struct { + CannedAcl string `json:"canned_acl,omitempty"` + Destination string `json:"destination"` + EnableEncryption bool `json:"enable_encryption,omitempty"` + EncryptionType string `json:"encryption_type,omitempty"` + Endpoint string `json:"endpoint,omitempty"` + KmsKey string `json:"kms_key,omitempty"` + Region string `json:"region,omitempty"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsVolumes struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScriptsWorkspace struct { + Destination string `json:"destination"` +} + +type DataSourceClusterClusterInfoSpecInitScripts struct { + Abfss *DataSourceClusterClusterInfoSpecInitScriptsAbfss `json:"abfss,omitempty"` + Dbfs *DataSourceClusterClusterInfoSpecInitScriptsDbfs `json:"dbfs,omitempty"` + File *DataSourceClusterClusterInfoSpecInitScriptsFile `json:"file,omitempty"` + Gcs *DataSourceClusterClusterInfoSpecInitScriptsGcs `json:"gcs,omitempty"` + S3 *DataSourceClusterClusterInfoSpecInitScriptsS3 `json:"s3,omitempty"` + Volumes *DataSourceClusterClusterInfoSpecInitScriptsVolumes `json:"volumes,omitempty"` + Workspace *DataSourceClusterClusterInfoSpecInitScriptsWorkspace `json:"workspace,omitempty"` +} + +type DataSourceClusterClusterInfoSpecLibraryCran struct { + Package string `json:"package"` + Repo string `json:"repo,omitempty"` +} + +type DataSourceClusterClusterInfoSpecLibraryMaven struct { + Coordinates string `json:"coordinates"` + Exclusions []string `json:"exclusions,omitempty"` + Repo string `json:"repo,omitempty"` +} + +type DataSourceClusterClusterInfoSpecLibraryPypi struct { + Package string `json:"package"` + Repo string `json:"repo,omitempty"` +} + +type DataSourceClusterClusterInfoSpecLibrary struct { + Egg string `json:"egg,omitempty"` + Jar string `json:"jar,omitempty"` + Requirements string `json:"requirements,omitempty"` + Whl string `json:"whl,omitempty"` + Cran *DataSourceClusterClusterInfoSpecLibraryCran `json:"cran,omitempty"` + Maven *DataSourceClusterClusterInfoSpecLibraryMaven `json:"maven,omitempty"` + Pypi *DataSourceClusterClusterInfoSpecLibraryPypi `json:"pypi,omitempty"` +} + +type DataSourceClusterClusterInfoSpecWorkloadTypeClients struct { + Jobs bool `json:"jobs,omitempty"` + Notebooks bool `json:"notebooks,omitempty"` +} + +type DataSourceClusterClusterInfoSpecWorkloadType struct { + Clients *DataSourceClusterClusterInfoSpecWorkloadTypeClients `json:"clients,omitempty"` +} + +type DataSourceClusterClusterInfoSpec struct { + ApplyPolicyDefaultValues bool `json:"apply_policy_default_values,omitempty"` + ClusterId string `json:"cluster_id,omitempty"` + ClusterName string `json:"cluster_name,omitempty"` + CustomTags map[string]string `json:"custom_tags,omitempty"` + DataSecurityMode string `json:"data_security_mode,omitempty"` + DriverInstancePoolId string `json:"driver_instance_pool_id,omitempty"` + DriverNodeTypeId string `json:"driver_node_type_id,omitempty"` + EnableElasticDisk bool `json:"enable_elastic_disk,omitempty"` + EnableLocalDiskEncryption bool `json:"enable_local_disk_encryption,omitempty"` + IdempotencyToken string `json:"idempotency_token,omitempty"` + InstancePoolId string `json:"instance_pool_id,omitempty"` + NodeTypeId string `json:"node_type_id,omitempty"` + NumWorkers int `json:"num_workers,omitempty"` + PolicyId string `json:"policy_id,omitempty"` + RuntimeEngine string `json:"runtime_engine,omitempty"` + SingleUserName string `json:"single_user_name,omitempty"` + SparkConf map[string]string `json:"spark_conf,omitempty"` + SparkEnvVars map[string]string `json:"spark_env_vars,omitempty"` + SparkVersion string `json:"spark_version"` + SshPublicKeys []string `json:"ssh_public_keys,omitempty"` + Autoscale *DataSourceClusterClusterInfoSpecAutoscale `json:"autoscale,omitempty"` + AwsAttributes *DataSourceClusterClusterInfoSpecAwsAttributes `json:"aws_attributes,omitempty"` + AzureAttributes *DataSourceClusterClusterInfoSpecAzureAttributes `json:"azure_attributes,omitempty"` + ClusterLogConf *DataSourceClusterClusterInfoSpecClusterLogConf `json:"cluster_log_conf,omitempty"` + ClusterMountInfo []DataSourceClusterClusterInfoSpecClusterMountInfo `json:"cluster_mount_info,omitempty"` + DockerImage *DataSourceClusterClusterInfoSpecDockerImage `json:"docker_image,omitempty"` + GcpAttributes *DataSourceClusterClusterInfoSpecGcpAttributes `json:"gcp_attributes,omitempty"` + InitScripts []DataSourceClusterClusterInfoSpecInitScripts `json:"init_scripts,omitempty"` + Library []DataSourceClusterClusterInfoSpecLibrary `json:"library,omitempty"` + WorkloadType *DataSourceClusterClusterInfoSpecWorkloadType `json:"workload_type,omitempty"` +} + type DataSourceClusterClusterInfoTerminationReason struct { Code string `json:"code,omitempty"` Parameters map[string]string `json:"parameters,omitempty"` Type string `json:"type,omitempty"` } +type DataSourceClusterClusterInfoWorkloadTypeClients struct { + Jobs bool `json:"jobs,omitempty"` + Notebooks bool `json:"notebooks,omitempty"` +} + +type DataSourceClusterClusterInfoWorkloadType struct { + Clients *DataSourceClusterClusterInfoWorkloadTypeClients `json:"clients,omitempty"` +} + type DataSourceClusterClusterInfo struct { AutoterminationMinutes int `json:"autotermination_minutes,omitempty"` ClusterCores int `json:"cluster_cores,omitempty"` @@ -155,14 +363,14 @@ type DataSourceClusterClusterInfo struct { CreatorUserName string `json:"creator_user_name,omitempty"` CustomTags map[string]string `json:"custom_tags,omitempty"` DataSecurityMode string `json:"data_security_mode,omitempty"` - DefaultTags map[string]string `json:"default_tags"` + DefaultTags map[string]string `json:"default_tags,omitempty"` DriverInstancePoolId string `json:"driver_instance_pool_id,omitempty"` DriverNodeTypeId string `json:"driver_node_type_id,omitempty"` EnableElasticDisk bool `json:"enable_elastic_disk,omitempty"` EnableLocalDiskEncryption bool `json:"enable_local_disk_encryption,omitempty"` InstancePoolId string `json:"instance_pool_id,omitempty"` JdbcPort int `json:"jdbc_port,omitempty"` - LastActivityTime int `json:"last_activity_time,omitempty"` + LastRestartedTime int `json:"last_restarted_time,omitempty"` LastStateLossTime int `json:"last_state_loss_time,omitempty"` NodeTypeId string `json:"node_type_id,omitempty"` NumWorkers int `json:"num_workers,omitempty"` @@ -172,12 +380,12 @@ type DataSourceClusterClusterInfo struct { SparkConf map[string]string `json:"spark_conf,omitempty"` SparkContextId int `json:"spark_context_id,omitempty"` SparkEnvVars map[string]string `json:"spark_env_vars,omitempty"` - SparkVersion string `json:"spark_version"` + SparkVersion string `json:"spark_version,omitempty"` SshPublicKeys []string `json:"ssh_public_keys,omitempty"` StartTime int `json:"start_time,omitempty"` - State string `json:"state"` + State string `json:"state,omitempty"` StateMessage string `json:"state_message,omitempty"` - TerminateTime int `json:"terminate_time,omitempty"` + TerminatedTime int `json:"terminated_time,omitempty"` Autoscale *DataSourceClusterClusterInfoAutoscale `json:"autoscale,omitempty"` AwsAttributes *DataSourceClusterClusterInfoAwsAttributes `json:"aws_attributes,omitempty"` AzureAttributes *DataSourceClusterClusterInfoAzureAttributes `json:"azure_attributes,omitempty"` @@ -188,7 +396,9 @@ type DataSourceClusterClusterInfo struct { Executors []DataSourceClusterClusterInfoExecutors `json:"executors,omitempty"` GcpAttributes *DataSourceClusterClusterInfoGcpAttributes `json:"gcp_attributes,omitempty"` InitScripts []DataSourceClusterClusterInfoInitScripts `json:"init_scripts,omitempty"` + Spec *DataSourceClusterClusterInfoSpec `json:"spec,omitempty"` TerminationReason *DataSourceClusterClusterInfoTerminationReason `json:"termination_reason,omitempty"` + WorkloadType *DataSourceClusterClusterInfoWorkloadType `json:"workload_type,omitempty"` } type DataSourceCluster struct { diff --git a/bundle/internal/tf/schema/data_source_schema.go b/bundle/internal/tf/schema/data_source_schema.go new file mode 100644 index 00000000..9d778cc8 --- /dev/null +++ b/bundle/internal/tf/schema/data_source_schema.go @@ -0,0 +1,36 @@ +// Generated from Databricks Terraform provider schema. DO NOT EDIT. + +package schema + +type DataSourceSchemaSchemaInfoEffectivePredictiveOptimizationFlag struct { + InheritedFromName string `json:"inherited_from_name,omitempty"` + InheritedFromType string `json:"inherited_from_type,omitempty"` + Value string `json:"value"` +} + +type DataSourceSchemaSchemaInfo struct { + BrowseOnly bool `json:"browse_only,omitempty"` + CatalogName string `json:"catalog_name,omitempty"` + CatalogType string `json:"catalog_type,omitempty"` + Comment string `json:"comment,omitempty"` + CreatedAt int `json:"created_at,omitempty"` + CreatedBy string `json:"created_by,omitempty"` + EnablePredictiveOptimization string `json:"enable_predictive_optimization,omitempty"` + FullName string `json:"full_name,omitempty"` + MetastoreId string `json:"metastore_id,omitempty"` + Name string `json:"name,omitempty"` + Owner string `json:"owner,omitempty"` + Properties map[string]string `json:"properties,omitempty"` + SchemaId string `json:"schema_id,omitempty"` + StorageLocation string `json:"storage_location,omitempty"` + StorageRoot string `json:"storage_root,omitempty"` + UpdatedAt int `json:"updated_at,omitempty"` + UpdatedBy string `json:"updated_by,omitempty"` + EffectivePredictiveOptimizationFlag *DataSourceSchemaSchemaInfoEffectivePredictiveOptimizationFlag `json:"effective_predictive_optimization_flag,omitempty"` +} + +type DataSourceSchema struct { + Id string `json:"id,omitempty"` + Name string `json:"name"` + SchemaInfo *DataSourceSchemaSchemaInfo `json:"schema_info,omitempty"` +} diff --git a/bundle/internal/tf/schema/data_source_volume.go b/bundle/internal/tf/schema/data_source_volume.go new file mode 100644 index 00000000..67e6100f --- /dev/null +++ b/bundle/internal/tf/schema/data_source_volume.go @@ -0,0 +1,38 @@ +// Generated from Databricks Terraform provider schema. DO NOT EDIT. + +package schema + +type DataSourceVolumeVolumeInfoEncryptionDetailsSseEncryptionDetails struct { + Algorithm string `json:"algorithm,omitempty"` + AwsKmsKeyArn string `json:"aws_kms_key_arn,omitempty"` +} + +type DataSourceVolumeVolumeInfoEncryptionDetails struct { + SseEncryptionDetails *DataSourceVolumeVolumeInfoEncryptionDetailsSseEncryptionDetails `json:"sse_encryption_details,omitempty"` +} + +type DataSourceVolumeVolumeInfo struct { + AccessPoint string `json:"access_point,omitempty"` + BrowseOnly bool `json:"browse_only,omitempty"` + CatalogName string `json:"catalog_name,omitempty"` + Comment string `json:"comment,omitempty"` + CreatedAt int `json:"created_at,omitempty"` + CreatedBy string `json:"created_by,omitempty"` + FullName string `json:"full_name,omitempty"` + MetastoreId string `json:"metastore_id,omitempty"` + Name string `json:"name,omitempty"` + Owner string `json:"owner,omitempty"` + SchemaName string `json:"schema_name,omitempty"` + StorageLocation string `json:"storage_location,omitempty"` + UpdatedAt int `json:"updated_at,omitempty"` + UpdatedBy string `json:"updated_by,omitempty"` + VolumeId string `json:"volume_id,omitempty"` + VolumeType string `json:"volume_type,omitempty"` + EncryptionDetails *DataSourceVolumeVolumeInfoEncryptionDetails `json:"encryption_details,omitempty"` +} + +type DataSourceVolume struct { + Id string `json:"id,omitempty"` + Name string `json:"name"` + VolumeInfo *DataSourceVolumeVolumeInfo `json:"volume_info,omitempty"` +} diff --git a/bundle/internal/tf/schema/data_sources.go b/bundle/internal/tf/schema/data_sources.go index b68df2b4..4ac78613 100644 --- a/bundle/internal/tf/schema/data_sources.go +++ b/bundle/internal/tf/schema/data_sources.go @@ -36,6 +36,7 @@ type DataSources struct { Notebook map[string]any `json:"databricks_notebook,omitempty"` NotebookPaths map[string]any `json:"databricks_notebook_paths,omitempty"` Pipelines map[string]any `json:"databricks_pipelines,omitempty"` + Schema map[string]any `json:"databricks_schema,omitempty"` Schemas map[string]any `json:"databricks_schemas,omitempty"` ServicePrincipal map[string]any `json:"databricks_service_principal,omitempty"` ServicePrincipals map[string]any `json:"databricks_service_principals,omitempty"` @@ -50,6 +51,7 @@ type DataSources struct { Tables map[string]any `json:"databricks_tables,omitempty"` User map[string]any `json:"databricks_user,omitempty"` Views map[string]any `json:"databricks_views,omitempty"` + Volume map[string]any `json:"databricks_volume,omitempty"` Volumes map[string]any `json:"databricks_volumes,omitempty"` Zones map[string]any `json:"databricks_zones,omitempty"` } @@ -89,6 +91,7 @@ func NewDataSources() *DataSources { Notebook: make(map[string]any), NotebookPaths: make(map[string]any), Pipelines: make(map[string]any), + Schema: make(map[string]any), Schemas: make(map[string]any), ServicePrincipal: make(map[string]any), ServicePrincipals: make(map[string]any), @@ -103,6 +106,7 @@ func NewDataSources() *DataSources { Tables: make(map[string]any), User: make(map[string]any), Views: make(map[string]any), + Volume: make(map[string]any), Volumes: make(map[string]any), Zones: make(map[string]any), } diff --git a/bundle/internal/tf/schema/resource_dashboard.go b/bundle/internal/tf/schema/resource_dashboard.go new file mode 100644 index 00000000..0c2fa4a0 --- /dev/null +++ b/bundle/internal/tf/schema/resource_dashboard.go @@ -0,0 +1,21 @@ +// Generated from Databricks Terraform provider schema. DO NOT EDIT. + +package schema + +type ResourceDashboard struct { + CreateTime string `json:"create_time,omitempty"` + DashboardChangeDetected bool `json:"dashboard_change_detected,omitempty"` + DashboardId string `json:"dashboard_id,omitempty"` + DisplayName string `json:"display_name"` + EmbedCredentials bool `json:"embed_credentials,omitempty"` + Etag string `json:"etag,omitempty"` + FilePath string `json:"file_path,omitempty"` + Id string `json:"id,omitempty"` + LifecycleState string `json:"lifecycle_state,omitempty"` + Md5 string `json:"md5,omitempty"` + ParentPath string `json:"parent_path"` + Path string `json:"path,omitempty"` + SerializedDashboard string `json:"serialized_dashboard,omitempty"` + UpdateTime string `json:"update_time,omitempty"` + WarehouseId string `json:"warehouse_id"` +} diff --git a/bundle/internal/tf/schema/resource_permissions.go b/bundle/internal/tf/schema/resource_permissions.go index 5d8df11e..ee94a1a8 100644 --- a/bundle/internal/tf/schema/resource_permissions.go +++ b/bundle/internal/tf/schema/resource_permissions.go @@ -13,6 +13,7 @@ type ResourcePermissions struct { Authorization string `json:"authorization,omitempty"` ClusterId string `json:"cluster_id,omitempty"` ClusterPolicyId string `json:"cluster_policy_id,omitempty"` + DashboardId string `json:"dashboard_id,omitempty"` DirectoryId string `json:"directory_id,omitempty"` DirectoryPath string `json:"directory_path,omitempty"` ExperimentId string `json:"experiment_id,omitempty"` diff --git a/bundle/internal/tf/schema/resource_workspace_binding.go b/bundle/internal/tf/schema/resource_workspace_binding.go new file mode 100644 index 00000000..f0be7a41 --- /dev/null +++ b/bundle/internal/tf/schema/resource_workspace_binding.go @@ -0,0 +1,12 @@ +// Generated from Databricks Terraform provider schema. DO NOT EDIT. + +package schema + +type ResourceWorkspaceBinding struct { + BindingType string `json:"binding_type,omitempty"` + CatalogName string `json:"catalog_name,omitempty"` + Id string `json:"id,omitempty"` + SecurableName string `json:"securable_name,omitempty"` + SecurableType string `json:"securable_type,omitempty"` + WorkspaceId int `json:"workspace_id,omitempty"` +} diff --git a/bundle/internal/tf/schema/resources.go b/bundle/internal/tf/schema/resources.go index 79d71a65..79c1b32b 100644 --- a/bundle/internal/tf/schema/resources.go +++ b/bundle/internal/tf/schema/resources.go @@ -16,6 +16,7 @@ type Resources struct { ClusterPolicy map[string]any `json:"databricks_cluster_policy,omitempty"` ComplianceSecurityProfileWorkspaceSetting map[string]any `json:"databricks_compliance_security_profile_workspace_setting,omitempty"` Connection map[string]any `json:"databricks_connection,omitempty"` + Dashboard map[string]any `json:"databricks_dashboard,omitempty"` DbfsFile map[string]any `json:"databricks_dbfs_file,omitempty"` DefaultNamespaceSetting map[string]any `json:"databricks_default_namespace_setting,omitempty"` Directory map[string]any `json:"databricks_directory,omitempty"` @@ -96,6 +97,7 @@ type Resources struct { VectorSearchEndpoint map[string]any `json:"databricks_vector_search_endpoint,omitempty"` VectorSearchIndex map[string]any `json:"databricks_vector_search_index,omitempty"` Volume map[string]any `json:"databricks_volume,omitempty"` + WorkspaceBinding map[string]any `json:"databricks_workspace_binding,omitempty"` WorkspaceConf map[string]any `json:"databricks_workspace_conf,omitempty"` WorkspaceFile map[string]any `json:"databricks_workspace_file,omitempty"` } @@ -115,6 +117,7 @@ func NewResources() *Resources { ClusterPolicy: make(map[string]any), ComplianceSecurityProfileWorkspaceSetting: make(map[string]any), Connection: make(map[string]any), + Dashboard: make(map[string]any), DbfsFile: make(map[string]any), DefaultNamespaceSetting: make(map[string]any), Directory: make(map[string]any), @@ -195,6 +198,7 @@ func NewResources() *Resources { VectorSearchEndpoint: make(map[string]any), VectorSearchIndex: make(map[string]any), Volume: make(map[string]any), + WorkspaceBinding: make(map[string]any), WorkspaceConf: make(map[string]any), WorkspaceFile: make(map[string]any), } diff --git a/bundle/internal/tf/schema/root.go b/bundle/internal/tf/schema/root.go index a79e998c..8401d8da 100644 --- a/bundle/internal/tf/schema/root.go +++ b/bundle/internal/tf/schema/root.go @@ -21,7 +21,7 @@ type Root struct { const ProviderHost = "registry.terraform.io" const ProviderSource = "databricks/databricks" -const ProviderVersion = "1.48.3" +const ProviderVersion = "1.49.0" func NewRoot() *Root { return &Root{