@pulumi/databricks 1.68.0-alpha.1745473147 → 1.68.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/types/input.d.ts CHANGED
@@ -1626,7 +1626,7 @@ export interface GetClusterClusterInfoSpec {
1626
1626
  /**
1627
1627
  * [Runtime version](https://docs.databricks.com/runtime/index.html) of the cluster.
1628
1628
  */
1629
- sparkVersion: string;
1629
+ sparkVersion?: string;
1630
1630
  /**
1631
1631
  * SSH public key contents that will be added to each Spark node in this cluster.
1632
1632
  */
@@ -1721,7 +1721,7 @@ export interface GetClusterClusterInfoSpecArgs {
1721
1721
  /**
1722
1722
  * [Runtime version](https://docs.databricks.com/runtime/index.html) of the cluster.
1723
1723
  */
1724
- sparkVersion: pulumi.Input<string>;
1724
+ sparkVersion?: pulumi.Input<string>;
1725
1725
  /**
1726
1726
  * SSH public key contents that will be added to each Spark node in this cluster.
1727
1727
  */
@@ -3287,7 +3287,7 @@ export interface GetJobJobSettingsSettingsJobClusterNewCluster {
3287
3287
  sparkEnvVars?: {
3288
3288
  [key: string]: string;
3289
3289
  };
3290
- sparkVersion: string;
3290
+ sparkVersion?: string;
3291
3291
  sshPublicKeys?: string[];
3292
3292
  workloadType?: inputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadType;
3293
3293
  }
@@ -3325,7 +3325,7 @@ export interface GetJobJobSettingsSettingsJobClusterNewClusterArgs {
3325
3325
  sparkEnvVars?: pulumi.Input<{
3326
3326
  [key: string]: pulumi.Input<string>;
3327
3327
  }>;
3328
- sparkVersion: pulumi.Input<string>;
3328
+ sparkVersion?: pulumi.Input<string>;
3329
3329
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
3330
3330
  workloadType?: pulumi.Input<inputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadTypeArgs>;
3331
3331
  }
@@ -3613,7 +3613,7 @@ export interface GetJobJobSettingsSettingsNewCluster {
3613
3613
  sparkEnvVars?: {
3614
3614
  [key: string]: string;
3615
3615
  };
3616
- sparkVersion: string;
3616
+ sparkVersion?: string;
3617
3617
  sshPublicKeys?: string[];
3618
3618
  workloadType?: inputs.GetJobJobSettingsSettingsNewClusterWorkloadType;
3619
3619
  }
@@ -3651,7 +3651,7 @@ export interface GetJobJobSettingsSettingsNewClusterArgs {
3651
3651
  sparkEnvVars?: pulumi.Input<{
3652
3652
  [key: string]: pulumi.Input<string>;
3653
3653
  }>;
3654
- sparkVersion: pulumi.Input<string>;
3654
+ sparkVersion?: pulumi.Input<string>;
3655
3655
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
3656
3656
  workloadType?: pulumi.Input<inputs.GetJobJobSettingsSettingsNewClusterWorkloadTypeArgs>;
3657
3657
  }
@@ -3987,6 +3987,7 @@ export interface GetJobJobSettingsSettingsSparkSubmitTaskArgs {
3987
3987
  }
3988
3988
  export interface GetJobJobSettingsSettingsTask {
3989
3989
  conditionTask?: inputs.GetJobJobSettingsSettingsTaskConditionTask;
3990
+ dashboardTask?: inputs.GetJobJobSettingsSettingsTaskDashboardTask;
3990
3991
  dbtTask?: inputs.GetJobJobSettingsSettingsTaskDbtTask;
3991
3992
  dependsOns?: inputs.GetJobJobSettingsSettingsTaskDependsOn[];
3992
3993
  description?: string;
@@ -4003,6 +4004,7 @@ export interface GetJobJobSettingsSettingsTask {
4003
4004
  notebookTask?: inputs.GetJobJobSettingsSettingsTaskNotebookTask;
4004
4005
  notificationSettings?: inputs.GetJobJobSettingsSettingsTaskNotificationSettings;
4005
4006
  pipelineTask?: inputs.GetJobJobSettingsSettingsTaskPipelineTask;
4007
+ powerBiTask?: inputs.GetJobJobSettingsSettingsTaskPowerBiTask;
4006
4008
  pythonWheelTask?: inputs.GetJobJobSettingsSettingsTaskPythonWheelTask;
4007
4009
  retryOnTimeout?: boolean;
4008
4010
  runIf?: string;
@@ -4017,6 +4019,7 @@ export interface GetJobJobSettingsSettingsTask {
4017
4019
  }
4018
4020
  export interface GetJobJobSettingsSettingsTaskArgs {
4019
4021
  conditionTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskConditionTaskArgs>;
4022
+ dashboardTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskDashboardTaskArgs>;
4020
4023
  dbtTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskDbtTaskArgs>;
4021
4024
  dependsOns?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskDependsOnArgs>[]>;
4022
4025
  description?: pulumi.Input<string>;
@@ -4033,6 +4036,7 @@ export interface GetJobJobSettingsSettingsTaskArgs {
4033
4036
  notebookTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskNotebookTaskArgs>;
4034
4037
  notificationSettings?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskNotificationSettingsArgs>;
4035
4038
  pipelineTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskPipelineTaskArgs>;
4039
+ powerBiTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskPowerBiTaskArgs>;
4036
4040
  pythonWheelTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskPythonWheelTaskArgs>;
4037
4041
  retryOnTimeout?: pulumi.Input<boolean>;
4038
4042
  runIf?: pulumi.Input<string>;
@@ -4055,6 +4059,34 @@ export interface GetJobJobSettingsSettingsTaskConditionTaskArgs {
4055
4059
  op: pulumi.Input<string>;
4056
4060
  right: pulumi.Input<string>;
4057
4061
  }
4062
+ export interface GetJobJobSettingsSettingsTaskDashboardTask {
4063
+ dashboardId?: string;
4064
+ subscription?: inputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscription;
4065
+ warehouseId?: string;
4066
+ }
4067
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskArgs {
4068
+ dashboardId?: pulumi.Input<string>;
4069
+ subscription?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionArgs>;
4070
+ warehouseId?: pulumi.Input<string>;
4071
+ }
4072
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscription {
4073
+ customSubject?: string;
4074
+ paused?: boolean;
4075
+ subscribers?: inputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber[];
4076
+ }
4077
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionArgs {
4078
+ customSubject?: pulumi.Input<string>;
4079
+ paused?: pulumi.Input<boolean>;
4080
+ subscribers?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriberArgs>[]>;
4081
+ }
4082
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber {
4083
+ destinationId?: string;
4084
+ userName?: string;
4085
+ }
4086
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriberArgs {
4087
+ destinationId?: pulumi.Input<string>;
4088
+ userName?: pulumi.Input<string>;
4089
+ }
4058
4090
  export interface GetJobJobSettingsSettingsTaskDbtTask {
4059
4091
  catalog?: string;
4060
4092
  commands: string[];
@@ -4109,6 +4141,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskArgs {
4109
4141
  }
4110
4142
  export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
4111
4143
  conditionTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTask;
4144
+ dashboardTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask;
4112
4145
  dbtTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask;
4113
4146
  dependsOns?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOn[];
4114
4147
  description?: string;
@@ -4124,6 +4157,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
4124
4157
  notebookTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotebookTask;
4125
4158
  notificationSettings?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotificationSettings;
4126
4159
  pipelineTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTask;
4160
+ powerBiTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask;
4127
4161
  pythonWheelTask?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask;
4128
4162
  retryOnTimeout?: boolean;
4129
4163
  runIf?: string;
@@ -4138,6 +4172,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
4138
4172
  }
4139
4173
  export interface GetJobJobSettingsSettingsTaskForEachTaskTaskArgs {
4140
4174
  conditionTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskArgs>;
4175
+ dashboardTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskArgs>;
4141
4176
  dbtTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTaskArgs>;
4142
4177
  dependsOns?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOnArgs>[]>;
4143
4178
  description?: pulumi.Input<string>;
@@ -4153,6 +4188,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskArgs {
4153
4188
  notebookTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotebookTaskArgs>;
4154
4189
  notificationSettings?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotificationSettingsArgs>;
4155
4190
  pipelineTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTaskArgs>;
4191
+ powerBiTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskArgs>;
4156
4192
  pythonWheelTask?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTaskArgs>;
4157
4193
  retryOnTimeout?: pulumi.Input<boolean>;
4158
4194
  runIf?: pulumi.Input<string>;
@@ -4175,6 +4211,34 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTaskArgs {
4175
4211
  op: pulumi.Input<string>;
4176
4212
  right: pulumi.Input<string>;
4177
4213
  }
4214
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask {
4215
+ dashboardId?: string;
4216
+ subscription?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription;
4217
+ warehouseId?: string;
4218
+ }
4219
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskArgs {
4220
+ dashboardId?: pulumi.Input<string>;
4221
+ subscription?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionArgs>;
4222
+ warehouseId?: pulumi.Input<string>;
4223
+ }
4224
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription {
4225
+ customSubject?: string;
4226
+ paused?: boolean;
4227
+ subscribers?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber[];
4228
+ }
4229
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionArgs {
4230
+ customSubject?: pulumi.Input<string>;
4231
+ paused?: pulumi.Input<boolean>;
4232
+ subscribers?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriberArgs>[]>;
4233
+ }
4234
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
4235
+ destinationId?: string;
4236
+ userName?: string;
4237
+ }
4238
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriberArgs {
4239
+ destinationId?: pulumi.Input<string>;
4240
+ userName?: pulumi.Input<string>;
4241
+ }
4178
4242
  export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask {
4179
4243
  catalog?: string;
4180
4244
  commands: string[];
@@ -4311,7 +4375,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskNewCluster {
4311
4375
  sparkEnvVars?: {
4312
4376
  [key: string]: string;
4313
4377
  };
4314
- sparkVersion: string;
4378
+ sparkVersion?: string;
4315
4379
  sshPublicKeys?: string[];
4316
4380
  workloadType?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNewClusterWorkloadType;
4317
4381
  }
@@ -4349,7 +4413,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskNewClusterArgs {
4349
4413
  sparkEnvVars?: pulumi.Input<{
4350
4414
  [key: string]: pulumi.Input<string>;
4351
4415
  }>;
4352
- sparkVersion: pulumi.Input<string>;
4416
+ sparkVersion?: pulumi.Input<string>;
4353
4417
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
4354
4418
  workloadType?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNewClusterWorkloadTypeArgs>;
4355
4419
  }
@@ -4593,6 +4657,52 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTaskArgs {
4593
4657
  fullRefresh?: pulumi.Input<boolean>;
4594
4658
  pipelineId: pulumi.Input<string>;
4595
4659
  }
4660
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask {
4661
+ connectionResourceName?: string;
4662
+ powerBiModel?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel;
4663
+ refreshAfterUpdate?: boolean;
4664
+ tables?: inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable[];
4665
+ warehouseId?: string;
4666
+ }
4667
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskArgs {
4668
+ connectionResourceName?: pulumi.Input<string>;
4669
+ powerBiModel?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModelArgs>;
4670
+ refreshAfterUpdate?: pulumi.Input<boolean>;
4671
+ tables?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTableArgs>[]>;
4672
+ warehouseId?: pulumi.Input<string>;
4673
+ }
4674
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel {
4675
+ authenticationMethod?: string;
4676
+ modelName?: string;
4677
+ overwriteExisting?: boolean;
4678
+ storageMode?: string;
4679
+ workspaceName?: string;
4680
+ }
4681
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModelArgs {
4682
+ authenticationMethod?: pulumi.Input<string>;
4683
+ modelName?: pulumi.Input<string>;
4684
+ overwriteExisting?: pulumi.Input<boolean>;
4685
+ storageMode?: pulumi.Input<string>;
4686
+ workspaceName?: pulumi.Input<string>;
4687
+ }
4688
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable {
4689
+ catalog?: string;
4690
+ /**
4691
+ * the job name of databricks.Job if the resource was matched by id.
4692
+ */
4693
+ name?: string;
4694
+ schema?: string;
4695
+ storageMode?: string;
4696
+ }
4697
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTableArgs {
4698
+ catalog?: pulumi.Input<string>;
4699
+ /**
4700
+ * the job name of databricks.Job if the resource was matched by id.
4701
+ */
4702
+ name?: pulumi.Input<string>;
4703
+ schema?: pulumi.Input<string>;
4704
+ storageMode?: pulumi.Input<string>;
4705
+ }
4596
4706
  export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask {
4597
4707
  entryPoint?: string;
4598
4708
  namedParameters?: {
@@ -4887,7 +4997,7 @@ export interface GetJobJobSettingsSettingsTaskNewCluster {
4887
4997
  sparkEnvVars?: {
4888
4998
  [key: string]: string;
4889
4999
  };
4890
- sparkVersion: string;
5000
+ sparkVersion?: string;
4891
5001
  sshPublicKeys?: string[];
4892
5002
  workloadType?: inputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadType;
4893
5003
  }
@@ -4925,7 +5035,7 @@ export interface GetJobJobSettingsSettingsTaskNewClusterArgs {
4925
5035
  sparkEnvVars?: pulumi.Input<{
4926
5036
  [key: string]: pulumi.Input<string>;
4927
5037
  }>;
4928
- sparkVersion: pulumi.Input<string>;
5038
+ sparkVersion?: pulumi.Input<string>;
4929
5039
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
4930
5040
  workloadType?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadTypeArgs>;
4931
5041
  }
@@ -5169,6 +5279,52 @@ export interface GetJobJobSettingsSettingsTaskPipelineTaskArgs {
5169
5279
  fullRefresh?: pulumi.Input<boolean>;
5170
5280
  pipelineId: pulumi.Input<string>;
5171
5281
  }
5282
+ export interface GetJobJobSettingsSettingsTaskPowerBiTask {
5283
+ connectionResourceName?: string;
5284
+ powerBiModel?: inputs.GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel;
5285
+ refreshAfterUpdate?: boolean;
5286
+ tables?: inputs.GetJobJobSettingsSettingsTaskPowerBiTaskTable[];
5287
+ warehouseId?: string;
5288
+ }
5289
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskArgs {
5290
+ connectionResourceName?: pulumi.Input<string>;
5291
+ powerBiModel?: pulumi.Input<inputs.GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModelArgs>;
5292
+ refreshAfterUpdate?: pulumi.Input<boolean>;
5293
+ tables?: pulumi.Input<pulumi.Input<inputs.GetJobJobSettingsSettingsTaskPowerBiTaskTableArgs>[]>;
5294
+ warehouseId?: pulumi.Input<string>;
5295
+ }
5296
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel {
5297
+ authenticationMethod?: string;
5298
+ modelName?: string;
5299
+ overwriteExisting?: boolean;
5300
+ storageMode?: string;
5301
+ workspaceName?: string;
5302
+ }
5303
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModelArgs {
5304
+ authenticationMethod?: pulumi.Input<string>;
5305
+ modelName?: pulumi.Input<string>;
5306
+ overwriteExisting?: pulumi.Input<boolean>;
5307
+ storageMode?: pulumi.Input<string>;
5308
+ workspaceName?: pulumi.Input<string>;
5309
+ }
5310
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskTable {
5311
+ catalog?: string;
5312
+ /**
5313
+ * the job name of databricks.Job if the resource was matched by id.
5314
+ */
5315
+ name?: string;
5316
+ schema?: string;
5317
+ storageMode?: string;
5318
+ }
5319
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskTableArgs {
5320
+ catalog?: pulumi.Input<string>;
5321
+ /**
5322
+ * the job name of databricks.Job if the resource was matched by id.
5323
+ */
5324
+ name?: pulumi.Input<string>;
5325
+ schema?: pulumi.Input<string>;
5326
+ storageMode?: pulumi.Input<string>;
5327
+ }
5172
5328
  export interface GetJobJobSettingsSettingsTaskPythonWheelTask {
5173
5329
  entryPoint?: string;
5174
5330
  namedParameters?: {
@@ -7698,8 +7854,6 @@ export interface JobEmailNotifications {
7698
7854
  noAlertForSkippedRuns?: pulumi.Input<boolean>;
7699
7855
  /**
7700
7856
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
7701
- *
7702
- * The following parameter is only available for the job level configuration.
7703
7857
  */
7704
7858
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<string>[]>;
7705
7859
  /**
@@ -7710,6 +7864,11 @@ export interface JobEmailNotifications {
7710
7864
  * (List) list of emails to notify when the run starts.
7711
7865
  */
7712
7866
  onStarts?: pulumi.Input<pulumi.Input<string>[]>;
7867
+ /**
7868
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
7869
+ *
7870
+ * The following parameter is only available for the job level configuration.
7871
+ */
7713
7872
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<string>[]>;
7714
7873
  /**
7715
7874
  * (List) list of emails to notify when the run completes successfully.
@@ -7838,7 +7997,7 @@ export interface JobJobClusterNewCluster {
7838
7997
  sparkEnvVars?: pulumi.Input<{
7839
7998
  [key: string]: pulumi.Input<string>;
7840
7999
  }>;
7841
- sparkVersion: pulumi.Input<string>;
8000
+ sparkVersion?: pulumi.Input<string>;
7842
8001
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
7843
8002
  useMlRuntime?: pulumi.Input<boolean>;
7844
8003
  /**
@@ -8052,7 +8211,7 @@ export interface JobNewCluster {
8052
8211
  sparkEnvVars?: pulumi.Input<{
8053
8212
  [key: string]: pulumi.Input<string>;
8054
8213
  }>;
8055
- sparkVersion: pulumi.Input<string>;
8214
+ sparkVersion?: pulumi.Input<string>;
8056
8215
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
8057
8216
  useMlRuntime?: pulumi.Input<boolean>;
8058
8217
  /**
@@ -8351,11 +8510,13 @@ export interface JobSparkPythonTask {
8351
8510
  */
8352
8511
  parameters?: pulumi.Input<pulumi.Input<string>[]>;
8353
8512
  /**
8354
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
8513
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
8355
8514
  */
8356
8515
  pythonFile: pulumi.Input<string>;
8357
8516
  /**
8358
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
8517
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
8518
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
8519
+ * * `GIT`: The Python file is located in a remote Git repository.
8359
8520
  */
8360
8521
  source?: pulumi.Input<string>;
8361
8522
  }
@@ -8495,24 +8656,39 @@ export interface JobTaskConditionTask {
8495
8656
  }
8496
8657
  export interface JobTaskDashboardTask {
8497
8658
  /**
8498
- * (String) identifier of the Databricks SQL Dashboard databricks_sql_dashboard.
8659
+ * The identifier of the dashboard to refresh
8499
8660
  */
8500
8661
  dashboardId?: pulumi.Input<string>;
8662
+ /**
8663
+ * Represents a subscription configuration for scheduled dashboard snapshots.
8664
+ */
8501
8665
  subscription?: pulumi.Input<inputs.JobTaskDashboardTaskSubscription>;
8666
+ /**
8667
+ * The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
8668
+ */
8502
8669
  warehouseId?: pulumi.Input<string>;
8503
8670
  }
8504
8671
  export interface JobTaskDashboardTaskSubscription {
8505
8672
  /**
8506
- * string specifying a custom subject of email sent.
8673
+ * Allows users to specify a custom subject line on the email sent to subscribers.
8507
8674
  */
8508
8675
  customSubject?: pulumi.Input<string>;
8676
+ /**
8677
+ * When true, the subscription will not send emails.
8678
+ */
8509
8679
  paused?: pulumi.Input<boolean>;
8680
+ /**
8681
+ * The list of subscribers to send the snapshot of the dashboard to.
8682
+ */
8510
8683
  subscribers?: pulumi.Input<pulumi.Input<inputs.JobTaskDashboardTaskSubscriptionSubscriber>[]>;
8511
8684
  }
8512
8685
  export interface JobTaskDashboardTaskSubscriptionSubscriber {
8686
+ /**
8687
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
8688
+ */
8513
8689
  destinationId?: pulumi.Input<string>;
8514
8690
  /**
8515
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
8691
+ * A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
8516
8692
  */
8517
8693
  userName?: pulumi.Input<string>;
8518
8694
  }
@@ -8569,8 +8745,6 @@ export interface JobTaskEmailNotifications {
8569
8745
  noAlertForSkippedRuns?: pulumi.Input<boolean>;
8570
8746
  /**
8571
8747
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
8572
- *
8573
- * The following parameter is only available for the job level configuration.
8574
8748
  */
8575
8749
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<string>[]>;
8576
8750
  /**
@@ -8581,6 +8755,11 @@ export interface JobTaskEmailNotifications {
8581
8755
  * (List) list of emails to notify when the run starts.
8582
8756
  */
8583
8757
  onStarts?: pulumi.Input<pulumi.Input<string>[]>;
8758
+ /**
8759
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
8760
+ *
8761
+ * The following parameter is only available for the job level configuration.
8762
+ */
8584
8763
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<string>[]>;
8585
8764
  /**
8586
8765
  * (List) list of emails to notify when the run completes successfully.
@@ -8730,24 +8909,39 @@ export interface JobTaskForEachTaskTaskConditionTask {
8730
8909
  }
8731
8910
  export interface JobTaskForEachTaskTaskDashboardTask {
8732
8911
  /**
8733
- * (String) identifier of the Databricks SQL Dashboard databricks_sql_dashboard.
8912
+ * The identifier of the dashboard to refresh
8734
8913
  */
8735
8914
  dashboardId?: pulumi.Input<string>;
8915
+ /**
8916
+ * Represents a subscription configuration for scheduled dashboard snapshots.
8917
+ */
8736
8918
  subscription?: pulumi.Input<inputs.JobTaskForEachTaskTaskDashboardTaskSubscription>;
8919
+ /**
8920
+ * The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
8921
+ */
8737
8922
  warehouseId?: pulumi.Input<string>;
8738
8923
  }
8739
8924
  export interface JobTaskForEachTaskTaskDashboardTaskSubscription {
8740
8925
  /**
8741
- * string specifying a custom subject of email sent.
8926
+ * Allows users to specify a custom subject line on the email sent to subscribers.
8742
8927
  */
8743
8928
  customSubject?: pulumi.Input<string>;
8929
+ /**
8930
+ * When true, the subscription will not send emails.
8931
+ */
8744
8932
  paused?: pulumi.Input<boolean>;
8933
+ /**
8934
+ * The list of subscribers to send the snapshot of the dashboard to.
8935
+ */
8745
8936
  subscribers?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber>[]>;
8746
8937
  }
8747
8938
  export interface JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
8939
+ /**
8940
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
8941
+ */
8748
8942
  destinationId?: pulumi.Input<string>;
8749
8943
  /**
8750
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
8944
+ * A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
8751
8945
  */
8752
8946
  userName?: pulumi.Input<string>;
8753
8947
  }
@@ -8804,8 +8998,6 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
8804
8998
  noAlertForSkippedRuns?: pulumi.Input<boolean>;
8805
8999
  /**
8806
9000
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
8807
- *
8808
- * The following parameter is only available for the job level configuration.
8809
9001
  */
8810
9002
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<string>[]>;
8811
9003
  /**
@@ -8816,6 +9008,11 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
8816
9008
  * (List) list of emails to notify when the run starts.
8817
9009
  */
8818
9010
  onStarts?: pulumi.Input<pulumi.Input<string>[]>;
9011
+ /**
9012
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
9013
+ *
9014
+ * The following parameter is only available for the job level configuration.
9015
+ */
8819
9016
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<string>[]>;
8820
9017
  /**
8821
9018
  * (List) list of emails to notify when the run completes successfully.
@@ -8918,7 +9115,7 @@ export interface JobTaskForEachTaskTaskNewCluster {
8918
9115
  sparkEnvVars?: pulumi.Input<{
8919
9116
  [key: string]: pulumi.Input<string>;
8920
9117
  }>;
8921
- sparkVersion: pulumi.Input<string>;
9118
+ sparkVersion?: pulumi.Input<string>;
8922
9119
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
8923
9120
  useMlRuntime?: pulumi.Input<boolean>;
8924
9121
  /**
@@ -9120,32 +9317,65 @@ export interface JobTaskForEachTaskTaskPipelineTask {
9120
9317
  pipelineId: pulumi.Input<string>;
9121
9318
  }
9122
9319
  export interface JobTaskForEachTaskTaskPowerBiTask {
9320
+ /**
9321
+ * The resource name of the UC connection to authenticate from Databricks to Power BI
9322
+ */
9123
9323
  connectionResourceName?: pulumi.Input<string>;
9324
+ /**
9325
+ * The semantic model to update. Block consists of following fields:
9326
+ */
9124
9327
  powerBiModel?: pulumi.Input<inputs.JobTaskForEachTaskTaskPowerBiTaskPowerBiModel>;
9328
+ /**
9329
+ * Whether the model should be refreshed after the update. Default is false
9330
+ */
9125
9331
  refreshAfterUpdate?: pulumi.Input<boolean>;
9332
+ /**
9333
+ * The tables to be exported to Power BI. Block consists of following fields:
9334
+ */
9126
9335
  tables?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskPowerBiTaskTable>[]>;
9336
+ /**
9337
+ * The SQL warehouse ID to use as the Power BI data source
9338
+ */
9127
9339
  warehouseId?: pulumi.Input<string>;
9128
9340
  }
9129
9341
  export interface JobTaskForEachTaskTaskPowerBiTaskPowerBiModel {
9342
+ /**
9343
+ * How the published Power BI model authenticates to Databricks
9344
+ */
9130
9345
  authenticationMethod?: pulumi.Input<string>;
9346
+ /**
9347
+ * The name of the Power BI model
9348
+ */
9131
9349
  modelName?: pulumi.Input<string>;
9350
+ /**
9351
+ * Whether to overwrite existing Power BI models. Default is false
9352
+ */
9132
9353
  overwriteExisting?: pulumi.Input<boolean>;
9354
+ /**
9355
+ * The default storage mode of the Power BI model
9356
+ */
9133
9357
  storageMode?: pulumi.Input<string>;
9358
+ /**
9359
+ * The name of the Power BI workspace of the model
9360
+ */
9134
9361
  workspaceName?: pulumi.Input<string>;
9135
9362
  }
9136
9363
  export interface JobTaskForEachTaskTaskPowerBiTaskTable {
9137
9364
  /**
9138
- * The name of the catalog to use inside Unity Catalog.
9365
+ * The catalog name in Databricks
9139
9366
  */
9140
9367
  catalog?: pulumi.Input<string>;
9141
9368
  /**
9142
- * An optional name for the job. The default value is Untitled.
9369
+ * The table name in Databricks. If empty, all tables under the schema are selected.
9143
9370
  */
9144
9371
  name?: pulumi.Input<string>;
9145
9372
  /**
9146
- * The name of the schema dbt should run in. Defaults to `default`.
9373
+ * The schema name in Databricks
9147
9374
  */
9148
9375
  schema?: pulumi.Input<string>;
9376
+ /**
9377
+ * The Power BI storage mode of the table
9378
+ */
9149
9379
  storageMode?: pulumi.Input<string>;
9150
9380
  }
9151
9381
  export interface JobTaskForEachTaskTaskPythonWheelTask {
@@ -9220,11 +9450,13 @@ export interface JobTaskForEachTaskTaskSparkPythonTask {
9220
9450
  */
9221
9451
  parameters?: pulumi.Input<pulumi.Input<string>[]>;
9222
9452
  /**
9223
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
9453
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
9224
9454
  */
9225
9455
  pythonFile: pulumi.Input<string>;
9226
9456
  /**
9227
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
9457
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
9458
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
9459
+ * * `GIT`: The Python file is located in a remote Git repository.
9228
9460
  */
9229
9461
  source?: pulumi.Input<string>;
9230
9462
  }
@@ -9277,10 +9509,10 @@ export interface JobTaskForEachTaskTaskSqlTaskAlert {
9277
9509
  subscriptions?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskSqlTaskAlertSubscription>[]>;
9278
9510
  }
9279
9511
  export interface JobTaskForEachTaskTaskSqlTaskAlertSubscription {
9280
- destinationId?: pulumi.Input<string>;
9281
9512
  /**
9282
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
9513
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
9283
9514
  */
9515
+ destinationId?: pulumi.Input<string>;
9284
9516
  userName?: pulumi.Input<string>;
9285
9517
  }
9286
9518
  export interface JobTaskForEachTaskTaskSqlTaskDashboard {
@@ -9302,10 +9534,10 @@ export interface JobTaskForEachTaskTaskSqlTaskDashboard {
9302
9534
  subscriptions?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskSqlTaskDashboardSubscription>[]>;
9303
9535
  }
9304
9536
  export interface JobTaskForEachTaskTaskSqlTaskDashboardSubscription {
9305
- destinationId?: pulumi.Input<string>;
9306
9537
  /**
9307
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
9538
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
9308
9539
  */
9540
+ destinationId?: pulumi.Input<string>;
9309
9541
  userName?: pulumi.Input<string>;
9310
9542
  }
9311
9543
  export interface JobTaskForEachTaskTaskSqlTaskFile {
@@ -9370,10 +9602,6 @@ export interface JobTaskForEachTaskTaskSqlTaskQuery {
9370
9602
  export interface JobTaskForEachTaskTaskWebhookNotifications {
9371
9603
  /**
9372
9604
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
9373
- *
9374
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
9375
- *
9376
- * Example
9377
9605
  */
9378
9606
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskWebhookNotificationsOnDurationWarningThresholdExceeded>[]>;
9379
9607
  /**
@@ -9384,6 +9612,13 @@ export interface JobTaskForEachTaskTaskWebhookNotifications {
9384
9612
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
9385
9613
  */
9386
9614
  onStarts?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskWebhookNotificationsOnStart>[]>;
9615
+ /**
9616
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
9617
+ *
9618
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
9619
+ *
9620
+ * Example
9621
+ */
9387
9622
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<inputs.JobTaskForEachTaskTaskWebhookNotificationsOnStreamingBacklogExceeded>[]>;
9388
9623
  /**
9389
9624
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -9516,7 +9751,7 @@ export interface JobTaskNewCluster {
9516
9751
  sparkEnvVars?: pulumi.Input<{
9517
9752
  [key: string]: pulumi.Input<string>;
9518
9753
  }>;
9519
- sparkVersion: pulumi.Input<string>;
9754
+ sparkVersion?: pulumi.Input<string>;
9520
9755
  sshPublicKeys?: pulumi.Input<pulumi.Input<string>[]>;
9521
9756
  useMlRuntime?: pulumi.Input<boolean>;
9522
9757
  /**
@@ -9718,32 +9953,65 @@ export interface JobTaskPipelineTask {
9718
9953
  pipelineId: pulumi.Input<string>;
9719
9954
  }
9720
9955
  export interface JobTaskPowerBiTask {
9956
+ /**
9957
+ * The resource name of the UC connection to authenticate from Databricks to Power BI
9958
+ */
9721
9959
  connectionResourceName?: pulumi.Input<string>;
9960
+ /**
9961
+ * The semantic model to update. Block consists of following fields:
9962
+ */
9722
9963
  powerBiModel?: pulumi.Input<inputs.JobTaskPowerBiTaskPowerBiModel>;
9964
+ /**
9965
+ * Whether the model should be refreshed after the update. Default is false
9966
+ */
9723
9967
  refreshAfterUpdate?: pulumi.Input<boolean>;
9968
+ /**
9969
+ * The tables to be exported to Power BI. Block consists of following fields:
9970
+ */
9724
9971
  tables?: pulumi.Input<pulumi.Input<inputs.JobTaskPowerBiTaskTable>[]>;
9972
+ /**
9973
+ * The SQL warehouse ID to use as the Power BI data source
9974
+ */
9725
9975
  warehouseId?: pulumi.Input<string>;
9726
9976
  }
9727
9977
  export interface JobTaskPowerBiTaskPowerBiModel {
9978
+ /**
9979
+ * How the published Power BI model authenticates to Databricks
9980
+ */
9728
9981
  authenticationMethod?: pulumi.Input<string>;
9982
+ /**
9983
+ * The name of the Power BI model
9984
+ */
9729
9985
  modelName?: pulumi.Input<string>;
9986
+ /**
9987
+ * Whether to overwrite existing Power BI models. Default is false
9988
+ */
9730
9989
  overwriteExisting?: pulumi.Input<boolean>;
9990
+ /**
9991
+ * The default storage mode of the Power BI model
9992
+ */
9731
9993
  storageMode?: pulumi.Input<string>;
9994
+ /**
9995
+ * The name of the Power BI workspace of the model
9996
+ */
9732
9997
  workspaceName?: pulumi.Input<string>;
9733
9998
  }
9734
9999
  export interface JobTaskPowerBiTaskTable {
9735
10000
  /**
9736
- * The name of the catalog to use inside Unity Catalog.
10001
+ * The catalog name in Databricks
9737
10002
  */
9738
10003
  catalog?: pulumi.Input<string>;
9739
10004
  /**
9740
- * An optional name for the job. The default value is Untitled.
10005
+ * The table name in Databricks. If empty, all tables under the schema are selected.
9741
10006
  */
9742
10007
  name?: pulumi.Input<string>;
9743
10008
  /**
9744
- * The name of the schema dbt should run in. Defaults to `default`.
10009
+ * The schema name in Databricks
9745
10010
  */
9746
10011
  schema?: pulumi.Input<string>;
10012
+ /**
10013
+ * The Power BI storage mode of the table
10014
+ */
9747
10015
  storageMode?: pulumi.Input<string>;
9748
10016
  }
9749
10017
  export interface JobTaskPythonWheelTask {
@@ -9818,11 +10086,13 @@ export interface JobTaskSparkPythonTask {
9818
10086
  */
9819
10087
  parameters?: pulumi.Input<pulumi.Input<string>[]>;
9820
10088
  /**
9821
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
10089
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
9822
10090
  */
9823
10091
  pythonFile: pulumi.Input<string>;
9824
10092
  /**
9825
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
10093
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
10094
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
10095
+ * * `GIT`: The Python file is located in a remote Git repository.
9826
10096
  */
9827
10097
  source?: pulumi.Input<string>;
9828
10098
  }
@@ -9875,10 +10145,10 @@ export interface JobTaskSqlTaskAlert {
9875
10145
  subscriptions?: pulumi.Input<pulumi.Input<inputs.JobTaskSqlTaskAlertSubscription>[]>;
9876
10146
  }
9877
10147
  export interface JobTaskSqlTaskAlertSubscription {
9878
- destinationId?: pulumi.Input<string>;
9879
10148
  /**
9880
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
10149
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
9881
10150
  */
10151
+ destinationId?: pulumi.Input<string>;
9882
10152
  userName?: pulumi.Input<string>;
9883
10153
  }
9884
10154
  export interface JobTaskSqlTaskDashboard {
@@ -9900,10 +10170,10 @@ export interface JobTaskSqlTaskDashboard {
9900
10170
  subscriptions?: pulumi.Input<pulumi.Input<inputs.JobTaskSqlTaskDashboardSubscription>[]>;
9901
10171
  }
9902
10172
  export interface JobTaskSqlTaskDashboardSubscription {
9903
- destinationId?: pulumi.Input<string>;
9904
10173
  /**
9905
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
10174
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
9906
10175
  */
10176
+ destinationId?: pulumi.Input<string>;
9907
10177
  userName?: pulumi.Input<string>;
9908
10178
  }
9909
10179
  export interface JobTaskSqlTaskFile {
@@ -9968,10 +10238,6 @@ export interface JobTaskSqlTaskQuery {
9968
10238
  export interface JobTaskWebhookNotifications {
9969
10239
  /**
9970
10240
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
9971
- *
9972
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
9973
- *
9974
- * Example
9975
10241
  */
9976
10242
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<inputs.JobTaskWebhookNotificationsOnDurationWarningThresholdExceeded>[]>;
9977
10243
  /**
@@ -9982,6 +10248,13 @@ export interface JobTaskWebhookNotifications {
9982
10248
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
9983
10249
  */
9984
10250
  onStarts?: pulumi.Input<pulumi.Input<inputs.JobTaskWebhookNotificationsOnStart>[]>;
10251
+ /**
10252
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
10253
+ *
10254
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
10255
+ *
10256
+ * Example
10257
+ */
9985
10258
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<inputs.JobTaskWebhookNotificationsOnStreamingBacklogExceeded>[]>;
9986
10259
  /**
9987
10260
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -10085,10 +10358,6 @@ export interface JobTriggerTableUpdate {
10085
10358
  export interface JobWebhookNotifications {
10086
10359
  /**
10087
10360
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
10088
- *
10089
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
10090
- *
10091
- * Example
10092
10361
  */
10093
10362
  onDurationWarningThresholdExceededs?: pulumi.Input<pulumi.Input<inputs.JobWebhookNotificationsOnDurationWarningThresholdExceeded>[]>;
10094
10363
  /**
@@ -10099,6 +10368,13 @@ export interface JobWebhookNotifications {
10099
10368
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
10100
10369
  */
10101
10370
  onStarts?: pulumi.Input<pulumi.Input<inputs.JobWebhookNotificationsOnStart>[]>;
10371
+ /**
10372
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
10373
+ *
10374
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
10375
+ *
10376
+ * Example
10377
+ */
10102
10378
  onStreamingBacklogExceededs?: pulumi.Input<pulumi.Input<inputs.JobWebhookNotificationsOnStreamingBacklogExceeded>[]>;
10103
10379
  /**
10104
10380
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -10711,7 +10987,7 @@ export interface ModelServingConfigServedEntityExternalModelOpenaiConfig {
10711
10987
  */
10712
10988
  microsoftEntraTenantId?: pulumi.Input<string>;
10713
10989
  /**
10714
- * This is the base URL for the OpenAI API (default: "https://api.openai.com/v1"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
10990
+ * This is the base URL for the OpenAI API (default: "<https://api.openai.com/v1>"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
10715
10991
  */
10716
10992
  openaiApiBase?: pulumi.Input<string>;
10717
10993
  /**
@@ -10958,13 +11234,13 @@ export interface MwsNetworksGcpNetworkInfo {
10958
11234
  /**
10959
11235
  * The name of the secondary IP range for pods. A Databricks-managed GKE cluster uses this IP range for its pods. This secondary IP range can only be used by one workspace.
10960
11236
  *
10961
- * @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-vpc
11237
+ * @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
10962
11238
  */
10963
11239
  podIpRangeName?: pulumi.Input<string>;
10964
11240
  /**
10965
11241
  * The name of the secondary IP range for services. A Databricks-managed GKE cluster uses this IP range for its services. This secondary IP range can only be used by one workspace.
10966
11242
  *
10967
- * @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-vpc
11243
+ * @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
10968
11244
  */
10969
11245
  serviceIpRangeName?: pulumi.Input<string>;
10970
11246
  /**
@@ -11025,11 +11301,11 @@ export interface MwsWorkspacesExternalCustomerInfo {
11025
11301
  }
11026
11302
  export interface MwsWorkspacesGcpManagedNetworkConfig {
11027
11303
  /**
11028
- * @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
11304
+ * @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
11029
11305
  */
11030
11306
  gkeClusterPodIpRange?: pulumi.Input<string>;
11031
11307
  /**
11032
- * @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
11308
+ * @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
11033
11309
  */
11034
11310
  gkeClusterServiceIpRange?: pulumi.Input<string>;
11035
11311
  subnetCidr: pulumi.Input<string>;