@pulumi/databricks 1.68.0-alpha.1745473147 → 1.68.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/types/output.d.ts CHANGED
@@ -1756,7 +1756,7 @@ export interface GetClusterClusterInfoSpec {
1756
1756
  /**
1757
1757
  * [Runtime version](https://docs.databricks.com/runtime/index.html) of the cluster.
1758
1758
  */
1759
- sparkVersion: string;
1759
+ sparkVersion?: string;
1760
1760
  /**
1761
1761
  * SSH public key contents that will be added to each Spark node in this cluster.
1762
1762
  */
@@ -2588,7 +2588,7 @@ export interface GetJobJobSettingsSettingsJobClusterNewCluster {
2588
2588
  sparkEnvVars?: {
2589
2589
  [key: string]: string;
2590
2590
  };
2591
- sparkVersion: string;
2591
+ sparkVersion?: string;
2592
2592
  sshPublicKeys?: string[];
2593
2593
  workloadType?: outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadType;
2594
2594
  }
@@ -2751,7 +2751,7 @@ export interface GetJobJobSettingsSettingsNewCluster {
2751
2751
  sparkEnvVars?: {
2752
2752
  [key: string]: string;
2753
2753
  };
2754
- sparkVersion: string;
2754
+ sparkVersion?: string;
2755
2755
  sshPublicKeys?: string[];
2756
2756
  workloadType?: outputs.GetJobJobSettingsSettingsNewClusterWorkloadType;
2757
2757
  }
@@ -2922,6 +2922,7 @@ export interface GetJobJobSettingsSettingsSparkSubmitTask {
2922
2922
  }
2923
2923
  export interface GetJobJobSettingsSettingsTask {
2924
2924
  conditionTask?: outputs.GetJobJobSettingsSettingsTaskConditionTask;
2925
+ dashboardTask?: outputs.GetJobJobSettingsSettingsTaskDashboardTask;
2925
2926
  dbtTask?: outputs.GetJobJobSettingsSettingsTaskDbtTask;
2926
2927
  dependsOns?: outputs.GetJobJobSettingsSettingsTaskDependsOn[];
2927
2928
  description?: string;
@@ -2938,6 +2939,7 @@ export interface GetJobJobSettingsSettingsTask {
2938
2939
  notebookTask?: outputs.GetJobJobSettingsSettingsTaskNotebookTask;
2939
2940
  notificationSettings?: outputs.GetJobJobSettingsSettingsTaskNotificationSettings;
2940
2941
  pipelineTask?: outputs.GetJobJobSettingsSettingsTaskPipelineTask;
2942
+ powerBiTask?: outputs.GetJobJobSettingsSettingsTaskPowerBiTask;
2941
2943
  pythonWheelTask?: outputs.GetJobJobSettingsSettingsTaskPythonWheelTask;
2942
2944
  retryOnTimeout: boolean;
2943
2945
  runIf?: string;
@@ -2955,6 +2957,20 @@ export interface GetJobJobSettingsSettingsTaskConditionTask {
2955
2957
  op: string;
2956
2958
  right: string;
2957
2959
  }
2960
+ export interface GetJobJobSettingsSettingsTaskDashboardTask {
2961
+ dashboardId?: string;
2962
+ subscription?: outputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscription;
2963
+ warehouseId?: string;
2964
+ }
2965
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscription {
2966
+ customSubject?: string;
2967
+ paused?: boolean;
2968
+ subscribers?: outputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber[];
2969
+ }
2970
+ export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber {
2971
+ destinationId?: string;
2972
+ userName?: string;
2973
+ }
2958
2974
  export interface GetJobJobSettingsSettingsTaskDbtTask {
2959
2975
  catalog?: string;
2960
2976
  commands: string[];
@@ -2983,6 +2999,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTask {
2983
2999
  }
2984
3000
  export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
2985
3001
  conditionTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTask;
3002
+ dashboardTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask;
2986
3003
  dbtTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask;
2987
3004
  dependsOns?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOn[];
2988
3005
  description?: string;
@@ -2998,6 +3015,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
2998
3015
  notebookTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotebookTask;
2999
3016
  notificationSettings?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotificationSettings;
3000
3017
  pipelineTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTask;
3018
+ powerBiTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask;
3001
3019
  pythonWheelTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask;
3002
3020
  retryOnTimeout: boolean;
3003
3021
  runIf?: string;
@@ -3015,6 +3033,20 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTask {
3015
3033
  op: string;
3016
3034
  right: string;
3017
3035
  }
3036
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask {
3037
+ dashboardId?: string;
3038
+ subscription?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription;
3039
+ warehouseId?: string;
3040
+ }
3041
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription {
3042
+ customSubject?: string;
3043
+ paused?: boolean;
3044
+ subscribers?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber[];
3045
+ }
3046
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
3047
+ destinationId?: string;
3048
+ userName?: string;
3049
+ }
3018
3050
  export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask {
3019
3051
  catalog?: string;
3020
3052
  commands: string[];
@@ -3100,7 +3132,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskNewCluster {
3100
3132
  sparkEnvVars?: {
3101
3133
  [key: string]: string;
3102
3134
  };
3103
- sparkVersion: string;
3135
+ sparkVersion?: string;
3104
3136
  sshPublicKeys?: string[];
3105
3137
  workloadType?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNewClusterWorkloadType;
3106
3138
  }
@@ -3224,6 +3256,29 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTask {
3224
3256
  fullRefresh?: boolean;
3225
3257
  pipelineId: string;
3226
3258
  }
3259
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask {
3260
+ connectionResourceName?: string;
3261
+ powerBiModel?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel;
3262
+ refreshAfterUpdate?: boolean;
3263
+ tables?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable[];
3264
+ warehouseId?: string;
3265
+ }
3266
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel {
3267
+ authenticationMethod?: string;
3268
+ modelName?: string;
3269
+ overwriteExisting?: boolean;
3270
+ storageMode?: string;
3271
+ workspaceName?: string;
3272
+ }
3273
+ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable {
3274
+ catalog?: string;
3275
+ /**
3276
+ * the job name of databricks.Job if the resource was matched by id.
3277
+ */
3278
+ name?: string;
3279
+ schema?: string;
3280
+ storageMode?: string;
3281
+ }
3227
3282
  export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask {
3228
3283
  entryPoint?: string;
3229
3284
  namedParameters?: {
@@ -3388,7 +3443,7 @@ export interface GetJobJobSettingsSettingsTaskNewCluster {
3388
3443
  sparkEnvVars?: {
3389
3444
  [key: string]: string;
3390
3445
  };
3391
- sparkVersion: string;
3446
+ sparkVersion?: string;
3392
3447
  sshPublicKeys?: string[];
3393
3448
  workloadType?: outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadType;
3394
3449
  }
@@ -3512,6 +3567,29 @@ export interface GetJobJobSettingsSettingsTaskPipelineTask {
3512
3567
  fullRefresh?: boolean;
3513
3568
  pipelineId: string;
3514
3569
  }
3570
+ export interface GetJobJobSettingsSettingsTaskPowerBiTask {
3571
+ connectionResourceName?: string;
3572
+ powerBiModel?: outputs.GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel;
3573
+ refreshAfterUpdate?: boolean;
3574
+ tables?: outputs.GetJobJobSettingsSettingsTaskPowerBiTaskTable[];
3575
+ warehouseId?: string;
3576
+ }
3577
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel {
3578
+ authenticationMethod?: string;
3579
+ modelName?: string;
3580
+ overwriteExisting?: boolean;
3581
+ storageMode?: string;
3582
+ workspaceName?: string;
3583
+ }
3584
+ export interface GetJobJobSettingsSettingsTaskPowerBiTaskTable {
3585
+ catalog?: string;
3586
+ /**
3587
+ * the job name of databricks.Job if the resource was matched by id.
3588
+ */
3589
+ name?: string;
3590
+ schema?: string;
3591
+ storageMode?: string;
3592
+ }
3515
3593
  export interface GetJobJobSettingsSettingsTaskPythonWheelTask {
3516
3594
  entryPoint?: string;
3517
3595
  namedParameters?: {
@@ -4877,8 +4955,6 @@ export interface JobEmailNotifications {
4877
4955
  noAlertForSkippedRuns?: boolean;
4878
4956
  /**
4879
4957
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
4880
- *
4881
- * The following parameter is only available for the job level configuration.
4882
4958
  */
4883
4959
  onDurationWarningThresholdExceededs?: string[];
4884
4960
  /**
@@ -4889,6 +4965,11 @@ export interface JobEmailNotifications {
4889
4965
  * (List) list of emails to notify when the run starts.
4890
4966
  */
4891
4967
  onStarts?: string[];
4968
+ /**
4969
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
4970
+ *
4971
+ * The following parameter is only available for the job level configuration.
4972
+ */
4892
4973
  onStreamingBacklogExceededs?: string[];
4893
4974
  /**
4894
4975
  * (List) list of emails to notify when the run completes successfully.
@@ -5017,7 +5098,7 @@ export interface JobJobClusterNewCluster {
5017
5098
  sparkEnvVars?: {
5018
5099
  [key: string]: string;
5019
5100
  };
5020
- sparkVersion: string;
5101
+ sparkVersion?: string;
5021
5102
  sshPublicKeys?: string[];
5022
5103
  useMlRuntime?: boolean;
5023
5104
  /**
@@ -5231,7 +5312,7 @@ export interface JobNewCluster {
5231
5312
  sparkEnvVars?: {
5232
5313
  [key: string]: string;
5233
5314
  };
5234
- sparkVersion: string;
5315
+ sparkVersion?: string;
5235
5316
  sshPublicKeys?: string[];
5236
5317
  useMlRuntime?: boolean;
5237
5318
  /**
@@ -5530,11 +5611,13 @@ export interface JobSparkPythonTask {
5530
5611
  */
5531
5612
  parameters?: string[];
5532
5613
  /**
5533
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
5614
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
5534
5615
  */
5535
5616
  pythonFile: string;
5536
5617
  /**
5537
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
5618
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
5619
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
5620
+ * * `GIT`: The Python file is located in a remote Git repository.
5538
5621
  */
5539
5622
  source?: string;
5540
5623
  }
@@ -5674,24 +5757,39 @@ export interface JobTaskConditionTask {
5674
5757
  }
5675
5758
  export interface JobTaskDashboardTask {
5676
5759
  /**
5677
- * (String) identifier of the Databricks SQL Dashboard databricks_sql_dashboard.
5760
+ * The identifier of the dashboard to refresh
5678
5761
  */
5679
5762
  dashboardId?: string;
5763
+ /**
5764
+ * Represents a subscription configuration for scheduled dashboard snapshots.
5765
+ */
5680
5766
  subscription?: outputs.JobTaskDashboardTaskSubscription;
5767
+ /**
5768
+ * The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
5769
+ */
5681
5770
  warehouseId?: string;
5682
5771
  }
5683
5772
  export interface JobTaskDashboardTaskSubscription {
5684
5773
  /**
5685
- * string specifying a custom subject of email sent.
5774
+ * Allows users to specify a custom subject line on the email sent to subscribers.
5686
5775
  */
5687
5776
  customSubject?: string;
5777
+ /**
5778
+ * When true, the subscription will not send emails.
5779
+ */
5688
5780
  paused?: boolean;
5781
+ /**
5782
+ * The list of subscribers to send the snapshot of the dashboard to.
5783
+ */
5689
5784
  subscribers?: outputs.JobTaskDashboardTaskSubscriptionSubscriber[];
5690
5785
  }
5691
5786
  export interface JobTaskDashboardTaskSubscriptionSubscriber {
5787
+ /**
5788
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
5789
+ */
5692
5790
  destinationId?: string;
5693
5791
  /**
5694
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
5792
+ * A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
5695
5793
  */
5696
5794
  userName?: string;
5697
5795
  }
@@ -5748,8 +5846,6 @@ export interface JobTaskEmailNotifications {
5748
5846
  noAlertForSkippedRuns?: boolean;
5749
5847
  /**
5750
5848
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
5751
- *
5752
- * The following parameter is only available for the job level configuration.
5753
5849
  */
5754
5850
  onDurationWarningThresholdExceededs?: string[];
5755
5851
  /**
@@ -5760,6 +5856,11 @@ export interface JobTaskEmailNotifications {
5760
5856
  * (List) list of emails to notify when the run starts.
5761
5857
  */
5762
5858
  onStarts?: string[];
5859
+ /**
5860
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
5861
+ *
5862
+ * The following parameter is only available for the job level configuration.
5863
+ */
5763
5864
  onStreamingBacklogExceededs?: string[];
5764
5865
  /**
5765
5866
  * (List) list of emails to notify when the run completes successfully.
@@ -5909,24 +6010,39 @@ export interface JobTaskForEachTaskTaskConditionTask {
5909
6010
  }
5910
6011
  export interface JobTaskForEachTaskTaskDashboardTask {
5911
6012
  /**
5912
- * (String) identifier of the Databricks SQL Dashboard databricks_sql_dashboard.
6013
+ * The identifier of the dashboard to refresh
5913
6014
  */
5914
6015
  dashboardId?: string;
6016
+ /**
6017
+ * Represents a subscription configuration for scheduled dashboard snapshots.
6018
+ */
5915
6019
  subscription?: outputs.JobTaskForEachTaskTaskDashboardTaskSubscription;
6020
+ /**
6021
+ * The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
6022
+ */
5916
6023
  warehouseId?: string;
5917
6024
  }
5918
6025
  export interface JobTaskForEachTaskTaskDashboardTaskSubscription {
5919
6026
  /**
5920
- * string specifying a custom subject of email sent.
6027
+ * Allows users to specify a custom subject line on the email sent to subscribers.
5921
6028
  */
5922
6029
  customSubject?: string;
6030
+ /**
6031
+ * When true, the subscription will not send emails.
6032
+ */
5923
6033
  paused?: boolean;
6034
+ /**
6035
+ * The list of subscribers to send the snapshot of the dashboard to.
6036
+ */
5924
6037
  subscribers?: outputs.JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber[];
5925
6038
  }
5926
6039
  export interface JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
6040
+ /**
6041
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
6042
+ */
5927
6043
  destinationId?: string;
5928
6044
  /**
5929
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
6045
+ * A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
5930
6046
  */
5931
6047
  userName?: string;
5932
6048
  }
@@ -5983,8 +6099,6 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
5983
6099
  noAlertForSkippedRuns?: boolean;
5984
6100
  /**
5985
6101
  * (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
5986
- *
5987
- * The following parameter is only available for the job level configuration.
5988
6102
  */
5989
6103
  onDurationWarningThresholdExceededs?: string[];
5990
6104
  /**
@@ -5995,6 +6109,11 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
5995
6109
  * (List) list of emails to notify when the run starts.
5996
6110
  */
5997
6111
  onStarts?: string[];
6112
+ /**
6113
+ * (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
6114
+ *
6115
+ * The following parameter is only available for the job level configuration.
6116
+ */
5998
6117
  onStreamingBacklogExceededs?: string[];
5999
6118
  /**
6000
6119
  * (List) list of emails to notify when the run completes successfully.
@@ -6097,7 +6216,7 @@ export interface JobTaskForEachTaskTaskNewCluster {
6097
6216
  sparkEnvVars?: {
6098
6217
  [key: string]: string;
6099
6218
  };
6100
- sparkVersion: string;
6219
+ sparkVersion?: string;
6101
6220
  sshPublicKeys?: string[];
6102
6221
  useMlRuntime?: boolean;
6103
6222
  /**
@@ -6299,32 +6418,65 @@ export interface JobTaskForEachTaskTaskPipelineTask {
6299
6418
  pipelineId: string;
6300
6419
  }
6301
6420
  export interface JobTaskForEachTaskTaskPowerBiTask {
6421
+ /**
6422
+ * The resource name of the UC connection to authenticate from Databricks to Power BI
6423
+ */
6302
6424
  connectionResourceName?: string;
6425
+ /**
6426
+ * The semantic model to update. Block consists of following fields:
6427
+ */
6303
6428
  powerBiModel?: outputs.JobTaskForEachTaskTaskPowerBiTaskPowerBiModel;
6429
+ /**
6430
+ * Whether the model should be refreshed after the update. Default is false
6431
+ */
6304
6432
  refreshAfterUpdate?: boolean;
6433
+ /**
6434
+ * The tables to be exported to Power BI. Block consists of following fields:
6435
+ */
6305
6436
  tables?: outputs.JobTaskForEachTaskTaskPowerBiTaskTable[];
6437
+ /**
6438
+ * The SQL warehouse ID to use as the Power BI data source
6439
+ */
6306
6440
  warehouseId?: string;
6307
6441
  }
6308
6442
  export interface JobTaskForEachTaskTaskPowerBiTaskPowerBiModel {
6443
+ /**
6444
+ * How the published Power BI model authenticates to Databricks
6445
+ */
6309
6446
  authenticationMethod?: string;
6447
+ /**
6448
+ * The name of the Power BI model
6449
+ */
6310
6450
  modelName?: string;
6451
+ /**
6452
+ * Whether to overwrite existing Power BI models. Default is false
6453
+ */
6311
6454
  overwriteExisting?: boolean;
6455
+ /**
6456
+ * The default storage mode of the Power BI model
6457
+ */
6312
6458
  storageMode?: string;
6459
+ /**
6460
+ * The name of the Power BI workspace of the model
6461
+ */
6313
6462
  workspaceName?: string;
6314
6463
  }
6315
6464
  export interface JobTaskForEachTaskTaskPowerBiTaskTable {
6316
6465
  /**
6317
- * The name of the catalog to use inside Unity Catalog.
6466
+ * The catalog name in Databricks
6318
6467
  */
6319
6468
  catalog?: string;
6320
6469
  /**
6321
- * An optional name for the job. The default value is Untitled.
6470
+ * The table name in Databricks. If empty, all tables under the schema are selected.
6322
6471
  */
6323
6472
  name?: string;
6324
6473
  /**
6325
- * The name of the schema dbt should run in. Defaults to `default`.
6474
+ * The schema name in Databricks
6326
6475
  */
6327
6476
  schema?: string;
6477
+ /**
6478
+ * The Power BI storage mode of the table
6479
+ */
6328
6480
  storageMode?: string;
6329
6481
  }
6330
6482
  export interface JobTaskForEachTaskTaskPythonWheelTask {
@@ -6399,11 +6551,13 @@ export interface JobTaskForEachTaskTaskSparkPythonTask {
6399
6551
  */
6400
6552
  parameters?: string[];
6401
6553
  /**
6402
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
6554
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
6403
6555
  */
6404
6556
  pythonFile: string;
6405
6557
  /**
6406
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
6558
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
6559
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
6560
+ * * `GIT`: The Python file is located in a remote Git repository.
6407
6561
  */
6408
6562
  source?: string;
6409
6563
  }
@@ -6456,10 +6610,10 @@ export interface JobTaskForEachTaskTaskSqlTaskAlert {
6456
6610
  subscriptions?: outputs.JobTaskForEachTaskTaskSqlTaskAlertSubscription[];
6457
6611
  }
6458
6612
  export interface JobTaskForEachTaskTaskSqlTaskAlertSubscription {
6459
- destinationId?: string;
6460
6613
  /**
6461
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
6614
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
6462
6615
  */
6616
+ destinationId?: string;
6463
6617
  userName?: string;
6464
6618
  }
6465
6619
  export interface JobTaskForEachTaskTaskSqlTaskDashboard {
@@ -6481,10 +6635,10 @@ export interface JobTaskForEachTaskTaskSqlTaskDashboard {
6481
6635
  subscriptions?: outputs.JobTaskForEachTaskTaskSqlTaskDashboardSubscription[];
6482
6636
  }
6483
6637
  export interface JobTaskForEachTaskTaskSqlTaskDashboardSubscription {
6484
- destinationId?: string;
6485
6638
  /**
6486
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
6639
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
6487
6640
  */
6641
+ destinationId?: string;
6488
6642
  userName?: string;
6489
6643
  }
6490
6644
  export interface JobTaskForEachTaskTaskSqlTaskFile {
@@ -6549,10 +6703,6 @@ export interface JobTaskForEachTaskTaskSqlTaskQuery {
6549
6703
  export interface JobTaskForEachTaskTaskWebhookNotifications {
6550
6704
  /**
6551
6705
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
6552
- *
6553
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
6554
- *
6555
- * Example
6556
6706
  */
6557
6707
  onDurationWarningThresholdExceededs?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnDurationWarningThresholdExceeded[];
6558
6708
  /**
@@ -6563,6 +6713,13 @@ export interface JobTaskForEachTaskTaskWebhookNotifications {
6563
6713
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
6564
6714
  */
6565
6715
  onStarts?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnStart[];
6716
+ /**
6717
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
6718
+ *
6719
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
6720
+ *
6721
+ * Example
6722
+ */
6566
6723
  onStreamingBacklogExceededs?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnStreamingBacklogExceeded[];
6567
6724
  /**
6568
6725
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -6695,7 +6852,7 @@ export interface JobTaskNewCluster {
6695
6852
  sparkEnvVars?: {
6696
6853
  [key: string]: string;
6697
6854
  };
6698
- sparkVersion: string;
6855
+ sparkVersion?: string;
6699
6856
  sshPublicKeys?: string[];
6700
6857
  useMlRuntime?: boolean;
6701
6858
  /**
@@ -6897,32 +7054,65 @@ export interface JobTaskPipelineTask {
6897
7054
  pipelineId: string;
6898
7055
  }
6899
7056
  export interface JobTaskPowerBiTask {
7057
+ /**
7058
+ * The resource name of the UC connection to authenticate from Databricks to Power BI
7059
+ */
6900
7060
  connectionResourceName?: string;
7061
+ /**
7062
+ * The semantic model to update. Block consists of following fields:
7063
+ */
6901
7064
  powerBiModel?: outputs.JobTaskPowerBiTaskPowerBiModel;
7065
+ /**
7066
+ * Whether the model should be refreshed after the update. Default is false
7067
+ */
6902
7068
  refreshAfterUpdate?: boolean;
7069
+ /**
7070
+ * The tables to be exported to Power BI. Block consists of following fields:
7071
+ */
6903
7072
  tables?: outputs.JobTaskPowerBiTaskTable[];
7073
+ /**
7074
+ * The SQL warehouse ID to use as the Power BI data source
7075
+ */
6904
7076
  warehouseId?: string;
6905
7077
  }
6906
7078
  export interface JobTaskPowerBiTaskPowerBiModel {
7079
+ /**
7080
+ * How the published Power BI model authenticates to Databricks
7081
+ */
6907
7082
  authenticationMethod?: string;
7083
+ /**
7084
+ * The name of the Power BI model
7085
+ */
6908
7086
  modelName?: string;
7087
+ /**
7088
+ * Whether to overwrite existing Power BI models. Default is false
7089
+ */
6909
7090
  overwriteExisting?: boolean;
7091
+ /**
7092
+ * The default storage mode of the Power BI model
7093
+ */
6910
7094
  storageMode?: string;
7095
+ /**
7096
+ * The name of the Power BI workspace of the model
7097
+ */
6911
7098
  workspaceName?: string;
6912
7099
  }
6913
7100
  export interface JobTaskPowerBiTaskTable {
6914
7101
  /**
6915
- * The name of the catalog to use inside Unity Catalog.
7102
+ * The catalog name in Databricks
6916
7103
  */
6917
7104
  catalog?: string;
6918
7105
  /**
6919
- * An optional name for the job. The default value is Untitled.
7106
+ * The table name in Databricks. If empty, all tables under the schema are selected.
6920
7107
  */
6921
7108
  name?: string;
6922
7109
  /**
6923
- * The name of the schema dbt should run in. Defaults to `default`.
7110
+ * The schema name in Databricks
6924
7111
  */
6925
7112
  schema?: string;
7113
+ /**
7114
+ * The Power BI storage mode of the table
7115
+ */
6926
7116
  storageMode?: string;
6927
7117
  }
6928
7118
  export interface JobTaskPythonWheelTask {
@@ -6997,11 +7187,13 @@ export interface JobTaskSparkPythonTask {
6997
7187
  */
6998
7188
  parameters?: string[];
6999
7189
  /**
7000
- * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/Repos`. For files stored in a remote repository, the path must be relative. This field is required.
7190
+ * The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
7001
7191
  */
7002
7192
  pythonFile: string;
7003
7193
  /**
7004
- * Location type of the Python file, can only be `GIT`. When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
7194
+ * Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
7195
+ * * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
7196
+ * * `GIT`: The Python file is located in a remote Git repository.
7005
7197
  */
7006
7198
  source?: string;
7007
7199
  }
@@ -7054,10 +7246,10 @@ export interface JobTaskSqlTaskAlert {
7054
7246
  subscriptions?: outputs.JobTaskSqlTaskAlertSubscription[];
7055
7247
  }
7056
7248
  export interface JobTaskSqlTaskAlertSubscription {
7057
- destinationId?: string;
7058
7249
  /**
7059
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
7250
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
7060
7251
  */
7252
+ destinationId?: string;
7061
7253
  userName?: string;
7062
7254
  }
7063
7255
  export interface JobTaskSqlTaskDashboard {
@@ -7079,10 +7271,10 @@ export interface JobTaskSqlTaskDashboard {
7079
7271
  subscriptions?: outputs.JobTaskSqlTaskDashboardSubscription[];
7080
7272
  }
7081
7273
  export interface JobTaskSqlTaskDashboardSubscription {
7082
- destinationId?: string;
7083
7274
  /**
7084
- * The email of an active workspace user. Non-admin users can only set this field to their own email.
7275
+ * A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
7085
7276
  */
7277
+ destinationId?: string;
7086
7278
  userName?: string;
7087
7279
  }
7088
7280
  export interface JobTaskSqlTaskFile {
@@ -7147,10 +7339,6 @@ export interface JobTaskSqlTaskQuery {
7147
7339
  export interface JobTaskWebhookNotifications {
7148
7340
  /**
7149
7341
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
7150
- *
7151
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
7152
- *
7153
- * Example
7154
7342
  */
7155
7343
  onDurationWarningThresholdExceededs?: outputs.JobTaskWebhookNotificationsOnDurationWarningThresholdExceeded[];
7156
7344
  /**
@@ -7161,6 +7349,13 @@ export interface JobTaskWebhookNotifications {
7161
7349
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
7162
7350
  */
7163
7351
  onStarts?: outputs.JobTaskWebhookNotificationsOnStart[];
7352
+ /**
7353
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
7354
+ *
7355
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
7356
+ *
7357
+ * Example
7358
+ */
7164
7359
  onStreamingBacklogExceededs?: outputs.JobTaskWebhookNotificationsOnStreamingBacklogExceeded[];
7165
7360
  /**
7166
7361
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -7264,10 +7459,6 @@ export interface JobTriggerTableUpdate {
7264
7459
  export interface JobWebhookNotifications {
7265
7460
  /**
7266
7461
  * (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
7267
- *
7268
- * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
7269
- *
7270
- * Example
7271
7462
  */
7272
7463
  onDurationWarningThresholdExceededs?: outputs.JobWebhookNotificationsOnDurationWarningThresholdExceeded[];
7273
7464
  /**
@@ -7278,6 +7469,13 @@ export interface JobWebhookNotifications {
7278
7469
  * (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
7279
7470
  */
7280
7471
  onStarts?: outputs.JobWebhookNotificationsOnStart[];
7472
+ /**
7473
+ * (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
7474
+ *
7475
+ * Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
7476
+ *
7477
+ * Example
7478
+ */
7281
7479
  onStreamingBacklogExceededs?: outputs.JobWebhookNotificationsOnStreamingBacklogExceeded[];
7282
7480
  /**
7283
7481
  * (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
@@ -7890,7 +8088,7 @@ export interface ModelServingConfigServedEntityExternalModelOpenaiConfig {
7890
8088
  */
7891
8089
  microsoftEntraTenantId?: string;
7892
8090
  /**
7893
- * This is the base URL for the OpenAI API (default: "https://api.openai.com/v1"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
8091
+ * This is the base URL for the OpenAI API (default: "<https://api.openai.com/v1>"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
7894
8092
  */
7895
8093
  openaiApiBase?: string;
7896
8094
  /**
@@ -8137,13 +8335,13 @@ export interface MwsNetworksGcpNetworkInfo {
8137
8335
  /**
8138
8336
  * The name of the secondary IP range for pods. A Databricks-managed GKE cluster uses this IP range for its pods. This secondary IP range can only be used by one workspace.
8139
8337
  *
8140
- * @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-vpc
8338
+ * @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
8141
8339
  */
8142
8340
  podIpRangeName?: string;
8143
8341
  /**
8144
8342
  * The name of the secondary IP range for services. A Databricks-managed GKE cluster uses this IP range for its services. This secondary IP range can only be used by one workspace.
8145
8343
  *
8146
- * @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-vpc
8344
+ * @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
8147
8345
  */
8148
8346
  serviceIpRangeName?: string;
8149
8347
  /**
@@ -8204,11 +8402,11 @@ export interface MwsWorkspacesExternalCustomerInfo {
8204
8402
  }
8205
8403
  export interface MwsWorkspacesGcpManagedNetworkConfig {
8206
8404
  /**
8207
- * @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
8405
+ * @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
8208
8406
  */
8209
8407
  gkeClusterPodIpRange?: string;
8210
8408
  /**
8211
- * @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.74.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
8409
+ * @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
8212
8410
  */
8213
8411
  gkeClusterServiceIpRange?: string;
8214
8412
  subnetCidr: string;