@pulumi/databricks 1.68.0-alpha.1745473147 → 1.68.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/job.d.ts +15 -0
- package/job.js.map +1 -1
- package/modelServing.d.ts +80 -0
- package/modelServing.js +80 -0
- package/modelServing.js.map +1 -1
- package/mwsWorkspaces.d.ts +57 -10
- package/mwsWorkspaces.js +26 -4
- package/mwsWorkspaces.js.map +1 -1
- package/package.json +2 -2
- package/types/input.d.ts +339 -63
- package/types/output.d.ts +256 -58
package/types/output.d.ts
CHANGED
|
@@ -1756,7 +1756,7 @@ export interface GetClusterClusterInfoSpec {
|
|
|
1756
1756
|
/**
|
|
1757
1757
|
* [Runtime version](https://docs.databricks.com/runtime/index.html) of the cluster.
|
|
1758
1758
|
*/
|
|
1759
|
-
sparkVersion
|
|
1759
|
+
sparkVersion?: string;
|
|
1760
1760
|
/**
|
|
1761
1761
|
* SSH public key contents that will be added to each Spark node in this cluster.
|
|
1762
1762
|
*/
|
|
@@ -2588,7 +2588,7 @@ export interface GetJobJobSettingsSettingsJobClusterNewCluster {
|
|
|
2588
2588
|
sparkEnvVars?: {
|
|
2589
2589
|
[key: string]: string;
|
|
2590
2590
|
};
|
|
2591
|
-
sparkVersion
|
|
2591
|
+
sparkVersion?: string;
|
|
2592
2592
|
sshPublicKeys?: string[];
|
|
2593
2593
|
workloadType?: outputs.GetJobJobSettingsSettingsJobClusterNewClusterWorkloadType;
|
|
2594
2594
|
}
|
|
@@ -2751,7 +2751,7 @@ export interface GetJobJobSettingsSettingsNewCluster {
|
|
|
2751
2751
|
sparkEnvVars?: {
|
|
2752
2752
|
[key: string]: string;
|
|
2753
2753
|
};
|
|
2754
|
-
sparkVersion
|
|
2754
|
+
sparkVersion?: string;
|
|
2755
2755
|
sshPublicKeys?: string[];
|
|
2756
2756
|
workloadType?: outputs.GetJobJobSettingsSettingsNewClusterWorkloadType;
|
|
2757
2757
|
}
|
|
@@ -2922,6 +2922,7 @@ export interface GetJobJobSettingsSettingsSparkSubmitTask {
|
|
|
2922
2922
|
}
|
|
2923
2923
|
export interface GetJobJobSettingsSettingsTask {
|
|
2924
2924
|
conditionTask?: outputs.GetJobJobSettingsSettingsTaskConditionTask;
|
|
2925
|
+
dashboardTask?: outputs.GetJobJobSettingsSettingsTaskDashboardTask;
|
|
2925
2926
|
dbtTask?: outputs.GetJobJobSettingsSettingsTaskDbtTask;
|
|
2926
2927
|
dependsOns?: outputs.GetJobJobSettingsSettingsTaskDependsOn[];
|
|
2927
2928
|
description?: string;
|
|
@@ -2938,6 +2939,7 @@ export interface GetJobJobSettingsSettingsTask {
|
|
|
2938
2939
|
notebookTask?: outputs.GetJobJobSettingsSettingsTaskNotebookTask;
|
|
2939
2940
|
notificationSettings?: outputs.GetJobJobSettingsSettingsTaskNotificationSettings;
|
|
2940
2941
|
pipelineTask?: outputs.GetJobJobSettingsSettingsTaskPipelineTask;
|
|
2942
|
+
powerBiTask?: outputs.GetJobJobSettingsSettingsTaskPowerBiTask;
|
|
2941
2943
|
pythonWheelTask?: outputs.GetJobJobSettingsSettingsTaskPythonWheelTask;
|
|
2942
2944
|
retryOnTimeout: boolean;
|
|
2943
2945
|
runIf?: string;
|
|
@@ -2955,6 +2957,20 @@ export interface GetJobJobSettingsSettingsTaskConditionTask {
|
|
|
2955
2957
|
op: string;
|
|
2956
2958
|
right: string;
|
|
2957
2959
|
}
|
|
2960
|
+
export interface GetJobJobSettingsSettingsTaskDashboardTask {
|
|
2961
|
+
dashboardId?: string;
|
|
2962
|
+
subscription?: outputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscription;
|
|
2963
|
+
warehouseId?: string;
|
|
2964
|
+
}
|
|
2965
|
+
export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscription {
|
|
2966
|
+
customSubject?: string;
|
|
2967
|
+
paused?: boolean;
|
|
2968
|
+
subscribers?: outputs.GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber[];
|
|
2969
|
+
}
|
|
2970
|
+
export interface GetJobJobSettingsSettingsTaskDashboardTaskSubscriptionSubscriber {
|
|
2971
|
+
destinationId?: string;
|
|
2972
|
+
userName?: string;
|
|
2973
|
+
}
|
|
2958
2974
|
export interface GetJobJobSettingsSettingsTaskDbtTask {
|
|
2959
2975
|
catalog?: string;
|
|
2960
2976
|
commands: string[];
|
|
@@ -2983,6 +2999,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTask {
|
|
|
2983
2999
|
}
|
|
2984
3000
|
export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
|
|
2985
3001
|
conditionTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTask;
|
|
3002
|
+
dashboardTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask;
|
|
2986
3003
|
dbtTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask;
|
|
2987
3004
|
dependsOns?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDependsOn[];
|
|
2988
3005
|
description?: string;
|
|
@@ -2998,6 +3015,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTask {
|
|
|
2998
3015
|
notebookTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotebookTask;
|
|
2999
3016
|
notificationSettings?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNotificationSettings;
|
|
3000
3017
|
pipelineTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTask;
|
|
3018
|
+
powerBiTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask;
|
|
3001
3019
|
pythonWheelTask?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask;
|
|
3002
3020
|
retryOnTimeout: boolean;
|
|
3003
3021
|
runIf?: string;
|
|
@@ -3015,6 +3033,20 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskConditionTask {
|
|
|
3015
3033
|
op: string;
|
|
3016
3034
|
right: string;
|
|
3017
3035
|
}
|
|
3036
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTask {
|
|
3037
|
+
dashboardId?: string;
|
|
3038
|
+
subscription?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription;
|
|
3039
|
+
warehouseId?: string;
|
|
3040
|
+
}
|
|
3041
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscription {
|
|
3042
|
+
customSubject?: string;
|
|
3043
|
+
paused?: boolean;
|
|
3044
|
+
subscribers?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber[];
|
|
3045
|
+
}
|
|
3046
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
|
|
3047
|
+
destinationId?: string;
|
|
3048
|
+
userName?: string;
|
|
3049
|
+
}
|
|
3018
3050
|
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskDbtTask {
|
|
3019
3051
|
catalog?: string;
|
|
3020
3052
|
commands: string[];
|
|
@@ -3100,7 +3132,7 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskNewCluster {
|
|
|
3100
3132
|
sparkEnvVars?: {
|
|
3101
3133
|
[key: string]: string;
|
|
3102
3134
|
};
|
|
3103
|
-
sparkVersion
|
|
3135
|
+
sparkVersion?: string;
|
|
3104
3136
|
sshPublicKeys?: string[];
|
|
3105
3137
|
workloadType?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskNewClusterWorkloadType;
|
|
3106
3138
|
}
|
|
@@ -3224,6 +3256,29 @@ export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPipelineTask {
|
|
|
3224
3256
|
fullRefresh?: boolean;
|
|
3225
3257
|
pipelineId: string;
|
|
3226
3258
|
}
|
|
3259
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTask {
|
|
3260
|
+
connectionResourceName?: string;
|
|
3261
|
+
powerBiModel?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel;
|
|
3262
|
+
refreshAfterUpdate?: boolean;
|
|
3263
|
+
tables?: outputs.GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable[];
|
|
3264
|
+
warehouseId?: string;
|
|
3265
|
+
}
|
|
3266
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskPowerBiModel {
|
|
3267
|
+
authenticationMethod?: string;
|
|
3268
|
+
modelName?: string;
|
|
3269
|
+
overwriteExisting?: boolean;
|
|
3270
|
+
storageMode?: string;
|
|
3271
|
+
workspaceName?: string;
|
|
3272
|
+
}
|
|
3273
|
+
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPowerBiTaskTable {
|
|
3274
|
+
catalog?: string;
|
|
3275
|
+
/**
|
|
3276
|
+
* the job name of databricks.Job if the resource was matched by id.
|
|
3277
|
+
*/
|
|
3278
|
+
name?: string;
|
|
3279
|
+
schema?: string;
|
|
3280
|
+
storageMode?: string;
|
|
3281
|
+
}
|
|
3227
3282
|
export interface GetJobJobSettingsSettingsTaskForEachTaskTaskPythonWheelTask {
|
|
3228
3283
|
entryPoint?: string;
|
|
3229
3284
|
namedParameters?: {
|
|
@@ -3388,7 +3443,7 @@ export interface GetJobJobSettingsSettingsTaskNewCluster {
|
|
|
3388
3443
|
sparkEnvVars?: {
|
|
3389
3444
|
[key: string]: string;
|
|
3390
3445
|
};
|
|
3391
|
-
sparkVersion
|
|
3446
|
+
sparkVersion?: string;
|
|
3392
3447
|
sshPublicKeys?: string[];
|
|
3393
3448
|
workloadType?: outputs.GetJobJobSettingsSettingsTaskNewClusterWorkloadType;
|
|
3394
3449
|
}
|
|
@@ -3512,6 +3567,29 @@ export interface GetJobJobSettingsSettingsTaskPipelineTask {
|
|
|
3512
3567
|
fullRefresh?: boolean;
|
|
3513
3568
|
pipelineId: string;
|
|
3514
3569
|
}
|
|
3570
|
+
export interface GetJobJobSettingsSettingsTaskPowerBiTask {
|
|
3571
|
+
connectionResourceName?: string;
|
|
3572
|
+
powerBiModel?: outputs.GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel;
|
|
3573
|
+
refreshAfterUpdate?: boolean;
|
|
3574
|
+
tables?: outputs.GetJobJobSettingsSettingsTaskPowerBiTaskTable[];
|
|
3575
|
+
warehouseId?: string;
|
|
3576
|
+
}
|
|
3577
|
+
export interface GetJobJobSettingsSettingsTaskPowerBiTaskPowerBiModel {
|
|
3578
|
+
authenticationMethod?: string;
|
|
3579
|
+
modelName?: string;
|
|
3580
|
+
overwriteExisting?: boolean;
|
|
3581
|
+
storageMode?: string;
|
|
3582
|
+
workspaceName?: string;
|
|
3583
|
+
}
|
|
3584
|
+
export interface GetJobJobSettingsSettingsTaskPowerBiTaskTable {
|
|
3585
|
+
catalog?: string;
|
|
3586
|
+
/**
|
|
3587
|
+
* the job name of databricks.Job if the resource was matched by id.
|
|
3588
|
+
*/
|
|
3589
|
+
name?: string;
|
|
3590
|
+
schema?: string;
|
|
3591
|
+
storageMode?: string;
|
|
3592
|
+
}
|
|
3515
3593
|
export interface GetJobJobSettingsSettingsTaskPythonWheelTask {
|
|
3516
3594
|
entryPoint?: string;
|
|
3517
3595
|
namedParameters?: {
|
|
@@ -4877,8 +4955,6 @@ export interface JobEmailNotifications {
|
|
|
4877
4955
|
noAlertForSkippedRuns?: boolean;
|
|
4878
4956
|
/**
|
|
4879
4957
|
* (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
4880
|
-
*
|
|
4881
|
-
* The following parameter is only available for the job level configuration.
|
|
4882
4958
|
*/
|
|
4883
4959
|
onDurationWarningThresholdExceededs?: string[];
|
|
4884
4960
|
/**
|
|
@@ -4889,6 +4965,11 @@ export interface JobEmailNotifications {
|
|
|
4889
4965
|
* (List) list of emails to notify when the run starts.
|
|
4890
4966
|
*/
|
|
4891
4967
|
onStarts?: string[];
|
|
4968
|
+
/**
|
|
4969
|
+
* (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
|
|
4970
|
+
*
|
|
4971
|
+
* The following parameter is only available for the job level configuration.
|
|
4972
|
+
*/
|
|
4892
4973
|
onStreamingBacklogExceededs?: string[];
|
|
4893
4974
|
/**
|
|
4894
4975
|
* (List) list of emails to notify when the run completes successfully.
|
|
@@ -5017,7 +5098,7 @@ export interface JobJobClusterNewCluster {
|
|
|
5017
5098
|
sparkEnvVars?: {
|
|
5018
5099
|
[key: string]: string;
|
|
5019
5100
|
};
|
|
5020
|
-
sparkVersion
|
|
5101
|
+
sparkVersion?: string;
|
|
5021
5102
|
sshPublicKeys?: string[];
|
|
5022
5103
|
useMlRuntime?: boolean;
|
|
5023
5104
|
/**
|
|
@@ -5231,7 +5312,7 @@ export interface JobNewCluster {
|
|
|
5231
5312
|
sparkEnvVars?: {
|
|
5232
5313
|
[key: string]: string;
|
|
5233
5314
|
};
|
|
5234
|
-
sparkVersion
|
|
5315
|
+
sparkVersion?: string;
|
|
5235
5316
|
sshPublicKeys?: string[];
|
|
5236
5317
|
useMlRuntime?: boolean;
|
|
5237
5318
|
/**
|
|
@@ -5530,11 +5611,13 @@ export interface JobSparkPythonTask {
|
|
|
5530
5611
|
*/
|
|
5531
5612
|
parameters?: string[];
|
|
5532
5613
|
/**
|
|
5533
|
-
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with
|
|
5614
|
+
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
|
|
5534
5615
|
*/
|
|
5535
5616
|
pythonFile: string;
|
|
5536
5617
|
/**
|
|
5537
|
-
* Location type of the Python file,
|
|
5618
|
+
* Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
|
|
5619
|
+
* * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
|
|
5620
|
+
* * `GIT`: The Python file is located in a remote Git repository.
|
|
5538
5621
|
*/
|
|
5539
5622
|
source?: string;
|
|
5540
5623
|
}
|
|
@@ -5674,24 +5757,39 @@ export interface JobTaskConditionTask {
|
|
|
5674
5757
|
}
|
|
5675
5758
|
export interface JobTaskDashboardTask {
|
|
5676
5759
|
/**
|
|
5677
|
-
*
|
|
5760
|
+
* The identifier of the dashboard to refresh
|
|
5678
5761
|
*/
|
|
5679
5762
|
dashboardId?: string;
|
|
5763
|
+
/**
|
|
5764
|
+
* Represents a subscription configuration for scheduled dashboard snapshots.
|
|
5765
|
+
*/
|
|
5680
5766
|
subscription?: outputs.JobTaskDashboardTaskSubscription;
|
|
5767
|
+
/**
|
|
5768
|
+
* The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
|
|
5769
|
+
*/
|
|
5681
5770
|
warehouseId?: string;
|
|
5682
5771
|
}
|
|
5683
5772
|
export interface JobTaskDashboardTaskSubscription {
|
|
5684
5773
|
/**
|
|
5685
|
-
*
|
|
5774
|
+
* Allows users to specify a custom subject line on the email sent to subscribers.
|
|
5686
5775
|
*/
|
|
5687
5776
|
customSubject?: string;
|
|
5777
|
+
/**
|
|
5778
|
+
* When true, the subscription will not send emails.
|
|
5779
|
+
*/
|
|
5688
5780
|
paused?: boolean;
|
|
5781
|
+
/**
|
|
5782
|
+
* The list of subscribers to send the snapshot of the dashboard to.
|
|
5783
|
+
*/
|
|
5689
5784
|
subscribers?: outputs.JobTaskDashboardTaskSubscriptionSubscriber[];
|
|
5690
5785
|
}
|
|
5691
5786
|
export interface JobTaskDashboardTaskSubscriptionSubscriber {
|
|
5787
|
+
/**
|
|
5788
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
5789
|
+
*/
|
|
5692
5790
|
destinationId?: string;
|
|
5693
5791
|
/**
|
|
5694
|
-
*
|
|
5792
|
+
* A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
|
|
5695
5793
|
*/
|
|
5696
5794
|
userName?: string;
|
|
5697
5795
|
}
|
|
@@ -5748,8 +5846,6 @@ export interface JobTaskEmailNotifications {
|
|
|
5748
5846
|
noAlertForSkippedRuns?: boolean;
|
|
5749
5847
|
/**
|
|
5750
5848
|
* (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
5751
|
-
*
|
|
5752
|
-
* The following parameter is only available for the job level configuration.
|
|
5753
5849
|
*/
|
|
5754
5850
|
onDurationWarningThresholdExceededs?: string[];
|
|
5755
5851
|
/**
|
|
@@ -5760,6 +5856,11 @@ export interface JobTaskEmailNotifications {
|
|
|
5760
5856
|
* (List) list of emails to notify when the run starts.
|
|
5761
5857
|
*/
|
|
5762
5858
|
onStarts?: string[];
|
|
5859
|
+
/**
|
|
5860
|
+
* (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
|
|
5861
|
+
*
|
|
5862
|
+
* The following parameter is only available for the job level configuration.
|
|
5863
|
+
*/
|
|
5763
5864
|
onStreamingBacklogExceededs?: string[];
|
|
5764
5865
|
/**
|
|
5765
5866
|
* (List) list of emails to notify when the run completes successfully.
|
|
@@ -5909,24 +6010,39 @@ export interface JobTaskForEachTaskTaskConditionTask {
|
|
|
5909
6010
|
}
|
|
5910
6011
|
export interface JobTaskForEachTaskTaskDashboardTask {
|
|
5911
6012
|
/**
|
|
5912
|
-
*
|
|
6013
|
+
* The identifier of the dashboard to refresh
|
|
5913
6014
|
*/
|
|
5914
6015
|
dashboardId?: string;
|
|
6016
|
+
/**
|
|
6017
|
+
* Represents a subscription configuration for scheduled dashboard snapshots.
|
|
6018
|
+
*/
|
|
5915
6019
|
subscription?: outputs.JobTaskForEachTaskTaskDashboardTaskSubscription;
|
|
6020
|
+
/**
|
|
6021
|
+
* The warehouse id to execute the dashboard with for the schedule. If not specified, will use the default warehouse of dashboard
|
|
6022
|
+
*/
|
|
5916
6023
|
warehouseId?: string;
|
|
5917
6024
|
}
|
|
5918
6025
|
export interface JobTaskForEachTaskTaskDashboardTaskSubscription {
|
|
5919
6026
|
/**
|
|
5920
|
-
*
|
|
6027
|
+
* Allows users to specify a custom subject line on the email sent to subscribers.
|
|
5921
6028
|
*/
|
|
5922
6029
|
customSubject?: string;
|
|
6030
|
+
/**
|
|
6031
|
+
* When true, the subscription will not send emails.
|
|
6032
|
+
*/
|
|
5923
6033
|
paused?: boolean;
|
|
6034
|
+
/**
|
|
6035
|
+
* The list of subscribers to send the snapshot of the dashboard to.
|
|
6036
|
+
*/
|
|
5924
6037
|
subscribers?: outputs.JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber[];
|
|
5925
6038
|
}
|
|
5926
6039
|
export interface JobTaskForEachTaskTaskDashboardTaskSubscriptionSubscriber {
|
|
6040
|
+
/**
|
|
6041
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
6042
|
+
*/
|
|
5927
6043
|
destinationId?: string;
|
|
5928
6044
|
/**
|
|
5929
|
-
*
|
|
6045
|
+
* A snapshot of the dashboard will be sent to the user's email when the `userName` field is present.
|
|
5930
6046
|
*/
|
|
5931
6047
|
userName?: string;
|
|
5932
6048
|
}
|
|
@@ -5983,8 +6099,6 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
|
|
|
5983
6099
|
noAlertForSkippedRuns?: boolean;
|
|
5984
6100
|
/**
|
|
5985
6101
|
* (List) list of emails to notify when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
5986
|
-
*
|
|
5987
|
-
* The following parameter is only available for the job level configuration.
|
|
5988
6102
|
*/
|
|
5989
6103
|
onDurationWarningThresholdExceededs?: string[];
|
|
5990
6104
|
/**
|
|
@@ -5995,6 +6109,11 @@ export interface JobTaskForEachTaskTaskEmailNotifications {
|
|
|
5995
6109
|
* (List) list of emails to notify when the run starts.
|
|
5996
6110
|
*/
|
|
5997
6111
|
onStarts?: string[];
|
|
6112
|
+
/**
|
|
6113
|
+
* (List) list of emails to notify when any streaming backlog thresholds are exceeded for any stream.
|
|
6114
|
+
*
|
|
6115
|
+
* The following parameter is only available for the job level configuration.
|
|
6116
|
+
*/
|
|
5998
6117
|
onStreamingBacklogExceededs?: string[];
|
|
5999
6118
|
/**
|
|
6000
6119
|
* (List) list of emails to notify when the run completes successfully.
|
|
@@ -6097,7 +6216,7 @@ export interface JobTaskForEachTaskTaskNewCluster {
|
|
|
6097
6216
|
sparkEnvVars?: {
|
|
6098
6217
|
[key: string]: string;
|
|
6099
6218
|
};
|
|
6100
|
-
sparkVersion
|
|
6219
|
+
sparkVersion?: string;
|
|
6101
6220
|
sshPublicKeys?: string[];
|
|
6102
6221
|
useMlRuntime?: boolean;
|
|
6103
6222
|
/**
|
|
@@ -6299,32 +6418,65 @@ export interface JobTaskForEachTaskTaskPipelineTask {
|
|
|
6299
6418
|
pipelineId: string;
|
|
6300
6419
|
}
|
|
6301
6420
|
export interface JobTaskForEachTaskTaskPowerBiTask {
|
|
6421
|
+
/**
|
|
6422
|
+
* The resource name of the UC connection to authenticate from Databricks to Power BI
|
|
6423
|
+
*/
|
|
6302
6424
|
connectionResourceName?: string;
|
|
6425
|
+
/**
|
|
6426
|
+
* The semantic model to update. Block consists of following fields:
|
|
6427
|
+
*/
|
|
6303
6428
|
powerBiModel?: outputs.JobTaskForEachTaskTaskPowerBiTaskPowerBiModel;
|
|
6429
|
+
/**
|
|
6430
|
+
* Whether the model should be refreshed after the update. Default is false
|
|
6431
|
+
*/
|
|
6304
6432
|
refreshAfterUpdate?: boolean;
|
|
6433
|
+
/**
|
|
6434
|
+
* The tables to be exported to Power BI. Block consists of following fields:
|
|
6435
|
+
*/
|
|
6305
6436
|
tables?: outputs.JobTaskForEachTaskTaskPowerBiTaskTable[];
|
|
6437
|
+
/**
|
|
6438
|
+
* The SQL warehouse ID to use as the Power BI data source
|
|
6439
|
+
*/
|
|
6306
6440
|
warehouseId?: string;
|
|
6307
6441
|
}
|
|
6308
6442
|
export interface JobTaskForEachTaskTaskPowerBiTaskPowerBiModel {
|
|
6443
|
+
/**
|
|
6444
|
+
* How the published Power BI model authenticates to Databricks
|
|
6445
|
+
*/
|
|
6309
6446
|
authenticationMethod?: string;
|
|
6447
|
+
/**
|
|
6448
|
+
* The name of the Power BI model
|
|
6449
|
+
*/
|
|
6310
6450
|
modelName?: string;
|
|
6451
|
+
/**
|
|
6452
|
+
* Whether to overwrite existing Power BI models. Default is false
|
|
6453
|
+
*/
|
|
6311
6454
|
overwriteExisting?: boolean;
|
|
6455
|
+
/**
|
|
6456
|
+
* The default storage mode of the Power BI model
|
|
6457
|
+
*/
|
|
6312
6458
|
storageMode?: string;
|
|
6459
|
+
/**
|
|
6460
|
+
* The name of the Power BI workspace of the model
|
|
6461
|
+
*/
|
|
6313
6462
|
workspaceName?: string;
|
|
6314
6463
|
}
|
|
6315
6464
|
export interface JobTaskForEachTaskTaskPowerBiTaskTable {
|
|
6316
6465
|
/**
|
|
6317
|
-
* The
|
|
6466
|
+
* The catalog name in Databricks
|
|
6318
6467
|
*/
|
|
6319
6468
|
catalog?: string;
|
|
6320
6469
|
/**
|
|
6321
|
-
*
|
|
6470
|
+
* The table name in Databricks. If empty, all tables under the schema are selected.
|
|
6322
6471
|
*/
|
|
6323
6472
|
name?: string;
|
|
6324
6473
|
/**
|
|
6325
|
-
* The
|
|
6474
|
+
* The schema name in Databricks
|
|
6326
6475
|
*/
|
|
6327
6476
|
schema?: string;
|
|
6477
|
+
/**
|
|
6478
|
+
* The Power BI storage mode of the table
|
|
6479
|
+
*/
|
|
6328
6480
|
storageMode?: string;
|
|
6329
6481
|
}
|
|
6330
6482
|
export interface JobTaskForEachTaskTaskPythonWheelTask {
|
|
@@ -6399,11 +6551,13 @@ export interface JobTaskForEachTaskTaskSparkPythonTask {
|
|
|
6399
6551
|
*/
|
|
6400
6552
|
parameters?: string[];
|
|
6401
6553
|
/**
|
|
6402
|
-
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with
|
|
6554
|
+
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
|
|
6403
6555
|
*/
|
|
6404
6556
|
pythonFile: string;
|
|
6405
6557
|
/**
|
|
6406
|
-
* Location type of the Python file,
|
|
6558
|
+
* Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
|
|
6559
|
+
* * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
|
|
6560
|
+
* * `GIT`: The Python file is located in a remote Git repository.
|
|
6407
6561
|
*/
|
|
6408
6562
|
source?: string;
|
|
6409
6563
|
}
|
|
@@ -6456,10 +6610,10 @@ export interface JobTaskForEachTaskTaskSqlTaskAlert {
|
|
|
6456
6610
|
subscriptions?: outputs.JobTaskForEachTaskTaskSqlTaskAlertSubscription[];
|
|
6457
6611
|
}
|
|
6458
6612
|
export interface JobTaskForEachTaskTaskSqlTaskAlertSubscription {
|
|
6459
|
-
destinationId?: string;
|
|
6460
6613
|
/**
|
|
6461
|
-
*
|
|
6614
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
6462
6615
|
*/
|
|
6616
|
+
destinationId?: string;
|
|
6463
6617
|
userName?: string;
|
|
6464
6618
|
}
|
|
6465
6619
|
export interface JobTaskForEachTaskTaskSqlTaskDashboard {
|
|
@@ -6481,10 +6635,10 @@ export interface JobTaskForEachTaskTaskSqlTaskDashboard {
|
|
|
6481
6635
|
subscriptions?: outputs.JobTaskForEachTaskTaskSqlTaskDashboardSubscription[];
|
|
6482
6636
|
}
|
|
6483
6637
|
export interface JobTaskForEachTaskTaskSqlTaskDashboardSubscription {
|
|
6484
|
-
destinationId?: string;
|
|
6485
6638
|
/**
|
|
6486
|
-
*
|
|
6639
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
6487
6640
|
*/
|
|
6641
|
+
destinationId?: string;
|
|
6488
6642
|
userName?: string;
|
|
6489
6643
|
}
|
|
6490
6644
|
export interface JobTaskForEachTaskTaskSqlTaskFile {
|
|
@@ -6549,10 +6703,6 @@ export interface JobTaskForEachTaskTaskSqlTaskQuery {
|
|
|
6549
6703
|
export interface JobTaskForEachTaskTaskWebhookNotifications {
|
|
6550
6704
|
/**
|
|
6551
6705
|
* (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
6552
|
-
*
|
|
6553
|
-
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
6554
|
-
*
|
|
6555
|
-
* Example
|
|
6556
6706
|
*/
|
|
6557
6707
|
onDurationWarningThresholdExceededs?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnDurationWarningThresholdExceeded[];
|
|
6558
6708
|
/**
|
|
@@ -6563,6 +6713,13 @@ export interface JobTaskForEachTaskTaskWebhookNotifications {
|
|
|
6563
6713
|
* (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
|
|
6564
6714
|
*/
|
|
6565
6715
|
onStarts?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnStart[];
|
|
6716
|
+
/**
|
|
6717
|
+
* (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
|
|
6718
|
+
*
|
|
6719
|
+
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
6720
|
+
*
|
|
6721
|
+
* Example
|
|
6722
|
+
*/
|
|
6566
6723
|
onStreamingBacklogExceededs?: outputs.JobTaskForEachTaskTaskWebhookNotificationsOnStreamingBacklogExceeded[];
|
|
6567
6724
|
/**
|
|
6568
6725
|
* (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
|
|
@@ -6695,7 +6852,7 @@ export interface JobTaskNewCluster {
|
|
|
6695
6852
|
sparkEnvVars?: {
|
|
6696
6853
|
[key: string]: string;
|
|
6697
6854
|
};
|
|
6698
|
-
sparkVersion
|
|
6855
|
+
sparkVersion?: string;
|
|
6699
6856
|
sshPublicKeys?: string[];
|
|
6700
6857
|
useMlRuntime?: boolean;
|
|
6701
6858
|
/**
|
|
@@ -6897,32 +7054,65 @@ export interface JobTaskPipelineTask {
|
|
|
6897
7054
|
pipelineId: string;
|
|
6898
7055
|
}
|
|
6899
7056
|
export interface JobTaskPowerBiTask {
|
|
7057
|
+
/**
|
|
7058
|
+
* The resource name of the UC connection to authenticate from Databricks to Power BI
|
|
7059
|
+
*/
|
|
6900
7060
|
connectionResourceName?: string;
|
|
7061
|
+
/**
|
|
7062
|
+
* The semantic model to update. Block consists of following fields:
|
|
7063
|
+
*/
|
|
6901
7064
|
powerBiModel?: outputs.JobTaskPowerBiTaskPowerBiModel;
|
|
7065
|
+
/**
|
|
7066
|
+
* Whether the model should be refreshed after the update. Default is false
|
|
7067
|
+
*/
|
|
6902
7068
|
refreshAfterUpdate?: boolean;
|
|
7069
|
+
/**
|
|
7070
|
+
* The tables to be exported to Power BI. Block consists of following fields:
|
|
7071
|
+
*/
|
|
6903
7072
|
tables?: outputs.JobTaskPowerBiTaskTable[];
|
|
7073
|
+
/**
|
|
7074
|
+
* The SQL warehouse ID to use as the Power BI data source
|
|
7075
|
+
*/
|
|
6904
7076
|
warehouseId?: string;
|
|
6905
7077
|
}
|
|
6906
7078
|
export interface JobTaskPowerBiTaskPowerBiModel {
|
|
7079
|
+
/**
|
|
7080
|
+
* How the published Power BI model authenticates to Databricks
|
|
7081
|
+
*/
|
|
6907
7082
|
authenticationMethod?: string;
|
|
7083
|
+
/**
|
|
7084
|
+
* The name of the Power BI model
|
|
7085
|
+
*/
|
|
6908
7086
|
modelName?: string;
|
|
7087
|
+
/**
|
|
7088
|
+
* Whether to overwrite existing Power BI models. Default is false
|
|
7089
|
+
*/
|
|
6909
7090
|
overwriteExisting?: boolean;
|
|
7091
|
+
/**
|
|
7092
|
+
* The default storage mode of the Power BI model
|
|
7093
|
+
*/
|
|
6910
7094
|
storageMode?: string;
|
|
7095
|
+
/**
|
|
7096
|
+
* The name of the Power BI workspace of the model
|
|
7097
|
+
*/
|
|
6911
7098
|
workspaceName?: string;
|
|
6912
7099
|
}
|
|
6913
7100
|
export interface JobTaskPowerBiTaskTable {
|
|
6914
7101
|
/**
|
|
6915
|
-
* The
|
|
7102
|
+
* The catalog name in Databricks
|
|
6916
7103
|
*/
|
|
6917
7104
|
catalog?: string;
|
|
6918
7105
|
/**
|
|
6919
|
-
*
|
|
7106
|
+
* The table name in Databricks. If empty, all tables under the schema are selected.
|
|
6920
7107
|
*/
|
|
6921
7108
|
name?: string;
|
|
6922
7109
|
/**
|
|
6923
|
-
* The
|
|
7110
|
+
* The schema name in Databricks
|
|
6924
7111
|
*/
|
|
6925
7112
|
schema?: string;
|
|
7113
|
+
/**
|
|
7114
|
+
* The Power BI storage mode of the table
|
|
7115
|
+
*/
|
|
6926
7116
|
storageMode?: string;
|
|
6927
7117
|
}
|
|
6928
7118
|
export interface JobTaskPythonWheelTask {
|
|
@@ -6997,11 +7187,13 @@ export interface JobTaskSparkPythonTask {
|
|
|
6997
7187
|
*/
|
|
6998
7188
|
parameters?: string[];
|
|
6999
7189
|
/**
|
|
7000
|
-
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with
|
|
7190
|
+
* The URI of the Python file to be executed. databricks_dbfs_file, cloud file URIs (e.g. `s3:/`, `abfss:/`, `gs:/`), workspace paths and remote repository are supported. For Python files stored in the Databricks workspace, the path must be absolute and begin with `/`. For files stored in a remote repository, the path must be relative. This field is required.
|
|
7001
7191
|
*/
|
|
7002
7192
|
pythonFile: string;
|
|
7003
7193
|
/**
|
|
7004
|
-
* Location type of the Python file,
|
|
7194
|
+
* Location type of the Python file. When set to `WORKSPACE` or not specified, the file will be retrieved from the local Databricks workspace or cloud location (if the pythonFile has a URI format). When set to `GIT`, the Python file will be retrieved from a Git repository defined in `gitSource`.
|
|
7195
|
+
* * `WORKSPACE`: The Python file is located in a Databricks workspace or at a cloud filesystem URI.
|
|
7196
|
+
* * `GIT`: The Python file is located in a remote Git repository.
|
|
7005
7197
|
*/
|
|
7006
7198
|
source?: string;
|
|
7007
7199
|
}
|
|
@@ -7054,10 +7246,10 @@ export interface JobTaskSqlTaskAlert {
|
|
|
7054
7246
|
subscriptions?: outputs.JobTaskSqlTaskAlertSubscription[];
|
|
7055
7247
|
}
|
|
7056
7248
|
export interface JobTaskSqlTaskAlertSubscription {
|
|
7057
|
-
destinationId?: string;
|
|
7058
7249
|
/**
|
|
7059
|
-
*
|
|
7250
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
7060
7251
|
*/
|
|
7252
|
+
destinationId?: string;
|
|
7061
7253
|
userName?: string;
|
|
7062
7254
|
}
|
|
7063
7255
|
export interface JobTaskSqlTaskDashboard {
|
|
@@ -7079,10 +7271,10 @@ export interface JobTaskSqlTaskDashboard {
|
|
|
7079
7271
|
subscriptions?: outputs.JobTaskSqlTaskDashboardSubscription[];
|
|
7080
7272
|
}
|
|
7081
7273
|
export interface JobTaskSqlTaskDashboardSubscription {
|
|
7082
|
-
destinationId?: string;
|
|
7083
7274
|
/**
|
|
7084
|
-
*
|
|
7275
|
+
* A snapshot of the dashboard will be sent to the destination when the `destinationId` field is present.
|
|
7085
7276
|
*/
|
|
7277
|
+
destinationId?: string;
|
|
7086
7278
|
userName?: string;
|
|
7087
7279
|
}
|
|
7088
7280
|
export interface JobTaskSqlTaskFile {
|
|
@@ -7147,10 +7339,6 @@ export interface JobTaskSqlTaskQuery {
|
|
|
7147
7339
|
export interface JobTaskWebhookNotifications {
|
|
7148
7340
|
/**
|
|
7149
7341
|
* (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
7150
|
-
*
|
|
7151
|
-
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
7152
|
-
*
|
|
7153
|
-
* Example
|
|
7154
7342
|
*/
|
|
7155
7343
|
onDurationWarningThresholdExceededs?: outputs.JobTaskWebhookNotificationsOnDurationWarningThresholdExceeded[];
|
|
7156
7344
|
/**
|
|
@@ -7161,6 +7349,13 @@ export interface JobTaskWebhookNotifications {
|
|
|
7161
7349
|
* (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
|
|
7162
7350
|
*/
|
|
7163
7351
|
onStarts?: outputs.JobTaskWebhookNotificationsOnStart[];
|
|
7352
|
+
/**
|
|
7353
|
+
* (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
|
|
7354
|
+
*
|
|
7355
|
+
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
7356
|
+
*
|
|
7357
|
+
* Example
|
|
7358
|
+
*/
|
|
7164
7359
|
onStreamingBacklogExceededs?: outputs.JobTaskWebhookNotificationsOnStreamingBacklogExceeded[];
|
|
7165
7360
|
/**
|
|
7166
7361
|
* (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
|
|
@@ -7264,10 +7459,6 @@ export interface JobTriggerTableUpdate {
|
|
|
7264
7459
|
export interface JobWebhookNotifications {
|
|
7265
7460
|
/**
|
|
7266
7461
|
* (List) list of notification IDs to call when the duration of a run exceeds the threshold specified by the `RUN_DURATION_SECONDS` metric in the `health` block.
|
|
7267
|
-
*
|
|
7268
|
-
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
7269
|
-
*
|
|
7270
|
-
* Example
|
|
7271
7462
|
*/
|
|
7272
7463
|
onDurationWarningThresholdExceededs?: outputs.JobWebhookNotificationsOnDurationWarningThresholdExceeded[];
|
|
7273
7464
|
/**
|
|
@@ -7278,6 +7469,13 @@ export interface JobWebhookNotifications {
|
|
|
7278
7469
|
* (List) list of notification IDs to call when the run starts. A maximum of 3 destinations can be specified.
|
|
7279
7470
|
*/
|
|
7280
7471
|
onStarts?: outputs.JobWebhookNotificationsOnStart[];
|
|
7472
|
+
/**
|
|
7473
|
+
* (List) list of notification IDs to call when any streaming backlog thresholds are exceeded for any stream.
|
|
7474
|
+
*
|
|
7475
|
+
* Note that the `id` is not to be confused with the name of the alert destination. The `id` can be retrieved through the API or the URL of Databricks UI `https://<workspace host>/sql/destinations/<notification id>?o=<workspace id>`
|
|
7476
|
+
*
|
|
7477
|
+
* Example
|
|
7478
|
+
*/
|
|
7281
7479
|
onStreamingBacklogExceededs?: outputs.JobWebhookNotificationsOnStreamingBacklogExceeded[];
|
|
7282
7480
|
/**
|
|
7283
7481
|
* (List) list of notification IDs to call when the run completes successfully. A maximum of 3 destinations can be specified.
|
|
@@ -7890,7 +8088,7 @@ export interface ModelServingConfigServedEntityExternalModelOpenaiConfig {
|
|
|
7890
8088
|
*/
|
|
7891
8089
|
microsoftEntraTenantId?: string;
|
|
7892
8090
|
/**
|
|
7893
|
-
* This is the base URL for the OpenAI API (default: "https://api.openai.com/v1"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
|
|
8091
|
+
* This is the base URL for the OpenAI API (default: "<https://api.openai.com/v1>"). For Azure OpenAI, this field is required and is the base URL for the Azure OpenAI API service provided by Azure.
|
|
7894
8092
|
*/
|
|
7895
8093
|
openaiApiBase?: string;
|
|
7896
8094
|
/**
|
|
@@ -8137,13 +8335,13 @@ export interface MwsNetworksGcpNetworkInfo {
|
|
|
8137
8335
|
/**
|
|
8138
8336
|
* The name of the secondary IP range for pods. A Databricks-managed GKE cluster uses this IP range for its pods. This secondary IP range can only be used by one workspace.
|
|
8139
8337
|
*
|
|
8140
|
-
* @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.
|
|
8338
|
+
* @deprecated gcp_network_info.pod_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
|
|
8141
8339
|
*/
|
|
8142
8340
|
podIpRangeName?: string;
|
|
8143
8341
|
/**
|
|
8144
8342
|
* The name of the secondary IP range for services. A Databricks-managed GKE cluster uses this IP range for its services. This secondary IP range can only be used by one workspace.
|
|
8145
8343
|
*
|
|
8146
|
-
* @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.
|
|
8344
|
+
* @deprecated gcp_network_info.service_ip_range_name is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-vpc
|
|
8147
8345
|
*/
|
|
8148
8346
|
serviceIpRangeName?: string;
|
|
8149
8347
|
/**
|
|
@@ -8204,11 +8402,11 @@ export interface MwsWorkspacesExternalCustomerInfo {
|
|
|
8204
8402
|
}
|
|
8205
8403
|
export interface MwsWorkspacesGcpManagedNetworkConfig {
|
|
8206
8404
|
/**
|
|
8207
|
-
* @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.
|
|
8405
|
+
* @deprecated gcp_managed_network_config.gke_cluster_pod_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
|
|
8208
8406
|
*/
|
|
8209
8407
|
gkeClusterPodIpRange?: string;
|
|
8210
8408
|
/**
|
|
8211
|
-
* @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.
|
|
8409
|
+
* @deprecated gcp_managed_network_config.gke_cluster_service_ip_range is deprecated and will be removed in a future release. For more information, review the documentation at https://registry.terraform.io/providers/databricks/databricks/1.75.0/docs/guides/gcp-workspace#creating-a-databricks-workspace
|
|
8212
8410
|
*/
|
|
8213
8411
|
gkeClusterServiceIpRange?: string;
|
|
8214
8412
|
subnetCidr: string;
|