pulumi-databricks 1.74.0a1753335781__py3-none-any.whl → 1.77.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (252) hide show
  1. pulumi_databricks/__init__.py +185 -12
  2. pulumi_databricks/_inputs.py +30467 -20225
  3. pulumi_databricks/access_control_rule_set.py +28 -29
  4. pulumi_databricks/account_federation_policy.py +473 -0
  5. pulumi_databricks/account_network_policy.py +74 -43
  6. pulumi_databricks/account_setting_v2.py +759 -0
  7. pulumi_databricks/aibi_dashboard_embedding_access_policy_setting.py +30 -31
  8. pulumi_databricks/aibi_dashboard_embedding_approved_domains_setting.py +30 -31
  9. pulumi_databricks/alert.py +185 -186
  10. pulumi_databricks/alert_v2.py +369 -211
  11. pulumi_databricks/app.py +315 -209
  12. pulumi_databricks/apps_settings_custom_template.py +531 -0
  13. pulumi_databricks/artifact_allowlist.py +72 -73
  14. pulumi_databricks/automatic_cluster_update_workspace_setting.py +30 -31
  15. pulumi_databricks/budget.py +84 -85
  16. pulumi_databricks/budget_policy.py +75 -53
  17. pulumi_databricks/catalog.py +544 -225
  18. pulumi_databricks/catalog_workspace_binding.py +82 -83
  19. pulumi_databricks/cluster.py +561 -683
  20. pulumi_databricks/cluster_policy.py +115 -116
  21. pulumi_databricks/compliance_security_profile_workspace_setting.py +30 -31
  22. pulumi_databricks/config/__init__.py +1 -1
  23. pulumi_databricks/config/__init__.pyi +3 -2
  24. pulumi_databricks/config/vars.py +40 -37
  25. pulumi_databricks/connection.py +346 -211
  26. pulumi_databricks/credential.py +262 -263
  27. pulumi_databricks/custom_app_integration.py +175 -176
  28. pulumi_databricks/dashboard.py +207 -208
  29. pulumi_databricks/data_quality_monitor.py +453 -0
  30. pulumi_databricks/data_quality_refresh.py +492 -0
  31. pulumi_databricks/database_database_catalog.py +437 -0
  32. pulumi_databricks/database_instance.py +938 -137
  33. pulumi_databricks/database_synced_database_table.py +589 -0
  34. pulumi_databricks/dbfs_file.py +87 -84
  35. pulumi_databricks/default_namespace_setting.py +30 -31
  36. pulumi_databricks/directory.py +61 -62
  37. pulumi_databricks/disable_legacy_access_setting.py +109 -33
  38. pulumi_databricks/disable_legacy_dbfs_setting.py +58 -39
  39. pulumi_databricks/disable_legacy_features_setting.py +62 -37
  40. pulumi_databricks/enhanced_security_monitoring_workspace_setting.py +30 -31
  41. pulumi_databricks/entitlements.py +148 -149
  42. pulumi_databricks/entity_tag_assignment.py +409 -0
  43. pulumi_databricks/external_location.py +276 -290
  44. pulumi_databricks/external_metadata.py +726 -0
  45. pulumi_databricks/feature_engineering_feature.py +480 -0
  46. pulumi_databricks/feature_engineering_materialized_feature.py +397 -0
  47. pulumi_databricks/file.py +96 -97
  48. pulumi_databricks/get_account_federation_policies.py +127 -0
  49. pulumi_databricks/get_account_federation_policy.py +214 -0
  50. pulumi_databricks/get_account_network_policies.py +36 -7
  51. pulumi_databricks/get_account_network_policy.py +33 -29
  52. pulumi_databricks/get_account_setting_v2.py +331 -0
  53. pulumi_databricks/get_alert_v2.py +100 -84
  54. pulumi_databricks/get_alerts_v2.py +35 -17
  55. pulumi_databricks/get_app.py +31 -20
  56. pulumi_databricks/get_apps.py +26 -15
  57. pulumi_databricks/get_apps_settings_custom_template.py +207 -0
  58. pulumi_databricks/get_apps_settings_custom_templates.py +133 -0
  59. pulumi_databricks/get_aws_assume_role_policy.py +27 -28
  60. pulumi_databricks/get_aws_bucket_policy.py +39 -40
  61. pulumi_databricks/get_aws_cross_account_policy.py +47 -48
  62. pulumi_databricks/get_aws_unity_catalog_assume_role_policy.py +35 -36
  63. pulumi_databricks/get_aws_unity_catalog_policy.py +35 -36
  64. pulumi_databricks/get_budget_policies.py +74 -11
  65. pulumi_databricks/get_budget_policy.py +43 -39
  66. pulumi_databricks/get_catalog.py +35 -19
  67. pulumi_databricks/get_catalogs.py +32 -14
  68. pulumi_databricks/get_cluster.py +41 -25
  69. pulumi_databricks/get_cluster_policy.py +72 -54
  70. pulumi_databricks/get_clusters.py +41 -25
  71. pulumi_databricks/get_current_config.py +23 -24
  72. pulumi_databricks/get_current_metastore.py +29 -13
  73. pulumi_databricks/get_current_user.py +17 -18
  74. pulumi_databricks/get_dashboards.py +34 -15
  75. pulumi_databricks/get_data_quality_monitor.py +210 -0
  76. pulumi_databricks/get_data_quality_monitors.py +143 -0
  77. pulumi_databricks/get_data_quality_refresh.py +270 -0
  78. pulumi_databricks/get_data_quality_refreshes.py +207 -0
  79. pulumi_databricks/get_database_database_catalog.py +176 -0
  80. pulumi_databricks/get_database_database_catalogs.py +120 -0
  81. pulumi_databricks/get_database_instance.py +274 -50
  82. pulumi_databricks/get_database_instances.py +35 -11
  83. pulumi_databricks/get_database_synced_database_table.py +225 -0
  84. pulumi_databricks/get_database_synced_database_tables.py +120 -0
  85. pulumi_databricks/get_dbfs_file.py +19 -20
  86. pulumi_databricks/get_dbfs_file_paths.py +16 -17
  87. pulumi_databricks/get_directory.py +43 -25
  88. pulumi_databricks/get_entity_tag_assignment.py +202 -0
  89. pulumi_databricks/get_entity_tag_assignments.py +187 -0
  90. pulumi_databricks/get_external_location.py +35 -19
  91. pulumi_databricks/get_external_locations.py +32 -14
  92. pulumi_databricks/get_external_metadata.py +292 -0
  93. pulumi_databricks/get_external_metadatas.py +135 -0
  94. pulumi_databricks/get_feature_engineering_feature.py +179 -0
  95. pulumi_databricks/get_feature_engineering_features.py +103 -0
  96. pulumi_databricks/get_feature_engineering_materialized_feature.py +180 -0
  97. pulumi_databricks/get_feature_engineering_materialized_features.py +123 -0
  98. pulumi_databricks/get_functions.py +38 -24
  99. pulumi_databricks/get_group.py +85 -86
  100. pulumi_databricks/get_instance_pool.py +10 -11
  101. pulumi_databricks/get_instance_profiles.py +25 -9
  102. pulumi_databricks/get_job.py +22 -23
  103. pulumi_databricks/get_jobs.py +44 -26
  104. pulumi_databricks/get_materialized_features_feature_tag.py +102 -0
  105. pulumi_databricks/get_materialized_features_feature_tags.py +133 -0
  106. pulumi_databricks/get_metastore.py +28 -29
  107. pulumi_databricks/get_metastores.py +9 -10
  108. pulumi_databricks/get_mlflow_experiment.py +62 -46
  109. pulumi_databricks/get_mlflow_model.py +47 -31
  110. pulumi_databricks/get_mlflow_models.py +32 -14
  111. pulumi_databricks/get_mws_credentials.py +9 -10
  112. pulumi_databricks/get_mws_network_connectivity_config.py +40 -41
  113. pulumi_databricks/get_mws_network_connectivity_configs.py +15 -16
  114. pulumi_databricks/get_mws_workspaces.py +5 -6
  115. pulumi_databricks/get_node_type.py +123 -87
  116. pulumi_databricks/get_notebook.py +37 -38
  117. pulumi_databricks/get_notebook_paths.py +16 -17
  118. pulumi_databricks/get_notification_destinations.py +33 -18
  119. pulumi_databricks/get_online_store.py +36 -27
  120. pulumi_databricks/get_online_stores.py +33 -13
  121. pulumi_databricks/get_pipelines.py +49 -31
  122. pulumi_databricks/get_policy_info.py +338 -0
  123. pulumi_databricks/get_policy_infos.py +169 -0
  124. pulumi_databricks/get_quality_monitor_v2.py +21 -18
  125. pulumi_databricks/get_quality_monitors_v2.py +26 -8
  126. pulumi_databricks/get_registered_model.py +41 -27
  127. pulumi_databricks/get_registered_model_versions.py +29 -15
  128. pulumi_databricks/get_rfa_access_request_destinations.py +126 -0
  129. pulumi_databricks/get_schema.py +32 -16
  130. pulumi_databricks/get_schemas.py +38 -20
  131. pulumi_databricks/get_service_principal.py +80 -58
  132. pulumi_databricks/get_service_principal_federation_policies.py +151 -0
  133. pulumi_databricks/get_service_principal_federation_policy.py +220 -0
  134. pulumi_databricks/get_service_principals.py +16 -17
  135. pulumi_databricks/get_serving_endpoints.py +23 -9
  136. pulumi_databricks/get_share.py +130 -32
  137. pulumi_databricks/get_shares.py +29 -11
  138. pulumi_databricks/get_spark_version.py +87 -69
  139. pulumi_databricks/get_sql_warehouse.py +124 -110
  140. pulumi_databricks/get_sql_warehouses.py +37 -19
  141. pulumi_databricks/get_storage_credential.py +32 -16
  142. pulumi_databricks/get_storage_credentials.py +32 -14
  143. pulumi_databricks/get_table.py +30 -14
  144. pulumi_databricks/get_tables.py +41 -23
  145. pulumi_databricks/get_tag_policies.py +139 -0
  146. pulumi_databricks/get_tag_policy.py +175 -0
  147. pulumi_databricks/get_user.py +33 -34
  148. pulumi_databricks/get_views.py +41 -23
  149. pulumi_databricks/get_volume.py +32 -16
  150. pulumi_databricks/get_volumes.py +41 -23
  151. pulumi_databricks/get_workspace_network_option.py +45 -26
  152. pulumi_databricks/get_workspace_setting_v2.py +331 -0
  153. pulumi_databricks/get_zones.py +39 -21
  154. pulumi_databricks/git_credential.py +242 -76
  155. pulumi_databricks/global_init_script.py +99 -100
  156. pulumi_databricks/grant.py +215 -212
  157. pulumi_databricks/grants.py +200 -211
  158. pulumi_databricks/group.py +176 -177
  159. pulumi_databricks/group_instance_profile.py +37 -38
  160. pulumi_databricks/group_member.py +37 -38
  161. pulumi_databricks/group_role.py +37 -38
  162. pulumi_databricks/instance_pool.py +168 -169
  163. pulumi_databricks/instance_profile.py +69 -70
  164. pulumi_databricks/ip_access_list.py +71 -72
  165. pulumi_databricks/job.py +426 -346
  166. pulumi_databricks/lakehouse_monitor.py +199 -200
  167. pulumi_databricks/library.py +264 -99
  168. pulumi_databricks/materialized_features_feature_tag.py +213 -0
  169. pulumi_databricks/metastore.py +254 -255
  170. pulumi_databricks/metastore_assignment.py +52 -53
  171. pulumi_databricks/metastore_data_access.py +153 -154
  172. pulumi_databricks/metastore_provider.py +69 -70
  173. pulumi_databricks/mlflow_experiment.py +108 -109
  174. pulumi_databricks/mlflow_model.py +45 -46
  175. pulumi_databricks/mlflow_webhook.py +75 -76
  176. pulumi_databricks/model_serving.py +195 -74
  177. pulumi_databricks/model_serving_provisioned_throughput.py +100 -54
  178. pulumi_databricks/mount.py +103 -104
  179. pulumi_databricks/mws_credentials.py +99 -100
  180. pulumi_databricks/mws_customer_managed_keys.py +75 -76
  181. pulumi_databricks/mws_log_delivery.py +188 -189
  182. pulumi_databricks/mws_ncc_binding.py +35 -36
  183. pulumi_databricks/mws_ncc_private_endpoint_rule.py +312 -289
  184. pulumi_databricks/mws_network_connectivity_config.py +136 -98
  185. pulumi_databricks/mws_networks.py +159 -160
  186. pulumi_databricks/mws_permission_assignment.py +70 -71
  187. pulumi_databricks/mws_private_access_settings.py +116 -117
  188. pulumi_databricks/mws_storage_configurations.py +74 -75
  189. pulumi_databricks/mws_vpc_endpoint.py +149 -150
  190. pulumi_databricks/mws_workspaces.py +464 -418
  191. pulumi_databricks/notebook.py +143 -144
  192. pulumi_databricks/notification_destination.py +38 -39
  193. pulumi_databricks/obo_token.py +63 -64
  194. pulumi_databricks/online_store.py +121 -70
  195. pulumi_databricks/online_table.py +41 -42
  196. pulumi_databricks/outputs.py +47821 -30802
  197. pulumi_databricks/permission_assignment.py +298 -57
  198. pulumi_databricks/permissions.py +523 -362
  199. pulumi_databricks/pipeline.py +515 -480
  200. pulumi_databricks/policy_info.py +916 -0
  201. pulumi_databricks/provider.py +283 -269
  202. pulumi_databricks/pulumi-plugin.json +1 -1
  203. pulumi_databricks/quality_monitor.py +263 -217
  204. pulumi_databricks/quality_monitor_v2.py +49 -46
  205. pulumi_databricks/query.py +227 -228
  206. pulumi_databricks/recipient.py +208 -209
  207. pulumi_databricks/registered_model.py +393 -122
  208. pulumi_databricks/repo.py +117 -118
  209. pulumi_databricks/restrict_workspace_admins_setting.py +30 -31
  210. pulumi_databricks/rfa_access_request_destinations.py +286 -0
  211. pulumi_databricks/schema.py +159 -160
  212. pulumi_databricks/secret.py +72 -73
  213. pulumi_databricks/secret_acl.py +52 -53
  214. pulumi_databricks/secret_scope.py +55 -56
  215. pulumi_databricks/service_principal.py +279 -278
  216. pulumi_databricks/service_principal_federation_policy.py +469 -0
  217. pulumi_databricks/service_principal_role.py +35 -36
  218. pulumi_databricks/service_principal_secret.py +156 -157
  219. pulumi_databricks/share.py +186 -187
  220. pulumi_databricks/sql_alert.py +98 -99
  221. pulumi_databricks/sql_dashboard.py +94 -95
  222. pulumi_databricks/sql_endpoint.py +298 -266
  223. pulumi_databricks/sql_global_config.py +103 -104
  224. pulumi_databricks/sql_permissions.py +121 -122
  225. pulumi_databricks/sql_query.py +150 -151
  226. pulumi_databricks/sql_table.py +310 -283
  227. pulumi_databricks/sql_visualization.py +92 -93
  228. pulumi_databricks/sql_widget.py +91 -88
  229. pulumi_databricks/storage_credential.py +236 -184
  230. pulumi_databricks/system_schema.py +50 -51
  231. pulumi_databricks/table.py +147 -148
  232. pulumi_databricks/tag_policy.py +357 -0
  233. pulumi_databricks/token.py +83 -84
  234. pulumi_databricks/user.py +284 -285
  235. pulumi_databricks/user_instance_profile.py +35 -36
  236. pulumi_databricks/user_role.py +35 -36
  237. pulumi_databricks/vector_search_endpoint.py +117 -118
  238. pulumi_databricks/vector_search_index.py +86 -87
  239. pulumi_databricks/volume.py +129 -130
  240. pulumi_databricks/workspace_binding.py +82 -131
  241. pulumi_databricks/workspace_conf.py +18 -19
  242. pulumi_databricks/workspace_file.py +100 -101
  243. pulumi_databricks/workspace_network_option.py +89 -40
  244. pulumi_databricks/workspace_setting_v2.py +759 -0
  245. {pulumi_databricks-1.74.0a1753335781.dist-info → pulumi_databricks-1.77.0.dist-info}/METADATA +1 -1
  246. pulumi_databricks-1.77.0.dist-info/RECORD +250 -0
  247. pulumi_databricks/get_recipient_federation_policies.py +0 -83
  248. pulumi_databricks/get_recipient_federation_policy.py +0 -166
  249. pulumi_databricks/recipient_federation_policy.py +0 -347
  250. pulumi_databricks-1.74.0a1753335781.dist-info/RECORD +0 -205
  251. {pulumi_databricks-1.74.0a1753335781.dist-info → pulumi_databricks-1.77.0.dist-info}/WHEEL +0 -0
  252. {pulumi_databricks-1.74.0a1753335781.dist-info → pulumi_databricks-1.77.0.dist-info}/top_level.txt +0 -0
@@ -2,8 +2,7 @@
2
2
  # *** WARNING: this file was generated by pulumi-language-python. ***
3
3
  # *** Do not edit by hand unless you're certain you know what you are doing! ***
4
4
 
5
- import builtins
6
- import copy
5
+ import builtins as _builtins
7
6
  import warnings
8
7
  import sys
9
8
  import pulumi
@@ -22,31 +21,31 @@ __all__ = ['InstancePoolArgs', 'InstancePool']
22
21
  @pulumi.input_type
23
22
  class InstancePoolArgs:
24
23
  def __init__(__self__, *,
25
- idle_instance_autotermination_minutes: pulumi.Input[builtins.int],
26
- instance_pool_name: pulumi.Input[builtins.str],
24
+ idle_instance_autotermination_minutes: pulumi.Input[_builtins.int],
25
+ instance_pool_name: pulumi.Input[_builtins.str],
27
26
  aws_attributes: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']] = None,
28
27
  azure_attributes: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']] = None,
29
- custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]] = None,
28
+ custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]] = None,
30
29
  disk_spec: Optional[pulumi.Input['InstancePoolDiskSpecArgs']] = None,
31
- enable_elastic_disk: Optional[pulumi.Input[builtins.bool]] = None,
30
+ enable_elastic_disk: Optional[pulumi.Input[_builtins.bool]] = None,
32
31
  gcp_attributes: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']] = None,
33
32
  instance_pool_fleet_attributes: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']] = None,
34
- instance_pool_id: Optional[pulumi.Input[builtins.str]] = None,
35
- max_capacity: Optional[pulumi.Input[builtins.int]] = None,
36
- min_idle_instances: Optional[pulumi.Input[builtins.int]] = None,
37
- node_type_id: Optional[pulumi.Input[builtins.str]] = None,
33
+ instance_pool_id: Optional[pulumi.Input[_builtins.str]] = None,
34
+ max_capacity: Optional[pulumi.Input[_builtins.int]] = None,
35
+ min_idle_instances: Optional[pulumi.Input[_builtins.int]] = None,
36
+ node_type_id: Optional[pulumi.Input[_builtins.str]] = None,
38
37
  preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]] = None,
39
- preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]] = None):
38
+ preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]] = None):
40
39
  """
41
40
  The set of arguments for constructing a InstancePool resource.
42
- :param pulumi.Input[builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
43
- :param pulumi.Input[builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
44
- :param pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
45
- :param pulumi.Input[builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
46
- :param pulumi.Input[builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
47
- :param pulumi.Input[builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
48
- :param pulumi.Input[builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
49
- :param pulumi.Input[Sequence[pulumi.Input[builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
41
+ :param pulumi.Input[_builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
42
+ :param pulumi.Input[_builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
43
+ :param pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
44
+ :param pulumi.Input[_builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
45
+ :param pulumi.Input[_builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
46
+ :param pulumi.Input[_builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
47
+ :param pulumi.Input[_builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
48
+ :param pulumi.Input[Sequence[pulumi.Input[_builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
50
49
  """
51
50
  pulumi.set(__self__, "idle_instance_autotermination_minutes", idle_instance_autotermination_minutes)
52
51
  pulumi.set(__self__, "instance_pool_name", instance_pool_name)
@@ -77,31 +76,31 @@ class InstancePoolArgs:
77
76
  if preloaded_spark_versions is not None:
78
77
  pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions)
79
78
 
80
- @property
79
+ @_builtins.property
81
80
  @pulumi.getter(name="idleInstanceAutoterminationMinutes")
82
- def idle_instance_autotermination_minutes(self) -> pulumi.Input[builtins.int]:
81
+ def idle_instance_autotermination_minutes(self) -> pulumi.Input[_builtins.int]:
83
82
  """
84
83
  (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
85
84
  """
86
85
  return pulumi.get(self, "idle_instance_autotermination_minutes")
87
86
 
88
87
  @idle_instance_autotermination_minutes.setter
89
- def idle_instance_autotermination_minutes(self, value: pulumi.Input[builtins.int]):
88
+ def idle_instance_autotermination_minutes(self, value: pulumi.Input[_builtins.int]):
90
89
  pulumi.set(self, "idle_instance_autotermination_minutes", value)
91
90
 
92
- @property
91
+ @_builtins.property
93
92
  @pulumi.getter(name="instancePoolName")
94
- def instance_pool_name(self) -> pulumi.Input[builtins.str]:
93
+ def instance_pool_name(self) -> pulumi.Input[_builtins.str]:
95
94
  """
96
95
  (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
97
96
  """
98
97
  return pulumi.get(self, "instance_pool_name")
99
98
 
100
99
  @instance_pool_name.setter
101
- def instance_pool_name(self, value: pulumi.Input[builtins.str]):
100
+ def instance_pool_name(self, value: pulumi.Input[_builtins.str]):
102
101
  pulumi.set(self, "instance_pool_name", value)
103
102
 
104
- @property
103
+ @_builtins.property
105
104
  @pulumi.getter(name="awsAttributes")
106
105
  def aws_attributes(self) -> Optional[pulumi.Input['InstancePoolAwsAttributesArgs']]:
107
106
  return pulumi.get(self, "aws_attributes")
@@ -110,7 +109,7 @@ class InstancePoolArgs:
110
109
  def aws_attributes(self, value: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']]):
111
110
  pulumi.set(self, "aws_attributes", value)
112
111
 
113
- @property
112
+ @_builtins.property
114
113
  @pulumi.getter(name="azureAttributes")
115
114
  def azure_attributes(self) -> Optional[pulumi.Input['InstancePoolAzureAttributesArgs']]:
116
115
  return pulumi.get(self, "azure_attributes")
@@ -119,19 +118,19 @@ class InstancePoolArgs:
119
118
  def azure_attributes(self, value: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']]):
120
119
  pulumi.set(self, "azure_attributes", value)
121
120
 
122
- @property
121
+ @_builtins.property
123
122
  @pulumi.getter(name="customTags")
124
- def custom_tags(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]]:
123
+ def custom_tags(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]]:
125
124
  """
126
125
  (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
127
126
  """
128
127
  return pulumi.get(self, "custom_tags")
129
128
 
130
129
  @custom_tags.setter
131
- def custom_tags(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]]):
130
+ def custom_tags(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]]):
132
131
  pulumi.set(self, "custom_tags", value)
133
132
 
134
- @property
133
+ @_builtins.property
135
134
  @pulumi.getter(name="diskSpec")
136
135
  def disk_spec(self) -> Optional[pulumi.Input['InstancePoolDiskSpecArgs']]:
137
136
  return pulumi.get(self, "disk_spec")
@@ -140,19 +139,19 @@ class InstancePoolArgs:
140
139
  def disk_spec(self, value: Optional[pulumi.Input['InstancePoolDiskSpecArgs']]):
141
140
  pulumi.set(self, "disk_spec", value)
142
141
 
143
- @property
142
+ @_builtins.property
144
143
  @pulumi.getter(name="enableElasticDisk")
145
- def enable_elastic_disk(self) -> Optional[pulumi.Input[builtins.bool]]:
144
+ def enable_elastic_disk(self) -> Optional[pulumi.Input[_builtins.bool]]:
146
145
  """
147
146
  (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
148
147
  """
149
148
  return pulumi.get(self, "enable_elastic_disk")
150
149
 
151
150
  @enable_elastic_disk.setter
152
- def enable_elastic_disk(self, value: Optional[pulumi.Input[builtins.bool]]):
151
+ def enable_elastic_disk(self, value: Optional[pulumi.Input[_builtins.bool]]):
153
152
  pulumi.set(self, "enable_elastic_disk", value)
154
153
 
155
- @property
154
+ @_builtins.property
156
155
  @pulumi.getter(name="gcpAttributes")
157
156
  def gcp_attributes(self) -> Optional[pulumi.Input['InstancePoolGcpAttributesArgs']]:
158
157
  return pulumi.get(self, "gcp_attributes")
@@ -161,7 +160,7 @@ class InstancePoolArgs:
161
160
  def gcp_attributes(self, value: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']]):
162
161
  pulumi.set(self, "gcp_attributes", value)
163
162
 
164
- @property
163
+ @_builtins.property
165
164
  @pulumi.getter(name="instancePoolFleetAttributes")
166
165
  def instance_pool_fleet_attributes(self) -> Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']]:
167
166
  return pulumi.get(self, "instance_pool_fleet_attributes")
@@ -170,52 +169,52 @@ class InstancePoolArgs:
170
169
  def instance_pool_fleet_attributes(self, value: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']]):
171
170
  pulumi.set(self, "instance_pool_fleet_attributes", value)
172
171
 
173
- @property
172
+ @_builtins.property
174
173
  @pulumi.getter(name="instancePoolId")
175
- def instance_pool_id(self) -> Optional[pulumi.Input[builtins.str]]:
174
+ def instance_pool_id(self) -> Optional[pulumi.Input[_builtins.str]]:
176
175
  return pulumi.get(self, "instance_pool_id")
177
176
 
178
177
  @instance_pool_id.setter
179
- def instance_pool_id(self, value: Optional[pulumi.Input[builtins.str]]):
178
+ def instance_pool_id(self, value: Optional[pulumi.Input[_builtins.str]]):
180
179
  pulumi.set(self, "instance_pool_id", value)
181
180
 
182
- @property
181
+ @_builtins.property
183
182
  @pulumi.getter(name="maxCapacity")
184
- def max_capacity(self) -> Optional[pulumi.Input[builtins.int]]:
183
+ def max_capacity(self) -> Optional[pulumi.Input[_builtins.int]]:
185
184
  """
186
185
  (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
187
186
  """
188
187
  return pulumi.get(self, "max_capacity")
189
188
 
190
189
  @max_capacity.setter
191
- def max_capacity(self, value: Optional[pulumi.Input[builtins.int]]):
190
+ def max_capacity(self, value: Optional[pulumi.Input[_builtins.int]]):
192
191
  pulumi.set(self, "max_capacity", value)
193
192
 
194
- @property
193
+ @_builtins.property
195
194
  @pulumi.getter(name="minIdleInstances")
196
- def min_idle_instances(self) -> Optional[pulumi.Input[builtins.int]]:
195
+ def min_idle_instances(self) -> Optional[pulumi.Input[_builtins.int]]:
197
196
  """
198
197
  (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
199
198
  """
200
199
  return pulumi.get(self, "min_idle_instances")
201
200
 
202
201
  @min_idle_instances.setter
203
- def min_idle_instances(self, value: Optional[pulumi.Input[builtins.int]]):
202
+ def min_idle_instances(self, value: Optional[pulumi.Input[_builtins.int]]):
204
203
  pulumi.set(self, "min_idle_instances", value)
205
204
 
206
- @property
205
+ @_builtins.property
207
206
  @pulumi.getter(name="nodeTypeId")
208
- def node_type_id(self) -> Optional[pulumi.Input[builtins.str]]:
207
+ def node_type_id(self) -> Optional[pulumi.Input[_builtins.str]]:
209
208
  """
210
209
  (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
211
210
  """
212
211
  return pulumi.get(self, "node_type_id")
213
212
 
214
213
  @node_type_id.setter
215
- def node_type_id(self, value: Optional[pulumi.Input[builtins.str]]):
214
+ def node_type_id(self, value: Optional[pulumi.Input[_builtins.str]]):
216
215
  pulumi.set(self, "node_type_id", value)
217
216
 
218
- @property
217
+ @_builtins.property
219
218
  @pulumi.getter(name="preloadedDockerImages")
220
219
  def preloaded_docker_images(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]]:
221
220
  return pulumi.get(self, "preloaded_docker_images")
@@ -224,16 +223,16 @@ class InstancePoolArgs:
224
223
  def preloaded_docker_images(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]]):
225
224
  pulumi.set(self, "preloaded_docker_images", value)
226
225
 
227
- @property
226
+ @_builtins.property
228
227
  @pulumi.getter(name="preloadedSparkVersions")
229
- def preloaded_spark_versions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]]:
228
+ def preloaded_spark_versions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]]:
230
229
  """
231
230
  (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
232
231
  """
233
232
  return pulumi.get(self, "preloaded_spark_versions")
234
233
 
235
234
  @preloaded_spark_versions.setter
236
- def preloaded_spark_versions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]]):
235
+ def preloaded_spark_versions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]]):
237
236
  pulumi.set(self, "preloaded_spark_versions", value)
238
237
 
239
238
 
@@ -242,29 +241,29 @@ class _InstancePoolState:
242
241
  def __init__(__self__, *,
243
242
  aws_attributes: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']] = None,
244
243
  azure_attributes: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']] = None,
245
- custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]] = None,
244
+ custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]] = None,
246
245
  disk_spec: Optional[pulumi.Input['InstancePoolDiskSpecArgs']] = None,
247
- enable_elastic_disk: Optional[pulumi.Input[builtins.bool]] = None,
246
+ enable_elastic_disk: Optional[pulumi.Input[_builtins.bool]] = None,
248
247
  gcp_attributes: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']] = None,
249
- idle_instance_autotermination_minutes: Optional[pulumi.Input[builtins.int]] = None,
248
+ idle_instance_autotermination_minutes: Optional[pulumi.Input[_builtins.int]] = None,
250
249
  instance_pool_fleet_attributes: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']] = None,
251
- instance_pool_id: Optional[pulumi.Input[builtins.str]] = None,
252
- instance_pool_name: Optional[pulumi.Input[builtins.str]] = None,
253
- max_capacity: Optional[pulumi.Input[builtins.int]] = None,
254
- min_idle_instances: Optional[pulumi.Input[builtins.int]] = None,
255
- node_type_id: Optional[pulumi.Input[builtins.str]] = None,
250
+ instance_pool_id: Optional[pulumi.Input[_builtins.str]] = None,
251
+ instance_pool_name: Optional[pulumi.Input[_builtins.str]] = None,
252
+ max_capacity: Optional[pulumi.Input[_builtins.int]] = None,
253
+ min_idle_instances: Optional[pulumi.Input[_builtins.int]] = None,
254
+ node_type_id: Optional[pulumi.Input[_builtins.str]] = None,
256
255
  preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]] = None,
257
- preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]] = None):
256
+ preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]] = None):
258
257
  """
259
258
  Input properties used for looking up and filtering InstancePool resources.
260
- :param pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
261
- :param pulumi.Input[builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
262
- :param pulumi.Input[builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
263
- :param pulumi.Input[builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
264
- :param pulumi.Input[builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
265
- :param pulumi.Input[builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
266
- :param pulumi.Input[builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
267
- :param pulumi.Input[Sequence[pulumi.Input[builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
259
+ :param pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
260
+ :param pulumi.Input[_builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
261
+ :param pulumi.Input[_builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
262
+ :param pulumi.Input[_builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
263
+ :param pulumi.Input[_builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
264
+ :param pulumi.Input[_builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
265
+ :param pulumi.Input[_builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
266
+ :param pulumi.Input[Sequence[pulumi.Input[_builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
268
267
  """
269
268
  if aws_attributes is not None:
270
269
  pulumi.set(__self__, "aws_attributes", aws_attributes)
@@ -297,7 +296,7 @@ class _InstancePoolState:
297
296
  if preloaded_spark_versions is not None:
298
297
  pulumi.set(__self__, "preloaded_spark_versions", preloaded_spark_versions)
299
298
 
300
- @property
299
+ @_builtins.property
301
300
  @pulumi.getter(name="awsAttributes")
302
301
  def aws_attributes(self) -> Optional[pulumi.Input['InstancePoolAwsAttributesArgs']]:
303
302
  return pulumi.get(self, "aws_attributes")
@@ -306,7 +305,7 @@ class _InstancePoolState:
306
305
  def aws_attributes(self, value: Optional[pulumi.Input['InstancePoolAwsAttributesArgs']]):
307
306
  pulumi.set(self, "aws_attributes", value)
308
307
 
309
- @property
308
+ @_builtins.property
310
309
  @pulumi.getter(name="azureAttributes")
311
310
  def azure_attributes(self) -> Optional[pulumi.Input['InstancePoolAzureAttributesArgs']]:
312
311
  return pulumi.get(self, "azure_attributes")
@@ -315,19 +314,19 @@ class _InstancePoolState:
315
314
  def azure_attributes(self, value: Optional[pulumi.Input['InstancePoolAzureAttributesArgs']]):
316
315
  pulumi.set(self, "azure_attributes", value)
317
316
 
318
- @property
317
+ @_builtins.property
319
318
  @pulumi.getter(name="customTags")
320
- def custom_tags(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]]:
319
+ def custom_tags(self) -> Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]]:
321
320
  """
322
321
  (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
323
322
  """
324
323
  return pulumi.get(self, "custom_tags")
325
324
 
326
325
  @custom_tags.setter
327
- def custom_tags(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]]):
326
+ def custom_tags(self, value: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]]):
328
327
  pulumi.set(self, "custom_tags", value)
329
328
 
330
- @property
329
+ @_builtins.property
331
330
  @pulumi.getter(name="diskSpec")
332
331
  def disk_spec(self) -> Optional[pulumi.Input['InstancePoolDiskSpecArgs']]:
333
332
  return pulumi.get(self, "disk_spec")
@@ -336,19 +335,19 @@ class _InstancePoolState:
336
335
  def disk_spec(self, value: Optional[pulumi.Input['InstancePoolDiskSpecArgs']]):
337
336
  pulumi.set(self, "disk_spec", value)
338
337
 
339
- @property
338
+ @_builtins.property
340
339
  @pulumi.getter(name="enableElasticDisk")
341
- def enable_elastic_disk(self) -> Optional[pulumi.Input[builtins.bool]]:
340
+ def enable_elastic_disk(self) -> Optional[pulumi.Input[_builtins.bool]]:
342
341
  """
343
342
  (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
344
343
  """
345
344
  return pulumi.get(self, "enable_elastic_disk")
346
345
 
347
346
  @enable_elastic_disk.setter
348
- def enable_elastic_disk(self, value: Optional[pulumi.Input[builtins.bool]]):
347
+ def enable_elastic_disk(self, value: Optional[pulumi.Input[_builtins.bool]]):
349
348
  pulumi.set(self, "enable_elastic_disk", value)
350
349
 
351
- @property
350
+ @_builtins.property
352
351
  @pulumi.getter(name="gcpAttributes")
353
352
  def gcp_attributes(self) -> Optional[pulumi.Input['InstancePoolGcpAttributesArgs']]:
354
353
  return pulumi.get(self, "gcp_attributes")
@@ -357,19 +356,19 @@ class _InstancePoolState:
357
356
  def gcp_attributes(self, value: Optional[pulumi.Input['InstancePoolGcpAttributesArgs']]):
358
357
  pulumi.set(self, "gcp_attributes", value)
359
358
 
360
- @property
359
+ @_builtins.property
361
360
  @pulumi.getter(name="idleInstanceAutoterminationMinutes")
362
- def idle_instance_autotermination_minutes(self) -> Optional[pulumi.Input[builtins.int]]:
361
+ def idle_instance_autotermination_minutes(self) -> Optional[pulumi.Input[_builtins.int]]:
363
362
  """
364
363
  (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
365
364
  """
366
365
  return pulumi.get(self, "idle_instance_autotermination_minutes")
367
366
 
368
367
  @idle_instance_autotermination_minutes.setter
369
- def idle_instance_autotermination_minutes(self, value: Optional[pulumi.Input[builtins.int]]):
368
+ def idle_instance_autotermination_minutes(self, value: Optional[pulumi.Input[_builtins.int]]):
370
369
  pulumi.set(self, "idle_instance_autotermination_minutes", value)
371
370
 
372
- @property
371
+ @_builtins.property
373
372
  @pulumi.getter(name="instancePoolFleetAttributes")
374
373
  def instance_pool_fleet_attributes(self) -> Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']]:
375
374
  return pulumi.get(self, "instance_pool_fleet_attributes")
@@ -378,64 +377,64 @@ class _InstancePoolState:
378
377
  def instance_pool_fleet_attributes(self, value: Optional[pulumi.Input['InstancePoolInstancePoolFleetAttributesArgs']]):
379
378
  pulumi.set(self, "instance_pool_fleet_attributes", value)
380
379
 
381
- @property
380
+ @_builtins.property
382
381
  @pulumi.getter(name="instancePoolId")
383
- def instance_pool_id(self) -> Optional[pulumi.Input[builtins.str]]:
382
+ def instance_pool_id(self) -> Optional[pulumi.Input[_builtins.str]]:
384
383
  return pulumi.get(self, "instance_pool_id")
385
384
 
386
385
  @instance_pool_id.setter
387
- def instance_pool_id(self, value: Optional[pulumi.Input[builtins.str]]):
386
+ def instance_pool_id(self, value: Optional[pulumi.Input[_builtins.str]]):
388
387
  pulumi.set(self, "instance_pool_id", value)
389
388
 
390
- @property
389
+ @_builtins.property
391
390
  @pulumi.getter(name="instancePoolName")
392
- def instance_pool_name(self) -> Optional[pulumi.Input[builtins.str]]:
391
+ def instance_pool_name(self) -> Optional[pulumi.Input[_builtins.str]]:
393
392
  """
394
393
  (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
395
394
  """
396
395
  return pulumi.get(self, "instance_pool_name")
397
396
 
398
397
  @instance_pool_name.setter
399
- def instance_pool_name(self, value: Optional[pulumi.Input[builtins.str]]):
398
+ def instance_pool_name(self, value: Optional[pulumi.Input[_builtins.str]]):
400
399
  pulumi.set(self, "instance_pool_name", value)
401
400
 
402
- @property
401
+ @_builtins.property
403
402
  @pulumi.getter(name="maxCapacity")
404
- def max_capacity(self) -> Optional[pulumi.Input[builtins.int]]:
403
+ def max_capacity(self) -> Optional[pulumi.Input[_builtins.int]]:
405
404
  """
406
405
  (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
407
406
  """
408
407
  return pulumi.get(self, "max_capacity")
409
408
 
410
409
  @max_capacity.setter
411
- def max_capacity(self, value: Optional[pulumi.Input[builtins.int]]):
410
+ def max_capacity(self, value: Optional[pulumi.Input[_builtins.int]]):
412
411
  pulumi.set(self, "max_capacity", value)
413
412
 
414
- @property
413
+ @_builtins.property
415
414
  @pulumi.getter(name="minIdleInstances")
416
- def min_idle_instances(self) -> Optional[pulumi.Input[builtins.int]]:
415
+ def min_idle_instances(self) -> Optional[pulumi.Input[_builtins.int]]:
417
416
  """
418
417
  (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
419
418
  """
420
419
  return pulumi.get(self, "min_idle_instances")
421
420
 
422
421
  @min_idle_instances.setter
423
- def min_idle_instances(self, value: Optional[pulumi.Input[builtins.int]]):
422
+ def min_idle_instances(self, value: Optional[pulumi.Input[_builtins.int]]):
424
423
  pulumi.set(self, "min_idle_instances", value)
425
424
 
426
- @property
425
+ @_builtins.property
427
426
  @pulumi.getter(name="nodeTypeId")
428
- def node_type_id(self) -> Optional[pulumi.Input[builtins.str]]:
427
+ def node_type_id(self) -> Optional[pulumi.Input[_builtins.str]]:
429
428
  """
430
429
  (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
431
430
  """
432
431
  return pulumi.get(self, "node_type_id")
433
432
 
434
433
  @node_type_id.setter
435
- def node_type_id(self, value: Optional[pulumi.Input[builtins.str]]):
434
+ def node_type_id(self, value: Optional[pulumi.Input[_builtins.str]]):
436
435
  pulumi.set(self, "node_type_id", value)
437
436
 
438
- @property
437
+ @_builtins.property
439
438
  @pulumi.getter(name="preloadedDockerImages")
440
439
  def preloaded_docker_images(self) -> Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]]:
441
440
  return pulumi.get(self, "preloaded_docker_images")
@@ -444,16 +443,16 @@ class _InstancePoolState:
444
443
  def preloaded_docker_images(self, value: Optional[pulumi.Input[Sequence[pulumi.Input['InstancePoolPreloadedDockerImageArgs']]]]):
445
444
  pulumi.set(self, "preloaded_docker_images", value)
446
445
 
447
- @property
446
+ @_builtins.property
448
447
  @pulumi.getter(name="preloadedSparkVersions")
449
- def preloaded_spark_versions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]]:
448
+ def preloaded_spark_versions(self) -> Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]]:
450
449
  """
451
450
  (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
452
451
  """
453
452
  return pulumi.get(self, "preloaded_spark_versions")
454
453
 
455
454
  @preloaded_spark_versions.setter
456
- def preloaded_spark_versions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]]):
455
+ def preloaded_spark_versions(self, value: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]]):
457
456
  pulumi.set(self, "preloaded_spark_versions", value)
458
457
 
459
458
 
@@ -465,19 +464,19 @@ class InstancePool(pulumi.CustomResource):
465
464
  opts: Optional[pulumi.ResourceOptions] = None,
466
465
  aws_attributes: Optional[pulumi.Input[Union['InstancePoolAwsAttributesArgs', 'InstancePoolAwsAttributesArgsDict']]] = None,
467
466
  azure_attributes: Optional[pulumi.Input[Union['InstancePoolAzureAttributesArgs', 'InstancePoolAzureAttributesArgsDict']]] = None,
468
- custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]] = None,
467
+ custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]] = None,
469
468
  disk_spec: Optional[pulumi.Input[Union['InstancePoolDiskSpecArgs', 'InstancePoolDiskSpecArgsDict']]] = None,
470
- enable_elastic_disk: Optional[pulumi.Input[builtins.bool]] = None,
469
+ enable_elastic_disk: Optional[pulumi.Input[_builtins.bool]] = None,
471
470
  gcp_attributes: Optional[pulumi.Input[Union['InstancePoolGcpAttributesArgs', 'InstancePoolGcpAttributesArgsDict']]] = None,
472
- idle_instance_autotermination_minutes: Optional[pulumi.Input[builtins.int]] = None,
471
+ idle_instance_autotermination_minutes: Optional[pulumi.Input[_builtins.int]] = None,
473
472
  instance_pool_fleet_attributes: Optional[pulumi.Input[Union['InstancePoolInstancePoolFleetAttributesArgs', 'InstancePoolInstancePoolFleetAttributesArgsDict']]] = None,
474
- instance_pool_id: Optional[pulumi.Input[builtins.str]] = None,
475
- instance_pool_name: Optional[pulumi.Input[builtins.str]] = None,
476
- max_capacity: Optional[pulumi.Input[builtins.int]] = None,
477
- min_idle_instances: Optional[pulumi.Input[builtins.int]] = None,
478
- node_type_id: Optional[pulumi.Input[builtins.str]] = None,
473
+ instance_pool_id: Optional[pulumi.Input[_builtins.str]] = None,
474
+ instance_pool_name: Optional[pulumi.Input[_builtins.str]] = None,
475
+ max_capacity: Optional[pulumi.Input[_builtins.int]] = None,
476
+ min_idle_instances: Optional[pulumi.Input[_builtins.int]] = None,
477
+ node_type_id: Optional[pulumi.Input[_builtins.str]] = None,
479
478
  preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input[Union['InstancePoolPreloadedDockerImageArgs', 'InstancePoolPreloadedDockerImageArgsDict']]]]] = None,
480
- preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]] = None,
479
+ preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]] = None,
481
480
  __props__=None):
482
481
  """
483
482
  This resource allows you to manage [instance pools](https://docs.databricks.com/clusters/instance-pools/index.html) to reduce cluster start and auto-scaling times by maintaining a set of idle, ready-to-use instances. An instance pool reduces cluster start and auto-scaling times by maintaining a set of idle, ready-to-use cloud instances. When a cluster attached to a pool needs an instance, it first attempts to allocate one of the pool's idle instances. If the pool has no idle instances, it expands by allocating a new instance from the instance provider in order to accommodate the cluster's request. When a cluster releases an instance, it returns to the pool and is free for another cluster to use. Only clusters attached to a pool can use that pool's idle instances.
@@ -544,14 +543,14 @@ class InstancePool(pulumi.CustomResource):
544
543
 
545
544
  :param str resource_name: The name of the resource.
546
545
  :param pulumi.ResourceOptions opts: Options for the resource.
547
- :param pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
548
- :param pulumi.Input[builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
549
- :param pulumi.Input[builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
550
- :param pulumi.Input[builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
551
- :param pulumi.Input[builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
552
- :param pulumi.Input[builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
553
- :param pulumi.Input[builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
554
- :param pulumi.Input[Sequence[pulumi.Input[builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
546
+ :param pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
547
+ :param pulumi.Input[_builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
548
+ :param pulumi.Input[_builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
549
+ :param pulumi.Input[_builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
550
+ :param pulumi.Input[_builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
551
+ :param pulumi.Input[_builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
552
+ :param pulumi.Input[_builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
553
+ :param pulumi.Input[Sequence[pulumi.Input[_builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
555
554
  """
556
555
  ...
557
556
  @overload
@@ -639,19 +638,19 @@ class InstancePool(pulumi.CustomResource):
639
638
  opts: Optional[pulumi.ResourceOptions] = None,
640
639
  aws_attributes: Optional[pulumi.Input[Union['InstancePoolAwsAttributesArgs', 'InstancePoolAwsAttributesArgsDict']]] = None,
641
640
  azure_attributes: Optional[pulumi.Input[Union['InstancePoolAzureAttributesArgs', 'InstancePoolAzureAttributesArgsDict']]] = None,
642
- custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]] = None,
641
+ custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]] = None,
643
642
  disk_spec: Optional[pulumi.Input[Union['InstancePoolDiskSpecArgs', 'InstancePoolDiskSpecArgsDict']]] = None,
644
- enable_elastic_disk: Optional[pulumi.Input[builtins.bool]] = None,
643
+ enable_elastic_disk: Optional[pulumi.Input[_builtins.bool]] = None,
645
644
  gcp_attributes: Optional[pulumi.Input[Union['InstancePoolGcpAttributesArgs', 'InstancePoolGcpAttributesArgsDict']]] = None,
646
- idle_instance_autotermination_minutes: Optional[pulumi.Input[builtins.int]] = None,
645
+ idle_instance_autotermination_minutes: Optional[pulumi.Input[_builtins.int]] = None,
647
646
  instance_pool_fleet_attributes: Optional[pulumi.Input[Union['InstancePoolInstancePoolFleetAttributesArgs', 'InstancePoolInstancePoolFleetAttributesArgsDict']]] = None,
648
- instance_pool_id: Optional[pulumi.Input[builtins.str]] = None,
649
- instance_pool_name: Optional[pulumi.Input[builtins.str]] = None,
650
- max_capacity: Optional[pulumi.Input[builtins.int]] = None,
651
- min_idle_instances: Optional[pulumi.Input[builtins.int]] = None,
652
- node_type_id: Optional[pulumi.Input[builtins.str]] = None,
647
+ instance_pool_id: Optional[pulumi.Input[_builtins.str]] = None,
648
+ instance_pool_name: Optional[pulumi.Input[_builtins.str]] = None,
649
+ max_capacity: Optional[pulumi.Input[_builtins.int]] = None,
650
+ min_idle_instances: Optional[pulumi.Input[_builtins.int]] = None,
651
+ node_type_id: Optional[pulumi.Input[_builtins.str]] = None,
653
652
  preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input[Union['InstancePoolPreloadedDockerImageArgs', 'InstancePoolPreloadedDockerImageArgsDict']]]]] = None,
654
- preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]] = None,
653
+ preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]] = None,
655
654
  __props__=None):
656
655
  opts = pulumi.ResourceOptions.merge(_utilities.get_resource_opts_defaults(), opts)
657
656
  if not isinstance(opts, pulumi.ResourceOptions):
@@ -692,19 +691,19 @@ class InstancePool(pulumi.CustomResource):
692
691
  opts: Optional[pulumi.ResourceOptions] = None,
693
692
  aws_attributes: Optional[pulumi.Input[Union['InstancePoolAwsAttributesArgs', 'InstancePoolAwsAttributesArgsDict']]] = None,
694
693
  azure_attributes: Optional[pulumi.Input[Union['InstancePoolAzureAttributesArgs', 'InstancePoolAzureAttributesArgsDict']]] = None,
695
- custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]]] = None,
694
+ custom_tags: Optional[pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]]] = None,
696
695
  disk_spec: Optional[pulumi.Input[Union['InstancePoolDiskSpecArgs', 'InstancePoolDiskSpecArgsDict']]] = None,
697
- enable_elastic_disk: Optional[pulumi.Input[builtins.bool]] = None,
696
+ enable_elastic_disk: Optional[pulumi.Input[_builtins.bool]] = None,
698
697
  gcp_attributes: Optional[pulumi.Input[Union['InstancePoolGcpAttributesArgs', 'InstancePoolGcpAttributesArgsDict']]] = None,
699
- idle_instance_autotermination_minutes: Optional[pulumi.Input[builtins.int]] = None,
698
+ idle_instance_autotermination_minutes: Optional[pulumi.Input[_builtins.int]] = None,
700
699
  instance_pool_fleet_attributes: Optional[pulumi.Input[Union['InstancePoolInstancePoolFleetAttributesArgs', 'InstancePoolInstancePoolFleetAttributesArgsDict']]] = None,
701
- instance_pool_id: Optional[pulumi.Input[builtins.str]] = None,
702
- instance_pool_name: Optional[pulumi.Input[builtins.str]] = None,
703
- max_capacity: Optional[pulumi.Input[builtins.int]] = None,
704
- min_idle_instances: Optional[pulumi.Input[builtins.int]] = None,
705
- node_type_id: Optional[pulumi.Input[builtins.str]] = None,
700
+ instance_pool_id: Optional[pulumi.Input[_builtins.str]] = None,
701
+ instance_pool_name: Optional[pulumi.Input[_builtins.str]] = None,
702
+ max_capacity: Optional[pulumi.Input[_builtins.int]] = None,
703
+ min_idle_instances: Optional[pulumi.Input[_builtins.int]] = None,
704
+ node_type_id: Optional[pulumi.Input[_builtins.str]] = None,
706
705
  preloaded_docker_images: Optional[pulumi.Input[Sequence[pulumi.Input[Union['InstancePoolPreloadedDockerImageArgs', 'InstancePoolPreloadedDockerImageArgsDict']]]]] = None,
707
- preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[builtins.str]]]] = None) -> 'InstancePool':
706
+ preloaded_spark_versions: Optional[pulumi.Input[Sequence[pulumi.Input[_builtins.str]]]] = None) -> 'InstancePool':
708
707
  """
709
708
  Get an existing InstancePool resource's state with the given name, id, and optional extra
710
709
  properties used to qualify the lookup.
@@ -712,14 +711,14 @@ class InstancePool(pulumi.CustomResource):
712
711
  :param str resource_name: The unique name of the resulting resource.
713
712
  :param pulumi.Input[str] id: The unique provider ID of the resource to lookup.
714
713
  :param pulumi.ResourceOptions opts: Options for the resource.
715
- :param pulumi.Input[Mapping[str, pulumi.Input[builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
716
- :param pulumi.Input[builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
717
- :param pulumi.Input[builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
718
- :param pulumi.Input[builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
719
- :param pulumi.Input[builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
720
- :param pulumi.Input[builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
721
- :param pulumi.Input[builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
722
- :param pulumi.Input[Sequence[pulumi.Input[builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
714
+ :param pulumi.Input[Mapping[str, pulumi.Input[_builtins.str]]] custom_tags: (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
715
+ :param pulumi.Input[_builtins.bool] enable_elastic_disk: (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
716
+ :param pulumi.Input[_builtins.int] idle_instance_autotermination_minutes: (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
717
+ :param pulumi.Input[_builtins.str] instance_pool_name: (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
718
+ :param pulumi.Input[_builtins.int] max_capacity: (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
719
+ :param pulumi.Input[_builtins.int] min_idle_instances: (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
720
+ :param pulumi.Input[_builtins.str] node_type_id: (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
721
+ :param pulumi.Input[Sequence[pulumi.Input[_builtins.str]]] preloaded_spark_versions: (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
723
722
  """
724
723
  opts = pulumi.ResourceOptions.merge(opts, pulumi.ResourceOptions(id=id))
725
724
 
@@ -742,100 +741,100 @@ class InstancePool(pulumi.CustomResource):
742
741
  __props__.__dict__["preloaded_spark_versions"] = preloaded_spark_versions
743
742
  return InstancePool(resource_name, opts=opts, __props__=__props__)
744
743
 
745
- @property
744
+ @_builtins.property
746
745
  @pulumi.getter(name="awsAttributes")
747
746
  def aws_attributes(self) -> pulumi.Output[Optional['outputs.InstancePoolAwsAttributes']]:
748
747
  return pulumi.get(self, "aws_attributes")
749
748
 
750
- @property
749
+ @_builtins.property
751
750
  @pulumi.getter(name="azureAttributes")
752
751
  def azure_attributes(self) -> pulumi.Output[Optional['outputs.InstancePoolAzureAttributes']]:
753
752
  return pulumi.get(self, "azure_attributes")
754
753
 
755
- @property
754
+ @_builtins.property
756
755
  @pulumi.getter(name="customTags")
757
- def custom_tags(self) -> pulumi.Output[Optional[Mapping[str, builtins.str]]]:
756
+ def custom_tags(self) -> pulumi.Output[Optional[Mapping[str, _builtins.str]]]:
758
757
  """
759
758
  (Map) Additional tags for instance pool resources. Databricks tags all pool resources (e.g. AWS & Azure instances and Disk volumes). The tags of the instance pool will propagate to the clusters using the pool (see the [official documentation](https://docs.databricks.com/administration-guide/account-settings/usage-detail-tags-aws.html#tag-propagation)). Attempting to set the same tags in both cluster and instance pool will raise an error. *Databricks allows at most 43 custom tags.*
760
759
  """
761
760
  return pulumi.get(self, "custom_tags")
762
761
 
763
- @property
762
+ @_builtins.property
764
763
  @pulumi.getter(name="diskSpec")
765
764
  def disk_spec(self) -> pulumi.Output[Optional['outputs.InstancePoolDiskSpec']]:
766
765
  return pulumi.get(self, "disk_spec")
767
766
 
768
- @property
767
+ @_builtins.property
769
768
  @pulumi.getter(name="enableElasticDisk")
770
- def enable_elastic_disk(self) -> pulumi.Output[Optional[builtins.bool]]:
769
+ def enable_elastic_disk(self) -> pulumi.Output[Optional[_builtins.bool]]:
771
770
  """
772
771
  (Bool) Autoscaling Local Storage: when enabled, the instances in the pool dynamically acquire additional disk space when they are running low on disk space.
773
772
  """
774
773
  return pulumi.get(self, "enable_elastic_disk")
775
774
 
776
- @property
775
+ @_builtins.property
777
776
  @pulumi.getter(name="gcpAttributes")
778
777
  def gcp_attributes(self) -> pulumi.Output[Optional['outputs.InstancePoolGcpAttributes']]:
779
778
  return pulumi.get(self, "gcp_attributes")
780
779
 
781
- @property
780
+ @_builtins.property
782
781
  @pulumi.getter(name="idleInstanceAutoterminationMinutes")
783
- def idle_instance_autotermination_minutes(self) -> pulumi.Output[builtins.int]:
782
+ def idle_instance_autotermination_minutes(self) -> pulumi.Output[_builtins.int]:
784
783
  """
785
784
  (Integer) The number of minutes that idle instances in excess of the min_idle_instances are maintained by the pool before being terminated. If not specified, excess idle instances are terminated automatically after a default timeout period. If specified, the time must be between 0 and 10000 minutes. If you specify 0, excess idle instances are removed as soon as possible.
786
785
  """
787
786
  return pulumi.get(self, "idle_instance_autotermination_minutes")
788
787
 
789
- @property
788
+ @_builtins.property
790
789
  @pulumi.getter(name="instancePoolFleetAttributes")
791
790
  def instance_pool_fleet_attributes(self) -> pulumi.Output[Optional['outputs.InstancePoolInstancePoolFleetAttributes']]:
792
791
  return pulumi.get(self, "instance_pool_fleet_attributes")
793
792
 
794
- @property
793
+ @_builtins.property
795
794
  @pulumi.getter(name="instancePoolId")
796
- def instance_pool_id(self) -> pulumi.Output[builtins.str]:
795
+ def instance_pool_id(self) -> pulumi.Output[_builtins.str]:
797
796
  return pulumi.get(self, "instance_pool_id")
798
797
 
799
- @property
798
+ @_builtins.property
800
799
  @pulumi.getter(name="instancePoolName")
801
- def instance_pool_name(self) -> pulumi.Output[builtins.str]:
800
+ def instance_pool_name(self) -> pulumi.Output[_builtins.str]:
802
801
  """
803
802
  (String) The name of the instance pool. This is required for create and edit operations. It must be unique, non-empty, and less than 100 characters.
804
803
  """
805
804
  return pulumi.get(self, "instance_pool_name")
806
805
 
807
- @property
806
+ @_builtins.property
808
807
  @pulumi.getter(name="maxCapacity")
809
- def max_capacity(self) -> pulumi.Output[Optional[builtins.int]]:
808
+ def max_capacity(self) -> pulumi.Output[Optional[_builtins.int]]:
810
809
  """
811
810
  (Integer) The maximum number of instances the pool can contain, including both idle instances and ones in use by clusters. Once the maximum capacity is reached, you cannot create new clusters from the pool and existing clusters cannot autoscale up until some instances are made idle in the pool via cluster termination or down-scaling. There is no default limit, but as a [best practice](https://docs.databricks.com/clusters/instance-pools/pool-best-practices.html#configure-pools-to-control-cost), this should be set based on anticipated usage.
812
811
  """
813
812
  return pulumi.get(self, "max_capacity")
814
813
 
815
- @property
814
+ @_builtins.property
816
815
  @pulumi.getter(name="minIdleInstances")
817
- def min_idle_instances(self) -> pulumi.Output[Optional[builtins.int]]:
816
+ def min_idle_instances(self) -> pulumi.Output[Optional[_builtins.int]]:
818
817
  """
819
818
  (Integer) The minimum number of idle instances maintained by the pool. This is in addition to any instances in use by active clusters.
820
819
  """
821
820
  return pulumi.get(self, "min_idle_instances")
822
821
 
823
- @property
822
+ @_builtins.property
824
823
  @pulumi.getter(name="nodeTypeId")
825
- def node_type_id(self) -> pulumi.Output[Optional[builtins.str]]:
824
+ def node_type_id(self) -> pulumi.Output[Optional[_builtins.str]]:
826
825
  """
827
826
  (String) The node type for the instances in the pool. All clusters attached to the pool inherit this node type and the pool's idle instances are allocated based on this type. You can retrieve a list of available node types by using the [List Node Types API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistnodetypes) call.
828
827
  """
829
828
  return pulumi.get(self, "node_type_id")
830
829
 
831
- @property
830
+ @_builtins.property
832
831
  @pulumi.getter(name="preloadedDockerImages")
833
832
  def preloaded_docker_images(self) -> pulumi.Output[Optional[Sequence['outputs.InstancePoolPreloadedDockerImage']]]:
834
833
  return pulumi.get(self, "preloaded_docker_images")
835
834
 
836
- @property
835
+ @_builtins.property
837
836
  @pulumi.getter(name="preloadedSparkVersions")
838
- def preloaded_spark_versions(self) -> pulumi.Output[Optional[Sequence[builtins.str]]]:
837
+ def preloaded_spark_versions(self) -> pulumi.Output[Optional[Sequence[_builtins.str]]]:
839
838
  """
840
839
  (List) A list with at most one runtime version the pool installs on each instance. Pool clusters that use a preloaded runtime version start faster as they do not have to wait for the image to download. You can retrieve them via get_spark_version data source or via [Runtime Versions API](https://docs.databricks.com/dev-tools/api/latest/clusters.html#clusterclusterservicelistsparkversions) call.
841
840
  """