databricks-bundles 0.274.0__py3-none-any.whl → 0.276.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -27,10 +27,9 @@ class Resources:
27
27
 
28
28
  .. code-block:: yaml
29
29
 
30
- experimental:
31
- python:
32
- resources:
33
- - "resources:load_resources"
30
+ python:
31
+ resources:
32
+ - "resources:load_resources"
34
33
 
35
34
  `load_resources` function can be implemented using built-in functions:
36
35
 
@@ -15,7 +15,10 @@ if TYPE_CHECKING:
15
15
 
16
16
  @dataclass(kw_only=True)
17
17
  class CleanRoomsNotebookTask:
18
- """"""
18
+ """
19
+ Clean Rooms notebook task for V1 Clean Room service (GA).
20
+ Replaces the deprecated CleanRoomNotebookTask (defined above) which was for V0 service.
21
+ """
19
22
 
20
23
  clean_room_name: VariableOr[str]
21
24
  """
@@ -57,8 +57,6 @@ __all__ = [
57
57
  "IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValue",
58
58
  "IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValueDict",
59
59
  "IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValueParam",
60
- "IngestionSourceType",
61
- "IngestionSourceTypeParam",
62
60
  "InitScriptInfo",
63
61
  "InitScriptInfoDict",
64
62
  "InitScriptInfoParam",
@@ -246,10 +244,6 @@ from databricks.bundles.pipelines._models.ingestion_pipeline_definition_workday_
246
244
  IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValueDict,
247
245
  IngestionPipelineDefinitionWorkdayReportParametersQueryKeyValueParam,
248
246
  )
249
- from databricks.bundles.pipelines._models.ingestion_source_type import (
250
- IngestionSourceType,
251
- IngestionSourceTypeParam,
252
- )
253
247
  from databricks.bundles.pipelines._models.init_script_info import (
254
248
  InitScriptInfo,
255
249
  InitScriptInfoDict,
@@ -6,7 +6,7 @@ class DayOfWeek(Enum):
6
6
  """
7
7
  :meta private: [EXPERIMENTAL]
8
8
 
9
- Days of week in which the restart is allowed to happen (within a five-hour window starting at start_hour).
9
+ Days of week in which the window is allowed to happen.
10
10
  If not specified all days of the week will be used.
11
11
  """
12
12
 
@@ -8,10 +8,6 @@ from databricks.bundles.pipelines._models.ingestion_config import (
8
8
  IngestionConfig,
9
9
  IngestionConfigParam,
10
10
  )
11
- from databricks.bundles.pipelines._models.ingestion_source_type import (
12
- IngestionSourceType,
13
- IngestionSourceTypeParam,
14
- )
15
11
  from databricks.bundles.pipelines._models.source_config import (
16
12
  SourceConfig,
17
13
  SourceConfigParam,
@@ -60,13 +56,6 @@ class IngestionPipelineDefinition:
60
56
  Top-level source configurations
61
57
  """
62
58
 
63
- source_type: VariableOrOptional[IngestionSourceType] = None
64
- """
65
- The type of the foreign source.
66
- The source type will be inferred from the source connection or ingestion gateway.
67
- This field is output only and will be ignored if provided.
68
- """
69
-
70
59
  table_configuration: VariableOrOptional[TableSpecificConfig] = None
71
60
  """
72
61
  Configuration settings to control the ingestion of tables. These settings are applied to all tables in the pipeline.
@@ -114,13 +103,6 @@ class IngestionPipelineDefinitionDict(TypedDict, total=False):
114
103
  Top-level source configurations
115
104
  """
116
105
 
117
- source_type: VariableOrOptional[IngestionSourceTypeParam]
118
- """
119
- The type of the foreign source.
120
- The source type will be inferred from the source connection or ingestion gateway.
121
- This field is output only and will be ignored if provided.
122
- """
123
-
124
106
  table_configuration: VariableOrOptional[TableSpecificConfigParam]
125
107
  """
126
108
  Configuration settings to control the ingestion of tables. These settings are applied to all tables in the pipeline.
@@ -70,8 +70,6 @@ class Pipeline(Resource):
70
70
 
71
71
  budget_policy_id: VariableOrOptional[str] = None
72
72
  """
73
- :meta private: [EXPERIMENTAL]
74
-
75
73
  Budget policy of this pipeline.
76
74
  """
77
75
 
@@ -212,6 +210,13 @@ class Pipeline(Resource):
212
210
  [DEPRECATED] Target schema (database) to add tables in this pipeline to. Exactly one of `schema` or `target` must be specified. To publish to Unity Catalog, also specify `catalog`. This legacy field is deprecated for pipeline creation in favor of the `schema` field.
213
211
  """
214
212
 
213
+ usage_policy_id: VariableOrOptional[str] = None
214
+ """
215
+ :meta private: [EXPERIMENTAL]
216
+
217
+ Usage policy of this pipeline.
218
+ """
219
+
215
220
  @classmethod
216
221
  def from_dict(cls, value: "PipelineDict") -> "Self":
217
222
  return _transform(cls, value)
@@ -230,8 +235,6 @@ class PipelineDict(TypedDict, total=False):
230
235
 
231
236
  budget_policy_id: VariableOrOptional[str]
232
237
  """
233
- :meta private: [EXPERIMENTAL]
234
-
235
238
  Budget policy of this pipeline.
236
239
  """
237
240
 
@@ -372,5 +375,12 @@ class PipelineDict(TypedDict, total=False):
372
375
  [DEPRECATED] Target schema (database) to add tables in this pipeline to. Exactly one of `schema` or `target` must be specified. To publish to Unity Catalog, also specify `catalog`. This legacy field is deprecated for pipeline creation in favor of the `schema` field.
373
376
  """
374
377
 
378
+ usage_policy_id: VariableOrOptional[str]
379
+ """
380
+ :meta private: [EXPERIMENTAL]
381
+
382
+ Usage policy of this pipeline.
383
+ """
384
+
375
385
 
376
386
  PipelineParam = PipelineDict | Pipeline
@@ -1 +1 @@
1
- __version__ = "0.274.0"
1
+ __version__ = "0.276.0"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: databricks-bundles
3
- Version: 0.274.0
3
+ Version: 0.276.0
4
4
  Summary: Python support for Databricks Asset Bundles
5
5
  Author-email: Gleb Kanterov <gleb.kanterov@databricks.com>
6
6
  Requires-Python: >=3.10
@@ -22,7 +22,7 @@ Reference documentation is available at https://databricks.github.io/cli/experim
22
22
 
23
23
  To use `databricks-bundles`, you must first:
24
24
 
25
- 1. Install the [Databricks CLI](https://github.com/databricks/cli), version 0.274.0 or above
25
+ 1. Install the [Databricks CLI](https://github.com/databricks/cli), version 0.276.0 or above
26
26
  2. Authenticate to your Databricks workspace if you have not done so already:
27
27
 
28
28
  ```bash
@@ -2,7 +2,7 @@ databricks/__init__.py,sha256=CF2MJcZFwbpn9TwQER8qnCDhkPooBGQNVkX4v7g6p3g,537
2
2
  databricks/bundles/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
3
3
  databricks/bundles/build.py,sha256=Qx40q9qEtIKK7e6OyQHUZGdY045JBchjjvOPBB8TgbM,17830
4
4
  databricks/bundles/py.typed,sha256=8PjyZ1aVoQpRVvt71muvuq5qE-jTFZkK-GLHkhdebmc,26
5
- databricks/bundles/version.py,sha256=lJmiFMuCAbQPG7lhxILEKX6BtNHojVeMeiSNRIJb-Z0,24
5
+ databricks/bundles/version.py,sha256=OlkUF5JvR8KKM2wnWYT3gyANfvd2yNbToqDHJGCWKFA,24
6
6
  databricks/bundles/core/__init__.py,sha256=eivW2LIitfT-doDfPWac26vhdhkSACqoRi_48ZJE8zQ,1357
7
7
  databricks/bundles/core/_bundle.py,sha256=B5qINwnoRDitkofFaoCWDCS1tbB3a7nufqK0iDRfC6g,3287
8
8
  databricks/bundles/core/_diagnostics.py,sha256=arhPuRpjfOUjr8_T91-b-YM-ZtpkrLWeJv0BcLfTIOo,5879
@@ -11,7 +11,7 @@ databricks/bundles/core/_location.py,sha256=tMa-wGKOLCqOCk6bxVphZA4GWJxvjzAkDu8w
11
11
  databricks/bundles/core/_resource.py,sha256=ymi21F7e0Xy1fMrtL9ZMRXWQaqrK9F-t24w5vYtodqo,137
12
12
  databricks/bundles/core/_resource_mutator.py,sha256=agL3HOIubJCWuRi_AgSKC3Y-9kEYQRqocdbC7NoD9_U,5644
13
13
  databricks/bundles/core/_resource_type.py,sha256=fmcXqFPjBLC__HRfF_e4VN4lae2Oi37f1EI8XeMzQZ8,1797
14
- databricks/bundles/core/_resources.py,sha256=mJvqyjnOAG1GniRwKYhDSpRnsdqf2QNFA6rJwy6kEbc,11537
14
+ databricks/bundles/core/_resources.py,sha256=cvW_QbKeazTEJeD1FvmafPvWk29iqx5ruB3WmTbqb0k,11509
15
15
  databricks/bundles/core/_transform.py,sha256=ISupC7du3JnqrQQMgHf5Mt7fpZKEcUNZ5qwgMWZQfAE,8836
16
16
  databricks/bundles/core/_transform_to_json.py,sha256=aGiIBVx7pSO9LiJZ5CPYsTA5U6XhCbn1Ew15dhqC0yc,1890
17
17
  databricks/bundles/core/_variable.py,sha256=16g6vSLs5J-Ek2u2LNSnbqK-vpgbeirCddWjS-AiLGU,3617
@@ -23,7 +23,7 @@ databricks/bundles/jobs/_models/aws_attributes.py,sha256=pfZIHLVty93bMtfwkMKLPuU
23
23
  databricks/bundles/jobs/_models/aws_availability.py,sha256=C4yzZLt_CSIxlZ3MoCV4MPdQRjVRgRU_vkrLIkL6hkQ,477
24
24
  databricks/bundles/jobs/_models/azure_attributes.py,sha256=0w-lwBbr-6Xo2i5Ok6-9cwBoetwmzI3DA7kG81XRv6U,3764
25
25
  databricks/bundles/jobs/_models/azure_availability.py,sha256=KY8dt1vWA8RHBY3nq5CeqPjqhLyn5RvRfoHQ-Y2vlbk,540
26
- databricks/bundles/jobs/_models/clean_rooms_notebook_task.py,sha256=zgzDes0v3AEhd_2WXaq1xUo_hKtG927x4qzFzwvfbaY,2016
26
+ databricks/bundles/jobs/_models/clean_rooms_notebook_task.py,sha256=okf7ai_BbstscYTjWoA6xGmkm2pYTnaDPXUdcnS4Eaw,2175
27
27
  databricks/bundles/jobs/_models/clients_types.py,sha256=awwSptxl3zVXgTyoa_lO4JSp3QGVdRaxJBnxEUsmGwU,1199
28
28
  databricks/bundles/jobs/_models/cluster_log_conf.py,sha256=CP6C3nTXugvSm9-NlMPfENEcHn6Wl1q_9wqAFesos5o,2682
29
29
  databricks/bundles/jobs/_models/cluster_spec.py,sha256=u2u4xxksKD7qS552DgFUoKLH_AL2h91ayEN2inUfHcQ,19595
@@ -113,14 +113,14 @@ databricks/bundles/jobs/_models/webhook.py,sha256=S209r8QqufJLRoACU6a0MnTzuKOvn3
113
113
  databricks/bundles/jobs/_models/webhook_notifications.py,sha256=4FrMTYy4tDeMe3VqSbn9jjNYISTAmYTzENpGvnQGju4,4349
114
114
  databricks/bundles/jobs/_models/workload_type.py,sha256=A8KViUIB4x_gEXVS2p4KTGZ9Lr50Z3LLzIYxyE676xw,1162
115
115
  databricks/bundles/jobs/_models/workspace_storage_info.py,sha256=Qnm6lsw9rwXB7Te_Um0c7TvIH4Vv7ndKKYYV0pxJ6q8,1100
116
- databricks/bundles/pipelines/__init__.py,sha256=QoEMqHXXj87R5kobtE2GJv7GRu-1nYhsSUC3HsxKQRY,12067
116
+ databricks/bundles/pipelines/__init__.py,sha256=J9t4B5vd6iGPlyNfHqplwlH2rDi4liCEOesFAqnBiUM,11878
117
117
  databricks/bundles/pipelines/_models/adlsgen2_info.py,sha256=_eGe6ivi2VmokxKxKUji9-fSZLBubAr43uBDnN7vvlY,1104
118
118
  databricks/bundles/pipelines/_models/aws_attributes.py,sha256=sKKsOBfsg21soMJrdv3ETHIKg40LVGZWirlqSlQ2n8o,10191
119
119
  databricks/bundles/pipelines/_models/aws_availability.py,sha256=C4yzZLt_CSIxlZ3MoCV4MPdQRjVRgRU_vkrLIkL6hkQ,477
120
120
  databricks/bundles/pipelines/_models/azure_attributes.py,sha256=lh4epnkK9_8issxpxyLpnJ3uNwMNCsr9wBdaX8NfMlQ,3774
121
121
  databricks/bundles/pipelines/_models/azure_availability.py,sha256=KY8dt1vWA8RHBY3nq5CeqPjqhLyn5RvRfoHQ-Y2vlbk,540
122
122
  databricks/bundles/pipelines/_models/cluster_log_conf.py,sha256=EvCwe_ohAdv-mC5Veiqp28VFUaU9S5BeQA3KLLfEQGg,2697
123
- databricks/bundles/pipelines/_models/day_of_week.py,sha256=7pseZqzzKd3wlUPnUImoXWfC47Ntfqut6e0-vFXK8aU,606
123
+ databricks/bundles/pipelines/_models/day_of_week.py,sha256=dtri8UkWRZeuZm7o82pxFlJMPMcyYj-BhKzxWhYEJ60,554
124
124
  databricks/bundles/pipelines/_models/dbfs_storage_info.py,sha256=JLc9yrL3YOOVlYbQdA07tqF2wxkJdDvAKhQmAKURxaQ,972
125
125
  databricks/bundles/pipelines/_models/ebs_volume_type.py,sha256=-93BcybklhLyQEZvF8C1BhnHMeCjfDwI2qwimZ6X5eU,415
126
126
  databricks/bundles/pipelines/_models/event_log_spec.py,sha256=diWtjzD4xBbfgeEuVl_XYaOXEyTvSK40VLRue_1Ads0,1428
@@ -131,11 +131,10 @@ databricks/bundles/pipelines/_models/gcp_availability.py,sha256=a2ayWsyEQDpIDx-m
131
131
  databricks/bundles/pipelines/_models/gcs_storage_info.py,sha256=hwOowyNKCBhzsUiCQSrtmQPxrMINEq5jg2EefkrE2fQ,1020
132
132
  databricks/bundles/pipelines/_models/ingestion_config.py,sha256=ERT5ySyVIVJ0T3r3tqjuCFzV5FEIJsfWZvbMmRTHNMk,1679
133
133
  databricks/bundles/pipelines/_models/ingestion_gateway_pipeline_definition.py,sha256=UkpHUaelJYcEZ-3r8FRwEpUcQxHrvQoSVsmadmlQaBU,3030
134
- databricks/bundles/pipelines/_models/ingestion_pipeline_definition.py,sha256=vTzpSDIHfmp5U8s1yjX56VLwkqHzN5ynZEQkYAkIPZY,4603
134
+ databricks/bundles/pipelines/_models/ingestion_pipeline_definition.py,sha256=KHYododL_q3Z8yDvog793nyOJG6Q1ID9odFakRO6SVs,3943
135
135
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_table_specific_config_query_based_connector_config.py,sha256=IXJqQjQ2aORHJZy1PZ3ST4hx4HVh7O4b-JoKoxYR9QE,5181
136
136
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_workday_report_parameters.py,sha256=3foG3H-6eOTpQbH0JqK3EY27isg6ZCm8w88cufh6HRU,3434
137
137
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_workday_report_parameters_query_key_value.py,sha256=mC5r259VRZKnd0WKi1Toeb1qG2VZglumzVos4Wv8BoQ,2261
138
- databricks/bundles/pipelines/_models/ingestion_source_type.py,sha256=WjyeHeyW4iNOXlMsoXGoO57qKLuXYg6BkSvB5d5mLxA,1147
139
138
  databricks/bundles/pipelines/_models/init_script_info.py,sha256=PM1qnkkC_fR97DSYML2-UJLGFKl1bGvttESfKosFSM8,4546
140
139
  databricks/bundles/pipelines/_models/lifecycle.py,sha256=_0h3QZ8LPYTnKC8YDRmW0NbQBxwukG1WsqBu8wZrtO4,985
141
140
  databricks/bundles/pipelines/_models/local_file_info.py,sha256=yH12cJKjPrFQxG69DgAdp87PIYVGfjECvabbRPqKZjI,952
@@ -144,7 +143,7 @@ databricks/bundles/pipelines/_models/maven_library.py,sha256=xz7BIo3XZ4xfp9S3sov
144
143
  databricks/bundles/pipelines/_models/notebook_library.py,sha256=YFEBdlvoNfB3oLPz-w-n_HBQrRVzFD9pbu-BPza88Rk,951
145
144
  databricks/bundles/pipelines/_models/notifications.py,sha256=Q7xHA5Bii9Zhgr2TISYF9mWKqu-6RzGO76gLexLMM3c,1987
146
145
  databricks/bundles/pipelines/_models/path_pattern.py,sha256=X3DRx7GiZzaUFC_lHRcZFGdbmUB3YyZfrgu9TWZH3LM,935
147
- databricks/bundles/pipelines/_models/pipeline.py,sha256=596tfHmrSiR5s3SwEw3ocS2-dNXWjPo6dPu5pAbc3hg,11371
146
+ databricks/bundles/pipelines/_models/pipeline.py,sha256=G0jJKu6a1jHWZc4QS7PZYnetps6kwii-nUv4S5NNJL0,11572
148
147
  databricks/bundles/pipelines/_models/pipeline_cluster.py,sha256=O7orq6-TZex8UpHrQLozM8Eq55bGlcDw2byB7wV4p_k,13306
149
148
  databricks/bundles/pipelines/_models/pipeline_cluster_autoscale.py,sha256=tKrqppzu25MVaPo5tSPjhHuTGXBYwvXLJ6pSGcTUaes,2685
150
149
  databricks/bundles/pipelines/_models/pipeline_cluster_autoscale_mode.py,sha256=WhhfyIZEI4jlqoj9ks9lLYeTxRxLVLcP6I30Wqx6p8A,619
@@ -177,7 +176,7 @@ databricks/bundles/volumes/_models/volume.py,sha256=ALGmeXW3rGH424pp6SaXPT1I87XX
177
176
  databricks/bundles/volumes/_models/volume_grant.py,sha256=U_-4-KL8LM3n5xJBLHj_wjPsqiVjCDRj8ttiUYqFRmI,1083
178
177
  databricks/bundles/volumes/_models/volume_grant_privilege.py,sha256=fCA0LVE9Q3sbHvTAj7e62E9ASq9jH5oK1iREQdp1TxQ,384
179
178
  databricks/bundles/volumes/_models/volume_type.py,sha256=kdczwT3EJ0n5ZBV75SgX-6766igJBLTh6ywaaxIbLDk,189
180
- databricks_bundles-0.274.0.dist-info/licenses/LICENSE,sha256=QKOZO8KtzbS_Qt3Tbl0dfGnidaeilKe0UiIjnEq1tjc,3790
181
- databricks_bundles-0.274.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
182
- databricks_bundles-0.274.0.dist-info/METADATA,sha256=f_D7z0w4JX72Sa0YX7G5UKwxq2vv_aVLYpptV1lYn-U,1541
183
- databricks_bundles-0.274.0.dist-info/RECORD,,
179
+ databricks_bundles-0.276.0.dist-info/licenses/LICENSE,sha256=QKOZO8KtzbS_Qt3Tbl0dfGnidaeilKe0UiIjnEq1tjc,3790
180
+ databricks_bundles-0.276.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
181
+ databricks_bundles-0.276.0.dist-info/METADATA,sha256=IQm3Sc_tCi7KFGB1H0eO9opWw6pBRVfOwfJ8Vj0fLso,1541
182
+ databricks_bundles-0.276.0.dist-info/RECORD,,
@@ -1,50 +0,0 @@
1
- from enum import Enum
2
- from typing import Literal
3
-
4
-
5
- class IngestionSourceType(Enum):
6
- MYSQL = "MYSQL"
7
- POSTGRESQL = "POSTGRESQL"
8
- REDSHIFT = "REDSHIFT"
9
- SQLDW = "SQLDW"
10
- SQLSERVER = "SQLSERVER"
11
- SALESFORCE = "SALESFORCE"
12
- BIGQUERY = "BIGQUERY"
13
- NETSUITE = "NETSUITE"
14
- WORKDAY_RAAS = "WORKDAY_RAAS"
15
- GA4_RAW_DATA = "GA4_RAW_DATA"
16
- SERVICENOW = "SERVICENOW"
17
- MANAGED_POSTGRESQL = "MANAGED_POSTGRESQL"
18
- ORACLE = "ORACLE"
19
- TERADATA = "TERADATA"
20
- SHAREPOINT = "SHAREPOINT"
21
- DYNAMICS365 = "DYNAMICS365"
22
- CONFLUENCE = "CONFLUENCE"
23
- META_MARKETING = "META_MARKETING"
24
- FOREIGN_CATALOG = "FOREIGN_CATALOG"
25
-
26
-
27
- IngestionSourceTypeParam = (
28
- Literal[
29
- "MYSQL",
30
- "POSTGRESQL",
31
- "REDSHIFT",
32
- "SQLDW",
33
- "SQLSERVER",
34
- "SALESFORCE",
35
- "BIGQUERY",
36
- "NETSUITE",
37
- "WORKDAY_RAAS",
38
- "GA4_RAW_DATA",
39
- "SERVICENOW",
40
- "MANAGED_POSTGRESQL",
41
- "ORACLE",
42
- "TERADATA",
43
- "SHAREPOINT",
44
- "DYNAMICS365",
45
- "CONFLUENCE",
46
- "META_MARKETING",
47
- "FOREIGN_CATALOG",
48
- ]
49
- | IngestionSourceType
50
- )