databricks-bundles 0.284.0__py3-none-any.whl → 0.285.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -153,6 +153,9 @@ __all__ = [
153
153
  "ModelTriggerConfigurationConditionParam",
154
154
  "ModelTriggerConfigurationDict",
155
155
  "ModelTriggerConfigurationParam",
156
+ "NodeTypeFlexibility",
157
+ "NodeTypeFlexibilityDict",
158
+ "NodeTypeFlexibilityParam",
156
159
  "NotebookTask",
157
160
  "NotebookTaskDict",
158
161
  "NotebookTaskParam",
@@ -522,6 +525,11 @@ from databricks.bundles.jobs._models.model_trigger_configuration_condition impor
522
525
  ModelTriggerConfigurationCondition,
523
526
  ModelTriggerConfigurationConditionParam,
524
527
  )
528
+ from databricks.bundles.jobs._models.node_type_flexibility import (
529
+ NodeTypeFlexibility,
530
+ NodeTypeFlexibilityDict,
531
+ NodeTypeFlexibilityParam,
532
+ )
525
533
  from databricks.bundles.jobs._models.notebook_task import (
526
534
  NotebookTask,
527
535
  NotebookTaskDict,
@@ -38,6 +38,10 @@ from databricks.bundles.jobs._models.init_script_info import (
38
38
  InitScriptInfoParam,
39
39
  )
40
40
  from databricks.bundles.jobs._models.kind import Kind, KindParam
41
+ from databricks.bundles.jobs._models.node_type_flexibility import (
42
+ NodeTypeFlexibility,
43
+ NodeTypeFlexibilityParam,
44
+ )
41
45
  from databricks.bundles.jobs._models.runtime_engine import (
42
46
  RuntimeEngine,
43
47
  RuntimeEngineParam,
@@ -125,6 +129,11 @@ class ClusterSpec:
125
129
  assigned.
126
130
  """
127
131
 
132
+ driver_node_type_flexibility: VariableOrOptional[NodeTypeFlexibility] = None
133
+ """
134
+ Flexible node type configuration for the driver node.
135
+ """
136
+
128
137
  driver_node_type_id: VariableOrOptional[str] = None
129
138
  """
130
139
  The node type of the Spark driver.
@@ -258,6 +267,11 @@ class ClusterSpec:
258
267
  `effective_spark_version` is determined by `spark_version` (DBR release), this field `use_ml_runtime`, and whether `node_type_id` is gpu node or not.
259
268
  """
260
269
 
270
+ worker_node_type_flexibility: VariableOrOptional[NodeTypeFlexibility] = None
271
+ """
272
+ Flexible node type configuration for worker nodes.
273
+ """
274
+
261
275
  workload_type: VariableOrOptional[WorkloadType] = None
262
276
 
263
277
  @classmethod
@@ -339,6 +353,11 @@ class ClusterSpecDict(TypedDict, total=False):
339
353
  assigned.
340
354
  """
341
355
 
356
+ driver_node_type_flexibility: VariableOrOptional[NodeTypeFlexibilityParam]
357
+ """
358
+ Flexible node type configuration for the driver node.
359
+ """
360
+
342
361
  driver_node_type_id: VariableOrOptional[str]
343
362
  """
344
363
  The node type of the Spark driver.
@@ -472,6 +491,11 @@ class ClusterSpecDict(TypedDict, total=False):
472
491
  `effective_spark_version` is determined by `spark_version` (DBR release), this field `use_ml_runtime`, and whether `node_type_id` is gpu node or not.
473
492
  """
474
493
 
494
+ worker_node_type_flexibility: VariableOrOptional[NodeTypeFlexibilityParam]
495
+ """
496
+ Flexible node type configuration for worker nodes.
497
+ """
498
+
475
499
  workload_type: VariableOrOptional[WorkloadTypeParam]
476
500
 
477
501
 
@@ -4,10 +4,7 @@ from typing import TYPE_CHECKING, TypedDict
4
4
  from databricks.bundles.core._transform import _transform
5
5
  from databricks.bundles.core._transform_to_json import _transform_to_json_value
6
6
  from databricks.bundles.core._variable import VariableOrOptional
7
- from databricks.bundles.jobs._models.subscription import (
8
- Subscription,
9
- SubscriptionParam,
10
- )
7
+ from databricks.bundles.jobs._models.subscription import Subscription, SubscriptionParam
11
8
 
12
9
  if TYPE_CHECKING:
13
10
  from typing_extensions import Self
@@ -9,10 +9,7 @@ from databricks.bundles.core._variable import (
9
9
  VariableOrList,
10
10
  VariableOrOptional,
11
11
  )
12
- from databricks.bundles.jobs._models.continuous import (
13
- Continuous,
14
- ContinuousParam,
15
- )
12
+ from databricks.bundles.jobs._models.continuous import Continuous, ContinuousParam
16
13
  from databricks.bundles.jobs._models.cron_schedule import (
17
14
  CronSchedule,
18
15
  CronScheduleParam,
@@ -0,0 +1,40 @@
1
+ from dataclasses import dataclass, field
2
+ from typing import TYPE_CHECKING, TypedDict
3
+
4
+ from databricks.bundles.core._transform import _transform
5
+ from databricks.bundles.core._transform_to_json import _transform_to_json_value
6
+ from databricks.bundles.core._variable import VariableOrList
7
+
8
+ if TYPE_CHECKING:
9
+ from typing_extensions import Self
10
+
11
+
12
+ @dataclass(kw_only=True)
13
+ class NodeTypeFlexibility:
14
+ """
15
+ Configuration for flexible node types, allowing fallback to alternate node types during cluster launch and upscale.
16
+ """
17
+
18
+ alternate_node_type_ids: VariableOrList[str] = field(default_factory=list)
19
+ """
20
+ A list of node type IDs to use as fallbacks when the primary node type is unavailable.
21
+ """
22
+
23
+ @classmethod
24
+ def from_dict(cls, value: "NodeTypeFlexibilityDict") -> "Self":
25
+ return _transform(cls, value)
26
+
27
+ def as_dict(self) -> "NodeTypeFlexibilityDict":
28
+ return _transform_to_json_value(self) # type:ignore
29
+
30
+
31
+ class NodeTypeFlexibilityDict(TypedDict, total=False):
32
+ """"""
33
+
34
+ alternate_node_type_ids: VariableOrList[str]
35
+ """
36
+ A list of node type IDs to use as fallbacks when the primary node type is unavailable.
37
+ """
38
+
39
+
40
+ NodeTypeFlexibilityParam = NodeTypeFlexibilityDict | NodeTypeFlexibility
@@ -2,6 +2,9 @@ __all__ = [
2
2
  "Adlsgen2Info",
3
3
  "Adlsgen2InfoDict",
4
4
  "Adlsgen2InfoParam",
5
+ "AutoFullRefreshPolicy",
6
+ "AutoFullRefreshPolicyDict",
7
+ "AutoFullRefreshPolicyParam",
5
8
  "AwsAttributes",
6
9
  "AwsAttributesDict",
7
10
  "AwsAttributesParam",
@@ -81,6 +84,9 @@ __all__ = [
81
84
  "Notifications",
82
85
  "NotificationsDict",
83
86
  "NotificationsParam",
87
+ "OperationTimeWindow",
88
+ "OperationTimeWindowDict",
89
+ "OperationTimeWindowParam",
84
90
  "PathPattern",
85
91
  "PathPatternDict",
86
92
  "PathPatternParam",
@@ -155,6 +161,11 @@ from databricks.bundles.pipelines._models.adlsgen2_info import (
155
161
  Adlsgen2InfoDict,
156
162
  Adlsgen2InfoParam,
157
163
  )
164
+ from databricks.bundles.pipelines._models.auto_full_refresh_policy import (
165
+ AutoFullRefreshPolicy,
166
+ AutoFullRefreshPolicyDict,
167
+ AutoFullRefreshPolicyParam,
168
+ )
158
169
  from databricks.bundles.pipelines._models.aws_attributes import (
159
170
  AwsAttributes,
160
171
  AwsAttributesDict,
@@ -287,6 +298,11 @@ from databricks.bundles.pipelines._models.notifications import (
287
298
  NotificationsDict,
288
299
  NotificationsParam,
289
300
  )
301
+ from databricks.bundles.pipelines._models.operation_time_window import (
302
+ OperationTimeWindow,
303
+ OperationTimeWindowDict,
304
+ OperationTimeWindowParam,
305
+ )
290
306
  from databricks.bundles.pipelines._models.path_pattern import (
291
307
  PathPattern,
292
308
  PathPatternDict,
@@ -0,0 +1,54 @@
1
+ from dataclasses import dataclass
2
+ from typing import TYPE_CHECKING, TypedDict
3
+
4
+ from databricks.bundles.core._transform import _transform
5
+ from databricks.bundles.core._transform_to_json import _transform_to_json_value
6
+ from databricks.bundles.core._variable import VariableOr, VariableOrOptional
7
+
8
+ if TYPE_CHECKING:
9
+ from typing_extensions import Self
10
+
11
+
12
+ @dataclass(kw_only=True)
13
+ class AutoFullRefreshPolicy:
14
+ """
15
+ Policy for auto full refresh.
16
+ """
17
+
18
+ enabled: VariableOr[bool]
19
+ """
20
+ (Required, Mutable) Whether to enable auto full refresh or not.
21
+ """
22
+
23
+ min_interval_hours: VariableOrOptional[int] = None
24
+ """
25
+ (Optional, Mutable) Specify the minimum interval in hours between the timestamp
26
+ at which a table was last full refreshed and the current timestamp for triggering auto full
27
+ If unspecified and autoFullRefresh is enabled then by default min_interval_hours is 24 hours.
28
+ """
29
+
30
+ @classmethod
31
+ def from_dict(cls, value: "AutoFullRefreshPolicyDict") -> "Self":
32
+ return _transform(cls, value)
33
+
34
+ def as_dict(self) -> "AutoFullRefreshPolicyDict":
35
+ return _transform_to_json_value(self) # type:ignore
36
+
37
+
38
+ class AutoFullRefreshPolicyDict(TypedDict, total=False):
39
+ """"""
40
+
41
+ enabled: VariableOr[bool]
42
+ """
43
+ (Required, Mutable) Whether to enable auto full refresh or not.
44
+ """
45
+
46
+ min_interval_hours: VariableOrOptional[int]
47
+ """
48
+ (Optional, Mutable) Specify the minimum interval in hours between the timestamp
49
+ at which a table was last full refreshed and the current timestamp for triggering auto full
50
+ If unspecified and autoFullRefresh is enabled then by default min_interval_hours is 24 hours.
51
+ """
52
+
53
+
54
+ AutoFullRefreshPolicyParam = AutoFullRefreshPolicyDict | AutoFullRefreshPolicy
@@ -4,8 +4,6 @@ from typing import Literal
4
4
 
5
5
  class DayOfWeek(Enum):
6
6
  """
7
- :meta private: [EXPERIMENTAL]
8
-
9
7
  Days of week in which the window is allowed to happen.
10
8
  If not specified all days of the week will be used.
11
9
  """
@@ -8,6 +8,10 @@ from databricks.bundles.pipelines._models.ingestion_config import (
8
8
  IngestionConfig,
9
9
  IngestionConfigParam,
10
10
  )
11
+ from databricks.bundles.pipelines._models.operation_time_window import (
12
+ OperationTimeWindow,
13
+ OperationTimeWindowParam,
14
+ )
11
15
  from databricks.bundles.pipelines._models.source_config import (
12
16
  SourceConfig,
13
17
  SourceConfigParam,
@@ -30,6 +34,11 @@ class IngestionPipelineDefinition:
30
34
  Immutable. The Unity Catalog connection that this ingestion pipeline uses to communicate with the source. This is used with connectors for applications like Salesforce, Workday, and so on.
31
35
  """
32
36
 
37
+ full_refresh_window: VariableOrOptional[OperationTimeWindow] = None
38
+ """
39
+ (Optional) A window that specifies a set of time ranges for snapshot queries in CDC.
40
+ """
41
+
33
42
  ingest_from_uc_foreign_catalog: VariableOrOptional[bool] = None
34
43
  """
35
44
  :meta private: [EXPERIMENTAL]
@@ -81,6 +90,11 @@ class IngestionPipelineDefinitionDict(TypedDict, total=False):
81
90
  Immutable. The Unity Catalog connection that this ingestion pipeline uses to communicate with the source. This is used with connectors for applications like Salesforce, Workday, and so on.
82
91
  """
83
92
 
93
+ full_refresh_window: VariableOrOptional[OperationTimeWindowParam]
94
+ """
95
+ (Optional) A window that specifies a set of time ranges for snapshot queries in CDC.
96
+ """
97
+
84
98
  ingest_from_uc_foreign_catalog: VariableOrOptional[bool]
85
99
  """
86
100
  :meta private: [EXPERIMENTAL]
@@ -0,0 +1,69 @@
1
+ from dataclasses import dataclass, field
2
+ from typing import TYPE_CHECKING, TypedDict
3
+
4
+ from databricks.bundles.core._transform import _transform
5
+ from databricks.bundles.core._transform_to_json import _transform_to_json_value
6
+ from databricks.bundles.core._variable import (
7
+ VariableOr,
8
+ VariableOrList,
9
+ VariableOrOptional,
10
+ )
11
+ from databricks.bundles.pipelines._models.day_of_week import DayOfWeek, DayOfWeekParam
12
+
13
+ if TYPE_CHECKING:
14
+ from typing_extensions import Self
15
+
16
+
17
+ @dataclass(kw_only=True)
18
+ class OperationTimeWindow:
19
+ """
20
+ Proto representing a window
21
+ """
22
+
23
+ start_hour: VariableOr[int]
24
+ """
25
+ An integer between 0 and 23 denoting the start hour for the window in the 24-hour day.
26
+ """
27
+
28
+ days_of_week: VariableOrList[DayOfWeek] = field(default_factory=list)
29
+ """
30
+ Days of week in which the window is allowed to happen
31
+ If not specified all days of the week will be used.
32
+ """
33
+
34
+ time_zone_id: VariableOrOptional[str] = None
35
+ """
36
+ Time zone id of window. See https://docs.databricks.com/sql/language-manual/sql-ref-syntax-aux-conf-mgmt-set-timezone.html for details.
37
+ If not specified, UTC will be used.
38
+ """
39
+
40
+ @classmethod
41
+ def from_dict(cls, value: "OperationTimeWindowDict") -> "Self":
42
+ return _transform(cls, value)
43
+
44
+ def as_dict(self) -> "OperationTimeWindowDict":
45
+ return _transform_to_json_value(self) # type:ignore
46
+
47
+
48
+ class OperationTimeWindowDict(TypedDict, total=False):
49
+ """"""
50
+
51
+ start_hour: VariableOr[int]
52
+ """
53
+ An integer between 0 and 23 denoting the start hour for the window in the 24-hour day.
54
+ """
55
+
56
+ days_of_week: VariableOrList[DayOfWeekParam]
57
+ """
58
+ Days of week in which the window is allowed to happen
59
+ If not specified all days of the week will be used.
60
+ """
61
+
62
+ time_zone_id: VariableOrOptional[str]
63
+ """
64
+ Time zone id of window. See https://docs.databricks.com/sql/language-manual/sql-ref-syntax-aux-conf-mgmt-set-timezone.html for details.
65
+ If not specified, UTC will be used.
66
+ """
67
+
68
+
69
+ OperationTimeWindowParam = OperationTimeWindowDict | OperationTimeWindow
@@ -4,6 +4,10 @@ from typing import TYPE_CHECKING, TypedDict
4
4
  from databricks.bundles.core._transform import _transform
5
5
  from databricks.bundles.core._transform_to_json import _transform_to_json_value
6
6
  from databricks.bundles.core._variable import VariableOrList, VariableOrOptional
7
+ from databricks.bundles.pipelines._models.auto_full_refresh_policy import (
8
+ AutoFullRefreshPolicy,
9
+ AutoFullRefreshPolicyParam,
10
+ )
7
11
  from databricks.bundles.pipelines._models.ingestion_pipeline_definition_table_specific_config_query_based_connector_config import (
8
12
  IngestionPipelineDefinitionTableSpecificConfigQueryBasedConnectorConfig,
9
13
  IngestionPipelineDefinitionTableSpecificConfigQueryBasedConnectorConfigParam,
@@ -25,6 +29,21 @@ if TYPE_CHECKING:
25
29
  class TableSpecificConfig:
26
30
  """"""
27
31
 
32
+ auto_full_refresh_policy: VariableOrOptional[AutoFullRefreshPolicy] = None
33
+ """
34
+ (Optional, Mutable) Policy for auto full refresh, if enabled pipeline will automatically try
35
+ to fix issues by doing a full refresh on the table in the retry run. auto_full_refresh_policy
36
+ in table configuration will override the above level auto_full_refresh_policy.
37
+ For example,
38
+ {
39
+ "auto_full_refresh_policy": {
40
+ "enabled": true,
41
+ "min_interval_hours": 23,
42
+ }
43
+ }
44
+ If unspecified, auto full refresh is disabled.
45
+ """
46
+
28
47
  exclude_columns: VariableOrList[str] = field(default_factory=list)
29
48
  """
30
49
  A list of column names to be excluded for the ingestion.
@@ -102,6 +121,21 @@ class TableSpecificConfig:
102
121
  class TableSpecificConfigDict(TypedDict, total=False):
103
122
  """"""
104
123
 
124
+ auto_full_refresh_policy: VariableOrOptional[AutoFullRefreshPolicyParam]
125
+ """
126
+ (Optional, Mutable) Policy for auto full refresh, if enabled pipeline will automatically try
127
+ to fix issues by doing a full refresh on the table in the retry run. auto_full_refresh_policy
128
+ in table configuration will override the above level auto_full_refresh_policy.
129
+ For example,
130
+ {
131
+ "auto_full_refresh_policy": {
132
+ "enabled": true,
133
+ "min_interval_hours": 23,
134
+ }
135
+ }
136
+ If unspecified, auto full refresh is disabled.
137
+ """
138
+
105
139
  exclude_columns: VariableOrList[str]
106
140
  """
107
141
  A list of column names to be excluded for the ingestion.
@@ -1 +1 @@
1
- __version__ = "0.284.0"
1
+ __version__ = "0.285.0"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: databricks-bundles
3
- Version: 0.284.0
3
+ Version: 0.285.0
4
4
  Summary: Python support for Databricks Asset Bundles
5
5
  Author-email: Gleb Kanterov <gleb.kanterov@databricks.com>
6
6
  Requires-Python: >=3.10
@@ -22,7 +22,7 @@ Reference documentation is available at https://databricks.github.io/cli/python/
22
22
 
23
23
  To use `databricks-bundles`, you must first:
24
24
 
25
- 1. Install the [Databricks CLI](https://github.com/databricks/cli), version 0.284.0 or above
25
+ 1. Install the [Databricks CLI](https://github.com/databricks/cli), version 0.285.0 or above
26
26
  2. Authenticate to your Databricks workspace if you have not done so already:
27
27
 
28
28
  ```bash
@@ -2,7 +2,7 @@ databricks/__init__.py,sha256=CF2MJcZFwbpn9TwQER8qnCDhkPooBGQNVkX4v7g6p3g,537
2
2
  databricks/bundles/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
3
3
  databricks/bundles/build.py,sha256=WpMZiPcyYqmUeimmYlXlV3RLTUFFUQrUzG83q4Zb4xA,17817
4
4
  databricks/bundles/py.typed,sha256=8PjyZ1aVoQpRVvt71muvuq5qE-jTFZkK-GLHkhdebmc,26
5
- databricks/bundles/version.py,sha256=V1_n1mF3NPAXSq7h4G-DWecwhIzZCIYCQoFdmkK70Fg,24
5
+ databricks/bundles/version.py,sha256=g93r5O01Gj5nsOSfCgPbONbiZqSeNNjMhlNIf2xrigc,24
6
6
  databricks/bundles/core/__init__.py,sha256=eivW2LIitfT-doDfPWac26vhdhkSACqoRi_48ZJE8zQ,1357
7
7
  databricks/bundles/core/_bundle.py,sha256=B5qINwnoRDitkofFaoCWDCS1tbB3a7nufqK0iDRfC6g,3287
8
8
  databricks/bundles/core/_diagnostics.py,sha256=arhPuRpjfOUjr8_T91-b-YM-ZtpkrLWeJv0BcLfTIOo,5879
@@ -15,7 +15,7 @@ databricks/bundles/core/_resources.py,sha256=cvW_QbKeazTEJeD1FvmafPvWk29iqx5ruB3
15
15
  databricks/bundles/core/_transform.py,sha256=ISupC7du3JnqrQQMgHf5Mt7fpZKEcUNZ5qwgMWZQfAE,8836
16
16
  databricks/bundles/core/_transform_to_json.py,sha256=aGiIBVx7pSO9LiJZ5CPYsTA5U6XhCbn1Ew15dhqC0yc,1890
17
17
  databricks/bundles/core/_variable.py,sha256=16g6vSLs5J-Ek2u2LNSnbqK-vpgbeirCddWjS-AiLGU,3617
18
- databricks/bundles/jobs/__init__.py,sha256=iYkL1EP3lcaXLrFkAe6EkZ1a1MhbEViL_E2IMU6sOPg,19969
18
+ databricks/bundles/jobs/__init__.py,sha256=Ur44V10jRuf9kYXE9C3qookfWCc8c58b7wbv2DDj57I,20213
19
19
  databricks/bundles/jobs/_models/adlsgen2_info.py,sha256=_eGe6ivi2VmokxKxKUji9-fSZLBubAr43uBDnN7vvlY,1104
20
20
  databricks/bundles/jobs/_models/authentication_method.py,sha256=XI8hU5fhPlGTsZdePZtR6FIjyT4iEt2URb61q1MsgNI,198
21
21
  databricks/bundles/jobs/_models/auto_scale.py,sha256=Z5vZa0bZi6LJ_Ac-VJfqqCJAtI-zY1_auTGhsV8khvA,1624
@@ -26,14 +26,14 @@ databricks/bundles/jobs/_models/azure_availability.py,sha256=KY8dt1vWA8RHBY3nq5C
26
26
  databricks/bundles/jobs/_models/clean_rooms_notebook_task.py,sha256=okf7ai_BbstscYTjWoA6xGmkm2pYTnaDPXUdcnS4Eaw,2175
27
27
  databricks/bundles/jobs/_models/clients_types.py,sha256=awwSptxl3zVXgTyoa_lO4JSp3QGVdRaxJBnxEUsmGwU,1199
28
28
  databricks/bundles/jobs/_models/cluster_log_conf.py,sha256=CP6C3nTXugvSm9-NlMPfENEcHn6Wl1q_9wqAFesos5o,2682
29
- databricks/bundles/jobs/_models/cluster_spec.py,sha256=yFX_TMnBENE37Blw4XaxvfzOV-tWreBm361Zzc_761E,19361
29
+ databricks/bundles/jobs/_models/cluster_spec.py,sha256=-J_U62Ugj03bqfU9z7S8EexbqjozjDCsmOPgiB8HKBQ,20100
30
30
  databricks/bundles/jobs/_models/compute_config.py,sha256=61-BdovRQ48n93GlPc3bjfy_U5TJ6ERQrViQzTnAg9s,1267
31
31
  databricks/bundles/jobs/_models/condition.py,sha256=iZhVmAJCXRbAWGhusB3No7QTMZnfoQmp_hvYpNvjkf8,207
32
32
  databricks/bundles/jobs/_models/condition_task.py,sha256=4A71BNRZNk9_iS8IFiHr69RLgx_qmabnQUwcy_g7e0E,2814
33
33
  databricks/bundles/jobs/_models/condition_task_op.py,sha256=mj5tx5qmvUCIkDvFYbo1-6fvVP_Q1zYzxv011ufHbx4,1141
34
34
  databricks/bundles/jobs/_models/continuous.py,sha256=ILsXONBdaHnXoNofLJj5CWu4Hk3NGpvS4izBKOwZHL0,1609
35
35
  databricks/bundles/jobs/_models/cron_schedule.py,sha256=_dUQ8vdfnZvIPvNyB-qdn_9cKKGrys249Wv_SWDdP7Q,2186
36
- databricks/bundles/jobs/_models/dashboard_task.py,sha256=sh-yHCUoDEj-WAe1w5gDz6luRNg2dgKTK42DlKiNZbo,1546
36
+ databricks/bundles/jobs/_models/dashboard_task.py,sha256=HTbz7Tra_gvnPNW2Skug-k-bxjwbY7xu0cq5nlGRYh8,1533
37
37
  databricks/bundles/jobs/_models/data_security_mode.py,sha256=fuelzF06CfyhG9OmK-2orH5EifQnEtPTEmh7NZcL7zA,2660
38
38
  databricks/bundles/jobs/_models/dbfs_storage_info.py,sha256=JLc9yrL3YOOVlYbQdA07tqF2wxkJdDvAKhQmAKURxaQ,972
39
39
  databricks/bundles/jobs/_models/dbt_platform_task.py,sha256=a_cR8fCEIpN0rAcoLEAy_UCJQAgu1q0XkPZlVhy8PAQ,1491
@@ -51,7 +51,7 @@ databricks/bundles/jobs/_models/gen_ai_compute_task.py,sha256=WNjYn_s4bVlB-msh7V
51
51
  databricks/bundles/jobs/_models/git_provider.py,sha256=VzNKsIrOJauZjCTCeq3zVUumrnZoamYXxpgSz0kPsEI,664
52
52
  databricks/bundles/jobs/_models/git_source.py,sha256=0WVNzvbdMRvFC4N-fH2Jw1sh-kuNCt7WVkO5E1Jy5a8,3031
53
53
  databricks/bundles/jobs/_models/init_script_info.py,sha256=Skv-u-naj3GRdmwvnxbMIDdOj3YU9PARTBMCbTNKfN8,4511
54
- databricks/bundles/jobs/_models/job.py,sha256=I-iTxHccP0thklivztWBmy_mqw-lApKXy1NEi_rD-7Q,16069
54
+ databricks/bundles/jobs/_models/job.py,sha256=mFeIy9IUYFWsf0rukmoKnWU2hbblEoFwejPIokoS_EA,16056
55
55
  databricks/bundles/jobs/_models/job_cluster.py,sha256=b9mQ993-tmY_KZem8vRCT7MLcw07KhKEfvmyWxoGav0,1595
56
56
  databricks/bundles/jobs/_models/job_email_notifications.py,sha256=rnkjkZ06YTsHLkbjSDGV9Lm5o9r6fNSG9KwGA2rlyjY,5308
57
57
  databricks/bundles/jobs/_models/job_environment.py,sha256=ekMKwai2LtB9Dzg3LyGmoVjLCC6ShTWBjW54_B-qOBo,1185
@@ -72,6 +72,7 @@ databricks/bundles/jobs/_models/log_analytics_info.py,sha256=JeEeyNcldckin93yy4x
72
72
  databricks/bundles/jobs/_models/maven_library.py,sha256=xz7BIo3XZ4xfp9S3sovADLxDY_DVcZbqduS8VTo1La4,2002
73
73
  databricks/bundles/jobs/_models/model_trigger_configuration.py,sha256=_eIyPPfBfHQkQPN7MpQOaA3_EvUIFDc8vmvg7TVF6oI,3416
74
74
  databricks/bundles/jobs/_models/model_trigger_configuration_condition.py,sha256=QdaXDlfMtuEV5K8Wh2hSn5AHfzdMsemz2InFQCq_2zw,434
75
+ databricks/bundles/jobs/_models/node_type_flexibility.py,sha256=xRuhfDNg4op5MnnI1-UXuyyxnlLnXRdI-b0xR7qnpAI,1258
75
76
  databricks/bundles/jobs/_models/notebook_task.py,sha256=e1u7uNHTgW8Y9fAA-tcrAeo__bIN_KyWaca41DYzEOc,5097
76
77
  databricks/bundles/jobs/_models/pause_status.py,sha256=4Zsd3vXwRXLBXO1LLauX6TkdbrfRJYzJxVeTSfaapNk,189
77
78
  databricks/bundles/jobs/_models/performance_target.py,sha256=pSefsgLvGSZ723xQEmHDHt-scofmsrQQGCHhr_vAddI,572
@@ -115,15 +116,16 @@ databricks/bundles/jobs/_models/webhook.py,sha256=S209r8QqufJLRoACU6a0MnTzuKOvn3
115
116
  databricks/bundles/jobs/_models/webhook_notifications.py,sha256=4FrMTYy4tDeMe3VqSbn9jjNYISTAmYTzENpGvnQGju4,4349
116
117
  databricks/bundles/jobs/_models/workload_type.py,sha256=A8KViUIB4x_gEXVS2p4KTGZ9Lr50Z3LLzIYxyE676xw,1162
117
118
  databricks/bundles/jobs/_models/workspace_storage_info.py,sha256=Qnm6lsw9rwXB7Te_Um0c7TvIH4Vv7ndKKYYV0pxJ6q8,1100
118
- databricks/bundles/pipelines/__init__.py,sha256=_BOWtW0nBP1ndeBH08uTOvMCEJEtAiIj-MWba2ZThRk,12133
119
+ databricks/bundles/pipelines/__init__.py,sha256=dJXqxxH3QAlfs1B8goqa54uVUzAstgv15ODszQODdNE,12646
119
120
  databricks/bundles/pipelines/_models/adlsgen2_info.py,sha256=_eGe6ivi2VmokxKxKUji9-fSZLBubAr43uBDnN7vvlY,1104
121
+ databricks/bundles/pipelines/_models/auto_full_refresh_policy.py,sha256=G9ga-LW_GOUn_jdr0evdt4bnvGDC_-YefE_ybGEDP44,1774
120
122
  databricks/bundles/pipelines/_models/aws_attributes.py,sha256=_YVduqrylaoZr_JQh15VIKtu5b9rzG1-1f5sAvxElbA,10193
121
123
  databricks/bundles/pipelines/_models/aws_availability.py,sha256=C4yzZLt_CSIxlZ3MoCV4MPdQRjVRgRU_vkrLIkL6hkQ,477
122
124
  databricks/bundles/pipelines/_models/azure_attributes.py,sha256=lh4epnkK9_8issxpxyLpnJ3uNwMNCsr9wBdaX8NfMlQ,3774
123
125
  databricks/bundles/pipelines/_models/azure_availability.py,sha256=KY8dt1vWA8RHBY3nq5CeqPjqhLyn5RvRfoHQ-Y2vlbk,540
124
126
  databricks/bundles/pipelines/_models/cluster_log_conf.py,sha256=EvCwe_ohAdv-mC5Veiqp28VFUaU9S5BeQA3KLLfEQGg,2697
125
127
  databricks/bundles/pipelines/_models/connection_parameters.py,sha256=235CXyN25AVcBxzMEWtxNeIb1voVqqhASVNlRv9MVSw,1561
126
- databricks/bundles/pipelines/_models/day_of_week.py,sha256=dtri8UkWRZeuZm7o82pxFlJMPMcyYj-BhKzxWhYEJ60,554
128
+ databricks/bundles/pipelines/_models/day_of_week.py,sha256=oDPKjuIkOzRxGJ3Wjddv1C-1mB-vRiy-00cNpjfNQ_o,519
127
129
  databricks/bundles/pipelines/_models/dbfs_storage_info.py,sha256=JLc9yrL3YOOVlYbQdA07tqF2wxkJdDvAKhQmAKURxaQ,972
128
130
  databricks/bundles/pipelines/_models/ebs_volume_type.py,sha256=-93BcybklhLyQEZvF8C1BhnHMeCjfDwI2qwimZ6X5eU,415
129
131
  databricks/bundles/pipelines/_models/event_log_spec.py,sha256=diWtjzD4xBbfgeEuVl_XYaOXEyTvSK40VLRue_1Ads0,1428
@@ -134,7 +136,7 @@ databricks/bundles/pipelines/_models/gcp_availability.py,sha256=a2ayWsyEQDpIDx-m
134
136
  databricks/bundles/pipelines/_models/gcs_storage_info.py,sha256=hwOowyNKCBhzsUiCQSrtmQPxrMINEq5jg2EefkrE2fQ,1020
135
137
  databricks/bundles/pipelines/_models/ingestion_config.py,sha256=78ZNABAt7du-ri05y4fumS1HrwsThggZMkYrTga0cuQ,1666
136
138
  databricks/bundles/pipelines/_models/ingestion_gateway_pipeline_definition.py,sha256=6_rqYTz-gn9JUdRHbnfFbbLrr37RrPxS_9Y2pYCVmUQ,3634
137
- databricks/bundles/pipelines/_models/ingestion_pipeline_definition.py,sha256=536lJt42KWdsnObEsc8vm1wGYqTpljixIOR_QtV8mAw,4318
139
+ databricks/bundles/pipelines/_models/ingestion_pipeline_definition.py,sha256=9sMptxvErUrSyJjWInoIIHf9rj287zSe9-FuPqeQ7uk,4802
138
140
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_table_specific_config_query_based_connector_config.py,sha256=IXJqQjQ2aORHJZy1PZ3ST4hx4HVh7O4b-JoKoxYR9QE,5181
139
141
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_workday_report_parameters.py,sha256=3foG3H-6eOTpQbH0JqK3EY27isg6ZCm8w88cufh6HRU,3434
140
142
  databricks/bundles/pipelines/_models/ingestion_pipeline_definition_workday_report_parameters_query_key_value.py,sha256=mC5r259VRZKnd0WKi1Toeb1qG2VZglumzVos4Wv8BoQ,2261
@@ -145,6 +147,7 @@ databricks/bundles/pipelines/_models/log_analytics_info.py,sha256=JeEeyNcldckin9
145
147
  databricks/bundles/pipelines/_models/maven_library.py,sha256=xz7BIo3XZ4xfp9S3sovADLxDY_DVcZbqduS8VTo1La4,2002
146
148
  databricks/bundles/pipelines/_models/notebook_library.py,sha256=YFEBdlvoNfB3oLPz-w-n_HBQrRVzFD9pbu-BPza88Rk,951
147
149
  databricks/bundles/pipelines/_models/notifications.py,sha256=Q7xHA5Bii9Zhgr2TISYF9mWKqu-6RzGO76gLexLMM3c,1987
150
+ databricks/bundles/pipelines/_models/operation_time_window.py,sha256=Ee3olezKAUFN8wg1ru0MjzOjfolsNa4MGn_NckYohKU,2112
148
151
  databricks/bundles/pipelines/_models/path_pattern.py,sha256=X3DRx7GiZzaUFC_lHRcZFGdbmUB3YyZfrgu9TWZH3LM,935
149
152
  databricks/bundles/pipelines/_models/pipeline.py,sha256=G0jJKu6a1jHWZc4QS7PZYnetps6kwii-nUv4S5NNJL0,11572
150
153
  databricks/bundles/pipelines/_models/pipeline_cluster.py,sha256=O7orq6-TZex8UpHrQLozM8Eq55bGlcDw2byB7wV4p_k,13306
@@ -164,7 +167,7 @@ databricks/bundles/pipelines/_models/schema_spec.py,sha256=nNXx-JK2jTPDWJ490yy8D
164
167
  databricks/bundles/pipelines/_models/source_catalog_config.py,sha256=4WfC1DfLR8biXXRHSG-1Wbhvzi3TtcHt8p8ugZT-hzM,1479
165
168
  databricks/bundles/pipelines/_models/source_config.py,sha256=Z29mIf0XL7TrreIEgzmc5EQBbmgX98saXK0lJo8GgoI,1119
166
169
  databricks/bundles/pipelines/_models/table_spec.py,sha256=3w9nTGzOKDhUgEtfx04i6tN3c4UDCsSaXW-zlwXgqGQ,3033
167
- databricks/bundles/pipelines/_models/table_specific_config.py,sha256=9GBOy6-_OIQZr0OLYlBHsMwIkL2rXl7FYK-gLvffFXs,6189
170
+ databricks/bundles/pipelines/_models/table_specific_config.py,sha256=agnEZufrKrwdP85GQIklJ-CJhRTxrXTEP-HE5AZMBVQ,7414
168
171
  databricks/bundles/pipelines/_models/table_specific_config_scd_type.py,sha256=_RO5oXr_b4ibygpeWXmkil24TnRQZKxbpjTx-g5qc2Q,404
169
172
  databricks/bundles/pipelines/_models/volumes_storage_info.py,sha256=31pQ9fnqQGhT2mD_ScjEhy-dm0307ne7iP_gxfcJXDY,1253
170
173
  databricks/bundles/pipelines/_models/workspace_storage_info.py,sha256=Qnm6lsw9rwXB7Te_Um0c7TvIH4Vv7ndKKYYV0pxJ6q8,1100
@@ -179,7 +182,7 @@ databricks/bundles/volumes/_models/volume.py,sha256=ALGmeXW3rGH424pp6SaXPT1I87XX
179
182
  databricks/bundles/volumes/_models/volume_grant.py,sha256=U_-4-KL8LM3n5xJBLHj_wjPsqiVjCDRj8ttiUYqFRmI,1083
180
183
  databricks/bundles/volumes/_models/volume_grant_privilege.py,sha256=fCA0LVE9Q3sbHvTAj7e62E9ASq9jH5oK1iREQdp1TxQ,384
181
184
  databricks/bundles/volumes/_models/volume_type.py,sha256=kdczwT3EJ0n5ZBV75SgX-6766igJBLTh6ywaaxIbLDk,189
182
- databricks_bundles-0.284.0.dist-info/licenses/LICENSE,sha256=QKOZO8KtzbS_Qt3Tbl0dfGnidaeilKe0UiIjnEq1tjc,3790
183
- databricks_bundles-0.284.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
184
- databricks_bundles-0.284.0.dist-info/METADATA,sha256=ly8SWyt_JvzL9lH-Zs8y7Mc5w4jPLFlZcZFAjF8BJW4,1528
185
- databricks_bundles-0.284.0.dist-info/RECORD,,
185
+ databricks_bundles-0.285.0.dist-info/licenses/LICENSE,sha256=QKOZO8KtzbS_Qt3Tbl0dfGnidaeilKe0UiIjnEq1tjc,3790
186
+ databricks_bundles-0.285.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
187
+ databricks_bundles-0.285.0.dist-info/METADATA,sha256=2chVb_Ak323s6uq4bALScDPr0KwANkP47NJczG3Z_RM,1528
188
+ databricks_bundles-0.285.0.dist-info/RECORD,,