apache-airflow-providers-databricks 7.2.1__py3-none-any.whl → 7.2.2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of apache-airflow-providers-databricks might be problematic. Click here for more details.

@@ -29,7 +29,7 @@ from airflow import __version__ as airflow_version
29
29
 
30
30
  __all__ = ["__version__"]
31
31
 
32
- __version__ = "7.2.1"
32
+ __version__ = "7.2.2"
33
33
 
34
34
  if packaging.version.parse(packaging.version.parse(airflow_version).base_version) < packaging.version.parse(
35
35
  "2.9.0"
@@ -27,8 +27,9 @@ def get_provider_info():
27
27
  "name": "Databricks",
28
28
  "description": "`Databricks <https://databricks.com/>`__\n",
29
29
  "state": "ready",
30
- "source-date-epoch": 1741508706,
30
+ "source-date-epoch": 1742979996,
31
31
  "versions": [
32
+ "7.2.2",
32
33
  "7.2.1",
33
34
  "7.2.0",
34
35
  "7.0.0",
@@ -29,7 +29,7 @@ from typing import TYPE_CHECKING, Any
29
29
 
30
30
  from airflow.configuration import conf
31
31
  from airflow.exceptions import AirflowException
32
- from airflow.models import BaseOperator, XCom
32
+ from airflow.models import BaseOperator
33
33
  from airflow.providers.databricks.hooks.databricks import DatabricksHook, RunLifeCycleState, RunState
34
34
  from airflow.providers.databricks.operators.databricks_workflow import (
35
35
  DatabricksWorkflowTaskGroup,
@@ -50,7 +50,9 @@ if TYPE_CHECKING:
50
50
 
51
51
  if AIRFLOW_V_3_0_PLUS:
52
52
  from airflow.sdk import BaseOperatorLink
53
+ from airflow.sdk.execution_time.xcom import XCom
53
54
  else:
55
+ from airflow.models import XCom # type: ignore[no-redef]
54
56
  from airflow.models.baseoperatorlink import BaseOperatorLink # type: ignore[no-redef]
55
57
 
56
58
  DEFER_METHOD_NAME = "execute_complete"
@@ -31,7 +31,6 @@ from airflow.models import DagBag
31
31
  from airflow.models.dag import DAG, clear_task_instances
32
32
  from airflow.models.dagrun import DagRun
33
33
  from airflow.models.taskinstance import TaskInstance, TaskInstanceKey
34
- from airflow.models.xcom import XCom
35
34
  from airflow.plugins_manager import AirflowPlugin
36
35
  from airflow.providers.databricks.hooks.databricks import DatabricksHook
37
36
  from airflow.providers.databricks.version_compat import AIRFLOW_V_3_0_PLUS
@@ -53,7 +52,9 @@ if TYPE_CHECKING:
53
52
 
54
53
  if AIRFLOW_V_3_0_PLUS:
55
54
  from airflow.sdk import BaseOperatorLink
55
+ from airflow.sdk.execution_time.xcom import XCom
56
56
  else:
57
+ from airflow.models import XCom # type: ignore[no-redef]
57
58
  from airflow.models.baseoperatorlink import BaseOperatorLink # type: ignore[no-redef]
58
59
 
59
60
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: apache-airflow-providers-databricks
3
- Version: 7.2.1
3
+ Version: 7.2.2
4
4
  Summary: Provider package apache-airflow-providers-databricks for Apache Airflow
5
5
  Keywords: airflow-provider,databricks,airflow,integration
6
6
  Author-email: Apache Software Foundation <dev@airflow.apache.org>
@@ -33,11 +33,11 @@ Requires-Dist: apache-airflow-providers-fab ; extra == "fab"
33
33
  Requires-Dist: databricks-sdk==0.10.0 ; extra == "sdk"
34
34
  Requires-Dist: apache-airflow-providers-standard ; extra == "standard"
35
35
  Project-URL: Bug Tracker, https://github.com/apache/airflow/issues
36
- Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.1/changelog.html
37
- Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.1
36
+ Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.2/changelog.html
37
+ Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.2
38
+ Project-URL: Mastodon, https://fosstodon.org/@airflow
38
39
  Project-URL: Slack Chat, https://s.apache.org/airflow-slack
39
40
  Project-URL: Source Code, https://github.com/apache/airflow
40
- Project-URL: Twitter, https://x.com/ApacheAirflow
41
41
  Project-URL: YouTube, https://www.youtube.com/channel/UCSXwxpWZQ7XZ1WL3wqevChA/
42
42
  Provides-Extra: azure-identity
43
43
  Provides-Extra: fab
@@ -69,7 +69,7 @@ Provides-Extra: standard
69
69
 
70
70
  Package ``apache-airflow-providers-databricks``
71
71
 
72
- Release: ``7.2.1``
72
+ Release: ``7.2.2``
73
73
 
74
74
 
75
75
  `Databricks <https://databricks.com/>`__
@@ -82,7 +82,7 @@ This is a provider package for ``databricks`` provider. All classes for this pro
82
82
  are in ``airflow.providers.databricks`` python package.
83
83
 
84
84
  You can find package information and changelog for the provider
85
- in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.1/>`_.
85
+ in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.2/>`_.
86
86
 
87
87
  Installation
88
88
  ------------
@@ -113,7 +113,7 @@ Cross provider package dependencies
113
113
  -----------------------------------
114
114
 
115
115
  Those are dependencies that might be needed in order to use all the features of the package.
116
- You need to install the specified provider packages in order to use them.
116
+ You need to install the specified providers in order to use them.
117
117
 
118
118
  You can install such cross-provider dependencies when installing from PyPI. For example:
119
119
 
@@ -130,5 +130,5 @@ Dependent package
130
130
  ============================================================================================================ ==============
131
131
 
132
132
  The changelog for the provider package can be found in the
133
- `changelog <https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.1/changelog.html>`_.
133
+ `changelog <https://airflow.apache.org/docs/apache-airflow-providers-databricks/7.2.2/changelog.html>`_.
134
134
 
@@ -1,19 +1,19 @@
1
1
  airflow/providers/databricks/LICENSE,sha256=gXPVwptPlW1TJ4HSuG5OMPg-a3h43OGMkZRR1rpwfJA,10850
2
- airflow/providers/databricks/__init__.py,sha256=vu5UIOrg11AdDubzxLAhf2giA_m3-vhs8qGQxa6v1nk,1497
2
+ airflow/providers/databricks/__init__.py,sha256=-094H4znLyAeiEWK6WoEtwB7J3mRF_mh1HRqSC6_X2E,1497
3
3
  airflow/providers/databricks/exceptions.py,sha256=85RklmLOI_PnTzfXNIUd5fAu2aMMUhelwumQAX0wANE,1261
4
- airflow/providers/databricks/get_provider_info.py,sha256=hGBW2LAM9o9NPZI_RbAUkHA_LCi1CAzRyQ2Ch0QafnU,7350
4
+ airflow/providers/databricks/get_provider_info.py,sha256=zfvrXwvvfkp7sneZrK7Mhslxv9oZRG7efWLUnndGLNE,7371
5
5
  airflow/providers/databricks/version_compat.py,sha256=aHg90_DtgoSnQvILFICexMyNlHlALBdaeWqkX3dFDug,1605
6
6
  airflow/providers/databricks/hooks/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
7
7
  airflow/providers/databricks/hooks/databricks.py,sha256=-rgK_sMc2_BjTvSvMh1Md3XanVayOmcxijQfs1vRCPw,24774
8
8
  airflow/providers/databricks/hooks/databricks_base.py,sha256=8KVRF-ty20UQpJP3kgE6RDLAYqXk7ZjI07ZpwFIcGB8,34917
9
9
  airflow/providers/databricks/hooks/databricks_sql.py,sha256=Giy0XBz0t6PHepREIXzxlFZhaNYgnS4bzBT_I04v5rM,13119
10
10
  airflow/providers/databricks/operators/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
11
- airflow/providers/databricks/operators/databricks.py,sha256=AsASQe4kcWJ1Ks4dptHCMio0mm0PYgwZ9aC6E1CKe5c,69926
11
+ airflow/providers/databricks/operators/databricks.py,sha256=YCZKJE4-r1Q18tHYhT7Fp6Nv18l_KyqyYte8ev0B9P4,70035
12
12
  airflow/providers/databricks/operators/databricks_repos.py,sha256=m_72OnnU9df7UB-8SK2Tp5VjfNyjYeAnil3dCKs9SbA,13282
13
13
  airflow/providers/databricks/operators/databricks_sql.py,sha256=thBHpt9_LMLJZ0PN-eLCI3AaT8IFq3NAHLDWDFP-Jiw,17031
14
14
  airflow/providers/databricks/operators/databricks_workflow.py,sha256=0vFu4w6O4tlStZ_Jhk1iswKFcTk-g_dthGFeDpXGZlw,14146
15
15
  airflow/providers/databricks/plugins/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
16
- airflow/providers/databricks/plugins/databricks_workflow.py,sha256=BPB-Cz50s_Zv4Paa7p01xpES3nErQ32lFHzDfeIYhyM,17209
16
+ airflow/providers/databricks/plugins/databricks_workflow.py,sha256=OVrW3DcqjClBt07RoI0nWbLKK5LfTaVJwjd9na44ugk,17287
17
17
  airflow/providers/databricks/sensors/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
18
18
  airflow/providers/databricks/sensors/databricks_partition.py,sha256=hS6Q2fb84_vASb7Ai50-pmjVukX6G8xIwdaZVIE17oM,10045
19
19
  airflow/providers/databricks/sensors/databricks_sql.py,sha256=jIA9oGBUCAlXzyrqigxlg7JQDsBFuNIF8ZUEJM8gPxg,5766
@@ -21,7 +21,7 @@ airflow/providers/databricks/triggers/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvR
21
21
  airflow/providers/databricks/triggers/databricks.py,sha256=xk9aEfdZnG33a4WSFfg6SZF4FfROV8B4HOyBYBvZR_Q,5104
22
22
  airflow/providers/databricks/utils/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
23
23
  airflow/providers/databricks/utils/databricks.py,sha256=EICTPZTD0R0dy9UGKgv8srkrBTgzCQrcYNL9oBWuhzk,2890
24
- apache_airflow_providers_databricks-7.2.1.dist-info/entry_points.txt,sha256=hjmZm3ab2cteTR4t9eE28oKixHwNIKtLCThd6sx3XRQ,227
25
- apache_airflow_providers_databricks-7.2.1.dist-info/WHEEL,sha256=_2ozNFCLWc93bK4WKHCO-eDUENDlo-dgc9cU3qokYO4,82
26
- apache_airflow_providers_databricks-7.2.1.dist-info/METADATA,sha256=20VpPbM6oJBp_i6NJgqUqOyBXBa8Eaq-O0z5jDmvPYE,6083
27
- apache_airflow_providers_databricks-7.2.1.dist-info/RECORD,,
24
+ apache_airflow_providers_databricks-7.2.2.dist-info/entry_points.txt,sha256=hjmZm3ab2cteTR4t9eE28oKixHwNIKtLCThd6sx3XRQ,227
25
+ apache_airflow_providers_databricks-7.2.2.dist-info/WHEEL,sha256=_2ozNFCLWc93bK4WKHCO-eDUENDlo-dgc9cU3qokYO4,82
26
+ apache_airflow_providers_databricks-7.2.2.dist-info/METADATA,sha256=wwDGW5Upxlsc9mAOQABc5QfMhd6Iptuc7cv2vDoMZP4,6079
27
+ apache_airflow_providers_databricks-7.2.2.dist-info/RECORD,,