apache-airflow-providers-apache-hive 9.1.2__py3-none-any.whl → 9.1.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- airflow/providers/apache/hive/__init__.py +1 -1
- airflow/providers/apache/hive/hooks/hive.py +17 -20
- airflow/providers/apache/hive/operators/hive.py +10 -4
- airflow/providers/apache/hive/operators/hive_stats.py +2 -2
- airflow/providers/apache/hive/sensors/hive_partition.py +2 -2
- airflow/providers/apache/hive/sensors/metastore_partition.py +1 -1
- airflow/providers/apache/hive/sensors/named_hive_partition.py +2 -2
- airflow/providers/apache/hive/transfers/hive_to_mysql.py +2 -2
- airflow/providers/apache/hive/transfers/hive_to_samba.py +2 -2
- airflow/providers/apache/hive/transfers/mssql_to_hive.py +2 -2
- airflow/providers/apache/hive/transfers/mysql_to_hive.py +2 -2
- airflow/providers/apache/hive/transfers/s3_to_hive.py +2 -2
- airflow/providers/apache/hive/transfers/vertica_to_hive.py +2 -2
- airflow/providers/apache/hive/version_compat.py +8 -18
- {apache_airflow_providers_apache_hive-9.1.2.dist-info → apache_airflow_providers_apache_hive-9.1.3.dist-info}/METADATA +36 -20
- {apache_airflow_providers_apache_hive-9.1.2.dist-info → apache_airflow_providers_apache_hive-9.1.3.dist-info}/RECORD +18 -18
- {apache_airflow_providers_apache_hive-9.1.2.dist-info → apache_airflow_providers_apache_hive-9.1.3.dist-info}/WHEEL +0 -0
- {apache_airflow_providers_apache_hive-9.1.2.dist-info → apache_airflow_providers_apache_hive-9.1.3.dist-info}/entry_points.txt +0 -0
|
@@ -29,7 +29,7 @@ from airflow import __version__ as airflow_version
|
|
|
29
29
|
|
|
30
30
|
__all__ = ["__version__"]
|
|
31
31
|
|
|
32
|
-
__version__ = "9.1.
|
|
32
|
+
__version__ = "9.1.3"
|
|
33
33
|
|
|
34
34
|
if packaging.version.parse(packaging.version.parse(airflow_version).base_version) < packaging.version.parse(
|
|
35
35
|
"2.10.0"
|
|
@@ -18,6 +18,7 @@
|
|
|
18
18
|
from __future__ import annotations
|
|
19
19
|
|
|
20
20
|
import contextlib
|
|
21
|
+
import csv
|
|
21
22
|
import os
|
|
22
23
|
import re
|
|
23
24
|
import socket
|
|
@@ -30,19 +31,21 @@ from typing import TYPE_CHECKING, Any, Literal
|
|
|
30
31
|
from deprecated import deprecated
|
|
31
32
|
from typing_extensions import overload
|
|
32
33
|
|
|
33
|
-
if TYPE_CHECKING:
|
|
34
|
-
import pandas as pd
|
|
35
|
-
import polars as pl
|
|
36
|
-
|
|
37
|
-
import csv
|
|
38
|
-
|
|
39
34
|
from airflow.configuration import conf
|
|
40
35
|
from airflow.exceptions import AirflowException, AirflowProviderDeprecationWarning
|
|
41
|
-
from airflow.providers.
|
|
36
|
+
from airflow.providers.common.compat.sdk import (
|
|
37
|
+
AIRFLOW_VAR_NAME_FORMAT_MAPPING,
|
|
38
|
+
BaseHook,
|
|
39
|
+
)
|
|
42
40
|
from airflow.providers.common.sql.hooks.sql import DbApiHook
|
|
43
41
|
from airflow.security import utils
|
|
44
42
|
from airflow.utils.helpers import as_flattened_list
|
|
45
43
|
|
|
44
|
+
if TYPE_CHECKING:
|
|
45
|
+
import pandas as pd
|
|
46
|
+
import polars as pl
|
|
47
|
+
|
|
48
|
+
|
|
46
49
|
HIVE_QUEUE_PRIORITIES = ["VERY_HIGH", "HIGH", "NORMAL", "LOW", "VERY_LOW"]
|
|
47
50
|
|
|
48
51
|
|
|
@@ -573,21 +576,15 @@ class HiveMetastoreHook(BaseHook):
|
|
|
573
576
|
conn_socket = TSocket.TSocket(host, conn.port)
|
|
574
577
|
|
|
575
578
|
if conf.get("core", "security") == "kerberos" and auth_mechanism == "GSSAPI":
|
|
576
|
-
|
|
577
|
-
import saslwrapper as sasl
|
|
578
|
-
except ImportError:
|
|
579
|
-
import sasl
|
|
580
|
-
|
|
581
|
-
def sasl_factory() -> sasl.Client:
|
|
582
|
-
sasl_client = sasl.Client()
|
|
583
|
-
sasl_client.setAttr("host", host)
|
|
584
|
-
sasl_client.setAttr("service", kerberos_service_name)
|
|
585
|
-
sasl_client.init()
|
|
586
|
-
return sasl_client
|
|
587
|
-
|
|
579
|
+
from pyhive.hive import get_installed_sasl
|
|
588
580
|
from thrift_sasl import TSaslClientTransport
|
|
589
581
|
|
|
590
|
-
|
|
582
|
+
sasl_auth = "GSSAPI"
|
|
583
|
+
transport = TSaslClientTransport(
|
|
584
|
+
lambda: get_installed_sasl(host=host, sasl_auth=sasl_auth, service=kerberos_service_name),
|
|
585
|
+
sasl_auth,
|
|
586
|
+
conn_socket,
|
|
587
|
+
)
|
|
591
588
|
else:
|
|
592
589
|
transport = TTransport.TBufferedTransport(conn_socket)
|
|
593
590
|
|
|
@@ -25,14 +25,14 @@ from typing import TYPE_CHECKING, Any
|
|
|
25
25
|
|
|
26
26
|
from airflow.configuration import conf
|
|
27
27
|
from airflow.providers.apache.hive.hooks.hive import HiveCliHook
|
|
28
|
-
from airflow.providers.
|
|
28
|
+
from airflow.providers.common.compat.sdk import (
|
|
29
29
|
AIRFLOW_VAR_NAME_FORMAT_MAPPING,
|
|
30
30
|
BaseOperator,
|
|
31
31
|
context_to_airflow_vars,
|
|
32
32
|
)
|
|
33
33
|
|
|
34
34
|
if TYPE_CHECKING:
|
|
35
|
-
from airflow.
|
|
35
|
+
from airflow.providers.common.compat.sdk import Context
|
|
36
36
|
|
|
37
37
|
|
|
38
38
|
class HiveOperator(BaseOperator):
|
|
@@ -143,9 +143,15 @@ class HiveOperator(BaseOperator):
|
|
|
143
143
|
# set the mapred_job_name if it's not set with dag, task, execution time info
|
|
144
144
|
if not self.mapred_job_name:
|
|
145
145
|
ti = context["ti"]
|
|
146
|
-
logical_date = context
|
|
146
|
+
logical_date = context.get("logical_date", None)
|
|
147
147
|
if logical_date is None:
|
|
148
|
-
raise RuntimeError(
|
|
148
|
+
raise RuntimeError(
|
|
149
|
+
"logical_date is not available. Please make sure the task is not used in an asset-triggered Dag. "
|
|
150
|
+
"HiveOperator was designed to work with timetable scheduled Dags, "
|
|
151
|
+
"and an asset-triggered Dag run does not have a logical_date. "
|
|
152
|
+
"If you need to use HiveOperator in an asset-triggered Dag,"
|
|
153
|
+
"please open an issue on the Airflow project."
|
|
154
|
+
)
|
|
149
155
|
hostname = ti.hostname or ""
|
|
150
156
|
self.hook.mapred_job_name = self.mapred_job_name_template.format(
|
|
151
157
|
dag_id=ti.dag_id,
|
|
@@ -23,12 +23,12 @@ from typing import TYPE_CHECKING, Any
|
|
|
23
23
|
|
|
24
24
|
from airflow.exceptions import AirflowException
|
|
25
25
|
from airflow.providers.apache.hive.hooks.hive import HiveMetastoreHook
|
|
26
|
-
from airflow.providers.
|
|
26
|
+
from airflow.providers.common.compat.sdk import BaseOperator
|
|
27
27
|
from airflow.providers.mysql.hooks.mysql import MySqlHook
|
|
28
28
|
from airflow.providers.presto.hooks.presto import PrestoHook
|
|
29
29
|
|
|
30
30
|
if TYPE_CHECKING:
|
|
31
|
-
from airflow.
|
|
31
|
+
from airflow.providers.common.compat.sdk import Context
|
|
32
32
|
|
|
33
33
|
|
|
34
34
|
class HiveStatsCollectionOperator(BaseOperator):
|
|
@@ -21,10 +21,10 @@ from collections.abc import Sequence
|
|
|
21
21
|
from typing import TYPE_CHECKING, Any
|
|
22
22
|
|
|
23
23
|
from airflow.providers.apache.hive.hooks.hive import HiveMetastoreHook
|
|
24
|
-
from airflow.providers.
|
|
24
|
+
from airflow.providers.common.compat.sdk import BaseSensorOperator
|
|
25
25
|
|
|
26
26
|
if TYPE_CHECKING:
|
|
27
|
-
from airflow.
|
|
27
|
+
from airflow.providers.common.compat.sdk import Context
|
|
28
28
|
|
|
29
29
|
|
|
30
30
|
class HivePartitionSensor(BaseSensorOperator):
|
|
@@ -23,7 +23,7 @@ from typing import TYPE_CHECKING, Any
|
|
|
23
23
|
from airflow.providers.common.sql.sensors.sql import SqlSensor
|
|
24
24
|
|
|
25
25
|
if TYPE_CHECKING:
|
|
26
|
-
from airflow.
|
|
26
|
+
from airflow.providers.common.compat.sdk import Context
|
|
27
27
|
|
|
28
28
|
|
|
29
29
|
class MetastorePartitionSensor(SqlSensor):
|
|
@@ -20,10 +20,10 @@ from __future__ import annotations
|
|
|
20
20
|
from collections.abc import Sequence
|
|
21
21
|
from typing import TYPE_CHECKING, Any
|
|
22
22
|
|
|
23
|
-
from airflow.providers.
|
|
23
|
+
from airflow.providers.common.compat.sdk import BaseSensorOperator
|
|
24
24
|
|
|
25
25
|
if TYPE_CHECKING:
|
|
26
|
-
from airflow.
|
|
26
|
+
from airflow.providers.common.compat.sdk import Context
|
|
27
27
|
|
|
28
28
|
|
|
29
29
|
class NamedHivePartitionSensor(BaseSensorOperator):
|
|
@@ -24,11 +24,11 @@ from tempfile import NamedTemporaryFile
|
|
|
24
24
|
from typing import TYPE_CHECKING
|
|
25
25
|
|
|
26
26
|
from airflow.providers.apache.hive.hooks.hive import HiveServer2Hook
|
|
27
|
-
from airflow.providers.
|
|
27
|
+
from airflow.providers.common.compat.sdk import BaseOperator, context_to_airflow_vars
|
|
28
28
|
from airflow.providers.mysql.hooks.mysql import MySqlHook
|
|
29
29
|
|
|
30
30
|
if TYPE_CHECKING:
|
|
31
|
-
from airflow.
|
|
31
|
+
from airflow.providers.common.compat.sdk import Context
|
|
32
32
|
|
|
33
33
|
|
|
34
34
|
class HiveToMySqlOperator(BaseOperator):
|
|
@@ -24,11 +24,11 @@ from tempfile import NamedTemporaryFile
|
|
|
24
24
|
from typing import TYPE_CHECKING
|
|
25
25
|
|
|
26
26
|
from airflow.providers.apache.hive.hooks.hive import HiveServer2Hook
|
|
27
|
-
from airflow.providers.
|
|
27
|
+
from airflow.providers.common.compat.sdk import BaseOperator, context_to_airflow_vars
|
|
28
28
|
from airflow.providers.samba.hooks.samba import SambaHook
|
|
29
29
|
|
|
30
30
|
if TYPE_CHECKING:
|
|
31
|
-
from airflow.
|
|
31
|
+
from airflow.providers.common.compat.sdk import Context
|
|
32
32
|
|
|
33
33
|
|
|
34
34
|
class HiveToSambaOperator(BaseOperator):
|
|
@@ -27,11 +27,11 @@ from typing import TYPE_CHECKING
|
|
|
27
27
|
import pymssql
|
|
28
28
|
|
|
29
29
|
from airflow.providers.apache.hive.hooks.hive import HiveCliHook
|
|
30
|
-
from airflow.providers.
|
|
30
|
+
from airflow.providers.common.compat.sdk import BaseOperator
|
|
31
31
|
from airflow.providers.microsoft.mssql.hooks.mssql import MsSqlHook
|
|
32
32
|
|
|
33
33
|
if TYPE_CHECKING:
|
|
34
|
-
from airflow.
|
|
34
|
+
from airflow.providers.common.compat.sdk import Context
|
|
35
35
|
|
|
36
36
|
|
|
37
37
|
class MsSqlToHiveOperator(BaseOperator):
|
|
@@ -37,11 +37,11 @@ except ImportError:
|
|
|
37
37
|
|
|
38
38
|
|
|
39
39
|
from airflow.providers.apache.hive.hooks.hive import HiveCliHook
|
|
40
|
-
from airflow.providers.
|
|
40
|
+
from airflow.providers.common.compat.sdk import BaseOperator
|
|
41
41
|
from airflow.providers.mysql.hooks.mysql import MySqlHook
|
|
42
42
|
|
|
43
43
|
if TYPE_CHECKING:
|
|
44
|
-
from airflow.
|
|
44
|
+
from airflow.providers.common.compat.sdk import Context
|
|
45
45
|
|
|
46
46
|
|
|
47
47
|
class MySqlToHiveOperator(BaseOperator):
|
|
@@ -31,10 +31,10 @@ from typing import TYPE_CHECKING, Any
|
|
|
31
31
|
from airflow.exceptions import AirflowException
|
|
32
32
|
from airflow.providers.amazon.aws.hooks.s3 import S3Hook
|
|
33
33
|
from airflow.providers.apache.hive.hooks.hive import HiveCliHook
|
|
34
|
-
from airflow.providers.
|
|
34
|
+
from airflow.providers.common.compat.sdk import BaseOperator
|
|
35
35
|
|
|
36
36
|
if TYPE_CHECKING:
|
|
37
|
-
from airflow.
|
|
37
|
+
from airflow.providers.common.compat.sdk import Context
|
|
38
38
|
|
|
39
39
|
|
|
40
40
|
class S3ToHiveOperator(BaseOperator):
|
|
@@ -25,11 +25,11 @@ from tempfile import NamedTemporaryFile
|
|
|
25
25
|
from typing import TYPE_CHECKING, Any
|
|
26
26
|
|
|
27
27
|
from airflow.providers.apache.hive.hooks.hive import HiveCliHook
|
|
28
|
-
from airflow.providers.
|
|
28
|
+
from airflow.providers.common.compat.sdk import BaseOperator
|
|
29
29
|
from airflow.providers.vertica.hooks.vertica import VerticaHook
|
|
30
30
|
|
|
31
31
|
if TYPE_CHECKING:
|
|
32
|
-
from airflow.
|
|
32
|
+
from airflow.providers.common.compat.sdk import Context
|
|
33
33
|
|
|
34
34
|
|
|
35
35
|
class VerticaToHiveOperator(BaseOperator):
|
|
@@ -22,6 +22,14 @@
|
|
|
22
22
|
#
|
|
23
23
|
from __future__ import annotations
|
|
24
24
|
|
|
25
|
+
# Re-export from common.compat for backward compatibility
|
|
26
|
+
from airflow.providers.common.compat.sdk import (
|
|
27
|
+
AIRFLOW_VAR_NAME_FORMAT_MAPPING,
|
|
28
|
+
BaseOperator,
|
|
29
|
+
BaseSensorOperator,
|
|
30
|
+
context_to_airflow_vars,
|
|
31
|
+
)
|
|
32
|
+
|
|
25
33
|
|
|
26
34
|
def get_base_airflow_version_tuple() -> tuple[int, int, int]:
|
|
27
35
|
from packaging.version import Version
|
|
@@ -35,27 +43,9 @@ def get_base_airflow_version_tuple() -> tuple[int, int, int]:
|
|
|
35
43
|
AIRFLOW_V_3_0_PLUS = get_base_airflow_version_tuple() >= (3, 0, 0)
|
|
36
44
|
AIRFLOW_V_3_1_PLUS: bool = get_base_airflow_version_tuple() >= (3, 1, 0)
|
|
37
45
|
|
|
38
|
-
if AIRFLOW_V_3_1_PLUS:
|
|
39
|
-
from airflow.sdk import BaseHook
|
|
40
|
-
else:
|
|
41
|
-
from airflow.hooks.base import BaseHook # type: ignore[attr-defined,no-redef]
|
|
42
|
-
|
|
43
|
-
if AIRFLOW_V_3_0_PLUS:
|
|
44
|
-
from airflow.sdk import BaseOperator, BaseSensorOperator
|
|
45
|
-
from airflow.sdk.execution_time.context import AIRFLOW_VAR_NAME_FORMAT_MAPPING, context_to_airflow_vars
|
|
46
|
-
else:
|
|
47
|
-
from airflow.models import BaseOperator
|
|
48
|
-
from airflow.sensors.base import BaseSensorOperator # type: ignore[no-redef]
|
|
49
|
-
from airflow.utils.operator_helpers import ( # type: ignore[no-redef, attr-defined]
|
|
50
|
-
AIRFLOW_VAR_NAME_FORMAT_MAPPING,
|
|
51
|
-
context_to_airflow_vars,
|
|
52
|
-
)
|
|
53
|
-
|
|
54
|
-
|
|
55
46
|
__all__ = [
|
|
56
47
|
"AIRFLOW_V_3_0_PLUS",
|
|
57
48
|
"AIRFLOW_V_3_1_PLUS",
|
|
58
|
-
"BaseHook",
|
|
59
49
|
"BaseOperator",
|
|
60
50
|
"BaseSensorOperator",
|
|
61
51
|
"AIRFLOW_VAR_NAME_FORMAT_MAPPING",
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: apache-airflow-providers-apache-hive
|
|
3
|
-
Version: 9.1.
|
|
3
|
+
Version: 9.1.3
|
|
4
4
|
Summary: Provider package apache-airflow-providers-apache-hive for Apache Airflow
|
|
5
5
|
Keywords: airflow-provider,apache.hive,airflow,integration
|
|
6
6
|
Author-email: Apache Software Foundation <dev@airflow.apache.org>
|
|
@@ -21,6 +21,7 @@ Classifier: Programming Language :: Python :: 3.12
|
|
|
21
21
|
Classifier: Programming Language :: Python :: 3.13
|
|
22
22
|
Classifier: Topic :: System :: Monitoring
|
|
23
23
|
Requires-Dist: apache-airflow>=2.10.0
|
|
24
|
+
Requires-Dist: apache-airflow-providers-common-compat>=1.8.0
|
|
24
25
|
Requires-Dist: apache-airflow-providers-common-sql>=1.26.0
|
|
25
26
|
Requires-Dist: hmsclient>=0.1.0
|
|
26
27
|
Requires-Dist: pandas>=2.1.2; python_version <"3.13"
|
|
@@ -36,8 +37,8 @@ Requires-Dist: apache-airflow-providers-presto ; extra == "presto"
|
|
|
36
37
|
Requires-Dist: apache-airflow-providers-samba ; extra == "samba"
|
|
37
38
|
Requires-Dist: apache-airflow-providers-vertica ; extra == "vertica"
|
|
38
39
|
Project-URL: Bug Tracker, https://github.com/apache/airflow/issues
|
|
39
|
-
Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.
|
|
40
|
-
Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.
|
|
40
|
+
Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.3/changelog.html
|
|
41
|
+
Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.3
|
|
41
42
|
Project-URL: Mastodon, https://fosstodon.org/@airflow
|
|
42
43
|
Project-URL: Slack Chat, https://s.apache.org/airflow-slack
|
|
43
44
|
Project-URL: Source Code, https://github.com/apache/airflow
|
|
@@ -75,9 +76,8 @@ Provides-Extra: vertica
|
|
|
75
76
|
|
|
76
77
|
Package ``apache-airflow-providers-apache-hive``
|
|
77
78
|
|
|
78
|
-
Release: ``9.1.
|
|
79
|
+
Release: ``9.1.3``
|
|
79
80
|
|
|
80
|
-
Release Date: ``|PypiReleaseDate|``
|
|
81
81
|
|
|
82
82
|
`Apache Hive <https://hive.apache.org/>`__
|
|
83
83
|
|
|
@@ -89,12 +89,12 @@ This is a provider package for ``apache.hive`` provider. All classes for this pr
|
|
|
89
89
|
are in ``airflow.providers.apache.hive`` python package.
|
|
90
90
|
|
|
91
91
|
You can find package information and changelog for the provider
|
|
92
|
-
in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.
|
|
92
|
+
in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.3/>`_.
|
|
93
93
|
|
|
94
94
|
Installation
|
|
95
95
|
------------
|
|
96
96
|
|
|
97
|
-
You can install this package on top of an existing Airflow
|
|
97
|
+
You can install this package on top of an existing Airflow installation (see ``Requirements`` below
|
|
98
98
|
for the minimum Airflow version supported) via
|
|
99
99
|
``pip install apache-airflow-providers-apache-hive``
|
|
100
100
|
|
|
@@ -103,18 +103,19 @@ The package supports the following python versions: 3.10,3.11,3.12,3.13
|
|
|
103
103
|
Requirements
|
|
104
104
|
------------
|
|
105
105
|
|
|
106
|
-
|
|
107
|
-
PIP package
|
|
108
|
-
|
|
109
|
-
``apache-airflow``
|
|
110
|
-
``apache-airflow-providers-common-
|
|
111
|
-
``
|
|
112
|
-
``
|
|
113
|
-
``pandas``
|
|
114
|
-
``
|
|
115
|
-
``
|
|
116
|
-
``
|
|
117
|
-
|
|
106
|
+
========================================== =====================================
|
|
107
|
+
PIP package Version required
|
|
108
|
+
========================================== =====================================
|
|
109
|
+
``apache-airflow`` ``>=2.10.0``
|
|
110
|
+
``apache-airflow-providers-common-compat`` ``>=1.7.4``
|
|
111
|
+
``apache-airflow-providers-common-sql`` ``>=1.26.0``
|
|
112
|
+
``hmsclient`` ``>=0.1.0``
|
|
113
|
+
``pandas`` ``>=2.1.2; python_version < "3.13"``
|
|
114
|
+
``pandas`` ``>=2.2.3; python_version >= "3.13"``
|
|
115
|
+
``pyhive[hive_pure_sasl]`` ``>=0.7.0``
|
|
116
|
+
``thrift`` ``>=0.11.0``
|
|
117
|
+
``jmespath`` ``>=0.7.0``
|
|
118
|
+
========================================== =====================================
|
|
118
119
|
|
|
119
120
|
Cross provider package dependencies
|
|
120
121
|
-----------------------------------
|
|
@@ -141,6 +142,21 @@ Dependent package
|
|
|
141
142
|
`apache-airflow-providers-vertica <https://airflow.apache.org/docs/apache-airflow-providers-vertica>`_ ``vertica``
|
|
142
143
|
====================================================================================================================== ===================
|
|
143
144
|
|
|
145
|
+
Optional dependencies
|
|
146
|
+
----------------------
|
|
147
|
+
|
|
148
|
+
=================== ============================================
|
|
149
|
+
Extra Dependencies
|
|
150
|
+
=================== ============================================
|
|
151
|
+
``amazon`` ``apache-airflow-providers-amazon``
|
|
152
|
+
``microsoft.mssql`` ``apache-airflow-providers-microsoft-mssql``
|
|
153
|
+
``mysql`` ``apache-airflow-providers-mysql``
|
|
154
|
+
``presto`` ``apache-airflow-providers-presto``
|
|
155
|
+
``samba`` ``apache-airflow-providers-samba``
|
|
156
|
+
``vertica`` ``apache-airflow-providers-vertica``
|
|
157
|
+
``common.compat`` ``apache-airflow-providers-common-compat``
|
|
158
|
+
=================== ============================================
|
|
159
|
+
|
|
144
160
|
The changelog for the provider package can be found in the
|
|
145
|
-
`changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.
|
|
161
|
+
`changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.3/changelog.html>`_.
|
|
146
162
|
|
|
@@ -1,28 +1,28 @@
|
|
|
1
1
|
airflow/providers/apache/hive/LICENSE,sha256=gXPVwptPlW1TJ4HSuG5OMPg-a3h43OGMkZRR1rpwfJA,10850
|
|
2
|
-
airflow/providers/apache/hive/__init__.py,sha256=
|
|
2
|
+
airflow/providers/apache/hive/__init__.py,sha256=agFgo_XRmaHGNjcsHBhmlUYJo8UggvlWpYH3qF6Klbo,1500
|
|
3
3
|
airflow/providers/apache/hive/get_provider_info.py,sha256=PdxdiZ7fdK1D-BTyZqj1oMS_1GS8_kbz2hZVgWMTkFI,5569
|
|
4
|
-
airflow/providers/apache/hive/version_compat.py,sha256=
|
|
4
|
+
airflow/providers/apache/hive/version_compat.py,sha256=_GXtl-AYDBq0f1jpNovs_Vw1QwHZsXbHuWE8FVAOrT4,2011
|
|
5
5
|
airflow/providers/apache/hive/hooks/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
|
|
6
|
-
airflow/providers/apache/hive/hooks/hive.py,sha256=
|
|
6
|
+
airflow/providers/apache/hive/hooks/hive.py,sha256=cE94TbK7ckbW9M4Q1IJcFZEYHqlhA0EviveR7pbHz_Y,44796
|
|
7
7
|
airflow/providers/apache/hive/macros/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
|
|
8
8
|
airflow/providers/apache/hive/macros/hive.py,sha256=FEMCR9rEWKFOKtKAvhVWhD5jRJnSYeHepVGgJLzUa2k,4548
|
|
9
9
|
airflow/providers/apache/hive/operators/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
|
|
10
|
-
airflow/providers/apache/hive/operators/hive.py,sha256=
|
|
11
|
-
airflow/providers/apache/hive/operators/hive_stats.py,sha256=
|
|
10
|
+
airflow/providers/apache/hive/operators/hive.py,sha256=YO_Q3l-SskEqAUhPdmpimZsuNbSXF1l3bwb42MC5oRU,7517
|
|
11
|
+
airflow/providers/apache/hive/operators/hive_stats.py,sha256=QYkdUPu_FKh2n791yW278hOIgTHRGAKEs19t0ZaWiRw,7109
|
|
12
12
|
airflow/providers/apache/hive/plugins/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
|
|
13
13
|
airflow/providers/apache/hive/plugins/hive.py,sha256=rrGccro6DEdnHSmII3goBYD9te-XdhPaPMjWpF-36HU,1146
|
|
14
14
|
airflow/providers/apache/hive/sensors/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
|
|
15
|
-
airflow/providers/apache/hive/sensors/hive_partition.py,sha256=
|
|
16
|
-
airflow/providers/apache/hive/sensors/metastore_partition.py,sha256=
|
|
17
|
-
airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=
|
|
15
|
+
airflow/providers/apache/hive/sensors/hive_partition.py,sha256=j7TVA-ZAGDVpi6HpuaAwSr_PBMz-v3Qf7SzNeCwwI4Q,3095
|
|
16
|
+
airflow/providers/apache/hive/sensors/metastore_partition.py,sha256=x0X9xt50cIQucfi2-4p30C17CUQgEQqyWI7npW50YJQ,3379
|
|
17
|
+
airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=TChDCPRzdiSE9ymrkKOJdqyeNI92Ji0zAUUfKSSfOmU,4196
|
|
18
18
|
airflow/providers/apache/hive/transfers/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
|
|
19
|
-
airflow/providers/apache/hive/transfers/hive_to_mysql.py,sha256=
|
|
20
|
-
airflow/providers/apache/hive/transfers/hive_to_samba.py,sha256=
|
|
21
|
-
airflow/providers/apache/hive/transfers/mssql_to_hive.py,sha256=
|
|
22
|
-
airflow/providers/apache/hive/transfers/mysql_to_hive.py,sha256=
|
|
23
|
-
airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=
|
|
24
|
-
airflow/providers/apache/hive/transfers/vertica_to_hive.py,sha256=
|
|
25
|
-
apache_airflow_providers_apache_hive-9.1.
|
|
26
|
-
apache_airflow_providers_apache_hive-9.1.
|
|
27
|
-
apache_airflow_providers_apache_hive-9.1.
|
|
28
|
-
apache_airflow_providers_apache_hive-9.1.
|
|
19
|
+
airflow/providers/apache/hive/transfers/hive_to_mysql.py,sha256=SOV3soZX-KmtEjWULJBo7CYwqkwDwdWxY0tWmMS1PVU,5301
|
|
20
|
+
airflow/providers/apache/hive/transfers/hive_to_samba.py,sha256=i9d9BV4bfUr5DabSB5UpOJ5S60Gz17smJnK96Vkhqlc,3003
|
|
21
|
+
airflow/providers/apache/hive/transfers/mssql_to_hive.py,sha256=w-B1hqVzCGPex6-KlaPyEK4bePYFpZkVrl3z1r4dQWI,5791
|
|
22
|
+
airflow/providers/apache/hive/transfers/mysql_to_hive.py,sha256=Q-qJ3sX75RGzYpbU_RXVo6HBKkFeig0-agbr6uG3HsA,7067
|
|
23
|
+
airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=lZ2TcAJJkNpP84nxrTZJ1mqNeCS_7PRC0CPNAVtEQj8,12452
|
|
24
|
+
airflow/providers/apache/hive/transfers/vertica_to_hive.py,sha256=W2sDdp9nfYjr-FYK-v3t33BM0yD_GRqgp8JXuCaXqrY,5590
|
|
25
|
+
apache_airflow_providers_apache_hive-9.1.3.dist-info/entry_points.txt,sha256=Hzixt33mYYldwmwswarArUB7ZU0xbmUtd3tFViZ414s,185
|
|
26
|
+
apache_airflow_providers_apache_hive-9.1.3.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
|
|
27
|
+
apache_airflow_providers_apache_hive-9.1.3.dist-info/METADATA,sha256=Bwj8nM9mCZedmMXxqRYZctCVQGNqT0krBgL249m98qs,8134
|
|
28
|
+
apache_airflow_providers_apache_hive-9.1.3.dist-info/RECORD,,
|
|
File without changes
|