apache-airflow-providers-apache-hive 9.1.0__py3-none-any.whl → 9.1.1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of apache-airflow-providers-apache-hive might be problematic. Click here for more details.

@@ -29,7 +29,7 @@ from airflow import __version__ as airflow_version
29
29
 
30
30
  __all__ = ["__version__"]
31
31
 
32
- __version__ = "9.1.0"
32
+ __version__ = "9.1.1"
33
33
 
34
34
  if packaging.version.parse(packaging.version.parse(airflow_version).base_version) < packaging.version.parse(
35
35
  "2.10.0"
@@ -25,10 +25,10 @@ import subprocess
25
25
  import time
26
26
  from collections.abc import Iterable, Mapping
27
27
  from tempfile import NamedTemporaryFile, TemporaryDirectory
28
- from typing import TYPE_CHECKING, Any
28
+ from typing import TYPE_CHECKING, Any, Literal
29
29
 
30
30
  from deprecated import deprecated
31
- from typing_extensions import Literal, overload
31
+ from typing_extensions import overload
32
32
 
33
33
  if TYPE_CHECKING:
34
34
  import pandas as pd
@@ -38,19 +38,16 @@ import csv
38
38
 
39
39
  from airflow.configuration import conf
40
40
  from airflow.exceptions import AirflowException, AirflowProviderDeprecationWarning
41
- from airflow.hooks.base import BaseHook
42
- from airflow.providers.common.compat.version_compat import AIRFLOW_V_3_0_PLUS
41
+ from airflow.providers.apache.hive.version_compat import AIRFLOW_VAR_NAME_FORMAT_MAPPING
43
42
  from airflow.providers.common.sql.hooks.sql import DbApiHook
43
+
44
+ try:
45
+ from airflow.sdk import BaseHook
46
+ except ImportError:
47
+ from airflow.hooks.base import BaseHook # type: ignore[attr-defined,no-redef]
44
48
  from airflow.security import utils
45
49
  from airflow.utils.helpers import as_flattened_list
46
50
 
47
- if AIRFLOW_V_3_0_PLUS:
48
- from airflow.sdk.execution_time.context import AIRFLOW_VAR_NAME_FORMAT_MAPPING
49
- else:
50
- from airflow.utils.operator_helpers import ( # type: ignore[no-redef, attr-defined]
51
- AIRFLOW_VAR_NAME_FORMAT_MAPPING,
52
- )
53
-
54
51
  HIVE_QUEUE_PRIORITIES = ["VERY_HIGH", "HIGH", "NORMAL", "LOW", "VERY_LOW"]
55
52
 
56
53
 
@@ -277,7 +274,7 @@ class HiveCliHook(BaseHook):
277
274
  True
278
275
  """
279
276
  conn = self.conn
280
- schema = schema or conn.schema
277
+ schema = schema or conn.schema or ""
281
278
 
282
279
  invalid_chars_list = re.findall(r"[^a-z0-9_]", schema)
283
280
  if invalid_chars_list:
@@ -605,7 +602,9 @@ class HiveMetastoreHook(BaseHook):
605
602
 
606
603
  def _find_valid_host(self) -> Any:
607
604
  conn = self.conn
608
- hosts = conn.host.split(",")
605
+ hosts = []
606
+ if conn.host:
607
+ hosts = conn.host.split(",")
609
608
  for host in hosts:
610
609
  host_socket = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
611
610
  self.log.info("Trying to connect to %s:%s", host, conn.port)
@@ -24,17 +24,12 @@ from functools import cached_property
24
24
  from typing import TYPE_CHECKING, Any
25
25
 
26
26
  from airflow.configuration import conf
27
- from airflow.models import BaseOperator
28
27
  from airflow.providers.apache.hive.hooks.hive import HiveCliHook
29
- from airflow.providers.common.compat.version_compat import AIRFLOW_V_3_0_PLUS
30
-
31
- if AIRFLOW_V_3_0_PLUS:
32
- from airflow.sdk.execution_time.context import AIRFLOW_VAR_NAME_FORMAT_MAPPING, context_to_airflow_vars
33
- else:
34
- from airflow.utils.operator_helpers import ( # type: ignore[no-redef, attr-defined]
35
- AIRFLOW_VAR_NAME_FORMAT_MAPPING,
36
- context_to_airflow_vars,
37
- )
28
+ from airflow.providers.apache.hive.version_compat import (
29
+ AIRFLOW_VAR_NAME_FORMAT_MAPPING,
30
+ BaseOperator,
31
+ context_to_airflow_vars,
32
+ )
38
33
 
39
34
  if TYPE_CHECKING:
40
35
  from airflow.utils.context import Context
@@ -18,12 +18,12 @@
18
18
  from __future__ import annotations
19
19
 
20
20
  import json
21
- from collections.abc import Sequence
22
- from typing import TYPE_CHECKING, Any, Callable
21
+ from collections.abc import Callable, Sequence
22
+ from typing import TYPE_CHECKING, Any
23
23
 
24
24
  from airflow.exceptions import AirflowException
25
- from airflow.models import BaseOperator
26
25
  from airflow.providers.apache.hive.hooks.hive import HiveMetastoreHook
26
+ from airflow.providers.apache.hive.version_compat import BaseOperator
27
27
  from airflow.providers.mysql.hooks.mysql import MySqlHook
28
28
  from airflow.providers.presto.hooks.presto import PrestoHook
29
29
 
@@ -21,7 +21,7 @@ from collections.abc import Sequence
21
21
  from typing import TYPE_CHECKING, Any
22
22
 
23
23
  from airflow.providers.apache.hive.hooks.hive import HiveMetastoreHook
24
- from airflow.sensors.base import BaseSensorOperator
24
+ from airflow.providers.apache.hive.version_compat import BaseSensorOperator
25
25
 
26
26
  if TYPE_CHECKING:
27
27
  from airflow.utils.context import Context
@@ -20,7 +20,7 @@ from __future__ import annotations
20
20
  from collections.abc import Sequence
21
21
  from typing import TYPE_CHECKING, Any
22
22
 
23
- from airflow.sensors.base import BaseSensorOperator
23
+ from airflow.providers.apache.hive.version_compat import BaseSensorOperator
24
24
 
25
25
  if TYPE_CHECKING:
26
26
  from airflow.utils.context import Context
@@ -23,17 +23,10 @@ from collections.abc import Sequence
23
23
  from tempfile import NamedTemporaryFile
24
24
  from typing import TYPE_CHECKING
25
25
 
26
- from airflow.models import BaseOperator
27
26
  from airflow.providers.apache.hive.hooks.hive import HiveServer2Hook
28
- from airflow.providers.common.compat.version_compat import AIRFLOW_V_3_0_PLUS
27
+ from airflow.providers.apache.hive.version_compat import BaseOperator, context_to_airflow_vars
29
28
  from airflow.providers.mysql.hooks.mysql import MySqlHook
30
29
 
31
- if AIRFLOW_V_3_0_PLUS:
32
- from airflow.sdk.execution_time.context import context_to_airflow_vars
33
- else:
34
- from airflow.utils.operator_helpers import context_to_airflow_vars # type: ignore[no-redef, attr-defined]
35
-
36
-
37
30
  if TYPE_CHECKING:
38
31
  from airflow.utils.context import Context
39
32
 
@@ -23,16 +23,10 @@ from collections.abc import Sequence
23
23
  from tempfile import NamedTemporaryFile
24
24
  from typing import TYPE_CHECKING
25
25
 
26
- from airflow.models import BaseOperator
27
26
  from airflow.providers.apache.hive.hooks.hive import HiveServer2Hook
28
- from airflow.providers.common.compat.version_compat import AIRFLOW_V_3_0_PLUS
27
+ from airflow.providers.apache.hive.version_compat import BaseOperator, context_to_airflow_vars
29
28
  from airflow.providers.samba.hooks.samba import SambaHook
30
29
 
31
- if AIRFLOW_V_3_0_PLUS:
32
- from airflow.sdk.execution_time.context import context_to_airflow_vars
33
- else:
34
- from airflow.utils.operator_helpers import context_to_airflow_vars # type: ignore[no-redef, attr-defined]
35
-
36
30
  if TYPE_CHECKING:
37
31
  from airflow.utils.context import Context
38
32
 
@@ -26,8 +26,8 @@ from typing import TYPE_CHECKING
26
26
 
27
27
  import pymssql
28
28
 
29
- from airflow.models import BaseOperator
30
29
  from airflow.providers.apache.hive.hooks.hive import HiveCliHook
30
+ from airflow.providers.apache.hive.version_compat import BaseOperator
31
31
  from airflow.providers.microsoft.mssql.hooks.mssql import MsSqlHook
32
32
 
33
33
  if TYPE_CHECKING:
@@ -36,8 +36,8 @@ except ImportError:
36
36
  )
37
37
 
38
38
 
39
- from airflow.models import BaseOperator
40
39
  from airflow.providers.apache.hive.hooks.hive import HiveCliHook
40
+ from airflow.providers.apache.hive.version_compat import BaseOperator
41
41
  from airflow.providers.mysql.hooks.mysql import MySqlHook
42
42
 
43
43
  if TYPE_CHECKING:
@@ -29,9 +29,9 @@ from tempfile import NamedTemporaryFile, TemporaryDirectory
29
29
  from typing import TYPE_CHECKING, Any
30
30
 
31
31
  from airflow.exceptions import AirflowException
32
- from airflow.models import BaseOperator
33
32
  from airflow.providers.amazon.aws.hooks.s3 import S3Hook
34
33
  from airflow.providers.apache.hive.hooks.hive import HiveCliHook
34
+ from airflow.providers.apache.hive.version_compat import BaseOperator
35
35
 
36
36
  if TYPE_CHECKING:
37
37
  from airflow.utils.context import Context
@@ -24,8 +24,8 @@ from collections.abc import Sequence
24
24
  from tempfile import NamedTemporaryFile
25
25
  from typing import TYPE_CHECKING, Any
26
26
 
27
- from airflow.models import BaseOperator
28
27
  from airflow.providers.apache.hive.hooks.hive import HiveCliHook
28
+ from airflow.providers.apache.hive.version_compat import BaseOperator
29
29
  from airflow.providers.vertica.hooks.vertica import VerticaHook
30
30
 
31
31
  if TYPE_CHECKING:
@@ -0,0 +1,55 @@
1
+ # Licensed to the Apache Software Foundation (ASF) under one
2
+ # or more contributor license agreements. See the NOTICE file
3
+ # distributed with this work for additional information
4
+ # regarding copyright ownership. The ASF licenses this file
5
+ # to you under the Apache License, Version 2.0 (the
6
+ # "License"); you may not use this file except in compliance
7
+ # with the License. You may obtain a copy of the License at
8
+ #
9
+ # http://www.apache.org/licenses/LICENSE-2.0
10
+ #
11
+ # Unless required by applicable law or agreed to in writing,
12
+ # software distributed under the License is distributed on an
13
+ # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
14
+ # KIND, either express or implied. See the License for the
15
+ # specific language governing permissions and limitations
16
+ # under the License.
17
+ #
18
+ # NOTE! THIS FILE IS COPIED MANUALLY IN OTHER PROVIDERS DELIBERATELY TO AVOID ADDING UNNECESSARY
19
+ # DEPENDENCIES BETWEEN PROVIDERS. IF YOU WANT TO ADD CONDITIONAL CODE IN YOUR PROVIDER THAT DEPENDS
20
+ # ON AIRFLOW VERSION, PLEASE COPY THIS FILE TO THE ROOT PACKAGE OF YOUR PROVIDER AND IMPORT
21
+ # THOSE CONSTANTS FROM IT RATHER THAN IMPORTING THEM FROM ANOTHER PROVIDER OR TEST CODE
22
+ #
23
+ from __future__ import annotations
24
+
25
+
26
+ def get_base_airflow_version_tuple() -> tuple[int, int, int]:
27
+ from packaging.version import Version
28
+
29
+ from airflow import __version__
30
+
31
+ airflow_version = Version(__version__)
32
+ return airflow_version.major, airflow_version.minor, airflow_version.micro
33
+
34
+
35
+ AIRFLOW_V_3_0_PLUS = get_base_airflow_version_tuple() >= (3, 0, 0)
36
+
37
+ if AIRFLOW_V_3_0_PLUS:
38
+ from airflow.sdk import BaseOperator, BaseSensorOperator
39
+ from airflow.sdk.execution_time.context import AIRFLOW_VAR_NAME_FORMAT_MAPPING, context_to_airflow_vars
40
+ else:
41
+ from airflow.models import BaseOperator
42
+ from airflow.sensors.base import BaseSensorOperator # type: ignore[no-redef]
43
+ from airflow.utils.operator_helpers import ( # type: ignore[no-redef, attr-defined]
44
+ AIRFLOW_VAR_NAME_FORMAT_MAPPING,
45
+ context_to_airflow_vars,
46
+ )
47
+
48
+
49
+ __all__ = [
50
+ "AIRFLOW_V_3_0_PLUS",
51
+ "BaseOperator",
52
+ "BaseSensorOperator",
53
+ "AIRFLOW_VAR_NAME_FORMAT_MAPPING",
54
+ "context_to_airflow_vars",
55
+ ]
@@ -1,11 +1,11 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: apache-airflow-providers-apache-hive
3
- Version: 9.1.0
3
+ Version: 9.1.1
4
4
  Summary: Provider package apache-airflow-providers-apache-hive for Apache Airflow
5
5
  Keywords: airflow-provider,apache.hive,airflow,integration
6
6
  Author-email: Apache Software Foundation <dev@airflow.apache.org>
7
7
  Maintainer-email: Apache Software Foundation <dev@airflow.apache.org>
8
- Requires-Python: ~=3.9
8
+ Requires-Python: ~=3.10
9
9
  Description-Content-Type: text/x-rst
10
10
  Classifier: Development Status :: 5 - Production/Stable
11
11
  Classifier: Environment :: Console
@@ -15,7 +15,6 @@ Classifier: Intended Audience :: System Administrators
15
15
  Classifier: Framework :: Apache Airflow
16
16
  Classifier: Framework :: Apache Airflow :: Provider
17
17
  Classifier: License :: OSI Approved :: Apache Software License
18
- Classifier: Programming Language :: Python :: 3.9
19
18
  Classifier: Programming Language :: Python :: 3.10
20
19
  Classifier: Programming Language :: Python :: 3.11
21
20
  Classifier: Programming Language :: Python :: 3.12
@@ -34,8 +33,8 @@ Requires-Dist: apache-airflow-providers-presto ; extra == "presto"
34
33
  Requires-Dist: apache-airflow-providers-samba ; extra == "samba"
35
34
  Requires-Dist: apache-airflow-providers-vertica ; extra == "vertica"
36
35
  Project-URL: Bug Tracker, https://github.com/apache/airflow/issues
37
- Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.0/changelog.html
38
- Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.0
36
+ Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.1/changelog.html
37
+ Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.1
39
38
  Project-URL: Mastodon, https://fosstodon.org/@airflow
40
39
  Project-URL: Slack Chat, https://s.apache.org/airflow-slack
41
40
  Project-URL: Source Code, https://github.com/apache/airflow
@@ -73,7 +72,7 @@ Provides-Extra: vertica
73
72
 
74
73
  Package ``apache-airflow-providers-apache-hive``
75
74
 
76
- Release: ``9.1.0``
75
+ Release: ``9.1.1``
77
76
 
78
77
 
79
78
  `Apache Hive <https://hive.apache.org/>`__
@@ -86,7 +85,7 @@ This is a provider package for ``apache.hive`` provider. All classes for this pr
86
85
  are in ``airflow.providers.apache.hive`` python package.
87
86
 
88
87
  You can find package information and changelog for the provider
89
- in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.0/>`_.
88
+ in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.1/>`_.
90
89
 
91
90
  Installation
92
91
  ------------
@@ -95,7 +94,7 @@ You can install this package on top of an existing Airflow 2 installation (see `
95
94
  for the minimum Airflow version supported) via
96
95
  ``pip install apache-airflow-providers-apache-hive``
97
96
 
98
- The package supports the following python versions: 3.9,3.10,3.11,3.12
97
+ The package supports the following python versions: 3.10,3.11,3.12
99
98
 
100
99
  Requirements
101
100
  ------------
@@ -138,5 +137,5 @@ Dependent package
138
137
  ====================================================================================================================== ===================
139
138
 
140
139
  The changelog for the provider package can be found in the
141
- `changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.0/changelog.html>`_.
140
+ `changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.1.1/changelog.html>`_.
142
141
 
@@ -1,27 +1,28 @@
1
1
  airflow/providers/apache/hive/LICENSE,sha256=gXPVwptPlW1TJ4HSuG5OMPg-a3h43OGMkZRR1rpwfJA,10850
2
- airflow/providers/apache/hive/__init__.py,sha256=4A8vByNcISGrsZVhWBPVTcf7lp2FsDeWev2kifiVVxM,1500
2
+ airflow/providers/apache/hive/__init__.py,sha256=8F3zH4JTiVVmllb0y0zN0f2fEIrsSq7aNYavyXYRc0E,1500
3
3
  airflow/providers/apache/hive/get_provider_info.py,sha256=PdxdiZ7fdK1D-BTyZqj1oMS_1GS8_kbz2hZVgWMTkFI,5569
4
+ airflow/providers/apache/hive/version_compat.py,sha256=vpHPRbg97lVT6nQO6dsHw_qeLPUT07UG4wAjWQ1thu8,2188
4
5
  airflow/providers/apache/hive/hooks/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
5
- airflow/providers/apache/hive/hooks/hive.py,sha256=ay_4PN2DI4NDfk4z88OgbQTeQ5_FlajBZVoYaaEvUhM,45145
6
+ airflow/providers/apache/hive/hooks/hive.py,sha256=MCxceSCyOihazaeeZLm0fmFr4tmkBY4SpigE1-GQe0c,45064
6
7
  airflow/providers/apache/hive/macros/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
7
8
  airflow/providers/apache/hive/macros/hive.py,sha256=FEMCR9rEWKFOKtKAvhVWhD5jRJnSYeHepVGgJLzUa2k,4548
8
9
  airflow/providers/apache/hive/operators/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
9
- airflow/providers/apache/hive/operators/hive.py,sha256=QU7-zSc57bk8mNpFVuAmBmIcHiyqiN-OMCzyFwceETY,7348
10
- airflow/providers/apache/hive/operators/hive_stats.py,sha256=8fNlR9o8J9DKMDBpB849apqA2iGbMovlHOWute4TFCY,7074
10
+ airflow/providers/apache/hive/operators/hive.py,sha256=_FpAflhrj_VY1_ifd2pCFFKqXKAeh0FuDdWKcbiBM6E,7068
11
+ airflow/providers/apache/hive/operators/hive_stats.py,sha256=AU4uL-nKyYAfay9-XGjhUek3V5Egm7Z6wV-FRJ0jmZY,7104
11
12
  airflow/providers/apache/hive/plugins/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
12
13
  airflow/providers/apache/hive/plugins/hive.py,sha256=rrGccro6DEdnHSmII3goBYD9te-XdhPaPMjWpF-36HU,1146
13
14
  airflow/providers/apache/hive/sensors/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
14
- airflow/providers/apache/hive/sensors/hive_partition.py,sha256=UIafLmwTUXdNfWAhLOOSZQc6v6So7lz_0bYA2Wyunw4,3066
15
+ airflow/providers/apache/hive/sensors/hive_partition.py,sha256=xYBw7hPI2mii01otkzImNy8S4yEyesX-F7t6X2Trf_M,3090
15
16
  airflow/providers/apache/hive/sensors/metastore_partition.py,sha256=H-luvQ96TRp5Oax0LIoag4rlnKiDlkJCsYNh80-IzXU,3365
16
- airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=D_jewkK_O-LuR5DXyN-fw_gIhW3C7fmLhrkOfwfgoi0,4167
17
+ airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=UpkokOVUD4W_8zxpu_96bpAlwyguMpfTAPktHnomG4Y,4191
17
18
  airflow/providers/apache/hive/transfers/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
18
- airflow/providers/apache/hive/transfers/hive_to_mysql.py,sha256=PX4clGkdNTyXDFi_Hj5WxaVzY-L8Z6DVEVTCyH6amD4,5536
19
- airflow/providers/apache/hive/transfers/hive_to_samba.py,sha256=tfyIt2I0IIQcYtubKJvhmwEjyo3m2-FffEuFc3aCHks,3237
20
- airflow/providers/apache/hive/transfers/mssql_to_hive.py,sha256=9NVgJESAwPvxl3D1e7WNhW0LDoWMiL-Fdr6Q9Yyc2GU,5756
21
- airflow/providers/apache/hive/transfers/mysql_to_hive.py,sha256=6BkAoYrn9wDfSC0BKHp86EgoUVtqwHGz9nZsielCydQ,7022
22
- airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=UDmTXp1mMNWl31qBe_Z3Sqp7VLJym6kQk3gsxkAxP3g,12417
23
- airflow/providers/apache/hive/transfers/vertica_to_hive.py,sha256=IBiMGF7322rOrwDbWMVB89oKO8hFhhBmc6rQwPkFA34,5555
24
- apache_airflow_providers_apache_hive-9.1.0.dist-info/entry_points.txt,sha256=Hzixt33mYYldwmwswarArUB7ZU0xbmUtd3tFViZ414s,185
25
- apache_airflow_providers_apache_hive-9.1.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
26
- apache_airflow_providers_apache_hive-9.1.0.dist-info/METADATA,sha256=RhQwm3ChWLCvmAuQOh5OetppPh8LVWOgxJR0VjZ3mNs,7104
27
- apache_airflow_providers_apache_hive-9.1.0.dist-info/RECORD,,
19
+ airflow/providers/apache/hive/transfers/hive_to_mysql.py,sha256=OypRFwRMYWvxdsI1-BbPLeuA7bPbW9ADsv6ppGGw9YQ,5296
20
+ airflow/providers/apache/hive/transfers/hive_to_samba.py,sha256=tgBzCSKYp27WfGr0P-ttkAZ4mNl_zcSkwpYAIUmKPUM,2998
21
+ airflow/providers/apache/hive/transfers/mssql_to_hive.py,sha256=Vwy8lfpN7uVOQ2IMAuZWgnS18K-TKZJDcaRs8gWnEco,5786
22
+ airflow/providers/apache/hive/transfers/mysql_to_hive.py,sha256=2V8mD38injLk9FYpSRqI9KfkQ62EFGGS9B6hLidlD74,7052
23
+ airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=qImzkrOD_XHKmlkcjMWZ-i08kOqncyKa42dnIyyoLkU,12447
24
+ airflow/providers/apache/hive/transfers/vertica_to_hive.py,sha256=ogMltscmRKQ-EG1io0Z2cJdHlN_oKrFaXMP7P2ubAkw,5585
25
+ apache_airflow_providers_apache_hive-9.1.1.dist-info/entry_points.txt,sha256=Hzixt33mYYldwmwswarArUB7ZU0xbmUtd3tFViZ414s,185
26
+ apache_airflow_providers_apache_hive-9.1.1.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
27
+ apache_airflow_providers_apache_hive-9.1.1.dist-info/METADATA,sha256=n1rS9I1ekjwBhmBAoOOmCA7eQIZENotISeXBGIyea7s,7051
28
+ apache_airflow_providers_apache_hive-9.1.1.dist-info/RECORD,,