apache-airflow-providers-apache-hive 9.0.5__py3-none-any.whl → 9.0.6__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -29,7 +29,7 @@ from airflow import __version__ as airflow_version
29
29
 
30
30
  __all__ = ["__version__"]
31
31
 
32
- __version__ = "9.0.5"
32
+ __version__ = "9.0.6"
33
33
 
34
34
  if packaging.version.parse(packaging.version.parse(airflow_version).base_version) < packaging.version.parse(
35
35
  "2.9.0"
@@ -26,61 +26,6 @@ def get_provider_info():
26
26
  "package-name": "apache-airflow-providers-apache-hive",
27
27
  "name": "Apache Hive",
28
28
  "description": "`Apache Hive <https://hive.apache.org/>`__\n",
29
- "state": "ready",
30
- "source-date-epoch": 1743835719,
31
- "versions": [
32
- "9.0.5",
33
- "9.0.4",
34
- "9.0.3",
35
- "9.0.2",
36
- "9.0.0",
37
- "8.2.1",
38
- "8.2.0",
39
- "8.1.2",
40
- "8.1.1",
41
- "8.1.0",
42
- "8.0.0",
43
- "7.0.1",
44
- "7.0.0",
45
- "6.4.2",
46
- "6.4.1",
47
- "6.4.0",
48
- "6.3.0",
49
- "6.2.0",
50
- "6.1.6",
51
- "6.1.5",
52
- "6.1.4",
53
- "6.1.3",
54
- "6.1.2",
55
- "6.1.1",
56
- "6.1.0",
57
- "6.0.0",
58
- "5.1.3",
59
- "5.1.2",
60
- "5.1.1",
61
- "5.1.0",
62
- "5.0.0",
63
- "4.1.1",
64
- "4.1.0",
65
- "4.0.1",
66
- "4.0.0",
67
- "3.1.0",
68
- "3.0.0",
69
- "2.3.3",
70
- "2.3.2",
71
- "2.3.1",
72
- "2.3.0",
73
- "2.2.0",
74
- "2.1.0",
75
- "2.0.3",
76
- "2.0.2",
77
- "2.0.1",
78
- "2.0.0",
79
- "1.0.3",
80
- "1.0.2",
81
- "1.0.1",
82
- "1.0.0",
83
- ],
84
29
  "integrations": [
85
30
  {
86
31
  "integration-name": "Apache Hive",
@@ -185,23 +130,4 @@ def get_provider_info():
185
130
  },
186
131
  }
187
132
  },
188
- "dependencies": [
189
- "apache-airflow>=2.9.0",
190
- "apache-airflow-providers-common-sql>=1.20.0",
191
- "hmsclient>=0.1.0",
192
- "pandas>=2.1.2,<2.2",
193
- "pyhive[hive_pure_sasl]>=0.7.0",
194
- "thrift>=0.11.0",
195
- "jmespath>=0.7.0",
196
- ],
197
- "optional-dependencies": {
198
- "amazon": ["apache-airflow-providers-amazon"],
199
- "microsoft.mssql": ["apache-airflow-providers-microsoft-mssql"],
200
- "mysql": ["apache-airflow-providers-mysql"],
201
- "presto": ["apache-airflow-providers-presto"],
202
- "samba": ["apache-airflow-providers-samba"],
203
- "vertica": ["apache-airflow-providers-vertica"],
204
- "common.compat": ["apache-airflow-providers-common-compat"],
205
- },
206
- "devel-dependencies": [],
207
133
  }
@@ -209,7 +209,7 @@ class HiveCliHook(BaseHook):
209
209
  f"The schema used in beeline command ({conn.schema}) should not contain ';' character)"
210
210
  )
211
211
  return
212
- elif ":" in conn.host or "/" in conn.host or ";" in conn.host:
212
+ if ":" in conn.host or "/" in conn.host or ";" in conn.host:
213
213
  raise ValueError(
214
214
  f"The host used in beeline command ({conn.host}) should not contain ':/;' characters)"
215
215
  )
@@ -609,8 +609,7 @@ class HiveMetastoreHook(BaseHook):
609
609
  self.log.info("Connected to %s:%s", host, conn.port)
610
610
  host_socket.close()
611
611
  return host
612
- else:
613
- self.log.error("Could not connect to %s:%s", host, conn.port)
612
+ self.log.error("Could not connect to %s:%s", host, conn.port)
614
613
  return None
615
614
 
616
615
  def get_conn(self) -> Any:
@@ -713,8 +712,7 @@ class HiveMetastoreHook(BaseHook):
713
712
 
714
713
  pnames = [p.name for p in table.partitionKeys]
715
714
  return [dict(zip(pnames, p.values)) for p in parts]
716
- else:
717
- raise AirflowException("The table isn't partitioned")
715
+ raise AirflowException("The table isn't partitioned")
718
716
 
719
717
  @staticmethod
720
718
  def _get_max_partition_from_part_specs(
@@ -75,8 +75,7 @@ def _closest_date(target_dt, date_list, before_target=None) -> datetime.date | N
75
75
  return min(date_list, key=any_time).date()
76
76
  if before_target:
77
77
  return min(date_list, key=time_before).date()
78
- else:
79
- return min(date_list, key=time_after).date()
78
+ return min(date_list, key=time_after).date()
80
79
 
81
80
 
82
81
  def closest_ds_partition(
@@ -109,10 +108,9 @@ def closest_ds_partition(
109
108
  part_vals = [next(iter(p.values())) for p in partitions]
110
109
  if ds in part_vals:
111
110
  return ds
112
- else:
113
- parts = [datetime.datetime.strptime(pv, "%Y-%m-%d") for pv in part_vals]
114
- target_dt = datetime.datetime.strptime(ds, "%Y-%m-%d")
115
- closest_ds = _closest_date(target_dt, parts, before_target=before)
116
- if closest_ds is not None:
117
- return closest_ds.isoformat()
111
+ parts = [datetime.datetime.strptime(pv, "%Y-%m-%d") for pv in part_vals]
112
+ target_dt = datetime.datetime.strptime(ds, "%Y-%m-%d")
113
+ closest_ds = _closest_date(target_dt, parts, before_target=before)
114
+ if closest_ds is not None:
115
+ return closest_ds.isoformat()
118
116
  return None
@@ -79,8 +79,7 @@ class NamedHivePartitionSensor(BaseSensorOperator):
79
79
  second_split = table_partition.split("/", 1)
80
80
  if len(second_split) == 1:
81
81
  raise ValueError(f"Could not parse {partition}into table, partition")
82
- else:
83
- table, partition = second_split
82
+ table, partition = second_split
84
83
  return schema, table, partition
85
84
 
86
85
  def poke_partition(self, partition: str) -> Any:
@@ -253,13 +253,12 @@ class S3ToHiveOperator(BaseOperator):
253
253
  test_field_match = all(h1.lower() == h2.lower() for h1, h2 in zip(header_list, field_names))
254
254
  if test_field_match:
255
255
  return True
256
- else:
257
- self.log.warning(
258
- "Headers do not match field names File headers:\n %s\nField names: \n %s\n",
259
- header_list,
260
- field_names,
261
- )
262
- return False
256
+ self.log.warning(
257
+ "Headers do not match field names File headers:\n %s\nField names: \n %s\n",
258
+ header_list,
259
+ field_names,
260
+ )
261
+ return False
263
262
 
264
263
  @staticmethod
265
264
  def _delete_top_row_and_compress(input_file_name, output_file_ext, dest_dir):
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: apache-airflow-providers-apache-hive
3
- Version: 9.0.5
3
+ Version: 9.0.6
4
4
  Summary: Provider package apache-airflow-providers-apache-hive for Apache Airflow
5
5
  Keywords: airflow-provider,apache.hive,airflow,integration
6
6
  Author-email: Apache Software Foundation <dev@airflow.apache.org>
@@ -35,8 +35,8 @@ Requires-Dist: apache-airflow-providers-presto ; extra == "presto"
35
35
  Requires-Dist: apache-airflow-providers-samba ; extra == "samba"
36
36
  Requires-Dist: apache-airflow-providers-vertica ; extra == "vertica"
37
37
  Project-URL: Bug Tracker, https://github.com/apache/airflow/issues
38
- Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.5/changelog.html
39
- Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.5
38
+ Project-URL: Changelog, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.6/changelog.html
39
+ Project-URL: Documentation, https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.6
40
40
  Project-URL: Mastodon, https://fosstodon.org/@airflow
41
41
  Project-URL: Slack Chat, https://s.apache.org/airflow-slack
42
42
  Project-URL: Source Code, https://github.com/apache/airflow
@@ -74,7 +74,7 @@ Provides-Extra: vertica
74
74
 
75
75
  Package ``apache-airflow-providers-apache-hive``
76
76
 
77
- Release: ``9.0.5``
77
+ Release: ``9.0.6``
78
78
 
79
79
 
80
80
  `Apache Hive <https://hive.apache.org/>`__
@@ -87,7 +87,7 @@ This is a provider package for ``apache.hive`` provider. All classes for this pr
87
87
  are in ``airflow.providers.apache.hive`` python package.
88
88
 
89
89
  You can find package information and changelog for the provider
90
- in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.5/>`_.
90
+ in the `documentation <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.6/>`_.
91
91
 
92
92
  Installation
93
93
  ------------
@@ -140,5 +140,5 @@ Dependent package
140
140
  ====================================================================================================================== ===================
141
141
 
142
142
  The changelog for the provider package can be found in the
143
- `changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.5/changelog.html>`_.
143
+ `changelog <https://airflow.apache.org/docs/apache-airflow-providers-apache-hive/9.0.6/changelog.html>`_.
144
144
 
@@ -1,10 +1,10 @@
1
1
  airflow/providers/apache/hive/LICENSE,sha256=gXPVwptPlW1TJ4HSuG5OMPg-a3h43OGMkZRR1rpwfJA,10850
2
- airflow/providers/apache/hive/__init__.py,sha256=6jsYbyQlPUwqObw9uAE8YaSLhIzanjZIg0UinkJojBs,1498
3
- airflow/providers/apache/hive/get_provider_info.py,sha256=2kYTOB-iJwO00I1Xq6qg2c25ldMl_mcY2yJaTcv7cAI,7568
2
+ airflow/providers/apache/hive/__init__.py,sha256=vNoiAoYLF_9C7nqGSP4r3R85TLGsvMfUPet08HE40YM,1498
3
+ airflow/providers/apache/hive/get_provider_info.py,sha256=PdxdiZ7fdK1D-BTyZqj1oMS_1GS8_kbz2hZVgWMTkFI,5569
4
4
  airflow/providers/apache/hive/hooks/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
5
- airflow/providers/apache/hive/hooks/hive.py,sha256=oDLegXFihuehkxZSAePvY3pP7Huc7vsiVJBkRkTL9jo,42825
5
+ airflow/providers/apache/hive/hooks/hive.py,sha256=GZYlNhwI0HF9RqTzU3X7BBs-TwVqndoXTEvcT5Yphdw,42779
6
6
  airflow/providers/apache/hive/macros/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
7
- airflow/providers/apache/hive/macros/hive.py,sha256=p3LUIb74pbthh8lcUXV6f3m-cZrcRXjV8aGiazdpnzo,4592
7
+ airflow/providers/apache/hive/macros/hive.py,sha256=FEMCR9rEWKFOKtKAvhVWhD5jRJnSYeHepVGgJLzUa2k,4548
8
8
  airflow/providers/apache/hive/operators/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
9
9
  airflow/providers/apache/hive/operators/hive.py,sha256=QU7-zSc57bk8mNpFVuAmBmIcHiyqiN-OMCzyFwceETY,7348
10
10
  airflow/providers/apache/hive/operators/hive_stats.py,sha256=8fNlR9o8J9DKMDBpB849apqA2iGbMovlHOWute4TFCY,7074
@@ -13,15 +13,15 @@ airflow/providers/apache/hive/plugins/hive.py,sha256=rrGccro6DEdnHSmII3goBYD9te-
13
13
  airflow/providers/apache/hive/sensors/__init__.py,sha256=mlJxuZLkd5x-iq2SBwD3mvRQpt3YR7wjz_nceyF1IaI,787
14
14
  airflow/providers/apache/hive/sensors/hive_partition.py,sha256=UIafLmwTUXdNfWAhLOOSZQc6v6So7lz_0bYA2Wyunw4,3066
15
15
  airflow/providers/apache/hive/sensors/metastore_partition.py,sha256=H-luvQ96TRp5Oax0LIoag4rlnKiDlkJCsYNh80-IzXU,3365
16
- airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=5R3NJ_Vi9nqROD9VPs-YXnziYhWd-X5ukgelLwfngy4,4185
16
+ airflow/providers/apache/hive/sensors/named_hive_partition.py,sha256=D_jewkK_O-LuR5DXyN-fw_gIhW3C7fmLhrkOfwfgoi0,4167
17
17
  airflow/providers/apache/hive/transfers/__init__.py,sha256=9hdXHABrVpkbpjZgUft39kOFL2xSGeG4GEua0Hmelus,785
18
18
  airflow/providers/apache/hive/transfers/hive_to_mysql.py,sha256=PX4clGkdNTyXDFi_Hj5WxaVzY-L8Z6DVEVTCyH6amD4,5536
19
19
  airflow/providers/apache/hive/transfers/hive_to_samba.py,sha256=tfyIt2I0IIQcYtubKJvhmwEjyo3m2-FffEuFc3aCHks,3237
20
20
  airflow/providers/apache/hive/transfers/mssql_to_hive.py,sha256=9NVgJESAwPvxl3D1e7WNhW0LDoWMiL-Fdr6Q9Yyc2GU,5756
21
21
  airflow/providers/apache/hive/transfers/mysql_to_hive.py,sha256=6BkAoYrn9wDfSC0BKHp86EgoUVtqwHGz9nZsielCydQ,7022
22
- airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=K5Zke6CF9Rgw7YDkOtWIYXC3_2TuXLTuwYxu8uVrGP0,12455
22
+ airflow/providers/apache/hive/transfers/s3_to_hive.py,sha256=UDmTXp1mMNWl31qBe_Z3Sqp7VLJym6kQk3gsxkAxP3g,12417
23
23
  airflow/providers/apache/hive/transfers/vertica_to_hive.py,sha256=IBiMGF7322rOrwDbWMVB89oKO8hFhhBmc6rQwPkFA34,5555
24
- apache_airflow_providers_apache_hive-9.0.5.dist-info/entry_points.txt,sha256=Hzixt33mYYldwmwswarArUB7ZU0xbmUtd3tFViZ414s,185
25
- apache_airflow_providers_apache_hive-9.0.5.dist-info/WHEEL,sha256=_2ozNFCLWc93bK4WKHCO-eDUENDlo-dgc9cU3qokYO4,82
26
- apache_airflow_providers_apache_hive-9.0.5.dist-info/METADATA,sha256=ni641jvy7YGcw5LwF8y9ok2TdYawCKBLGZirpVG2Yrw,7194
27
- apache_airflow_providers_apache_hive-9.0.5.dist-info/RECORD,,
24
+ apache_airflow_providers_apache_hive-9.0.6.dist-info/entry_points.txt,sha256=Hzixt33mYYldwmwswarArUB7ZU0xbmUtd3tFViZ414s,185
25
+ apache_airflow_providers_apache_hive-9.0.6.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
26
+ apache_airflow_providers_apache_hive-9.0.6.dist-info/METADATA,sha256=rcSnAiHtHCbTln0U1bEn-AW6oJHI9XxtBqbD-ijE0Ns,7194
27
+ apache_airflow_providers_apache_hive-9.0.6.dist-info/RECORD,,
@@ -1,4 +1,4 @@
1
1
  Wheel-Version: 1.0
2
- Generator: flit 3.11.0
2
+ Generator: flit 3.12.0
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any