dlt-utils-lib 0.3.0__tar.gz → 0.3.2__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -0,0 +1,11 @@
1
+ Metadata-Version: 2.1
2
+ Name: dlt_utils_lib
3
+ Version: 0.3.2
4
+ Summary: UNKNOWN
5
+ Home-page: UNKNOWN
6
+ License: UNKNOWN
7
+ Platform: UNKNOWN
8
+ Provides-Extra: dev
9
+
10
+ UNKNOWN
11
+
@@ -66,7 +66,7 @@ def create_bronze_table_definition(spark, dlt, table_name: str, files_path: str,
66
66
  .withColumn('ar_h_change_seq', col('ar_h_change_seq').cast('string')) \
67
67
  .transform(lambda df: update_cdc_timestamp(df, time_diff_history_cdc_timestamp)) \
68
68
  .transform(lambda df: apply_partitions(df, partitions)) \
69
- .transform(lambda df: add_default_value_for_removed_col(df, default_value_for_removed_col)) \
69
+ .transform(lambda df: add_default_value_for_removed_col(df, default_value_for_removed_col))
70
70
 
71
71
  return transform_cdc_to_bronze
72
72
 
@@ -0,0 +1,11 @@
1
+ Metadata-Version: 2.1
2
+ Name: dlt-utils-lib
3
+ Version: 0.3.2
4
+ Summary: UNKNOWN
5
+ Home-page: UNKNOWN
6
+ License: UNKNOWN
7
+ Platform: UNKNOWN
8
+ Provides-Extra: dev
9
+
10
+ UNKNOWN
11
+
@@ -9,4 +9,5 @@ dlt_utils_lib.egg-info/requires.txt
9
9
  dlt_utils_lib.egg-info/tmpz2kz6zic
10
10
  dlt_utils_lib.egg-info/top_level.txt
11
11
  tests/__init__.py
12
+ tests/conftest.py
12
13
  tests/test_transformation_module.py
File without changes
@@ -2,7 +2,7 @@ from setuptools import setup, find_packages
2
2
 
3
3
  setup(
4
4
  name='dlt_utils_lib',
5
- version='0.3.0',
5
+ version='0.3.2',
6
6
  packages=find_packages(),
7
7
  install_requires=[
8
8
  'pyspark',
@@ -0,0 +1,12 @@
1
+ import pytest
2
+ from pyspark.sql import SparkSession
3
+
4
+ @pytest.fixture(scope='session')
5
+ def spark():
6
+ """Fixture for creating a Spark session."""
7
+ spark = SparkSession.builder \
8
+ .appName("PySpark Test") \
9
+ .master("local[*]") \
10
+ .getOrCreate()
11
+ yield spark
12
+ spark.stop()
@@ -1,21 +1,26 @@
1
1
  import sys
2
2
  import os
3
- from databricks.sdk.runtime import *
4
3
  import pytest
5
- from pyspark.sql import SparkSession
6
- from pyspark.sql import Row
4
+ from unittest import mock
5
+ from pyspark.sql import SparkSession, Row
7
6
  from pyspark.sql.functions import expr
8
7
  from pyspark.sql.types import TimestampType, StringType
9
8
  from datetime import datetime
10
- import warnings
11
-
12
- # import for databricks execution whivh not support normal PYTHONPATH :()
13
- sys.path.append(os.path.abspath(os.path.join(os.getcwd(), "../dlt_utils_lib")))
14
- from dlt_utils.dlt_transformations import update_cdc_timestamp, apply_partitions, add_default_value_for_removed_col, rename_columns
15
-
16
- @pytest.fixture(scope="session")
17
- def spark() -> SparkSession:
18
- return SparkSession.builder.getOrCreate()
9
+ from dlt_utils.dlt_transformations import rename_columns, apply_partitions, update_cdc_timestamp, add_default_value_for_removed_col, rename_columns
10
+ # Mock databricks.sdk.runtime if not in a Databricks environment
11
+ try:
12
+ from databricks.sdk.runtime import *
13
+ except ImportError:
14
+ print("Databricks runtime not available. Mocking required modules.")
15
+ mock_dbutils = mock.Mock()
16
+ sys.modules['databricks.sdk.runtime'] = mock.Mock()
17
+ sys.modules['dbruntime'] = mock.Mock()
18
+ sys.modules['databricks'] = mock.Mock()
19
+
20
+ # Mocking any specific methods you might call in your code
21
+ mock_dbutils.jobs = mock.Mock()
22
+ mock_dbutils.widgets = mock.Mock()
23
+ mock_dbutils.notebook = mock.Mock()
19
24
 
20
25
 
21
26
  def test_apply_partitions(spark):
@@ -1,4 +0,0 @@
1
- Metadata-Version: 2.1
2
- Name: dlt_utils_lib
3
- Version: 0.3.0
4
- Provides-Extra: dev
@@ -1,4 +0,0 @@
1
- Metadata-Version: 2.1
2
- Name: dlt-utils-lib
3
- Version: 0.3.0
4
- Provides-Extra: dev