dlt-utils-lib 0.3.0__tar.gz → 0.3.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- dlt_utils_lib-0.3.2/PKG-INFO +11 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils/__init__.py +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils/dlt_transformations.py +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils/main_cdc_utils.py +1 -1
- dlt_utils_lib-0.3.2/dlt_utils_lib.egg-info/PKG-INFO +11 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils_lib.egg-info/SOURCES.txt +1 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils_lib.egg-info/dependency_links.txt +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils_lib.egg-info/requires.txt +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils_lib.egg-info/tmpz2kz6zic +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/dlt_utils_lib.egg-info/top_level.txt +1 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/setup.cfg +0 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/setup.py +1 -1
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/tests/__init__.py +0 -0
- dlt_utils_lib-0.3.2/tests/conftest.py +12 -0
- {dlt_utils_lib-0.3.0 → dlt_utils_lib-0.3.2}/tests/test_transformation_module.py +17 -12
- dlt_utils_lib-0.3.0/PKG-INFO +0 -4
- dlt_utils_lib-0.3.0/dlt_utils_lib.egg-info/PKG-INFO +0 -4
|
File without changes
|
|
File without changes
|
|
@@ -66,7 +66,7 @@ def create_bronze_table_definition(spark, dlt, table_name: str, files_path: str,
|
|
|
66
66
|
.withColumn('ar_h_change_seq', col('ar_h_change_seq').cast('string')) \
|
|
67
67
|
.transform(lambda df: update_cdc_timestamp(df, time_diff_history_cdc_timestamp)) \
|
|
68
68
|
.transform(lambda df: apply_partitions(df, partitions)) \
|
|
69
|
-
.transform(lambda df: add_default_value_for_removed_col(df, default_value_for_removed_col))
|
|
69
|
+
.transform(lambda df: add_default_value_for_removed_col(df, default_value_for_removed_col))
|
|
70
70
|
|
|
71
71
|
return transform_cdc_to_bronze
|
|
72
72
|
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
@@ -0,0 +1,12 @@
|
|
|
1
|
+
import pytest
|
|
2
|
+
from pyspark.sql import SparkSession
|
|
3
|
+
|
|
4
|
+
@pytest.fixture(scope='session')
|
|
5
|
+
def spark():
|
|
6
|
+
"""Fixture for creating a Spark session."""
|
|
7
|
+
spark = SparkSession.builder \
|
|
8
|
+
.appName("PySpark Test") \
|
|
9
|
+
.master("local[*]") \
|
|
10
|
+
.getOrCreate()
|
|
11
|
+
yield spark
|
|
12
|
+
spark.stop()
|
|
@@ -1,21 +1,26 @@
|
|
|
1
1
|
import sys
|
|
2
2
|
import os
|
|
3
|
-
from databricks.sdk.runtime import *
|
|
4
3
|
import pytest
|
|
5
|
-
from
|
|
6
|
-
from pyspark.sql import Row
|
|
4
|
+
from unittest import mock
|
|
5
|
+
from pyspark.sql import SparkSession, Row
|
|
7
6
|
from pyspark.sql.functions import expr
|
|
8
7
|
from pyspark.sql.types import TimestampType, StringType
|
|
9
8
|
from datetime import datetime
|
|
10
|
-
import
|
|
11
|
-
|
|
12
|
-
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
|
|
18
|
-
|
|
9
|
+
from dlt_utils.dlt_transformations import rename_columns, apply_partitions, update_cdc_timestamp, add_default_value_for_removed_col, rename_columns
|
|
10
|
+
# Mock databricks.sdk.runtime if not in a Databricks environment
|
|
11
|
+
try:
|
|
12
|
+
from databricks.sdk.runtime import *
|
|
13
|
+
except ImportError:
|
|
14
|
+
print("Databricks runtime not available. Mocking required modules.")
|
|
15
|
+
mock_dbutils = mock.Mock()
|
|
16
|
+
sys.modules['databricks.sdk.runtime'] = mock.Mock()
|
|
17
|
+
sys.modules['dbruntime'] = mock.Mock()
|
|
18
|
+
sys.modules['databricks'] = mock.Mock()
|
|
19
|
+
|
|
20
|
+
# Mocking any specific methods you might call in your code
|
|
21
|
+
mock_dbutils.jobs = mock.Mock()
|
|
22
|
+
mock_dbutils.widgets = mock.Mock()
|
|
23
|
+
mock_dbutils.notebook = mock.Mock()
|
|
19
24
|
|
|
20
25
|
|
|
21
26
|
def test_apply_partitions(spark):
|
dlt_utils_lib-0.3.0/PKG-INFO
DELETED