databricks-sqlalchemy 2.0.7__py3-none-any.whl → 2.0.8__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5,6 +5,14 @@ from databricks.sqlalchemy._types import (
5
5
  TIMESTAMP_NTZ,
6
6
  DatabricksArray,
7
7
  DatabricksMap,
8
+ DatabricksVariant,
8
9
  )
9
10
 
10
- __all__ = ["TINYINT", "TIMESTAMP", "TIMESTAMP_NTZ", "DatabricksArray", "DatabricksMap"]
11
+ __all__ = [
12
+ "TINYINT",
13
+ "TIMESTAMP",
14
+ "TIMESTAMP_NTZ",
15
+ "DatabricksArray",
16
+ "DatabricksMap",
17
+ "DatabricksVariant",
18
+ ]
@@ -318,6 +318,7 @@ GET_COLUMNS_TYPE_MAP = {
318
318
  "map": sqlalchemy.types.String,
319
319
  "struct": sqlalchemy.types.String,
320
320
  "uniontype": sqlalchemy.types.String,
321
+ "variant": type_overrides.DatabricksVariant,
321
322
  "decimal": sqlalchemy.types.Numeric,
322
323
  "timestamp": type_overrides.TIMESTAMP,
323
324
  "timestamp_ntz": type_overrides.TIMESTAMP_NTZ,
@@ -9,6 +9,9 @@ from sqlalchemy.types import TypeDecorator, UserDefinedType
9
9
 
10
10
  from databricks.sql.utils import ParamEscaper
11
11
 
12
+ from sqlalchemy.sql import expression
13
+ import json
14
+
12
15
 
13
16
  def process_literal_param_hack(value: Any):
14
17
  """This method is supposed to accept a Python type and return a string representation of that type.
@@ -397,3 +400,60 @@ def compile_databricks_map(type_, compiler, **kw):
397
400
  key_type = compiler.process(type_.key_type, **kw)
398
401
  value_type = compiler.process(type_.value_type, **kw)
399
402
  return f"MAP<{key_type},{value_type}>"
403
+
404
+
405
+ class DatabricksVariant(UserDefinedType):
406
+ """
407
+ A custom variant type for storing semi-structured data including STRUCT, ARRAY, MAP, and scalar types.
408
+ Note: VARIANT MAP types can only have STRING keys.
409
+
410
+ Examples:
411
+ DatabricksVariant() -> VARIANT
412
+
413
+ Usage:
414
+ Column('data', DatabricksVariant())
415
+ """
416
+
417
+ cache_ok = True
418
+
419
+ def __init__(self):
420
+ self.pe = ParamEscaper()
421
+
422
+ def bind_processor(self, dialect):
423
+ """Process values before sending to database."""
424
+
425
+ def process(value):
426
+ if value is None:
427
+ return None
428
+ try:
429
+ return json.dumps(value, ensure_ascii=False, separators=(",", ":"))
430
+ except (TypeError, ValueError) as e:
431
+ raise ValueError(f"Cannot serialize value {value} to JSON: {e}")
432
+
433
+ return process
434
+
435
+ def bind_expression(self, bindvalue):
436
+ """Wrap with PARSE_JSON() in SQL"""
437
+ return expression.func.PARSE_JSON(bindvalue)
438
+
439
+ def literal_processor(self, dialect):
440
+ """Process literal values for SQL generation.
441
+ For VARIANT columns, use PARSE_JSON() to properly insert data.
442
+ """
443
+
444
+ def process(value):
445
+ if value is None:
446
+ return "NULL"
447
+ try:
448
+ return self.pe.escape_string(
449
+ json.dumps(value, ensure_ascii=False, separators=(",", ":"))
450
+ )
451
+ except (TypeError, ValueError) as e:
452
+ raise ValueError(f"Cannot serialize value {value} to JSON: {e}")
453
+
454
+ return process
455
+
456
+
457
+ @compiles(DatabricksVariant, "databricks")
458
+ def compile_variant(type_, compiler, **kw):
459
+ return "VARIANT"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: databricks-sqlalchemy
3
- Version: 2.0.7
3
+ Version: 2.0.8
4
4
  Summary: Databricks SQLAlchemy plugin for Python
5
5
  License: Apache-2.0
6
6
  Author: Databricks
@@ -1,15 +1,15 @@
1
1
  databricks/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
- databricks/sqlalchemy/__init__.py,sha256=IL3_QTbL9vbdQQ09Dzrol352VXU8v3oUilpiCkjvxQw,278
2
+ databricks/sqlalchemy/__init__.py,sha256=PArcuIQ1nGvovXlY3P-1ly766kh_9MgXsXB_Qgoglp4,349
3
3
  databricks/sqlalchemy/_ddl.py,sha256=c0_GwfmnrFVr4-Ls14fmdGUUFyUok_GW4Uo45hLABFc,3983
4
- databricks/sqlalchemy/_parse.py,sha256=aFpwcLowSDP1R7BY3G-yuEXiPFL-_VaIGvqKNDMehcQ,13049
5
- databricks/sqlalchemy/_types.py,sha256=nESbFFzf0qFYyomOOBnCAvRB710ABkPgxWMu7qc3mvw,14216
4
+ databricks/sqlalchemy/_parse.py,sha256=oGhErqLLDyGNvVOybfIqYAMJoAa9FhjU6hQODAD8a1w,13098
5
+ databricks/sqlalchemy/_types.py,sha256=SRK0OFn4OORilxZMAnA6QlQ54OGJnt81THTZvRAGa6I,15953
6
6
  databricks/sqlalchemy/base.py,sha256=KcjfHMH0NsceYE2NRxrePtf5T1uw9u8JHofRdbnAKS4,15619
7
7
  databricks/sqlalchemy/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
8
8
  databricks/sqlalchemy/pytest.ini,sha256=ImutflUjkhByVNWCQ18Todj6XTvgJAQX_v7fD-gWHhU,106
9
9
  databricks/sqlalchemy/requirements.py,sha256=OobunAEwZ9y2dvSQLOmdgJciVn9xGlY9NAFfszPCTU0,9018
10
10
  databricks/sqlalchemy/setup.cfg,sha256=ImutflUjkhByVNWCQ18Todj6XTvgJAQX_v7fD-gWHhU,106
11
- databricks_sqlalchemy-2.0.7.dist-info/LICENSE,sha256=WgVm2VpfZ3CsUfPndD2NeCrEIcFA4UB-YnnW4ejxcbE,11346
12
- databricks_sqlalchemy-2.0.7.dist-info/METADATA,sha256=QIT9EVR2yGQy5N5Z8uDJ9TaMw0li8OXZ1PAS-U2WB5I,12764
13
- databricks_sqlalchemy-2.0.7.dist-info/WHEEL,sha256=b4K_helf-jlQoXBBETfwnf4B04YC67LOev0jo4fX5m8,88
14
- databricks_sqlalchemy-2.0.7.dist-info/entry_points.txt,sha256=AAjpsvZbVcoMAcWLIesoAT5FNZhBEcIhxdKknVua3jw,74
15
- databricks_sqlalchemy-2.0.7.dist-info/RECORD,,
11
+ databricks_sqlalchemy-2.0.8.dist-info/LICENSE,sha256=WgVm2VpfZ3CsUfPndD2NeCrEIcFA4UB-YnnW4ejxcbE,11346
12
+ databricks_sqlalchemy-2.0.8.dist-info/METADATA,sha256=YgQ-w-A6zMEVSeTneL_YHJUK1VmUuj7-c4CCUZqBh04,12764
13
+ databricks_sqlalchemy-2.0.8.dist-info/WHEEL,sha256=b4K_helf-jlQoXBBETfwnf4B04YC67LOev0jo4fX5m8,88
14
+ databricks_sqlalchemy-2.0.8.dist-info/entry_points.txt,sha256=AAjpsvZbVcoMAcWLIesoAT5FNZhBEcIhxdKknVua3jw,74
15
+ databricks_sqlalchemy-2.0.8.dist-info/RECORD,,