databricks-sqlalchemy 2.0.6__tar.gz → 2.0.8__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/CHANGELOG.md +8 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/PKG-INFO +3 -2
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/pyproject.toml +1 -1
- databricks_sqlalchemy-2.0.8/src/databricks/sqlalchemy/__init__.py +18 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_parse.py +1 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_types.py +136 -0
- databricks_sqlalchemy-2.0.6/src/databricks/sqlalchemy/__init__.py +0 -4
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/LICENSE +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/README.md +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/__init__.py +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_ddl.py +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/base.py +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/py.typed +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/pytest.ini +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/requirements.py +0 -0
- {databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/setup.cfg +0 -0
@@ -1,5 +1,13 @@
|
|
1
1
|
# Release History
|
2
2
|
|
3
|
+
# 2.0.8 (2025-09-08)
|
4
|
+
|
5
|
+
- Feature: Added support for variant datatype (databricks/databricks-sqlalchemy#42 by @msrathore-db)
|
6
|
+
|
7
|
+
# 2.0.7 (2025-06-23)
|
8
|
+
|
9
|
+
- Feature: Added support for complex data types such as DatabricksArray and DatabricksMap [Private Preview] (databricks/databricks-sqlalchemy#30 by @jprakash-db)
|
10
|
+
|
3
11
|
# 2.0.6 (2025-04-29)
|
4
12
|
|
5
13
|
- Relaxed pin for `pyarrow` (databricks/databricks-sqlalchemy#20 by @dhirschfeld)
|
@@ -1,6 +1,6 @@
|
|
1
|
-
Metadata-Version: 2.
|
1
|
+
Metadata-Version: 2.3
|
2
2
|
Name: databricks-sqlalchemy
|
3
|
-
Version: 2.0.
|
3
|
+
Version: 2.0.8
|
4
4
|
Summary: Databricks SQLAlchemy plugin for Python
|
5
5
|
License: Apache-2.0
|
6
6
|
Author: Databricks
|
@@ -13,6 +13,7 @@ Classifier: Programming Language :: Python :: 3.9
|
|
13
13
|
Classifier: Programming Language :: Python :: 3.10
|
14
14
|
Classifier: Programming Language :: Python :: 3.11
|
15
15
|
Classifier: Programming Language :: Python :: 3.12
|
16
|
+
Classifier: Programming Language :: Python :: 3.13
|
16
17
|
Requires-Dist: databricks_sql_connector (>=4.0.0)
|
17
18
|
Requires-Dist: pyarrow (>=14.0.1)
|
18
19
|
Requires-Dist: sqlalchemy (>=2.0.21)
|
@@ -0,0 +1,18 @@
|
|
1
|
+
from databricks.sqlalchemy.base import DatabricksDialect
|
2
|
+
from databricks.sqlalchemy._types import (
|
3
|
+
TINYINT,
|
4
|
+
TIMESTAMP,
|
5
|
+
TIMESTAMP_NTZ,
|
6
|
+
DatabricksArray,
|
7
|
+
DatabricksMap,
|
8
|
+
DatabricksVariant,
|
9
|
+
)
|
10
|
+
|
11
|
+
__all__ = [
|
12
|
+
"TINYINT",
|
13
|
+
"TIMESTAMP",
|
14
|
+
"TIMESTAMP_NTZ",
|
15
|
+
"DatabricksArray",
|
16
|
+
"DatabricksMap",
|
17
|
+
"DatabricksVariant",
|
18
|
+
]
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_parse.py
RENAMED
@@ -318,6 +318,7 @@ GET_COLUMNS_TYPE_MAP = {
|
|
318
318
|
"map": sqlalchemy.types.String,
|
319
319
|
"struct": sqlalchemy.types.String,
|
320
320
|
"uniontype": sqlalchemy.types.String,
|
321
|
+
"variant": type_overrides.DatabricksVariant,
|
321
322
|
"decimal": sqlalchemy.types.Numeric,
|
322
323
|
"timestamp": type_overrides.TIMESTAMP,
|
323
324
|
"timestamp_ntz": type_overrides.TIMESTAMP_NTZ,
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_types.py
RENAMED
@@ -5,9 +5,13 @@ from typing import Any, Union, Optional
|
|
5
5
|
import sqlalchemy
|
6
6
|
from sqlalchemy.engine.interfaces import Dialect
|
7
7
|
from sqlalchemy.ext.compiler import compiles
|
8
|
+
from sqlalchemy.types import TypeDecorator, UserDefinedType
|
8
9
|
|
9
10
|
from databricks.sql.utils import ParamEscaper
|
10
11
|
|
12
|
+
from sqlalchemy.sql import expression
|
13
|
+
import json
|
14
|
+
|
11
15
|
|
12
16
|
def process_literal_param_hack(value: Any):
|
13
17
|
"""This method is supposed to accept a Python type and return a string representation of that type.
|
@@ -26,6 +30,11 @@ def process_literal_param_hack(value: Any):
|
|
26
30
|
return value
|
27
31
|
|
28
32
|
|
33
|
+
def identity_processor(value):
|
34
|
+
"""This method returns the value itself, when no other processor is provided"""
|
35
|
+
return value
|
36
|
+
|
37
|
+
|
29
38
|
@compiles(sqlalchemy.types.Enum, "databricks")
|
30
39
|
@compiles(sqlalchemy.types.String, "databricks")
|
31
40
|
@compiles(sqlalchemy.types.Text, "databricks")
|
@@ -321,3 +330,130 @@ class TINYINT(sqlalchemy.types.TypeDecorator):
|
|
321
330
|
@compiles(TINYINT, "databricks")
|
322
331
|
def compile_tinyint(type_, compiler, **kw):
|
323
332
|
return "TINYINT"
|
333
|
+
|
334
|
+
|
335
|
+
class DatabricksArray(UserDefinedType):
|
336
|
+
"""
|
337
|
+
A custom array type that can wrap any other SQLAlchemy type.
|
338
|
+
|
339
|
+
Examples:
|
340
|
+
DatabricksArray(String) -> ARRAY<STRING>
|
341
|
+
DatabricksArray(Integer) -> ARRAY<INT>
|
342
|
+
DatabricksArray(CustomType) -> ARRAY<CUSTOM_TYPE>
|
343
|
+
"""
|
344
|
+
|
345
|
+
def __init__(self, item_type):
|
346
|
+
self.item_type = item_type() if isinstance(item_type, type) else item_type
|
347
|
+
|
348
|
+
def bind_processor(self, dialect):
|
349
|
+
item_processor = self.item_type.bind_processor(dialect)
|
350
|
+
if item_processor is None:
|
351
|
+
item_processor = identity_processor
|
352
|
+
|
353
|
+
def process(value):
|
354
|
+
return [item_processor(val) for val in value]
|
355
|
+
|
356
|
+
return process
|
357
|
+
|
358
|
+
|
359
|
+
@compiles(DatabricksArray, "databricks")
|
360
|
+
def compile_databricks_array(type_, compiler, **kw):
|
361
|
+
inner = compiler.process(type_.item_type, **kw)
|
362
|
+
|
363
|
+
return f"ARRAY<{inner}>"
|
364
|
+
|
365
|
+
|
366
|
+
class DatabricksMap(UserDefinedType):
|
367
|
+
"""
|
368
|
+
A custom map type that can wrap any other SQLAlchemy types for both key and value.
|
369
|
+
|
370
|
+
Examples:
|
371
|
+
DatabricksMap(String, String) -> MAP<STRING,STRING>
|
372
|
+
DatabricksMap(Integer, String) -> MAP<INT,STRING>
|
373
|
+
DatabricksMap(String, DatabricksArray(Integer)) -> MAP<STRING,ARRAY<INT>>
|
374
|
+
"""
|
375
|
+
|
376
|
+
def __init__(self, key_type, value_type):
|
377
|
+
self.key_type = key_type() if isinstance(key_type, type) else key_type
|
378
|
+
self.value_type = value_type() if isinstance(value_type, type) else value_type
|
379
|
+
|
380
|
+
def bind_processor(self, dialect):
|
381
|
+
key_processor = self.key_type.bind_processor(dialect)
|
382
|
+
value_processor = self.value_type.bind_processor(dialect)
|
383
|
+
|
384
|
+
if key_processor is None:
|
385
|
+
key_processor = identity_processor
|
386
|
+
if value_processor is None:
|
387
|
+
value_processor = identity_processor
|
388
|
+
|
389
|
+
def process(value):
|
390
|
+
return {
|
391
|
+
key_processor(key): value_processor(value)
|
392
|
+
for key, value in value.items()
|
393
|
+
}
|
394
|
+
|
395
|
+
return process
|
396
|
+
|
397
|
+
|
398
|
+
@compiles(DatabricksMap, "databricks")
|
399
|
+
def compile_databricks_map(type_, compiler, **kw):
|
400
|
+
key_type = compiler.process(type_.key_type, **kw)
|
401
|
+
value_type = compiler.process(type_.value_type, **kw)
|
402
|
+
return f"MAP<{key_type},{value_type}>"
|
403
|
+
|
404
|
+
|
405
|
+
class DatabricksVariant(UserDefinedType):
|
406
|
+
"""
|
407
|
+
A custom variant type for storing semi-structured data including STRUCT, ARRAY, MAP, and scalar types.
|
408
|
+
Note: VARIANT MAP types can only have STRING keys.
|
409
|
+
|
410
|
+
Examples:
|
411
|
+
DatabricksVariant() -> VARIANT
|
412
|
+
|
413
|
+
Usage:
|
414
|
+
Column('data', DatabricksVariant())
|
415
|
+
"""
|
416
|
+
|
417
|
+
cache_ok = True
|
418
|
+
|
419
|
+
def __init__(self):
|
420
|
+
self.pe = ParamEscaper()
|
421
|
+
|
422
|
+
def bind_processor(self, dialect):
|
423
|
+
"""Process values before sending to database."""
|
424
|
+
|
425
|
+
def process(value):
|
426
|
+
if value is None:
|
427
|
+
return None
|
428
|
+
try:
|
429
|
+
return json.dumps(value, ensure_ascii=False, separators=(",", ":"))
|
430
|
+
except (TypeError, ValueError) as e:
|
431
|
+
raise ValueError(f"Cannot serialize value {value} to JSON: {e}")
|
432
|
+
|
433
|
+
return process
|
434
|
+
|
435
|
+
def bind_expression(self, bindvalue):
|
436
|
+
"""Wrap with PARSE_JSON() in SQL"""
|
437
|
+
return expression.func.PARSE_JSON(bindvalue)
|
438
|
+
|
439
|
+
def literal_processor(self, dialect):
|
440
|
+
"""Process literal values for SQL generation.
|
441
|
+
For VARIANT columns, use PARSE_JSON() to properly insert data.
|
442
|
+
"""
|
443
|
+
|
444
|
+
def process(value):
|
445
|
+
if value is None:
|
446
|
+
return "NULL"
|
447
|
+
try:
|
448
|
+
return self.pe.escape_string(
|
449
|
+
json.dumps(value, ensure_ascii=False, separators=(",", ":"))
|
450
|
+
)
|
451
|
+
except (TypeError, ValueError) as e:
|
452
|
+
raise ValueError(f"Cannot serialize value {value} to JSON: {e}")
|
453
|
+
|
454
|
+
return process
|
455
|
+
|
456
|
+
|
457
|
+
@compiles(DatabricksVariant, "databricks")
|
458
|
+
def compile_variant(type_, compiler, **kw):
|
459
|
+
return "VARIANT"
|
File without changes
|
File without changes
|
File without changes
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/_ddl.py
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/base.py
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/py.typed
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/pytest.ini
RENAMED
File without changes
|
File without changes
|
{databricks_sqlalchemy-2.0.6 → databricks_sqlalchemy-2.0.8}/src/databricks/sqlalchemy/setup.cfg
RENAMED
File without changes
|