databricks-sqlalchemy 2.0.5__tar.gz → 2.0.7__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- databricks_sqlalchemy-2.0.7/CHANGELOG.md +19 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/PKG-INFO +4 -3
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/pyproject.toml +2 -2
- databricks_sqlalchemy-2.0.7/src/databricks/sqlalchemy/__init__.py +10 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_types.py +76 -0
- databricks_sqlalchemy-2.0.5/CHANGELOG.md +0 -10
- databricks_sqlalchemy-2.0.5/src/databricks/sqlalchemy/__init__.py +0 -4
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/LICENSE +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/README.md +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/__init__.py +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_ddl.py +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_parse.py +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/base.py +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/py.typed +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/pytest.ini +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/requirements.py +0 -0
- {databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/setup.cfg +0 -0
@@ -0,0 +1,19 @@
|
|
1
|
+
# Release History
|
2
|
+
|
3
|
+
# 2.0.7 (2025-06-23)
|
4
|
+
|
5
|
+
- Feature: Added support for complex data types such as DatabricksArray and DatabricksMap [Private Preview] (databricks/databricks-sqlalchemy#30 by @jprakash-db)
|
6
|
+
|
7
|
+
|
8
|
+
# 2.0.6 (2025-04-29)
|
9
|
+
|
10
|
+
- Relaxed pin for `pyarrow` (databricks/databricks-sqlalchemy#20 by @dhirschfeld)
|
11
|
+
|
12
|
+
# 2.0.5 (2025-02-22)
|
13
|
+
|
14
|
+
- Added support for double column types (databricks/databricks-sqlalchemy#19 by @up-stevesloan)
|
15
|
+
|
16
|
+
# 2.0.4 (2025-01-27)
|
17
|
+
|
18
|
+
- All the SQLAlchemy features from `databricks-sql-connector>=4.0.0` have been moved to this `databricks-sqlalchemy` library
|
19
|
+
- Support for SQLAlchemy v2 dialect is provided
|
@@ -1,6 +1,6 @@
|
|
1
|
-
Metadata-Version: 2.
|
1
|
+
Metadata-Version: 2.3
|
2
2
|
Name: databricks-sqlalchemy
|
3
|
-
Version: 2.0.
|
3
|
+
Version: 2.0.7
|
4
4
|
Summary: Databricks SQLAlchemy plugin for Python
|
5
5
|
License: Apache-2.0
|
6
6
|
Author: Databricks
|
@@ -13,8 +13,9 @@ Classifier: Programming Language :: Python :: 3.9
|
|
13
13
|
Classifier: Programming Language :: Python :: 3.10
|
14
14
|
Classifier: Programming Language :: Python :: 3.11
|
15
15
|
Classifier: Programming Language :: Python :: 3.12
|
16
|
+
Classifier: Programming Language :: Python :: 3.13
|
16
17
|
Requires-Dist: databricks_sql_connector (>=4.0.0)
|
17
|
-
Requires-Dist: pyarrow (>=14.0.1
|
18
|
+
Requires-Dist: pyarrow (>=14.0.1)
|
18
19
|
Requires-Dist: sqlalchemy (>=2.0.21)
|
19
20
|
Project-URL: Bug Tracker, https://github.com/databricks/databricks-sqlalchemy/issues
|
20
21
|
Project-URL: Homepage, https://github.com/databricks/databricks-sqlalchemy
|
@@ -1,6 +1,6 @@
|
|
1
1
|
[tool.poetry]
|
2
2
|
name = "databricks-sqlalchemy"
|
3
|
-
version = "2.0.
|
3
|
+
version = "2.0.7"
|
4
4
|
description = "Databricks SQLAlchemy plugin for Python"
|
5
5
|
authors = ["Databricks <databricks-sql-connector-maintainers@databricks.com>"]
|
6
6
|
license = "Apache-2.0"
|
@@ -11,7 +11,7 @@ include = ["CHANGELOG.md"]
|
|
11
11
|
[tool.poetry.dependencies]
|
12
12
|
python = "^3.8.0"
|
13
13
|
databricks_sql_connector = { version = ">=4.0.0"}
|
14
|
-
pyarrow = { version = ">=14.0.1
|
14
|
+
pyarrow = { version = ">=14.0.1"}
|
15
15
|
sqlalchemy = { version = ">=2.0.21" }
|
16
16
|
|
17
17
|
[tool.poetry.dev-dependencies]
|
@@ -0,0 +1,10 @@
|
|
1
|
+
from databricks.sqlalchemy.base import DatabricksDialect
|
2
|
+
from databricks.sqlalchemy._types import (
|
3
|
+
TINYINT,
|
4
|
+
TIMESTAMP,
|
5
|
+
TIMESTAMP_NTZ,
|
6
|
+
DatabricksArray,
|
7
|
+
DatabricksMap,
|
8
|
+
)
|
9
|
+
|
10
|
+
__all__ = ["TINYINT", "TIMESTAMP", "TIMESTAMP_NTZ", "DatabricksArray", "DatabricksMap"]
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_types.py
RENAMED
@@ -5,6 +5,7 @@ from typing import Any, Union, Optional
|
|
5
5
|
import sqlalchemy
|
6
6
|
from sqlalchemy.engine.interfaces import Dialect
|
7
7
|
from sqlalchemy.ext.compiler import compiles
|
8
|
+
from sqlalchemy.types import TypeDecorator, UserDefinedType
|
8
9
|
|
9
10
|
from databricks.sql.utils import ParamEscaper
|
10
11
|
|
@@ -26,6 +27,11 @@ def process_literal_param_hack(value: Any):
|
|
26
27
|
return value
|
27
28
|
|
28
29
|
|
30
|
+
def identity_processor(value):
|
31
|
+
"""This method returns the value itself, when no other processor is provided"""
|
32
|
+
return value
|
33
|
+
|
34
|
+
|
29
35
|
@compiles(sqlalchemy.types.Enum, "databricks")
|
30
36
|
@compiles(sqlalchemy.types.String, "databricks")
|
31
37
|
@compiles(sqlalchemy.types.Text, "databricks")
|
@@ -321,3 +327,73 @@ class TINYINT(sqlalchemy.types.TypeDecorator):
|
|
321
327
|
@compiles(TINYINT, "databricks")
|
322
328
|
def compile_tinyint(type_, compiler, **kw):
|
323
329
|
return "TINYINT"
|
330
|
+
|
331
|
+
|
332
|
+
class DatabricksArray(UserDefinedType):
|
333
|
+
"""
|
334
|
+
A custom array type that can wrap any other SQLAlchemy type.
|
335
|
+
|
336
|
+
Examples:
|
337
|
+
DatabricksArray(String) -> ARRAY<STRING>
|
338
|
+
DatabricksArray(Integer) -> ARRAY<INT>
|
339
|
+
DatabricksArray(CustomType) -> ARRAY<CUSTOM_TYPE>
|
340
|
+
"""
|
341
|
+
|
342
|
+
def __init__(self, item_type):
|
343
|
+
self.item_type = item_type() if isinstance(item_type, type) else item_type
|
344
|
+
|
345
|
+
def bind_processor(self, dialect):
|
346
|
+
item_processor = self.item_type.bind_processor(dialect)
|
347
|
+
if item_processor is None:
|
348
|
+
item_processor = identity_processor
|
349
|
+
|
350
|
+
def process(value):
|
351
|
+
return [item_processor(val) for val in value]
|
352
|
+
|
353
|
+
return process
|
354
|
+
|
355
|
+
|
356
|
+
@compiles(DatabricksArray, "databricks")
|
357
|
+
def compile_databricks_array(type_, compiler, **kw):
|
358
|
+
inner = compiler.process(type_.item_type, **kw)
|
359
|
+
|
360
|
+
return f"ARRAY<{inner}>"
|
361
|
+
|
362
|
+
|
363
|
+
class DatabricksMap(UserDefinedType):
|
364
|
+
"""
|
365
|
+
A custom map type that can wrap any other SQLAlchemy types for both key and value.
|
366
|
+
|
367
|
+
Examples:
|
368
|
+
DatabricksMap(String, String) -> MAP<STRING,STRING>
|
369
|
+
DatabricksMap(Integer, String) -> MAP<INT,STRING>
|
370
|
+
DatabricksMap(String, DatabricksArray(Integer)) -> MAP<STRING,ARRAY<INT>>
|
371
|
+
"""
|
372
|
+
|
373
|
+
def __init__(self, key_type, value_type):
|
374
|
+
self.key_type = key_type() if isinstance(key_type, type) else key_type
|
375
|
+
self.value_type = value_type() if isinstance(value_type, type) else value_type
|
376
|
+
|
377
|
+
def bind_processor(self, dialect):
|
378
|
+
key_processor = self.key_type.bind_processor(dialect)
|
379
|
+
value_processor = self.value_type.bind_processor(dialect)
|
380
|
+
|
381
|
+
if key_processor is None:
|
382
|
+
key_processor = identity_processor
|
383
|
+
if value_processor is None:
|
384
|
+
value_processor = identity_processor
|
385
|
+
|
386
|
+
def process(value):
|
387
|
+
return {
|
388
|
+
key_processor(key): value_processor(value)
|
389
|
+
for key, value in value.items()
|
390
|
+
}
|
391
|
+
|
392
|
+
return process
|
393
|
+
|
394
|
+
|
395
|
+
@compiles(DatabricksMap, "databricks")
|
396
|
+
def compile_databricks_map(type_, compiler, **kw):
|
397
|
+
key_type = compiler.process(type_.key_type, **kw)
|
398
|
+
value_type = compiler.process(type_.value_type, **kw)
|
399
|
+
return f"MAP<{key_type},{value_type}>"
|
@@ -1,10 +0,0 @@
|
|
1
|
-
# Release History
|
2
|
-
|
3
|
-
# 2.0.5 (2025-02-22)
|
4
|
-
|
5
|
-
- Added support for double column types (databricks/databricks-sqlalchemy#19 by @up-stevesloan)
|
6
|
-
|
7
|
-
# 2.0.4 (2025-01-27)
|
8
|
-
|
9
|
-
- All the SQLAlchemy features from `databricks-sql-connector>=4.0.0` have been moved to this `databricks-sqlalchemy` library
|
10
|
-
- Support for SQLAlchemy v2 dialect is provided
|
File without changes
|
File without changes
|
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_ddl.py
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/_parse.py
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/base.py
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/py.typed
RENAMED
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/pytest.ini
RENAMED
File without changes
|
File without changes
|
{databricks_sqlalchemy-2.0.5 → databricks_sqlalchemy-2.0.7}/src/databricks/sqlalchemy/setup.cfg
RENAMED
File without changes
|