duckdb 1.4.1.dev125__cp39-cp39-win_amd64.whl → 1.5.0.dev44__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of duckdb might be problematic. Click here for more details.
- _duckdb.cp39-win_amd64.pyd +0 -0
- duckdb/__init__.py +374 -373
- duckdb/__init__.pyi +180 -604
- duckdb/bytes_io_wrapper.py +7 -6
- duckdb/experimental/__init__.py +1 -2
- duckdb/experimental/spark/__init__.py +4 -3
- duckdb/experimental/spark/_globals.py +8 -8
- duckdb/experimental/spark/_typing.py +9 -7
- duckdb/experimental/spark/conf.py +15 -16
- duckdb/experimental/spark/context.py +44 -60
- duckdb/experimental/spark/errors/__init__.py +35 -33
- duckdb/experimental/spark/errors/error_classes.py +1 -1
- duckdb/experimental/spark/errors/exceptions/__init__.py +1 -1
- duckdb/experimental/spark/errors/exceptions/base.py +88 -39
- duckdb/experimental/spark/errors/utils.py +16 -11
- duckdb/experimental/spark/exception.py +6 -9
- duckdb/experimental/spark/sql/__init__.py +5 -5
- duckdb/experimental/spark/sql/_typing.py +15 -8
- duckdb/experimental/spark/sql/catalog.py +20 -21
- duckdb/experimental/spark/sql/column.py +54 -47
- duckdb/experimental/spark/sql/conf.py +8 -9
- duckdb/experimental/spark/sql/dataframe.py +233 -185
- duckdb/experimental/spark/sql/functions.py +1248 -1222
- duckdb/experimental/spark/sql/group.py +52 -56
- duckdb/experimental/spark/sql/readwriter.py +94 -80
- duckdb/experimental/spark/sql/session.py +59 -64
- duckdb/experimental/spark/sql/streaming.py +10 -9
- duckdb/experimental/spark/sql/type_utils.py +64 -66
- duckdb/experimental/spark/sql/types.py +344 -308
- duckdb/experimental/spark/sql/udf.py +6 -6
- duckdb/filesystem.py +8 -13
- duckdb/functional/__init__.py +16 -2
- duckdb/polars_io.py +57 -66
- duckdb/query_graph/__main__.py +96 -91
- duckdb/typing/__init__.py +8 -8
- duckdb/typing/__init__.pyi +2 -4
- duckdb/udf.py +5 -10
- duckdb/value/__init__.py +0 -1
- duckdb/value/constant/__init__.py +59 -61
- duckdb/value/constant/__init__.pyi +4 -3
- duckdb-1.5.0.dev44.dist-info/METADATA +80 -0
- duckdb-1.5.0.dev44.dist-info/RECORD +47 -0
- adbc_driver_duckdb/__init__.py +0 -50
- adbc_driver_duckdb/dbapi.py +0 -115
- duckdb-1.4.1.dev125.dist-info/METADATA +0 -326
- duckdb-1.4.1.dev125.dist-info/RECORD +0 -49
- {duckdb-1.4.1.dev125.dist-info → duckdb-1.5.0.dev44.dist-info}/WHEEL +0 -0
- {duckdb-1.4.1.dev125.dist-info → duckdb-1.5.0.dev44.dist-info}/licenses/LICENSE +0 -0
duckdb/__init__.pyi
CHANGED
|
@@ -41,7 +41,6 @@ from duckdb.value.constant import (
|
|
|
41
41
|
|
|
42
42
|
# We also run this in python3.7, where this is needed
|
|
43
43
|
from typing_extensions import Literal
|
|
44
|
-
|
|
45
44
|
# stubgen override - missing import of Set
|
|
46
45
|
from typing import Any, ClassVar, Set, Optional, Callable
|
|
47
46
|
from io import StringIO, TextIOBase
|
|
@@ -49,13 +48,11 @@ from pathlib import Path
|
|
|
49
48
|
|
|
50
49
|
from typing import overload, Dict, List, Union, Tuple
|
|
51
50
|
import pandas
|
|
52
|
-
|
|
53
51
|
# stubgen override - unfortunately we need this for version checks
|
|
54
52
|
import sys
|
|
55
53
|
import fsspec
|
|
56
54
|
import pyarrow.lib
|
|
57
55
|
import polars
|
|
58
|
-
|
|
59
56
|
# stubgen override - This should probably not be exposed
|
|
60
57
|
apilevel: str
|
|
61
58
|
comment: token_type
|
|
@@ -81,10 +78,15 @@ __jupyter__: bool
|
|
|
81
78
|
__formatted_python_version__: str
|
|
82
79
|
|
|
83
80
|
class BinderException(ProgrammingError): ...
|
|
81
|
+
|
|
84
82
|
class CatalogException(ProgrammingError): ...
|
|
83
|
+
|
|
85
84
|
class ConnectionException(OperationalError): ...
|
|
85
|
+
|
|
86
86
|
class ConstraintException(IntegrityError): ...
|
|
87
|
+
|
|
87
88
|
class ConversionException(DataError): ...
|
|
89
|
+
|
|
88
90
|
class DataError(Error): ...
|
|
89
91
|
|
|
90
92
|
class ExplainType:
|
|
@@ -202,37 +204,46 @@ class Statement:
|
|
|
202
204
|
class Expression:
|
|
203
205
|
def __init__(self, *args, **kwargs) -> None: ...
|
|
204
206
|
def __neg__(self) -> "Expression": ...
|
|
207
|
+
|
|
205
208
|
def __add__(self, expr: "Expression") -> "Expression": ...
|
|
206
209
|
def __radd__(self, expr: "Expression") -> "Expression": ...
|
|
210
|
+
|
|
207
211
|
def __sub__(self, expr: "Expression") -> "Expression": ...
|
|
208
212
|
def __rsub__(self, expr: "Expression") -> "Expression": ...
|
|
213
|
+
|
|
209
214
|
def __mul__(self, expr: "Expression") -> "Expression": ...
|
|
210
215
|
def __rmul__(self, expr: "Expression") -> "Expression": ...
|
|
216
|
+
|
|
211
217
|
def __div__(self, expr: "Expression") -> "Expression": ...
|
|
212
218
|
def __rdiv__(self, expr: "Expression") -> "Expression": ...
|
|
219
|
+
|
|
213
220
|
def __truediv__(self, expr: "Expression") -> "Expression": ...
|
|
214
221
|
def __rtruediv__(self, expr: "Expression") -> "Expression": ...
|
|
222
|
+
|
|
215
223
|
def __floordiv__(self, expr: "Expression") -> "Expression": ...
|
|
216
224
|
def __rfloordiv__(self, expr: "Expression") -> "Expression": ...
|
|
225
|
+
|
|
217
226
|
def __mod__(self, expr: "Expression") -> "Expression": ...
|
|
218
227
|
def __rmod__(self, expr: "Expression") -> "Expression": ...
|
|
228
|
+
|
|
219
229
|
def __pow__(self, expr: "Expression") -> "Expression": ...
|
|
220
230
|
def __rpow__(self, expr: "Expression") -> "Expression": ...
|
|
231
|
+
|
|
221
232
|
def __and__(self, expr: "Expression") -> "Expression": ...
|
|
222
233
|
def __rand__(self, expr: "Expression") -> "Expression": ...
|
|
223
234
|
def __or__(self, expr: "Expression") -> "Expression": ...
|
|
224
235
|
def __ror__(self, expr: "Expression") -> "Expression": ...
|
|
225
236
|
def __invert__(self) -> "Expression": ...
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
def __ne__(
|
|
230
|
-
self, expr: "Expression"
|
|
231
|
-
) -> "Expression": ...
|
|
237
|
+
|
|
238
|
+
def __eq__(# type: ignore[override]
|
|
239
|
+
self, expr: "Expression") -> "Expression": ...
|
|
240
|
+
def __ne__(# type: ignore[override]
|
|
241
|
+
self, expr: "Expression") -> "Expression": ...
|
|
232
242
|
def __gt__(self, expr: "Expression") -> "Expression": ...
|
|
233
243
|
def __ge__(self, expr: "Expression") -> "Expression": ...
|
|
234
244
|
def __lt__(self, expr: "Expression") -> "Expression": ...
|
|
235
245
|
def __le__(self, expr: "Expression") -> "Expression": ...
|
|
246
|
+
|
|
236
247
|
def show(self) -> None: ...
|
|
237
248
|
def __repr__(self) -> str: ...
|
|
238
249
|
def get_name(self) -> str: ...
|
|
@@ -280,18 +291,7 @@ class DuckDBPyConnection:
|
|
|
280
291
|
def unregister_filesystem(self, name: str) -> None: ...
|
|
281
292
|
def list_filesystems(self) -> list: ...
|
|
282
293
|
def filesystem_is_registered(self, name: str) -> bool: ...
|
|
283
|
-
def create_function(
|
|
284
|
-
self,
|
|
285
|
-
name: str,
|
|
286
|
-
function: function,
|
|
287
|
-
parameters: Optional[List[DuckDBPyType]] = None,
|
|
288
|
-
return_type: Optional[DuckDBPyType] = None,
|
|
289
|
-
*,
|
|
290
|
-
type: Optional[PythonUDFType] = PythonUDFType.NATIVE,
|
|
291
|
-
null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT,
|
|
292
|
-
exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT,
|
|
293
|
-
side_effects: bool = False,
|
|
294
|
-
) -> DuckDBPyConnection: ...
|
|
294
|
+
def create_function(self, name: str, function: function, parameters: Optional[List[DuckDBPyType]] = None, return_type: Optional[DuckDBPyType] = None, *, type: Optional[PythonUDFType] = PythonUDFType.NATIVE, null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT, exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT, side_effects: bool = False) -> DuckDBPyConnection: ...
|
|
295
295
|
def remove_function(self, name: str) -> DuckDBPyConnection: ...
|
|
296
296
|
def sqltype(self, type_str: str) -> DuckDBPyType: ...
|
|
297
297
|
def dtype(self, type_str: str) -> DuckDBPyType: ...
|
|
@@ -334,152 +334,21 @@ class DuckDBPyConnection:
|
|
|
334
334
|
def unregister(self, view_name: str) -> DuckDBPyConnection: ...
|
|
335
335
|
def table(self, table_name: str) -> DuckDBPyRelation: ...
|
|
336
336
|
def view(self, view_name: str) -> DuckDBPyRelation: ...
|
|
337
|
-
def values(self, *args: Union[List[Any],
|
|
337
|
+
def values(self, *args: Union[List[Any],Expression, Tuple[Expression]]) -> DuckDBPyRelation: ...
|
|
338
338
|
def table_function(self, name: str, parameters: object = None) -> DuckDBPyRelation: ...
|
|
339
|
-
def read_json(
|
|
340
|
-
self,
|
|
341
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
342
|
-
*,
|
|
343
|
-
columns: Optional[Dict[str, str]] = None,
|
|
344
|
-
sample_size: Optional[int] = None,
|
|
345
|
-
maximum_depth: Optional[int] = None,
|
|
346
|
-
records: Optional[str] = None,
|
|
347
|
-
format: Optional[str] = None,
|
|
348
|
-
date_format: Optional[str] = None,
|
|
349
|
-
timestamp_format: Optional[str] = None,
|
|
350
|
-
compression: Optional[str] = None,
|
|
351
|
-
maximum_object_size: Optional[int] = None,
|
|
352
|
-
ignore_errors: Optional[bool] = None,
|
|
353
|
-
convert_strings_to_integers: Optional[bool] = None,
|
|
354
|
-
field_appearance_threshold: Optional[float] = None,
|
|
355
|
-
map_inference_threshold: Optional[int] = None,
|
|
356
|
-
maximum_sample_files: Optional[int] = None,
|
|
357
|
-
filename: Optional[Union[bool, str]] = None,
|
|
358
|
-
hive_partitioning: Optional[bool] = None,
|
|
359
|
-
union_by_name: Optional[bool] = None,
|
|
360
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
361
|
-
hive_types_autocast: Optional[bool] = None,
|
|
362
|
-
) -> DuckDBPyRelation: ...
|
|
339
|
+
def read_json(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, columns: Optional[Dict[str,str]] = None, sample_size: Optional[int] = None, maximum_depth: Optional[int] = None, records: Optional[str] = None, format: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, compression: Optional[str] = None, maximum_object_size: Optional[int] = None, ignore_errors: Optional[bool] = None, convert_strings_to_integers: Optional[bool] = None, field_appearance_threshold: Optional[float] = None, map_inference_threshold: Optional[int] = None, maximum_sample_files: Optional[int] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
|
|
363
340
|
def extract_statements(self, query: str) -> List[Statement]: ...
|
|
364
341
|
def sql(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
|
|
365
342
|
def query(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
|
|
366
343
|
def from_query(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
|
|
367
|
-
def read_csv(
|
|
368
|
-
|
|
369
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
370
|
-
*,
|
|
371
|
-
header: Optional[Union[bool, int]] = None,
|
|
372
|
-
compression: Optional[str] = None,
|
|
373
|
-
sep: Optional[str] = None,
|
|
374
|
-
delimiter: Optional[str] = None,
|
|
375
|
-
dtype: Optional[Union[Dict[str, str], List[str]]] = None,
|
|
376
|
-
na_values: Optional[Union[str, List[str]]] = None,
|
|
377
|
-
skiprows: Optional[int] = None,
|
|
378
|
-
quotechar: Optional[str] = None,
|
|
379
|
-
escapechar: Optional[str] = None,
|
|
380
|
-
encoding: Optional[str] = None,
|
|
381
|
-
parallel: Optional[bool] = None,
|
|
382
|
-
date_format: Optional[str] = None,
|
|
383
|
-
timestamp_format: Optional[str] = None,
|
|
384
|
-
sample_size: Optional[int] = None,
|
|
385
|
-
all_varchar: Optional[bool] = None,
|
|
386
|
-
normalize_names: Optional[bool] = None,
|
|
387
|
-
null_padding: Optional[bool] = None,
|
|
388
|
-
names: Optional[List[str]] = None,
|
|
389
|
-
lineterminator: Optional[str] = None,
|
|
390
|
-
columns: Optional[Dict[str, str]] = None,
|
|
391
|
-
auto_type_candidates: Optional[List[str]] = None,
|
|
392
|
-
max_line_size: Optional[int] = None,
|
|
393
|
-
ignore_errors: Optional[bool] = None,
|
|
394
|
-
store_rejects: Optional[bool] = None,
|
|
395
|
-
rejects_table: Optional[str] = None,
|
|
396
|
-
rejects_scan: Optional[str] = None,
|
|
397
|
-
rejects_limit: Optional[int] = None,
|
|
398
|
-
force_not_null: Optional[List[str]] = None,
|
|
399
|
-
buffer_size: Optional[int] = None,
|
|
400
|
-
decimal: Optional[str] = None,
|
|
401
|
-
allow_quoted_nulls: Optional[bool] = None,
|
|
402
|
-
filename: Optional[Union[bool, str]] = None,
|
|
403
|
-
hive_partitioning: Optional[bool] = None,
|
|
404
|
-
union_by_name: Optional[bool] = None,
|
|
405
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
406
|
-
hive_types_autocast: Optional[bool] = None,
|
|
407
|
-
) -> DuckDBPyRelation: ...
|
|
408
|
-
def from_csv_auto(
|
|
409
|
-
self,
|
|
410
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
411
|
-
*,
|
|
412
|
-
header: Optional[Union[bool, int]] = None,
|
|
413
|
-
compression: Optional[str] = None,
|
|
414
|
-
sep: Optional[str] = None,
|
|
415
|
-
delimiter: Optional[str] = None,
|
|
416
|
-
dtype: Optional[Union[Dict[str, str], List[str]]] = None,
|
|
417
|
-
na_values: Optional[Union[str, List[str]]] = None,
|
|
418
|
-
skiprows: Optional[int] = None,
|
|
419
|
-
quotechar: Optional[str] = None,
|
|
420
|
-
escapechar: Optional[str] = None,
|
|
421
|
-
encoding: Optional[str] = None,
|
|
422
|
-
parallel: Optional[bool] = None,
|
|
423
|
-
date_format: Optional[str] = None,
|
|
424
|
-
timestamp_format: Optional[str] = None,
|
|
425
|
-
sample_size: Optional[int] = None,
|
|
426
|
-
all_varchar: Optional[bool] = None,
|
|
427
|
-
normalize_names: Optional[bool] = None,
|
|
428
|
-
null_padding: Optional[bool] = None,
|
|
429
|
-
names: Optional[List[str]] = None,
|
|
430
|
-
lineterminator: Optional[str] = None,
|
|
431
|
-
columns: Optional[Dict[str, str]] = None,
|
|
432
|
-
auto_type_candidates: Optional[List[str]] = None,
|
|
433
|
-
max_line_size: Optional[int] = None,
|
|
434
|
-
ignore_errors: Optional[bool] = None,
|
|
435
|
-
store_rejects: Optional[bool] = None,
|
|
436
|
-
rejects_table: Optional[str] = None,
|
|
437
|
-
rejects_scan: Optional[str] = None,
|
|
438
|
-
rejects_limit: Optional[int] = None,
|
|
439
|
-
force_not_null: Optional[List[str]] = None,
|
|
440
|
-
buffer_size: Optional[int] = None,
|
|
441
|
-
decimal: Optional[str] = None,
|
|
442
|
-
allow_quoted_nulls: Optional[bool] = None,
|
|
443
|
-
filename: Optional[Union[bool, str]] = None,
|
|
444
|
-
hive_partitioning: Optional[bool] = None,
|
|
445
|
-
union_by_name: Optional[bool] = None,
|
|
446
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
447
|
-
hive_types_autocast: Optional[bool] = None,
|
|
448
|
-
) -> DuckDBPyRelation: ...
|
|
344
|
+
def read_csv(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
|
|
345
|
+
def from_csv_auto(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
|
|
449
346
|
def from_df(self, df: pandas.DataFrame) -> DuckDBPyRelation: ...
|
|
450
347
|
def from_arrow(self, arrow_object: object) -> DuckDBPyRelation: ...
|
|
451
|
-
def from_parquet(
|
|
452
|
-
|
|
453
|
-
file_glob: str,
|
|
454
|
-
binary_as_string: bool = False,
|
|
455
|
-
*,
|
|
456
|
-
file_row_number: bool = False,
|
|
457
|
-
filename: bool = False,
|
|
458
|
-
hive_partitioning: bool = False,
|
|
459
|
-
union_by_name: bool = False,
|
|
460
|
-
compression: Optional[str] = None,
|
|
461
|
-
) -> DuckDBPyRelation: ...
|
|
462
|
-
def read_parquet(
|
|
463
|
-
self,
|
|
464
|
-
file_glob: str,
|
|
465
|
-
binary_as_string: bool = False,
|
|
466
|
-
*,
|
|
467
|
-
file_row_number: bool = False,
|
|
468
|
-
filename: bool = False,
|
|
469
|
-
hive_partitioning: bool = False,
|
|
470
|
-
union_by_name: bool = False,
|
|
471
|
-
compression: Optional[str] = None,
|
|
472
|
-
) -> DuckDBPyRelation: ...
|
|
348
|
+
def from_parquet(self, file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None) -> DuckDBPyRelation: ...
|
|
349
|
+
def read_parquet(self, file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None) -> DuckDBPyRelation: ...
|
|
473
350
|
def get_table_names(self, query: str, *, qualified: bool = False) -> Set[str]: ...
|
|
474
|
-
def install_extension(
|
|
475
|
-
self,
|
|
476
|
-
extension: str,
|
|
477
|
-
*,
|
|
478
|
-
force_install: bool = False,
|
|
479
|
-
repository: Optional[str] = None,
|
|
480
|
-
repository_url: Optional[str] = None,
|
|
481
|
-
version: Optional[str] = None,
|
|
482
|
-
) -> None: ...
|
|
351
|
+
def install_extension(self, extension: str, *, force_install: bool = False, repository: Optional[str] = None, repository_url: Optional[str] = None, version: Optional[str] = None) -> None: ...
|
|
483
352
|
def load_extension(self, extension: str) -> None: ...
|
|
484
353
|
# END OF CONNECTION METHODS
|
|
485
354
|
|
|
@@ -490,41 +359,19 @@ class DuckDBPyRelation:
|
|
|
490
359
|
def __init__(self, *args, **kwargs) -> None: ...
|
|
491
360
|
def __contains__(self, name: str) -> bool: ...
|
|
492
361
|
def aggregate(self, aggr_expr: str, group_expr: str = ...) -> DuckDBPyRelation: ...
|
|
493
|
-
def apply(
|
|
494
|
-
|
|
495
|
-
function_name: str,
|
|
496
|
-
function_aggr: str,
|
|
497
|
-
group_expr: str = ...,
|
|
498
|
-
function_parameter: str = ...,
|
|
499
|
-
projected_columns: str = ...,
|
|
500
|
-
) -> DuckDBPyRelation: ...
|
|
362
|
+
def apply(self, function_name: str, function_aggr: str, group_expr: str = ..., function_parameter: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
363
|
+
|
|
501
364
|
def cume_dist(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
502
365
|
def dense_rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
503
366
|
def percent_rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
504
367
|
def rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
505
368
|
def rank_dense(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
506
369
|
def row_number(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
507
|
-
|
|
508
|
-
|
|
509
|
-
|
|
510
|
-
|
|
511
|
-
|
|
512
|
-
default_value: str,
|
|
513
|
-
ignore_nulls: bool,
|
|
514
|
-
projected_columns: str = ...,
|
|
515
|
-
) -> DuckDBPyRelation: ...
|
|
516
|
-
def lead(
|
|
517
|
-
self,
|
|
518
|
-
column: str,
|
|
519
|
-
window_spec: str,
|
|
520
|
-
offset: int,
|
|
521
|
-
default_value: str,
|
|
522
|
-
ignore_nulls: bool,
|
|
523
|
-
projected_columns: str = ...,
|
|
524
|
-
) -> DuckDBPyRelation: ...
|
|
525
|
-
def nth_value(
|
|
526
|
-
self, column: str, window_spec: str, offset: int, ignore_nulls: bool = ..., projected_columns: str = ...
|
|
527
|
-
) -> DuckDBPyRelation: ...
|
|
370
|
+
|
|
371
|
+
def lag(self, column: str, window_spec: str, offset: int, default_value: str, ignore_nulls: bool, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
372
|
+
def lead(self, column: str, window_spec: str, offset: int, default_value: str, ignore_nulls: bool, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
373
|
+
def nth_value(self, column: str, window_spec: str, offset: int, ignore_nulls: bool = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
374
|
+
|
|
528
375
|
def value_counts(self, column: str, groups: str = ...) -> DuckDBPyRelation: ...
|
|
529
376
|
def geomean(self, column: str, groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
530
377
|
def first(self, column: str, groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
@@ -533,119 +380,41 @@ class DuckDBPyRelation:
|
|
|
533
380
|
def last_value(self, column: str, window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
534
381
|
def mode(self, aggregation_columns: str, group_columns: str = ...) -> DuckDBPyRelation: ...
|
|
535
382
|
def n_tile(self, window_spec: str, num_buckets: int, projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
536
|
-
def quantile_cont(
|
|
537
|
-
|
|
538
|
-
) -> DuckDBPyRelation: ...
|
|
539
|
-
def quantile_disc(
|
|
540
|
-
self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...
|
|
541
|
-
) -> DuckDBPyRelation: ...
|
|
383
|
+
def quantile_cont(self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
384
|
+
def quantile_disc(self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
542
385
|
def sum(self, sum_aggr: str, group_expr: str = ...) -> DuckDBPyRelation: ...
|
|
543
|
-
|
|
544
|
-
|
|
545
|
-
) -> DuckDBPyRelation: ...
|
|
546
|
-
def
|
|
547
|
-
|
|
548
|
-
|
|
549
|
-
|
|
550
|
-
|
|
551
|
-
|
|
552
|
-
|
|
553
|
-
) -> DuckDBPyRelation: ...
|
|
554
|
-
def
|
|
555
|
-
|
|
556
|
-
|
|
557
|
-
|
|
558
|
-
|
|
559
|
-
|
|
560
|
-
|
|
561
|
-
) -> DuckDBPyRelation: ...
|
|
562
|
-
def
|
|
563
|
-
|
|
564
|
-
) -> DuckDBPyRelation: ...
|
|
565
|
-
def
|
|
566
|
-
|
|
567
|
-
) -> DuckDBPyRelation: ...
|
|
568
|
-
def
|
|
569
|
-
|
|
570
|
-
) -> DuckDBPyRelation: ...
|
|
571
|
-
def
|
|
572
|
-
|
|
573
|
-
) -> DuckDBPyRelation: ...
|
|
574
|
-
|
|
575
|
-
self,
|
|
576
|
-
column: str,
|
|
577
|
-
min: Optional[int],
|
|
578
|
-
max: Optional[int],
|
|
579
|
-
groups: str = ...,
|
|
580
|
-
window_spec: str = ...,
|
|
581
|
-
projected_columns: str = ...,
|
|
582
|
-
) -> DuckDBPyRelation: ...
|
|
583
|
-
def bool_and(
|
|
584
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
585
|
-
) -> DuckDBPyRelation: ...
|
|
586
|
-
def bool_or(
|
|
587
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
588
|
-
) -> DuckDBPyRelation: ...
|
|
589
|
-
def count(
|
|
590
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
591
|
-
) -> DuckDBPyRelation: ...
|
|
592
|
-
def favg(
|
|
593
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
594
|
-
) -> DuckDBPyRelation: ...
|
|
595
|
-
def fsum(
|
|
596
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
597
|
-
) -> DuckDBPyRelation: ...
|
|
598
|
-
def histogram(
|
|
599
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
600
|
-
) -> DuckDBPyRelation: ...
|
|
601
|
-
def max(
|
|
602
|
-
self, max_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
603
|
-
) -> DuckDBPyRelation: ...
|
|
604
|
-
def min(
|
|
605
|
-
self, min_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
606
|
-
) -> DuckDBPyRelation: ...
|
|
607
|
-
def mean(
|
|
608
|
-
self, mean_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
609
|
-
) -> DuckDBPyRelation: ...
|
|
610
|
-
def median(
|
|
611
|
-
self, median_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
612
|
-
) -> DuckDBPyRelation: ...
|
|
613
|
-
def product(
|
|
614
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
615
|
-
) -> DuckDBPyRelation: ...
|
|
616
|
-
def quantile(
|
|
617
|
-
self, q: str, quantile_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
618
|
-
) -> DuckDBPyRelation: ...
|
|
619
|
-
def std(
|
|
620
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
621
|
-
) -> DuckDBPyRelation: ...
|
|
622
|
-
def stddev(
|
|
623
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
624
|
-
) -> DuckDBPyRelation: ...
|
|
625
|
-
def stddev_pop(
|
|
626
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
627
|
-
) -> DuckDBPyRelation: ...
|
|
628
|
-
def stddev_samp(
|
|
629
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
630
|
-
) -> DuckDBPyRelation: ...
|
|
631
|
-
def string_agg(
|
|
632
|
-
self, column: str, sep: str = ..., groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
633
|
-
) -> DuckDBPyRelation: ...
|
|
634
|
-
def var(
|
|
635
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
636
|
-
) -> DuckDBPyRelation: ...
|
|
637
|
-
def var_pop(
|
|
638
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
639
|
-
) -> DuckDBPyRelation: ...
|
|
640
|
-
def var_samp(
|
|
641
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
642
|
-
) -> DuckDBPyRelation: ...
|
|
643
|
-
def variance(
|
|
644
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
645
|
-
) -> DuckDBPyRelation: ...
|
|
646
|
-
def list(
|
|
647
|
-
self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
|
|
648
|
-
) -> DuckDBPyRelation: ...
|
|
386
|
+
|
|
387
|
+
def any_value(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
388
|
+
def arg_max(self, arg_column: str, value_column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
389
|
+
def arg_min(self, arg_column: str, value_column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
390
|
+
def avg(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
391
|
+
def bit_and(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
392
|
+
def bit_or(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
393
|
+
def bit_xor(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
394
|
+
def bitstring_agg(self, column: str, min: Optional[int], max: Optional[int], groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
395
|
+
def bool_and(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
396
|
+
def bool_or(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
397
|
+
def count(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
398
|
+
def favg(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
399
|
+
def fsum(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
400
|
+
def histogram(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
401
|
+
def max(self, max_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
402
|
+
def min(self, min_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
403
|
+
def mean(self, mean_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
404
|
+
def median(self, median_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
405
|
+
def product(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
406
|
+
def quantile(self, q: str, quantile_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
407
|
+
def std(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
408
|
+
def stddev(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
409
|
+
def stddev_pop(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
410
|
+
def stddev_samp(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
411
|
+
def string_agg(self, column: str, sep: str = ..., groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
412
|
+
def var(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
413
|
+
def var_pop(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
414
|
+
def var_samp(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
415
|
+
def variance(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
416
|
+
def list(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
|
|
417
|
+
|
|
649
418
|
def arrow(self, batch_size: int = ...) -> pyarrow.lib.RecordBatchReader: ...
|
|
650
419
|
def __arrow_c_stream__(self, requested_schema: Optional[object] = None) -> object: ...
|
|
651
420
|
def create(self, table_name: str) -> None: ...
|
|
@@ -655,7 +424,7 @@ class DuckDBPyRelation:
|
|
|
655
424
|
def distinct(self) -> DuckDBPyRelation: ...
|
|
656
425
|
def except_(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
|
|
657
426
|
def execute(self, *args, **kwargs) -> DuckDBPyRelation: ...
|
|
658
|
-
def explain(self, type: Optional[
|
|
427
|
+
def explain(self, type: Optional[Literal['standard', 'analyze'] | int] = 'standard') -> str: ...
|
|
659
428
|
def fetchall(self) -> List[Any]: ...
|
|
660
429
|
def fetchmany(self, size: int = ...) -> List[Any]: ...
|
|
661
430
|
def fetchnumpy(self) -> dict: ...
|
|
@@ -668,9 +437,7 @@ class DuckDBPyRelation:
|
|
|
668
437
|
def update(self, set: Dict[str, Expression], condition: Optional[Expression] = None) -> None: ...
|
|
669
438
|
def insert_into(self, table_name: str) -> None: ...
|
|
670
439
|
def intersect(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
|
|
671
|
-
def join(
|
|
672
|
-
self, other_rel: DuckDBPyRelation, condition: Union[str, Expression], how: str = ...
|
|
673
|
-
) -> DuckDBPyRelation: ...
|
|
440
|
+
def join(self, other_rel: DuckDBPyRelation, condition: Union[str, Expression], how: str = ...) -> DuckDBPyRelation: ...
|
|
674
441
|
def cross(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
|
|
675
442
|
def limit(self, n: int, offset: int = ...) -> DuckDBPyRelation: ...
|
|
676
443
|
def map(self, map_function: function, schema: Optional[Dict[str, DuckDBPyType]] = None) -> DuckDBPyRelation: ...
|
|
@@ -681,55 +448,46 @@ class DuckDBPyRelation:
|
|
|
681
448
|
def pl(self, rows_per_batch: int = ..., connection: DuckDBPyConnection = ...) -> polars.DataFrame: ...
|
|
682
449
|
def query(self, virtual_table_name: str, sql_query: str) -> DuckDBPyRelation: ...
|
|
683
450
|
def record_batch(self, batch_size: int = ...) -> pyarrow.lib.RecordBatchReader: ...
|
|
684
|
-
def fetch_record_batch(
|
|
685
|
-
self, rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...
|
|
686
|
-
) -> pyarrow.lib.RecordBatchReader: ...
|
|
451
|
+
def fetch_record_batch(self, rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
|
|
687
452
|
def select_types(self, types: List[Union[str, DuckDBPyType]]) -> DuckDBPyRelation: ...
|
|
688
453
|
def select_dtypes(self, types: List[Union[str, DuckDBPyType]]) -> DuckDBPyRelation: ...
|
|
689
454
|
def set_alias(self, alias: str) -> DuckDBPyRelation: ...
|
|
690
|
-
def show(
|
|
691
|
-
self,
|
|
692
|
-
max_width: Optional[int] = None,
|
|
693
|
-
max_rows: Optional[int] = None,
|
|
694
|
-
max_col_width: Optional[int] = None,
|
|
695
|
-
null_value: Optional[str] = None,
|
|
696
|
-
render_mode: Optional[RenderMode] = None,
|
|
697
|
-
) -> None: ...
|
|
455
|
+
def show(self, max_width: Optional[int] = None, max_rows: Optional[int] = None, max_col_width: Optional[int] = None, null_value: Optional[str] = None, render_mode: Optional[RenderMode] = None) -> None: ...
|
|
698
456
|
def sql_query(self) -> str: ...
|
|
699
457
|
def to_arrow_table(self, batch_size: int = ...) -> pyarrow.lib.Table: ...
|
|
700
458
|
def to_csv(
|
|
701
|
-
|
|
702
|
-
|
|
703
|
-
|
|
704
|
-
|
|
705
|
-
|
|
706
|
-
|
|
707
|
-
|
|
708
|
-
|
|
709
|
-
|
|
710
|
-
|
|
711
|
-
|
|
712
|
-
|
|
713
|
-
|
|
714
|
-
|
|
715
|
-
|
|
716
|
-
|
|
717
|
-
|
|
459
|
+
self,
|
|
460
|
+
file_name: str,
|
|
461
|
+
sep: Optional[str] = None,
|
|
462
|
+
na_rep: Optional[str] = None,
|
|
463
|
+
header: Optional[bool] = None,
|
|
464
|
+
quotechar: Optional[str] = None,
|
|
465
|
+
escapechar: Optional[str] = None,
|
|
466
|
+
date_format: Optional[str] = None,
|
|
467
|
+
timestamp_format: Optional[str] = None,
|
|
468
|
+
quoting: Optional[str | int] = None,
|
|
469
|
+
encoding: Optional[str] = None,
|
|
470
|
+
compression: Optional[str] = None,
|
|
471
|
+
write_partition_columns: Optional[bool] = None,
|
|
472
|
+
overwrite: Optional[bool] = None,
|
|
473
|
+
per_thread_output: Optional[bool] = None,
|
|
474
|
+
use_tmp_file: Optional[bool] = None,
|
|
475
|
+
partition_by: Optional[List[str]] = None
|
|
718
476
|
) -> None: ...
|
|
719
477
|
def to_df(self, *args, **kwargs) -> pandas.DataFrame: ...
|
|
720
478
|
def to_parquet(
|
|
721
|
-
|
|
722
|
-
|
|
723
|
-
|
|
724
|
-
|
|
725
|
-
|
|
726
|
-
|
|
727
|
-
|
|
728
|
-
|
|
729
|
-
|
|
730
|
-
|
|
731
|
-
|
|
732
|
-
|
|
479
|
+
self,
|
|
480
|
+
file_name: str,
|
|
481
|
+
compression: Optional[str] = None,
|
|
482
|
+
field_ids: Optional[dict | str] = None,
|
|
483
|
+
row_group_size_bytes: Optional[int | str] = None,
|
|
484
|
+
row_group_size: Optional[int] = None,
|
|
485
|
+
partition_by: Optional[List[str]] = None,
|
|
486
|
+
write_partition_columns: Optional[bool] = None,
|
|
487
|
+
overwrite: Optional[bool] = None,
|
|
488
|
+
per_thread_output: Optional[bool] = None,
|
|
489
|
+
use_tmp_file: Optional[bool] = None,
|
|
490
|
+
append: Optional[bool] = None
|
|
733
491
|
) -> None: ...
|
|
734
492
|
def fetch_df_chunk(self, vectors_per_chunk: int = 1, *, date_as_object: bool = False) -> pandas.DataFrame: ...
|
|
735
493
|
def to_table(self, table_name: str) -> None: ...
|
|
@@ -739,37 +497,37 @@ class DuckDBPyRelation:
|
|
|
739
497
|
def union(self, union_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
|
|
740
498
|
def unique(self, unique_aggr: str) -> DuckDBPyRelation: ...
|
|
741
499
|
def write_csv(
|
|
742
|
-
|
|
743
|
-
|
|
744
|
-
|
|
745
|
-
|
|
746
|
-
|
|
747
|
-
|
|
748
|
-
|
|
749
|
-
|
|
750
|
-
|
|
751
|
-
|
|
752
|
-
|
|
753
|
-
|
|
754
|
-
|
|
755
|
-
|
|
756
|
-
|
|
757
|
-
|
|
758
|
-
|
|
500
|
+
self,
|
|
501
|
+
file_name: str,
|
|
502
|
+
sep: Optional[str] = None,
|
|
503
|
+
na_rep: Optional[str] = None,
|
|
504
|
+
header: Optional[bool] = None,
|
|
505
|
+
quotechar: Optional[str] = None,
|
|
506
|
+
escapechar: Optional[str] = None,
|
|
507
|
+
date_format: Optional[str] = None,
|
|
508
|
+
timestamp_format: Optional[str] = None,
|
|
509
|
+
quoting: Optional[str | int] = None,
|
|
510
|
+
encoding: Optional[str] = None,
|
|
511
|
+
compression: Optional[str] = None,
|
|
512
|
+
write_partition_columns: Optional[bool] = None,
|
|
513
|
+
overwrite: Optional[bool] = None,
|
|
514
|
+
per_thread_output: Optional[bool] = None,
|
|
515
|
+
use_tmp_file: Optional[bool] = None,
|
|
516
|
+
partition_by: Optional[List[str]] = None
|
|
759
517
|
) -> None: ...
|
|
760
518
|
def write_parquet(
|
|
761
|
-
|
|
762
|
-
|
|
763
|
-
|
|
764
|
-
|
|
765
|
-
|
|
766
|
-
|
|
767
|
-
|
|
768
|
-
|
|
769
|
-
|
|
770
|
-
|
|
771
|
-
|
|
772
|
-
|
|
519
|
+
self,
|
|
520
|
+
file_name: str,
|
|
521
|
+
compression: Optional[str] = None,
|
|
522
|
+
field_ids: Optional[dict | str] = None,
|
|
523
|
+
row_group_size_bytes: Optional[int | str] = None,
|
|
524
|
+
row_group_size: Optional[int] = None,
|
|
525
|
+
partition_by: Optional[List[str]] = None,
|
|
526
|
+
write_partition_columns: Optional[bool] = None,
|
|
527
|
+
overwrite: Optional[bool] = None,
|
|
528
|
+
per_thread_output: Optional[bool] = None,
|
|
529
|
+
use_tmp_file: Optional[bool] = None,
|
|
530
|
+
append: Optional[bool] = None
|
|
773
531
|
) -> None: ...
|
|
774
532
|
def __len__(self) -> int: ...
|
|
775
533
|
@property
|
|
@@ -788,6 +546,7 @@ class DuckDBPyRelation:
|
|
|
788
546
|
def types(self) -> List[DuckDBPyType]: ...
|
|
789
547
|
|
|
790
548
|
class Error(Exception): ...
|
|
549
|
+
|
|
791
550
|
class FatalException(Error): ...
|
|
792
551
|
|
|
793
552
|
class HTTPException(IOException):
|
|
@@ -797,31 +556,51 @@ class HTTPException(IOException):
|
|
|
797
556
|
headers: Dict[str, str]
|
|
798
557
|
|
|
799
558
|
class IOException(OperationalError): ...
|
|
559
|
+
|
|
800
560
|
class IntegrityError(Error): ...
|
|
561
|
+
|
|
801
562
|
class InternalError(Error): ...
|
|
563
|
+
|
|
802
564
|
class InternalException(InternalError): ...
|
|
565
|
+
|
|
803
566
|
class InterruptException(Error): ...
|
|
567
|
+
|
|
804
568
|
class InvalidInputException(ProgrammingError): ...
|
|
569
|
+
|
|
805
570
|
class InvalidTypeException(ProgrammingError): ...
|
|
571
|
+
|
|
806
572
|
class NotImplementedException(NotSupportedError): ...
|
|
573
|
+
|
|
807
574
|
class NotSupportedError(Error): ...
|
|
575
|
+
|
|
808
576
|
class OperationalError(Error): ...
|
|
577
|
+
|
|
809
578
|
class OutOfMemoryException(OperationalError): ...
|
|
579
|
+
|
|
810
580
|
class OutOfRangeException(DataError): ...
|
|
581
|
+
|
|
811
582
|
class ParserException(ProgrammingError): ...
|
|
583
|
+
|
|
812
584
|
class PermissionException(Error): ...
|
|
585
|
+
|
|
813
586
|
class ProgrammingError(Error): ...
|
|
587
|
+
|
|
814
588
|
class SequenceException(Error): ...
|
|
589
|
+
|
|
815
590
|
class SerializationException(OperationalError): ...
|
|
591
|
+
|
|
816
592
|
class SyntaxException(ProgrammingError): ...
|
|
593
|
+
|
|
817
594
|
class TransactionException(OperationalError): ...
|
|
595
|
+
|
|
818
596
|
class TypeMismatchException(DataError): ...
|
|
597
|
+
|
|
819
598
|
class Warning(Exception): ...
|
|
820
599
|
|
|
821
600
|
class token_type:
|
|
822
601
|
# stubgen override - these make mypy sad
|
|
823
|
-
#
|
|
824
|
-
#
|
|
602
|
+
#__doc__: ClassVar[str] = ... # read-only
|
|
603
|
+
#__members__: ClassVar[dict] = ... # read-only
|
|
825
604
|
__entries: ClassVar[dict] = ...
|
|
826
605
|
comment: ClassVar[token_type] = ...
|
|
827
606
|
identifier: ClassVar[token_type] = ...
|
|
@@ -861,18 +640,7 @@ def register_filesystem(filesystem: fsspec.AbstractFileSystem, *, connection: Du
|
|
|
861
640
|
def unregister_filesystem(name: str, *, connection: DuckDBPyConnection = ...) -> None: ...
|
|
862
641
|
def list_filesystems(*, connection: DuckDBPyConnection = ...) -> list: ...
|
|
863
642
|
def filesystem_is_registered(name: str, *, connection: DuckDBPyConnection = ...) -> bool: ...
|
|
864
|
-
def create_function(
|
|
865
|
-
name: str,
|
|
866
|
-
function: function,
|
|
867
|
-
parameters: Optional[List[DuckDBPyType]] = None,
|
|
868
|
-
return_type: Optional[DuckDBPyType] = None,
|
|
869
|
-
*,
|
|
870
|
-
type: Optional[PythonUDFType] = PythonUDFType.NATIVE,
|
|
871
|
-
null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT,
|
|
872
|
-
exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT,
|
|
873
|
-
side_effects: bool = False,
|
|
874
|
-
connection: DuckDBPyConnection = ...,
|
|
875
|
-
) -> DuckDBPyConnection: ...
|
|
643
|
+
def create_function(name: str, function: function, parameters: Optional[List[DuckDBPyType]] = None, return_type: Optional[DuckDBPyType] = None, *, type: Optional[PythonUDFType] = PythonUDFType.NATIVE, null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT, exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT, side_effects: bool = False, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
876
644
|
def remove_function(name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
877
645
|
def sqltype(type_str: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
878
646
|
def dtype(type_str: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
@@ -881,24 +649,14 @@ def array_type(type: DuckDBPyType, size: int, *, connection: DuckDBPyConnection
|
|
|
881
649
|
def list_type(type: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
882
650
|
def union_type(members: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
883
651
|
def string_type(collation: str = "", *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
884
|
-
def enum_type(
|
|
885
|
-
name: str, type: DuckDBPyType, values: List[Any], *, connection: DuckDBPyConnection = ...
|
|
886
|
-
) -> DuckDBPyType: ...
|
|
652
|
+
def enum_type(name: str, type: DuckDBPyType, values: List[Any], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
887
653
|
def decimal_type(width: int, scale: int, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
888
|
-
def struct_type(
|
|
889
|
-
|
|
890
|
-
) -> DuckDBPyType: ...
|
|
891
|
-
def row_type(
|
|
892
|
-
fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...
|
|
893
|
-
) -> DuckDBPyType: ...
|
|
654
|
+
def struct_type(fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
655
|
+
def row_type(fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
894
656
|
def map_type(key: DuckDBPyType, value: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
|
|
895
657
|
def duplicate(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
896
|
-
def execute(
|
|
897
|
-
|
|
898
|
-
) -> DuckDBPyConnection: ...
|
|
899
|
-
def executemany(
|
|
900
|
-
query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...
|
|
901
|
-
) -> DuckDBPyConnection: ...
|
|
658
|
+
def execute(query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
659
|
+
def executemany(query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
902
660
|
def close(*, connection: DuckDBPyConnection = ...) -> None: ...
|
|
903
661
|
def interrupt(*, connection: DuckDBPyConnection = ...) -> None: ...
|
|
904
662
|
def query_progress(*, connection: DuckDBPyConnection = ...) -> float: ...
|
|
@@ -909,16 +667,10 @@ def fetchnumpy(*, connection: DuckDBPyConnection = ...) -> dict: ...
|
|
|
909
667
|
def fetchdf(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
|
|
910
668
|
def fetch_df(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
|
|
911
669
|
def df(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
|
|
912
|
-
def fetch_df_chunk(
|
|
913
|
-
|
|
914
|
-
) -> pandas.DataFrame: ...
|
|
915
|
-
def pl(
|
|
916
|
-
rows_per_batch: int = 1000000, *, lazy: bool = False, connection: DuckDBPyConnection = ...
|
|
917
|
-
) -> polars.DataFrame: ...
|
|
670
|
+
def fetch_df_chunk(vectors_per_chunk: int = 1, *, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
|
|
671
|
+
def pl(rows_per_batch: int = 1000000, *, lazy: bool = False, connection: DuckDBPyConnection = ...) -> polars.DataFrame: ...
|
|
918
672
|
def fetch_arrow_table(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.Table: ...
|
|
919
|
-
def fetch_record_batch(
|
|
920
|
-
rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...
|
|
921
|
-
) -> pyarrow.lib.RecordBatchReader: ...
|
|
673
|
+
def fetch_record_batch(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
|
|
922
674
|
def arrow(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
|
|
923
675
|
def torch(*, connection: DuckDBPyConnection = ...) -> dict: ...
|
|
924
676
|
def tf(*, connection: DuckDBPyConnection = ...) -> dict: ...
|
|
@@ -926,212 +678,36 @@ def begin(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
|
926
678
|
def commit(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
927
679
|
def rollback(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
928
680
|
def checkpoint(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
929
|
-
def append(
|
|
930
|
-
table_name: str, df: pandas.DataFrame, *, by_name: bool = False, connection: DuckDBPyConnection = ...
|
|
931
|
-
) -> DuckDBPyConnection: ...
|
|
681
|
+
def append(table_name: str, df: pandas.DataFrame, *, by_name: bool = False, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
932
682
|
def register(view_name: str, python_object: object, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
933
683
|
def unregister(view_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
|
|
934
684
|
def table(table_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
935
685
|
def view(view_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
936
|
-
def values(
|
|
937
|
-
|
|
938
|
-
) -> DuckDBPyRelation: ...
|
|
939
|
-
def table_function(
|
|
940
|
-
name: str, parameters: object = None, *, connection: DuckDBPyConnection = ...
|
|
941
|
-
) -> DuckDBPyRelation: ...
|
|
942
|
-
def read_json(
|
|
943
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
944
|
-
*,
|
|
945
|
-
columns: Optional[Dict[str, str]] = None,
|
|
946
|
-
sample_size: Optional[int] = None,
|
|
947
|
-
maximum_depth: Optional[int] = None,
|
|
948
|
-
records: Optional[str] = None,
|
|
949
|
-
format: Optional[str] = None,
|
|
950
|
-
date_format: Optional[str] = None,
|
|
951
|
-
timestamp_format: Optional[str] = None,
|
|
952
|
-
compression: Optional[str] = None,
|
|
953
|
-
maximum_object_size: Optional[int] = None,
|
|
954
|
-
ignore_errors: Optional[bool] = None,
|
|
955
|
-
convert_strings_to_integers: Optional[bool] = None,
|
|
956
|
-
field_appearance_threshold: Optional[float] = None,
|
|
957
|
-
map_inference_threshold: Optional[int] = None,
|
|
958
|
-
maximum_sample_files: Optional[int] = None,
|
|
959
|
-
filename: Optional[Union[bool, str]] = None,
|
|
960
|
-
hive_partitioning: Optional[bool] = None,
|
|
961
|
-
union_by_name: Optional[bool] = None,
|
|
962
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
963
|
-
hive_types_autocast: Optional[bool] = None,
|
|
964
|
-
connection: DuckDBPyConnection = ...,
|
|
965
|
-
) -> DuckDBPyRelation: ...
|
|
686
|
+
def values(*args: Union[List[Any],Expression, Tuple[Expression]], connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
687
|
+
def table_function(name: str, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
688
|
+
def read_json(path_or_buffer: Union[str, StringIO, TextIOBase], *, columns: Optional[Dict[str,str]] = None, sample_size: Optional[int] = None, maximum_depth: Optional[int] = None, records: Optional[str] = None, format: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, compression: Optional[str] = None, maximum_object_size: Optional[int] = None, ignore_errors: Optional[bool] = None, convert_strings_to_integers: Optional[bool] = None, field_appearance_threshold: Optional[float] = None, map_inference_threshold: Optional[int] = None, maximum_sample_files: Optional[int] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
966
689
|
def extract_statements(query: str, *, connection: DuckDBPyConnection = ...) -> List[Statement]: ...
|
|
967
|
-
def sql(
|
|
968
|
-
|
|
969
|
-
) -> DuckDBPyRelation: ...
|
|
970
|
-
def
|
|
971
|
-
|
|
972
|
-
) -> DuckDBPyRelation: ...
|
|
973
|
-
def from_query(
|
|
974
|
-
query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...
|
|
975
|
-
) -> DuckDBPyRelation: ...
|
|
976
|
-
def read_csv(
|
|
977
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
978
|
-
*,
|
|
979
|
-
header: Optional[Union[bool, int]] = None,
|
|
980
|
-
compression: Optional[str] = None,
|
|
981
|
-
sep: Optional[str] = None,
|
|
982
|
-
delimiter: Optional[str] = None,
|
|
983
|
-
dtype: Optional[Union[Dict[str, str], List[str]]] = None,
|
|
984
|
-
na_values: Optional[Union[str, List[str]]] = None,
|
|
985
|
-
skiprows: Optional[int] = None,
|
|
986
|
-
quotechar: Optional[str] = None,
|
|
987
|
-
escapechar: Optional[str] = None,
|
|
988
|
-
encoding: Optional[str] = None,
|
|
989
|
-
parallel: Optional[bool] = None,
|
|
990
|
-
date_format: Optional[str] = None,
|
|
991
|
-
timestamp_format: Optional[str] = None,
|
|
992
|
-
sample_size: Optional[int] = None,
|
|
993
|
-
all_varchar: Optional[bool] = None,
|
|
994
|
-
normalize_names: Optional[bool] = None,
|
|
995
|
-
null_padding: Optional[bool] = None,
|
|
996
|
-
names: Optional[List[str]] = None,
|
|
997
|
-
lineterminator: Optional[str] = None,
|
|
998
|
-
columns: Optional[Dict[str, str]] = None,
|
|
999
|
-
auto_type_candidates: Optional[List[str]] = None,
|
|
1000
|
-
max_line_size: Optional[int] = None,
|
|
1001
|
-
ignore_errors: Optional[bool] = None,
|
|
1002
|
-
store_rejects: Optional[bool] = None,
|
|
1003
|
-
rejects_table: Optional[str] = None,
|
|
1004
|
-
rejects_scan: Optional[str] = None,
|
|
1005
|
-
rejects_limit: Optional[int] = None,
|
|
1006
|
-
force_not_null: Optional[List[str]] = None,
|
|
1007
|
-
buffer_size: Optional[int] = None,
|
|
1008
|
-
decimal: Optional[str] = None,
|
|
1009
|
-
allow_quoted_nulls: Optional[bool] = None,
|
|
1010
|
-
filename: Optional[Union[bool, str]] = None,
|
|
1011
|
-
hive_partitioning: Optional[bool] = None,
|
|
1012
|
-
union_by_name: Optional[bool] = None,
|
|
1013
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
1014
|
-
hive_types_autocast: Optional[bool] = None,
|
|
1015
|
-
connection: DuckDBPyConnection = ...,
|
|
1016
|
-
) -> DuckDBPyRelation: ...
|
|
1017
|
-
def from_csv_auto(
|
|
1018
|
-
path_or_buffer: Union[str, StringIO, TextIOBase],
|
|
1019
|
-
*,
|
|
1020
|
-
header: Optional[Union[bool, int]] = None,
|
|
1021
|
-
compression: Optional[str] = None,
|
|
1022
|
-
sep: Optional[str] = None,
|
|
1023
|
-
delimiter: Optional[str] = None,
|
|
1024
|
-
dtype: Optional[Union[Dict[str, str], List[str]]] = None,
|
|
1025
|
-
na_values: Optional[Union[str, List[str]]] = None,
|
|
1026
|
-
skiprows: Optional[int] = None,
|
|
1027
|
-
quotechar: Optional[str] = None,
|
|
1028
|
-
escapechar: Optional[str] = None,
|
|
1029
|
-
encoding: Optional[str] = None,
|
|
1030
|
-
parallel: Optional[bool] = None,
|
|
1031
|
-
date_format: Optional[str] = None,
|
|
1032
|
-
timestamp_format: Optional[str] = None,
|
|
1033
|
-
sample_size: Optional[int] = None,
|
|
1034
|
-
all_varchar: Optional[bool] = None,
|
|
1035
|
-
normalize_names: Optional[bool] = None,
|
|
1036
|
-
null_padding: Optional[bool] = None,
|
|
1037
|
-
names: Optional[List[str]] = None,
|
|
1038
|
-
lineterminator: Optional[str] = None,
|
|
1039
|
-
columns: Optional[Dict[str, str]] = None,
|
|
1040
|
-
auto_type_candidates: Optional[List[str]] = None,
|
|
1041
|
-
max_line_size: Optional[int] = None,
|
|
1042
|
-
ignore_errors: Optional[bool] = None,
|
|
1043
|
-
store_rejects: Optional[bool] = None,
|
|
1044
|
-
rejects_table: Optional[str] = None,
|
|
1045
|
-
rejects_scan: Optional[str] = None,
|
|
1046
|
-
rejects_limit: Optional[int] = None,
|
|
1047
|
-
force_not_null: Optional[List[str]] = None,
|
|
1048
|
-
buffer_size: Optional[int] = None,
|
|
1049
|
-
decimal: Optional[str] = None,
|
|
1050
|
-
allow_quoted_nulls: Optional[bool] = None,
|
|
1051
|
-
filename: Optional[Union[bool, str]] = None,
|
|
1052
|
-
hive_partitioning: Optional[bool] = None,
|
|
1053
|
-
union_by_name: Optional[bool] = None,
|
|
1054
|
-
hive_types: Optional[Dict[str, str]] = None,
|
|
1055
|
-
hive_types_autocast: Optional[bool] = None,
|
|
1056
|
-
connection: DuckDBPyConnection = ...,
|
|
1057
|
-
) -> DuckDBPyRelation: ...
|
|
690
|
+
def sql(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
691
|
+
def query(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
692
|
+
def from_query(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
693
|
+
def read_csv(path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
694
|
+
def from_csv_auto(path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1058
695
|
def from_df(df: pandas.DataFrame, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1059
696
|
def from_arrow(arrow_object: object, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1060
|
-
def from_parquet(
|
|
1061
|
-
|
|
1062
|
-
binary_as_string: bool = False,
|
|
1063
|
-
*,
|
|
1064
|
-
file_row_number: bool = False,
|
|
1065
|
-
filename: bool = False,
|
|
1066
|
-
hive_partitioning: bool = False,
|
|
1067
|
-
union_by_name: bool = False,
|
|
1068
|
-
compression: Optional[str] = None,
|
|
1069
|
-
connection: DuckDBPyConnection = ...,
|
|
1070
|
-
) -> DuckDBPyRelation: ...
|
|
1071
|
-
def read_parquet(
|
|
1072
|
-
file_glob: str,
|
|
1073
|
-
binary_as_string: bool = False,
|
|
1074
|
-
*,
|
|
1075
|
-
file_row_number: bool = False,
|
|
1076
|
-
filename: bool = False,
|
|
1077
|
-
hive_partitioning: bool = False,
|
|
1078
|
-
union_by_name: bool = False,
|
|
1079
|
-
compression: Optional[str] = None,
|
|
1080
|
-
connection: DuckDBPyConnection = ...,
|
|
1081
|
-
) -> DuckDBPyRelation: ...
|
|
697
|
+
def from_parquet(file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
698
|
+
def read_parquet(file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1082
699
|
def get_table_names(query: str, *, qualified: bool = False, connection: DuckDBPyConnection = ...) -> Set[str]: ...
|
|
1083
|
-
def install_extension(
|
|
1084
|
-
extension: str,
|
|
1085
|
-
*,
|
|
1086
|
-
force_install: bool = False,
|
|
1087
|
-
repository: Optional[str] = None,
|
|
1088
|
-
repository_url: Optional[str] = None,
|
|
1089
|
-
version: Optional[str] = None,
|
|
1090
|
-
connection: DuckDBPyConnection = ...,
|
|
1091
|
-
) -> None: ...
|
|
700
|
+
def install_extension(extension: str, *, force_install: bool = False, repository: Optional[str] = None, repository_url: Optional[str] = None, version: Optional[str] = None, connection: DuckDBPyConnection = ...) -> None: ...
|
|
1092
701
|
def load_extension(extension: str, *, connection: DuckDBPyConnection = ...) -> None: ...
|
|
1093
|
-
def project(
|
|
1094
|
-
df: pandas.DataFrame, *args: str, groups: str = "", connection: DuckDBPyConnection = ...
|
|
1095
|
-
) -> DuckDBPyRelation: ...
|
|
702
|
+
def project(df: pandas.DataFrame, *args: str, groups: str = "", connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1096
703
|
def distinct(df: pandas.DataFrame, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1097
|
-
def write_csv(
|
|
1098
|
-
|
|
1099
|
-
filename: str,
|
|
1100
|
-
*,
|
|
1101
|
-
sep: Optional[str] = None,
|
|
1102
|
-
na_rep: Optional[str] = None,
|
|
1103
|
-
header: Optional[bool] = None,
|
|
1104
|
-
quotechar: Optional[str] = None,
|
|
1105
|
-
escapechar: Optional[str] = None,
|
|
1106
|
-
date_format: Optional[str] = None,
|
|
1107
|
-
timestamp_format: Optional[str] = None,
|
|
1108
|
-
quoting: Optional[Union[str, int]] = None,
|
|
1109
|
-
encoding: Optional[str] = None,
|
|
1110
|
-
compression: Optional[str] = None,
|
|
1111
|
-
overwrite: Optional[bool] = None,
|
|
1112
|
-
per_thread_output: Optional[bool] = None,
|
|
1113
|
-
use_tmp_file: Optional[bool] = None,
|
|
1114
|
-
partition_by: Optional[List[str]] = None,
|
|
1115
|
-
write_partition_columns: Optional[bool] = None,
|
|
1116
|
-
connection: DuckDBPyConnection = ...,
|
|
1117
|
-
) -> None: ...
|
|
1118
|
-
def aggregate(
|
|
1119
|
-
df: pandas.DataFrame,
|
|
1120
|
-
aggr_expr: Union[str, List[Expression]],
|
|
1121
|
-
group_expr: str = "",
|
|
1122
|
-
*,
|
|
1123
|
-
connection: DuckDBPyConnection = ...,
|
|
1124
|
-
) -> DuckDBPyRelation: ...
|
|
704
|
+
def write_csv(df: pandas.DataFrame, filename: str, *, sep: Optional[str] = None, na_rep: Optional[str] = None, header: Optional[bool] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, quoting: Optional[str | int] = None, encoding: Optional[str] = None, compression: Optional[str] = None, overwrite: Optional[bool] = None, per_thread_output: Optional[bool] = None, use_tmp_file: Optional[bool] = None, partition_by: Optional[List[str]] = None, write_partition_columns: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> None: ...
|
|
705
|
+
def aggregate(df: pandas.DataFrame, aggr_expr: str | List[Expression], group_expr: str = "", *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1125
706
|
def alias(df: pandas.DataFrame, alias: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1126
707
|
def filter(df: pandas.DataFrame, filter_expr: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1127
|
-
def limit(
|
|
1128
|
-
df: pandas.DataFrame, n: int, offset: int = 0, *, connection: DuckDBPyConnection = ...
|
|
1129
|
-
) -> DuckDBPyRelation: ...
|
|
708
|
+
def limit(df: pandas.DataFrame, n: int, offset: int = 0, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1130
709
|
def order(df: pandas.DataFrame, order_expr: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1131
|
-
def query_df(
|
|
1132
|
-
df: pandas.DataFrame, virtual_table_name: str, sql_query: str, *, connection: DuckDBPyConnection = ...
|
|
1133
|
-
) -> DuckDBPyRelation: ...
|
|
710
|
+
def query_df(df: pandas.DataFrame, virtual_table_name: str, sql_query: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
|
|
1134
711
|
def description(*, connection: DuckDBPyConnection = ...) -> Optional[List[Any]]: ...
|
|
1135
712
|
def rowcount(*, connection: DuckDBPyConnection = ...) -> int: ...
|
|
1136
|
-
|
|
1137
713
|
# END OF CONNECTION WRAPPER
|