duckdb 1.4.1.dev116__cp311-cp311-macosx_11_0_arm64.whl → 1.5.0.dev44__cp311-cp311-macosx_11_0_arm64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of duckdb might be problematic. Click here for more details.

Files changed (46) hide show
  1. _duckdb.cpython-311-darwin.so +0 -0
  2. duckdb/__init__.py +374 -373
  3. duckdb/__init__.pyi +180 -604
  4. duckdb/bytes_io_wrapper.py +7 -6
  5. duckdb/experimental/__init__.py +1 -2
  6. duckdb/experimental/spark/__init__.py +4 -3
  7. duckdb/experimental/spark/_globals.py +8 -8
  8. duckdb/experimental/spark/_typing.py +9 -7
  9. duckdb/experimental/spark/conf.py +15 -16
  10. duckdb/experimental/spark/context.py +44 -60
  11. duckdb/experimental/spark/errors/__init__.py +35 -33
  12. duckdb/experimental/spark/errors/error_classes.py +1 -1
  13. duckdb/experimental/spark/errors/exceptions/__init__.py +1 -1
  14. duckdb/experimental/spark/errors/exceptions/base.py +88 -39
  15. duckdb/experimental/spark/errors/utils.py +16 -11
  16. duckdb/experimental/spark/exception.py +6 -9
  17. duckdb/experimental/spark/sql/__init__.py +5 -5
  18. duckdb/experimental/spark/sql/_typing.py +15 -8
  19. duckdb/experimental/spark/sql/catalog.py +20 -21
  20. duckdb/experimental/spark/sql/column.py +54 -47
  21. duckdb/experimental/spark/sql/conf.py +8 -9
  22. duckdb/experimental/spark/sql/dataframe.py +233 -185
  23. duckdb/experimental/spark/sql/functions.py +1248 -1222
  24. duckdb/experimental/spark/sql/group.py +52 -56
  25. duckdb/experimental/spark/sql/readwriter.py +94 -80
  26. duckdb/experimental/spark/sql/session.py +59 -64
  27. duckdb/experimental/spark/sql/streaming.py +10 -9
  28. duckdb/experimental/spark/sql/type_utils.py +64 -66
  29. duckdb/experimental/spark/sql/types.py +344 -308
  30. duckdb/experimental/spark/sql/udf.py +6 -6
  31. duckdb/filesystem.py +8 -13
  32. duckdb/functional/__init__.py +16 -2
  33. duckdb/polars_io.py +57 -66
  34. duckdb/query_graph/__main__.py +96 -91
  35. duckdb/typing/__init__.py +8 -8
  36. duckdb/typing/__init__.pyi +2 -4
  37. duckdb/udf.py +5 -10
  38. duckdb/value/__init__.py +0 -1
  39. duckdb/value/constant/__init__.py +59 -61
  40. duckdb/value/constant/__init__.pyi +4 -3
  41. duckdb-1.5.0.dev44.dist-info/METADATA +80 -0
  42. duckdb-1.5.0.dev44.dist-info/RECORD +47 -0
  43. duckdb-1.4.1.dev116.dist-info/METADATA +0 -326
  44. duckdb-1.4.1.dev116.dist-info/RECORD +0 -47
  45. {duckdb-1.4.1.dev116.dist-info → duckdb-1.5.0.dev44.dist-info}/WHEEL +0 -0
  46. {duckdb-1.4.1.dev116.dist-info → duckdb-1.5.0.dev44.dist-info}/licenses/LICENSE +0 -0
duckdb/__init__.pyi CHANGED
@@ -41,7 +41,6 @@ from duckdb.value.constant import (
41
41
 
42
42
  # We also run this in python3.7, where this is needed
43
43
  from typing_extensions import Literal
44
-
45
44
  # stubgen override - missing import of Set
46
45
  from typing import Any, ClassVar, Set, Optional, Callable
47
46
  from io import StringIO, TextIOBase
@@ -49,13 +48,11 @@ from pathlib import Path
49
48
 
50
49
  from typing import overload, Dict, List, Union, Tuple
51
50
  import pandas
52
-
53
51
  # stubgen override - unfortunately we need this for version checks
54
52
  import sys
55
53
  import fsspec
56
54
  import pyarrow.lib
57
55
  import polars
58
-
59
56
  # stubgen override - This should probably not be exposed
60
57
  apilevel: str
61
58
  comment: token_type
@@ -81,10 +78,15 @@ __jupyter__: bool
81
78
  __formatted_python_version__: str
82
79
 
83
80
  class BinderException(ProgrammingError): ...
81
+
84
82
  class CatalogException(ProgrammingError): ...
83
+
85
84
  class ConnectionException(OperationalError): ...
85
+
86
86
  class ConstraintException(IntegrityError): ...
87
+
87
88
  class ConversionException(DataError): ...
89
+
88
90
  class DataError(Error): ...
89
91
 
90
92
  class ExplainType:
@@ -202,37 +204,46 @@ class Statement:
202
204
  class Expression:
203
205
  def __init__(self, *args, **kwargs) -> None: ...
204
206
  def __neg__(self) -> "Expression": ...
207
+
205
208
  def __add__(self, expr: "Expression") -> "Expression": ...
206
209
  def __radd__(self, expr: "Expression") -> "Expression": ...
210
+
207
211
  def __sub__(self, expr: "Expression") -> "Expression": ...
208
212
  def __rsub__(self, expr: "Expression") -> "Expression": ...
213
+
209
214
  def __mul__(self, expr: "Expression") -> "Expression": ...
210
215
  def __rmul__(self, expr: "Expression") -> "Expression": ...
216
+
211
217
  def __div__(self, expr: "Expression") -> "Expression": ...
212
218
  def __rdiv__(self, expr: "Expression") -> "Expression": ...
219
+
213
220
  def __truediv__(self, expr: "Expression") -> "Expression": ...
214
221
  def __rtruediv__(self, expr: "Expression") -> "Expression": ...
222
+
215
223
  def __floordiv__(self, expr: "Expression") -> "Expression": ...
216
224
  def __rfloordiv__(self, expr: "Expression") -> "Expression": ...
225
+
217
226
  def __mod__(self, expr: "Expression") -> "Expression": ...
218
227
  def __rmod__(self, expr: "Expression") -> "Expression": ...
228
+
219
229
  def __pow__(self, expr: "Expression") -> "Expression": ...
220
230
  def __rpow__(self, expr: "Expression") -> "Expression": ...
231
+
221
232
  def __and__(self, expr: "Expression") -> "Expression": ...
222
233
  def __rand__(self, expr: "Expression") -> "Expression": ...
223
234
  def __or__(self, expr: "Expression") -> "Expression": ...
224
235
  def __ror__(self, expr: "Expression") -> "Expression": ...
225
236
  def __invert__(self) -> "Expression": ...
226
- def __eq__( # type: ignore[override]
227
- self, expr: "Expression"
228
- ) -> "Expression": ...
229
- def __ne__( # type: ignore[override]
230
- self, expr: "Expression"
231
- ) -> "Expression": ...
237
+
238
+ def __eq__(# type: ignore[override]
239
+ self, expr: "Expression") -> "Expression": ...
240
+ def __ne__(# type: ignore[override]
241
+ self, expr: "Expression") -> "Expression": ...
232
242
  def __gt__(self, expr: "Expression") -> "Expression": ...
233
243
  def __ge__(self, expr: "Expression") -> "Expression": ...
234
244
  def __lt__(self, expr: "Expression") -> "Expression": ...
235
245
  def __le__(self, expr: "Expression") -> "Expression": ...
246
+
236
247
  def show(self) -> None: ...
237
248
  def __repr__(self) -> str: ...
238
249
  def get_name(self) -> str: ...
@@ -280,18 +291,7 @@ class DuckDBPyConnection:
280
291
  def unregister_filesystem(self, name: str) -> None: ...
281
292
  def list_filesystems(self) -> list: ...
282
293
  def filesystem_is_registered(self, name: str) -> bool: ...
283
- def create_function(
284
- self,
285
- name: str,
286
- function: function,
287
- parameters: Optional[List[DuckDBPyType]] = None,
288
- return_type: Optional[DuckDBPyType] = None,
289
- *,
290
- type: Optional[PythonUDFType] = PythonUDFType.NATIVE,
291
- null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT,
292
- exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT,
293
- side_effects: bool = False,
294
- ) -> DuckDBPyConnection: ...
294
+ def create_function(self, name: str, function: function, parameters: Optional[List[DuckDBPyType]] = None, return_type: Optional[DuckDBPyType] = None, *, type: Optional[PythonUDFType] = PythonUDFType.NATIVE, null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT, exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT, side_effects: bool = False) -> DuckDBPyConnection: ...
295
295
  def remove_function(self, name: str) -> DuckDBPyConnection: ...
296
296
  def sqltype(self, type_str: str) -> DuckDBPyType: ...
297
297
  def dtype(self, type_str: str) -> DuckDBPyType: ...
@@ -334,152 +334,21 @@ class DuckDBPyConnection:
334
334
  def unregister(self, view_name: str) -> DuckDBPyConnection: ...
335
335
  def table(self, table_name: str) -> DuckDBPyRelation: ...
336
336
  def view(self, view_name: str) -> DuckDBPyRelation: ...
337
- def values(self, *args: Union[List[Any], Expression, Tuple[Expression]]) -> DuckDBPyRelation: ...
337
+ def values(self, *args: Union[List[Any],Expression, Tuple[Expression]]) -> DuckDBPyRelation: ...
338
338
  def table_function(self, name: str, parameters: object = None) -> DuckDBPyRelation: ...
339
- def read_json(
340
- self,
341
- path_or_buffer: Union[str, StringIO, TextIOBase],
342
- *,
343
- columns: Optional[Dict[str, str]] = None,
344
- sample_size: Optional[int] = None,
345
- maximum_depth: Optional[int] = None,
346
- records: Optional[str] = None,
347
- format: Optional[str] = None,
348
- date_format: Optional[str] = None,
349
- timestamp_format: Optional[str] = None,
350
- compression: Optional[str] = None,
351
- maximum_object_size: Optional[int] = None,
352
- ignore_errors: Optional[bool] = None,
353
- convert_strings_to_integers: Optional[bool] = None,
354
- field_appearance_threshold: Optional[float] = None,
355
- map_inference_threshold: Optional[int] = None,
356
- maximum_sample_files: Optional[int] = None,
357
- filename: Optional[Union[bool, str]] = None,
358
- hive_partitioning: Optional[bool] = None,
359
- union_by_name: Optional[bool] = None,
360
- hive_types: Optional[Dict[str, str]] = None,
361
- hive_types_autocast: Optional[bool] = None,
362
- ) -> DuckDBPyRelation: ...
339
+ def read_json(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, columns: Optional[Dict[str,str]] = None, sample_size: Optional[int] = None, maximum_depth: Optional[int] = None, records: Optional[str] = None, format: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, compression: Optional[str] = None, maximum_object_size: Optional[int] = None, ignore_errors: Optional[bool] = None, convert_strings_to_integers: Optional[bool] = None, field_appearance_threshold: Optional[float] = None, map_inference_threshold: Optional[int] = None, maximum_sample_files: Optional[int] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
363
340
  def extract_statements(self, query: str) -> List[Statement]: ...
364
341
  def sql(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
365
342
  def query(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
366
343
  def from_query(self, query: str, *, alias: str = "", params: object = None) -> DuckDBPyRelation: ...
367
- def read_csv(
368
- self,
369
- path_or_buffer: Union[str, StringIO, TextIOBase],
370
- *,
371
- header: Optional[Union[bool, int]] = None,
372
- compression: Optional[str] = None,
373
- sep: Optional[str] = None,
374
- delimiter: Optional[str] = None,
375
- dtype: Optional[Union[Dict[str, str], List[str]]] = None,
376
- na_values: Optional[Union[str, List[str]]] = None,
377
- skiprows: Optional[int] = None,
378
- quotechar: Optional[str] = None,
379
- escapechar: Optional[str] = None,
380
- encoding: Optional[str] = None,
381
- parallel: Optional[bool] = None,
382
- date_format: Optional[str] = None,
383
- timestamp_format: Optional[str] = None,
384
- sample_size: Optional[int] = None,
385
- all_varchar: Optional[bool] = None,
386
- normalize_names: Optional[bool] = None,
387
- null_padding: Optional[bool] = None,
388
- names: Optional[List[str]] = None,
389
- lineterminator: Optional[str] = None,
390
- columns: Optional[Dict[str, str]] = None,
391
- auto_type_candidates: Optional[List[str]] = None,
392
- max_line_size: Optional[int] = None,
393
- ignore_errors: Optional[bool] = None,
394
- store_rejects: Optional[bool] = None,
395
- rejects_table: Optional[str] = None,
396
- rejects_scan: Optional[str] = None,
397
- rejects_limit: Optional[int] = None,
398
- force_not_null: Optional[List[str]] = None,
399
- buffer_size: Optional[int] = None,
400
- decimal: Optional[str] = None,
401
- allow_quoted_nulls: Optional[bool] = None,
402
- filename: Optional[Union[bool, str]] = None,
403
- hive_partitioning: Optional[bool] = None,
404
- union_by_name: Optional[bool] = None,
405
- hive_types: Optional[Dict[str, str]] = None,
406
- hive_types_autocast: Optional[bool] = None,
407
- ) -> DuckDBPyRelation: ...
408
- def from_csv_auto(
409
- self,
410
- path_or_buffer: Union[str, StringIO, TextIOBase],
411
- *,
412
- header: Optional[Union[bool, int]] = None,
413
- compression: Optional[str] = None,
414
- sep: Optional[str] = None,
415
- delimiter: Optional[str] = None,
416
- dtype: Optional[Union[Dict[str, str], List[str]]] = None,
417
- na_values: Optional[Union[str, List[str]]] = None,
418
- skiprows: Optional[int] = None,
419
- quotechar: Optional[str] = None,
420
- escapechar: Optional[str] = None,
421
- encoding: Optional[str] = None,
422
- parallel: Optional[bool] = None,
423
- date_format: Optional[str] = None,
424
- timestamp_format: Optional[str] = None,
425
- sample_size: Optional[int] = None,
426
- all_varchar: Optional[bool] = None,
427
- normalize_names: Optional[bool] = None,
428
- null_padding: Optional[bool] = None,
429
- names: Optional[List[str]] = None,
430
- lineterminator: Optional[str] = None,
431
- columns: Optional[Dict[str, str]] = None,
432
- auto_type_candidates: Optional[List[str]] = None,
433
- max_line_size: Optional[int] = None,
434
- ignore_errors: Optional[bool] = None,
435
- store_rejects: Optional[bool] = None,
436
- rejects_table: Optional[str] = None,
437
- rejects_scan: Optional[str] = None,
438
- rejects_limit: Optional[int] = None,
439
- force_not_null: Optional[List[str]] = None,
440
- buffer_size: Optional[int] = None,
441
- decimal: Optional[str] = None,
442
- allow_quoted_nulls: Optional[bool] = None,
443
- filename: Optional[Union[bool, str]] = None,
444
- hive_partitioning: Optional[bool] = None,
445
- union_by_name: Optional[bool] = None,
446
- hive_types: Optional[Dict[str, str]] = None,
447
- hive_types_autocast: Optional[bool] = None,
448
- ) -> DuckDBPyRelation: ...
344
+ def read_csv(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
345
+ def from_csv_auto(self, path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None) -> DuckDBPyRelation: ...
449
346
  def from_df(self, df: pandas.DataFrame) -> DuckDBPyRelation: ...
450
347
  def from_arrow(self, arrow_object: object) -> DuckDBPyRelation: ...
451
- def from_parquet(
452
- self,
453
- file_glob: str,
454
- binary_as_string: bool = False,
455
- *,
456
- file_row_number: bool = False,
457
- filename: bool = False,
458
- hive_partitioning: bool = False,
459
- union_by_name: bool = False,
460
- compression: Optional[str] = None,
461
- ) -> DuckDBPyRelation: ...
462
- def read_parquet(
463
- self,
464
- file_glob: str,
465
- binary_as_string: bool = False,
466
- *,
467
- file_row_number: bool = False,
468
- filename: bool = False,
469
- hive_partitioning: bool = False,
470
- union_by_name: bool = False,
471
- compression: Optional[str] = None,
472
- ) -> DuckDBPyRelation: ...
348
+ def from_parquet(self, file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None) -> DuckDBPyRelation: ...
349
+ def read_parquet(self, file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None) -> DuckDBPyRelation: ...
473
350
  def get_table_names(self, query: str, *, qualified: bool = False) -> Set[str]: ...
474
- def install_extension(
475
- self,
476
- extension: str,
477
- *,
478
- force_install: bool = False,
479
- repository: Optional[str] = None,
480
- repository_url: Optional[str] = None,
481
- version: Optional[str] = None,
482
- ) -> None: ...
351
+ def install_extension(self, extension: str, *, force_install: bool = False, repository: Optional[str] = None, repository_url: Optional[str] = None, version: Optional[str] = None) -> None: ...
483
352
  def load_extension(self, extension: str) -> None: ...
484
353
  # END OF CONNECTION METHODS
485
354
 
@@ -490,41 +359,19 @@ class DuckDBPyRelation:
490
359
  def __init__(self, *args, **kwargs) -> None: ...
491
360
  def __contains__(self, name: str) -> bool: ...
492
361
  def aggregate(self, aggr_expr: str, group_expr: str = ...) -> DuckDBPyRelation: ...
493
- def apply(
494
- self,
495
- function_name: str,
496
- function_aggr: str,
497
- group_expr: str = ...,
498
- function_parameter: str = ...,
499
- projected_columns: str = ...,
500
- ) -> DuckDBPyRelation: ...
362
+ def apply(self, function_name: str, function_aggr: str, group_expr: str = ..., function_parameter: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
363
+
501
364
  def cume_dist(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
502
365
  def dense_rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
503
366
  def percent_rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
504
367
  def rank(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
505
368
  def rank_dense(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
506
369
  def row_number(self, window_spec: str, projected_columns: str = ...) -> DuckDBPyRelation: ...
507
- def lag(
508
- self,
509
- column: str,
510
- window_spec: str,
511
- offset: int,
512
- default_value: str,
513
- ignore_nulls: bool,
514
- projected_columns: str = ...,
515
- ) -> DuckDBPyRelation: ...
516
- def lead(
517
- self,
518
- column: str,
519
- window_spec: str,
520
- offset: int,
521
- default_value: str,
522
- ignore_nulls: bool,
523
- projected_columns: str = ...,
524
- ) -> DuckDBPyRelation: ...
525
- def nth_value(
526
- self, column: str, window_spec: str, offset: int, ignore_nulls: bool = ..., projected_columns: str = ...
527
- ) -> DuckDBPyRelation: ...
370
+
371
+ def lag(self, column: str, window_spec: str, offset: int, default_value: str, ignore_nulls: bool, projected_columns: str = ...) -> DuckDBPyRelation: ...
372
+ def lead(self, column: str, window_spec: str, offset: int, default_value: str, ignore_nulls: bool, projected_columns: str = ...) -> DuckDBPyRelation: ...
373
+ def nth_value(self, column: str, window_spec: str, offset: int, ignore_nulls: bool = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
374
+
528
375
  def value_counts(self, column: str, groups: str = ...) -> DuckDBPyRelation: ...
529
376
  def geomean(self, column: str, groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
530
377
  def first(self, column: str, groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
@@ -533,119 +380,41 @@ class DuckDBPyRelation:
533
380
  def last_value(self, column: str, window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
534
381
  def mode(self, aggregation_columns: str, group_columns: str = ...) -> DuckDBPyRelation: ...
535
382
  def n_tile(self, window_spec: str, num_buckets: int, projected_columns: str = ...) -> DuckDBPyRelation: ...
536
- def quantile_cont(
537
- self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...
538
- ) -> DuckDBPyRelation: ...
539
- def quantile_disc(
540
- self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...
541
- ) -> DuckDBPyRelation: ...
383
+ def quantile_cont(self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
384
+ def quantile_disc(self, column: str, q: Union[float, List[float]] = ..., groups: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
542
385
  def sum(self, sum_aggr: str, group_expr: str = ...) -> DuckDBPyRelation: ...
543
- def any_value(
544
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
545
- ) -> DuckDBPyRelation: ...
546
- def arg_max(
547
- self,
548
- arg_column: str,
549
- value_column: str,
550
- groups: str = ...,
551
- window_spec: str = ...,
552
- projected_columns: str = ...,
553
- ) -> DuckDBPyRelation: ...
554
- def arg_min(
555
- self,
556
- arg_column: str,
557
- value_column: str,
558
- groups: str = ...,
559
- window_spec: str = ...,
560
- projected_columns: str = ...,
561
- ) -> DuckDBPyRelation: ...
562
- def avg(
563
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
564
- ) -> DuckDBPyRelation: ...
565
- def bit_and(
566
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
567
- ) -> DuckDBPyRelation: ...
568
- def bit_or(
569
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
570
- ) -> DuckDBPyRelation: ...
571
- def bit_xor(
572
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
573
- ) -> DuckDBPyRelation: ...
574
- def bitstring_agg(
575
- self,
576
- column: str,
577
- min: Optional[int],
578
- max: Optional[int],
579
- groups: str = ...,
580
- window_spec: str = ...,
581
- projected_columns: str = ...,
582
- ) -> DuckDBPyRelation: ...
583
- def bool_and(
584
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
585
- ) -> DuckDBPyRelation: ...
586
- def bool_or(
587
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
588
- ) -> DuckDBPyRelation: ...
589
- def count(
590
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
591
- ) -> DuckDBPyRelation: ...
592
- def favg(
593
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
594
- ) -> DuckDBPyRelation: ...
595
- def fsum(
596
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
597
- ) -> DuckDBPyRelation: ...
598
- def histogram(
599
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
600
- ) -> DuckDBPyRelation: ...
601
- def max(
602
- self, max_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
603
- ) -> DuckDBPyRelation: ...
604
- def min(
605
- self, min_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
606
- ) -> DuckDBPyRelation: ...
607
- def mean(
608
- self, mean_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
609
- ) -> DuckDBPyRelation: ...
610
- def median(
611
- self, median_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
612
- ) -> DuckDBPyRelation: ...
613
- def product(
614
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
615
- ) -> DuckDBPyRelation: ...
616
- def quantile(
617
- self, q: str, quantile_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
618
- ) -> DuckDBPyRelation: ...
619
- def std(
620
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
621
- ) -> DuckDBPyRelation: ...
622
- def stddev(
623
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
624
- ) -> DuckDBPyRelation: ...
625
- def stddev_pop(
626
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
627
- ) -> DuckDBPyRelation: ...
628
- def stddev_samp(
629
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
630
- ) -> DuckDBPyRelation: ...
631
- def string_agg(
632
- self, column: str, sep: str = ..., groups: str = ..., window_spec: str = ..., projected_columns: str = ...
633
- ) -> DuckDBPyRelation: ...
634
- def var(
635
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
636
- ) -> DuckDBPyRelation: ...
637
- def var_pop(
638
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
639
- ) -> DuckDBPyRelation: ...
640
- def var_samp(
641
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
642
- ) -> DuckDBPyRelation: ...
643
- def variance(
644
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
645
- ) -> DuckDBPyRelation: ...
646
- def list(
647
- self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...
648
- ) -> DuckDBPyRelation: ...
386
+
387
+ def any_value(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
388
+ def arg_max(self, arg_column: str, value_column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
389
+ def arg_min(self, arg_column: str, value_column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
390
+ def avg(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
391
+ def bit_and(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
392
+ def bit_or(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
393
+ def bit_xor(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
394
+ def bitstring_agg(self, column: str, min: Optional[int], max: Optional[int], groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
395
+ def bool_and(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
396
+ def bool_or(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
397
+ def count(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
398
+ def favg(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
399
+ def fsum(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
400
+ def histogram(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
401
+ def max(self, max_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
402
+ def min(self, min_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
403
+ def mean(self, mean_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
404
+ def median(self, median_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
405
+ def product(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
406
+ def quantile(self, q: str, quantile_aggr: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
407
+ def std(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
408
+ def stddev(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
409
+ def stddev_pop(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
410
+ def stddev_samp(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
411
+ def string_agg(self, column: str, sep: str = ..., groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
412
+ def var(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
413
+ def var_pop(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
414
+ def var_samp(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
415
+ def variance(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
416
+ def list(self, column: str, groups: str = ..., window_spec: str = ..., projected_columns: str = ...) -> DuckDBPyRelation: ...
417
+
649
418
  def arrow(self, batch_size: int = ...) -> pyarrow.lib.RecordBatchReader: ...
650
419
  def __arrow_c_stream__(self, requested_schema: Optional[object] = None) -> object: ...
651
420
  def create(self, table_name: str) -> None: ...
@@ -655,7 +424,7 @@ class DuckDBPyRelation:
655
424
  def distinct(self) -> DuckDBPyRelation: ...
656
425
  def except_(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
657
426
  def execute(self, *args, **kwargs) -> DuckDBPyRelation: ...
658
- def explain(self, type: Optional[Union[Literal["standard", "analyze"], int]] = "standard") -> str: ...
427
+ def explain(self, type: Optional[Literal['standard', 'analyze'] | int] = 'standard') -> str: ...
659
428
  def fetchall(self) -> List[Any]: ...
660
429
  def fetchmany(self, size: int = ...) -> List[Any]: ...
661
430
  def fetchnumpy(self) -> dict: ...
@@ -668,9 +437,7 @@ class DuckDBPyRelation:
668
437
  def update(self, set: Dict[str, Expression], condition: Optional[Expression] = None) -> None: ...
669
438
  def insert_into(self, table_name: str) -> None: ...
670
439
  def intersect(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
671
- def join(
672
- self, other_rel: DuckDBPyRelation, condition: Union[str, Expression], how: str = ...
673
- ) -> DuckDBPyRelation: ...
440
+ def join(self, other_rel: DuckDBPyRelation, condition: Union[str, Expression], how: str = ...) -> DuckDBPyRelation: ...
674
441
  def cross(self, other_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
675
442
  def limit(self, n: int, offset: int = ...) -> DuckDBPyRelation: ...
676
443
  def map(self, map_function: function, schema: Optional[Dict[str, DuckDBPyType]] = None) -> DuckDBPyRelation: ...
@@ -681,55 +448,46 @@ class DuckDBPyRelation:
681
448
  def pl(self, rows_per_batch: int = ..., connection: DuckDBPyConnection = ...) -> polars.DataFrame: ...
682
449
  def query(self, virtual_table_name: str, sql_query: str) -> DuckDBPyRelation: ...
683
450
  def record_batch(self, batch_size: int = ...) -> pyarrow.lib.RecordBatchReader: ...
684
- def fetch_record_batch(
685
- self, rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...
686
- ) -> pyarrow.lib.RecordBatchReader: ...
451
+ def fetch_record_batch(self, rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
687
452
  def select_types(self, types: List[Union[str, DuckDBPyType]]) -> DuckDBPyRelation: ...
688
453
  def select_dtypes(self, types: List[Union[str, DuckDBPyType]]) -> DuckDBPyRelation: ...
689
454
  def set_alias(self, alias: str) -> DuckDBPyRelation: ...
690
- def show(
691
- self,
692
- max_width: Optional[int] = None,
693
- max_rows: Optional[int] = None,
694
- max_col_width: Optional[int] = None,
695
- null_value: Optional[str] = None,
696
- render_mode: Optional[RenderMode] = None,
697
- ) -> None: ...
455
+ def show(self, max_width: Optional[int] = None, max_rows: Optional[int] = None, max_col_width: Optional[int] = None, null_value: Optional[str] = None, render_mode: Optional[RenderMode] = None) -> None: ...
698
456
  def sql_query(self) -> str: ...
699
457
  def to_arrow_table(self, batch_size: int = ...) -> pyarrow.lib.Table: ...
700
458
  def to_csv(
701
- self,
702
- file_name: str,
703
- sep: Optional[str] = None,
704
- na_rep: Optional[str] = None,
705
- header: Optional[bool] = None,
706
- quotechar: Optional[str] = None,
707
- escapechar: Optional[str] = None,
708
- date_format: Optional[str] = None,
709
- timestamp_format: Optional[str] = None,
710
- quoting: Optional[Union[str, int]] = None,
711
- encoding: Optional[str] = None,
712
- compression: Optional[str] = None,
713
- write_partition_columns: Optional[bool] = None,
714
- overwrite: Optional[bool] = None,
715
- per_thread_output: Optional[bool] = None,
716
- use_tmp_file: Optional[bool] = None,
717
- partition_by: Optional[List[str]] = None,
459
+ self,
460
+ file_name: str,
461
+ sep: Optional[str] = None,
462
+ na_rep: Optional[str] = None,
463
+ header: Optional[bool] = None,
464
+ quotechar: Optional[str] = None,
465
+ escapechar: Optional[str] = None,
466
+ date_format: Optional[str] = None,
467
+ timestamp_format: Optional[str] = None,
468
+ quoting: Optional[str | int] = None,
469
+ encoding: Optional[str] = None,
470
+ compression: Optional[str] = None,
471
+ write_partition_columns: Optional[bool] = None,
472
+ overwrite: Optional[bool] = None,
473
+ per_thread_output: Optional[bool] = None,
474
+ use_tmp_file: Optional[bool] = None,
475
+ partition_by: Optional[List[str]] = None
718
476
  ) -> None: ...
719
477
  def to_df(self, *args, **kwargs) -> pandas.DataFrame: ...
720
478
  def to_parquet(
721
- self,
722
- file_name: str,
723
- compression: Optional[str] = None,
724
- field_ids: Optional[Union[dict, str]] = None,
725
- row_group_size_bytes: Optional[Union[int, str]] = None,
726
- row_group_size: Optional[int] = None,
727
- partition_by: Optional[List[str]] = None,
728
- write_partition_columns: Optional[bool] = None,
729
- overwrite: Optional[bool] = None,
730
- per_thread_output: Optional[bool] = None,
731
- use_tmp_file: Optional[bool] = None,
732
- append: Optional[bool] = None,
479
+ self,
480
+ file_name: str,
481
+ compression: Optional[str] = None,
482
+ field_ids: Optional[dict | str] = None,
483
+ row_group_size_bytes: Optional[int | str] = None,
484
+ row_group_size: Optional[int] = None,
485
+ partition_by: Optional[List[str]] = None,
486
+ write_partition_columns: Optional[bool] = None,
487
+ overwrite: Optional[bool] = None,
488
+ per_thread_output: Optional[bool] = None,
489
+ use_tmp_file: Optional[bool] = None,
490
+ append: Optional[bool] = None
733
491
  ) -> None: ...
734
492
  def fetch_df_chunk(self, vectors_per_chunk: int = 1, *, date_as_object: bool = False) -> pandas.DataFrame: ...
735
493
  def to_table(self, table_name: str) -> None: ...
@@ -739,37 +497,37 @@ class DuckDBPyRelation:
739
497
  def union(self, union_rel: DuckDBPyRelation) -> DuckDBPyRelation: ...
740
498
  def unique(self, unique_aggr: str) -> DuckDBPyRelation: ...
741
499
  def write_csv(
742
- self,
743
- file_name: str,
744
- sep: Optional[str] = None,
745
- na_rep: Optional[str] = None,
746
- header: Optional[bool] = None,
747
- quotechar: Optional[str] = None,
748
- escapechar: Optional[str] = None,
749
- date_format: Optional[str] = None,
750
- timestamp_format: Optional[str] = None,
751
- quoting: Optional[Union[str, int]] = None,
752
- encoding: Optional[str] = None,
753
- compression: Optional[str] = None,
754
- write_partition_columns: Optional[bool] = None,
755
- overwrite: Optional[bool] = None,
756
- per_thread_output: Optional[bool] = None,
757
- use_tmp_file: Optional[bool] = None,
758
- partition_by: Optional[List[str]] = None,
500
+ self,
501
+ file_name: str,
502
+ sep: Optional[str] = None,
503
+ na_rep: Optional[str] = None,
504
+ header: Optional[bool] = None,
505
+ quotechar: Optional[str] = None,
506
+ escapechar: Optional[str] = None,
507
+ date_format: Optional[str] = None,
508
+ timestamp_format: Optional[str] = None,
509
+ quoting: Optional[str | int] = None,
510
+ encoding: Optional[str] = None,
511
+ compression: Optional[str] = None,
512
+ write_partition_columns: Optional[bool] = None,
513
+ overwrite: Optional[bool] = None,
514
+ per_thread_output: Optional[bool] = None,
515
+ use_tmp_file: Optional[bool] = None,
516
+ partition_by: Optional[List[str]] = None
759
517
  ) -> None: ...
760
518
  def write_parquet(
761
- self,
762
- file_name: str,
763
- compression: Optional[str] = None,
764
- field_ids: Optional[Union[dict, str]] = None,
765
- row_group_size_bytes: Optional[Union[int, str]] = None,
766
- row_group_size: Optional[int] = None,
767
- partition_by: Optional[List[str]] = None,
768
- write_partition_columns: Optional[bool] = None,
769
- overwrite: Optional[bool] = None,
770
- per_thread_output: Optional[bool] = None,
771
- use_tmp_file: Optional[bool] = None,
772
- append: Optional[bool] = None,
519
+ self,
520
+ file_name: str,
521
+ compression: Optional[str] = None,
522
+ field_ids: Optional[dict | str] = None,
523
+ row_group_size_bytes: Optional[int | str] = None,
524
+ row_group_size: Optional[int] = None,
525
+ partition_by: Optional[List[str]] = None,
526
+ write_partition_columns: Optional[bool] = None,
527
+ overwrite: Optional[bool] = None,
528
+ per_thread_output: Optional[bool] = None,
529
+ use_tmp_file: Optional[bool] = None,
530
+ append: Optional[bool] = None
773
531
  ) -> None: ...
774
532
  def __len__(self) -> int: ...
775
533
  @property
@@ -788,6 +546,7 @@ class DuckDBPyRelation:
788
546
  def types(self) -> List[DuckDBPyType]: ...
789
547
 
790
548
  class Error(Exception): ...
549
+
791
550
  class FatalException(Error): ...
792
551
 
793
552
  class HTTPException(IOException):
@@ -797,31 +556,51 @@ class HTTPException(IOException):
797
556
  headers: Dict[str, str]
798
557
 
799
558
  class IOException(OperationalError): ...
559
+
800
560
  class IntegrityError(Error): ...
561
+
801
562
  class InternalError(Error): ...
563
+
802
564
  class InternalException(InternalError): ...
565
+
803
566
  class InterruptException(Error): ...
567
+
804
568
  class InvalidInputException(ProgrammingError): ...
569
+
805
570
  class InvalidTypeException(ProgrammingError): ...
571
+
806
572
  class NotImplementedException(NotSupportedError): ...
573
+
807
574
  class NotSupportedError(Error): ...
575
+
808
576
  class OperationalError(Error): ...
577
+
809
578
  class OutOfMemoryException(OperationalError): ...
579
+
810
580
  class OutOfRangeException(DataError): ...
581
+
811
582
  class ParserException(ProgrammingError): ...
583
+
812
584
  class PermissionException(Error): ...
585
+
813
586
  class ProgrammingError(Error): ...
587
+
814
588
  class SequenceException(Error): ...
589
+
815
590
  class SerializationException(OperationalError): ...
591
+
816
592
  class SyntaxException(ProgrammingError): ...
593
+
817
594
  class TransactionException(OperationalError): ...
595
+
818
596
  class TypeMismatchException(DataError): ...
597
+
819
598
  class Warning(Exception): ...
820
599
 
821
600
  class token_type:
822
601
  # stubgen override - these make mypy sad
823
- # __doc__: ClassVar[str] = ... # read-only
824
- # __members__: ClassVar[dict] = ... # read-only
602
+ #__doc__: ClassVar[str] = ... # read-only
603
+ #__members__: ClassVar[dict] = ... # read-only
825
604
  __entries: ClassVar[dict] = ...
826
605
  comment: ClassVar[token_type] = ...
827
606
  identifier: ClassVar[token_type] = ...
@@ -861,18 +640,7 @@ def register_filesystem(filesystem: fsspec.AbstractFileSystem, *, connection: Du
861
640
  def unregister_filesystem(name: str, *, connection: DuckDBPyConnection = ...) -> None: ...
862
641
  def list_filesystems(*, connection: DuckDBPyConnection = ...) -> list: ...
863
642
  def filesystem_is_registered(name: str, *, connection: DuckDBPyConnection = ...) -> bool: ...
864
- def create_function(
865
- name: str,
866
- function: function,
867
- parameters: Optional[List[DuckDBPyType]] = None,
868
- return_type: Optional[DuckDBPyType] = None,
869
- *,
870
- type: Optional[PythonUDFType] = PythonUDFType.NATIVE,
871
- null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT,
872
- exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT,
873
- side_effects: bool = False,
874
- connection: DuckDBPyConnection = ...,
875
- ) -> DuckDBPyConnection: ...
643
+ def create_function(name: str, function: function, parameters: Optional[List[DuckDBPyType]] = None, return_type: Optional[DuckDBPyType] = None, *, type: Optional[PythonUDFType] = PythonUDFType.NATIVE, null_handling: Optional[FunctionNullHandling] = FunctionNullHandling.DEFAULT, exception_handling: Optional[PythonExceptionHandling] = PythonExceptionHandling.DEFAULT, side_effects: bool = False, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
876
644
  def remove_function(name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
877
645
  def sqltype(type_str: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
878
646
  def dtype(type_str: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
@@ -881,24 +649,14 @@ def array_type(type: DuckDBPyType, size: int, *, connection: DuckDBPyConnection
881
649
  def list_type(type: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
882
650
  def union_type(members: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
883
651
  def string_type(collation: str = "", *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
884
- def enum_type(
885
- name: str, type: DuckDBPyType, values: List[Any], *, connection: DuckDBPyConnection = ...
886
- ) -> DuckDBPyType: ...
652
+ def enum_type(name: str, type: DuckDBPyType, values: List[Any], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
887
653
  def decimal_type(width: int, scale: int, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
888
- def struct_type(
889
- fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...
890
- ) -> DuckDBPyType: ...
891
- def row_type(
892
- fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...
893
- ) -> DuckDBPyType: ...
654
+ def struct_type(fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
655
+ def row_type(fields: Union[Dict[str, DuckDBPyType], List[str]], *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
894
656
  def map_type(key: DuckDBPyType, value: DuckDBPyType, *, connection: DuckDBPyConnection = ...) -> DuckDBPyType: ...
895
657
  def duplicate(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
896
- def execute(
897
- query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...
898
- ) -> DuckDBPyConnection: ...
899
- def executemany(
900
- query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...
901
- ) -> DuckDBPyConnection: ...
658
+ def execute(query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
659
+ def executemany(query: object, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
902
660
  def close(*, connection: DuckDBPyConnection = ...) -> None: ...
903
661
  def interrupt(*, connection: DuckDBPyConnection = ...) -> None: ...
904
662
  def query_progress(*, connection: DuckDBPyConnection = ...) -> float: ...
@@ -909,16 +667,10 @@ def fetchnumpy(*, connection: DuckDBPyConnection = ...) -> dict: ...
909
667
  def fetchdf(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
910
668
  def fetch_df(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
911
669
  def df(*, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
912
- def fetch_df_chunk(
913
- vectors_per_chunk: int = 1, *, date_as_object: bool = False, connection: DuckDBPyConnection = ...
914
- ) -> pandas.DataFrame: ...
915
- def pl(
916
- rows_per_batch: int = 1000000, *, lazy: bool = False, connection: DuckDBPyConnection = ...
917
- ) -> polars.DataFrame: ...
670
+ def fetch_df_chunk(vectors_per_chunk: int = 1, *, date_as_object: bool = False, connection: DuckDBPyConnection = ...) -> pandas.DataFrame: ...
671
+ def pl(rows_per_batch: int = 1000000, *, lazy: bool = False, connection: DuckDBPyConnection = ...) -> polars.DataFrame: ...
918
672
  def fetch_arrow_table(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.Table: ...
919
- def fetch_record_batch(
920
- rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...
921
- ) -> pyarrow.lib.RecordBatchReader: ...
673
+ def fetch_record_batch(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
922
674
  def arrow(rows_per_batch: int = 1000000, *, connection: DuckDBPyConnection = ...) -> pyarrow.lib.RecordBatchReader: ...
923
675
  def torch(*, connection: DuckDBPyConnection = ...) -> dict: ...
924
676
  def tf(*, connection: DuckDBPyConnection = ...) -> dict: ...
@@ -926,212 +678,36 @@ def begin(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
926
678
  def commit(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
927
679
  def rollback(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
928
680
  def checkpoint(*, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
929
- def append(
930
- table_name: str, df: pandas.DataFrame, *, by_name: bool = False, connection: DuckDBPyConnection = ...
931
- ) -> DuckDBPyConnection: ...
681
+ def append(table_name: str, df: pandas.DataFrame, *, by_name: bool = False, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
932
682
  def register(view_name: str, python_object: object, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
933
683
  def unregister(view_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyConnection: ...
934
684
  def table(table_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
935
685
  def view(view_name: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
936
- def values(
937
- *args: Union[List[Any], Expression, Tuple[Expression]], connection: DuckDBPyConnection = ...
938
- ) -> DuckDBPyRelation: ...
939
- def table_function(
940
- name: str, parameters: object = None, *, connection: DuckDBPyConnection = ...
941
- ) -> DuckDBPyRelation: ...
942
- def read_json(
943
- path_or_buffer: Union[str, StringIO, TextIOBase],
944
- *,
945
- columns: Optional[Dict[str, str]] = None,
946
- sample_size: Optional[int] = None,
947
- maximum_depth: Optional[int] = None,
948
- records: Optional[str] = None,
949
- format: Optional[str] = None,
950
- date_format: Optional[str] = None,
951
- timestamp_format: Optional[str] = None,
952
- compression: Optional[str] = None,
953
- maximum_object_size: Optional[int] = None,
954
- ignore_errors: Optional[bool] = None,
955
- convert_strings_to_integers: Optional[bool] = None,
956
- field_appearance_threshold: Optional[float] = None,
957
- map_inference_threshold: Optional[int] = None,
958
- maximum_sample_files: Optional[int] = None,
959
- filename: Optional[Union[bool, str]] = None,
960
- hive_partitioning: Optional[bool] = None,
961
- union_by_name: Optional[bool] = None,
962
- hive_types: Optional[Dict[str, str]] = None,
963
- hive_types_autocast: Optional[bool] = None,
964
- connection: DuckDBPyConnection = ...,
965
- ) -> DuckDBPyRelation: ...
686
+ def values(*args: Union[List[Any],Expression, Tuple[Expression]], connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
687
+ def table_function(name: str, parameters: object = None, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
688
+ def read_json(path_or_buffer: Union[str, StringIO, TextIOBase], *, columns: Optional[Dict[str,str]] = None, sample_size: Optional[int] = None, maximum_depth: Optional[int] = None, records: Optional[str] = None, format: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, compression: Optional[str] = None, maximum_object_size: Optional[int] = None, ignore_errors: Optional[bool] = None, convert_strings_to_integers: Optional[bool] = None, field_appearance_threshold: Optional[float] = None, map_inference_threshold: Optional[int] = None, maximum_sample_files: Optional[int] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
966
689
  def extract_statements(query: str, *, connection: DuckDBPyConnection = ...) -> List[Statement]: ...
967
- def sql(
968
- query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...
969
- ) -> DuckDBPyRelation: ...
970
- def query(
971
- query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...
972
- ) -> DuckDBPyRelation: ...
973
- def from_query(
974
- query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...
975
- ) -> DuckDBPyRelation: ...
976
- def read_csv(
977
- path_or_buffer: Union[str, StringIO, TextIOBase],
978
- *,
979
- header: Optional[Union[bool, int]] = None,
980
- compression: Optional[str] = None,
981
- sep: Optional[str] = None,
982
- delimiter: Optional[str] = None,
983
- dtype: Optional[Union[Dict[str, str], List[str]]] = None,
984
- na_values: Optional[Union[str, List[str]]] = None,
985
- skiprows: Optional[int] = None,
986
- quotechar: Optional[str] = None,
987
- escapechar: Optional[str] = None,
988
- encoding: Optional[str] = None,
989
- parallel: Optional[bool] = None,
990
- date_format: Optional[str] = None,
991
- timestamp_format: Optional[str] = None,
992
- sample_size: Optional[int] = None,
993
- all_varchar: Optional[bool] = None,
994
- normalize_names: Optional[bool] = None,
995
- null_padding: Optional[bool] = None,
996
- names: Optional[List[str]] = None,
997
- lineterminator: Optional[str] = None,
998
- columns: Optional[Dict[str, str]] = None,
999
- auto_type_candidates: Optional[List[str]] = None,
1000
- max_line_size: Optional[int] = None,
1001
- ignore_errors: Optional[bool] = None,
1002
- store_rejects: Optional[bool] = None,
1003
- rejects_table: Optional[str] = None,
1004
- rejects_scan: Optional[str] = None,
1005
- rejects_limit: Optional[int] = None,
1006
- force_not_null: Optional[List[str]] = None,
1007
- buffer_size: Optional[int] = None,
1008
- decimal: Optional[str] = None,
1009
- allow_quoted_nulls: Optional[bool] = None,
1010
- filename: Optional[Union[bool, str]] = None,
1011
- hive_partitioning: Optional[bool] = None,
1012
- union_by_name: Optional[bool] = None,
1013
- hive_types: Optional[Dict[str, str]] = None,
1014
- hive_types_autocast: Optional[bool] = None,
1015
- connection: DuckDBPyConnection = ...,
1016
- ) -> DuckDBPyRelation: ...
1017
- def from_csv_auto(
1018
- path_or_buffer: Union[str, StringIO, TextIOBase],
1019
- *,
1020
- header: Optional[Union[bool, int]] = None,
1021
- compression: Optional[str] = None,
1022
- sep: Optional[str] = None,
1023
- delimiter: Optional[str] = None,
1024
- dtype: Optional[Union[Dict[str, str], List[str]]] = None,
1025
- na_values: Optional[Union[str, List[str]]] = None,
1026
- skiprows: Optional[int] = None,
1027
- quotechar: Optional[str] = None,
1028
- escapechar: Optional[str] = None,
1029
- encoding: Optional[str] = None,
1030
- parallel: Optional[bool] = None,
1031
- date_format: Optional[str] = None,
1032
- timestamp_format: Optional[str] = None,
1033
- sample_size: Optional[int] = None,
1034
- all_varchar: Optional[bool] = None,
1035
- normalize_names: Optional[bool] = None,
1036
- null_padding: Optional[bool] = None,
1037
- names: Optional[List[str]] = None,
1038
- lineterminator: Optional[str] = None,
1039
- columns: Optional[Dict[str, str]] = None,
1040
- auto_type_candidates: Optional[List[str]] = None,
1041
- max_line_size: Optional[int] = None,
1042
- ignore_errors: Optional[bool] = None,
1043
- store_rejects: Optional[bool] = None,
1044
- rejects_table: Optional[str] = None,
1045
- rejects_scan: Optional[str] = None,
1046
- rejects_limit: Optional[int] = None,
1047
- force_not_null: Optional[List[str]] = None,
1048
- buffer_size: Optional[int] = None,
1049
- decimal: Optional[str] = None,
1050
- allow_quoted_nulls: Optional[bool] = None,
1051
- filename: Optional[Union[bool, str]] = None,
1052
- hive_partitioning: Optional[bool] = None,
1053
- union_by_name: Optional[bool] = None,
1054
- hive_types: Optional[Dict[str, str]] = None,
1055
- hive_types_autocast: Optional[bool] = None,
1056
- connection: DuckDBPyConnection = ...,
1057
- ) -> DuckDBPyRelation: ...
690
+ def sql(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
691
+ def query(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
692
+ def from_query(query: str, *, alias: str = "", params: object = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
693
+ def read_csv(path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
694
+ def from_csv_auto(path_or_buffer: Union[str, StringIO, TextIOBase], *, header: Optional[bool | int] = None, compression: Optional[str] = None, sep: Optional[str] = None, delimiter: Optional[str] = None, dtype: Optional[Dict[str, str] | List[str]] = None, na_values: Optional[str| List[str]] = None, skiprows: Optional[int] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, encoding: Optional[str] = None, parallel: Optional[bool] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, sample_size: Optional[int] = None, all_varchar: Optional[bool] = None, normalize_names: Optional[bool] = None, null_padding: Optional[bool] = None, names: Optional[List[str]] = None, lineterminator: Optional[str] = None, columns: Optional[Dict[str, str]] = None, auto_type_candidates: Optional[List[str]] = None, max_line_size: Optional[int] = None, ignore_errors: Optional[bool] = None, store_rejects: Optional[bool] = None, rejects_table: Optional[str] = None, rejects_scan: Optional[str] = None, rejects_limit: Optional[int] = None, force_not_null: Optional[List[str]] = None, buffer_size: Optional[int] = None, decimal: Optional[str] = None, allow_quoted_nulls: Optional[bool] = None, filename: Optional[bool | str] = None, hive_partitioning: Optional[bool] = None, union_by_name: Optional[bool] = None, hive_types: Optional[Dict[str, str]] = None, hive_types_autocast: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1058
695
  def from_df(df: pandas.DataFrame, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1059
696
  def from_arrow(arrow_object: object, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1060
- def from_parquet(
1061
- file_glob: str,
1062
- binary_as_string: bool = False,
1063
- *,
1064
- file_row_number: bool = False,
1065
- filename: bool = False,
1066
- hive_partitioning: bool = False,
1067
- union_by_name: bool = False,
1068
- compression: Optional[str] = None,
1069
- connection: DuckDBPyConnection = ...,
1070
- ) -> DuckDBPyRelation: ...
1071
- def read_parquet(
1072
- file_glob: str,
1073
- binary_as_string: bool = False,
1074
- *,
1075
- file_row_number: bool = False,
1076
- filename: bool = False,
1077
- hive_partitioning: bool = False,
1078
- union_by_name: bool = False,
1079
- compression: Optional[str] = None,
1080
- connection: DuckDBPyConnection = ...,
1081
- ) -> DuckDBPyRelation: ...
697
+ def from_parquet(file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
698
+ def read_parquet(file_glob: str, binary_as_string: bool = False, *, file_row_number: bool = False, filename: bool = False, hive_partitioning: bool = False, union_by_name: bool = False, compression: Optional[str] = None, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1082
699
  def get_table_names(query: str, *, qualified: bool = False, connection: DuckDBPyConnection = ...) -> Set[str]: ...
1083
- def install_extension(
1084
- extension: str,
1085
- *,
1086
- force_install: bool = False,
1087
- repository: Optional[str] = None,
1088
- repository_url: Optional[str] = None,
1089
- version: Optional[str] = None,
1090
- connection: DuckDBPyConnection = ...,
1091
- ) -> None: ...
700
+ def install_extension(extension: str, *, force_install: bool = False, repository: Optional[str] = None, repository_url: Optional[str] = None, version: Optional[str] = None, connection: DuckDBPyConnection = ...) -> None: ...
1092
701
  def load_extension(extension: str, *, connection: DuckDBPyConnection = ...) -> None: ...
1093
- def project(
1094
- df: pandas.DataFrame, *args: str, groups: str = "", connection: DuckDBPyConnection = ...
1095
- ) -> DuckDBPyRelation: ...
702
+ def project(df: pandas.DataFrame, *args: str, groups: str = "", connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1096
703
  def distinct(df: pandas.DataFrame, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1097
- def write_csv(
1098
- df: pandas.DataFrame,
1099
- filename: str,
1100
- *,
1101
- sep: Optional[str] = None,
1102
- na_rep: Optional[str] = None,
1103
- header: Optional[bool] = None,
1104
- quotechar: Optional[str] = None,
1105
- escapechar: Optional[str] = None,
1106
- date_format: Optional[str] = None,
1107
- timestamp_format: Optional[str] = None,
1108
- quoting: Optional[Union[str, int]] = None,
1109
- encoding: Optional[str] = None,
1110
- compression: Optional[str] = None,
1111
- overwrite: Optional[bool] = None,
1112
- per_thread_output: Optional[bool] = None,
1113
- use_tmp_file: Optional[bool] = None,
1114
- partition_by: Optional[List[str]] = None,
1115
- write_partition_columns: Optional[bool] = None,
1116
- connection: DuckDBPyConnection = ...,
1117
- ) -> None: ...
1118
- def aggregate(
1119
- df: pandas.DataFrame,
1120
- aggr_expr: Union[str, List[Expression]],
1121
- group_expr: str = "",
1122
- *,
1123
- connection: DuckDBPyConnection = ...,
1124
- ) -> DuckDBPyRelation: ...
704
+ def write_csv(df: pandas.DataFrame, filename: str, *, sep: Optional[str] = None, na_rep: Optional[str] = None, header: Optional[bool] = None, quotechar: Optional[str] = None, escapechar: Optional[str] = None, date_format: Optional[str] = None, timestamp_format: Optional[str] = None, quoting: Optional[str | int] = None, encoding: Optional[str] = None, compression: Optional[str] = None, overwrite: Optional[bool] = None, per_thread_output: Optional[bool] = None, use_tmp_file: Optional[bool] = None, partition_by: Optional[List[str]] = None, write_partition_columns: Optional[bool] = None, connection: DuckDBPyConnection = ...) -> None: ...
705
+ def aggregate(df: pandas.DataFrame, aggr_expr: str | List[Expression], group_expr: str = "", *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1125
706
  def alias(df: pandas.DataFrame, alias: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1126
707
  def filter(df: pandas.DataFrame, filter_expr: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1127
- def limit(
1128
- df: pandas.DataFrame, n: int, offset: int = 0, *, connection: DuckDBPyConnection = ...
1129
- ) -> DuckDBPyRelation: ...
708
+ def limit(df: pandas.DataFrame, n: int, offset: int = 0, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1130
709
  def order(df: pandas.DataFrame, order_expr: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1131
- def query_df(
1132
- df: pandas.DataFrame, virtual_table_name: str, sql_query: str, *, connection: DuckDBPyConnection = ...
1133
- ) -> DuckDBPyRelation: ...
710
+ def query_df(df: pandas.DataFrame, virtual_table_name: str, sql_query: str, *, connection: DuckDBPyConnection = ...) -> DuckDBPyRelation: ...
1134
711
  def description(*, connection: DuckDBPyConnection = ...) -> Optional[List[Any]]: ...
1135
712
  def rowcount(*, connection: DuckDBPyConnection = ...) -> int: ...
1136
-
1137
713
  # END OF CONNECTION WRAPPER