etlplus 0.12.9__py3-none-any.whl → 0.12.11__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (59) hide show
  1. etlplus/file/accdb.py +78 -0
  2. etlplus/file/arrow.py +78 -0
  3. etlplus/file/avro.py +13 -1
  4. etlplus/file/bson.py +77 -0
  5. etlplus/file/cbor.py +78 -0
  6. etlplus/file/cfg.py +79 -0
  7. etlplus/file/conf.py +80 -0
  8. etlplus/file/csv.py +13 -1
  9. etlplus/file/dat.py +13 -1
  10. etlplus/file/dta.py +77 -0
  11. etlplus/file/duckdb.py +78 -0
  12. etlplus/file/enums.py +10 -4
  13. etlplus/file/feather.py +13 -1
  14. etlplus/file/fwf.py +12 -1
  15. etlplus/file/hbs.py +78 -0
  16. etlplus/file/hdf5.py +78 -0
  17. etlplus/file/ini.py +79 -0
  18. etlplus/file/ion.py +78 -0
  19. etlplus/file/jinja2.py +78 -0
  20. etlplus/file/json.py +13 -1
  21. etlplus/file/log.py +78 -0
  22. etlplus/file/mat.py +78 -0
  23. etlplus/file/mdb.py +78 -0
  24. etlplus/file/msgpack.py +78 -0
  25. etlplus/file/mustache.py +78 -0
  26. etlplus/file/nc.py +78 -0
  27. etlplus/file/ndjson.py +12 -6
  28. etlplus/file/numbers.py +75 -0
  29. etlplus/file/ods.py +79 -0
  30. etlplus/file/orc.py +13 -1
  31. etlplus/file/parquet.py +13 -1
  32. etlplus/file/pb.py +78 -0
  33. etlplus/file/pbf.py +77 -0
  34. etlplus/file/properties.py +78 -0
  35. etlplus/file/proto.py +77 -0
  36. etlplus/file/psv.py +14 -1
  37. etlplus/file/rda.py +78 -0
  38. etlplus/file/rds.py +78 -0
  39. etlplus/file/sas7bdat.py +78 -0
  40. etlplus/file/sav.py +77 -0
  41. etlplus/file/sqlite.py +78 -0
  42. etlplus/file/sylk.py +77 -0
  43. etlplus/file/tab.py +3 -4
  44. etlplus/file/toml.py +78 -0
  45. etlplus/file/tsv.py +14 -1
  46. etlplus/file/txt.py +11 -1
  47. etlplus/file/vm.py +78 -0
  48. etlplus/file/wks.py +77 -0
  49. etlplus/file/xlsm.py +79 -0
  50. etlplus/file/xml.py +12 -1
  51. etlplus/file/xpt.py +78 -0
  52. etlplus/file/yaml.py +12 -1
  53. etlplus/file/zsav.py +77 -0
  54. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/METADATA +31 -1
  55. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/RECORD +59 -22
  56. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/WHEEL +0 -0
  57. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/entry_points.txt +0 -0
  58. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/licenses/LICENSE +0 -0
  59. {etlplus-0.12.9.dist-info → etlplus-0.12.11.dist-info}/top_level.txt +0 -0
etlplus/file/accdb.py ADDED
@@ -0,0 +1,78 @@
1
+ """
2
+ :mod:`etlplus.file.accdb` module.
3
+
4
+ Helpers for reading/writing newer Microsoft Access database (ACCDB) files.
5
+
6
+ Notes
7
+ -----
8
+ - An ACCDB file is a proprietary database file format used by Microsoft Access
9
+ 2007 and later.
10
+ - Common cases:
11
+ - Storing relational data for small to medium-sized applications.
12
+ - Desktop database applications.
13
+ - Data management for non-enterprise solutions.
14
+ - Rule of thumb:
15
+ - If the file follows the ACCDB specification, use this module for reading
16
+ and writing.
17
+ """
18
+
19
+ from __future__ import annotations
20
+
21
+ from pathlib import Path
22
+
23
+ from ..types import JSONData
24
+ from ..types import JSONList
25
+ from . import stub
26
+
27
+ # SECTION: EXPORTS ========================================================== #
28
+
29
+
30
+ __all__ = [
31
+ 'read',
32
+ 'write',
33
+ ]
34
+
35
+
36
+ # SECTION: FUNCTIONS ======================================================== #
37
+
38
+
39
+ def read(
40
+ path: Path,
41
+ ) -> JSONList:
42
+ """
43
+ Read ACCDB content from ``path``.
44
+
45
+ Parameters
46
+ ----------
47
+ path : Path
48
+ Path to the ACCDB file on disk.
49
+
50
+ Returns
51
+ -------
52
+ JSONList
53
+ The list of dictionaries read from the ACCDB file.
54
+ """
55
+ return stub.read(path, format_name='ACCDB')
56
+
57
+
58
+ def write(
59
+ path: Path,
60
+ data: JSONData,
61
+ ) -> int:
62
+ """
63
+ Write ``data`` to ACCDB at ``path`` and return record count.
64
+
65
+ Parameters
66
+ ----------
67
+ path : Path
68
+ Path to the ACCDB file on disk.
69
+ data : JSONData
70
+ Data to write as ACCDB. Should be a list of dictionaries or a single
71
+ dictionary.
72
+
73
+ Returns
74
+ -------
75
+ int
76
+ The number of rows written to the ACCDB file.
77
+ """
78
+ return stub.write(path, data, format_name='ACCDB')
etlplus/file/arrow.py ADDED
@@ -0,0 +1,78 @@
1
+ """
2
+ :mod:`etlplus.file.arrow` module.
3
+
4
+ Helpers for reading/writing Apache Arrow (ARROW) files.
5
+
6
+ Notes
7
+ -----
8
+ - An ARROW file is a binary file format designed for efficient
9
+ columnar data storage and processing.
10
+ - Common cases:
11
+ - High-performance data analytics.
12
+ - Interoperability between different data processing systems.
13
+ - In-memory data representation for fast computations.
14
+ - Rule of thumb:
15
+ - If the file follows the Apache Arrow specification, use this module for
16
+ reading and writing.
17
+ """
18
+
19
+ from __future__ import annotations
20
+
21
+ from pathlib import Path
22
+
23
+ from ..types import JSONData
24
+ from ..types import JSONList
25
+ from . import stub
26
+
27
+ # SECTION: EXPORTS ========================================================== #
28
+
29
+
30
+ __all__ = [
31
+ 'read',
32
+ 'write',
33
+ ]
34
+
35
+
36
+ # SECTION: FUNCTIONS ======================================================== #
37
+
38
+
39
+ def read(
40
+ path: Path,
41
+ ) -> JSONList:
42
+ """
43
+ Read ARROW content from ``path``.
44
+
45
+ Parameters
46
+ ----------
47
+ path : Path
48
+ Path to the Apache Arrow file on disk.
49
+
50
+ Returns
51
+ -------
52
+ JSONList
53
+ The list of dictionaries read from the Apache Arrow file.
54
+ """
55
+ return stub.read(path, format_name='ARROW')
56
+
57
+
58
+ def write(
59
+ path: Path,
60
+ data: JSONData,
61
+ ) -> int:
62
+ """
63
+ Write ``data`` to ARROW at ``path`` and return record count.
64
+
65
+ Parameters
66
+ ----------
67
+ path : Path
68
+ Path to the ARROW file on disk.
69
+ data : JSONData
70
+ Data to write as ARROW. Should be a list of dictionaries or a
71
+ single dictionary.
72
+
73
+ Returns
74
+ -------
75
+ int
76
+ The number of rows written to the ARROW file.
77
+ """
78
+ return stub.write(path, data, format_name='ARROW')
etlplus/file/avro.py CHANGED
@@ -1,7 +1,19 @@
1
1
  """
2
2
  :mod:`etlplus.file.avro` module.
3
3
 
4
- Helpers for reading/writing Avro files.
4
+ Helpers for reading/writing Apache Avro (AVRO) files.
5
+
6
+ Notes
7
+ -----
8
+ - An AVRO file is a binary file format designed for efficient
9
+ on-disk storage of data, with a schema definition.
10
+ - Common cases:
11
+ - Data serialization for distributed systems.
12
+ - Interoperability between different programming languages.
13
+ - Storage of large datasets with schema evolution support.
14
+ - Rule of thumb:
15
+ - If the file follows the Apache Avro specification, use this module for
16
+ reading and writing.
5
17
  """
6
18
 
7
19
  from __future__ import annotations
etlplus/file/bson.py ADDED
@@ -0,0 +1,77 @@
1
+ """
2
+ :mod:`etlplus.file.bson` module.
3
+
4
+ Helpers for reading/writing Binary JSON (BSON) files.
5
+
6
+ Notes
7
+ -----
8
+ - A BSON file is a binary-encoded serialization of JSON-like documents.
9
+ - Common cases:
10
+ - Data storage in MongoDB.
11
+ - Efficient data interchange between systems.
12
+ - Handling of complex data types not supported in standard JSON.
13
+ - Rule of thumb:
14
+ - If the file follows the BSON specification, use this module for reading
15
+ and writing.
16
+ """
17
+
18
+ from __future__ import annotations
19
+
20
+ from pathlib import Path
21
+
22
+ from ..types import JSONData
23
+ from ..types import JSONList
24
+ from . import stub
25
+
26
+ # SECTION: EXPORTS ========================================================== #
27
+
28
+
29
+ __all__ = [
30
+ 'read',
31
+ 'write',
32
+ ]
33
+
34
+
35
+ # SECTION: FUNCTIONS ======================================================== #
36
+
37
+
38
+ def read(
39
+ path: Path,
40
+ ) -> JSONList:
41
+ """
42
+ Read BSON content from ``path``.
43
+
44
+ Parameters
45
+ ----------
46
+ path : Path
47
+ Path to the BSON file on disk.
48
+
49
+ Returns
50
+ -------
51
+ JSONList
52
+ The list of dictionaries read from the BSON file.
53
+ """
54
+ return stub.read(path, format_name='BSON')
55
+
56
+
57
+ def write(
58
+ path: Path,
59
+ data: JSONData,
60
+ ) -> int:
61
+ """
62
+ Write ``data`` to BSON at ``path`` and return record count.
63
+
64
+ Parameters
65
+ ----------
66
+ path : Path
67
+ Path to the BSON file on disk.
68
+ data : JSONData
69
+ Data to write as BSON. Should be a list of dictionaries or a
70
+ single dictionary.
71
+
72
+ Returns
73
+ -------
74
+ int
75
+ The number of rows written to the BSON file.
76
+ """
77
+ return stub.write(path, data, format_name='BSON')
etlplus/file/cbor.py ADDED
@@ -0,0 +1,78 @@
1
+ """
2
+ :mod:`etlplus.file.cbor` module.
3
+
4
+ Helpers for reading/writing Concise Binary Object Representation (CBOR) files.
5
+
6
+ Notes
7
+ -----
8
+ - A CBOR file is a binary data format designed for small code size and message
9
+ size, suitable for constrained environments.
10
+ - Common cases:
11
+ - IoT data interchange.
12
+ - Efficient data serialization.
13
+ - Storage of structured data in a compact binary form.
14
+ - Rule of thumb:
15
+ - If the file follows the CBOR specification, use this module for reading
16
+ and writing.
17
+ """
18
+
19
+ from __future__ import annotations
20
+
21
+ from pathlib import Path
22
+
23
+ from ..types import JSONData
24
+ from ..types import JSONList
25
+ from . import stub
26
+
27
+ # SECTION: EXPORTS ========================================================== #
28
+
29
+
30
+ __all__ = [
31
+ 'read',
32
+ 'write',
33
+ ]
34
+
35
+
36
+ # SECTION: FUNCTIONS ======================================================== #
37
+
38
+
39
+ def read(
40
+ path: Path,
41
+ ) -> JSONList:
42
+ """
43
+ Read CBOR content from ``path``.
44
+
45
+ Parameters
46
+ ----------
47
+ path : Path
48
+ Path to the CBOR file on disk.
49
+
50
+ Returns
51
+ -------
52
+ JSONList
53
+ The list of dictionaries read from the CBOR file.
54
+ """
55
+ return stub.read(path, format_name='CBOR')
56
+
57
+
58
+ def write(
59
+ path: Path,
60
+ data: JSONData,
61
+ ) -> int:
62
+ """
63
+ Write ``data`` to CBOR at ``path`` and return record count.
64
+
65
+ Parameters
66
+ ----------
67
+ path : Path
68
+ Path to the CBOR file on disk.
69
+ data : JSONData
70
+ Data to write as CBOR. Should be a list of dictionaries or a
71
+ single dictionary.
72
+
73
+ Returns
74
+ -------
75
+ int
76
+ The number of rows written to the CBOR file.
77
+ """
78
+ return stub.write(path, data, format_name='CBOR')
etlplus/file/cfg.py ADDED
@@ -0,0 +1,79 @@
1
+ """
2
+ :mod:`etlplus.file.cfg` module.
3
+
4
+ Helpers for reading/writing config (CFG) files.
5
+
6
+ Notes
7
+ -----
8
+ - A CFG file is a configuration file that may use various syntaxes, such as
9
+ INI, YAML, or custom formats.
10
+ - Common cases:
11
+ - INI-style key-value pairs with sections (such as in Python ecosystems,
12
+ using ``configparser``).
13
+ - YAML-like structures with indentation.
14
+ - Custom formats specific to certain applications.
15
+ - Rule of thumb:
16
+ - If the file follows a standard format like INI or YAML, consider using
17
+ dedicated parsers.
18
+ """
19
+
20
+ from __future__ import annotations
21
+
22
+ from pathlib import Path
23
+
24
+ from ..types import JSONData
25
+ from ..types import JSONList
26
+ from . import stub
27
+
28
+ # SECTION: EXPORTS ========================================================== #
29
+
30
+
31
+ __all__ = [
32
+ 'read',
33
+ 'write',
34
+ ]
35
+
36
+
37
+ # SECTION: FUNCTIONS ======================================================== #
38
+
39
+
40
+ def read(
41
+ path: Path,
42
+ ) -> JSONList:
43
+ """
44
+ Read CFG content from ``path``.
45
+
46
+ Parameters
47
+ ----------
48
+ path : Path
49
+ Path to the CFG file on disk.
50
+
51
+ Returns
52
+ -------
53
+ JSONList
54
+ The list of dictionaries read from the CFG file.
55
+ """
56
+ return stub.read(path, format_name='CFG')
57
+
58
+
59
+ def write(
60
+ path: Path,
61
+ data: JSONData,
62
+ ) -> int:
63
+ """
64
+ Write ``data`` to CFG file at ``path`` and return record count.
65
+
66
+ Parameters
67
+ ----------
68
+ path : Path
69
+ Path to the CFG file on disk.
70
+ data : JSONData
71
+ Data to write as CFG file. Should be a list of dictionaries or a
72
+ single dictionary.
73
+
74
+ Returns
75
+ -------
76
+ int
77
+ The number of rows written to the CFG file.
78
+ """
79
+ return stub.write(path, data, format_name='CFG')
etlplus/file/conf.py ADDED
@@ -0,0 +1,80 @@
1
+ """
2
+ :mod:`etlplus.file.conf` module.
3
+
4
+ Helpers for reading/writing config (CONF) files.
5
+
6
+ Notes
7
+ -----
8
+ - A CONF file is a configuration file that may use various syntaxes, such as
9
+ INI, YAML, or custom formats.
10
+ - Common cases:
11
+ - INI-style key-value pairs with sections.
12
+ - YAML-like structures with indentation.
13
+ - Custom formats specific to certain applications (such as Unix-like
14
+ systems, where ``.conf`` is a strong convention for "This is a
15
+ configuration file").
16
+ - Rule of thumb:
17
+ - If the file follows a standard format like INI or YAML, consider using
18
+ dedicated parsers.
19
+ """
20
+
21
+ from __future__ import annotations
22
+
23
+ from pathlib import Path
24
+
25
+ from ..types import JSONData
26
+ from ..types import JSONList
27
+ from . import stub
28
+
29
+ # SECTION: EXPORTS ========================================================== #
30
+
31
+
32
+ __all__ = [
33
+ 'read',
34
+ 'write',
35
+ ]
36
+
37
+
38
+ # SECTION: FUNCTIONS ======================================================== #
39
+
40
+
41
+ def read(
42
+ path: Path,
43
+ ) -> JSONList:
44
+ """
45
+ Read CONF content from ``path``.
46
+
47
+ Parameters
48
+ ----------
49
+ path : Path
50
+ Path to the CONF file on disk.
51
+
52
+ Returns
53
+ -------
54
+ JSONList
55
+ The list of dictionaries read from the CONF file.
56
+ """
57
+ return stub.read(path, format_name='CONF')
58
+
59
+
60
+ def write(
61
+ path: Path,
62
+ data: JSONData,
63
+ ) -> int:
64
+ """
65
+ Write ``data`` to CONF at ``path`` and return record count.
66
+
67
+ Parameters
68
+ ----------
69
+ path : Path
70
+ Path to the CONF file on disk.
71
+ data : JSONData
72
+ Data to write as CONF. Should be a list of dictionaries or a
73
+ single dictionary.
74
+
75
+ Returns
76
+ -------
77
+ int
78
+ The number of rows written to the CONF file.
79
+ """
80
+ return stub.write(path, data, format_name='CONF')
etlplus/file/csv.py CHANGED
@@ -1,7 +1,19 @@
1
1
  """
2
2
  :mod:`etlplus.file.csv` module.
3
3
 
4
- Helpers for reading/writing CSV files.
4
+ Helpers for reading/writing Comma-Separated Values (CSV) files.
5
+
6
+ Notes
7
+ -----
8
+ - A CSV file is a plain text file that uses commas to separate values.
9
+ - Common cases:
10
+ - Each line in the file represents a single record.
11
+ - The first line often contains headers that define the column names.
12
+ - Values may be enclosed in quotes, especially if they contain commas
13
+ or special characters.
14
+ - Rule of thumb:
15
+ - If the file follows the CSV specification, use this module for
16
+ reading and writing.
5
17
  """
6
18
 
7
19
  from __future__ import annotations
etlplus/file/dat.py CHANGED
@@ -1,7 +1,19 @@
1
1
  """
2
2
  :mod:`etlplus.file.dat` module.
3
3
 
4
- Helpers for reading/writing DAT (data) files.
4
+ Helpers for reading/writing data (DAT) files.
5
+
6
+ Notes
7
+ -----
8
+ - A “DAT-formatted” file is a generic data file that may use various
9
+ delimiters or fixed-width formats.
10
+ - Common cases:
11
+ - Delimited text files (e.g., CSV, TSV).
12
+ - Fixed-width formatted files.
13
+ - Custom formats specific to certain applications.
14
+ - Rule of thumb:
15
+ - If the file does not follow a specific standard format, use this module
16
+ for reading and writing.
5
17
  """
6
18
 
7
19
  from __future__ import annotations
etlplus/file/dta.py ADDED
@@ -0,0 +1,77 @@
1
+ """
2
+ :mod:`etlplus.file.dta` module.
3
+
4
+ Helpers for reading/writing Stata (DTA) data files.
5
+
6
+ Notes
7
+ -----
8
+ - Stata DTA files are binary files used by Stata statistical software that
9
+ store datasets with variables, labels, and data types.
10
+ - Common cases:
11
+ - Reading data for analysis in Python.
12
+ - Writing processed data back to Stata format.
13
+ - Rule of thumb:
14
+ - If you need to work with Stata data files, use this module for reading
15
+ and writing.
16
+ """
17
+
18
+ from __future__ import annotations
19
+
20
+ from pathlib import Path
21
+
22
+ from ..types import JSONData
23
+ from ..types import JSONList
24
+ from . import stub
25
+
26
+ # SECTION: EXPORTS ========================================================== #
27
+
28
+
29
+ __all__ = [
30
+ 'read',
31
+ 'write',
32
+ ]
33
+
34
+
35
+ # SECTION: FUNCTIONS ======================================================== #
36
+
37
+
38
+ def read(
39
+ path: Path,
40
+ ) -> JSONList:
41
+ """
42
+ Read DTA content from ``path``.
43
+
44
+ Parameters
45
+ ----------
46
+ path : Path
47
+ Path to the DTA file on disk.
48
+
49
+ Returns
50
+ -------
51
+ JSONList
52
+ The list of dictionaries read from the DTA file.
53
+ """
54
+ return stub.read(path, format_name='DTA')
55
+
56
+
57
+ def write(
58
+ path: Path,
59
+ data: JSONData,
60
+ ) -> int:
61
+ """
62
+ Write ``data`` to DTA file at ``path`` and return record count.
63
+
64
+ Parameters
65
+ ----------
66
+ path : Path
67
+ Path to the DTA file on disk.
68
+ data : JSONData
69
+ Data to write as DTA file. Should be a list of dictionaries or a
70
+ single dictionary.
71
+
72
+ Returns
73
+ -------
74
+ int
75
+ The number of rows written to the DTA file.
76
+ """
77
+ return stub.write(path, data, format_name='DTA')
etlplus/file/duckdb.py ADDED
@@ -0,0 +1,78 @@
1
+ """
2
+ :mod:`etlplus.file.duckdb` module.
3
+
4
+ Helpers for reading/writing DuckDB database (DUCKDB) files.
5
+
6
+ Notes
7
+ -----
8
+ - A DUCKDB file is a self-contained, serverless database file format used by
9
+ DuckDB.
10
+ - Common cases:
11
+ - Analytical data storage and processing.
12
+ - Embedded database applications.
13
+ - Fast querying of large datasets.
14
+ - Rule of thumb:
15
+ - If the file follows the DUCKDB specification, use this module for reading
16
+ and writing.
17
+ """
18
+
19
+ from __future__ import annotations
20
+
21
+ from pathlib import Path
22
+
23
+ from ..types import JSONData
24
+ from ..types import JSONList
25
+ from . import stub
26
+
27
+ # SECTION: EXPORTS ========================================================== #
28
+
29
+
30
+ __all__ = [
31
+ 'read',
32
+ 'write',
33
+ ]
34
+
35
+
36
+ # SECTION: FUNCTIONS ======================================================== #
37
+
38
+
39
+ def read(
40
+ path: Path,
41
+ ) -> JSONList:
42
+ """
43
+ Read DUCKDB content from ``path``.
44
+
45
+ Parameters
46
+ ----------
47
+ path : Path
48
+ Path to the DUCKDB file on disk.
49
+
50
+ Returns
51
+ -------
52
+ JSONList
53
+ The list of dictionaries read from the DUCKDB file.
54
+ """
55
+ return stub.read(path, format_name='DUCKDB')
56
+
57
+
58
+ def write(
59
+ path: Path,
60
+ data: JSONData,
61
+ ) -> int:
62
+ """
63
+ Write ``data`` to DUCKDB at ``path`` and return record count.
64
+
65
+ Parameters
66
+ ----------
67
+ path : Path
68
+ Path to the DUCKDB file on disk.
69
+ data : JSONData
70
+ Data to write as DUCKDB. Should be a list of dictionaries or a
71
+ single dictionary.
72
+
73
+ Returns
74
+ -------
75
+ int
76
+ The number of rows written to the DUCKDB file.
77
+ """
78
+ return stub.write(path, data, format_name='DUCKDB')