omlish 0.0.0.dev136__py3-none-any.whl → 0.0.0.dev138__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- omlish/__about__.py +2 -2
- omlish/cached.py +2 -2
- omlish/collections/mappings.py +1 -1
- omlish/configs/flattening.py +1 -1
- omlish/diag/_pycharm/runhack.py +3 -0
- omlish/formats/json/stream/errors.py +2 -0
- omlish/formats/json/stream/lex.py +11 -5
- omlish/formats/json/stream/parse.py +37 -21
- omlish/funcs/genmachine.py +5 -4
- omlish/io/compress/__init__.py +0 -0
- omlish/io/compress/abc.py +104 -0
- omlish/io/compress/adapters.py +147 -0
- omlish/io/compress/bz2.py +42 -0
- omlish/io/compress/gzip.py +306 -0
- omlish/io/compress/lzma.py +32 -0
- omlish/io/compress/types.py +29 -0
- omlish/io/generators/__init__.py +0 -0
- omlish/io/generators/readers.py +183 -0
- omlish/io/generators/stepped.py +104 -0
- omlish/lang/__init__.py +11 -1
- omlish/lang/functions.py +0 -2
- omlish/lang/generators.py +243 -0
- omlish/lang/iterables.py +28 -51
- omlish/lang/maybes.py +4 -4
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/METADATA +1 -1
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/RECORD +34 -22
- /omlish/collections/{_abc.py → abc.py} +0 -0
- /omlish/io/{_abc.py → abc.py} +0 -0
- /omlish/logs/{_abc.py → abc.py} +0 -0
- /omlish/sql/{_abc.py → abc.py} +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/LICENSE +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/WHEEL +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/entry_points.txt +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/top_level.txt +0 -0
omlish/__about__.py
CHANGED
omlish/cached.py
CHANGED
@@ -10,9 +10,9 @@ builtins and thus not distinguish it from a normal property.
|
|
10
10
|
|
11
11
|
"""
|
12
12
|
from .lang.cached import _CachedProperty # noqa
|
13
|
-
from .lang.cached import cached_function
|
13
|
+
from .lang.cached import cached_function as _cached_function
|
14
14
|
|
15
|
-
function =
|
15
|
+
function = _cached_function
|
16
16
|
|
17
17
|
property = property # noqa
|
18
18
|
|
omlish/collections/mappings.py
CHANGED
@@ -86,7 +86,7 @@ class DynamicTypeMap(ta.Generic[V]):
|
|
86
86
|
self._items = list(items)
|
87
87
|
self._weak = bool(weak)
|
88
88
|
|
89
|
-
self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary()
|
89
|
+
self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary() if weak else {}
|
90
90
|
|
91
91
|
@property
|
92
92
|
def items(self) -> ta.Sequence[V]:
|
omlish/configs/flattening.py
CHANGED
omlish/diag/_pycharm/runhack.py
CHANGED
@@ -3,7 +3,10 @@
|
|
3
3
|
.venv/bin/python $(curl -LsSf https://raw.githubusercontent.com/wrmsr/omlish/master/omlish/diag/_pycharm/runhack.py -o $(mktemp) && echo "$_") install
|
4
4
|
|
5
5
|
==
|
6
|
+
TODO:
|
7
|
+
- check for existing files - can't run regular dep entrypoints now
|
6
8
|
|
9
|
+
==
|
7
10
|
See:
|
8
11
|
- https://github.com/JetBrains/intellij-community/blob/6400f70dde6f743e39a257a5a78cc51b644c835e/python/helpers/pycharm/_jb_pytest_runner.py
|
9
12
|
- https://github.com/JetBrains/intellij-community/blob/5a4e584aa59767f2e7cf4bd377adfaaf7503984b/python/helpers/pycharm/_jb_runner_tools.py
|
@@ -3,6 +3,7 @@ TODO:
|
|
3
3
|
- max buf size
|
4
4
|
- max recursion depth
|
5
5
|
- mark start pos of tokens, currently returning end
|
6
|
+
- _do_string inner loop optimization somehow
|
6
7
|
"""
|
7
8
|
import dataclasses as dc
|
8
9
|
import io
|
@@ -12,6 +13,7 @@ import typing as ta
|
|
12
13
|
|
13
14
|
from .... import check
|
14
15
|
from ....funcs.genmachine import GenMachine
|
16
|
+
from .errors import JsonStreamError
|
15
17
|
|
16
18
|
|
17
19
|
##
|
@@ -95,7 +97,7 @@ CONST_TOKENS: ta.Mapping[str, tuple[TokenKind, str | float | None]] = {
|
|
95
97
|
|
96
98
|
|
97
99
|
@dc.dataclass()
|
98
|
-
class
|
100
|
+
class JsonStreamLexError(JsonStreamError):
|
99
101
|
message: str
|
100
102
|
|
101
103
|
pos: Position
|
@@ -160,8 +162,8 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
160
162
|
self._buf.truncate()
|
161
163
|
return raw
|
162
164
|
|
163
|
-
def _raise(self, msg: str) -> ta.NoReturn:
|
164
|
-
raise
|
165
|
+
def _raise(self, msg: str, src: Exception | None = None) -> ta.NoReturn:
|
166
|
+
raise JsonStreamLexError(msg, self.pos) from src
|
165
167
|
|
166
168
|
def _do_main(self):
|
167
169
|
while True:
|
@@ -202,7 +204,7 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
202
204
|
self._raise('Unexpected end of input')
|
203
205
|
|
204
206
|
if not c:
|
205
|
-
|
207
|
+
self._raise(f'Unterminated string literal: {self._buf.getvalue()}')
|
206
208
|
|
207
209
|
self._buf.write(c)
|
208
210
|
if c == '"' and last != '\\':
|
@@ -210,7 +212,11 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
210
212
|
last = c
|
211
213
|
|
212
214
|
raw = self._flip_buf()
|
213
|
-
|
215
|
+
try:
|
216
|
+
sv = json.loads(raw)
|
217
|
+
except json.JSONDecodeError as e:
|
218
|
+
self._raise(f'Invalid string literal: {raw!r}', e)
|
219
|
+
|
214
220
|
yield self._make_tok('STRING', sv, raw, pos)
|
215
221
|
|
216
222
|
return self._do_main()
|
@@ -1,9 +1,12 @@
|
|
1
|
+
import dataclasses as dc
|
1
2
|
import typing as ta
|
2
3
|
|
3
4
|
from .... import lang
|
4
5
|
from ....funcs.genmachine import GenMachine
|
6
|
+
from .errors import JsonStreamError
|
5
7
|
from .lex import SCALAR_VALUE_TYPES
|
6
8
|
from .lex import VALUE_TOKEN_KINDS
|
9
|
+
from .lex import Position
|
7
10
|
from .lex import ScalarValue
|
8
11
|
from .lex import Token
|
9
12
|
|
@@ -79,6 +82,13 @@ def yield_parser_events(obj: ta.Any) -> ta.Generator[JsonStreamParserEvent, None
|
|
79
82
|
##
|
80
83
|
|
81
84
|
|
85
|
+
@dc.dataclass()
|
86
|
+
class JsonStreamParseError(JsonStreamError):
|
87
|
+
message: str
|
88
|
+
|
89
|
+
pos: Position | None = None
|
90
|
+
|
91
|
+
|
82
92
|
class JsonStreamObject(list):
|
83
93
|
def __repr__(self) -> str:
|
84
94
|
return f'{self.__class__.__name__}({super().__repr__()})'
|
@@ -100,29 +110,29 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
100
110
|
if tt == 'KEY':
|
101
111
|
self._stack.pop()
|
102
112
|
if not self._stack:
|
103
|
-
raise
|
113
|
+
raise JsonStreamParseError('Unexpected key')
|
104
114
|
|
105
115
|
tt2 = self._stack[-1]
|
106
116
|
if tt2 == 'OBJECT':
|
107
117
|
return ((v,), self._do_after_pair())
|
108
118
|
|
109
119
|
else:
|
110
|
-
raise
|
120
|
+
raise JsonStreamParseError('Unexpected key')
|
111
121
|
|
112
122
|
elif tt == 'ARRAY':
|
113
123
|
return ((v,), self._do_after_element())
|
114
124
|
|
115
125
|
else:
|
116
|
-
raise
|
126
|
+
raise JsonStreamParseError(f'Unexpected value: {v!r}')
|
117
127
|
|
118
128
|
#
|
119
129
|
|
120
|
-
def _do_value(self):
|
130
|
+
def _do_value(self, *, must_be_present: bool = False):
|
121
131
|
try:
|
122
132
|
tok = yield None
|
123
133
|
except GeneratorExit:
|
124
134
|
if self._stack:
|
125
|
-
raise
|
135
|
+
raise JsonStreamParseError('Expected value') from None
|
126
136
|
else:
|
127
137
|
raise
|
128
138
|
|
@@ -141,13 +151,16 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
141
151
|
yield y
|
142
152
|
return r
|
143
153
|
|
154
|
+
elif must_be_present:
|
155
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
156
|
+
|
144
157
|
elif tok.kind == 'RBRACKET':
|
145
158
|
y, r = self._emit_end_array()
|
146
159
|
yield y
|
147
160
|
return r
|
148
161
|
|
149
162
|
else:
|
150
|
-
raise
|
163
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
151
164
|
|
152
165
|
#
|
153
166
|
|
@@ -157,19 +170,19 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
157
170
|
|
158
171
|
def _emit_end_object(self):
|
159
172
|
if not self._stack:
|
160
|
-
raise
|
173
|
+
raise JsonStreamParseError('Unexpected end object')
|
161
174
|
|
162
175
|
tt = self._stack.pop()
|
163
176
|
if tt != 'OBJECT':
|
164
|
-
raise
|
177
|
+
raise JsonStreamParseError('Unexpected end object')
|
165
178
|
|
166
179
|
return self._emit_event(EndObject)
|
167
180
|
|
168
|
-
def _do_object_body(self):
|
181
|
+
def _do_object_body(self, *, must_be_present: bool = False):
|
169
182
|
try:
|
170
183
|
tok = yield None
|
171
184
|
except GeneratorExit:
|
172
|
-
raise
|
185
|
+
raise JsonStreamParseError('Expected object body') from None
|
173
186
|
|
174
187
|
if tok.kind == 'STRING':
|
175
188
|
k = tok.value
|
@@ -177,30 +190,33 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
177
190
|
try:
|
178
191
|
tok = yield None
|
179
192
|
except GeneratorExit:
|
180
|
-
raise
|
193
|
+
raise JsonStreamParseError('Expected key') from None
|
181
194
|
if tok.kind != 'COLON':
|
182
|
-
raise
|
195
|
+
raise JsonStreamParseError('Expected colon', tok.pos)
|
183
196
|
|
184
197
|
yield (Key(k),)
|
185
198
|
self._stack.append('KEY')
|
186
199
|
return self._do_value()
|
187
200
|
|
201
|
+
elif must_be_present:
|
202
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
203
|
+
|
188
204
|
elif tok.kind == 'RBRACE':
|
189
205
|
y, r = self._emit_end_object()
|
190
206
|
yield y
|
191
207
|
return r
|
192
208
|
|
193
209
|
else:
|
194
|
-
raise
|
210
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
195
211
|
|
196
212
|
def _do_after_pair(self):
|
197
213
|
try:
|
198
214
|
tok = yield None
|
199
215
|
except GeneratorExit:
|
200
|
-
raise
|
216
|
+
raise JsonStreamParseError('Expected continuation') from None
|
201
217
|
|
202
218
|
if tok.kind == 'COMMA':
|
203
|
-
return self._do_object_body()
|
219
|
+
return self._do_object_body(must_be_present=True)
|
204
220
|
|
205
221
|
elif tok.kind == 'RBRACE':
|
206
222
|
y, r = self._emit_end_object()
|
@@ -208,7 +224,7 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
208
224
|
return r
|
209
225
|
|
210
226
|
else:
|
211
|
-
raise
|
227
|
+
raise JsonStreamParseError('Expected continuation', tok.pos)
|
212
228
|
|
213
229
|
#
|
214
230
|
|
@@ -218,11 +234,11 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
218
234
|
|
219
235
|
def _emit_end_array(self):
|
220
236
|
if not self._stack:
|
221
|
-
raise
|
237
|
+
raise JsonStreamParseError('Expected end array')
|
222
238
|
|
223
239
|
tt = self._stack.pop()
|
224
240
|
if tt != 'ARRAY':
|
225
|
-
raise
|
241
|
+
raise JsonStreamParseError('Unexpected end array')
|
226
242
|
|
227
243
|
return self._emit_event(EndArray)
|
228
244
|
|
@@ -230,10 +246,10 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
230
246
|
try:
|
231
247
|
tok = yield None
|
232
248
|
except GeneratorExit:
|
233
|
-
raise
|
249
|
+
raise JsonStreamParseError('Expected continuation') from None
|
234
250
|
|
235
251
|
if tok.kind == 'COMMA':
|
236
|
-
return self._do_value()
|
252
|
+
return self._do_value(must_be_present=True)
|
237
253
|
|
238
254
|
elif tok.kind == 'RBRACKET':
|
239
255
|
y, r = self._emit_end_array()
|
@@ -241,4 +257,4 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
241
257
|
return r
|
242
258
|
|
243
259
|
else:
|
244
|
-
raise
|
260
|
+
raise JsonStreamParseError('Expected continuation', tok.pos)
|
omlish/funcs/genmachine.py
CHANGED
@@ -12,7 +12,7 @@ import typing as ta
|
|
12
12
|
I = ta.TypeVar('I')
|
13
13
|
O = ta.TypeVar('O')
|
14
14
|
|
15
|
-
# MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I, ta.Optional[MachineGen[I, O]]]
|
15
|
+
# MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I | None, ta.Optional[MachineGen[I, O]]]
|
16
16
|
MachineGen: ta.TypeAlias = ta.Generator[ta.Any, ta.Any, ta.Any]
|
17
17
|
|
18
18
|
|
@@ -67,8 +67,7 @@ class GenMachine(ta.Generic[I, O]):
|
|
67
67
|
return self
|
68
68
|
|
69
69
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
70
|
-
|
71
|
-
self.close()
|
70
|
+
self.close()
|
72
71
|
|
73
72
|
#
|
74
73
|
|
@@ -93,8 +92,10 @@ class GenMachine(ta.Generic[I, O]):
|
|
93
92
|
if self._gen is None:
|
94
93
|
raise GenMachine.ClosedError
|
95
94
|
|
95
|
+
gi: I | None = i
|
96
96
|
try:
|
97
|
-
while (o := self._gen.send(
|
97
|
+
while (o := self._gen.send(gi)) is not None:
|
98
|
+
gi = None
|
98
99
|
yield from o
|
99
100
|
|
100
101
|
except StopIteration as s:
|
File without changes
|
@@ -0,0 +1,104 @@
|
|
1
|
+
"""
|
2
|
+
https://docs.python.org/3/library/bz2.html#bz2.BZ2Compressor
|
3
|
+
https://docs.python.org/3/library/zlib.html#zlib.decompressobj
|
4
|
+
https://docs.python.org/3/library/lzma.html#lzma.LZMADecompressor
|
5
|
+
"""
|
6
|
+
import abc
|
7
|
+
|
8
|
+
|
9
|
+
##
|
10
|
+
|
11
|
+
|
12
|
+
class Compressor(abc.ABC):
|
13
|
+
@abc.abstractmethod
|
14
|
+
def compress(self, data: bytes) -> bytes:
|
15
|
+
"""
|
16
|
+
Provide data to the compressor object. Returns a chunk of compressed data if possible, or an empty byte string
|
17
|
+
otherwise.
|
18
|
+
|
19
|
+
When you have finished providing data to the compressor, call the flush() method to finish the compression
|
20
|
+
process.
|
21
|
+
"""
|
22
|
+
|
23
|
+
raise NotImplementedError
|
24
|
+
|
25
|
+
@abc.abstractmethod
|
26
|
+
def flush(self) -> bytes:
|
27
|
+
"""
|
28
|
+
Finish the compression process. Returns the compressed data left in internal buffers.
|
29
|
+
|
30
|
+
The compressor object may not be used after this method has been called.
|
31
|
+
"""
|
32
|
+
|
33
|
+
raise NotImplementedError
|
34
|
+
|
35
|
+
|
36
|
+
##
|
37
|
+
|
38
|
+
|
39
|
+
class Decompressor(abc.ABC):
|
40
|
+
@property
|
41
|
+
@abc.abstractmethod
|
42
|
+
def unused_data(self) -> bytes:
|
43
|
+
"""
|
44
|
+
Data found after the end of the compressed stream.
|
45
|
+
|
46
|
+
If this attribute is accessed before the end of the stream has been reached, its value will be b''.
|
47
|
+
"""
|
48
|
+
|
49
|
+
raise NotImplementedError
|
50
|
+
|
51
|
+
@property
|
52
|
+
@abc.abstractmethod
|
53
|
+
def eof(self) -> bool:
|
54
|
+
"""True if the end-of-stream marker has been reached."""
|
55
|
+
|
56
|
+
raise NotImplementedError
|
57
|
+
|
58
|
+
@abc.abstractmethod
|
59
|
+
def decompress(self, data: bytes, *max_length: int) -> bytes:
|
60
|
+
"""
|
61
|
+
Decompress data, returning a bytes object containing the uncompressed data corresponding to at least part of the
|
62
|
+
data in string. This data should be concatenated to the output produced by any preceding calls to the
|
63
|
+
decompress() method. Some of the input data may be preserved in internal buffers for later processing.
|
64
|
+
|
65
|
+
If the optional parameter max_length is non-zero then the return value will be no longer than max_length.
|
66
|
+
"""
|
67
|
+
|
68
|
+
raise NotImplementedError
|
69
|
+
|
70
|
+
|
71
|
+
class NeedsInputDecompressor(Decompressor):
|
72
|
+
"""
|
73
|
+
Used by:
|
74
|
+
- bz2.BZ2Decompressor
|
75
|
+
- lzma.LZMADecompressor
|
76
|
+
"""
|
77
|
+
|
78
|
+
@property
|
79
|
+
@abc.abstractmethod
|
80
|
+
def needs_input(self) -> bool:
|
81
|
+
"""
|
82
|
+
False if the decompress() method can provide more decompressed data before requiring new uncompressed input.
|
83
|
+
"""
|
84
|
+
|
85
|
+
raise NotImplementedError
|
86
|
+
|
87
|
+
|
88
|
+
class UnconsumedTailDecompressor(Decompressor):
|
89
|
+
"""
|
90
|
+
Used by:
|
91
|
+
- zlib.decompressobj
|
92
|
+
"""
|
93
|
+
|
94
|
+
@property
|
95
|
+
@abc.abstractmethod
|
96
|
+
def unconsumed_tail(self) -> bytes:
|
97
|
+
"""
|
98
|
+
A bytes object that contains any data that was not consumed by the last decompress() call because it exceeded
|
99
|
+
the limit for the uncompressed data buffer. This data has not yet been seen by the zlib machinery, so you must
|
100
|
+
feed it (possibly with further data concatenated to it) back to a subsequent decompress() method call in order
|
101
|
+
to get correct output.
|
102
|
+
"""
|
103
|
+
|
104
|
+
raise NotImplementedError
|
@@ -0,0 +1,147 @@
|
|
1
|
+
# PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2
|
2
|
+
# --------------------------------------------
|
3
|
+
#
|
4
|
+
# 1. This LICENSE AGREEMENT is between the Python Software Foundation ("PSF"), and the Individual or Organization
|
5
|
+
# ("Licensee") accessing and otherwise using this software ("Python") in source or binary form and its associated
|
6
|
+
# documentation.
|
7
|
+
#
|
8
|
+
# 2. Subject to the terms and conditions of this License Agreement, PSF hereby grants Licensee a nonexclusive,
|
9
|
+
# royalty-free, world-wide license to reproduce, analyze, test, perform and/or display publicly, prepare derivative
|
10
|
+
# works, distribute, and otherwise use Python alone or in any derivative version, provided, however, that PSF's License
|
11
|
+
# Agreement and PSF's notice of copyright, i.e., "Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009,
|
12
|
+
# 2010, 2011, 2012, 2013, 2014, 2015, 2016, 2017 Python Software Foundation; All Rights Reserved" are retained in Python
|
13
|
+
# alone or in any derivative version prepared by Licensee.
|
14
|
+
#
|
15
|
+
# 3. In the event Licensee prepares a derivative work that is based on or incorporates Python or any part thereof, and
|
16
|
+
# wants to make the derivative work available to others as provided herein, then Licensee hereby agrees to include in
|
17
|
+
# any such work a brief summary of the changes made to Python.
|
18
|
+
#
|
19
|
+
# 4. PSF is making Python available to Licensee on an "AS IS" basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES,
|
20
|
+
# EXPRESS OR IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND DISCLAIMS ANY REPRESENTATION OR WARRANTY
|
21
|
+
# OF MERCHANTABILITY OR FITNESS FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT INFRINGE ANY THIRD PARTY
|
22
|
+
# RIGHTS.
|
23
|
+
#
|
24
|
+
# 5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL
|
25
|
+
# DAMAGES OR LOSS AS A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, OR ANY DERIVATIVE THEREOF, EVEN IF
|
26
|
+
# ADVISED OF THE POSSIBILITY THEREOF.
|
27
|
+
#
|
28
|
+
# 6. This License Agreement will automatically terminate upon a material breach of its terms and conditions.
|
29
|
+
#
|
30
|
+
# 7. Nothing in this License Agreement shall be deemed to create any relationship of agency, partnership, or joint
|
31
|
+
# venture between PSF and Licensee. This License Agreement does not grant permission to use PSF trademarks or trade
|
32
|
+
# name in a trademark sense to endorse or promote products or services of Licensee, or any third party.
|
33
|
+
#
|
34
|
+
# 8. By copying, installing or otherwise using Python, Licensee agrees to be bound by the terms and conditions of this
|
35
|
+
# License Agreement.
|
36
|
+
# ~> https://github.com/python/cpython/blob/f19c50a4817ffebb26132182ed8cec6a72342cc0/Lib/_compression.py
|
37
|
+
import typing as ta
|
38
|
+
|
39
|
+
from ... import check
|
40
|
+
from .abc import Compressor
|
41
|
+
from .abc import NeedsInputDecompressor
|
42
|
+
from .abc import UnconsumedTailDecompressor
|
43
|
+
from .types import IncrementalCompressor
|
44
|
+
from .types import IncrementalDecompressor
|
45
|
+
|
46
|
+
|
47
|
+
##
|
48
|
+
|
49
|
+
|
50
|
+
class CompressorIncrementalAdapter:
|
51
|
+
def __init__(
|
52
|
+
self,
|
53
|
+
factory: ta.Callable[..., Compressor],
|
54
|
+
) -> None:
|
55
|
+
super().__init__()
|
56
|
+
|
57
|
+
self._factory = factory
|
58
|
+
|
59
|
+
def __call__(self) -> IncrementalCompressor:
|
60
|
+
compressor = self._factory()
|
61
|
+
|
62
|
+
while True:
|
63
|
+
data = check.isinstance((yield None), bytes)
|
64
|
+
if not data:
|
65
|
+
break
|
66
|
+
|
67
|
+
compressed = compressor.compress(data)
|
68
|
+
if compressed:
|
69
|
+
check.none((yield compressed))
|
70
|
+
|
71
|
+
if (fl := compressor.flush()):
|
72
|
+
check.none((yield fl))
|
73
|
+
|
74
|
+
check.none((yield b''))
|
75
|
+
|
76
|
+
|
77
|
+
##
|
78
|
+
|
79
|
+
|
80
|
+
class DecompressorIncrementalAdapter:
|
81
|
+
def __init__(
|
82
|
+
self,
|
83
|
+
factory: ta.Callable[..., NeedsInputDecompressor | UnconsumedTailDecompressor],
|
84
|
+
*,
|
85
|
+
trailing_error: type[BaseException] | tuple[type[BaseException], ...] = (),
|
86
|
+
) -> None:
|
87
|
+
super().__init__()
|
88
|
+
|
89
|
+
self._factory = factory
|
90
|
+
self._trailing_error = trailing_error
|
91
|
+
|
92
|
+
def __call__(self) -> IncrementalDecompressor:
|
93
|
+
pos = 0
|
94
|
+
|
95
|
+
data = None # Default if EOF is encountered
|
96
|
+
|
97
|
+
decompressor = self._factory()
|
98
|
+
|
99
|
+
while True:
|
100
|
+
# Depending on the input data, our call to the decompressor may not return any data. In this case, try again
|
101
|
+
# after reading another block.
|
102
|
+
while True:
|
103
|
+
if decompressor.eof:
|
104
|
+
rawblock = decompressor.unused_data
|
105
|
+
if not rawblock:
|
106
|
+
rawblock = check.isinstance((yield None), bytes)
|
107
|
+
if not rawblock:
|
108
|
+
break
|
109
|
+
|
110
|
+
# Continue to next stream.
|
111
|
+
decompressor = self._factory()
|
112
|
+
|
113
|
+
try:
|
114
|
+
data = decompressor.decompress(rawblock)
|
115
|
+
except self._trailing_error:
|
116
|
+
# Trailing data isn't a valid compressed stream; ignore it.
|
117
|
+
break
|
118
|
+
|
119
|
+
else:
|
120
|
+
if hasattr(decompressor, 'needs_input'):
|
121
|
+
if decompressor.needs_input:
|
122
|
+
rawblock = check.isinstance((yield None), bytes)
|
123
|
+
if not rawblock:
|
124
|
+
raise EOFError('Compressed file ended before the end-of-stream marker was reached')
|
125
|
+
else:
|
126
|
+
rawblock = b''
|
127
|
+
|
128
|
+
elif hasattr(decompressor, 'unconsumed_tail'):
|
129
|
+
if not (rawblock := decompressor.unconsumed_tail):
|
130
|
+
rawblock = check.isinstance((yield None), bytes)
|
131
|
+
if not rawblock:
|
132
|
+
raise EOFError('Compressed file ended before the end-of-stream marker was reached')
|
133
|
+
|
134
|
+
else:
|
135
|
+
raise TypeError(decompressor)
|
136
|
+
|
137
|
+
data = decompressor.decompress(rawblock)
|
138
|
+
|
139
|
+
if data:
|
140
|
+
break
|
141
|
+
|
142
|
+
if not data:
|
143
|
+
check.none((yield b''))
|
144
|
+
return
|
145
|
+
|
146
|
+
pos += len(data)
|
147
|
+
check.none((yield data))
|
@@ -0,0 +1,42 @@
|
|
1
|
+
import functools
|
2
|
+
import typing as ta
|
3
|
+
|
4
|
+
from ... import lang
|
5
|
+
from .adapters import CompressorIncrementalAdapter
|
6
|
+
from .adapters import DecompressorIncrementalAdapter
|
7
|
+
from .types import IncrementalCompressor
|
8
|
+
from .types import IncrementalDecompressor
|
9
|
+
|
10
|
+
|
11
|
+
if ta.TYPE_CHECKING:
|
12
|
+
import bz2
|
13
|
+
else:
|
14
|
+
bz2 = lang.proxy_import('bz2')
|
15
|
+
|
16
|
+
|
17
|
+
class IncrementalBz2Compressor:
|
18
|
+
def __init__(
|
19
|
+
self,
|
20
|
+
*,
|
21
|
+
compresslevel: int = 9,
|
22
|
+
) -> None:
|
23
|
+
super().__init__()
|
24
|
+
|
25
|
+
self._compresslevel = compresslevel
|
26
|
+
|
27
|
+
@lang.autostart
|
28
|
+
def __call__(self) -> IncrementalCompressor:
|
29
|
+
return CompressorIncrementalAdapter(
|
30
|
+
functools.partial(
|
31
|
+
bz2.BZ2Compressor, # type: ignore
|
32
|
+
self._compresslevel,
|
33
|
+
),
|
34
|
+
)()
|
35
|
+
|
36
|
+
|
37
|
+
class IncrementalBz2Decompressor:
|
38
|
+
def __call__(self) -> IncrementalDecompressor:
|
39
|
+
return DecompressorIncrementalAdapter(
|
40
|
+
bz2.BZ2Decompressor, # type: ignore
|
41
|
+
trailing_error=OSError,
|
42
|
+
)()
|