omlish 0.0.0.dev136__py3-none-any.whl → 0.0.0.dev138__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- omlish/__about__.py +2 -2
- omlish/cached.py +2 -2
- omlish/collections/mappings.py +1 -1
- omlish/configs/flattening.py +1 -1
- omlish/diag/_pycharm/runhack.py +3 -0
- omlish/formats/json/stream/errors.py +2 -0
- omlish/formats/json/stream/lex.py +11 -5
- omlish/formats/json/stream/parse.py +37 -21
- omlish/funcs/genmachine.py +5 -4
- omlish/io/compress/__init__.py +0 -0
- omlish/io/compress/abc.py +104 -0
- omlish/io/compress/adapters.py +147 -0
- omlish/io/compress/bz2.py +42 -0
- omlish/io/compress/gzip.py +306 -0
- omlish/io/compress/lzma.py +32 -0
- omlish/io/compress/types.py +29 -0
- omlish/io/generators/__init__.py +0 -0
- omlish/io/generators/readers.py +183 -0
- omlish/io/generators/stepped.py +104 -0
- omlish/lang/__init__.py +11 -1
- omlish/lang/functions.py +0 -2
- omlish/lang/generators.py +243 -0
- omlish/lang/iterables.py +28 -51
- omlish/lang/maybes.py +4 -4
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/METADATA +1 -1
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/RECORD +34 -22
- /omlish/collections/{_abc.py → abc.py} +0 -0
- /omlish/io/{_abc.py → abc.py} +0 -0
- /omlish/logs/{_abc.py → abc.py} +0 -0
- /omlish/sql/{_abc.py → abc.py} +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/LICENSE +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/WHEEL +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/entry_points.txt +0 -0
- {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/top_level.txt +0 -0
omlish/__about__.py
CHANGED
omlish/cached.py
CHANGED
@@ -10,9 +10,9 @@ builtins and thus not distinguish it from a normal property.
|
|
10
10
|
|
11
11
|
"""
|
12
12
|
from .lang.cached import _CachedProperty # noqa
|
13
|
-
from .lang.cached import cached_function
|
13
|
+
from .lang.cached import cached_function as _cached_function
|
14
14
|
|
15
|
-
function =
|
15
|
+
function = _cached_function
|
16
16
|
|
17
17
|
property = property # noqa
|
18
18
|
|
omlish/collections/mappings.py
CHANGED
@@ -86,7 +86,7 @@ class DynamicTypeMap(ta.Generic[V]):
|
|
86
86
|
self._items = list(items)
|
87
87
|
self._weak = bool(weak)
|
88
88
|
|
89
|
-
self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary()
|
89
|
+
self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary() if weak else {}
|
90
90
|
|
91
91
|
@property
|
92
92
|
def items(self) -> ta.Sequence[V]:
|
omlish/configs/flattening.py
CHANGED
omlish/diag/_pycharm/runhack.py
CHANGED
@@ -3,7 +3,10 @@
|
|
3
3
|
.venv/bin/python $(curl -LsSf https://raw.githubusercontent.com/wrmsr/omlish/master/omlish/diag/_pycharm/runhack.py -o $(mktemp) && echo "$_") install
|
4
4
|
|
5
5
|
==
|
6
|
+
TODO:
|
7
|
+
- check for existing files - can't run regular dep entrypoints now
|
6
8
|
|
9
|
+
==
|
7
10
|
See:
|
8
11
|
- https://github.com/JetBrains/intellij-community/blob/6400f70dde6f743e39a257a5a78cc51b644c835e/python/helpers/pycharm/_jb_pytest_runner.py
|
9
12
|
- https://github.com/JetBrains/intellij-community/blob/5a4e584aa59767f2e7cf4bd377adfaaf7503984b/python/helpers/pycharm/_jb_runner_tools.py
|
@@ -3,6 +3,7 @@ TODO:
|
|
3
3
|
- max buf size
|
4
4
|
- max recursion depth
|
5
5
|
- mark start pos of tokens, currently returning end
|
6
|
+
- _do_string inner loop optimization somehow
|
6
7
|
"""
|
7
8
|
import dataclasses as dc
|
8
9
|
import io
|
@@ -12,6 +13,7 @@ import typing as ta
|
|
12
13
|
|
13
14
|
from .... import check
|
14
15
|
from ....funcs.genmachine import GenMachine
|
16
|
+
from .errors import JsonStreamError
|
15
17
|
|
16
18
|
|
17
19
|
##
|
@@ -95,7 +97,7 @@ CONST_TOKENS: ta.Mapping[str, tuple[TokenKind, str | float | None]] = {
|
|
95
97
|
|
96
98
|
|
97
99
|
@dc.dataclass()
|
98
|
-
class
|
100
|
+
class JsonStreamLexError(JsonStreamError):
|
99
101
|
message: str
|
100
102
|
|
101
103
|
pos: Position
|
@@ -160,8 +162,8 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
160
162
|
self._buf.truncate()
|
161
163
|
return raw
|
162
164
|
|
163
|
-
def _raise(self, msg: str) -> ta.NoReturn:
|
164
|
-
raise
|
165
|
+
def _raise(self, msg: str, src: Exception | None = None) -> ta.NoReturn:
|
166
|
+
raise JsonStreamLexError(msg, self.pos) from src
|
165
167
|
|
166
168
|
def _do_main(self):
|
167
169
|
while True:
|
@@ -202,7 +204,7 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
202
204
|
self._raise('Unexpected end of input')
|
203
205
|
|
204
206
|
if not c:
|
205
|
-
|
207
|
+
self._raise(f'Unterminated string literal: {self._buf.getvalue()}')
|
206
208
|
|
207
209
|
self._buf.write(c)
|
208
210
|
if c == '"' and last != '\\':
|
@@ -210,7 +212,11 @@ class JsonStreamLexer(GenMachine[str, Token]):
|
|
210
212
|
last = c
|
211
213
|
|
212
214
|
raw = self._flip_buf()
|
213
|
-
|
215
|
+
try:
|
216
|
+
sv = json.loads(raw)
|
217
|
+
except json.JSONDecodeError as e:
|
218
|
+
self._raise(f'Invalid string literal: {raw!r}', e)
|
219
|
+
|
214
220
|
yield self._make_tok('STRING', sv, raw, pos)
|
215
221
|
|
216
222
|
return self._do_main()
|
@@ -1,9 +1,12 @@
|
|
1
|
+
import dataclasses as dc
|
1
2
|
import typing as ta
|
2
3
|
|
3
4
|
from .... import lang
|
4
5
|
from ....funcs.genmachine import GenMachine
|
6
|
+
from .errors import JsonStreamError
|
5
7
|
from .lex import SCALAR_VALUE_TYPES
|
6
8
|
from .lex import VALUE_TOKEN_KINDS
|
9
|
+
from .lex import Position
|
7
10
|
from .lex import ScalarValue
|
8
11
|
from .lex import Token
|
9
12
|
|
@@ -79,6 +82,13 @@ def yield_parser_events(obj: ta.Any) -> ta.Generator[JsonStreamParserEvent, None
|
|
79
82
|
##
|
80
83
|
|
81
84
|
|
85
|
+
@dc.dataclass()
|
86
|
+
class JsonStreamParseError(JsonStreamError):
|
87
|
+
message: str
|
88
|
+
|
89
|
+
pos: Position | None = None
|
90
|
+
|
91
|
+
|
82
92
|
class JsonStreamObject(list):
|
83
93
|
def __repr__(self) -> str:
|
84
94
|
return f'{self.__class__.__name__}({super().__repr__()})'
|
@@ -100,29 +110,29 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
100
110
|
if tt == 'KEY':
|
101
111
|
self._stack.pop()
|
102
112
|
if not self._stack:
|
103
|
-
raise
|
113
|
+
raise JsonStreamParseError('Unexpected key')
|
104
114
|
|
105
115
|
tt2 = self._stack[-1]
|
106
116
|
if tt2 == 'OBJECT':
|
107
117
|
return ((v,), self._do_after_pair())
|
108
118
|
|
109
119
|
else:
|
110
|
-
raise
|
120
|
+
raise JsonStreamParseError('Unexpected key')
|
111
121
|
|
112
122
|
elif tt == 'ARRAY':
|
113
123
|
return ((v,), self._do_after_element())
|
114
124
|
|
115
125
|
else:
|
116
|
-
raise
|
126
|
+
raise JsonStreamParseError(f'Unexpected value: {v!r}')
|
117
127
|
|
118
128
|
#
|
119
129
|
|
120
|
-
def _do_value(self):
|
130
|
+
def _do_value(self, *, must_be_present: bool = False):
|
121
131
|
try:
|
122
132
|
tok = yield None
|
123
133
|
except GeneratorExit:
|
124
134
|
if self._stack:
|
125
|
-
raise
|
135
|
+
raise JsonStreamParseError('Expected value') from None
|
126
136
|
else:
|
127
137
|
raise
|
128
138
|
|
@@ -141,13 +151,16 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
141
151
|
yield y
|
142
152
|
return r
|
143
153
|
|
154
|
+
elif must_be_present:
|
155
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
156
|
+
|
144
157
|
elif tok.kind == 'RBRACKET':
|
145
158
|
y, r = self._emit_end_array()
|
146
159
|
yield y
|
147
160
|
return r
|
148
161
|
|
149
162
|
else:
|
150
|
-
raise
|
163
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
151
164
|
|
152
165
|
#
|
153
166
|
|
@@ -157,19 +170,19 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
157
170
|
|
158
171
|
def _emit_end_object(self):
|
159
172
|
if not self._stack:
|
160
|
-
raise
|
173
|
+
raise JsonStreamParseError('Unexpected end object')
|
161
174
|
|
162
175
|
tt = self._stack.pop()
|
163
176
|
if tt != 'OBJECT':
|
164
|
-
raise
|
177
|
+
raise JsonStreamParseError('Unexpected end object')
|
165
178
|
|
166
179
|
return self._emit_event(EndObject)
|
167
180
|
|
168
|
-
def _do_object_body(self):
|
181
|
+
def _do_object_body(self, *, must_be_present: bool = False):
|
169
182
|
try:
|
170
183
|
tok = yield None
|
171
184
|
except GeneratorExit:
|
172
|
-
raise
|
185
|
+
raise JsonStreamParseError('Expected object body') from None
|
173
186
|
|
174
187
|
if tok.kind == 'STRING':
|
175
188
|
k = tok.value
|
@@ -177,30 +190,33 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
177
190
|
try:
|
178
191
|
tok = yield None
|
179
192
|
except GeneratorExit:
|
180
|
-
raise
|
193
|
+
raise JsonStreamParseError('Expected key') from None
|
181
194
|
if tok.kind != 'COLON':
|
182
|
-
raise
|
195
|
+
raise JsonStreamParseError('Expected colon', tok.pos)
|
183
196
|
|
184
197
|
yield (Key(k),)
|
185
198
|
self._stack.append('KEY')
|
186
199
|
return self._do_value()
|
187
200
|
|
201
|
+
elif must_be_present:
|
202
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
203
|
+
|
188
204
|
elif tok.kind == 'RBRACE':
|
189
205
|
y, r = self._emit_end_object()
|
190
206
|
yield y
|
191
207
|
return r
|
192
208
|
|
193
209
|
else:
|
194
|
-
raise
|
210
|
+
raise JsonStreamParseError('Expected value', tok.pos)
|
195
211
|
|
196
212
|
def _do_after_pair(self):
|
197
213
|
try:
|
198
214
|
tok = yield None
|
199
215
|
except GeneratorExit:
|
200
|
-
raise
|
216
|
+
raise JsonStreamParseError('Expected continuation') from None
|
201
217
|
|
202
218
|
if tok.kind == 'COMMA':
|
203
|
-
return self._do_object_body()
|
219
|
+
return self._do_object_body(must_be_present=True)
|
204
220
|
|
205
221
|
elif tok.kind == 'RBRACE':
|
206
222
|
y, r = self._emit_end_object()
|
@@ -208,7 +224,7 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
208
224
|
return r
|
209
225
|
|
210
226
|
else:
|
211
|
-
raise
|
227
|
+
raise JsonStreamParseError('Expected continuation', tok.pos)
|
212
228
|
|
213
229
|
#
|
214
230
|
|
@@ -218,11 +234,11 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
218
234
|
|
219
235
|
def _emit_end_array(self):
|
220
236
|
if not self._stack:
|
221
|
-
raise
|
237
|
+
raise JsonStreamParseError('Expected end array')
|
222
238
|
|
223
239
|
tt = self._stack.pop()
|
224
240
|
if tt != 'ARRAY':
|
225
|
-
raise
|
241
|
+
raise JsonStreamParseError('Unexpected end array')
|
226
242
|
|
227
243
|
return self._emit_event(EndArray)
|
228
244
|
|
@@ -230,10 +246,10 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
230
246
|
try:
|
231
247
|
tok = yield None
|
232
248
|
except GeneratorExit:
|
233
|
-
raise
|
249
|
+
raise JsonStreamParseError('Expected continuation') from None
|
234
250
|
|
235
251
|
if tok.kind == 'COMMA':
|
236
|
-
return self._do_value()
|
252
|
+
return self._do_value(must_be_present=True)
|
237
253
|
|
238
254
|
elif tok.kind == 'RBRACKET':
|
239
255
|
y, r = self._emit_end_array()
|
@@ -241,4 +257,4 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
|
|
241
257
|
return r
|
242
258
|
|
243
259
|
else:
|
244
|
-
raise
|
260
|
+
raise JsonStreamParseError('Expected continuation', tok.pos)
|
omlish/funcs/genmachine.py
CHANGED
@@ -12,7 +12,7 @@ import typing as ta
|
|
12
12
|
I = ta.TypeVar('I')
|
13
13
|
O = ta.TypeVar('O')
|
14
14
|
|
15
|
-
# MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I, ta.Optional[MachineGen[I, O]]]
|
15
|
+
# MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I | None, ta.Optional[MachineGen[I, O]]]
|
16
16
|
MachineGen: ta.TypeAlias = ta.Generator[ta.Any, ta.Any, ta.Any]
|
17
17
|
|
18
18
|
|
@@ -67,8 +67,7 @@ class GenMachine(ta.Generic[I, O]):
|
|
67
67
|
return self
|
68
68
|
|
69
69
|
def __exit__(self, exc_type, exc_val, exc_tb):
|
70
|
-
|
71
|
-
self.close()
|
70
|
+
self.close()
|
72
71
|
|
73
72
|
#
|
74
73
|
|
@@ -93,8 +92,10 @@ class GenMachine(ta.Generic[I, O]):
|
|
93
92
|
if self._gen is None:
|
94
93
|
raise GenMachine.ClosedError
|
95
94
|
|
95
|
+
gi: I | None = i
|
96
96
|
try:
|
97
|
-
while (o := self._gen.send(
|
97
|
+
while (o := self._gen.send(gi)) is not None:
|
98
|
+
gi = None
|
98
99
|
yield from o
|
99
100
|
|
100
101
|
except StopIteration as s:
|
File without changes
|
@@ -0,0 +1,104 @@
|
|
1
|
+
"""
|
2
|
+
https://docs.python.org/3/library/bz2.html#bz2.BZ2Compressor
|
3
|
+
https://docs.python.org/3/library/zlib.html#zlib.decompressobj
|
4
|
+
https://docs.python.org/3/library/lzma.html#lzma.LZMADecompressor
|
5
|
+
"""
|
6
|
+
import abc
|
7
|
+
|
8
|
+
|
9
|
+
##
|
10
|
+
|
11
|
+
|
12
|
+
class Compressor(abc.ABC):
|
13
|
+
@abc.abstractmethod
|
14
|
+
def compress(self, data: bytes) -> bytes:
|
15
|
+
"""
|
16
|
+
Provide data to the compressor object. Returns a chunk of compressed data if possible, or an empty byte string
|
17
|
+
otherwise.
|
18
|
+
|
19
|
+
When you have finished providing data to the compressor, call the flush() method to finish the compression
|
20
|
+
process.
|
21
|
+
"""
|
22
|
+
|
23
|
+
raise NotImplementedError
|
24
|
+
|
25
|
+
@abc.abstractmethod
|
26
|
+
def flush(self) -> bytes:
|
27
|
+
"""
|
28
|
+
Finish the compression process. Returns the compressed data left in internal buffers.
|
29
|
+
|
30
|
+
The compressor object may not be used after this method has been called.
|
31
|
+
"""
|
32
|
+
|
33
|
+
raise NotImplementedError
|
34
|
+
|
35
|
+
|
36
|
+
##
|
37
|
+
|
38
|
+
|
39
|
+
class Decompressor(abc.ABC):
|
40
|
+
@property
|
41
|
+
@abc.abstractmethod
|
42
|
+
def unused_data(self) -> bytes:
|
43
|
+
"""
|
44
|
+
Data found after the end of the compressed stream.
|
45
|
+
|
46
|
+
If this attribute is accessed before the end of the stream has been reached, its value will be b''.
|
47
|
+
"""
|
48
|
+
|
49
|
+
raise NotImplementedError
|
50
|
+
|
51
|
+
@property
|
52
|
+
@abc.abstractmethod
|
53
|
+
def eof(self) -> bool:
|
54
|
+
"""True if the end-of-stream marker has been reached."""
|
55
|
+
|
56
|
+
raise NotImplementedError
|
57
|
+
|
58
|
+
@abc.abstractmethod
|
59
|
+
def decompress(self, data: bytes, *max_length: int) -> bytes:
|
60
|
+
"""
|
61
|
+
Decompress data, returning a bytes object containing the uncompressed data corresponding to at least part of the
|
62
|
+
data in string. This data should be concatenated to the output produced by any preceding calls to the
|
63
|
+
decompress() method. Some of the input data may be preserved in internal buffers for later processing.
|
64
|
+
|
65
|
+
If the optional parameter max_length is non-zero then the return value will be no longer than max_length.
|
66
|
+
"""
|
67
|
+
|
68
|
+
raise NotImplementedError
|
69
|
+
|
70
|
+
|
71
|
+
class NeedsInputDecompressor(Decompressor):
|
72
|
+
"""
|
73
|
+
Used by:
|
74
|
+
- bz2.BZ2Decompressor
|
75
|
+
- lzma.LZMADecompressor
|
76
|
+
"""
|
77
|
+
|
78
|
+
@property
|
79
|
+
@abc.abstractmethod
|
80
|
+
def needs_input(self) -> bool:
|
81
|
+
"""
|
82
|
+
False if the decompress() method can provide more decompressed data before requiring new uncompressed input.
|
83
|
+
"""
|
84
|
+
|
85
|
+
raise NotImplementedError
|
86
|
+
|
87
|
+
|
88
|
+
class UnconsumedTailDecompressor(Decompressor):
|
89
|
+
"""
|
90
|
+
Used by:
|
91
|
+
- zlib.decompressobj
|
92
|
+
"""
|
93
|
+
|
94
|
+
@property
|
95
|
+
@abc.abstractmethod
|
96
|
+
def unconsumed_tail(self) -> bytes:
|
97
|
+
"""
|
98
|
+
A bytes object that contains any data that was not consumed by the last decompress() call because it exceeded
|
99
|
+
the limit for the uncompressed data buffer. This data has not yet been seen by the zlib machinery, so you must
|
100
|
+
feed it (possibly with further data concatenated to it) back to a subsequent decompress() method call in order
|
101
|
+
to get correct output.
|
102
|
+
"""
|
103
|
+
|
104
|
+
raise NotImplementedError
|
@@ -0,0 +1,147 @@
|
|
1
|
+
# PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2
|
2
|
+
# --------------------------------------------
|
3
|
+
#
|
4
|
+
# 1. This LICENSE AGREEMENT is between the Python Software Foundation ("PSF"), and the Individual or Organization
|
5
|
+
# ("Licensee") accessing and otherwise using this software ("Python") in source or binary form and its associated
|
6
|
+
# documentation.
|
7
|
+
#
|
8
|
+
# 2. Subject to the terms and conditions of this License Agreement, PSF hereby grants Licensee a nonexclusive,
|
9
|
+
# royalty-free, world-wide license to reproduce, analyze, test, perform and/or display publicly, prepare derivative
|
10
|
+
# works, distribute, and otherwise use Python alone or in any derivative version, provided, however, that PSF's License
|
11
|
+
# Agreement and PSF's notice of copyright, i.e., "Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009,
|
12
|
+
# 2010, 2011, 2012, 2013, 2014, 2015, 2016, 2017 Python Software Foundation; All Rights Reserved" are retained in Python
|
13
|
+
# alone or in any derivative version prepared by Licensee.
|
14
|
+
#
|
15
|
+
# 3. In the event Licensee prepares a derivative work that is based on or incorporates Python or any part thereof, and
|
16
|
+
# wants to make the derivative work available to others as provided herein, then Licensee hereby agrees to include in
|
17
|
+
# any such work a brief summary of the changes made to Python.
|
18
|
+
#
|
19
|
+
# 4. PSF is making Python available to Licensee on an "AS IS" basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES,
|
20
|
+
# EXPRESS OR IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND DISCLAIMS ANY REPRESENTATION OR WARRANTY
|
21
|
+
# OF MERCHANTABILITY OR FITNESS FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT INFRINGE ANY THIRD PARTY
|
22
|
+
# RIGHTS.
|
23
|
+
#
|
24
|
+
# 5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL
|
25
|
+
# DAMAGES OR LOSS AS A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, OR ANY DERIVATIVE THEREOF, EVEN IF
|
26
|
+
# ADVISED OF THE POSSIBILITY THEREOF.
|
27
|
+
#
|
28
|
+
# 6. This License Agreement will automatically terminate upon a material breach of its terms and conditions.
|
29
|
+
#
|
30
|
+
# 7. Nothing in this License Agreement shall be deemed to create any relationship of agency, partnership, or joint
|
31
|
+
# venture between PSF and Licensee. This License Agreement does not grant permission to use PSF trademarks or trade
|
32
|
+
# name in a trademark sense to endorse or promote products or services of Licensee, or any third party.
|
33
|
+
#
|
34
|
+
# 8. By copying, installing or otherwise using Python, Licensee agrees to be bound by the terms and conditions of this
|
35
|
+
# License Agreement.
|
36
|
+
# ~> https://github.com/python/cpython/blob/f19c50a4817ffebb26132182ed8cec6a72342cc0/Lib/_compression.py
|
37
|
+
import typing as ta
|
38
|
+
|
39
|
+
from ... import check
|
40
|
+
from .abc import Compressor
|
41
|
+
from .abc import NeedsInputDecompressor
|
42
|
+
from .abc import UnconsumedTailDecompressor
|
43
|
+
from .types import IncrementalCompressor
|
44
|
+
from .types import IncrementalDecompressor
|
45
|
+
|
46
|
+
|
47
|
+
##
|
48
|
+
|
49
|
+
|
50
|
+
class CompressorIncrementalAdapter:
|
51
|
+
def __init__(
|
52
|
+
self,
|
53
|
+
factory: ta.Callable[..., Compressor],
|
54
|
+
) -> None:
|
55
|
+
super().__init__()
|
56
|
+
|
57
|
+
self._factory = factory
|
58
|
+
|
59
|
+
def __call__(self) -> IncrementalCompressor:
|
60
|
+
compressor = self._factory()
|
61
|
+
|
62
|
+
while True:
|
63
|
+
data = check.isinstance((yield None), bytes)
|
64
|
+
if not data:
|
65
|
+
break
|
66
|
+
|
67
|
+
compressed = compressor.compress(data)
|
68
|
+
if compressed:
|
69
|
+
check.none((yield compressed))
|
70
|
+
|
71
|
+
if (fl := compressor.flush()):
|
72
|
+
check.none((yield fl))
|
73
|
+
|
74
|
+
check.none((yield b''))
|
75
|
+
|
76
|
+
|
77
|
+
##
|
78
|
+
|
79
|
+
|
80
|
+
class DecompressorIncrementalAdapter:
|
81
|
+
def __init__(
|
82
|
+
self,
|
83
|
+
factory: ta.Callable[..., NeedsInputDecompressor | UnconsumedTailDecompressor],
|
84
|
+
*,
|
85
|
+
trailing_error: type[BaseException] | tuple[type[BaseException], ...] = (),
|
86
|
+
) -> None:
|
87
|
+
super().__init__()
|
88
|
+
|
89
|
+
self._factory = factory
|
90
|
+
self._trailing_error = trailing_error
|
91
|
+
|
92
|
+
def __call__(self) -> IncrementalDecompressor:
|
93
|
+
pos = 0
|
94
|
+
|
95
|
+
data = None # Default if EOF is encountered
|
96
|
+
|
97
|
+
decompressor = self._factory()
|
98
|
+
|
99
|
+
while True:
|
100
|
+
# Depending on the input data, our call to the decompressor may not return any data. In this case, try again
|
101
|
+
# after reading another block.
|
102
|
+
while True:
|
103
|
+
if decompressor.eof:
|
104
|
+
rawblock = decompressor.unused_data
|
105
|
+
if not rawblock:
|
106
|
+
rawblock = check.isinstance((yield None), bytes)
|
107
|
+
if not rawblock:
|
108
|
+
break
|
109
|
+
|
110
|
+
# Continue to next stream.
|
111
|
+
decompressor = self._factory()
|
112
|
+
|
113
|
+
try:
|
114
|
+
data = decompressor.decompress(rawblock)
|
115
|
+
except self._trailing_error:
|
116
|
+
# Trailing data isn't a valid compressed stream; ignore it.
|
117
|
+
break
|
118
|
+
|
119
|
+
else:
|
120
|
+
if hasattr(decompressor, 'needs_input'):
|
121
|
+
if decompressor.needs_input:
|
122
|
+
rawblock = check.isinstance((yield None), bytes)
|
123
|
+
if not rawblock:
|
124
|
+
raise EOFError('Compressed file ended before the end-of-stream marker was reached')
|
125
|
+
else:
|
126
|
+
rawblock = b''
|
127
|
+
|
128
|
+
elif hasattr(decompressor, 'unconsumed_tail'):
|
129
|
+
if not (rawblock := decompressor.unconsumed_tail):
|
130
|
+
rawblock = check.isinstance((yield None), bytes)
|
131
|
+
if not rawblock:
|
132
|
+
raise EOFError('Compressed file ended before the end-of-stream marker was reached')
|
133
|
+
|
134
|
+
else:
|
135
|
+
raise TypeError(decompressor)
|
136
|
+
|
137
|
+
data = decompressor.decompress(rawblock)
|
138
|
+
|
139
|
+
if data:
|
140
|
+
break
|
141
|
+
|
142
|
+
if not data:
|
143
|
+
check.none((yield b''))
|
144
|
+
return
|
145
|
+
|
146
|
+
pos += len(data)
|
147
|
+
check.none((yield data))
|
@@ -0,0 +1,42 @@
|
|
1
|
+
import functools
|
2
|
+
import typing as ta
|
3
|
+
|
4
|
+
from ... import lang
|
5
|
+
from .adapters import CompressorIncrementalAdapter
|
6
|
+
from .adapters import DecompressorIncrementalAdapter
|
7
|
+
from .types import IncrementalCompressor
|
8
|
+
from .types import IncrementalDecompressor
|
9
|
+
|
10
|
+
|
11
|
+
if ta.TYPE_CHECKING:
|
12
|
+
import bz2
|
13
|
+
else:
|
14
|
+
bz2 = lang.proxy_import('bz2')
|
15
|
+
|
16
|
+
|
17
|
+
class IncrementalBz2Compressor:
|
18
|
+
def __init__(
|
19
|
+
self,
|
20
|
+
*,
|
21
|
+
compresslevel: int = 9,
|
22
|
+
) -> None:
|
23
|
+
super().__init__()
|
24
|
+
|
25
|
+
self._compresslevel = compresslevel
|
26
|
+
|
27
|
+
@lang.autostart
|
28
|
+
def __call__(self) -> IncrementalCompressor:
|
29
|
+
return CompressorIncrementalAdapter(
|
30
|
+
functools.partial(
|
31
|
+
bz2.BZ2Compressor, # type: ignore
|
32
|
+
self._compresslevel,
|
33
|
+
),
|
34
|
+
)()
|
35
|
+
|
36
|
+
|
37
|
+
class IncrementalBz2Decompressor:
|
38
|
+
def __call__(self) -> IncrementalDecompressor:
|
39
|
+
return DecompressorIncrementalAdapter(
|
40
|
+
bz2.BZ2Decompressor, # type: ignore
|
41
|
+
trailing_error=OSError,
|
42
|
+
)()
|