omlish 0.0.0.dev136__py3-none-any.whl → 0.0.0.dev138__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (34) hide show
  1. omlish/__about__.py +2 -2
  2. omlish/cached.py +2 -2
  3. omlish/collections/mappings.py +1 -1
  4. omlish/configs/flattening.py +1 -1
  5. omlish/diag/_pycharm/runhack.py +3 -0
  6. omlish/formats/json/stream/errors.py +2 -0
  7. omlish/formats/json/stream/lex.py +11 -5
  8. omlish/formats/json/stream/parse.py +37 -21
  9. omlish/funcs/genmachine.py +5 -4
  10. omlish/io/compress/__init__.py +0 -0
  11. omlish/io/compress/abc.py +104 -0
  12. omlish/io/compress/adapters.py +147 -0
  13. omlish/io/compress/bz2.py +42 -0
  14. omlish/io/compress/gzip.py +306 -0
  15. omlish/io/compress/lzma.py +32 -0
  16. omlish/io/compress/types.py +29 -0
  17. omlish/io/generators/__init__.py +0 -0
  18. omlish/io/generators/readers.py +183 -0
  19. omlish/io/generators/stepped.py +104 -0
  20. omlish/lang/__init__.py +11 -1
  21. omlish/lang/functions.py +0 -2
  22. omlish/lang/generators.py +243 -0
  23. omlish/lang/iterables.py +28 -51
  24. omlish/lang/maybes.py +4 -4
  25. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/METADATA +1 -1
  26. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/RECORD +34 -22
  27. /omlish/collections/{_abc.py → abc.py} +0 -0
  28. /omlish/io/{_abc.py → abc.py} +0 -0
  29. /omlish/logs/{_abc.py → abc.py} +0 -0
  30. /omlish/sql/{_abc.py → abc.py} +0 -0
  31. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/LICENSE +0 -0
  32. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/WHEEL +0 -0
  33. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/entry_points.txt +0 -0
  34. {omlish-0.0.0.dev136.dist-info → omlish-0.0.0.dev138.dist-info}/top_level.txt +0 -0
omlish/__about__.py CHANGED
@@ -1,5 +1,5 @@
1
- __version__ = '0.0.0.dev136'
2
- __revision__ = '345b1984bd3c96dc1b3ed4db8402e21d989f38ee'
1
+ __version__ = '0.0.0.dev138'
2
+ __revision__ = 'd7ba57140bc56bdfb990684087dcda294928835f'
3
3
 
4
4
 
5
5
  #
omlish/cached.py CHANGED
@@ -10,9 +10,9 @@ builtins and thus not distinguish it from a normal property.
10
10
 
11
11
  """
12
12
  from .lang.cached import _CachedProperty # noqa
13
- from .lang.cached import cached_function
13
+ from .lang.cached import cached_function as _cached_function
14
14
 
15
- function = cached_function
15
+ function = _cached_function
16
16
 
17
17
  property = property # noqa
18
18
 
@@ -86,7 +86,7 @@ class DynamicTypeMap(ta.Generic[V]):
86
86
  self._items = list(items)
87
87
  self._weak = bool(weak)
88
88
 
89
- self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary()
89
+ self._cache: ta.MutableMapping[type, ta.Any] = weakref.WeakKeyDictionary() if weak else {}
90
90
 
91
91
  @property
92
92
  def items(self) -> ta.Sequence[V]:
@@ -130,7 +130,7 @@ class Flattening:
130
130
  .split(self._index_close + self._index_open):
131
131
  yield int(p)
132
132
  else:
133
- check.state(')' not in part)
133
+ check.state(self._index_close not in part)
134
134
  yield part
135
135
 
136
136
  for fk, v in flattened.items():
@@ -3,7 +3,10 @@
3
3
  .venv/bin/python $(curl -LsSf https://raw.githubusercontent.com/wrmsr/omlish/master/omlish/diag/_pycharm/runhack.py -o $(mktemp) && echo "$_") install
4
4
 
5
5
  ==
6
+ TODO:
7
+ - check for existing files - can't run regular dep entrypoints now
6
8
 
9
+ ==
7
10
  See:
8
11
  - https://github.com/JetBrains/intellij-community/blob/6400f70dde6f743e39a257a5a78cc51b644c835e/python/helpers/pycharm/_jb_pytest_runner.py
9
12
  - https://github.com/JetBrains/intellij-community/blob/5a4e584aa59767f2e7cf4bd377adfaaf7503984b/python/helpers/pycharm/_jb_runner_tools.py
@@ -0,0 +1,2 @@
1
+ class JsonStreamError(Exception):
2
+ pass
@@ -3,6 +3,7 @@ TODO:
3
3
  - max buf size
4
4
  - max recursion depth
5
5
  - mark start pos of tokens, currently returning end
6
+ - _do_string inner loop optimization somehow
6
7
  """
7
8
  import dataclasses as dc
8
9
  import io
@@ -12,6 +13,7 @@ import typing as ta
12
13
 
13
14
  from .... import check
14
15
  from ....funcs.genmachine import GenMachine
16
+ from .errors import JsonStreamError
15
17
 
16
18
 
17
19
  ##
@@ -95,7 +97,7 @@ CONST_TOKENS: ta.Mapping[str, tuple[TokenKind, str | float | None]] = {
95
97
 
96
98
 
97
99
  @dc.dataclass()
98
- class JsonLexError(Exception):
100
+ class JsonStreamLexError(JsonStreamError):
99
101
  message: str
100
102
 
101
103
  pos: Position
@@ -160,8 +162,8 @@ class JsonStreamLexer(GenMachine[str, Token]):
160
162
  self._buf.truncate()
161
163
  return raw
162
164
 
163
- def _raise(self, msg: str) -> ta.NoReturn:
164
- raise JsonLexError(msg, self.pos)
165
+ def _raise(self, msg: str, src: Exception | None = None) -> ta.NoReturn:
166
+ raise JsonStreamLexError(msg, self.pos) from src
165
167
 
166
168
  def _do_main(self):
167
169
  while True:
@@ -202,7 +204,7 @@ class JsonStreamLexer(GenMachine[str, Token]):
202
204
  self._raise('Unexpected end of input')
203
205
 
204
206
  if not c:
205
- raise NotImplementedError
207
+ self._raise(f'Unterminated string literal: {self._buf.getvalue()}')
206
208
 
207
209
  self._buf.write(c)
208
210
  if c == '"' and last != '\\':
@@ -210,7 +212,11 @@ class JsonStreamLexer(GenMachine[str, Token]):
210
212
  last = c
211
213
 
212
214
  raw = self._flip_buf()
213
- sv = json.loads(raw)
215
+ try:
216
+ sv = json.loads(raw)
217
+ except json.JSONDecodeError as e:
218
+ self._raise(f'Invalid string literal: {raw!r}', e)
219
+
214
220
  yield self._make_tok('STRING', sv, raw, pos)
215
221
 
216
222
  return self._do_main()
@@ -1,9 +1,12 @@
1
+ import dataclasses as dc
1
2
  import typing as ta
2
3
 
3
4
  from .... import lang
4
5
  from ....funcs.genmachine import GenMachine
6
+ from .errors import JsonStreamError
5
7
  from .lex import SCALAR_VALUE_TYPES
6
8
  from .lex import VALUE_TOKEN_KINDS
9
+ from .lex import Position
7
10
  from .lex import ScalarValue
8
11
  from .lex import Token
9
12
 
@@ -79,6 +82,13 @@ def yield_parser_events(obj: ta.Any) -> ta.Generator[JsonStreamParserEvent, None
79
82
  ##
80
83
 
81
84
 
85
+ @dc.dataclass()
86
+ class JsonStreamParseError(JsonStreamError):
87
+ message: str
88
+
89
+ pos: Position | None = None
90
+
91
+
82
92
  class JsonStreamObject(list):
83
93
  def __repr__(self) -> str:
84
94
  return f'{self.__class__.__name__}({super().__repr__()})'
@@ -100,29 +110,29 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
100
110
  if tt == 'KEY':
101
111
  self._stack.pop()
102
112
  if not self._stack:
103
- raise self.StateError
113
+ raise JsonStreamParseError('Unexpected key')
104
114
 
105
115
  tt2 = self._stack[-1]
106
116
  if tt2 == 'OBJECT':
107
117
  return ((v,), self._do_after_pair())
108
118
 
109
119
  else:
110
- raise self.StateError
120
+ raise JsonStreamParseError('Unexpected key')
111
121
 
112
122
  elif tt == 'ARRAY':
113
123
  return ((v,), self._do_after_element())
114
124
 
115
125
  else:
116
- raise self.StateError
126
+ raise JsonStreamParseError(f'Unexpected value: {v!r}')
117
127
 
118
128
  #
119
129
 
120
- def _do_value(self):
130
+ def _do_value(self, *, must_be_present: bool = False):
121
131
  try:
122
132
  tok = yield None
123
133
  except GeneratorExit:
124
134
  if self._stack:
125
- raise self.StateError from None
135
+ raise JsonStreamParseError('Expected value') from None
126
136
  else:
127
137
  raise
128
138
 
@@ -141,13 +151,16 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
141
151
  yield y
142
152
  return r
143
153
 
154
+ elif must_be_present:
155
+ raise JsonStreamParseError('Expected value', tok.pos)
156
+
144
157
  elif tok.kind == 'RBRACKET':
145
158
  y, r = self._emit_end_array()
146
159
  yield y
147
160
  return r
148
161
 
149
162
  else:
150
- raise self.StateError
163
+ raise JsonStreamParseError('Expected value', tok.pos)
151
164
 
152
165
  #
153
166
 
@@ -157,19 +170,19 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
157
170
 
158
171
  def _emit_end_object(self):
159
172
  if not self._stack:
160
- raise self.StateError
173
+ raise JsonStreamParseError('Unexpected end object')
161
174
 
162
175
  tt = self._stack.pop()
163
176
  if tt != 'OBJECT':
164
- raise self.StateError
177
+ raise JsonStreamParseError('Unexpected end object')
165
178
 
166
179
  return self._emit_event(EndObject)
167
180
 
168
- def _do_object_body(self):
181
+ def _do_object_body(self, *, must_be_present: bool = False):
169
182
  try:
170
183
  tok = yield None
171
184
  except GeneratorExit:
172
- raise self.StateError from None
185
+ raise JsonStreamParseError('Expected object body') from None
173
186
 
174
187
  if tok.kind == 'STRING':
175
188
  k = tok.value
@@ -177,30 +190,33 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
177
190
  try:
178
191
  tok = yield None
179
192
  except GeneratorExit:
180
- raise self.StateError from None
193
+ raise JsonStreamParseError('Expected key') from None
181
194
  if tok.kind != 'COLON':
182
- raise self.StateError
195
+ raise JsonStreamParseError('Expected colon', tok.pos)
183
196
 
184
197
  yield (Key(k),)
185
198
  self._stack.append('KEY')
186
199
  return self._do_value()
187
200
 
201
+ elif must_be_present:
202
+ raise JsonStreamParseError('Expected value', tok.pos)
203
+
188
204
  elif tok.kind == 'RBRACE':
189
205
  y, r = self._emit_end_object()
190
206
  yield y
191
207
  return r
192
208
 
193
209
  else:
194
- raise self.StateError
210
+ raise JsonStreamParseError('Expected value', tok.pos)
195
211
 
196
212
  def _do_after_pair(self):
197
213
  try:
198
214
  tok = yield None
199
215
  except GeneratorExit:
200
- raise self.StateError from None
216
+ raise JsonStreamParseError('Expected continuation') from None
201
217
 
202
218
  if tok.kind == 'COMMA':
203
- return self._do_object_body()
219
+ return self._do_object_body(must_be_present=True)
204
220
 
205
221
  elif tok.kind == 'RBRACE':
206
222
  y, r = self._emit_end_object()
@@ -208,7 +224,7 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
208
224
  return r
209
225
 
210
226
  else:
211
- raise self.StateError
227
+ raise JsonStreamParseError('Expected continuation', tok.pos)
212
228
 
213
229
  #
214
230
 
@@ -218,11 +234,11 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
218
234
 
219
235
  def _emit_end_array(self):
220
236
  if not self._stack:
221
- raise self.StateError
237
+ raise JsonStreamParseError('Expected end array')
222
238
 
223
239
  tt = self._stack.pop()
224
240
  if tt != 'ARRAY':
225
- raise self.StateError
241
+ raise JsonStreamParseError('Unexpected end array')
226
242
 
227
243
  return self._emit_event(EndArray)
228
244
 
@@ -230,10 +246,10 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
230
246
  try:
231
247
  tok = yield None
232
248
  except GeneratorExit:
233
- raise self.StateError from None
249
+ raise JsonStreamParseError('Expected continuation') from None
234
250
 
235
251
  if tok.kind == 'COMMA':
236
- return self._do_value()
252
+ return self._do_value(must_be_present=True)
237
253
 
238
254
  elif tok.kind == 'RBRACKET':
239
255
  y, r = self._emit_end_array()
@@ -241,4 +257,4 @@ class JsonStreamParser(GenMachine[Token, JsonStreamParserEvent]):
241
257
  return r
242
258
 
243
259
  else:
244
- raise self.StateError
260
+ raise JsonStreamParseError('Expected continuation', tok.pos)
@@ -12,7 +12,7 @@ import typing as ta
12
12
  I = ta.TypeVar('I')
13
13
  O = ta.TypeVar('O')
14
14
 
15
- # MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I, ta.Optional[MachineGen[I, O]]]
15
+ # MachineGen: ta.TypeAlias = ta.Generator[ta.Iterable[O] | None, I | None, ta.Optional[MachineGen[I, O]]]
16
16
  MachineGen: ta.TypeAlias = ta.Generator[ta.Any, ta.Any, ta.Any]
17
17
 
18
18
 
@@ -67,8 +67,7 @@ class GenMachine(ta.Generic[I, O]):
67
67
  return self
68
68
 
69
69
  def __exit__(self, exc_type, exc_val, exc_tb):
70
- if exc_type is None:
71
- self.close()
70
+ self.close()
72
71
 
73
72
  #
74
73
 
@@ -93,8 +92,10 @@ class GenMachine(ta.Generic[I, O]):
93
92
  if self._gen is None:
94
93
  raise GenMachine.ClosedError
95
94
 
95
+ gi: I | None = i
96
96
  try:
97
- while (o := self._gen.send(i)) is not None:
97
+ while (o := self._gen.send(gi)) is not None:
98
+ gi = None
98
99
  yield from o
99
100
 
100
101
  except StopIteration as s:
File without changes
@@ -0,0 +1,104 @@
1
+ """
2
+ https://docs.python.org/3/library/bz2.html#bz2.BZ2Compressor
3
+ https://docs.python.org/3/library/zlib.html#zlib.decompressobj
4
+ https://docs.python.org/3/library/lzma.html#lzma.LZMADecompressor
5
+ """
6
+ import abc
7
+
8
+
9
+ ##
10
+
11
+
12
+ class Compressor(abc.ABC):
13
+ @abc.abstractmethod
14
+ def compress(self, data: bytes) -> bytes:
15
+ """
16
+ Provide data to the compressor object. Returns a chunk of compressed data if possible, or an empty byte string
17
+ otherwise.
18
+
19
+ When you have finished providing data to the compressor, call the flush() method to finish the compression
20
+ process.
21
+ """
22
+
23
+ raise NotImplementedError
24
+
25
+ @abc.abstractmethod
26
+ def flush(self) -> bytes:
27
+ """
28
+ Finish the compression process. Returns the compressed data left in internal buffers.
29
+
30
+ The compressor object may not be used after this method has been called.
31
+ """
32
+
33
+ raise NotImplementedError
34
+
35
+
36
+ ##
37
+
38
+
39
+ class Decompressor(abc.ABC):
40
+ @property
41
+ @abc.abstractmethod
42
+ def unused_data(self) -> bytes:
43
+ """
44
+ Data found after the end of the compressed stream.
45
+
46
+ If this attribute is accessed before the end of the stream has been reached, its value will be b''.
47
+ """
48
+
49
+ raise NotImplementedError
50
+
51
+ @property
52
+ @abc.abstractmethod
53
+ def eof(self) -> bool:
54
+ """True if the end-of-stream marker has been reached."""
55
+
56
+ raise NotImplementedError
57
+
58
+ @abc.abstractmethod
59
+ def decompress(self, data: bytes, *max_length: int) -> bytes:
60
+ """
61
+ Decompress data, returning a bytes object containing the uncompressed data corresponding to at least part of the
62
+ data in string. This data should be concatenated to the output produced by any preceding calls to the
63
+ decompress() method. Some of the input data may be preserved in internal buffers for later processing.
64
+
65
+ If the optional parameter max_length is non-zero then the return value will be no longer than max_length.
66
+ """
67
+
68
+ raise NotImplementedError
69
+
70
+
71
+ class NeedsInputDecompressor(Decompressor):
72
+ """
73
+ Used by:
74
+ - bz2.BZ2Decompressor
75
+ - lzma.LZMADecompressor
76
+ """
77
+
78
+ @property
79
+ @abc.abstractmethod
80
+ def needs_input(self) -> bool:
81
+ """
82
+ False if the decompress() method can provide more decompressed data before requiring new uncompressed input.
83
+ """
84
+
85
+ raise NotImplementedError
86
+
87
+
88
+ class UnconsumedTailDecompressor(Decompressor):
89
+ """
90
+ Used by:
91
+ - zlib.decompressobj
92
+ """
93
+
94
+ @property
95
+ @abc.abstractmethod
96
+ def unconsumed_tail(self) -> bytes:
97
+ """
98
+ A bytes object that contains any data that was not consumed by the last decompress() call because it exceeded
99
+ the limit for the uncompressed data buffer. This data has not yet been seen by the zlib machinery, so you must
100
+ feed it (possibly with further data concatenated to it) back to a subsequent decompress() method call in order
101
+ to get correct output.
102
+ """
103
+
104
+ raise NotImplementedError
@@ -0,0 +1,147 @@
1
+ # PYTHON SOFTWARE FOUNDATION LICENSE VERSION 2
2
+ # --------------------------------------------
3
+ #
4
+ # 1. This LICENSE AGREEMENT is between the Python Software Foundation ("PSF"), and the Individual or Organization
5
+ # ("Licensee") accessing and otherwise using this software ("Python") in source or binary form and its associated
6
+ # documentation.
7
+ #
8
+ # 2. Subject to the terms and conditions of this License Agreement, PSF hereby grants Licensee a nonexclusive,
9
+ # royalty-free, world-wide license to reproduce, analyze, test, perform and/or display publicly, prepare derivative
10
+ # works, distribute, and otherwise use Python alone or in any derivative version, provided, however, that PSF's License
11
+ # Agreement and PSF's notice of copyright, i.e., "Copyright (c) 2001, 2002, 2003, 2004, 2005, 2006, 2007, 2008, 2009,
12
+ # 2010, 2011, 2012, 2013, 2014, 2015, 2016, 2017 Python Software Foundation; All Rights Reserved" are retained in Python
13
+ # alone or in any derivative version prepared by Licensee.
14
+ #
15
+ # 3. In the event Licensee prepares a derivative work that is based on or incorporates Python or any part thereof, and
16
+ # wants to make the derivative work available to others as provided herein, then Licensee hereby agrees to include in
17
+ # any such work a brief summary of the changes made to Python.
18
+ #
19
+ # 4. PSF is making Python available to Licensee on an "AS IS" basis. PSF MAKES NO REPRESENTATIONS OR WARRANTIES,
20
+ # EXPRESS OR IMPLIED. BY WAY OF EXAMPLE, BUT NOT LIMITATION, PSF MAKES NO AND DISCLAIMS ANY REPRESENTATION OR WARRANTY
21
+ # OF MERCHANTABILITY OR FITNESS FOR ANY PARTICULAR PURPOSE OR THAT THE USE OF PYTHON WILL NOT INFRINGE ANY THIRD PARTY
22
+ # RIGHTS.
23
+ #
24
+ # 5. PSF SHALL NOT BE LIABLE TO LICENSEE OR ANY OTHER USERS OF PYTHON FOR ANY INCIDENTAL, SPECIAL, OR CONSEQUENTIAL
25
+ # DAMAGES OR LOSS AS A RESULT OF MODIFYING, DISTRIBUTING, OR OTHERWISE USING PYTHON, OR ANY DERIVATIVE THEREOF, EVEN IF
26
+ # ADVISED OF THE POSSIBILITY THEREOF.
27
+ #
28
+ # 6. This License Agreement will automatically terminate upon a material breach of its terms and conditions.
29
+ #
30
+ # 7. Nothing in this License Agreement shall be deemed to create any relationship of agency, partnership, or joint
31
+ # venture between PSF and Licensee. This License Agreement does not grant permission to use PSF trademarks or trade
32
+ # name in a trademark sense to endorse or promote products or services of Licensee, or any third party.
33
+ #
34
+ # 8. By copying, installing or otherwise using Python, Licensee agrees to be bound by the terms and conditions of this
35
+ # License Agreement.
36
+ # ~> https://github.com/python/cpython/blob/f19c50a4817ffebb26132182ed8cec6a72342cc0/Lib/_compression.py
37
+ import typing as ta
38
+
39
+ from ... import check
40
+ from .abc import Compressor
41
+ from .abc import NeedsInputDecompressor
42
+ from .abc import UnconsumedTailDecompressor
43
+ from .types import IncrementalCompressor
44
+ from .types import IncrementalDecompressor
45
+
46
+
47
+ ##
48
+
49
+
50
+ class CompressorIncrementalAdapter:
51
+ def __init__(
52
+ self,
53
+ factory: ta.Callable[..., Compressor],
54
+ ) -> None:
55
+ super().__init__()
56
+
57
+ self._factory = factory
58
+
59
+ def __call__(self) -> IncrementalCompressor:
60
+ compressor = self._factory()
61
+
62
+ while True:
63
+ data = check.isinstance((yield None), bytes)
64
+ if not data:
65
+ break
66
+
67
+ compressed = compressor.compress(data)
68
+ if compressed:
69
+ check.none((yield compressed))
70
+
71
+ if (fl := compressor.flush()):
72
+ check.none((yield fl))
73
+
74
+ check.none((yield b''))
75
+
76
+
77
+ ##
78
+
79
+
80
+ class DecompressorIncrementalAdapter:
81
+ def __init__(
82
+ self,
83
+ factory: ta.Callable[..., NeedsInputDecompressor | UnconsumedTailDecompressor],
84
+ *,
85
+ trailing_error: type[BaseException] | tuple[type[BaseException], ...] = (),
86
+ ) -> None:
87
+ super().__init__()
88
+
89
+ self._factory = factory
90
+ self._trailing_error = trailing_error
91
+
92
+ def __call__(self) -> IncrementalDecompressor:
93
+ pos = 0
94
+
95
+ data = None # Default if EOF is encountered
96
+
97
+ decompressor = self._factory()
98
+
99
+ while True:
100
+ # Depending on the input data, our call to the decompressor may not return any data. In this case, try again
101
+ # after reading another block.
102
+ while True:
103
+ if decompressor.eof:
104
+ rawblock = decompressor.unused_data
105
+ if not rawblock:
106
+ rawblock = check.isinstance((yield None), bytes)
107
+ if not rawblock:
108
+ break
109
+
110
+ # Continue to next stream.
111
+ decompressor = self._factory()
112
+
113
+ try:
114
+ data = decompressor.decompress(rawblock)
115
+ except self._trailing_error:
116
+ # Trailing data isn't a valid compressed stream; ignore it.
117
+ break
118
+
119
+ else:
120
+ if hasattr(decompressor, 'needs_input'):
121
+ if decompressor.needs_input:
122
+ rawblock = check.isinstance((yield None), bytes)
123
+ if not rawblock:
124
+ raise EOFError('Compressed file ended before the end-of-stream marker was reached')
125
+ else:
126
+ rawblock = b''
127
+
128
+ elif hasattr(decompressor, 'unconsumed_tail'):
129
+ if not (rawblock := decompressor.unconsumed_tail):
130
+ rawblock = check.isinstance((yield None), bytes)
131
+ if not rawblock:
132
+ raise EOFError('Compressed file ended before the end-of-stream marker was reached')
133
+
134
+ else:
135
+ raise TypeError(decompressor)
136
+
137
+ data = decompressor.decompress(rawblock)
138
+
139
+ if data:
140
+ break
141
+
142
+ if not data:
143
+ check.none((yield b''))
144
+ return
145
+
146
+ pos += len(data)
147
+ check.none((yield data))
@@ -0,0 +1,42 @@
1
+ import functools
2
+ import typing as ta
3
+
4
+ from ... import lang
5
+ from .adapters import CompressorIncrementalAdapter
6
+ from .adapters import DecompressorIncrementalAdapter
7
+ from .types import IncrementalCompressor
8
+ from .types import IncrementalDecompressor
9
+
10
+
11
+ if ta.TYPE_CHECKING:
12
+ import bz2
13
+ else:
14
+ bz2 = lang.proxy_import('bz2')
15
+
16
+
17
+ class IncrementalBz2Compressor:
18
+ def __init__(
19
+ self,
20
+ *,
21
+ compresslevel: int = 9,
22
+ ) -> None:
23
+ super().__init__()
24
+
25
+ self._compresslevel = compresslevel
26
+
27
+ @lang.autostart
28
+ def __call__(self) -> IncrementalCompressor:
29
+ return CompressorIncrementalAdapter(
30
+ functools.partial(
31
+ bz2.BZ2Compressor, # type: ignore
32
+ self._compresslevel,
33
+ ),
34
+ )()
35
+
36
+
37
+ class IncrementalBz2Decompressor:
38
+ def __call__(self) -> IncrementalDecompressor:
39
+ return DecompressorIncrementalAdapter(
40
+ bz2.BZ2Decompressor, # type: ignore
41
+ trailing_error=OSError,
42
+ )()