omlish 0.0.0.dev80__py3-none-any.whl → 0.0.0.dev82__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- omlish/__about__.py +4 -4
- omlish/dataclasses/impl/__init__.py +8 -0
- omlish/dataclasses/impl/params.py +3 -0
- omlish/dataclasses/impl/slots.py +61 -7
- omlish/formats/json/__init__.py +8 -1
- omlish/formats/json/backends/__init__.py +7 -0
- omlish/formats/json/backends/base.py +38 -0
- omlish/formats/json/backends/default.py +10 -0
- omlish/formats/json/backends/jiter.py +25 -0
- omlish/formats/json/backends/orjson.py +46 -2
- omlish/formats/json/backends/std.py +39 -0
- omlish/formats/json/backends/ujson.py +49 -0
- omlish/formats/json/cli.py +125 -31
- omlish/formats/json/consts.py +22 -0
- omlish/formats/json/encoding.py +17 -0
- omlish/formats/json/json.py +9 -39
- omlish/formats/json/render.py +49 -24
- omlish/formats/json/stream/__init__.py +0 -0
- omlish/formats/json/stream/build.py +113 -0
- omlish/formats/json/stream/lex.py +285 -0
- omlish/formats/json/stream/parse.py +244 -0
- omlish/formats/json/stream/render.py +119 -0
- omlish/genmachine.py +56 -10
- omlish/lang/resources.py +6 -1
- omlish/marshal/base.py +2 -0
- omlish/marshal/newtypes.py +24 -0
- omlish/marshal/standard.py +4 -0
- omlish/reflect/__init__.py +1 -0
- omlish/reflect/types.py +6 -1
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/METADATA +5 -5
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/RECORD +35 -24
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/LICENSE +0 -0
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/WHEEL +0 -0
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/entry_points.txt +0 -0
- {omlish-0.0.0.dev80.dist-info → omlish-0.0.0.dev82.dist-info}/top_level.txt +0 -0
omlish/__about__.py
CHANGED
@@ -1,5 +1,5 @@
|
|
1
|
-
__version__ = '0.0.0.
|
2
|
-
__revision__ = '
|
1
|
+
__version__ = '0.0.0.dev82'
|
2
|
+
__revision__ = '025f44593a2c0c59d06340e1d7dbcd15b1ab7684'
|
3
3
|
|
4
4
|
|
5
5
|
#
|
@@ -45,7 +45,7 @@ class Project(ProjectBase):
|
|
45
45
|
'lz4 ~= 4.3',
|
46
46
|
# 'lz4 @ git+https://github.com/wrmsr/python-lz4@wrmsr_20240830_GIL_NOT_USED'
|
47
47
|
|
48
|
-
'python-snappy ~= 0.7
|
48
|
+
'python-snappy ~= 0.7',
|
49
49
|
|
50
50
|
'zstd ~= 1.5',
|
51
51
|
],
|
@@ -97,7 +97,7 @@ class Project(ProjectBase):
|
|
97
97
|
|
98
98
|
'aiomysql ~= 0.2',
|
99
99
|
'aiosqlite ~= 0.20',
|
100
|
-
'asyncpg ~= 0.30
|
100
|
+
'asyncpg ~= 0.30',
|
101
101
|
|
102
102
|
'apsw ~= 3.46',
|
103
103
|
|
@@ -22,4 +22,12 @@ TODO:
|
|
22
22
|
- enums
|
23
23
|
- nodal
|
24
24
|
- embedding? forward kwargs in general? or only for replace?
|
25
|
+
|
26
|
+
TODO refs:
|
27
|
+
- batch up exec calls
|
28
|
+
- https://github.com/python/cpython/commit/8945b7ff55b87d11c747af2dad0e3e4d631e62d6
|
29
|
+
- add doc parameter to dataclasses.field
|
30
|
+
- https://github.com/python/cpython/commit/9c7657f09914254724683d91177aed7947637be5
|
31
|
+
- add decorator argument to make_dataclass
|
32
|
+
- https://github.com/python/cpython/commit/3e3a4d231518f91ff2f3c5a085b3849e32f1d548
|
25
33
|
"""
|
omlish/dataclasses/impl/slots.py
CHANGED
@@ -1,5 +1,7 @@
|
|
1
1
|
import dataclasses as dc
|
2
|
+
import inspect
|
2
3
|
import itertools
|
4
|
+
import types
|
3
5
|
|
4
6
|
|
5
7
|
MISSING = dc.MISSING
|
@@ -25,7 +27,7 @@ def _get_slots(cls):
|
|
25
27
|
slots = []
|
26
28
|
if getattr(cls, '__weakrefoffset__', -1) != 0:
|
27
29
|
slots.append('__weakref__')
|
28
|
-
if getattr(cls, '
|
30
|
+
if getattr(cls, '__dictoffset__', -1) != 0:
|
29
31
|
slots.append('__dict__')
|
30
32
|
yield from slots
|
31
33
|
case str(slot):
|
@@ -37,44 +39,96 @@ def _get_slots(cls):
|
|
37
39
|
raise TypeError(f"Slots of '{cls.__name__}' cannot be determined")
|
38
40
|
|
39
41
|
|
42
|
+
def _update_func_cell_for__class__(f, oldcls, newcls):
|
43
|
+
# Returns True if we update a cell, else False.
|
44
|
+
if f is None:
|
45
|
+
# f will be None in the case of a property where not all of fget, fset, and fdel are used. Nothing to do in
|
46
|
+
# that case.
|
47
|
+
return False
|
48
|
+
try:
|
49
|
+
idx = f.__code__.co_freevars.index('__class__')
|
50
|
+
except ValueError:
|
51
|
+
# This function doesn't reference __class__, so nothing to do.
|
52
|
+
return False
|
53
|
+
# Fix the cell to point to the new class, if it's already pointing at the old class. I'm not convinced that the "is
|
54
|
+
# oldcls" test is needed, but other than performance can't hurt.
|
55
|
+
closure = f.__closure__[idx]
|
56
|
+
if closure.cell_contents is oldcls:
|
57
|
+
closure.cell_contents = newcls
|
58
|
+
return True
|
59
|
+
return False
|
60
|
+
|
61
|
+
|
40
62
|
def add_slots(
|
41
63
|
cls: type,
|
42
64
|
is_frozen: bool,
|
43
65
|
weakref_slot: bool,
|
44
66
|
) -> type:
|
67
|
+
# Need to create a new class, since we can't set __slots__ after a class has been created, and the @dataclass
|
68
|
+
# decorator is called after the class is created.
|
69
|
+
|
70
|
+
# Make sure __slots__ isn't already set.
|
45
71
|
if '__slots__' in cls.__dict__:
|
46
72
|
raise TypeError(f'{cls.__name__} already specifies __slots__')
|
47
73
|
|
74
|
+
# Create a new dict for our new class.
|
48
75
|
cls_dict = dict(cls.__dict__)
|
49
76
|
field_names = tuple(f.name for f in dc.fields(cls)) # noqa
|
50
77
|
|
78
|
+
# Make sure slots don't overlap with those in base classes.
|
51
79
|
inherited_slots = set(itertools.chain.from_iterable(map(_get_slots, cls.__mro__[1:-1])))
|
52
80
|
|
81
|
+
# The slots for our class. Remove slots from our base classes. Add '__weakref__' if weakref_slot was given, unless
|
82
|
+
# it is already present.
|
53
83
|
cls_dict['__slots__'] = tuple(
|
54
84
|
itertools.filterfalse(
|
55
85
|
inherited_slots.__contains__,
|
56
86
|
itertools.chain(
|
57
87
|
field_names,
|
88
|
+
# gh-93521: '__weakref__' also needs to be filtered out if already present in inherited_slots
|
58
89
|
('__weakref__',) if weakref_slot else (),
|
59
90
|
),
|
60
91
|
),
|
61
92
|
)
|
62
93
|
|
63
94
|
for field_name in field_names:
|
95
|
+
# Remove our attributes, if present. They'll still be available in _MARKER.
|
64
96
|
cls_dict.pop(field_name, None)
|
65
97
|
|
98
|
+
# Remove __dict__ itself.
|
66
99
|
cls_dict.pop('__dict__', None)
|
100
|
+
|
101
|
+
# Clear existing `__weakref__` descriptor, it belongs to a previous type:
|
67
102
|
cls_dict.pop('__weakref__', None)
|
68
103
|
|
69
104
|
qualname = getattr(cls, '__qualname__', None)
|
70
|
-
|
105
|
+
newcls = type(cls)(cls.__name__, cls.__bases__, cls_dict)
|
71
106
|
if qualname is not None:
|
72
|
-
|
107
|
+
newcls.__qualname__ = qualname
|
73
108
|
|
74
109
|
if is_frozen:
|
110
|
+
# Need this for pickling frozen classes with slots.
|
75
111
|
if '__getstate__' not in cls_dict:
|
76
|
-
|
112
|
+
newcls.__getstate__ = _dataclass_getstate # type: ignore
|
77
113
|
if '__setstate__' not in cls_dict:
|
78
|
-
|
79
|
-
|
80
|
-
|
114
|
+
newcls.__setstate__ = _dataclass_setstate # type: ignore
|
115
|
+
|
116
|
+
# Fix up any closures which reference __class__. This is used to fix zero argument super so that it points to the
|
117
|
+
# correct class (the newly created one, which we're returning) and not the original class. We can break out of this
|
118
|
+
# loop as soon as we make an update, since all closures for a class will share a given cell.
|
119
|
+
for member in newcls.__dict__.values():
|
120
|
+
# If this is a wrapped function, unwrap it.
|
121
|
+
member = inspect.unwrap(member)
|
122
|
+
if isinstance(member, types.FunctionType):
|
123
|
+
if _update_func_cell_for__class__(member, cls, newcls):
|
124
|
+
break
|
125
|
+
|
126
|
+
elif isinstance(member, property):
|
127
|
+
if (
|
128
|
+
_update_func_cell_for__class__(member.fget, cls, newcls) or
|
129
|
+
_update_func_cell_for__class__(member.fset, cls, newcls) or
|
130
|
+
_update_func_cell_for__class__(member.fdel, cls, newcls)
|
131
|
+
):
|
132
|
+
break
|
133
|
+
|
134
|
+
return newcls
|
omlish/formats/json/__init__.py
CHANGED
@@ -1,9 +1,16 @@
|
|
1
|
-
from .
|
1
|
+
from .consts import ( # noqa
|
2
2
|
COMPACT_KWARGS,
|
3
3
|
COMPACT_SEPARATORS,
|
4
4
|
PRETTY_INDENT,
|
5
5
|
PRETTY_KWARGS,
|
6
|
+
)
|
7
|
+
|
8
|
+
from .encoding import ( # noqa
|
9
|
+
decodes,
|
6
10
|
detect_encoding,
|
11
|
+
)
|
12
|
+
|
13
|
+
from .json import ( # noqa
|
7
14
|
dump,
|
8
15
|
dump_compact,
|
9
16
|
dump_pretty,
|
@@ -0,0 +1,38 @@
|
|
1
|
+
import abc
|
2
|
+
import typing as ta
|
3
|
+
|
4
|
+
from .... import lang
|
5
|
+
|
6
|
+
|
7
|
+
class Backend(lang.Abstract):
|
8
|
+
@abc.abstractmethod
|
9
|
+
def dump(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
10
|
+
raise NotImplementedError
|
11
|
+
|
12
|
+
@abc.abstractmethod
|
13
|
+
def dumps(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
14
|
+
raise NotImplementedError
|
15
|
+
|
16
|
+
@abc.abstractmethod
|
17
|
+
def load(self, fp: ta.Any, **kwargs: ta.Any) -> ta.Any:
|
18
|
+
raise NotImplementedError
|
19
|
+
|
20
|
+
@abc.abstractmethod
|
21
|
+
def loads(self, s: str | bytes | bytearray, **kwargs: ta.Any) -> ta.Any:
|
22
|
+
raise NotImplementedError
|
23
|
+
|
24
|
+
@abc.abstractmethod
|
25
|
+
def dump_pretty(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
26
|
+
raise NotImplementedError
|
27
|
+
|
28
|
+
@abc.abstractmethod
|
29
|
+
def dumps_pretty(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
30
|
+
raise NotImplementedError
|
31
|
+
|
32
|
+
@abc.abstractmethod
|
33
|
+
def dump_compact(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
34
|
+
raise NotImplementedError
|
35
|
+
|
36
|
+
@abc.abstractmethod
|
37
|
+
def dumps_compact(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
38
|
+
raise NotImplementedError
|
@@ -0,0 +1,25 @@
|
|
1
|
+
"""
|
2
|
+
from_json(
|
3
|
+
json_data: bytes,
|
4
|
+
/,
|
5
|
+
*,
|
6
|
+
allow_inf_nan: bool = True,
|
7
|
+
cache_mode: Literal[True, False, "all", "keys", "none"] = "all",
|
8
|
+
partial_mode: Literal[True, False, "off", "on", "trailing-strings"] = False,
|
9
|
+
catch_duplicate_keys: bool = False,
|
10
|
+
lossless_floats: bool = False,
|
11
|
+
) -> Any:
|
12
|
+
json_data: The JSON data to parse
|
13
|
+
allow_inf_nan: Whether to allow infinity (`Infinity` an `-Infinity`) and `NaN` values to float fields.
|
14
|
+
Defaults to True.
|
15
|
+
cache_mode: cache Python strings to improve performance at the cost of some memory usage
|
16
|
+
- True / 'all' - cache all strings
|
17
|
+
- 'keys' - cache only object keys
|
18
|
+
- False / 'none' - cache nothing
|
19
|
+
partial_mode: How to handle incomplete strings:
|
20
|
+
- False / 'off' - raise an exception if the input is incomplete
|
21
|
+
- True / 'on' - allow incomplete JSON but discard the last string if it is incomplete
|
22
|
+
- 'trailing-strings' - allow incomplete JSON, and include the last incomplete string in the output
|
23
|
+
catch_duplicate_keys: if True, raise an exception if objects contain the same key multiple times
|
24
|
+
lossless_floats: if True, preserve full detail on floats using `LosslessFloat`
|
25
|
+
"""
|
@@ -1,11 +1,16 @@
|
|
1
1
|
"""
|
2
|
-
|
3
|
-
|
2
|
+
loads(obj: str | bytes | bytearray | memoryview) -> ta.Any | oj.JSONDEcodeError
|
3
|
+
dumps(
|
4
|
+
obj: ta.Any,
|
5
|
+
default: ta.Callable[[ta.Any], Ata.ny] | None = ...,
|
6
|
+
option: int | None = ...,
|
7
|
+
) -> bytes
|
4
8
|
"""
|
5
9
|
import dataclasses as dc
|
6
10
|
import typing as ta
|
7
11
|
|
8
12
|
from .... import lang
|
13
|
+
from .base import Backend
|
9
14
|
|
10
15
|
|
11
16
|
if ta.TYPE_CHECKING:
|
@@ -14,6 +19,9 @@ else:
|
|
14
19
|
oj = lang.proxy_import('orjson')
|
15
20
|
|
16
21
|
|
22
|
+
##
|
23
|
+
|
24
|
+
|
17
25
|
@dc.dataclass(frozen=True, kw_only=True)
|
18
26
|
class Options:
|
19
27
|
append_newline: bool = False # append \n to the output
|
@@ -70,3 +78,39 @@ class Options:
|
|
70
78
|
class DumpOpts:
|
71
79
|
default: ta.Callable[[ta.Any], ta.Any] | None = None
|
72
80
|
option: Options = Options()
|
81
|
+
|
82
|
+
|
83
|
+
##
|
84
|
+
|
85
|
+
|
86
|
+
class OrjsonBackend(Backend):
|
87
|
+
def dump(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
88
|
+
fp.write(self.dumps(obj, **kwargs))
|
89
|
+
|
90
|
+
def dumps(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
91
|
+
return oj.dumps(obj, **kwargs).decode('utf-8')
|
92
|
+
|
93
|
+
def load(self, fp: ta.Any, **kwargs: ta.Any) -> ta.Any:
|
94
|
+
return oj.loads(fp.read(), **kwargs)
|
95
|
+
|
96
|
+
def loads(self, s: str | bytes | bytearray, **kwargs: ta.Any) -> ta.Any:
|
97
|
+
return oj.loads(s, **kwargs)
|
98
|
+
|
99
|
+
def dump_pretty(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
100
|
+
fp.write(self.dumps_pretty(obj, **kwargs))
|
101
|
+
|
102
|
+
def dumps_pretty(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
103
|
+
return self.dumps(obj, option=kwargs.pop('option', 0) | oj.OPT_INDENT_2, **kwargs)
|
104
|
+
|
105
|
+
def dump_compact(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
106
|
+
return self.dump(obj, fp, **kwargs)
|
107
|
+
|
108
|
+
def dumps_compact(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
109
|
+
return self.dumps(obj, **kwargs)
|
110
|
+
|
111
|
+
|
112
|
+
ORJSON_BACKEND: OrjsonBackend | None
|
113
|
+
if lang.can_import('orjson'):
|
114
|
+
ORJSON_BACKEND = OrjsonBackend()
|
115
|
+
else:
|
116
|
+
ORJSON_BACKEND = None
|
@@ -8,6 +8,13 @@ import dataclasses as dc
|
|
8
8
|
import json
|
9
9
|
import typing as ta
|
10
10
|
|
11
|
+
from ..consts import COMPACT_KWARGS
|
12
|
+
from ..consts import PRETTY_KWARGS
|
13
|
+
from .base import Backend
|
14
|
+
|
15
|
+
|
16
|
+
##
|
17
|
+
|
11
18
|
|
12
19
|
@dc.dataclass(frozen=True, kw_only=True)
|
13
20
|
class DumpOpts:
|
@@ -35,3 +42,35 @@ class LoadOpts:
|
|
35
42
|
|
36
43
|
# called with the result of any object literal decoded with an ordered list of pairs, by default dict # noqa
|
37
44
|
object_pairs_hook: ta.Callable[[list[tuple[str, ta.Any]]], ta.Any] | None = None
|
45
|
+
|
46
|
+
|
47
|
+
##
|
48
|
+
|
49
|
+
|
50
|
+
class StdBackend(Backend):
|
51
|
+
def dump(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
52
|
+
json.dump(obj, fp, **kwargs)
|
53
|
+
|
54
|
+
def dumps(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
55
|
+
return json.dumps(obj, **kwargs)
|
56
|
+
|
57
|
+
def load(self, fp: ta.Any, **kwargs: ta.Any) -> ta.Any:
|
58
|
+
return json.load(fp, **kwargs)
|
59
|
+
|
60
|
+
def loads(self, s: str | bytes | bytearray, **kwargs: ta.Any) -> ta.Any:
|
61
|
+
return json.loads(s, **kwargs)
|
62
|
+
|
63
|
+
def dump_pretty(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
64
|
+
json.dump(obj, fp, **PRETTY_KWARGS, **kwargs)
|
65
|
+
|
66
|
+
def dumps_pretty(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
67
|
+
return json.dumps(obj, **PRETTY_KWARGS, **kwargs)
|
68
|
+
|
69
|
+
def dump_compact(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
70
|
+
json.dump(obj, fp, **COMPACT_KWARGS, **kwargs)
|
71
|
+
|
72
|
+
def dumps_compact(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
73
|
+
return json.dumps(obj, **COMPACT_KWARGS, **kwargs)
|
74
|
+
|
75
|
+
|
76
|
+
STD_BACKEND = StdBackend()
|
@@ -7,6 +7,19 @@ dumps(obj: ta.Any, **DumpOpts) -> None
|
|
7
7
|
import dataclasses as dc
|
8
8
|
import typing as ta
|
9
9
|
|
10
|
+
from .... import lang
|
11
|
+
from ..consts import PRETTY_INDENT
|
12
|
+
from .base import Backend
|
13
|
+
|
14
|
+
|
15
|
+
if ta.TYPE_CHECKING:
|
16
|
+
import ujson as uj
|
17
|
+
else:
|
18
|
+
uj = lang.proxy_import('ujson')
|
19
|
+
|
20
|
+
|
21
|
+
##
|
22
|
+
|
10
23
|
|
11
24
|
@dc.dataclass(frozen=True, kw_only=True)
|
12
25
|
class DumpOpts:
|
@@ -23,3 +36,39 @@ class DumpOpts:
|
|
23
36
|
reject_bytes: bool = True
|
24
37
|
default: ta.Callable[[ta.Any], ta.Any] | None = None # should return a serializable version of obj or raise TypeError # noqa
|
25
38
|
separators: tuple[str, str] | None = None
|
39
|
+
|
40
|
+
|
41
|
+
##
|
42
|
+
|
43
|
+
|
44
|
+
class UjsonBackend(Backend):
|
45
|
+
def dump(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
46
|
+
uj.dump(obj, fp, **kwargs)
|
47
|
+
|
48
|
+
def dumps(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
49
|
+
return uj.dumps(obj, **kwargs)
|
50
|
+
|
51
|
+
def load(self, fp: ta.Any, **kwargs: ta.Any) -> ta.Any:
|
52
|
+
return uj.load(fp, **kwargs)
|
53
|
+
|
54
|
+
def loads(self, s: str | bytes | bytearray, **kwargs: ta.Any) -> ta.Any:
|
55
|
+
return uj.loads(s, **kwargs)
|
56
|
+
|
57
|
+
def dump_pretty(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
58
|
+
uj.dump(obj, fp, indent=PRETTY_INDENT, **kwargs)
|
59
|
+
|
60
|
+
def dumps_pretty(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
61
|
+
return uj.dumps(obj, indent=PRETTY_INDENT, **kwargs)
|
62
|
+
|
63
|
+
def dump_compact(self, obj: ta.Any, fp: ta.Any, **kwargs: ta.Any) -> None:
|
64
|
+
uj.dump(obj, fp, **kwargs)
|
65
|
+
|
66
|
+
def dumps_compact(self, obj: ta.Any, **kwargs: ta.Any) -> str:
|
67
|
+
return uj.dumps(obj, **kwargs)
|
68
|
+
|
69
|
+
|
70
|
+
UJSON_BACKEND: UjsonBackend | None
|
71
|
+
if lang.can_import('ujson'):
|
72
|
+
UJSON_BACKEND = UjsonBackend()
|
73
|
+
else:
|
74
|
+
UJSON_BACKEND = None
|
omlish/formats/json/cli.py
CHANGED
@@ -1,15 +1,28 @@
|
|
1
|
+
"""
|
2
|
+
TODO:
|
3
|
+
- xml - [{"att", {"el", {"cdata", ...
|
4
|
+
- csv - dict if headers, array if not
|
5
|
+
"""
|
1
6
|
import argparse
|
7
|
+
import codecs
|
2
8
|
import contextlib
|
3
9
|
import dataclasses as dc
|
4
10
|
import enum
|
11
|
+
import io
|
5
12
|
import json
|
13
|
+
import os
|
6
14
|
import subprocess
|
7
15
|
import sys
|
8
16
|
import typing as ta
|
9
17
|
|
18
|
+
from ... import check
|
10
19
|
from ... import lang
|
11
20
|
from ... import term
|
12
21
|
from .render import JsonRenderer
|
22
|
+
from .stream.build import JsonObjectBuilder
|
23
|
+
from .stream.lex import JsonStreamLexer
|
24
|
+
from .stream.parse import JsonStreamParser
|
25
|
+
from .stream.render import StreamJsonRenderer
|
13
26
|
|
14
27
|
|
15
28
|
if ta.TYPE_CHECKING:
|
@@ -67,15 +80,26 @@ def _main() -> None:
|
|
67
80
|
parser = argparse.ArgumentParser()
|
68
81
|
|
69
82
|
parser.add_argument('file', nargs='?')
|
83
|
+
|
84
|
+
parser.add_argument('--stream', action='store_true')
|
85
|
+
parser.add_argument('--stream-build', action='store_true')
|
86
|
+
parser.add_argument('--stream-buffer-size', type=int, default=0x4000)
|
87
|
+
|
70
88
|
parser.add_argument('-f', '--format')
|
89
|
+
|
71
90
|
parser.add_argument('-z', '--compact', action='store_true')
|
72
91
|
parser.add_argument('-p', '--pretty', action='store_true')
|
73
92
|
parser.add_argument('-i', '--indent')
|
74
93
|
parser.add_argument('-s', '--sort-keys', action='store_true')
|
94
|
+
|
75
95
|
parser.add_argument('-c', '--color', action='store_true')
|
96
|
+
|
76
97
|
parser.add_argument('-l', '--less', action='store_true')
|
98
|
+
|
77
99
|
args = parser.parse_args()
|
78
100
|
|
101
|
+
#
|
102
|
+
|
79
103
|
separators = None
|
80
104
|
if args.compact:
|
81
105
|
separators = (',', ':')
|
@@ -89,6 +113,28 @@ def _main() -> None:
|
|
89
113
|
except ValueError:
|
90
114
|
indent = args.indent
|
91
115
|
|
116
|
+
kw: dict[str, ta.Any] = dict(
|
117
|
+
indent=indent,
|
118
|
+
separators=separators,
|
119
|
+
sort_keys=args.sort_keys,
|
120
|
+
)
|
121
|
+
|
122
|
+
def render_one(v: ta.Any) -> str:
|
123
|
+
if args.color:
|
124
|
+
return JsonRenderer.render_str(
|
125
|
+
v,
|
126
|
+
**kw,
|
127
|
+
style=term_color,
|
128
|
+
)
|
129
|
+
|
130
|
+
else:
|
131
|
+
return json.dumps(
|
132
|
+
v,
|
133
|
+
**kw,
|
134
|
+
)
|
135
|
+
|
136
|
+
#
|
137
|
+
|
92
138
|
fmt_name = args.format
|
93
139
|
if fmt_name is None:
|
94
140
|
if args.file is not None:
|
@@ -99,45 +145,93 @@ def _main() -> None:
|
|
99
145
|
fmt_name = 'json'
|
100
146
|
fmt = FORMATS_BY_NAME[fmt_name]
|
101
147
|
|
148
|
+
if args.stream:
|
149
|
+
check.arg(fmt is Formats.JSON.value)
|
150
|
+
|
151
|
+
#
|
152
|
+
|
102
153
|
with contextlib.ExitStack() as es:
|
103
154
|
if args.file is None:
|
104
|
-
in_file = sys.stdin
|
155
|
+
in_file = sys.stdin.buffer
|
156
|
+
|
105
157
|
else:
|
106
|
-
in_file = es.enter_context(open(args.file))
|
158
|
+
in_file = es.enter_context(open(args.file, 'rb'))
|
107
159
|
|
108
|
-
|
160
|
+
#
|
109
161
|
|
110
|
-
|
111
|
-
|
112
|
-
|
113
|
-
|
114
|
-
|
162
|
+
if args.less:
|
163
|
+
less = subprocess.Popen(
|
164
|
+
[
|
165
|
+
'less',
|
166
|
+
*(['-R'] if args.color else []),
|
167
|
+
],
|
168
|
+
stdin=subprocess.PIPE,
|
169
|
+
encoding='utf-8',
|
170
|
+
)
|
171
|
+
out = check.not_none(less.stdin)
|
115
172
|
|
116
|
-
|
117
|
-
|
118
|
-
|
119
|
-
**kw,
|
120
|
-
style=term_color,
|
121
|
-
)
|
173
|
+
def close_less():
|
174
|
+
out.close()
|
175
|
+
less.wait()
|
122
176
|
|
123
|
-
|
124
|
-
out = json.dumps(
|
125
|
-
data,
|
126
|
-
**kw,
|
127
|
-
)
|
128
|
-
|
129
|
-
if args.less:
|
130
|
-
subprocess.run(
|
131
|
-
[
|
132
|
-
'less',
|
133
|
-
*(['-R'] if args.color else []),
|
134
|
-
],
|
135
|
-
input=out.encode(),
|
136
|
-
check=True,
|
137
|
-
)
|
177
|
+
es.enter_context(lang.defer(close_less)) # noqa
|
138
178
|
|
139
|
-
|
140
|
-
|
179
|
+
else:
|
180
|
+
out = sys.stdout
|
181
|
+
|
182
|
+
#
|
183
|
+
|
184
|
+
if args.stream:
|
185
|
+
fd = in_file.fileno()
|
186
|
+
decoder = codecs.getincrementaldecoder('utf-8')()
|
187
|
+
|
188
|
+
with contextlib.ExitStack() as es2:
|
189
|
+
lex = es2.enter_context(JsonStreamLexer())
|
190
|
+
parse = es2.enter_context(JsonStreamParser())
|
191
|
+
|
192
|
+
if args.stream_build:
|
193
|
+
build = es2.enter_context(JsonObjectBuilder())
|
194
|
+
renderer = None
|
195
|
+
|
196
|
+
else:
|
197
|
+
renderer = StreamJsonRenderer(
|
198
|
+
out,
|
199
|
+
StreamJsonRenderer.Options(
|
200
|
+
**kw,
|
201
|
+
style=term_color if args.color else None,
|
202
|
+
),
|
203
|
+
)
|
204
|
+
build = None
|
205
|
+
|
206
|
+
while True:
|
207
|
+
buf = os.read(fd, args.stream_buffer_size)
|
208
|
+
|
209
|
+
for s in decoder.decode(buf, not buf):
|
210
|
+
n = 0
|
211
|
+
for c in s:
|
212
|
+
for t in lex(c):
|
213
|
+
for e in parse(t):
|
214
|
+
if renderer is not None:
|
215
|
+
renderer.render((e,))
|
216
|
+
|
217
|
+
if build is not None:
|
218
|
+
for v in build(e):
|
219
|
+
print(render_one(v), file=out)
|
220
|
+
|
221
|
+
n += 1
|
222
|
+
|
223
|
+
if n:
|
224
|
+
out.flush()
|
225
|
+
|
226
|
+
if not buf:
|
227
|
+
break
|
228
|
+
|
229
|
+
out.write('\n')
|
230
|
+
|
231
|
+
else:
|
232
|
+
with io.TextIOWrapper(in_file) as tw:
|
233
|
+
v = fmt.load(tw)
|
234
|
+
print(render_one(v), file=out)
|
141
235
|
|
142
236
|
|
143
237
|
if __name__ == '__main__':
|
@@ -0,0 +1,22 @@
|
|
1
|
+
import typing as ta
|
2
|
+
|
3
|
+
|
4
|
+
##
|
5
|
+
|
6
|
+
|
7
|
+
PRETTY_INDENT = 2
|
8
|
+
|
9
|
+
PRETTY_KWARGS: ta.Mapping[str, ta.Any] = dict(
|
10
|
+
indent=PRETTY_INDENT,
|
11
|
+
)
|
12
|
+
|
13
|
+
|
14
|
+
##
|
15
|
+
|
16
|
+
|
17
|
+
COMPACT_SEPARATORS = (',', ':')
|
18
|
+
|
19
|
+
COMPACT_KWARGS: ta.Mapping[str, ta.Any] = dict(
|
20
|
+
indent=None,
|
21
|
+
separators=COMPACT_SEPARATORS,
|
22
|
+
)
|