py2docfx 0.1.11.dev1824276__py3-none-any.whl → 0.1.11.dev1830301__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- py2docfx/convert_prepare/environment.py +1 -1
- py2docfx/convert_prepare/package_info.py +1 -1
- py2docfx/convert_prepare/pip_utils.py +1 -1
- py2docfx/convert_prepare/tests/test_package_info.py +7 -22
- py2docfx/convert_prepare/tests/test_params.py +5 -0
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/_collections.py +0 -145
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/bcppcompiler.py +1 -2
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/ccompiler.py +7 -11
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/bdist.py +4 -3
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/build_ext.py +1 -4
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/check.py +3 -4
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/install_data.py +39 -29
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/install_lib.py +1 -3
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/command/sdist.py +4 -4
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/compat/py38.py +1 -0
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/cygwinccompiler.py +21 -42
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/dist.py +3 -12
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/extension.py +9 -4
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/msvc9compiler.py +1 -3
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/msvccompiler.py +1 -3
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/spawn.py +0 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/sysconfig.py +44 -25
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_archive_util.py +1 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_build.py +1 -2
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_build_ext.py +1 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_cygwinccompiler.py +0 -42
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_dist.py +1 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_extension.py +4 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_install_data.py +17 -9
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_mingwccompiler.py +5 -4
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_spawn.py +1 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_sysconfig.py +6 -3
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_unixccompiler.py +35 -1
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/tests/test_util.py +4 -24
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/unixccompiler.py +19 -20
- py2docfx/venv/0/Lib/site-packages/setuptools/_distutils/util.py +20 -26
- py2docfx/venv/0/Lib/site-packages/wheel/__init__.py +3 -0
- py2docfx/venv/0/Lib/site-packages/wheel/__main__.py +23 -0
- py2docfx/venv/0/Lib/site-packages/wheel/_bdist_wheel.py +604 -0
- py2docfx/venv/0/Lib/site-packages/wheel/_setuptools_logging.py +26 -0
- py2docfx/venv/0/Lib/site-packages/wheel/bdist_wheel.py +11 -0
- py2docfx/venv/0/Lib/site-packages/wheel/cli/__init__.py +155 -0
- py2docfx/venv/0/Lib/site-packages/wheel/cli/convert.py +273 -0
- py2docfx/venv/0/Lib/site-packages/wheel/cli/pack.py +85 -0
- py2docfx/venv/0/Lib/site-packages/wheel/cli/tags.py +139 -0
- py2docfx/venv/0/Lib/site-packages/wheel/cli/unpack.py +30 -0
- py2docfx/venv/0/Lib/site-packages/wheel/macosx_libfile.py +482 -0
- py2docfx/venv/0/Lib/site-packages/wheel/metadata.py +183 -0
- py2docfx/venv/0/Lib/site-packages/wheel/util.py +26 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/__init__.py +0 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/__init__.py +0 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_elffile.py +108 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_manylinux.py +260 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_musllinux.py +83 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_parser.py +356 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_structures.py +61 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/_tokenizer.py +192 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/markers.py +253 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/requirements.py +90 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/specifiers.py +1011 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/tags.py +571 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/utils.py +172 -0
- py2docfx/venv/0/Lib/site-packages/wheel/vendored/packaging/version.py +561 -0
- py2docfx/venv/0/Lib/site-packages/wheel/wheelfile.py +227 -0
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/_collections.py +0 -145
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/bcppcompiler.py +1 -2
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/ccompiler.py +7 -11
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/bdist.py +4 -3
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/build_ext.py +1 -4
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/check.py +3 -4
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/install_data.py +39 -29
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/install_lib.py +1 -3
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/command/sdist.py +4 -4
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/compat/py38.py +1 -0
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/cygwinccompiler.py +21 -42
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/dist.py +3 -12
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/extension.py +9 -4
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/msvc9compiler.py +1 -3
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/msvccompiler.py +1 -3
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/spawn.py +0 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/sysconfig.py +44 -25
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_archive_util.py +1 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_build.py +1 -2
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_build_ext.py +1 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_cygwinccompiler.py +0 -42
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_dist.py +1 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_extension.py +4 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_install_data.py +17 -9
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_mingwccompiler.py +5 -4
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_spawn.py +1 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_sysconfig.py +6 -3
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_unixccompiler.py +35 -1
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/tests/test_util.py +4 -24
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/unixccompiler.py +19 -20
- py2docfx/venv/template/Lib/site-packages/setuptools/_distutils/util.py +20 -26
- py2docfx/venv/template/Lib/site-packages/wheel/__init__.py +3 -0
- py2docfx/venv/template/Lib/site-packages/wheel/__main__.py +23 -0
- py2docfx/venv/template/Lib/site-packages/wheel/_bdist_wheel.py +604 -0
- py2docfx/venv/template/Lib/site-packages/wheel/_setuptools_logging.py +26 -0
- py2docfx/venv/template/Lib/site-packages/wheel/bdist_wheel.py +11 -0
- py2docfx/venv/template/Lib/site-packages/wheel/cli/__init__.py +155 -0
- py2docfx/venv/template/Lib/site-packages/wheel/cli/convert.py +273 -0
- py2docfx/venv/template/Lib/site-packages/wheel/cli/pack.py +85 -0
- py2docfx/venv/template/Lib/site-packages/wheel/cli/tags.py +139 -0
- py2docfx/venv/template/Lib/site-packages/wheel/cli/unpack.py +30 -0
- py2docfx/venv/template/Lib/site-packages/wheel/macosx_libfile.py +482 -0
- py2docfx/venv/template/Lib/site-packages/wheel/metadata.py +183 -0
- py2docfx/venv/template/Lib/site-packages/wheel/util.py +26 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/__init__.py +0 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/__init__.py +0 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_elffile.py +108 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_manylinux.py +260 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_musllinux.py +83 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_parser.py +356 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_structures.py +61 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/_tokenizer.py +192 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/markers.py +253 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/requirements.py +90 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/specifiers.py +1011 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/tags.py +571 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/utils.py +172 -0
- py2docfx/venv/template/Lib/site-packages/wheel/vendored/packaging/version.py +561 -0
- py2docfx/venv/template/Lib/site-packages/wheel/wheelfile.py +227 -0
- {py2docfx-0.1.11.dev1824276.dist-info → py2docfx-0.1.11.dev1830301.dist-info}/METADATA +1 -1
- {py2docfx-0.1.11.dev1824276.dist-info → py2docfx-0.1.11.dev1830301.dist-info}/RECORD +127 -71
- {py2docfx-0.1.11.dev1824276.dist-info → py2docfx-0.1.11.dev1830301.dist-info}/WHEEL +0 -0
- {py2docfx-0.1.11.dev1824276.dist-info → py2docfx-0.1.11.dev1830301.dist-info}/top_level.txt +0 -0
@@ -0,0 +1,356 @@
|
|
1
|
+
"""Handwritten parser of dependency specifiers.
|
2
|
+
|
3
|
+
The docstring for each __parse_* function contains EBNF-inspired grammar representing
|
4
|
+
the implementation.
|
5
|
+
"""
|
6
|
+
|
7
|
+
import ast
|
8
|
+
from typing import Any, List, NamedTuple, Optional, Tuple, Union
|
9
|
+
|
10
|
+
from ._tokenizer import DEFAULT_RULES, Tokenizer
|
11
|
+
|
12
|
+
|
13
|
+
class Node:
|
14
|
+
def __init__(self, value: str) -> None:
|
15
|
+
self.value = value
|
16
|
+
|
17
|
+
def __str__(self) -> str:
|
18
|
+
return self.value
|
19
|
+
|
20
|
+
def __repr__(self) -> str:
|
21
|
+
return f"<{self.__class__.__name__}('{self}')>"
|
22
|
+
|
23
|
+
def serialize(self) -> str:
|
24
|
+
raise NotImplementedError
|
25
|
+
|
26
|
+
|
27
|
+
class Variable(Node):
|
28
|
+
def serialize(self) -> str:
|
29
|
+
return str(self)
|
30
|
+
|
31
|
+
|
32
|
+
class Value(Node):
|
33
|
+
def serialize(self) -> str:
|
34
|
+
return f'"{self}"'
|
35
|
+
|
36
|
+
|
37
|
+
class Op(Node):
|
38
|
+
def serialize(self) -> str:
|
39
|
+
return str(self)
|
40
|
+
|
41
|
+
|
42
|
+
MarkerVar = Union[Variable, Value]
|
43
|
+
MarkerItem = Tuple[MarkerVar, Op, MarkerVar]
|
44
|
+
# MarkerAtom = Union[MarkerItem, List["MarkerAtom"]]
|
45
|
+
# MarkerList = List[Union["MarkerList", MarkerAtom, str]]
|
46
|
+
# mypy does not support recursive type definition
|
47
|
+
# https://github.com/python/mypy/issues/731
|
48
|
+
MarkerAtom = Any
|
49
|
+
MarkerList = List[Any]
|
50
|
+
|
51
|
+
|
52
|
+
class ParsedRequirement(NamedTuple):
|
53
|
+
name: str
|
54
|
+
url: str
|
55
|
+
extras: List[str]
|
56
|
+
specifier: str
|
57
|
+
marker: Optional[MarkerList]
|
58
|
+
|
59
|
+
|
60
|
+
# --------------------------------------------------------------------------------------
|
61
|
+
# Recursive descent parser for dependency specifier
|
62
|
+
# --------------------------------------------------------------------------------------
|
63
|
+
def parse_requirement(source: str) -> ParsedRequirement:
|
64
|
+
return _parse_requirement(Tokenizer(source, rules=DEFAULT_RULES))
|
65
|
+
|
66
|
+
|
67
|
+
def _parse_requirement(tokenizer: Tokenizer) -> ParsedRequirement:
|
68
|
+
"""
|
69
|
+
requirement = WS? IDENTIFIER WS? extras WS? requirement_details
|
70
|
+
"""
|
71
|
+
tokenizer.consume("WS")
|
72
|
+
|
73
|
+
name_token = tokenizer.expect(
|
74
|
+
"IDENTIFIER", expected="package name at the start of dependency specifier"
|
75
|
+
)
|
76
|
+
name = name_token.text
|
77
|
+
tokenizer.consume("WS")
|
78
|
+
|
79
|
+
extras = _parse_extras(tokenizer)
|
80
|
+
tokenizer.consume("WS")
|
81
|
+
|
82
|
+
url, specifier, marker = _parse_requirement_details(tokenizer)
|
83
|
+
tokenizer.expect("END", expected="end of dependency specifier")
|
84
|
+
|
85
|
+
return ParsedRequirement(name, url, extras, specifier, marker)
|
86
|
+
|
87
|
+
|
88
|
+
def _parse_requirement_details(
|
89
|
+
tokenizer: Tokenizer,
|
90
|
+
) -> Tuple[str, str, Optional[MarkerList]]:
|
91
|
+
"""
|
92
|
+
requirement_details = AT URL (WS requirement_marker?)?
|
93
|
+
| specifier WS? (requirement_marker)?
|
94
|
+
"""
|
95
|
+
|
96
|
+
specifier = ""
|
97
|
+
url = ""
|
98
|
+
marker = None
|
99
|
+
|
100
|
+
if tokenizer.check("AT"):
|
101
|
+
tokenizer.read()
|
102
|
+
tokenizer.consume("WS")
|
103
|
+
|
104
|
+
url_start = tokenizer.position
|
105
|
+
url = tokenizer.expect("URL", expected="URL after @").text
|
106
|
+
if tokenizer.check("END", peek=True):
|
107
|
+
return (url, specifier, marker)
|
108
|
+
|
109
|
+
tokenizer.expect("WS", expected="whitespace after URL")
|
110
|
+
|
111
|
+
# The input might end after whitespace.
|
112
|
+
if tokenizer.check("END", peek=True):
|
113
|
+
return (url, specifier, marker)
|
114
|
+
|
115
|
+
marker = _parse_requirement_marker(
|
116
|
+
tokenizer, span_start=url_start, after="URL and whitespace"
|
117
|
+
)
|
118
|
+
else:
|
119
|
+
specifier_start = tokenizer.position
|
120
|
+
specifier = _parse_specifier(tokenizer)
|
121
|
+
tokenizer.consume("WS")
|
122
|
+
|
123
|
+
if tokenizer.check("END", peek=True):
|
124
|
+
return (url, specifier, marker)
|
125
|
+
|
126
|
+
marker = _parse_requirement_marker(
|
127
|
+
tokenizer,
|
128
|
+
span_start=specifier_start,
|
129
|
+
after=(
|
130
|
+
"version specifier"
|
131
|
+
if specifier
|
132
|
+
else "name and no valid version specifier"
|
133
|
+
),
|
134
|
+
)
|
135
|
+
|
136
|
+
return (url, specifier, marker)
|
137
|
+
|
138
|
+
|
139
|
+
def _parse_requirement_marker(
|
140
|
+
tokenizer: Tokenizer, *, span_start: int, after: str
|
141
|
+
) -> MarkerList:
|
142
|
+
"""
|
143
|
+
requirement_marker = SEMICOLON marker WS?
|
144
|
+
"""
|
145
|
+
|
146
|
+
if not tokenizer.check("SEMICOLON"):
|
147
|
+
tokenizer.raise_syntax_error(
|
148
|
+
f"Expected end or semicolon (after {after})",
|
149
|
+
span_start=span_start,
|
150
|
+
)
|
151
|
+
tokenizer.read()
|
152
|
+
|
153
|
+
marker = _parse_marker(tokenizer)
|
154
|
+
tokenizer.consume("WS")
|
155
|
+
|
156
|
+
return marker
|
157
|
+
|
158
|
+
|
159
|
+
def _parse_extras(tokenizer: Tokenizer) -> List[str]:
|
160
|
+
"""
|
161
|
+
extras = (LEFT_BRACKET wsp* extras_list? wsp* RIGHT_BRACKET)?
|
162
|
+
"""
|
163
|
+
if not tokenizer.check("LEFT_BRACKET", peek=True):
|
164
|
+
return []
|
165
|
+
|
166
|
+
with tokenizer.enclosing_tokens(
|
167
|
+
"LEFT_BRACKET",
|
168
|
+
"RIGHT_BRACKET",
|
169
|
+
around="extras",
|
170
|
+
):
|
171
|
+
tokenizer.consume("WS")
|
172
|
+
extras = _parse_extras_list(tokenizer)
|
173
|
+
tokenizer.consume("WS")
|
174
|
+
|
175
|
+
return extras
|
176
|
+
|
177
|
+
|
178
|
+
def _parse_extras_list(tokenizer: Tokenizer) -> List[str]:
|
179
|
+
"""
|
180
|
+
extras_list = identifier (wsp* ',' wsp* identifier)*
|
181
|
+
"""
|
182
|
+
extras: List[str] = []
|
183
|
+
|
184
|
+
if not tokenizer.check("IDENTIFIER"):
|
185
|
+
return extras
|
186
|
+
|
187
|
+
extras.append(tokenizer.read().text)
|
188
|
+
|
189
|
+
while True:
|
190
|
+
tokenizer.consume("WS")
|
191
|
+
if tokenizer.check("IDENTIFIER", peek=True):
|
192
|
+
tokenizer.raise_syntax_error("Expected comma between extra names")
|
193
|
+
elif not tokenizer.check("COMMA"):
|
194
|
+
break
|
195
|
+
|
196
|
+
tokenizer.read()
|
197
|
+
tokenizer.consume("WS")
|
198
|
+
|
199
|
+
extra_token = tokenizer.expect("IDENTIFIER", expected="extra name after comma")
|
200
|
+
extras.append(extra_token.text)
|
201
|
+
|
202
|
+
return extras
|
203
|
+
|
204
|
+
|
205
|
+
def _parse_specifier(tokenizer: Tokenizer) -> str:
|
206
|
+
"""
|
207
|
+
specifier = LEFT_PARENTHESIS WS? version_many WS? RIGHT_PARENTHESIS
|
208
|
+
| WS? version_many WS?
|
209
|
+
"""
|
210
|
+
with tokenizer.enclosing_tokens(
|
211
|
+
"LEFT_PARENTHESIS",
|
212
|
+
"RIGHT_PARENTHESIS",
|
213
|
+
around="version specifier",
|
214
|
+
):
|
215
|
+
tokenizer.consume("WS")
|
216
|
+
parsed_specifiers = _parse_version_many(tokenizer)
|
217
|
+
tokenizer.consume("WS")
|
218
|
+
|
219
|
+
return parsed_specifiers
|
220
|
+
|
221
|
+
|
222
|
+
def _parse_version_many(tokenizer: Tokenizer) -> str:
|
223
|
+
"""
|
224
|
+
version_many = (SPECIFIER (WS? COMMA WS? SPECIFIER)*)?
|
225
|
+
"""
|
226
|
+
parsed_specifiers = ""
|
227
|
+
while tokenizer.check("SPECIFIER"):
|
228
|
+
span_start = tokenizer.position
|
229
|
+
parsed_specifiers += tokenizer.read().text
|
230
|
+
if tokenizer.check("VERSION_PREFIX_TRAIL", peek=True):
|
231
|
+
tokenizer.raise_syntax_error(
|
232
|
+
".* suffix can only be used with `==` or `!=` operators",
|
233
|
+
span_start=span_start,
|
234
|
+
span_end=tokenizer.position + 1,
|
235
|
+
)
|
236
|
+
if tokenizer.check("VERSION_LOCAL_LABEL_TRAIL", peek=True):
|
237
|
+
tokenizer.raise_syntax_error(
|
238
|
+
"Local version label can only be used with `==` or `!=` operators",
|
239
|
+
span_start=span_start,
|
240
|
+
span_end=tokenizer.position,
|
241
|
+
)
|
242
|
+
tokenizer.consume("WS")
|
243
|
+
if not tokenizer.check("COMMA"):
|
244
|
+
break
|
245
|
+
parsed_specifiers += tokenizer.read().text
|
246
|
+
tokenizer.consume("WS")
|
247
|
+
|
248
|
+
return parsed_specifiers
|
249
|
+
|
250
|
+
|
251
|
+
# --------------------------------------------------------------------------------------
|
252
|
+
# Recursive descent parser for marker expression
|
253
|
+
# --------------------------------------------------------------------------------------
|
254
|
+
def parse_marker(source: str) -> MarkerList:
|
255
|
+
return _parse_full_marker(Tokenizer(source, rules=DEFAULT_RULES))
|
256
|
+
|
257
|
+
|
258
|
+
def _parse_full_marker(tokenizer: Tokenizer) -> MarkerList:
|
259
|
+
retval = _parse_marker(tokenizer)
|
260
|
+
tokenizer.expect("END", expected="end of marker expression")
|
261
|
+
return retval
|
262
|
+
|
263
|
+
|
264
|
+
def _parse_marker(tokenizer: Tokenizer) -> MarkerList:
|
265
|
+
"""
|
266
|
+
marker = marker_atom (BOOLOP marker_atom)+
|
267
|
+
"""
|
268
|
+
expression = [_parse_marker_atom(tokenizer)]
|
269
|
+
while tokenizer.check("BOOLOP"):
|
270
|
+
token = tokenizer.read()
|
271
|
+
expr_right = _parse_marker_atom(tokenizer)
|
272
|
+
expression.extend((token.text, expr_right))
|
273
|
+
return expression
|
274
|
+
|
275
|
+
|
276
|
+
def _parse_marker_atom(tokenizer: Tokenizer) -> MarkerAtom:
|
277
|
+
"""
|
278
|
+
marker_atom = WS? LEFT_PARENTHESIS WS? marker WS? RIGHT_PARENTHESIS WS?
|
279
|
+
| WS? marker_item WS?
|
280
|
+
"""
|
281
|
+
|
282
|
+
tokenizer.consume("WS")
|
283
|
+
if tokenizer.check("LEFT_PARENTHESIS", peek=True):
|
284
|
+
with tokenizer.enclosing_tokens(
|
285
|
+
"LEFT_PARENTHESIS",
|
286
|
+
"RIGHT_PARENTHESIS",
|
287
|
+
around="marker expression",
|
288
|
+
):
|
289
|
+
tokenizer.consume("WS")
|
290
|
+
marker: MarkerAtom = _parse_marker(tokenizer)
|
291
|
+
tokenizer.consume("WS")
|
292
|
+
else:
|
293
|
+
marker = _parse_marker_item(tokenizer)
|
294
|
+
tokenizer.consume("WS")
|
295
|
+
return marker
|
296
|
+
|
297
|
+
|
298
|
+
def _parse_marker_item(tokenizer: Tokenizer) -> MarkerItem:
|
299
|
+
"""
|
300
|
+
marker_item = WS? marker_var WS? marker_op WS? marker_var WS?
|
301
|
+
"""
|
302
|
+
tokenizer.consume("WS")
|
303
|
+
marker_var_left = _parse_marker_var(tokenizer)
|
304
|
+
tokenizer.consume("WS")
|
305
|
+
marker_op = _parse_marker_op(tokenizer)
|
306
|
+
tokenizer.consume("WS")
|
307
|
+
marker_var_right = _parse_marker_var(tokenizer)
|
308
|
+
tokenizer.consume("WS")
|
309
|
+
return (marker_var_left, marker_op, marker_var_right)
|
310
|
+
|
311
|
+
|
312
|
+
def _parse_marker_var(tokenizer: Tokenizer) -> MarkerVar:
|
313
|
+
"""
|
314
|
+
marker_var = VARIABLE | QUOTED_STRING
|
315
|
+
"""
|
316
|
+
if tokenizer.check("VARIABLE"):
|
317
|
+
return process_env_var(tokenizer.read().text.replace(".", "_"))
|
318
|
+
elif tokenizer.check("QUOTED_STRING"):
|
319
|
+
return process_python_str(tokenizer.read().text)
|
320
|
+
else:
|
321
|
+
tokenizer.raise_syntax_error(
|
322
|
+
message="Expected a marker variable or quoted string"
|
323
|
+
)
|
324
|
+
|
325
|
+
|
326
|
+
def process_env_var(env_var: str) -> Variable:
|
327
|
+
if env_var in ("platform_python_implementation", "python_implementation"):
|
328
|
+
return Variable("platform_python_implementation")
|
329
|
+
else:
|
330
|
+
return Variable(env_var)
|
331
|
+
|
332
|
+
|
333
|
+
def process_python_str(python_str: str) -> Value:
|
334
|
+
value = ast.literal_eval(python_str)
|
335
|
+
return Value(str(value))
|
336
|
+
|
337
|
+
|
338
|
+
def _parse_marker_op(tokenizer: Tokenizer) -> Op:
|
339
|
+
"""
|
340
|
+
marker_op = IN | NOT IN | OP
|
341
|
+
"""
|
342
|
+
if tokenizer.check("IN"):
|
343
|
+
tokenizer.read()
|
344
|
+
return Op("in")
|
345
|
+
elif tokenizer.check("NOT"):
|
346
|
+
tokenizer.read()
|
347
|
+
tokenizer.expect("WS", expected="whitespace after 'not'")
|
348
|
+
tokenizer.expect("IN", expected="'in' after 'not'")
|
349
|
+
return Op("not in")
|
350
|
+
elif tokenizer.check("OP"):
|
351
|
+
return Op(tokenizer.read().text)
|
352
|
+
else:
|
353
|
+
return tokenizer.raise_syntax_error(
|
354
|
+
"Expected marker operator, one of "
|
355
|
+
"<=, <, !=, ==, >=, >, ~=, ===, in, not in"
|
356
|
+
)
|
@@ -0,0 +1,61 @@
|
|
1
|
+
# This file is dual licensed under the terms of the Apache License, Version
|
2
|
+
# 2.0, and the BSD License. See the LICENSE file in the root of this repository
|
3
|
+
# for complete details.
|
4
|
+
|
5
|
+
|
6
|
+
class InfinityType:
|
7
|
+
def __repr__(self) -> str:
|
8
|
+
return "Infinity"
|
9
|
+
|
10
|
+
def __hash__(self) -> int:
|
11
|
+
return hash(repr(self))
|
12
|
+
|
13
|
+
def __lt__(self, other: object) -> bool:
|
14
|
+
return False
|
15
|
+
|
16
|
+
def __le__(self, other: object) -> bool:
|
17
|
+
return False
|
18
|
+
|
19
|
+
def __eq__(self, other: object) -> bool:
|
20
|
+
return isinstance(other, self.__class__)
|
21
|
+
|
22
|
+
def __gt__(self, other: object) -> bool:
|
23
|
+
return True
|
24
|
+
|
25
|
+
def __ge__(self, other: object) -> bool:
|
26
|
+
return True
|
27
|
+
|
28
|
+
def __neg__(self: object) -> "NegativeInfinityType":
|
29
|
+
return NegativeInfinity
|
30
|
+
|
31
|
+
|
32
|
+
Infinity = InfinityType()
|
33
|
+
|
34
|
+
|
35
|
+
class NegativeInfinityType:
|
36
|
+
def __repr__(self) -> str:
|
37
|
+
return "-Infinity"
|
38
|
+
|
39
|
+
def __hash__(self) -> int:
|
40
|
+
return hash(repr(self))
|
41
|
+
|
42
|
+
def __lt__(self, other: object) -> bool:
|
43
|
+
return True
|
44
|
+
|
45
|
+
def __le__(self, other: object) -> bool:
|
46
|
+
return True
|
47
|
+
|
48
|
+
def __eq__(self, other: object) -> bool:
|
49
|
+
return isinstance(other, self.__class__)
|
50
|
+
|
51
|
+
def __gt__(self, other: object) -> bool:
|
52
|
+
return False
|
53
|
+
|
54
|
+
def __ge__(self, other: object) -> bool:
|
55
|
+
return False
|
56
|
+
|
57
|
+
def __neg__(self: object) -> InfinityType:
|
58
|
+
return Infinity
|
59
|
+
|
60
|
+
|
61
|
+
NegativeInfinity = NegativeInfinityType()
|
@@ -0,0 +1,192 @@
|
|
1
|
+
import contextlib
|
2
|
+
import re
|
3
|
+
from dataclasses import dataclass
|
4
|
+
from typing import Dict, Iterator, NoReturn, Optional, Tuple, Union
|
5
|
+
|
6
|
+
from .specifiers import Specifier
|
7
|
+
|
8
|
+
|
9
|
+
@dataclass
|
10
|
+
class Token:
|
11
|
+
name: str
|
12
|
+
text: str
|
13
|
+
position: int
|
14
|
+
|
15
|
+
|
16
|
+
class ParserSyntaxError(Exception):
|
17
|
+
"""The provided source text could not be parsed correctly."""
|
18
|
+
|
19
|
+
def __init__(
|
20
|
+
self,
|
21
|
+
message: str,
|
22
|
+
*,
|
23
|
+
source: str,
|
24
|
+
span: Tuple[int, int],
|
25
|
+
) -> None:
|
26
|
+
self.span = span
|
27
|
+
self.message = message
|
28
|
+
self.source = source
|
29
|
+
|
30
|
+
super().__init__()
|
31
|
+
|
32
|
+
def __str__(self) -> str:
|
33
|
+
marker = " " * self.span[0] + "~" * (self.span[1] - self.span[0]) + "^"
|
34
|
+
return "\n ".join([self.message, self.source, marker])
|
35
|
+
|
36
|
+
|
37
|
+
DEFAULT_RULES: "Dict[str, Union[str, re.Pattern[str]]]" = {
|
38
|
+
"LEFT_PARENTHESIS": r"\(",
|
39
|
+
"RIGHT_PARENTHESIS": r"\)",
|
40
|
+
"LEFT_BRACKET": r"\[",
|
41
|
+
"RIGHT_BRACKET": r"\]",
|
42
|
+
"SEMICOLON": r";",
|
43
|
+
"COMMA": r",",
|
44
|
+
"QUOTED_STRING": re.compile(
|
45
|
+
r"""
|
46
|
+
(
|
47
|
+
('[^']*')
|
48
|
+
|
|
49
|
+
("[^"]*")
|
50
|
+
)
|
51
|
+
""",
|
52
|
+
re.VERBOSE,
|
53
|
+
),
|
54
|
+
"OP": r"(===|==|~=|!=|<=|>=|<|>)",
|
55
|
+
"BOOLOP": r"\b(or|and)\b",
|
56
|
+
"IN": r"\bin\b",
|
57
|
+
"NOT": r"\bnot\b",
|
58
|
+
"VARIABLE": re.compile(
|
59
|
+
r"""
|
60
|
+
\b(
|
61
|
+
python_version
|
62
|
+
|python_full_version
|
63
|
+
|os[._]name
|
64
|
+
|sys[._]platform
|
65
|
+
|platform_(release|system)
|
66
|
+
|platform[._](version|machine|python_implementation)
|
67
|
+
|python_implementation
|
68
|
+
|implementation_(name|version)
|
69
|
+
|extra
|
70
|
+
)\b
|
71
|
+
""",
|
72
|
+
re.VERBOSE,
|
73
|
+
),
|
74
|
+
"SPECIFIER": re.compile(
|
75
|
+
Specifier._operator_regex_str + Specifier._version_regex_str,
|
76
|
+
re.VERBOSE | re.IGNORECASE,
|
77
|
+
),
|
78
|
+
"AT": r"\@",
|
79
|
+
"URL": r"[^ \t]+",
|
80
|
+
"IDENTIFIER": r"\b[a-zA-Z0-9][a-zA-Z0-9._-]*\b",
|
81
|
+
"VERSION_PREFIX_TRAIL": r"\.\*",
|
82
|
+
"VERSION_LOCAL_LABEL_TRAIL": r"\+[a-z0-9]+(?:[-_\.][a-z0-9]+)*",
|
83
|
+
"WS": r"[ \t]+",
|
84
|
+
"END": r"$",
|
85
|
+
}
|
86
|
+
|
87
|
+
|
88
|
+
class Tokenizer:
|
89
|
+
"""Context-sensitive token parsing.
|
90
|
+
|
91
|
+
Provides methods to examine the input stream to check whether the next token
|
92
|
+
matches.
|
93
|
+
"""
|
94
|
+
|
95
|
+
def __init__(
|
96
|
+
self,
|
97
|
+
source: str,
|
98
|
+
*,
|
99
|
+
rules: "Dict[str, Union[str, re.Pattern[str]]]",
|
100
|
+
) -> None:
|
101
|
+
self.source = source
|
102
|
+
self.rules: Dict[str, re.Pattern[str]] = {
|
103
|
+
name: re.compile(pattern) for name, pattern in rules.items()
|
104
|
+
}
|
105
|
+
self.next_token: Optional[Token] = None
|
106
|
+
self.position = 0
|
107
|
+
|
108
|
+
def consume(self, name: str) -> None:
|
109
|
+
"""Move beyond provided token name, if at current position."""
|
110
|
+
if self.check(name):
|
111
|
+
self.read()
|
112
|
+
|
113
|
+
def check(self, name: str, *, peek: bool = False) -> bool:
|
114
|
+
"""Check whether the next token has the provided name.
|
115
|
+
|
116
|
+
By default, if the check succeeds, the token *must* be read before
|
117
|
+
another check. If `peek` is set to `True`, the token is not loaded and
|
118
|
+
would need to be checked again.
|
119
|
+
"""
|
120
|
+
assert (
|
121
|
+
self.next_token is None
|
122
|
+
), f"Cannot check for {name!r}, already have {self.next_token!r}"
|
123
|
+
assert name in self.rules, f"Unknown token name: {name!r}"
|
124
|
+
|
125
|
+
expression = self.rules[name]
|
126
|
+
|
127
|
+
match = expression.match(self.source, self.position)
|
128
|
+
if match is None:
|
129
|
+
return False
|
130
|
+
if not peek:
|
131
|
+
self.next_token = Token(name, match[0], self.position)
|
132
|
+
return True
|
133
|
+
|
134
|
+
def expect(self, name: str, *, expected: str) -> Token:
|
135
|
+
"""Expect a certain token name next, failing with a syntax error otherwise.
|
136
|
+
|
137
|
+
The token is *not* read.
|
138
|
+
"""
|
139
|
+
if not self.check(name):
|
140
|
+
raise self.raise_syntax_error(f"Expected {expected}")
|
141
|
+
return self.read()
|
142
|
+
|
143
|
+
def read(self) -> Token:
|
144
|
+
"""Consume the next token and return it."""
|
145
|
+
token = self.next_token
|
146
|
+
assert token is not None
|
147
|
+
|
148
|
+
self.position += len(token.text)
|
149
|
+
self.next_token = None
|
150
|
+
|
151
|
+
return token
|
152
|
+
|
153
|
+
def raise_syntax_error(
|
154
|
+
self,
|
155
|
+
message: str,
|
156
|
+
*,
|
157
|
+
span_start: Optional[int] = None,
|
158
|
+
span_end: Optional[int] = None,
|
159
|
+
) -> NoReturn:
|
160
|
+
"""Raise ParserSyntaxError at the given position."""
|
161
|
+
span = (
|
162
|
+
self.position if span_start is None else span_start,
|
163
|
+
self.position if span_end is None else span_end,
|
164
|
+
)
|
165
|
+
raise ParserSyntaxError(
|
166
|
+
message,
|
167
|
+
source=self.source,
|
168
|
+
span=span,
|
169
|
+
)
|
170
|
+
|
171
|
+
@contextlib.contextmanager
|
172
|
+
def enclosing_tokens(
|
173
|
+
self, open_token: str, close_token: str, *, around: str
|
174
|
+
) -> Iterator[None]:
|
175
|
+
if self.check(open_token):
|
176
|
+
open_position = self.position
|
177
|
+
self.read()
|
178
|
+
else:
|
179
|
+
open_position = None
|
180
|
+
|
181
|
+
yield
|
182
|
+
|
183
|
+
if open_position is None:
|
184
|
+
return
|
185
|
+
|
186
|
+
if not self.check(close_token):
|
187
|
+
self.raise_syntax_error(
|
188
|
+
f"Expected matching {close_token} for {open_token}, after {around}",
|
189
|
+
span_start=open_position,
|
190
|
+
)
|
191
|
+
|
192
|
+
self.read()
|