pygmentize 0.0.1
Sign up to get free protection for your applications and to get access to all the features.
- data/LICENSE +19 -0
- data/lib/pygments.rb +23 -0
- data/pygmentize.gemspec +11 -0
- data/test/pygments.rb +19 -0
- data/vendor/pygmentize.py +7 -0
- data/vendor/pygments/AUTHORS +73 -0
- data/vendor/pygments/LICENSE +25 -0
- data/vendor/pygments/__init__.py +91 -0
- data/vendor/pygments/__init__.pyc +0 -0
- data/vendor/pygments/cmdline.py +430 -0
- data/vendor/pygments/cmdline.pyc +0 -0
- data/vendor/pygments/console.py +74 -0
- data/vendor/pygments/console.pyc +0 -0
- data/vendor/pygments/filter.py +74 -0
- data/vendor/pygments/filter.pyc +0 -0
- data/vendor/pygments/filters/__init__.py +357 -0
- data/vendor/pygments/filters/__init__.pyc +0 -0
- data/vendor/pygments/formatter.py +92 -0
- data/vendor/pygments/formatter.pyc +0 -0
- data/vendor/pygments/formatters/__init__.py +68 -0
- data/vendor/pygments/formatters/__init__.pyc +0 -0
- data/vendor/pygments/formatters/_mapping.py +92 -0
- data/vendor/pygments/formatters/_mapping.pyc +0 -0
- data/vendor/pygments/formatters/bbcode.py +109 -0
- data/vendor/pygments/formatters/bbcode.pyc +0 -0
- data/vendor/pygments/formatters/html.py +723 -0
- data/vendor/pygments/formatters/html.pyc +0 -0
- data/vendor/pygments/formatters/img.py +553 -0
- data/vendor/pygments/formatters/img.pyc +0 -0
- data/vendor/pygments/formatters/latex.py +354 -0
- data/vendor/pygments/formatters/latex.pyc +0 -0
- data/vendor/pygments/formatters/other.py +117 -0
- data/vendor/pygments/formatters/other.pyc +0 -0
- data/vendor/pygments/formatters/rtf.py +136 -0
- data/vendor/pygments/formatters/rtf.pyc +0 -0
- data/vendor/pygments/formatters/svg.py +154 -0
- data/vendor/pygments/formatters/svg.pyc +0 -0
- data/vendor/pygments/formatters/terminal.py +109 -0
- data/vendor/pygments/formatters/terminal.pyc +0 -0
- data/vendor/pygments/formatters/terminal256.py +219 -0
- data/vendor/pygments/formatters/terminal256.pyc +0 -0
- data/vendor/pygments/lexer.py +660 -0
- data/vendor/pygments/lexer.pyc +0 -0
- data/vendor/pygments/lexers/__init__.py +226 -0
- data/vendor/pygments/lexers/__init__.pyc +0 -0
- data/vendor/pygments/lexers/_asybuiltins.py +1645 -0
- data/vendor/pygments/lexers/_clbuiltins.py +232 -0
- data/vendor/pygments/lexers/_luabuiltins.py +256 -0
- data/vendor/pygments/lexers/_mapping.py +234 -0
- data/vendor/pygments/lexers/_mapping.pyc +0 -0
- data/vendor/pygments/lexers/_phpbuiltins.py +3389 -0
- data/vendor/pygments/lexers/_vimbuiltins.py +3 -0
- data/vendor/pygments/lexers/agile.py +1485 -0
- data/vendor/pygments/lexers/agile.pyc +0 -0
- data/vendor/pygments/lexers/asm.py +353 -0
- data/vendor/pygments/lexers/compiled.py +2365 -0
- data/vendor/pygments/lexers/dotnet.py +355 -0
- data/vendor/pygments/lexers/functional.py +756 -0
- data/vendor/pygments/lexers/functional.pyc +0 -0
- data/vendor/pygments/lexers/math.py +461 -0
- data/vendor/pygments/lexers/other.py +2297 -0
- data/vendor/pygments/lexers/parsers.py +695 -0
- data/vendor/pygments/lexers/special.py +100 -0
- data/vendor/pygments/lexers/special.pyc +0 -0
- data/vendor/pygments/lexers/templates.py +1387 -0
- data/vendor/pygments/lexers/text.py +1586 -0
- data/vendor/pygments/lexers/web.py +1619 -0
- data/vendor/pygments/lexers/web.pyc +0 -0
- data/vendor/pygments/plugin.py +74 -0
- data/vendor/pygments/plugin.pyc +0 -0
- data/vendor/pygments/scanner.py +104 -0
- data/vendor/pygments/style.py +117 -0
- data/vendor/pygments/style.pyc +0 -0
- data/vendor/pygments/styles/__init__.py +68 -0
- data/vendor/pygments/styles/__init__.pyc +0 -0
- data/vendor/pygments/styles/autumn.py +65 -0
- data/vendor/pygments/styles/borland.py +51 -0
- data/vendor/pygments/styles/bw.py +49 -0
- data/vendor/pygments/styles/colorful.py +81 -0
- data/vendor/pygments/styles/default.py +73 -0
- data/vendor/pygments/styles/default.pyc +0 -0
- data/vendor/pygments/styles/emacs.py +72 -0
- data/vendor/pygments/styles/friendly.py +72 -0
- data/vendor/pygments/styles/fruity.py +43 -0
- data/vendor/pygments/styles/manni.py +75 -0
- data/vendor/pygments/styles/monokai.py +106 -0
- data/vendor/pygments/styles/murphy.py +80 -0
- data/vendor/pygments/styles/native.py +65 -0
- data/vendor/pygments/styles/pastie.py +75 -0
- data/vendor/pygments/styles/perldoc.py +69 -0
- data/vendor/pygments/styles/tango.py +141 -0
- data/vendor/pygments/styles/trac.py +63 -0
- data/vendor/pygments/styles/vim.py +63 -0
- data/vendor/pygments/styles/vs.py +38 -0
- data/vendor/pygments/token.py +198 -0
- data/vendor/pygments/token.pyc +0 -0
- data/vendor/pygments/unistring.py +130 -0
- data/vendor/pygments/unistring.pyc +0 -0
- data/vendor/pygments/util.py +226 -0
- data/vendor/pygments/util.pyc +0 -0
- metadata +166 -0
Binary file
|
@@ -0,0 +1,660 @@
|
|
1
|
+
# -*- coding: utf-8 -*-
|
2
|
+
"""
|
3
|
+
pygments.lexer
|
4
|
+
~~~~~~~~~~~~~~
|
5
|
+
|
6
|
+
Base lexer classes.
|
7
|
+
|
8
|
+
:copyright: Copyright 2006-2010 by the Pygments team, see AUTHORS.
|
9
|
+
:license: BSD, see LICENSE for details.
|
10
|
+
"""
|
11
|
+
import re
|
12
|
+
|
13
|
+
from pygments.filter import apply_filters, Filter
|
14
|
+
from pygments.filters import get_filter_by_name
|
15
|
+
from pygments.token import Error, Text, Other, _TokenType
|
16
|
+
from pygments.util import get_bool_opt, get_int_opt, get_list_opt, \
|
17
|
+
make_analysator
|
18
|
+
|
19
|
+
|
20
|
+
__all__ = ['Lexer', 'RegexLexer', 'ExtendedRegexLexer', 'DelegatingLexer',
|
21
|
+
'LexerContext', 'include', 'bygroups', 'using', 'this']
|
22
|
+
|
23
|
+
|
24
|
+
_default_analyse = staticmethod(lambda x: 0.0)
|
25
|
+
|
26
|
+
|
27
|
+
class LexerMeta(type):
|
28
|
+
"""
|
29
|
+
This metaclass automagically converts ``analyse_text`` methods into
|
30
|
+
static methods which always return float values.
|
31
|
+
"""
|
32
|
+
|
33
|
+
def __new__(cls, name, bases, d):
|
34
|
+
if 'analyse_text' in d:
|
35
|
+
d['analyse_text'] = make_analysator(d['analyse_text'])
|
36
|
+
return type.__new__(cls, name, bases, d)
|
37
|
+
|
38
|
+
|
39
|
+
class Lexer(object):
|
40
|
+
"""
|
41
|
+
Lexer for a specific language.
|
42
|
+
|
43
|
+
Basic options recognized:
|
44
|
+
``stripnl``
|
45
|
+
Strip leading and trailing newlines from the input (default: True).
|
46
|
+
``stripall``
|
47
|
+
Strip all leading and trailing whitespace from the input
|
48
|
+
(default: False).
|
49
|
+
``ensurenl``
|
50
|
+
Make sure that the input ends with a newline (default: True). This
|
51
|
+
is required for some lexers that consume input linewise.
|
52
|
+
*New in Pygments 1.3.*
|
53
|
+
``tabsize``
|
54
|
+
If given and greater than 0, expand tabs in the input (default: 0).
|
55
|
+
``encoding``
|
56
|
+
If given, must be an encoding name. This encoding will be used to
|
57
|
+
convert the input string to Unicode, if it is not already a Unicode
|
58
|
+
string (default: ``'latin1'``).
|
59
|
+
Can also be ``'guess'`` to use a simple UTF-8 / Latin1 detection, or
|
60
|
+
``'chardet'`` to use the chardet library, if it is installed.
|
61
|
+
"""
|
62
|
+
|
63
|
+
#: Name of the lexer
|
64
|
+
name = None
|
65
|
+
|
66
|
+
#: Shortcuts for the lexer
|
67
|
+
aliases = []
|
68
|
+
|
69
|
+
#: fn match rules
|
70
|
+
filenames = []
|
71
|
+
|
72
|
+
#: fn alias filenames
|
73
|
+
alias_filenames = []
|
74
|
+
|
75
|
+
#: mime types
|
76
|
+
mimetypes = []
|
77
|
+
|
78
|
+
__metaclass__ = LexerMeta
|
79
|
+
|
80
|
+
def __init__(self, **options):
|
81
|
+
self.options = options
|
82
|
+
self.stripnl = get_bool_opt(options, 'stripnl', True)
|
83
|
+
self.stripall = get_bool_opt(options, 'stripall', False)
|
84
|
+
self.ensurenl = get_bool_opt(options, 'ensurenl', True)
|
85
|
+
self.tabsize = get_int_opt(options, 'tabsize', 0)
|
86
|
+
self.encoding = options.get('encoding', 'latin1')
|
87
|
+
# self.encoding = options.get('inencoding', None) or self.encoding
|
88
|
+
self.filters = []
|
89
|
+
for filter_ in get_list_opt(options, 'filters', ()):
|
90
|
+
self.add_filter(filter_)
|
91
|
+
|
92
|
+
def __repr__(self):
|
93
|
+
if self.options:
|
94
|
+
return '<pygments.lexers.%s with %r>' % (self.__class__.__name__,
|
95
|
+
self.options)
|
96
|
+
else:
|
97
|
+
return '<pygments.lexers.%s>' % self.__class__.__name__
|
98
|
+
|
99
|
+
def add_filter(self, filter_, **options):
|
100
|
+
"""
|
101
|
+
Add a new stream filter to this lexer.
|
102
|
+
"""
|
103
|
+
if not isinstance(filter_, Filter):
|
104
|
+
filter_ = get_filter_by_name(filter_, **options)
|
105
|
+
self.filters.append(filter_)
|
106
|
+
|
107
|
+
def analyse_text(text):
|
108
|
+
"""
|
109
|
+
Has to return a float between ``0`` and ``1`` that indicates
|
110
|
+
if a lexer wants to highlight this text. Used by ``guess_lexer``.
|
111
|
+
If this method returns ``0`` it won't highlight it in any case, if
|
112
|
+
it returns ``1`` highlighting with this lexer is guaranteed.
|
113
|
+
|
114
|
+
The `LexerMeta` metaclass automatically wraps this function so
|
115
|
+
that it works like a static method (no ``self`` or ``cls``
|
116
|
+
parameter) and the return value is automatically converted to
|
117
|
+
`float`. If the return value is an object that is boolean `False`
|
118
|
+
it's the same as if the return values was ``0.0``.
|
119
|
+
"""
|
120
|
+
|
121
|
+
def get_tokens(self, text, unfiltered=False):
|
122
|
+
"""
|
123
|
+
Return an iterable of (tokentype, value) pairs generated from
|
124
|
+
`text`. If `unfiltered` is set to `True`, the filtering mechanism
|
125
|
+
is bypassed even if filters are defined.
|
126
|
+
|
127
|
+
Also preprocess the text, i.e. expand tabs and strip it if
|
128
|
+
wanted and applies registered filters.
|
129
|
+
"""
|
130
|
+
if not isinstance(text, unicode):
|
131
|
+
if self.encoding == 'guess':
|
132
|
+
try:
|
133
|
+
text = text.decode('utf-8')
|
134
|
+
if text.startswith(u'\ufeff'):
|
135
|
+
text = text[len(u'\ufeff'):]
|
136
|
+
except UnicodeDecodeError:
|
137
|
+
text = text.decode('latin1')
|
138
|
+
elif self.encoding == 'chardet':
|
139
|
+
try:
|
140
|
+
import chardet
|
141
|
+
except ImportError:
|
142
|
+
raise ImportError('To enable chardet encoding guessing, '
|
143
|
+
'please install the chardet library '
|
144
|
+
'from http://chardet.feedparser.org/')
|
145
|
+
enc = chardet.detect(text)
|
146
|
+
text = text.decode(enc['encoding'])
|
147
|
+
else:
|
148
|
+
text = text.decode(self.encoding)
|
149
|
+
# text now *is* a unicode string
|
150
|
+
text = text.replace('\r\n', '\n')
|
151
|
+
text = text.replace('\r', '\n')
|
152
|
+
if self.stripall:
|
153
|
+
text = text.strip()
|
154
|
+
elif self.stripnl:
|
155
|
+
text = text.strip('\n')
|
156
|
+
if self.tabsize > 0:
|
157
|
+
text = text.expandtabs(self.tabsize)
|
158
|
+
if self.ensurenl and not text.endswith('\n'):
|
159
|
+
text += '\n'
|
160
|
+
|
161
|
+
def streamer():
|
162
|
+
for i, t, v in self.get_tokens_unprocessed(text):
|
163
|
+
yield t, v
|
164
|
+
stream = streamer()
|
165
|
+
if not unfiltered:
|
166
|
+
stream = apply_filters(stream, self.filters, self)
|
167
|
+
return stream
|
168
|
+
|
169
|
+
def get_tokens_unprocessed(self, text):
|
170
|
+
"""
|
171
|
+
Return an iterable of (tokentype, value) pairs.
|
172
|
+
In subclasses, implement this method as a generator to
|
173
|
+
maximize effectiveness.
|
174
|
+
"""
|
175
|
+
raise NotImplementedError
|
176
|
+
|
177
|
+
|
178
|
+
class DelegatingLexer(Lexer):
|
179
|
+
"""
|
180
|
+
This lexer takes two lexer as arguments. A root lexer and
|
181
|
+
a language lexer. First everything is scanned using the language
|
182
|
+
lexer, afterwards all ``Other`` tokens are lexed using the root
|
183
|
+
lexer.
|
184
|
+
|
185
|
+
The lexers from the ``template`` lexer package use this base lexer.
|
186
|
+
"""
|
187
|
+
|
188
|
+
def __init__(self, _root_lexer, _language_lexer, _needle=Other, **options):
|
189
|
+
self.root_lexer = _root_lexer(**options)
|
190
|
+
self.language_lexer = _language_lexer(**options)
|
191
|
+
self.needle = _needle
|
192
|
+
Lexer.__init__(self, **options)
|
193
|
+
|
194
|
+
def get_tokens_unprocessed(self, text):
|
195
|
+
buffered = ''
|
196
|
+
insertions = []
|
197
|
+
lng_buffer = []
|
198
|
+
for i, t, v in self.language_lexer.get_tokens_unprocessed(text):
|
199
|
+
if t is self.needle:
|
200
|
+
if lng_buffer:
|
201
|
+
insertions.append((len(buffered), lng_buffer))
|
202
|
+
lng_buffer = []
|
203
|
+
buffered += v
|
204
|
+
else:
|
205
|
+
lng_buffer.append((i, t, v))
|
206
|
+
if lng_buffer:
|
207
|
+
insertions.append((len(buffered), lng_buffer))
|
208
|
+
return do_insertions(insertions,
|
209
|
+
self.root_lexer.get_tokens_unprocessed(buffered))
|
210
|
+
|
211
|
+
|
212
|
+
#-------------------------------------------------------------------------------
|
213
|
+
# RegexLexer and ExtendedRegexLexer
|
214
|
+
#
|
215
|
+
|
216
|
+
|
217
|
+
class include(str):
|
218
|
+
"""
|
219
|
+
Indicates that a state should include rules from another state.
|
220
|
+
"""
|
221
|
+
pass
|
222
|
+
|
223
|
+
|
224
|
+
class combined(tuple):
|
225
|
+
"""
|
226
|
+
Indicates a state combined from multiple states.
|
227
|
+
"""
|
228
|
+
|
229
|
+
def __new__(cls, *args):
|
230
|
+
return tuple.__new__(cls, args)
|
231
|
+
|
232
|
+
def __init__(self, *args):
|
233
|
+
# tuple.__init__ doesn't do anything
|
234
|
+
pass
|
235
|
+
|
236
|
+
|
237
|
+
class _PseudoMatch(object):
|
238
|
+
"""
|
239
|
+
A pseudo match object constructed from a string.
|
240
|
+
"""
|
241
|
+
|
242
|
+
def __init__(self, start, text):
|
243
|
+
self._text = text
|
244
|
+
self._start = start
|
245
|
+
|
246
|
+
def start(self, arg=None):
|
247
|
+
return self._start
|
248
|
+
|
249
|
+
def end(self, arg=None):
|
250
|
+
return self._start + len(self._text)
|
251
|
+
|
252
|
+
def group(self, arg=None):
|
253
|
+
if arg:
|
254
|
+
raise IndexError('No such group')
|
255
|
+
return self._text
|
256
|
+
|
257
|
+
def groups(self):
|
258
|
+
return (self._text,)
|
259
|
+
|
260
|
+
def groupdict(self):
|
261
|
+
return {}
|
262
|
+
|
263
|
+
|
264
|
+
def bygroups(*args):
|
265
|
+
"""
|
266
|
+
Callback that yields multiple actions for each group in the match.
|
267
|
+
"""
|
268
|
+
def callback(lexer, match, ctx=None):
|
269
|
+
for i, action in enumerate(args):
|
270
|
+
if action is None:
|
271
|
+
continue
|
272
|
+
elif type(action) is _TokenType:
|
273
|
+
data = match.group(i + 1)
|
274
|
+
if data:
|
275
|
+
yield match.start(i + 1), action, data
|
276
|
+
else:
|
277
|
+
if ctx:
|
278
|
+
ctx.pos = match.start(i + 1)
|
279
|
+
for item in action(lexer, _PseudoMatch(match.start(i + 1),
|
280
|
+
match.group(i + 1)), ctx):
|
281
|
+
if item:
|
282
|
+
yield item
|
283
|
+
if ctx:
|
284
|
+
ctx.pos = match.end()
|
285
|
+
return callback
|
286
|
+
|
287
|
+
|
288
|
+
class _This(object):
|
289
|
+
"""
|
290
|
+
Special singleton used for indicating the caller class.
|
291
|
+
Used by ``using``.
|
292
|
+
"""
|
293
|
+
this = _This()
|
294
|
+
|
295
|
+
|
296
|
+
def using(_other, **kwargs):
|
297
|
+
"""
|
298
|
+
Callback that processes the match with a different lexer.
|
299
|
+
|
300
|
+
The keyword arguments are forwarded to the lexer, except `state` which
|
301
|
+
is handled separately.
|
302
|
+
|
303
|
+
`state` specifies the state that the new lexer will start in, and can
|
304
|
+
be an enumerable such as ('root', 'inline', 'string') or a simple
|
305
|
+
string which is assumed to be on top of the root state.
|
306
|
+
|
307
|
+
Note: For that to work, `_other` must not be an `ExtendedRegexLexer`.
|
308
|
+
"""
|
309
|
+
gt_kwargs = {}
|
310
|
+
if 'state' in kwargs:
|
311
|
+
s = kwargs.pop('state')
|
312
|
+
if isinstance(s, (list, tuple)):
|
313
|
+
gt_kwargs['stack'] = s
|
314
|
+
else:
|
315
|
+
gt_kwargs['stack'] = ('root', s)
|
316
|
+
|
317
|
+
if _other is this:
|
318
|
+
def callback(lexer, match, ctx=None):
|
319
|
+
# if keyword arguments are given the callback
|
320
|
+
# function has to create a new lexer instance
|
321
|
+
if kwargs:
|
322
|
+
# XXX: cache that somehow
|
323
|
+
kwargs.update(lexer.options)
|
324
|
+
lx = lexer.__class__(**kwargs)
|
325
|
+
else:
|
326
|
+
lx = lexer
|
327
|
+
s = match.start()
|
328
|
+
for i, t, v in lx.get_tokens_unprocessed(match.group(), **gt_kwargs):
|
329
|
+
yield i + s, t, v
|
330
|
+
if ctx:
|
331
|
+
ctx.pos = match.end()
|
332
|
+
else:
|
333
|
+
def callback(lexer, match, ctx=None):
|
334
|
+
# XXX: cache that somehow
|
335
|
+
kwargs.update(lexer.options)
|
336
|
+
lx = _other(**kwargs)
|
337
|
+
|
338
|
+
s = match.start()
|
339
|
+
for i, t, v in lx.get_tokens_unprocessed(match.group(), **gt_kwargs):
|
340
|
+
yield i + s, t, v
|
341
|
+
if ctx:
|
342
|
+
ctx.pos = match.end()
|
343
|
+
return callback
|
344
|
+
|
345
|
+
|
346
|
+
class RegexLexerMeta(LexerMeta):
|
347
|
+
"""
|
348
|
+
Metaclass for RegexLexer, creates the self._tokens attribute from
|
349
|
+
self.tokens on the first instantiation.
|
350
|
+
"""
|
351
|
+
|
352
|
+
def _process_state(cls, unprocessed, processed, state):
|
353
|
+
assert type(state) is str, "wrong state name %r" % state
|
354
|
+
assert state[0] != '#', "invalid state name %r" % state
|
355
|
+
if state in processed:
|
356
|
+
return processed[state]
|
357
|
+
tokens = processed[state] = []
|
358
|
+
rflags = cls.flags
|
359
|
+
for tdef in unprocessed[state]:
|
360
|
+
if isinstance(tdef, include):
|
361
|
+
# it's a state reference
|
362
|
+
assert tdef != state, "circular state reference %r" % state
|
363
|
+
tokens.extend(cls._process_state(unprocessed, processed, str(tdef)))
|
364
|
+
continue
|
365
|
+
|
366
|
+
assert type(tdef) is tuple, "wrong rule def %r" % tdef
|
367
|
+
|
368
|
+
try:
|
369
|
+
rex = re.compile(tdef[0], rflags).match
|
370
|
+
except Exception, err:
|
371
|
+
raise ValueError("uncompilable regex %r in state %r of %r: %s" %
|
372
|
+
(tdef[0], state, cls, err))
|
373
|
+
|
374
|
+
assert type(tdef[1]) is _TokenType or callable(tdef[1]), \
|
375
|
+
'token type must be simple type or callable, not %r' % (tdef[1],)
|
376
|
+
|
377
|
+
if len(tdef) == 2:
|
378
|
+
new_state = None
|
379
|
+
else:
|
380
|
+
tdef2 = tdef[2]
|
381
|
+
if isinstance(tdef2, str):
|
382
|
+
# an existing state
|
383
|
+
if tdef2 == '#pop':
|
384
|
+
new_state = -1
|
385
|
+
elif tdef2 in unprocessed:
|
386
|
+
new_state = (tdef2,)
|
387
|
+
elif tdef2 == '#push':
|
388
|
+
new_state = tdef2
|
389
|
+
elif tdef2[:5] == '#pop:':
|
390
|
+
new_state = -int(tdef2[5:])
|
391
|
+
else:
|
392
|
+
assert False, 'unknown new state %r' % tdef2
|
393
|
+
elif isinstance(tdef2, combined):
|
394
|
+
# combine a new state from existing ones
|
395
|
+
new_state = '_tmp_%d' % cls._tmpname
|
396
|
+
cls._tmpname += 1
|
397
|
+
itokens = []
|
398
|
+
for istate in tdef2:
|
399
|
+
assert istate != state, 'circular state ref %r' % istate
|
400
|
+
itokens.extend(cls._process_state(unprocessed,
|
401
|
+
processed, istate))
|
402
|
+
processed[new_state] = itokens
|
403
|
+
new_state = (new_state,)
|
404
|
+
elif isinstance(tdef2, tuple):
|
405
|
+
# push more than one state
|
406
|
+
for state in tdef2:
|
407
|
+
assert (state in unprocessed or
|
408
|
+
state in ('#pop', '#push')), \
|
409
|
+
'unknown new state ' + state
|
410
|
+
new_state = tdef2
|
411
|
+
else:
|
412
|
+
assert False, 'unknown new state def %r' % tdef2
|
413
|
+
tokens.append((rex, tdef[1], new_state))
|
414
|
+
return tokens
|
415
|
+
|
416
|
+
def process_tokendef(cls, name, tokendefs=None):
|
417
|
+
processed = cls._all_tokens[name] = {}
|
418
|
+
tokendefs = tokendefs or cls.tokens[name]
|
419
|
+
for state in tokendefs.keys():
|
420
|
+
cls._process_state(tokendefs, processed, state)
|
421
|
+
return processed
|
422
|
+
|
423
|
+
def __call__(cls, *args, **kwds):
|
424
|
+
if not hasattr(cls, '_tokens'):
|
425
|
+
cls._all_tokens = {}
|
426
|
+
cls._tmpname = 0
|
427
|
+
if hasattr(cls, 'token_variants') and cls.token_variants:
|
428
|
+
# don't process yet
|
429
|
+
pass
|
430
|
+
else:
|
431
|
+
cls._tokens = cls.process_tokendef('', cls.tokens)
|
432
|
+
|
433
|
+
return type.__call__(cls, *args, **kwds)
|
434
|
+
|
435
|
+
|
436
|
+
class RegexLexer(Lexer):
|
437
|
+
"""
|
438
|
+
Base for simple stateful regular expression-based lexers.
|
439
|
+
Simplifies the lexing process so that you need only
|
440
|
+
provide a list of states and regular expressions.
|
441
|
+
"""
|
442
|
+
__metaclass__ = RegexLexerMeta
|
443
|
+
|
444
|
+
#: Flags for compiling the regular expressions.
|
445
|
+
#: Defaults to MULTILINE.
|
446
|
+
flags = re.MULTILINE
|
447
|
+
|
448
|
+
#: Dict of ``{'state': [(regex, tokentype, new_state), ...], ...}``
|
449
|
+
#:
|
450
|
+
#: The initial state is 'root'.
|
451
|
+
#: ``new_state`` can be omitted to signify no state transition.
|
452
|
+
#: If it is a string, the state is pushed on the stack and changed.
|
453
|
+
#: If it is a tuple of strings, all states are pushed on the stack and
|
454
|
+
#: the current state will be the topmost.
|
455
|
+
#: It can also be ``combined('state1', 'state2', ...)``
|
456
|
+
#: to signify a new, anonymous state combined from the rules of two
|
457
|
+
#: or more existing ones.
|
458
|
+
#: Furthermore, it can be '#pop' to signify going back one step in
|
459
|
+
#: the state stack, or '#push' to push the current state on the stack
|
460
|
+
#: again.
|
461
|
+
#:
|
462
|
+
#: The tuple can also be replaced with ``include('state')``, in which
|
463
|
+
#: case the rules from the state named by the string are included in the
|
464
|
+
#: current one.
|
465
|
+
tokens = {}
|
466
|
+
|
467
|
+
def get_tokens_unprocessed(self, text, stack=('root',)):
|
468
|
+
"""
|
469
|
+
Split ``text`` into (tokentype, text) pairs.
|
470
|
+
|
471
|
+
``stack`` is the inital stack (default: ``['root']``)
|
472
|
+
"""
|
473
|
+
pos = 0
|
474
|
+
tokendefs = self._tokens
|
475
|
+
statestack = list(stack)
|
476
|
+
statetokens = tokendefs[statestack[-1]]
|
477
|
+
while 1:
|
478
|
+
for rexmatch, action, new_state in statetokens:
|
479
|
+
m = rexmatch(text, pos)
|
480
|
+
if m:
|
481
|
+
if type(action) is _TokenType:
|
482
|
+
yield pos, action, m.group()
|
483
|
+
else:
|
484
|
+
for item in action(self, m):
|
485
|
+
yield item
|
486
|
+
pos = m.end()
|
487
|
+
if new_state is not None:
|
488
|
+
# state transition
|
489
|
+
if isinstance(new_state, tuple):
|
490
|
+
for state in new_state:
|
491
|
+
if state == '#pop':
|
492
|
+
statestack.pop()
|
493
|
+
elif state == '#push':
|
494
|
+
statestack.append(statestack[-1])
|
495
|
+
else:
|
496
|
+
statestack.append(state)
|
497
|
+
elif isinstance(new_state, int):
|
498
|
+
# pop
|
499
|
+
del statestack[new_state:]
|
500
|
+
elif new_state == '#push':
|
501
|
+
statestack.append(statestack[-1])
|
502
|
+
else:
|
503
|
+
assert False, "wrong state def: %r" % new_state
|
504
|
+
statetokens = tokendefs[statestack[-1]]
|
505
|
+
break
|
506
|
+
else:
|
507
|
+
try:
|
508
|
+
if text[pos] == '\n':
|
509
|
+
# at EOL, reset state to "root"
|
510
|
+
pos += 1
|
511
|
+
statestack = ['root']
|
512
|
+
statetokens = tokendefs['root']
|
513
|
+
yield pos, Text, u'\n'
|
514
|
+
continue
|
515
|
+
yield pos, Error, text[pos]
|
516
|
+
pos += 1
|
517
|
+
except IndexError:
|
518
|
+
break
|
519
|
+
|
520
|
+
|
521
|
+
class LexerContext(object):
|
522
|
+
"""
|
523
|
+
A helper object that holds lexer position data.
|
524
|
+
"""
|
525
|
+
|
526
|
+
def __init__(self, text, pos, stack=None, end=None):
|
527
|
+
self.text = text
|
528
|
+
self.pos = pos
|
529
|
+
self.end = end or len(text) # end=0 not supported ;-)
|
530
|
+
self.stack = stack or ['root']
|
531
|
+
|
532
|
+
def __repr__(self):
|
533
|
+
return 'LexerContext(%r, %r, %r)' % (
|
534
|
+
self.text, self.pos, self.stack)
|
535
|
+
|
536
|
+
|
537
|
+
class ExtendedRegexLexer(RegexLexer):
|
538
|
+
"""
|
539
|
+
A RegexLexer that uses a context object to store its state.
|
540
|
+
"""
|
541
|
+
|
542
|
+
def get_tokens_unprocessed(self, text=None, context=None):
|
543
|
+
"""
|
544
|
+
Split ``text`` into (tokentype, text) pairs.
|
545
|
+
If ``context`` is given, use this lexer context instead.
|
546
|
+
"""
|
547
|
+
tokendefs = self._tokens
|
548
|
+
if not context:
|
549
|
+
ctx = LexerContext(text, 0)
|
550
|
+
statetokens = tokendefs['root']
|
551
|
+
else:
|
552
|
+
ctx = context
|
553
|
+
statetokens = tokendefs[ctx.stack[-1]]
|
554
|
+
text = ctx.text
|
555
|
+
while 1:
|
556
|
+
for rexmatch, action, new_state in statetokens:
|
557
|
+
m = rexmatch(text, ctx.pos, ctx.end)
|
558
|
+
if m:
|
559
|
+
if type(action) is _TokenType:
|
560
|
+
yield ctx.pos, action, m.group()
|
561
|
+
ctx.pos = m.end()
|
562
|
+
else:
|
563
|
+
for item in action(self, m, ctx):
|
564
|
+
yield item
|
565
|
+
if not new_state:
|
566
|
+
# altered the state stack?
|
567
|
+
statetokens = tokendefs[ctx.stack[-1]]
|
568
|
+
# CAUTION: callback must set ctx.pos!
|
569
|
+
if new_state is not None:
|
570
|
+
# state transition
|
571
|
+
if isinstance(new_state, tuple):
|
572
|
+
ctx.stack.extend(new_state)
|
573
|
+
elif isinstance(new_state, int):
|
574
|
+
# pop
|
575
|
+
del ctx.stack[new_state:]
|
576
|
+
elif new_state == '#push':
|
577
|
+
ctx.stack.append(ctx.stack[-1])
|
578
|
+
else:
|
579
|
+
assert False, "wrong state def: %r" % new_state
|
580
|
+
statetokens = tokendefs[ctx.stack[-1]]
|
581
|
+
break
|
582
|
+
else:
|
583
|
+
try:
|
584
|
+
if ctx.pos >= ctx.end:
|
585
|
+
break
|
586
|
+
if text[ctx.pos] == '\n':
|
587
|
+
# at EOL, reset state to "root"
|
588
|
+
ctx.pos += 1
|
589
|
+
ctx.stack = ['root']
|
590
|
+
statetokens = tokendefs['root']
|
591
|
+
yield ctx.pos, Text, u'\n'
|
592
|
+
continue
|
593
|
+
yield ctx.pos, Error, text[ctx.pos]
|
594
|
+
ctx.pos += 1
|
595
|
+
except IndexError:
|
596
|
+
break
|
597
|
+
|
598
|
+
|
599
|
+
def do_insertions(insertions, tokens):
|
600
|
+
"""
|
601
|
+
Helper for lexers which must combine the results of several
|
602
|
+
sublexers.
|
603
|
+
|
604
|
+
``insertions`` is a list of ``(index, itokens)`` pairs.
|
605
|
+
Each ``itokens`` iterable should be inserted at position
|
606
|
+
``index`` into the token stream given by the ``tokens``
|
607
|
+
argument.
|
608
|
+
|
609
|
+
The result is a combined token stream.
|
610
|
+
|
611
|
+
TODO: clean up the code here.
|
612
|
+
"""
|
613
|
+
insertions = iter(insertions)
|
614
|
+
try:
|
615
|
+
index, itokens = insertions.next()
|
616
|
+
except StopIteration:
|
617
|
+
# no insertions
|
618
|
+
for item in tokens:
|
619
|
+
yield item
|
620
|
+
return
|
621
|
+
|
622
|
+
realpos = None
|
623
|
+
insleft = True
|
624
|
+
|
625
|
+
# iterate over the token stream where we want to insert
|
626
|
+
# the tokens from the insertion list.
|
627
|
+
for i, t, v in tokens:
|
628
|
+
# first iteration. store the postition of first item
|
629
|
+
if realpos is None:
|
630
|
+
realpos = i
|
631
|
+
oldi = 0
|
632
|
+
while insleft and i + len(v) >= index:
|
633
|
+
tmpval = v[oldi:index - i]
|
634
|
+
yield realpos, t, tmpval
|
635
|
+
realpos += len(tmpval)
|
636
|
+
for it_index, it_token, it_value in itokens:
|
637
|
+
yield realpos, it_token, it_value
|
638
|
+
realpos += len(it_value)
|
639
|
+
oldi = index - i
|
640
|
+
try:
|
641
|
+
index, itokens = insertions.next()
|
642
|
+
except StopIteration:
|
643
|
+
insleft = False
|
644
|
+
break # not strictly necessary
|
645
|
+
yield realpos, t, v[oldi:]
|
646
|
+
realpos += len(v) - oldi
|
647
|
+
|
648
|
+
# leftover tokens
|
649
|
+
while insleft:
|
650
|
+
# no normal tokens, set realpos to zero
|
651
|
+
realpos = realpos or 0
|
652
|
+
for p, t, v in itokens:
|
653
|
+
yield realpos, t, v
|
654
|
+
realpos += len(v)
|
655
|
+
try:
|
656
|
+
index, itokens = insertions.next()
|
657
|
+
except StopIteration:
|
658
|
+
insleft = False
|
659
|
+
break # not strictly necessary
|
660
|
+
|