944 lines
34 KiB
Python
944 lines
34 KiB
Python
|
"""
|
||
|
pygments.lexer
|
||
|
~~~~~~~~~~~~~~
|
||
|
|
||
|
Base lexer classes.
|
||
|
|
||
|
:copyright: Copyright 2006-2023 by the Pygments team, see AUTHORS.
|
||
|
:license: BSD, see LICENSE for details.
|
||
|
"""
|
||
|
|
||
|
import re
|
||
|
import sys
|
||
|
import time
|
||
|
|
||
|
from pip._vendor.pygments.filter import apply_filters, Filter
|
||
|
from pip._vendor.pygments.filters import get_filter_by_name
|
||
|
from pip._vendor.pygments.token import Error, Text, Other, Whitespace, _TokenType
|
||
|
from pip._vendor.pygments.util import get_bool_opt, get_int_opt, get_list_opt, \
|
||
|
make_analysator, Future, guess_decode
|
||
|
from pip._vendor.pygments.regexopt import regex_opt
|
||
|
|
||
|
__all__ = ['Lexer', 'RegexLexer', 'ExtendedRegexLexer', 'DelegatingLexer',
|
||
|
'LexerContext', 'include', 'inherit', 'bygroups', 'using', 'this',
|
||
|
'default', 'words', 'line_re']
|
||
|
|
||
|
line_re = re.compile('.*?\n')
|
||
|
|
||
|
_encoding_map = [(b'\xef\xbb\xbf', 'utf-8'),
|
||
|
(b'\xff\xfe\0\0', 'utf-32'),
|
||
|
(b'\0\0\xfe\xff', 'utf-32be'),
|
||
|
(b'\xff\xfe', 'utf-16'),
|
||
|
(b'\xfe\xff', 'utf-16be')]
|
||
|
|
||
|
_default_analyse = staticmethod(lambda x: 0.0)
|
||
|
|
||
|
|
||
|
class LexerMeta(type):
|
||
|
"""
|
||
|
This metaclass automagically converts ``analyse_text`` methods into
|
||
|
static methods which always return float values.
|
||
|
"""
|
||
|
|
||
|
def __new__(mcs, name, bases, d):
|
||
|
if 'analyse_text' in d:
|
||
|
d['analyse_text'] = make_analysator(d['analyse_text'])
|
||
|
return type.__new__(mcs, name, bases, d)
|
||
|
|
||
|
|
||
|
class Lexer(metaclass=LexerMeta):
|
||
|
"""
|
||
|
Lexer for a specific language.
|
||
|
|
||
|
See also :doc:`lexerdevelopment`, a high-level guide to writing
|
||
|
lexers.
|
||
|
|
||
|
Lexer classes have attributes used for choosing the most appropriate
|
||
|
lexer based on various criteria.
|
||
|
|
||
|
.. autoattribute:: name
|
||
|
:no-value:
|
||
|
.. autoattribute:: aliases
|
||
|
:no-value:
|
||
|
.. autoattribute:: filenames
|
||
|
:no-value:
|
||
|
.. autoattribute:: alias_filenames
|
||
|
.. autoattribute:: mimetypes
|
||
|
:no-value:
|
||
|
.. autoattribute:: priority
|
||
|
|
||
|
Lexers included in Pygments should have an additional attribute:
|
||
|
|
||
|
.. autoattribute:: url
|
||
|
:no-value:
|
||
|
|
||
|
You can pass options to the constructor. The basic options recognized
|
||
|
by all lexers and processed by the base `Lexer` class are:
|
||
|
|
||
|
``stripnl``
|
||
|
Strip leading and trailing newlines from the input (default: True).
|
||
|
``stripall``
|
||
|
Strip all leading and trailing whitespace from the input
|
||
|
(default: False).
|
||
|
``ensurenl``
|
||
|
Make sure that the input ends with a newline (default: True). This
|
||
|
is required for some lexers that consume input linewise.
|
||
|
|
||
|
.. versionadded:: 1.3
|
||
|
|
||
|
``tabsize``
|
||
|
If given and greater than 0, expand tabs in the input (default: 0).
|
||
|
``encoding``
|
||
|
If given, must be an encoding name. This encoding will be used to
|
||
|
convert the input string to Unicode, if it is not already a Unicode
|
||
|
string (default: ``'guess'``, which uses a simple UTF-8 / Locale /
|
||
|
Latin1 detection. Can also be ``'chardet'`` to use the chardet
|
||
|
library, if it is installed.
|
||
|
``inencoding``
|
||
|
Overrides the ``encoding`` if given.
|
||
|
"""
|
||
|
|
||
|
#: Full name of the lexer, in human-readable form
|
||
|
name = None
|
||
|
|
||
|
#: A list of short, unique identifiers that can be used to look
|
||
|
#: up the lexer from a list, e.g., using `get_lexer_by_name()`.
|
||
|
aliases = []
|
||
|
|
||
|
#: A list of `fnmatch` patterns that match filenames which contain
|
||
|
#: content for this lexer. The patterns in this list should be unique among
|
||
|
#: all lexers.
|
||
|
filenames = []
|
||
|
|
||
|
#: A list of `fnmatch` patterns that match filenames which may or may not
|
||
|
#: contain content for this lexer. This list is used by the
|
||
|
#: :func:`.guess_lexer_for_filename()` function, to determine which lexers
|
||
|
#: are then included in guessing the correct one. That means that
|
||
|
#: e.g. every lexer for HTML and a template language should include
|
||
|
#: ``\*.html`` in this list.
|
||
|
alias_filenames = []
|
||
|
|
||
|
#: A list of MIME types for content that can be lexed with this lexer.
|
||
|
mimetypes = []
|
||
|
|
||
|
#: Priority, should multiple lexers match and no content is provided
|
||
|
priority = 0
|
||
|
|
||
|
#: URL of the language specification/definition. Used in the Pygments
|
||
|
#: documentation.
|
||
|
url = None
|
||
|
|
||
|
def __init__(self, **options):
|
||
|
"""
|
||
|
This constructor takes arbitrary options as keyword arguments.
|
||
|
Every subclass must first process its own options and then call
|
||
|
the `Lexer` constructor, since it processes the basic
|
||
|
options like `stripnl`.
|
||
|
|
||
|
An example looks like this:
|
||
|
|
||
|
.. sourcecode:: python
|
||
|
|
||
|
def __init__(self, **options):
|
||
|
self.compress = options.get('compress', '')
|
||
|
Lexer.__init__(self, **options)
|
||
|
|
||
|
As these options must all be specifiable as strings (due to the
|
||
|
command line usage), there are various utility functions
|
||
|
available to help with that, see `Utilities`_.
|
||
|
"""
|
||
|
self.options = options
|
||
|
self.stripnl = get_bool_opt(options, 'stripnl', True)
|
||
|
self.stripall = get_bool_opt(options, 'stripall', False)
|
||
|
self.ensurenl = get_bool_opt(options, 'ensurenl', True)
|
||
|
self.tabsize = get_int_opt(options, 'tabsize', 0)
|
||
|
self.encoding = options.get('encoding', 'guess')
|
||
|
self.encoding = options.get('inencoding') or self.encoding
|
||
|
self.filters = []
|
||
|
for filter_ in get_list_opt(options, 'filters', ()):
|
||
|
self.add_filter(filter_)
|
||
|
|
||
|
def __repr__(self):
|
||
|
if self.options:
|
||
|
return '<pygments.lexers.%s with %r>' % (self.__class__.__name__,
|
||
|
self.options)
|
||
|
else:
|
||
|
return '<pygments.lexers.%s>' % self.__class__.__name__
|
||
|
|
||
|
def add_filter(self, filter_, **options):
|
||
|
"""
|
||
|
Add a new stream filter to this lexer.
|
||
|
"""
|
||
|
if not isinstance(filter_, Filter):
|
||
|
filter_ = get_filter_by_name(filter_, **options)
|
||
|
self.filters.append(filter_)
|
||
|
|
||
|
def analyse_text(text):
|
||
|
"""
|
||
|
A static method which is called for lexer guessing.
|
||
|
|
||
|
It should analyse the text and return a float in the range
|
||
|
from ``0.0`` to ``1.0``. If it returns ``0.0``, the lexer
|
||
|
will not be selected as the most probable one, if it returns
|
||
|
``1.0``, it will be selected immediately. This is used by
|
||
|
`guess_lexer`.
|
||
|
|
||
|
The `LexerMeta` metaclass automatically wraps this function so
|
||
|
that it works like a static method (no ``self`` or ``cls``
|
||
|
parameter) and the return value is automatically converted to
|
||
|
`float`. If the return value is an object that is boolean `False`
|
||
|
it's the same as if the return values was ``0.0``.
|
||
|
"""
|
||
|
|
||
|
def get_tokens(self, text, unfiltered=False):
|
||
|
"""
|
||
|
This method is the basic interface of a lexer. It is called by
|
||
|
the `highlight()` function. It must process the text and return an
|
||
|
iterable of ``(tokentype, value)`` pairs from `text`.
|
||
|
|
||
|
Normally, you don't need to override this method. The default
|
||
|
implementation processes the options recognized by all lexers
|
||
|
(`stripnl`, `stripall` and so on), and then yields all tokens
|
||
|
from `get_tokens_unprocessed()`, with the ``index`` dropped.
|
||
|
|
||
|
If `unfiltered` is set to `True`, the filtering mechanism is
|
||
|
bypassed even if filters are defined.
|
||
|
"""
|
||
|
if not isinstance(text, str):
|
||
|
if self.encoding == 'guess':
|
||
|
text, _ = guess_decode(text)
|
||
|
elif self.encoding == 'chardet':
|
||
|
try:
|
||
|
from pip._vendor import chardet
|
||
|
except ImportError as e:
|
||
|
raise ImportError('To enable chardet encoding guessing, '
|
||
|
'please install the chardet library '
|
||
|
'from http://chardet.feedparser.org/') from e
|
||
|
# check for BOM first
|
||
|
decoded = None
|
||
|
for bom, encoding in _encoding_map:
|
||
|
if text.startswith(bom):
|
||
|
decoded = text[len(bom):].decode(encoding, 'replace')
|
||
|
break
|
||
|
# no BOM found, so use chardet
|
||
|
if decoded is None:
|
||
|
enc = chardet.detect(text[:1024]) # Guess using first 1KB
|
||
|
decoded = text.decode(enc.get('encoding') or 'utf-8',
|
||
|
'replace')
|
||
|
text = decoded
|
||
|
else:
|
||
|
text = text.decode(self.encoding)
|
||
|
if text.startswith('\ufeff'):
|
||
|
text = text[len('\ufeff'):]
|
||
|
else:
|
||
|
if text.startswith('\ufeff'):
|
||
|
text = text[len('\ufeff'):]
|
||
|
|
||
|
# text now *is* a unicode string
|
||
|
text = text.replace('\r\n', '\n')
|
||
|
text = text.replace('\r', '\n')
|
||
|
if self.stripall:
|
||
|
text = text.strip()
|
||
|
elif self.stripnl:
|
||
|
text = text.strip('\n')
|
||
|
if self.tabsize > 0:
|
||
|
text = text.expandtabs(self.tabsize)
|
||
|
if self.ensurenl and not text.endswith('\n'):
|
||
|
text += '\n'
|
||
|
|
||
|
def streamer():
|
||
|
for _, t, v in self.get_tokens_unprocessed(text):
|
||
|
yield t, v
|
||
|
stream = streamer()
|
||
|
if not unfiltered:
|
||
|
stream = apply_filters(stream, self.filters, self)
|
||
|
return stream
|
||
|
|
||
|
def get_tokens_unprocessed(self, text):
|
||
|
"""
|
||
|
This method should process the text and return an iterable of
|
||
|
``(index, tokentype, value)`` tuples where ``index`` is the starting
|
||
|
position of the token within the input text.
|
||
|
|
||
|
It must be overridden by subclasses. It is recommended to
|
||
|
implement it as a generator to maximize effectiveness.
|
||
|
"""
|
||
|
raise NotImplementedError
|
||
|
|
||
|
|
||
|
class DelegatingLexer(Lexer):
|
||
|
"""
|
||
|
This lexer takes two lexer as arguments. A root lexer and
|
||
|
a language lexer. First everything is scanned using the language
|
||
|
lexer, afterwards all ``Other`` tokens are lexed using the root
|
||
|
lexer.
|
||
|
|
||
|
The lexers from the ``template`` lexer package use this base lexer.
|
||
|
"""
|
||
|
|
||
|
def __init__(self, _root_lexer, _language_lexer, _needle=Other, **options):
|
||
|
self.root_lexer = _root_lexer(**options)
|
||
|
self.language_lexer = _language_lexer(**options)
|
||
|
self.needle = _needle
|
||
|
Lexer.__init__(self, **options)
|
||
|
|
||
|
def get_tokens_unprocessed(self, text):
|
||
|
buffered = ''
|
||
|
insertions = []
|
||
|
lng_buffer = []
|
||
|
for i, t, v in self.language_lexer.get_tokens_unprocessed(text):
|
||
|
if t is self.needle:
|
||
|
if lng_buffer:
|
||
|
insertions.append((len(buffered), lng_buffer))
|
||
|
lng_buffer = []
|
||
|
buffered += v
|
||
|
else:
|
||
|
lng_buffer.append((i, t, v))
|
||
|
if lng_buffer:
|
||
|
insertions.append((len(buffered), lng_buffer))
|
||
|
return do_insertions(insertions,
|
||
|
self.root_lexer.get_tokens_unprocessed(buffered))
|
||
|
|
||
|
|
||
|
# ------------------------------------------------------------------------------
|
||
|
# RegexLexer and ExtendedRegexLexer
|
||
|
#
|
||
|
|
||
|
|
||
|
class include(str): # pylint: disable=invalid-name
|
||
|
"""
|
||
|
Indicates that a state should include rules from another state.
|
||
|
"""
|
||
|
pass
|
||
|
|
||
|
|
||
|
class _inherit:
|
||
|
"""
|
||
|
Indicates the a state should inherit from its superclass.
|
||
|
"""
|
||
|
def __repr__(self):
|
||
|
return 'inherit'
|
||
|
|
||
|
inherit = _inherit() # pylint: disable=invalid-name
|
||
|
|
||
|
|
||
|
class combined(tuple): # pylint: disable=invalid-name
|
||
|
"""
|
||
|
Indicates a state combined from multiple states.
|
||
|
"""
|
||
|
|
||
|
def __new__(cls, *args):
|
||
|
return tuple.__new__(cls, args)
|
||
|
|
||
|
def __init__(self, *args):
|
||
|
# tuple.__init__ doesn't do anything
|
||
|
pass
|
||
|
|
||
|
|
||
|
class _PseudoMatch:
|
||
|
"""
|
||
|
A pseudo match object constructed from a string.
|
||
|
"""
|
||
|
|
||
|
def __init__(self, start, text):
|
||
|
self._text = text
|
||
|
self._start = start
|
||
|
|
||
|
def start(self, arg=None):
|
||
|
return self._start
|
||
|
|
||
|
def end(self, arg=None):
|
||
|
return self._start + len(self._text)
|
||
|
|
||
|
def group(self, arg=None):
|
||
|
if arg:
|
||
|
raise IndexError('No such group')
|
||
|
return self._text
|
||
|
|
||
|
def groups(self):
|
||
|
return (self._text,)
|
||
|
|
||
|
def groupdict(self):
|
||
|
return {}
|
||
|
|
||
|
|
||
|
def bygroups(*args):
|
||
|
"""
|
||
|
Callback that yields multiple actions for each group in the match.
|
||
|
"""
|
||
|
def callback(lexer, match, ctx=None):
|
||
|
for i, action in enumerate(args):
|
||
|
if action is None:
|
||
|
continue
|
||
|
elif type(action) is _TokenType:
|
||
|
data = match.group(i + 1)
|
||
|
if data:
|
||
|
yield match.start(i + 1), action, data
|
||
|
else:
|
||
|
data = match.group(i + 1)
|
||
|
if data is not None:
|
||
|
if ctx:
|
||
|
ctx.pos = match.start(i + 1)
|
||
|
for item in action(lexer,
|
||
|
_PseudoMatch(match.start(i + 1), data), ctx):
|
||
|
if item:
|
||
|
yield item
|
||
|
if ctx:
|
||
|
ctx.pos = match.end()
|
||
|
return callback
|
||
|
|
||
|
|
||
|
class _This:
|
||
|
"""
|
||
|
Special singleton used for indicating the caller class.
|
||
|
Used by ``using``.
|
||
|
"""
|
||
|
|
||
|
this = _This()
|
||
|
|
||
|
|
||
|
def using(_other, **kwargs):
|
||
|
"""
|
||
|
Callback that processes the match with a different lexer.
|
||
|
|
||
|
The keyword arguments are forwarded to the lexer, except `state` which
|
||
|
is handled separately.
|
||
|
|
||
|
`state` specifies the state that the new lexer will start in, and can
|
||
|
be an enumerable such as ('root', 'inline', 'string') or a simple
|
||
|
string which is assumed to be on top of the root state.
|
||
|
|
||
|
Note: For that to work, `_other` must not be an `ExtendedRegexLexer`.
|
||
|
"""
|
||
|
gt_kwargs = {}
|
||
|
if 'state' in kwargs:
|
||
|
s = kwargs.pop('state')
|
||
|
if isinstance(s, (list, tuple)):
|
||
|
gt_kwargs['stack'] = s
|
||
|
else:
|
||
|
gt_kwargs['stack'] = ('root', s)
|
||
|
|
||
|
if _other is this:
|
||
|
def callback(lexer, match, ctx=None):
|
||
|
# if keyword arguments are given the callback
|
||
|
# function has to create a new lexer instance
|
||
|
if kwargs:
|
||
|
# XXX: cache that somehow
|
||
|
kwargs.update(lexer.options)
|
||
|
lx = lexer.__class__(**kwargs)
|
||
|
else:
|
||
|
lx = lexer
|
||
|
s = match.start()
|
||
|
for i, t, v in lx.get_tokens_unprocessed(match.group(), **gt_kwargs):
|
||
|
yield i + s, t, v
|
||
|
if ctx:
|
||
|
ctx.pos = match.end()
|
||
|
else:
|
||
|
def callback(lexer, match, ctx=None):
|
||
|
# XXX: cache that somehow
|
||
|
kwargs.update(lexer.options)
|
||
|
lx = _other(**kwargs)
|
||
|
|
||
|
s = match.start()
|
||
|
for i, t, v in lx.get_tokens_unprocessed(match.group(), **gt_kwargs):
|
||
|
yield i + s, t, v
|
||
|
if ctx:
|
||
|
ctx.pos = match.end()
|
||
|
return callback
|
||
|
|
||
|
|
||
|
class default:
|
||
|
"""
|
||
|
Indicates a state or state action (e.g. #pop) to apply.
|
||
|
For example default('#pop') is equivalent to ('', Token, '#pop')
|
||
|
Note that state tuples may be used as well.
|
||
|
|
||
|
.. versionadded:: 2.0
|
||
|
"""
|
||
|
def __init__(self, state):
|
||
|
self.state = state
|
||
|
|
||
|
|
||
|
class words(Future):
|
||
|
"""
|
||
|
Indicates a list of literal words that is transformed into an optimized
|
||
|
regex that matches any of the words.
|
||
|
|
||
|
.. versionadded:: 2.0
|
||
|
"""
|
||
|
def __init__(self, words, prefix='', suffix=''):
|
||
|
self.words = words
|
||
|
self.prefix = prefix
|
||
|
self.suffix = suffix
|
||
|
|
||
|
def get(self):
|
||
|
return regex_opt(self.words, prefix=self.prefix, suffix=self.suffix)
|
||
|
|
||
|
|
||
|
class RegexLexerMeta(LexerMeta):
|
||
|
"""
|
||
|
Metaclass for RegexLexer, creates the self._tokens attribute from
|
||
|
self.tokens on the first instantiation.
|
||
|
"""
|
||
|
|
||
|
def _process_regex(cls, regex, rflags, state):
|
||
|
"""Preprocess the regular expression component of a token definition."""
|
||
|
if isinstance(regex, Future):
|
||
|
regex = regex.get()
|
||
|
return re.compile(regex, rflags).match
|
||
|
|
||
|
def _process_token(cls, token):
|
||
|
"""Preprocess the token component of a token definition."""
|
||
|
assert type(token) is _TokenType or callable(token), \
|
||
|
'token type must be simple type or callable, not %r' % (token,)
|
||
|
return token
|
||
|
|
||
|
def _process_new_state(cls, new_state, unprocessed, processed):
|
||
|
"""Preprocess the state transition action of a token definition."""
|
||
|
if isinstance(new_state, str):
|
||
|
# an existing state
|
||
|
if new_state == '#pop':
|
||
|
return -1
|
||
|
elif new_state in unprocessed:
|
||
|
return (new_state,)
|
||
|
elif new_state == '#push':
|
||
|
return new_state
|
||
|
elif new_state[:5] == '#pop:':
|
||
|
return -int(new_state[5:])
|
||
|
else:
|
||
|
assert False, 'unknown new state %r' % new_state
|
||
|
elif isinstance(new_state, combined):
|
||
|
# combine a new state from existing ones
|
||
|
tmp_state = '_tmp_%d' % cls._tmpname
|
||
|
cls._tmpname += 1
|
||
|
itokens = []
|
||
|
for istate in new_state:
|
||
|
assert istate != new_state, 'circular state ref %r' % istate
|
||
|
itokens.extend(cls._process_state(unprocessed,
|
||
|
processed, istate))
|
||
|
processed[tmp_state] = itokens
|
||
|
return (tmp_state,)
|
||
|
elif isinstance(new_state, tuple):
|
||
|
# push more than one state
|
||
|
for istate in new_state:
|
||
|
assert (istate in unprocessed or
|
||
|
istate in ('#pop', '#push')), \
|
||
|
'unknown new state ' + istate
|
||
|
return new_state
|
||
|
else:
|
||
|
assert False, 'unknown new state def %r' % new_state
|
||
|
|
||
|
def _process_state(cls, unprocessed, processed, state):
|
||
|
"""Preprocess a single state definition."""
|
||
|
assert type(state) is str, "wrong state name %r" % state
|
||
|
assert state[0] != '#', "invalid state name %r" % state
|
||
|
if state in processed:
|
||
|
return processed[state]
|
||
|
tokens = processed[state] = []
|
||
|
rflags = cls.flags
|
||
|
for tdef in unprocessed[state]:
|
||
|
if isinstance(tdef, include):
|
||
|
# it's a state reference
|
||
|
assert tdef != state, "circular state reference %r" % state
|
||
|
tokens.extend(cls._process_state(unprocessed, processed,
|
||
|
str(tdef)))
|
||
|
continue
|
||
|
if isinstance(tdef, _inherit):
|
||
|
# should be processed already, but may not in the case of:
|
||
|
# 1. the state has no counterpart in any parent
|
||
|
# 2. the state includes more than one 'inherit'
|
||
|
continue
|
||
|
if isinstance(tdef, default):
|
||
|
new_state = cls._process_new_state(tdef.state, unprocessed, processed)
|
||
|
tokens.append((re.compile('').match, None, new_state))
|
||
|
continue
|
||
|
|
||
|
assert type(tdef) is tuple, "wrong rule def %r" % tdef
|
||
|
|
||
|
try:
|
||
|
rex = cls._process_regex(tdef[0], rflags, state)
|
||
|
except Exception as err:
|
||
|
raise ValueError("uncompilable regex %r in state %r of %r: %s" %
|
||
|
(tdef[0], state, cls, err)) from err
|
||
|
|
||
|
token = cls._process_token(tdef[1])
|
||
|
|
||
|
if len(tdef) == 2:
|
||
|
new_state = None
|
||
|
else:
|
||
|
new_state = cls._process_new_state(tdef[2],
|
||
|
unprocessed, processed)
|
||
|
|
||
|
tokens.append((rex, token, new_state))
|
||
|
return tokens
|
||
|
|
||
|
def process_tokendef(cls, name, tokendefs=None):
|
||
|
"""Preprocess a dictionary of token definitions."""
|
||
|
processed = cls._all_tokens[name] = {}
|
||
|
tokendefs = tokendefs or cls.tokens[name]
|
||
|
for state in list(tokendefs):
|
||
|
cls._process_state(tokendefs, processed, state)
|
||
|
return processed
|
||
|
|
||
|
def get_tokendefs(cls):
|
||
|
"""
|
||
|
Merge tokens from superclasses in MRO order, returning a single tokendef
|
||
|
dictionary.
|
||
|
|
||
|
Any state that is not defined by a subclass will be inherited
|
||
|
automatically. States that *are* defined by subclasses will, by
|
||
|
default, override that state in the superclass. If a subclass wishes to
|
||
|
inherit definitions from a superclass, it can use the special value
|
||
|
"inherit", which will cause the superclass' state definition to be
|
||
|
included at that point in the state.
|
||
|
"""
|
||
|
tokens = {}
|
||
|
inheritable = {}
|
||
|
for c in cls.__mro__:
|
||
|
toks = c.__dict__.get('tokens', {})
|
||
|
|
||
|
for state, items in toks.items():
|
||
|
curitems = tokens.get(state)
|
||
|
if curitems is None:
|
||
|
# N.b. because this is assigned by reference, sufficiently
|
||
|
# deep hierarchies are processed incrementally (e.g. for
|
||
|
# A(B), B(C), C(RegexLexer), B will be premodified so X(B)
|
||
|
# will not see any inherits in B).
|
||
|
tokens[state] = items
|
||
|
try:
|
||
|
inherit_ndx = items.index(inherit)
|
||
|
except ValueError:
|
||
|
continue
|
||
|
inheritable[state] = inherit_ndx
|
||
|
continue
|
||
|
|
||
|
inherit_ndx = inheritable.pop(state, None)
|
||
|
if inherit_ndx is None:
|
||
|
continue
|
||
|
|
||
|
# Replace the "inherit" value with the items
|
||
|
curitems[inherit_ndx:inherit_ndx+1] = items
|
||
|
try:
|
||
|
# N.b. this is the index in items (that is, the superclass
|
||
|
# copy), so offset required when storing below.
|
||
|
new_inh_ndx = items.index(inherit)
|
||
|
except ValueError:
|
||
|
pass
|
||
|
else:
|
||
|
inheritable[state] = inherit_ndx + new_inh_ndx
|
||
|
|
||
|
return tokens
|
||
|
|
||
|
def __call__(cls, *args, **kwds):
|
||
|
"""Instantiate cls after preprocessing its token definitions."""
|
||
|
if '_tokens' not in cls.__dict__:
|
||
|
cls._all_tokens = {}
|
||
|
cls._tmpname = 0
|
||
|
if hasattr(cls, 'token_variants') and cls.token_variants:
|
||
|
# don't process yet
|
||
|
pass
|
||
|
else:
|
||
|
cls._tokens = cls.process_tokendef('', cls.get_tokendefs())
|
||
|
|
||
|
return type.__call__(cls, *args, **kwds)
|
||
|
|
||
|
|
||
|
class RegexLexer(Lexer, metaclass=RegexLexerMeta):
|
||
|
"""
|
||
|
Base for simple stateful regular expression-based lexers.
|
||
|
Simplifies the lexing process so that you need only
|
||
|
provide a list of states and regular expressions.
|
||
|
"""
|
||
|
|
||
|
#: Flags for compiling the regular expressions.
|
||
|
#: Defaults to MULTILINE.
|
||
|
flags = re.MULTILINE
|
||
|
|
||
|
#: At all time there is a stack of states. Initially, the stack contains
|
||
|
#: a single state 'root'. The top of the stack is called "the current state".
|
||
|
#:
|
||
|
#: Dict of ``{'state': [(regex, tokentype, new_state), ...], ...}``
|
||
|
#:
|
||
|
#: ``new_state`` can be omitted to signify no state transition.
|
||
|
#: If ``new_state`` is a string, it is pushed on the stack. This ensure
|
||
|
#: the new current state is ``new_state``.
|
||
|
#: If ``new_state`` is a tuple of strings, all of those strings are pushed
|
||
|
#: on the stack and the current state will be the last element of the list.
|
||
|
#: ``new_state`` can also be ``combined('state1', 'state2', ...)``
|
||
|
#: to signify a new, anonymous state combined from the rules of two
|
||
|
#: or more existing ones.
|
||
|
#: Furthermore, it can be '#pop' to signify going back one step in
|
||
|
#: the state stack, or '#push' to push the current state on the stack
|
||
|
#: again. Note that if you push while in a combined state, the combined
|
||
|
#: state itself is pushed, and not only the state in which the rule is
|
||
|
#: defined.
|
||
|
#:
|
||
|
#: The tuple can also be replaced with ``include('state')``, in which
|
||
|
#: case the rules from the state named by the string are included in the
|
||
|
#: current one.
|
||
|
tokens = {}
|
||
|
|
||
|
def get_tokens_unprocessed(self, text, stack=('root',)):
|
||
|
"""
|
||
|
Split ``text`` into (tokentype, text) pairs.
|
||
|
|
||
|
``stack`` is the initial stack (default: ``['root']``)
|
||
|
"""
|
||
|
pos = 0
|
||
|
tokendefs = self._tokens
|
||
|
statestack = list(stack)
|
||
|
statetokens = tokendefs[statestack[-1]]
|
||
|
while 1:
|
||
|
for rexmatch, action, new_state in statetokens:
|
||
|
m = rexmatch(text, pos)
|
||
|
if m:
|
||
|
if action is not None:
|
||
|
if type(action) is _TokenType:
|
||
|
yield pos, action, m.group()
|
||
|
else:
|
||
|
yield from action(self, m)
|
||
|
pos = m.end()
|
||
|
if new_state is not None:
|
||
|
# state transition
|
||
|
if isinstance(new_state, tuple):
|
||
|
for state in new_state:
|
||
|
if state == '#pop':
|
||
|
if len(statestack) > 1:
|
||
|
statestack.pop()
|
||
|
elif state == '#push':
|
||
|
statestack.append(statestack[-1])
|
||
|
else:
|
||
|
statestack.append(state)
|
||
|
elif isinstance(new_state, int):
|
||
|
# pop, but keep at least one state on the stack
|
||
|
# (random code leading to unexpected pops should
|
||
|
# not allow exceptions)
|
||
|
if abs(new_state) >= len(statestack):
|
||
|
del statestack[1:]
|
||
|
else:
|
||
|
del statestack[new_state:]
|
||
|
elif new_state == '#push':
|
||
|
statestack.append(statestack[-1])
|
||
|
else:
|
||
|
assert False, "wrong state def: %r" % new_state
|
||
|
statetokens = tokendefs[statestack[-1]]
|
||
|
break
|
||
|
else:
|
||
|
# We are here only if all state tokens have been considered
|
||
|
# and there was not a match on any of them.
|
||
|
try:
|
||
|
if text[pos] == '\n':
|
||
|
# at EOL, reset state to "root"
|
||
|
statestack = ['root']
|
||
|
statetokens = tokendefs['root']
|
||
|
yield pos, Whitespace, '\n'
|
||
|
pos += 1
|
||
|
continue
|
||
|
yield pos, Error, text[pos]
|
||
|
pos += 1
|
||
|
except IndexError:
|
||
|
break
|
||
|
|
||
|
|
||
|
class LexerContext:
|
||
|
"""
|
||
|
A helper object that holds lexer position data.
|
||
|
"""
|
||
|
|
||
|
def __init__(self, text, pos, stack=None, end=None):
|
||
|
self.text = text
|
||
|
self.pos = pos
|
||
|
self.end = end or len(text) # end=0 not supported ;-)
|
||
|
self.stack = stack or ['root']
|
||
|
|
||
|
def __repr__(self):
|
||
|
return 'LexerContext(%r, %r, %r)' % (
|
||
|
self.text, self.pos, self.stack)
|
||
|
|
||
|
|
||
|
class ExtendedRegexLexer(RegexLexer):
|
||
|
"""
|
||
|
A RegexLexer that uses a context object to store its state.
|
||
|
"""
|
||
|
|
||
|
def get_tokens_unprocessed(self, text=None, context=None):
|
||
|
"""
|
||
|
Split ``text`` into (tokentype, text) pairs.
|
||
|
If ``context`` is given, use this lexer context instead.
|
||
|
"""
|
||
|
tokendefs = self._tokens
|
||
|
if not context:
|
||
|
ctx = LexerContext(text, 0)
|
||
|
statetokens = tokendefs['root']
|
||
|
else:
|
||
|
ctx = context
|
||
|
statetokens = tokendefs[ctx.stack[-1]]
|
||
|
text = ctx.text
|
||
|
while 1:
|
||
|
for rexmatch, action, new_state in statetokens:
|
||
|
m = rexmatch(text, ctx.pos, ctx.end)
|
||
|
if m:
|
||
|
if action is not None:
|
||
|
if type(action) is _TokenType:
|
||
|
yield ctx.pos, action, m.group()
|
||
|
ctx.pos = m.end()
|
||
|
else:
|
||
|
yield from action(self, m, ctx)
|
||
|
if not new_state:
|
||
|
# altered the state stack?
|
||
|
statetokens = tokendefs[ctx.stack[-1]]
|
||
|
# CAUTION: callback must set ctx.pos!
|
||
|
if new_state is not None:
|
||
|
# state transition
|
||
|
if isinstance(new_state, tuple):
|
||
|
for state in new_state:
|
||
|
if state == '#pop':
|
||
|
if len(ctx.stack) > 1:
|
||
|
ctx.stack.pop()
|
||
|
elif state == '#push':
|
||
|
ctx.stack.append(ctx.stack[-1])
|
||
|
else:
|
||
|
ctx.stack.append(state)
|
||
|
elif isinstance(new_state, int):
|
||
|
# see RegexLexer for why this check is made
|
||
|
if abs(new_state) >= len(ctx.stack):
|
||
|
del ctx.stack[1:]
|
||
|
else:
|
||
|
del ctx.stack[new_state:]
|
||
|
elif new_state == '#push':
|
||
|
ctx.stack.append(ctx.stack[-1])
|
||
|
else:
|
||
|
assert False, "wrong state def: %r" % new_state
|
||
|
statetokens = tokendefs[ctx.stack[-1]]
|
||
|
break
|
||
|
else:
|
||
|
try:
|
||
|
if ctx.pos >= ctx.end:
|
||
|
break
|
||
|
if text[ctx.pos] == '\n':
|
||
|
# at EOL, reset state to "root"
|
||
|
ctx.stack = ['root']
|
||
|
statetokens = tokendefs['root']
|
||
|
yield ctx.pos, Text, '\n'
|
||
|
ctx.pos += 1
|
||
|
continue
|
||
|
yield ctx.pos, Error, text[ctx.pos]
|
||
|
ctx.pos += 1
|
||
|
except IndexError:
|
||
|
break
|
||
|
|
||
|
|
||
|
def do_insertions(insertions, tokens):
|
||
|
"""
|
||
|
Helper for lexers which must combine the results of several
|
||
|
sublexers.
|
||
|
|
||
|
``insertions`` is a list of ``(index, itokens)`` pairs.
|
||
|
Each ``itokens`` iterable should be inserted at position
|
||
|
``index`` into the token stream given by the ``tokens``
|
||
|
argument.
|
||
|
|
||
|
The result is a combined token stream.
|
||
|
|
||
|
TODO: clean up the code here.
|
||
|
"""
|
||
|
insertions = iter(insertions)
|
||
|
try:
|
||
|
index, itokens = next(insertions)
|
||
|
except StopIteration:
|
||
|
# no insertions
|
||
|
yield from tokens
|
||
|
return
|
||
|
|
||
|
realpos = None
|
||
|
insleft = True
|
||
|
|
||
|
# iterate over the token stream where we want to insert
|
||
|
# the tokens from the insertion list.
|
||
|
for i, t, v in tokens:
|
||
|
# first iteration. store the position of first item
|
||
|
if realpos is None:
|
||
|
realpos = i
|
||
|
oldi = 0
|
||
|
while insleft and i + len(v) >= index:
|
||
|
tmpval = v[oldi:index - i]
|
||
|
if tmpval:
|
||
|
yield realpos, t, tmpval
|
||
|
realpos += len(tmpval)
|
||
|
for it_index, it_token, it_value in itokens:
|
||
|
yield realpos, it_token, it_value
|
||
|
realpos += len(it_value)
|
||
|
oldi = index - i
|
||
|
try:
|
||
|
index, itokens = next(insertions)
|
||
|
except StopIteration:
|
||
|
insleft = False
|
||
|
break # not strictly necessary
|
||
|
if oldi < len(v):
|
||
|
yield realpos, t, v[oldi:]
|
||
|
realpos += len(v) - oldi
|
||
|
|
||
|
# leftover tokens
|
||
|
while insleft:
|
||
|
# no normal tokens, set realpos to zero
|
||
|
realpos = realpos or 0
|
||
|
for p, t, v in itokens:
|
||
|
yield realpos, t, v
|
||
|
realpos += len(v)
|
||
|
try:
|
||
|
index, itokens = next(insertions)
|
||
|
except StopIteration:
|
||
|
insleft = False
|
||
|
break # not strictly necessary
|
||
|
|
||
|
|
||
|
class ProfilingRegexLexerMeta(RegexLexerMeta):
|
||
|
"""Metaclass for ProfilingRegexLexer, collects regex timing info."""
|
||
|
|
||
|
def _process_regex(cls, regex, rflags, state):
|
||
|
if isinstance(regex, words):
|
||
|
rex = regex_opt(regex.words, prefix=regex.prefix,
|
||
|
suffix=regex.suffix)
|
||
|
else:
|
||
|
rex = regex
|
||
|
compiled = re.compile(rex, rflags)
|
||
|
|
||
|
def match_func(text, pos, endpos=sys.maxsize):
|
||
|
info = cls._prof_data[-1].setdefault((state, rex), [0, 0.0])
|
||
|
t0 = time.time()
|
||
|
res = compiled.match(text, pos, endpos)
|
||
|
t1 = time.time()
|
||
|
info[0] += 1
|
||
|
info[1] += t1 - t0
|
||
|
return res
|
||
|
return match_func
|
||
|
|
||
|
|
||
|
class ProfilingRegexLexer(RegexLexer, metaclass=ProfilingRegexLexerMeta):
|
||
|
"""Drop-in replacement for RegexLexer that does profiling of its regexes."""
|
||
|
|
||
|
_prof_data = []
|
||
|
_prof_sort_index = 4 # defaults to time per call
|
||
|
|
||
|
def get_tokens_unprocessed(self, text, stack=('root',)):
|
||
|
# this needs to be a stack, since using(this) will produce nested calls
|
||
|
self.__class__._prof_data.append({})
|
||
|
yield from RegexLexer.get_tokens_unprocessed(self, text, stack)
|
||
|
rawdata = self.__class__._prof_data.pop()
|
||
|
data = sorted(((s, repr(r).strip('u\'').replace('\\\\', '\\')[:65],
|
||
|
n, 1000 * t, 1000 * t / n)
|
||
|
for ((s, r), (n, t)) in rawdata.items()),
|
||
|
key=lambda x: x[self._prof_sort_index],
|
||
|
reverse=True)
|
||
|
sum_total = sum(x[3] for x in data)
|
||
|
|
||
|
print()
|
||
|
print('Profiling result for %s lexing %d chars in %.3f ms' %
|
||
|
(self.__class__.__name__, len(text), sum_total))
|
||
|
print('=' * 110)
|
||
|
print('%-20s %-64s ncalls tottime percall' % ('state', 'regex'))
|
||
|
print('-' * 110)
|
||
|
for d in data:
|
||
|
print('%-20s %-65s %5d %8.4f %8.4f' % d)
|
||
|
print('=' * 110)
|