You cannot select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
867 lines
29 KiB
Python
867 lines
29 KiB
Python
4 months ago
|
"""Implements a Jinja / Python combination lexer. The ``Lexer`` class
|
||
|
is used to do some preprocessing. It filters out invalid operators like
|
||
|
the bitshift operators we don't allow in templates. It separates
|
||
|
template code and python code in expressions.
|
||
|
"""
|
||
|
import re
|
||
|
import typing as t
|
||
|
from ast import literal_eval
|
||
|
from collections import deque
|
||
|
from sys import intern
|
||
|
|
||
|
from ._identifier import pattern as name_re
|
||
|
from .exceptions import TemplateSyntaxError
|
||
|
from .utils import LRUCache
|
||
|
|
||
|
if t.TYPE_CHECKING:
|
||
|
import typing_extensions as te
|
||
|
from .environment import Environment
|
||
|
|
||
|
# cache for the lexers. Exists in order to be able to have multiple
|
||
|
# environments with the same lexer
|
||
|
_lexer_cache: t.MutableMapping[t.Tuple, "Lexer"] = LRUCache(50) # type: ignore
|
||
|
|
||
|
# static regular expressions
|
||
|
whitespace_re = re.compile(r"\s+")
|
||
|
newline_re = re.compile(r"(\r\n|\r|\n)")
|
||
|
string_re = re.compile(
|
||
|
r"('([^'\\]*(?:\\.[^'\\]*)*)'" r'|"([^"\\]*(?:\\.[^"\\]*)*)")', re.S
|
||
|
)
|
||
|
integer_re = re.compile(
|
||
|
r"""
|
||
|
(
|
||
|
0b(_?[0-1])+ # binary
|
||
|
|
|
||
|
0o(_?[0-7])+ # octal
|
||
|
|
|
||
|
0x(_?[\da-f])+ # hex
|
||
|
|
|
||
|
[1-9](_?\d)* # decimal
|
||
|
|
|
||
|
0(_?0)* # decimal zero
|
||
|
)
|
||
|
""",
|
||
|
re.IGNORECASE | re.VERBOSE,
|
||
|
)
|
||
|
float_re = re.compile(
|
||
|
r"""
|
||
|
(?<!\.) # doesn't start with a .
|
||
|
(\d+_)*\d+ # digits, possibly _ separated
|
||
|
(
|
||
|
(\.(\d+_)*\d+)? # optional fractional part
|
||
|
e[+\-]?(\d+_)*\d+ # exponent part
|
||
|
|
|
||
|
\.(\d+_)*\d+ # required fractional part
|
||
|
)
|
||
|
""",
|
||
|
re.IGNORECASE | re.VERBOSE,
|
||
|
)
|
||
|
|
||
|
# internal the tokens and keep references to them
|
||
|
TOKEN_ADD = intern("add")
|
||
|
TOKEN_ASSIGN = intern("assign")
|
||
|
TOKEN_COLON = intern("colon")
|
||
|
TOKEN_COMMA = intern("comma")
|
||
|
TOKEN_DIV = intern("div")
|
||
|
TOKEN_DOT = intern("dot")
|
||
|
TOKEN_EQ = intern("eq")
|
||
|
TOKEN_FLOORDIV = intern("floordiv")
|
||
|
TOKEN_GT = intern("gt")
|
||
|
TOKEN_GTEQ = intern("gteq")
|
||
|
TOKEN_LBRACE = intern("lbrace")
|
||
|
TOKEN_LBRACKET = intern("lbracket")
|
||
|
TOKEN_LPAREN = intern("lparen")
|
||
|
TOKEN_LT = intern("lt")
|
||
|
TOKEN_LTEQ = intern("lteq")
|
||
|
TOKEN_MOD = intern("mod")
|
||
|
TOKEN_MUL = intern("mul")
|
||
|
TOKEN_NE = intern("ne")
|
||
|
TOKEN_PIPE = intern("pipe")
|
||
|
TOKEN_POW = intern("pow")
|
||
|
TOKEN_RBRACE = intern("rbrace")
|
||
|
TOKEN_RBRACKET = intern("rbracket")
|
||
|
TOKEN_RPAREN = intern("rparen")
|
||
|
TOKEN_SEMICOLON = intern("semicolon")
|
||
|
TOKEN_SUB = intern("sub")
|
||
|
TOKEN_TILDE = intern("tilde")
|
||
|
TOKEN_WHITESPACE = intern("whitespace")
|
||
|
TOKEN_FLOAT = intern("float")
|
||
|
TOKEN_INTEGER = intern("integer")
|
||
|
TOKEN_NAME = intern("name")
|
||
|
TOKEN_STRING = intern("string")
|
||
|
TOKEN_OPERATOR = intern("operator")
|
||
|
TOKEN_BLOCK_BEGIN = intern("block_begin")
|
||
|
TOKEN_BLOCK_END = intern("block_end")
|
||
|
TOKEN_VARIABLE_BEGIN = intern("variable_begin")
|
||
|
TOKEN_VARIABLE_END = intern("variable_end")
|
||
|
TOKEN_RAW_BEGIN = intern("raw_begin")
|
||
|
TOKEN_RAW_END = intern("raw_end")
|
||
|
TOKEN_COMMENT_BEGIN = intern("comment_begin")
|
||
|
TOKEN_COMMENT_END = intern("comment_end")
|
||
|
TOKEN_COMMENT = intern("comment")
|
||
|
TOKEN_LINESTATEMENT_BEGIN = intern("linestatement_begin")
|
||
|
TOKEN_LINESTATEMENT_END = intern("linestatement_end")
|
||
|
TOKEN_LINECOMMENT_BEGIN = intern("linecomment_begin")
|
||
|
TOKEN_LINECOMMENT_END = intern("linecomment_end")
|
||
|
TOKEN_LINECOMMENT = intern("linecomment")
|
||
|
TOKEN_DATA = intern("data")
|
||
|
TOKEN_INITIAL = intern("initial")
|
||
|
TOKEN_EOF = intern("eof")
|
||
|
|
||
|
# bind operators to token types
|
||
|
operators = {
|
||
|
"+": TOKEN_ADD,
|
||
|
"-": TOKEN_SUB,
|
||
|
"/": TOKEN_DIV,
|
||
|
"//": TOKEN_FLOORDIV,
|
||
|
"*": TOKEN_MUL,
|
||
|
"%": TOKEN_MOD,
|
||
|
"**": TOKEN_POW,
|
||
|
"~": TOKEN_TILDE,
|
||
|
"[": TOKEN_LBRACKET,
|
||
|
"]": TOKEN_RBRACKET,
|
||
|
"(": TOKEN_LPAREN,
|
||
|
")": TOKEN_RPAREN,
|
||
|
"{": TOKEN_LBRACE,
|
||
|
"}": TOKEN_RBRACE,
|
||
|
"==": TOKEN_EQ,
|
||
|
"!=": TOKEN_NE,
|
||
|
">": TOKEN_GT,
|
||
|
">=": TOKEN_GTEQ,
|
||
|
"<": TOKEN_LT,
|
||
|
"<=": TOKEN_LTEQ,
|
||
|
"=": TOKEN_ASSIGN,
|
||
|
".": TOKEN_DOT,
|
||
|
":": TOKEN_COLON,
|
||
|
"|": TOKEN_PIPE,
|
||
|
",": TOKEN_COMMA,
|
||
|
";": TOKEN_SEMICOLON,
|
||
|
}
|
||
|
|
||
|
reverse_operators = {v: k for k, v in operators.items()}
|
||
|
assert len(operators) == len(reverse_operators), "operators dropped"
|
||
|
operator_re = re.compile(
|
||
|
f"({'|'.join(re.escape(x) for x in sorted(operators, key=lambda x: -len(x)))})"
|
||
|
)
|
||
|
|
||
|
ignored_tokens = frozenset(
|
||
|
[
|
||
|
TOKEN_COMMENT_BEGIN,
|
||
|
TOKEN_COMMENT,
|
||
|
TOKEN_COMMENT_END,
|
||
|
TOKEN_WHITESPACE,
|
||
|
TOKEN_LINECOMMENT_BEGIN,
|
||
|
TOKEN_LINECOMMENT_END,
|
||
|
TOKEN_LINECOMMENT,
|
||
|
]
|
||
|
)
|
||
|
ignore_if_empty = frozenset(
|
||
|
[TOKEN_WHITESPACE, TOKEN_DATA, TOKEN_COMMENT, TOKEN_LINECOMMENT]
|
||
|
)
|
||
|
|
||
|
|
||
|
def _describe_token_type(token_type: str) -> str:
|
||
|
if token_type in reverse_operators:
|
||
|
return reverse_operators[token_type]
|
||
|
|
||
|
return {
|
||
|
TOKEN_COMMENT_BEGIN: "begin of comment",
|
||
|
TOKEN_COMMENT_END: "end of comment",
|
||
|
TOKEN_COMMENT: "comment",
|
||
|
TOKEN_LINECOMMENT: "comment",
|
||
|
TOKEN_BLOCK_BEGIN: "begin of statement block",
|
||
|
TOKEN_BLOCK_END: "end of statement block",
|
||
|
TOKEN_VARIABLE_BEGIN: "begin of print statement",
|
||
|
TOKEN_VARIABLE_END: "end of print statement",
|
||
|
TOKEN_LINESTATEMENT_BEGIN: "begin of line statement",
|
||
|
TOKEN_LINESTATEMENT_END: "end of line statement",
|
||
|
TOKEN_DATA: "template data / text",
|
||
|
TOKEN_EOF: "end of template",
|
||
|
}.get(token_type, token_type)
|
||
|
|
||
|
|
||
|
def describe_token(token: "Token") -> str:
|
||
|
"""Returns a description of the token."""
|
||
|
if token.type == TOKEN_NAME:
|
||
|
return token.value
|
||
|
|
||
|
return _describe_token_type(token.type)
|
||
|
|
||
|
|
||
|
def describe_token_expr(expr: str) -> str:
|
||
|
"""Like `describe_token` but for token expressions."""
|
||
|
if ":" in expr:
|
||
|
type, value = expr.split(":", 1)
|
||
|
|
||
|
if type == TOKEN_NAME:
|
||
|
return value
|
||
|
else:
|
||
|
type = expr
|
||
|
|
||
|
return _describe_token_type(type)
|
||
|
|
||
|
|
||
|
def count_newlines(value: str) -> int:
|
||
|
"""Count the number of newline characters in the string. This is
|
||
|
useful for extensions that filter a stream.
|
||
|
"""
|
||
|
return len(newline_re.findall(value))
|
||
|
|
||
|
|
||
|
def compile_rules(environment: "Environment") -> t.List[t.Tuple[str, str]]:
|
||
|
"""Compiles all the rules from the environment into a list of rules."""
|
||
|
e = re.escape
|
||
|
rules = [
|
||
|
(
|
||
|
len(environment.comment_start_string),
|
||
|
TOKEN_COMMENT_BEGIN,
|
||
|
e(environment.comment_start_string),
|
||
|
),
|
||
|
(
|
||
|
len(environment.block_start_string),
|
||
|
TOKEN_BLOCK_BEGIN,
|
||
|
e(environment.block_start_string),
|
||
|
),
|
||
|
(
|
||
|
len(environment.variable_start_string),
|
||
|
TOKEN_VARIABLE_BEGIN,
|
||
|
e(environment.variable_start_string),
|
||
|
),
|
||
|
]
|
||
|
|
||
|
if environment.line_statement_prefix is not None:
|
||
|
rules.append(
|
||
|
(
|
||
|
len(environment.line_statement_prefix),
|
||
|
TOKEN_LINESTATEMENT_BEGIN,
|
||
|
r"^[ \t\v]*" + e(environment.line_statement_prefix),
|
||
|
)
|
||
|
)
|
||
|
if environment.line_comment_prefix is not None:
|
||
|
rules.append(
|
||
|
(
|
||
|
len(environment.line_comment_prefix),
|
||
|
TOKEN_LINECOMMENT_BEGIN,
|
||
|
r"(?:^|(?<=\S))[^\S\r\n]*" + e(environment.line_comment_prefix),
|
||
|
)
|
||
|
)
|
||
|
|
||
|
return [x[1:] for x in sorted(rules, reverse=True)]
|
||
|
|
||
|
|
||
|
class Failure:
|
||
|
"""Class that raises a `TemplateSyntaxError` if called.
|
||
|
Used by the `Lexer` to specify known errors.
|
||
|
"""
|
||
|
|
||
|
def __init__(
|
||
|
self, message: str, cls: t.Type[TemplateSyntaxError] = TemplateSyntaxError
|
||
|
) -> None:
|
||
|
self.message = message
|
||
|
self.error_class = cls
|
||
|
|
||
|
def __call__(self, lineno: int, filename: str) -> "te.NoReturn":
|
||
|
raise self.error_class(self.message, lineno, filename)
|
||
|
|
||
|
|
||
|
class Token(t.NamedTuple):
|
||
|
lineno: int
|
||
|
type: str
|
||
|
value: str
|
||
|
|
||
|
def __str__(self) -> str:
|
||
|
return describe_token(self)
|
||
|
|
||
|
def test(self, expr: str) -> bool:
|
||
|
"""Test a token against a token expression. This can either be a
|
||
|
token type or ``'token_type:token_value'``. This can only test
|
||
|
against string values and types.
|
||
|
"""
|
||
|
# here we do a regular string equality check as test_any is usually
|
||
|
# passed an iterable of not interned strings.
|
||
|
if self.type == expr:
|
||
|
return True
|
||
|
|
||
|
if ":" in expr:
|
||
|
return expr.split(":", 1) == [self.type, self.value]
|
||
|
|
||
|
return False
|
||
|
|
||
|
def test_any(self, *iterable: str) -> bool:
|
||
|
"""Test against multiple token expressions."""
|
||
|
return any(self.test(expr) for expr in iterable)
|
||
|
|
||
|
|
||
|
class TokenStreamIterator:
|
||
|
"""The iterator for tokenstreams. Iterate over the stream
|
||
|
until the eof token is reached.
|
||
|
"""
|
||
|
|
||
|
def __init__(self, stream: "TokenStream") -> None:
|
||
|
self.stream = stream
|
||
|
|
||
|
def __iter__(self) -> "TokenStreamIterator":
|
||
|
return self
|
||
|
|
||
|
def __next__(self) -> Token:
|
||
|
token = self.stream.current
|
||
|
|
||
|
if token.type is TOKEN_EOF:
|
||
|
self.stream.close()
|
||
|
raise StopIteration
|
||
|
|
||
|
next(self.stream)
|
||
|
return token
|
||
|
|
||
|
|
||
|
class TokenStream:
|
||
|
"""A token stream is an iterable that yields :class:`Token`\\s. The
|
||
|
parser however does not iterate over it but calls :meth:`next` to go
|
||
|
one token ahead. The current active token is stored as :attr:`current`.
|
||
|
"""
|
||
|
|
||
|
def __init__(
|
||
|
self,
|
||
|
generator: t.Iterable[Token],
|
||
|
name: t.Optional[str],
|
||
|
filename: t.Optional[str],
|
||
|
):
|
||
|
self._iter = iter(generator)
|
||
|
self._pushed: "te.Deque[Token]" = deque()
|
||
|
self.name = name
|
||
|
self.filename = filename
|
||
|
self.closed = False
|
||
|
self.current = Token(1, TOKEN_INITIAL, "")
|
||
|
next(self)
|
||
|
|
||
|
def __iter__(self) -> TokenStreamIterator:
|
||
|
return TokenStreamIterator(self)
|
||
|
|
||
|
def __bool__(self) -> bool:
|
||
|
return bool(self._pushed) or self.current.type is not TOKEN_EOF
|
||
|
|
||
|
@property
|
||
|
def eos(self) -> bool:
|
||
|
"""Are we at the end of the stream?"""
|
||
|
return not self
|
||
|
|
||
|
def push(self, token: Token) -> None:
|
||
|
"""Push a token back to the stream."""
|
||
|
self._pushed.append(token)
|
||
|
|
||
|
def look(self) -> Token:
|
||
|
"""Look at the next token."""
|
||
|
old_token = next(self)
|
||
|
result = self.current
|
||
|
self.push(result)
|
||
|
self.current = old_token
|
||
|
return result
|
||
|
|
||
|
def skip(self, n: int = 1) -> None:
|
||
|
"""Got n tokens ahead."""
|
||
|
for _ in range(n):
|
||
|
next(self)
|
||
|
|
||
|
def next_if(self, expr: str) -> t.Optional[Token]:
|
||
|
"""Perform the token test and return the token if it matched.
|
||
|
Otherwise the return value is `None`.
|
||
|
"""
|
||
|
if self.current.test(expr):
|
||
|
return next(self)
|
||
|
|
||
|
return None
|
||
|
|
||
|
def skip_if(self, expr: str) -> bool:
|
||
|
"""Like :meth:`next_if` but only returns `True` or `False`."""
|
||
|
return self.next_if(expr) is not None
|
||
|
|
||
|
def __next__(self) -> Token:
|
||
|
"""Go one token ahead and return the old one.
|
||
|
|
||
|
Use the built-in :func:`next` instead of calling this directly.
|
||
|
"""
|
||
|
rv = self.current
|
||
|
|
||
|
if self._pushed:
|
||
|
self.current = self._pushed.popleft()
|
||
|
elif self.current.type is not TOKEN_EOF:
|
||
|
try:
|
||
|
self.current = next(self._iter)
|
||
|
except StopIteration:
|
||
|
self.close()
|
||
|
|
||
|
return rv
|
||
|
|
||
|
def close(self) -> None:
|
||
|
"""Close the stream."""
|
||
|
self.current = Token(self.current.lineno, TOKEN_EOF, "")
|
||
|
self._iter = iter(())
|
||
|
self.closed = True
|
||
|
|
||
|
def expect(self, expr: str) -> Token:
|
||
|
"""Expect a given token type and return it. This accepts the same
|
||
|
argument as :meth:`jinja2.lexer.Token.test`.
|
||
|
"""
|
||
|
if not self.current.test(expr):
|
||
|
expr = describe_token_expr(expr)
|
||
|
|
||
|
if self.current.type is TOKEN_EOF:
|
||
|
raise TemplateSyntaxError(
|
||
|
f"unexpected end of template, expected {expr!r}.",
|
||
|
self.current.lineno,
|
||
|
self.name,
|
||
|
self.filename,
|
||
|
)
|
||
|
|
||
|
raise TemplateSyntaxError(
|
||
|
f"expected token {expr!r}, got {describe_token(self.current)!r}",
|
||
|
self.current.lineno,
|
||
|
self.name,
|
||
|
self.filename,
|
||
|
)
|
||
|
|
||
|
return next(self)
|
||
|
|
||
|
|
||
|
def get_lexer(environment: "Environment") -> "Lexer":
|
||
|
"""Return a lexer which is probably cached."""
|
||
|
key = (
|
||
|
environment.block_start_string,
|
||
|
environment.block_end_string,
|
||
|
environment.variable_start_string,
|
||
|
environment.variable_end_string,
|
||
|
environment.comment_start_string,
|
||
|
environment.comment_end_string,
|
||
|
environment.line_statement_prefix,
|
||
|
environment.line_comment_prefix,
|
||
|
environment.trim_blocks,
|
||
|
environment.lstrip_blocks,
|
||
|
environment.newline_sequence,
|
||
|
environment.keep_trailing_newline,
|
||
|
)
|
||
|
lexer = _lexer_cache.get(key)
|
||
|
|
||
|
if lexer is None:
|
||
|
_lexer_cache[key] = lexer = Lexer(environment)
|
||
|
|
||
|
return lexer
|
||
|
|
||
|
|
||
|
class OptionalLStrip(tuple):
|
||
|
"""A special tuple for marking a point in the state that can have
|
||
|
lstrip applied.
|
||
|
"""
|
||
|
|
||
|
__slots__ = ()
|
||
|
|
||
|
# Even though it looks like a no-op, creating instances fails
|
||
|
# without this.
|
||
|
def __new__(cls, *members, **kwargs): # type: ignore
|
||
|
return super().__new__(cls, members)
|
||
|
|
||
|
|
||
|
class _Rule(t.NamedTuple):
|
||
|
pattern: t.Pattern[str]
|
||
|
tokens: t.Union[str, t.Tuple[str, ...], t.Tuple[Failure]]
|
||
|
command: t.Optional[str]
|
||
|
|
||
|
|
||
|
class Lexer:
|
||
|
"""Class that implements a lexer for a given environment. Automatically
|
||
|
created by the environment class, usually you don't have to do that.
|
||
|
|
||
|
Note that the lexer is not automatically bound to an environment.
|
||
|
Multiple environments can share the same lexer.
|
||
|
"""
|
||
|
|
||
|
def __init__(self, environment: "Environment") -> None:
|
||
|
# shortcuts
|
||
|
e = re.escape
|
||
|
|
||
|
def c(x: str) -> t.Pattern[str]:
|
||
|
return re.compile(x, re.M | re.S)
|
||
|
|
||
|
# lexing rules for tags
|
||
|
tag_rules: t.List[_Rule] = [
|
||
|
_Rule(whitespace_re, TOKEN_WHITESPACE, None),
|
||
|
_Rule(float_re, TOKEN_FLOAT, None),
|
||
|
_Rule(integer_re, TOKEN_INTEGER, None),
|
||
|
_Rule(name_re, TOKEN_NAME, None),
|
||
|
_Rule(string_re, TOKEN_STRING, None),
|
||
|
_Rule(operator_re, TOKEN_OPERATOR, None),
|
||
|
]
|
||
|
|
||
|
# assemble the root lexing rule. because "|" is ungreedy
|
||
|
# we have to sort by length so that the lexer continues working
|
||
|
# as expected when we have parsing rules like <% for block and
|
||
|
# <%= for variables. (if someone wants asp like syntax)
|
||
|
# variables are just part of the rules if variable processing
|
||
|
# is required.
|
||
|
root_tag_rules = compile_rules(environment)
|
||
|
|
||
|
block_start_re = e(environment.block_start_string)
|
||
|
block_end_re = e(environment.block_end_string)
|
||
|
comment_end_re = e(environment.comment_end_string)
|
||
|
variable_end_re = e(environment.variable_end_string)
|
||
|
|
||
|
# block suffix if trimming is enabled
|
||
|
block_suffix_re = "\\n?" if environment.trim_blocks else ""
|
||
|
|
||
|
self.lstrip_blocks = environment.lstrip_blocks
|
||
|
|
||
|
self.newline_sequence = environment.newline_sequence
|
||
|
self.keep_trailing_newline = environment.keep_trailing_newline
|
||
|
|
||
|
root_raw_re = (
|
||
|
rf"(?P<raw_begin>{block_start_re}(\-|\+|)\s*raw\s*"
|
||
|
rf"(?:\-{block_end_re}\s*|{block_end_re}))"
|
||
|
)
|
||
|
root_parts_re = "|".join(
|
||
|
[root_raw_re] + [rf"(?P<{n}>{r}(\-|\+|))" for n, r in root_tag_rules]
|
||
|
)
|
||
|
|
||
|
# global lexing rules
|
||
|
self.rules: t.Dict[str, t.List[_Rule]] = {
|
||
|
"root": [
|
||
|
# directives
|
||
|
_Rule(
|
||
|
c(rf"(.*?)(?:{root_parts_re})"),
|
||
|
OptionalLStrip(TOKEN_DATA, "#bygroup"), # type: ignore
|
||
|
"#bygroup",
|
||
|
),
|
||
|
# data
|
||
|
_Rule(c(".+"), TOKEN_DATA, None),
|
||
|
],
|
||
|
# comments
|
||
|
TOKEN_COMMENT_BEGIN: [
|
||
|
_Rule(
|
||
|
c(
|
||
|
rf"(.*?)((?:\+{comment_end_re}|\-{comment_end_re}\s*"
|
||
|
rf"|{comment_end_re}{block_suffix_re}))"
|
||
|
),
|
||
|
(TOKEN_COMMENT, TOKEN_COMMENT_END),
|
||
|
"#pop",
|
||
|
),
|
||
|
_Rule(c(r"(.)"), (Failure("Missing end of comment tag"),), None),
|
||
|
],
|
||
|
# blocks
|
||
|
TOKEN_BLOCK_BEGIN: [
|
||
|
_Rule(
|
||
|
c(
|
||
|
rf"(?:\+{block_end_re}|\-{block_end_re}\s*"
|
||
|
rf"|{block_end_re}{block_suffix_re})"
|
||
|
),
|
||
|
TOKEN_BLOCK_END,
|
||
|
"#pop",
|
||
|
),
|
||
|
]
|
||
|
+ tag_rules,
|
||
|
# variables
|
||
|
TOKEN_VARIABLE_BEGIN: [
|
||
|
_Rule(
|
||
|
c(rf"\-{variable_end_re}\s*|{variable_end_re}"),
|
||
|
TOKEN_VARIABLE_END,
|
||
|
"#pop",
|
||
|
)
|
||
|
]
|
||
|
+ tag_rules,
|
||
|
# raw block
|
||
|
TOKEN_RAW_BEGIN: [
|
||
|
_Rule(
|
||
|
c(
|
||
|
rf"(.*?)((?:{block_start_re}(\-|\+|))\s*endraw\s*"
|
||
|
rf"(?:\+{block_end_re}|\-{block_end_re}\s*"
|
||
|
rf"|{block_end_re}{block_suffix_re}))"
|
||
|
),
|
||
|
OptionalLStrip(TOKEN_DATA, TOKEN_RAW_END), # type: ignore
|
||
|
"#pop",
|
||
|
),
|
||
|
_Rule(c(r"(.)"), (Failure("Missing end of raw directive"),), None),
|
||
|
],
|
||
|
# line statements
|
||
|
TOKEN_LINESTATEMENT_BEGIN: [
|
||
|
_Rule(c(r"\s*(\n|$)"), TOKEN_LINESTATEMENT_END, "#pop")
|
||
|
]
|
||
|
+ tag_rules,
|
||
|
# line comments
|
||
|
TOKEN_LINECOMMENT_BEGIN: [
|
||
|
_Rule(
|
||
|
c(r"(.*?)()(?=\n|$)"),
|
||
|
(TOKEN_LINECOMMENT, TOKEN_LINECOMMENT_END),
|
||
|
"#pop",
|
||
|
)
|
||
|
],
|
||
|
}
|
||
|
|
||
|
def _normalize_newlines(self, value: str) -> str:
|
||
|
"""Replace all newlines with the configured sequence in strings
|
||
|
and template data.
|
||
|
"""
|
||
|
return newline_re.sub(self.newline_sequence, value)
|
||
|
|
||
|
def tokenize(
|
||
|
self,
|
||
|
source: str,
|
||
|
name: t.Optional[str] = None,
|
||
|
filename: t.Optional[str] = None,
|
||
|
state: t.Optional[str] = None,
|
||
|
) -> TokenStream:
|
||
|
"""Calls tokeniter + tokenize and wraps it in a token stream."""
|
||
|
stream = self.tokeniter(source, name, filename, state)
|
||
|
return TokenStream(self.wrap(stream, name, filename), name, filename)
|
||
|
|
||
|
def wrap(
|
||
|
self,
|
||
|
stream: t.Iterable[t.Tuple[int, str, str]],
|
||
|
name: t.Optional[str] = None,
|
||
|
filename: t.Optional[str] = None,
|
||
|
) -> t.Iterator[Token]:
|
||
|
"""This is called with the stream as returned by `tokenize` and wraps
|
||
|
every token in a :class:`Token` and converts the value.
|
||
|
"""
|
||
|
for lineno, token, value_str in stream:
|
||
|
if token in ignored_tokens:
|
||
|
continue
|
||
|
|
||
|
value: t.Any = value_str
|
||
|
|
||
|
if token == TOKEN_LINESTATEMENT_BEGIN:
|
||
|
token = TOKEN_BLOCK_BEGIN
|
||
|
elif token == TOKEN_LINESTATEMENT_END:
|
||
|
token = TOKEN_BLOCK_END
|
||
|
# we are not interested in those tokens in the parser
|
||
|
elif token in (TOKEN_RAW_BEGIN, TOKEN_RAW_END):
|
||
|
continue
|
||
|
elif token == TOKEN_DATA:
|
||
|
value = self._normalize_newlines(value_str)
|
||
|
elif token == "keyword":
|
||
|
token = value_str
|
||
|
elif token == TOKEN_NAME:
|
||
|
value = value_str
|
||
|
|
||
|
if not value.isidentifier():
|
||
|
raise TemplateSyntaxError(
|
||
|
"Invalid character in identifier", lineno, name, filename
|
||
|
)
|
||
|
elif token == TOKEN_STRING:
|
||
|
# try to unescape string
|
||
|
try:
|
||
|
value = (
|
||
|
self._normalize_newlines(value_str[1:-1])
|
||
|
.encode("ascii", "backslashreplace")
|
||
|
.decode("unicode-escape")
|
||
|
)
|
||
|
except Exception as e:
|
||
|
msg = str(e).split(":")[-1].strip()
|
||
|
raise TemplateSyntaxError(msg, lineno, name, filename) from e
|
||
|
elif token == TOKEN_INTEGER:
|
||
|
value = int(value_str.replace("_", ""), 0)
|
||
|
elif token == TOKEN_FLOAT:
|
||
|
# remove all "_" first to support more Python versions
|
||
|
value = literal_eval(value_str.replace("_", ""))
|
||
|
elif token == TOKEN_OPERATOR:
|
||
|
token = operators[value_str]
|
||
|
|
||
|
yield Token(lineno, token, value)
|
||
|
|
||
|
def tokeniter(
|
||
|
self,
|
||
|
source: str,
|
||
|
name: t.Optional[str],
|
||
|
filename: t.Optional[str] = None,
|
||
|
state: t.Optional[str] = None,
|
||
|
) -> t.Iterator[t.Tuple[int, str, str]]:
|
||
|
"""This method tokenizes the text and returns the tokens in a
|
||
|
generator. Use this method if you just want to tokenize a template.
|
||
|
|
||
|
.. versionchanged:: 3.0
|
||
|
Only ``\\n``, ``\\r\\n`` and ``\\r`` are treated as line
|
||
|
breaks.
|
||
|
"""
|
||
|
lines = newline_re.split(source)[::2]
|
||
|
|
||
|
if not self.keep_trailing_newline and lines[-1] == "":
|
||
|
del lines[-1]
|
||
|
|
||
|
source = "\n".join(lines)
|
||
|
pos = 0
|
||
|
lineno = 1
|
||
|
stack = ["root"]
|
||
|
|
||
|
if state is not None and state != "root":
|
||
|
assert state in ("variable", "block"), "invalid state"
|
||
|
stack.append(state + "_begin")
|
||
|
|
||
|
statetokens = self.rules[stack[-1]]
|
||
|
source_length = len(source)
|
||
|
balancing_stack: t.List[str] = []
|
||
|
newlines_stripped = 0
|
||
|
line_starting = True
|
||
|
|
||
|
while True:
|
||
|
# tokenizer loop
|
||
|
for regex, tokens, new_state in statetokens:
|
||
|
m = regex.match(source, pos)
|
||
|
|
||
|
# if no match we try again with the next rule
|
||
|
if m is None:
|
||
|
continue
|
||
|
|
||
|
# we only match blocks and variables if braces / parentheses
|
||
|
# are balanced. continue parsing with the lower rule which
|
||
|
# is the operator rule. do this only if the end tags look
|
||
|
# like operators
|
||
|
if balancing_stack and tokens in (
|
||
|
TOKEN_VARIABLE_END,
|
||
|
TOKEN_BLOCK_END,
|
||
|
TOKEN_LINESTATEMENT_END,
|
||
|
):
|
||
|
continue
|
||
|
|
||
|
# tuples support more options
|
||
|
if isinstance(tokens, tuple):
|
||
|
groups: t.Sequence[str] = m.groups()
|
||
|
|
||
|
if isinstance(tokens, OptionalLStrip):
|
||
|
# Rule supports lstrip. Match will look like
|
||
|
# text, block type, whitespace control, type, control, ...
|
||
|
text = groups[0]
|
||
|
# Skipping the text and first type, every other group is the
|
||
|
# whitespace control for each type. One of the groups will be
|
||
|
# -, +, or empty string instead of None.
|
||
|
strip_sign = next(g for g in groups[2::2] if g is not None)
|
||
|
|
||
|
if strip_sign == "-":
|
||
|
# Strip all whitespace between the text and the tag.
|
||
|
stripped = text.rstrip()
|
||
|
newlines_stripped = text[len(stripped) :].count("\n")
|
||
|
groups = [stripped, *groups[1:]]
|
||
|
elif (
|
||
|
# Not marked for preserving whitespace.
|
||
|
strip_sign != "+"
|
||
|
# lstrip is enabled.
|
||
|
and self.lstrip_blocks
|
||
|
# Not a variable expression.
|
||
|
and not m.groupdict().get(TOKEN_VARIABLE_BEGIN)
|
||
|
):
|
||
|
# The start of text between the last newline and the tag.
|
||
|
l_pos = text.rfind("\n") + 1
|
||
|
|
||
|
if l_pos > 0 or line_starting:
|
||
|
# If there's only whitespace between the newline and the
|
||
|
# tag, strip it.
|
||
|
if whitespace_re.fullmatch(text, l_pos):
|
||
|
groups = [text[:l_pos], *groups[1:]]
|
||
|
|
||
|
for idx, token in enumerate(tokens):
|
||
|
# failure group
|
||
|
if token.__class__ is Failure:
|
||
|
raise token(lineno, filename)
|
||
|
# bygroup is a bit more complex, in that case we
|
||
|
# yield for the current token the first named
|
||
|
# group that matched
|
||
|
elif token == "#bygroup":
|
||
|
for key, value in m.groupdict().items():
|
||
|
if value is not None:
|
||
|
yield lineno, key, value
|
||
|
lineno += value.count("\n")
|
||
|
break
|
||
|
else:
|
||
|
raise RuntimeError(
|
||
|
f"{regex!r} wanted to resolve the token dynamically"
|
||
|
" but no group matched"
|
||
|
)
|
||
|
# normal group
|
||
|
else:
|
||
|
data = groups[idx]
|
||
|
|
||
|
if data or token not in ignore_if_empty:
|
||
|
yield lineno, token, data
|
||
|
|
||
|
lineno += data.count("\n") + newlines_stripped
|
||
|
newlines_stripped = 0
|
||
|
|
||
|
# strings as token just are yielded as it.
|
||
|
else:
|
||
|
data = m.group()
|
||
|
|
||
|
# update brace/parentheses balance
|
||
|
if tokens == TOKEN_OPERATOR:
|
||
|
if data == "{":
|
||
|
balancing_stack.append("}")
|
||
|
elif data == "(":
|
||
|
balancing_stack.append(")")
|
||
|
elif data == "[":
|
||
|
balancing_stack.append("]")
|
||
|
elif data in ("}", ")", "]"):
|
||
|
if not balancing_stack:
|
||
|
raise TemplateSyntaxError(
|
||
|
f"unexpected '{data}'", lineno, name, filename
|
||
|
)
|
||
|
|
||
|
expected_op = balancing_stack.pop()
|
||
|
|
||
|
if expected_op != data:
|
||
|
raise TemplateSyntaxError(
|
||
|
f"unexpected '{data}', expected '{expected_op}'",
|
||
|
lineno,
|
||
|
name,
|
||
|
filename,
|
||
|
)
|
||
|
|
||
|
# yield items
|
||
|
if data or tokens not in ignore_if_empty:
|
||
|
yield lineno, tokens, data
|
||
|
|
||
|
lineno += data.count("\n")
|
||
|
|
||
|
line_starting = m.group()[-1:] == "\n"
|
||
|
# fetch new position into new variable so that we can check
|
||
|
# if there is a internal parsing error which would result
|
||
|
# in an infinite loop
|
||
|
pos2 = m.end()
|
||
|
|
||
|
# handle state changes
|
||
|
if new_state is not None:
|
||
|
# remove the uppermost state
|
||
|
if new_state == "#pop":
|
||
|
stack.pop()
|
||
|
# resolve the new state by group checking
|
||
|
elif new_state == "#bygroup":
|
||
|
for key, value in m.groupdict().items():
|
||
|
if value is not None:
|
||
|
stack.append(key)
|
||
|
break
|
||
|
else:
|
||
|
raise RuntimeError(
|
||
|
f"{regex!r} wanted to resolve the new state dynamically"
|
||
|
f" but no group matched"
|
||
|
)
|
||
|
# direct state name given
|
||
|
else:
|
||
|
stack.append(new_state)
|
||
|
|
||
|
statetokens = self.rules[stack[-1]]
|
||
|
# we are still at the same position and no stack change.
|
||
|
# this means a loop without break condition, avoid that and
|
||
|
# raise error
|
||
|
elif pos2 == pos:
|
||
|
raise RuntimeError(
|
||
|
f"{regex!r} yielded empty string without stack change"
|
||
|
)
|
||
|
|
||
|
# publish new function and start again
|
||
|
pos = pos2
|
||
|
break
|
||
|
# if loop terminated without break we haven't found a single match
|
||
|
# either we are at the end of the file or we have a problem
|
||
|
else:
|
||
|
# end of text
|
||
|
if pos >= source_length:
|
||
|
return
|
||
|
|
||
|
# something went wrong
|
||
|
raise TemplateSyntaxError(
|
||
|
f"unexpected char {source[pos]!r} at {pos}", lineno, name, filename
|
||
|
)
|