mirror of https://github.com/django/django.git
Refs #32990 -- Added tests for DebugLexer/Lexer.tokenize().
This commit is contained in:
parent
259b28706e
commit
c99aaf14ee
|
@ -1,8 +1,48 @@
|
||||||
from django.template import Context, Template, Variable, VariableDoesNotExist
|
from django.template import Context, Template, Variable, VariableDoesNotExist
|
||||||
|
from django.template.base import DebugLexer, Lexer, TokenType
|
||||||
from django.test import SimpleTestCase
|
from django.test import SimpleTestCase
|
||||||
from django.utils.translation import gettext_lazy
|
from django.utils.translation import gettext_lazy
|
||||||
|
|
||||||
|
|
||||||
|
class LexerTestMixin:
|
||||||
|
template_string = (
|
||||||
|
'text\n'
|
||||||
|
'{% if test %}{{ varvalue }}{% endif %}'
|
||||||
|
'{#comment {{not a var}} %{not a block}% #}'
|
||||||
|
)
|
||||||
|
expected_token_tuples = [
|
||||||
|
# (token_type, contents, lineno, position)
|
||||||
|
(TokenType.TEXT, 'text\n', 1, (0, 5)),
|
||||||
|
(TokenType.BLOCK, 'if test', 2, (5, 18)),
|
||||||
|
(TokenType.VAR, 'varvalue', 2, (18, 32)),
|
||||||
|
(TokenType.BLOCK, 'endif', 2, (32, 43)),
|
||||||
|
(TokenType.COMMENT, 'comment {{not a var}} %{not a block}%', 2, (43, 85)),
|
||||||
|
]
|
||||||
|
|
||||||
|
def test_tokenize(self):
|
||||||
|
tokens = self.lexer_class(self.template_string).tokenize()
|
||||||
|
token_tuples = [(t.token_type, t.contents, t.lineno, t.position) for t in tokens]
|
||||||
|
self.assertEqual(token_tuples, self.make_expected())
|
||||||
|
|
||||||
|
def make_expected(self):
|
||||||
|
raise NotImplementedError('This method must be implemented by a subclass.')
|
||||||
|
|
||||||
|
|
||||||
|
class LexerTests(LexerTestMixin, SimpleTestCase):
|
||||||
|
lexer_class = Lexer
|
||||||
|
|
||||||
|
def make_expected(self):
|
||||||
|
# The non-debug lexer does not record position.
|
||||||
|
return [t[:-1] + (None,) for t in self.expected_token_tuples]
|
||||||
|
|
||||||
|
|
||||||
|
class DebugLexerTests(LexerTestMixin, SimpleTestCase):
|
||||||
|
lexer_class = DebugLexer
|
||||||
|
|
||||||
|
def make_expected(self):
|
||||||
|
return self.expected_token_tuples
|
||||||
|
|
||||||
|
|
||||||
class TemplateTests(SimpleTestCase):
|
class TemplateTests(SimpleTestCase):
|
||||||
def test_lazy_template_string(self):
|
def test_lazy_template_string(self):
|
||||||
template_string = gettext_lazy('lazy string')
|
template_string = gettext_lazy('lazy string')
|
||||||
|
|
Loading…
Reference in New Issue