Consolidated some text utils into the utils_tests test package.
This commit is contained in:
parent
0dab07e5da
commit
3c6ac0bab8
|
@ -1,105 +0,0 @@
|
||||||
# -*- coding: utf-8 -*-
|
|
||||||
from __future__ import unicode_literals
|
|
||||||
|
|
||||||
from django.test import TestCase
|
|
||||||
from django.utils.encoding import iri_to_uri, force_text
|
|
||||||
from django.utils.functional import lazy
|
|
||||||
from django.utils.http import (cookie_date, http_date,
|
|
||||||
urlquote, urlquote_plus, urlunquote, urlunquote_plus)
|
|
||||||
from django.utils import six
|
|
||||||
from django.utils.text import get_text_list, smart_split
|
|
||||||
from django.utils.translation import override
|
|
||||||
|
|
||||||
lazystr = lazy(force_text, six.text_type)
|
|
||||||
|
|
||||||
|
|
||||||
class TextTests(TestCase):
|
|
||||||
"""
|
|
||||||
Tests for stuff in django.utils.text and other text munging util functions.
|
|
||||||
"""
|
|
||||||
|
|
||||||
def test_get_text_list(self):
|
|
||||||
self.assertEqual(get_text_list(['a', 'b', 'c', 'd']), 'a, b, c or d')
|
|
||||||
self.assertEqual(get_text_list(['a', 'b', 'c'], 'and'), 'a, b and c')
|
|
||||||
self.assertEqual(get_text_list(['a', 'b'], 'and'), 'a and b')
|
|
||||||
self.assertEqual(get_text_list(['a']), 'a')
|
|
||||||
self.assertEqual(get_text_list([]), '')
|
|
||||||
with override('ar'):
|
|
||||||
self.assertEqual(get_text_list(['a', 'b', 'c']), "a، b أو c")
|
|
||||||
|
|
||||||
def test_smart_split(self):
|
|
||||||
|
|
||||||
testdata = [
|
|
||||||
('This is "a person" test.',
|
|
||||||
['This', 'is', '"a person"', 'test.']),
|
|
||||||
('This is "a person\'s" test.',
|
|
||||||
['This', 'is', '"a person\'s"', 'test.']),
|
|
||||||
('This is "a person\\"s" test.',
|
|
||||||
['This', 'is', '"a person\\"s"', 'test.']),
|
|
||||||
('"a \'one',
|
|
||||||
['"a', "'one"]),
|
|
||||||
('all friends\' tests',
|
|
||||||
['all', 'friends\'', 'tests']),
|
|
||||||
('url search_page words="something else"',
|
|
||||||
['url', 'search_page', 'words="something else"']),
|
|
||||||
("url search_page words='something else'",
|
|
||||||
['url', 'search_page', "words='something else'"]),
|
|
||||||
('url search_page words "something else"',
|
|
||||||
['url', 'search_page', 'words', '"something else"']),
|
|
||||||
('url search_page words-"something else"',
|
|
||||||
['url', 'search_page', 'words-"something else"']),
|
|
||||||
('url search_page words=hello',
|
|
||||||
['url', 'search_page', 'words=hello']),
|
|
||||||
('url search_page words="something else',
|
|
||||||
['url', 'search_page', 'words="something', 'else']),
|
|
||||||
("cut:','|cut:' '",
|
|
||||||
["cut:','|cut:' '"]),
|
|
||||||
(lazystr("a b c d"), # Test for #20231
|
|
||||||
['a', 'b', 'c', 'd']),
|
|
||||||
]
|
|
||||||
for test, expected in testdata:
|
|
||||||
self.assertEqual(list(smart_split(test)), expected)
|
|
||||||
|
|
||||||
def test_urlquote(self):
|
|
||||||
self.assertEqual(urlquote('Paris & Orl\xe9ans'),
|
|
||||||
'Paris%20%26%20Orl%C3%A9ans')
|
|
||||||
self.assertEqual(urlquote('Paris & Orl\xe9ans', safe="&"),
|
|
||||||
'Paris%20&%20Orl%C3%A9ans')
|
|
||||||
self.assertEqual(
|
|
||||||
urlunquote('Paris%20%26%20Orl%C3%A9ans'),
|
|
||||||
'Paris & Orl\xe9ans')
|
|
||||||
self.assertEqual(
|
|
||||||
urlunquote('Paris%20&%20Orl%C3%A9ans'),
|
|
||||||
'Paris & Orl\xe9ans')
|
|
||||||
self.assertEqual(urlquote_plus('Paris & Orl\xe9ans'),
|
|
||||||
'Paris+%26+Orl%C3%A9ans')
|
|
||||||
self.assertEqual(urlquote_plus('Paris & Orl\xe9ans', safe="&"),
|
|
||||||
'Paris+&+Orl%C3%A9ans')
|
|
||||||
self.assertEqual(
|
|
||||||
urlunquote_plus('Paris+%26+Orl%C3%A9ans'),
|
|
||||||
'Paris & Orl\xe9ans')
|
|
||||||
self.assertEqual(
|
|
||||||
urlunquote_plus('Paris+&+Orl%C3%A9ans'),
|
|
||||||
'Paris & Orl\xe9ans')
|
|
||||||
|
|
||||||
def test_cookie_date(self):
|
|
||||||
t = 1167616461.0
|
|
||||||
self.assertEqual(cookie_date(t), 'Mon, 01-Jan-2007 01:54:21 GMT')
|
|
||||||
|
|
||||||
def test_http_date(self):
|
|
||||||
t = 1167616461.0
|
|
||||||
self.assertEqual(http_date(t), 'Mon, 01 Jan 2007 01:54:21 GMT')
|
|
||||||
|
|
||||||
def test_iri_to_uri(self):
|
|
||||||
self.assertEqual(iri_to_uri('red%09ros\xe9#red'),
|
|
||||||
'red%09ros%C3%A9#red')
|
|
||||||
|
|
||||||
self.assertEqual(iri_to_uri('/blog/for/J\xfcrgen M\xfcnster/'),
|
|
||||||
'/blog/for/J%C3%BCrgen%20M%C3%BCnster/')
|
|
||||||
|
|
||||||
self.assertEqual(iri_to_uri('locations/%s' % urlquote_plus('Paris & Orl\xe9ans')),
|
|
||||||
'locations/Paris+%26+Orl%C3%A9ans')
|
|
||||||
|
|
||||||
def test_iri_to_uri_idempotent(self):
|
|
||||||
self.assertEqual(iri_to_uri(iri_to_uri('red%09ros\xe9#red')),
|
|
||||||
'red%09ros%C3%A9#red')
|
|
|
@ -5,8 +5,9 @@ import unittest
|
||||||
import datetime
|
import datetime
|
||||||
|
|
||||||
from django.utils import six
|
from django.utils import six
|
||||||
from django.utils.encoding import (force_bytes, force_text, filepath_to_uri,
|
from django.utils.encoding import (filepath_to_uri, force_bytes,
|
||||||
python_2_unicode_compatible)
|
force_text, iri_to_uri, python_2_unicode_compatible)
|
||||||
|
from django.utils.http import urlquote_plus
|
||||||
|
|
||||||
|
|
||||||
class TestEncodingUtils(unittest.TestCase):
|
class TestEncodingUtils(unittest.TestCase):
|
||||||
|
@ -45,6 +46,20 @@ class TestEncodingUtils(unittest.TestCase):
|
||||||
self.assertEqual(filepath_to_uri('upload\\чубака.mp4'.encode('utf-8')),
|
self.assertEqual(filepath_to_uri('upload\\чубака.mp4'.encode('utf-8')),
|
||||||
'upload/%D1%87%D1%83%D0%B1%D0%B0%D0%BA%D0%B0.mp4')
|
'upload/%D1%87%D1%83%D0%B1%D0%B0%D0%BA%D0%B0.mp4')
|
||||||
|
|
||||||
|
def test_iri_to_uri(self):
|
||||||
|
self.assertEqual(iri_to_uri('red%09ros\xe9#red'),
|
||||||
|
'red%09ros%C3%A9#red')
|
||||||
|
|
||||||
|
self.assertEqual(iri_to_uri('/blog/for/J\xfcrgen M\xfcnster/'),
|
||||||
|
'/blog/for/J%C3%BCrgen%20M%C3%BCnster/')
|
||||||
|
|
||||||
|
self.assertEqual(iri_to_uri('locations/%s' % urlquote_plus('Paris & Orl\xe9ans')),
|
||||||
|
'locations/Paris+%26+Orl%C3%A9ans')
|
||||||
|
|
||||||
|
def test_iri_to_uri_idempotent(self):
|
||||||
|
self.assertEqual(iri_to_uri(iri_to_uri('red%09ros\xe9#red')),
|
||||||
|
'red%09ros%C3%A9#red')
|
||||||
|
|
||||||
@unittest.skipIf(six.PY3, "tests a class not defining __str__ under Python 2")
|
@unittest.skipIf(six.PY3, "tests a class not defining __str__ under Python 2")
|
||||||
def test_decorated_class_without_str(self):
|
def test_decorated_class_without_str(self):
|
||||||
with self.assertRaises(ValueError):
|
with self.assertRaises(ValueError):
|
||||||
|
|
|
@ -1,3 +1,5 @@
|
||||||
|
from __future__ import unicode_literals
|
||||||
|
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
import sys
|
import sys
|
||||||
import unittest
|
import unittest
|
||||||
|
@ -125,6 +127,28 @@ class TestUtilsHttp(unittest.TestCase):
|
||||||
decoded = http.urlsafe_base64_decode(encoded)
|
decoded = http.urlsafe_base64_decode(encoded)
|
||||||
self.assertEqual(bytestring, decoded)
|
self.assertEqual(bytestring, decoded)
|
||||||
|
|
||||||
|
def test_urlquote(self):
|
||||||
|
self.assertEqual(http.urlquote('Paris & Orl\xe9ans'),
|
||||||
|
'Paris%20%26%20Orl%C3%A9ans')
|
||||||
|
self.assertEqual(http.urlquote('Paris & Orl\xe9ans', safe="&"),
|
||||||
|
'Paris%20&%20Orl%C3%A9ans')
|
||||||
|
self.assertEqual(
|
||||||
|
http.urlunquote('Paris%20%26%20Orl%C3%A9ans'),
|
||||||
|
'Paris & Orl\xe9ans')
|
||||||
|
self.assertEqual(
|
||||||
|
http.urlunquote('Paris%20&%20Orl%C3%A9ans'),
|
||||||
|
'Paris & Orl\xe9ans')
|
||||||
|
self.assertEqual(http.urlquote_plus('Paris & Orl\xe9ans'),
|
||||||
|
'Paris+%26+Orl%C3%A9ans')
|
||||||
|
self.assertEqual(http.urlquote_plus('Paris & Orl\xe9ans', safe="&"),
|
||||||
|
'Paris+&+Orl%C3%A9ans')
|
||||||
|
self.assertEqual(
|
||||||
|
http.urlunquote_plus('Paris+%26+Orl%C3%A9ans'),
|
||||||
|
'Paris & Orl\xe9ans')
|
||||||
|
self.assertEqual(
|
||||||
|
http.urlunquote_plus('Paris+&+Orl%C3%A9ans'),
|
||||||
|
'Paris & Orl\xe9ans')
|
||||||
|
|
||||||
|
|
||||||
class ETagProcessingTests(unittest.TestCase):
|
class ETagProcessingTests(unittest.TestCase):
|
||||||
def test_parsing(self):
|
def test_parsing(self):
|
||||||
|
@ -137,6 +161,14 @@ class ETagProcessingTests(unittest.TestCase):
|
||||||
|
|
||||||
|
|
||||||
class HttpDateProcessingTests(unittest.TestCase):
|
class HttpDateProcessingTests(unittest.TestCase):
|
||||||
|
def test_http_date(self):
|
||||||
|
t = 1167616461.0
|
||||||
|
self.assertEqual(http.http_date(t), 'Mon, 01 Jan 2007 01:54:21 GMT')
|
||||||
|
|
||||||
|
def test_cookie_date(self):
|
||||||
|
t = 1167616461.0
|
||||||
|
self.assertEqual(http.cookie_date(t), 'Mon, 01-Jan-2007 01:54:21 GMT')
|
||||||
|
|
||||||
def test_parsing_rfc1123(self):
|
def test_parsing_rfc1123(self):
|
||||||
parsed = http.parse_http_date('Sun, 06 Nov 1994 08:49:37 GMT')
|
parsed = http.parse_http_date('Sun, 06 Nov 1994 08:49:37 GMT')
|
||||||
self.assertEqual(datetime.utcfromtimestamp(parsed),
|
self.assertEqual(datetime.utcfromtimestamp(parsed),
|
||||||
|
|
|
@ -7,12 +7,58 @@ import warnings
|
||||||
from django.test import SimpleTestCase
|
from django.test import SimpleTestCase
|
||||||
from django.utils import six, text
|
from django.utils import six, text
|
||||||
from django.utils.deprecation import RemovedInDjango19Warning
|
from django.utils.deprecation import RemovedInDjango19Warning
|
||||||
|
from django.utils.encoding import force_text
|
||||||
|
from django.utils.functional import lazy
|
||||||
|
from django.utils.translation import override
|
||||||
|
|
||||||
|
lazystr = lazy(force_text, six.text_type)
|
||||||
|
|
||||||
IS_WIDE_BUILD = (len('\U0001F4A9') == 1)
|
IS_WIDE_BUILD = (len('\U0001F4A9') == 1)
|
||||||
|
|
||||||
|
|
||||||
class TestUtilsText(SimpleTestCase):
|
class TestUtilsText(SimpleTestCase):
|
||||||
|
|
||||||
|
def test_get_text_list(self):
|
||||||
|
self.assertEqual(text.get_text_list(['a', 'b', 'c', 'd']), 'a, b, c or d')
|
||||||
|
self.assertEqual(text.get_text_list(['a', 'b', 'c'], 'and'), 'a, b and c')
|
||||||
|
self.assertEqual(text.get_text_list(['a', 'b'], 'and'), 'a and b')
|
||||||
|
self.assertEqual(text.get_text_list(['a']), 'a')
|
||||||
|
self.assertEqual(text.get_text_list([]), '')
|
||||||
|
with override('ar'):
|
||||||
|
self.assertEqual(text.get_text_list(['a', 'b', 'c']), "a، b أو c")
|
||||||
|
|
||||||
|
def test_smart_split(self):
|
||||||
|
testdata = [
|
||||||
|
('This is "a person" test.',
|
||||||
|
['This', 'is', '"a person"', 'test.']),
|
||||||
|
('This is "a person\'s" test.',
|
||||||
|
['This', 'is', '"a person\'s"', 'test.']),
|
||||||
|
('This is "a person\\"s" test.',
|
||||||
|
['This', 'is', '"a person\\"s"', 'test.']),
|
||||||
|
('"a \'one',
|
||||||
|
['"a', "'one"]),
|
||||||
|
('all friends\' tests',
|
||||||
|
['all', 'friends\'', 'tests']),
|
||||||
|
('url search_page words="something else"',
|
||||||
|
['url', 'search_page', 'words="something else"']),
|
||||||
|
("url search_page words='something else'",
|
||||||
|
['url', 'search_page', "words='something else'"]),
|
||||||
|
('url search_page words "something else"',
|
||||||
|
['url', 'search_page', 'words', '"something else"']),
|
||||||
|
('url search_page words-"something else"',
|
||||||
|
['url', 'search_page', 'words-"something else"']),
|
||||||
|
('url search_page words=hello',
|
||||||
|
['url', 'search_page', 'words=hello']),
|
||||||
|
('url search_page words="something else',
|
||||||
|
['url', 'search_page', 'words="something', 'else']),
|
||||||
|
("cut:','|cut:' '",
|
||||||
|
["cut:','|cut:' '"]),
|
||||||
|
(lazystr("a b c d"), # Test for #20231
|
||||||
|
['a', 'b', 'c', 'd']),
|
||||||
|
]
|
||||||
|
for test, expected in testdata:
|
||||||
|
self.assertEqual(list(text.smart_split(test)), expected)
|
||||||
|
|
||||||
def test_truncate_chars(self):
|
def test_truncate_chars(self):
|
||||||
truncator = text.Truncator(
|
truncator = text.Truncator(
|
||||||
'The quick brown fox jumped over the lazy dog.'
|
'The quick brown fox jumped over the lazy dog.'
|
||||||
|
|
Loading…
Reference in New Issue