832 lines
31 KiB
Python
832 lines
31 KiB
Python
import logging
|
|
import multiprocessing
|
|
import os
|
|
import unittest.loader
|
|
from argparse import ArgumentParser
|
|
from contextlib import contextmanager
|
|
from importlib import import_module
|
|
from unittest import TestSuite, TextTestRunner, defaultTestLoader, mock
|
|
|
|
from django.db import connections
|
|
from django.test import SimpleTestCase
|
|
from django.test.runner import DiscoverRunner, get_max_test_processes
|
|
from django.test.utils import (
|
|
NullTimeKeeper,
|
|
TimeKeeper,
|
|
captured_stderr,
|
|
captured_stdout,
|
|
)
|
|
|
|
|
|
@contextmanager
|
|
def change_cwd(directory):
|
|
current_dir = os.path.abspath(os.path.dirname(__file__))
|
|
new_dir = os.path.join(current_dir, directory)
|
|
old_cwd = os.getcwd()
|
|
os.chdir(new_dir)
|
|
try:
|
|
yield
|
|
finally:
|
|
os.chdir(old_cwd)
|
|
|
|
|
|
@contextmanager
|
|
def change_loader_patterns(patterns):
|
|
original_patterns = DiscoverRunner.test_loader.testNamePatterns
|
|
DiscoverRunner.test_loader.testNamePatterns = patterns
|
|
try:
|
|
yield
|
|
finally:
|
|
DiscoverRunner.test_loader.testNamePatterns = original_patterns
|
|
|
|
|
|
# Isolate from the real environment.
|
|
@mock.patch.dict(os.environ, {}, clear=True)
|
|
@mock.patch.object(multiprocessing, "cpu_count", return_value=12)
|
|
# Python 3.8 on macOS defaults to 'spawn' mode.
|
|
@mock.patch.object(multiprocessing, "get_start_method", return_value="fork")
|
|
class DiscoverRunnerParallelArgumentTests(SimpleTestCase):
|
|
def get_parser(self):
|
|
parser = ArgumentParser()
|
|
DiscoverRunner.add_arguments(parser)
|
|
return parser
|
|
|
|
def test_parallel_default(self, *mocked_objects):
|
|
result = self.get_parser().parse_args([])
|
|
self.assertEqual(result.parallel, 0)
|
|
|
|
def test_parallel_flag(self, *mocked_objects):
|
|
result = self.get_parser().parse_args(["--parallel"])
|
|
self.assertEqual(result.parallel, "auto")
|
|
|
|
def test_parallel_auto(self, *mocked_objects):
|
|
result = self.get_parser().parse_args(["--parallel", "auto"])
|
|
self.assertEqual(result.parallel, "auto")
|
|
|
|
def test_parallel_count(self, *mocked_objects):
|
|
result = self.get_parser().parse_args(["--parallel", "17"])
|
|
self.assertEqual(result.parallel, 17)
|
|
|
|
def test_parallel_invalid(self, *mocked_objects):
|
|
with self.assertRaises(SystemExit), captured_stderr() as stderr:
|
|
self.get_parser().parse_args(["--parallel", "unaccepted"])
|
|
msg = "argument --parallel: 'unaccepted' is not an integer or the string 'auto'"
|
|
self.assertIn(msg, stderr.getvalue())
|
|
|
|
def test_get_max_test_processes(self, *mocked_objects):
|
|
self.assertEqual(get_max_test_processes(), 12)
|
|
|
|
@mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"})
|
|
def test_get_max_test_processes_env_var(self, *mocked_objects):
|
|
self.assertEqual(get_max_test_processes(), 7)
|
|
|
|
def test_get_max_test_processes_spawn(
|
|
self,
|
|
mocked_get_start_method,
|
|
mocked_cpu_count,
|
|
):
|
|
mocked_get_start_method.return_value = "spawn"
|
|
self.assertEqual(get_max_test_processes(), 12)
|
|
with mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"}):
|
|
self.assertEqual(get_max_test_processes(), 7)
|
|
|
|
def test_get_max_test_processes_forkserver(
|
|
self,
|
|
mocked_get_start_method,
|
|
mocked_cpu_count,
|
|
):
|
|
mocked_get_start_method.return_value = "forkserver"
|
|
self.assertEqual(get_max_test_processes(), 1)
|
|
with mock.patch.dict(os.environ, {"DJANGO_TEST_PROCESSES": "7"}):
|
|
self.assertEqual(get_max_test_processes(), 1)
|
|
|
|
|
|
class DiscoverRunnerTests(SimpleTestCase):
|
|
@staticmethod
|
|
def get_test_methods_names(suite):
|
|
return [t.__class__.__name__ + "." + t._testMethodName for t in suite._tests]
|
|
|
|
def test_init_debug_mode(self):
|
|
runner = DiscoverRunner()
|
|
self.assertFalse(runner.debug_mode)
|
|
|
|
def test_add_arguments_shuffle(self):
|
|
parser = ArgumentParser()
|
|
DiscoverRunner.add_arguments(parser)
|
|
ns = parser.parse_args([])
|
|
self.assertIs(ns.shuffle, False)
|
|
ns = parser.parse_args(["--shuffle"])
|
|
self.assertIsNone(ns.shuffle)
|
|
ns = parser.parse_args(["--shuffle", "5"])
|
|
self.assertEqual(ns.shuffle, 5)
|
|
|
|
def test_add_arguments_debug_mode(self):
|
|
parser = ArgumentParser()
|
|
DiscoverRunner.add_arguments(parser)
|
|
|
|
ns = parser.parse_args([])
|
|
self.assertFalse(ns.debug_mode)
|
|
ns = parser.parse_args(["--debug-mode"])
|
|
self.assertTrue(ns.debug_mode)
|
|
|
|
def test_setup_shuffler_no_shuffle_argument(self):
|
|
runner = DiscoverRunner()
|
|
self.assertIs(runner.shuffle, False)
|
|
runner.setup_shuffler()
|
|
self.assertIsNone(runner.shuffle_seed)
|
|
|
|
def test_setup_shuffler_shuffle_none(self):
|
|
runner = DiscoverRunner(shuffle=None)
|
|
self.assertIsNone(runner.shuffle)
|
|
with mock.patch("random.randint", return_value=1):
|
|
with captured_stdout() as stdout:
|
|
runner.setup_shuffler()
|
|
self.assertEqual(stdout.getvalue(), "Using shuffle seed: 1 (generated)\n")
|
|
self.assertEqual(runner.shuffle_seed, 1)
|
|
|
|
def test_setup_shuffler_shuffle_int(self):
|
|
runner = DiscoverRunner(shuffle=2)
|
|
self.assertEqual(runner.shuffle, 2)
|
|
with captured_stdout() as stdout:
|
|
runner.setup_shuffler()
|
|
expected_out = "Using shuffle seed: 2 (given)\n"
|
|
self.assertEqual(stdout.getvalue(), expected_out)
|
|
self.assertEqual(runner.shuffle_seed, 2)
|
|
|
|
def test_load_tests_for_label_file_path(self):
|
|
with change_cwd("."):
|
|
msg = (
|
|
"One of the test labels is a path to a file: "
|
|
"'test_discover_runner.py', which is not supported. Use a "
|
|
"dotted module name or path to a directory instead."
|
|
)
|
|
with self.assertRaisesMessage(RuntimeError, msg):
|
|
DiscoverRunner().load_tests_for_label("test_discover_runner.py", {})
|
|
|
|
def test_dotted_test_module(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests_sample"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 4)
|
|
|
|
def test_dotted_test_class_vanilla_unittest(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests_sample.TestVanillaUnittest"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 1)
|
|
|
|
def test_dotted_test_class_django_testcase(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests_sample.TestDjangoTestCase"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 1)
|
|
|
|
def test_dotted_test_method_django_testcase(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests_sample.TestDjangoTestCase.test_sample"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 1)
|
|
|
|
def test_pattern(self):
|
|
count = (
|
|
DiscoverRunner(
|
|
pattern="*_tests.py",
|
|
verbosity=0,
|
|
)
|
|
.build_suite(["test_runner_apps.sample"])
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 1)
|
|
|
|
def test_name_patterns(self):
|
|
all_test_1 = [
|
|
"DjangoCase1.test_1",
|
|
"DjangoCase2.test_1",
|
|
"SimpleCase1.test_1",
|
|
"SimpleCase2.test_1",
|
|
"UnittestCase1.test_1",
|
|
"UnittestCase2.test_1",
|
|
]
|
|
all_test_2 = [
|
|
"DjangoCase1.test_2",
|
|
"DjangoCase2.test_2",
|
|
"SimpleCase1.test_2",
|
|
"SimpleCase2.test_2",
|
|
"UnittestCase1.test_2",
|
|
"UnittestCase2.test_2",
|
|
]
|
|
all_tests = sorted([*all_test_1, *all_test_2, "UnittestCase2.test_3_test"])
|
|
for pattern, expected in [
|
|
[["test_1"], all_test_1],
|
|
[["UnittestCase1"], ["UnittestCase1.test_1", "UnittestCase1.test_2"]],
|
|
[["*test"], ["UnittestCase2.test_3_test"]],
|
|
[["test*"], all_tests],
|
|
[["test"], all_tests],
|
|
[["test_1", "test_2"], sorted([*all_test_1, *all_test_2])],
|
|
[["test*1"], all_test_1],
|
|
]:
|
|
with self.subTest(pattern):
|
|
suite = DiscoverRunner(
|
|
test_name_patterns=pattern,
|
|
verbosity=0,
|
|
).build_suite(["test_runner_apps.simple"])
|
|
self.assertEqual(expected, self.get_test_methods_names(suite))
|
|
|
|
def test_loader_patterns_not_mutated(self):
|
|
runner = DiscoverRunner(test_name_patterns=["test_sample"], verbosity=0)
|
|
tests = [
|
|
("test_runner_apps.sample.tests", 1),
|
|
("test_runner_apps.sample.tests.Test.test_sample", 1),
|
|
("test_runner_apps.sample.empty", 0),
|
|
("test_runner_apps.sample.tests_sample.EmptyTestCase", 0),
|
|
]
|
|
for test_labels, tests_count in tests:
|
|
with self.subTest(test_labels=test_labels):
|
|
with change_loader_patterns(["UnittestCase1"]):
|
|
count = runner.build_suite([test_labels]).countTestCases()
|
|
self.assertEqual(count, tests_count)
|
|
self.assertEqual(
|
|
runner.test_loader.testNamePatterns, ["UnittestCase1"]
|
|
)
|
|
|
|
def test_loader_patterns_not_mutated_when_test_label_is_file_path(self):
|
|
runner = DiscoverRunner(test_name_patterns=["test_sample"], verbosity=0)
|
|
with change_cwd("."), change_loader_patterns(["UnittestCase1"]):
|
|
with self.assertRaises(RuntimeError):
|
|
runner.build_suite(["test_discover_runner.py"])
|
|
self.assertEqual(runner.test_loader.testNamePatterns, ["UnittestCase1"])
|
|
|
|
def test_file_path(self):
|
|
with change_cwd(".."):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps/sample/"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 5)
|
|
|
|
def test_empty_label(self):
|
|
"""
|
|
If the test label is empty, discovery should happen on the current
|
|
working directory.
|
|
"""
|
|
with change_cwd("."):
|
|
suite = DiscoverRunner(verbosity=0).build_suite([])
|
|
self.assertEqual(
|
|
suite._tests[0].id().split(".")[0],
|
|
os.path.basename(os.getcwd()),
|
|
)
|
|
|
|
def test_empty_test_case(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests_sample.EmptyTestCase"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 0)
|
|
|
|
def test_discovery_on_package(self):
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.tests"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 1)
|
|
|
|
def test_ignore_adjacent(self):
|
|
"""
|
|
When given a dotted path to a module, unittest discovery searches
|
|
not just the module, but also the directory containing the module.
|
|
|
|
This results in tests from adjacent modules being run when they
|
|
should not. The discover runner avoids this behavior.
|
|
"""
|
|
count = (
|
|
DiscoverRunner(verbosity=0)
|
|
.build_suite(
|
|
["test_runner_apps.sample.empty"],
|
|
)
|
|
.countTestCases()
|
|
)
|
|
|
|
self.assertEqual(count, 0)
|
|
|
|
def test_testcase_ordering(self):
|
|
with change_cwd(".."):
|
|
suite = DiscoverRunner(verbosity=0).build_suite(
|
|
["test_runner_apps/sample/"]
|
|
)
|
|
self.assertEqual(
|
|
suite._tests[0].__class__.__name__,
|
|
"TestDjangoTestCase",
|
|
msg="TestDjangoTestCase should be the first test case",
|
|
)
|
|
self.assertEqual(
|
|
suite._tests[1].__class__.__name__,
|
|
"TestZimpleTestCase",
|
|
msg="TestZimpleTestCase should be the second test case",
|
|
)
|
|
# All others can follow in unspecified order, including doctests
|
|
self.assertIn(
|
|
"DocTestCase", [t.__class__.__name__ for t in suite._tests[2:]]
|
|
)
|
|
|
|
def test_duplicates_ignored(self):
|
|
"""
|
|
Tests shouldn't be discovered twice when discovering on overlapping paths.
|
|
"""
|
|
base_app = "forms_tests"
|
|
sub_app = "forms_tests.field_tests"
|
|
runner = DiscoverRunner(verbosity=0)
|
|
with self.modify_settings(INSTALLED_APPS={"append": sub_app}):
|
|
single = runner.build_suite([base_app]).countTestCases()
|
|
dups = runner.build_suite([base_app, sub_app]).countTestCases()
|
|
self.assertEqual(single, dups)
|
|
|
|
def test_reverse(self):
|
|
"""
|
|
Reverse should reorder tests while maintaining the grouping specified
|
|
by ``DiscoverRunner.reorder_by``.
|
|
"""
|
|
runner = DiscoverRunner(reverse=True, verbosity=0)
|
|
suite = runner.build_suite(
|
|
test_labels=("test_runner_apps.sample", "test_runner_apps.simple")
|
|
)
|
|
self.assertIn(
|
|
"test_runner_apps.simple",
|
|
next(iter(suite)).id(),
|
|
msg="Test labels should be reversed.",
|
|
)
|
|
suite = runner.build_suite(test_labels=("test_runner_apps.simple",))
|
|
suite = tuple(suite)
|
|
self.assertIn(
|
|
"DjangoCase", suite[0].id(), msg="Test groups should not be reversed."
|
|
)
|
|
self.assertIn(
|
|
"SimpleCase", suite[4].id(), msg="Test groups order should be preserved."
|
|
)
|
|
self.assertIn(
|
|
"DjangoCase2", suite[0].id(), msg="Django test cases should be reversed."
|
|
)
|
|
self.assertIn(
|
|
"SimpleCase2", suite[4].id(), msg="Simple test cases should be reversed."
|
|
)
|
|
self.assertIn(
|
|
"UnittestCase2",
|
|
suite[8].id(),
|
|
msg="Unittest test cases should be reversed.",
|
|
)
|
|
self.assertIn(
|
|
"test_2", suite[0].id(), msg="Methods of Django cases should be reversed."
|
|
)
|
|
self.assertIn(
|
|
"test_2", suite[4].id(), msg="Methods of simple cases should be reversed."
|
|
)
|
|
self.assertIn(
|
|
"test_2", suite[9].id(), msg="Methods of unittest cases should be reversed."
|
|
)
|
|
|
|
def test_build_suite_failed_tests_first(self):
|
|
# The "doesnotexist" label results in a _FailedTest instance.
|
|
suite = DiscoverRunner(verbosity=0).build_suite(
|
|
test_labels=["test_runner_apps.sample", "doesnotexist"],
|
|
)
|
|
tests = list(suite)
|
|
self.assertIsInstance(tests[0], unittest.loader._FailedTest)
|
|
self.assertNotIsInstance(tests[-1], unittest.loader._FailedTest)
|
|
|
|
def test_build_suite_shuffling(self):
|
|
# These will result in unittest.loader._FailedTest instances rather
|
|
# than TestCase objects, but they are sufficient for testing.
|
|
labels = ["label1", "label2", "label3", "label4"]
|
|
cases = [
|
|
({}, ["label1", "label2", "label3", "label4"]),
|
|
({"reverse": True}, ["label4", "label3", "label2", "label1"]),
|
|
({"shuffle": 8}, ["label4", "label1", "label3", "label2"]),
|
|
({"shuffle": 8, "reverse": True}, ["label2", "label3", "label1", "label4"]),
|
|
]
|
|
for kwargs, expected in cases:
|
|
with self.subTest(kwargs=kwargs):
|
|
# Prevent writing the seed to stdout.
|
|
runner = DiscoverRunner(**kwargs, verbosity=0)
|
|
tests = runner.build_suite(test_labels=labels)
|
|
# The ids have the form "unittest.loader._FailedTest.label1".
|
|
names = [test.id().split(".")[-1] for test in tests]
|
|
self.assertEqual(names, expected)
|
|
|
|
def test_overridable_get_test_runner_kwargs(self):
|
|
self.assertIsInstance(DiscoverRunner().get_test_runner_kwargs(), dict)
|
|
|
|
def test_overridable_test_suite(self):
|
|
self.assertEqual(DiscoverRunner().test_suite, TestSuite)
|
|
|
|
def test_overridable_test_runner(self):
|
|
self.assertEqual(DiscoverRunner().test_runner, TextTestRunner)
|
|
|
|
def test_overridable_test_loader(self):
|
|
self.assertEqual(DiscoverRunner().test_loader, defaultTestLoader)
|
|
|
|
def test_tags(self):
|
|
runner = DiscoverRunner(tags=["core"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 1
|
|
)
|
|
runner = DiscoverRunner(tags=["fast"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 2
|
|
)
|
|
runner = DiscoverRunner(tags=["slow"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 2
|
|
)
|
|
|
|
def test_exclude_tags(self):
|
|
runner = DiscoverRunner(tags=["fast"], exclude_tags=["core"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 1
|
|
)
|
|
runner = DiscoverRunner(tags=["fast"], exclude_tags=["slow"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 0
|
|
)
|
|
runner = DiscoverRunner(exclude_tags=["slow"], verbosity=0)
|
|
self.assertEqual(
|
|
runner.build_suite(["test_runner_apps.tagged.tests"]).countTestCases(), 0
|
|
)
|
|
|
|
def test_tag_inheritance(self):
|
|
def count_tests(**kwargs):
|
|
kwargs.setdefault("verbosity", 0)
|
|
suite = DiscoverRunner(**kwargs).build_suite(
|
|
["test_runner_apps.tagged.tests_inheritance"]
|
|
)
|
|
return suite.countTestCases()
|
|
|
|
self.assertEqual(count_tests(tags=["foo"]), 4)
|
|
self.assertEqual(count_tests(tags=["bar"]), 2)
|
|
self.assertEqual(count_tests(tags=["baz"]), 2)
|
|
self.assertEqual(count_tests(tags=["foo"], exclude_tags=["bar"]), 2)
|
|
self.assertEqual(count_tests(tags=["foo"], exclude_tags=["bar", "baz"]), 1)
|
|
self.assertEqual(count_tests(exclude_tags=["foo"]), 0)
|
|
|
|
def test_tag_fail_to_load(self):
|
|
with self.assertRaises(SyntaxError):
|
|
import_module("test_runner_apps.tagged.tests_syntax_error")
|
|
runner = DiscoverRunner(tags=["syntax_error"], verbosity=0)
|
|
# A label that doesn't exist or cannot be loaded due to syntax errors
|
|
# is always considered matching.
|
|
suite = runner.build_suite(["doesnotexist", "test_runner_apps.tagged"])
|
|
self.assertEqual(
|
|
[test.id() for test in suite],
|
|
[
|
|
"unittest.loader._FailedTest.doesnotexist",
|
|
"unittest.loader._FailedTest.test_runner_apps.tagged."
|
|
"tests_syntax_error",
|
|
],
|
|
)
|
|
|
|
def test_included_tags_displayed(self):
|
|
runner = DiscoverRunner(tags=["foo", "bar"], verbosity=2)
|
|
with captured_stdout() as stdout:
|
|
runner.build_suite(["test_runner_apps.tagged.tests"])
|
|
self.assertIn("Including test tag(s): bar, foo.\n", stdout.getvalue())
|
|
|
|
def test_excluded_tags_displayed(self):
|
|
runner = DiscoverRunner(exclude_tags=["foo", "bar"], verbosity=3)
|
|
with captured_stdout() as stdout:
|
|
runner.build_suite(["test_runner_apps.tagged.tests"])
|
|
self.assertIn("Excluding test tag(s): bar, foo.\n", stdout.getvalue())
|
|
|
|
def test_number_of_tests_found_displayed(self):
|
|
runner = DiscoverRunner()
|
|
with captured_stdout() as stdout:
|
|
runner.build_suite(
|
|
[
|
|
"test_runner_apps.sample.tests_sample.TestDjangoTestCase",
|
|
"test_runner_apps.simple",
|
|
]
|
|
)
|
|
self.assertIn("Found 14 test(s).\n", stdout.getvalue())
|
|
|
|
def test_pdb_with_parallel(self):
|
|
msg = "You cannot use --pdb with parallel tests; pass --parallel=1 to use it."
|
|
with self.assertRaisesMessage(ValueError, msg):
|
|
DiscoverRunner(pdb=True, parallel=2)
|
|
|
|
def test_number_of_parallel_workers(self):
|
|
"""Number of processes doesn't exceed the number of TestCases."""
|
|
runner = DiscoverRunner(parallel=5, verbosity=0)
|
|
suite = runner.build_suite(["test_runner_apps.tagged"])
|
|
self.assertEqual(suite.processes, len(suite.subsuites))
|
|
|
|
def test_number_of_databases_parallel_test_suite(self):
|
|
"""
|
|
Number of databases doesn't exceed the number of TestCases with
|
|
parallel tests.
|
|
"""
|
|
runner = DiscoverRunner(parallel=8, verbosity=0)
|
|
suite = runner.build_suite(["test_runner_apps.tagged"])
|
|
self.assertEqual(suite.processes, len(suite.subsuites))
|
|
self.assertEqual(runner.parallel, suite.processes)
|
|
|
|
def test_number_of_databases_no_parallel_test_suite(self):
|
|
"""
|
|
Number of databases doesn't exceed the number of TestCases with
|
|
non-parallel tests.
|
|
"""
|
|
runner = DiscoverRunner(parallel=8, verbosity=0)
|
|
suite = runner.build_suite(["test_runner_apps.simple.tests.DjangoCase1"])
|
|
self.assertEqual(runner.parallel, 1)
|
|
self.assertIsInstance(suite, TestSuite)
|
|
|
|
def test_buffer_mode_test_pass(self):
|
|
runner = DiscoverRunner(buffer=True, verbosity=0)
|
|
with captured_stdout() as stdout, captured_stderr() as stderr:
|
|
suite = runner.build_suite(
|
|
[
|
|
"test_runner_apps.buffer.tests_buffer.WriteToStdoutStderrTestCase."
|
|
"test_pass",
|
|
]
|
|
)
|
|
runner.run_suite(suite)
|
|
self.assertNotIn("Write to stderr.", stderr.getvalue())
|
|
self.assertNotIn("Write to stdout.", stdout.getvalue())
|
|
|
|
def test_buffer_mode_test_fail(self):
|
|
runner = DiscoverRunner(buffer=True, verbosity=0)
|
|
with captured_stdout() as stdout, captured_stderr() as stderr:
|
|
suite = runner.build_suite(
|
|
[
|
|
"test_runner_apps.buffer.tests_buffer.WriteToStdoutStderrTestCase."
|
|
"test_fail",
|
|
]
|
|
)
|
|
runner.run_suite(suite)
|
|
self.assertIn("Write to stderr.", stderr.getvalue())
|
|
self.assertIn("Write to stdout.", stdout.getvalue())
|
|
|
|
def run_suite_with_runner(self, runner_class, **kwargs):
|
|
class MyRunner(DiscoverRunner):
|
|
def test_runner(self, *args, **kwargs):
|
|
return runner_class()
|
|
|
|
runner = MyRunner(**kwargs)
|
|
# Suppress logging "Using shuffle seed" to the console.
|
|
with captured_stdout():
|
|
runner.setup_shuffler()
|
|
with captured_stdout() as stdout:
|
|
try:
|
|
result = runner.run_suite(None)
|
|
except RuntimeError as exc:
|
|
result = str(exc)
|
|
output = stdout.getvalue()
|
|
return result, output
|
|
|
|
def test_run_suite_logs_seed(self):
|
|
class TestRunner:
|
|
def run(self, suite):
|
|
return "<fake-result>"
|
|
|
|
expected_prefix = "Used shuffle seed"
|
|
# Test with and without shuffling enabled.
|
|
result, output = self.run_suite_with_runner(TestRunner)
|
|
self.assertEqual(result, "<fake-result>")
|
|
self.assertNotIn(expected_prefix, output)
|
|
|
|
result, output = self.run_suite_with_runner(TestRunner, shuffle=2)
|
|
self.assertEqual(result, "<fake-result>")
|
|
expected_output = f"{expected_prefix}: 2 (given)\n"
|
|
self.assertEqual(output, expected_output)
|
|
|
|
def test_run_suite_logs_seed_exception(self):
|
|
"""
|
|
run_suite() logs the seed when TestRunner.run() raises an exception.
|
|
"""
|
|
|
|
class TestRunner:
|
|
def run(self, suite):
|
|
raise RuntimeError("my exception")
|
|
|
|
result, output = self.run_suite_with_runner(TestRunner, shuffle=2)
|
|
self.assertEqual(result, "my exception")
|
|
expected_output = "Used shuffle seed: 2 (given)\n"
|
|
self.assertEqual(output, expected_output)
|
|
|
|
@mock.patch("faulthandler.enable")
|
|
def test_faulthandler_enabled(self, mocked_enable):
|
|
with mock.patch("faulthandler.is_enabled", return_value=False):
|
|
DiscoverRunner(enable_faulthandler=True)
|
|
mocked_enable.assert_called()
|
|
|
|
@mock.patch("faulthandler.enable")
|
|
def test_faulthandler_already_enabled(self, mocked_enable):
|
|
with mock.patch("faulthandler.is_enabled", return_value=True):
|
|
DiscoverRunner(enable_faulthandler=True)
|
|
mocked_enable.assert_not_called()
|
|
|
|
@mock.patch("faulthandler.enable")
|
|
def test_faulthandler_enabled_fileno(self, mocked_enable):
|
|
# sys.stderr that is not an actual file.
|
|
with mock.patch(
|
|
"faulthandler.is_enabled", return_value=False
|
|
), captured_stderr():
|
|
DiscoverRunner(enable_faulthandler=True)
|
|
mocked_enable.assert_called()
|
|
|
|
@mock.patch("faulthandler.enable")
|
|
def test_faulthandler_disabled(self, mocked_enable):
|
|
with mock.patch("faulthandler.is_enabled", return_value=False):
|
|
DiscoverRunner(enable_faulthandler=False)
|
|
mocked_enable.assert_not_called()
|
|
|
|
def test_timings_not_captured(self):
|
|
runner = DiscoverRunner(timing=False)
|
|
with captured_stderr() as stderr:
|
|
with runner.time_keeper.timed("test"):
|
|
pass
|
|
runner.time_keeper.print_results()
|
|
self.assertIsInstance(runner.time_keeper, NullTimeKeeper)
|
|
self.assertNotIn("test", stderr.getvalue())
|
|
|
|
def test_timings_captured(self):
|
|
runner = DiscoverRunner(timing=True)
|
|
with captured_stderr() as stderr:
|
|
with runner.time_keeper.timed("test"):
|
|
pass
|
|
runner.time_keeper.print_results()
|
|
self.assertIsInstance(runner.time_keeper, TimeKeeper)
|
|
self.assertIn("test", stderr.getvalue())
|
|
|
|
def test_log(self):
|
|
custom_low_level = 5
|
|
custom_high_level = 45
|
|
msg = "logging message"
|
|
cases = [
|
|
(0, None, False),
|
|
(0, custom_low_level, False),
|
|
(0, logging.DEBUG, False),
|
|
(0, logging.INFO, False),
|
|
(0, logging.WARNING, False),
|
|
(0, custom_high_level, False),
|
|
(1, None, True),
|
|
(1, custom_low_level, False),
|
|
(1, logging.DEBUG, False),
|
|
(1, logging.INFO, True),
|
|
(1, logging.WARNING, True),
|
|
(1, custom_high_level, True),
|
|
(2, None, True),
|
|
(2, custom_low_level, True),
|
|
(2, logging.DEBUG, True),
|
|
(2, logging.INFO, True),
|
|
(2, logging.WARNING, True),
|
|
(2, custom_high_level, True),
|
|
(3, None, True),
|
|
(3, custom_low_level, True),
|
|
(3, logging.DEBUG, True),
|
|
(3, logging.INFO, True),
|
|
(3, logging.WARNING, True),
|
|
(3, custom_high_level, True),
|
|
]
|
|
for verbosity, level, output in cases:
|
|
with self.subTest(verbosity=verbosity, level=level):
|
|
with captured_stdout() as stdout:
|
|
runner = DiscoverRunner(verbosity=verbosity)
|
|
runner.log(msg, level)
|
|
self.assertEqual(stdout.getvalue(), f"{msg}\n" if output else "")
|
|
|
|
def test_log_logger(self):
|
|
logger = logging.getLogger("test.logging")
|
|
cases = [
|
|
(None, "INFO:test.logging:log message"),
|
|
# Test a low custom logging level.
|
|
(5, "Level 5:test.logging:log message"),
|
|
(logging.DEBUG, "DEBUG:test.logging:log message"),
|
|
(logging.INFO, "INFO:test.logging:log message"),
|
|
(logging.WARNING, "WARNING:test.logging:log message"),
|
|
# Test a high custom logging level.
|
|
(45, "Level 45:test.logging:log message"),
|
|
]
|
|
for level, expected in cases:
|
|
with self.subTest(level=level):
|
|
runner = DiscoverRunner(logger=logger)
|
|
# Pass a logging level smaller than the smallest level in cases
|
|
# in order to capture all messages.
|
|
with self.assertLogs("test.logging", level=1) as cm:
|
|
runner.log("log message", level)
|
|
self.assertEqual(cm.output, [expected])
|
|
|
|
def test_suite_result_with_failure(self):
|
|
cases = [
|
|
(1, "FailureTestCase"),
|
|
(1, "ErrorTestCase"),
|
|
(0, "ExpectedFailureTestCase"),
|
|
(1, "UnexpectedSuccessTestCase"),
|
|
]
|
|
runner = DiscoverRunner(verbosity=0)
|
|
for expected_failures, testcase in cases:
|
|
with self.subTest(testcase=testcase):
|
|
suite = runner.build_suite(
|
|
[
|
|
f"test_runner_apps.failures.tests_failures.{testcase}",
|
|
]
|
|
)
|
|
with captured_stderr():
|
|
result = runner.run_suite(suite)
|
|
failures = runner.suite_result(suite, result)
|
|
self.assertEqual(failures, expected_failures)
|
|
|
|
|
|
class DiscoverRunnerGetDatabasesTests(SimpleTestCase):
|
|
runner = DiscoverRunner(verbosity=2)
|
|
skip_msg = "Skipping setup of unused database(s): "
|
|
|
|
def get_databases(self, test_labels):
|
|
with captured_stdout() as stdout:
|
|
suite = self.runner.build_suite(test_labels)
|
|
databases = self.runner.get_databases(suite)
|
|
return databases, stdout.getvalue()
|
|
|
|
def assertSkippedDatabases(self, test_labels, expected_databases):
|
|
databases, output = self.get_databases(test_labels)
|
|
self.assertEqual(databases, expected_databases)
|
|
skipped_databases = set(connections) - set(expected_databases)
|
|
if skipped_databases:
|
|
self.assertIn(self.skip_msg + ", ".join(sorted(skipped_databases)), output)
|
|
else:
|
|
self.assertNotIn(self.skip_msg, output)
|
|
|
|
def test_mixed(self):
|
|
databases, output = self.get_databases(["test_runner_apps.databases.tests"])
|
|
self.assertEqual(databases, {"default": True, "other": False})
|
|
self.assertNotIn(self.skip_msg, output)
|
|
|
|
def test_all(self):
|
|
databases, output = self.get_databases(
|
|
["test_runner_apps.databases.tests.AllDatabasesTests"]
|
|
)
|
|
self.assertEqual(databases, {alias: False for alias in connections})
|
|
self.assertNotIn(self.skip_msg, output)
|
|
|
|
def test_default_and_other(self):
|
|
self.assertSkippedDatabases(
|
|
[
|
|
"test_runner_apps.databases.tests.DefaultDatabaseTests",
|
|
"test_runner_apps.databases.tests.OtherDatabaseTests",
|
|
],
|
|
{"default": False, "other": False},
|
|
)
|
|
|
|
def test_default_only(self):
|
|
self.assertSkippedDatabases(
|
|
[
|
|
"test_runner_apps.databases.tests.DefaultDatabaseTests",
|
|
],
|
|
{"default": False},
|
|
)
|
|
|
|
def test_other_only(self):
|
|
self.assertSkippedDatabases(
|
|
["test_runner_apps.databases.tests.OtherDatabaseTests"], {"other": False}
|
|
)
|
|
|
|
def test_no_databases_required(self):
|
|
self.assertSkippedDatabases(
|
|
["test_runner_apps.databases.tests.NoDatabaseTests"], {}
|
|
)
|
|
|
|
def test_serialize(self):
|
|
databases, _ = self.get_databases(
|
|
["test_runner_apps.databases.tests.DefaultDatabaseSerializedTests"]
|
|
)
|
|
self.assertEqual(databases, {"default": True})
|