992 lines
29 KiB
Python
992 lines
29 KiB
Python
# -*- coding: utf-8 -*-
|
|
from __future__ import absolute_import
|
|
from __future__ import division
|
|
from __future__ import print_function
|
|
|
|
import inspect
|
|
import os
|
|
import sys
|
|
import types
|
|
|
|
import py
|
|
|
|
import _pytest._code
|
|
import pytest
|
|
from _pytest import main
|
|
from _pytest import outcomes
|
|
from _pytest import reports
|
|
from _pytest import runner
|
|
|
|
|
|
class TestSetupState(object):
|
|
def test_setup(self, testdir):
|
|
ss = runner.SetupState()
|
|
item = testdir.getitem("def test_func(): pass")
|
|
values = [1]
|
|
ss.prepare(item)
|
|
ss.addfinalizer(values.pop, colitem=item)
|
|
assert values
|
|
ss._pop_and_teardown()
|
|
assert not values
|
|
|
|
def test_teardown_exact_stack_empty(self, testdir):
|
|
item = testdir.getitem("def test_func(): pass")
|
|
ss = runner.SetupState()
|
|
ss.teardown_exact(item, None)
|
|
ss.teardown_exact(item, None)
|
|
ss.teardown_exact(item, None)
|
|
|
|
def test_setup_fails_and_failure_is_cached(self, testdir):
|
|
item = testdir.getitem(
|
|
"""
|
|
def setup_module(mod):
|
|
raise ValueError(42)
|
|
def test_func(): pass
|
|
"""
|
|
)
|
|
ss = runner.SetupState()
|
|
pytest.raises(ValueError, lambda: ss.prepare(item))
|
|
pytest.raises(ValueError, lambda: ss.prepare(item))
|
|
|
|
def test_teardown_multiple_one_fails(self, testdir):
|
|
r = []
|
|
|
|
def fin1():
|
|
r.append("fin1")
|
|
|
|
def fin2():
|
|
raise Exception("oops")
|
|
|
|
def fin3():
|
|
r.append("fin3")
|
|
|
|
item = testdir.getitem("def test_func(): pass")
|
|
ss = runner.SetupState()
|
|
ss.addfinalizer(fin1, item)
|
|
ss.addfinalizer(fin2, item)
|
|
ss.addfinalizer(fin3, item)
|
|
with pytest.raises(Exception) as err:
|
|
ss._callfinalizers(item)
|
|
assert err.value.args == ("oops",)
|
|
assert r == ["fin3", "fin1"]
|
|
|
|
def test_teardown_multiple_fail(self, testdir):
|
|
# Ensure the first exception is the one which is re-raised.
|
|
# Ideally both would be reported however.
|
|
def fin1():
|
|
raise Exception("oops1")
|
|
|
|
def fin2():
|
|
raise Exception("oops2")
|
|
|
|
item = testdir.getitem("def test_func(): pass")
|
|
ss = runner.SetupState()
|
|
ss.addfinalizer(fin1, item)
|
|
ss.addfinalizer(fin2, item)
|
|
with pytest.raises(Exception) as err:
|
|
ss._callfinalizers(item)
|
|
assert err.value.args == ("oops2",)
|
|
|
|
def test_teardown_multiple_scopes_one_fails(self, testdir):
|
|
module_teardown = []
|
|
|
|
def fin_func():
|
|
raise Exception("oops1")
|
|
|
|
def fin_module():
|
|
module_teardown.append("fin_module")
|
|
|
|
item = testdir.getitem("def test_func(): pass")
|
|
ss = runner.SetupState()
|
|
ss.addfinalizer(fin_module, item.listchain()[-2])
|
|
ss.addfinalizer(fin_func, item)
|
|
ss.prepare(item)
|
|
with pytest.raises(Exception, match="oops1"):
|
|
ss.teardown_exact(item, None)
|
|
assert module_teardown
|
|
|
|
|
|
class BaseFunctionalTests(object):
|
|
def test_passfunction(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert rep.passed
|
|
assert not rep.failed
|
|
assert rep.outcome == "passed"
|
|
assert not rep.longrepr
|
|
|
|
def test_failfunction(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
assert 0
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert not rep.passed
|
|
assert not rep.skipped
|
|
assert rep.failed
|
|
assert rep.when == "call"
|
|
assert rep.outcome == "failed"
|
|
# assert isinstance(rep.longrepr, ReprExceptionInfo)
|
|
|
|
def test_skipfunction(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def test_func():
|
|
pytest.skip("hello")
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert not rep.failed
|
|
assert not rep.passed
|
|
assert rep.skipped
|
|
assert rep.outcome == "skipped"
|
|
# assert rep.skipped.when == "call"
|
|
# assert rep.skipped.when == "call"
|
|
# assert rep.skipped == "%sreason == "hello"
|
|
# assert rep.skipped.location.lineno == 3
|
|
# assert rep.skipped.location.path
|
|
# assert not rep.skipped.failurerepr
|
|
|
|
def test_skip_in_setup_function(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def setup_function(func):
|
|
pytest.skip("hello")
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
print(reports)
|
|
rep = reports[0]
|
|
assert not rep.failed
|
|
assert not rep.passed
|
|
assert rep.skipped
|
|
# assert rep.skipped.reason == "hello"
|
|
# assert rep.skipped.location.lineno == 3
|
|
# assert rep.skipped.location.lineno == 3
|
|
assert len(reports) == 2
|
|
assert reports[1].passed # teardown
|
|
|
|
def test_failure_in_setup_function(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def setup_function(func):
|
|
raise ValueError(42)
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
rep = reports[0]
|
|
assert not rep.skipped
|
|
assert not rep.passed
|
|
assert rep.failed
|
|
assert rep.when == "setup"
|
|
assert len(reports) == 2
|
|
|
|
def test_failure_in_teardown_function(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def teardown_function(func):
|
|
raise ValueError(42)
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
print(reports)
|
|
assert len(reports) == 3
|
|
rep = reports[2]
|
|
assert not rep.skipped
|
|
assert not rep.passed
|
|
assert rep.failed
|
|
assert rep.when == "teardown"
|
|
# assert rep.longrepr.reprcrash.lineno == 3
|
|
# assert rep.longrepr.reprtraceback.reprentries
|
|
|
|
def test_custom_failure_repr(self, testdir):
|
|
testdir.makepyfile(
|
|
conftest="""
|
|
import pytest
|
|
class Function(pytest.Function):
|
|
def repr_failure(self, excinfo):
|
|
return "hello"
|
|
"""
|
|
)
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def test_func():
|
|
assert 0
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert not rep.skipped
|
|
assert not rep.passed
|
|
assert rep.failed
|
|
# assert rep.outcome.when == "call"
|
|
# assert rep.failed.where.lineno == 3
|
|
# assert rep.failed.where.path.basename == "test_func.py"
|
|
# assert rep.failed.failurerepr == "hello"
|
|
|
|
def test_teardown_final_returncode(self, testdir):
|
|
rec = testdir.inline_runsource(
|
|
"""
|
|
def test_func():
|
|
pass
|
|
def teardown_function(func):
|
|
raise ValueError(42)
|
|
"""
|
|
)
|
|
assert rec.ret == 1
|
|
|
|
def test_logstart_logfinish_hooks(self, testdir):
|
|
rec = testdir.inline_runsource(
|
|
"""
|
|
import pytest
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
reps = rec.getcalls("pytest_runtest_logstart pytest_runtest_logfinish")
|
|
assert [x._name for x in reps] == [
|
|
"pytest_runtest_logstart",
|
|
"pytest_runtest_logfinish",
|
|
]
|
|
for rep in reps:
|
|
assert rep.nodeid == "test_logstart_logfinish_hooks.py::test_func"
|
|
assert rep.location == ("test_logstart_logfinish_hooks.py", 1, "test_func")
|
|
|
|
def test_exact_teardown_issue90(self, testdir):
|
|
rec = testdir.inline_runsource(
|
|
"""
|
|
import pytest
|
|
|
|
class TestClass(object):
|
|
def test_method(self):
|
|
pass
|
|
def teardown_class(cls):
|
|
raise Exception()
|
|
|
|
def test_func():
|
|
import sys
|
|
# on python2 exc_info is keept till a function exits
|
|
# so we would end up calling test functions while
|
|
# sys.exc_info would return the indexerror
|
|
# from guessing the lastitem
|
|
excinfo = sys.exc_info()
|
|
import traceback
|
|
assert excinfo[0] is None, \
|
|
traceback.format_exception(*excinfo)
|
|
def teardown_function(func):
|
|
raise ValueError(42)
|
|
"""
|
|
)
|
|
reps = rec.getreports("pytest_runtest_logreport")
|
|
print(reps)
|
|
for i in range(2):
|
|
assert reps[i].nodeid.endswith("test_method")
|
|
assert reps[i].passed
|
|
assert reps[2].when == "teardown"
|
|
assert reps[2].failed
|
|
assert len(reps) == 6
|
|
for i in range(3, 5):
|
|
assert reps[i].nodeid.endswith("test_func")
|
|
assert reps[i].passed
|
|
assert reps[5].when == "teardown"
|
|
assert reps[5].nodeid.endswith("test_func")
|
|
assert reps[5].failed
|
|
|
|
def test_exact_teardown_issue1206(self, testdir):
|
|
"""issue shadowing error with wrong number of arguments on teardown_method."""
|
|
rec = testdir.inline_runsource(
|
|
"""
|
|
import pytest
|
|
|
|
class TestClass(object):
|
|
def teardown_method(self, x, y, z):
|
|
pass
|
|
|
|
def test_method(self):
|
|
assert True
|
|
"""
|
|
)
|
|
reps = rec.getreports("pytest_runtest_logreport")
|
|
print(reps)
|
|
assert len(reps) == 3
|
|
#
|
|
assert reps[0].nodeid.endswith("test_method")
|
|
assert reps[0].passed
|
|
assert reps[0].when == "setup"
|
|
#
|
|
assert reps[1].nodeid.endswith("test_method")
|
|
assert reps[1].passed
|
|
assert reps[1].when == "call"
|
|
#
|
|
assert reps[2].nodeid.endswith("test_method")
|
|
assert reps[2].failed
|
|
assert reps[2].when == "teardown"
|
|
assert reps[2].longrepr.reprcrash.message in (
|
|
# python3 error
|
|
"TypeError: teardown_method() missing 2 required positional arguments: 'y' and 'z'",
|
|
# python2 error
|
|
"TypeError: teardown_method() takes exactly 4 arguments (2 given)",
|
|
)
|
|
|
|
def test_failure_in_setup_function_ignores_custom_repr(self, testdir):
|
|
testdir.makepyfile(
|
|
conftest="""
|
|
import pytest
|
|
class Function(pytest.Function):
|
|
def repr_failure(self, excinfo):
|
|
assert 0
|
|
"""
|
|
)
|
|
reports = testdir.runitem(
|
|
"""
|
|
def setup_function(func):
|
|
raise ValueError(42)
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
assert len(reports) == 2
|
|
rep = reports[0]
|
|
print(rep)
|
|
assert not rep.skipped
|
|
assert not rep.passed
|
|
assert rep.failed
|
|
# assert rep.outcome.when == "setup"
|
|
# assert rep.outcome.where.lineno == 3
|
|
# assert rep.outcome.where.path.basename == "test_func.py"
|
|
# assert instanace(rep.failed.failurerepr, PythonFailureRepr)
|
|
|
|
def test_systemexit_does_not_bail_out(self, testdir):
|
|
try:
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
raise SystemExit(42)
|
|
"""
|
|
)
|
|
except SystemExit:
|
|
pytest.fail("runner did not catch SystemExit")
|
|
rep = reports[1]
|
|
assert rep.failed
|
|
assert rep.when == "call"
|
|
|
|
def test_exit_propagates(self, testdir):
|
|
try:
|
|
testdir.runitem(
|
|
"""
|
|
import pytest
|
|
def test_func():
|
|
raise pytest.exit.Exception()
|
|
"""
|
|
)
|
|
except pytest.exit.Exception:
|
|
pass
|
|
else:
|
|
pytest.fail("did not raise")
|
|
|
|
|
|
class TestExecutionNonForked(BaseFunctionalTests):
|
|
def getrunner(self):
|
|
def f(item):
|
|
return runner.runtestprotocol(item, log=False)
|
|
|
|
return f
|
|
|
|
def test_keyboardinterrupt_propagates(self, testdir):
|
|
try:
|
|
testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
raise KeyboardInterrupt("fake")
|
|
"""
|
|
)
|
|
except KeyboardInterrupt:
|
|
pass
|
|
else:
|
|
pytest.fail("did not raise")
|
|
|
|
|
|
class TestExecutionForked(BaseFunctionalTests):
|
|
pytestmark = pytest.mark.skipif("not hasattr(os, 'fork')")
|
|
|
|
def getrunner(self):
|
|
# XXX re-arrange this test to live in pytest-xdist
|
|
boxed = pytest.importorskip("xdist.boxed")
|
|
return boxed.forked_run_report
|
|
|
|
def test_suicide(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
import os
|
|
os.kill(os.getpid(), 15)
|
|
"""
|
|
)
|
|
rep = reports[0]
|
|
assert rep.failed
|
|
assert rep.when == "???"
|
|
|
|
|
|
class TestSessionReports(object):
|
|
def test_collect_result(self, testdir):
|
|
col = testdir.getmodulecol(
|
|
"""
|
|
def test_func1():
|
|
pass
|
|
class TestClass(object):
|
|
pass
|
|
"""
|
|
)
|
|
rep = runner.collect_one_node(col)
|
|
assert not rep.failed
|
|
assert not rep.skipped
|
|
assert rep.passed
|
|
locinfo = rep.location
|
|
assert locinfo[0] == col.fspath.basename
|
|
assert not locinfo[1]
|
|
assert locinfo[2] == col.fspath.basename
|
|
res = rep.result
|
|
assert len(res) == 2
|
|
assert res[0].name == "test_func1"
|
|
assert res[1].name == "TestClass"
|
|
|
|
|
|
reporttypes = [
|
|
reports.BaseReport,
|
|
reports.TestReport,
|
|
reports.TeardownErrorReport,
|
|
reports.CollectReport,
|
|
]
|
|
|
|
|
|
@pytest.mark.parametrize(
|
|
"reporttype", reporttypes, ids=[x.__name__ for x in reporttypes]
|
|
)
|
|
def test_report_extra_parameters(reporttype):
|
|
if hasattr(inspect, "signature"):
|
|
args = list(inspect.signature(reporttype.__init__).parameters.keys())[1:]
|
|
else:
|
|
args = inspect.getargspec(reporttype.__init__)[0][1:]
|
|
basekw = dict.fromkeys(args, [])
|
|
report = reporttype(newthing=1, **basekw)
|
|
assert report.newthing == 1
|
|
|
|
|
|
def test_callinfo():
|
|
ci = runner.CallInfo(lambda: 0, "123")
|
|
assert ci.when == "123"
|
|
assert ci.result == 0
|
|
assert "result" in repr(ci)
|
|
assert repr(ci) == "<CallInfo when='123' result: 0>"
|
|
|
|
ci = runner.CallInfo(lambda: 0 / 0, "123")
|
|
assert ci.when == "123"
|
|
assert not hasattr(ci, "result")
|
|
assert repr(ci) == "<CallInfo when='123' exception: division by zero>"
|
|
assert ci.excinfo
|
|
assert "exc" in repr(ci)
|
|
|
|
|
|
def test_callinfo_repr_while_running():
|
|
def repr_while_running():
|
|
f = sys._getframe().f_back
|
|
assert "func" in f.f_locals
|
|
assert repr(f.f_locals["self"]) == "<CallInfo when='when' result: '<NOTSET>'>"
|
|
|
|
ci = runner.CallInfo(repr_while_running, "when")
|
|
assert repr(ci) == "<CallInfo when='when' result: None>"
|
|
|
|
|
|
# design question: do we want general hooks in python files?
|
|
# then something like the following functional tests makes sense
|
|
|
|
|
|
@pytest.mark.xfail
|
|
def test_runtest_in_module_ordering(testdir):
|
|
p1 = testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
def pytest_runtest_setup(item): # runs after class-level!
|
|
item.function.mylist.append("module")
|
|
class TestClass(object):
|
|
def pytest_runtest_setup(self, item):
|
|
assert not hasattr(item.function, 'mylist')
|
|
item.function.mylist = ['class']
|
|
@pytest.fixture
|
|
def mylist(self, request):
|
|
return request.function.mylist
|
|
def pytest_runtest_call(self, item, __multicall__):
|
|
try:
|
|
__multicall__.execute()
|
|
except ValueError:
|
|
pass
|
|
def test_hello1(self, mylist):
|
|
assert mylist == ['class', 'module'], mylist
|
|
raise ValueError()
|
|
def test_hello2(self, mylist):
|
|
assert mylist == ['class', 'module'], mylist
|
|
def pytest_runtest_teardown(item):
|
|
del item.function.mylist
|
|
"""
|
|
)
|
|
result = testdir.runpytest(p1)
|
|
result.stdout.fnmatch_lines(["*2 passed*"])
|
|
|
|
|
|
def test_outcomeexception_exceptionattributes():
|
|
outcome = outcomes.OutcomeException("test")
|
|
assert outcome.args[0] == outcome.msg
|
|
|
|
|
|
def test_outcomeexception_passes_except_Exception():
|
|
with pytest.raises(outcomes.OutcomeException):
|
|
try:
|
|
raise outcomes.OutcomeException("test")
|
|
except Exception:
|
|
pass
|
|
|
|
|
|
def test_pytest_exit():
|
|
try:
|
|
pytest.exit("hello")
|
|
except pytest.exit.Exception:
|
|
excinfo = _pytest._code.ExceptionInfo()
|
|
assert excinfo.errisinstance(KeyboardInterrupt)
|
|
|
|
|
|
def test_pytest_fail():
|
|
try:
|
|
pytest.fail("hello")
|
|
except pytest.fail.Exception:
|
|
excinfo = _pytest._code.ExceptionInfo()
|
|
s = excinfo.exconly(tryshort=True)
|
|
assert s.startswith("Failed")
|
|
|
|
|
|
def test_pytest_exit_msg(testdir):
|
|
testdir.makeconftest(
|
|
"""
|
|
import pytest
|
|
|
|
def pytest_configure(config):
|
|
pytest.exit('oh noes')
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
result.stderr.fnmatch_lines(["Exit: oh noes"])
|
|
|
|
|
|
def test_pytest_exit_returncode(testdir):
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
def test_foo():
|
|
pytest.exit("some exit msg", 99)
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
assert result.ret == 99
|
|
|
|
|
|
def test_pytest_fail_notrace_runtest(testdir):
|
|
"""Test pytest.fail(..., pytrace=False) does not show tracebacks during test run."""
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
def test_hello():
|
|
pytest.fail("hello", pytrace=False)
|
|
def teardown_function(function):
|
|
pytest.fail("world", pytrace=False)
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
result.stdout.fnmatch_lines(["world", "hello"])
|
|
assert "def teardown_function" not in result.stdout.str()
|
|
|
|
|
|
def test_pytest_fail_notrace_collection(testdir):
|
|
"""Test pytest.fail(..., pytrace=False) does not show tracebacks during collection."""
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
def some_internal_function():
|
|
pytest.fail("hello", pytrace=False)
|
|
some_internal_function()
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
result.stdout.fnmatch_lines(["hello"])
|
|
assert "def some_internal_function()" not in result.stdout.str()
|
|
|
|
|
|
@pytest.mark.parametrize("str_prefix", ["u", ""])
|
|
def test_pytest_fail_notrace_non_ascii(testdir, str_prefix):
|
|
"""Fix pytest.fail with pytrace=False with non-ascii characters (#1178).
|
|
|
|
This tests with native and unicode strings containing non-ascii chars.
|
|
"""
|
|
testdir.makepyfile(
|
|
u"""
|
|
# coding: utf-8
|
|
import pytest
|
|
|
|
def test_hello():
|
|
pytest.fail(%s'oh oh: ☺', pytrace=False)
|
|
"""
|
|
% str_prefix
|
|
)
|
|
result = testdir.runpytest()
|
|
if sys.version_info[0] >= 3:
|
|
result.stdout.fnmatch_lines(["*test_hello*", "oh oh: ☺"])
|
|
else:
|
|
result.stdout.fnmatch_lines(["*test_hello*", "oh oh: *"])
|
|
assert "def test_hello" not in result.stdout.str()
|
|
|
|
|
|
def test_pytest_no_tests_collected_exit_status(testdir):
|
|
result = testdir.runpytest()
|
|
result.stdout.fnmatch_lines("*collected 0 items*")
|
|
assert result.ret == main.EXIT_NOTESTSCOLLECTED
|
|
|
|
testdir.makepyfile(
|
|
test_foo="""
|
|
def test_foo():
|
|
assert 1
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
result.stdout.fnmatch_lines("*collected 1 item*")
|
|
result.stdout.fnmatch_lines("*1 passed*")
|
|
assert result.ret == main.EXIT_OK
|
|
|
|
result = testdir.runpytest("-k nonmatch")
|
|
result.stdout.fnmatch_lines("*collected 1 item*")
|
|
result.stdout.fnmatch_lines("*1 deselected*")
|
|
assert result.ret == main.EXIT_NOTESTSCOLLECTED
|
|
|
|
|
|
def test_exception_printing_skip():
|
|
try:
|
|
pytest.skip("hello")
|
|
except pytest.skip.Exception:
|
|
excinfo = _pytest._code.ExceptionInfo()
|
|
s = excinfo.exconly(tryshort=True)
|
|
assert s.startswith("Skipped")
|
|
|
|
|
|
def test_importorskip(monkeypatch):
|
|
importorskip = pytest.importorskip
|
|
|
|
def f():
|
|
importorskip("asdlkj")
|
|
|
|
try:
|
|
sysmod = importorskip("sys")
|
|
assert sysmod is sys
|
|
# path = pytest.importorskip("os.path")
|
|
# assert path == os.path
|
|
excinfo = pytest.raises(pytest.skip.Exception, f)
|
|
path = py.path.local(excinfo.getrepr().reprcrash.path)
|
|
# check that importorskip reports the actual call
|
|
# in this test the test_runner.py file
|
|
assert path.purebasename == "test_runner"
|
|
pytest.raises(SyntaxError, "pytest.importorskip('x y z')")
|
|
pytest.raises(SyntaxError, "pytest.importorskip('x=y')")
|
|
mod = types.ModuleType("hello123")
|
|
mod.__version__ = "1.3"
|
|
monkeypatch.setitem(sys.modules, "hello123", mod)
|
|
pytest.raises(
|
|
pytest.skip.Exception,
|
|
"""
|
|
pytest.importorskip("hello123", minversion="1.3.1")
|
|
""",
|
|
)
|
|
mod2 = pytest.importorskip("hello123", minversion="1.3")
|
|
assert mod2 == mod
|
|
except pytest.skip.Exception:
|
|
print(_pytest._code.ExceptionInfo())
|
|
pytest.fail("spurious skip")
|
|
|
|
|
|
def test_importorskip_imports_last_module_part():
|
|
ospath = pytest.importorskip("os.path")
|
|
assert os.path == ospath
|
|
|
|
|
|
def test_importorskip_dev_module(monkeypatch):
|
|
try:
|
|
mod = types.ModuleType("mockmodule")
|
|
mod.__version__ = "0.13.0.dev-43290"
|
|
monkeypatch.setitem(sys.modules, "mockmodule", mod)
|
|
mod2 = pytest.importorskip("mockmodule", minversion="0.12.0")
|
|
assert mod2 == mod
|
|
pytest.raises(
|
|
pytest.skip.Exception,
|
|
"""
|
|
pytest.importorskip('mockmodule1', minversion='0.14.0')""",
|
|
)
|
|
except pytest.skip.Exception:
|
|
print(_pytest._code.ExceptionInfo())
|
|
pytest.fail("spurious skip")
|
|
|
|
|
|
def test_importorskip_module_level(testdir):
|
|
"""importorskip must be able to skip entire modules when used at module level"""
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
foobarbaz = pytest.importorskip("foobarbaz")
|
|
|
|
def test_foo():
|
|
pass
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
result.stdout.fnmatch_lines(["*collected 0 items / 1 skipped*"])
|
|
|
|
|
|
def test_pytest_cmdline_main(testdir):
|
|
p = testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
def test_hello():
|
|
assert 1
|
|
if __name__ == '__main__':
|
|
pytest.cmdline.main([__file__])
|
|
"""
|
|
)
|
|
import subprocess
|
|
|
|
popen = subprocess.Popen([sys.executable, str(p)], stdout=subprocess.PIPE)
|
|
popen.communicate()
|
|
ret = popen.wait()
|
|
assert ret == 0
|
|
|
|
|
|
def test_unicode_in_longrepr(testdir):
|
|
testdir.makeconftest(
|
|
"""
|
|
# -*- coding: utf-8 -*-
|
|
import pytest
|
|
@pytest.hookimpl(hookwrapper=True)
|
|
def pytest_runtest_makereport():
|
|
outcome = yield
|
|
rep = outcome.get_result()
|
|
if rep.when == "call":
|
|
rep.longrepr = u'ä'
|
|
"""
|
|
)
|
|
testdir.makepyfile(
|
|
"""
|
|
def test_out():
|
|
assert 0
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
assert result.ret == 1
|
|
assert "UnicodeEncodeError" not in result.stderr.str()
|
|
|
|
|
|
def test_failure_in_setup(testdir):
|
|
testdir.makepyfile(
|
|
"""
|
|
def setup_module():
|
|
0/0
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
result = testdir.runpytest("--tb=line")
|
|
assert "def setup_module" not in result.stdout.str()
|
|
|
|
|
|
def test_makereport_getsource(testdir):
|
|
testdir.makepyfile(
|
|
"""
|
|
def test_foo():
|
|
if False: pass
|
|
else: assert False
|
|
"""
|
|
)
|
|
result = testdir.runpytest()
|
|
assert "INTERNALERROR" not in result.stdout.str()
|
|
result.stdout.fnmatch_lines(["*else: assert False*"])
|
|
|
|
|
|
def test_makereport_getsource_dynamic_code(testdir, monkeypatch):
|
|
"""Test that exception in dynamically generated code doesn't break getting the source line."""
|
|
import inspect
|
|
|
|
original_findsource = inspect.findsource
|
|
|
|
def findsource(obj, *args, **kwargs):
|
|
# Can be triggered by dynamically created functions
|
|
if obj.__name__ == "foo":
|
|
raise IndexError()
|
|
return original_findsource(obj, *args, **kwargs)
|
|
|
|
monkeypatch.setattr(inspect, "findsource", findsource)
|
|
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
|
|
@pytest.fixture
|
|
def foo(missing):
|
|
pass
|
|
|
|
def test_fix(foo):
|
|
assert False
|
|
"""
|
|
)
|
|
result = testdir.runpytest("-vv")
|
|
assert "INTERNALERROR" not in result.stdout.str()
|
|
result.stdout.fnmatch_lines(["*test_fix*", "*fixture*'missing'*not found*"])
|
|
|
|
|
|
def test_store_except_info_on_error():
|
|
""" Test that upon test failure, the exception info is stored on
|
|
sys.last_traceback and friends.
|
|
"""
|
|
# Simulate item that might raise a specific exception, depending on `raise_error` class var
|
|
class ItemMightRaise(object):
|
|
nodeid = "item_that_raises"
|
|
raise_error = True
|
|
|
|
def runtest(self):
|
|
if self.raise_error:
|
|
raise IndexError("TEST")
|
|
|
|
try:
|
|
runner.pytest_runtest_call(ItemMightRaise())
|
|
except IndexError:
|
|
pass
|
|
# Check that exception info is stored on sys
|
|
assert sys.last_type is IndexError
|
|
assert sys.last_value.args[0] == "TEST"
|
|
assert sys.last_traceback
|
|
|
|
# The next run should clear the exception info stored by the previous run
|
|
ItemMightRaise.raise_error = False
|
|
runner.pytest_runtest_call(ItemMightRaise())
|
|
assert sys.last_type is None
|
|
assert sys.last_value is None
|
|
assert sys.last_traceback is None
|
|
|
|
|
|
def test_current_test_env_var(testdir, monkeypatch):
|
|
pytest_current_test_vars = []
|
|
monkeypatch.setattr(
|
|
sys, "pytest_current_test_vars", pytest_current_test_vars, raising=False
|
|
)
|
|
testdir.makepyfile(
|
|
"""
|
|
import pytest
|
|
import sys
|
|
import os
|
|
|
|
@pytest.fixture
|
|
def fix():
|
|
sys.pytest_current_test_vars.append(('setup', os.environ['PYTEST_CURRENT_TEST']))
|
|
yield
|
|
sys.pytest_current_test_vars.append(('teardown', os.environ['PYTEST_CURRENT_TEST']))
|
|
|
|
def test(fix):
|
|
sys.pytest_current_test_vars.append(('call', os.environ['PYTEST_CURRENT_TEST']))
|
|
"""
|
|
)
|
|
result = testdir.runpytest_inprocess()
|
|
assert result.ret == 0
|
|
test_id = "test_current_test_env_var.py::test"
|
|
assert pytest_current_test_vars == [
|
|
("setup", test_id + " (setup)"),
|
|
("call", test_id + " (call)"),
|
|
("teardown", test_id + " (teardown)"),
|
|
]
|
|
assert "PYTEST_CURRENT_TEST" not in os.environ
|
|
|
|
|
|
class TestReportContents(object):
|
|
"""
|
|
Test user-level API of ``TestReport`` objects.
|
|
"""
|
|
|
|
def getrunner(self):
|
|
return lambda item: runner.runtestprotocol(item, log=False)
|
|
|
|
def test_longreprtext_pass(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert rep.longreprtext == ""
|
|
|
|
def test_longreprtext_failure(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
x = 1
|
|
assert x == 4
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert "assert 1 == 4" in rep.longreprtext
|
|
|
|
def test_captured_text(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
import pytest
|
|
import sys
|
|
|
|
@pytest.fixture
|
|
def fix():
|
|
sys.stdout.write('setup: stdout\\n')
|
|
sys.stderr.write('setup: stderr\\n')
|
|
yield
|
|
sys.stdout.write('teardown: stdout\\n')
|
|
sys.stderr.write('teardown: stderr\\n')
|
|
assert 0
|
|
|
|
def test_func(fix):
|
|
sys.stdout.write('call: stdout\\n')
|
|
sys.stderr.write('call: stderr\\n')
|
|
assert 0
|
|
"""
|
|
)
|
|
setup, call, teardown = reports
|
|
assert setup.capstdout == "setup: stdout\n"
|
|
assert call.capstdout == "setup: stdout\ncall: stdout\n"
|
|
assert teardown.capstdout == "setup: stdout\ncall: stdout\nteardown: stdout\n"
|
|
|
|
assert setup.capstderr == "setup: stderr\n"
|
|
assert call.capstderr == "setup: stderr\ncall: stderr\n"
|
|
assert teardown.capstderr == "setup: stderr\ncall: stderr\nteardown: stderr\n"
|
|
|
|
def test_no_captured_text(self, testdir):
|
|
reports = testdir.runitem(
|
|
"""
|
|
def test_func():
|
|
pass
|
|
"""
|
|
)
|
|
rep = reports[1]
|
|
assert rep.capstdout == ""
|
|
assert rep.capstderr == ""
|