2010-11-21 04:35:55 +08:00
2011-02-09 21:55:21 +08:00
.. _paramexamples:
2011-09-06 17:43:42 +08:00
Parametrizing tests
2010-11-21 04:35:55 +08:00
=================================================
2011-11-17 19:09:21 +08:00
.. currentmodule :: _pytest.python
2014-01-18 19:31:33 +08:00
`` pytest `` allows to easily parametrize test functions.
2012-10-18 18:24:50 +08:00
For basic docs, see :ref: `parametrize-basics` .
2011-11-17 19:09:21 +08:00
In the following we provide some examples using
the builtin mechanisms.
2011-09-06 17:43:42 +08:00
Generating parameters combinations, depending on command line
2011-02-09 21:55:21 +08:00
----------------------------------------------------------------------------
.. regendoc:wipe
2011-11-17 19:09:21 +08:00
Let's say we want to execute a test with different computation
parameters and the parameter range shall be determined by a command
line argument. Let's first write a simple (do-nothing) computation test::
2011-02-09 21:55:21 +08:00
# content of test_compute.py
def test_compute(param1):
assert param1 < 4
Now we add a test configuration like this::
# content of conftest.py
def pytest_addoption(parser):
parser.addoption("--all", action="store_true",
help="run all combinations")
def pytest_generate_tests(metafunc):
2012-10-05 20:24:44 +08:00
if 'param1' in metafunc.fixturenames:
2017-07-28 06:18:44 +08:00
if metafunc.config.getoption('all'):
2011-02-09 21:55:21 +08:00
end = 5
else:
end = 2
2011-11-17 19:09:21 +08:00
metafunc.parametrize("param1", range(end))
2011-02-09 21:55:21 +08:00
This means that we only run 2 tests if we do not pass `` --all `` ::
2016-06-21 22:16:57 +08:00
$ pytest -q test_compute.py
2017-11-23 23:33:41 +08:00
.. [100%]
2015-06-07 05:30:49 +08:00
2 passed in 0.12 seconds
2011-02-09 21:55:21 +08:00
We run only two computations, so we see two dots.
let's run the full monty::
2016-06-21 22:16:57 +08:00
$ pytest -q --all
2017-11-23 23:33:41 +08:00
....F [100%]
================================= FAILURES =================================
_____________________________ test_compute[4] ______________________________
2015-05-19 08:54:24 +08:00
2011-02-09 21:55:21 +08:00
param1 = 4
2015-05-19 08:54:24 +08:00
2011-02-09 21:55:21 +08:00
def test_compute(param1):
> assert param1 < 4
E assert 4 < 4
2015-05-19 08:54:24 +08:00
2011-02-09 21:55:21 +08:00
test_compute.py:3: AssertionError
2015-06-07 05:30:49 +08:00
1 failed, 4 passed in 0.12 seconds
2011-02-09 21:55:21 +08:00
As expected when running the full range of `` param1 `` values
we'll get an error on the last one.
2014-04-18 03:08:49 +08:00
Different options for test IDs
------------------------------------
pytest will build a string that is the test ID for each set of values in a
2014-10-08 07:43:27 +08:00
parametrized test. These IDs can be used with `` -k `` to select specific cases
2014-04-18 03:08:49 +08:00
to run, and they will also identify the specific case when one is failing.
2014-10-08 07:43:27 +08:00
Running pytest with `` --collect-only `` will show the generated IDs.
2014-04-18 03:08:49 +08:00
Numbers, strings, booleans and None will have their usual string representation
used in the test ID. For other objects, pytest will make a string based on
the argument name::
2015-07-25 23:48:12 +08:00
# content of test_time.py
2014-04-18 03:08:49 +08:00
2015-09-26 00:42:06 +08:00
import pytest
2014-04-18 03:08:49 +08:00
from datetime import datetime, timedelta
2015-09-26 00:42:06 +08:00
testdata = [
(datetime(2001, 12, 12), datetime(2001, 12, 11), timedelta(1)),
(datetime(2001, 12, 11), datetime(2001, 12, 12), timedelta(-1)),
]
2014-04-18 03:08:49 +08:00
@pytest.mark.parametrize("a,b,expected", testdata)
def test_timedistance_v0(a, b, expected):
2015-09-26 00:42:06 +08:00
diff = a - b
assert diff == expected
2014-04-18 03:08:49 +08:00
@pytest.mark.parametrize("a,b,expected", testdata, ids=["forward", "backward"])
def test_timedistance_v1(a, b, expected):
2015-09-26 00:42:06 +08:00
diff = a - b
assert diff == expected
2014-04-18 03:08:49 +08:00
def idfn(val):
2015-09-26 00:42:06 +08:00
if isinstance(val, (datetime,)):
# note this wouldn't show any hours/minutes/seconds
return val.strftime('%Y%m%d')
2014-04-18 03:08:49 +08:00
@pytest.mark.parametrize("a,b,expected", testdata, ids=idfn)
def test_timedistance_v2(a, b, expected):
2015-09-26 00:42:06 +08:00
diff = a - b
assert diff == expected
2014-04-18 03:08:49 +08:00
2017-05-23 13:57:34 +08:00
@pytest.mark.parametrize("a,b,expected", [
pytest.param(datetime(2001, 12, 12), datetime(2001, 12, 11),
timedelta(1), id='forward'),
pytest.param(datetime(2001, 12, 11), datetime(2001, 12, 12),
timedelta(-1), id='backward'),
])
def test_timedistance_v3(a, b, expected):
diff = a - b
assert diff == expected
2014-04-18 03:08:49 +08:00
In `` test_timedistance_v0 `` , we let pytest generate the test IDs.
In `` test_timedistance_v1 `` , we specified `` ids `` as a list of strings which were
used as the test IDs. These are succinct, but can be a pain to maintain.
In `` test_timedistance_v2 `` , we specified `` ids `` as a function that can generate a
string representation to make part of the test ID. So our `` datetime `` values use the
label generated by `` idfn `` , but because we didn't generate a label for `` timedelta ``
objects, they are still using the default pytest representation::
2016-06-21 22:16:57 +08:00
$ pytest test_time.py --collect-only
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2017-05-23 13:57:34 +08:00
collected 8 items
2015-09-26 00:42:06 +08:00
<Module 'test_time.py'>
<Function 'test_timedistance_v0[a0-b0-expected0]'>
<Function 'test_timedistance_v0[a1-b1-expected1]'>
<Function 'test_timedistance_v1[forward]'>
<Function 'test_timedistance_v1[backward]'>
<Function 'test_timedistance_v2[20011212-20011211-expected0]'>
<Function 'test_timedistance_v2[20011211-20011212-expected1]'>
2017-05-23 13:57:34 +08:00
<Function 'test_timedistance_v3[forward]'>
<Function 'test_timedistance_v3[backward]'>
2015-03-26 16:34:10 +08:00
2017-11-23 23:33:41 +08:00
======================= no tests ran in 0.12 seconds =======================
2014-04-18 03:08:49 +08:00
2017-05-23 13:57:34 +08:00
In `` test_timedistance_v3 `` , we used `` pytest.param `` to specify the test IDs
together with the actual data, instead of listing them separately.
2011-12-05 18:10:48 +08:00
A quick port of "testscenarios"
2011-11-17 19:09:21 +08:00
------------------------------------
2012-07-02 19:13:48 +08:00
.. _`test scenarios`: http://pypi.python.org/pypi/testscenarios/
2011-11-17 19:09:21 +08:00
2011-12-05 18:10:48 +08:00
Here is a quick port to run tests configured with `test scenarios`_ ,
2011-11-17 19:09:21 +08:00
an add-on from Robert Collins for the standard unittest framework. We
only have to work a bit to construct the correct arguments for pytest's
:py:func: `Metafunc.parametrize` ::
# content of test_scenarios.py
def pytest_generate_tests(metafunc):
idlist = []
argvalues = []
for scenario in metafunc.cls.scenarios:
idlist.append(scenario[0])
items = scenario[1].items()
argnames = [x[0] for x in items]
argvalues.append(([x[1] for x in items]))
2012-09-21 15:39:54 +08:00
metafunc.parametrize(argnames, argvalues, ids=idlist, scope="class")
2011-11-17 19:09:21 +08:00
scenario1 = ('basic', {'attribute': 'value'})
scenario2 = ('advanced', {'attribute': 'value2'})
2017-02-17 02:41:51 +08:00
class TestSampleWithScenarios(object):
2011-11-17 19:09:21 +08:00
scenarios = [scenario1, scenario2]
2012-09-21 15:39:54 +08:00
def test_demo1(self, attribute):
assert isinstance(attribute, str)
def test_demo2(self, attribute):
2011-11-17 19:09:21 +08:00
assert isinstance(attribute, str)
this is a fully self-contained example which you can run with::
2016-06-21 22:16:57 +08:00
$ pytest test_scenarios.py
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2012-10-07 19:06:17 +08:00
collected 4 items
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
test_scenarios.py .... [100%]
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
========================= 4 passed in 0.12 seconds =========================
2011-11-17 19:09:21 +08:00
If you just collect tests you'll also nicely see 'advanced' and 'basic' as variants for the test function::
2016-06-21 22:16:57 +08:00
$ pytest --collect-only test_scenarios.py
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2012-10-07 19:06:17 +08:00
collected 4 items
2011-11-17 19:09:21 +08:00
<Module 'test_scenarios.py'>
<Class 'TestSampleWithScenarios'>
<Instance '()'>
2012-09-21 15:39:54 +08:00
<Function 'test_demo1[basic]'>
<Function 'test_demo2[basic]'>
<Function 'test_demo1[advanced]'>
<Function 'test_demo2[advanced]'>
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
======================= no tests ran in 0.12 seconds =======================
2012-09-21 15:39:54 +08:00
Note that we told `` metafunc.parametrize() `` that your scenario values
should be considered class-scoped. With pytest-2.3 this leads to a
resource-based ordering.
2011-11-17 19:09:21 +08:00
Deferring the setup of parametrized resources
2011-02-09 21:55:21 +08:00
---------------------------------------------------
.. regendoc:wipe
The parametrization of test functions happens at collection
2011-11-17 19:09:21 +08:00
time. It is a good idea to setup expensive resources like DB
2011-12-05 18:10:48 +08:00
connections or subprocess only when the actual test is run.
Here is a simple example how you can achieve that, first
2011-11-17 19:09:21 +08:00
the actual test requiring a `` db `` object::
2011-02-09 21:55:21 +08:00
# content of test_backends.py
2011-12-05 18:10:48 +08:00
2011-02-09 21:55:21 +08:00
import pytest
def test_db_initialized(db):
# a dummy test
if db.__class__.__name__ == "DB2":
pytest.fail("deliberately failing for demo purposes")
2011-11-17 19:09:21 +08:00
We can now add a test configuration that generates two invocations of
the `` test_db_initialized `` function and also implements a factory that
creates a database object for the actual test invocations::
2011-02-09 21:55:21 +08:00
# content of conftest.py
2012-10-18 18:24:50 +08:00
import pytest
2011-02-09 21:55:21 +08:00
def pytest_generate_tests(metafunc):
2012-10-05 20:24:44 +08:00
if 'db' in metafunc.fixturenames:
2011-11-17 19:09:21 +08:00
metafunc.parametrize("db", ['d1', 'd2'], indirect=True)
2011-02-09 21:55:21 +08:00
2017-02-17 02:41:51 +08:00
class DB1(object):
2011-02-09 21:55:21 +08:00
"one database object"
2017-02-17 02:41:51 +08:00
class DB2(object):
2011-02-09 21:55:21 +08:00
"alternative database object"
2011-12-05 18:10:48 +08:00
2012-10-18 18:24:50 +08:00
@pytest.fixture
def db(request):
2011-02-09 21:55:21 +08:00
if request.param == "d1":
return DB1()
elif request.param == "d2":
return DB2()
else:
raise ValueError("invalid internal test config")
Let's first see how it looks like at collection time::
2016-06-21 22:16:57 +08:00
$ pytest test_backends.py --collect-only
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2012-10-07 19:06:17 +08:00
collected 2 items
2011-02-09 21:55:21 +08:00
<Module 'test_backends.py'>
2011-11-17 19:09:21 +08:00
<Function 'test_db_initialized[d1]'>
<Function 'test_db_initialized[d2]'>
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
======================= no tests ran in 0.12 seconds =======================
2011-02-09 21:55:21 +08:00
And then when we run the test::
2016-06-21 22:16:57 +08:00
$ pytest -q test_backends.py
2017-11-23 23:33:41 +08:00
.F [100%]
================================= FAILURES =================================
_________________________ test_db_initialized[d2] __________________________
2014-01-29 20:47:11 +08:00
2015-09-22 22:52:35 +08:00
db = <conftest.DB2 object at 0xdeadbeef>
2014-01-29 20:47:11 +08:00
2011-02-09 21:55:21 +08:00
def test_db_initialized(db):
# a dummy test
if db.__class__.__name__ == "DB2":
> pytest.fail("deliberately failing for demo purposes")
E Failed: deliberately failing for demo purposes
2014-01-29 20:47:11 +08:00
2011-02-09 21:55:21 +08:00
test_backends.py:6: Failed
2015-06-07 05:30:49 +08:00
1 failed, 1 passed in 0.12 seconds
2011-02-09 21:55:21 +08:00
2012-10-18 18:24:50 +08:00
The first invocation with `` db == "DB1" `` passed while the second with `` db == "DB2" `` failed. Our `` db `` fixture function has instantiated each of the DB values during the setup phase while the `` pytest_generate_tests `` generated two according calls to the `` test_db_initialized `` during the collection phase.
2011-11-17 19:09:21 +08:00
.. regendoc:wipe
2011-02-09 21:55:21 +08:00
2015-08-04 05:48:41 +08:00
Apply indirect on particular arguments
---------------------------------------------------
Very often parametrization uses more than one argument name. There is opportunity to apply `` indirect ``
parameter on particular arguments. It can be done by passing list or tuple of
arguments' names to `` indirect `` . In the example below there is a function `` test_indirect `` which uses
two fixtures: `` x `` and `` y `` . Here we give to indirect the list, which contains the name of the
fixture `` x `` . The indirect parameter will be applied to this argument only, and the value `` a ``
2015-09-26 00:43:33 +08:00
will be passed to respective fixture function::
2015-08-04 05:48:41 +08:00
# content of test_indirect_list.py
import pytest
@pytest.fixture(scope='function')
def x(request):
return request.param * 3
@pytest.fixture(scope='function')
def y(request):
return request.param * 2
@pytest.mark.parametrize('x, y', [('a', 'b')], indirect=['x'])
def test_indirect(x,y):
assert x == 'aaa'
assert y == 'b'
2015-09-26 00:43:33 +08:00
The result of this test will be successful::
2015-08-04 05:48:41 +08:00
2016-06-21 22:16:57 +08:00
$ pytest test_indirect_list.py --collect-only
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2017-07-04 07:29:13 +08:00
collected 1 item
2015-09-22 20:02:11 +08:00
<Module 'test_indirect_list.py'>
<Function 'test_indirect[a-b]'>
2017-11-23 23:33:41 +08:00
======================= no tests ran in 0.12 seconds =======================
2015-08-04 05:48:41 +08:00
.. regendoc:wipe
2010-11-21 04:35:55 +08:00
Parametrizing test methods through per-class configuration
--------------------------------------------------------------
2015-12-27 09:35:02 +08:00
.. _`unittest parametrizer`: https://github.com/testing-cabal/unittest-ext/blob/master/params.py
2010-11-21 04:35:55 +08:00
2011-11-17 19:09:21 +08:00
2017-10-16 06:55:30 +08:00
Here is an example `` pytest_generate_tests `` function implementing a
2011-12-05 18:10:48 +08:00
parametrization scheme similar to Michael Foord's `unittest
2015-11-28 14:46:45 +08:00
parametrizer`_ but in a lot less code::
2010-11-21 04:35:55 +08:00
# content of ./test_parametrize.py
import pytest
def pytest_generate_tests(metafunc):
# called once per each test function
2011-11-17 19:09:21 +08:00
funcarglist = metafunc.cls.params[metafunc.function.__name__]
2016-08-04 04:48:11 +08:00
argnames = sorted(funcarglist[0])
2011-12-05 18:10:48 +08:00
metafunc.parametrize(argnames, [[funcargs[name] for name in argnames]
2011-11-17 19:09:21 +08:00
for funcargs in funcarglist])
2010-11-21 04:35:55 +08:00
2017-02-17 02:41:51 +08:00
class TestClass(object):
2010-11-21 04:35:55 +08:00
# a map specifying multiple argument sets for a test method
params = {
'test_equals': [dict(a=1, b=2), dict(a=3, b=3), ],
2011-11-17 19:09:21 +08:00
'test_zerodivision': [dict(a=1, b=0), ],
2010-11-21 04:35:55 +08:00
}
def test_equals(self, a, b):
assert a == b
def test_zerodivision(self, a, b):
pytest.raises(ZeroDivisionError, "a/b")
2011-11-17 19:09:21 +08:00
Our test generator looks up a class-level definition which specifies which
argument sets to use for each test function. Let's run it::
2010-11-21 04:35:55 +08:00
2016-06-21 22:16:57 +08:00
$ pytest -q
2017-11-23 23:33:41 +08:00
F.. [100%]
================================= FAILURES =================================
________________________ TestClass.test_equals[1-2] ________________________
2014-01-29 20:47:11 +08:00
2015-09-22 22:52:35 +08:00
self = <test_parametrize.TestClass object at 0xdeadbeef>, a = 1, b = 2
2014-01-29 20:47:11 +08:00
2010-11-21 04:35:55 +08:00
def test_equals(self, a, b):
> assert a == b
E assert 1 == 2
2014-01-29 20:47:11 +08:00
2011-11-17 19:09:21 +08:00
test_parametrize.py:18: AssertionError
2015-06-07 05:30:49 +08:00
1 failed, 2 passed in 0.12 seconds
2010-11-21 04:35:55 +08:00
2012-10-18 18:24:50 +08:00
Indirect parametrization with multiple fixtures
2010-11-21 04:35:55 +08:00
--------------------------------------------------------------
Here is a stripped down real-life example of using parametrized
2012-10-18 18:24:50 +08:00
testing for testing serialization of objects between different python
interpreters. We define a `` test_basic_objects `` function which
is to be run with different sets of arguments for its three arguments:
2010-11-21 04:35:55 +08:00
2011-11-17 19:09:21 +08:00
* `` python1 `` : first python interpreter, run to pickle-dump an object to a file
2011-12-05 18:10:48 +08:00
* `` python2 `` : second interpreter, run to pickle-load an object from a file
2011-11-17 19:09:21 +08:00
* `` obj `` : object to be dumped/loaded
2010-11-21 04:35:55 +08:00
.. literalinclude :: multipython.py
2011-11-19 02:32:11 +08:00
Running it results in some skips if we don't have all the python interpreters installed and otherwise runs all combinations (5 interpreters times 5 interpreters times 3 objects to serialize/deserialize)::
2010-11-21 04:35:55 +08:00
2016-06-21 22:16:57 +08:00
. $ pytest -rs -q multipython.py
2017-11-24 04:25:36 +08:00
........................... [100%]
27 passed in 0.12 seconds
2012-12-20 22:57:07 +08:00
Indirect parametrization of optional implementations/imports
--------------------------------------------------------------------
If you want to compare the outcomes of several implementations of a given
API, you can write test functions that receive the already imported implementations
and get skipped in case the implementation is not importable/available. Let's
2014-01-18 19:31:33 +08:00
say we have a "base" implementation and the other (possibly optimized ones)
2012-12-20 22:57:07 +08:00
need to provide similar results::
# content of conftest.py
import pytest
@pytest.fixture(scope="session")
def basemod(request):
return pytest.importorskip("base")
@pytest.fixture(scope="session", params=["opt1", "opt2"])
def optmod(request):
return pytest.importorskip(request.param)
And then a base implementation of a simple function::
# content of base.py
def func1():
return 1
And an optimized version::
# content of opt1.py
def func1():
return 1.0001
And finally a little test module::
# content of test_module.py
def test_func1(basemod, optmod):
assert round(basemod.func1(), 3) == round(optmod.func1(), 3)
If you run this with reporting for skips enabled::
2016-06-21 22:16:57 +08:00
$ pytest -rs test_module.py
2017-11-23 23:33:41 +08:00
=========================== test session starts ============================
2017-05-13 04:17:40 +08:00
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
2017-03-14 06:41:20 +08:00
rootdir: $REGENDOC_TMPDIR, inifile:
2012-12-20 22:57:07 +08:00
collected 2 items
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
test_module.py .s [100%]
========================= short test summary info ==========================
2017-07-31 05:37:18 +08:00
SKIP [1] $REGENDOC_TMPDIR/conftest.py:11: could not import 'opt2'
2014-01-29 20:47:11 +08:00
2017-11-23 23:33:41 +08:00
=================== 1 passed, 1 skipped in 0.12 seconds ====================
2012-12-20 22:57:07 +08:00
You'll see that we don't have a `` opt2 `` module and thus the second test run
of our `` test_func1 `` was skipped. A few notes:
- the fixture functions in the `` conftest.py `` file are "session-scoped" because we
2014-01-18 19:31:33 +08:00
don't need to import more than once
2012-12-20 22:57:07 +08:00
- if you have multiple test functions and a skipped import, you will see
the `` [1] `` count increasing in the report
- you can put :ref: `@pytest.mark.parametrize <@pytest.mark.parametrize>` style
2014-01-18 19:31:33 +08:00
parametrization on the test functions to parametrize input/output
2012-12-20 22:57:07 +08:00
values as well.
2017-11-04 02:37:18 +08:00
Set marks or test ID for individual parametrized test
--------------------------------------------------------------------
Use `` pytest.param `` to apply marks or set test ID to individual parametrized test.
For example::
# content of test_pytest_param_example.py
import pytest
@pytest.mark.parametrize('test_input,expected', [
('3+5', 8),
pytest.param('1+7', 8,
marks=pytest.mark.basic),
pytest.param('2+4', 6,
marks=pytest.mark.basic,
id='basic_2+4'),
pytest.param('6*9', 42,
marks=[pytest.mark.basic, pytest.mark.xfail],
id='basic_6*9'),
])
def test_eval(test_input, expected):
assert eval(test_input) == expected
In this example, we have 4 parametrized tests. Except for the first test,
we mark the rest three parametrized tests with the custom marker `` basic `` ,
and for the fourth test we also use the built-in mark `` xfail `` to indicate this
test is expected to fail. For explicitness, we set test ids for some tests.
Then run `` pytest `` with verbose mode and with only the `` basic `` marker::
pytest -v -m basic
============================================ test session starts =============================================
platform linux -- Python 3.x.y, pytest-3.x.y, py-1.x.y, pluggy-0.x.y
rootdir: $REGENDOC_TMPDIR, inifile:
collected 4 items
test_pytest_param_example.py::test_eval[1+7-8] PASSED
test_pytest_param_example.py::test_eval[basic_2+4] PASSED
test_pytest_param_example.py::test_eval[basic_6*9] xfail
========================================== short test summary info ===========================================
XFAIL test_pytest_param_example.py::test_eval[basic_6*9]
============================================= 1 tests deselected =============================================
As the result:
2012-12-20 22:57:07 +08:00
2017-11-04 02:37:18 +08:00
- Four tests were collected
- One test was deselected because it doesn't have the `` basic `` mark.
- Three tests with the `` basic `` mark was selected.
- The test `` test_eval[1+7-8] `` passed, but the name is autogenerated and confusing.
- The test `` test_eval[basic_2+4] `` passed.
- The test `` test_eval[basic_6*9] `` was expected to fail and did fail.