2018-02-08 06:53:31 +08:00
|
|
|
|
|
|
|
Reference
|
|
|
|
=========
|
|
|
|
|
|
|
|
This page contains the full reference to pytest's API.
|
|
|
|
|
2018-02-27 07:13:33 +08:00
|
|
|
.. contents::
|
2018-02-28 04:32:34 +08:00
|
|
|
:depth: 3
|
2018-02-27 07:27:15 +08:00
|
|
|
:local:
|
|
|
|
|
|
|
|
|
2018-02-28 05:26:40 +08:00
|
|
|
pytest.approx
|
|
|
|
-------------
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.python_api.approx
|
|
|
|
|
2018-02-27 07:34:53 +08:00
|
|
|
outcomes
|
|
|
|
--------
|
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
You can use the following functions in your test, fixture or setup
|
|
|
|
functions to force a certain test outcome. Note that most often
|
|
|
|
you can rather use declarative marks, see :ref:`skipping`.
|
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.fail
|
|
|
|
~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.outcomes.fail
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.skip
|
|
|
|
~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.outcomes.skip
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.importorskip
|
|
|
|
~~~~~~~~~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.outcomes.importorskip
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.xfail
|
|
|
|
~~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.outcomes.xfail
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.exit
|
|
|
|
~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.outcomes.exit
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.main
|
|
|
|
~~~~~~~~~~~
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.config.main
|
|
|
|
|
2018-03-01 07:34:20 +08:00
|
|
|
pytest.param
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
|
|
|
.. autofunction:: _pytest.mark.param
|
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.raises
|
|
|
|
-------------
|
2018-02-27 07:57:02 +08:00
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`assertraises`.
|
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.python_api.raises
|
2018-02-27 07:57:02 +08:00
|
|
|
:with: excinfo
|
2018-02-27 07:27:15 +08:00
|
|
|
|
2018-02-28 04:15:08 +08:00
|
|
|
pytest.deprecated_call
|
|
|
|
----------------------
|
2018-02-27 07:34:53 +08:00
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`ensuring_function_triggers`.
|
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
.. autofunction:: _pytest.recwarn.deprecated_call
|
2018-02-27 07:57:02 +08:00
|
|
|
:with:
|
2018-02-08 06:53:31 +08:00
|
|
|
|
2018-02-28 05:26:40 +08:00
|
|
|
pytest.warns
|
|
|
|
------------
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`assertwarnings`
|
|
|
|
|
|
|
|
.. autofunction:: _pytest.recwarn.warns
|
|
|
|
:with:
|
|
|
|
|
|
|
|
|
2018-02-08 06:53:31 +08:00
|
|
|
.. _`hook-reference`:
|
|
|
|
|
|
|
|
Hooks
|
|
|
|
-----
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`writing_plugins`.
|
2018-02-08 06:53:31 +08:00
|
|
|
|
|
|
|
.. currentmodule:: _pytest.hookspec
|
|
|
|
|
|
|
|
Reference to all hooks which can be implemented by :ref:`conftest.py files <localplugin>` and :ref:`plugins <plugins>`.
|
|
|
|
|
|
|
|
Bootstrapping hooks
|
|
|
|
~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
Bootstrapping hooks called for plugins registered early enough (internal and setuptools plugins).
|
|
|
|
|
|
|
|
.. autofunction:: pytest_load_initial_conftests
|
|
|
|
.. autofunction:: pytest_cmdline_preparse
|
|
|
|
.. autofunction:: pytest_cmdline_parse
|
|
|
|
.. autofunction:: pytest_cmdline_main
|
|
|
|
|
|
|
|
Initialization hooks
|
|
|
|
~~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
Initialization hooks called for plugins and ``conftest.py`` files.
|
|
|
|
|
|
|
|
.. autofunction:: pytest_addoption
|
|
|
|
.. autofunction:: pytest_configure
|
|
|
|
.. autofunction:: pytest_unconfigure
|
|
|
|
|
|
|
|
Test running hooks
|
|
|
|
~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
All runtest related hooks receive a :py:class:`pytest.Item <_pytest.main.Item>` object.
|
|
|
|
|
|
|
|
.. autofunction:: pytest_runtestloop
|
|
|
|
.. autofunction:: pytest_runtest_protocol
|
|
|
|
.. autofunction:: pytest_runtest_logstart
|
|
|
|
.. autofunction:: pytest_runtest_logfinish
|
|
|
|
.. autofunction:: pytest_runtest_setup
|
|
|
|
.. autofunction:: pytest_runtest_call
|
|
|
|
.. autofunction:: pytest_runtest_teardown
|
|
|
|
.. autofunction:: pytest_runtest_makereport
|
|
|
|
|
|
|
|
For deeper understanding you may look at the default implementation of
|
|
|
|
these hooks in :py:mod:`_pytest.runner` and maybe also
|
|
|
|
in :py:mod:`_pytest.pdb` which interacts with :py:mod:`_pytest.capture`
|
|
|
|
and its input/output capturing in order to immediately drop
|
|
|
|
into interactive debugging when a test failure occurs.
|
|
|
|
|
|
|
|
The :py:mod:`_pytest.terminal` reported specifically uses
|
|
|
|
the reporting hook to print information about a test run.
|
|
|
|
|
|
|
|
Collection hooks
|
|
|
|
~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
``pytest`` calls the following hooks for collecting files and directories:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_collection
|
|
|
|
.. autofunction:: pytest_ignore_collect
|
|
|
|
.. autofunction:: pytest_collect_directory
|
|
|
|
.. autofunction:: pytest_collect_file
|
|
|
|
|
|
|
|
For influencing the collection of objects in Python modules
|
|
|
|
you can use the following hook:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_pycollect_makeitem
|
|
|
|
.. autofunction:: pytest_generate_tests
|
|
|
|
.. autofunction:: pytest_make_parametrize_id
|
|
|
|
|
|
|
|
After collection is complete, you can modify the order of
|
|
|
|
items, delete or otherwise amend the test items:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_collection_modifyitems
|
|
|
|
|
|
|
|
Reporting hooks
|
|
|
|
~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
Session related reporting hooks:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_collectstart
|
|
|
|
.. autofunction:: pytest_itemcollected
|
|
|
|
.. autofunction:: pytest_collectreport
|
|
|
|
.. autofunction:: pytest_deselected
|
|
|
|
.. autofunction:: pytest_report_header
|
|
|
|
.. autofunction:: pytest_report_collectionfinish
|
|
|
|
.. autofunction:: pytest_report_teststatus
|
|
|
|
.. autofunction:: pytest_terminal_summary
|
|
|
|
.. autofunction:: pytest_fixture_setup
|
|
|
|
.. autofunction:: pytest_fixture_post_finalizer
|
|
|
|
|
|
|
|
And here is the central hook for reporting about
|
|
|
|
test execution:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_runtest_logreport
|
|
|
|
|
|
|
|
You can also use this hook to customize assertion representation for some
|
|
|
|
types:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_assertrepr_compare
|
|
|
|
|
|
|
|
|
|
|
|
Debugging/Interaction hooks
|
|
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
There are few hooks which can be used for special
|
|
|
|
reporting or interaction with exceptions:
|
|
|
|
|
|
|
|
.. autofunction:: pytest_internalerror
|
|
|
|
.. autofunction:: pytest_keyboard_interrupt
|
|
|
|
.. autofunction:: pytest_exception_interact
|
|
|
|
.. autofunction:: pytest_enter_pdb
|
|
|
|
|
|
|
|
|
|
|
|
Objects
|
|
|
|
-------
|
|
|
|
|
|
|
|
Full reference to objects accessible from :ref:`fixtures <fixture>` or hooks
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.config.Config()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.config.Parser()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.nodes.Node()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.nodes.Collector()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
2018-02-28 04:48:19 +08:00
|
|
|
.. autoclass:: _pytest._code.ExceptionInfo
|
|
|
|
:members:
|
|
|
|
|
2018-02-08 06:53:31 +08:00
|
|
|
.. autoclass:: _pytest.nodes.FSCollector()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.main.Session()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.nodes.Item()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.python.Module()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.python.Class()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.python.Function()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.fixtures.FixtureDef()
|
|
|
|
:members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.runner.CallInfo()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.runner.TestReport()
|
|
|
|
:members:
|
|
|
|
:inherited-members:
|
|
|
|
|
2018-03-01 07:34:20 +08:00
|
|
|
.. currentmodule:: _pytest.python
|
|
|
|
.. autoclass:: Metafunc
|
|
|
|
:members:
|
|
|
|
|
2018-02-08 06:53:31 +08:00
|
|
|
.. autoclass:: pluggy._Result
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autofunction:: _pytest.config.get_plugin_manager()
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.config.PytestPluginManager()
|
|
|
|
:members:
|
|
|
|
:undoc-members:
|
|
|
|
:show-inheritance:
|
|
|
|
|
|
|
|
.. autoclass:: pluggy.PluginManager()
|
|
|
|
:members:
|
|
|
|
|
2018-03-01 07:34:20 +08:00
|
|
|
.. currentmodule:: _pytest.mark
|
|
|
|
|
|
|
|
.. autoclass:: MarkGenerator
|
2018-02-28 05:26:40 +08:00
|
|
|
:members:
|
|
|
|
|
2018-03-01 07:34:20 +08:00
|
|
|
.. autoclass:: MarkDecorator
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: MarkInfo
|
|
|
|
:members:
|
|
|
|
|
|
|
|
|
|
|
|
Marks
|
|
|
|
-----
|
|
|
|
|
|
|
|
Marks can be used apply meta data to *test functions* (but not fixtures), which can then be accessed by
|
|
|
|
fixtures or plugins.
|
|
|
|
|
|
|
|
|
|
|
|
.. _`pytest.mark.parametrize ref`:
|
|
|
|
|
|
|
|
pytest.mark.parametrize
|
|
|
|
~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :doc:`parametrize`.
|
|
|
|
|
|
|
|
.. automethod:: _pytest.python.Metafunc.parametrize
|
|
|
|
|
|
|
|
|
|
|
|
.. _`pytest.mark.skip ref`:
|
|
|
|
|
|
|
|
pytest.mark.skip
|
|
|
|
~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`skip`.
|
|
|
|
|
|
|
|
Unconditionally skip a test function.
|
|
|
|
|
|
|
|
.. py:function:: pytest.mark.skip(*, reason=None)
|
|
|
|
|
|
|
|
:keyword str reason: Reason why the test function is being skipped.
|
|
|
|
|
|
|
|
|
|
|
|
.. _`pytest.mark.skipif ref`:
|
|
|
|
|
|
|
|
pytest.mark.skipif
|
|
|
|
~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`xfail`.
|
|
|
|
|
|
|
|
Skip a test function if a condition is ``True``.
|
|
|
|
|
|
|
|
.. py:function:: pytest.mark.skipif(condition, *, reason=None)
|
|
|
|
|
|
|
|
:type condition: bool or str
|
|
|
|
:param condition: ``True/False`` if the condition should be skipped or a :ref:`condition string <string conditions>`.
|
|
|
|
:keyword str reason: Reason why the test function is being skipped.
|
|
|
|
|
|
|
|
|
|
|
|
.. _`pytest.mark.xfail ref`:
|
|
|
|
|
|
|
|
pytest.mark.xfail
|
|
|
|
~~~~~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`xfail`.
|
|
|
|
|
|
|
|
Marks a test function as *expected to fail*.
|
|
|
|
|
|
|
|
.. py:function:: pytest.mark.xfail(condition=None, *, reason=None, raises=None, run=True, strict=False)
|
|
|
|
|
|
|
|
:type condition: bool or str
|
|
|
|
:param condition: ``True/False`` if the condition should be marked as xfail or a :ref:`condition string <string conditions>`.
|
|
|
|
:keyword str reason: Reason why the test function is marked as xfail.
|
|
|
|
:keyword Exception raises: Exception subclass expected to be raised by the test function; other exceptions will fail the test.
|
|
|
|
:keyword bool run:
|
|
|
|
If the test function should actually be executed. If ``False``, the function will always xfail and will
|
|
|
|
not be executed (useful a function is segfaulting).
|
|
|
|
:keyword bool strict:
|
|
|
|
* If ``False`` (the default) the function will be shown in the terminal output as ``xfailed`` if it fails
|
|
|
|
and as ``xpass`` if it passes. In both cases this will not cause the test suite to fail as a whole. This
|
|
|
|
is particularly useful to mark *flaky* tests (tests that random at fail) to be tackled later.
|
|
|
|
* If ``True``, the function will be shown in the terminal output as ``xfailed`` if it fails, but if it
|
|
|
|
unexpectedly passes then it will **fail** the test suite. This is particularly useful to mark functions
|
|
|
|
that are always failing and there should be a clear indication if they unexpectedly start to pass (for example
|
|
|
|
a new release of a library fixes a known bug).
|
|
|
|
|
|
|
|
|
|
|
|
custom marks
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
|
|
|
Marks are created dynamically using the factory object ``pytest.mark`` and applied as a decorator.
|
|
|
|
|
|
|
|
For example:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
@pytest.mark.timeout(10, 'slow', method='thread')
|
|
|
|
def test_function():
|
|
|
|
...
|
|
|
|
|
|
|
|
Will create and attach a :class:`MarkInfo <_pytest.mark.MarkInfo>` object to the collected
|
|
|
|
:class:`Item <_pytest.nodes.Item>`, which can then be accessed by fixtures or hooks with
|
|
|
|
:meth:`Node.get_marker <_pytest.nodes.Node.get_marker>`. The ``mark`` object will have the following attributes:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
mark.args == (10, 'slow')
|
|
|
|
mark.kwargs == {'method': 'thread'}
|
2018-02-28 05:26:40 +08:00
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
Fixtures
|
|
|
|
--------
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`fixture`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
Fixtures are requested by test functions or other fixtures by declaring them as argument names.
|
|
|
|
|
|
|
|
|
|
|
|
Example of a test requiring a fixture:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
def test_output(capsys):
|
|
|
|
print('hello')
|
|
|
|
out, err = capsys.readouterr()
|
|
|
|
assert out == 'hello\n'
|
|
|
|
|
|
|
|
|
|
|
|
Example of a fixture requiring another fixture:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
@pytest.fixture
|
|
|
|
def db_session(tmpdir):
|
|
|
|
fn = tmpdir / 'db.file'
|
|
|
|
return connect(str(fn))
|
|
|
|
|
|
|
|
For more details, consult the full :ref:`fixtures docs <fixture>`.
|
|
|
|
|
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
fixture decorator
|
|
|
|
~~~~~~~~~~~~~~~~~
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
.. currentmodule:: _pytest.fixtures
|
|
|
|
.. autofunction:: fixture
|
|
|
|
:decorator:
|
|
|
|
|
|
|
|
|
|
|
|
.. _`cache-api`:
|
|
|
|
|
|
|
|
config.cache
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`cache`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
The ``config.cache`` object allows other plugins and fixtures
|
|
|
|
to store and retrieve values across test runs. To access it from fixtures
|
|
|
|
request ``pytestconfig`` into your fixture and get it with ``pytestconfig.cache``.
|
|
|
|
|
|
|
|
Under the hood, the cache plugin uses the simple
|
|
|
|
``dumps``/``loads`` API of the :py:mod:`json` stdlib module.
|
|
|
|
|
|
|
|
.. currentmodule:: _pytest.cacheprovider
|
|
|
|
|
|
|
|
.. automethod:: Cache.get
|
|
|
|
.. automethod:: Cache.set
|
|
|
|
.. automethod:: Cache.makedir
|
|
|
|
|
|
|
|
|
|
|
|
capsys
|
|
|
|
~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`capture`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
.. currentmodule:: _pytest.capture
|
|
|
|
|
|
|
|
.. autofunction:: capsys()
|
2018-02-09 06:29:04 +08:00
|
|
|
:no-auto-options:
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
Returns an instance of :py:class:`CaptureFixture`.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
def test_output(capsys):
|
|
|
|
print("hello")
|
|
|
|
captured = capsys.readouterr()
|
|
|
|
assert captured.out == "hello\n"
|
|
|
|
|
|
|
|
|
|
|
|
capsysbinary
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`capture`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
.. autofunction:: capsysbinary()
|
2018-02-09 06:29:04 +08:00
|
|
|
:no-auto-options:
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
Returns an instance of :py:class:`CaptureFixture`.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
def test_output(capsysbinary):
|
|
|
|
print("hello")
|
|
|
|
captured = capsysbinary.readouterr()
|
|
|
|
assert captured.out == b"hello\n"
|
|
|
|
|
|
|
|
|
|
|
|
capfd
|
|
|
|
~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`capture`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
.. autofunction:: capfd()
|
2018-02-09 06:29:04 +08:00
|
|
|
:no-auto-options:
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
Returns an instance of :py:class:`CaptureFixture`.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
def test_system_echo(capfd):
|
|
|
|
os.system('echo "hello"')
|
|
|
|
captured = capsys.readouterr()
|
|
|
|
assert captured.out == "hello\n"
|
|
|
|
|
|
|
|
|
|
|
|
capfdbinary
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`capture`.
|
|
|
|
|
2018-02-09 05:57:26 +08:00
|
|
|
.. autofunction:: capfdbinary()
|
2018-02-09 06:29:04 +08:00
|
|
|
:no-auto-options:
|
2018-02-09 05:57:26 +08:00
|
|
|
|
|
|
|
Returns an instance of :py:class:`CaptureFixture`.
|
|
|
|
|
|
|
|
Example:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
def test_system_echo(capfdbinary):
|
|
|
|
os.system('echo "hello"')
|
|
|
|
captured = capfdbinary.readouterr()
|
|
|
|
assert captured.out == b"hello\n"
|
|
|
|
|
|
|
|
|
2018-02-09 06:08:44 +08:00
|
|
|
doctest_namespace
|
|
|
|
~~~~~~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :doc:`doctest`.
|
|
|
|
|
2018-02-09 06:08:44 +08:00
|
|
|
.. autofunction:: _pytest.doctest.doctest_namespace()
|
|
|
|
|
|
|
|
Usually this fixture is used in conjunction with another ``autouse`` fixture:
|
|
|
|
|
|
|
|
.. code-block:: python
|
|
|
|
|
|
|
|
@pytest.fixture(autouse=True)
|
|
|
|
def add_np(doctest_namespace):
|
|
|
|
doctest_namespace['np'] = numpy
|
|
|
|
|
|
|
|
For more details: :ref:`doctest_namespace`.
|
2018-02-27 07:27:15 +08:00
|
|
|
|
|
|
|
|
|
|
|
request
|
|
|
|
~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`request example`.
|
|
|
|
|
2018-02-27 07:27:15 +08:00
|
|
|
The ``request`` fixture is a special fixture providing information of the requesting test function.
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.fixtures.FixtureRequest()
|
|
|
|
:members:
|
2018-02-28 04:13:14 +08:00
|
|
|
|
|
|
|
|
|
|
|
pytestconfig
|
|
|
|
~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:19:57 +08:00
|
|
|
.. autofunction:: _pytest.fixtures.pytestconfig()
|
2018-02-28 04:13:14 +08:00
|
|
|
|
2018-02-28 04:19:57 +08:00
|
|
|
|
|
|
|
record_xml_property
|
|
|
|
~~~~~~~~~~~~~~~~~~~
|
|
|
|
|
2018-02-28 04:58:51 +08:00
|
|
|
**Tutorial**: :ref:`record_xml_property example`.
|
|
|
|
|
2018-02-28 04:19:57 +08:00
|
|
|
.. autofunction:: _pytest.junitxml.record_xml_property()
|
2018-02-28 04:32:34 +08:00
|
|
|
|
|
|
|
caplog
|
|
|
|
~~~~~~
|
|
|
|
|
2018-02-28 04:48:19 +08:00
|
|
|
**Tutorial**: :doc:`logging`.
|
|
|
|
|
2018-02-28 04:32:34 +08:00
|
|
|
.. autofunction:: _pytest.logging.caplog()
|
|
|
|
:no-auto-options:
|
|
|
|
|
2018-02-28 04:48:19 +08:00
|
|
|
This returns a :class:`_pytest.logging.LogCaptureFixture` instance.
|
2018-02-28 04:32:34 +08:00
|
|
|
|
|
|
|
.. autoclass:: _pytest.logging.LogCaptureFixture
|
|
|
|
:members:
|
2018-02-28 04:48:19 +08:00
|
|
|
|
|
|
|
|
|
|
|
monkeypatch
|
|
|
|
~~~~~~~~~~~
|
|
|
|
|
2018-02-28 05:03:35 +08:00
|
|
|
.. currentmodule:: _pytest.monkeypatch
|
|
|
|
|
2018-02-28 04:48:19 +08:00
|
|
|
**Tutorial**: :doc:`monkeypatch`.
|
|
|
|
|
|
|
|
.. autofunction:: _pytest.monkeypatch.monkeypatch()
|
|
|
|
:no-auto-options:
|
|
|
|
|
2018-02-28 05:03:35 +08:00
|
|
|
This returns a :class:`MonkeyPatch` instance.
|
2018-02-28 04:48:19 +08:00
|
|
|
|
|
|
|
.. autoclass:: _pytest.monkeypatch.MonkeyPatch
|
|
|
|
:members:
|
|
|
|
|
2018-02-28 05:03:35 +08:00
|
|
|
testdir
|
|
|
|
~~~~~~~
|
|
|
|
|
|
|
|
.. currentmodule:: _pytest.pytester
|
|
|
|
|
|
|
|
This fixture provides a :class:`Testdir` instance useful for black-box testing of test files, making it ideal to
|
|
|
|
test plugins.
|
|
|
|
|
|
|
|
To use it, include in your top-most ``conftest.py`` file::
|
|
|
|
|
|
|
|
pytest_plugins = 'pytester'
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.. autoclass:: Testdir()
|
|
|
|
:members: runpytest,runpytest_subprocess,runpytest_inprocess,makeconftest,makepyfile
|
|
|
|
|
|
|
|
.. autoclass:: RunResult()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
.. autoclass:: LineMatcher()
|
|
|
|
:members:
|
2018-02-28 05:26:40 +08:00
|
|
|
|
|
|
|
|
|
|
|
recwarn
|
|
|
|
~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`assertwarnings`
|
|
|
|
|
|
|
|
.. currentmodule:: _pytest.recwarn
|
|
|
|
|
|
|
|
.. autofunction:: recwarn()
|
|
|
|
:no-auto-options:
|
|
|
|
|
|
|
|
.. autoclass:: _pytest.recwarn.WarningsRecorder()
|
|
|
|
:members:
|
|
|
|
|
|
|
|
Each recorded warning is an instance of :class:`warnings.WarningMessage`.
|
|
|
|
|
|
|
|
.. note::
|
|
|
|
:class:`RecordedWarning` was changed from a plain class to a namedtuple in pytest 3.1
|
|
|
|
|
|
|
|
.. note::
|
|
|
|
``DeprecationWarning`` and ``PendingDeprecationWarning`` are treated
|
|
|
|
differently; see :ref:`ensuring_function_triggers`.
|
|
|
|
|
|
|
|
|
|
|
|
tmpdir
|
|
|
|
~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :doc:`tmpdir`
|
|
|
|
|
|
|
|
.. currentmodule:: _pytest.tmpdir
|
|
|
|
|
|
|
|
.. autofunction:: tmpdir()
|
|
|
|
:no-auto-options:
|
|
|
|
|
|
|
|
|
|
|
|
tmpdir_factory
|
|
|
|
~~~~~~~~~~~~~~
|
|
|
|
|
|
|
|
**Tutorial**: :ref:`tmpdir factory example`
|
|
|
|
|
|
|
|
.. _`tmpdir factory api`:
|
|
|
|
|
|
|
|
``tmpdir_factory`` instances have the following methods:
|
|
|
|
|
|
|
|
.. currentmodule:: _pytest.tmpdir
|
|
|
|
|
|
|
|
.. automethod:: TempdirFactory.mktemp
|
|
|
|
.. automethod:: TempdirFactory.getbasetemp
|