.package create: /tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/.package .package installdeps: flit_core >=2,<3 py36 create: /tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/py36 py36 installdeps: coverage, pytest==6.0.1 py36 inst: /tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/.tmp/package/1/pytest_check-0.3.9.tar.gz py36 installed: attrs==20.1.0,coverage==5.2.1,importlib-metadata==1.7.0,iniconfig==1.0.1,more-itertools==8.5.0,packaging==20.4,pluggy==0.13.1,py==1.9.0,pyparsing==2.4.7,pytest==6.0.1,pytest-check @ file:///tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/.tmp/package/1/pytest_check-0.3.9.tar.gz,six==1.15.0,toml==0.10.1,zipp==3.1.0 py36 run-test-pre: PYTHONHASHSEED='2252279102' py36 run-test: commands[0] | coverage run --parallel --source=/tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/py36/lib/python3.6/site-packages/pytest_check,tests -m pytest tests ============================= test session starts ============================== platform linux -- Python 3.6.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 cachedir: .tox/py36/.pytest_cache rootdir: /tmp/tmpem5tjc7u/pytest_check-0.3.9, configfile: tox.ini plugins: pytest_check-0.3.9 collected 35 items tests/test_check.py .................FF.... [ 65%] tests/test_check_context_manager.py ...... [ 82%] tests/test_check_errors.py FF [ 88%] tests/test_check_fixture.py . [ 91%] tests/test_check_func_decorator.py ... [100%] =================================== FAILURES =================================== _______________________________ test_check_xfail _______________________________ testdir = def test_check_xfail(testdir): testdir.makepyfile( """ import pytest_check as check import pytest @pytest.mark.xfail() def test_fail(): check.equal(1, 2) """ ) result = testdir.runpytest() > result.assert_outcomes(xfailed=1) E AssertionError: assert {'errors': 0,...pped': 0, ...} == {'errors': 0,...pped': 0, ...} E Omitting 4 identical items, use -vv to show E Differing items: E {'xfailed': 0} != {'xfailed': 1} E {'xpassed': 1} != {'xpassed': 0} E Use -v to get the full diff /tmp/tmpem5tjc7u/pytest_check-0.3.9/tests/test_check.py:178: AssertionError ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.6.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_check_xfail0 plugins: pytest_check-0.3.9 collected 1 item test_check_xfail.py X [100%] ============================== 1 xpassed in 0.10s ============================== ___________________________ test_check_xfail_strict ____________________________ testdir = def test_check_xfail_strict(testdir): testdir.makepyfile( """ import pytest_check as check import pytest @pytest.mark.xfail(strict=True) def test_fail(): check.equal(1, 2) """ ) result = testdir.runpytest() > result.assert_outcomes(xfailed=1) E AssertionError: assert {'errors': 0,...pped': 0, ...} == {'errors': 0,...pped': 0, ...} E Omitting 4 identical items, use -vv to show E Differing items: E {'xfailed': 0} != {'xfailed': 1} E {'failed': 1} != {'failed': 0} E Use -v to get the full diff /tmp/tmpem5tjc7u/pytest_check-0.3.9/tests/test_check.py:194: AssertionError ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.6.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_check_xfail_strict0 plugins: pytest_check-0.3.9 collected 1 item test_check_xfail_strict.py F [100%] =================================== FAILURES =================================== __________________________________ test_fail ___________________________________ [XPASS(strict)] =========================== short test summary info ============================ FAILED test_check_xfail_strict.py::test_fail ============================== 1 failed in 0.11s =============================== ______________________________ test_setup_failure ______________________________ testdir = def test_setup_failure(testdir): testdir.makepyfile( """ import pytest import pytest_check as check @pytest.fixture() def a_fixture(): check.equal(1, 2) def test_1(a_fixture): pass """ ) result = testdir.runpytest() > result.assert_outcomes(error=1) E TypeError: assert_outcomes() got an unexpected keyword argument 'error' /tmp/tmpem5tjc7u/pytest_check-0.3.9/tests/test_check_errors.py:16: TypeError ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.6.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_setup_failure0 plugins: pytest_check-0.3.9 collected 1 item test_setup_failure.py E [100%] ==================================== ERRORS ==================================== ___________________________ ERROR at setup of test_1 ___________________________ FAILURE: assert 1 == 2 test_setup_failure.py:6 in a_fixture() -> check.equal(1, 2) ------------------------------------------------------------ Failed Checks: 1 =========================== short test summary info ============================ ERROR test_setup_failure.py::test_1 =============================== 1 error in 0.13s =============================== ____________________________ test_teardown_failure _____________________________ testdir = def test_teardown_failure(testdir): testdir.makepyfile( """ import pytest import pytest_check as check @pytest.fixture() def a_fixture(): yield check.equal(1, 2) def test_1(a_fixture): pass """ ) result = testdir.runpytest() > result.assert_outcomes(passed=1, error=1) E TypeError: assert_outcomes() got an unexpected keyword argument 'error' /tmp/tmpem5tjc7u/pytest_check-0.3.9/tests/test_check_errors.py:36: TypeError ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.6.7, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_teardown_failure0 plugins: pytest_check-0.3.9 collected 1 item test_teardown_failure.py .E [100%] ==================================== ERRORS ==================================== _________________________ ERROR at teardown of test_1 __________________________ FAILURE: assert 1 == 2 test_teardown_failure.py:7 in a_fixture() -> check.equal(1, 2) ------------------------------------------------------------ Failed Checks: 1 =========================== short test summary info ============================ ERROR test_teardown_failure.py::test_1 ========================== 1 passed, 1 error in 0.12s ========================== =========================== short test summary info ============================ FAILED tests/test_check.py::test_check_xfail - AssertionError: assert {'error... FAILED tests/test_check.py::test_check_xfail_strict - AssertionError: assert ... FAILED tests/test_check_errors.py::test_setup_failure - TypeError: assert_out... FAILED tests/test_check_errors.py::test_teardown_failure - TypeError: assert_... ========================= 4 failed, 31 passed in 8.21s ========================= ERROR: InvocationError for command /tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/py36/bin/coverage run --parallel --source=/tmp/tmpem5tjc7u/pytest_check-0.3.9/.tox/py36/lib/python3.6/site-packages/pytest_check,tests -m pytest tests (exited with code 1) ___________________________________ summary ____________________________________ ERROR: py36: commands failed write json report at: /tmp/tmpem5tjc7u/pytest_check-0.3.9/result.json Time: 65.9 seconds