GLOB sdist-make: /tmp/tmprhq4wk8a/pytest-cache-1.0/setup.py py38 create: /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38 py38 installdeps: pytest==6.0.1 py38 inst: /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/.tmp/package/1/pytest-cache-1.0.zip py38 installed: apipkg==1.5,attrs==20.1.0,execnet==1.7.1,iniconfig==1.0.1,more-itertools==8.5.0,packaging==20.4,pluggy==0.13.1,py==1.9.0,pyparsing==2.4.7,pytest==6.0.1,pytest-cache @ file:///tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/.tmp/package/1/pytest-cache-1.0.zip,six==1.15.0,toml==0.10.1 py38 run-test-pre: PYTHONHASHSEED='137068555' py38 run-test: commands[0] | py.test --junitxml=/tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/log/junit-py38.xml test_cache.py ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 cachedir: .tox/py38/.pytest_cache rootdir: /tmp/tmprhq4wk8a/pytest-cache-1.0, configfile: tox.ini collected 11 items test_cache.py .FF.F.FF..F [100%] =================================== FAILURES =================================== ___________________________ test_cache_reportheader ____________________________ testdir = def test_cache_reportheader(testdir): p = testdir.makepyfile(""" def test_hello(): pass """) cachedir = p.dirpath(".cache") result = testdir.runpytest("-v") > result.stdout.fnmatch_lines([ "cachedir: %s" % cachedir, ]) E Failed: nomatch: 'cachedir: /tmp/pytest-of-travis/pytest-4/test_cache_reportheader0/.cache' E and: '============================= test session starts ==============================' E and: 'platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 -- /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/bin/python' E and: 'cachedir: .pytest_cache' E and: 'rootdir: /tmp/pytest-of-travis/pytest-4/test_cache_reportheader0' E and: 'collecting ... collected 1 item' E and: '' E and: 'test_cache_reportheader.py::test_hello PASSED [100%]' E and: '' E and: '============================== 1 passed in 0.03s ===============================' E remains unmatched: 'cachedir: /tmp/pytest-of-travis/pytest-4/test_cache_reportheader0/.cache' /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:19: Failed ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 -- /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/bin/python cachedir: .pytest_cache rootdir: /tmp/pytest-of-travis/pytest-4/test_cache_reportheader0 collecting ... collected 1 item test_cache_reportheader.py::test_hello PASSED [100%] ============================== 1 passed in 0.03s =============================== _______________________________ test_cache_show ________________________________ testdir = def test_cache_show(testdir): result = testdir.runpytest("--cache") > assert result.ret == 0 E assert == 0 E + where = .ret /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:25: AssertionError ----------------------------- Captured stderr call ----------------------------- ERROR: usage: py.test [options] [file_or_dir] [file_or_dir] [...] py.test: error: unrecognized arguments: --cache inifile: None rootdir: /tmp/pytest-of-travis/pytest-4/test_cache_show0 ____________________ TestNewAPI.test_config_cache_dataerror ____________________ self = testdir = def test_config_cache_dataerror(self, testdir): testdir.makeini("[pytest]") config = testdir.parseconfigure() cache = config.cache > pytest.raises(ValueError, lambda: cache.set("key/name", cache)) /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:66: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:66: in pytest.raises(ValueError, lambda: cache.set("key/name", cache)) /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/lib/python3.8/site-packages/_pytest/cacheprovider.py:155: in set data = json.dumps(value, indent=2, sort_keys=True) /opt/python/3.8.0/lib/python3.8/json/__init__.py:234: in dumps return cls( /opt/python/3.8.0/lib/python3.8/json/encoder.py:201: in encode chunks = list(chunks) /opt/python/3.8.0/lib/python3.8/json/encoder.py:438: in _iterencode o = _default(o) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = , o = Cache() def default(self, o): """Implement this method in a subclass such that it returns a serializable object for ``o``, or calls the base implementation (to raise a ``TypeError``). For example, to support arbitrary iterators, you could implement default like this:: def default(self, o): try: iterable = iter(o) except TypeError: pass else: return list(iterable) # Let the base class default method raise the TypeError return JSONEncoder.default(self, o) """ > raise TypeError(f'Object of type {o.__class__.__name__} ' f'is not JSON serializable') E TypeError: Object of type Cache is not JSON serializable /opt/python/3.8.0/lib/python3.8/json/encoder.py:179: TypeError ____________________ TestLastFailed.test_lastfailed_usecase ____________________ self = testdir = monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x7f7c97f02730> @pytest.mark.skipif("sys.version_info < (2,6)") def test_lastfailed_usecase(self, testdir, monkeypatch): monkeypatch.setenv("PYTHONDONTWRITEBYTECODE", 1) p = testdir.makepyfile(""" def test_1(): assert 0 def test_2(): assert 0 def test_3(): assert 1 """) result = testdir.runpytest() result.stdout.fnmatch_lines([ "*2 failed*", ]) p.write(py.code.Source(""" def test_1(): assert 1 def test_2(): assert 1 def test_3(): assert 0 """)) result = testdir.runpytest("--lf") > result.stdout.fnmatch_lines([ "*2 passed*1 desel*", ]) E Failed: nomatch: '*2 passed*1 desel*' E and: '============================= test session starts ==============================' E and: 'platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1' E and: 'rootdir: /tmp/pytest-of-travis/pytest-4/test_lastfailed_usecase0' E and: 'collected 2 items' E and: 'run-last-failure: rerun previous 2 failures' E and: '' E and: 'test_lastfailed_usecase.py .. [100%]' E and: '' E and: '============================== 2 passed in 0.05s ===============================' E remains unmatched: '*2 passed*1 desel*' /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:129: Failed ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_lastfailed_usecase0 collected 3 items test_lastfailed_usecase.py FF. [100%] =================================== FAILURES =================================== ____________________________________ test_1 ____________________________________ def test_1(): > assert 0 E assert 0 test_lastfailed_usecase.py:2: AssertionError ____________________________________ test_2 ____________________________________ def test_2(): > assert 0 E assert 0 test_lastfailed_usecase.py:4: AssertionError =========================== short test summary info ============================ FAILED test_lastfailed_usecase.py::test_1 - assert 0 FAILED test_lastfailed_usecase.py::test_2 - assert 0 ========================= 2 failed, 1 passed in 0.07s ========================== ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_lastfailed_usecase0 collected 2 items run-last-failure: rerun previous 2 failures test_lastfailed_usecase.py .. [100%] ============================== 2 passed in 0.05s =============================== ____________________ TestLastFailed.test_failedfirst_order _____________________ self = testdir = def test_failedfirst_order(self, testdir): always_pass = testdir.tmpdir.join('test_a.py').write(py.code.Source(""" def test_always_passes(): assert 1 """)) always_fail = testdir.tmpdir.join('test_b.py').write(py.code.Source(""" def test_always_fails(): assert 0 """)) result = testdir.runpytest() # Test order will be collection order; alphabetical result.stdout.fnmatch_lines([ "test_a.py*", "test_b.py*", ]) result = testdir.runpytest("--lf", "--ff") # Test order will be failing tests firs > result.stdout.fnmatch_lines([ "test_b.py*", "test_a.py*", ]) E Failed: nomatch: 'test_b.py*' E and: '============================= test session starts ==============================' E and: 'platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1' E and: 'rootdir: /tmp/pytest-of-travis/pytest-4/test_failedfirst_order0' E and: 'collected 1 item' E and: 'run-last-failure: rerun previous 1 failure first (skipped 1 file)' E and: '' E fnmatch: 'test_b.py*' E with: 'test_b.py F [100%]' E nomatch: 'test_a.py*' E and: '' E and: '=================================== FAILURES ===================================' E and: '______________________________ test_always_fails _______________________________' E and: '' E and: ' def test_always_fails():' E and: '> assert 0' E and: 'E assert 0' E and: '' E and: 'test_b.py:3: AssertionError' E and: '=========================== short test summary info ============================' E and: 'FAILED test_b.py::test_always_fails - assert 0' E and: '============================== 1 failed in 0.05s ===============================' E remains unmatched: 'test_a.py*' /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:166: Failed ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_failedfirst_order0 collected 2 items test_a.py . [ 50%] test_b.py F [100%] =================================== FAILURES =================================== ______________________________ test_always_fails _______________________________ def test_always_fails(): > assert 0 E assert 0 test_b.py:3: AssertionError =========================== short test summary info ============================ FAILED test_b.py::test_always_fails - assert 0 ========================= 1 failed, 1 passed in 0.08s ========================== ============================= test session starts ============================== platform linux -- Python 3.8.0, pytest-6.0.1, py-1.9.0, pluggy-0.13.1 rootdir: /tmp/pytest-of-travis/pytest-4/test_failedfirst_order0 collected 1 item run-last-failure: rerun previous 1 failure first (skipped 1 file) test_b.py F [100%] =================================== FAILURES =================================== ______________________________ test_always_fails _______________________________ def test_always_fails(): > assert 0 E assert 0 test_b.py:3: AssertionError =========================== short test summary info ============================ FAILED test_b.py::test_always_fails - assert 0 ============================== 1 failed in 0.05s =============================== _____________________ TestLastFailed.test_lastfailed_xpass _____________________ self = testdir = def test_lastfailed_xpass(self, testdir): > rep = testdir.inline_runsource1(""" import pytest @pytest.mark.xfail def test_hello(): assert 1 """) E AttributeError: 'Testdir' object has no attribute 'inline_runsource1' /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:233: AttributeError =============================== warnings summary =============================== .tox/py38/lib/python3.8/site-packages/_pytest/config/__init__.py:1148 /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/lib/python3.8/site-packages/_pytest/config/__init__.py:1148: PytestConfigWarning: Unknown config ini key: pep8ignore self._warn_or_fail_if_strict("Unknown config ini key: {}\n".format(key)) .tox/py38/lib/python3.8/site-packages/_pytest/junitxml.py:446 /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/lib/python3.8/site-packages/_pytest/junitxml.py:446: PytestDeprecationWarning: The 'junit_family' default value will change to 'xunit2' in pytest 6.0. See: https://docs.pytest.org/en/stable/deprecations.html#junit-family-default-value-change-to-xunit2 for more information. _issue_warning_captured(deprecated.JUNIT_XML_DEFAULT_FAMILY, config.hook, 2) test_cache.py::TestLastFailed::test_lastfailed_usecase /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:105: PytestWarning: Value of environment variable PYTHONDONTWRITEBYTECODE type should be str, but got 1 (type: int); converted to str implicitly monkeypatch.setenv("PYTHONDONTWRITEBYTECODE", 1) test_cache.py::TestLastFailed::test_lastfailed_difference_invocations /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:173: PytestWarning: Value of environment variable PYTHONDONTWRITEBYTECODE type should be str, but got 1 (type: int); converted to str implicitly monkeypatch.setenv("PYTHONDONTWRITEBYTECODE", 1) test_cache.py::TestLastFailed::test_lastfailed_usecase_splice /tmp/tmprhq4wk8a/pytest-cache-1.0/test_cache.py:209: PytestWarning: Value of environment variable PYTHONDONTWRITEBYTECODE type should be str, but got 1 (type: int); converted to str implicitly monkeypatch.setenv("PYTHONDONTWRITEBYTECODE", 1) -- Docs: https://docs.pytest.org/en/stable/warnings.html - generated xml file: /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/log/junit-py38.xml - =========================== short test summary info ============================ FAILED test_cache.py::test_cache_reportheader - Failed: nomatch: 'cachedir: /... FAILED test_cache.py::test_cache_show - assert == 0 FAILED test_cache.py::TestNewAPI::test_config_cache_dataerror - TypeError: Ob... FAILED test_cache.py::TestLastFailed::test_lastfailed_usecase - Failed: nomat... FAILED test_cache.py::TestLastFailed::test_failedfirst_order - Failed: nomatc... FAILED test_cache.py::TestLastFailed::test_lastfailed_xpass - AttributeError:... =================== 6 failed, 5 passed, 5 warnings in 6.17s ==================== ERROR: InvocationError for command /tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/bin/py.test --junitxml=/tmp/tmprhq4wk8a/pytest-cache-1.0/.tox/py38/log/junit-py38.xml test_cache.py (exited with code 1) ___________________________________ summary ____________________________________ ERROR: py38: commands failed write json report at: /tmp/tmprhq4wk8a/pytest-cache-1.0/result.json Time: 35.3 seconds