GLOB sdist-make: /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/setup.py py37 create: /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37 py37 installdeps: pytest==5.2.4 py37 inst: /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/.tmp/package/1/pytest-blink1-0.0.8.zip py37 installed: atomicwrites==1.3.0,attrs==19.3.0,importlib-metadata==1.1.0,more-itertools==8.0.0,packaging==19.2,pluggy==0.13.1,py==1.8.0,pyparsing==2.4.5,pytest==5.2.4,pytest-blink1==0.0.8,six==1.13.0,wcwidth==0.1.7,zipp==0.6.0 py37 run-test-pre: PYTHONHASHSEED='1172066338' py37 run-test: commands[0] | py.test tests ============================= test session starts ============================== platform linux -- Python 3.7.1, pytest-5.2.4, py-1.8.0, pluggy-0.13.1 cachedir: .tox/py37/.pytest_cache rootdir: /tmp/tmptqkn_kiv/pytest-blink1-0.0.8 plugins: blink1-0.0.8 collected 3 items tests/test_blink1.py FFF [100%] =================================== FAILURES =================================== _______________________________ test_bar_fixture _______________________________ testdir = def test_bar_fixture(testdir): """Make sure that pytest accepts our fixture.""" # create a temporary pytest test module testdir.makepyfile(""" def test_sth(bar): assert bar == "europython2015" """) # run pytest with the following cmd args result = testdir.runpytest( '--foo=europython2015', '-v' ) # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ > '*::test_sth PASSED', ]) E Failed: nomatch: '*::test_sth PASSED' E and: '' E remains unmatched: '*::test_sth PASSED' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/tests/test_blink1.py:21: Failed ----------------------------- Captured stderr call ----------------------------- ERROR: usage: py.test [options] [file_or_dir] [file_or_dir] [...] py.test: error: unrecognized arguments: --foo=europython2015 inifile: None rootdir: /tmp/pytest-of-travis/pytest-1/test_bar_fixture0 ______________________________ test_help_message _______________________________ testdir = def test_help_message(testdir): result = testdir.runpytest( '--help', ) # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ 'blink1:', > '*--foo=DEST_FOO*Set the value for the fixture "bar".', ]) E Failed: nomatch: 'blink1:' E and: 'usage: py.test [options] [file_or_dir] [file_or_dir] [...]' E and: '' E and: 'positional arguments:' E and: ' file_or_dir' E and: '' E and: 'general:' E and: ' -k EXPRESSION only run tests which match the given substring' E and: ' expression. An expression is a python evaluatable' E and: ' expression where all names are substring-matched against' E and: ' test names and their parent classes. Example: -k' E and: " 'test_method or test_other' matches all test functions" E and: " and classes whose name contains 'test_method' or" E and: " 'test_other', while -k 'not test_method' matches those" E and: " that don't contain 'test_method' in their names. -k 'not" E and: " test_method and not test_other' will eliminate the" E and: ' matches. Additionally keywords are matched to classes' E and: ' and functions containing extra names in their' E and: " 'extra_keyword_matches' set, as well as functions which" E and: ' have names assigned directly to them.' E and: ' -m MARKEXPR only run tests matching given mark expression. example:' E and: " -m 'mark1 and not mark2'." E and: ' --markers show markers (builtin, plugin and per-project ones).' E and: ' -x, --exitfirst exit instantly on first error or failed test.' E and: ' --maxfail=num exit after first num failures or errors.' E and: ' --strict-markers, --strict' E and: ' markers not registered in the `markers` section of the' E and: ' configuration file raise errors.' E and: ' -c file load configuration from `file` instead of trying to' E and: ' locate one of the implicit configuration files.' E and: ' --continue-on-collection-errors' E and: ' Force test execution even if collection errors occur.' E and: ' --rootdir=ROOTDIR Define root directory for tests. Can be relative path:' E and: " 'root_dir', './root_dir', 'root_dir/another_dir/';" E and: " absolute path: '/home/user/root_dir'; path with" E and: " variables: '$HOME/root_dir'." E and: ' --fixtures, --funcargs' E and: ' show available fixtures, sorted by plugin appearance' E and: " (fixtures with leading '_' are only shown with '-v')" E and: ' --fixtures-per-test show fixtures per test' E and: ' --import-mode={prepend,append}' E and: ' prepend/append to sys.path when importing test modules,' E and: ' default is to prepend.' E and: ' --pdb start the interactive Python debugger on errors or' E and: ' KeyboardInterrupt.' E and: ' --pdbcls=modulename:classname' E and: ' start a custom interactive Python debugger on errors.' E and: ' For example:' E and: ' --pdbcls=IPython.terminal.debugger:TerminalPdb' E and: ' --trace Immediately break when running each test.' E and: ' --capture=method per-test capturing method: one of fd|sys|no.' E and: ' -s shortcut for --capture=no.' E and: ' --runxfail report the results of xfail tests as if they were not' E and: ' marked' E and: ' --lf, --last-failed rerun only the tests that failed at the last run (or all' E and: ' if none failed)' E and: ' --ff, --failed-first run all tests but run the last failures first. This may' E and: ' re-order tests and thus lead to repeated fixture' E and: ' setup/teardown' E and: ' --nf, --new-first run tests from new files first, then the rest of the' E and: ' tests sorted by file mtime' E and: ' --cache-show=[CACHESHOW]' E and: " show cache contents, don't perform collection or tests." E and: " Optional argument: glob (default: '*')." E and: ' --cache-clear remove all cache contents at start of test run.' E and: ' --lfnf={all,none}, --last-failed-no-failures={all,none}' E and: ' which tests to run with no previously (known) failures.' E and: ' --sw, --stepwise exit on test failure and continue from last failing test' E and: ' next time' E and: ' --stepwise-skip ignore the first failing test but stop on the next' E and: ' failing test' E and: '' E and: 'reporting:' E and: ' --durations=N show N slowest setup/test durations (N=0 for all).' E and: ' -v, --verbose increase verbosity.' E and: ' -q, --quiet decrease verbosity.' E and: ' --verbosity=VERBOSE set verbosity' E and: ' -r chars show extra test summary info as specified by chars:' E and: ' (f)ailed, (E)rror, (s)kipped, (x)failed, (X)passed,' E and: ' (p)assed, (P)assed with output, (a)ll except passed' E and: ' (p/P), or (A)ll. (w)arnings are enabled by default (see' E and: ' --disable-warnings).' E and: ' --disable-warnings, --disable-pytest-warnings' E and: ' disable warnings summary' E and: ' -l, --showlocals show locals in tracebacks (disabled by default).' E and: ' --tb=style traceback print mode (auto/long/short/line/native/no).' E and: ' --show-capture={no,stdout,stderr,log,all}' E and: ' Controls how captured stdout/stderr/log is shown on' E and: " failed tests. Default is 'all'." E and: " --full-trace don't cut any tracebacks (default is to cut)." E and: ' --color=color color terminal output (yes/no/auto).' E and: ' --pastebin=mode send failed|all info to bpaste.net pastebin service.' E and: ' --junit-xml=path create junit-xml style report file at given path.' E and: ' --junit-prefix=str prepend prefix to classnames in junit-xml output' E and: ' --result-log=path DEPRECATED path for machine-readable result log.' E and: '' E and: 'collection:' E and: " --collect-only only collect tests, don't execute them." E and: ' --pyargs try to interpret all arguments as python packages.' E and: ' --ignore=path ignore path during collection (multi-allowed).' E and: ' --ignore-glob=path ignore path pattern during collection (multi-allowed).' E and: ' --deselect=nodeid_prefix' E and: ' deselect item during collection (multi-allowed).' E and: " --confcutdir=dir only load conftest.py's relative to specified dir." E and: " --noconftest Don't load any conftest.py files." E and: ' --keep-duplicates Keep duplicate tests.' E and: ' --collect-in-virtualenv' E and: " Don't ignore tests in a local virtualenv directory" E and: ' --doctest-modules run doctests in all .py modules' E and: ' --doctest-report={none,cdiff,ndiff,udiff,only_first_failure}' E and: ' choose another output format for diffs on doctest' E and: ' failure' E and: ' --doctest-glob=pat doctests file matching pattern, default: test*.txt' E and: ' --doctest-ignore-import-errors' E and: ' ignore doctest ImportErrors' E and: ' --doctest-continue-on-failure' E and: ' for a given doctest, continue to run after the first' E and: ' failure' E and: '' E and: 'test session debugging and configuration:' E and: ' --basetemp=dir base temporary directory for this test run.(warning:' E and: ' this directory is removed if it exists)' E and: ' --version display pytest lib version and import information.' E and: ' -h, --help show help message and configuration info' E and: ' -p name early-load given plugin module name or entry point' E and: ' (multi-allowed). To avoid loading of plugins, use the' E and: ' `no:` prefix, e.g. `no:doctest`.' E and: ' --trace-config trace considerations of conftest.py files.' E and: ' --debug store internal tracing debug information in' E and: " 'pytestdebug.log'." E and: ' -o OVERRIDE_INI, --override-ini=OVERRIDE_INI' E and: ' override ini option with "option=value" style, e.g. `-o' E and: ' xfail_strict=True -o cache_dir=cache`.' E and: " --assert=MODE Control assertion debugging tools. 'plain' performs no" E and: " assertion debugging. 'rewrite' (the default) rewrites" E and: ' assert statements in test modules on import to provide' E and: ' assert expression information.' E and: ' --setup-only only setup fixtures, do not execute tests.' E and: ' --setup-show show setup of fixtures while executing tests.' E and: " --setup-plan show what fixtures and tests would be executed but don't" E and: ' execute anything.' E and: '' E and: 'pytest-warnings:' E and: ' -W PYTHONWARNINGS, --pythonwarnings=PYTHONWARNINGS' E and: ' set which warnings to report, see -W option of python' E and: ' itself.' E and: '' E and: 'logging:' E and: ' --no-print-logs disable printing caught logs on failed tests.' E and: ' --log-level=LOG_LEVEL' E and: ' logging level used by the logging module' E and: ' --log-format=LOG_FORMAT' E and: ' log format as used by the logging module.' E and: ' --log-date-format=LOG_DATE_FORMAT' E and: ' log date format as used by the logging module.' E and: ' --log-cli-level=LOG_CLI_LEVEL' E and: ' cli logging level.' E and: ' --log-cli-format=LOG_CLI_FORMAT' E and: ' log format as used by the logging module.' E and: ' --log-cli-date-format=LOG_CLI_DATE_FORMAT' E and: ' log date format as used by the logging module.' E and: ' --log-file=LOG_FILE path to a file when logging will be written to.' E and: ' --log-file-level=LOG_FILE_LEVEL' E and: ' log file logging level.' E and: ' --log-file-format=LOG_FILE_FORMAT' E and: ' log format as used by the logging module.' E and: ' --log-file-date-format=LOG_FILE_DATE_FORMAT' E and: ' log date format as used by the logging module.' E and: '' E and: '[pytest] ini-options in the first pytest.ini|tox.ini|setup.cfg file found:' E and: '' E and: ' markers (linelist): markers for test functions' E and: ' empty_parameter_set_mark (string):' E and: ' default marker for empty parametersets' E and: ' norecursedirs (args): directory patterns to avoid for recursion' E and: ' testpaths (args): directories to search for tests when no files or' E and: ' directories are given in the command line.' E and: ' usefixtures (args): list of default fixtures to be used with this project' E and: ' python_files (args): glob-style file patterns for Python test module' E and: ' discovery' E and: ' python_classes (args):' E and: ' prefixes or glob names for Python test class discovery' E and: ' python_functions (args):' E and: ' prefixes or glob names for Python test function and' E and: ' method discovery' E and: ' disable_test_id_escaping_and_forfeit_all_rights_to_community_support (bool):' E and: ' disable string escape non-ascii characters, might cause' E and: ' unwanted side effects(use at your own risk)' E and: ' console_output_style (string):' E and: ' console output: "classic", or with additional progress' E and: ' information ("progress" (percentage) | "count").' E and: ' xfail_strict (bool): default for the strict parameter of xfail markers when' E and: ' not given explicitly (default: False)' E and: ' enable_assertion_pass_hook (bool):' E and: ' Enables the pytest_assertion_pass hook.Make sure to' E and: ' delete any previously generated pyc cache files.' E and: ' junit_suite_name (string):' E and: ' Test suite name for JUnit report' E and: ' junit_logging (string):' E and: ' Write captured log messages to JUnit report: one of' E and: ' no|system-out|system-err' E and: ' junit_log_passing_tests (bool):' E and: ' Capture log information for passing tests to JUnit' E and: ' report:' E and: ' junit_duration_report (string):' E and: ' Duration time to report: one of total|call' E and: ' junit_family (string):' E and: ' Emit XML for schema: one of legacy|xunit1|xunit2' E and: ' doctest_optionflags (args):' E and: ' option flags for doctests' E and: ' doctest_encoding (string):' E and: ' encoding used for doctest files' E and: ' cache_dir (string): cache directory path.' E and: ' filterwarnings (linelist):' E and: ' Each line specifies a pattern for' E and: ' warnings.filterwarnings. Processed after -W and' E and: ' --pythonwarnings.' E and: ' log_print (bool): default value for --no-print-logs' E and: ' log_level (string): default value for --log-level' E and: ' log_format (string): default value for --log-format' E and: ' log_date_format (string):' E and: ' default value for --log-date-format' E and: ' log_cli (bool): enable log display during test run (also known as "live' E and: ' logging").' E and: ' log_cli_level (string):' E and: ' default value for --log-cli-level' E and: ' log_cli_format (string):' E and: ' default value for --log-cli-format' E and: ' log_cli_date_format (string):' E and: ' default value for --log-cli-date-format' E and: ' log_file (string): default value for --log-file' E and: ' log_file_level (string):' E and: ' default value for --log-file-level' E and: ' log_file_format (string):' E and: ' default value for --log-file-format' E and: ' log_file_date_format (string):' E and: ' default value for --log-file-date-format' E and: ' faulthandler_timeout (string):' E and: ' Dump the traceback of all threads if a test takes more' E and: ' than TIMEOUT seconds to finish. Not available on' E and: ' Windows.' E and: ' addopts (args): extra command line options' E and: ' minversion (string): minimally required pytest version' E and: '' E and: 'environment variables:' E and: ' PYTEST_ADDOPTS extra command line options' E and: ' PYTEST_PLUGINS comma-separated plugins to load during startup' E and: ' PYTEST_DISABLE_PLUGIN_AUTOLOAD set to disable plugin auto-loading' E and: " PYTEST_DEBUG set to enable debug tracing of pytest's internals" E and: '' E and: '' E and: 'to see available markers type: pytest --markers' E and: 'to see available fixtures type: pytest --fixtures' E and: "(shown according to specified file_or_dir or current dir if not specified; fixtures with leading '_' are only shown with the '-v' option" E and: '' E remains unmatched: 'blink1:' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/tests/test_blink1.py:35: Failed ----------------------------- Captured stdout call ----------------------------- usage: py.test [options] [file_or_dir] [file_or_dir] [...] positional arguments: file_or_dir general: -k EXPRESSION only run tests which match the given substring expression. An expression is a python evaluatable expression where all names are substring-matched against test names and their parent classes. Example: -k 'test_method or test_other' matches all test functions and classes whose name contains 'test_method' or 'test_other', while -k 'not test_method' matches those that don't contain 'test_method' in their names. -k 'not test_method and not test_other' will eliminate the matches. Additionally keywords are matched to classes and functions containing extra names in their 'extra_keyword_matches' set, as well as functions which have names assigned directly to them. -m MARKEXPR only run tests matching given mark expression. example: -m 'mark1 and not mark2'. --markers show markers (builtin, plugin and per-project ones). -x, --exitfirst exit instantly on first error or failed test. --maxfail=num exit after first num failures or errors. --strict-markers, --strict markers not registered in the `markers` section of the configuration file raise errors. -c file load configuration from `file` instead of trying to locate one of the implicit configuration files. --continue-on-collection-errors Force test execution even if collection errors occur. --rootdir=ROOTDIR Define root directory for tests. Can be relative path: 'root_dir', './root_dir', 'root_dir/another_dir/'; absolute path: '/home/user/root_dir'; path with variables: '$HOME/root_dir'. --fixtures, --funcargs show available fixtures, sorted by plugin appearance (fixtures with leading '_' are only shown with '-v') --fixtures-per-test show fixtures per test --import-mode={prepend,append} prepend/append to sys.path when importing test modules, default is to prepend. --pdb start the interactive Python debugger on errors or KeyboardInterrupt. --pdbcls=modulename:classname start a custom interactive Python debugger on errors. For example: --pdbcls=IPython.terminal.debugger:TerminalPdb --trace Immediately break when running each test. --capture=method per-test capturing method: one of fd|sys|no. -s shortcut for --capture=no. --runxfail report the results of xfail tests as if they were not marked --lf, --last-failed rerun only the tests that failed at the last run (or all if none failed) --ff, --failed-first run all tests but run the last failures first. This may re-order tests and thus lead to repeated fixture setup/teardown --nf, --new-first run tests from new files first, then the rest of the tests sorted by file mtime --cache-show=[CACHESHOW] show cache contents, don't perform collection or tests. Optional argument: glob (default: '*'). --cache-clear remove all cache contents at start of test run. --lfnf={all,none}, --last-failed-no-failures={all,none} which tests to run with no previously (known) failures. --sw, --stepwise exit on test failure and continue from last failing test next time --stepwise-skip ignore the first failing test but stop on the next failing test reporting: --durations=N show N slowest setup/test durations (N=0 for all). -v, --verbose increase verbosity. -q, --quiet decrease verbosity. --verbosity=VERBOSE set verbosity -r chars show extra test summary info as specified by chars: (f)ailed, (E)rror, (s)kipped, (x)failed, (X)passed, (p)assed, (P)assed with output, (a)ll except passed (p/P), or (A)ll. (w)arnings are enabled by default (see --disable-warnings). --disable-warnings, --disable-pytest-warnings disable warnings summary -l, --showlocals show locals in tracebacks (disabled by default). --tb=style traceback print mode (auto/long/short/line/native/no). --show-capture={no,stdout,stderr,log,all} Controls how captured stdout/stderr/log is shown on failed tests. Default is 'all'. --full-trace don't cut any tracebacks (default is to cut). --color=color color terminal output (yes/no/auto). --pastebin=mode send failed|all info to bpaste.net pastebin service. --junit-xml=path create junit-xml style report file at given path. --junit-prefix=str prepend prefix to classnames in junit-xml output --result-log=path DEPRECATED path for machine-readable result log. collection: --collect-only only collect tests, don't execute them. --pyargs try to interpret all arguments as python packages. --ignore=path ignore path during collection (multi-allowed). --ignore-glob=path ignore path pattern during collection (multi-allowed). --deselect=nodeid_prefix deselect item during collection (multi-allowed). --confcutdir=dir only load conftest.py's relative to specified dir. --noconftest Don't load any conftest.py files. --keep-duplicates Keep duplicate tests. --collect-in-virtualenv Don't ignore tests in a local virtualenv directory --doctest-modules run doctests in all .py modules --doctest-report={none,cdiff,ndiff,udiff,only_first_failure} choose another output format for diffs on doctest failure --doctest-glob=pat doctests file matching pattern, default: test*.txt --doctest-ignore-import-errors ignore doctest ImportErrors --doctest-continue-on-failure for a given doctest, continue to run after the first failure test session debugging and configuration: --basetemp=dir base temporary directory for this test run.(warning: this directory is removed if it exists) --version display pytest lib version and import information. -h, --help show help message and configuration info -p name early-load given plugin module name or entry point (multi-allowed). To avoid loading of plugins, use the `no:` prefix, e.g. `no:doctest`. --trace-config trace considerations of conftest.py files. --debug store internal tracing debug information in 'pytestdebug.log'. -o OVERRIDE_INI, --override-ini=OVERRIDE_INI override ini option with "option=value" style, e.g. `-o xfail_strict=True -o cache_dir=cache`. --assert=MODE Control assertion debugging tools. 'plain' performs no assertion debugging. 'rewrite' (the default) rewrites assert statements in test modules on import to provide assert expression information. --setup-only only setup fixtures, do not execute tests. --setup-show show setup of fixtures while executing tests. --setup-plan show what fixtures and tests would be executed but don't execute anything. pytest-warnings: -W PYTHONWARNINGS, --pythonwarnings=PYTHONWARNINGS set which warnings to report, see -W option of python itself. logging: --no-print-logs disable printing caught logs on failed tests. --log-level=LOG_LEVEL logging level used by the logging module --log-format=LOG_FORMAT log format as used by the logging module. --log-date-format=LOG_DATE_FORMAT log date format as used by the logging module. --log-cli-level=LOG_CLI_LEVEL cli logging level. --log-cli-format=LOG_CLI_FORMAT log format as used by the logging module. --log-cli-date-format=LOG_CLI_DATE_FORMAT log date format as used by the logging module. --log-file=LOG_FILE path to a file when logging will be written to. --log-file-level=LOG_FILE_LEVEL log file logging level. --log-file-format=LOG_FILE_FORMAT log format as used by the logging module. --log-file-date-format=LOG_FILE_DATE_FORMAT log date format as used by the logging module. [pytest] ini-options in the first pytest.ini|tox.ini|setup.cfg file found: markers (linelist): markers for test functions empty_parameter_set_mark (string): default marker for empty parametersets norecursedirs (args): directory patterns to avoid for recursion testpaths (args): directories to search for tests when no files or directories are given in the command line. usefixtures (args): list of default fixtures to be used with this project python_files (args): glob-style file patterns for Python test module discovery python_classes (args): prefixes or glob names for Python test class discovery python_functions (args): prefixes or glob names for Python test function and method discovery disable_test_id_escaping_and_forfeit_all_rights_to_community_support (bool): disable string escape non-ascii characters, might cause unwanted side effects(use at your own risk) console_output_style (string): console output: "classic", or with additional progress information ("progress" (percentage) | "count"). xfail_strict (bool): default for the strict parameter of xfail markers when not given explicitly (default: False) enable_assertion_pass_hook (bool): Enables the pytest_assertion_pass hook.Make sure to delete any previously generated pyc cache files. junit_suite_name (string): Test suite name for JUnit report junit_logging (string): Write captured log messages to JUnit report: one of no|system-out|system-err junit_log_passing_tests (bool): Capture log information for passing tests to JUnit report: junit_duration_report (string): Duration time to report: one of total|call junit_family (string): Emit XML for schema: one of legacy|xunit1|xunit2 doctest_optionflags (args): option flags for doctests doctest_encoding (string): encoding used for doctest files cache_dir (string): cache directory path. filterwarnings (linelist): Each line specifies a pattern for warnings.filterwarnings. Processed after -W and --pythonwarnings. log_print (bool): default value for --no-print-logs log_level (string): default value for --log-level log_format (string): default value for --log-format log_date_format (string): default value for --log-date-format log_cli (bool): enable log display during test run (also known as "live logging"). log_cli_level (string): default value for --log-cli-level log_cli_format (string): default value for --log-cli-format log_cli_date_format (string): default value for --log-cli-date-format log_file (string): default value for --log-file log_file_level (string): default value for --log-file-level log_file_format (string): default value for --log-file-format log_file_date_format (string): default value for --log-file-date-format faulthandler_timeout (string): Dump the traceback of all threads if a test takes more than TIMEOUT seconds to finish. Not available on Windows. addopts (args): extra command line options minversion (string): minimally required pytest version environment variables: PYTEST_ADDOPTS extra command line options PYTEST_PLUGINS comma-separated plugins to load during startup PYTEST_DISABLE_PLUGIN_AUTOLOAD set to disable plugin auto-loading PYTEST_DEBUG set to enable debug tracing of pytest's internals to see available markers type: pytest --markers to see available fixtures type: pytest --fixtures (shown according to specified file_or_dir or current dir if not specified; fixtures with leading '_' are only shown with the '-v' option ____________________________ test_hello_ini_setting ____________________________ testdir = def test_hello_ini_setting(testdir): testdir.makeini(""" [pytest] HELLO = world """) testdir.makepyfile(""" import pytest @pytest.fixture def hello(request): return request.config.getini('HELLO') def test_hello_world(hello): assert hello == 'world' """) result = testdir.runpytest('-v') # fnmatch_lines does an assertion internally result.stdout.fnmatch_lines([ > '*::test_hello_world PASSED', ]) E Failed: nomatch: '*::test_hello_world PASSED' E and: '============================= test session starts ==============================' E and: 'platform linux -- Python 3.7.1, pytest-5.2.4, py-1.8.0, pluggy-0.13.1 -- /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/bin/python' E and: 'cachedir: .pytest_cache' E and: 'rootdir: /tmp/pytest-of-travis/pytest-1/test_hello_ini_setting0, inifile: tox.ini' E and: 'plugins: blink1-0.0.8' E and: 'collecting ... collected 1 item' E and: '' E and: 'test_hello_ini_setting.py::test_hello_world ERROR [100%]' E and: '' E and: '==================================== ERRORS ====================================' E and: '______________________ ERROR at setup of test_hello_world ______________________' E and: '' E and: "self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO'" E and: '' E and: ' def getini(self, name: str):' E and: ' """ return configuration value from an :ref:`ini file `. If the' E and: " specified name hasn't been registered through a prior" E and: ' :py:func:`parser.addini <_pytest.config.Parser.addini>`' E and: ' call (usually from a plugin), a ValueError is raised. """' E and: ' try:' E and: '> return self._inicache[name]' E and: "E KeyError: 'HELLO'" E and: '' E and: '/tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:971: KeyError' E and: '' E and: 'During handling of the above exception, another exception occurred:' E and: '' E and: "self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO'" E and: '' E and: ' def _getini(self, name: str) -> Any:' E and: ' try:' E and: '> description, type, default = self._parser._inidict[name]' E and: "E KeyError: 'HELLO'" E and: '' E and: '/tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:978: KeyError' E and: '' E and: 'During handling of the above exception, another exception occurred:' E and: '' E and: "request = >" E and: '' E and: ' @pytest.fixture' E and: ' def hello(request):' E and: "> return request.config.getini('HELLO')" E and: '' E and: 'test_hello_ini_setting.py:5: ' E and: '_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ' E and: '/tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:973: in getini' E and: ' self._inicache[name] = val = self._getini(name)' E and: '_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ ' E and: '' E and: "self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO'" E and: '' E and: ' def _getini(self, name: str) -> Any:' E and: ' try:' E and: ' description, type, default = self._parser._inidict[name]' E and: ' except KeyError:' E and: '> raise ValueError("unknown configuration value: {!r}".format(name))' E and: "E ValueError: unknown configuration value: 'HELLO'" E and: '' E and: '/tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:980: ValueError' E and: '=============================== 1 error in 0.55s ===============================' E and: '' E remains unmatched: '*::test_hello_world PASSED' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/tests/test_blink1.py:60: Failed ----------------------------- Captured stdout call ----------------------------- ============================= test session starts ============================== platform linux -- Python 3.7.1, pytest-5.2.4, py-1.8.0, pluggy-0.13.1 -- /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/bin/python cachedir: .pytest_cache rootdir: /tmp/pytest-of-travis/pytest-1/test_hello_ini_setting0, inifile: tox.ini plugins: blink1-0.0.8 collecting ... collected 1 item test_hello_ini_setting.py::test_hello_world ERROR [100%] ==================================== ERRORS ==================================== ______________________ ERROR at setup of test_hello_world ______________________ self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO' def getini(self, name: str): """ return configuration value from an :ref:`ini file `. If the specified name hasn't been registered through a prior :py:func:`parser.addini <_pytest.config.Parser.addini>` call (usually from a plugin), a ValueError is raised. """ try: > return self._inicache[name] E KeyError: 'HELLO' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:971: KeyError During handling of the above exception, another exception occurred: self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO' def _getini(self, name: str) -> Any: try: > description, type, default = self._parser._inidict[name] E KeyError: 'HELLO' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:978: KeyError During handling of the above exception, another exception occurred: request = > @pytest.fixture def hello(request): > return request.config.getini('HELLO') test_hello_ini_setting.py:5: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:973: in getini self._inicache[name] = val = self._getini(name) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <_pytest.config.Config object at 0x7f325fa73828>, name = 'HELLO' def _getini(self, name: str) -> Any: try: description, type, default = self._parser._inidict[name] except KeyError: > raise ValueError("unknown configuration value: {!r}".format(name)) E ValueError: unknown configuration value: 'HELLO' /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/lib/python3.7/site-packages/_pytest/config/__init__.py:980: ValueError =============================== 1 error in 0.55s =============================== ============================== 3 failed in 2.35s =============================== ERROR: InvocationError for command /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/.tox/py37/bin/py.test tests (exited with code 1) ___________________________________ summary ____________________________________ ERROR: py37: commands failed write json report at: /tmp/tmptqkn_kiv/pytest-blink1-0.0.8/result.json Time: 49.4 seconds