1
0
mirror of https://github.com/pypiserver/pypiserver synced 2024-11-09 16:45:51 +01:00
pypiserver/tests/test_config.py
Mason Lin ae3dcf2bbd
feat: 🩺 allow customized health check endpoint (#442)
* feat: 🩺 allow customized health check endpoint

Get the liveness endpoint from the environment variable `HEALTH_ENDPOINT` and verify it. If the customized endpoint is invalied, it will fallback to the DEFAULT_HEALTH_ENDPOINT.

* test:  Test customized endpoint feature

* fix: 🚨 fix check

* feat: Use CLI interface to set health endpoint

* style: 💄 fix black format

* Separate 'build app' and 'add routes'

https://github.com/pypiserver/pypiserver/pull/442#discussion_r973771421

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* keep DEFAULTS in config.py

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* style alignment

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* make CLI arg description more clear

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* style: 🎨 style alignment

* refactor:  SRP, add routes after app created, instead of patching in app_from_config

* style: 🎨 format CLI help

* test:  add test_setup_routes_from_config

* fix: 🐛 test name doesn't work as expected because of using the wrong ids generator.

* test: 🧪 add config error cases for health endpoint

* test:  fix health_endpoint_arg tests

* fix:  Do not fallback to default silently, should raise error

* test: 🧪 add test_health_endpoint in test_main

* test:  setup routes in main

* docs: 📝 Update the help command output in the Quickstart

* docs: 🐛 missing space

* docs: 📝 Add 'Custom Health Check Endpoint' to 'Recipes'

* docs: 📝 refine README

* revert:  revert auto isoft

* build: 💚 fix mypy, missing return types

* Update README.rst

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* Update README.rst

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* Update pypiserver/config.py

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* Update README.rst

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

* style: 💄 black format

* Update README.rst

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>

Co-authored-by: Dmitrii Orlov <dmtree.dev@yahoo.com>
2022-11-02 12:32:20 +01:00

782 lines
26 KiB
Python

"""Tests for config parsing."""
import hashlib
import typing as t
import itertools
import pathlib
import sys
import pytest
from pypiserver.backend import SimpleFileBackend, BackendProxy
from pypiserver.config import DEFAULTS, Config, RunConfig, UpdateConfig
FILE_DIR = pathlib.Path(__file__).parent.resolve()
# Username and password stored in the htpasswd.a.a test file.
HTPASS_TEST_FILE = str(FILE_DIR / "../fixtures/htpasswd.a.a")
HTPASS_TEST_USER = "a"
HTPASS_TEST_PASS = "a"
TEST_WELCOME_FILE = str(pathlib.Path(__file__).parent / "sample_msg.html")
TEST_IGNORELIST_FILE = str(pathlib.Path(__file__).parent / "test-ignorelist")
class ConfigTestCase(t.NamedTuple):
# A description of the test case
case: str
# Arguments to pass to the Config constructor
args: t.List[str]
# Legacy arguments that should yield an equivalent config class
legacy_args: t.List[str]
# The config class the arguments should resolve to
exp_config_type: t.Type
# Expected values in the config. These don't necessarily need to be
# exclusive. Instead, they should just look at the attributes relevant
# to the test case at hand. A special "_test" key, if present, should
# map to a function that takes the config as an argument. If this
# returns a falsey value, the test will be failed.
exp_config_values: t.Dict[str, t.Any]
# The iterables generated by this function are designed to be unpacked
# into the _CONFIG_TEST_PARAMS constant.
def generate_subcommand_test_cases(
case: str,
extra_args: t.List[str] = None,
exp_config_values: t.Dict[str, t.Any] = None,
) -> t.Iterable[ConfigTestCase]:
"""Generate `run` and `update` test cases automatically.
Use to avoid having to specify individual cases for situations like
global arguments, where the resultant configs should have the same values.
These tests also help to ensure parity between legacy and modern orderings,
since generally the only difference between the two should be the presence
or absence of the subcommand.
:param case: the test case name. will be combined with the subcommand
to generate a case name for the resultant case.
:param extra_args: arguments to pass after the subcommand positional
arguments
:param extra_legacy_args: legacy arguments to pass in addition to the
subcommand arguments, if any
:param exp_config_values: the values that should be present on both
run and update test cases.
"""
extra_args = extra_args or []
extra_legacy_args = extra_args
exp_config_values = exp_config_values or {}
# The legacy "update" subcommand was specified with an optional `-U`
# argument. This allows us to map the subcommand to that argument, so
# we can include it in the resulting legacy args.
legacy_base_arg_map = {"update": ["-U"]}
# A mapping of subcommands to their expected Config types.
config_type_map = {
"run": RunConfig,
"update": UpdateConfig,
}
return (
ConfigTestCase(
case="{subcmd}: {case}",
args=[subcmd, *extra_args],
legacy_args=[
*legacy_base_arg_map.get(subcmd, []),
*extra_legacy_args,
],
exp_config_type=config_type_map[subcmd],
exp_config_values=exp_config_values,
)
for subcmd in ("run", "update")
)
# Define Config test parameters
_CONFIG_TEST_PARAMS: t.Tuple[ConfigTestCase, ...] = (
# ******************************************************************
# Raw subcommands
# ******************************************************************
*generate_subcommand_test_cases(
case="no arguments",
),
# ******************************************************************
# Global args
# ******************************************************************
# Package directories
*generate_subcommand_test_cases(
case="no package directory specified",
exp_config_values={"roots": DEFAULTS.PACKAGE_DIRECTORIES},
),
*generate_subcommand_test_cases(
case="single package directory specified",
extra_args=[str(FILE_DIR)],
exp_config_values={"roots": [FILE_DIR]},
),
*generate_subcommand_test_cases(
case="multiple package directory specified",
extra_args=[str(FILE_DIR), str(FILE_DIR.parent)],
exp_config_values={
"roots": [
FILE_DIR,
FILE_DIR.parent,
]
},
),
ConfigTestCase(
case="update with package directory (out-of-order legacy order)",
args=["update", str(FILE_DIR)],
legacy_args=[str(FILE_DIR), "-U"],
exp_config_type=UpdateConfig,
exp_config_values={"roots": [FILE_DIR]},
),
ConfigTestCase(
case="update with multiple package directories (weird ordering)",
args=["update", str(FILE_DIR), str(FILE_DIR.parent)],
legacy_args=[str(FILE_DIR), "-U", str(FILE_DIR.parent)],
exp_config_type=UpdateConfig,
exp_config_values={
"roots": [
FILE_DIR,
FILE_DIR.parent,
]
},
),
# verbosity
*(
# Generate verbosity test-cases for 0 through 5 -v arguments,
# for all subcommands.
itertools.chain.from_iterable(
# This inner iterable (generate(...) for verbosity in range(5))
# will be an iterable of 5 items, where each item is essentially
# an n-tuple, where n is the number of subcommands. Passing this
# iterable to chain.from_iterable() flattens it, so it is just
# one long iterable of cases. These are then unpacked into the
# test case tuple with the *, above.
generate_subcommand_test_cases(
case=f"verbosity {verbosity}",
extra_args=["-v" for _ in range(verbosity)],
exp_config_values={"verbosity": verbosity},
)
for verbosity in range(5)
)
),
# log-file
*generate_subcommand_test_cases(
case="log file unspecified", exp_config_values={"log_file": None}
),
*generate_subcommand_test_cases(
case="log file specified",
extra_args=["--log-file", "foo"],
exp_config_values={"log_file": "foo"},
),
# log-stream
*generate_subcommand_test_cases(
case="log stream unspecified",
exp_config_values={"log_stream": DEFAULTS.LOG_STREAM},
),
*generate_subcommand_test_cases(
case="log stream set to stdout",
extra_args=["--log-stream", "stdout"],
exp_config_values={"log_stream": sys.stdout},
),
*generate_subcommand_test_cases(
case="log stream set to stderr",
extra_args=["--log-stream", "stderr"],
exp_config_values={"log_stream": sys.stderr},
),
*generate_subcommand_test_cases(
case="log stream set to none",
extra_args=["--log-stream", "none"],
exp_config_values={"log_stream": None},
),
*generate_subcommand_test_cases(
case="log format unset",
exp_config_values={"log_frmt": DEFAULTS.LOG_FRMT},
),
*generate_subcommand_test_cases(
case="log format set",
extra_args=["--log-frmt", "foobar %(message)s"],
exp_config_values={"log_frmt": "foobar %(message)s"},
),
# ******************************************************************
# Run subcommand args
# ******************************************************************
# port
ConfigTestCase(
case="Run: port unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"port": DEFAULTS.PORT},
),
ConfigTestCase(
case="Run: port specified",
args=["run", "-p", "9900"],
legacy_args=["-p", "9900"],
exp_config_type=RunConfig,
exp_config_values={"port": 9900},
),
ConfigTestCase(
case="Run: port specified (long form)",
args=["run", "--port", "9900"],
legacy_args=["--port", "9900"],
exp_config_type=RunConfig,
exp_config_values={"port": 9900},
),
# interface
ConfigTestCase(
case="Run: interface unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"host": DEFAULTS.INTERFACE},
),
ConfigTestCase(
case="Run: interface specified",
args=["run", "-i", "1.1.1.1"],
legacy_args=["-i", "1.1.1.1"],
exp_config_type=RunConfig,
exp_config_values={"host": "1.1.1.1"},
),
ConfigTestCase(
case="Run: interface specified (long form)",
args=["run", "--interface", "1.1.1.1"],
legacy_args=["--interface", "1.1.1.1"],
exp_config_type=RunConfig,
exp_config_values={"host": "1.1.1.1"},
),
ConfigTestCase(
case="Run: host specified",
args=["run", "-H", "1.1.1.1"],
legacy_args=["-H", "1.1.1.1"],
exp_config_type=RunConfig,
exp_config_values={"host": "1.1.1.1"},
),
ConfigTestCase(
case="Run: host specified (long form)",
args=["run", "--host", "1.1.1.1"],
legacy_args=["--host", "1.1.1.1"],
exp_config_type=RunConfig,
exp_config_values={"host": "1.1.1.1"},
),
# authenticate
ConfigTestCase(
case="Run: authenticate unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"authenticate": DEFAULTS.AUTHENTICATE},
),
ConfigTestCase(
case="Run: authenticate specified as non-default value",
args=["run", "-a", "list"],
legacy_args=["-a", "list"],
exp_config_type=RunConfig,
exp_config_values={"authenticate": ["list"]},
),
ConfigTestCase(
case="Run: authenticate specified with multiple values",
args=["run", "-a", "list, update,download"],
legacy_args=["-a", "list, update,download"],
exp_config_type=RunConfig,
exp_config_values={"authenticate": ["download", "list", "update"]},
),
ConfigTestCase(
case="Run: authenticate specified with dot",
# both auth and pass must be specified as empty if one of them is empty.
args=["run", "-a", ".", "-P", "."],
legacy_args=["-a", ".", "-P", "."],
exp_config_type=RunConfig,
exp_config_values={
"authenticate": [],
"_test": lambda conf: bool(conf.auther("foo", "bar")) is True,
},
),
# passwords
ConfigTestCase(
case="Run: passwords file unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"password_file": None},
),
ConfigTestCase(
"Run: passwords file specified",
args=["run", "-P", HTPASS_TEST_FILE],
legacy_args=["-P", HTPASS_TEST_FILE],
exp_config_type=RunConfig,
exp_config_values={
"password_file": HTPASS_TEST_FILE,
"_test": lambda conf: (
bool(conf.auther("foo", "bar")) is False
and bool(conf.auther("a", "a")) is True
),
},
),
ConfigTestCase(
"Run: passwords file specified (long-form)",
args=["run", "--passwords", HTPASS_TEST_FILE],
legacy_args=["--passwords", HTPASS_TEST_FILE],
exp_config_type=RunConfig,
exp_config_values={
"password_file": HTPASS_TEST_FILE,
"_test": (
lambda conf: (
bool(conf.auther("foo", "bar")) is False
and conf.auther("a", "a") is True
)
),
},
),
ConfigTestCase(
"Run: passwords file empty ('.')",
# both auth and pass must be specified as empty if one of them is empty.
args=["run", "-P", ".", "-a", "."],
legacy_args=["-P", ".", "-a", "."],
exp_config_type=RunConfig,
exp_config_values={
"password_file": ".",
"_test": lambda conf: bool(conf.auther("foo", "bar")) is True,
},
),
# disable-fallback
ConfigTestCase(
case="Run: disable-fallback unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"disable_fallback": False},
),
ConfigTestCase(
case="Run: disable-fallback set",
args=["run", "--disable-fallback"],
legacy_args=["--disable-fallback"],
exp_config_type=RunConfig,
exp_config_values={"disable_fallback": True},
),
# fallback-url
ConfigTestCase(
case="Run: fallback-url unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"fallback_url": DEFAULTS.FALLBACK_URL},
),
ConfigTestCase(
case="Run: fallback-url specified",
args=["run", "--fallback-url", "foobar.com"],
legacy_args=["--fallback-url", "foobar.com"],
exp_config_type=RunConfig,
exp_config_values={"fallback_url": "foobar.com"},
),
# health-endpoint
ConfigTestCase(
case="Run: health-endpoint unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"health_endpoint": DEFAULTS.HEALTH_ENDPOINT},
),
ConfigTestCase(
case="Run: health-endpoint specified",
args=["run", "--health-endpoint", "/healthz"],
legacy_args=["--health-endpoint", "/healthz"],
exp_config_type=RunConfig,
exp_config_values={"health_endpoint": "/healthz"},
),
# server method
ConfigTestCase(
case="Run: server method unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"server_method": DEFAULTS.SERVER_METHOD},
),
*(
ConfigTestCase(
case="Run: server method set to {arg}",
args=["run", "--server", arg],
legacy_args=["--server", arg],
exp_config_type=RunConfig,
exp_config_values={"server_method": arg},
)
for arg in (
"auto",
"cherrypy",
"gevent",
"gunicorn",
"paste",
"twisted",
"wsgiref",
)
),
ConfigTestCase(
case="Run: server method is case insensitive",
args=["run", "--server", "CherryPy"],
legacy_args=["--server", "CherryPy"],
exp_config_type=RunConfig,
exp_config_values={"server_method": "cherrypy"},
),
# overwrite
ConfigTestCase(
"Run: overwrite unset",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"overwrite": False},
),
ConfigTestCase(
case="Run: overwrite set (long-form)",
args=["run", "-o"],
legacy_args=["-o"],
exp_config_type=RunConfig,
exp_config_values={"overwrite": True},
),
ConfigTestCase(
case="Run: overwrite set (long-form)",
args=["run", "--overwrite"],
legacy_args=["--overwrite"],
exp_config_type=RunConfig,
exp_config_values={"overwrite": True},
),
# hash-algo
ConfigTestCase(
case="Run: hash-algo unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"hash_algo": DEFAULTS.HASH_ALGO},
),
*(
ConfigTestCase(
case="Run: hash-algo {}",
args=["run", "--hash-algo", algo],
legacy_args=["--hash-algo", algo],
exp_config_type=RunConfig,
exp_config_values={"hash_algo": algo},
)
for algo in hashlib.algorithms_available
),
*(
ConfigTestCase(
case="Run: hash-algo disabled",
args=["run", "--hash-algo", off_value],
legacy_args=["--hash-algo", off_value],
exp_config_type=RunConfig,
exp_config_values={"hash_algo": None},
)
for off_value in ("0", "off", "false", "no", "NO")
),
# welcome file
ConfigTestCase(
case="Run: welcome file unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={
"_test": lambda conf: "Welcome to pypiserver" in conf.welcome_msg
},
),
ConfigTestCase(
case="Run: custom welcome file specified",
args=["run", "--welcome", TEST_WELCOME_FILE],
legacy_args=["--welcome", TEST_WELCOME_FILE],
exp_config_type=RunConfig,
exp_config_values={
"_test": lambda conf: "Hello pypiserver tester!" in conf.welcome_msg
},
),
# cache-control
ConfigTestCase(
case="Run: cache-control unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"cache_control": None},
),
ConfigTestCase(
case="Run: cache-control specified",
args=["run", "--cache-control", "1900"],
legacy_args=["--cache-control", "1900"],
exp_config_type=RunConfig,
exp_config_values={"cache_control": 1900},
),
# log-req-frmt
ConfigTestCase(
case="Run: log request format unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"log_req_frmt": DEFAULTS.LOG_REQ_FRMT},
),
ConfigTestCase(
case="Run: log request format specified",
args=["run", "--log-req-frmt", "foo"],
legacy_args=["--log-req-frmt", "foo"],
exp_config_type=RunConfig,
exp_config_values={"log_req_frmt": "foo"},
),
# log-res-frmt
ConfigTestCase(
case="Run: log response format unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"log_res_frmt": DEFAULTS.LOG_RES_FRMT},
),
ConfigTestCase(
case="Run: log response format specified",
args=["run", "--log-res-frmt", "foo"],
legacy_args=["--log-res-frmt", "foo"],
exp_config_type=RunConfig,
exp_config_values={"log_res_frmt": "foo"},
),
# log-err-frmt
ConfigTestCase(
case="Run: log error format unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={"log_err_frmt": DEFAULTS.LOG_ERR_FRMT},
),
ConfigTestCase(
case="Run: log error format specified",
args=["run", "--log-err-frmt", "foo"],
legacy_args=["--log-err-frmt", "foo"],
exp_config_type=RunConfig,
exp_config_values={"log_err_frmt": "foo"},
),
# backend
ConfigTestCase(
"Run: backend unspecified",
args=["run"],
legacy_args=[],
exp_config_type=RunConfig,
exp_config_values={
"backend_arg": "auto",
"_test": (
lambda conf: (
isinstance(conf.backend, BackendProxy)
and isinstance(conf.backend.backend, SimpleFileBackend)
)
),
},
),
ConfigTestCase(
"Run: simple backend specified",
args=["run", "--backend", "simple-dir"],
legacy_args=["--backend", "simple-dir"],
exp_config_type=RunConfig,
exp_config_values={
"_test": (
lambda conf: (
isinstance(conf.backend.backend, SimpleFileBackend)
)
),
},
),
# ******************************************************************
# Update subcommand args
# ******************************************************************
# execute
ConfigTestCase(
case="Update: execute not specified",
args=["update"],
legacy_args=["-U"],
exp_config_type=UpdateConfig,
exp_config_values={"execute": False},
),
ConfigTestCase(
case="Update: execute specified",
args=["update", "-x"],
legacy_args=["-U", "-x"],
exp_config_type=UpdateConfig,
exp_config_values={"execute": True},
),
ConfigTestCase(
case="Update: execute specified (long-form)",
args=["update", "--execute"],
legacy_args=["-U", "--execute"],
exp_config_type=UpdateConfig,
exp_config_values={"execute": True},
),
# download-directory
ConfigTestCase(
case="Update: download-directory not specified",
args=["update"],
legacy_args=["-U"],
exp_config_type=UpdateConfig,
exp_config_values={"download_directory": None},
),
ConfigTestCase(
case="Update: download-directory specified",
args=["update", "-d", "foo"],
legacy_args=["-U", "-d", "foo"],
exp_config_type=UpdateConfig,
exp_config_values={"download_directory": "foo"},
),
ConfigTestCase(
case="Update: download-directory specified (long-form)",
args=["update", "--download-directory", "foo"],
legacy_args=["-U", "--download-directory", "foo"],
exp_config_type=UpdateConfig,
exp_config_values={"download_directory": "foo"},
),
# allow-unstable
ConfigTestCase(
case="Update: allow-unstable not specified",
args=["update"],
legacy_args=["-U"],
exp_config_type=UpdateConfig,
exp_config_values={"allow_unstable": False},
),
ConfigTestCase(
case="Update: allow-unstable specified",
args=["update", "-u"],
legacy_args=["-U", "-u"],
exp_config_type=UpdateConfig,
exp_config_values={"allow_unstable": True},
),
ConfigTestCase(
case="Update: allow-unstable specified (long-form)",
args=["update", "--allow-unstable"],
legacy_args=["-U", "--allow-unstable"],
exp_config_type=UpdateConfig,
exp_config_values={"allow_unstable": True},
),
# ignorelist-file
ConfigTestCase(
case="Update: ignorelist-file not specified",
args=["update"],
legacy_args=["-U"],
exp_config_type=UpdateConfig,
exp_config_values={"ignorelist": []},
),
ConfigTestCase(
case="Update: ignorelist-file specified",
args=["update", "--ignorelist-file", TEST_IGNORELIST_FILE],
legacy_args=["-U", "--ignorelist-file", TEST_IGNORELIST_FILE],
exp_config_type=UpdateConfig,
exp_config_values={"ignorelist": ["mypiserver", "something"]},
),
ConfigTestCase(
case="Update: blacklist-file specified",
args=["update", "--blacklist-file", TEST_IGNORELIST_FILE],
legacy_args=["-U", "--blacklist-file", TEST_IGNORELIST_FILE],
exp_config_type=UpdateConfig,
exp_config_values={"ignorelist": ["mypiserver", "something"]},
),
)
# Split case names out from cases to use as pytest IDs.
# pylint: disable=unsubscriptable-object
CONFIG_TEST_PARAMS = (i[1:] for i in _CONFIG_TEST_PARAMS)
# pylint: enable=unsubscriptable-object
CONFIG_TEST_IDS = (i.case for i in _CONFIG_TEST_PARAMS)
class ConfigErrorCase(t.NamedTuple):
"""Configuration arguments that should cause errors.
The cases include a case descrpition, a list of arguments,
and, if desired, expected text that should be part of what
is printed out to stderr. If no text is provided, the content
of stderr will not be checked.
"""
case: str
args: t.List[str]
exp_txt: t.Optional[str]
_CONFIG_ERROR_CASES = (
*(
ConfigErrorCase(
case=f"Invalid hash algo: {val}",
args=["run", "--hash-algo", val],
exp_txt=f"Hash algorithm '{val}' is not available",
)
for val in ("true", "foo", "1", "md6")
),
*(
ConfigErrorCase(
case=f"Invalid health endpoint: {val}",
args=["run", "--health-endpoint", val],
exp_txt="Invalid path for the health endpoint",
)
for val in ("/", "health", "/health!", "/:health", "/health?check=True")
),
)
# pylint: disable=unsubscriptable-object
CONFIG_ERROR_PARAMS = (i[1:] for i in _CONFIG_ERROR_CASES)
# pylint: enable=unsubscriptable-object
CONFIG_ERROR_IDS = (i.case for i in _CONFIG_ERROR_CASES)
@pytest.mark.parametrize(
"args, legacy_args, exp_config_type, exp_config_values",
CONFIG_TEST_PARAMS,
ids=CONFIG_TEST_IDS,
)
def test_config(
args: t.List[str],
legacy_args: t.List[str],
exp_config_type: t.Type,
exp_config_values: t.Dict[str, t.Any],
) -> None:
"""Validate config test cases."""
conf = Config.from_args(args)
conf_legacy = Config.from_args(legacy_args)
assert isinstance(conf, exp_config_type)
assert all(
getattr(conf, k) == v
for k, v in exp_config_values.items()
if k != "_test"
), {
k: (getattr(conf, k), v)
for k, v in exp_config_values.items()
if k != "_test" and getattr(conf, k) != v
}
if "_test" in exp_config_values:
assert exp_config_values["_test"](conf)
assert conf == conf_legacy
@pytest.mark.parametrize(
"args, exp_txt",
CONFIG_ERROR_PARAMS,
ids=CONFIG_ERROR_IDS,
)
def test_config_error(
args: t.List[str],
exp_txt: t.Optional[str],
capsys,
) -> None:
"""Validate error cases."""
with pytest.raises(SystemExit):
Config.from_args(args)
# Unfortunately the error text is printed before the SystemExit is
# raised, rather than being raised _with_ the systemexit, so we
# need to capture stderr and check it for our expected text, if
# any was specified in the test case.
if exp_txt is not None:
assert exp_txt in capsys.readouterr().err
def test_argv_conf():
"""Config uses argv if no args are provided."""
orig_args = list(sys.argv)
sys.argv = [sys.argv[0], "run", "-v", "--disable-fallback"]
try:
conf = Config.from_args()
assert isinstance(conf, RunConfig)
assert conf.verbosity == 1
assert conf.disable_fallback is True
finally:
sys.argv = orig_args