04df6e42e0
* Simplify score_file() by using case in instead of nested ifs with regexps. * Merge record_score() and record_template(). * Alt condition processing no longer stops when a template condition is seen but continues processing to verify that all conditions are valid (as the documentation says it should). Fixes #478. * Support alt dirs with deeply nested tracked files (fixes #490). * Use git ls-files to filter out which tracked files to consider for alt processing. Should speed up auto-alt (#505). * Use nocasematch when comparing distro and distro_family. Fixed #455.
189 lines
5.5 KiB
Python
189 lines
5.5 KiB
Python
"""Unit tests: record_score"""
|
|
|
|
import pytest
|
|
|
|
INIT_VARS = """
|
|
score=0
|
|
local_class=testclass
|
|
local_system=testsystem
|
|
local_host=testhost
|
|
local_user=testuser
|
|
alt_scores=()
|
|
alt_targets=()
|
|
alt_sources=()
|
|
alt_template_cmds=()
|
|
"""
|
|
|
|
REPORT_RESULTS = """
|
|
echo "SIZE:${#alt_scores[@]}"
|
|
echo "SCORES:${alt_scores[@]}"
|
|
echo "TARGETS:${alt_targets[@]}"
|
|
echo "SOURCES:${alt_sources[@]}"
|
|
echo "TEMPLATE_CMDS:${alt_template_cmds[@]}"
|
|
"""
|
|
|
|
|
|
def test_dont_record_zeros(runner, yadm):
|
|
"""Record nothing if the score is zero"""
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
record_score "0" "testtgt" "testsrc"
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:0\n" in run.out
|
|
assert "SCORES:\n" in run.out
|
|
assert "TARGETS:\n" in run.out
|
|
assert "SOURCES:\n" in run.out
|
|
assert "TEMPLATE_CMDS:\n" in run.out
|
|
|
|
|
|
def test_new_scores(runner, yadm):
|
|
"""Test new scores"""
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
record_score "1" "tgt_one" "src_one" ""
|
|
record_score "2" "tgt_two" "src_two" ""
|
|
record_score "4" "tgt_three" "src_three" ""
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:3\n" in run.out
|
|
assert "SCORES:1 2 4\n" in run.out
|
|
assert "TARGETS:tgt_one tgt_two tgt_three\n" in run.out
|
|
assert "SOURCES:src_one src_two src_three\n" in run.out
|
|
assert "TEMPLATE_CMDS: \n" in run.out
|
|
|
|
|
|
@pytest.mark.parametrize("difference", ["lower", "equal", "higher"])
|
|
def test_existing_scores(runner, yadm, difference):
|
|
"""Test existing scores"""
|
|
|
|
expected_score = "2"
|
|
expected_src = "existing_src"
|
|
if difference == "lower":
|
|
score = "1"
|
|
elif difference == "equal":
|
|
score = "2"
|
|
else:
|
|
score = "4"
|
|
expected_score = "4"
|
|
expected_src = "new_src"
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
alt_scores=(2)
|
|
alt_targets=("testtgt")
|
|
alt_sources=("existing_src")
|
|
alt_template_cmds=("")
|
|
record_score "{score}" "testtgt" "new_src" ""
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:1\n" in run.out
|
|
assert f"SCORES:{expected_score}\n" in run.out
|
|
assert "TARGETS:testtgt\n" in run.out
|
|
assert f"SOURCES:{expected_src}\n" in run.out
|
|
assert "TEMPLATE_CMDS:\n" in run.out
|
|
|
|
|
|
def test_existing_template(runner, yadm):
|
|
"""Record nothing if a template command is registered for this target"""
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
alt_scores=(1)
|
|
alt_targets=("testtgt")
|
|
alt_sources=("src")
|
|
alt_template_cmds=("existing_template")
|
|
record_score "2" "testtgt" "new_src" ""
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:1\n" in run.out
|
|
assert "SCORES:1\n" in run.out
|
|
assert "TARGETS:testtgt\n" in run.out
|
|
assert "SOURCES:src\n" in run.out
|
|
assert "TEMPLATE_CMDS:existing_template\n" in run.out
|
|
|
|
|
|
def test_config_first(runner, yadm):
|
|
"""Verify YADM_CONFIG is always processed first"""
|
|
|
|
config = "yadm_config_file"
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
YADM_CONFIG={config}
|
|
record_score "1" "tgt_before" "src_before" ""
|
|
record_score "1" "tgt_tmp" "src_tmp" "cmd_tmp"
|
|
record_score "2" "{config}" "src_config" ""
|
|
record_score "3" "tgt_after" "src_after" ""
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:4\n" in run.out
|
|
assert "SCORES:2 1 1 3\n" in run.out
|
|
assert f"TARGETS:{config} tgt_before tgt_tmp tgt_after\n" in run.out
|
|
assert "SOURCES:src_config src_before src_tmp src_after\n" in run.out
|
|
assert "TEMPLATE_CMDS: cmd_tmp \n" in run.out
|
|
|
|
|
|
def test_new_template(runner, yadm):
|
|
"""Test new template"""
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
record_score 0 "tgt_one" "src_one" "cmd_one"
|
|
record_score 0 "tgt_two" "src_two" "cmd_two"
|
|
record_score 0 "tgt_three" "src_three" "cmd_three"
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:3\n" in run.out
|
|
assert "SCORES:0 0 0\n" in run.out
|
|
assert "TARGETS:tgt_one tgt_two tgt_three\n" in run.out
|
|
assert "SOURCES:src_one src_two src_three\n" in run.out
|
|
assert "TEMPLATE_CMDS:cmd_one cmd_two cmd_three\n" in run.out
|
|
|
|
|
|
def test_overwrite_existing_template(runner, yadm):
|
|
"""Overwrite existing templates"""
|
|
|
|
script = f"""
|
|
YADM_TEST=1 source {yadm}
|
|
{INIT_VARS}
|
|
alt_scores=(0)
|
|
alt_targets=("testtgt")
|
|
alt_template_cmds=("existing_cmd")
|
|
alt_sources=("existing_src")
|
|
record_score 0 "testtgt" "new_src" "new_cmd"
|
|
{REPORT_RESULTS}
|
|
"""
|
|
run = runner(command=["bash"], inp=script)
|
|
assert run.success
|
|
assert run.err == ""
|
|
assert "SIZE:1\n" in run.out
|
|
assert "SCORES:0\n" in run.out
|
|
assert "TARGETS:testtgt\n" in run.out
|
|
assert "SOURCES:new_src\n" in run.out
|
|
assert "TEMPLATE_CMDS:new_cmd\n" in run.out
|