diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 17:32:43 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-07 17:32:43 +0000 |
commit | 6bf0a5cb5034a7e684dcc3500e841785237ce2dd (patch) | |
tree | a68f146d7fa01f0134297619fbe7e33db084e0aa /taskcluster/test/test_mozilla_central.py | |
parent | Initial commit. (diff) | |
download | thunderbird-upstream.tar.xz thunderbird-upstream.zip |
Adding upstream version 1:115.7.0.upstream/1%115.7.0upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to '')
-rw-r--r-- | taskcluster/test/test_mozilla_central.py | 69 |
1 files changed, 69 insertions, 0 deletions
diff --git a/taskcluster/test/test_mozilla_central.py b/taskcluster/test/test_mozilla_central.py new file mode 100644 index 0000000000..696a108db6 --- /dev/null +++ b/taskcluster/test/test_mozilla_central.py @@ -0,0 +1,69 @@ +# Any copyright is dedicated to the public domain. +# http://creativecommons.org/publicdomain/zero/1.0/ + +import pytest +from mozunit import main + +pytestmark = pytest.mark.slow +PARAMS = { + "head_repository": "https://hg.mozilla.org/mozilla-central", + "project": "central", +} + + +def test_generate_graph(optimized_task_graph): + """Simply tests that generating the graph does not fail.""" + assert len(optimized_task_graph.tasks) > 0 + + +@pytest.mark.parametrize( + "func,min_expected", + ( + pytest.param( + lambda t: t.kind == "build" and "fuzzing" in t.attributes["build_platform"], + 5, + id="fuzzing builds", + ), + ), +) +def test_tasks_are_scheduled(optimized_task_graph, filter_tasks, func, min_expected): + """Ensure the specified tasks are scheduled on mozilla-central.""" + tasks = [t.label for t in filter_tasks(optimized_task_graph, func)] + print(tasks) + assert len(tasks) >= min_expected + + +def test_test_setting(full_task_graph, filter_tasks): + """Verify that all test tasks' ``test-setting`` object conforms to the schema.""" + from gecko_taskgraph.transforms.test.other import test_setting_description_schema + from taskgraph.util.schema import validate_schema + + tasks = filter_tasks(full_task_graph, lambda t: t.kind == "test") + + failures = [] + for task in tasks: + try: + validate_schema( + test_setting_description_schema, + dict(task.task["extra"]["test-setting"]), + task.label, + ) + except Exception as e: + failures.append(e) + + if failures: + more = None + # Only display a few failures at once. + if len(failures) > 10: + more = len(failures) - 10 + failures = failures[:10] + + failstr = "\n\n" + "\n\n".join(str(e) for e in failures) + if more: + failstr += "\n\n" + f"... and {more} more" + + pytest.fail(f"Task validation errors:{failstr}") + + +if __name__ == "__main__": + main() |