1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
import os
import pathlib
def run_perfdocs(config, logger=None, paths=None, generate=True):
"""
Build up performance testing documentation dynamically by combining
text data from YAML files that reside in `perfdoc` folders
across the `testing` directory. Each directory is expected to have
an `index.rst` file along with `config.yml` YAMLs defining what needs
to be added to the documentation.
The YAML must also define the name of the "framework" that should be
used in the main index.rst for the performance testing documentation.
The testing documentation list will be ordered alphabetically once
it's produced (to avoid unwanted shifts because of unordered dicts
and path searching).
Note that the suite name headings will be given the H4 (---) style so it
is suggested that you use H3 (===) style as the heading for your
test section. H5 will be used be used for individual tests within each
suite.
Usage for verification: "./mach lint -l perfdocs ."
Usage for generation: "./mach lint -l perfdocs --fix ."
For validation, see the Verifier class for a description of how
it works.
The run will fail if the valid result from validate_tree is not
False, implying some warning/problem was logged.
:param dict config: The configuration given by mozlint.
:param StructuredLogger logger: The StructuredLogger instance to be used to
output the linting warnings/errors.
:param list paths: The paths that are being tested. Used to filter
out errors from files outside of these paths.
:param bool generate: If true, the docs will be (re)generated.
"""
from perfdocs.logger import PerfDocLogger
if not os.environ.get("WORKSPACE", None):
floc = pathlib.Path(__file__).absolute()
top_dir = pathlib.Path(str(floc).split("tools")[0]).resolve()
else:
top_dir = pathlib.Path(os.environ.get("WORKSPACE")).resolve()
PerfDocLogger.LOGGER = logger
PerfDocLogger.TOP_DIR = top_dir
# Convert all the paths to relative ones
target_dir = [pathlib.Path(path) for path in paths]
rel_paths = []
for path in target_dir:
try:
rel_paths.append(path.relative_to(top_dir))
except ValueError:
rel_paths.append(path)
PerfDocLogger.PATHS = rel_paths
for path in target_dir:
if not path.exists():
raise Exception("Cannot locate directory at %s" % str(path))
decision_task_id = os.environ.get("DECISION_TASK_ID", None)
if decision_task_id:
from taskgraph.util.taskcluster import get_artifact
task_graph = get_artifact(decision_task_id, "public/full-task-graph.json")
else:
from tryselect.tasks import generate_tasks
task_graph = generate_tasks(
params=None, full=True, disable_target_task_filter=True
).tasks
# Late import because logger isn't defined until later
from perfdocs.generator import Generator
from perfdocs.verifier import Verifier
# Run the verifier first
verifier = Verifier(top_dir, task_graph)
verifier.validate_tree()
if not PerfDocLogger.FAILED:
# Even if the tree is valid, we need to check if the documentation
# needs to be regenerated, and if it does, we throw a linting error.
# `generate` dictates whether or not the documentation is generated.
generator = Generator(verifier, generate=generate, workspace=top_dir)
generator.generate_perfdocs()
|