summaryrefslogtreecommitdiffstats
path: root/js/src/tests/parsemark.py
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 19:33:14 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-07 19:33:14 +0000
commit36d22d82aa202bb199967e9512281e9a53db42c9 (patch)
tree105e8c98ddea1c1e4784a60a5a6410fa416be2de /js/src/tests/parsemark.py
parentInitial commit. (diff)
downloadfirefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.tar.xz
firefox-esr-36d22d82aa202bb199967e9512281e9a53db42c9.zip
Adding upstream version 115.7.0esr.upstream/115.7.0esrupstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'js/src/tests/parsemark.py')
-rw-r--r--js/src/tests/parsemark.py257
1 files changed, 257 insertions, 0 deletions
diff --git a/js/src/tests/parsemark.py b/js/src/tests/parsemark.py
new file mode 100644
index 0000000000..199747e7dd
--- /dev/null
+++ b/js/src/tests/parsemark.py
@@ -0,0 +1,257 @@
+#!/usr/bin/env python
+
+"""%prog [options] shellpath dirpath
+
+Pulls performance data on parsing via the js shell.
+Displays the average number of milliseconds it took to parse each file.
+
+For comparison, something apparently approximating a t-test is performed:
+"Faster" means that:
+
+ t_baseline_goodrun = (t_baseline_avg - t_baseline_stddev)
+ t_current_badrun = (t_current_avg + t_current_stddev)
+ t_current_badrun < t_baseline_goodrun
+
+Effectively, a bad run from the current data is better than a good run from the
+baseline data, we're probably faster. A similar computation is used for
+determining the "slower" designation.
+
+Arguments:
+ shellpath executable JavaScript shell
+ dirpath directory filled with parsilicious js files
+"""
+
+import json
+import math
+import optparse
+import os
+import subprocess as subp
+import sys
+from string import Template
+
+try:
+ import compare_bench
+except ImportError:
+ compare_bench = None
+
+
+_DIR = os.path.dirname(__file__)
+JS_CODE_TEMPLATE = Template(
+ """
+if (typeof snarf !== 'undefined') read = snarf
+var contents = read("$filepath");
+$prepare
+for (var i = 0; i < $warmup_run_count; i++)
+ $func(contents, $options);
+var results = [];
+for (var i = 0; i < $real_run_count; i++) {
+ var start = elapsed() / 1000;
+ $func(contents, $options);
+ var end = elapsed() / 1000;
+ results.push(end - start);
+}
+print(results);
+"""
+)
+
+
+def gen_filepaths(dirpath, target_ext=".js"):
+ for filename in os.listdir(dirpath):
+ if filename.endswith(target_ext):
+ yield os.path.join(dirpath, filename)
+
+
+def avg(seq):
+ return sum(seq) / len(seq)
+
+
+def stddev(seq, mean):
+ diffs = ((float(item) - mean) ** 2 for item in seq)
+ return math.sqrt(sum(diffs) / len(seq))
+
+
+def bench(
+ shellpath, filepath, warmup_runs, counted_runs, prepare, func, options, stfu=False
+):
+ """Return a list of milliseconds for the counted runs."""
+ assert '"' not in filepath
+ code = JS_CODE_TEMPLATE.substitute(
+ filepath=filepath,
+ warmup_run_count=warmup_runs,
+ real_run_count=counted_runs,
+ prepare=prepare,
+ func=func,
+ options=options,
+ )
+ proc = subp.Popen([shellpath, "-e", code], stdout=subp.PIPE)
+ stdout, _ = proc.communicate()
+ milliseconds = [float(val) for val in stdout.decode().split(",")]
+ mean = avg(milliseconds)
+ sigma = stddev(milliseconds, mean)
+ if not stfu:
+ print("Runs:", [int(ms) for ms in milliseconds])
+ print("Mean:", mean)
+ print("Stddev: {:.2f} ({:.2f}% of mean)".format(sigma, sigma / mean * 100))
+ return mean, sigma
+
+
+def parsemark(filepaths, fbench, stfu=False):
+ """:param fbench: fbench(filename) -> float"""
+ bench_map = {} # {filename: (avg, stddev)}
+ for filepath in filepaths:
+ filename = os.path.split(filepath)[-1]
+ if not stfu:
+ print("Parsemarking {}...".format(filename))
+ bench_map[filename] = fbench(filepath)
+ print("{")
+ for i, (filename, (avg, stddev)) in enumerate(iter(bench_map.items())):
+ assert '"' not in filename
+ fmt = ' {:30s}: {{"average_ms": {:6.2f}, "stddev_ms": {:6.2f}}}'
+ if i != len(bench_map) - 1:
+ fmt += ","
+ filename_str = '"{}"'.format(filename)
+ print(fmt.format(filename_str, avg, stddev))
+ print("}")
+ return dict(
+ (filename, dict(average_ms=avg, stddev_ms=stddev))
+ for filename, (avg, stddev) in iter(bench_map.items())
+ )
+
+
+def main():
+ parser = optparse.OptionParser(usage=__doc__.strip())
+ parser.add_option(
+ "-w",
+ "--warmup-runs",
+ metavar="COUNT",
+ type=int,
+ default=5,
+ help="used to minimize test instability [%default]",
+ )
+ parser.add_option(
+ "-c",
+ "--counted-runs",
+ metavar="COUNT",
+ type=int,
+ default=50,
+ help="timed data runs that count towards the average" " [%default]",
+ )
+ parser.add_option(
+ "-s",
+ "--shell",
+ metavar="PATH",
+ help="explicit shell location; when omitted, will look" " in likely places",
+ )
+ parser.add_option(
+ "-b",
+ "--baseline",
+ metavar="JSON_PATH",
+ dest="baseline_path",
+ help="json file with baseline values to " "compare against",
+ )
+ parser.add_option(
+ "--mode",
+ dest="mode",
+ type="choice",
+ choices=("parse", "dumpStencil", "compile", "decode"),
+ default="parse",
+ help="The target of the benchmark (parse/dumpStencil/compile/decode), defaults to parse",
+ )
+ parser.add_option(
+ "--lazy",
+ dest="lazy",
+ action="store_true",
+ default=False,
+ help="Use lazy parsing when compiling",
+ )
+ parser.add_option(
+ "-q",
+ "--quiet",
+ dest="stfu",
+ action="store_true",
+ default=False,
+ help="only print JSON to stdout [%default]",
+ )
+ options, args = parser.parse_args()
+ try:
+ shellpath = args.pop(0)
+ except IndexError:
+ parser.print_help()
+ print()
+ print("error: shellpath required", file=sys.stderr)
+ return -1
+ try:
+ dirpath = args.pop(0)
+ except IndexError:
+ parser.print_help()
+ print()
+ print("error: dirpath required", file=sys.stderr)
+ return -1
+ if not shellpath or not os.path.exists(shellpath):
+ print("error: could not find shell:", shellpath, file=sys.stderr)
+ return -1
+ if options.baseline_path:
+ if not os.path.isfile(options.baseline_path):
+ print("error: baseline file does not exist", file=sys.stderr)
+ return -1
+ if not compare_bench:
+ print(
+ "error: JSON support is missing, cannot compare benchmarks",
+ file=sys.stderr,
+ )
+ return -1
+
+ if options.lazy and options.mode == "parse":
+ print(
+ "error: parse mode doesn't support lazy",
+ file=sys.stderr,
+ )
+ return -1
+
+ funcOpt = {}
+ if options.mode == "decode":
+ encodeOpt = {}
+ encodeOpt["execute"] = False
+ encodeOpt["saveIncrementalBytecode"] = True
+ if not options.lazy:
+ encodeOpt["forceFullParse"] = True
+
+ # In order to test the decoding, we first have to encode the content.
+ prepare = Template(
+ """
+contents = cacheEntry(contents);
+evaluate(contents, $options);
+"""
+ ).substitute(options=json.dumps(encodeOpt))
+
+ func = "evaluate"
+ funcOpt["execute"] = False
+ funcOpt["loadBytecode"] = True
+ if not options.lazy:
+ funcOpt["forceFullParse"] = True
+ else:
+ prepare = ""
+ func = options.mode
+ if not options.lazy:
+ funcOpt["forceFullParse"] = True
+
+ def benchfile(filepath):
+ return bench(
+ shellpath,
+ filepath,
+ options.warmup_runs,
+ options.counted_runs,
+ prepare,
+ func,
+ json.dumps(funcOpt),
+ stfu=options.stfu,
+ )
+
+ bench_map = parsemark(gen_filepaths(dirpath), benchfile, options.stfu)
+ if options.baseline_path:
+ compare_bench.compare_immediate(bench_map, options.baseline_path)
+ return 0
+
+
+if __name__ == "__main__":
+ sys.exit(main())