diff options
Diffstat (limited to '')
-rw-r--r-- | testing/mochitest/runtests.py | 3932 |
1 files changed, 3932 insertions, 0 deletions
diff --git a/testing/mochitest/runtests.py b/testing/mochitest/runtests.py new file mode 100644 index 0000000000..58251c2f8c --- /dev/null +++ b/testing/mochitest/runtests.py @@ -0,0 +1,3932 @@ +# This Source Code Form is subject to the terms of the Mozilla Public +# License, v. 2.0. If a copy of the MPL was not distributed with this +# file, You can obtain one at http://mozilla.org/MPL/2.0/. + +""" +Runs the Mochitest test harness. +""" + +import os +import sys + +SCRIPT_DIR = os.path.abspath(os.path.realpath(os.path.dirname(__file__))) +sys.path.insert(0, SCRIPT_DIR) + +import ctypes +import glob +import json +import numbers +import platform +import re +import shlex +import shutil +import signal +import socket +import subprocess +import sys +import tempfile +import time +import traceback +import uuid +import zipfile +from argparse import Namespace +from collections import defaultdict +from contextlib import closing +from ctypes.util import find_library +from datetime import datetime, timedelta +from distutils import spawn + +import bisection +import mozcrash +import mozdebug +import mozinfo +import mozprocess +import mozrunner +from manifestparser import TestManifest +from manifestparser.filters import ( + chunk_by_dir, + chunk_by_runtime, + chunk_by_slice, + failures, + pathprefix, + subsuite, + tags, +) +from manifestparser.util import normsep +from mozgeckoprofiler import symbolicate_profile_json, view_gecko_profile + +try: + from marionette_driver.addons import Addons + from marionette_driver.marionette import Marionette +except ImportError as e: # noqa + # Defer ImportError until attempt to use Marionette + def reraise(*args, **kwargs): + raise (e) # noqa + + Marionette = reraise + +import mozleak +from leaks import LSANLeaks, ShutdownLeaks +from mochitest_options import ( + MochitestArgumentParser, + build_obj, + get_default_valgrind_suppression_files, +) +from mozlog import commandline, get_proxy_logger +from mozprofile import Profile +from mozprofile.cli import KeyValueParseError, parse_key_value, parse_preferences +from mozprofile.permissions import ServerLocations +from mozrunner.utils import get_stack_fixer_function, test_environment +from mozscreenshot import dump_screen + +HAVE_PSUTIL = False +try: + import psutil + + HAVE_PSUTIL = True +except ImportError: + pass + +import six +from six.moves.urllib.parse import quote_plus as encodeURIComponent +from six.moves.urllib_request import urlopen + +here = os.path.abspath(os.path.dirname(__file__)) + +NO_TESTS_FOUND = """ +No tests were found for flavor '{}' and the following manifest filters: +{} + +Make sure the test paths (if any) are spelt correctly and the corresponding +--flavor and --subsuite are being used. See `mach mochitest --help` for a +list of valid flavors. +""".lstrip() + + +######################################## +# Option for MOZ (former NSPR) logging # +######################################## + +# Set the desired log modules you want a log be produced +# by a try run for, or leave blank to disable the feature. +# This will be passed to MOZ_LOG environment variable. +# Try run will then put a download link for a zip archive +# of all the log files on treeherder. +MOZ_LOG = "" + +##################### +# Test log handling # +##################### + +# output processing + + +class MessageLogger(object): + + """File-like object for logging messages (structured logs)""" + + BUFFERING_THRESHOLD = 100 + # This is a delimiter used by the JS side to avoid logs interleaving + DELIMITER = "\ue175\uee31\u2c32\uacbf" + BUFFERED_ACTIONS = set(["test_status", "log"]) + VALID_ACTIONS = set( + [ + "suite_start", + "suite_end", + "test_start", + "test_end", + "test_status", + "log", + "assertion_count", + "buffering_on", + "buffering_off", + ] + ) + # Regexes that will be replaced with an empty string if found in a test + # name. We do this to normalize test names which may contain URLs and test + # package prefixes. + TEST_PATH_PREFIXES = [ + r"^/tests/", + r"^\w+://[\w\.]+(:\d+)?(/\w+)?/(tests?|a11y|chrome|browser)/", + ] + + def __init__(self, logger, buffering=True, structured=True): + self.logger = logger + self.structured = structured + self.gecko_id = "GECKO" + self.is_test_running = False + + # Even if buffering is enabled, we only want to buffer messages between + # TEST-START/TEST-END. So it is off to begin, but will be enabled after + # a TEST-START comes in. + self._buffering = False + self.restore_buffering = buffering + + # Guard to ensure we never buffer if this value was initially `False` + self._buffering_initially_enabled = buffering + + # Message buffering + self.buffered_messages = [] + + def validate(self, obj): + """Tests whether the given object is a valid structured message + (only does a superficial validation)""" + if not ( + isinstance(obj, dict) + and "action" in obj + and obj["action"] in MessageLogger.VALID_ACTIONS + ): + raise ValueError + + def _fix_subtest_name(self, message): + """Make sure subtest name is a string""" + if "subtest" in message and not isinstance( + message["subtest"], six.string_types + ): + message["subtest"] = str(message["subtest"]) + + def _fix_test_name(self, message): + """Normalize a logged test path to match the relative path from the sourcedir.""" + if message.get("test") is not None: + test = message["test"] + for pattern in MessageLogger.TEST_PATH_PREFIXES: + test = re.sub(pattern, "", test) + if test != message["test"]: + message["test"] = test + break + + def _fix_message_format(self, message): + if "message" in message: + if isinstance(message["message"], bytes): + message["message"] = message["message"].decode("utf-8", "replace") + elif not isinstance(message["message"], six.text_type): + message["message"] = six.text_type(message["message"]) + + def parse_line(self, line): + """Takes a given line of input (structured or not) and + returns a list of structured messages""" + if isinstance(line, six.binary_type): + # if line is a sequence of bytes, let's decode it + line = line.rstrip().decode("UTF-8", "replace") + else: + # line is in unicode - so let's use it as it is + line = line.rstrip() + + messages = [] + for fragment in line.split(MessageLogger.DELIMITER): + if not fragment: + continue + try: + message = json.loads(fragment) + self.validate(message) + except ValueError: + if self.structured: + message = dict( + action="process_output", + process=self.gecko_id, + data=fragment, + ) + else: + message = dict( + action="log", + level="info", + message=fragment, + ) + + self._fix_subtest_name(message) + self._fix_test_name(message) + self._fix_message_format(message) + messages.append(message) + + return messages + + @property + def buffering(self): + if not self._buffering_initially_enabled: + return False + return self._buffering + + @buffering.setter + def buffering(self, val): + self._buffering = val + + def process_message(self, message): + """Processes a structured message. Takes into account buffering, errors, ...""" + # Activation/deactivating message buffering from the JS side + if message["action"] == "buffering_on": + if self.is_test_running: + self.buffering = True + return + if message["action"] == "buffering_off": + self.buffering = False + return + + # Error detection also supports "raw" errors (in log messages) because some tests + # manually dump 'TEST-UNEXPECTED-FAIL'. + if "expected" in message or ( + message["action"] == "log" + and message.get("message", "").startswith("TEST-UNEXPECTED") + ): + self.restore_buffering = self.restore_buffering or self.buffering + self.buffering = False + if self.buffered_messages: + snipped = len(self.buffered_messages) - self.BUFFERING_THRESHOLD + if snipped > 0: + self.logger.info( + "<snipped {0} output lines - " + "if you need more context, please use " + "SimpleTest.requestCompleteLog() in your test>".format(snipped) + ) + # Dumping previously buffered messages + self.dump_buffered(limit=True) + + # Logging the error message + self.logger.log_raw(message) + # Determine if message should be buffered + elif ( + self.buffering + and self.structured + and message["action"] in self.BUFFERED_ACTIONS + ): + self.buffered_messages.append(message) + # Otherwise log the message directly + else: + self.logger.log_raw(message) + + # If a test ended, we clean the buffer + if message["action"] == "test_end": + self.is_test_running = False + self.buffered_messages = [] + self.restore_buffering = self.restore_buffering or self.buffering + self.buffering = False + + if message["action"] == "test_start": + self.is_test_running = True + if self.restore_buffering: + self.restore_buffering = False + self.buffering = True + + def write(self, line): + messages = self.parse_line(line) + for message in messages: + self.process_message(message) + return messages + + def flush(self): + sys.stdout.flush() + + def dump_buffered(self, limit=False): + if limit: + dumped_messages = self.buffered_messages[-self.BUFFERING_THRESHOLD :] + else: + dumped_messages = self.buffered_messages + + last_timestamp = None + for buf in dumped_messages: + # pylint --py3k W1619 + timestamp = datetime.fromtimestamp(buf["time"] / 1000).strftime("%H:%M:%S") + if timestamp != last_timestamp: + self.logger.info("Buffered messages logged at {}".format(timestamp)) + last_timestamp = timestamp + + self.logger.log_raw(buf) + self.logger.info("Buffered messages finished") + # Cleaning the list of buffered messages + self.buffered_messages = [] + + def finish(self): + self.dump_buffered() + self.buffering = False + self.logger.suite_end() + + +#################### +# PROCESS HANDLING # +#################### + + +def call(*args, **kwargs): + """front-end function to mozprocess.ProcessHandler""" + # TODO: upstream -> mozprocess + # https://bugzilla.mozilla.org/show_bug.cgi?id=791383 + log = get_proxy_logger("mochitest") + + def on_output(line): + log.process_output( + process=process.pid, + data=line.decode("utf8", "replace"), + command=process.commandline, + ) + + process = mozprocess.ProcessHandlerMixin( + *args, processOutputLine=on_output, **kwargs + ) + process.run() + return process.wait() + + +def killPid(pid, log): + # see also https://bugzilla.mozilla.org/show_bug.cgi?id=911249#c58 + + if HAVE_PSUTIL: + # Kill a process tree (including grandchildren) with signal.SIGTERM + if pid == os.getpid(): + raise RuntimeError("Error: trying to kill ourselves, not another process") + try: + parent = psutil.Process(pid) + children = parent.children(recursive=True) + children.append(parent) + for p in children: + p.send_signal(signal.SIGTERM) + gone, alive = psutil.wait_procs(children, timeout=30) + for p in gone: + log.info("psutil found pid %s dead" % p.pid) + for p in alive: + log.info("failed to kill pid %d after 30s" % p.pid) + except Exception as e: + log.info("Error: Failed to kill process %d: %s" % (pid, str(e))) + else: + try: + os.kill(pid, getattr(signal, "SIGKILL", signal.SIGTERM)) + except Exception as e: + log.info("Failed to kill process %d: %s" % (pid, str(e))) + + +if mozinfo.isWin: + import ctypes.wintypes + + def isPidAlive(pid): + STILL_ACTIVE = 259 + PROCESS_QUERY_LIMITED_INFORMATION = 0x1000 + pHandle = ctypes.windll.kernel32.OpenProcess( + PROCESS_QUERY_LIMITED_INFORMATION, 0, pid + ) + if not pHandle: + return False + + try: + pExitCode = ctypes.wintypes.DWORD() + ctypes.windll.kernel32.GetExitCodeProcess(pHandle, ctypes.byref(pExitCode)) + + if pExitCode.value != STILL_ACTIVE: + return False + + # We have a live process handle. But Windows aggressively + # re-uses pids, so let's attempt to verify that this is + # actually Firefox. + namesize = 1024 + pName = ctypes.create_string_buffer(namesize) + namelen = ctypes.windll.psapi.GetProcessImageFileNameA( + pHandle, pName, namesize + ) + if namelen == 0: + # Still an active process, so conservatively assume it's Firefox. + return True + + return pName.value.endswith((b"firefox.exe", b"plugin-container.exe")) + finally: + ctypes.windll.kernel32.CloseHandle(pHandle) + + +else: + import errno + + def isPidAlive(pid): + try: + # kill(pid, 0) checks for a valid PID without actually sending a signal + # The method throws OSError if the PID is invalid, which we catch + # below. + os.kill(pid, 0) + + # Wait on it to see if it's a zombie. This can throw OSError.ECHILD if + # the process terminates before we get to this point. + wpid, wstatus = os.waitpid(pid, os.WNOHANG) + return wpid == 0 + except OSError as err: + # Catch the errors we might expect from os.kill/os.waitpid, + # and re-raise any others + if err.errno in (errno.ESRCH, errno.ECHILD, errno.EPERM): + return False + raise + + +# TODO: ^ upstream isPidAlive to mozprocess + +####################### +# HTTP SERVER SUPPORT # +####################### + + +class MochitestServer(object): + + "Web server used to serve Mochitests, for closer fidelity to the real web." + + def __init__(self, options, logger): + if isinstance(options, Namespace): + options = vars(options) + self._log = logger + self._keep_open = bool(options["keep_open"]) + self._utilityPath = options["utilityPath"] + self._xrePath = options["xrePath"] + self._profileDir = options["profilePath"] + self.webServer = options["webServer"] + self.httpPort = options["httpPort"] + if options.get("remoteWebServer") == "10.0.2.2": + # probably running an Android emulator and 10.0.2.2 will + # not be visible from host + shutdownServer = "127.0.0.1" + else: + shutdownServer = self.webServer + self.shutdownURL = "http://%(server)s:%(port)s/server/shutdown" % { + "server": shutdownServer, + "port": self.httpPort, + } + self.testPrefix = "undefined" + + if options.get("httpdPath"): + self._httpdPath = options["httpdPath"] + else: + self._httpdPath = SCRIPT_DIR + self._httpdPath = os.path.abspath(self._httpdPath) + + def start(self): + "Run the Mochitest server, returning the process ID of the server." + + # get testing environment + env = test_environment(xrePath=self._xrePath, log=self._log) + env["XPCOM_DEBUG_BREAK"] = "warn" + if "LD_LIBRARY_PATH" not in env or env["LD_LIBRARY_PATH"] is None: + env["LD_LIBRARY_PATH"] = self._xrePath + else: + env["LD_LIBRARY_PATH"] = ":".join([self._xrePath, env["LD_LIBRARY_PATH"]]) + + # When running with an ASan build, our xpcshell server will also be ASan-enabled, + # thus consuming too much resources when running together with the browser on + # the test machines. Try to limit the amount of resources by disabling certain + # features. + env["ASAN_OPTIONS"] = "quarantine_size=1:redzone=32:malloc_context_size=5" + + # Likewise, when running with a TSan build, our xpcshell server will + # also be TSan-enabled. Except that in this case, we don't really + # care about races in xpcshell. So disable TSan for the server. + env["TSAN_OPTIONS"] = "report_bugs=0" + + # Don't use socket process for the xpcshell server. + env["MOZ_DISABLE_SOCKET_PROCESS"] = "1" + + if mozinfo.isWin: + env["PATH"] = env["PATH"] + ";" + str(self._xrePath) + + args = [ + "-g", + self._xrePath, + "-f", + os.path.join(self._httpdPath, "httpd.js"), + "-e", + "const _PROFILE_PATH = '%(profile)s'; const _SERVER_PORT = '%(port)s'; " + "const _SERVER_ADDR = '%(server)s'; const _TEST_PREFIX = %(testPrefix)s; " + "const _DISPLAY_RESULTS = %(displayResults)s;" + % { + "profile": self._profileDir.replace("\\", "\\\\"), + "port": self.httpPort, + "server": self.webServer, + "testPrefix": self.testPrefix, + "displayResults": str(self._keep_open).lower(), + }, + "-f", + os.path.join(SCRIPT_DIR, "server.js"), + ] + + xpcshell = os.path.join( + self._utilityPath, "xpcshell" + mozinfo.info["bin_suffix"] + ) + command = [xpcshell] + args + self._process = mozprocess.ProcessHandler(command, cwd=SCRIPT_DIR, env=env) + self._process.run() + self._log.info("%s : launching %s" % (self.__class__.__name__, command)) + pid = self._process.pid + self._log.info("runtests.py | Server pid: %d" % pid) + + def ensureReady(self, timeout): + assert timeout >= 0 + + aliveFile = os.path.join(self._profileDir, "server_alive.txt") + i = 0 + while i < timeout: + if os.path.exists(aliveFile): + break + time.sleep(0.05) + i += 0.05 + else: + self._log.error( + "TEST-UNEXPECTED-FAIL | runtests.py | Timed out while waiting for server startup." + ) + self.stop() + sys.exit(1) + + def stop(self): + try: + with closing(urlopen(self.shutdownURL)) as c: + self._log.info(six.ensure_text(c.read())) + except Exception: + self._log.info("Failed to stop web server on %s" % self.shutdownURL) + traceback.print_exc() + finally: + if self._process is not None: + # Kill the server immediately to avoid logging intermittent + # shutdown crashes, sometimes observed on Windows 10. + self._process.kill() + self._log.info("Web server killed.") + + +class WebSocketServer(object): + + "Class which encapsulates the mod_pywebsocket server" + + def __init__(self, options, scriptdir, logger, debuggerInfo=None): + self.port = options.webSocketPort + self.debuggerInfo = debuggerInfo + self._log = logger + self._scriptdir = scriptdir + + def start(self): + # Invoke pywebsocket through a wrapper which adds special SIGINT handling. + # + # If we're in an interactive debugger, the wrapper causes the server to + # ignore SIGINT so the server doesn't capture a ctrl+c meant for the + # debugger. + # + # If we're not in an interactive debugger, the wrapper causes the server to + # die silently upon receiving a SIGINT. + scriptPath = "pywebsocket_wrapper.py" + script = os.path.join(self._scriptdir, scriptPath) + + cmd = [sys.executable, script] + if self.debuggerInfo and self.debuggerInfo.interactive: + cmd += ["--interactive"] + cmd += [ + "-H", + "127.0.0.1", + "-p", + str(self.port), + "-w", + self._scriptdir, + "-l", + os.path.join(self._scriptdir, "websock.log"), + "--log-level=debug", + "--allow-handlers-outside-root-dir", + ] + env = dict(os.environ) + env["PYTHONPATH"] = os.pathsep.join(sys.path) + # Start the process. Ignore stderr so that exceptions from the server + # are not treated as failures when parsing the test log. + self._process = mozprocess.ProcessHandler( + cmd, + cwd=SCRIPT_DIR, + env=env, + processStderrLine=lambda _: None, + ) + self._process.run() + pid = self._process.pid + self._log.info("runtests.py | Websocket server pid: %d" % pid) + + def stop(self): + if self._process is not None: + self._process.kill() + + +class SSLTunnel: + def __init__(self, options, logger): + self.log = logger + self.process = None + self.utilityPath = options.utilityPath + self.xrePath = options.xrePath + self.certPath = options.certPath + self.sslPort = options.sslPort + self.httpPort = options.httpPort + self.webServer = options.webServer + self.webSocketPort = options.webSocketPort + + self.customCertRE = re.compile("^cert=(?P<nickname>[0-9a-zA-Z_ ]+)") + self.clientAuthRE = re.compile("^clientauth=(?P<clientauth>[a-z]+)") + self.redirRE = re.compile("^redir=(?P<redirhost>[0-9a-zA-Z_ .]+)") + + def writeLocation(self, config, loc): + for option in loc.options: + match = self.customCertRE.match(option) + if match: + customcert = match.group("nickname") + config.write( + "listen:%s:%s:%s:%s\n" + % (loc.host, loc.port, self.sslPort, customcert) + ) + + match = self.clientAuthRE.match(option) + if match: + clientauth = match.group("clientauth") + config.write( + "clientauth:%s:%s:%s:%s\n" + % (loc.host, loc.port, self.sslPort, clientauth) + ) + + match = self.redirRE.match(option) + if match: + redirhost = match.group("redirhost") + config.write( + "redirhost:%s:%s:%s:%s\n" + % (loc.host, loc.port, self.sslPort, redirhost) + ) + + if option in ( + "tls1", + "tls1_1", + "tls1_2", + "tls1_3", + "ssl3", + "3des", + "failHandshake", + ): + config.write( + "%s:%s:%s:%s\n" % (option, loc.host, loc.port, self.sslPort) + ) + + def buildConfig(self, locations, public=None): + """Create the ssltunnel configuration file""" + configFd, self.configFile = tempfile.mkstemp(prefix="ssltunnel", suffix=".cfg") + with os.fdopen(configFd, "w") as config: + config.write("httpproxy:1\n") + config.write("certdbdir:%s\n" % self.certPath) + config.write("forward:127.0.0.1:%s\n" % self.httpPort) + config.write( + "websocketserver:%s:%s\n" % (self.webServer, self.webSocketPort) + ) + # Use "*" to tell ssltunnel to listen on the public ip + # address instead of the loopback address 127.0.0.1. This + # may have the side-effect of causing firewall warnings on + # macOS and Windows. Use "127.0.0.1" to listen on the + # loopback address. Remote tests using physical or + # emulated Android devices must use the public ip address + # in order for the sslproxy to work but Desktop tests + # which run on the same host as ssltunnel may use the + # loopback address. + listen_address = "*" if public else "127.0.0.1" + config.write("listen:%s:%s:pgoserver\n" % (listen_address, self.sslPort)) + + for loc in locations: + if loc.scheme == "https" and "nocert" not in loc.options: + self.writeLocation(config, loc) + + def start(self): + """Starts the SSL Tunnel""" + + # start ssltunnel to provide https:// URLs capability + ssltunnel = os.path.join(self.utilityPath, "ssltunnel") + if os.name == "nt": + ssltunnel += ".exe" + if not os.path.exists(ssltunnel): + self.log.error( + "INFO | runtests.py | expected to find ssltunnel at %s" % ssltunnel + ) + exit(1) + + env = test_environment(xrePath=self.xrePath, log=self.log) + env["LD_LIBRARY_PATH"] = self.xrePath + self.process = mozprocess.ProcessHandler([ssltunnel, self.configFile], env=env) + self.process.run() + self.log.info("runtests.py | SSL tunnel pid: %d" % self.process.pid) + + def stop(self): + """Stops the SSL Tunnel and cleans up""" + if self.process is not None: + self.process.kill() + if os.path.exists(self.configFile): + os.remove(self.configFile) + + +def checkAndConfigureV4l2loopback(device): + """ + Determine if a given device path is a v4l2loopback device, and if so + toggle a few settings on it via fcntl. Very linux-specific. + + Returns (status, device name) where status is a boolean. + """ + if not mozinfo.isLinux: + return False, "" + + libc = ctypes.cdll.LoadLibrary(find_library("c")) + O_RDWR = 2 + # These are from linux/videodev2.h + + class v4l2_capability(ctypes.Structure): + _fields_ = [ + ("driver", ctypes.c_char * 16), + ("card", ctypes.c_char * 32), + ("bus_info", ctypes.c_char * 32), + ("version", ctypes.c_uint32), + ("capabilities", ctypes.c_uint32), + ("device_caps", ctypes.c_uint32), + ("reserved", ctypes.c_uint32 * 3), + ] + + VIDIOC_QUERYCAP = 0x80685600 + + fd = libc.open(six.ensure_binary(device), O_RDWR) + if fd < 0: + return False, "" + + vcap = v4l2_capability() + if libc.ioctl(fd, VIDIOC_QUERYCAP, ctypes.byref(vcap)) != 0: + return False, "" + + if six.ensure_text(vcap.driver) != "v4l2 loopback": + return False, "" + + class v4l2_control(ctypes.Structure): + _fields_ = [("id", ctypes.c_uint32), ("value", ctypes.c_int32)] + + # These are private v4l2 control IDs, see: + # https://github.com/umlaeute/v4l2loopback/blob/fd822cf0faaccdf5f548cddd9a5a3dcebb6d584d/v4l2loopback.c#L131 + KEEP_FORMAT = 0x8000000 + SUSTAIN_FRAMERATE = 0x8000001 + VIDIOC_S_CTRL = 0xC008561C + + control = v4l2_control() + control.id = KEEP_FORMAT + control.value = 1 + libc.ioctl(fd, VIDIOC_S_CTRL, ctypes.byref(control)) + + control.id = SUSTAIN_FRAMERATE + control.value = 1 + libc.ioctl(fd, VIDIOC_S_CTRL, ctypes.byref(control)) + libc.close(fd) + + return True, six.ensure_text(vcap.card) + + +def findTestMediaDevices(log): + """ + Find the test media devices configured on this system, and return a dict + containing information about them. The dict will have keys for 'audio' + and 'video', each containing the name of the media device to use. + + If audio and video devices could not be found, return None. + + This method is only currently implemented for Linux. + """ + if not mozinfo.isLinux: + return None + + info = {} + # Look for a v4l2loopback device. + name = None + device = None + for dev in sorted(glob.glob("/dev/video*")): + result, name_ = checkAndConfigureV4l2loopback(dev) + if result: + name = name_ + device = dev + break + + if not (name and device): + log.error("Couldn't find a v4l2loopback video device") + return None + + # Feed it a frame of output so it has something to display + gst01 = spawn.find_executable("gst-launch-0.1") + gst010 = spawn.find_executable("gst-launch-0.10") + gst10 = spawn.find_executable("gst-launch-1.0") + if gst01: + gst = gst01 + if gst010: + gst = gst010 + else: + gst = gst10 + process = mozprocess.ProcessHandler( + [ + gst, + "--no-fault", + "videotestsrc", + "pattern=green", + "num-buffers=1", + "!", + "v4l2sink", + "device=%s" % device, + ] + ) + process.run() + info["video"] = {"name": name, "process": process} + + # check if PulseAudio module-null-sink is loaded + pactl = spawn.find_executable("pactl") + + if not pactl: + log.error("Could not find pactl on system") + return None + + try: + o = subprocess.check_output([pactl, "list", "short", "modules"]) + except subprocess.CalledProcessError: + log.error("Could not list currently loaded modules") + return None + + null_sink = [x for x in o.splitlines() if b"module-null-sink" in x] + + if not null_sink: + try: + subprocess.check_call([pactl, "load-module", "module-null-sink"]) + except subprocess.CalledProcessError: + log.error("Could not load module-null-sink") + return None + + # Hardcode the name since it's always the same. + info["audio"] = {"name": "Monitor of Null Output"} + return info + + +def create_zip(path): + """ + Takes a `path` on disk and creates a zipfile with its contents. Returns a + path to the location of the temporary zip file. + """ + with tempfile.NamedTemporaryFile() as f: + # `shutil.make_archive` writes to "{f.name}.zip", so we're really just + # using `NamedTemporaryFile` as a way to get a random path. + return shutil.make_archive(f.name, "zip", path) + + +def update_mozinfo(): + """walk up directories to find mozinfo.json update the info""" + # TODO: This should go in a more generic place, e.g. mozinfo + + path = SCRIPT_DIR + dirs = set() + while path != os.path.expanduser("~"): + if path in dirs: + break + dirs.add(path) + path = os.path.split(path)[0] + + mozinfo.find_and_update_from_json(*dirs) + + +class MochitestDesktop(object): + """ + Mochitest class for desktop firefox. + """ + + oldcwd = os.getcwd() + + # Path to the test script on the server + TEST_PATH = "tests" + CHROME_PATH = "redirect.html" + + certdbNew = False + sslTunnel = None + DEFAULT_TIMEOUT = 60.0 + mediaDevices = None + mozinfo_variables_shown = False + + patternFiles = {} + + # XXX use automation.py for test name to avoid breaking legacy + # TODO: replace this with 'runtests.py' or 'mochitest' or the like + test_name = "automation.py" + + def __init__(self, flavor, logger_options, staged_addons=None, quiet=False): + update_mozinfo() + self.flavor = flavor + self.staged_addons = staged_addons + self.server = None + self.wsserver = None + self.websocketProcessBridge = None + self.sslTunnel = None + self.manifest = None + self.tests_by_manifest = defaultdict(list) + self.args_by_manifest = defaultdict(set) + self.prefs_by_manifest = defaultdict(set) + self.env_vars_by_manifest = defaultdict(set) + self.tests_dirs_by_manifest = defaultdict(set) + self._active_tests = None + self.currentTests = None + self._locations = None + self.browserEnv = None + + self.marionette = None + self.start_script = None + self.mozLogs = None + self.start_script_kwargs = {} + self.extraArgs = [] + self.extraPrefs = {} + self.extraEnv = {} + self.extraTestsDirs = [] + self.conditioned_profile_dir = None + + if logger_options.get("log"): + self.log = logger_options["log"] + else: + self.log = commandline.setup_logging( + "mochitest", logger_options, {"tbpl": sys.stdout} + ) + + self.message_logger = MessageLogger( + logger=self.log, buffering=quiet, structured=True + ) + + # Max time in seconds to wait for server startup before tests will fail -- if + # this seems big, it's mostly for debug machines where cold startup + # (particularly after a build) takes forever. + self.SERVER_STARTUP_TIMEOUT = 180 if mozinfo.info.get("debug") else 90 + + # metro browser sub process id + self.browserProcessId = None + + self.haveDumpedScreen = False + # Create variables to count the number of passes, fails, todos. + self.countpass = 0 + self.countfail = 0 + self.counttodo = 0 + + self.expectedError = {} + self.result = {} + + self.start_script = os.path.join(here, "start_desktop.js") + + # Used to temporarily serve a performance profile + self.profiler_tempdir = None + + def environment(self, **kwargs): + kwargs["log"] = self.log + return test_environment(**kwargs) + + def getFullPath(self, path): + "Get an absolute path relative to self.oldcwd." + return os.path.normpath(os.path.join(self.oldcwd, os.path.expanduser(path))) + + def getLogFilePath(self, logFile): + """return the log file path relative to the device we are testing on, in most cases + it will be the full path on the local system + """ + return self.getFullPath(logFile) + + @property + def locations(self): + if self._locations is not None: + return self._locations + locations_file = os.path.join(SCRIPT_DIR, "server-locations.txt") + self._locations = ServerLocations(locations_file) + return self._locations + + def buildURLOptions(self, options, env): + """Add test control options from the command line to the url + + URL parameters to test URL: + + autorun -- kick off tests automatically + closeWhenDone -- closes the browser after the tests + hideResultsTable -- hides the table of individual test results + logFile -- logs test run to an absolute path + startAt -- name of test to start at + endAt -- name of test to end at + timeout -- per-test timeout in seconds + repeat -- How many times to repeat the test, ie: repeat=1 will run the test twice. + """ + self.urlOpts = [] + + if not hasattr(options, "logFile"): + options.logFile = "" + if not hasattr(options, "fileLevel"): + options.fileLevel = "INFO" + + # allow relative paths for logFile + if options.logFile: + options.logFile = self.getLogFilePath(options.logFile) + + if options.flavor in ("a11y", "browser", "chrome"): + self.makeTestConfig(options) + else: + if options.autorun: + self.urlOpts.append("autorun=1") + if options.timeout: + self.urlOpts.append("timeout=%d" % options.timeout) + if options.maxTimeouts: + self.urlOpts.append("maxTimeouts=%d" % options.maxTimeouts) + if not options.keep_open: + self.urlOpts.append("closeWhenDone=1") + if options.logFile: + self.urlOpts.append("logFile=" + encodeURIComponent(options.logFile)) + self.urlOpts.append( + "fileLevel=" + encodeURIComponent(options.fileLevel) + ) + if options.consoleLevel: + self.urlOpts.append( + "consoleLevel=" + encodeURIComponent(options.consoleLevel) + ) + if options.startAt: + self.urlOpts.append("startAt=%s" % options.startAt) + if options.endAt: + self.urlOpts.append("endAt=%s" % options.endAt) + if options.shuffle: + self.urlOpts.append("shuffle=1") + if "MOZ_HIDE_RESULTS_TABLE" in env and env["MOZ_HIDE_RESULTS_TABLE"] == "1": + self.urlOpts.append("hideResultsTable=1") + if options.runUntilFailure: + self.urlOpts.append("runUntilFailure=1") + if options.repeat: + self.urlOpts.append("repeat=%d" % options.repeat) + if len(options.test_paths) == 1 and os.path.isfile( + os.path.join( + self.oldcwd, + os.path.dirname(__file__), + self.TEST_PATH, + options.test_paths[0], + ) + ): + self.urlOpts.append( + "testname=%s" % "/".join([self.TEST_PATH, options.test_paths[0]]) + ) + if options.manifestFile: + self.urlOpts.append("manifestFile=%s" % options.manifestFile) + if options.failureFile: + self.urlOpts.append( + "failureFile=%s" % self.getFullPath(options.failureFile) + ) + if options.runSlower: + self.urlOpts.append("runSlower=true") + if options.debugOnFailure: + self.urlOpts.append("debugOnFailure=true") + if options.dumpOutputDirectory: + self.urlOpts.append( + "dumpOutputDirectory=%s" + % encodeURIComponent(options.dumpOutputDirectory) + ) + if options.dumpAboutMemoryAfterTest: + self.urlOpts.append("dumpAboutMemoryAfterTest=true") + if options.dumpDMDAfterTest: + self.urlOpts.append("dumpDMDAfterTest=true") + if options.debugger: + self.urlOpts.append("interactiveDebugger=true") + if options.jscov_dir_prefix: + self.urlOpts.append("jscovDirPrefix=%s" % options.jscov_dir_prefix) + if options.cleanupCrashes: + self.urlOpts.append("cleanupCrashes=true") + if "MOZ_XORIGIN_MOCHITEST" in env and env["MOZ_XORIGIN_MOCHITEST"] == "1": + options.xOriginTests = True + if options.xOriginTests: + self.urlOpts.append("xOriginTests=true") + + def normflavor(self, flavor): + """ + In some places the string 'browser-chrome' is expected instead of + 'browser' and 'mochitest' instead of 'plain'. Normalize the flavor + strings for those instances. + """ + # TODO Use consistent flavor strings everywhere and remove this + if flavor == "browser": + return "browser-chrome" + elif flavor == "plain": + return "mochitest" + return flavor + + # This check can be removed when bug 983867 is fixed. + def isTest(self, options, filename): + allow_js_css = False + if options.flavor == "browser": + allow_js_css = True + testPattern = re.compile(r"browser_.+\.js") + elif options.flavor in ("a11y", "chrome"): + testPattern = re.compile(r"(browser|test)_.+\.(xul|html|js|xhtml)") + else: + testPattern = re.compile(r"test_") + + if not allow_js_css and (".js" in filename or ".css" in filename): + return False + + pathPieces = filename.split("/") + + return testPattern.match(pathPieces[-1]) and not re.search( + r"\^headers\^$", filename + ) + + def setTestRoot(self, options): + if options.flavor != "plain": + self.testRoot = options.flavor + else: + self.testRoot = self.TEST_PATH + self.testRootAbs = os.path.join(SCRIPT_DIR, self.testRoot) + + def buildTestURL(self, options, scheme="http"): + if scheme == "https": + testHost = "https://example.com:443" + elif options.xOriginTests: + testHost = "http://mochi.xorigin-test:8888" + else: + testHost = "http://mochi.test:8888" + testURL = "/".join([testHost, self.TEST_PATH]) + + if len(options.test_paths) == 1: + if os.path.isfile( + os.path.join( + self.oldcwd, + os.path.dirname(__file__), + self.TEST_PATH, + options.test_paths[0], + ) + ): + testURL = "/".join([testURL, os.path.dirname(options.test_paths[0])]) + else: + testURL = "/".join([testURL, options.test_paths[0]]) + + if options.flavor in ("a11y", "chrome"): + testURL = "/".join([testHost, self.CHROME_PATH]) + elif options.flavor == "browser": + testURL = "about:blank" + return testURL + + def parseAndCreateTestsDirs(self, m): + testsDirs = list(self.tests_dirs_by_manifest[m])[0] + self.extraTestsDirs = [] + if testsDirs: + self.extraTestsDirs = testsDirs.strip().split() + self.log.info( + "The following extra test directories will be created:\n {}".format( + "\n ".join(self.extraTestsDirs) + ) + ) + self.createExtraTestsDirs(self.extraTestsDirs, m) + + def createExtraTestsDirs(self, extraTestsDirs=None, manifest=None): + """Take a list of directories that might be needed to exist by the test + prior to even the main process be executed, and: + - verify it does not already exists + - create it if it does + Removal of those directories is handled in cleanup() + """ + if type(extraTestsDirs) != list: + return + + for d in extraTestsDirs: + if os.path.exists(d): + raise FileExistsError( + "Directory '{}' already exists. This is a member of " + "test-directories in manifest {}.".format(d, manifest) + ) + + created = [] + for d in extraTestsDirs: + os.makedirs(d) + created += [d] + + if created != extraTestsDirs: + raise EnvironmentError( + "Not all directories were created: extraTestsDirs={} -- created={}".format( + extraTestsDirs, created + ) + ) + + def getTestsByScheme( + self, options, testsToFilter=None, disabled=True, manifestToFilter=None + ): + """Build the url path to the specific test harness and test file or directory + Build a manifest of tests to run and write out a json file for the harness to read + testsToFilter option is used to filter/keep the tests provided in the list + + disabled -- This allows to add all disabled tests on the build side + and then on the run side to only run the enabled ones + """ + + tests = self.getActiveTests(options, disabled) + paths = [] + for test in tests: + if testsToFilter and (test["path"] not in testsToFilter): + continue + # If we are running a specific manifest, the previously computed set of active + # tests should be filtered out based on the manifest that contains that entry. + # + # This is especially important when a test file is listed in multiple + # manifests (e.g. because the same test runs under a different configuration, + # and so it is being included in multiple manifests), without filtering the + # active tests based on the current manifest (configuration) that we are + # running for each of the N manifests we would be executing the active tests + # exactly N times (and so NxN runs instead of the expected N runs, one for each + # manifest). + if manifestToFilter and (test["manifest"] not in manifestToFilter): + continue + paths.append(test) + + # Generate test by schemes + for (scheme, grouped_tests) in self.groupTestsByScheme(paths).items(): + # Bug 883865 - add this functionality into manifestparser + with open( + os.path.join(SCRIPT_DIR, options.testRunManifestFile), "w" + ) as manifestFile: + manifestFile.write(json.dumps({"tests": grouped_tests})) + options.manifestFile = options.testRunManifestFile + yield (scheme, grouped_tests) + + def startWebSocketServer(self, options, debuggerInfo): + """Launch the websocket server""" + self.wsserver = WebSocketServer(options, SCRIPT_DIR, self.log, debuggerInfo) + self.wsserver.start() + + def startWebServer(self, options): + """Create the webserver and start it up""" + + self.server = MochitestServer(options, self.log) + self.server.start() + + if options.pidFile != "": + with open(options.pidFile + ".xpcshell.pid", "w") as f: + f.write("%s" % self.server._process.pid) + + def startWebsocketProcessBridge(self, options): + """Create a websocket server that can launch various processes that + JS needs (eg; ICE server for webrtc testing) + """ + + command = [ + sys.executable, + os.path.join("websocketprocessbridge", "websocketprocessbridge.py"), + "--port", + options.websocket_process_bridge_port, + ] + self.websocketProcessBridge = mozprocess.ProcessHandler(command, cwd=SCRIPT_DIR) + self.websocketProcessBridge.run() + self.log.info( + "runtests.py | websocket/process bridge pid: %d" + % self.websocketProcessBridge.pid + ) + + # ensure the server is up, wait for at most ten seconds + for i in range(1, 100): + if self.websocketProcessBridge.proc.poll() is not None: + self.log.error( + "runtests.py | websocket/process bridge failed " + "to launch. Are all the dependencies installed?" + ) + return + + try: + sock = socket.create_connection(("127.0.0.1", 8191)) + sock.close() + break + except Exception: + time.sleep(0.1) + else: + self.log.error( + "runtests.py | Timed out while waiting for " + "websocket/process bridge startup." + ) + + def needsWebsocketProcessBridge(self, options): + """ + Returns a bool indicating if the current test configuration needs + to start the websocket process bridge or not. The boils down to if + WebRTC tests that need the bridge are present. + """ + tests = self.getActiveTests(options) + is_webrtc_tag_present = False + for test in tests: + if "webrtc" in test.get("tags", ""): + is_webrtc_tag_present = True + break + return is_webrtc_tag_present and options.subsuite in ["media"] + + def startServers(self, options, debuggerInfo, public=None): + # start servers and set ports + # TODO: pass these values, don't set on `self` + self.webServer = options.webServer + self.httpPort = options.httpPort + self.sslPort = options.sslPort + self.webSocketPort = options.webSocketPort + + # httpd-path is specified by standard makefile targets and may be specified + # on the command line to select a particular version of httpd.js. If not + # specified, try to select the one from hostutils.zip, as required in + # bug 882932. + if not options.httpdPath: + options.httpdPath = os.path.join(options.utilityPath, "components") + + self.startWebServer(options) + self.startWebSocketServer(options, debuggerInfo) + + # Only webrtc mochitests in the media suite need the websocketprocessbridge. + if self.needsWebsocketProcessBridge(options): + self.startWebsocketProcessBridge(options) + + # start SSL pipe + self.sslTunnel = SSLTunnel(options, logger=self.log) + self.sslTunnel.buildConfig(self.locations, public=public) + self.sslTunnel.start() + + # If we're lucky, the server has fully started by now, and all paths are + # ready, etc. However, xpcshell cold start times suck, at least for debug + # builds. We'll try to connect to the server for awhile, and if we fail, + # we'll try to kill the server and exit with an error. + if self.server is not None: + self.server.ensureReady(self.SERVER_STARTUP_TIMEOUT) + + def stopServers(self): + """Servers are no longer needed, and perhaps more importantly, anything they + might spew to console might confuse things.""" + if self.server is not None: + try: + self.log.info("Stopping web server") + self.server.stop() + except Exception: + self.log.critical("Exception when stopping web server") + + if self.wsserver is not None: + try: + self.log.info("Stopping web socket server") + self.wsserver.stop() + except Exception: + self.log.critical("Exception when stopping web socket server") + + if self.sslTunnel is not None: + try: + self.log.info("Stopping ssltunnel") + self.sslTunnel.stop() + except Exception: + self.log.critical("Exception stopping ssltunnel") + + if self.websocketProcessBridge is not None: + try: + self.websocketProcessBridge.kill() + self.websocketProcessBridge.wait() + self.log.info("Stopping websocket/process bridge") + except Exception: + self.log.critical("Exception stopping websocket/process bridge") + + if hasattr(self, "gstForV4l2loopbackProcess"): + try: + self.gstForV4l2loopbackProcess.kill() + self.gstForV4l2loopbackProcess.wait() + self.log.info("Stopping gst for v4l2loopback") + except Exception: + self.log.critical("Exception stopping gst for v4l2loopback") + + def copyExtraFilesToProfile(self, options): + "Copy extra files or dirs specified on the command line to the testing profile." + for f in options.extraProfileFiles: + abspath = self.getFullPath(f) + if os.path.isfile(abspath): + shutil.copy2(abspath, options.profilePath) + elif os.path.isdir(abspath): + dest = os.path.join(options.profilePath, os.path.basename(abspath)) + shutil.copytree(abspath, dest) + else: + self.log.warning("runtests.py | Failed to copy %s to profile" % abspath) + + def getChromeTestDir(self, options): + dir = os.path.join(os.path.abspath("."), SCRIPT_DIR) + "/" + if mozinfo.isWin: + dir = "file:///" + dir.replace("\\", "/") + return dir + + def writeChromeManifest(self, options): + manifest = os.path.join(options.profilePath, "tests.manifest") + with open(manifest, "w") as manifestFile: + # Register chrome directory. + chrometestDir = self.getChromeTestDir(options) + manifestFile.write( + "content mochitests %s contentaccessible=yes\n" % chrometestDir + ) + manifestFile.write( + "content mochitests-any %s contentaccessible=yes remoteenabled=yes\n" + % chrometestDir + ) + manifestFile.write( + "content mochitests-content %s contentaccessible=yes remoterequired=yes\n" + % chrometestDir + ) + + if options.testingModulesDir is not None: + manifestFile.write( + "resource testing-common file:///%s\n" % options.testingModulesDir + ) + if options.store_chrome_manifest: + shutil.copyfile(manifest, options.store_chrome_manifest) + return manifest + + def addChromeToProfile(self, options): + "Adds MochiKit chrome tests to the profile." + + # Create (empty) chrome directory. + chromedir = os.path.join(options.profilePath, "chrome") + os.mkdir(chromedir) + + # Write userChrome.css. + chrome = """ +/* set default namespace to XUL */ +@namespace url("http://www.mozilla.org/keymaster/gatekeeper/there.is.only.xul"); +toolbar, +toolbarpalette { + background-color: rgb(235, 235, 235) !important; +} +toolbar#nav-bar { + background-image: none !important; +} +""" + with open( + os.path.join(options.profilePath, "userChrome.css"), "a" + ) as chromeFile: + chromeFile.write(chrome) + + manifest = self.writeChromeManifest(options) + + return manifest + + def getExtensionsToInstall(self, options): + "Return a list of extensions to install in the profile" + extensions = [] + appDir = ( + options.app[: options.app.rfind(os.sep)] + if options.app + else options.utilityPath + ) + + extensionDirs = [ + # Extensions distributed with the test harness. + os.path.normpath(os.path.join(SCRIPT_DIR, "extensions")), + ] + if appDir: + # Extensions distributed with the application. + extensionDirs.append(os.path.join(appDir, "distribution", "extensions")) + + for extensionDir in extensionDirs: + if os.path.isdir(extensionDir): + for dirEntry in os.listdir(extensionDir): + if dirEntry not in options.extensionsToExclude: + path = os.path.join(extensionDir, dirEntry) + if os.path.isdir(path) or ( + os.path.isfile(path) and path.endswith(".xpi") + ): + extensions.append(path) + extensions.extend(options.extensionsToInstall) + return extensions + + def logPreamble(self, tests): + """Logs a suite_start message and test_start/test_end at the beginning of a run.""" + self.log.suite_start( + self.tests_by_manifest, name="mochitest-{}".format(self.flavor) + ) + for test in tests: + if "disabled" in test: + self.log.test_start(test["path"]) + self.log.test_end(test["path"], "SKIP", message=test["disabled"]) + + def loadFailurePatternFile(self, pat_file): + if pat_file in self.patternFiles: + return self.patternFiles[pat_file] + if not os.path.isfile(pat_file): + self.log.warning( + "runtests.py | Cannot find failure pattern file " + pat_file + ) + return None + + # Using ":error" to ensure it shows up in the failure summary. + self.log.warning( + "[runtests.py:error] Using {} to filter failures. If there " + "is any number mismatch below, you could have fixed " + "something documented in that file. Please reduce the " + "failure count appropriately.".format(pat_file) + ) + patternRE = re.compile( + r""" + ^\s*\*\s* # list bullet + (test_\S+|\.{3}) # test name + (?:\s*(`.+?`|asserts))? # failure pattern + (?::.+)? # optional description + \s*\[(\d+|\*)\] # expected count + \s*$ + """, + re.X, + ) + patterns = {} + with open(pat_file) as f: + last_name = None + for line in f: + match = patternRE.match(line) + if not match: + continue + name = match.group(1) + name = last_name if name == "..." else name + last_name = name + pat = match.group(2) + if pat is not None: + pat = "ASSERTION" if pat == "asserts" else pat[1:-1] + count = match.group(3) + count = None if count == "*" else int(count) + if name not in patterns: + patterns[name] = [] + patterns[name].append((pat, count)) + self.patternFiles[pat_file] = patterns + return patterns + + def getFailurePatterns(self, pat_file, test_name): + patterns = self.loadFailurePatternFile(pat_file) + if patterns: + return patterns.get(test_name, None) + + def getActiveTests(self, options, disabled=True): + """ + This method is used to parse the manifest and return active filtered tests. + """ + if self._active_tests: + return self._active_tests + + tests = [] + manifest = self.getTestManifest(options) + if manifest: + if options.extra_mozinfo_json: + mozinfo.update(options.extra_mozinfo_json) + + info = mozinfo.info + + filters = [ + subsuite(options.subsuite), + ] + + # Allow for only running tests/manifests which match this tag + if options.conditionedProfile: + if not options.test_tags: + options.test_tags = [] + options.test_tags.append("condprof") + + if options.test_tags: + filters.append(tags(options.test_tags)) + + if options.test_paths: + options.test_paths = self.normalize_paths(options.test_paths) + filters.append(pathprefix(options.test_paths)) + + # Add chunking filters if specified + if options.totalChunks: + if options.chunkByDir: + filters.append( + chunk_by_dir( + options.thisChunk, options.totalChunks, options.chunkByDir + ) + ) + elif options.chunkByRuntime: + if mozinfo.info["os"] == "android": + platkey = "android" + elif mozinfo.isWin: + platkey = "windows" + else: + platkey = "unix" + + runtime_file = os.path.join( + SCRIPT_DIR, + "runtimes", + "manifest-runtimes-{}.json".format(platkey), + ) + if not os.path.exists(runtime_file): + self.log.error("runtime file %s not found!" % runtime_file) + sys.exit(1) + + # Given the mochitest flavor, load the runtimes information + # for only that flavor due to manifest runtime format change in Bug 1637463. + with open(runtime_file, "r") as f: + if "suite_name" in options: + runtimes = json.load(f).get(options.suite_name, {}) + else: + runtimes = {} + + filters.append( + chunk_by_runtime( + options.thisChunk, options.totalChunks, runtimes + ) + ) + else: + filters.append( + chunk_by_slice(options.thisChunk, options.totalChunks) + ) + + noDefaultFilters = False + if options.runFailures: + filters.append(failures(options.runFailures)) + noDefaultFilters = True + + tests = manifest.active_tests( + exists=False, + disabled=disabled, + filters=filters, + noDefaultFilters=noDefaultFilters, + **info + ) + + if len(tests) == 0: + self.log.error( + NO_TESTS_FOUND.format(options.flavor, manifest.fmt_filters()) + ) + + paths = [] + for test in tests: + if len(tests) == 1 and "disabled" in test: + del test["disabled"] + + pathAbs = os.path.abspath(test["path"]) + assert os.path.normcase(pathAbs).startswith( + os.path.normcase(self.testRootAbs) + ) + tp = pathAbs[len(self.testRootAbs) :].replace("\\", "/").strip("/") + + if not self.isTest(options, tp): + self.log.warning( + "Warning: %s from manifest %s is not a valid test" + % (test["name"], test["manifest"]) + ) + continue + + manifest_key = test["manifest_relpath"] + # Ignore ancestor_manifests that live at the root (e.g, don't have a + # path separator). + if "ancestor_manifest" in test and "/" in normsep( + test["ancestor_manifest"] + ): + manifest_key = "{}:{}".format(test["ancestor_manifest"], manifest_key) + + self.tests_by_manifest[manifest_key.replace("\\", "/")].append(tp) + self.args_by_manifest[manifest_key].add(test.get("args")) + self.prefs_by_manifest[manifest_key].add(test.get("prefs")) + self.env_vars_by_manifest[manifest_key].add(test.get("environment")) + self.tests_dirs_by_manifest[manifest_key].add(test.get("test-directories")) + + for key in ["args", "prefs", "environment", "test-directories"]: + if key in test and not options.runByManifest and "disabled" not in test: + self.log.error( + "parsing {}: runByManifest mode must be enabled to " + "set the `{}` key".format(test["manifest_relpath"], key) + ) + sys.exit(1) + + testob = {"path": tp, "manifest": manifest_key} + if "disabled" in test: + testob["disabled"] = test["disabled"] + if "expected" in test: + testob["expected"] = test["expected"] + if "https_first_disabled" in test: + testob["https_first_disabled"] = test["https_first_disabled"] == "true" + if "allow_xul_xbl" in test: + testob["allow_xul_xbl"] = test["allow_xul_xbl"] == "true" + if "scheme" in test: + testob["scheme"] = test["scheme"] + if "tags" in test: + testob["tags"] = test["tags"] + if options.failure_pattern_file: + pat_file = os.path.join( + os.path.dirname(test["manifest"]), options.failure_pattern_file + ) + patterns = self.getFailurePatterns(pat_file, test["name"]) + if patterns: + testob["expected"] = patterns + paths.append(testob) + + # The 'args' key needs to be set in the DEFAULT section, unfortunately + # we can't tell what comes from DEFAULT or not. So to validate this, we + # stash all args from tests in the same manifest into a set. If the + # length of the set > 1, then we know 'args' didn't come from DEFAULT. + args_not_default = [ + m for m, p in six.iteritems(self.args_by_manifest) if len(p) > 1 + ] + if args_not_default: + self.log.error( + "The 'args' key must be set in the DEFAULT section of a " + "manifest. Fix the following manifests: {}".format( + "\n".join(args_not_default) + ) + ) + sys.exit(1) + + # The 'prefs' key needs to be set in the DEFAULT section too. + pref_not_default = [ + m for m, p in six.iteritems(self.prefs_by_manifest) if len(p) > 1 + ] + if pref_not_default: + self.log.error( + "The 'prefs' key must be set in the DEFAULT section of a " + "manifest. Fix the following manifests: {}".format( + "\n".join(pref_not_default) + ) + ) + sys.exit(1) + # The 'environment' key needs to be set in the DEFAULT section too. + env_not_default = [ + m for m, p in six.iteritems(self.env_vars_by_manifest) if len(p) > 1 + ] + if env_not_default: + self.log.error( + "The 'environment' key must be set in the DEFAULT section of a " + "manifest. Fix the following manifests: {}".format( + "\n".join(env_not_default) + ) + ) + sys.exit(1) + + paths.sort(key=lambda p: p["path"].split("/")) + if options.dump_tests: + options.dump_tests = os.path.expanduser(options.dump_tests) + assert os.path.exists(os.path.dirname(options.dump_tests)) + with open(options.dump_tests, "w") as dumpFile: + dumpFile.write(json.dumps({"active_tests": paths})) + + self.log.info("Dumping active_tests to %s file." % options.dump_tests) + sys.exit() + + # Upload a list of test manifests that were executed in this run. + if "MOZ_UPLOAD_DIR" in os.environ: + artifact = os.path.join(os.environ["MOZ_UPLOAD_DIR"], "manifests.list") + with open(artifact, "a") as fh: + fh.write("\n".join(sorted(self.tests_by_manifest.keys()))) + + self._active_tests = paths + return self._active_tests + + def getTestManifest(self, options): + if isinstance(options.manifestFile, TestManifest): + manifest = options.manifestFile + elif options.manifestFile and os.path.isfile(options.manifestFile): + manifestFileAbs = os.path.abspath(options.manifestFile) + assert manifestFileAbs.startswith(SCRIPT_DIR) + manifest = TestManifest([options.manifestFile], strict=False) + elif options.manifestFile and os.path.isfile( + os.path.join(SCRIPT_DIR, options.manifestFile) + ): + manifestFileAbs = os.path.abspath( + os.path.join(SCRIPT_DIR, options.manifestFile) + ) + assert manifestFileAbs.startswith(SCRIPT_DIR) + manifest = TestManifest([manifestFileAbs], strict=False) + else: + masterName = self.normflavor(options.flavor) + ".ini" + masterPath = os.path.join(SCRIPT_DIR, self.testRoot, masterName) + + if os.path.exists(masterPath): + manifest = TestManifest([masterPath], strict=False) + else: + manifest = None + self.log.warning( + "TestManifest masterPath %s does not exist" % masterPath + ) + + return manifest + + def makeTestConfig(self, options): + "Creates a test configuration file for customizing test execution." + options.logFile = options.logFile.replace("\\", "\\\\") + + if ( + "MOZ_HIDE_RESULTS_TABLE" in os.environ + and os.environ["MOZ_HIDE_RESULTS_TABLE"] == "1" + ): + options.hideResultsTable = True + + # strip certain unnecessary items to avoid serialization errors in json.dumps() + d = dict( + (k, v) + for k, v in options.__dict__.items() + if (v is None) or isinstance(v, (six.string_types, numbers.Number)) + ) + d["testRoot"] = self.testRoot + if options.jscov_dir_prefix: + d["jscovDirPrefix"] = options.jscov_dir_prefix + if not options.keep_open: + d["closeWhenDone"] = "1" + + d["runFailures"] = False + if options.runFailures: + d["runFailures"] = True + content = json.dumps(d) + + with open(os.path.join(options.profilePath, "testConfig.js"), "w") as config: + config.write(content) + + def buildBrowserEnv(self, options, debugger=False, env=None): + """build the environment variables for the specific test and operating system""" + if mozinfo.info["asan"] and mozinfo.isLinux and mozinfo.bits == 64: + useLSan = True + else: + useLSan = False + + browserEnv = self.environment( + xrePath=options.xrePath, env=env, debugger=debugger, useLSan=useLSan + ) + + if hasattr(options, "topsrcdir"): + browserEnv["MOZ_DEVELOPER_REPO_DIR"] = options.topsrcdir + if hasattr(options, "topobjdir"): + browserEnv["MOZ_DEVELOPER_OBJ_DIR"] = options.topobjdir + + if options.headless: + browserEnv["MOZ_HEADLESS"] = "1" + + if options.dmd: + browserEnv["DMD"] = os.environ.get("DMD", "1") + + # bug 1443327: do not set MOZ_CRASHREPORTER_SHUTDOWN during browser-chrome + # tests, since some browser-chrome tests test content process crashes; + # also exclude non-e10s since at least one non-e10s mochitest is problematic + if ( + options.flavor == "browser" or not options.e10s + ) and "MOZ_CRASHREPORTER_SHUTDOWN" in browserEnv: + del browserEnv["MOZ_CRASHREPORTER_SHUTDOWN"] + + try: + browserEnv.update( + dict( + parse_key_value( + self.extraEnv, context="environment variable in manifest" + ) + ) + ) + except KeyValueParseError as e: + self.log.error(str(e)) + return None + + # These variables are necessary for correct application startup; change + # via the commandline at your own risk. + browserEnv["XPCOM_DEBUG_BREAK"] = "stack" + + # interpolate environment passed with options + try: + browserEnv.update( + dict(parse_key_value(options.environment, context="--setenv")) + ) + except KeyValueParseError as e: + self.log.error(str(e)) + return None + + if ( + "MOZ_PROFILER_STARTUP_FEATURES" not in browserEnv + or "nativeallocations" + not in browserEnv["MOZ_PROFILER_STARTUP_FEATURES"].split(",") + ): + # Only turn on the bloat log if the profiler's native allocation feature is + # not enabled. The two are not compatible. + browserEnv["XPCOM_MEM_BLOAT_LOG"] = self.leak_report_file + + # If profiling options are enabled, turn on the gecko profiler by using the + # profiler environmental variables. + if options.profiler: + # The user wants to capture a profile, and automatically view it. The + # profile will be saved to a temporary folder, then deleted after + # opening in profiler.firefox.com. + self.profiler_tempdir = tempfile.mkdtemp() + browserEnv["MOZ_PROFILER_SHUTDOWN"] = os.path.join( + self.profiler_tempdir, "mochitest-profile.json" + ) + browserEnv["MOZ_PROFILER_STARTUP"] = "1" + + if options.profilerSaveOnly: + # The user wants to capture a profile, but only to save it. This defaults + # to the MOZ_UPLOAD_DIR. + browserEnv["MOZ_PROFILER_STARTUP"] = "1" + if "MOZ_UPLOAD_DIR" in browserEnv: + browserEnv["MOZ_PROFILER_SHUTDOWN"] = os.path.join( + browserEnv["MOZ_UPLOAD_DIR"], "mochitest-profile.json" + ) + else: + self.log.error( + "--profiler-save-only was specified, but no MOZ_UPLOAD_DIR " + "environment variable was provided. Please set this " + "environment variable to a directory path in order to save " + "a performance profile." + ) + return None + + try: + gmp_path = self.getGMPPluginPath(options) + if gmp_path is not None: + browserEnv["MOZ_GMP_PATH"] = gmp_path + except EnvironmentError: + self.log.error("Could not find path to gmp-fake plugin!") + return None + + if options.fatalAssertions: + browserEnv["XPCOM_DEBUG_BREAK"] = "stack-and-abort" + + # Produce a mozlog, if setup (see MOZ_LOG global at the top of + # this script). + self.mozLogs = MOZ_LOG and "MOZ_UPLOAD_DIR" in os.environ + if self.mozLogs: + browserEnv["MOZ_LOG"] = MOZ_LOG + + return browserEnv + + def killNamedProc(self, pname, orphans=True): + """Kill processes matching the given command name""" + self.log.info("Checking for %s processes..." % pname) + + if HAVE_PSUTIL: + for proc in psutil.process_iter(): + try: + if proc.name() == pname: + procd = proc.as_dict(attrs=["pid", "ppid", "name", "username"]) + if proc.ppid() == 1 or not orphans: + self.log.info("killing %s" % procd) + killPid(proc.pid, self.log) + else: + self.log.info("NOT killing %s (not an orphan?)" % procd) + except Exception as e: + self.log.info( + "Warning: Unable to kill process %s: %s" % (pname, str(e)) + ) + # may not be able to access process info for all processes + continue + else: + + def _psInfo(line): + if pname in line: + self.log.info(line) + + process = mozprocess.ProcessHandler( + ["ps", "-f"], processOutputLine=_psInfo, universal_newlines=True + ) + process.run() + process.wait() + + def _psKill(line): + parts = line.split() + if len(parts) == 3 and parts[0].isdigit(): + pid = int(parts[0]) + ppid = int(parts[1]) + if parts[2] == pname: + if ppid == 1 or not orphans: + self.log.info("killing %s (pid %d)" % (pname, pid)) + killPid(pid, self.log) + else: + self.log.info( + "NOT killing %s (pid %d) (not an orphan?)" + % (pname, pid) + ) + + process = mozprocess.ProcessHandler( + ["ps", "-o", "pid,ppid,comm"], + processOutputLine=_psKill, + universal_newlines=True, + ) + process.run() + process.wait() + + def execute_start_script(self): + if not self.start_script or not self.marionette: + return + + if os.path.isfile(self.start_script): + with open(self.start_script, "r") as fh: + script = fh.read() + else: + script = self.start_script + + with self.marionette.using_context("chrome"): + return self.marionette.execute_script( + script, script_args=(self.start_script_kwargs,) + ) + + def fillCertificateDB(self, options): + # TODO: move -> mozprofile: + # https://bugzilla.mozilla.org/show_bug.cgi?id=746243#c35 + + pwfilePath = os.path.join(options.profilePath, ".crtdbpw") + with open(pwfilePath, "w") as pwfile: + pwfile.write("\n") + + # Pre-create the certification database for the profile + env = self.environment(xrePath=options.xrePath) + env["LD_LIBRARY_PATH"] = options.xrePath + bin_suffix = mozinfo.info.get("bin_suffix", "") + certutil = os.path.join(options.utilityPath, "certutil" + bin_suffix) + pk12util = os.path.join(options.utilityPath, "pk12util" + bin_suffix) + toolsEnv = env + if mozinfo.info["asan"]: + # Disable leak checking when running these tools + toolsEnv["ASAN_OPTIONS"] = "detect_leaks=0" + if mozinfo.info["tsan"]: + # Disable race checking when running these tools + toolsEnv["TSAN_OPTIONS"] = "report_bugs=0" + + if self.certdbNew: + # android uses the new DB formats exclusively + certdbPath = "sql:" + options.profilePath + else: + # desktop seems to use the old + certdbPath = options.profilePath + + # certutil.exe depends on some DLLs in the app directory + # When running tests against an MSIX-installed Firefox, these DLLs + # cannot be used out of the install directory, they must be copied + # elsewhere first. + if "WindowsApps" in options.app: + install_dir = os.path.dirname(options.app) + for f in os.listdir(install_dir): + if f.endswith(".dll"): + shutil.copy(os.path.join(install_dir, f), options.utilityPath) + + status = call( + [certutil, "-N", "-d", certdbPath, "-f", pwfilePath], env=toolsEnv + ) + if status: + return status + + # Walk the cert directory and add custom CAs and client certs + files = os.listdir(options.certPath) + for item in files: + root, ext = os.path.splitext(item) + if ext == ".ca": + trustBits = "CT,," + if root.endswith("-object"): + trustBits = "CT,,CT" + call( + [ + certutil, + "-A", + "-i", + os.path.join(options.certPath, item), + "-d", + certdbPath, + "-f", + pwfilePath, + "-n", + root, + "-t", + trustBits, + ], + env=toolsEnv, + ) + elif ext == ".client": + call( + [ + pk12util, + "-i", + os.path.join(options.certPath, item), + "-w", + pwfilePath, + "-d", + certdbPath, + ], + env=toolsEnv, + ) + + os.unlink(pwfilePath) + return 0 + + def proxy(self, options): + # proxy + # use SSL port for legacy compatibility; see + # - https://bugzilla.mozilla.org/show_bug.cgi?id=688667#c66 + # - https://bugzilla.mozilla.org/show_bug.cgi?id=899221 + # - https://github.com/mozilla/mozbase/commit/43f9510e3d58bfed32790c82a57edac5f928474d + # 'ws': str(self.webSocketPort) + return { + "remote": options.webServer, + "http": options.httpPort, + "https": options.sslPort, + "ws": options.sslPort, + } + + def merge_base_profiles(self, options, category): + """Merge extra profile data from testing/profiles.""" + + # In test packages used in CI, the profile_data directory is installed + # in the SCRIPT_DIR. + profile_data_dir = os.path.join(SCRIPT_DIR, "profile_data") + # If possible, read profile data from topsrcdir. This prevents us from + # requiring a re-build to pick up newly added extensions in the + # <profile>/extensions directory. + if build_obj: + path = os.path.join(build_obj.topsrcdir, "testing", "profiles") + if os.path.isdir(path): + profile_data_dir = path + # Still not found? Look for testing/profiles relative to testing/mochitest. + if not os.path.isdir(profile_data_dir): + path = os.path.abspath(os.path.join(SCRIPT_DIR, "..", "profiles")) + if os.path.isdir(path): + profile_data_dir = path + + with open(os.path.join(profile_data_dir, "profiles.json"), "r") as fh: + base_profiles = json.load(fh)[category] + + # values to use when interpolating preferences + interpolation = { + "server": "%s:%s" % (options.webServer, options.httpPort), + } + + for profile in base_profiles: + path = os.path.join(profile_data_dir, profile) + self.profile.merge(path, interpolation=interpolation) + + @property + def conditioned_profile_copy(self): + """Returns a copy of the original conditioned profile that was created.""" + + condprof_copy = os.path.join(tempfile.mkdtemp(), "profile") + shutil.copytree( + self.conditioned_profile_dir, + condprof_copy, + ignore=shutil.ignore_patterns("lock"), + ) + self.log.info("Created a conditioned-profile copy: %s" % condprof_copy) + return condprof_copy + + def downloadConditionedProfile(self, profile_scenario, app): + from condprof.client import get_profile + from condprof.util import get_current_platform, get_version + + if self.conditioned_profile_dir: + # We already have a directory, so provide a copy that + # will get deleted after it's done with + return self.conditioned_profile_copy + + temp_download_dir = tempfile.mkdtemp() + + # Call condprof's client API to yield our platform-specific + # conditioned-profile binary + platform = get_current_platform() + + if not profile_scenario: + profile_scenario = "settled" + + version = get_version(app) + try: + cond_prof_target_dir = get_profile( + temp_download_dir, + platform, + profile_scenario, + repo="mozilla-central", + version=version, + retries=2, # quicker failure + ) + except Exception: + if version is None: + # any other error is a showstopper + self.log.critical("Could not get the conditioned profile") + traceback.print_exc() + raise + version = None + try: + self.log.info("retrying a profile with no version specified") + cond_prof_target_dir = get_profile( + temp_download_dir, + platform, + profile_scenario, + repo="mozilla-central", + version=version, + ) + except Exception: + self.log.critical("Could not get the conditioned profile") + traceback.print_exc() + raise + + # Now get the full directory path to our fetched conditioned profile + self.conditioned_profile_dir = os.path.join( + temp_download_dir, cond_prof_target_dir + ) + if not os.path.exists(cond_prof_target_dir): + self.log.critical( + "Can't find target_dir {}, from get_profile()" + "temp_download_dir {}, platform {}, scenario {}".format( + cond_prof_target_dir, temp_download_dir, platform, profile_scenario + ) + ) + raise OSError + + self.log.info( + "Original self.conditioned_profile_dir is now set: {}".format( + self.conditioned_profile_dir + ) + ) + return self.conditioned_profile_copy + + def buildProfile(self, options): + """create the profile and add optional chrome bits and files if requested""" + # get extensions to install + extensions = self.getExtensionsToInstall(options) + + # Whitelist the _tests directory (../..) so that TESTING_JS_MODULES work + tests_dir = os.path.dirname(os.path.dirname(SCRIPT_DIR)) + sandbox_whitelist_paths = [tests_dir] + options.sandboxReadWhitelist + if platform.system() == "Linux" or platform.system() in ( + "Windows", + "Microsoft", + ): + # Trailing slashes are needed to indicate directories on Linux and Windows + sandbox_whitelist_paths = [ + os.path.join(p, "") for p in sandbox_whitelist_paths + ] + + if options.conditionedProfile: + if options.profilePath and os.path.exists(options.profilePath): + shutil.rmtree(options.profilePath, ignore_errors=True) + options.profilePath = self.downloadConditionedProfile("full", options.app) + + # This is causing `certutil -N -d -f`` to not use -f (pwd file) + try: + os.remove(os.path.join(options.profilePath, "key4.db")) + except Exception as e: + self.log.info( + "Caught exception while removing key4.db" + "during setup of conditioned profile: %s" % e + ) + + # Create the profile + self.profile = Profile( + profile=options.profilePath, + addons=extensions, + locations=self.locations, + proxy=self.proxy(options), + whitelistpaths=sandbox_whitelist_paths, + ) + + # Fix options.profilePath for legacy consumers. + options.profilePath = self.profile.profile + + manifest = self.addChromeToProfile(options) + self.copyExtraFilesToProfile(options) + + # create certificate database for the profile + # TODO: this should really be upstreamed somewhere, maybe mozprofile + certificateStatus = self.fillCertificateDB(options) + if certificateStatus: + self.log.error( + "TEST-UNEXPECTED-FAIL | runtests.py | Certificate integration failed" + ) + return None + + # Set preferences in the following order (latter overrides former): + # 1) Preferences from base profile (e.g from testing/profiles) + # 2) Prefs hardcoded in this function + # 3) Prefs from --setpref + + # Prefs from base profiles + self.merge_base_profiles(options, "mochitest") + + # Hardcoded prefs (TODO move these into a base profile) + prefs = { + "browser.tabs.remote.autostart": options.e10s, + # Enable tracing output for detailed failures in case of + # failing connection attempts, and hangs (bug 1397201) + "marionette.log.level": "Trace", + # Disable async font fallback, because the unpredictable + # extra reflow it can trigger (potentially affecting a later + # test) results in spurious intermittent failures. + "gfx.font_rendering.fallback.async": False, + } + + if options.flavor == "browser" and options.timeout: + prefs["testing.browserTestHarness.timeout"] = options.timeout + + # browser-chrome tests use a fairly short default timeout of 45 seconds; + # this is sometimes too short on asan and debug, where we expect reduced + # performance. + if ( + (mozinfo.info["asan"] or mozinfo.info["debug"]) + and options.flavor == "browser" + and options.timeout is None + ): + self.log.info("Increasing default timeout to 90 seconds") + prefs["testing.browserTestHarness.timeout"] = 90 + + # tsan builds need even more time + if ( + mozinfo.info["tsan"] + and options.flavor == "browser" + and options.timeout is None + ): + self.log.info("Increasing default timeout to 120 seconds") + prefs["testing.browserTestHarness.timeout"] = 120 + + if mozinfo.info["os"] == "win" and mozinfo.info["processor"] == "aarch64": + extended_timeout = self.DEFAULT_TIMEOUT * 4 + self.log.info( + "Increasing default timeout to {} seconds".format(extended_timeout) + ) + prefs["testing.browserTestHarness.timeout"] = extended_timeout + + if getattr(self, "testRootAbs", None): + prefs["mochitest.testRoot"] = self.testRootAbs + + # See if we should use fake media devices. + if options.useTestMediaDevices: + prefs["media.audio_loopback_dev"] = self.mediaDevices["audio"]["name"] + prefs["media.video_loopback_dev"] = self.mediaDevices["video"]["name"] + prefs["media.cubeb.output_device"] = "Null Output" + prefs["media.volume_scale"] = "1.0" + self.gstForV4l2loopbackProcess = self.mediaDevices["video"]["process"] + + self.profile.set_preferences(prefs) + + # Extra prefs from --setpref + self.profile.set_preferences(self.extraPrefs) + return manifest + + def getGMPPluginPath(self, options): + if options.gmp_path: + return options.gmp_path + + gmp_parentdirs = [ + # For local builds, GMP plugins will be under dist/bin. + options.xrePath, + # For packaged builds, GMP plugins will get copied under + # $profile/plugins. + os.path.join(self.profile.profile, "plugins"), + ] + + gmp_subdirs = [ + os.path.join("gmp-fake", "1.0"), + os.path.join("gmp-fakeopenh264", "1.0"), + os.path.join("gmp-clearkey", "0.1"), + ] + + gmp_paths = [ + os.path.join(parent, sub) + for parent in gmp_parentdirs + for sub in gmp_subdirs + if os.path.isdir(os.path.join(parent, sub)) + ] + + if not gmp_paths: + # This is fatal for desktop environments. + raise EnvironmentError("Could not find test gmp plugins") + + return os.pathsep.join(gmp_paths) + + def cleanup(self, options, final=False): + """remove temporary files, profile and virtual audio input device""" + if hasattr(self, "manifest") and self.manifest is not None: + if os.path.exists(self.manifest): + os.remove(self.manifest) + if hasattr(self, "profile"): + del self.profile + if hasattr(self, "extraTestsDirs"): + for d in self.extraTestsDirs: + if os.path.exists(d): + shutil.rmtree(d) + if options.pidFile != "" and os.path.exists(options.pidFile): + try: + os.remove(options.pidFile) + if os.path.exists(options.pidFile + ".xpcshell.pid"): + os.remove(options.pidFile + ".xpcshell.pid") + except Exception: + self.log.warning( + "cleaning up pidfile '%s' was unsuccessful from the test harness" + % options.pidFile + ) + options.manifestFile = None + + if hasattr(self, "virtualInputDeviceIdList"): + pactl = spawn.find_executable("pactl") + + if not pactl: + self.log.error("Could not find pactl on system") + return None + + for id in self.virtualInputDeviceIdList: + try: + subprocess.check_call([pactl, "unload-module", str(id)]) + except subprocess.CalledProcessError: + self.log.error( + "Could not remove pulse module with id {}".format(id) + ) + return None + + self.virtualInputDeviceIdList = [] + + def dumpScreen(self, utilityPath): + if self.haveDumpedScreen: + self.log.info( + "Not taking screenshot here: see the one that was previously logged" + ) + return + self.haveDumpedScreen = True + dump_screen(utilityPath, self.log) + + def killAndGetStack(self, processPID, utilityPath, debuggerInfo, dump_screen=False): + """ + Kill the process, preferrably in a way that gets us a stack trace. + Also attempts to obtain a screenshot before killing the process + if specified. + """ + self.log.info("Killing process: %s" % processPID) + if dump_screen: + self.dumpScreen(utilityPath) + + if mozinfo.info.get("crashreporter", True) and not debuggerInfo: + try: + minidump_path = os.path.join(self.profile.profile, "minidumps") + mozcrash.kill_and_get_minidump(processPID, minidump_path, utilityPath) + except OSError: + # https://bugzilla.mozilla.org/show_bug.cgi?id=921509 + self.log.info("Can't trigger Breakpad, process no longer exists") + return + self.log.info("Can't trigger Breakpad, just killing process") + killPid(processPID, self.log) + + def extract_child_pids(self, process_log, parent_pid=None): + """Parses the given log file for the pids of any processes launched by + the main process and returns them as a list. + If parent_pid is provided, and psutil is available, returns children of + parent_pid according to psutil. + """ + rv = [] + if parent_pid and HAVE_PSUTIL: + self.log.info("Determining child pids from psutil...") + try: + rv = [p.pid for p in psutil.Process(parent_pid).children()] + self.log.info(str(rv)) + except psutil.NoSuchProcess: + self.log.warning("Failed to lookup children of pid %d" % parent_pid) + + rv = set(rv) + pid_re = re.compile(r"==> process \d+ launched child process (\d+)") + with open(process_log) as fd: + for line in fd: + self.log.info(line.rstrip()) + m = pid_re.search(line) + if m: + rv.add(int(m.group(1))) + return rv + + def checkForZombies(self, processLog, utilityPath, debuggerInfo): + """Look for hung processes""" + + if not os.path.exists(processLog): + self.log.info("Automation Error: PID log not found: %s" % processLog) + # Whilst no hung process was found, the run should still display as + # a failure + return True + + # scan processLog for zombies + self.log.info("zombiecheck | Reading PID log: %s" % processLog) + processList = self.extract_child_pids(processLog) + # kill zombies + foundZombie = False + for processPID in processList: + self.log.info( + "zombiecheck | Checking for orphan process with PID: %d" % processPID + ) + if isPidAlive(processPID): + foundZombie = True + self.log.error( + "TEST-UNEXPECTED-FAIL | zombiecheck | child process " + "%d still alive after shutdown" % processPID + ) + self.killAndGetStack( + processPID, utilityPath, debuggerInfo, dump_screen=not debuggerInfo + ) + + return foundZombie + + def checkForRunningBrowsers(self): + firefoxes = "" + if HAVE_PSUTIL: + attrs = ["pid", "ppid", "name", "cmdline", "username"] + for proc in psutil.process_iter(): + try: + if "firefox" in proc.name(): + firefoxes = "%s%s\n" % (firefoxes, proc.as_dict(attrs=attrs)) + except Exception: + # may not be able to access process info for all processes + continue + if len(firefoxes) > 0: + # In automation, this warning is unexpected and should be investigated. + # In local testing, this is probably okay, as long as the browser is not + # running a marionette server. + self.log.warning("Found 'firefox' running before starting test browser!") + self.log.warning(firefoxes) + + def runApp( + self, + testUrl, + env, + app, + profile, + extraArgs, + utilityPath, + debuggerInfo=None, + valgrindPath=None, + valgrindArgs=None, + valgrindSuppFiles=None, + symbolsPath=None, + timeout=-1, + detectShutdownLeaks=False, + screenshotOnFail=False, + bisectChunk=None, + marionette_args=None, + e10s=True, + runFailures=False, + crashAsPass=False, + ): + """ + Run the app, log the duration it took to execute, return the status code. + Kills the app if it runs for longer than |maxTime| seconds, or outputs nothing + for |timeout| seconds. + """ + # It can't be the case that both a with-debugger and an + # on-Valgrind run have been requested. doTests() should have + # already excluded this possibility. + assert not (valgrindPath and debuggerInfo) + + # debugger information + interactive = False + debug_args = None + if debuggerInfo: + interactive = debuggerInfo.interactive + debug_args = [debuggerInfo.path] + debuggerInfo.args + + # Set up Valgrind arguments. + if valgrindPath: + interactive = False + valgrindArgs_split = ( + [] if valgrindArgs is None else shlex.split(valgrindArgs) + ) + + valgrindSuppFiles_final = [] + if valgrindSuppFiles is not None: + valgrindSuppFiles_final = [ + "--suppressions=" + path for path in valgrindSuppFiles.split(",") + ] + + debug_args = ( + [valgrindPath] + + mozdebug.get_default_valgrind_args() + + valgrindArgs_split + + valgrindSuppFiles_final + ) + + # fix default timeout + if timeout == -1: + timeout = self.DEFAULT_TIMEOUT + + # Note in the log if running on Valgrind + if valgrindPath: + self.log.info( + "runtests.py | Running on Valgrind. " + + "Using timeout of %d seconds." % timeout + ) + + # copy env so we don't munge the caller's environment + env = env.copy() + + # Used to defer a possible IOError exception from Marionette + marionette_exception = None + + temp_file_paths = [] + + # make sure we clean up after ourselves. + try: + # set process log environment variable + tmpfd, processLog = tempfile.mkstemp(suffix="pidlog") + os.close(tmpfd) + env["MOZ_PROCESS_LOG"] = processLog + + if debuggerInfo: + # If a debugger is attached, don't use timeouts, and don't + # capture ctrl-c. + timeout = None + signal.signal(signal.SIGINT, lambda sigid, frame: None) + + # build command line + cmd = os.path.abspath(app) + args = list(extraArgs) + args.append("-marionette") + # TODO: mozrunner should use -foreground at least for mac + # https://bugzilla.mozilla.org/show_bug.cgi?id=916512 + args.append("-foreground") + self.start_script_kwargs["testUrl"] = testUrl or "about:blank" + + if detectShutdownLeaks: + env["MOZ_LOG"] = ( + env["MOZ_LOG"] + "," if env["MOZ_LOG"] else "" + ) + "DocShellAndDOMWindowLeak:3" + shutdownLeaks = ShutdownLeaks(self.log) + else: + shutdownLeaks = None + + if mozinfo.info["asan"] and mozinfo.isLinux and mozinfo.bits == 64: + lsanLeaks = LSANLeaks(self.log) + else: + lsanLeaks = None + + # create an instance to process the output + outputHandler = self.OutputHandler( + harness=self, + utilityPath=utilityPath, + symbolsPath=symbolsPath, + dump_screen_on_timeout=not debuggerInfo, + dump_screen_on_fail=screenshotOnFail, + shutdownLeaks=shutdownLeaks, + lsanLeaks=lsanLeaks, + bisectChunk=bisectChunk, + ) + + def timeoutHandler(): + browserProcessId = outputHandler.browserProcessId + self.handleTimeout( + timeout, + proc, + utilityPath, + debuggerInfo, + browserProcessId, + processLog, + ) + + kp_kwargs = { + "kill_on_timeout": False, + "cwd": SCRIPT_DIR, + "onTimeout": [timeoutHandler], + } + kp_kwargs["processOutputLine"] = [outputHandler] + + self.checkForRunningBrowsers() + + # create mozrunner instance and start the system under test process + self.lastTestSeen = self.test_name + startTime = datetime.now() + + runner_cls = mozrunner.runners.get( + mozinfo.info.get("appname", "firefox"), mozrunner.Runner + ) + runner = runner_cls( + profile=self.profile, + binary=cmd, + cmdargs=args, + env=env, + process_class=mozprocess.ProcessHandlerMixin, + process_args=kp_kwargs, + ) + + # start the runner + runner.start( + debug_args=debug_args, interactive=interactive, outputTimeout=timeout + ) + proc = runner.process_handler + self.log.info("runtests.py | Application pid: %d" % proc.pid) + + gecko_id = "GECKO(%d)" % proc.pid + self.log.process_start(gecko_id) + self.message_logger.gecko_id = gecko_id + + try: + # start marionette and kick off the tests + marionette_args = marionette_args or {} + self.marionette = Marionette(**marionette_args) + self.marionette.start_session() + + # install specialpowers and mochikit addons + addons = Addons(self.marionette) + + if self.staged_addons: + for addon_path in self.staged_addons: + if not os.path.isdir(addon_path): + self.log.error( + "TEST-UNEXPECTED-FAIL | invalid setup: missing extension at %s" + % addon_path + ) + return 1, self.lastTestSeen + temp_addon_path = create_zip(addon_path) + temp_file_paths.append(temp_addon_path) + addons.install(temp_addon_path) + + self.execute_start_script() + + # an open marionette session interacts badly with mochitest, + # delete it until we figure out why. + self.marionette.delete_session() + del self.marionette + + except IOError: + # Any IOError as thrown by Marionette means that something is + # wrong with the process, like a crash or the socket is no + # longer open. We defer raising this specific error so that + # post-test checks for leaks and crashes are performed and + # reported first. + marionette_exception = sys.exc_info() + + # wait until app is finished + # XXX copy functionality from + # https://github.com/mozilla/mozbase/blob/master/mozrunner/mozrunner/runner.py#L61 + # until bug 913970 is fixed regarding mozrunner `wait` not returning status + # see https://bugzilla.mozilla.org/show_bug.cgi?id=913970 + self.log.info("runtests.py | Waiting for browser...") + status = proc.wait() + if status is None: + self.log.warning( + "runtests.py | Failed to get app exit code - running/crashed?" + ) + # must report an integer to process_exit() + status = 0 + self.log.process_exit("Main app process", status) + runner.process_handler = None + + # finalize output handler + outputHandler.finish() + + # record post-test information + if status: + self.message_logger.dump_buffered() + msg = "application terminated with exit code %s" % status + # self.message_logger.is_test_running indicates we need to send a test_end + if crashAsPass and self.message_logger.is_test_running: + # this works for browser-chrome, mochitest-plain has status=0 + message = { + "action": "test_end", + "status": "CRASH", + "expected": "CRASH", + "thread": None, + "pid": None, + "source": "mochitest", + "time": int(time.time()) * 1000, + "test": self.lastTestSeen, + "message": msg, + } + # need to send a test_end in order to have mozharness process messages properly + # this requires a custom message vs log.error/log.warning/etc. + self.message_logger.process_message(message) + else: + self.log.error( + "TEST-UNEXPECTED-FAIL | %s | %s" % (self.lastTestSeen, msg) + ) + else: + self.lastTestSeen = "Main app process exited normally" + + self.log.info( + "runtests.py | Application ran for: %s" + % str(datetime.now() - startTime) + ) + + # Do a final check for zombie child processes. + zombieProcesses = self.checkForZombies( + processLog, utilityPath, debuggerInfo + ) + + # check for crashes + quiet = False + if crashAsPass: + quiet = True + + minidump_path = os.path.join(self.profile.profile, "minidumps") + crash_count = mozcrash.log_crashes( + self.log, + minidump_path, + symbolsPath, + test=self.lastTestSeen, + quiet=quiet, + ) + + expected = None + if crashAsPass: + # self.message_logger.is_test_running indicates we need a test_end message + if crash_count > 0 and self.message_logger.is_test_running: + # this works for browser-chrome, mochitest-plain has status=0 + expected = "CRASH" + status = 0 + elif crash_count or zombieProcesses: + if self.message_logger.is_test_running: + expected = "PASS" + status = 1 + + if expected: + # send this out so we always wrap up the test-end message + message = { + "action": "test_end", + "status": "CRASH", + "expected": expected, + "thread": None, + "pid": None, + "source": "mochitest", + "time": int(time.time()) * 1000, + "test": self.lastTestSeen, + "message": "application terminated with exit code 0", + } + # need to send a test_end in order to have mozharness process messages properly + # this requires a custom message vs log.error/log.warning/etc. + self.message_logger.process_message(message) + finally: + # cleanup + if os.path.exists(processLog): + os.remove(processLog) + for p in temp_file_paths: + os.remove(p) + + if marionette_exception is not None: + exc, value, tb = marionette_exception + six.reraise(exc, value, tb) + + return status, self.lastTestSeen + + def initializeLooping(self, options): + """ + This method is used to clear the contents before each run of for loop. + This method is used for --run-by-dir and --bisect-chunk. + """ + if options.conditionedProfile: + if options.profilePath and os.path.exists(options.profilePath): + shutil.rmtree(options.profilePath, ignore_errors=True) + if options.manifestFile and os.path.exists(options.manifestFile): + os.remove(options.manifestFile) + + self.expectedError.clear() + self.result.clear() + options.manifestFile = None + options.profilePath = None + + def initializeVirtualInputDevices(self): + """ + Configure the system to have a number of virtual audio input devices, that + each produce a tone at a particular frequency. + + This method is only currently implemented for Linux. + """ + if not mozinfo.isLinux: + return + + pactl = spawn.find_executable("pactl") + + if not pactl: + self.log.error("Could not find pactl on system") + return + + DEVICES_COUNT = 4 + DEVICES_BASE_FREQUENCY = 110 # Hz + self.virtualInputDeviceIdList = [] + # If the device are already present, find their id and return early + o = subprocess.check_output([pactl, "list", "modules", "short"]) + found_devices = 0 + for input in o.splitlines(): + device = input.decode().split("\t") + if device[1] == "module-sine-source": + self.virtualInputDeviceIdList.append(int(device[0])) + found_devices += 1 + + if found_devices == DEVICES_COUNT: + return + elif found_devices != 0: + # Remove all devices and reinitialize them properly + for id in self.virtualInputDeviceIdList: + try: + subprocess.check_call([pactl, "unload-module", str(id)]) + except subprocess.CalledProcessError: + log.error("Could not remove pulse module with id {}".format(id)) + return None + + # We want quite a number of input devices, each with a different tone + # frequency and device name so that we can recognize them easily during + # testing. + command = [pactl, "load-module", "module-sine-source", "rate=44100"] + for i in range(1, DEVICES_COUNT + 1): + freq = i * DEVICES_BASE_FREQUENCY + complete_command = command + [ + "source_name=sine-{}".format(freq), + "frequency={}".format(freq), + ] + try: + o = subprocess.check_output(complete_command) + self.virtualInputDeviceIdList.append(o) + + except subprocess.CalledProcessError: + self.log.error( + "Could not create device with module-sine-source" + " (freq={})".format(freq) + ) + + def normalize_paths(self, paths): + # Normalize test paths so they are relative to test root + norm_paths = [] + for p in paths: + abspath = os.path.abspath(os.path.join(self.oldcwd, p)) + if abspath.startswith(self.testRootAbs): + norm_paths.append(os.path.relpath(abspath, self.testRootAbs)) + else: + norm_paths.append(p) + return norm_paths + + def runMochitests(self, options, testsToRun, manifestToFilter=None): + "This is a base method for calling other methods in this class for --bisect-chunk." + # Making an instance of bisect class for --bisect-chunk option. + bisect = bisection.Bisect(self) + finished = False + status = 0 + bisection_log = 0 + while not finished: + if options.bisectChunk: + testsToRun = bisect.pre_test(options, testsToRun, status) + # To inform that we are in the process of bisection, and to + # look for bleedthrough + if options.bisectChunk != "default" and not bisection_log: + self.log.error( + "TEST-UNEXPECTED-FAIL | Bisection | Please ignore repeats " + "and look for 'Bleedthrough' (if any) at the end of " + "the failure list" + ) + bisection_log = 1 + + result = self.doTests(options, testsToRun, manifestToFilter) + if options.bisectChunk: + status = bisect.post_test(options, self.expectedError, self.result) + else: + status = -1 + + if status == -1: + finished = True + + # We need to print the summary only if options.bisectChunk has a value. + # Also we need to make sure that we do not print the summary in between + # running tests via --run-by-dir. + if options.bisectChunk and options.bisectChunk in self.result: + bisect.print_summary() + + return result + + def groupTestsByScheme(self, tests): + """ + split tests into groups by schemes. test is classified as http if + no scheme specified + """ + httpTests = [] + httpsTests = [] + for test in tests: + if not test.get("scheme") or test.get("scheme") == "http": + httpTests.append(test) + elif test.get("scheme") == "https": + httpsTests.append(test) + return {"http": httpTests, "https": httpsTests} + + def verifyTests(self, options): + """ + Support --verify mode: Run test(s) many times in a variety of + configurations/environments in an effort to find intermittent + failures. + """ + + # Number of times to repeat test(s) when running with --repeat + VERIFY_REPEAT = 10 + # Number of times to repeat test(s) when running test in + VERIFY_REPEAT_SINGLE_BROWSER = 5 + + def step1(): + options.repeat = VERIFY_REPEAT + options.keep_open = False + options.runUntilFailure = True + options.profilePath = None + result = self.runTests(options) + result = result or (-2 if self.countfail > 0 else 0) + self.message_logger.finish() + return result + + def step2(): + options.repeat = 0 + options.keep_open = False + options.runUntilFailure = False + for i in range(VERIFY_REPEAT_SINGLE_BROWSER): + options.profilePath = None + result = self.runTests(options) + result = result or (-2 if self.countfail > 0 else 0) + self.message_logger.finish() + if result != 0: + break + return result + + def step3(): + options.repeat = VERIFY_REPEAT + options.keep_open = False + options.runUntilFailure = True + options.environment.append("MOZ_CHAOSMODE=0xfb") + options.profilePath = None + result = self.runTests(options) + options.environment.remove("MOZ_CHAOSMODE=0xfb") + result = result or (-2 if self.countfail > 0 else 0) + self.message_logger.finish() + return result + + def step4(): + options.repeat = 0 + options.keep_open = False + options.runUntilFailure = False + options.environment.append("MOZ_CHAOSMODE=0xfb") + for i in range(VERIFY_REPEAT_SINGLE_BROWSER): + options.profilePath = None + result = self.runTests(options) + result = result or (-2 if self.countfail > 0 else 0) + self.message_logger.finish() + if result != 0: + break + options.environment.remove("MOZ_CHAOSMODE=0xfb") + return result + + def fission_step(fission_pref): + if fission_pref not in options.extraPrefs: + options.extraPrefs.append(fission_pref) + options.keep_open = False + options.runUntilFailure = True + options.profilePath = None + result = self.runTests(options) + result = result or (-2 if self.countfail > 0 else 0) + self.message_logger.finish() + return result + + def fission_step1(): + return fission_step("fission.autostart=false") + + def fission_step2(): + return fission_step("fission.autostart=true") + + if options.verify_fission: + steps = [ + ("1. Run each test without fission.", fission_step1), + ("2. Run each test with fission.", fission_step2), + ] + else: + steps = [ + ("1. Run each test %d times in one browser." % VERIFY_REPEAT, step1), + ( + "2. Run each test %d times in a new browser each time." + % VERIFY_REPEAT_SINGLE_BROWSER, + step2, + ), + ( + "3. Run each test %d times in one browser, in chaos mode." + % VERIFY_REPEAT, + step3, + ), + ( + "4. Run each test %d times in a new browser each time, " + "in chaos mode." % VERIFY_REPEAT_SINGLE_BROWSER, + step4, + ), + ] + + stepResults = {} + for (descr, step) in steps: + stepResults[descr] = "not run / incomplete" + + startTime = datetime.now() + maxTime = timedelta(seconds=options.verify_max_time) + finalResult = "PASSED" + for (descr, step) in steps: + if (datetime.now() - startTime) > maxTime: + self.log.info("::: Test verification is taking too long: Giving up!") + self.log.info( + "::: So far, all checks passed, but not all checks were run." + ) + break + self.log.info(":::") + self.log.info('::: Running test verification step "%s"...' % descr) + self.log.info(":::") + result = step() + if result != 0: + stepResults[descr] = "FAIL" + finalResult = "FAILED!" + break + stepResults[descr] = "Pass" + + self.logPreamble([]) + + self.log.info(":::") + self.log.info("::: Test verification summary for:") + self.log.info(":::") + tests = self.getActiveTests(options) + for test in tests: + self.log.info("::: " + test["path"]) + self.log.info(":::") + for descr in sorted(stepResults.keys()): + self.log.info("::: %s : %s" % (descr, stepResults[descr])) + self.log.info(":::") + self.log.info("::: Test verification %s" % finalResult) + self.log.info(":::") + + return 0 + + def runTests(self, options): + """Prepare, configure, run tests and cleanup""" + self.extraPrefs = parse_preferences(options.extraPrefs) + self.extraPrefs["fission.autostart"] = not options.disable_fission + + # for test manifest parsing. + mozinfo.update( + { + "a11y_checks": options.a11y_checks, + "e10s": options.e10s, + "fission": not options.disable_fission, + "headless": options.headless, + # Until the test harness can understand default pref values, + # (https://bugzilla.mozilla.org/show_bug.cgi?id=1577912) this value + # should by synchronized with the default pref value indicated in + # StaticPrefList.yaml. + # + # Currently for automation, the pref defaults to true (but can be + # overridden with --setpref). + "serviceworker_e10s": True, + "sessionHistoryInParent": self.extraPrefs.get( + "fission.sessionHistoryInParent", False + ) + or self.extraPrefs.get("fission.autostart", True), + "socketprocess_e10s": self.extraPrefs.get( + "network.process.enabled", False + ), + "socketprocess_networking": self.extraPrefs.get( + "network.http.network_access_on_socket_process.enabled", False + ), + "swgl": self.extraPrefs.get("gfx.webrender.software", False), + "verify": options.verify, + "verify_fission": options.verify_fission, + "webgl_ipc": self.extraPrefs.get("webgl.out-of-process", False), + "wmfme": ( + self.extraPrefs.get("media.wmf.media-engine.enabled", False) + and self.extraPrefs.get( + "media.wmf.media-engine.channel-decoder.enabled", False + ) + ), + "xorigin": options.xOriginTests, + "condprof": options.conditionedProfile, + "msix": "WindowsApps" in options.app, + } + ) + + if not self.mozinfo_variables_shown: + self.mozinfo_variables_shown = True + self.log.info( + "These variables are available in the mozinfo environment and " + "can be used to skip tests conditionally:" + ) + for info in sorted(mozinfo.info.items(), key=lambda item: item[0]): + self.log.info(" {key}: {value}".format(key=info[0], value=info[1])) + self.setTestRoot(options) + + # Despite our efforts to clean up servers started by this script, in practice + # we still see infrequent cases where a process is orphaned and interferes + # with future tests, typically because the old server is keeping the port in use. + # Try to avoid those failures by checking for and killing servers before + # trying to start new ones. + self.killNamedProc("ssltunnel") + self.killNamedProc("xpcshell") + + if options.cleanupCrashes: + mozcrash.cleanup_pending_crash_reports() + + tests = self.getActiveTests(options) + self.logPreamble(tests) + + if mozinfo.info["fission"] and not mozinfo.info["e10s"]: + # Make sure this is logged *after* suite_start so it gets associated with the + # current suite in the summary formatters. + self.log.error("Fission is not supported without e10s.") + return 1 + + tests = [t for t in tests if "disabled" not in t] + + # Until we have all green, this does not run on a11y (for perf reasons) + if not options.runByManifest: + result = self.runMochitests(options, [t["path"] for t in tests]) + self.handleShutdownProfile(options) + return result + + # code for --run-by-manifest + manifests = set(t["manifest"] for t in tests) + result = 0 + + origPrefs = self.extraPrefs.copy() + for m in sorted(manifests): + self.log.info("Running manifest: {}".format(m)) + + args = list(self.args_by_manifest[m])[0] + self.extraArgs = [] + if args: + for arg in args.strip().split(): + # Split off the argument value if available so that both + # name and value will be set individually + self.extraArgs.extend(arg.split("=")) + + self.log.info( + "The following arguments will be set:\n {}".format( + "\n ".join(self.extraArgs) + ) + ) + + prefs = list(self.prefs_by_manifest[m])[0] + self.extraPrefs = origPrefs.copy() + if prefs: + prefs = prefs.strip().split() + self.log.info( + "The following extra prefs will be set:\n {}".format( + "\n ".join(prefs) + ) + ) + self.extraPrefs.update(parse_preferences(prefs)) + + envVars = list(self.env_vars_by_manifest[m])[0] + self.extraEnv = {} + if envVars: + self.extraEnv = envVars.strip().split() + self.log.info( + "The following extra environment variables will be set:\n {}".format( + "\n ".join(self.extraEnv) + ) + ) + + self.parseAndCreateTestsDirs(m) + + # If we are using --run-by-manifest, we should not use the profile path (if) provided + # by the user, since we need to create a new directory for each run. We would face + # problems if we use the directory provided by the user. + tests_in_manifest = [t["path"] for t in tests if t["manifest"] == m] + res = self.runMochitests(options, tests_in_manifest, manifestToFilter=m) + result = result or res + + # Dump the logging buffer + self.message_logger.dump_buffered() + + if res == -1: + break + + if self.manifest is not None: + self.cleanup(options, True) + + e10s_mode = "e10s" if options.e10s else "non-e10s" + + # for failure mode: where browser window has crashed and we have no reported results + if ( + self.countpass == self.countfail == self.counttodo == 0 + and options.crashAsPass + ): + self.countpass = 1 + self.result = 0 + + # printing total number of tests + if options.flavor == "browser": + print("TEST-INFO | checking window state") + print("Browser Chrome Test Summary") + print("\tPassed: %s" % self.countpass) + print("\tFailed: %s" % self.countfail) + print("\tTodo: %s" % self.counttodo) + print("\tMode: %s" % e10s_mode) + print("*** End BrowserChrome Test Results ***") + else: + print("0 INFO TEST-START | Shutdown") + print("1 INFO Passed: %s" % self.countpass) + print("2 INFO Failed: %s" % self.countfail) + print("3 INFO Todo: %s" % self.counttodo) + print("4 INFO Mode: %s" % e10s_mode) + print("5 INFO SimpleTest FINISHED") + + self.handleShutdownProfile(options) + + if not result: + if self.countfail or not (self.countpass or self.counttodo): + # at least one test failed, or + # no tests passed, and no tests failed (possibly a crash) + result = 1 + + return result + + def handleShutdownProfile(self, options): + # If shutdown profiling was enabled, then the user will want to access the + # performance profile. The following code will display helpful log messages + # and automatically open the profile if it is requested. + if self.browserEnv and "MOZ_PROFILER_SHUTDOWN" in self.browserEnv: + profile_path = self.browserEnv["MOZ_PROFILER_SHUTDOWN"] + + profiler_logger = get_proxy_logger("profiler") + profiler_logger.info("Shutdown performance profiling was enabled") + profiler_logger.info("Profile saved locally to: %s" % profile_path) + + if options.profilerSaveOnly or options.profiler: + # Only do the extra work of symbolicating and viewing the profile if + # officially requested through a command line flag. The MOZ_PROFILER_* + # flags can be set by a user. + symbolicate_profile_json(profile_path, options.topobjdir) + view_gecko_profile_from_mochitest( + profile_path, options, profiler_logger + ) + else: + profiler_logger.info( + "The profiler was enabled outside of the mochitests. " + "Use --profiler instead of MOZ_PROFILER_SHUTDOWN to " + "symbolicate and open the profile automatically." + ) + + # Clean up the temporary file if it exists. + if self.profiler_tempdir: + shutil.rmtree(self.profiler_tempdir) + + def doTests(self, options, testsToFilter=None, manifestToFilter=None): + # A call to initializeLooping method is required in case of --run-by-dir or --bisect-chunk + # since we need to initialize variables for each loop. + if options.bisectChunk or options.runByManifest: + self.initializeLooping(options) + + # get debugger info, a dict of: + # {'path': path to the debugger (string), + # 'interactive': whether the debugger is interactive or not (bool) + # 'args': arguments to the debugger (list) + # TODO: use mozrunner.local.debugger_arguments: + # https://github.com/mozilla/mozbase/blob/master/mozrunner/mozrunner/local.py#L42 + + debuggerInfo = None + if options.debugger: + debuggerInfo = mozdebug.get_debugger_info( + options.debugger, options.debuggerArgs, options.debuggerInteractive + ) + + if options.useTestMediaDevices: + devices = findTestMediaDevices(self.log) + if not devices: + self.log.error("Could not find test media devices to use") + return 1 + self.mediaDevices = devices + self.initializeVirtualInputDevices() + + # See if we were asked to run on Valgrind + valgrindPath = None + valgrindArgs = None + valgrindSuppFiles = None + if options.valgrind: + valgrindPath = options.valgrind + if options.valgrindArgs: + valgrindArgs = options.valgrindArgs + if options.valgrindSuppFiles: + valgrindSuppFiles = options.valgrindSuppFiles + + if (valgrindArgs or valgrindSuppFiles) and not valgrindPath: + self.log.error( + "Specified --valgrind-args or --valgrind-supp-files," + " but not --valgrind" + ) + return 1 + + if valgrindPath and debuggerInfo: + self.log.error("Can't use both --debugger and --valgrind together") + return 1 + + if valgrindPath and not valgrindSuppFiles: + valgrindSuppFiles = ",".join(get_default_valgrind_suppression_files()) + + # buildProfile sets self.profile . + # This relies on sideeffects and isn't very stateful: + # https://bugzilla.mozilla.org/show_bug.cgi?id=919300 + self.manifest = self.buildProfile(options) + if self.manifest is None: + return 1 + + self.leak_report_file = os.path.join(options.profilePath, "runtests_leaks.log") + + self.browserEnv = self.buildBrowserEnv(options, debuggerInfo is not None) + + if self.browserEnv is None: + return 1 + + if self.mozLogs: + self.browserEnv["MOZ_LOG_FILE"] = "{}/moz-pid=%PID-uid={}.log".format( + self.browserEnv["MOZ_UPLOAD_DIR"], str(uuid.uuid4()) + ) + + status = 0 + try: + self.startServers(options, debuggerInfo) + + if options.jsconsole: + options.browserArgs.extend(["--jsconsole"]) + + if options.jsdebugger: + options.browserArgs.extend(["-wait-for-jsdebugger", "-jsdebugger"]) + + # -jsdebugger takes a binary path as an optional argument. + # Append jsdebuggerPath right after `-jsdebugger`. + if options.jsdebuggerPath: + options.browserArgs.extend([options.jsdebuggerPath]) + + # Remove the leak detection file so it can't "leak" to the tests run. + # The file is not there if leak logging was not enabled in the + # application build. + if os.path.exists(self.leak_report_file): + os.remove(self.leak_report_file) + + # then again to actually run mochitest + if options.timeout: + timeout = options.timeout + 30 + elif options.debugger or not options.autorun: + timeout = None + else: + # We generally want the JS harness or marionette to handle + # timeouts if they can. + # The default JS harness timeout is currently 300 seconds. + # The default Marionette socket timeout is currently 360 seconds. + # Wait a little (10 seconds) more before timing out here. + # See bug 479518 and bug 1414063. + timeout = 370.0 + + if "MOZ_CHAOSMODE=0xfb" in options.environment and timeout: + timeout *= 2 + + # Detect shutdown leaks for m-bc runs if + # code coverage is not enabled. + detectShutdownLeaks = False + if options.jscov_dir_prefix is None: + detectShutdownLeaks = ( + mozinfo.info["debug"] + and options.flavor == "browser" + and options.subsuite != "thunderbird" + and not options.crashAsPass + ) + + self.start_script_kwargs["flavor"] = self.normflavor(options.flavor) + marionette_args = { + "symbols_path": options.symbolsPath, + "socket_timeout": options.marionette_socket_timeout, + "startup_timeout": options.marionette_startup_timeout, + } + + if options.marionette: + host, port = options.marionette.split(":") + marionette_args["host"] = host + marionette_args["port"] = int(port) + + # testsToFilter parameter is used to filter out the test list that + # is sent to getTestsByScheme + for (scheme, tests) in self.getTestsByScheme( + options, testsToFilter, True, manifestToFilter + ): + # read the number of tests here, if we are not going to run any, + # terminate early + if not tests: + continue + + self.currentTests = [t["path"] for t in tests] + testURL = self.buildTestURL(options, scheme=scheme) + + self.buildURLOptions(options, self.browserEnv) + if self.urlOpts: + testURL += "?" + "&".join(self.urlOpts) + + if options.runFailures: + testURL += "&runFailures=true" + + if options.timeoutAsPass: + testURL += "&timeoutAsPass=true" + + if options.conditionedProfile: + testURL += "&conditionedProfile=true" + + self.log.info("runtests.py | Running with scheme: {}".format(scheme)) + self.log.info( + "runtests.py | Running with e10s: {}".format(options.e10s) + ) + self.log.info( + "runtests.py | Running with fission: {}".format( + mozinfo.info.get("fission", True) + ) + ) + self.log.info( + "runtests.py | Running with cross-origin iframes: {}".format( + mozinfo.info.get("xorigin", False) + ) + ) + self.log.info( + "runtests.py | Running with serviceworker_e10s: {}".format( + mozinfo.info.get("serviceworker_e10s", False) + ) + ) + self.log.info( + "runtests.py | Running with socketprocess_e10s: {}".format( + mozinfo.info.get("socketprocess_e10s", False) + ) + ) + self.log.info("runtests.py | Running tests: start.\n") + ret, _ = self.runApp( + testURL, + self.browserEnv, + options.app, + profile=self.profile, + extraArgs=options.browserArgs + self.extraArgs, + utilityPath=options.utilityPath, + debuggerInfo=debuggerInfo, + valgrindPath=valgrindPath, + valgrindArgs=valgrindArgs, + valgrindSuppFiles=valgrindSuppFiles, + symbolsPath=options.symbolsPath, + timeout=timeout, + detectShutdownLeaks=detectShutdownLeaks, + screenshotOnFail=options.screenshotOnFail, + bisectChunk=options.bisectChunk, + marionette_args=marionette_args, + e10s=options.e10s, + runFailures=options.runFailures, + crashAsPass=options.crashAsPass, + ) + status = ret or status + except KeyboardInterrupt: + self.log.info("runtests.py | Received keyboard interrupt.\n") + status = -1 + except Exception as e: + traceback.print_exc() + self.log.error( + "Automation Error: Received unexpected exception while running application\n" + ) + if "ADBTimeoutError" in repr(e): + self.log.info("runtests.py | Device disconnected. Aborting test.\n") + raise + status = 1 + finally: + self.stopServers() + + ignoreMissingLeaks = options.ignoreMissingLeaks + leakThresholds = options.leakThresholds + + if options.crashAsPass: + ignoreMissingLeaks.append("tab") + ignoreMissingLeaks.append("socket") + + # Stop leak detection if m-bc code coverage is enabled + # by maxing out the leak threshold for all processes. + if options.jscov_dir_prefix: + for processType in leakThresholds: + ignoreMissingLeaks.append(processType) + leakThresholds[processType] = sys.maxsize + + utilityPath = options.utilityPath or options.xrePath + mozleak.process_leak_log( + self.leak_report_file, + leak_thresholds=leakThresholds, + ignore_missing_leaks=ignoreMissingLeaks, + log=self.log, + stack_fixer=get_stack_fixer_function(utilityPath, options.symbolsPath), + ) + + self.log.info("runtests.py | Running tests: end.") + + if self.manifest is not None: + self.cleanup(options, False) + + return status + + def handleTimeout( + self, timeout, proc, utilityPath, debuggerInfo, browser_pid, processLog + ): + """handle process output timeout""" + # TODO: bug 913975 : _processOutput should call self.processOutputLine + # one more time one timeout (I think) + error_message = ( + "TEST-UNEXPECTED-TIMEOUT | %s | application timed out after " + "%d seconds with no output" + ) % (self.lastTestSeen, int(timeout)) + self.message_logger.dump_buffered() + self.message_logger.buffering = False + self.log.info(error_message) + self.log.error("Force-terminating active process(es).") + + browser_pid = browser_pid or proc.pid + child_pids = self.extract_child_pids(processLog, browser_pid) + self.log.info("Found child pids: %s" % child_pids) + + if HAVE_PSUTIL: + try: + browser_proc = [psutil.Process(browser_pid)] + except Exception: + self.log.info("Failed to get proc for pid %d" % browser_pid) + browser_proc = [] + try: + child_procs = [psutil.Process(pid) for pid in child_pids] + except Exception: + self.log.info("Failed to get child procs") + child_procs = [] + for pid in child_pids: + self.killAndGetStack( + pid, utilityPath, debuggerInfo, dump_screen=not debuggerInfo + ) + gone, alive = psutil.wait_procs(child_procs, timeout=30) + for p in gone: + self.log.info("psutil found pid %s dead" % p.pid) + for p in alive: + self.log.warning("failed to kill pid %d after 30s" % p.pid) + self.killAndGetStack( + browser_pid, utilityPath, debuggerInfo, dump_screen=not debuggerInfo + ) + gone, alive = psutil.wait_procs(browser_proc, timeout=30) + for p in gone: + self.log.info("psutil found pid %s dead" % p.pid) + for p in alive: + self.log.warning("failed to kill pid %d after 30s" % p.pid) + else: + self.log.error( + "psutil not available! Will wait 30s before " + "attempting to kill parent process. This should " + "not occur in mozilla automation. See bug 1143547." + ) + for pid in child_pids: + self.killAndGetStack( + pid, utilityPath, debuggerInfo, dump_screen=not debuggerInfo + ) + if child_pids: + time.sleep(30) + + self.killAndGetStack( + browser_pid, utilityPath, debuggerInfo, dump_screen=not debuggerInfo + ) + + def archiveMozLogs(self): + if self.mozLogs: + with zipfile.ZipFile( + "{}/mozLogs.zip".format(os.environ["MOZ_UPLOAD_DIR"]), + "w", + zipfile.ZIP_DEFLATED, + ) as logzip: + for logfile in glob.glob( + "{}/moz*.log*".format(os.environ["MOZ_UPLOAD_DIR"]) + ): + logzip.write(logfile, os.path.basename(logfile)) + os.remove(logfile) + logzip.close() + + class OutputHandler(object): + + """line output handler for mozrunner""" + + def __init__( + self, + harness, + utilityPath, + symbolsPath=None, + dump_screen_on_timeout=True, + dump_screen_on_fail=False, + shutdownLeaks=None, + lsanLeaks=None, + bisectChunk=None, + ): + """ + harness -- harness instance + dump_screen_on_timeout -- whether to dump the screen on timeout + """ + self.harness = harness + self.utilityPath = utilityPath + self.symbolsPath = symbolsPath + self.dump_screen_on_timeout = dump_screen_on_timeout + self.dump_screen_on_fail = dump_screen_on_fail + self.shutdownLeaks = shutdownLeaks + self.lsanLeaks = lsanLeaks + self.bisectChunk = bisectChunk + self.browserProcessId = None + self.stackFixerFunction = self.stackFixer() + + def processOutputLine(self, line): + """per line handler of output for mozprocess""" + # Parsing the line (by the structured messages logger). + messages = self.harness.message_logger.parse_line(line) + + for message in messages: + # Passing the message to the handlers + for handler in self.outputHandlers(): + message = handler(message) + + # Processing the message by the logger + self.harness.message_logger.process_message(message) + + __call__ = processOutputLine + + def outputHandlers(self): + """returns ordered list of output handlers""" + handlers = [ + self.fix_stack, + self.record_last_test, + self.dumpScreenOnTimeout, + self.dumpScreenOnFail, + self.trackShutdownLeaks, + self.trackLSANLeaks, + self.countline, + ] + if self.bisectChunk: + handlers.append(self.record_result) + handlers.append(self.first_error) + + return handlers + + def stackFixer(self): + """ + return get_stack_fixer_function, if any, to use on the output lines + """ + return get_stack_fixer_function(self.utilityPath, self.symbolsPath) + + def finish(self): + if self.shutdownLeaks: + self.harness.countfail += self.shutdownLeaks.process() + + if self.lsanLeaks: + self.harness.countfail += self.lsanLeaks.process() + + # output message handlers: + # these take a message and return a message + + def record_result(self, message): + # by default make the result key equal to pass. + if message["action"] == "test_start": + key = message["test"].split("/")[-1].strip() + self.harness.result[key] = "PASS" + elif message["action"] == "test_status": + if "expected" in message: + key = message["test"].split("/")[-1].strip() + self.harness.result[key] = "FAIL" + elif message["status"] == "FAIL": + key = message["test"].split("/")[-1].strip() + self.harness.result[key] = "TODO" + return message + + def first_error(self, message): + if ( + message["action"] == "test_status" + and "expected" in message + and message["status"] == "FAIL" + ): + key = message["test"].split("/")[-1].strip() + if key not in self.harness.expectedError: + self.harness.expectedError[key] = message.get( + "message", message["subtest"] + ).strip() + return message + + def countline(self, message): + if message["action"] == "log": + line = message.get("message", "") + elif message["action"] == "process_output": + line = message.get("data", "") + else: + return message + val = 0 + try: + val = int(line.split(":")[-1].strip()) + except (AttributeError, ValueError): + return message + + if "Passed:" in line: + self.harness.countpass += val + elif "Failed:" in line: + self.harness.countfail += val + elif "Todo:" in line: + self.harness.counttodo += val + return message + + def fix_stack(self, message): + if self.stackFixerFunction: + if message["action"] == "log": + message["message"] = self.stackFixerFunction(message["message"]) + elif message["action"] == "process_output": + message["data"] = self.stackFixerFunction(message["data"]) + return message + + def record_last_test(self, message): + """record last test on harness""" + if message["action"] == "test_start": + self.harness.lastTestSeen = message["test"] + elif message["action"] == "test_end": + if ( + self.harness.currentTests + and message["test"] == self.harness.currentTests[-1] + ): + self.harness.lastTestSeen = "Last test finished" + else: + self.harness.lastTestSeen = "{} (finished)".format(message["test"]) + return message + + def dumpScreenOnTimeout(self, message): + if ( + not self.dump_screen_on_fail + and self.dump_screen_on_timeout + and message["action"] == "test_status" + and "expected" in message + and "Test timed out" in message["subtest"] + ): + self.harness.dumpScreen(self.utilityPath) + return message + + def dumpScreenOnFail(self, message): + if ( + self.dump_screen_on_fail + and "expected" in message + and message["status"] == "FAIL" + ): + self.harness.dumpScreen(self.utilityPath) + return message + + def trackLSANLeaks(self, message): + if self.lsanLeaks and message["action"] in ("log", "process_output"): + line = ( + message.get("message", "") + if message["action"] == "log" + else message["data"] + ) + self.lsanLeaks.log(line) + return message + + def trackShutdownLeaks(self, message): + if self.shutdownLeaks: + self.shutdownLeaks.log(message) + return message + + +def view_gecko_profile_from_mochitest(profile_path, options, profiler_logger): + """Getting shutdown performance profiles from just the command line arguments is + difficult. This function makes the developer ergonomics a bit easier by taking the + generated Gecko profile, and automatically serving it to profiler.firefox.com. The + Gecko profile during shutdown is dumped to disk at: + + {objdir}/_tests/testing/mochitest/{profilename} + + This function takes that file, and launches a local webserver, and then points + a browser to profiler.firefox.com to view it. From there it's easy to publish + or save the profile. + """ + + if options.profilerSaveOnly: + # The user did not want this to automatically open, only share the location. + return + + if not os.path.exists(profile_path): + profiler_logger.error( + "No profile was found at the profile path, cannot " + "launch profiler.firefox.com." + ) + return + + profiler_logger.info("Loading this profile in the Firefox Profiler") + + view_gecko_profile(profile_path) + + +def run_test_harness(parser, options): + parser.validate(options) + + logger_options = { + key: value + for key, value in six.iteritems(vars(options)) + if key.startswith("log") or key == "valgrind" + } + + runner = MochitestDesktop( + options.flavor, logger_options, options.stagedAddons, quiet=options.quiet + ) + + options.runByManifest = False + if options.flavor in ("plain", "browser", "chrome"): + options.runByManifest = True + + if options.verify or options.verify_fission: + result = runner.verifyTests(options) + else: + result = runner.runTests(options) + + runner.archiveMozLogs() + runner.message_logger.finish() + return result + + +def cli(args=sys.argv[1:]): + # parse command line options + parser = MochitestArgumentParser(app="generic") + options = parser.parse_args(args) + if options is None: + # parsing error + sys.exit(1) + + return run_test_harness(parser, options) + + +if __name__ == "__main__": + sys.exit(cli()) |