#!/usr/bin/env vpython3 # Copyright (c) 2020 The WebRTC project authors. All Rights Reserved. # # Use of this source code is governed by a BSD-style license # that can be found in the LICENSE file in the root of the source # tree. An additional intellectual property rights grant can be found # in the file PATENTS. All contributing project authors may # be found in the AUTHORS file in the root of the source tree. import datetime import json import subprocess import time import zlib from typing import Optional import dataclasses import httplib2 from tracing.value import histogram from tracing.value import histogram_set from tracing.value.diagnostics import generic_set from tracing.value.diagnostics import reserved_infos @dataclasses.dataclass class UploaderOptions(): """Required information to upload perf metrics. Attributes: perf_dashboard_machine_group: The "master" the bots are grouped under. This string is the group in the the perf dashboard path group/bot/perf_id/metric/subtest. bot: The bot running the test (e.g. webrtc-win-large-tests). test_suite: The key for the test in the dashboard (i.e. what you select in the top-level test suite selector in the dashboard webrtc_git_hash: webrtc.googlesource.com commit hash. commit_position: Commit pos corresponding to the git hash. build_page_url: URL to the build page for this build. dashboard_url: Which dashboard to use. input_results_file: A HistogramSet proto file coming from WebRTC tests. output_json_file: Where to write the output (for debugging). wait_timeout_sec: Maximum amount of time in seconds that the script will wait for the confirmation. wait_polling_period_sec: Status will be requested from the Dashboard every wait_polling_period_sec seconds. """ perf_dashboard_machine_group: str bot: str test_suite: str webrtc_git_hash: str commit_position: int build_page_url: str dashboard_url: str input_results_file: str output_json_file: Optional[str] = None wait_timeout_sec: datetime.timedelta = datetime.timedelta(seconds=1200) wait_polling_period_sec: datetime.timedelta = datetime.timedelta(seconds=120) def _GenerateOauthToken(): args = ['luci-auth', 'token'] p = subprocess.Popen(args, universal_newlines=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE) if p.wait() == 0: output = p.stdout.read() return output.strip() raise RuntimeError( 'Error generating authentication token.\nStdout: %s\nStderr:%s' % (p.stdout.read(), p.stderr.read())) def _CreateHeaders(oauth_token): return {'Authorization': 'Bearer %s' % oauth_token} def _SendHistogramSet(url, histograms): """Make a HTTP POST with the given JSON to the Performance Dashboard. Args: url: URL of Performance Dashboard instance, e.g. "https://chromeperf.appspot.com". histograms: a histogram set object that contains the data to be sent. """ headers = _CreateHeaders(_GenerateOauthToken()) serialized = json.dumps(_ApplyHacks(histograms.AsDicts()), indent=4) if url.startswith('http://localhost'): # The catapult server turns off compression in developer mode. data = serialized else: data = zlib.compress(serialized.encode('utf-8')) print('Sending %d bytes to %s.' % (len(data), url + '/add_histograms')) http = httplib2.Http() response, content = http.request(url + '/add_histograms', method='POST', body=data, headers=headers) return response, content def _WaitForUploadConfirmation(url, upload_token, wait_timeout, wait_polling_period): """Make a HTTP GET requests to the Performance Dashboard untill upload status is known or the time is out. Args: url: URL of Performance Dashboard instance, e.g. "https://chromeperf.appspot.com". upload_token: String that identifies Performance Dashboard and can be used for the status check. wait_timeout: (datetime.timedelta) Maximum time to wait for the confirmation. wait_polling_period: (datetime.timedelta) Performance Dashboard will be polled every wait_polling_period amount of time. """ assert wait_polling_period <= wait_timeout headers = _CreateHeaders(_GenerateOauthToken()) http = httplib2.Http() oauth_refreshed = False response = None resp_json = None current_time = datetime.datetime.now() end_time = current_time + wait_timeout next_poll_time = current_time + wait_polling_period while datetime.datetime.now() < end_time: current_time = datetime.datetime.now() if next_poll_time > current_time: time.sleep((next_poll_time - current_time).total_seconds()) next_poll_time = datetime.datetime.now() + wait_polling_period response, content = http.request(url + '/uploads/' + upload_token, method='GET', headers=headers) print('Upload state polled. Response: %r.' % content) if not oauth_refreshed and response.status == 403: print('Oauth token refreshed. Continue polling.') headers = _CreateHeaders(_GenerateOauthToken()) oauth_refreshed = True continue if response.status != 200: break resp_json = json.loads(content) if resp_json['state'] == 'COMPLETED' or resp_json['state'] == 'FAILED': break return response, resp_json # Because of an issues on the Dashboard side few measurements over a large set # can fail to upload. That would lead to the whole upload to be marked as # failed. Check it, so it doesn't increase flakiness of our tests. # TODO(crbug.com/1145904): Remove check after fixed. def _CheckFullUploadInfo(url, upload_token, min_measurements_amount=50, max_failed_measurements_percent=0.03): """Make a HTTP GET requests to the Performance Dashboard to get full info about upload (including measurements). Checks if upload is correct despite not having status "COMPLETED". Args: url: URL of Performance Dashboard instance, e.g. "https://chromeperf.appspot.com". upload_token: String that identifies Performance Dashboard and can be used for the status check. min_measurements_amount: minimal amount of measurements that the upload should have to start tolerating failures in particular measurements. max_failed_measurements_percent: maximal percent of failured measurements to tolerate. """ headers = _CreateHeaders(_GenerateOauthToken()) http = httplib2.Http() response, content = http.request(url + '/uploads/' + upload_token + '?additional_info=measurements', method='GET', headers=headers) if response.status != 200: print('Failed to reach the dashboard to get full upload info.') return False resp_json = json.loads(content) print('Full upload info: %s.' % json.dumps(resp_json, indent=4)) if 'measurements' in resp_json: measurements_cnt = len(resp_json['measurements']) not_completed_state_cnt = len( [m for m in resp_json['measurements'] if m['state'] != 'COMPLETED']) if (measurements_cnt >= min_measurements_amount and (not_completed_state_cnt / (measurements_cnt * 1.0) <= max_failed_measurements_percent)): print(('Not all measurements were confirmed to upload. ' 'Measurements count: %d, failed to upload or timed out: %d' % (measurements_cnt, not_completed_state_cnt))) return True return False # TODO(https://crbug.com/1029452): HACKHACK # Remove once we have doubles in the proto and handle -infinity correctly. def _ApplyHacks(dicts): def _NoInf(value): if value == float('inf'): return histogram.JS_MAX_VALUE if value == float('-inf'): return -histogram.JS_MAX_VALUE return value for d in dicts: if 'running' in d: d['running'] = [_NoInf(value) for value in d['running']] if 'sampleValues' in d: d['sampleValues'] = [_NoInf(value) for value in d['sampleValues']] return dicts def _LoadHistogramSetFromProto(options): hs = histogram_set.HistogramSet() with open(options.input_results_file, 'rb') as f: hs.ImportProto(f.read()) return hs def _AddBuildInfo(histograms, options): common_diagnostics = { reserved_infos.MASTERS: options.perf_dashboard_machine_group, reserved_infos.BOTS: options.bot, reserved_infos.POINT_ID: options.commit_position, reserved_infos.BENCHMARKS: options.test_suite, reserved_infos.WEBRTC_REVISIONS: str(options.webrtc_git_hash), reserved_infos.BUILD_URLS: options.build_page_url, } for k, v in list(common_diagnostics.items()): histograms.AddSharedDiagnosticToAllHistograms(k.name, generic_set.GenericSet([v])) def _DumpOutput(histograms, output_file): with open(output_file, 'w') as f: json.dump(_ApplyHacks(histograms.AsDicts()), f, indent=4) def UploadToDashboardImpl(options): histograms = _LoadHistogramSetFromProto(options) _AddBuildInfo(histograms, options) if options.output_json_file: _DumpOutput(histograms, options.output_json_file) response, content = _SendHistogramSet(options.dashboard_url, histograms) if response.status != 200: print(('Upload failed with %d: %s\n\n%s' % (response.status, response.reason, content))) return 1 upload_token = json.loads(content).get('token') if not upload_token: print(('Received 200 from dashboard. ', 'Not waiting for the upload status confirmation.')) return 0 response, resp_json = _WaitForUploadConfirmation( options.dashboard_url, upload_token, options.wait_timeout_sec, options.wait_polling_period_sec) if ((resp_json and resp_json['state'] == 'COMPLETED') or _CheckFullUploadInfo(options.dashboard_url, upload_token)): print('Upload completed.') return 0 if response.status != 200: print(('Upload status poll failed with %d: %s' % (response.status, response.reason))) return 1 if resp_json['state'] == 'FAILED': print('Upload failed.') return 1 print(('Upload wasn\'t completed in a given time: %s seconds.' % options.wait_timeout_sec)) return 1 def UploadToDashboard(options): try: exit_code = UploadToDashboardImpl(options) except RuntimeError as e: print(e) return 1 return exit_code