# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/. from __future__ import absolute_import, division, print_function import datetime import errno import json import os import posixpath import re import requests import six.moves.urllib_parse as urlparse import subprocess import mozpack.path as mozpath from moztest.resolve import TestResolver, TestManifestLoader from mozfile import which from mozbuild.base import MozbuildObject, MachCommandConditions as conditions REFERER = "https://wiki.developer.mozilla.org/en-US/docs/Mozilla/Test-Info" class TestInfo(object): """ Support 'mach test-info'. """ def __init__(self, verbose): self.verbose = verbose here = os.path.abspath(os.path.dirname(__file__)) self.build_obj = MozbuildObject.from_environment(cwd=here) def log_verbose(self, what): if self.verbose: print(what) class TestInfoTests(TestInfo): """ Support 'mach test-info tests': Detailed report of specified tests. """ def __init__(self, verbose): TestInfo.__init__(self, verbose) self._hg = None if conditions.is_hg(self.build_obj): self._hg = which("hg") if not self._hg: raise OSError(errno.ENOENT, "Could not find 'hg' on PATH.") self._git = None if conditions.is_git(self.build_obj): self._git = which("git") if not self._git: raise OSError(errno.ENOENT, "Could not find 'git' on PATH.") def find_in_hg_or_git(self, test_name): if self._hg: cmd = [self._hg, "files", "-I", test_name] elif self._git: cmd = [self._git, "ls-files", test_name] else: return None try: out = subprocess.check_output(cmd, universal_newlines=True).splitlines() except subprocess.CalledProcessError: out = None return out def set_test_name(self): # Generating a unified report for a specific test is complicated # by differences in the test name used in various data sources. # Consider: # - It is often convenient to request a report based only on # a short file name, rather than the full path; # - Bugs may be filed in bugzilla against a simple, short test # name or the full path to the test; # This function attempts to find appropriate names for different # queries based on the specified test name. # full_test_name is full path to file in hg (or git) self.full_test_name = None out = self.find_in_hg_or_git(self.test_name) if out and len(out) == 1: self.full_test_name = out[0] elif out and len(out) > 1: print("Ambiguous test name specified. Found:") for line in out: print(line) else: out = self.find_in_hg_or_git("**/%s*" % self.test_name) if out and len(out) == 1: self.full_test_name = out[0] elif out and len(out) > 1: print("Ambiguous test name. Found:") for line in out: print(line) if self.full_test_name: self.full_test_name.replace(os.sep, posixpath.sep) print("Found %s in source control." % self.full_test_name) else: print("Unable to validate test name '%s'!" % self.test_name) self.full_test_name = self.test_name # search for full_test_name in test manifests here = os.path.abspath(os.path.dirname(__file__)) resolver = TestResolver.from_environment( cwd=here, loader_cls=TestManifestLoader ) relpath = self.build_obj._wrap_path_argument(self.full_test_name).relpath() tests = list(resolver.resolve_tests(paths=[relpath])) if len(tests) == 1: relpath = self.build_obj._wrap_path_argument(tests[0]["manifest"]).relpath() print("%s found in manifest %s" % (self.full_test_name, relpath)) if tests[0].get("flavor"): print(" flavor: %s" % tests[0]["flavor"]) if tests[0].get("skip-if"): print(" skip-if: %s" % tests[0]["skip-if"]) if tests[0].get("fail-if"): print(" fail-if: %s" % tests[0]["fail-if"]) elif len(tests) == 0: print("%s not found in any test manifest!" % self.full_test_name) else: print("%s found in more than one manifest!" % self.full_test_name) # short_name is full_test_name without path self.short_name = None name_idx = self.full_test_name.rfind("/") if name_idx > 0: self.short_name = self.full_test_name[name_idx + 1 :] if self.short_name and self.short_name == self.test_name: self.short_name = None def get_platform(self, record): if "platform" in record["build"]: platform = record["build"]["platform"] else: platform = "-" platform_words = platform.split("-") types_label = "" # combine run and build types and eliminate duplicates run_types = [] if "run" in record and "type" in record["run"]: run_types = record["run"]["type"] run_types = run_types if isinstance(run_types, list) else [run_types] build_types = [] if "build" in record and "type" in record["build"]: build_types = record["build"]["type"] build_types = ( build_types if isinstance(build_types, list) else [build_types] ) run_types = list(set(run_types + build_types)) # '1proc' is used as a treeherder label but does not appear in run types if "e10s" not in run_types: run_types = run_types + ["1proc"] for run_type in run_types: # chunked is not interesting if run_type == "chunked": continue # e10s is the default: implied if run_type == "e10s": continue # sometimes a build/run type is already present in the build platform if run_type in platform_words: continue if types_label: types_label += "-" types_label += run_type return "%s/%s:" % (platform, types_label) def report_bugs(self): # Report open bugs matching test name search = self.full_test_name if self.test_name: search = "%s,%s" % (search, self.test_name) if self.short_name: search = "%s,%s" % (search, self.short_name) payload = {"quicksearch": search, "include_fields": "id,summary"} response = requests.get("https://bugzilla.mozilla.org/rest/bug", payload) response.raise_for_status() json_response = response.json() print("\nBugzilla quick search for '%s':" % search) if "bugs" in json_response: for bug in json_response["bugs"]: print("Bug %s: %s" % (bug["id"], bug["summary"])) else: print("No bugs found.") def report( self, test_names, start, end, show_info, show_bugs, ): self.start = start self.end = end self.show_info = show_info if not self.show_info and not show_bugs: # by default, show everything self.show_info = True show_bugs = True for test_name in test_names: print("===== %s =====" % test_name) self.test_name = test_name if len(self.test_name) < 6: print("'%s' is too short for a test name!" % self.test_name) continue self.set_test_name() if show_bugs: self.report_bugs() class TestInfoReport(TestInfo): """ Support 'mach test-info report': Report of test runs summarized by manifest and component. """ def __init__(self, verbose): TestInfo.__init__(self, verbose) self.threads = [] def update_report(self, by_component, result, path_mod): def update_item(item, label, value): # It is important to include any existing item value in case ActiveData # returns multiple records for the same test; that can happen if the report # sometimes maps more than one ActiveData record to the same path. new_value = item.get(label, 0) + value if type(new_value) == int: item[label] = new_value else: item[label] = float(round(new_value, 2)) # pylint: disable=W1633 if "test" in result and "tests" in by_component: test = result["test"] if path_mod: test = path_mod(test) for bc in by_component["tests"]: for item in by_component["tests"][bc]: if test == item["test"]: # pylint: disable=W1633 seconds = float(round(result.get("duration", 0), 2)) update_item(item, "total run time, seconds", seconds) update_item(item, "total runs", result.get("count", 0)) update_item(item, "skipped runs", result.get("skips", 0)) update_item(item, "failed runs", result.get("failures", 0)) return True return False def path_mod_reftest(self, path): # " == " -> "" path = path.split(" ")[0] # "?" -> "" path = path.split("?")[0] # "#" -> "" path = path.split("#")[0] return path def path_mod_jsreftest(self, path): # ";assert" -> "" path = path.split(";")[0] return path def path_mod_marionette(self, path): # " " -> "" path = path.split(" ")[0] # "part1\part2" -> "part1/part2" path = path.replace("\\", os.path.sep) return path def path_mod_wpt(self, path): if path[0] == os.path.sep: # "/" -> "" path = path[1:] # "" -> "testing/web-platform/tests/" path = os.path.join("testing", "web-platform", "tests", path) # "?" -> "" path = path.split("?")[0] return path def path_mod_jittest(self, path): # "part1\part2" -> "part1/part2" path = path.replace("\\", os.path.sep) # "" -> "js/src/jit-test/tests/" return os.path.join("js", "src", "jit-test", "tests", path) def path_mod_xpcshell(self, path): # .ini: -> "" path = path.split(".ini:")[-1] return path def description( self, components, flavor, subsuite, paths, show_manifests, show_tests, show_summary, show_annotations, filter_values, filter_keys, ): # provide a natural language description of the report options what = [] if show_manifests: what.append("test manifests") if show_tests: what.append("tests") if show_annotations: what.append("test manifest annotations") if show_summary and len(what) == 0: what.append("summary of tests only") if len(what) > 1: what[-1] = "and " + what[-1] what = ", ".join(what) d = "Test summary report for " + what if components: d += ", in specified components (%s)" % components else: d += ", in all components" if flavor: d += ", in specified flavor (%s)" % flavor if subsuite: d += ", in specified subsuite (%s)" % subsuite if paths: d += ", on specified paths (%s)" % paths if filter_values: d += ", containing '%s'" % filter_values if filter_keys: d += " in manifest keys '%s'" % filter_keys else: d += " in any part of manifest entry" d += " as of %s." % datetime.datetime.now().strftime("%Y-%m-%d %H:%M") return d def report( self, components, flavor, subsuite, paths, show_manifests, show_tests, show_summary, show_annotations, filter_values, filter_keys, show_components, output_file, ): def matches_filters(test): """ Return True if all of the requested filter_values are found in this test; if filter_keys are specified, restrict search to those test keys. """ for value in filter_values: value_found = False for key in test: if not filter_keys or key in filter_keys: if re.search(value, test[key]): value_found = True break if not value_found: return False return True start_time = datetime.datetime.now() # Ensure useful report by default if ( not show_manifests and not show_tests and not show_summary and not show_annotations ): show_manifests = True show_summary = True by_component = {} if components: components = components.split(",") if filter_keys: filter_keys = filter_keys.split(",") if filter_values: filter_values = filter_values.split(",") else: filter_values = [] display_keys = (filter_keys or []) + ["skip-if", "fail-if", "fails-if"] display_keys = set(display_keys) print("Finding tests...") here = os.path.abspath(os.path.dirname(__file__)) resolver = TestResolver.from_environment( cwd=here, loader_cls=TestManifestLoader ) tests = list( resolver.resolve_tests(paths=paths, flavor=flavor, subsuite=subsuite) ) manifest_paths = set() for t in tests: if "manifest" in t and t["manifest"] is not None: manifest_paths.add(t["manifest"]) manifest_count = len(manifest_paths) print( "Resolver found {} tests, {} manifests".format(len(tests), manifest_count) ) if show_manifests: topsrcdir = self.build_obj.topsrcdir by_component["manifests"] = {} manifest_paths = list(manifest_paths) manifest_paths.sort() relpaths = [] for manifest_path in manifest_paths: relpath = mozpath.relpath(manifest_path, topsrcdir) if mozpath.commonprefix((manifest_path, topsrcdir)) != topsrcdir: continue relpaths.append(relpath) reader = self.build_obj.mozbuild_reader(config_mode="empty") files_info = reader.files_info(relpaths) for manifest_path in manifest_paths: relpath = mozpath.relpath(manifest_path, topsrcdir) if mozpath.commonprefix((manifest_path, topsrcdir)) != topsrcdir: continue manifest_info = None if relpath in files_info: bug_component = files_info[relpath].get("BUG_COMPONENT") if bug_component: key = "{}::{}".format( bug_component.product, bug_component.component ) else: key = "" if (not components) or (key in components): manifest_info = {"manifest": relpath, "tests": 0, "skipped": 0} rkey = key if show_components else "all" if rkey in by_component["manifests"]: by_component["manifests"][rkey].append(manifest_info) else: by_component["manifests"][rkey] = [manifest_info] if manifest_info: for t in tests: if t["manifest"] == manifest_path: manifest_info["tests"] += 1 if t.get("skip-if"): manifest_info["skipped"] += 1 for key in by_component["manifests"]: by_component["manifests"][key].sort(key=lambda k: k["manifest"]) if show_tests: by_component["tests"] = {} if show_tests or show_summary or show_annotations: test_count = 0 failed_count = 0 skipped_count = 0 annotation_count = 0 condition_count = 0 component_set = set() relpaths = [] conditions = {} known_unconditional_annotations = ["skip", "fail", "asserts", "random"] known_conditional_annotations = [ "skip-if", "fail-if", "run-if", "fails-if", "fuzzy-if", "random-if", "asserts-if", ] for t in tests: relpath = t.get("srcdir_relpath") relpaths.append(relpath) reader = self.build_obj.mozbuild_reader(config_mode="empty") files_info = reader.files_info(relpaths) for t in tests: if not matches_filters(t): continue if "referenced-test" in t: # Avoid double-counting reftests: disregard reference file entries continue if show_annotations: for key in t: if key in known_unconditional_annotations: annotation_count += 1 if key in known_conditional_annotations: annotation_count += 1 # Here 'key' is a manifest annotation type like 'skip-if' and t[key] # is the associated condition. For example, the manifestparser # manifest annotation, "skip-if = os == 'win'", is expected to be # encoded as t['skip-if'] = "os == 'win'". # To allow for reftest manifests, t[key] may have multiple entries # separated by ';', each corresponding to a condition for that test # and annotation type. For example, # "skip-if(Android&&webrender) skip-if(OSX)", would be # encoded as t['skip-if'] = "Android&&webrender;OSX". annotation_conditions = t[key].split(";") for condition in annotation_conditions: condition_count += 1 # Trim reftest fuzzy-if ranges: everything after the first comma # eg. "Android,0-2,1-3" -> "Android" condition = condition.split(",")[0] if condition not in conditions: conditions[condition] = 0 conditions[condition] += 1 test_count += 1 relpath = t.get("srcdir_relpath") if relpath in files_info: bug_component = files_info[relpath].get("BUG_COMPONENT") if bug_component: key = "{}::{}".format( bug_component.product, bug_component.component ) else: key = "" if (not components) or (key in components): component_set.add(key) test_info = {"test": relpath} for test_key in display_keys: value = t.get(test_key) if value: test_info[test_key] = value if t.get("fail-if"): failed_count += 1 if t.get("fails-if"): failed_count += 1 if t.get("skip-if"): skipped_count += 1 if show_tests: rkey = key if show_components else "all" if rkey in by_component["tests"]: # Avoid duplicates: Some test paths have multiple TestResolver # entries, as when a test is included by multiple manifests. found = False for ctest in by_component["tests"][rkey]: if ctest["test"] == test_info["test"]: found = True break if not found: by_component["tests"][rkey].append(test_info) else: by_component["tests"][rkey] = [test_info] if show_tests: for key in by_component["tests"]: by_component["tests"][key].sort(key=lambda k: k["test"]) by_component["description"] = self.description( components, flavor, subsuite, paths, show_manifests, show_tests, show_summary, show_annotations, filter_values, filter_keys, ) if show_summary: by_component["summary"] = {} by_component["summary"]["components"] = len(component_set) by_component["summary"]["manifests"] = manifest_count by_component["summary"]["tests"] = test_count by_component["summary"]["failed tests"] = failed_count by_component["summary"]["skipped tests"] = skipped_count if show_annotations: by_component["annotations"] = {} by_component["annotations"]["total annotations"] = annotation_count by_component["annotations"]["total conditions"] = condition_count by_component["annotations"]["unique conditions"] = len(conditions) by_component["annotations"]["conditions"] = conditions self.write_report(by_component, output_file) end_time = datetime.datetime.now() self.log_verbose( "%d seconds total to generate report" % (end_time - start_time).total_seconds() ) def write_report(self, by_component, output_file): json_report = json.dumps(by_component, indent=2, sort_keys=True) if output_file: output_file = os.path.abspath(output_file) output_dir = os.path.dirname(output_file) if not os.path.isdir(output_dir): os.makedirs(output_dir) with open(output_file, "w") as f: f.write(json_report) else: print(json_report) def report_diff(self, before, after, output_file): """ Support for 'mach test-info report-diff'. """ def get_file(path_or_url): if urlparse.urlparse(path_or_url).scheme: response = requests.get(path_or_url) response.raise_for_status() return json.loads(response.text) with open(path_or_url) as f: return json.load(f) report1 = get_file(before) report2 = get_file(after) by_component = {"tests": {}, "summary": {}} self.diff_summaries(by_component, report1["summary"], report2["summary"]) self.diff_all_components(by_component, report1["tests"], report2["tests"]) self.write_report(by_component, output_file) def diff_summaries(self, by_component, summary1, summary2): """ Update by_component with comparison of summaries. """ all_keys = set(summary1.keys()) | set(summary2.keys()) for key in all_keys: delta = summary2.get(key, 0) - summary1.get(key, 0) by_component["summary"]["%s delta" % key] = delta def diff_all_components(self, by_component, tests1, tests2): """ Update by_component with any added/deleted tests, for all components. """ self.added_count = 0 self.deleted_count = 0 for component in tests1: component1 = tests1[component] component2 = [] if component not in tests2 else tests2[component] self.diff_component(by_component, component, component1, component2) for component in tests2: if component not in tests1: component2 = tests2[component] self.diff_component(by_component, component, [], component2) by_component["summary"]["added tests"] = self.added_count by_component["summary"]["deleted tests"] = self.deleted_count def diff_component(self, by_component, component, component1, component2): """ Update by_component[component] with any added/deleted tests for the named component. "added": tests found in component2 but missing from component1. "deleted": tests found in component1 but missing from component2. """ tests1 = set([t["test"] for t in component1]) tests2 = set([t["test"] for t in component2]) deleted = tests1 - tests2 added = tests2 - tests1 if deleted or added: by_component["tests"][component] = {} if deleted: by_component["tests"][component]["deleted"] = sorted(list(deleted)) if added: by_component["tests"][component]["added"] = sorted(list(added)) self.added_count += len(added) self.deleted_count += len(deleted) common = len(tests1.intersection(tests2)) self.log_verbose( "%s: %d deleted, %d added, %d common" % (component, len(deleted), len(added), common) )