зеркало из https://github.com/mozilla/gecko-dev.git
691 строка
28 KiB
Python
691 строка
28 KiB
Python
# This Source Code Form is subject to the terms of the Mozilla Public
|
|
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
|
|
|
from __future__ import absolute_import, print_function, unicode_literals
|
|
|
|
import argparse
|
|
import copy
|
|
import logging
|
|
import re
|
|
import shlex
|
|
import six
|
|
from collections import defaultdict
|
|
|
|
logger = logging.getLogger(__name__)
|
|
|
|
# The build type aliases are very cryptic and only used in try flags these are
|
|
# mappings from the single char alias to a longer more recognizable form.
|
|
BUILD_TYPE_ALIASES = {
|
|
'o': 'opt',
|
|
'd': 'debug'
|
|
}
|
|
|
|
# consider anything in this whitelist of kinds to be governed by -b/-p
|
|
BUILD_KINDS = set([
|
|
'build',
|
|
'artifact-build',
|
|
'hazard',
|
|
'l10n',
|
|
'valgrind',
|
|
'spidermonkey',
|
|
])
|
|
|
|
|
|
# mapping from shortcut name (usable with -u) to a boolean function identifying
|
|
# matching test names
|
|
def alias_prefix(prefix):
|
|
return lambda name: name.startswith(prefix)
|
|
|
|
|
|
def alias_contains(infix):
|
|
return lambda name: infix in name
|
|
|
|
|
|
def alias_matches(pattern):
|
|
pattern = re.compile(pattern)
|
|
return lambda name: pattern.match(name)
|
|
|
|
|
|
UNITTEST_ALIASES = {
|
|
# Aliases specify shorthands that can be used in try syntax. The shorthand
|
|
# is the dictionary key, with the value representing a pattern for matching
|
|
# unittest_try_names.
|
|
#
|
|
# Note that alias expansion is performed in the absence of any chunk
|
|
# prefixes. For example, the first example above would replace "foo-7"
|
|
# with "foobar-7". Note that a few aliases allowed chunks to be specified
|
|
# without a leading `-`, for example 'mochitest-dt1'. That's no longer
|
|
# supported.
|
|
'cppunit': alias_prefix('cppunit'),
|
|
'crashtest': alias_prefix('crashtest'),
|
|
'crashtest-e10s': alias_prefix('crashtest-e10s'),
|
|
'e10s': alias_contains('e10s'),
|
|
'firefox-ui-functional': alias_prefix('firefox-ui-functional'),
|
|
'gaia-js-integration': alias_contains('gaia-js-integration'),
|
|
'gtest': alias_prefix('gtest'),
|
|
'jittest': alias_prefix('jittest'),
|
|
'jittests': alias_prefix('jittest'),
|
|
'jsreftest': alias_prefix('jsreftest'),
|
|
'jsreftest-e10s': alias_prefix('jsreftest-e10s'),
|
|
'marionette': alias_prefix('marionette'),
|
|
'mochitest': alias_prefix('mochitest'),
|
|
'mochitests': alias_prefix('mochitest'),
|
|
'mochitest-e10s': alias_prefix('mochitest-e10s'),
|
|
'mochitests-e10s': alias_prefix('mochitest-e10s'),
|
|
'mochitest-debug': alias_prefix('mochitest-debug-'),
|
|
'mochitest-a11y': alias_contains('mochitest-a11y'),
|
|
'mochitest-bc': alias_prefix('mochitest-browser-chrome'),
|
|
'mochitest-e10s-bc': alias_prefix('mochitest-browser-chrome-e10s'),
|
|
'mochitest-browser-chrome': alias_prefix('mochitest-browser-chrome'),
|
|
'mochitest-e10s-browser-chrome': alias_prefix('mochitest-browser-chrome-e10s'),
|
|
'mochitest-chrome': alias_contains('mochitest-chrome'),
|
|
'mochitest-dt': alias_prefix('mochitest-devtools-chrome'),
|
|
'mochitest-e10s-dt': alias_prefix('mochitest-devtools-chrome-e10s'),
|
|
'mochitest-gl': alias_prefix('mochitest-webgl'),
|
|
'mochitest-gl-e10s': alias_prefix('mochitest-webgl-e10s'),
|
|
'mochitest-gpu': alias_prefix('mochitest-gpu'),
|
|
'mochitest-gpu-e10s': alias_prefix('mochitest-gpu-e10s'),
|
|
'mochitest-media': alias_prefix('mochitest-media'),
|
|
'mochitest-media-e10s': alias_prefix('mochitest-media-e10s'),
|
|
'mochitest-vg': alias_prefix('mochitest-valgrind'),
|
|
'reftest': alias_matches(r'^(plain-)?reftest.*$'),
|
|
'reftest-no-accel': alias_matches(r'^(plain-)?reftest-no-accel.*$'),
|
|
'reftests': alias_matches(r'^(plain-)?reftest.*$'),
|
|
'reftests-e10s': alias_matches(r'^(plain-)?reftest-e10s.*$'),
|
|
'reftest-gpu': alias_matches(r'^(plain-)?reftest-gpu.*$'),
|
|
'robocop': alias_prefix('robocop'),
|
|
'web-platform-test': alias_prefix('web-platform-tests'),
|
|
'web-platform-tests': alias_prefix('web-platform-tests'),
|
|
'web-platform-tests-e10s': alias_prefix('web-platform-tests-e10s'),
|
|
'web-platform-tests-crashtests': alias_prefix('web-platform-tests-crashtest'),
|
|
'web-platform-tests-print-reftest': alias_prefix('web-platform-tests-print-reftest'),
|
|
'web-platform-tests-reftests': alias_prefix('web-platform-tests-reftest'),
|
|
'web-platform-tests-reftests-e10s': alias_prefix('web-platform-tests-reftest-e10s'),
|
|
'web-platform-tests-wdspec': alias_prefix('web-platform-tests-wdspec'),
|
|
'web-platform-tests-wdspec-e10s': alias_prefix('web-platform-tests-wdspec-e10s'),
|
|
'xpcshell': alias_prefix('xpcshell'),
|
|
}
|
|
|
|
# unittest platforms can be specified by substring of the "pretty name", which
|
|
# is basically the old Buildbot builder name. This dict has {pretty name,
|
|
# [test_platforms]} translations, This includes only the most commonly-used
|
|
# substrings. It is OK to add new test platforms to various shorthands here;
|
|
# if you add a new Linux64 test platform for instance, people will expect that
|
|
# their previous methods of requesting "all linux64 tests" will include this
|
|
# new platform, and they shouldn't have to explicitly spell out the new platform
|
|
# every time for such cases.
|
|
#
|
|
# Note that the test platforms here are only the prefix up to the `/`.
|
|
UNITTEST_PLATFORM_PRETTY_NAMES = {
|
|
'Ubuntu': [
|
|
'linux32',
|
|
'linux64',
|
|
'linux64-asan',
|
|
'linux1804-64',
|
|
'linux1804-64-asan'
|
|
],
|
|
'x64': [
|
|
'linux64',
|
|
'linux64-asan',
|
|
'linux1804-64',
|
|
'linux1804-64-asan'
|
|
],
|
|
'Android 7.0 Moto G5 32bit': ['android-hw-g5-7.0-arm7-api-16'],
|
|
'Android 8.0 Google Pixel 2 32bit': ['android-hw-p2-8.0-arm7-api-16'],
|
|
'Android 8.0 Google Pixel 2 64bit': ['android-hw-p2-8.0-android-aarch64'],
|
|
'10.14': ['macosx1014-64'],
|
|
'Windows 7': ['windows7-32'],
|
|
'Windows 7 VM': ['windows7-32-vm'],
|
|
'Windows 10': ['windows10-64'],
|
|
}
|
|
|
|
TEST_CHUNK_SUFFIX = re.compile('(.*)-([0-9]+)$')
|
|
|
|
|
|
def escape_whitespace_in_brackets(input_str):
|
|
'''
|
|
In tests you may restrict them by platform [] inside of the brackets
|
|
whitespace may occur this is typically invalid shell syntax so we escape it
|
|
with backslash sequences .
|
|
'''
|
|
result = ""
|
|
in_brackets = False
|
|
for char in input_str:
|
|
if char == '[':
|
|
in_brackets = True
|
|
result += char
|
|
continue
|
|
|
|
if char == ']':
|
|
in_brackets = False
|
|
result += char
|
|
continue
|
|
|
|
if char == ' ' and in_brackets:
|
|
result += '\ '
|
|
continue
|
|
|
|
result += char
|
|
|
|
return result
|
|
|
|
|
|
def split_try_msg(message):
|
|
try:
|
|
try_idx = message.index('try:')
|
|
except ValueError:
|
|
return []
|
|
message = message[try_idx:].split('\n')[0]
|
|
# shlex used to ensure we split correctly when giving values to argparse.
|
|
return shlex.split(escape_whitespace_in_brackets(message))
|
|
|
|
|
|
def parse_message(message):
|
|
parts = split_try_msg(message)
|
|
|
|
# Argument parser based on try flag flags
|
|
parser = argparse.ArgumentParser()
|
|
parser.add_argument('-b', '--build', dest='build_types')
|
|
parser.add_argument('-p', '--platform', nargs='?',
|
|
dest='platforms', const='all', default='all')
|
|
parser.add_argument('-u', '--unittests', nargs='?',
|
|
dest='unittests', const='all', default='all')
|
|
parser.add_argument('-t', '--talos', nargs='?', dest='talos', const='all', default='none')
|
|
parser.add_argument('-r', '--raptor', nargs='?', dest='raptor', const='all', default='none')
|
|
parser.add_argument('-i', '--interactive',
|
|
dest='interactive', action='store_true', default=False)
|
|
parser.add_argument('-e', '--all-emails',
|
|
dest='notifications', action='store_const', const='all')
|
|
parser.add_argument('-f', '--failure-emails',
|
|
dest='notifications', action='store_const', const='failure')
|
|
parser.add_argument('-j', '--job', dest='jobs', action='append')
|
|
parser.add_argument('--rebuild-talos', dest='talos_trigger_tests', action='store',
|
|
type=int, default=1)
|
|
parser.add_argument('--rebuild-raptor', dest='raptor_trigger_tests', action='store',
|
|
type=int, default=1)
|
|
parser.add_argument('--setenv', dest='env', action='append')
|
|
parser.add_argument('--gecko-profile', dest='profile', action='store_true')
|
|
parser.add_argument('--tag', dest='tag', action='store', default=None)
|
|
parser.add_argument('--no-retry', dest='no_retry', action='store_true')
|
|
parser.add_argument('--include-nightly', dest='include_nightly', action='store_true')
|
|
parser.add_argument('--artifact', dest='artifact', action='store_true')
|
|
|
|
# While we are transitioning from BB to TC, we want to push jobs to tc-worker
|
|
# machines but not overload machines with every try push. Therefore, we add
|
|
# this temporary option to be able to push jobs to tc-worker.
|
|
parser.add_argument('-w', '--taskcluster-worker',
|
|
dest='taskcluster_worker', action='store_true', default=False)
|
|
|
|
# In order to run test jobs multiple times
|
|
parser.add_argument('--rebuild', dest='trigger_tests', type=int, default=1)
|
|
args, _ = parser.parse_known_args(parts)
|
|
|
|
try_options = vars(args)
|
|
try_task_config = {
|
|
"use-artifact-builds": try_options.pop("artifact"),
|
|
"gecko-profile": try_options.pop("profile"),
|
|
"env": dict(arg.split("=") for arg in try_options.pop("env") or [])
|
|
}
|
|
return {
|
|
"try_options": try_options,
|
|
"try_task_config": try_task_config,
|
|
}
|
|
|
|
|
|
class TryOptionSyntax(object):
|
|
|
|
def __init__(self, parameters, full_task_graph, graph_config):
|
|
"""
|
|
Apply the try options in parameters.
|
|
|
|
The resulting object has attributes:
|
|
|
|
- build_types: a list containing zero or more of 'opt' and 'debug'
|
|
- platforms: a list of selected platform names, or None for all
|
|
- unittests: a list of tests, of the form given below, or None for all
|
|
- jobs: a list of requested job names, or None for all
|
|
- trigger_tests: the number of times tests should be triggered (--rebuild)
|
|
- interactive: true if --interactive
|
|
- notifications: either None if no notifications or one of 'all' or 'failure'
|
|
- talos_trigger_tests: the number of time talos tests should be triggered (--rebuild-talos)
|
|
- tag: restrict tests to the specified tag
|
|
- no_retry: do not retry failed jobs
|
|
|
|
The unittests and talos lists contain dictionaries of the form:
|
|
|
|
{
|
|
'test': '<suite name>',
|
|
'platforms': [..platform names..], # to limit to only certain platforms
|
|
'only_chunks': set([..chunk numbers..]), # to limit only to certain chunks
|
|
}
|
|
"""
|
|
self.full_task_graph = full_task_graph
|
|
self.graph_config = graph_config
|
|
self.jobs = []
|
|
self.build_types = []
|
|
self.platforms = []
|
|
self.unittests = []
|
|
self.talos = []
|
|
self.raptor = []
|
|
self.trigger_tests = 0
|
|
self.interactive = False
|
|
self.notifications = None
|
|
self.talos_trigger_tests = 0
|
|
self.raptor_trigger_tests = 0
|
|
self.tag = None
|
|
self.no_retry = False
|
|
|
|
options = parameters['try_options']
|
|
if not options:
|
|
return None
|
|
self.jobs = self.parse_jobs(options['jobs'])
|
|
self.build_types = self.parse_build_types(options['build_types'], full_task_graph)
|
|
self.platforms = self.parse_platforms(options, full_task_graph)
|
|
self.unittests = self.parse_test_option(
|
|
"unittest_try_name", options['unittests'], full_task_graph)
|
|
self.talos = self.parse_test_option("talos_try_name", options['talos'], full_task_graph)
|
|
self.raptor = self.parse_test_option("raptor_try_name", options['raptor'], full_task_graph)
|
|
self.trigger_tests = options['trigger_tests']
|
|
self.interactive = options['interactive']
|
|
self.notifications = options['notifications']
|
|
self.talos_trigger_tests = options['talos_trigger_tests']
|
|
self.raptor_trigger_tests = options['raptor_trigger_tests']
|
|
self.tag = options['tag']
|
|
self.no_retry = options['no_retry']
|
|
self.include_nightly = options['include_nightly']
|
|
|
|
self.test_tiers = self.generate_test_tiers(full_task_graph)
|
|
|
|
def generate_test_tiers(self, full_task_graph):
|
|
retval = defaultdict(set)
|
|
for t in six.itervalues(full_task_graph.tasks):
|
|
if t.attributes.get('kind') == 'test':
|
|
try:
|
|
tier = t.task['extra']['treeherder']['tier']
|
|
name = t.attributes.get('unittest_try_name')
|
|
retval[name].add(tier)
|
|
except KeyError:
|
|
pass
|
|
|
|
return retval
|
|
|
|
def parse_jobs(self, jobs_arg):
|
|
if not jobs_arg or jobs_arg == ['none']:
|
|
return [] # default is `-j none`
|
|
if jobs_arg == ['all']:
|
|
return None
|
|
expanded = []
|
|
for job in jobs_arg:
|
|
expanded.extend(j.strip() for j in job.split(','))
|
|
return expanded
|
|
|
|
def parse_build_types(self, build_types_arg, full_task_graph):
|
|
if build_types_arg is None:
|
|
build_types_arg = []
|
|
|
|
build_types = [_f for _f in (BUILD_TYPE_ALIASES.get(build_type) for
|
|
build_type in build_types_arg) if _f]
|
|
|
|
all_types = set(t.attributes['build_type']
|
|
for t in six.itervalues(full_task_graph.tasks)
|
|
if 'build_type' in t.attributes)
|
|
bad_types = set(build_types) - all_types
|
|
if bad_types:
|
|
raise Exception("Unknown build type(s) [%s] specified for try" % ','.join(bad_types))
|
|
|
|
return build_types
|
|
|
|
def parse_platforms(self, options, full_task_graph):
|
|
platform_arg = options['platforms']
|
|
if platform_arg == 'all':
|
|
return None
|
|
|
|
RIDEALONG_BUILDS = self.graph_config['try']['ridealong-builds']
|
|
results = []
|
|
for build in platform_arg.split(','):
|
|
if build in ('macosx64',):
|
|
# Regular opt builds are faster than shippable ones, but we don't run
|
|
# tests against them.
|
|
# We want to choose them (and only them) if no tests were requested.
|
|
if options['unittests'] == 'none' and options['talos'] == 'none' and \
|
|
options['raptor'] == 'none':
|
|
results.append('macosx64')
|
|
logger.info("adding macosx64 for try syntax using macosx64.")
|
|
# Otherwise, use _just_ the shippable builds.
|
|
else:
|
|
results.append('macosx64-shippable')
|
|
logger.info("adding macosx64-shippable for try syntax using macosx64.")
|
|
else:
|
|
results.append(build)
|
|
if build in RIDEALONG_BUILDS:
|
|
results.extend(RIDEALONG_BUILDS[build])
|
|
logger.info("platform %s triggers ridealong builds %s" %
|
|
(build, ', '.join(RIDEALONG_BUILDS[build])))
|
|
|
|
test_platforms = set(t.attributes['test_platform']
|
|
for t in six.itervalues(full_task_graph.tasks)
|
|
if 'test_platform' in t.attributes)
|
|
build_platforms = set(t.attributes['build_platform']
|
|
for t in six.itervalues(full_task_graph.tasks)
|
|
if 'build_platform' in t.attributes)
|
|
all_platforms = test_platforms | build_platforms
|
|
bad_platforms = set(results) - all_platforms
|
|
if bad_platforms:
|
|
raise Exception("Unknown platform(s) [%s] specified for try" % ','.join(bad_platforms))
|
|
|
|
return results
|
|
|
|
def parse_test_option(self, attr_name, test_arg, full_task_graph):
|
|
'''
|
|
|
|
Parse a unittest (-u) or talos (-t) option, in the context of a full
|
|
task graph containing available `unittest_try_name` or `talos_try_name`
|
|
attributes. There are three cases:
|
|
|
|
- test_arg is == 'none' (meaning an empty list)
|
|
- test_arg is == 'all' (meaning use the list of jobs for that job type)
|
|
- test_arg is comma string which needs to be parsed
|
|
'''
|
|
|
|
# Empty job list case...
|
|
if test_arg is None or test_arg == 'none':
|
|
return []
|
|
|
|
all_platforms = set(t.attributes['test_platform'].split('/')[0]
|
|
for t in six.itervalues(full_task_graph.tasks)
|
|
if 'test_platform' in t.attributes)
|
|
|
|
tests = self.parse_test_opts(test_arg, all_platforms)
|
|
|
|
if not tests:
|
|
return []
|
|
|
|
all_tests = set(t.attributes[attr_name]
|
|
for t in six.itervalues(full_task_graph.tasks)
|
|
if attr_name in t.attributes)
|
|
|
|
# Special case where tests is 'all' and must be expanded
|
|
if tests[0]['test'] == 'all':
|
|
results = []
|
|
all_entry = tests[0]
|
|
for test in all_tests:
|
|
entry = {'test': test}
|
|
# If there are platform restrictions copy them across the list.
|
|
if 'platforms' in all_entry:
|
|
entry['platforms'] = list(all_entry['platforms'])
|
|
results.append(entry)
|
|
return self.parse_test_chunks(all_tests, results)
|
|
else:
|
|
return self.parse_test_chunks(all_tests, tests)
|
|
|
|
def parse_test_opts(self, input_str, all_platforms):
|
|
'''
|
|
Parse `testspec,testspec,..`, where each testspec is a test name
|
|
optionally followed by a list of test platforms or negated platforms in
|
|
`[]`.
|
|
|
|
No brackets indicates that tests should run on all platforms for which
|
|
builds are available. If testspecs are provided, then each is treated,
|
|
from left to right, as an instruction to include or (if negated)
|
|
exclude a set of test platforms. A single spec may expand to multiple
|
|
test platforms via UNITTEST_PLATFORM_PRETTY_NAMES. If the first test
|
|
spec is negated, processing begins with the full set of available test
|
|
platforms; otherwise, processing begins with an empty set of test
|
|
platforms.
|
|
'''
|
|
|
|
# Final results which we will return.
|
|
tests = []
|
|
|
|
cur_test = {}
|
|
token = ''
|
|
in_platforms = False
|
|
|
|
def normalize_platforms():
|
|
if 'platforms' not in cur_test:
|
|
return
|
|
# if the first spec is a negation, start with all platforms
|
|
if cur_test['platforms'][0][0] == '-':
|
|
platforms = all_platforms.copy()
|
|
else:
|
|
platforms = []
|
|
for platform in cur_test['platforms']:
|
|
if platform[0] == '-':
|
|
platforms = [p for p in platforms if p != platform[1:]]
|
|
else:
|
|
platforms.append(platform)
|
|
cur_test['platforms'] = platforms
|
|
|
|
def add_test(value):
|
|
normalize_platforms()
|
|
cur_test['test'] = value.strip()
|
|
tests.insert(0, cur_test)
|
|
|
|
def add_platform(value):
|
|
platform = value.strip()
|
|
if platform[0] == '-':
|
|
negated = True
|
|
platform = platform[1:]
|
|
else:
|
|
negated = False
|
|
platforms = UNITTEST_PLATFORM_PRETTY_NAMES.get(platform, [platform])
|
|
if negated:
|
|
platforms = ["-" + p for p in platforms]
|
|
cur_test['platforms'] = platforms + cur_test.get('platforms', [])
|
|
|
|
# This might be somewhat confusing but we parse the string _backwards_ so
|
|
# there is no ambiguity over what state we are in.
|
|
for char in reversed(input_str):
|
|
|
|
# , indicates exiting a state
|
|
if char == ',':
|
|
|
|
# Exit a particular platform.
|
|
if in_platforms:
|
|
add_platform(token)
|
|
|
|
# Exit a particular test.
|
|
else:
|
|
add_test(token)
|
|
cur_test = {}
|
|
|
|
# Token must always be reset after we exit a state
|
|
token = ''
|
|
elif char == '[':
|
|
# Exiting platform state entering test state.
|
|
add_platform(token)
|
|
token = ''
|
|
in_platforms = False
|
|
elif char == ']':
|
|
# Entering platform state.
|
|
in_platforms = True
|
|
else:
|
|
# Accumulator.
|
|
token = char + token
|
|
|
|
# Handle any left over tokens.
|
|
if token:
|
|
add_test(token)
|
|
|
|
return tests
|
|
|
|
def handle_alias(self, test, all_tests):
|
|
'''
|
|
Expand a test if its name refers to an alias, returning a list of test
|
|
dictionaries cloned from the first (to maintain any metadata).
|
|
'''
|
|
if test['test'] not in UNITTEST_ALIASES:
|
|
return [test]
|
|
|
|
alias = UNITTEST_ALIASES[test['test']]
|
|
|
|
def mktest(name):
|
|
newtest = copy.deepcopy(test)
|
|
newtest['test'] = name
|
|
return newtest
|
|
|
|
def exprmatch(alias):
|
|
return [t for t in all_tests if alias(t)]
|
|
|
|
return [mktest(t) for t in exprmatch(alias)]
|
|
|
|
def parse_test_chunks(self, all_tests, tests):
|
|
'''
|
|
Test flags may include parameters to narrow down the number of chunks in a
|
|
given push. We don't model 1 chunk = 1 job in taskcluster so we must check
|
|
each test flag to see if it is actually specifying a chunk.
|
|
'''
|
|
results = []
|
|
seen_chunks = {}
|
|
for test in tests:
|
|
matches = TEST_CHUNK_SUFFIX.match(test['test'])
|
|
if matches:
|
|
name = matches.group(1)
|
|
chunk = matches.group(2)
|
|
if name in seen_chunks:
|
|
seen_chunks[name].add(chunk)
|
|
else:
|
|
seen_chunks[name] = {chunk}
|
|
test['test'] = name
|
|
test['only_chunks'] = seen_chunks[name]
|
|
results.append(test)
|
|
else:
|
|
results.extend(self.handle_alias(test, all_tests))
|
|
|
|
# uniquify the results over the test names
|
|
results = sorted(
|
|
{test["test"]: test for test in results}.values(),
|
|
key=lambda test: test["test"],
|
|
)
|
|
return results
|
|
|
|
def find_all_attribute_suffixes(self, graph, prefix):
|
|
rv = set()
|
|
for t in six.itervalues(graph.tasks):
|
|
for a in t.attributes:
|
|
if a.startswith(prefix):
|
|
rv.add(a[len(prefix):])
|
|
return sorted(rv)
|
|
|
|
def task_matches(self, task):
|
|
attr = task.attributes.get
|
|
|
|
def check_run_on_projects():
|
|
return {'all'} & set(attr('run_on_projects', []))
|
|
|
|
# Don't schedule fission tests when try option syntax is used
|
|
if attr('unittest_variant') == 'fission':
|
|
return False
|
|
|
|
def match_test(try_spec, attr_name):
|
|
run_by_default = True
|
|
if attr('build_type') not in self.build_types:
|
|
return False
|
|
|
|
if self.platforms is not None and attr('build_platform') not in self.platforms:
|
|
return False
|
|
elif not check_run_on_projects():
|
|
run_by_default = False
|
|
|
|
if try_spec is None:
|
|
return run_by_default
|
|
|
|
# TODO: optimize this search a bit
|
|
for test in try_spec:
|
|
if attr(attr_name) == test['test']:
|
|
break
|
|
else:
|
|
return False
|
|
|
|
if 'only_chunks' in test and attr('test_chunk') not in test['only_chunks']:
|
|
return False
|
|
|
|
tier = task.task['extra']['treeherder']['tier']
|
|
if 'platforms' in test:
|
|
if 'all' in test['platforms']:
|
|
return True
|
|
platform = attr('test_platform', '').split('/')[0]
|
|
# Platforms can be forced by syntax like "-u xpcshell[Windows 8]"
|
|
return platform in test['platforms']
|
|
elif tier != 1:
|
|
# Run Tier 2/3 tests if their build task is Tier 2/3 OR if there is
|
|
# no tier 1 test of that name.
|
|
build_task = self.full_task_graph.tasks[task.dependencies['build']]
|
|
build_task_tier = build_task.task['extra']['treeherder']['tier']
|
|
|
|
name = attr('unittest_try_name')
|
|
test_tiers = self.test_tiers.get(name)
|
|
|
|
if tier <= build_task_tier:
|
|
logger.debug("not skipping tier {} test {} because build task {} "
|
|
"is tier {}"
|
|
.format(tier, task.label, build_task.label,
|
|
build_task_tier))
|
|
return True
|
|
elif 1 not in test_tiers:
|
|
logger.debug("not skipping tier {} test {} without explicit inclusion; "
|
|
"it is configured to run on tiers {}"
|
|
.format(tier, task.label, test_tiers))
|
|
return True
|
|
else:
|
|
logger.debug("skipping tier {} test {} because build task {} is "
|
|
"tier {} and there is a higher-tier test of the same name"
|
|
.format(tier, task.label, build_task.label, build_task_tier))
|
|
return False
|
|
elif run_by_default:
|
|
return check_run_on_projects()
|
|
else:
|
|
return False
|
|
|
|
if attr('job_try_name'):
|
|
# Beware the subtle distinction between [] and None for self.jobs and self.platforms.
|
|
# They will be [] if there was no try syntax, and None if try syntax was detected but
|
|
# they remained unspecified.
|
|
if self.jobs is not None:
|
|
return attr('job_try_name') in self.jobs
|
|
|
|
# User specified `-j all`
|
|
if self.platforms is not None and attr('build_platform') not in self.platforms:
|
|
return False # honor -p for jobs governed by a platform
|
|
# "all" means "everything with `try` in run_on_projects"
|
|
return check_run_on_projects()
|
|
elif attr('kind') == 'test':
|
|
return match_test(self.unittests, 'unittest_try_name') \
|
|
or match_test(self.talos, 'talos_try_name') \
|
|
or match_test(self.raptor, 'raptor_try_name')
|
|
elif attr('kind') in BUILD_KINDS:
|
|
if attr('build_type') not in self.build_types:
|
|
return False
|
|
elif self.platforms is None:
|
|
# for "-p all", look for try in the 'run_on_projects' attribute
|
|
return check_run_on_projects()
|
|
else:
|
|
if attr('build_platform') not in self.platforms:
|
|
return False
|
|
return True
|
|
else:
|
|
return False
|
|
|
|
def __str__(self):
|
|
def none_for_all(list):
|
|
if list is None:
|
|
return '<all>'
|
|
return ', '.join(str(e) for e in list)
|
|
|
|
return "\n".join([
|
|
"build_types: " + ", ".join(self.build_types),
|
|
"platforms: " + none_for_all(self.platforms),
|
|
"unittests: " + none_for_all(self.unittests),
|
|
"talos: " + none_for_all(self.talos),
|
|
"raptor" + none_for_all(self.raptor),
|
|
"jobs: " + none_for_all(self.jobs),
|
|
"trigger_tests: " + str(self.trigger_tests),
|
|
"interactive: " + str(self.interactive),
|
|
"notifications: " + str(self.notifications),
|
|
"talos_trigger_tests: " + str(self.talos_trigger_tests),
|
|
"raptor_trigger_tests: " + str(self.raptor_trigger_tests),
|
|
"tag: " + str(self.tag),
|
|
"no_retry: " + str(self.no_retry),
|
|
])
|