gecko-dev/testing/xpcshell/runxpcshelltests.py

1595 строки
63 KiB
Python

#!/usr/bin/env python
#
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
import copy
import json
import math
import os
import os.path
import random
import shutil
import signal
import socket
import sys
import time
import traceback
import xml.dom.minidom
from collections import deque
from distutils import dir_util
from multiprocessing import cpu_count
from optparse import OptionParser
from subprocess import Popen, PIPE, STDOUT
from tempfile import mkdtemp, gettempdir
from threading import Timer, Thread, Event, RLock
try:
import psutil
HAVE_PSUTIL = True
except ImportError:
HAVE_PSUTIL = False
from automation import Automation, getGlobalLog, resetGlobalLog
from automationutils import *
# Printing buffered output in case of a failure or verbose mode will result
# in buffered output interleaved with other threads' output.
# To prevent his, each call to the logger as well as any blocks of output that
# are intended to be continuous are protected by the same lock.
LOG_MUTEX = RLock()
HARNESS_TIMEOUT = 5 * 60
# benchmarking on tbpl revealed that this works best for now
NUM_THREADS = int(cpu_count() * 4)
FAILURE_ACTIONS = set(['test_unexpected_fail',
'test_unexpected_pass',
'javascript_error'])
ACTION_STRINGS = {
"test_unexpected_fail": "TEST-UNEXPECTED-FAIL",
"test_known_fail": "TEST-KNOWN-FAIL",
"test_unexpected_pass": "TEST-UNEXPECTED-PASS",
"javascript_error": "TEST-UNEXPECTED-FAIL",
"test_pass": "TEST-PASS",
"test_info": "TEST-INFO"
}
# --------------------------------------------------------------
# TODO: this is a hack for mozbase without virtualenv, remove with bug 849900
#
here = os.path.dirname(__file__)
mozbase = os.path.realpath(os.path.join(os.path.dirname(here), 'mozbase'))
if os.path.isdir(mozbase):
for package in os.listdir(mozbase):
sys.path.append(os.path.join(mozbase, package))
import manifestparser
import mozcrash
import mozinfo
# ---------------------------------------------------------------
#TODO: replace this with json.loads when Python 2.6 is required.
def parse_json(j):
"""
Awful hack to parse a restricted subset of JSON strings into Python dicts.
"""
return eval(j, {'true':True,'false':False,'null':None})
""" Control-C handling """
gotSIGINT = False
def markGotSIGINT(signum, stackFrame):
global gotSIGINT
gotSIGINT = True
class XPCShellTestThread(Thread):
def __init__(self, test_object, event, cleanup_dir_list, retry=True,
tests_root_dir=None, app_dir_key=None, interactive=False,
verbose=False, pStdout=None, pStderr=None, keep_going=False,
log=None, **kwargs):
Thread.__init__(self)
self.daemon = True
self.test_object = test_object
self.cleanup_dir_list = cleanup_dir_list
self.retry = retry
self.appPath = kwargs.get('appPath')
self.xrePath = kwargs.get('xrePath')
self.testingModulesDir = kwargs.get('testingModulesDir')
self.debuggerInfo = kwargs.get('debuggerInfo')
self.pluginsPath = kwargs.get('pluginsPath')
self.httpdManifest = kwargs.get('httpdManifest')
self.httpdJSPath = kwargs.get('httpdJSPath')
self.headJSPath = kwargs.get('headJSPath')
self.testharnessdir = kwargs.get('testharnessdir')
self.profileName = kwargs.get('profileName')
self.singleFile = kwargs.get('singleFile')
self.env = copy.deepcopy(kwargs.get('env'))
self.symbolsPath = kwargs.get('symbolsPath')
self.logfiles = kwargs.get('logfiles')
self.xpcshell = kwargs.get('xpcshell')
self.xpcsRunArgs = kwargs.get('xpcsRunArgs')
self.failureManifest = kwargs.get('failureManifest')
self.on_message = kwargs.get('on_message')
self.tests_root_dir = tests_root_dir
self.app_dir_key = app_dir_key
self.interactive = interactive
self.verbose = verbose
self.pStdout = pStdout
self.pStderr = pStderr
self.keep_going = keep_going
self.log = log
# only one of these will be set to 1. adding them to the totals in
# the harness
self.passCount = 0
self.todoCount = 0
self.failCount = 0
self.output_lines = []
self.has_failure_output = False
self.saw_proc_start = False
self.saw_proc_end = False
# event from main thread to signal work done
self.event = event
self.done = False # explicitly set flag so we don't rely on thread.isAlive
def run(self):
try:
self.run_test()
except Exception as e:
self.exception = e
self.traceback = traceback.format_exc()
else:
self.exception = None
self.traceback = None
if self.retry:
self.log.info("TEST-INFO | %s | Test failed or timed out, will retry."
% self.test_object['name'])
self.done = True
self.event.set()
def kill(self, proc):
"""
Simple wrapper to kill a process.
On a remote system, this is overloaded to handle remote process communication.
"""
return proc.kill()
def removeDir(self, dirname):
"""
Simple wrapper to remove (recursively) a given directory.
On a remote system, we need to overload this to work on the remote filesystem.
"""
shutil.rmtree(dirname)
def poll(self, proc):
"""
Simple wrapper to check if a process has terminated.
On a remote system, this is overloaded to handle remote process communication.
"""
return proc.poll()
def createLogFile(self, test_file, stdout):
"""
For a given test file and stdout buffer, create a log file.
On a remote system we have to fix the test name since it can contain directories.
"""
with open(test_file + ".log", "w") as f:
f.write(stdout)
def getReturnCode(self, proc):
"""
Simple wrapper to get the return code for a given process.
On a remote system we overload this to work with the remote process management.
"""
return proc.returncode
def communicate(self, proc):
"""
Simple wrapper to communicate with a process.
On a remote system, this is overloaded to handle remote process communication.
"""
# Processing of incremental output put here to
# sidestep issues on remote platforms, where what we know
# as proc is a file pulled off of a device.
if proc.stdout:
while True:
line = proc.stdout.readline()
if not line:
break
self.process_line(line)
if self.saw_proc_start and not self.saw_proc_end:
self.has_failure_output = True
return proc.communicate()
def launchProcess(self, cmd, stdout, stderr, env, cwd):
"""
Simple wrapper to launch a process.
On a remote system, this is more complex and we need to overload this function.
"""
if HAVE_PSUTIL:
popen_func = psutil.Popen
else:
popen_func = Popen
proc = popen_func(cmd, stdout=stdout, stderr=stderr,
env=env, cwd=cwd)
return proc
def checkForCrashes(self,
dump_directory,
symbols_path,
test_name=None):
"""
Simple wrapper to check for crashes.
On a remote system, this is more complex and we need to overload this function.
"""
return mozcrash.check_for_crashes(dump_directory, symbols_path, test_name=test_name)
def logCommand(self, name, completeCmd, testdir):
self.log.info("TEST-INFO | %s | full command: %r" % (name, completeCmd))
self.log.info("TEST-INFO | %s | current directory: %r" % (name, testdir))
# Show only those environment variables that are changed from
# the ambient environment.
changedEnv = (set("%s=%s" % i for i in self.env.iteritems())
- set("%s=%s" % i for i in os.environ.iteritems()))
self.log.info("TEST-INFO | %s | environment: %s" % (name, list(changedEnv)))
def testTimeout(self, test_file, processPID):
if not self.retry:
self.log.error("TEST-UNEXPECTED-FAIL | %s | Test timed out" % test_file)
self.done = True
Automation().killAndGetStackNoScreenshot(processPID, self.appPath, self.debuggerInfo)
def buildCmdTestFile(self, name):
"""
Build the command line arguments for the test file.
On a remote system, this may be overloaded to use a remote path structure.
"""
return ['-e', 'const _TEST_FILE = ["%s"];' %
replaceBackSlashes(name)]
def setupTempDir(self):
tempDir = mkdtemp()
self.env["XPCSHELL_TEST_TEMP_DIR"] = tempDir
if self.interactive:
self.log.info("TEST-INFO | temp dir is %s" % tempDir)
return tempDir
def setupPluginsDir(self):
if not os.path.isdir(self.pluginsPath):
return None
pluginsDir = mkdtemp()
# shutil.copytree requires dst to not exist. Deleting the tempdir
# would make a race condition possible in a concurrent environment,
# so we are using dir_utils.copy_tree which accepts an existing dst
dir_util.copy_tree(self.pluginsPath, pluginsDir)
if self.interactive:
self.log.info("TEST-INFO | plugins dir is %s" % pluginsDir)
return pluginsDir
def setupProfileDir(self):
"""
Create a temporary folder for the profile and set appropriate environment variables.
When running check-interactive and check-one, the directory is well-defined and
retained for inspection once the tests complete.
On a remote system, this may be overloaded to use a remote path structure.
"""
if self.interactive or self.singleFile:
profileDir = os.path.join(gettempdir(), self.profileName, "xpcshellprofile")
try:
# This could be left over from previous runs
self.removeDir(profileDir)
except:
pass
os.makedirs(profileDir)
else:
profileDir = mkdtemp()
self.env["XPCSHELL_TEST_PROFILE_DIR"] = profileDir
if self.interactive or self.singleFile:
self.log.info("TEST-INFO | profile dir is %s" % profileDir)
return profileDir
def buildCmdHead(self, headfiles, tailfiles, xpcscmd):
"""
Build the command line arguments for the head and tail files,
along with the address of the webserver which some tests require.
On a remote system, this is overloaded to resolve quoting issues over a secondary command line.
"""
cmdH = ", ".join(['"' + replaceBackSlashes(f) + '"'
for f in headfiles])
cmdT = ", ".join(['"' + replaceBackSlashes(f) + '"'
for f in tailfiles])
return xpcscmd + \
['-e', 'const _SERVER_ADDR = "localhost"',
'-e', 'const _HEAD_FILES = [%s];' % cmdH,
'-e', 'const _TAIL_FILES = [%s];' % cmdT]
def getHeadAndTailFiles(self, test_object):
"""Obtain the list of head and tail files.
Returns a 2-tuple. The first element is a list of head files. The second
is a list of tail files.
"""
def sanitize_list(s, kind):
for f in s.strip().split(' '):
f = f.strip()
if len(f) < 1:
continue
path = os.path.normpath(os.path.join(test_object['here'], f))
if not os.path.exists(path):
raise Exception('%s file does not exist: %s' % (kind, path))
if not os.path.isfile(path):
raise Exception('%s file is not a file: %s' % (kind, path))
yield path
return (list(sanitize_list(test_object['head'], 'head')),
list(sanitize_list(test_object['tail'], 'tail')))
def buildXpcsCmd(self, testdir):
"""
Load the root head.js file as the first file in our test path, before other head, test, and tail files.
On a remote system, we overload this to add additional command line arguments, so this gets overloaded.
"""
# - NOTE: if you rename/add any of the constants set here, update
# do_load_child_test_harness() in head.js
if not self.appPath:
self.appPath = self.xrePath
self.xpcsCmd = [
self.xpcshell,
'-g', self.xrePath,
'-a', self.appPath,
'-r', self.httpdManifest,
'-m',
'-n',
'-s',
'-e', 'const _HTTPD_JS_PATH = "%s";' % self.httpdJSPath,
'-e', 'const _HEAD_JS_PATH = "%s";' % self.headJSPath
]
if self.testingModulesDir:
# Escape backslashes in string literal.
sanitized = self.testingModulesDir.replace('\\', '\\\\')
self.xpcsCmd.extend([
'-e',
'const _TESTING_MODULES_DIR = "%s";' % sanitized
])
self.xpcsCmd.extend(['-f', os.path.join(self.testharnessdir, 'head.js')])
if self.debuggerInfo:
self.xpcsCmd = [self.debuggerInfo["path"]] + self.debuggerInfo["args"] + self.xpcsCmd
# Automation doesn't specify a pluginsPath and xpcshell defaults to
# $APPDIR/plugins. We do the same here so we can carry on with
# setting up every test with its own plugins directory.
if not self.pluginsPath:
self.pluginsPath = os.path.join(self.appPath, 'plugins')
self.pluginsDir = self.setupPluginsDir()
if self.pluginsDir:
self.xpcsCmd.extend(['-p', self.pluginsDir])
def cleanupDir(self, directory, name, xunit_result):
if not os.path.exists(directory):
return
TRY_LIMIT = 25 # up to TRY_LIMIT attempts (one every second), because
# the Windows filesystem is slow to react to the changes
try_count = 0
while try_count < TRY_LIMIT:
try:
self.removeDir(directory)
except OSError:
self.log.info("TEST-INFO | Failed to remove directory: %s. Waiting." % directory)
# We suspect the filesystem may still be making changes. Wait a
# little bit and try again.
time.sleep(1)
try_count += 1
else:
# removed fine
return
# we try cleaning up again later at the end of the run
self.cleanup_dir_list.append(directory)
def clean_temp_dirs(self, name, stdout):
# We don't want to delete the profile when running check-interactive
# or check-one.
if self.profileDir and not self.interactive and not self.singleFile:
self.cleanupDir(self.profileDir, name, self.xunit_result)
self.cleanupDir(self.tempDir, name, self.xunit_result)
if self.pluginsDir:
self.cleanupDir(self.pluginsDir, name, self.xunit_result)
def message_from_line(self, line):
""" Given a line of raw output, convert to a string message. """
if isinstance(line, basestring):
# This function has received unstructured output.
if line:
if 'TEST-UNEXPECTED-' in line:
self.has_failure_output = True
return line
msg = ['%s: ' % line['process'] if 'process' in line else '']
# Each call to the logger in head.js either specified '_message'
# or both 'source_file' and 'diagnostic'. If either of these are
# missing, they ended up being undefined as a result of the way
# the test was run.
if '_message' in line:
msg.append(line['_message'])
if 'diagnostic' in line:
msg.append('\nDiagnostic: %s' % line['diagnostic'])
else:
msg.append('%s | %s | %s' % (ACTION_STRINGS[line['action']],
line.get('source_file', 'undefined'),
line.get('diagnostic', 'undefined')))
msg.append('\n%s' % line['stack'] if 'stack' in line else '')
return ''.join(msg)
def parse_output(self, output):
"""Parses process output for structured messages and saves output as it is
read. Sets self.has_failure_output in case of evidence of a failure"""
for line_string in output.splitlines():
self.process_line(line_string)
if self.saw_proc_start and not self.saw_proc_end:
self.has_failure_output = True
def report_message(self, line):
""" Reports a message to a consumer, both as a strucutured and
human-readable log message. """
message = self.message_from_line(line)
if message.endswith('\n'):
# A new line is always added by head.js to delimit messages,
# however consumers will want to supply their own.
message = message[:-1]
if self.on_message:
self.on_message(line, message)
else:
self.output_lines.append(message)
def process_line(self, line_string):
""" Parses a single line of output, determining its significance and
reporting a message.
"""
try:
line_object = json.loads(line_string)
if not isinstance(line_object, dict):
self.report_message(line_string)
return
except ValueError:
self.report_message(line_string)
return
if 'action' not in line_object:
# In case a test outputs something that happens to be valid
# JSON.
self.report_message(line_string)
return
action = line_object['action']
self.report_message(line_object)
if action in FAILURE_ACTIONS:
self.has_failure_output = True
elif action == 'child_test_start':
self.saw_proc_start = True
elif action == 'child_test_end':
self.saw_proc_end = True
def log_output(self, output):
"""Prints given output line-by-line to avoid overflowing buffers."""
self.log.info(">>>>>>>")
if output:
if isinstance(output, basestring):
output = output.splitlines()
for part in output:
# For multi-line output, such as a stack trace
for line in part.splitlines():
try:
line = line.decode('utf-8')
except UnicodeDecodeError:
self.log.info("TEST-INFO | %s | Detected non UTF-8 output."\
" Please modify the test to only print UTF-8." %
self.test_object['name'])
# add '?' instead of funky bytes
line = line.decode('utf-8', 'replace')
self.log.info(line)
self.log.info("<<<<<<<")
def run_test(self):
"""Run an individual xpcshell test."""
global gotSIGINT
name = self.test_object['path']
self.xunit_result = {'name': self.test_object['name'], 'classname': 'xpcshell'}
# The xUnit package is defined as the path component between the root
# dir and the test with path characters replaced with '.' (using Java
# class notation).
if self.tests_root_dir is not None:
self.tests_root_dir = os.path.normpath(self.tests_root_dir)
if self.test_object['here'].find(self.tests_root_dir) != 0:
raise Exception('tests_root_dir is not a parent path of %s' %
self.test_object['here'])
relpath = self.test_object['here'][len(self.tests_root_dir):].lstrip('/\\')
self.xunit_result['classname'] = relpath.replace('/', '.').replace('\\', '.')
# Check for skipped tests
if 'disabled' in self.test_object:
self.log.info('TEST-INFO | skipping %s | %s' %
(name, self.test_object['disabled']))
self.xunit_result['skipped'] = True
self.retry = False
self.keep_going = True
return
# Check for known-fail tests
expected = self.test_object['expected'] == 'pass'
# By default self.appPath will equal the gre dir. If specified in the
# xpcshell.ini file, set a different app dir for this test.
if self.app_dir_key and self.app_dir_key in self.test_object:
rel_app_dir = self.test_object[self.app_dir_key]
rel_app_dir = os.path.join(self.xrePath, rel_app_dir)
self.appPath = os.path.abspath(rel_app_dir)
else:
self.appPath = None
test_dir = os.path.dirname(name)
self.buildXpcsCmd(test_dir)
head_files, tail_files = self.getHeadAndTailFiles(self.test_object)
cmdH = self.buildCmdHead(head_files, tail_files, self.xpcsCmd)
# Create a profile and a temp dir that the JS harness can stick
# a profile and temporary data in
self.profileDir = self.setupProfileDir()
self.tempDir = self.setupTempDir()
# The test file will have to be loaded after the head files.
cmdT = self.buildCmdTestFile(name)
args = self.xpcsRunArgs[:]
if 'debug' in self.test_object:
args.insert(0, '-d')
completeCmd = cmdH + cmdT + args
testTimeoutInterval = HARNESS_TIMEOUT
# Allow a test to request a multiple of the timeout if it is expected to take long
if 'requesttimeoutfactor' in self.test_object:
testTimeoutInterval *= int(self.test_object['requesttimeoutfactor'])
testTimer = None
if not self.interactive and not self.debuggerInfo:
testTimer = Timer(testTimeoutInterval, lambda: self.testTimeout(name, proc.pid))
testTimer.start()
proc = None
stdout = None
stderr = None
try:
self.log.info("TEST-INFO | %s | running test ..." % name)
if self.verbose:
self.logCommand(name, completeCmd, test_dir)
startTime = time.time()
proc = self.launchProcess(completeCmd,
stdout=self.pStdout, stderr=self.pStderr, env=self.env, cwd=test_dir)
if self.interactive:
self.log.info("TEST-INFO | %s | Process ID: %d" % (name, proc.pid))
stdout, stderr = self.communicate(proc)
if self.interactive:
# Not sure what else to do here...
self.keep_going = True
return
if testTimer:
testTimer.cancel()
if stdout:
self.parse_output(stdout)
result = not (self.has_failure_output or
(self.getReturnCode(proc) != 0))
if result != expected:
if self.retry:
self.clean_temp_dirs(name, stdout)
return
failureType = "TEST-UNEXPECTED-%s" % ("FAIL" if expected else "PASS")
message = "%s | %s | test failed (with xpcshell return code: %d)" % (
failureType, name, self.getReturnCode(proc))
if self.output_lines:
message += ", see following log:"
with LOG_MUTEX:
self.log.error(message)
self.log_output(self.output_lines)
self.failCount += 1
self.xunit_result["passed"] = False
self.xunit_result["failure"] = {
"type": failureType,
"message": message,
"text": stdout
}
if self.failureManifest:
with open(self.failureManifest, 'a') as f:
f.write('[%s]\n' % self.test_object['path'])
for k, v in self.test_object.items():
f.write('%s = %s\n' % (k, v))
else:
now = time.time()
timeTaken = (now - startTime) * 1000
self.xunit_result["time"] = now - startTime
with LOG_MUTEX:
self.log.info("TEST-%s | %s | test passed (time: %.3fms)" % ("PASS" if expected else "KNOWN-FAIL", name, timeTaken))
if self.verbose:
self.log_output(self.output_lines)
self.xunit_result["passed"] = True
self.retry = False
if expected:
self.passCount = 1
else:
self.todoCount = 1
self.xunit_result["todo"] = True
if self.checkForCrashes(self.tempDir, self.symbolsPath, test_name=name):
if self.retry:
self.clean_temp_dirs(name, stdout)
return
message = "PROCESS-CRASH | %s | application crashed" % name
self.failCount = 1
self.xunit_result["passed"] = False
self.xunit_result["failure"] = {
"type": "PROCESS-CRASH",
"message": message,
"text": stdout
}
if self.logfiles and stdout:
self.createLogFile(name, stdout)
finally:
# We can sometimes get here before the process has terminated, which would
# cause removeDir() to fail - so check for the process & kill it it needed.
if proc and self.poll(proc) is None:
self.kill(proc)
if self.retry:
self.clean_temp_dirs(name, stdout)
return
with LOG_MUTEX:
message = "TEST-UNEXPECTED-FAIL | %s | Process still running after test!" % name
self.log.error(message)
self.log_output(self.output_lines)
self.failCount = 1
self.xunit_result["passed"] = False
self.xunit_result["failure"] = {
"type": "TEST-UNEXPECTED-FAIL",
"message": message,
"text": stdout
}
self.clean_temp_dirs(name, stdout)
if gotSIGINT:
self.xunit_result["passed"] = False
self.xunit_result["time"] = "0.0"
self.xunit_result["failure"] = {
"type": "SIGINT",
"message": "Received SIGINT",
"text": "Received SIGINT (control-C) during test execution."
}
self.log.error("TEST-UNEXPECTED-FAIL | Received SIGINT (control-C) during test execution")
if self.keep_going:
gotSIGINT = False
else:
self.keep_going = False
return
self.keep_going = True
class XPCShellTests(object):
log = getGlobalLog()
oldcwd = os.getcwd()
def __init__(self, log=None):
""" Init logging and node status """
if log:
resetGlobalLog(log)
# Each method of the underlying logger must acquire the log
# mutex before writing to stdout.
log_funs = ['debug', 'info', 'warning', 'error', 'critical', 'log']
for fun_name in log_funs:
unwrapped = getattr(self.log, fun_name, None)
if unwrapped:
def wrap(fn):
def wrapped(*args, **kwargs):
with LOG_MUTEX:
fn(*args, **kwargs)
return wrapped
setattr(self.log, fun_name, wrap(unwrapped))
self.nodeProc = None
def buildTestList(self):
"""
read the xpcshell.ini manifest and set self.alltests to be
an array of test objects.
if we are chunking tests, it will be done here as well
"""
mp = manifestparser.TestManifest(strict=False)
if self.manifest is None:
for testdir in self.testdirs:
if testdir:
mp.read(os.path.join(testdir, 'xpcshell.ini'))
else:
mp.read(self.manifest)
self.buildTestPath()
self.alltests = mp.active_tests(**mozinfo.info)
if self.singleFile is None and self.totalChunks > 1:
self.chunkTests()
def chunkTests(self):
"""
Split the list of tests up into [totalChunks] pieces and filter the
self.alltests based on thisChunk, so we only run a subset.
"""
totalTests = 0
for dir in self.alltests:
totalTests += len(self.alltests[dir])
testsPerChunk = math.ceil(totalTests / float(self.totalChunks))
start = int(round((self.thisChunk-1) * testsPerChunk))
end = start + testsPerChunk
currentCount = 0
templist = {}
for dir in self.alltests:
startPosition = 0
dirCount = len(self.alltests[dir])
endPosition = dirCount
if currentCount < start and currentCount + dirCount >= start:
startPosition = int(start - currentCount)
if currentCount + dirCount > end:
endPosition = int(end - currentCount)
if end - currentCount < 0 or (currentCount + dirCount < start):
endPosition = 0
if startPosition is not endPosition:
templist[dir] = self.alltests[dir][startPosition:endPosition]
currentCount += dirCount
self.alltests = templist
def setAbsPath(self):
"""
Set the absolute path for xpcshell, httpdjspath and xrepath.
These 3 variables depend on input from the command line and we need to allow for absolute paths.
This function is overloaded for a remote solution as os.path* won't work remotely.
"""
self.testharnessdir = os.path.dirname(os.path.abspath(__file__))
self.headJSPath = self.testharnessdir.replace("\\", "/") + "/head.js"
self.xpcshell = os.path.abspath(self.xpcshell)
# we assume that httpd.js lives in components/ relative to xpcshell
self.httpdJSPath = os.path.join(os.path.dirname(self.xpcshell), 'components', 'httpd.js')
self.httpdJSPath = replaceBackSlashes(self.httpdJSPath)
self.httpdManifest = os.path.join(os.path.dirname(self.xpcshell), 'components', 'httpd.manifest')
self.httpdManifest = replaceBackSlashes(self.httpdManifest)
if self.xrePath is None:
self.xrePath = os.path.dirname(self.xpcshell)
else:
self.xrePath = os.path.abspath(self.xrePath)
if self.mozInfo is None:
self.mozInfo = os.path.join(self.testharnessdir, "mozinfo.json")
def buildCoreEnvironment(self):
"""
Add environment variables likely to be used across all platforms, including remote systems.
"""
# Make assertions fatal
self.env["XPCOM_DEBUG_BREAK"] = "stack-and-abort"
# Crash reporting interferes with debugging
if not self.debuggerInfo:
self.env["MOZ_CRASHREPORTER"] = "1"
# Don't launch the crash reporter client
self.env["MOZ_CRASHREPORTER_NO_REPORT"] = "1"
# Capturing backtraces is very slow on some platforms, and it's
# disabled by automation.py too
self.env["NS_TRACE_MALLOC_DISABLE_STACKS"] = "1"
def buildEnvironment(self):
"""
Create and returns a dictionary of self.env to include all the appropriate env variables and values.
On a remote system, we overload this to set different values and are missing things like os.environ and PATH.
"""
self.env = dict(os.environ)
self.buildCoreEnvironment()
if sys.platform == 'win32':
self.env["PATH"] = self.env["PATH"] + ";" + self.xrePath
elif sys.platform in ('os2emx', 'os2knix'):
os.environ["BEGINLIBPATH"] = self.xrePath + ";" + self.env["BEGINLIBPATH"]
os.environ["LIBPATHSTRICT"] = "T"
elif sys.platform == 'osx' or sys.platform == "darwin":
self.env["DYLD_LIBRARY_PATH"] = self.xrePath
else: # unix or linux?
if not "LD_LIBRARY_PATH" in self.env or self.env["LD_LIBRARY_PATH"] is None:
self.env["LD_LIBRARY_PATH"] = self.xrePath
else:
self.env["LD_LIBRARY_PATH"] = ":".join([self.xrePath, self.env["LD_LIBRARY_PATH"]])
if "asan" in self.mozInfo and self.mozInfo["asan"]:
# ASan symbolizer support
llvmsym = os.path.join(self.xrePath, "llvm-symbolizer")
if os.path.isfile(llvmsym):
self.env["ASAN_SYMBOLIZER_PATH"] = llvmsym
self.log.info("INFO | runxpcshelltests.py | ASan using symbolizer at %s", llvmsym)
else:
self.log.info("INFO | runxpcshelltests.py | ASan symbolizer binary not found: %s", llvmsym)
return self.env
def getPipes(self):
"""
Determine the value of the stdout and stderr for the test.
Return value is a list (pStdout, pStderr).
"""
if self.interactive:
pStdout = None
pStderr = None
else:
if (self.debuggerInfo and self.debuggerInfo["interactive"]):
pStdout = None
pStderr = None
else:
if sys.platform == 'os2emx':
pStdout = None
else:
pStdout = PIPE
pStderr = STDOUT
return pStdout, pStderr
def buildTestPath(self):
"""
If we specifiy a testpath, set the self.testPath variable to be the given directory or file.
|testPath| will be the optional path only, or |None|.
|singleFile| will be the optional test only, or |None|.
"""
self.singleFile = None
if self.testPath is not None:
if self.testPath.endswith('.js'):
# Split into path and file.
if self.testPath.find('/') == -1:
# Test only.
self.singleFile = self.testPath
else:
# Both path and test.
# Reuse |testPath| temporarily.
self.testPath = self.testPath.rsplit('/', 1)
self.singleFile = self.testPath[1]
self.testPath = self.testPath[0]
else:
# Path only.
# Simply remove optional ending separator.
self.testPath = self.testPath.rstrip("/")
def verifyDirPath(self, dirname):
"""
Simple wrapper to get the absolute path for a given directory name.
On a remote system, we need to overload this to work on the remote filesystem.
"""
return os.path.abspath(dirname)
def trySetupNode(self):
"""
Run node for SPDY tests, if available, and updates mozinfo as appropriate.
"""
nodeMozInfo = {'hasNode': False} # Assume the worst
nodeBin = None
# We try to find the node executable in the path given to us by the user in
# the MOZ_NODE_PATH environment variable
localPath = os.getenv('MOZ_NODE_PATH', None)
if localPath and os.path.exists(localPath) and os.path.isfile(localPath):
nodeBin = localPath
if nodeBin:
self.log.info('Found node at %s' % (nodeBin,))
myDir = os.path.split(os.path.abspath(__file__))[0]
mozSpdyJs = os.path.join(myDir, 'moz-spdy', 'moz-spdy.js')
if os.path.exists(mozSpdyJs):
# OK, we found our SPDY server, let's try to get it running
self.log.info('Found moz-spdy at %s' % (mozSpdyJs,))
stdout, stderr = self.getPipes()
try:
# We pipe stdin to node because the spdy server will exit when its
# stdin reaches EOF
self.nodeProc = Popen([nodeBin, mozSpdyJs], stdin=PIPE, stdout=PIPE,
stderr=STDOUT, env=self.env, cwd=os.getcwd())
# Check to make sure the server starts properly by waiting for it to
# tell us it's started
msg = self.nodeProc.stdout.readline()
if msg.startswith('SPDY server listening'):
nodeMozInfo['hasNode'] = True
except OSError, e:
# This occurs if the subprocess couldn't be started
self.log.error('Could not run node SPDY server: %s' % (str(e),))
mozinfo.update(nodeMozInfo)
def shutdownNode(self):
"""
Shut down our node process, if it exists
"""
if self.nodeProc:
self.log.info('Node SPDY server shutting down ...')
# moz-spdy exits when its stdin reaches EOF, so force that to happen here
self.nodeProc.communicate()
def writeXunitResults(self, results, name=None, filename=None, fh=None):
"""
Write Xunit XML from results.
The function receives an iterable of results dicts. Each dict must have
the following keys:
classname - The "class" name of the test.
name - The simple name of the test.
In addition, it must have one of the following saying how the test
executed:
passed - Boolean indicating whether the test passed. False if it
failed.
skipped - True if the test was skipped.
The following keys are optional:
time - Execution time of the test in decimal seconds.
failure - Dict describing test failure. Requires keys:
type - String type of failure.
message - String describing basic failure.
text - Verbose string describing failure.
Arguments:
|name|, Name of the test suite. Many tools expect Java class dot notation
e.g. dom.simple.foo. A directory with '/' converted to '.' is a good
choice.
|fh|, File handle to write XML to.
|filename|, File name to write XML to.
|results|, Iterable of tuples describing the results.
"""
if filename is None and fh is None:
raise Exception("One of filename or fh must be defined.")
if name is None:
name = "xpcshell"
else:
assert isinstance(name, basestring)
if filename is not None:
fh = open(filename, 'wb')
doc = xml.dom.minidom.Document()
testsuite = doc.createElement("testsuite")
testsuite.setAttribute("name", name)
doc.appendChild(testsuite)
total = 0
passed = 0
failed = 0
skipped = 0
for result in results:
total += 1
if result.get("skipped", None):
skipped += 1
elif result["passed"]:
passed += 1
else:
failed += 1
testcase = doc.createElement("testcase")
testcase.setAttribute("classname", result["classname"])
testcase.setAttribute("name", result["name"])
if "time" in result:
testcase.setAttribute("time", str(result["time"]))
else:
# It appears most tools expect the time attribute to be present.
testcase.setAttribute("time", "0")
if "failure" in result:
failure = doc.createElement("failure")
failure.setAttribute("type", str(result["failure"]["type"]))
failure.setAttribute("message", result["failure"]["message"])
# Lossy translation but required to not break CDATA. Also, text could
# be None and Python 2.5's minidom doesn't accept None. Later versions
# do, however.
cdata = result["failure"]["text"]
if not isinstance(cdata, str):
cdata = ""
cdata = cdata.replace("]]>", "]] >")
text = doc.createCDATASection(cdata)
failure.appendChild(text)
testcase.appendChild(failure)
if result.get("skipped", None):
e = doc.createElement("skipped")
testcase.appendChild(e)
testsuite.appendChild(testcase)
testsuite.setAttribute("tests", str(total))
testsuite.setAttribute("failures", str(failed))
testsuite.setAttribute("skip", str(skipped))
doc.writexml(fh, addindent=" ", newl="\n", encoding="utf-8")
def post_to_autolog(self, results, name):
from moztest.results import TestContext, TestResult, TestResultCollection
from moztest.output.autolog import AutologOutput
context = TestContext(
testgroup='b2g xpcshell testsuite',
operating_system='android',
arch='emulator',
harness='xpcshell',
hostname=socket.gethostname(),
tree='b2g',
buildtype='opt',
)
collection = TestResultCollection('b2g emulator testsuite')
for result in results:
duration = result.get('time', 0)
if 'skipped' in result:
outcome = 'SKIPPED'
elif 'todo' in result:
outcome = 'KNOWN-FAIL'
elif result['passed']:
outcome = 'PASS'
else:
outcome = 'UNEXPECTED-FAIL'
output = None
if 'failure' in result:
output = result['failure']['text']
t = TestResult(name=result['name'], test_class=name,
time_start=0, context=context)
t.finish(result=outcome, time_end=duration, output=output)
collection.append(t)
collection.time_taken += duration
out = AutologOutput()
out.post(out.make_testgroups(collection))
def buildXpcsRunArgs(self):
"""
Add arguments to run the test or make it interactive.
"""
if self.interactive:
self.xpcsRunArgs = [
'-e', 'print("To start the test, type |_execute_test();|.");',
'-i']
else:
self.xpcsRunArgs = ['-e', '_execute_test(); quit(0);']
def addTestResults(self, test):
self.passCount += test.passCount
self.failCount += test.failCount
self.todoCount += test.todoCount
self.xunitResults.append(test.xunit_result)
def runTests(self, xpcshell, xrePath=None, appPath=None, symbolsPath=None,
manifest=None, testdirs=None, testPath=None, mobileArgs=None,
interactive=False, verbose=False, keepGoing=False, logfiles=True,
thisChunk=1, totalChunks=1, debugger=None,
debuggerArgs=None, debuggerInteractive=False,
profileName=None, mozInfo=None, sequential=False, shuffle=False,
testsRootDir=None, xunitFilename=None, xunitName=None,
testingModulesDir=None, autolog=False, pluginsPath=None,
testClass=XPCShellTestThread, failureManifest=None,
on_message=None, **otherOptions):
"""Run xpcshell tests.
|xpcshell|, is the xpcshell executable to use to run the tests.
|xrePath|, if provided, is the path to the XRE to use.
|appPath|, if provided, is the path to an application directory.
|symbolsPath|, if provided is the path to a directory containing
breakpad symbols for processing crashes in tests.
|manifest|, if provided, is a file containing a list of
test directories to run.
|testdirs|, if provided, is a list of absolute paths of test directories.
No-manifest only option.
|testPath|, if provided, indicates a single path and/or test to run.
|pluginsPath|, if provided, custom plugins directory to be returned from
the xpcshell dir svc provider for NS_APP_PLUGINS_DIR_LIST.
|interactive|, if set to True, indicates to provide an xpcshell prompt
instead of automatically executing the test.
|verbose|, if set to True, will cause stdout/stderr from tests to
be printed always
|logfiles|, if set to False, indicates not to save output to log files.
Non-interactive only option.
|debuggerInfo|, if set, specifies the debugger and debugger arguments
that will be used to launch xpcshell.
|profileName|, if set, specifies the name of the application for the profile
directory if running only a subset of tests.
|mozInfo|, if set, specifies specifies build configuration information, either as a filename containing JSON, or a dict.
|shuffle|, if True, execute tests in random order.
|testsRootDir|, absolute path to root directory of all tests. This is used
by xUnit generation to determine the package name of the tests.
|xunitFilename|, if set, specifies the filename to which to write xUnit XML
results.
|xunitName|, if outputting an xUnit XML file, the str value to use for the
testsuite name.
|testingModulesDir|, if provided, specifies where JS modules reside.
xpcshell will register a resource handler mapping this path.
|otherOptions| may be present for the convenience of subclasses
"""
global gotSIGINT
if testdirs is None:
testdirs = []
if xunitFilename is not None or xunitName is not None:
if not isinstance(testsRootDir, basestring):
raise Exception("testsRootDir must be a str when outputting xUnit.")
if not os.path.isabs(testsRootDir):
testsRootDir = os.path.abspath(testsRootDir)
if not os.path.exists(testsRootDir):
raise Exception("testsRootDir path does not exists: %s" %
testsRootDir)
# Try to guess modules directory.
# This somewhat grotesque hack allows the buildbot machines to find the
# modules directory without having to configure the buildbot hosts. This
# code path should never be executed in local runs because the build system
# should always set this argument.
if not testingModulesDir:
ourDir = os.path.dirname(__file__)
possible = os.path.join(ourDir, os.path.pardir, 'modules')
if os.path.isdir(possible):
testingModulesDir = possible
if testingModulesDir:
# The resource loader expects native paths. Depending on how we were
# invoked, a UNIX style path may sneak in on Windows. We try to
# normalize that.
testingModulesDir = os.path.normpath(testingModulesDir)
if not os.path.isabs(testingModulesDir):
testingModulesDir = os.path.abspath(testingModulesDir)
if not testingModulesDir.endswith(os.path.sep):
testingModulesDir += os.path.sep
self.xpcshell = xpcshell
self.xrePath = xrePath
self.appPath = appPath
self.symbolsPath = symbolsPath
self.manifest = manifest
self.testdirs = testdirs
self.testPath = testPath
self.interactive = interactive
self.verbose = verbose
self.keepGoing = keepGoing
self.logfiles = logfiles
self.on_message = on_message
self.totalChunks = totalChunks
self.thisChunk = thisChunk
self.debuggerInfo = getDebuggerInfo(self.oldcwd, debugger, debuggerArgs, debuggerInteractive)
self.profileName = profileName or "xpcshell"
self.mozInfo = mozInfo
self.testingModulesDir = testingModulesDir
self.pluginsPath = pluginsPath
self.sequential = sequential
if not testdirs and not manifest:
# nothing to test!
self.log.error("Error: No test dirs or test manifest specified!")
return False
self.testCount = 0
self.passCount = 0
self.failCount = 0
self.todoCount = 0
self.setAbsPath()
self.buildXpcsRunArgs()
self.event = Event()
# Handle filenames in mozInfo
if not isinstance(self.mozInfo, dict):
mozInfoFile = self.mozInfo
if not os.path.isfile(mozInfoFile):
self.log.error("Error: couldn't find mozinfo.json at '%s'. Perhaps you need to use --build-info-json?" % mozInfoFile)
return False
self.mozInfo = parse_json(open(mozInfoFile).read())
mozinfo.update(self.mozInfo)
# buildEnvironment() needs mozInfo, so we call it after mozInfo is initialized.
self.buildEnvironment()
# The appDirKey is a optional entry in either the default or individual test
# sections that defines a relative application directory for test runs. If
# defined we pass 'grePath/$appDirKey' for the -a parameter of the xpcshell
# test harness.
appDirKey = None
if "appname" in self.mozInfo:
appDirKey = self.mozInfo["appname"] + "-appdir"
# We have to do this before we build the test list so we know whether or
# not to run tests that depend on having the node spdy server
self.trySetupNode()
pStdout, pStderr = self.getPipes()
self.buildTestList()
if self.singleFile:
self.sequential = True
if shuffle:
random.shuffle(self.alltests)
self.xunitResults = []
self.cleanup_dir_list = []
self.try_again_list = []
kwargs = {
'appPath': self.appPath,
'xrePath': self.xrePath,
'testingModulesDir': self.testingModulesDir,
'debuggerInfo': self.debuggerInfo,
'pluginsPath': self.pluginsPath,
'httpdManifest': self.httpdManifest,
'httpdJSPath': self.httpdJSPath,
'headJSPath': self.headJSPath,
'testharnessdir': self.testharnessdir,
'profileName': self.profileName,
'singleFile': self.singleFile,
'env': self.env, # making a copy of this in the testthreads
'symbolsPath': self.symbolsPath,
'logfiles': self.logfiles,
'xpcshell': self.xpcshell,
'xpcsRunArgs': self.xpcsRunArgs,
'failureManifest': failureManifest,
'on_message': self.on_message,
}
if self.sequential:
# Allow user to kill hung xpcshell subprocess with SIGINT
# when we are only running tests sequentially.
signal.signal(signal.SIGINT, markGotSIGINT)
if self.debuggerInfo:
# Force a sequential run
self.sequential = True
# If we have an interactive debugger, disable SIGINT entirely.
if self.debuggerInfo["interactive"]:
signal.signal(signal.SIGINT, lambda signum, frame: None)
# create a queue of all tests that will run
tests_queue = deque()
# also a list for the tests that need to be run sequentially
sequential_tests = []
for test_object in self.alltests:
name = test_object['path']
if self.singleFile and not name.endswith(self.singleFile):
continue
if self.testPath and name.find(self.testPath) == -1:
continue
self.testCount += 1
test = testClass(test_object, self.event, self.cleanup_dir_list,
tests_root_dir=testsRootDir, app_dir_key=appDirKey,
interactive=interactive, verbose=verbose, pStdout=pStdout,
pStderr=pStderr, keep_going=keepGoing, log=self.log,
mobileArgs=mobileArgs, **kwargs)
if 'run-sequentially' in test_object or self.sequential:
sequential_tests.append(test)
else:
tests_queue.append(test)
if self.sequential:
self.log.info("INFO | Running tests sequentially.")
else:
self.log.info("INFO | Using at most %d threads." % NUM_THREADS)
# keep a set of NUM_THREADS running tests and start running the
# tests in the queue at most NUM_THREADS at a time
running_tests = set()
keep_going = True
exceptions = []
tracebacks = []
while tests_queue or running_tests:
# if we're not supposed to continue and all of the running tests
# are done, stop
if not keep_going and not running_tests:
break
# if there's room to run more tests, start running them
while keep_going and tests_queue and (len(running_tests) < NUM_THREADS):
test = tests_queue.popleft()
running_tests.add(test)
test.start()
# queue is full (for now) or no more new tests,
# process the finished tests so far
# wait for at least one of the tests to finish
self.event.wait(1)
self.event.clear()
# find what tests are done (might be more than 1)
done_tests = set()
for test in running_tests:
if test.done:
done_tests.add(test)
test.join(1) # join with timeout so we don't hang on blocked threads
# if the test had trouble, we will try running it again
# at the end of the run
if test.retry or test.is_alive():
# if the join call timed out, test.is_alive => True
self.try_again_list.append(test.test_object)
continue
# did the test encounter any exception?
if test.exception:
exceptions.append(test.exception)
tracebacks.append(test.traceback)
# we won't add any more tests, will just wait for
# the currently running ones to finish
keep_going = False
keep_going = keep_going and test.keep_going
self.addTestResults(test)
# make room for new tests to run
running_tests.difference_update(done_tests)
if keep_going:
# run the other tests sequentially
for test in sequential_tests:
if not keep_going:
self.log.error("TEST-UNEXPECTED-FAIL | Received SIGINT (control-C), so stopped run. " \
"(Use --keep-going to keep running tests after killing one with SIGINT)")
break
# we don't want to retry these tests
test.retry = False
test.start()
test.join()
self.addTestResults(test)
# did the test encounter any exception?
if test.exception:
exceptions.append(test.exception)
tracebacks.append(test.traceback)
break
keep_going = test.keep_going
# retry tests that failed when run in parallel
if self.try_again_list:
self.log.info("Retrying tests that failed when run in parallel.")
for test_object in self.try_again_list:
test = testClass(test_object, self.event, self.cleanup_dir_list,
retry=False, tests_root_dir=testsRootDir,
app_dir_key=appDirKey, interactive=interactive,
verbose=verbose, pStdout=pStdout, pStderr=pStderr,
keep_going=keepGoing, log=self.log, mobileArgs=mobileArgs,
**kwargs)
test.start()
test.join()
self.addTestResults(test)
# did the test encounter any exception?
if test.exception:
exceptions.append(test.exception)
tracebacks.append(test.traceback)
break
keep_going = test.keep_going
# restore default SIGINT behaviour
signal.signal(signal.SIGINT, signal.SIG_DFL)
self.shutdownNode()
# Clean up any slacker directories that might be lying around
# Some might fail because of windows taking too long to unlock them.
# We don't do anything if this fails because the test slaves will have
# their $TEMP dirs cleaned up on reboot anyway.
for directory in self.cleanup_dir_list:
try:
shutil.rmtree(directory)
except:
self.log.info("INFO | %s could not be cleaned up." % directory)
if exceptions:
self.log.info("INFO | Following exceptions were raised:")
for t in tracebacks:
self.log.error(t)
raise exceptions[0]
if self.testCount == 0:
self.log.error("TEST-UNEXPECTED-FAIL | runxpcshelltests.py | No tests run. Did you pass an invalid --test-path?")
self.failCount = 1
self.log.info("INFO | Result summary:")
self.log.info("INFO | Passed: %d" % self.passCount)
self.log.info("INFO | Failed: %d" % self.failCount)
self.log.info("INFO | Todo: %d" % self.todoCount)
self.log.info("INFO | Retried: %d" % len(self.try_again_list))
if autolog:
self.post_to_autolog(self.xunitResults, xunitName)
if xunitFilename is not None:
self.writeXunitResults(filename=xunitFilename, results=self.xunitResults,
name=xunitName)
if gotSIGINT and not keepGoing:
self.log.error("TEST-UNEXPECTED-FAIL | Received SIGINT (control-C), so stopped run. " \
"(Use --keep-going to keep running tests after killing one with SIGINT)")
return False
return self.failCount == 0
class XPCShellOptions(OptionParser):
def __init__(self):
"""Process command line arguments and call runTests() to do the real work."""
OptionParser.__init__(self)
addCommonOptions(self)
self.add_option("--app-path",
type="string", dest="appPath", default=None,
help="application directory (as opposed to XRE directory)")
self.add_option("--autolog",
action="store_true", dest="autolog", default=False,
help="post to autolog")
self.add_option("--interactive",
action="store_true", dest="interactive", default=False,
help="don't automatically run tests, drop to an xpcshell prompt")
self.add_option("--verbose",
action="store_true", dest="verbose", default=False,
help="always print stdout and stderr from tests")
self.add_option("--keep-going",
action="store_true", dest="keepGoing", default=False,
help="continue running tests after test killed with control-C (SIGINT)")
self.add_option("--logfiles",
action="store_true", dest="logfiles", default=True,
help="create log files (default, only used to override --no-logfiles)")
self.add_option("--manifest",
type="string", dest="manifest", default=None,
help="Manifest of test directories to use")
self.add_option("--no-logfiles",
action="store_false", dest="logfiles",
help="don't create log files")
self.add_option("--sequential",
action="store_true", dest="sequential", default=False,
help="Run all tests sequentially")
self.add_option("--test-path",
type="string", dest="testPath", default=None,
help="single path and/or test filename to test")
self.add_option("--tests-root-dir",
type="string", dest="testsRootDir", default=None,
help="absolute path to directory where all tests are located. this is typically $(objdir)/_tests")
self.add_option("--testing-modules-dir",
dest="testingModulesDir", default=None,
help="Directory where testing modules are located.")
self.add_option("--test-plugin-path",
type="string", dest="pluginsPath", default=None,
help="Path to the location of a plugins directory containing the test plugin or plugins required for tests. "
"By default xpcshell's dir svc provider returns gre/plugins. Use test-plugin-path to add a directory "
"to return for NS_APP_PLUGINS_DIR_LIST when queried.")
self.add_option("--total-chunks",
type = "int", dest = "totalChunks", default=1,
help = "how many chunks to split the tests up into")
self.add_option("--this-chunk",
type = "int", dest = "thisChunk", default=1,
help = "which chunk to run between 1 and --total-chunks")
self.add_option("--profile-name",
type = "string", dest="profileName", default=None,
help="name of application profile being tested")
self.add_option("--build-info-json",
type = "string", dest="mozInfo", default=None,
help="path to a mozinfo.json including information about the build configuration. defaults to looking for mozinfo.json next to the script.")
self.add_option("--shuffle",
action="store_true", dest="shuffle", default=False,
help="Execute tests in random order")
self.add_option("--xunit-file", dest="xunitFilename",
help="path to file where xUnit results will be written.")
self.add_option("--xunit-suite-name", dest="xunitName",
help="name to record for this xUnit test suite. Many "
"tools expect Java class notation, e.g. "
"dom.basic.foo")
self.add_option("--failure-manifest", dest="failureManifest",
action="store",
help="path to file where failure manifest will be written.")
def main():
parser = XPCShellOptions()
options, args = parser.parse_args()
if len(args) < 2 and options.manifest is None or \
(len(args) < 1 and options.manifest is not None):
print >>sys.stderr, """Usage: %s <path to xpcshell> <test dirs>
or: %s --manifest=test.manifest <path to xpcshell>""" % (sys.argv[0],
sys.argv[0])
sys.exit(1)
xpcsh = XPCShellTests()
if options.interactive and not options.testPath:
print >>sys.stderr, "Error: You must specify a test filename in interactive mode!"
sys.exit(1)
if not xpcsh.runTests(args[0], testdirs=args[1:], **options.__dict__):
sys.exit(1)
if __name__ == '__main__':
main()