2009-09-06 05:08:49 +04:00
|
|
|
#
|
2012-05-21 15:12:37 +04:00
|
|
|
# This Source Code Form is subject to the terms of the Mozilla Public
|
|
|
|
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
|
|
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
2009-09-06 05:08:49 +04:00
|
|
|
|
2011-11-15 08:33:21 +04:00
|
|
|
from __future__ import with_statement
|
|
|
|
import glob, logging, os, platform, shutil, subprocess, sys, tempfile, urllib2, zipfile
|
2009-09-06 05:08:49 +04:00
|
|
|
import re
|
2010-05-06 16:06:09 +04:00
|
|
|
from urlparse import urlparse
|
2012-02-25 15:48:05 +04:00
|
|
|
from operator import itemgetter
|
2009-07-22 23:24:54 +04:00
|
|
|
|
|
|
|
__all__ = [
|
2011-11-15 08:33:21 +04:00
|
|
|
"ZipFileReader",
|
2009-09-06 05:08:49 +04:00
|
|
|
"addCommonOptions",
|
|
|
|
"checkForCrashes",
|
|
|
|
"dumpLeakLog",
|
2010-05-06 16:06:09 +04:00
|
|
|
"isURL",
|
2009-09-06 05:08:49 +04:00
|
|
|
"processLeakLog",
|
2009-10-20 03:12:09 +04:00
|
|
|
"getDebuggerInfo",
|
|
|
|
"DEBUGGER_INFO",
|
2010-03-13 01:57:29 +03:00
|
|
|
"replaceBackSlashes",
|
2010-09-23 20:19:31 +04:00
|
|
|
"wrapCommand",
|
2012-02-25 15:48:05 +04:00
|
|
|
"ShutdownLeakLogger"
|
2009-09-06 05:08:49 +04:00
|
|
|
]
|
2009-07-22 23:24:54 +04:00
|
|
|
|
2009-10-20 03:12:09 +04:00
|
|
|
# Map of debugging programs to information about them, like default arguments
|
|
|
|
# and whether or not they are interactive.
|
|
|
|
DEBUGGER_INFO = {
|
|
|
|
# gdb requires that you supply the '--args' flag in order to pass arguments
|
|
|
|
# after the executable name to the executable.
|
|
|
|
"gdb": {
|
|
|
|
"interactive": True,
|
|
|
|
"args": "-q --args"
|
|
|
|
},
|
|
|
|
|
|
|
|
# valgrind doesn't explain much about leaks unless you set the
|
|
|
|
# '--leak-check=full' flag.
|
|
|
|
"valgrind": {
|
|
|
|
"interactive": False,
|
|
|
|
"args": "--leak-check=full"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-11-15 08:33:21 +04:00
|
|
|
class ZipFileReader(object):
|
|
|
|
"""
|
|
|
|
Class to read zip files in Python 2.5 and later. Limited to only what we
|
|
|
|
actually use.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, filename):
|
|
|
|
self._zipfile = zipfile.ZipFile(filename, "r")
|
|
|
|
|
|
|
|
def __del__(self):
|
|
|
|
self._zipfile.close()
|
|
|
|
|
|
|
|
def _getnormalizedpath(self, path):
|
|
|
|
"""
|
|
|
|
Gets a normalized path from 'path' (or the current working directory if
|
|
|
|
'path' is None). Also asserts that the path exists.
|
|
|
|
"""
|
|
|
|
if path is None:
|
|
|
|
path = os.curdir
|
|
|
|
path = os.path.normpath(os.path.expanduser(path))
|
|
|
|
assert os.path.isdir(path)
|
|
|
|
return path
|
|
|
|
|
|
|
|
def _extractname(self, name, path):
|
|
|
|
"""
|
|
|
|
Extracts a file with the given name from the zip file to the given path.
|
|
|
|
Also creates any directories needed along the way.
|
|
|
|
"""
|
|
|
|
filename = os.path.normpath(os.path.join(path, name))
|
|
|
|
if name.endswith("/"):
|
|
|
|
os.makedirs(filename)
|
|
|
|
else:
|
|
|
|
path = os.path.split(filename)[0]
|
|
|
|
if not os.path.isdir(path):
|
|
|
|
os.makedirs(path)
|
|
|
|
with open(filename, "wb") as dest:
|
|
|
|
dest.write(self._zipfile.read(name))
|
|
|
|
|
|
|
|
def namelist(self):
|
|
|
|
return self._zipfile.namelist()
|
|
|
|
|
|
|
|
def read(self, name):
|
|
|
|
return self._zipfile.read(name)
|
|
|
|
|
|
|
|
def extract(self, name, path = None):
|
|
|
|
if hasattr(self._zipfile, "extract"):
|
|
|
|
return self._zipfile.extract(name, path)
|
|
|
|
|
|
|
|
# This will throw if name is not part of the zip file.
|
|
|
|
self._zipfile.getinfo(name)
|
|
|
|
|
|
|
|
self._extractname(name, self._getnormalizedpath(path))
|
|
|
|
|
|
|
|
def extractall(self, path = None):
|
|
|
|
if hasattr(self._zipfile, "extractall"):
|
|
|
|
return self._zipfile.extractall(path)
|
|
|
|
|
|
|
|
path = self._getnormalizedpath(path)
|
|
|
|
|
|
|
|
for name in self._zipfile.namelist():
|
|
|
|
self._extractname(name, path)
|
|
|
|
|
2009-07-22 23:24:54 +04:00
|
|
|
log = logging.getLogger()
|
|
|
|
|
2010-05-06 16:06:09 +04:00
|
|
|
def isURL(thing):
|
|
|
|
"""Return True if |thing| looks like a URL."""
|
2011-11-30 02:13:08 +04:00
|
|
|
# We want to download URLs like http://... but not Windows paths like c:\...
|
|
|
|
return len(urlparse(thing).scheme) >= 2
|
2010-05-06 16:06:09 +04:00
|
|
|
|
2009-07-22 23:24:54 +04:00
|
|
|
def addCommonOptions(parser, defaults={}):
|
2009-09-06 05:08:49 +04:00
|
|
|
parser.add_option("--xre-path",
|
|
|
|
action = "store", type = "string", dest = "xrePath",
|
|
|
|
# individual scripts will set a sane default
|
|
|
|
default = None,
|
|
|
|
help = "absolute path to directory containing XRE (probably xulrunner)")
|
|
|
|
if 'SYMBOLS_PATH' not in defaults:
|
|
|
|
defaults['SYMBOLS_PATH'] = None
|
|
|
|
parser.add_option("--symbols-path",
|
|
|
|
action = "store", type = "string", dest = "symbolsPath",
|
|
|
|
default = defaults['SYMBOLS_PATH'],
|
2010-05-06 16:06:09 +04:00
|
|
|
help = "absolute path to directory containing breakpad symbols, or the URL of a zip file containing symbols")
|
2009-10-20 03:12:09 +04:00
|
|
|
parser.add_option("--debugger",
|
|
|
|
action = "store", dest = "debugger",
|
|
|
|
help = "use the given debugger to launch the application")
|
|
|
|
parser.add_option("--debugger-args",
|
|
|
|
action = "store", dest = "debuggerArgs",
|
|
|
|
help = "pass the given args to the debugger _before_ "
|
|
|
|
"the application on the command line")
|
|
|
|
parser.add_option("--debugger-interactive",
|
|
|
|
action = "store_true", dest = "debuggerInteractive",
|
|
|
|
help = "prevents the test harness from redirecting "
|
|
|
|
"stdout and stderr for interactive debuggers")
|
2009-07-22 23:24:54 +04:00
|
|
|
|
|
|
|
def checkForCrashes(dumpDir, symbolsPath, testName=None):
|
|
|
|
stackwalkPath = os.environ.get('MINIDUMP_STACKWALK', None)
|
|
|
|
# try to get the caller's filename if no test name is given
|
|
|
|
if testName is None:
|
2009-09-06 05:08:49 +04:00
|
|
|
try:
|
|
|
|
testName = os.path.basename(sys._getframe(1).f_code.co_filename)
|
|
|
|
except:
|
|
|
|
testName = "unknown"
|
2009-07-22 23:24:54 +04:00
|
|
|
|
2011-11-15 08:33:21 +04:00
|
|
|
# Check preconditions
|
2009-07-22 23:24:54 +04:00
|
|
|
dumps = glob.glob(os.path.join(dumpDir, '*.dmp'))
|
2011-11-15 08:33:21 +04:00
|
|
|
if len(dumps) == 0:
|
|
|
|
return False
|
|
|
|
|
|
|
|
foundCrash = False
|
|
|
|
removeSymbolsPath = False
|
|
|
|
|
|
|
|
# If our symbols are at a remote URL, download them now
|
|
|
|
if isURL(symbolsPath):
|
|
|
|
print "Downloading symbols from: " + symbolsPath
|
|
|
|
removeSymbolsPath = True
|
|
|
|
# Get the symbols and write them to a temporary zipfile
|
|
|
|
data = urllib2.urlopen(symbolsPath)
|
|
|
|
symbolsFile = tempfile.TemporaryFile()
|
|
|
|
symbolsFile.write(data.read())
|
|
|
|
# extract symbols to a temporary directory (which we'll delete after
|
|
|
|
# processing all crashes)
|
|
|
|
symbolsPath = tempfile.mkdtemp()
|
|
|
|
zfile = ZipFileReader(symbolsFile)
|
|
|
|
zfile.extractall(symbolsPath)
|
|
|
|
|
|
|
|
try:
|
|
|
|
for d in dumps:
|
|
|
|
log.info("PROCESS-CRASH | %s | application crashed (minidump found)", testName)
|
|
|
|
print "Crash dump filename: " + d
|
|
|
|
if symbolsPath and stackwalkPath and os.path.exists(stackwalkPath):
|
|
|
|
# run minidump stackwalk
|
|
|
|
p = subprocess.Popen([stackwalkPath, d, symbolsPath],
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE)
|
|
|
|
(out, err) = p.communicate()
|
|
|
|
if len(out) > 3:
|
|
|
|
# minidump_stackwalk is chatty, so ignore stderr when it succeeds.
|
|
|
|
print out
|
2010-11-06 23:30:32 +03:00
|
|
|
else:
|
2011-11-15 08:33:21 +04:00
|
|
|
print "stderr from minidump_stackwalk:"
|
|
|
|
print err
|
|
|
|
if p.returncode != 0:
|
|
|
|
print "minidump_stackwalk exited with return code %d" % p.returncode
|
2009-10-20 21:27:02 +04:00
|
|
|
else:
|
2011-11-15 08:33:21 +04:00
|
|
|
if not symbolsPath:
|
|
|
|
print "No symbols path given, can't process dump."
|
|
|
|
if not stackwalkPath:
|
|
|
|
print "MINIDUMP_STACKWALK not set, can't process dump."
|
|
|
|
elif stackwalkPath and not os.path.exists(stackwalkPath):
|
2009-10-20 21:27:02 +04:00
|
|
|
print "MINIDUMP_STACKWALK binary not found: %s" % stackwalkPath
|
2011-11-15 08:33:21 +04:00
|
|
|
dumpSavePath = os.environ.get('MINIDUMP_SAVE_PATH', None)
|
|
|
|
if dumpSavePath:
|
|
|
|
shutil.move(d, dumpSavePath)
|
|
|
|
print "Saved dump as %s" % os.path.join(dumpSavePath,
|
|
|
|
os.path.basename(d))
|
|
|
|
else:
|
|
|
|
os.remove(d)
|
|
|
|
extra = os.path.splitext(d)[0] + ".extra"
|
|
|
|
if os.path.exists(extra):
|
|
|
|
os.remove(extra)
|
|
|
|
foundCrash = True
|
|
|
|
finally:
|
|
|
|
if removeSymbolsPath:
|
|
|
|
shutil.rmtree(symbolsPath)
|
2009-09-06 05:08:49 +04:00
|
|
|
|
2009-07-22 23:24:54 +04:00
|
|
|
return foundCrash
|
2011-11-15 08:33:21 +04:00
|
|
|
|
2009-10-20 03:12:09 +04:00
|
|
|
def getFullPath(directory, path):
|
|
|
|
"Get an absolute path relative to 'directory'."
|
|
|
|
return os.path.normpath(os.path.join(directory, os.path.expanduser(path)))
|
|
|
|
|
|
|
|
def searchPath(directory, path):
|
|
|
|
"Go one step beyond getFullPath and try the various folders in PATH"
|
|
|
|
# Try looking in the current working directory first.
|
|
|
|
newpath = getFullPath(directory, path)
|
2010-01-21 23:02:51 +03:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-20 03:12:09 +04:00
|
|
|
return newpath
|
|
|
|
|
|
|
|
# At this point we have to fail if a directory was given (to prevent cases
|
|
|
|
# like './gdb' from matching '/usr/bin/./gdb').
|
|
|
|
if not os.path.dirname(path):
|
|
|
|
for dir in os.environ['PATH'].split(os.pathsep):
|
|
|
|
newpath = os.path.join(dir, path)
|
2010-01-21 23:02:51 +03:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-20 03:12:09 +04:00
|
|
|
return newpath
|
|
|
|
return None
|
|
|
|
|
|
|
|
def getDebuggerInfo(directory, debugger, debuggerArgs, debuggerInteractive = False):
|
|
|
|
|
|
|
|
debuggerInfo = None
|
|
|
|
|
|
|
|
if debugger:
|
|
|
|
debuggerPath = searchPath(directory, debugger)
|
|
|
|
if not debuggerPath:
|
|
|
|
print "Error: Path %s doesn't exist." % debugger
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
debuggerName = os.path.basename(debuggerPath).lower()
|
|
|
|
|
|
|
|
def getDebuggerInfo(type, default):
|
|
|
|
if debuggerName in DEBUGGER_INFO and type in DEBUGGER_INFO[debuggerName]:
|
|
|
|
return DEBUGGER_INFO[debuggerName][type]
|
|
|
|
return default
|
|
|
|
|
|
|
|
debuggerInfo = {
|
|
|
|
"path": debuggerPath,
|
|
|
|
"interactive" : getDebuggerInfo("interactive", False),
|
|
|
|
"args": getDebuggerInfo("args", "").split()
|
|
|
|
}
|
|
|
|
|
|
|
|
if debuggerArgs:
|
|
|
|
debuggerInfo["args"] = debuggerArgs.split()
|
|
|
|
if debuggerInteractive:
|
|
|
|
debuggerInfo["interactive"] = debuggerInteractive
|
2011-11-30 02:13:08 +04:00
|
|
|
|
2009-10-20 03:12:09 +04:00
|
|
|
return debuggerInfo
|
|
|
|
|
2009-09-06 05:08:49 +04:00
|
|
|
|
|
|
|
def dumpLeakLog(leakLogFile, filter = False):
|
|
|
|
"""Process the leak log, without parsing it.
|
|
|
|
|
|
|
|
Use this function if you want the raw log only.
|
|
|
|
Use it preferably with the |XPCOM_MEM_LEAK_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Don't warn (nor "info") if the log file is not there.
|
|
|
|
if not os.path.exists(leakLogFile):
|
|
|
|
return
|
|
|
|
|
|
|
|
leaks = open(leakLogFile, "r")
|
|
|
|
leakReport = leaks.read()
|
|
|
|
leaks.close()
|
|
|
|
|
|
|
|
# Only |XPCOM_MEM_LEAK_LOG| reports can be actually filtered out.
|
|
|
|
# Only check whether an actual leak was reported.
|
|
|
|
if filter and not "0 TOTAL " in leakReport:
|
|
|
|
return
|
|
|
|
|
|
|
|
# Simply copy the log.
|
|
|
|
log.info(leakReport.rstrip("\n"))
|
|
|
|
|
2009-11-10 22:16:12 +03:00
|
|
|
def processSingleLeakFile(leakLogFileName, PID, processType, leakThreshold):
|
|
|
|
"""Process a single leak log, corresponding to the specified
|
|
|
|
process PID and type.
|
2009-09-06 05:08:49 +04:00
|
|
|
"""
|
|
|
|
|
|
|
|
# Per-Inst Leaked Total Rem ...
|
|
|
|
# 0 TOTAL 17 192 419115886 2 ...
|
|
|
|
# 833 nsTimerImpl 60 120 24726 2 ...
|
|
|
|
lineRe = re.compile(r"^\s*\d+\s+(?P<name>\S+)\s+"
|
|
|
|
r"(?P<size>-?\d+)\s+(?P<bytesLeaked>-?\d+)\s+"
|
|
|
|
r"-?\d+\s+(?P<numLeaked>-?\d+)")
|
|
|
|
|
2009-11-10 22:16:12 +03:00
|
|
|
processString = ""
|
|
|
|
if PID and processType:
|
|
|
|
processString = "| %s process %s " % (processType, PID)
|
|
|
|
leaks = open(leakLogFileName, "r")
|
2009-09-06 05:08:49 +04:00
|
|
|
for line in leaks:
|
|
|
|
matches = lineRe.match(line)
|
|
|
|
if (matches and
|
|
|
|
int(matches.group("numLeaked")) == 0 and
|
|
|
|
matches.group("name") != "TOTAL"):
|
|
|
|
continue
|
|
|
|
log.info(line.rstrip())
|
|
|
|
leaks.close()
|
|
|
|
|
2009-11-10 22:16:12 +03:00
|
|
|
leaks = open(leakLogFileName, "r")
|
2009-09-06 05:08:49 +04:00
|
|
|
seenTotal = False
|
2009-12-09 21:51:38 +03:00
|
|
|
crashedOnPurpose = False
|
2009-09-06 05:08:49 +04:00
|
|
|
prefix = "TEST-PASS"
|
2010-06-10 16:27:47 +04:00
|
|
|
numObjects = 0
|
2009-09-06 05:08:49 +04:00
|
|
|
for line in leaks:
|
2009-12-09 21:51:38 +03:00
|
|
|
if line.find("purposefully crash") > -1:
|
|
|
|
crashedOnPurpose = True
|
2009-09-06 05:08:49 +04:00
|
|
|
matches = lineRe.match(line)
|
|
|
|
if not matches:
|
|
|
|
continue
|
|
|
|
name = matches.group("name")
|
|
|
|
size = int(matches.group("size"))
|
|
|
|
bytesLeaked = int(matches.group("bytesLeaked"))
|
|
|
|
numLeaked = int(matches.group("numLeaked"))
|
|
|
|
if size < 0 or bytesLeaked < 0 or numLeaked < 0:
|
2009-11-10 22:16:12 +03:00
|
|
|
log.info("TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | negative leaks caught!" %
|
|
|
|
processString)
|
2009-09-06 05:08:49 +04:00
|
|
|
if name == "TOTAL":
|
|
|
|
seenTotal = True
|
|
|
|
elif name == "TOTAL":
|
|
|
|
seenTotal = True
|
|
|
|
# Check for leaks.
|
|
|
|
if bytesLeaked < 0 or bytesLeaked > leakThreshold:
|
|
|
|
prefix = "TEST-UNEXPECTED-FAIL"
|
2009-11-10 22:16:12 +03:00
|
|
|
leakLog = "TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | leaked" \
|
|
|
|
" %d bytes during test execution" % (processString, bytesLeaked)
|
2009-09-06 05:08:49 +04:00
|
|
|
elif bytesLeaked > 0:
|
2009-11-10 22:16:12 +03:00
|
|
|
leakLog = "TEST-PASS %s| automationutils.processLeakLog() | WARNING leaked" \
|
|
|
|
" %d bytes during test execution" % (processString, bytesLeaked)
|
2009-09-06 05:08:49 +04:00
|
|
|
else:
|
2009-11-10 22:16:12 +03:00
|
|
|
leakLog = "TEST-PASS %s| automationutils.processLeakLog() | no leaks detected!" \
|
|
|
|
% processString
|
2009-09-06 05:08:49 +04:00
|
|
|
# Remind the threshold if it is not 0, which is the default/goal.
|
|
|
|
if leakThreshold != 0:
|
|
|
|
leakLog += " (threshold set at %d bytes)" % leakThreshold
|
|
|
|
# Log the information.
|
|
|
|
log.info(leakLog)
|
|
|
|
else:
|
|
|
|
if numLeaked != 0:
|
|
|
|
if numLeaked > 1:
|
|
|
|
instance = "instances"
|
|
|
|
rest = " each (%s bytes total)" % matches.group("bytesLeaked")
|
|
|
|
else:
|
|
|
|
instance = "instance"
|
|
|
|
rest = ""
|
2010-06-10 16:27:47 +04:00
|
|
|
numObjects += 1
|
|
|
|
if numObjects > 5:
|
|
|
|
# don't spam brief tinderbox logs with tons of leak output
|
|
|
|
prefix = "TEST-INFO"
|
2009-11-10 22:16:12 +03:00
|
|
|
log.info("%(prefix)s %(process)s| automationutils.processLeakLog() | leaked %(numLeaked)d %(instance)s of %(name)s "
|
2009-09-06 05:08:49 +04:00
|
|
|
"with size %(size)s bytes%(rest)s" %
|
|
|
|
{ "prefix": prefix,
|
2009-11-10 22:16:12 +03:00
|
|
|
"process": processString,
|
2009-09-06 05:08:49 +04:00
|
|
|
"numLeaked": numLeaked,
|
|
|
|
"instance": instance,
|
|
|
|
"name": name,
|
|
|
|
"size": matches.group("size"),
|
|
|
|
"rest": rest })
|
|
|
|
if not seenTotal:
|
2009-12-09 21:51:38 +03:00
|
|
|
if crashedOnPurpose:
|
|
|
|
log.info("INFO | automationutils.processLeakLog() | process %s was " \
|
|
|
|
"deliberately crashed and thus has no leak log" % PID)
|
|
|
|
else:
|
|
|
|
log.info("TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | missing output line for total leaks!" %
|
2009-11-10 22:16:12 +03:00
|
|
|
processString)
|
2009-09-06 05:08:49 +04:00
|
|
|
leaks.close()
|
2009-11-10 22:16:12 +03:00
|
|
|
|
|
|
|
|
|
|
|
def processLeakLog(leakLogFile, leakThreshold = 0):
|
|
|
|
"""Process the leak log, including separate leak logs created
|
|
|
|
by child processes.
|
|
|
|
|
|
|
|
Use this function if you want an additional PASS/FAIL summary.
|
|
|
|
It must be used with the |XPCOM_MEM_BLOAT_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
if not os.path.exists(leakLogFile):
|
|
|
|
log.info("WARNING | automationutils.processLeakLog() | refcount logging is off, so leaks can't be detected!")
|
|
|
|
return
|
|
|
|
|
|
|
|
(leakLogFileDir, leakFileBase) = os.path.split(leakLogFile)
|
|
|
|
pidRegExp = re.compile(r".*?_([a-z]*)_pid(\d*)$")
|
|
|
|
if leakFileBase[-4:] == ".log":
|
|
|
|
leakFileBase = leakFileBase[:-4]
|
|
|
|
pidRegExp = re.compile(r".*?_([a-z]*)_pid(\d*).log$")
|
|
|
|
|
|
|
|
for fileName in os.listdir(leakLogFileDir):
|
|
|
|
if fileName.find(leakFileBase) != -1:
|
|
|
|
thisFile = os.path.join(leakLogFileDir, fileName)
|
|
|
|
processPID = 0
|
|
|
|
processType = None
|
|
|
|
m = pidRegExp.search(fileName)
|
|
|
|
if m:
|
|
|
|
processType = m.group(1)
|
|
|
|
processPID = m.group(2)
|
|
|
|
processSingleLeakFile(thisFile, processPID, processType, leakThreshold)
|
2010-03-13 01:57:29 +03:00
|
|
|
|
|
|
|
def replaceBackSlashes(input):
|
|
|
|
return input.replace('\\', '/')
|
2010-09-23 20:19:31 +04:00
|
|
|
|
|
|
|
def wrapCommand(cmd):
|
|
|
|
"""
|
|
|
|
If running on OS X 10.5 or older, wrap |cmd| so that it will
|
|
|
|
be executed as an i386 binary, in case it's a 32-bit/64-bit universal
|
|
|
|
binary.
|
|
|
|
"""
|
|
|
|
if platform.system() == "Darwin" and \
|
|
|
|
hasattr(platform, 'mac_ver') and \
|
|
|
|
platform.mac_ver()[0][:4] < '10.6':
|
|
|
|
return ["arch", "-arch", "i386"] + cmd
|
|
|
|
# otherwise just execute the command normally
|
|
|
|
return cmd
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
class ShutdownLeakLogger(object):
|
|
|
|
"""
|
|
|
|
Parses the mochitest run log when running a debug build, assigns all leaked
|
|
|
|
DOM windows (that are still around after test suite shutdown, despite running
|
|
|
|
the GC) to the tests that created them and prints leak statistics.
|
|
|
|
"""
|
2012-07-21 21:37:56 +04:00
|
|
|
MAX_LEAK_COUNT = 3
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
def __init__(self, logger):
|
|
|
|
self.logger = logger
|
|
|
|
self.tests = []
|
|
|
|
self.leakedWindows = {}
|
|
|
|
self.leakedDocShells = set()
|
|
|
|
self.currentTest = None
|
|
|
|
self.seenShutdown = False
|
|
|
|
|
|
|
|
def log(self, line):
|
|
|
|
if line[2:11] == "DOMWINDOW":
|
|
|
|
self._logWindow(line)
|
|
|
|
elif line[2:10] == "DOCSHELL":
|
|
|
|
self._logDocShell(line)
|
|
|
|
elif line.startswith("TEST-START"):
|
|
|
|
fileName = line.split(" ")[-1].strip().replace("chrome://mochitests/content/browser/", "")
|
|
|
|
self.currentTest = {"fileName": fileName, "windows": set(), "docShells": set()}
|
|
|
|
elif line.startswith("INFO TEST-END"):
|
|
|
|
# don't track a test if no windows or docShells leaked
|
2012-05-14 19:02:26 +04:00
|
|
|
if self.currentTest and (self.currentTest["windows"] or self.currentTest["docShells"]):
|
2012-02-25 15:48:05 +04:00
|
|
|
self.tests.append(self.currentTest)
|
|
|
|
self.currentTest = None
|
|
|
|
elif line.startswith("INFO TEST-START | Shutdown"):
|
|
|
|
self.seenShutdown = True
|
|
|
|
|
|
|
|
def parse(self):
|
|
|
|
leakingTests = self._parseLeakingTests()
|
|
|
|
|
|
|
|
if leakingTests:
|
|
|
|
totalWindows = sum(len(test["leakedWindows"]) for test in leakingTests)
|
|
|
|
totalDocShells = sum(len(test["leakedDocShells"]) for test in leakingTests)
|
2012-07-10 16:19:48 +04:00
|
|
|
msgType = "TEST-INFO" if totalWindows + totalDocShells <= self.MAX_LEAK_COUNT else "TEST-UNEXPECTED-FAIL"
|
|
|
|
self.logger.info("%s | ShutdownLeaks | leaked %d DOMWindow(s) and %d DocShell(s) until shutdown", msgType, totalWindows, totalDocShells)
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
for test in leakingTests:
|
|
|
|
for url, count in self._zipLeakedWindows(test["leakedWindows"]):
|
2012-07-10 16:19:48 +04:00
|
|
|
self.logger.info("%s | %s | leaked %d window(s) until shutdown [url = %s]", msgType, test["fileName"], count, url)
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
if test["leakedDocShells"]:
|
2012-07-10 16:19:48 +04:00
|
|
|
self.logger.info("%s | %s | leaked %d docShell(s) until shutdown", msgType, test["fileName"], len(test["leakedDocShells"]))
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
def _logWindow(self, line):
|
|
|
|
created = line[:2] == "++"
|
|
|
|
id = self._parseValue(line, "serial")
|
|
|
|
|
2012-05-04 18:21:24 +04:00
|
|
|
# log line has invalid format
|
|
|
|
if not id:
|
|
|
|
return
|
|
|
|
|
2012-02-25 15:48:05 +04:00
|
|
|
if self.currentTest:
|
|
|
|
windows = self.currentTest["windows"]
|
|
|
|
if created:
|
|
|
|
windows.add(id)
|
|
|
|
else:
|
|
|
|
windows.discard(id)
|
|
|
|
elif self.seenShutdown and not created:
|
|
|
|
self.leakedWindows[id] = self._parseValue(line, "url")
|
|
|
|
|
|
|
|
def _logDocShell(self, line):
|
|
|
|
created = line[:2] == "++"
|
|
|
|
id = self._parseValue(line, "id")
|
|
|
|
|
2012-05-04 18:21:24 +04:00
|
|
|
# log line has invalid format
|
|
|
|
if not id:
|
|
|
|
return
|
|
|
|
|
2012-02-25 15:48:05 +04:00
|
|
|
if self.currentTest:
|
|
|
|
docShells = self.currentTest["docShells"]
|
|
|
|
if created:
|
|
|
|
docShells.add(id)
|
|
|
|
else:
|
|
|
|
docShells.discard(id)
|
|
|
|
elif self.seenShutdown and not created:
|
|
|
|
self.leakedDocShells.add(id)
|
|
|
|
|
|
|
|
def _parseValue(self, line, name):
|
2012-05-04 18:21:24 +04:00
|
|
|
match = re.search("\[%s = (.+?)\]" % name, line)
|
|
|
|
if match:
|
|
|
|
return match.group(1)
|
|
|
|
return None
|
2012-02-25 15:48:05 +04:00
|
|
|
|
|
|
|
def _parseLeakingTests(self):
|
|
|
|
leakingTests = []
|
|
|
|
|
|
|
|
for test in self.tests:
|
|
|
|
test["leakedWindows"] = [self.leakedWindows[id] for id in test["windows"] if id in self.leakedWindows]
|
|
|
|
test["leakedDocShells"] = [id for id in test["docShells"] if id in self.leakedDocShells]
|
|
|
|
test["leakCount"] = len(test["leakedWindows"]) + len(test["leakedDocShells"])
|
|
|
|
|
|
|
|
if test["leakCount"]:
|
|
|
|
leakingTests.append(test)
|
|
|
|
|
|
|
|
return sorted(leakingTests, key=itemgetter("leakCount"), reverse=True)
|
|
|
|
|
|
|
|
def _zipLeakedWindows(self, leakedWindows):
|
|
|
|
counts = []
|
|
|
|
counted = set()
|
|
|
|
|
|
|
|
for url in leakedWindows:
|
|
|
|
if not url in counted:
|
|
|
|
counts.append((url, leakedWindows.count(url)))
|
|
|
|
counted.add(url)
|
|
|
|
|
|
|
|
return sorted(counts, key=itemgetter(1), reverse=True)
|