forked from OSchip/llvm-project
1231 lines
47 KiB
Python
Executable File
1231 lines
47 KiB
Python
Executable File
#!/usr/bin/env python
|
|
|
|
"""
|
|
A simple testing framework for lldb using python's unit testing framework.
|
|
|
|
Tests for lldb are written as python scripts which take advantage of the script
|
|
bridging provided by LLDB.framework to interact with lldb core.
|
|
|
|
A specific naming pattern is followed by the .py script to be recognized as
|
|
a module which implements a test scenario, namely, Test*.py.
|
|
|
|
To specify the directories where "Test*.py" python test scripts are located,
|
|
you need to pass in a list of directory names. By default, the current
|
|
working directory is searched if nothing is specified on the command line.
|
|
|
|
Type:
|
|
|
|
./dotest.py -h
|
|
|
|
for available options.
|
|
"""
|
|
|
|
import os, signal, sys, time
|
|
import subprocess
|
|
import unittest2
|
|
|
|
def is_exe(fpath):
|
|
"""Returns true if fpath is an executable."""
|
|
return os.path.isfile(fpath) and os.access(fpath, os.X_OK)
|
|
|
|
def which(program):
|
|
"""Returns the full path to a program; None otherwise."""
|
|
fpath, fname = os.path.split(program)
|
|
if fpath:
|
|
if is_exe(program):
|
|
return program
|
|
else:
|
|
for path in os.environ["PATH"].split(os.pathsep):
|
|
exe_file = os.path.join(path, program)
|
|
if is_exe(exe_file):
|
|
return exe_file
|
|
return None
|
|
|
|
class _WritelnDecorator(object):
|
|
"""Used to decorate file-like objects with a handy 'writeln' method"""
|
|
def __init__(self,stream):
|
|
self.stream = stream
|
|
|
|
def __getattr__(self, attr):
|
|
if attr in ('stream', '__getstate__'):
|
|
raise AttributeError(attr)
|
|
return getattr(self.stream,attr)
|
|
|
|
def writeln(self, arg=None):
|
|
if arg:
|
|
self.write(arg)
|
|
self.write('\n') # text-mode streams translate to \r\n if needed
|
|
|
|
#
|
|
# Global variables:
|
|
#
|
|
|
|
# The test suite.
|
|
suite = unittest2.TestSuite()
|
|
|
|
# By default, both command line and Python API tests are performed.
|
|
# Use @python_api_test decorator, defined in lldbtest.py, to mark a test as
|
|
# a Python API test.
|
|
dont_do_python_api_test = False
|
|
|
|
# By default, both command line and Python API tests are performed.
|
|
just_do_python_api_test = False
|
|
|
|
# By default, benchmarks tests are not run.
|
|
just_do_benchmarks_test = False
|
|
|
|
# The blacklist is optional (-b blacklistFile) and allows a central place to skip
|
|
# testclass's and/or testclass.testmethod's.
|
|
blacklist = None
|
|
|
|
# The dictionary as a result of sourcing blacklistFile.
|
|
blacklistConfig = {}
|
|
|
|
# The config file is optional.
|
|
configFile = None
|
|
|
|
# Test suite repeat count. Can be overwritten with '-# count'.
|
|
count = 1
|
|
|
|
# The dictionary as a result of sourcing configFile.
|
|
config = {}
|
|
|
|
# The 'archs' and 'compilers' can be specified via either command line or configFile,
|
|
# with the command line overriding the configFile. When specified, they should be
|
|
# of the list type. For example, "-A x86_64^i386" => archs=['x86_64', 'i386'] and
|
|
# "-C gcc^clang" => compilers=['gcc', 'clang'].
|
|
archs = None
|
|
compilers = None
|
|
|
|
# Delay startup in order for the debugger to attach.
|
|
delay = False
|
|
|
|
# Dump the Python sys.path variable. Use '-D' to dump sys.path.
|
|
dumpSysPath = False
|
|
|
|
# Full path of the benchmark executable, as specified by the '-e' option.
|
|
bmExecutable = None
|
|
# The breakpoint specification of bmExecutable, as specified by the '-x' option.
|
|
bmBreakpointSpec = None
|
|
# The benchamrk iteration count, as specified by the '-y' option.
|
|
bmIterationCount = -1
|
|
|
|
# By default, don't exclude any directories. Use '-X' to add one excluded directory.
|
|
excluded = set(['.svn', '.git'])
|
|
|
|
# By default, failfast is False. Use '-F' to overwrite it.
|
|
failfast = False
|
|
|
|
# The filters (testclass.testmethod) used to admit tests into our test suite.
|
|
filters = []
|
|
|
|
# The runhooks is a list of lldb commands specifically for the debugger.
|
|
# Use '-k' to specify a runhook.
|
|
runHooks = []
|
|
|
|
# If '-g' is specified, the filterspec is not exclusive. If a test module does
|
|
# not contain testclass.testmethod which matches the filterspec, the whole test
|
|
# module is still admitted into our test suite. fs4all flag defaults to True.
|
|
fs4all = True
|
|
|
|
# Ignore the build search path relative to this script to locate the lldb.py module.
|
|
ignore = False
|
|
|
|
# By default, we do not skip build and cleanup. Use '-S' option to override.
|
|
skip_build_and_cleanup = False
|
|
|
|
# By default, we skip long running test case. Use '-l' option to override.
|
|
skip_long_running_test = True
|
|
|
|
# By default, we print the build dir, lldb version, and svn info. Use '-n' option to
|
|
# turn it off.
|
|
noHeaders = False
|
|
|
|
# The regular expression pattern to match against eligible filenames as our test cases.
|
|
regexp = None
|
|
|
|
# By default, tests are executed in place and cleanups are performed afterwards.
|
|
# Use '-r dir' option to relocate the tests and their intermediate files to a
|
|
# different directory and to forgo any cleanups. The directory specified must
|
|
# not exist yet.
|
|
rdir = None
|
|
|
|
# By default, recorded session info for errored/failed test are dumped into its
|
|
# own file under a session directory named after the timestamp of the test suite
|
|
# run. Use '-s session-dir-name' to specify a specific dir name.
|
|
sdir_name = None
|
|
|
|
# Set this flag if there is any session info dumped during the test run.
|
|
sdir_has_content = False
|
|
|
|
# svn_info stores the output from 'svn info lldb.base.dir'.
|
|
svn_info = ''
|
|
|
|
# Default verbosity is 0.
|
|
verbose = 0
|
|
|
|
# Set to True only if verbose is 0 and LLDB trace mode is off.
|
|
progress_bar = False
|
|
|
|
# By default, search from the script directory.
|
|
testdirs = [ sys.path[0] ]
|
|
|
|
# Separator string.
|
|
separator = '-' * 70
|
|
|
|
|
|
def usage():
|
|
print """
|
|
Usage: dotest.py [option] [args]
|
|
where options:
|
|
-h : print this help message and exit. Add '-v' for more detailed help.
|
|
-A : specify the architecture(s) to launch for the inferior process
|
|
-A i386 => launch inferior with i386 architecture
|
|
-A x86_64^i386 => launch inferior with x86_64 and i386 architectures
|
|
-C : specify the compiler(s) used to build the inferior executable
|
|
-C clang => build debuggee using clang compiler
|
|
-C /my/full/path/to/clang => specify a full path to the clang binary
|
|
-C clang^gcc => build debuggee using clang and gcc compilers
|
|
-D : dump the Python sys.path variable
|
|
-a : don't do lldb Python API tests
|
|
use @python_api_test to decorate a test case as lldb Python API test
|
|
+a : just do lldb Python API tests
|
|
do not specify both '-a' and '+a' at the same time
|
|
+b : just do benchmark tests
|
|
use @benchmark_test to decorate a test case as such
|
|
-b : read a blacklist file specified after this option
|
|
-c : read a config file specified after this option
|
|
the architectures and compilers (note the plurals) specified via '-A' and '-C'
|
|
will override those specified via a config file
|
|
(see also lldb-trunk/example/test/usage-config)
|
|
-d : delay startup for 10 seconds (in order for the debugger to attach)
|
|
-e : specify the full path of an executable used for benchmark purpose;
|
|
see also '-x', which provides the breakpoint sepcification
|
|
-F : failfast, stop the test suite on the first error/failure
|
|
-f : specify a filter, which consists of the test class name, a dot, followed by
|
|
the test method, to only admit such test into the test suite
|
|
e.g., -f 'ClassTypesTestCase.test_with_dwarf_and_python_api'
|
|
-g : if specified, the filterspec by -f is not exclusive, i.e., if a test module
|
|
does not match the filterspec (testclass.testmethod), the whole module is
|
|
still admitted to the test suite
|
|
-i : ignore (don't bailout) if 'lldb.py' module cannot be located in the build
|
|
tree relative to this script; use PYTHONPATH to locate the module
|
|
-k : specify a runhook, which is an lldb command to be executed by the debugger;
|
|
'-k' option can occur multiple times, the commands are executed one after the
|
|
other to bring the debugger to a desired state, so that, for example, further
|
|
benchmarking can be done
|
|
-l : don't skip long running test
|
|
-n : don't print the headers like build dir, lldb version, and svn info at all
|
|
-p : specify a regexp filename pattern for inclusion in the test suite
|
|
-r : specify a dir to relocate the tests and their intermediate files to;
|
|
the directory must not exist before running this test driver;
|
|
no cleanup of intermediate test files is performed in this case
|
|
-S : skip the build and cleanup while running the test
|
|
use this option with care as you would need to build the inferior(s) by hand
|
|
and build the executable(s) with the correct name(s)
|
|
this can be used with '-# n' to stress test certain test cases for n number of
|
|
times
|
|
-s : specify the name of the dir created to store the session files of tests
|
|
with errored or failed status; if not specified, the test driver uses the
|
|
timestamp as the session dir name
|
|
-t : turn on tracing of lldb command and other detailed test executions
|
|
-v : do verbose mode of unittest framework (print out each test case invocation)
|
|
-X : exclude a directory from consideration for test discovery
|
|
-X types => if 'types' appear in the pathname components of a potential testfile
|
|
it will be ignored
|
|
-x : specify the breakpoint specification for the benchmark executable;
|
|
see also '-e', which provides the full path of the executable
|
|
-y : specify the iteration count used to collect our benchmarks; an example is
|
|
the number of times to do 'thread step-over' to measure stepping speed
|
|
see also '-e' and '-x' options
|
|
-w : insert some wait time (currently 0.5 sec) between consecutive test cases
|
|
-# : Repeat the test suite for a specified number of times
|
|
|
|
and:
|
|
args : specify a list of directory names to search for test modules named after
|
|
Test*.py (test discovery)
|
|
if empty, search from the current working directory, instead
|
|
"""
|
|
|
|
if verbose > 0:
|
|
print """
|
|
Examples:
|
|
|
|
This is an example of using the -f option to pinpoint to a specfic test class
|
|
and test method to be run:
|
|
|
|
$ ./dotest.py -f ClassTypesTestCase.test_with_dsym_and_run_command
|
|
----------------------------------------------------------------------
|
|
Collected 1 test
|
|
|
|
test_with_dsym_and_run_command (TestClassTypes.ClassTypesTestCase)
|
|
Test 'frame variable this' when stopped on a class constructor. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 1 test in 1.396s
|
|
|
|
OK
|
|
|
|
And this is an example of using the -p option to run a single file (the filename
|
|
matches the pattern 'ObjC' and it happens to be 'TestObjCMethods.py'):
|
|
|
|
$ ./dotest.py -v -p ObjC
|
|
----------------------------------------------------------------------
|
|
Collected 4 tests
|
|
|
|
test_break_with_dsym (TestObjCMethods.FoundationTestCase)
|
|
Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok
|
|
test_break_with_dwarf (TestObjCMethods.FoundationTestCase)
|
|
Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok
|
|
test_data_type_and_expr_with_dsym (TestObjCMethods.FoundationTestCase)
|
|
Lookup objective-c data types and evaluate expressions. ... ok
|
|
test_data_type_and_expr_with_dwarf (TestObjCMethods.FoundationTestCase)
|
|
Lookup objective-c data types and evaluate expressions. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 4 tests in 16.661s
|
|
|
|
OK
|
|
|
|
Running of this script also sets up the LLDB_TEST environment variable so that
|
|
individual test cases can locate their supporting files correctly. The script
|
|
tries to set up Python's search paths for modules by looking at the build tree
|
|
relative to this script. See also the '-i' option in the following example.
|
|
|
|
Finally, this is an example of using the lldb.py module distributed/installed by
|
|
Xcode4 to run against the tests under the 'forward' directory, and with the '-w'
|
|
option to add some delay between two tests. It uses ARCH=x86_64 to specify that
|
|
as the architecture and CC=clang to specify the compiler used for the test run:
|
|
|
|
$ PYTHONPATH=/Xcode4/Library/PrivateFrameworks/LLDB.framework/Versions/A/Resources/Python ARCH=x86_64 CC=clang ./dotest.py -v -w -i forward
|
|
|
|
Session logs for test failures/errors will go into directory '2010-11-11-13_56_16'
|
|
----------------------------------------------------------------------
|
|
Collected 2 tests
|
|
|
|
test_with_dsym_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase)
|
|
Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok
|
|
test_with_dwarf_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase)
|
|
Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 2 tests in 5.659s
|
|
|
|
OK
|
|
|
|
The 'Session ...' verbiage is recently introduced (see also the '-s' option) to
|
|
notify the directory containing the session logs for test failures or errors.
|
|
In case there is any test failure/error, a similar message is appended at the
|
|
end of the stderr output for your convenience.
|
|
|
|
Environment variables related to loggings:
|
|
|
|
o LLDB_LOG: if defined, specifies the log file pathname for the 'lldb' subsystem
|
|
with a default option of 'event process' if LLDB_LOG_OPTION is not defined.
|
|
|
|
o GDB_REMOTE_LOG: if defined, specifies the log file pathname for the
|
|
'process.gdb-remote' subsystem with a default option of 'packets' if
|
|
GDB_REMOTE_LOG_OPTION is not defined.
|
|
"""
|
|
sys.exit(0)
|
|
|
|
|
|
def parseOptionsAndInitTestdirs():
|
|
"""Initialize the list of directories containing our unittest scripts.
|
|
|
|
'-h/--help as the first option prints out usage info and exit the program.
|
|
"""
|
|
|
|
global dont_do_python_api_test
|
|
global just_do_python_api_test
|
|
global just_do_benchmarks_test
|
|
global blacklist
|
|
global blacklistConfig
|
|
global configFile
|
|
global archs
|
|
global compilers
|
|
global count
|
|
global delay
|
|
global dumpSysPath
|
|
global bmExecutable
|
|
global bmBreakpointSpec
|
|
global bmIterationCount
|
|
global failfast
|
|
global filters
|
|
global fs4all
|
|
global ignore
|
|
global progress_bar
|
|
global runHooks
|
|
global skip_build_and_cleanup
|
|
global skip_long_running_test
|
|
global noHeaders
|
|
global regexp
|
|
global rdir
|
|
global sdir_name
|
|
global verbose
|
|
global testdirs
|
|
|
|
do_help = False
|
|
|
|
if len(sys.argv) == 1:
|
|
return
|
|
|
|
# Process possible trace and/or verbose flag, among other things.
|
|
index = 1
|
|
while index < len(sys.argv):
|
|
if sys.argv[index].startswith('-') or sys.argv[index].startswith('+'):
|
|
# We should continue processing...
|
|
pass
|
|
else:
|
|
# End of option processing.
|
|
break
|
|
|
|
if sys.argv[index].find('-h') != -1:
|
|
index += 1
|
|
do_help = True
|
|
elif sys.argv[index].startswith('-A'):
|
|
# Increment by 1 to fetch the ARCH spec.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
archs = sys.argv[index].split('^')
|
|
index += 1
|
|
elif sys.argv[index].startswith('-C'):
|
|
# Increment by 1 to fetch the CC spec.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
compilers = sys.argv[index].split('^')
|
|
index += 1
|
|
elif sys.argv[index].startswith('-D'):
|
|
dumpSysPath = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-a'):
|
|
dont_do_python_api_test = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('+a'):
|
|
just_do_python_api_test = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('+b'):
|
|
just_do_benchmarks_test = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-b'):
|
|
# Increment by 1 to fetch the blacklist file name option argument.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
blacklistFile = sys.argv[index]
|
|
if not os.path.isfile(blacklistFile):
|
|
print "Blacklist file:", blacklistFile, "does not exist!"
|
|
usage()
|
|
index += 1
|
|
# Now read the blacklist contents and assign it to blacklist.
|
|
execfile(blacklistFile, globals(), blacklistConfig)
|
|
blacklist = blacklistConfig.get('blacklist')
|
|
elif sys.argv[index].startswith('-c'):
|
|
# Increment by 1 to fetch the config file name option argument.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
configFile = sys.argv[index]
|
|
if not os.path.isfile(configFile):
|
|
print "Config file:", configFile, "does not exist!"
|
|
usage()
|
|
index += 1
|
|
elif sys.argv[index].startswith('-d'):
|
|
delay = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-e'):
|
|
# Increment by 1 to fetch the full path of the benchmark executable.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
bmExecutable = sys.argv[index]
|
|
if not is_exe(bmExecutable):
|
|
usage()
|
|
index += 1
|
|
elif sys.argv[index].startswith('-F'):
|
|
failfast = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-f'):
|
|
# Increment by 1 to fetch the filter spec.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
filters.append(sys.argv[index])
|
|
index += 1
|
|
elif sys.argv[index].startswith('-g'):
|
|
fs4all = False
|
|
index += 1
|
|
elif sys.argv[index].startswith('-i'):
|
|
ignore = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-k'):
|
|
# Increment by 1 to fetch the runhook lldb command.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
runHooks.append(sys.argv[index])
|
|
index += 1
|
|
elif sys.argv[index].startswith('-l'):
|
|
skip_long_running_test = False
|
|
index += 1
|
|
elif sys.argv[index].startswith('-n'):
|
|
noHeaders = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-p'):
|
|
# Increment by 1 to fetch the reg exp pattern argument.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
regexp = sys.argv[index]
|
|
index += 1
|
|
elif sys.argv[index].startswith('-r'):
|
|
# Increment by 1 to fetch the relocated directory argument.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
rdir = os.path.abspath(sys.argv[index])
|
|
if os.path.exists(rdir):
|
|
print "Relocated directory:", rdir, "must not exist!"
|
|
usage()
|
|
index += 1
|
|
elif sys.argv[index].startswith('-S'):
|
|
skip_build_and_cleanup = True
|
|
index += 1
|
|
elif sys.argv[index].startswith('-s'):
|
|
# Increment by 1 to fetch the session dir name.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
sdir_name = sys.argv[index]
|
|
index += 1
|
|
elif sys.argv[index].startswith('-t'):
|
|
os.environ["LLDB_COMMAND_TRACE"] = "YES"
|
|
index += 1
|
|
elif sys.argv[index].startswith('-v'):
|
|
verbose = 2
|
|
index += 1
|
|
elif sys.argv[index].startswith('-w'):
|
|
os.environ["LLDB_WAIT_BETWEEN_TEST_CASES"] = 'YES'
|
|
index += 1
|
|
elif sys.argv[index].startswith('-X'):
|
|
# Increment by 1 to fetch an excluded directory.
|
|
index += 1
|
|
if index >= len(sys.argv):
|
|
usage()
|
|
excluded.add(sys.argv[index])
|
|
index += 1
|
|
elif sys.argv[index].startswith('-x'):
|
|
# Increment by 1 to fetch the breakpoint specification of the benchmark executable.
|
|
index += 1
|
|
if index >= len(sys.argv):
|
|
usage()
|
|
bmBreakpointSpec = sys.argv[index]
|
|
index += 1
|
|
elif sys.argv[index].startswith('-y'):
|
|
# Increment by 1 to fetch the the benchmark iteration count.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
bmIterationCount = int(sys.argv[index])
|
|
index += 1
|
|
elif sys.argv[index].startswith('-#'):
|
|
# Increment by 1 to fetch the repeat count argument.
|
|
index += 1
|
|
if index >= len(sys.argv) or sys.argv[index].startswith('-'):
|
|
usage()
|
|
count = int(sys.argv[index])
|
|
index += 1
|
|
else:
|
|
print "Unknown option: ", sys.argv[index]
|
|
usage()
|
|
|
|
if do_help == True:
|
|
usage()
|
|
|
|
# Do not specify both '-a' and '+a' at the same time.
|
|
if dont_do_python_api_test and just_do_python_api_test:
|
|
usage()
|
|
|
|
# The simple progress bar is turned on only if verbose == 0 and LLDB_COMMAND_TRACE is not 'YES'
|
|
if ("LLDB_COMMAND_TRACE" not in os.environ or os.environ["LLDB_COMMAND_TRACE"]!="YES") and verbose==0:
|
|
progress_bar = True
|
|
|
|
# Gather all the dirs passed on the command line.
|
|
if len(sys.argv) > index:
|
|
testdirs = map(os.path.abspath, sys.argv[index:])
|
|
|
|
# If '-r dir' is specified, the tests should be run under the relocated
|
|
# directory. Let's copy the testdirs over.
|
|
if rdir:
|
|
from shutil import copytree, ignore_patterns
|
|
|
|
tmpdirs = []
|
|
for srcdir in testdirs:
|
|
dstdir = os.path.join(rdir, os.path.basename(srcdir))
|
|
# Don't copy the *.pyc and .svn stuffs.
|
|
copytree(srcdir, dstdir, ignore=ignore_patterns('*.pyc', '.svn'))
|
|
tmpdirs.append(dstdir)
|
|
|
|
# This will be our modified testdirs.
|
|
testdirs = tmpdirs
|
|
|
|
# With '-r dir' specified, there's no cleanup of intermediate test files.
|
|
os.environ["LLDB_DO_CLEANUP"] = 'NO'
|
|
|
|
# If testdirs is ['test'], the make directory has already been copied
|
|
# recursively and is contained within the rdir/test dir. For anything
|
|
# else, we would need to copy over the make directory and its contents,
|
|
# so that, os.listdir(rdir) looks like, for example:
|
|
#
|
|
# array_types conditional_break make
|
|
#
|
|
# where the make directory contains the Makefile.rules file.
|
|
if len(testdirs) != 1 or os.path.basename(testdirs[0]) != 'test':
|
|
# Don't copy the .svn stuffs.
|
|
copytree('make', os.path.join(rdir, 'make'),
|
|
ignore=ignore_patterns('.svn'))
|
|
|
|
#print "testdirs:", testdirs
|
|
|
|
# Source the configFile if specified.
|
|
# The side effect, if any, will be felt from this point on. An example
|
|
# config file may be these simple two lines:
|
|
#
|
|
# sys.stderr = open("/tmp/lldbtest-stderr", "w")
|
|
# sys.stdout = open("/tmp/lldbtest-stdout", "w")
|
|
#
|
|
# which will reassign the two file objects to sys.stderr and sys.stdout,
|
|
# respectively.
|
|
#
|
|
# See also lldb-trunk/example/test/usage-config.
|
|
global config
|
|
if configFile:
|
|
# Pass config (a dictionary) as the locals namespace for side-effect.
|
|
execfile(configFile, globals(), config)
|
|
#print "config:", config
|
|
#print "sys.stderr:", sys.stderr
|
|
#print "sys.stdout:", sys.stdout
|
|
|
|
|
|
def setupSysPath():
|
|
"""
|
|
Add LLDB.framework/Resources/Python to the search paths for modules.
|
|
As a side effect, we also discover the 'lldb' executable and export it here.
|
|
"""
|
|
|
|
global rdir
|
|
global testdirs
|
|
global dumpSysPath
|
|
global noHeaders
|
|
global svn_info
|
|
|
|
# Get the directory containing the current script.
|
|
if ("DOTEST_PROFILE" in os.environ or "DOTEST_PDB" in os.environ) and "DOTEST_SCRIPT_DIR" in os.environ:
|
|
scriptPath = os.environ["DOTEST_SCRIPT_DIR"]
|
|
else:
|
|
scriptPath = sys.path[0]
|
|
if not scriptPath.endswith('test'):
|
|
print "This script expects to reside in lldb's test directory."
|
|
sys.exit(-1)
|
|
|
|
if rdir:
|
|
# Set up the LLDB_TEST environment variable appropriately, so that the
|
|
# individual tests can be located relatively.
|
|
#
|
|
# See also lldbtest.TestBase.setUpClass(cls).
|
|
if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test':
|
|
os.environ["LLDB_TEST"] = os.path.join(rdir, 'test')
|
|
else:
|
|
os.environ["LLDB_TEST"] = rdir
|
|
else:
|
|
os.environ["LLDB_TEST"] = scriptPath
|
|
|
|
# Set up the LLDB_SRC environment variable, so that the tests can locate
|
|
# the LLDB source code.
|
|
os.environ["LLDB_SRC"] = os.path.join(sys.path[0], os.pardir)
|
|
|
|
pluginPath = os.path.join(scriptPath, 'plugins')
|
|
pexpectPath = os.path.join(scriptPath, 'pexpect-2.4')
|
|
|
|
# Append script dir, plugin dir, and pexpect dir to the sys.path.
|
|
sys.path.append(scriptPath)
|
|
sys.path.append(pluginPath)
|
|
sys.path.append(pexpectPath)
|
|
|
|
# This is our base name component.
|
|
base = os.path.abspath(os.path.join(scriptPath, os.pardir))
|
|
|
|
# These are for xcode build directories.
|
|
xcode3_build_dir = ['build']
|
|
xcode4_build_dir = ['build', 'lldb', 'Build', 'Products']
|
|
dbg = ['Debug']
|
|
rel = ['Release']
|
|
bai = ['BuildAndIntegration']
|
|
python_resource_dir = ['LLDB.framework', 'Resources', 'Python']
|
|
|
|
# Some of the tests can invoke the 'lldb' command directly.
|
|
# We'll try to locate the appropriate executable right here.
|
|
|
|
# First, you can define an environment variable LLDB_EXEC specifying the
|
|
# full pathname of the lldb executable.
|
|
if "LLDB_EXEC" in os.environ and is_exe(os.environ["LLDB_EXEC"]):
|
|
lldbExec = os.environ["LLDB_EXEC"]
|
|
else:
|
|
lldbExec = None
|
|
|
|
executable = ['lldb']
|
|
dbgExec = os.path.join(base, *(xcode3_build_dir + dbg + executable))
|
|
dbgExec2 = os.path.join(base, *(xcode4_build_dir + dbg + executable))
|
|
relExec = os.path.join(base, *(xcode3_build_dir + rel + executable))
|
|
relExec2 = os.path.join(base, *(xcode4_build_dir + rel + executable))
|
|
baiExec = os.path.join(base, *(xcode3_build_dir + bai + executable))
|
|
baiExec2 = os.path.join(base, *(xcode4_build_dir + bai + executable))
|
|
|
|
# The 'lldb' executable built here in the source tree.
|
|
lldbHere = None
|
|
if is_exe(dbgExec):
|
|
lldbHere = dbgExec
|
|
elif is_exe(dbgExec2):
|
|
lldbHere = dbgExec2
|
|
elif is_exe(relExec):
|
|
lldbHere = relExec
|
|
elif is_exe(relExec2):
|
|
lldbHere = relExec2
|
|
elif is_exe(baiExec):
|
|
lldbHere = baiExec
|
|
elif is_exe(baiExec2):
|
|
lldbHere = baiExec2
|
|
elif lldbExec:
|
|
lldbHere = lldbExec
|
|
|
|
if lldbHere:
|
|
os.environ["LLDB_HERE"] = lldbHere
|
|
os.environ["LLDB_BUILD_DIR"] = os.path.split(lldbHere)[0]
|
|
if not noHeaders:
|
|
print "LLDB build dir:", os.environ["LLDB_BUILD_DIR"]
|
|
os.system('%s -v' % lldbHere)
|
|
|
|
# One last chance to locate the 'lldb' executable.
|
|
if not lldbExec:
|
|
lldbExec = which('lldb')
|
|
if lldbHere and not lldbExec:
|
|
lldbExec = lldbHere
|
|
|
|
|
|
if not lldbExec:
|
|
print "The 'lldb' executable cannot be located. Some of the tests may not be run as a result."
|
|
else:
|
|
os.environ["LLDB_EXEC"] = lldbExec
|
|
#print "The 'lldb' from PATH env variable", lldbExec
|
|
|
|
if os.path.isdir(os.path.join(base, '.svn')):
|
|
pipe = subprocess.Popen(["svn", "info", base], stdout = subprocess.PIPE)
|
|
svn_info = pipe.stdout.read()
|
|
elif os.path.isdir(os.path.join(base, '.git')):
|
|
pipe = subprocess.Popen(["git", "svn", "info", base], stdout = subprocess.PIPE)
|
|
svn_info = pipe.stdout.read()
|
|
if not noHeaders:
|
|
print svn_info
|
|
|
|
global ignore
|
|
|
|
# The '-i' option is used to skip looking for lldb.py in the build tree.
|
|
if ignore:
|
|
return
|
|
|
|
dbgPath = os.path.join(base, *(xcode3_build_dir + dbg + python_resource_dir))
|
|
dbgPath2 = os.path.join(base, *(xcode4_build_dir + dbg + python_resource_dir))
|
|
relPath = os.path.join(base, *(xcode3_build_dir + rel + python_resource_dir))
|
|
relPath2 = os.path.join(base, *(xcode4_build_dir + rel + python_resource_dir))
|
|
baiPath = os.path.join(base, *(xcode3_build_dir + bai + python_resource_dir))
|
|
baiPath2 = os.path.join(base, *(xcode4_build_dir + bai + python_resource_dir))
|
|
|
|
lldbPath = None
|
|
if os.path.isfile(os.path.join(dbgPath, 'lldb.py')):
|
|
lldbPath = dbgPath
|
|
elif os.path.isfile(os.path.join(dbgPath2, 'lldb.py')):
|
|
lldbPath = dbgPath2
|
|
elif os.path.isfile(os.path.join(relPath, 'lldb.py')):
|
|
lldbPath = relPath
|
|
elif os.path.isfile(os.path.join(relPath2, 'lldb.py')):
|
|
lldbPath = relPath2
|
|
elif os.path.isfile(os.path.join(baiPath, 'lldb.py')):
|
|
lldbPath = baiPath
|
|
elif os.path.isfile(os.path.join(baiPath2, 'lldb.py')):
|
|
lldbPath = baiPath2
|
|
|
|
if not lldbPath:
|
|
print 'This script requires lldb.py to be in either ' + dbgPath + ',',
|
|
print relPath + ', or ' + baiPath
|
|
sys.exit(-1)
|
|
|
|
# This is to locate the lldb.py module. Insert it right after sys.path[0].
|
|
sys.path[1:1] = [lldbPath]
|
|
if dumpSysPath:
|
|
print "sys.path:", sys.path
|
|
|
|
|
|
def doDelay(delta):
|
|
"""Delaying startup for delta-seconds to facilitate debugger attachment."""
|
|
def alarm_handler(*args):
|
|
raise Exception("timeout")
|
|
|
|
signal.signal(signal.SIGALRM, alarm_handler)
|
|
signal.alarm(delta)
|
|
sys.stdout.write("pid=%d\n" % os.getpid())
|
|
sys.stdout.write("Enter RET to proceed (or timeout after %d seconds):" %
|
|
delta)
|
|
sys.stdout.flush()
|
|
try:
|
|
text = sys.stdin.readline()
|
|
except:
|
|
text = ""
|
|
signal.alarm(0)
|
|
sys.stdout.write("proceeding...\n")
|
|
pass
|
|
|
|
|
|
def visit(prefix, dir, names):
|
|
"""Visitor function for os.path.walk(path, visit, arg)."""
|
|
|
|
global suite
|
|
global regexp
|
|
global filters
|
|
global fs4all
|
|
global excluded
|
|
|
|
if set(dir.split(os.sep)).intersection(excluded):
|
|
#print "Detected an excluded dir component: %s" % dir
|
|
return
|
|
|
|
for name in names:
|
|
if os.path.isdir(os.path.join(dir, name)):
|
|
continue
|
|
|
|
if '.py' == os.path.splitext(name)[1] and name.startswith(prefix):
|
|
# Try to match the regexp pattern, if specified.
|
|
if regexp:
|
|
import re
|
|
if re.search(regexp, name):
|
|
#print "Filename: '%s' matches pattern: '%s'" % (name, regexp)
|
|
pass
|
|
else:
|
|
#print "Filename: '%s' does not match pattern: '%s'" % (name, regexp)
|
|
continue
|
|
|
|
# We found a match for our test. Add it to the suite.
|
|
|
|
# Update the sys.path first.
|
|
if not sys.path.count(dir):
|
|
sys.path.insert(0, dir)
|
|
base = os.path.splitext(name)[0]
|
|
|
|
# Thoroughly check the filterspec against the base module and admit
|
|
# the (base, filterspec) combination only when it makes sense.
|
|
filterspec = None
|
|
for filterspec in filters:
|
|
# Optimistically set the flag to True.
|
|
filtered = True
|
|
module = __import__(base)
|
|
parts = filterspec.split('.')
|
|
obj = module
|
|
for part in parts:
|
|
try:
|
|
parent, obj = obj, getattr(obj, part)
|
|
except AttributeError:
|
|
# The filterspec has failed.
|
|
filtered = False
|
|
break
|
|
|
|
# If filtered, we have a good filterspec. Add it.
|
|
if filtered:
|
|
#print "adding filter spec %s to module %s" % (filterspec, module)
|
|
suite.addTests(
|
|
unittest2.defaultTestLoader.loadTestsFromName(filterspec, module))
|
|
continue
|
|
|
|
# Forgo this module if the (base, filterspec) combo is invalid
|
|
# and no '-g' option is specified
|
|
if filters and fs4all and not filtered:
|
|
continue
|
|
|
|
# Add either the filtered test case(s) (which is done before) or the entire test class.
|
|
if not filterspec or not filtered:
|
|
# A simple case of just the module name. Also the failover case
|
|
# from the filterspec branch when the (base, filterspec) combo
|
|
# doesn't make sense.
|
|
suite.addTests(unittest2.defaultTestLoader.loadTestsFromName(base))
|
|
|
|
|
|
def lldbLoggings():
|
|
"""Check and do lldb loggings if necessary."""
|
|
|
|
# Turn on logging for debugging purposes if ${LLDB_LOG} environment variable is
|
|
# defined. Use ${LLDB_LOG} to specify the log file.
|
|
ci = lldb.DBG.GetCommandInterpreter()
|
|
res = lldb.SBCommandReturnObject()
|
|
if ("LLDB_LOG" in os.environ):
|
|
if ("LLDB_LOG_OPTION" in os.environ):
|
|
lldb_log_option = os.environ["LLDB_LOG_OPTION"]
|
|
else:
|
|
lldb_log_option = "event process expr state api"
|
|
ci.HandleCommand(
|
|
"log enable -n -f " + os.environ["LLDB_LOG"] + " lldb " + lldb_log_option,
|
|
res)
|
|
if not res.Succeeded():
|
|
raise Exception('log enable failed (check LLDB_LOG env variable.')
|
|
# Ditto for gdb-remote logging if ${GDB_REMOTE_LOG} environment variable is defined.
|
|
# Use ${GDB_REMOTE_LOG} to specify the log file.
|
|
if ("GDB_REMOTE_LOG" in os.environ):
|
|
if ("GDB_REMOTE_LOG_OPTION" in os.environ):
|
|
gdb_remote_log_option = os.environ["GDB_REMOTE_LOG_OPTION"]
|
|
else:
|
|
gdb_remote_log_option = "packets process"
|
|
ci.HandleCommand(
|
|
"log enable -n -f " + os.environ["GDB_REMOTE_LOG"] + " gdb-remote "
|
|
+ gdb_remote_log_option,
|
|
res)
|
|
if not res.Succeeded():
|
|
raise Exception('log enable failed (check GDB_REMOTE_LOG env variable.')
|
|
|
|
def getMyCommandLine():
|
|
ps = subprocess.Popen(['ps', '-o', "command=CMD", str(os.getpid())], stdout=subprocess.PIPE).communicate()[0]
|
|
lines = ps.split('\n')
|
|
cmd_line = lines[1]
|
|
return cmd_line
|
|
|
|
# ======================================== #
|
|
# #
|
|
# Execution of the test driver starts here #
|
|
# #
|
|
# ======================================== #
|
|
|
|
def checkDsymForUUIDIsNotOn():
|
|
cmd = ["defaults", "read", "com.apple.DebugSymbols"]
|
|
pipe = subprocess.Popen(cmd, stdout = subprocess.PIPE, stderr = subprocess.STDOUT)
|
|
cmd_output = pipe.stdout.read()
|
|
if cmd_output and "DBGFileMappedPaths = " in cmd_output:
|
|
print "%s =>" % ' '.join(cmd)
|
|
print cmd_output
|
|
print "Disable automatic lookup and caching of dSYMs before running the test suite!"
|
|
print "Exiting..."
|
|
sys.exit(0)
|
|
|
|
# On MacOS X, check to make sure that domain for com.apple.DebugSymbols defaults
|
|
# does not exist before proceeding to running the test suite.
|
|
if sys.platform.startswith("darwin"):
|
|
checkDsymForUUIDIsNotOn()
|
|
|
|
#
|
|
# Start the actions by first parsing the options while setting up the test
|
|
# directories, followed by setting up the search paths for lldb utilities;
|
|
# then, we walk the directory trees and collect the tests into our test suite.
|
|
#
|
|
parseOptionsAndInitTestdirs()
|
|
setupSysPath()
|
|
|
|
#
|
|
# If '-d' is specified, do a delay of 10 seconds for the debugger to attach.
|
|
#
|
|
if delay:
|
|
doDelay(10)
|
|
|
|
#
|
|
# If '-l' is specified, do not skip the long running tests.
|
|
if not skip_long_running_test:
|
|
os.environ["LLDB_SKIP_LONG_RUNNING_TEST"] = "NO"
|
|
|
|
#
|
|
# Walk through the testdirs while collecting tests.
|
|
#
|
|
for testdir in testdirs:
|
|
os.path.walk(testdir, visit, 'Test')
|
|
|
|
#
|
|
# Now that we have loaded all the test cases, run the whole test suite.
|
|
#
|
|
|
|
# For the time being, let's bracket the test runner within the
|
|
# lldb.SBDebugger.Initialize()/Terminate() pair.
|
|
import lldb, atexit
|
|
# Update: the act of importing lldb now executes lldb.SBDebugger.Initialize(),
|
|
# there's no need to call it a second time.
|
|
#lldb.SBDebugger.Initialize()
|
|
atexit.register(lambda: lldb.SBDebugger.Terminate())
|
|
|
|
# Create a singleton SBDebugger in the lldb namespace.
|
|
lldb.DBG = lldb.SBDebugger.Create()
|
|
|
|
# Put the blacklist in the lldb namespace, to be used by lldb.TestBase.
|
|
lldb.blacklist = blacklist
|
|
|
|
# Put dont/just_do_python_api_test in the lldb namespace.
|
|
lldb.dont_do_python_api_test = dont_do_python_api_test
|
|
lldb.just_do_python_api_test = just_do_python_api_test
|
|
lldb.just_do_benchmarks_test = just_do_benchmarks_test
|
|
|
|
# Do we need to skip build and cleanup?
|
|
lldb.skip_build_and_cleanup = skip_build_and_cleanup
|
|
|
|
# Put bmExecutable, bmBreakpointSpec, and bmIterationCount into the lldb namespace, too.
|
|
lldb.bmExecutable = bmExecutable
|
|
lldb.bmBreakpointSpec = bmBreakpointSpec
|
|
lldb.bmIterationCount = bmIterationCount
|
|
|
|
# And don't forget the runHooks!
|
|
lldb.runHooks = runHooks
|
|
|
|
# Turn on lldb loggings if necessary.
|
|
lldbLoggings()
|
|
|
|
# Install the control-c handler.
|
|
unittest2.signals.installHandler()
|
|
|
|
# If sdir_name is not specified through the '-s sdir_name' option, get a
|
|
# timestamp string and export it as LLDB_SESSION_DIR environment var. This will
|
|
# be used when/if we want to dump the session info of individual test cases
|
|
# later on.
|
|
#
|
|
# See also TestBase.dumpSessionInfo() in lldbtest.py.
|
|
if not sdir_name:
|
|
import datetime
|
|
# The windows platforms don't like ':' in the pathname.
|
|
timestamp = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S")
|
|
sdir_name = timestamp
|
|
os.environ["LLDB_SESSION_DIRNAME"] = os.path.join(os.getcwd(), sdir_name)
|
|
|
|
if not noHeaders:
|
|
sys.stderr.write("\nSession logs for test failures/errors/unexpected successes"
|
|
" will go into directory '%s'\n" % sdir_name)
|
|
sys.stderr.write("Command invoked: %s\n" % getMyCommandLine())
|
|
|
|
if not os.path.isdir(sdir_name):
|
|
os.mkdir(sdir_name)
|
|
fname = os.path.join(sdir_name, "svn-info")
|
|
with open(fname, "w") as f:
|
|
print >> f, svn_info
|
|
print >> f, "Command invoked: %s\n" % getMyCommandLine()
|
|
|
|
#
|
|
# Invoke the default TextTestRunner to run the test suite, possibly iterating
|
|
# over different configurations.
|
|
#
|
|
|
|
iterArchs = False
|
|
iterCompilers = False
|
|
|
|
if not archs and "archs" in config:
|
|
archs = config["archs"]
|
|
|
|
if isinstance(archs, list) and len(archs) >= 1:
|
|
iterArchs = True
|
|
|
|
if not compilers and "compilers" in config:
|
|
compilers = config["compilers"]
|
|
|
|
if isinstance(compilers, list) and len(compilers) >= 1:
|
|
iterCompilers = True
|
|
|
|
# Make a shallow copy of sys.path, we need to manipulate the search paths later.
|
|
# This is only necessary if we are relocated and with different configurations.
|
|
if rdir:
|
|
old_sys_path = sys.path[:]
|
|
# If we iterate on archs or compilers, there is a chance we want to split stderr/stdout.
|
|
if iterArchs or iterCompilers:
|
|
old_stderr = sys.stderr
|
|
old_stdout = sys.stdout
|
|
new_stderr = None
|
|
new_stdout = None
|
|
|
|
# Iterating over all possible architecture and compiler combinations.
|
|
for ia in range(len(archs) if iterArchs else 1):
|
|
archConfig = ""
|
|
if iterArchs:
|
|
os.environ["ARCH"] = archs[ia]
|
|
archConfig = "arch=%s" % archs[ia]
|
|
for ic in range(len(compilers) if iterCompilers else 1):
|
|
if iterCompilers:
|
|
os.environ["CC"] = compilers[ic]
|
|
configString = "%s compiler=%s" % (archConfig, compilers[ic])
|
|
else:
|
|
configString = archConfig
|
|
|
|
if iterArchs or iterCompilers:
|
|
# Translate ' ' to '-' for pathname component.
|
|
from string import maketrans
|
|
tbl = maketrans(' ', '-')
|
|
configPostfix = configString.translate(tbl)
|
|
|
|
# Check whether we need to split stderr/stdout into configuration
|
|
# specific files.
|
|
if old_stderr.name != '<stderr>' and config.get('split_stderr'):
|
|
if new_stderr:
|
|
new_stderr.close()
|
|
new_stderr = open("%s.%s" % (old_stderr.name, configPostfix), "w")
|
|
sys.stderr = new_stderr
|
|
if old_stdout.name != '<stdout>' and config.get('split_stdout'):
|
|
if new_stdout:
|
|
new_stdout.close()
|
|
new_stdout = open("%s.%s" % (old_stdout.name, configPostfix), "w")
|
|
sys.stdout = new_stdout
|
|
|
|
# If we specified a relocated directory to run the test suite, do
|
|
# the extra housekeeping to copy the testdirs to a configStringified
|
|
# directory and to update sys.path before invoking the test runner.
|
|
# The purpose is to separate the configuration-specific directories
|
|
# from each other.
|
|
if rdir:
|
|
from shutil import copytree, ignore_patterns
|
|
|
|
newrdir = "%s.%s" % (rdir, configPostfix)
|
|
|
|
# Copy the tree to a new directory with postfix name configPostfix.
|
|
copytree(rdir, newrdir, ignore=ignore_patterns('*.pyc', '*.o', '*.d'))
|
|
|
|
# Update the LLDB_TEST environment variable to reflect new top
|
|
# level test directory.
|
|
#
|
|
# See also lldbtest.TestBase.setUpClass(cls).
|
|
if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test':
|
|
os.environ["LLDB_TEST"] = os.path.join(newrdir, 'test')
|
|
else:
|
|
os.environ["LLDB_TEST"] = newrdir
|
|
|
|
# And update the Python search paths for modules.
|
|
sys.path = [x.replace(rdir, newrdir, 1) for x in old_sys_path]
|
|
|
|
# Output the configuration.
|
|
sys.stderr.write("\nConfiguration: " + configString + "\n")
|
|
|
|
#print "sys.stderr name is", sys.stderr.name
|
|
#print "sys.stdout name is", sys.stdout.name
|
|
|
|
# First, write out the number of collected test cases.
|
|
sys.stderr.write(separator + "\n")
|
|
sys.stderr.write("Collected %d test%s\n\n"
|
|
% (suite.countTestCases(),
|
|
suite.countTestCases() != 1 and "s" or ""))
|
|
|
|
class LLDBTestResult(unittest2.TextTestResult):
|
|
"""
|
|
Enforce a singleton pattern to allow introspection of test progress.
|
|
|
|
Overwrite addError(), addFailure(), and addExpectedFailure() methods
|
|
to enable each test instance to track its failure/error status. It
|
|
is used in the LLDB test framework to emit detailed trace messages
|
|
to a log file for easier human inspection of test failres/errors.
|
|
"""
|
|
__singleton__ = None
|
|
__ignore_singleton__ = False
|
|
|
|
def __init__(self, *args):
|
|
if not LLDBTestResult.__ignore_singleton__ and LLDBTestResult.__singleton__:
|
|
raise Exception("LLDBTestResult instantiated more than once")
|
|
super(LLDBTestResult, self).__init__(*args)
|
|
LLDBTestResult.__singleton__ = self
|
|
# Now put this singleton into the lldb module namespace.
|
|
lldb.test_result = self
|
|
# Computes the format string for displaying the counter.
|
|
global suite
|
|
counterWidth = len(str(suite.countTestCases()))
|
|
self.fmt = "%" + str(counterWidth) + "d: "
|
|
self.indentation = ' ' * (counterWidth + 2)
|
|
# This counts from 1 .. suite.countTestCases().
|
|
self.counter = 0
|
|
|
|
def getDescription(self, test):
|
|
doc_first_line = test.shortDescription()
|
|
if self.descriptions and doc_first_line:
|
|
return '\n'.join((str(test), self.indentation + doc_first_line))
|
|
else:
|
|
return str(test)
|
|
|
|
def startTest(self, test):
|
|
self.counter += 1
|
|
if self.showAll:
|
|
self.stream.write(self.fmt % self.counter)
|
|
super(LLDBTestResult, self).startTest(test)
|
|
|
|
def stopTest(self, test):
|
|
"""Called when the given test has been run"""
|
|
if progress_bar:
|
|
sys.__stdout__.write('.')
|
|
sys.__stdout__.flush()
|
|
if self.counter == suite.countTestCases():
|
|
sys.__stdout__.write('\n')
|
|
|
|
super(LLDBTestResult, self).stopTest(test)
|
|
|
|
def addError(self, test, err):
|
|
global sdir_has_content
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addError(test, err)
|
|
method = getattr(test, "markError", None)
|
|
if method:
|
|
method()
|
|
|
|
def addFailure(self, test, err):
|
|
global sdir_has_content
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addFailure(test, err)
|
|
method = getattr(test, "markFailure", None)
|
|
if method:
|
|
method()
|
|
|
|
def addExpectedFailure(self, test, err):
|
|
global sdir_has_content
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addExpectedFailure(test, err)
|
|
method = getattr(test, "markExpectedFailure", None)
|
|
if method:
|
|
method()
|
|
|
|
def addSkip(self, test, reason):
|
|
global sdir_has_content
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addSkip(test, reason)
|
|
method = getattr(test, "markSkippedTest", None)
|
|
if method:
|
|
method()
|
|
|
|
def addUnexpectedSuccess(self, test):
|
|
global sdir_has_content
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addUnexpectedSuccess(test)
|
|
method = getattr(test, "markUnexpectedSuccess", None)
|
|
if method:
|
|
method()
|
|
|
|
# Invoke the test runner.
|
|
if count == 1:
|
|
result = unittest2.TextTestRunner(stream=sys.stderr,
|
|
verbosity=verbose,
|
|
failfast=failfast,
|
|
resultclass=LLDBTestResult).run(suite)
|
|
else:
|
|
# We are invoking the same test suite more than once. In this case,
|
|
# mark __ignore_singleton__ flag as True so the signleton pattern is
|
|
# not enforced.
|
|
LLDBTestResult.__ignore_singleton__ = True
|
|
for i in range(count):
|
|
result = unittest2.TextTestRunner(stream=sys.stderr,
|
|
verbosity=verbose,
|
|
failfast=failfast,
|
|
resultclass=LLDBTestResult).run(suite)
|
|
|
|
|
|
if sdir_has_content:
|
|
sys.stderr.write("Session logs for test failures/errors/unexpected successes"
|
|
" can be found in directory '%s'\n" % sdir_name)
|
|
|
|
# Terminate the test suite if ${LLDB_TESTSUITE_FORCE_FINISH} is defined.
|
|
# This should not be necessary now.
|
|
if ("LLDB_TESTSUITE_FORCE_FINISH" in os.environ):
|
|
print "Terminating Test suite..."
|
|
subprocess.Popen(["/bin/sh", "-c", "kill %s; exit 0" % (os.getpid())])
|
|
|
|
# Exiting.
|
|
sys.exit(not result.wasSuccessful)
|