forked from OSchip/llvm-project
2004 lines
78 KiB
Python
Executable File
2004 lines
78 KiB
Python
Executable File
#!/usr/bin/env python
|
|
|
|
"""
|
|
A simple testing framework for lldb using python's unit testing framework.
|
|
|
|
Tests for lldb are written as python scripts which take advantage of the script
|
|
bridging provided by LLDB.framework to interact with lldb core.
|
|
|
|
A specific naming pattern is followed by the .py script to be recognized as
|
|
a module which implements a test scenario, namely, Test*.py.
|
|
|
|
To specify the directories where "Test*.py" python test scripts are located,
|
|
you need to pass in a list of directory names. By default, the current
|
|
working directory is searched if nothing is specified on the command line.
|
|
|
|
Type:
|
|
|
|
./dotest.py -h
|
|
|
|
for available options.
|
|
"""
|
|
|
|
from __future__ import print_function
|
|
|
|
import lldb_shared
|
|
|
|
import atexit
|
|
import commands
|
|
import importlib
|
|
import os
|
|
import dotest_args
|
|
import errno
|
|
import platform
|
|
import progress
|
|
import signal
|
|
import socket
|
|
import subprocess
|
|
import sys
|
|
import test_results
|
|
from test_results import EventBuilder
|
|
import inspect
|
|
import unittest2
|
|
import lldbtest_config
|
|
import test_categories
|
|
|
|
import six
|
|
|
|
def is_exe(fpath):
|
|
"""Returns true if fpath is an executable."""
|
|
return os.path.isfile(fpath) and os.access(fpath, os.X_OK)
|
|
|
|
def which(program):
|
|
"""Returns the full path to a program; None otherwise."""
|
|
fpath, fname = os.path.split(program)
|
|
if fpath:
|
|
if is_exe(program):
|
|
return program
|
|
else:
|
|
for path in os.environ["PATH"].split(os.pathsep):
|
|
exe_file = os.path.join(path, program)
|
|
if is_exe(exe_file):
|
|
return exe_file
|
|
return None
|
|
|
|
class _WritelnDecorator(object):
|
|
"""Used to decorate file-like objects with a handy 'writeln' method"""
|
|
def __init__(self,stream):
|
|
self.stream = stream
|
|
|
|
def __getattr__(self, attr):
|
|
if attr in ('stream', '__getstate__'):
|
|
raise AttributeError(attr)
|
|
return getattr(self.stream,attr)
|
|
|
|
def writeln(self, arg=None):
|
|
if arg:
|
|
self.write(arg)
|
|
self.write('\n') # text-mode streams translate to \r\n if needed
|
|
|
|
#
|
|
# Global variables:
|
|
#
|
|
|
|
# The test suite.
|
|
suite = unittest2.TestSuite()
|
|
|
|
# By default, lldb-mi tests are performed if lldb-mi can be found.
|
|
# Use @lldbmi_test decorator, defined in lldbtest.py, to mark a test as
|
|
# a lldb-mi test.
|
|
dont_do_lldbmi_test = False
|
|
just_do_lldbmi_test = False
|
|
|
|
# By default, benchmarks tests are not run.
|
|
just_do_benchmarks_test = False
|
|
|
|
dont_do_dsym_test = False
|
|
dont_do_dwarf_test = False
|
|
dont_do_dwo_test = False
|
|
|
|
# The blacklist is optional (-b blacklistFile) and allows a central place to skip
|
|
# testclass's and/or testclass.testmethod's.
|
|
blacklist = None
|
|
|
|
# The dictionary as a result of sourcing blacklistFile.
|
|
blacklistConfig = {}
|
|
|
|
# The list of categories we said we care about
|
|
categoriesList = None
|
|
# set to true if we are going to use categories for cherry-picking test cases
|
|
useCategories = False
|
|
# Categories we want to skip
|
|
skipCategories = []
|
|
# use this to track per-category failures
|
|
failuresPerCategory = {}
|
|
|
|
# The path to LLDB.framework is optional.
|
|
lldbFrameworkPath = None
|
|
|
|
# The config file is optional.
|
|
configFile = None
|
|
|
|
# Test suite repeat count. Can be overwritten with '-# count'.
|
|
count = 1
|
|
|
|
# The dictionary as a result of sourcing configFile.
|
|
config = {}
|
|
# The pre_flight and post_flight functions come from reading a config file.
|
|
pre_flight = None
|
|
post_flight = None
|
|
# So do the lldbtest_remote_sandbox and lldbtest_remote_shell_template variables.
|
|
lldbtest_remote_sandbox = None
|
|
lldbtest_remote_shell_template = None
|
|
|
|
# The 'archs' and 'compilers' can be specified via either command line or configFile,
|
|
# with the command line overriding the configFile. The corresponding options can be
|
|
# specified more than once. For example, "-A x86_64 -A i386" => archs=['x86_64', 'i386']
|
|
# and "-C gcc -C clang" => compilers=['gcc', 'clang'].
|
|
archs = None # Must be initialized after option parsing
|
|
compilers = None # Must be initialized after option parsing
|
|
|
|
# The arch might dictate some specific CFLAGS to be passed to the toolchain to build
|
|
# the inferior programs. The global variable cflags_extras provides a hook to do
|
|
# just that.
|
|
cflags_extras = ''
|
|
|
|
# Dump the Python sys.path variable. Use '-D' to dump sys.path.
|
|
dumpSysPath = False
|
|
|
|
# Full path of the benchmark executable, as specified by the '-e' option.
|
|
bmExecutable = None
|
|
# The breakpoint specification of bmExecutable, as specified by the '-x' option.
|
|
bmBreakpointSpec = None
|
|
# The benchmark iteration count, as specified by the '-y' option.
|
|
bmIterationCount = -1
|
|
|
|
# By default, don't exclude any directories. Use '-X' to add one excluded directory.
|
|
excluded = set(['.svn', '.git'])
|
|
|
|
# By default, failfast is False. Use '-F' to overwrite it.
|
|
failfast = False
|
|
|
|
# The filters (testclass.testmethod) used to admit tests into our test suite.
|
|
filters = []
|
|
|
|
# The runhooks is a list of lldb commands specifically for the debugger.
|
|
# Use '-k' to specify a runhook.
|
|
runHooks = []
|
|
|
|
# If '-g' is specified, the filterspec is not exclusive. If a test module does
|
|
# not contain testclass.testmethod which matches the filterspec, the whole test
|
|
# module is still admitted into our test suite. fs4all flag defaults to True.
|
|
fs4all = True
|
|
|
|
# Ignore the build search path relative to this script to locate the lldb.py module.
|
|
ignore = False
|
|
|
|
# By default, we do not skip build and cleanup. Use '-S' option to override.
|
|
skip_build_and_cleanup = False
|
|
|
|
# By default, we skip long running test case. Use '-l' option to override.
|
|
skip_long_running_test = True
|
|
|
|
# By default, we print the build dir, lldb version, and svn info. Use '-n' option to
|
|
# turn it off.
|
|
noHeaders = False
|
|
|
|
# Parsable mode silences headers, and any other output this script might generate, and instead
|
|
# prints machine-readable output similar to what clang tests produce.
|
|
parsable = False
|
|
|
|
# The regular expression pattern to match against eligible filenames as our test cases.
|
|
regexp = None
|
|
|
|
# By default, tests are executed in place and cleanups are performed afterwards.
|
|
# Use '-r dir' option to relocate the tests and their intermediate files to a
|
|
# different directory and to forgo any cleanups. The directory specified must
|
|
# not exist yet.
|
|
rdir = None
|
|
|
|
# By default, recorded session info for errored/failed test are dumped into its
|
|
# own file under a session directory named after the timestamp of the test suite
|
|
# run. Use '-s session-dir-name' to specify a specific dir name.
|
|
sdir_name = None
|
|
|
|
# Set this flag if there is any session info dumped during the test run.
|
|
sdir_has_content = False
|
|
|
|
# svn_info stores the output from 'svn info lldb.base.dir'.
|
|
svn_info = ''
|
|
|
|
# svn_silent means do not try to obtain svn status
|
|
svn_silent = True
|
|
|
|
# Default verbosity is 0.
|
|
verbose = 1
|
|
|
|
# Set to True only if verbose is 0 and LLDB trace mode is off.
|
|
progress_bar = False
|
|
|
|
# By default, search from the script directory.
|
|
# We can't use sys.path[0] to determine the script directory
|
|
# because it doesn't work under a debugger
|
|
testdirs = [ os.path.dirname(os.path.realpath(__file__)) ]
|
|
|
|
# Separator string.
|
|
separator = '-' * 70
|
|
|
|
failed = False
|
|
|
|
# LLDB Remote platform setting
|
|
lldb_platform_name = None
|
|
lldb_platform_url = None
|
|
lldb_platform_working_dir = None
|
|
|
|
# Parallel execution settings
|
|
is_inferior_test_runner = False
|
|
multiprocess_test_subdir = None
|
|
num_threads = None
|
|
output_on_success = False
|
|
no_multiprocess_test_runner = False
|
|
test_runner_name = None
|
|
|
|
# Test results handling globals
|
|
results_filename = None
|
|
results_port = None
|
|
results_file_object = None
|
|
results_formatter_name = None
|
|
results_formatter_object = None
|
|
results_formatter_options = None
|
|
|
|
# The names of all tests. Used to assert we don't have two tests with the same base name.
|
|
all_tests = set()
|
|
|
|
def usage(parser):
|
|
parser.print_help()
|
|
if verbose > 0:
|
|
print("""
|
|
Examples:
|
|
|
|
This is an example of using the -f option to pinpoint to a specific test class
|
|
and test method to be run:
|
|
|
|
$ ./dotest.py -f ClassTypesTestCase.test_with_dsym_and_run_command
|
|
----------------------------------------------------------------------
|
|
Collected 1 test
|
|
|
|
test_with_dsym_and_run_command (TestClassTypes.ClassTypesTestCase)
|
|
Test 'frame variable this' when stopped on a class constructor. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 1 test in 1.396s
|
|
|
|
OK
|
|
|
|
And this is an example of using the -p option to run a single file (the filename
|
|
matches the pattern 'ObjC' and it happens to be 'TestObjCMethods.py'):
|
|
|
|
$ ./dotest.py -v -p ObjC
|
|
----------------------------------------------------------------------
|
|
Collected 4 tests
|
|
|
|
test_break_with_dsym (TestObjCMethods.FoundationTestCase)
|
|
Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok
|
|
test_break_with_dwarf (TestObjCMethods.FoundationTestCase)
|
|
Test setting objc breakpoints using '_regexp-break' and 'breakpoint set'. ... ok
|
|
test_data_type_and_expr_with_dsym (TestObjCMethods.FoundationTestCase)
|
|
Lookup objective-c data types and evaluate expressions. ... ok
|
|
test_data_type_and_expr_with_dwarf (TestObjCMethods.FoundationTestCase)
|
|
Lookup objective-c data types and evaluate expressions. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 4 tests in 16.661s
|
|
|
|
OK
|
|
|
|
Running of this script also sets up the LLDB_TEST environment variable so that
|
|
individual test cases can locate their supporting files correctly. The script
|
|
tries to set up Python's search paths for modules by looking at the build tree
|
|
relative to this script. See also the '-i' option in the following example.
|
|
|
|
Finally, this is an example of using the lldb.py module distributed/installed by
|
|
Xcode4 to run against the tests under the 'forward' directory, and with the '-w'
|
|
option to add some delay between two tests. It uses ARCH=x86_64 to specify that
|
|
as the architecture and CC=clang to specify the compiler used for the test run:
|
|
|
|
$ PYTHONPATH=/Xcode4/Library/PrivateFrameworks/LLDB.framework/Versions/A/Resources/Python ARCH=x86_64 CC=clang ./dotest.py -v -w -i forward
|
|
|
|
Session logs for test failures/errors will go into directory '2010-11-11-13_56_16'
|
|
----------------------------------------------------------------------
|
|
Collected 2 tests
|
|
|
|
test_with_dsym_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase)
|
|
Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok
|
|
test_with_dwarf_and_run_command (TestForwardDeclaration.ForwardDeclarationTestCase)
|
|
Display *bar_ptr when stopped on a function with forward declaration of struct bar. ... ok
|
|
|
|
----------------------------------------------------------------------
|
|
Ran 2 tests in 5.659s
|
|
|
|
OK
|
|
|
|
The 'Session ...' verbiage is recently introduced (see also the '-s' option) to
|
|
notify the directory containing the session logs for test failures or errors.
|
|
In case there is any test failure/error, a similar message is appended at the
|
|
end of the stderr output for your convenience.
|
|
|
|
ENABLING LOGS FROM TESTS
|
|
|
|
Option 1:
|
|
|
|
Writing logs into different files per test case::
|
|
|
|
This option is particularly useful when multiple dotest instances are created
|
|
by dosep.py
|
|
|
|
$ ./dotest.py --channel "lldb all"
|
|
|
|
$ ./dotest.py --channel "lldb all" --channel "gdb-remote packets"
|
|
|
|
These log files are written to:
|
|
|
|
<session-dir>/<test-id>-host.log (logs from lldb host process)
|
|
<session-dir>/<test-id>-server.log (logs from debugserver/lldb-server)
|
|
<session-dir>/<test-id>-<test-result>.log (console logs)
|
|
|
|
By default, logs from successful runs are deleted. Use the --log-success flag
|
|
to create reference logs for debugging.
|
|
|
|
$ ./dotest.py --log-success
|
|
|
|
Option 2: (DEPRECATED)
|
|
|
|
The following options can only enable logs from the host lldb process.
|
|
Only categories from the "lldb" or "gdb-remote" channels can be enabled
|
|
They also do not automatically enable logs in locally running debug servers.
|
|
Also, logs from all test case are written into each log file
|
|
|
|
o LLDB_LOG: if defined, specifies the log file pathname for the 'lldb' subsystem
|
|
with a default option of 'event process' if LLDB_LOG_OPTION is not defined.
|
|
|
|
o GDB_REMOTE_LOG: if defined, specifies the log file pathname for the
|
|
'process.gdb-remote' subsystem with a default option of 'packets' if
|
|
GDB_REMOTE_LOG_OPTION is not defined.
|
|
|
|
""")
|
|
sys.exit(0)
|
|
|
|
|
|
def setCrashInfoHook_Mac(text):
|
|
import crashinfo
|
|
crashinfo.setCrashReporterDescription(text)
|
|
|
|
# implement this in some suitable way for your platform, and then bind it
|
|
# to setCrashInfoHook
|
|
def setCrashInfoHook_NonMac(text):
|
|
pass
|
|
|
|
setCrashInfoHook = None
|
|
|
|
def deleteCrashInfoDylib(dylib_path):
|
|
try:
|
|
# Need to modify this to handle multiple tests running at the same time. If we move this
|
|
# to the test's real dir, all should be we run sequentially within a test directory.
|
|
# os.remove(dylib_path)
|
|
None
|
|
finally:
|
|
pass
|
|
|
|
def setupCrashInfoHook():
|
|
global setCrashInfoHook
|
|
setCrashInfoHook = setCrashInfoHook_NonMac # safe default
|
|
if platform.system() == "Darwin":
|
|
import lock
|
|
test_dir = os.environ['LLDB_TEST']
|
|
if not test_dir or not os.path.exists(test_dir):
|
|
return
|
|
dylib_lock = os.path.join(test_dir,"crashinfo.lock")
|
|
dylib_src = os.path.join(test_dir,"crashinfo.c")
|
|
dylib_dst = os.path.join(test_dir,"crashinfo.so")
|
|
try:
|
|
compile_lock = lock.Lock(dylib_lock)
|
|
compile_lock.acquire()
|
|
if not os.path.isfile(dylib_dst) or os.path.getmtime(dylib_dst) < os.path.getmtime(dylib_src):
|
|
# we need to compile
|
|
cmd = "SDKROOT= xcrun clang %s -o %s -framework Python -Xlinker -dylib -iframework /System/Library/Frameworks/ -Xlinker -F /System/Library/Frameworks/" % (dylib_src,dylib_dst)
|
|
if subprocess.call(cmd,shell=True) != 0 or not os.path.isfile(dylib_dst):
|
|
raise Exception('command failed: "{}"'.format(cmd))
|
|
finally:
|
|
compile_lock.release()
|
|
del compile_lock
|
|
|
|
setCrashInfoHook = setCrashInfoHook_Mac
|
|
|
|
else:
|
|
pass
|
|
|
|
def parseOptionsAndInitTestdirs():
|
|
"""Initialize the list of directories containing our unittest scripts.
|
|
|
|
'-h/--help as the first option prints out usage info and exit the program.
|
|
"""
|
|
|
|
global dont_do_lldbmi_test
|
|
global just_do_lldbmi_test
|
|
global just_do_benchmarks_test
|
|
global dont_do_dsym_test
|
|
global dont_do_dwarf_test
|
|
global dont_do_dwo_test
|
|
global blacklist
|
|
global blacklistConfig
|
|
global categoriesList
|
|
global validCategories
|
|
global useCategories
|
|
global skipCategories
|
|
global lldbFrameworkPath
|
|
global configFile
|
|
global archs
|
|
global compilers
|
|
global count
|
|
global dumpSysPath
|
|
global bmExecutable
|
|
global bmBreakpointSpec
|
|
global bmIterationCount
|
|
global failfast
|
|
global filters
|
|
global fs4all
|
|
global ignore
|
|
global progress_bar
|
|
global runHooks
|
|
global skip_build_and_cleanup
|
|
global skip_long_running_test
|
|
global noHeaders
|
|
global parsable
|
|
global regexp
|
|
global rdir
|
|
global sdir_name
|
|
global svn_silent
|
|
global verbose
|
|
global testdirs
|
|
global lldb_platform_name
|
|
global lldb_platform_url
|
|
global lldb_platform_working_dir
|
|
global setCrashInfoHook
|
|
global is_inferior_test_runner
|
|
global multiprocess_test_subdir
|
|
global num_threads
|
|
global output_on_success
|
|
global no_multiprocess_test_runner
|
|
global test_runner_name
|
|
global results_filename
|
|
global results_formatter_name
|
|
global results_formatter_options
|
|
global results_port
|
|
|
|
do_help = False
|
|
|
|
platform_system = platform.system()
|
|
platform_machine = platform.machine()
|
|
|
|
parser = dotest_args.create_parser()
|
|
args = dotest_args.parse_args(parser, sys.argv[1:])
|
|
|
|
if args.unset_env_varnames:
|
|
for env_var in args.unset_env_varnames:
|
|
if env_var in os.environ:
|
|
# From Python Doc: When unsetenv() is supported, deletion of items in os.environ
|
|
# is automatically translated into a corresponding call to unsetenv().
|
|
del os.environ[env_var]
|
|
#os.unsetenv(env_var)
|
|
|
|
if args.set_env_vars:
|
|
for env_var in args.set_env_vars:
|
|
parts = env_var.split('=', 1)
|
|
if len(parts) == 1:
|
|
os.environ[parts[0]] = ""
|
|
else:
|
|
os.environ[parts[0]] = parts[1]
|
|
|
|
# only print the args if being verbose (and parsable is off)
|
|
if args.v and not args.q:
|
|
print(sys.argv)
|
|
|
|
if args.h:
|
|
do_help = True
|
|
|
|
if args.compilers:
|
|
compilers = args.compilers
|
|
else:
|
|
# Use a compiler appropriate appropriate for the Apple SDK if one was specified
|
|
if platform_system == 'Darwin' and args.apple_sdk:
|
|
compilers = [commands.getoutput('xcrun -sdk "%s" -find clang 2> /dev/null' % (args.apple_sdk))]
|
|
else:
|
|
# 'clang' on ubuntu 14.04 is 3.4 so we try clang-3.5 first
|
|
candidateCompilers = ['clang-3.5', 'clang', 'gcc']
|
|
for candidate in candidateCompilers:
|
|
if which(candidate):
|
|
compilers = [candidate]
|
|
break
|
|
|
|
if args.channels:
|
|
lldbtest_config.channels = args.channels
|
|
|
|
if args.log_success:
|
|
lldbtest_config.log_success = args.log_success
|
|
|
|
# Set SDKROOT if we are using an Apple SDK
|
|
if platform_system == 'Darwin' and args.apple_sdk:
|
|
os.environ['SDKROOT'] = commands.getoutput('xcrun --sdk "%s" --show-sdk-path 2> /dev/null' % (args.apple_sdk))
|
|
|
|
if args.archs:
|
|
archs = args.archs
|
|
for arch in archs:
|
|
if arch.startswith('arm') and platform_system == 'Darwin' and not args.apple_sdk:
|
|
os.environ['SDKROOT'] = commands.getoutput('xcrun --sdk iphoneos.internal --show-sdk-path 2> /dev/null')
|
|
if not os.path.exists(os.environ['SDKROOT']):
|
|
os.environ['SDKROOT'] = commands.getoutput('xcrun --sdk iphoneos --show-sdk-path 2> /dev/null')
|
|
else:
|
|
archs = [platform_machine]
|
|
|
|
if args.categoriesList:
|
|
categoriesList = set(test_categories.validate(args.categoriesList, False))
|
|
useCategories = True
|
|
else:
|
|
categoriesList = []
|
|
|
|
if args.skipCategories:
|
|
skipCategories = test_categories.validate(args.skipCategories, False)
|
|
|
|
if args.D:
|
|
dumpSysPath = True
|
|
|
|
if args.E:
|
|
cflags_extras = args.E
|
|
os.environ['CFLAGS_EXTRAS'] = cflags_extras
|
|
|
|
# argparse makes sure we have correct options
|
|
if args.N == 'dwarf':
|
|
dont_do_dwarf_test = True
|
|
elif args.N == 'dwo':
|
|
dont_do_dwo_test = True
|
|
elif args.N == 'dsym':
|
|
dont_do_dsym_test = True
|
|
|
|
if args.a or args.plus_a:
|
|
print("Options '-a' and '+a' have been deprecated. Please use the test category\n"
|
|
"functionality (-G pyapi, --skip-category pyapi) instead.")
|
|
sys.exit(1)
|
|
|
|
if args.plus_b:
|
|
just_do_benchmarks_test = True
|
|
|
|
if args.b:
|
|
if args.b.startswith('-'):
|
|
usage(parser)
|
|
blacklistFile = args.b
|
|
if not os.path.isfile(blacklistFile):
|
|
print('Blacklist file:', blacklistFile, 'does not exist!')
|
|
usage(parser)
|
|
# Now read the blacklist contents and assign it to blacklist.
|
|
execfile(blacklistFile, globals(), blacklistConfig)
|
|
blacklist = blacklistConfig.get('blacklist')
|
|
|
|
if args.c:
|
|
if args.c.startswith('-'):
|
|
usage(parser)
|
|
configFile = args.c
|
|
if not os.path.isfile(configFile):
|
|
print('Config file:', configFile, 'does not exist!')
|
|
usage(parser)
|
|
|
|
if args.d:
|
|
sys.stdout.write("Suspending the process %d to wait for debugger to attach...\n" % os.getpid())
|
|
sys.stdout.flush()
|
|
os.kill(os.getpid(), signal.SIGSTOP)
|
|
|
|
if args.e:
|
|
if args.e.startswith('-'):
|
|
usage(parser)
|
|
bmExecutable = args.e
|
|
if not is_exe(bmExecutable):
|
|
usage(parser)
|
|
|
|
if args.F:
|
|
failfast = True
|
|
|
|
if args.f:
|
|
if any([x.startswith('-') for x in args.f]):
|
|
usage(parser)
|
|
filters.extend(args.f)
|
|
# Shut off multiprocessing mode when additional filters are specified.
|
|
# The rational is that the user is probably going after a very specific
|
|
# test and doesn't need a bunch of parallel test runners all looking for
|
|
# it in a frenzy. Also, '-v' now spits out all test run output even
|
|
# on success, so the standard recipe for redoing a failing test (with -v
|
|
# and a -f to filter to the specific test) now causes all test scanning
|
|
# (in parallel) to print results for do-nothing runs in a very distracting
|
|
# manner. If we really need filtered parallel runs in the future, consider
|
|
# adding a --no-output-on-success that prevents -v from setting
|
|
# output-on-success.
|
|
no_multiprocess_test_runner = True
|
|
|
|
if args.g:
|
|
fs4all = False
|
|
|
|
if args.i:
|
|
ignore = True
|
|
|
|
if args.k:
|
|
runHooks.extend(args.k)
|
|
|
|
if args.l:
|
|
skip_long_running_test = False
|
|
|
|
if args.m:
|
|
dont_do_lldbmi_test = True
|
|
|
|
if args.plus_m:
|
|
if dont_do_lldbmi_test:
|
|
print("Warning: -m and +m can't both be specified! Using only -m")
|
|
else:
|
|
just_do_lldbmi_test = True
|
|
|
|
if args.framework:
|
|
lldbFrameworkPath = args.framework
|
|
|
|
if args.executable:
|
|
lldbtest_config.lldbExec = args.executable
|
|
|
|
if args.libcxx:
|
|
os.environ["LIBCXX_PATH"] = args.libcxx
|
|
|
|
if args.n:
|
|
noHeaders = True
|
|
|
|
if args.p:
|
|
if args.p.startswith('-'):
|
|
usage(parser)
|
|
regexp = args.p
|
|
|
|
if args.q:
|
|
noHeaders = True
|
|
parsable = True
|
|
|
|
if args.P and not args.v:
|
|
progress_bar = True
|
|
verbose = 0
|
|
|
|
if args.R:
|
|
if args.R.startswith('-'):
|
|
usage(parser)
|
|
rdir = os.path.abspath(args.R)
|
|
if os.path.exists(rdir):
|
|
import shutil
|
|
print('Removing tree:', rdir)
|
|
shutil.rmtree(rdir)
|
|
|
|
if args.r:
|
|
if args.r.startswith('-'):
|
|
usage(parser)
|
|
rdir = os.path.abspath(args.r)
|
|
if os.path.exists(rdir):
|
|
print('Relocated directory:', rdir, 'must not exist!')
|
|
usage(parser)
|
|
|
|
if args.S:
|
|
skip_build_and_cleanup = True
|
|
|
|
if args.s:
|
|
if args.s.startswith('-'):
|
|
usage(parser)
|
|
sdir_name = args.s
|
|
|
|
if args.t:
|
|
os.environ['LLDB_COMMAND_TRACE'] = 'YES'
|
|
|
|
if args.T:
|
|
svn_silent = False
|
|
|
|
if args.v:
|
|
verbose = 2
|
|
|
|
if args.w:
|
|
os.environ['LLDB_WAIT_BETWEEN_TEST_CASES'] = 'YES'
|
|
|
|
if args.X:
|
|
if args.X.startswith('-'):
|
|
usage(parser)
|
|
excluded.add(args.X)
|
|
|
|
if args.x:
|
|
if args.x.startswith('-'):
|
|
usage(parser)
|
|
bmBreakpointSpec = args.x
|
|
|
|
# argparse makes sure we have a number
|
|
if args.y:
|
|
bmIterationCount = args.y
|
|
|
|
# argparse makes sure we have a number
|
|
if args.sharp:
|
|
count = args.sharp
|
|
|
|
if sys.platform.startswith('win32'):
|
|
os.environ['LLDB_DISABLE_CRASH_DIALOG'] = str(args.disable_crash_dialog)
|
|
os.environ['LLDB_LAUNCH_INFERIORS_WITHOUT_CONSOLE'] = str(args.hide_inferior_console)
|
|
|
|
if do_help == True:
|
|
usage(parser)
|
|
|
|
# Do not specify both '-m' and '+m' at the same time.
|
|
if dont_do_lldbmi_test and just_do_lldbmi_test:
|
|
usage(parser)
|
|
|
|
if args.no_multiprocess:
|
|
no_multiprocess_test_runner = True
|
|
|
|
if args.inferior:
|
|
is_inferior_test_runner = True
|
|
|
|
# Turn on output_on_sucess if either explicitly added or -v specified.
|
|
if args.output_on_success or args.v:
|
|
output_on_success = True
|
|
|
|
if args.num_threads:
|
|
num_threads = args.num_threads
|
|
|
|
if args.test_subdir:
|
|
multiprocess_test_subdir = args.test_subdir
|
|
|
|
if args.test_runner_name:
|
|
test_runner_name = args.test_runner_name
|
|
|
|
# Capture test results-related args.
|
|
if args.results_file:
|
|
results_filename = args.results_file
|
|
|
|
if args.results_port:
|
|
results_port = args.results_port
|
|
|
|
if args.results_file and args.results_port:
|
|
sys.stderr.write(
|
|
"only one of --results-file and --results-port should "
|
|
"be specified\n")
|
|
usage(args)
|
|
|
|
if args.results_formatter:
|
|
results_formatter_name = args.results_formatter
|
|
if args.results_formatter_options:
|
|
results_formatter_options = args.results_formatter_options
|
|
|
|
if args.lldb_platform_name:
|
|
lldb_platform_name = args.lldb_platform_name
|
|
if args.lldb_platform_url:
|
|
lldb_platform_url = args.lldb_platform_url
|
|
if args.lldb_platform_working_dir:
|
|
lldb_platform_working_dir = args.lldb_platform_working_dir
|
|
|
|
if args.event_add_entries and len(args.event_add_entries) > 0:
|
|
entries = {}
|
|
# Parse out key=val pairs, separated by comma
|
|
for keyval in args.event_add_entries.split(","):
|
|
key_val_entry = keyval.split("=")
|
|
if len(key_val_entry) == 2:
|
|
(key, val) = key_val_entry
|
|
val_parts = val.split(':')
|
|
if len(val_parts) > 1:
|
|
(val, val_type) = val_parts
|
|
if val_type == 'int':
|
|
val = int(val)
|
|
entries[key] = val
|
|
# Tell the event builder to create all events with these
|
|
# key/val pairs in them.
|
|
if len(entries) > 0:
|
|
test_results.EventBuilder.add_entries_to_all_events(entries)
|
|
|
|
# Gather all the dirs passed on the command line.
|
|
if len(args.args) > 0:
|
|
testdirs = list(map(os.path.abspath, args.args))
|
|
# Shut off multiprocessing mode when test directories are specified.
|
|
no_multiprocess_test_runner = True
|
|
|
|
# If '-r dir' is specified, the tests should be run under the relocated
|
|
# directory. Let's copy the testdirs over.
|
|
if rdir:
|
|
from shutil import copytree, ignore_patterns
|
|
|
|
tmpdirs = []
|
|
orig_testdirs = testdirs[:]
|
|
for srcdir in testdirs:
|
|
# For example, /Volumes/data/lldb/svn/ToT/test/functionalities/watchpoint/hello_watchpoint
|
|
# shall be split into ['/Volumes/data/lldb/svn/ToT/', 'functionalities/watchpoint/hello_watchpoint'].
|
|
# Utilize the relative path to the 'test' directory to make our destination dir path.
|
|
if ("test" + os.sep) in srcdir:
|
|
to_split_on = "test" + os.sep
|
|
else:
|
|
to_split_on = "test"
|
|
dstdir = os.path.join(rdir, srcdir.split(to_split_on)[1])
|
|
dstdir = dstdir.rstrip(os.sep)
|
|
# Don't copy the *.pyc and .svn stuffs.
|
|
copytree(srcdir, dstdir, ignore=ignore_patterns('*.pyc', '.svn'))
|
|
tmpdirs.append(dstdir)
|
|
|
|
# This will be our modified testdirs.
|
|
testdirs = tmpdirs
|
|
|
|
# With '-r dir' specified, there's no cleanup of intermediate test files.
|
|
os.environ["LLDB_DO_CLEANUP"] = 'NO'
|
|
|
|
# If the original testdirs is ['test'], the make directory has already been copied
|
|
# recursively and is contained within the rdir/test dir. For anything
|
|
# else, we would need to copy over the make directory and its contents,
|
|
# so that, os.listdir(rdir) looks like, for example:
|
|
#
|
|
# array_types conditional_break make
|
|
#
|
|
# where the make directory contains the Makefile.rules file.
|
|
if len(testdirs) != 1 or os.path.basename(orig_testdirs[0]) != 'test':
|
|
scriptdir = os.path.dirname(__file__)
|
|
# Don't copy the .svn stuffs.
|
|
copytree(os.path.join(scriptdir, 'make'), os.path.join(rdir, 'make'),
|
|
ignore=ignore_patterns('.svn'))
|
|
|
|
#print("testdirs:", testdirs)
|
|
|
|
# Source the configFile if specified.
|
|
# The side effect, if any, will be felt from this point on. An example
|
|
# config file may be these simple two lines:
|
|
#
|
|
# sys.stderr = open("/tmp/lldbtest-stderr", "w")
|
|
# sys.stdout = open("/tmp/lldbtest-stdout", "w")
|
|
#
|
|
# which will reassign the two file objects to sys.stderr and sys.stdout,
|
|
# respectively.
|
|
#
|
|
# See also lldb-trunk/examples/test/usage-config.
|
|
global config, pre_flight, post_flight, lldbtest_remote_sandbox, lldbtest_remote_shell_template
|
|
if configFile:
|
|
# Pass config (a dictionary) as the locals namespace for side-effect.
|
|
execfile(configFile, globals(), config)
|
|
#print("config:", config)
|
|
if "pre_flight" in config:
|
|
pre_flight = config["pre_flight"]
|
|
if not six.callable(pre_flight):
|
|
print("fatal error: pre_flight is not callable, exiting.")
|
|
sys.exit(1)
|
|
if "post_flight" in config:
|
|
post_flight = config["post_flight"]
|
|
if not six.callable(post_flight):
|
|
print("fatal error: post_flight is not callable, exiting.")
|
|
sys.exit(1)
|
|
if "lldbtest_remote_sandbox" in config:
|
|
lldbtest_remote_sandbox = config["lldbtest_remote_sandbox"]
|
|
if "lldbtest_remote_shell_template" in config:
|
|
lldbtest_remote_shell_template = config["lldbtest_remote_shell_template"]
|
|
#print("sys.stderr:", sys.stderr)
|
|
#print("sys.stdout:", sys.stdout)
|
|
|
|
def getXcodeOutputPaths(lldbRootDirectory):
|
|
result = []
|
|
|
|
# These are for xcode build directories.
|
|
xcode3_build_dir = ['build']
|
|
xcode4_build_dir = ['build', 'lldb', 'Build', 'Products']
|
|
|
|
configurations = [['Debug'], ['DebugClang'], ['Release'], ['BuildAndIntegration']]
|
|
xcode_build_dirs = [xcode3_build_dir, xcode4_build_dir]
|
|
for configuration in configurations:
|
|
for xcode_build_dir in xcode_build_dirs:
|
|
outputPath = os.path.join(lldbRootDirectory, *(xcode_build_dir+configuration) )
|
|
result.append(outputPath)
|
|
|
|
return result
|
|
|
|
|
|
def createSocketToLocalPort(port):
|
|
def socket_closer(s):
|
|
"""Close down an opened socket properly."""
|
|
s.shutdown(socket.SHUT_RDWR)
|
|
s.close()
|
|
|
|
sock = socket.socket(socket.AF_INET, socket.SOCK_STREAM)
|
|
sock.connect(("localhost", port))
|
|
return (sock, lambda: socket_closer(sock))
|
|
|
|
|
|
def setupTestResults():
|
|
"""Sets up test results-related objects based on arg settings."""
|
|
global results_filename
|
|
global results_file_object
|
|
global results_formatter_name
|
|
global results_formatter_object
|
|
global results_formatter_options
|
|
global results_port
|
|
|
|
default_formatter_name = None
|
|
cleanup_func = None
|
|
|
|
if results_filename:
|
|
# Open the results file for writing.
|
|
if results_filename == 'stdout':
|
|
results_file_object = sys.stdout
|
|
cleanup_func = None
|
|
elif results_filename == 'stderr':
|
|
results_file_object = sys.stderr
|
|
cleanup_func = None
|
|
else:
|
|
results_file_object = open(results_filename, "w")
|
|
cleanup_func = results_file_object.close
|
|
default_formatter_name = "test_results.XunitFormatter"
|
|
elif results_port:
|
|
# Connect to the specified localhost port.
|
|
results_file_object, cleanup_func = createSocketToLocalPort(
|
|
results_port)
|
|
default_formatter_name = "test_results.RawPickledFormatter"
|
|
|
|
if results_file_object:
|
|
# We care about the formatter. Choose user-specified or, if
|
|
# none specified, use the default for the output type.
|
|
if results_formatter_name:
|
|
formatter_name = results_formatter_name
|
|
else:
|
|
formatter_name = default_formatter_name
|
|
|
|
# Create an instance of the class. First figure out the package/module.
|
|
components = formatter_name.split(".")
|
|
module = importlib.import_module(".".join(components[:-1]))
|
|
|
|
# Create the class name we need to load.
|
|
clazz = getattr(module, components[-1])
|
|
|
|
# Handle formatter options for the results formatter class.
|
|
formatter_arg_parser = clazz.arg_parser()
|
|
if results_formatter_options and len(results_formatter_options) > 0:
|
|
command_line_options = results_formatter_options
|
|
else:
|
|
command_line_options = []
|
|
|
|
formatter_options = formatter_arg_parser.parse_args(
|
|
command_line_options)
|
|
|
|
# Create the TestResultsFormatter given the processed options.
|
|
results_formatter_object = clazz(results_file_object, formatter_options)
|
|
|
|
# Start the results formatter session - we'll only have one
|
|
# during a given dotest process invocation.
|
|
initialize_event = EventBuilder.bare_event("initialize")
|
|
if isMultiprocessTestRunner():
|
|
if test_runner_name is not None and test_runner_name == "serial":
|
|
# Only one worker queue here.
|
|
worker_count = 1
|
|
else:
|
|
# Workers will be the number of threads specified.
|
|
worker_count = num_threads
|
|
else:
|
|
worker_count = 1
|
|
initialize_event["worker_count"] = worker_count
|
|
|
|
results_formatter_object.handle_event(initialize_event)
|
|
|
|
def shutdown_formatter():
|
|
# Tell the formatter to write out anything it may have
|
|
# been saving until the very end (e.g. xUnit results
|
|
# can't complete its output until this point).
|
|
results_formatter_object.send_terminate_as_needed()
|
|
|
|
# And now close out the output file-like object.
|
|
if cleanup_func is not None:
|
|
cleanup_func()
|
|
|
|
atexit.register(shutdown_formatter)
|
|
|
|
|
|
def getOutputPaths(lldbRootDirectory):
|
|
"""
|
|
Returns typical build output paths for the lldb executable
|
|
|
|
lldbDirectory - path to the root of the lldb svn/git repo
|
|
"""
|
|
result = []
|
|
|
|
if sys.platform == 'darwin':
|
|
result.extend(getXcodeOutputPaths(lldbRootDirectory))
|
|
|
|
# cmake builds? look for build or build/host folder next to llvm directory
|
|
# lldb is located in llvm/tools/lldb so we need to go up three levels
|
|
llvmParentDir = os.path.abspath(os.path.join(lldbRootDirectory, os.pardir, os.pardir, os.pardir))
|
|
result.append(os.path.join(llvmParentDir, 'build', 'bin'))
|
|
result.append(os.path.join(llvmParentDir, 'build', 'host', 'bin'))
|
|
|
|
# some cmake developers keep their build directory beside their lldb directory
|
|
lldbParentDir = os.path.abspath(os.path.join(lldbRootDirectory, os.pardir))
|
|
result.append(os.path.join(lldbParentDir, 'build', 'bin'))
|
|
result.append(os.path.join(lldbParentDir, 'build', 'host', 'bin'))
|
|
|
|
return result
|
|
|
|
def setupSysPath():
|
|
"""
|
|
Add LLDB.framework/Resources/Python to the search paths for modules.
|
|
As a side effect, we also discover the 'lldb' executable and export it here.
|
|
"""
|
|
|
|
global rdir
|
|
global testdirs
|
|
global dumpSysPath
|
|
global noHeaders
|
|
global svn_info
|
|
global svn_silent
|
|
global lldbFrameworkPath
|
|
|
|
# Get the directory containing the current script.
|
|
if "DOTEST_PROFILE" in os.environ and "DOTEST_SCRIPT_DIR" in os.environ:
|
|
scriptPath = os.environ["DOTEST_SCRIPT_DIR"]
|
|
else:
|
|
scriptPath = os.path.dirname(os.path.realpath(__file__))
|
|
if not scriptPath.endswith('test'):
|
|
print("This script expects to reside in lldb's test directory.")
|
|
sys.exit(-1)
|
|
|
|
if rdir:
|
|
# Set up the LLDB_TEST environment variable appropriately, so that the
|
|
# individual tests can be located relatively.
|
|
#
|
|
# See also lldbtest.TestBase.setUpClass(cls).
|
|
if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test':
|
|
os.environ["LLDB_TEST"] = os.path.join(rdir, 'test')
|
|
else:
|
|
os.environ["LLDB_TEST"] = rdir
|
|
else:
|
|
os.environ["LLDB_TEST"] = scriptPath
|
|
|
|
# Set up the LLDB_SRC environment variable, so that the tests can locate
|
|
# the LLDB source code.
|
|
os.environ["LLDB_SRC"] = os.path.join(scriptPath, os.pardir)
|
|
|
|
pluginPath = os.path.join(scriptPath, 'plugins')
|
|
toolsLLDBMIPath = os.path.join(scriptPath, 'tools', 'lldb-mi')
|
|
toolsLLDBServerPath = os.path.join(scriptPath, 'tools', 'lldb-server')
|
|
|
|
# Insert script dir, plugin dir, lldb-mi dir and lldb-server dir to the sys.path.
|
|
sys.path.insert(0, scriptPath)
|
|
sys.path.insert(0, pluginPath)
|
|
sys.path.insert(0, toolsLLDBMIPath) # Adding test/tools/lldb-mi to the path makes it easy
|
|
# to "import lldbmi_testcase" from the MI tests
|
|
sys.path.insert(0, toolsLLDBServerPath) # Adding test/tools/lldb-server to the path makes it easy
|
|
# to "import lldbgdbserverutils" from the lldb-server tests
|
|
|
|
# This is the root of the lldb git/svn checkout
|
|
lldbRootDirectory = os.path.abspath(os.path.join(scriptPath, os.pardir))
|
|
|
|
# Some of the tests can invoke the 'lldb' command directly.
|
|
# We'll try to locate the appropriate executable right here.
|
|
|
|
lldbMiExec = None
|
|
|
|
# The lldb executable can be set from the command line
|
|
# if it's not set, we try to find it now
|
|
# first, we try the environment
|
|
if not lldbtest_config.lldbExec:
|
|
# First, you can define an environment variable LLDB_EXEC specifying the
|
|
# full pathname of the lldb executable.
|
|
if "LLDB_EXEC" in os.environ:
|
|
lldbtest_config.lldbExec = os.environ["LLDB_EXEC"]
|
|
|
|
if not lldbtest_config.lldbExec:
|
|
outputPaths = getOutputPaths(lldbRootDirectory)
|
|
for outputPath in outputPaths:
|
|
candidatePath = os.path.join(outputPath, 'lldb')
|
|
if is_exe(candidatePath):
|
|
lldbtest_config.lldbExec = candidatePath
|
|
break
|
|
|
|
if not lldbtest_config.lldbExec:
|
|
# Last, check the path
|
|
lldbtest_config.lldbExec = which('lldb')
|
|
|
|
if lldbtest_config.lldbExec and not is_exe(lldbtest_config.lldbExec):
|
|
print("'{}' is not a path to a valid executable".format(lldbtest_config.lldbExec))
|
|
lldbtest_config.lldbExec = None
|
|
|
|
if not lldbtest_config.lldbExec:
|
|
print("The 'lldb' executable cannot be located. Some of the tests may not be run as a result.")
|
|
sys.exit(-1)
|
|
|
|
lldbLibDir = os.path.dirname(lldbtest_config.lldbExec) # confusingly, this is the "bin" directory
|
|
os.environ["LLDB_LIB_DIR"] = lldbLibDir
|
|
lldbImpLibDir = os.path.join(lldbLibDir, '..', 'lib') if sys.platform.startswith('win32') else lldbLibDir
|
|
os.environ["LLDB_IMPLIB_DIR"] = lldbImpLibDir
|
|
if not noHeaders:
|
|
print("LLDB library dir:", os.environ["LLDB_LIB_DIR"])
|
|
print("LLDB import library dir:", os.environ["LLDB_IMPLIB_DIR"])
|
|
os.system('%s -v' % lldbtest_config.lldbExec)
|
|
|
|
# Assume lldb-mi is in same place as lldb
|
|
# If not found, disable the lldb-mi tests
|
|
global dont_do_lldbmi_test
|
|
if lldbtest_config.lldbExec and is_exe(lldbtest_config.lldbExec + "-mi"):
|
|
lldbMiExec = lldbtest_config.lldbExec + "-mi"
|
|
if not lldbMiExec:
|
|
dont_do_lldbmi_test = True
|
|
if just_do_lldbmi_test:
|
|
print("The 'lldb-mi' executable cannot be located. The lldb-mi tests can not be run as a result.")
|
|
else:
|
|
print("The 'lldb-mi' executable cannot be located. Some of the tests may not be run as a result.")
|
|
else:
|
|
os.environ["LLDBMI_EXEC"] = lldbMiExec
|
|
|
|
# Skip printing svn/git information when running in parsable (lit-test compatibility) mode
|
|
if not svn_silent and not parsable:
|
|
if os.path.isdir(os.path.join(lldbRootDirectory, '.svn')) and which("svn") is not None:
|
|
pipe = subprocess.Popen([which("svn"), "info", lldbRootDirectory], stdout = subprocess.PIPE)
|
|
svn_info = pipe.stdout.read()
|
|
elif os.path.isdir(os.path.join(lldbRootDirectory, '.git')) and which("git") is not None:
|
|
pipe = subprocess.Popen([which("git"), "svn", "info", lldbRootDirectory], stdout = subprocess.PIPE)
|
|
svn_info = pipe.stdout.read()
|
|
if not noHeaders:
|
|
print(svn_info)
|
|
|
|
global ignore
|
|
|
|
lldbPythonDir = None # The directory that contains 'lldb/__init__.py'
|
|
if lldbFrameworkPath:
|
|
candidatePath = os.path.join(lldbFrameworkPath, 'Resources', 'Python')
|
|
if os.path.isfile(os.path.join(candidatePath, 'lldb/__init__.py')):
|
|
lldbPythonDir = candidatePath
|
|
if not lldbPythonDir:
|
|
print('Resources/Python/lldb/__init__.py was not found in ' + lldbFrameworkPath)
|
|
sys.exit(-1)
|
|
else:
|
|
# The '-i' option is used to skip looking for lldb.py in the build tree.
|
|
if ignore:
|
|
return
|
|
|
|
# If our lldb supports the -P option, use it to find the python path:
|
|
init_in_python_dir = os.path.join('lldb', '__init__.py')
|
|
|
|
lldb_dash_p_result = subprocess.check_output([lldbtest_config.lldbExec, "-P"], stderr=subprocess.STDOUT)
|
|
|
|
if lldb_dash_p_result and not lldb_dash_p_result.startswith(("<", "lldb: invalid option:")) \
|
|
and not lldb_dash_p_result.startswith("Traceback"):
|
|
lines = lldb_dash_p_result.splitlines()
|
|
|
|
# Workaround for readline vs libedit issue on FreeBSD. If stdout
|
|
# is not a terminal Python executes
|
|
# rl_variable_bind ("enable-meta-key", "off");
|
|
# This produces a warning with FreeBSD's libedit because the
|
|
# enable-meta-key variable is unknown. Not an issue on Apple
|
|
# because cpython commit f0ab6f9f0603 added a #ifndef __APPLE__
|
|
# around the call. See http://bugs.python.org/issue19884 for more
|
|
# information. For now we just discard the warning output.
|
|
if len(lines) >= 1 and lines[0].startswith("bind: Invalid command"):
|
|
lines.pop(0)
|
|
|
|
# Taking the last line because lldb outputs
|
|
# 'Cannot read termcap database;\nusing dumb terminal settings.\n'
|
|
# before the path
|
|
if len(lines) >= 1 and os.path.isfile(os.path.join(lines[-1], init_in_python_dir)):
|
|
lldbPythonDir = lines[-1]
|
|
if "freebsd" in sys.platform or "linux" in sys.platform:
|
|
os.environ['LLDB_LIB_DIR'] = os.path.join(lldbPythonDir, '..', '..')
|
|
|
|
if not lldbPythonDir:
|
|
if platform.system() == "Darwin":
|
|
python_resource_dir = ['LLDB.framework', 'Resources', 'Python']
|
|
outputPaths = getXcodeOutputPaths()
|
|
for outputPath in outputPaths:
|
|
candidatePath = os.path.join(outputPath, python_resource_dir)
|
|
if os.path.isfile(os.path.join(candidatePath, init_in_python_dir)):
|
|
lldbPythonDir = candidatePath
|
|
break
|
|
|
|
if not lldbPythonDir:
|
|
print('This script requires lldb.py to be in either ' + dbgPath + ',', end=' ')
|
|
print(relPath + ', or ' + baiPath + '. Some tests might fail.')
|
|
else:
|
|
print("Unable to load lldb extension module. Possible reasons for this include:")
|
|
print(" 1) LLDB was built with LLDB_DISABLE_PYTHON=1")
|
|
print(" 2) PYTHONPATH and PYTHONHOME are not set correctly. PYTHONHOME should refer to")
|
|
print(" the version of Python that LLDB built and linked against, and PYTHONPATH")
|
|
print(" should contain the Lib directory for the same python distro, as well as the")
|
|
print(" location of LLDB\'s site-packages folder.")
|
|
print(" 3) A different version of Python than that which was built against is exported in")
|
|
print(" the system\'s PATH environment variable, causing conflicts.")
|
|
print(" 4) The executable '%s' could not be found. Please check " % lldbExecutable)
|
|
print(" that it exists and is executable.")
|
|
|
|
if lldbPythonDir:
|
|
lldbPythonDir = os.path.normpath(lldbPythonDir)
|
|
# Some of the code that uses this path assumes it hasn't resolved the Versions... link.
|
|
# If the path we've constructed looks like that, then we'll strip out the Versions/A part.
|
|
(before, frameWithVersion, after) = lldbPythonDir.rpartition("LLDB.framework/Versions/A")
|
|
if frameWithVersion != "" :
|
|
lldbPythonDir = before + "LLDB.framework" + after
|
|
|
|
lldbPythonDir = os.path.abspath(lldbPythonDir)
|
|
|
|
# If tests need to find LLDB_FRAMEWORK, now they can do it
|
|
os.environ["LLDB_FRAMEWORK"] = os.path.dirname(os.path.dirname(lldbPythonDir))
|
|
|
|
# This is to locate the lldb.py module. Insert it right after sys.path[0].
|
|
sys.path[1:1] = [lldbPythonDir]
|
|
if dumpSysPath:
|
|
print("sys.path:", sys.path)
|
|
|
|
def visit(prefix, dir, names):
|
|
"""Visitor function for os.path.walk(path, visit, arg)."""
|
|
|
|
global suite
|
|
global regexp
|
|
global filters
|
|
global fs4all
|
|
global excluded
|
|
global all_tests
|
|
|
|
if set(dir.split(os.sep)).intersection(excluded):
|
|
#print("Detected an excluded dir component: %s" % dir)
|
|
return
|
|
|
|
for name in names:
|
|
if os.path.isdir(os.path.join(dir, name)):
|
|
continue
|
|
|
|
if '.py' == os.path.splitext(name)[1] and name.startswith(prefix):
|
|
|
|
if name in all_tests:
|
|
raise Exception("Found multiple tests with the name %s" % name)
|
|
all_tests.add(name)
|
|
|
|
# Try to match the regexp pattern, if specified.
|
|
if regexp:
|
|
import re
|
|
if re.search(regexp, name):
|
|
#print("Filename: '%s' matches pattern: '%s'" % (name, regexp))
|
|
pass
|
|
else:
|
|
#print("Filename: '%s' does not match pattern: '%s'" % (name, regexp))
|
|
continue
|
|
|
|
# We found a match for our test. Add it to the suite.
|
|
|
|
# Update the sys.path first.
|
|
if not sys.path.count(dir):
|
|
sys.path.insert(0, dir)
|
|
base = os.path.splitext(name)[0]
|
|
|
|
# Thoroughly check the filterspec against the base module and admit
|
|
# the (base, filterspec) combination only when it makes sense.
|
|
filterspec = None
|
|
for filterspec in filters:
|
|
# Optimistically set the flag to True.
|
|
filtered = True
|
|
module = __import__(base)
|
|
parts = filterspec.split('.')
|
|
obj = module
|
|
for part in parts:
|
|
try:
|
|
parent, obj = obj, getattr(obj, part)
|
|
except AttributeError:
|
|
# The filterspec has failed.
|
|
filtered = False
|
|
break
|
|
|
|
# If filtered, we have a good filterspec. Add it.
|
|
if filtered:
|
|
#print("adding filter spec %s to module %s" % (filterspec, module))
|
|
suite.addTests(
|
|
unittest2.defaultTestLoader.loadTestsFromName(filterspec, module))
|
|
continue
|
|
|
|
# Forgo this module if the (base, filterspec) combo is invalid
|
|
# and no '-g' option is specified
|
|
if filters and fs4all and not filtered:
|
|
continue
|
|
|
|
# Add either the filtered test case(s) (which is done before) or the entire test class.
|
|
if not filterspec or not filtered:
|
|
# A simple case of just the module name. Also the failover case
|
|
# from the filterspec branch when the (base, filterspec) combo
|
|
# doesn't make sense.
|
|
suite.addTests(unittest2.defaultTestLoader.loadTestsFromName(base))
|
|
|
|
|
|
def disabledynamics():
|
|
ci = lldb.DBG.GetCommandInterpreter()
|
|
res = lldb.SBCommandReturnObject()
|
|
ci.HandleCommand("setting set target.prefer-dynamic-value no-dynamic-values", res, False)
|
|
if not res.Succeeded():
|
|
raise Exception('disabling dynamic type support failed')
|
|
|
|
def lldbLoggings():
|
|
"""Check and do lldb loggings if necessary."""
|
|
|
|
# Turn on logging for debugging purposes if ${LLDB_LOG} environment variable is
|
|
# defined. Use ${LLDB_LOG} to specify the log file.
|
|
ci = lldb.DBG.GetCommandInterpreter()
|
|
res = lldb.SBCommandReturnObject()
|
|
if ("LLDB_LOG" in os.environ):
|
|
open(os.environ["LLDB_LOG"], 'w').close()
|
|
if ("LLDB_LOG_OPTION" in os.environ):
|
|
lldb_log_option = os.environ["LLDB_LOG_OPTION"]
|
|
else:
|
|
lldb_log_option = "event process expr state api"
|
|
ci.HandleCommand(
|
|
"log enable -n -f " + os.environ["LLDB_LOG"] + " lldb " + lldb_log_option,
|
|
res)
|
|
if not res.Succeeded():
|
|
raise Exception('log enable failed (check LLDB_LOG env variable)')
|
|
|
|
if ("LLDB_LINUX_LOG" in os.environ):
|
|
open(os.environ["LLDB_LINUX_LOG"], 'w').close()
|
|
if ("LLDB_LINUX_LOG_OPTION" in os.environ):
|
|
lldb_log_option = os.environ["LLDB_LINUX_LOG_OPTION"]
|
|
else:
|
|
lldb_log_option = "event process expr state api"
|
|
ci.HandleCommand(
|
|
"log enable -n -f " + os.environ["LLDB_LINUX_LOG"] + " linux " + lldb_log_option,
|
|
res)
|
|
if not res.Succeeded():
|
|
raise Exception('log enable failed (check LLDB_LINUX_LOG env variable)')
|
|
|
|
# Ditto for gdb-remote logging if ${GDB_REMOTE_LOG} environment variable is defined.
|
|
# Use ${GDB_REMOTE_LOG} to specify the log file.
|
|
if ("GDB_REMOTE_LOG" in os.environ):
|
|
if ("GDB_REMOTE_LOG_OPTION" in os.environ):
|
|
gdb_remote_log_option = os.environ["GDB_REMOTE_LOG_OPTION"]
|
|
else:
|
|
gdb_remote_log_option = "packets process"
|
|
ci.HandleCommand(
|
|
"log enable -n -f " + os.environ["GDB_REMOTE_LOG"] + " gdb-remote "
|
|
+ gdb_remote_log_option,
|
|
res)
|
|
if not res.Succeeded():
|
|
raise Exception('log enable failed (check GDB_REMOTE_LOG env variable)')
|
|
|
|
def getMyCommandLine():
|
|
return ' '.join(sys.argv)
|
|
|
|
# ======================================== #
|
|
# #
|
|
# Execution of the test driver starts here #
|
|
# #
|
|
# ======================================== #
|
|
|
|
def checkDsymForUUIDIsNotOn():
|
|
cmd = ["defaults", "read", "com.apple.DebugSymbols"]
|
|
pipe = subprocess.Popen(cmd, stdout = subprocess.PIPE, stderr = subprocess.STDOUT)
|
|
cmd_output = pipe.stdout.read()
|
|
if cmd_output and "DBGFileMappedPaths = " in cmd_output:
|
|
print("%s =>" % ' '.join(cmd))
|
|
print(cmd_output)
|
|
print("Disable automatic lookup and caching of dSYMs before running the test suite!")
|
|
print("Exiting...")
|
|
sys.exit(0)
|
|
|
|
def exitTestSuite(exitCode = None):
|
|
lldb.SBDebugger.Terminate()
|
|
if exitCode:
|
|
sys.exit(exitCode)
|
|
|
|
|
|
def isMultiprocessTestRunner():
|
|
# We're not multiprocess when we're either explicitly
|
|
# the inferior (as specified by the multiprocess test
|
|
# runner) OR we've been told to skip using the multiprocess
|
|
# test runner
|
|
return not (is_inferior_test_runner or no_multiprocess_test_runner)
|
|
|
|
if __name__ == "__main__":
|
|
# On MacOS X, check to make sure that domain for com.apple.DebugSymbols defaults
|
|
# does not exist before proceeding to running the test suite.
|
|
if sys.platform.startswith("darwin"):
|
|
checkDsymForUUIDIsNotOn()
|
|
|
|
#
|
|
# Start the actions by first parsing the options while setting up the test
|
|
# directories, followed by setting up the search paths for lldb utilities;
|
|
# then, we walk the directory trees and collect the tests into our test suite.
|
|
#
|
|
parseOptionsAndInitTestdirs()
|
|
|
|
# Setup test results (test results formatter and output handling).
|
|
setupTestResults()
|
|
|
|
# If we are running as the multiprocess test runner, kick off the
|
|
# multiprocess test runner here.
|
|
if isMultiprocessTestRunner():
|
|
import dosep
|
|
dosep.main(output_on_success, num_threads, multiprocess_test_subdir,
|
|
test_runner_name, results_formatter_object)
|
|
raise Exception("should never get here")
|
|
elif is_inferior_test_runner:
|
|
# Shut off Ctrl-C processing in inferiors. The parallel
|
|
# test runner handles this more holistically.
|
|
signal.signal(signal.SIGINT, signal.SIG_IGN)
|
|
|
|
setupSysPath()
|
|
setupCrashInfoHook()
|
|
|
|
#
|
|
# If '-l' is specified, do not skip the long running tests.
|
|
if not skip_long_running_test:
|
|
os.environ["LLDB_SKIP_LONG_RUNNING_TEST"] = "NO"
|
|
|
|
# For the time being, let's bracket the test runner within the
|
|
# lldb.SBDebugger.Initialize()/Terminate() pair.
|
|
import lldb
|
|
|
|
# Create a singleton SBDebugger in the lldb namespace.
|
|
lldb.DBG = lldb.SBDebugger.Create()
|
|
|
|
if lldb_platform_name:
|
|
print("Setting up remote platform '%s'" % (lldb_platform_name))
|
|
lldb.remote_platform = lldb.SBPlatform(lldb_platform_name)
|
|
if not lldb.remote_platform.IsValid():
|
|
print("error: unable to create the LLDB platform named '%s'." % (lldb_platform_name))
|
|
exitTestSuite(1)
|
|
if lldb_platform_url:
|
|
# We must connect to a remote platform if a LLDB platform URL was specified
|
|
print("Connecting to remote platform '%s' at '%s'..." % (lldb_platform_name, lldb_platform_url))
|
|
lldb.platform_url = lldb_platform_url
|
|
platform_connect_options = lldb.SBPlatformConnectOptions(lldb_platform_url)
|
|
err = lldb.remote_platform.ConnectRemote(platform_connect_options)
|
|
if err.Success():
|
|
print("Connected.")
|
|
else:
|
|
print("error: failed to connect to remote platform using URL '%s': %s" % (lldb_platform_url, err))
|
|
exitTestSuite(1)
|
|
else:
|
|
lldb.platform_url = None
|
|
|
|
if lldb_platform_working_dir:
|
|
print("Setting remote platform working directory to '%s'..." % (lldb_platform_working_dir))
|
|
lldb.remote_platform.SetWorkingDirectory(lldb_platform_working_dir)
|
|
|
|
lldb.remote_platform_working_dir = lldb_platform_working_dir
|
|
lldb.DBG.SetSelectedPlatform(lldb.remote_platform)
|
|
else:
|
|
lldb.remote_platform = None
|
|
lldb.remote_platform_working_dir = None
|
|
lldb.platform_url = None
|
|
|
|
target_platform = lldb.DBG.GetSelectedPlatform().GetTriple().split('-')[2]
|
|
|
|
# By default, both dsym, dwarf and dwo tests are performed.
|
|
# Use @dsym_test, @dwarf_test or @dwo_test decorators, defined in lldbtest.py, to mark a test as
|
|
# a dsym, dwarf or dwo test. Use '-N dsym', '-N dwarf' or '-N dwo' to exclude dsym, dwarf or
|
|
# dwo tests from running.
|
|
dont_do_dsym_test = dont_do_dsym_test or any(platform in target_platform for platform in ["linux", "freebsd", "windows"])
|
|
dont_do_dwo_test = dont_do_dwo_test or any(platform in target_platform for platform in ["darwin", "macosx", "ios"])
|
|
|
|
# Don't do debugserver tests on everything except OS X.
|
|
dont_do_debugserver_test = "linux" in target_platform or "freebsd" in target_platform or "windows" in target_platform
|
|
|
|
# Don't do lldb-server (llgs) tests on anything except Linux.
|
|
dont_do_llgs_test = not ("linux" in target_platform)
|
|
|
|
#
|
|
# Walk through the testdirs while collecting tests.
|
|
#
|
|
for testdir in testdirs:
|
|
os.path.walk(testdir, visit, 'Test')
|
|
|
|
#
|
|
# Now that we have loaded all the test cases, run the whole test suite.
|
|
#
|
|
|
|
# Put the blacklist in the lldb namespace, to be used by lldb.TestBase.
|
|
lldb.blacklist = blacklist
|
|
|
|
# The pre_flight and post_flight come from reading a config file.
|
|
lldb.pre_flight = pre_flight
|
|
lldb.post_flight = post_flight
|
|
def getsource_if_available(obj):
|
|
"""
|
|
Return the text of the source code for an object if available. Otherwise,
|
|
a print representation is returned.
|
|
"""
|
|
import inspect
|
|
try:
|
|
return inspect.getsource(obj)
|
|
except:
|
|
return repr(obj)
|
|
|
|
if not noHeaders:
|
|
print("lldb.pre_flight:", getsource_if_available(lldb.pre_flight))
|
|
print("lldb.post_flight:", getsource_if_available(lldb.post_flight))
|
|
|
|
# If either pre_flight or post_flight is defined, set lldb.test_remote to True.
|
|
if lldb.pre_flight or lldb.post_flight:
|
|
lldb.test_remote = True
|
|
else:
|
|
lldb.test_remote = False
|
|
|
|
# So do the lldbtest_remote_sandbox and lldbtest_remote_shell_template variables.
|
|
lldb.lldbtest_remote_sandbox = lldbtest_remote_sandbox
|
|
lldb.lldbtest_remote_sandboxed_executable = None
|
|
lldb.lldbtest_remote_shell_template = lldbtest_remote_shell_template
|
|
|
|
# Put all these test decorators in the lldb namespace.
|
|
lldb.dont_do_lldbmi_test = dont_do_lldbmi_test
|
|
lldb.just_do_lldbmi_test = just_do_lldbmi_test
|
|
lldb.just_do_benchmarks_test = just_do_benchmarks_test
|
|
lldb.dont_do_dsym_test = dont_do_dsym_test
|
|
lldb.dont_do_dwarf_test = dont_do_dwarf_test
|
|
lldb.dont_do_dwo_test = dont_do_dwo_test
|
|
lldb.dont_do_debugserver_test = dont_do_debugserver_test
|
|
lldb.dont_do_llgs_test = dont_do_llgs_test
|
|
|
|
# Do we need to skip build and cleanup?
|
|
lldb.skip_build_and_cleanup = skip_build_and_cleanup
|
|
|
|
# Put bmExecutable, bmBreakpointSpec, and bmIterationCount into the lldb namespace, too.
|
|
lldb.bmExecutable = bmExecutable
|
|
lldb.bmBreakpointSpec = bmBreakpointSpec
|
|
lldb.bmIterationCount = bmIterationCount
|
|
|
|
# And don't forget the runHooks!
|
|
lldb.runHooks = runHooks
|
|
|
|
# Turn on lldb loggings if necessary.
|
|
lldbLoggings()
|
|
|
|
# Disable default dynamic types for testing purposes
|
|
disabledynamics()
|
|
|
|
# Install the control-c handler.
|
|
unittest2.signals.installHandler()
|
|
|
|
# If sdir_name is not specified through the '-s sdir_name' option, get a
|
|
# timestamp string and export it as LLDB_SESSION_DIR environment var. This will
|
|
# be used when/if we want to dump the session info of individual test cases
|
|
# later on.
|
|
#
|
|
# See also TestBase.dumpSessionInfo() in lldbtest.py.
|
|
import datetime
|
|
# The windows platforms don't like ':' in the pathname.
|
|
timestamp_started = datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S")
|
|
if not sdir_name:
|
|
sdir_name = timestamp_started
|
|
os.environ["LLDB_SESSION_DIRNAME"] = os.path.join(os.getcwd(), sdir_name)
|
|
|
|
if not noHeaders:
|
|
sys.stderr.write("\nSession logs for test failures/errors/unexpected successes"
|
|
" will go into directory '%s'\n" % sdir_name)
|
|
sys.stderr.write("Command invoked: %s\n" % getMyCommandLine())
|
|
|
|
if not os.path.isdir(sdir_name):
|
|
try:
|
|
os.mkdir(sdir_name)
|
|
except OSError as exception:
|
|
if exception.errno != errno.EEXIST:
|
|
raise
|
|
where_to_save_session = os.getcwd()
|
|
fname = os.path.join(sdir_name, "TestStarted-%d" % os.getpid())
|
|
with open(fname, "w") as f:
|
|
print("Test started at: %s\n" % timestamp_started, file=f)
|
|
print(svn_info, file=f)
|
|
print("Command invoked: %s\n" % getMyCommandLine(), file=f)
|
|
|
|
#
|
|
# Invoke the default TextTestRunner to run the test suite, possibly iterating
|
|
# over different configurations.
|
|
#
|
|
|
|
iterArchs = False
|
|
iterCompilers = False
|
|
|
|
if not archs and "archs" in config:
|
|
archs = config["archs"]
|
|
|
|
if isinstance(archs, list) and len(archs) >= 1:
|
|
iterArchs = True
|
|
|
|
if not compilers and "compilers" in config:
|
|
compilers = config["compilers"]
|
|
|
|
#
|
|
# Add some intervention here to sanity check that the compilers requested are sane.
|
|
# If found not to be an executable program, the invalid one is dropped from the list.
|
|
for i in range(len(compilers)):
|
|
c = compilers[i]
|
|
if which(c):
|
|
continue
|
|
else:
|
|
if sys.platform.startswith("darwin"):
|
|
pipe = subprocess.Popen(['xcrun', '-find', c], stdout = subprocess.PIPE, stderr = subprocess.STDOUT)
|
|
cmd_output = pipe.stdout.read()
|
|
if cmd_output:
|
|
if "not found" in cmd_output:
|
|
print("dropping %s from the compilers used" % c)
|
|
compilers.remove(i)
|
|
else:
|
|
compilers[i] = cmd_output.split('\n')[0]
|
|
print("'xcrun -find %s' returning %s" % (c, compilers[i]))
|
|
|
|
if not parsable:
|
|
print("compilers=%s" % str(compilers))
|
|
|
|
if not compilers or len(compilers) == 0:
|
|
print("No eligible compiler found, exiting.")
|
|
exitTestSuite(1)
|
|
|
|
if isinstance(compilers, list) and len(compilers) >= 1:
|
|
iterCompilers = True
|
|
|
|
# Make a shallow copy of sys.path, we need to manipulate the search paths later.
|
|
# This is only necessary if we are relocated and with different configurations.
|
|
if rdir:
|
|
old_sys_path = sys.path[:]
|
|
# If we iterate on archs or compilers, there is a chance we want to split stderr/stdout.
|
|
if iterArchs or iterCompilers:
|
|
old_stderr = sys.stderr
|
|
old_stdout = sys.stdout
|
|
new_stderr = None
|
|
new_stdout = None
|
|
|
|
# Iterating over all possible architecture and compiler combinations.
|
|
for ia in range(len(archs) if iterArchs else 1):
|
|
archConfig = ""
|
|
if iterArchs:
|
|
os.environ["ARCH"] = archs[ia]
|
|
archConfig = "arch=%s" % archs[ia]
|
|
for ic in range(len(compilers) if iterCompilers else 1):
|
|
if iterCompilers:
|
|
os.environ["CC"] = compilers[ic]
|
|
configString = "%s compiler=%s" % (archConfig, compilers[ic])
|
|
else:
|
|
configString = archConfig
|
|
|
|
if iterArchs or iterCompilers:
|
|
# Translate ' ' to '-' for pathname component.
|
|
from string import maketrans
|
|
tbl = maketrans(' ', '-')
|
|
configPostfix = configString.translate(tbl)
|
|
|
|
# Check whether we need to split stderr/stdout into configuration
|
|
# specific files.
|
|
if old_stderr.name != '<stderr>' and config.get('split_stderr'):
|
|
if new_stderr:
|
|
new_stderr.close()
|
|
new_stderr = open("%s.%s" % (old_stderr.name, configPostfix), "w")
|
|
sys.stderr = new_stderr
|
|
if old_stdout.name != '<stdout>' and config.get('split_stdout'):
|
|
if new_stdout:
|
|
new_stdout.close()
|
|
new_stdout = open("%s.%s" % (old_stdout.name, configPostfix), "w")
|
|
sys.stdout = new_stdout
|
|
|
|
# If we specified a relocated directory to run the test suite, do
|
|
# the extra housekeeping to copy the testdirs to a configStringified
|
|
# directory and to update sys.path before invoking the test runner.
|
|
# The purpose is to separate the configuration-specific directories
|
|
# from each other.
|
|
if rdir:
|
|
from shutil import copytree, rmtree, ignore_patterns
|
|
|
|
newrdir = "%s.%s" % (rdir, configPostfix)
|
|
|
|
# Copy the tree to a new directory with postfix name configPostfix.
|
|
if os.path.exists(newrdir):
|
|
rmtree(newrdir)
|
|
copytree(rdir, newrdir, ignore=ignore_patterns('*.pyc', '*.o', '*.d'))
|
|
|
|
# Update the LLDB_TEST environment variable to reflect new top
|
|
# level test directory.
|
|
#
|
|
# See also lldbtest.TestBase.setUpClass(cls).
|
|
if len(testdirs) == 1 and os.path.basename(testdirs[0]) == 'test':
|
|
os.environ["LLDB_TEST"] = os.path.join(newrdir, 'test')
|
|
else:
|
|
os.environ["LLDB_TEST"] = newrdir
|
|
|
|
# And update the Python search paths for modules.
|
|
sys.path = [x.replace(rdir, newrdir, 1) for x in old_sys_path]
|
|
|
|
# Output the configuration.
|
|
if not parsable:
|
|
sys.stderr.write("\nConfiguration: " + configString + "\n")
|
|
|
|
#print("sys.stderr name is", sys.stderr.name)
|
|
#print("sys.stdout name is", sys.stdout.name)
|
|
|
|
# First, write out the number of collected test cases.
|
|
if not parsable:
|
|
sys.stderr.write(separator + "\n")
|
|
sys.stderr.write("Collected %d test%s\n\n"
|
|
% (suite.countTestCases(),
|
|
suite.countTestCases() != 1 and "s" or ""))
|
|
|
|
class LLDBTestResult(unittest2.TextTestResult):
|
|
"""
|
|
Enforce a singleton pattern to allow introspection of test progress.
|
|
|
|
Overwrite addError(), addFailure(), and addExpectedFailure() methods
|
|
to enable each test instance to track its failure/error status. It
|
|
is used in the LLDB test framework to emit detailed trace messages
|
|
to a log file for easier human inspection of test failures/errors.
|
|
"""
|
|
__singleton__ = None
|
|
__ignore_singleton__ = False
|
|
|
|
@staticmethod
|
|
def getTerminalSize():
|
|
import os
|
|
env = os.environ
|
|
def ioctl_GWINSZ(fd):
|
|
try:
|
|
import fcntl, termios, struct, os
|
|
cr = struct.unpack('hh', fcntl.ioctl(fd, termios.TIOCGWINSZ,
|
|
'1234'))
|
|
except:
|
|
return
|
|
return cr
|
|
cr = ioctl_GWINSZ(0) or ioctl_GWINSZ(1) or ioctl_GWINSZ(2)
|
|
if not cr:
|
|
try:
|
|
fd = os.open(os.ctermid(), os.O_RDONLY)
|
|
cr = ioctl_GWINSZ(fd)
|
|
os.close(fd)
|
|
except:
|
|
pass
|
|
if not cr:
|
|
cr = (env.get('LINES', 25), env.get('COLUMNS', 80))
|
|
return int(cr[1]), int(cr[0])
|
|
|
|
def __init__(self, *args):
|
|
if not LLDBTestResult.__ignore_singleton__ and LLDBTestResult.__singleton__:
|
|
raise Exception("LLDBTestResult instantiated more than once")
|
|
super(LLDBTestResult, self).__init__(*args)
|
|
LLDBTestResult.__singleton__ = self
|
|
# Now put this singleton into the lldb module namespace.
|
|
lldb.test_result = self
|
|
# Computes the format string for displaying the counter.
|
|
global suite
|
|
counterWidth = len(str(suite.countTestCases()))
|
|
self.fmt = "%" + str(counterWidth) + "d: "
|
|
self.indentation = ' ' * (counterWidth + 2)
|
|
# This counts from 1 .. suite.countTestCases().
|
|
self.counter = 0
|
|
(width, height) = LLDBTestResult.getTerminalSize()
|
|
self.progressbar = None
|
|
global progress_bar
|
|
if width > 10 and not parsable and progress_bar:
|
|
try:
|
|
self.progressbar = progress.ProgressWithEvents(stdout=self.stream,start=0,end=suite.countTestCases(),width=width-10)
|
|
except:
|
|
self.progressbar = None
|
|
self.results_formatter = results_formatter_object
|
|
|
|
def _config_string(self, test):
|
|
compiler = getattr(test, "getCompiler", None)
|
|
arch = getattr(test, "getArchitecture", None)
|
|
return "%s-%s" % (compiler() if compiler else "", arch() if arch else "")
|
|
|
|
def _exc_info_to_string(self, err, test):
|
|
"""Overrides superclass TestResult's method in order to append
|
|
our test config info string to the exception info string."""
|
|
if hasattr(test, "getArchitecture") and hasattr(test, "getCompiler"):
|
|
return '%sConfig=%s-%s' % (super(LLDBTestResult, self)._exc_info_to_string(err, test),
|
|
test.getArchitecture(),
|
|
test.getCompiler())
|
|
else:
|
|
return super(LLDBTestResult, self)._exc_info_to_string(err, test)
|
|
|
|
def getDescription(self, test):
|
|
doc_first_line = test.shortDescription()
|
|
if self.descriptions and doc_first_line:
|
|
return '\n'.join((str(test), self.indentation + doc_first_line))
|
|
else:
|
|
return str(test)
|
|
|
|
def getCategoriesForTest(self,test):
|
|
if hasattr(test,"_testMethodName"):
|
|
test_method = getattr(test,"_testMethodName")
|
|
test_method = getattr(test,test_method)
|
|
else:
|
|
test_method = None
|
|
if test_method != None and hasattr(test_method,"getCategories"):
|
|
test_categories = test_method.getCategories(test)
|
|
elif hasattr(test,"getCategories"):
|
|
test_categories = test.getCategories()
|
|
elif inspect.ismethod(test) and test.__self__ != None and hasattr(test.__self__,"getCategories"):
|
|
test_categories = test.__self__.getCategories()
|
|
else:
|
|
test_categories = []
|
|
if test_categories == None:
|
|
test_categories = []
|
|
return test_categories
|
|
|
|
def shouldSkipBecauseOfCategories(self,test):
|
|
global useCategories
|
|
import inspect
|
|
if useCategories:
|
|
global categoriesList
|
|
test_categories = self.getCategoriesForTest(test)
|
|
if len(test_categories) == 0 or len(categoriesList & set(test_categories)) == 0:
|
|
return True
|
|
|
|
global skipCategories
|
|
for category in skipCategories:
|
|
if category in self.getCategoriesForTest(test):
|
|
return True
|
|
|
|
return False
|
|
|
|
def hardMarkAsSkipped(self,test):
|
|
getattr(test, test._testMethodName).__func__.__unittest_skip__ = True
|
|
getattr(test, test._testMethodName).__func__.__unittest_skip_why__ = "test case does not fall in any category of interest for this run"
|
|
test.__class__.__unittest_skip__ = True
|
|
test.__class__.__unittest_skip_why__ = "test case does not fall in any category of interest for this run"
|
|
|
|
def startTest(self, test):
|
|
if self.shouldSkipBecauseOfCategories(test):
|
|
self.hardMarkAsSkipped(test)
|
|
global setCrashInfoHook
|
|
setCrashInfoHook("%s at %s" % (str(test),inspect.getfile(test.__class__)))
|
|
self.counter += 1
|
|
#if self.counter == 4:
|
|
# import crashinfo
|
|
# crashinfo.testCrashReporterDescription(None)
|
|
test.test_number = self.counter
|
|
if self.showAll:
|
|
self.stream.write(self.fmt % self.counter)
|
|
super(LLDBTestResult, self).startTest(test)
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_start(test))
|
|
|
|
def addSuccess(self, test):
|
|
global parsable
|
|
super(LLDBTestResult, self).addSuccess(test)
|
|
if parsable:
|
|
self.stream.write("PASS: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_success(test))
|
|
|
|
def addError(self, test, err):
|
|
global sdir_has_content
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addError(test, err)
|
|
method = getattr(test, "markError", None)
|
|
if method:
|
|
method()
|
|
if parsable:
|
|
self.stream.write("FAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_error(test, err))
|
|
|
|
def addCleanupError(self, test, err):
|
|
global sdir_has_content
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addCleanupError(test, err)
|
|
method = getattr(test, "markCleanupError", None)
|
|
if method:
|
|
method()
|
|
if parsable:
|
|
self.stream.write("CLEANUP ERROR: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_cleanup_error(
|
|
test, err))
|
|
|
|
def addFailure(self, test, err):
|
|
global sdir_has_content
|
|
global failuresPerCategory
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addFailure(test, err)
|
|
method = getattr(test, "markFailure", None)
|
|
if method:
|
|
method()
|
|
if parsable:
|
|
self.stream.write("FAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if useCategories:
|
|
test_categories = self.getCategoriesForTest(test)
|
|
for category in test_categories:
|
|
if category in failuresPerCategory:
|
|
failuresPerCategory[category] = failuresPerCategory[category] + 1
|
|
else:
|
|
failuresPerCategory[category] = 1
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_failure(test, err))
|
|
|
|
|
|
def addExpectedFailure(self, test, err, bugnumber):
|
|
global sdir_has_content
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addExpectedFailure(test, err, bugnumber)
|
|
method = getattr(test, "markExpectedFailure", None)
|
|
if method:
|
|
method(err, bugnumber)
|
|
if parsable:
|
|
self.stream.write("XFAIL: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_expected_failure(
|
|
test, err, bugnumber))
|
|
|
|
def addSkip(self, test, reason):
|
|
global sdir_has_content
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addSkip(test, reason)
|
|
method = getattr(test, "markSkippedTest", None)
|
|
if method:
|
|
method()
|
|
if parsable:
|
|
self.stream.write("UNSUPPORTED: LLDB (%s) :: %s (%s) \n" % (self._config_string(test), str(test), reason))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_skip(test, reason))
|
|
|
|
def addUnexpectedSuccess(self, test, bugnumber):
|
|
global sdir_has_content
|
|
global parsable
|
|
sdir_has_content = True
|
|
super(LLDBTestResult, self).addUnexpectedSuccess(test, bugnumber)
|
|
method = getattr(test, "markUnexpectedSuccess", None)
|
|
if method:
|
|
method(bugnumber)
|
|
if parsable:
|
|
self.stream.write("XPASS: LLDB (%s) :: %s\n" % (self._config_string(test), str(test)))
|
|
if self.results_formatter:
|
|
self.results_formatter.handle_event(
|
|
EventBuilder.event_for_unexpected_success(
|
|
test, bugnumber))
|
|
|
|
|
|
if parsable:
|
|
v = 0
|
|
elif progress_bar:
|
|
v = 1
|
|
else:
|
|
v = verbose
|
|
|
|
# Invoke the test runner.
|
|
if count == 1:
|
|
result = unittest2.TextTestRunner(stream=sys.stderr,
|
|
verbosity=v,
|
|
failfast=failfast,
|
|
resultclass=LLDBTestResult).run(suite)
|
|
else:
|
|
# We are invoking the same test suite more than once. In this case,
|
|
# mark __ignore_singleton__ flag as True so the signleton pattern is
|
|
# not enforced.
|
|
LLDBTestResult.__ignore_singleton__ = True
|
|
for i in range(count):
|
|
|
|
result = unittest2.TextTestRunner(stream=sys.stderr,
|
|
verbosity=v,
|
|
failfast=failfast,
|
|
resultclass=LLDBTestResult).run(suite)
|
|
|
|
failed = failed or not result.wasSuccessful()
|
|
|
|
if sdir_has_content and not parsable:
|
|
sys.stderr.write("Session logs for test failures/errors/unexpected successes"
|
|
" can be found in directory '%s'\n" % sdir_name)
|
|
|
|
if useCategories and len(failuresPerCategory) > 0:
|
|
sys.stderr.write("Failures per category:\n")
|
|
for category in failuresPerCategory:
|
|
sys.stderr.write("%s - %d\n" % (category,failuresPerCategory[category]))
|
|
|
|
os.chdir(where_to_save_session)
|
|
fname = os.path.join(sdir_name, "TestFinished-%d" % os.getpid())
|
|
with open(fname, "w") as f:
|
|
print("Test finished at: %s\n" % datetime.datetime.now().strftime("%Y-%m-%d-%H_%M_%S"), file=f)
|
|
|
|
# Terminate the test suite if ${LLDB_TESTSUITE_FORCE_FINISH} is defined.
|
|
# This should not be necessary now.
|
|
if ("LLDB_TESTSUITE_FORCE_FINISH" in os.environ):
|
|
print("Terminating Test suite...")
|
|
subprocess.Popen(["/bin/sh", "-c", "kill %s; exit 0" % (os.getpid())])
|
|
|
|
# Exiting.
|
|
exitTestSuite(failed)
|