Add bench.py as a driver script to run some benchmarks on lldb.

Add benchmarks for expression evaluations (TestExpressionCmd.py) and disassembly (TestDoAttachThenDisassembly.py).

An example:
[17:45:55] johnny:/Volumes/data/lldb/svn/trunk/test $ ./bench.py 2>&1 | grep -P '^lldb.*benchmark:'
lldb startup delay (create fresh target) benchmark: Avg: 0.104274 (Laps: 30, Total Elapsed Time: 3.128214)
lldb startup delay (set first breakpoint) benchmark: Avg: 0.102216 (Laps: 30, Total Elapsed Time: 3.066470)
lldb frame variable benchmark: Avg: 1.649162 (Laps: 20, Total Elapsed Time: 32.983245)
lldb stepping benchmark: Avg: 0.104409 (Laps: 50, Total Elapsed Time: 5.220461)
lldb expr cmd benchmark: Avg: 0.206774 (Laps: 25, Total Elapsed Time: 5.169350)
lldb disassembly benchmark: Avg: 0.089086 (Laps: 10, Total Elapsed Time: 0.890859)

llvm-svn: 142708
This commit is contained in:
Johnny Chen 2011-10-22 00:57:05 +00:00
parent f385f12e17
commit b8da426285
5 changed files with 196 additions and 5 deletions

48
lldb/test/bench.py Executable file
View File

@ -0,0 +1,48 @@
#!/usr/bin/env python
"""
A simple bench runner which delegates to the ./dotest.py test driver to run the
benchmarks defined in the list named 'benches'.
You need to hand edit 'benches' to modify/change the command lines passed to the
test driver.
Use the following to get only the benchmark results in your terminal output:
./bench.py 2>&1 | grep -P '^lldb.*benchmark:'
"""
import os, sys
import re
# dotest.py invocation with no '-e exe-path' uses lldb as the inferior program,
# unless there is a mentioning of custom executable program.
benches = [
# Measure startup delays creating a target and setting a breakpoint at main.
'./dotest.py -v +b -n -p TestStartupDelays.py',
# Measure 'frame variable' response after stopping at Driver::MainLoop().
'./dotest.py -v +b -x "-F Driver::MainLoop()" -n -p TestFrameVariableResponse.py',
# Measure stepping speed after stopping at Driver::MainLoop().
'./dotest.py -v +b -x "-F Driver::MainLoop()" -n -p TestSteppingSpeed.py',
# Measure expression cmd response with a simple custom executable program.
'./dotest.py +b -n -p TestExpressionCmd.py',
# Attach to a spawned lldb process then run disassembly benchmarks.
'./dotest.py -v +b -n -p TestDoAttachThenDisassembly.py'
]
def main():
"""Read the items from 'benches' and run the command line one by one."""
print "Starting bench runner...."
for command in benches:
print "Running %s" % (command)
os.system(command)
print "Bench runner done."
if __name__ == '__main__':
main()

View File

@ -0,0 +1,66 @@
"""Test lldb's disassemblt speed."""
import os, sys
import unittest2
import lldb
import pexpect
from lldbbench import *
class AttachThenDisassemblyBench(BenchBase):
mydir = os.path.join("benchmarks", "disassembly")
def setUp(self):
BenchBase.setUp(self)
@benchmarks_test
def test_attach_then_disassembly(self):
"""Attach to a spawned lldb process then run disassembly benchmarks."""
print
self.run_lldb_attach_then_disassembly(10)
print "lldb disassembly benchmark:", self.stopwatch
def run_lldb_attach_then_disassembly(self, count):
target = self.dbg.CreateTarget(self.lldbHere)
# Spawn a new process and don't display the stdout if not in TraceOn() mode.
import subprocess
popen = subprocess.Popen([self.lldbHere, self.lldbOption],
stdout = open(os.devnull, 'w') if not self.TraceOn() else None)
if self.TraceOn():
print "pid of spawned process: %d" % popen.pid
# Attach to the launched lldb process.
listener = lldb.SBListener("my.attach.listener")
error = lldb.SBError()
process = target.AttachToProcessWithID(listener, popen.pid, error)
# Set thread0 as the selected thread, followed by the 'MainLoop' frame
# as the selected frame. Then do disassembly on the function.
thread0 = process.GetThreadAtIndex(0)
process.SetSelectedThread(thread0)
i = 0
found = False
for f in thread0:
#print "frame#%d %s" % (i, f.GetFunctionName())
if "MainLoop" in f.GetFunctionName():
found = True
thread0.SetSelectedFrame(i)
if self.TraceOn():
print "Found frame#%d for function 'MainLoop'" % i
break
i += 1
# Reset the stopwatch now.
self.stopwatch.reset()
for i in range(count):
with self.stopwatch:
# Disassemble the function.
self.runCmd("disassemble -f")
if __name__ == '__main__':
import atexit
lldb.SBDebugger.Initialize()
atexit.register(lambda: lldb.SBDebugger.Terminate())
unittest2.main()

View File

@ -0,0 +1,76 @@
"""Test lldb's expression evaluations and collect statistics."""
import os, sys
import unittest2
import lldb
import pexpect
from lldbbench import *
class ExpressionEvaluationCase(BenchBase):
mydir = os.path.join("benchmarks", "expression")
def setUp(self):
BenchBase.setUp(self)
self.source = 'main.cpp'
self.line_to_break = line_number(self.source, '// Set breakpoint here.')
self.count = lldb.bmIterationCount
if self.count <= 0:
self.count = 25
@benchmarks_test
def test_expr_cmd(self):
"""Test lldb's expression commands and collect statistics."""
self.buildDefault()
self.exe_name = 'a.out'
print
self.run_lldb_repeated_exprs(self.exe_name, self.count)
print "lldb expr cmd benchmark:", self.stopwatch
def run_lldb_repeated_exprs(self, exe_name, count):
exe = os.path.join(os.getcwd(), exe_name)
# Set self.child_prompt, which is "(lldb) ".
self.child_prompt = '(lldb) '
prompt = self.child_prompt
# Reset the stopwatch now.
self.stopwatch.reset()
for i in range(count):
# So that the child gets torn down after the test.
self.child = pexpect.spawn('%s %s %s' % (self.lldbExec, self.lldbOption, exe))
child = self.child
# Turn on logging for what the child sends back.
if self.TraceOn():
child.logfile_read = sys.stdout
child.expect_exact(prompt)
child.sendline('breakpoint set -f %s -l %d' % (self.source, self.line_to_break))
child.expect_exact(prompt)
child.sendline('run')
child.expect_exact(prompt)
expr_cmd1 = 'expr ptr[j]->point.x'
expr_cmd2 = 'expr ptr[j]->point.y'
with self.stopwatch:
child.sendline(expr_cmd1)
child.expect_exact(prompt)
child.sendline(expr_cmd2)
child.expect_exact(prompt)
child.sendline('quit')
try:
self.child.expect(pexpect.EOF)
except:
pass
self.child = None
if __name__ == '__main__':
import atexit
lldb.SBDebugger.Initialize()
atexit.register(lambda: lldb.SBDebugger.Terminate())
unittest2.main()

View File

@ -26,7 +26,7 @@ class StartupDelaysBench(BenchBase):
self.count = lldb.bmIterationCount
if self.count <= 0:
self.count = 15
self.count = 30
@benchmarks_test
def test_startup_delay(self):

View File

@ -1062,10 +1062,11 @@ for ia in range(len(archs) if iterArchs else 1):
#print "sys.stdout name is", sys.stdout.name
# First, write out the number of collected test cases.
sys.stderr.write(separator + "\n")
sys.stderr.write("Collected %d test%s\n\n"
% (suite.countTestCases(),
suite.countTestCases() != 1 and "s" or ""))
if not noHeaders:
sys.stderr.write(separator + "\n")
sys.stderr.write("Collected %d test%s\n\n"
% (suite.countTestCases(),
suite.countTestCases() != 1 and "s" or ""))
class LLDBTestResult(unittest2.TextTestResult):
"""