forked from OSchip/llvm-project
430 lines
18 KiB
ReStructuredText
430 lines
18 KiB
ReStructuredText
Testing
|
||
=======
|
||
|
||
.. contents::
|
||
:local:
|
||
|
||
Test Suite Structure
|
||
--------------------
|
||
|
||
The LLDB test suite consists of three different kinds of test:
|
||
|
||
* **Unit tests**: written in C++ using the googletest unit testing library.
|
||
* **Shell tests**: Integration tests that test the debugger through the command
|
||
line. These tests interact with the debugger either through the command line
|
||
driver or through ``lldb-test`` which is a tool that exposes the internal
|
||
data structures in an easy-to-parse way for testing. Most people will know
|
||
these as *lit tests* in LLVM, although lit is the test driver and ShellTest
|
||
is the test format that uses ``RUN:`` lines. `FileCheck
|
||
<https://llvm.org/docs/CommandGuide/FileCheck.html>`_ is used to verify
|
||
the output.
|
||
* **API tests**: Integration tests that interact with the debugger through the
|
||
SB API. These are written in Python and use LLDB's ``dotest.py`` testing
|
||
framework on top of Python's `unittest2
|
||
<https://docs.python.org/2/library/unittest.html>`_.
|
||
|
||
All three test suites use ``lit`` (`LLVM Integrated Tester
|
||
<https://llvm.org/docs/CommandGuide/lit.html>`_ ) as the test driver. The test
|
||
suites can be run as a whole or separately.
|
||
|
||
|
||
Unit Tests
|
||
``````````
|
||
|
||
Unit tests are located under ``lldb/unittests``. If it's possible to test
|
||
something in isolation or as a single unit, you should make it a unit test.
|
||
|
||
Often you need instances of the core objects such as a debugger, target or
|
||
process, in order to test something meaningful. We already have a handful of
|
||
tests that have the necessary boiler plate, but this is something we could
|
||
abstract away and make it more user friendly.
|
||
|
||
Shell Tests
|
||
```````````
|
||
|
||
Shell tests are located under ``lldb/test/Shell``. These tests are generally
|
||
built around checking the output of ``lldb`` (the command line driver) or
|
||
``lldb-test`` using ``FileCheck``. Shell tests are generally small and fast to
|
||
write because they require little boilerplate.
|
||
|
||
``lldb-test`` is a relatively new addition to the test suite. It was the first
|
||
tool that was added that is designed for testing. Since then it has been
|
||
continuously extended with new subcommands, improving our test coverage. Among
|
||
other things you can use it to query lldb for symbol files, for object files
|
||
and breakpoints.
|
||
|
||
Obviously shell tests are great for testing the command line driver itself or
|
||
the subcomponents already exposed by lldb-test. But when it comes to LLDB's
|
||
vast functionality, most things can be tested both through the driver as well
|
||
as the Python API. For example, to test setting a breakpoint, you could do it
|
||
from the command line driver with ``b main`` or you could use the SB API and do
|
||
something like ``target.BreakpointCreateByName`` [#]_.
|
||
|
||
A good rule of thumb is to prefer shell tests when what is being tested is
|
||
relatively simple. Expressivity is limited compared to the API tests, which
|
||
means that you have to have a well-defined test scenario that you can easily
|
||
match with ``FileCheck``.
|
||
|
||
Another thing to consider are the binaries being debugged, which we call
|
||
inferiors. For shell tests, they have to be relatively simple. The
|
||
``dotest.py`` test framework has extensive support for complex build scenarios
|
||
and different variants, which is described in more detail below, while shell
|
||
tests are limited to single lines of shell commands with compiler and linker
|
||
invocations.
|
||
|
||
On the same topic, another interesting aspect of the shell tests is that there
|
||
you can often get away with a broken or incomplete binary, whereas the API
|
||
tests almost always require a fully functional executable. This enables testing
|
||
of (some) aspects of handling of binaries with non-native architectures or
|
||
operating systems.
|
||
|
||
Finally, the shell tests always run in batch mode. You start with some input
|
||
and the test verifies the output. The debugger can be sensitive to its
|
||
environment, such as the the platform it runs on. It can be hard to express
|
||
that the same test might behave slightly differently on macOS and Linux.
|
||
Additionally, the debugger is an interactive tool, and the shell test provide
|
||
no good way of testing those interactive aspects, such as tab completion for
|
||
example.
|
||
|
||
API Tests
|
||
`````````
|
||
|
||
API tests are located under ``lldb/test/API``. They are run with the
|
||
``dotest.py``. Tests are written in Python and test binaries (inferiors) are
|
||
compiled with Make. The majority of API tests are end-to-end tests that compile
|
||
programs from source, run them, and debug the processes.
|
||
|
||
As mentioned before, ``dotest.py`` is LLDB's testing framework. The
|
||
implementation is located under ``lldb/packages/Python/lldbsuite``. We have
|
||
several extensions and custom test primitives on top of what's offered by
|
||
`unittest2 <https://docs.python.org/2/library/unittest.html>`_. Those can be
|
||
found in
|
||
`lldbtest.py <https://github.com/llvm/llvm-project/blob/master/lldb/packages/Python/lldbsuite/test/lldbtest.py>`_.
|
||
|
||
Below is the directory layout of the `example API test
|
||
<https://github.com/llvm/llvm-project/tree/master/lldb/test/API/sample_test>`_.
|
||
The test directory will always contain a python file, starting with ``Test``.
|
||
Most of the tests are structured as a binary being debugged, so there will be
|
||
one or more source files and a ``Makefile``.
|
||
|
||
::
|
||
|
||
sample_test
|
||
├── Makefile
|
||
├── TestSampleTest.py
|
||
└── main.c
|
||
|
||
Let's start with the Python test file. Every test is its own class and can have
|
||
one or more test methods, that start with ``test_``. Many tests define
|
||
multiple test methods and share a bunch of common code. For example, for a
|
||
fictive test that makes sure we can set breakpoints we might have one test
|
||
method that ensures we can set a breakpoint by address, on that sets a
|
||
breakpoint by name and another that sets the same breakpoint by file and line
|
||
number. The setup, teardown and everything else other than setting the
|
||
breakpoint could be shared.
|
||
|
||
Our testing framework also has a bunch of utilities that abstract common
|
||
operations, such as creating targets, setting breakpoints etc. When code is
|
||
shared across tests, we extract it into a utility in ``lldbutil``. It's always
|
||
worth taking a look at `lldbutil
|
||
<https://github.com/llvm/llvm-project/blob/master/lldb/packages/Python/lldbsuite/test/lldbutil.py>`_
|
||
to see if there's a utility to simplify some of the testing boiler plate.
|
||
Because we can't always audit every existing test, this is doubly true when
|
||
looking at an existing test for inspiration.
|
||
|
||
It's possible to skip or `XFAIL
|
||
<https://ftp.gnu.org/old-gnu/Manuals/dejagnu-1.3/html_node/dejagnu_6.html>`_
|
||
tests using decorators. You'll see them a lot. The debugger can be sensitive to
|
||
things like the architecture, the host and target platform, the compiler
|
||
version etc. LLDB comes with a range of predefined decorators for these
|
||
configurations.
|
||
|
||
::
|
||
|
||
@expectedFailureAll(archs=["aarch64"], oslist=["linux"]
|
||
|
||
Another great thing about these decorators is that they're very easy to extend,
|
||
it's even possible to define a function in a test case that determines whether
|
||
the test should be run or not.
|
||
|
||
::
|
||
|
||
@expectedFailure(checking_function_name)
|
||
|
||
In addition to providing a lot more flexibility when it comes to writing the
|
||
test, the API test also allow for much more complex scenarios when it comes to
|
||
building inferiors. Every test has its own ``Makefile``, most of them only a
|
||
few lines long. A shared ``Makefile`` (``Makefile.rules``) with about a
|
||
thousand lines of rules takes care of most if not all of the boiler plate,
|
||
while individual make files can be used to build more advanced tests.
|
||
|
||
Here's an example of a simple ``Makefile`` used by the example test.
|
||
|
||
::
|
||
|
||
C_SOURCES := main.c
|
||
CFLAGS_EXTRAS := -std=c99
|
||
|
||
include Makefile.rules
|
||
|
||
Finding the right variables to set can be tricky. You can always take a look at
|
||
`Makefile.rules <https://github.com/llvm/llvm-project/blob/master/lldb/packages/Python/lldbsuite/test/make/Makefile.rules>`_
|
||
but often it's easier to find an existing ``Makefile`` that does something
|
||
similar to what you want to do.
|
||
|
||
Another thing this enables is having different variants for the same test
|
||
case. By default, we run every test for all 3 debug info formats, so once with
|
||
DWARF from the object files, once with gmodules and finally with a dSYM on
|
||
macOS or split DWARF (DWO) on Linux. But there are many more things we can test
|
||
that are orthogonal to the test itself. On GreenDragon we have a matrix bot
|
||
that runs the test suite under different configurations, with older host
|
||
compilers and different DWARF versions.
|
||
|
||
As you can imagine, this quickly lead to combinatorial explosion in the number
|
||
of variants. It's very tempting to add more variants because it's an easy way
|
||
to increase test coverage. It doesn't scale. It's easy to set up, but increases
|
||
the runtime of the tests and has a large ongoing cost.
|
||
|
||
The key take away is that the different variants don't obviate the need for
|
||
focused tests. So relying on it to test say DWARF5 is a really bad idea.
|
||
Instead you should write tests that check the specific DWARF5 feature, and have
|
||
the variant as a nice-to-have.
|
||
|
||
In conclusion, you'll want to opt for an API test to test the API itself or
|
||
when you need the expressivity, either for the test case itself or for the
|
||
program being debugged. The fact that the API tests work with different
|
||
variants mean that more general tests should be API tests, so that they can be
|
||
run against the different variants.
|
||
|
||
Running The Tests
|
||
-----------------
|
||
|
||
.. note::
|
||
|
||
On Windows any invocations of python should be replaced with python_d, the
|
||
debug interpreter, when running the test suite against a debug version of
|
||
LLDB.
|
||
|
||
.. note::
|
||
|
||
On NetBSD you must export ``LD_LIBRARY_PATH=$PWD/lib`` in your environment.
|
||
This is due to lack of the ``$ORIGIN`` linker feature.
|
||
|
||
Running the Full Test Suite
|
||
```````````````````````````
|
||
|
||
The easiest way to run the LLDB test suite is to use the ``check-lldb`` build
|
||
target.
|
||
|
||
By default, the ``check-lldb`` target builds the test programs with the same
|
||
compiler that was used to build LLDB. To build the tests with a different
|
||
compiler, you can set the ``LLDB_TEST_COMPILER`` CMake variable.
|
||
|
||
It is possible to customize the architecture of the test binaries and compiler
|
||
used by appending ``-A`` and ``-C`` options respectively to the CMake variable
|
||
``LLDB_TEST_USER_ARGS``. For example, to test LLDB against 32-bit binaries
|
||
built with a custom version of clang, do:
|
||
|
||
::
|
||
|
||
$ cmake -DLLDB_TEST_USER_ARGS="-A i386 -C /path/to/custom/clang" -G Ninja
|
||
$ ninja check-lldb
|
||
|
||
Note that multiple ``-A`` and ``-C`` flags can be specified to
|
||
``LLDB_TEST_USER_ARGS``.
|
||
|
||
Running a Single Test Suite
|
||
```````````````````````````
|
||
|
||
Each test suite can be run separately, similar to running the whole test suite
|
||
with ``check-lldb``.
|
||
|
||
* Use ``check-lldb-unit`` to run just the unit tests.
|
||
* Use ``check-lldb-api`` to run just the SB API tests.
|
||
* Use ``check-lldb-shell`` to run just the shell tests.
|
||
|
||
You can run specific subdirectories by appending the directory name to the
|
||
target. For example, to run all the tests in ``ObjectFile``, you can use the
|
||
target ``check-lldb-shell-objectfile``. However, because the unit tests and API
|
||
tests don't actually live under ``lldb/test``, this convenience is only
|
||
available for the shell tests.
|
||
|
||
Running a Single Test
|
||
`````````````````````
|
||
|
||
The recommended way to run a single test is by invoking the lit driver with a
|
||
filter. This ensures that the test is run with the same configuration as when
|
||
run as part of a test suite.
|
||
|
||
::
|
||
|
||
$ ./bin/llvm-lit -sv tools/lldb/test --filter <test>
|
||
|
||
|
||
Because lit automatically scans a directory for tests, it's also possible to
|
||
pass a subdirectory to run a specific subset of the tests.
|
||
|
||
::
|
||
|
||
$ ./bin/llvm-lit -sv tools/lldb/test/Shell/Commands/CommandScriptImmediateOutput
|
||
|
||
|
||
For the SB API tests it is possible to forward arguments to ``dotest.py`` by
|
||
passing ``--param`` to lit and setting a value for ``dotest-args``.
|
||
|
||
::
|
||
|
||
$ ./bin/llvm-lit -sv tools/lldb/test --param dotest-args='-C gcc'
|
||
|
||
|
||
Below is an overview of running individual test in the unit and API test suites
|
||
without going through the lit driver.
|
||
|
||
Running a Specific Test or Set of Tests: API Tests
|
||
``````````````````````````````````````````````````
|
||
|
||
In addition to running all the LLDB test suites with the ``check-lldb`` CMake
|
||
target above, it is possible to run individual LLDB tests. If you have a CMake
|
||
build you can use the ``lldb-dotest`` binary, which is a wrapper around
|
||
``dotest.py`` that passes all the arguments configured by CMake.
|
||
|
||
Alternatively, you can use ``dotest.py`` directly, if you want to run a test
|
||
one-off with a different configuration.
|
||
|
||
For example, to run the test cases defined in TestInferiorCrashing.py, run:
|
||
|
||
::
|
||
|
||
$ ./bin/lldb-dotest -p TestInferiorCrashing.py
|
||
|
||
::
|
||
|
||
$ cd $lldb/test
|
||
$ python dotest.py --executable <path-to-lldb> -p TestInferiorCrashing.py ../packages/Python/lldbsuite/test
|
||
|
||
If the test is not specified by name (e.g. if you leave the ``-p`` argument
|
||
off), all tests in that directory will be executed:
|
||
|
||
|
||
::
|
||
|
||
$ ./bin/lldb-dotest functionalities/data-formatter
|
||
|
||
::
|
||
|
||
$ python dotest.py --executable <path-to-lldb> functionalities/data-formatter
|
||
|
||
Many more options that are available. To see a list of all of them, run:
|
||
|
||
::
|
||
|
||
$ python dotest.py -h
|
||
|
||
|
||
Running a Specific Test or Set of Tests: Unit Tests
|
||
```````````````````````````````````````````````````
|
||
|
||
The unit tests are simple executables, located in the build directory under ``tools/lldb/unittests``.
|
||
|
||
To run them, just run the test binary, for example, to run all the Host tests:
|
||
|
||
::
|
||
|
||
$ ./tools/lldb/unittests/Host/HostTests
|
||
|
||
|
||
To run a specific test, pass a filter, for example:
|
||
|
||
::
|
||
|
||
$ ./tools/lldb/unittests/Host/HostTests --gtest_filter=SocketTest.DomainListenConnectAccept
|
||
|
||
|
||
Running the Test Suite Remotely
|
||
```````````````````````````````
|
||
|
||
Running the test-suite remotely is similar to the process of running a local
|
||
test suite, but there are two things to have in mind:
|
||
|
||
1. You must have the lldb-server running on the remote system, ready to accept
|
||
multiple connections. For more information on how to setup remote debugging
|
||
see the Remote debugging page.
|
||
2. You must tell the test-suite how to connect to the remote system. This is
|
||
achieved using the ``--platform-name``, ``--platform-url`` and
|
||
``--platform-working-dir`` parameters to ``dotest.py``. These parameters
|
||
correspond to the platform select and platform connect LLDB commands. You
|
||
will usually also need to specify the compiler and architecture for the
|
||
remote system.
|
||
|
||
Currently, running the remote test suite is supported only with ``dotest.py`` (or
|
||
dosep.py with a single thread), but we expect this issue to be addressed in the
|
||
near future.
|
||
|
||
Running tests in QEMU System Emulation Environment
|
||
``````````````````````````````````````````````````
|
||
|
||
QEMU can be used to test LLDB in an emulation environment in the absence of
|
||
actual hardware. `QEMU based testing <https://lldb.llvm.org/use/qemu-testing.html>`_
|
||
page describes how to setup a emulation environment using QEMU helper scripts
|
||
found under llvm-project/lldb/scripts/lldb-test-qemu. These scripts currently
|
||
work with Arm or AArch64, but support for other architectures can be added easily.
|
||
|
||
Debugging Test Failures
|
||
-----------------------
|
||
|
||
On non-Windows platforms, you can use the ``-d`` option to ``dotest.py`` which
|
||
will cause the script to wait for a while until a debugger is attached.
|
||
|
||
Debugging Test Failures on Windows
|
||
``````````````````````````````````
|
||
|
||
On Windows, it is strongly recommended to use Python Tools for Visual Studio
|
||
for debugging test failures. It can seamlessly step between native and managed
|
||
code, which is very helpful when you need to step through the test itself, and
|
||
then into the LLDB code that backs the operations the test is performing.
|
||
|
||
A quick guide to getting started with PTVS is as follows:
|
||
|
||
#. Install PTVS
|
||
#. Create a Visual Studio Project for the Python code.
|
||
#. Go to File -> New -> Project -> Python -> From Existing Python Code.
|
||
#. Choose llvm/tools/lldb as the directory containing the Python code.
|
||
#. When asked where to save the .pyproj file, choose the folder ``llvm/tools/lldb/pyproj``. This is a special folder that is ignored by the ``.gitignore`` file, since it is not checked in.
|
||
#. Set test/dotest.py as the startup file
|
||
#. Make sure there is a Python Environment installed for your distribution. For example, if you installed Python to ``C:\Python35``, PTVS needs to know that this is the interpreter you want to use for running the test suite.
|
||
#. Go to Tools -> Options -> Python Tools -> Environment Options
|
||
#. Click Add Environment, and enter Python 3.5 Debug for the name. Fill out the values correctly.
|
||
#. Configure the project to use this debug interpreter.
|
||
#. Right click the Project node in Solution Explorer.
|
||
#. In the General tab, Make sure Python 3.5 Debug is the selected Interpreter.
|
||
#. In Debug/Search Paths, enter the path to your ninja/lib/site-packages directory.
|
||
#. In Debug/Environment Variables, enter ``VCINSTALLDIR=C:\Program Files (x86)\Microsoft Visual Studio 14.0\VC\``.
|
||
#. If you want to enabled mixed mode debugging, check Enable native code debugging (this slows down debugging, so enable it only on an as-needed basis.)
|
||
#. Set the command line for the test suite to run.
|
||
#. Right click the project in solution explorer and choose the Debug tab.
|
||
#. Enter the arguments to dotest.py.
|
||
#. Example command options:
|
||
|
||
::
|
||
|
||
--arch=i686
|
||
# Path to debug lldb.exe
|
||
--executable D:/src/llvmbuild/ninja/bin/lldb.exe
|
||
# Directory to store log files
|
||
-s D:/src/llvmbuild/ninja/lldb-test-traces
|
||
-u CXXFLAGS -u CFLAGS
|
||
# If a test crashes, show JIT debugging dialog.
|
||
--enable-crash-dialog
|
||
# Path to release clang.exe
|
||
-C d:\src\llvmbuild\ninja_release\bin\clang.exe
|
||
# Path to the particular test you want to debug.
|
||
-p TestPaths.py
|
||
# Root of test tree
|
||
D:\src\llvm\tools\lldb\packages\Python\lldbsuite\test
|
||
|
||
::
|
||
|
||
--arch=i686 --executable D:/src/llvmbuild/ninja/bin/lldb.exe -s D:/src/llvmbuild/ninja/lldb-test-traces -u CXXFLAGS -u CFLAGS --enable-crash-dialog -C d:\src\llvmbuild\ninja_release\bin\clang.exe -p TestPaths.py D:\src\llvm\tools\lldb\packages\Python\lldbsuite\test --no-multiprocess
|
||
|
||
.. [#] `https://lldb.llvm.org/python_reference/lldb.SBTarget-class.html#BreakpointCreateByName <https://lldb.llvm.org/python_reference/lldb.SBTarget-class.html#BreakpointCreateByName>`_
|