Rizin uses both regression and unit tests.
- db/: The regressions tests sources
- unit/: Unit tests (written in C, using minunit).
- fuzz/: Fuzzing helper scripts
- bins/: Sample binaries (fetched from the external repository)
- bench/: Benchmarks
- rizin installed and in
$PATH(you can also use a rizin not in$PATH, but other files like calling convention files, format files, etc. must have been installed). - rz-test compiled and/or installed, which is done by default automatically when building Rizin.
To run regressions tests use rz-test from within the test directory.
By default it will run all tests under the db subdirectory, however you can
also specify which tests you want to run, by providing its name as argument to
rz-test.
For example, to run only the asm tests for x86_64, you can do rz-test db/asm/x86_64. rz-test provides other interesting options that you can check
out by doing rz-test -h.
An option that you may find interesting, in particular when doing changes that
may affect the output of multiple tests, is the -i option, which enables
interactive mode. When running tests in this mode, rz-test will warn you for
each failed test and it will ask for your input on how to treat the issue. It
can automatically fix the test so that it matches the new output (if that is the
right behaviour!) or it can mark it as broken for you.
To run unit tests, just use ninja -C build test (or meson test -C build)
from the top directory (replace build with the name of the directory you used
to build Rizin).
You can run one specific testcase category (e.g. the whole test_bin.c file) using meson test -C build bin.
If you are using meson test, you should consider using the --print-errorlogs flag.
In order to be able to run the benchmarks, the -Denable_benchmarks=true switch needs to be specified
when setting up the build directory (e.g. meson setup build -Denable_benchmarks=true).
Afterwards use ninja -C build test --benchmark (or meson test -C build --benchmark)
to run the benchmarks from the top directory (replace build with the name of the directory
you used to build Rizin).
Running a specific set of benchmarks (e.g. bitvector) can be done with ninja -C build test bitvector --benchmark.
A test can have one of the following results:
- success: The test passed, and that was expected.
- fixed: The test passed, but failure was expected.
- broken: Failure was expected, and happened.
- failed: The test failed unexpectedly. This is a regression.
Tests for the assembly and disassembly (in db/asm/*) have a different format:
General format:
type "assembly" opcode [offset] [IL]
where type can be any of:
- a meaning "assemble"
- d meaning "disassemble"
- B meaning "broken"
- E stands for cfg.bigendian=true
Some architectures are going to assemble an instruction differently depending on the offset it's written to. Optional.
Examples:
a "ret" c3
d "ret" c3
a "nop" 90 # Assembly is correct
dB "nopppp" 90 # Disassembly test is broken
Some instructions change if they appear together with another instructions. To test two or more instructions in a single test the instructions' assembly text and IL representations can be concatinated with a simicolon.
Example: Branch delay in Sparc:
dE "call g1;nop" 9fc0600001000000 0x40 (set o7 (bv 64 0x40));(seq nop (jmp (var g1)))
Example: ARM conditional blocks with it.
d "ite eq;moveq r0, 1" 0cbf0120 0x0 nop;(branch (var zf) (set r0 (bv 32 0x1)) nop)
To also test lifting an instruction to RzIL, you can append the readable IL representation like so:
d "inc ptr" 3e 0 set(v:ptr, x:add(x:var(v:ptr), y:bitv(bits:0x0000000000000001, len:64)))
This means that rz-test will also perform the lifting from bytes to RzIL, run the validation pass on the result and compare it against the given string.
In this case, passing an offset is mandatory, otherwise the argument would be ambiguous.
You can merge lines:
adB "nop" 90
acts the same as
aB "nop" 90
dB "nop" 90
The filename is very important. It is used to tell rizin which architecture to use: arch[[_cpu]_bits].
Examples:
x86_32means-a x86 -b 32arm_v7_64means-a arm -b 64
The JSON tests db/json are executed on 3 standard files (1 ELF, 1 MachO, 1 PE). The tests need to be working on the 3 files to pass.
Example commands tests for the other db/ folders:
NAME=test_db
FILE=bins/elf/ls
CMDS=<<EOF
pd 4
EOF
EXPECT=<<EOF
;-- main:
;-- entry0:
;-- func.100001174:
0x100001174 55 Push rbp
0x100001175 4889e5 Mov rbp, rsp
0x100001178 4157 Push r15
EOF
RUNIt is also possible to match specific parts of the output in EXPECT and EXPECT_ERR using regex (with
REGEXP_FILTER_OUT and REGEXP_FILTER_ERR respectively) in case some of the test's output is dynamic:
NAME=bp rebase
FILE=bins/elf/analysis/pie
ARGS=-d
CMDS=<<EOF
aa
db @ main
dbl~main
doc
dbl~main
EOF
REGEXP_FILTER_OUT=([a-zA-Z="]+\s+)
EXPECT=<<EOF
0x000005c5 0x000005c6 1 --x sw break enabled valid main
0x000005c5 0x000005c6 1 --x sw break enabled valid main
EOF
RUNWithout the regex that filtered out the non-deterministic file path and addresses, the expected output would have been the following:
0x566495c5 - 0x566495c6 1 --x sw break enabled valid cmd="" cond="" name="main" module="/home/user/rizin/test/bins/elf/analysis/pie"
0x000005c5 - 0x000005c6 1 --x sw break enabled valid cmd="" cond="" name="main" module="/home/user/rizin/test/bins/elf/analysis/pie"
- NAME is the name of the test, it must be unique
- FILE is the path of the file used for the test
- ARGS (optional) are the command line argument passed to rizin (e.g -b 16)
- CMDS are the commands to be executed by the test
- EXPECT is the expected output of the test from stdout. If
REGEXP_FILTER_OUTis used,EXPECTmatches only the filtered output. - EXPECT_ERR (optional) is the expected output of the test from stderr. Can be specified in addition or instead of
EXPECT - BROKEN (optional) is 1 if the tests is expected to be fail, 0 or unspecified otherwise
- TIMEOUT (optional) is the number of seconds to wait before considering the test timeout
- REGEXP_FILTER_OUT (optional) apply given regex on stdout before comparing the output to
EXPECT(e.g.REGEXP_FILTER_OUT=([a-zA-Z]+)). This is similar to piping stdout togrep -E "<regex>"and then comparing the matched text withEXPECT. - REGEXP_FILTER_ERR (optional) apply given regex on stderr before comparing the ouput to
EXPECT_ERR
You must end the test by adding RUN keyword
- For portability reasons do not use shell pipes, use
~ - dont use
pdif not necessary, usepi - All tests use the UTC timezone for consistency.
Assembly, JSON and commands tests are useful to test the overall behaviour of Rizin, but to test new API or new code we suggest to write small unit tests.
The basic structure of a unit test is the following:
#include <rz_XXXXX.h>
#include "minunit.h" // Place at the bottom of includes.
static bool test_my_feature(void) {
// code to test the behaviour
mu_end;
}
static bool all_tests() {
mu_run_test(test_my_feature);
return tests_passed != tests_run;
}
mu_main (all_tests)Minunit provides various functions to check the actual output of a function with the expected one. For example:
mu_assert_true(actual, message)checks thatactualevaluates to true, otherwise it printsmessageon stderr.mu_assert_false(actual, message)checks thatactualevaluates to false, otherwise it printsmessageon stderr.mu_assert_eq(actual, expected, message)checks that the integer (ut64 at most)actualis equal to the integerexpected, otherwise it printsmessageon stderr.mu_assert_ptreq(actual, expected, message)checks that the pointeractualis equal toexpected.mu_assert_null(actual, message)mu_assert_streq(actual, expected, message)mu_assert_memeq(actual, expected, len, message)- etc.
If you add a unit test file, be sure to also add it to unit/meson.build, so it
is compiled when you compile Rizin.
The test files are licensed under GPL 3 (or later).