mirror of
https://github.com/dashpay/dash.git
synced 2024-12-25 12:02:48 +01:00
87de1bc7c7
ffff9dcdc3cbe427739cc19cc7a53f032474fa2a test: Explain why test logging should be used (MarcoFalke)
Pull request description:
Background is that some tests don't have any `self.log` call at all. Thus there are no "anchor points" and those tests are hard to debug because the logs can't easily be parsed by a human.
ACKs for top commit:
jonatack:
ACK ffff9dcdc3cbe427739cc19cc7a53f032474fa2a
instagibbs:
ACK ffff9dcdc3
fanquake:
re-ACK ffff9dcdc3cbe427739cc19cc7a53f032474fa2a
Tree-SHA512: 08d962e85c4892c2a0c58feb5dc697c680a9d68e41a79417da6fcd415e0c5c735c4533a985cf225bb89deb5ca717d9bedf990657958079185804caa512b10f5a
300 lines
9.8 KiB
Markdown
300 lines
9.8 KiB
Markdown
This directory contains integration tests that test dashd and its
|
|
utilities in their entirety. It does not contain unit tests, which
|
|
can be found in [/src/test](/src/test), [/src/wallet/test](/src/wallet/test),
|
|
etc.
|
|
|
|
This directory contains the following sets of tests:
|
|
|
|
- [functional](/test/functional) which test the functionality of
|
|
dashd and dash-qt by interacting with them through the RPC and P2P
|
|
interfaces.
|
|
- [util](/test/util) which tests the dash utilities, currently only
|
|
dash-tx.
|
|
- [lint](/test/lint/) which perform various static analysis checks.
|
|
|
|
The util tests are run as part of `make check` target. The functional
|
|
tests and lint scripts can be run as explained in the sections below.
|
|
|
|
# Running tests locally
|
|
|
|
Before tests can be run locally, Dash Core must be built. See the [building instructions](/doc#building) for help.
|
|
|
|
|
|
### Functional tests
|
|
|
|
#### Dependencies and prerequisites
|
|
|
|
Many Dash specific tests require dash_hash. To install it:
|
|
|
|
- Clone the repo `git clone https://github.com/dashpay/dash_hash`
|
|
- Install dash_hash `cd dash_hash && pip3 install -r requirements.txt .`
|
|
|
|
The ZMQ functional test requires a python ZMQ library. To install it:
|
|
|
|
- on Unix, run `sudo apt-get install python3-zmq`
|
|
- on mac OS, run `pip3 install pyzmq`
|
|
|
|
|
|
On Windows the `PYTHONUTF8` environment variable must be set to 1:
|
|
|
|
```cmd
|
|
set PYTHONUTF8=1
|
|
```
|
|
|
|
#### Running the tests
|
|
|
|
Individual tests can be run by directly calling the test script, e.g.:
|
|
|
|
```
|
|
test/functional/wallet_hd.py
|
|
```
|
|
|
|
or can be run through the test_runner harness, eg:
|
|
|
|
```
|
|
test/functional/test_runner.py wallet_hd.py
|
|
```
|
|
|
|
You can run any combination (incl. duplicates) of tests by calling:
|
|
|
|
```
|
|
test/functional/test_runner.py <testname1> <testname2> <testname3> ...
|
|
```
|
|
|
|
Wildcard test names can be passed, if the paths are coherent and the test runner
|
|
is called from a `bash` shell or similar that does the globbing. For example,
|
|
to run all the wallet tests:
|
|
|
|
```
|
|
test/functional/test_runner.py test/functional/wallet*
|
|
functional/test_runner.py functional/wallet* (called from the test/ directory)
|
|
test_runner.py wallet* (called from the test/functional/ directory)
|
|
```
|
|
|
|
but not
|
|
|
|
```
|
|
test/functional/test_runner.py wallet*
|
|
```
|
|
|
|
Combinations of wildcards can be passed:
|
|
|
|
```
|
|
test/functional/test_runner.py ./test/functional/tool* test/functional/mempool*
|
|
test_runner.py tool* mempool*
|
|
```
|
|
|
|
Run the regression test suite with:
|
|
|
|
```
|
|
test/functional/test_runner.py
|
|
```
|
|
|
|
Run all possible tests with
|
|
|
|
```
|
|
test/functional/test_runner.py --extended
|
|
```
|
|
|
|
By default, up to 4 tests will be run in parallel by test_runner. To specify
|
|
how many jobs to run, append `--jobs=n`
|
|
|
|
The individual tests and the test_runner harness have many command-line
|
|
options. Run `test/functional/test_runner.py -h` to see them all.
|
|
|
|
#### Troubleshooting and debugging test failures
|
|
|
|
##### Resource contention
|
|
|
|
The P2P and RPC ports used by the dashd nodes-under-test are chosen to make
|
|
conflicts with other processes unlikely. However, if there is another dashd
|
|
process running on the system (perhaps from a previous test which hasn't successfully
|
|
killed all its dashd nodes), then there may be a port conflict which will
|
|
cause the test to fail. It is recommended that you run the tests on a system
|
|
where no other dashd processes are running.
|
|
|
|
On linux, the test framework will warn if there is another
|
|
dashd process running when the tests are started.
|
|
|
|
If there are zombie dashd processes after test failure, you can kill them
|
|
by running the following commands. **Note that these commands will kill all
|
|
dashd processes running on the system, so should not be used if any non-test
|
|
dashd processes are being run.**
|
|
|
|
```bash
|
|
killall dashd
|
|
```
|
|
|
|
or
|
|
|
|
```bash
|
|
pkill -9 dashd
|
|
```
|
|
|
|
|
|
##### Data directory cache
|
|
|
|
A pre-mined blockchain with 200 blocks is generated the first time a
|
|
functional test is run and is stored in test/cache. This speeds up
|
|
test startup times since new blockchains don't need to be generated for
|
|
each test. However, the cache may get into a bad state, in which case
|
|
tests will fail. If this happens, remove the cache directory (and make
|
|
sure dashd processes are stopped as above):
|
|
|
|
```bash
|
|
rm -rf test/cache
|
|
killall dashd
|
|
```
|
|
|
|
##### Test logging
|
|
|
|
The tests contain logging at five different levels (DEBUG, INFO, WARNING, ERROR
|
|
and CRITICAL). From within your functional tests you can log to these different
|
|
levels using the logger included in the test_framework, e.g.
|
|
`self.log.debug(object)`. By default:
|
|
|
|
- when run through the test_runner harness, *all* logs are written to
|
|
`test_framework.log` and no logs are output to the console.
|
|
- when run directly, *all* logs are written to `test_framework.log` and INFO
|
|
level and above are output to the console.
|
|
- when run by [our CI (Continuous Integration)](/ci/README.md), no logs are output to the console. However, if a test
|
|
fails, the `test_framework.log` and dashd `debug.log`s will all be dumped
|
|
to the console to help troubleshooting.
|
|
|
|
These log files can be located under the test data directory (which is always
|
|
printed in the first line of test output):
|
|
- `<test data directory>/test_framework.log`
|
|
- `<test data directory>/node<node number>/regtest/debug.log`.
|
|
|
|
The node number identifies the relevant test node, starting from `node0`, which
|
|
corresponds to its position in the nodes list of the specific test,
|
|
e.g. `self.nodes[0]`.
|
|
|
|
To change the level of logs output to the console, use the `-l` command line
|
|
argument.
|
|
|
|
`test_framework.log` and dashd `debug.log`s can be combined into a single
|
|
aggregate log by running the `combine_logs.py` script. The output can be plain
|
|
text, colorized text or html. For example:
|
|
|
|
```
|
|
test/functional/combine_logs.py -c <test data directory> | less -r
|
|
```
|
|
|
|
will pipe the colorized logs from the test into less.
|
|
|
|
Use `--tracerpc` to trace out all the RPC calls and responses to the console. For
|
|
some tests (eg any that use `submitblock` to submit a full block over RPC),
|
|
this can result in a lot of screen output.
|
|
|
|
By default, the test data directory will be deleted after a successful run.
|
|
Use `--nocleanup` to leave the test data directory intact. The test data
|
|
directory is never deleted after a failed test.
|
|
|
|
##### Attaching a debugger
|
|
|
|
A python debugger can be attached to tests at any point. Just add the line:
|
|
|
|
```py
|
|
import pdb; pdb.set_trace()
|
|
```
|
|
|
|
anywhere in the test. You will then be able to inspect variables, as well as
|
|
call methods that interact with the dashd nodes-under-test.
|
|
|
|
If further introspection of the dashd instances themselves becomes
|
|
necessary, this can be accomplished by first setting a pdb breakpoint
|
|
at an appropriate location, running the test to that point, then using
|
|
`gdb` (or `lldb` on macOS) to attach to the process and debug.
|
|
|
|
For instance, to attach to `self.node[1]` during a run you can get
|
|
the pid of the node within `pdb`.
|
|
|
|
```
|
|
(pdb) self.node[1].process.pid
|
|
```
|
|
|
|
Alternatively, you can find the pid by inspecting the temp folder for the specific test
|
|
you are running. The path to that folder is printed at the beginning of every
|
|
test run:
|
|
|
|
```bash
|
|
2017-06-27 14:13:56.686000 TestFramework (INFO): Initializing test directory /tmp/user/1000/testo9vsdjo3
|
|
```
|
|
|
|
Use the path to find the pid file in the temp folder:
|
|
|
|
```bash
|
|
cat /tmp/user/1000/testo9vsdjo3/node1/regtest/dashd.pid
|
|
```
|
|
|
|
Then you can use the pid to start `gdb`:
|
|
|
|
```bash
|
|
gdb /home/example/dashd <pid>
|
|
```
|
|
|
|
Note: gdb attach step may require ptrace_scope to be modified, or `sudo` preceding the `gdb`.
|
|
See this link for considerations: https://www.kernel.org/doc/Documentation/security/Yama.txt
|
|
|
|
Often while debugging rpc calls from functional tests, the test might reach timeout before
|
|
process can return a response. Use `--timeout-factor 0` to disable all rpc timeouts for that partcular
|
|
functional test. Ex: `test/functional/wallet_hd.py --timeout-factor 0`.
|
|
|
|
##### Profiling
|
|
|
|
An easy way to profile node performance during functional tests is provided
|
|
for Linux platforms using `perf`.
|
|
|
|
Perf will sample the running node and will generate profile data in the node's
|
|
datadir. The profile data can then be presented using `perf report` or a graphical
|
|
tool like [hotspot](https://github.com/KDAB/hotspot).
|
|
|
|
To generate a profile during test suite runs, use the `--perf` flag.
|
|
|
|
To see render the output to text, run
|
|
|
|
```sh
|
|
perf report -i /path/to/datadir/send-big-msgs.perf.data.xxxx --stdio | c++filt | less
|
|
```
|
|
|
|
For ways to generate more granular profiles, see the README in
|
|
[test/functional](/test/functional).
|
|
|
|
### Util tests
|
|
|
|
Util tests can be run locally by running `test/util/bitcoin-util-test.py`.
|
|
Use the `-v` option for verbose output.
|
|
|
|
### Lint tests
|
|
|
|
#### Dependencies
|
|
|
|
| Lint test | Dependency | Version [used by CI](../ci/lint/04_install.sh) | Installation
|
|
|-----------|:----------:|:-------------------------------------------:|--------------
|
|
| [`lint-python.sh`](lint/lint-python.sh) | [flake8](https://gitlab.com/pycqa/flake8) | [3.8.3](https://github.com/bitcoin/bitcoin/pull/19348) | `pip3 install flake8==3.8.3`
|
|
| [`lint-shell.sh`](lint/lint-shell.sh) | [ShellCheck](https://github.com/koalaman/shellcheck) | [0.7.1](https://github.com/bitcoin/bitcoin/pull/19348) | [details...](https://github.com/koalaman/shellcheck#installing)
|
|
| [`lint-spelling.sh`](lint/lint-spelling.sh) | [codespell](https://github.com/codespell-project/codespell) | [1.17.1](https://github.com/bitcoin/bitcoin/pull/19348) | `pip3 install codespell==1.17.1`
|
|
|
|
Please be aware that on Linux distributions all dependencies are usually available as packages, but could be outdated.
|
|
|
|
#### Running the tests
|
|
|
|
Individual tests can be run by directly calling the test script, e.g.:
|
|
|
|
```
|
|
test/lint/lint-filenames.sh
|
|
```
|
|
|
|
You can run all the shell-based lint tests by running:
|
|
|
|
```
|
|
test/lint/lint-all.sh
|
|
```
|
|
|
|
# Writing functional tests
|
|
|
|
You are encouraged to write functional tests for new or existing features.
|
|
Further information about the functional test framework and individual
|
|
tests is found in [test/functional](/test/functional).
|