mirror of
https://github.com/dashpay/dash.git
synced 2024-12-26 20:42:59 +01:00
e197f976e1
* Merge #16509: test: Adapt test framework for chains other than "regtest" faf36838bdba7393960fce6ad0c56dc1f93f5870 test: Avoid hardcoding the chain name in combine_logs (MarcoFalke) fa8a1d7ba30040f8c74f93fc41a61276c255a6a6 test: Adapt test framework for chains other than "regtest" (MarcoFalke) 68f546635d5de2ccfedadeabc7bc79e12e5eca6a test: Fix “local variable 'e' is assigned to but never used” (Ben Woosley) Pull request description: This is required for various work in progress: * testchains #8994 * signet #16411 * some of my locally written tests While it will be unused in the master branch as of now, it will make all of those pull requests shorter. Thus review for non-regtest tests can focus on the actual changes and not some test framework changes. ACKs for top commit: jonatack: ACK faf36838bdba7393960fce6ad0c56dc1f93f5870, ran tests and reviewed the code. Tree-SHA512: 35add66c12cab68f2fac8f7c7d47c604d3f24eae9336ff78f83e2c92b3dc08a25e7f4217199bac5393dd3fb72f945bba9c001d6fbb8efd298c88858075fcb3d6 * Add devnet support for tests * test: make sure devnet can connect to each other and start * Partial merge bitcoin/bitcoin#16681: Tests: Use self.chain instead of 'regtest' in almost all current tests, revert one TODO while at it Co-authored-by: MarcoFalke <falke.marco@gmail.com> Co-authored-by: Jorge Timón <jtimon@jtimon.cc>
379 lines
15 KiB
Python
Executable File
379 lines
15 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# Copyright (c) 2017 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
"""Class for dashd node under test"""
|
|
|
|
import contextlib
|
|
import decimal
|
|
import errno
|
|
import http.client
|
|
import json
|
|
import logging
|
|
import os.path
|
|
import re
|
|
import subprocess
|
|
import tempfile
|
|
import time
|
|
import urllib.parse
|
|
|
|
from .authproxy import JSONRPCException
|
|
from .messages import MY_SUBVERSION
|
|
from .util import (
|
|
append_config,
|
|
delete_cookie_file,
|
|
get_rpc_proxy,
|
|
rpc_url,
|
|
wait_until,
|
|
p2p_port,
|
|
get_chain_folder,
|
|
)
|
|
|
|
# For Python 3.4 compatibility
|
|
JSONDecodeError = getattr(json, "JSONDecodeError", ValueError)
|
|
|
|
BITCOIND_PROC_WAIT_TIMEOUT = 60
|
|
|
|
|
|
class FailedToStartError(Exception):
|
|
"""Raised when a node fails to start correctly."""
|
|
|
|
|
|
class TestNode():
|
|
"""A class for representing a dashd node under test.
|
|
|
|
This class contains:
|
|
|
|
- state about the node (whether it's running, etc)
|
|
- a Python subprocess.Popen object representing the running process
|
|
- an RPC connection to the node
|
|
- one or more P2P connections to the node
|
|
|
|
|
|
To make things easier for the test writer, any unrecognised messages will
|
|
be dispatched to the RPC connection."""
|
|
|
|
def __init__(self, i, datadir, extra_args_from_options, chain, rpchost, timewait, bitcoind, bitcoin_cli, stderr, mocktime, coverage_dir, extra_conf=None, extra_args=None, use_cli=False):
|
|
self.index = i
|
|
self.datadir = datadir
|
|
self.chain = chain
|
|
self.rpchost = rpchost
|
|
if timewait:
|
|
self.rpc_timeout = timewait
|
|
else:
|
|
# Wait for up to 60 seconds for the RPC server to respond
|
|
self.rpc_timeout = 60
|
|
self.binary = bitcoind
|
|
self.stderr = stderr
|
|
self.coverage_dir = coverage_dir
|
|
self.mocktime = mocktime
|
|
if extra_conf != None:
|
|
append_config(datadir, extra_conf)
|
|
# Most callers will just need to add extra args to the standard list below.
|
|
# For those callers that need more flexibity, they can just set the args property directly.
|
|
# Note that common args are set in the config file (see initialize_datadir)
|
|
self.extra_args = extra_args
|
|
self.extra_args_from_options = extra_args_from_options
|
|
self.args = [self.binary, "-datadir=" + self.datadir, "-logtimemicros", "-debug", "-debugexclude=libevent", "-debugexclude=leveldb", "-mocktime=" + str(mocktime), "-uacomment=testnode%d" % i]
|
|
|
|
self.cli = TestNodeCLI(bitcoin_cli, self.datadir)
|
|
self.use_cli = use_cli
|
|
|
|
# Don't try auto backups (they fail a lot when running tests)
|
|
self.args.append("-createwalletbackups=0")
|
|
|
|
self.running = False
|
|
self.process = None
|
|
self.rpc_connected = False
|
|
self.rpc = None
|
|
self.url = None
|
|
self.log = logging.getLogger('TestFramework.node%d' % i)
|
|
self.cleanup_on_exit = True # Whether to kill the node when this object goes away
|
|
|
|
self.p2ps = []
|
|
|
|
def _node_msg(self, msg: str) -> str:
|
|
"""Return a modified msg that identifies this node by its index as a debugging aid."""
|
|
return "[node %d] %s" % (self.index, msg)
|
|
|
|
def _raise_assertion_error(self, msg: str):
|
|
"""Raise an AssertionError with msg modified to identify this node."""
|
|
raise AssertionError(self._node_msg(msg))
|
|
|
|
def __del__(self):
|
|
# Ensure that we don't leave any dashd processes lying around after
|
|
# the test ends
|
|
if self.process and self.cleanup_on_exit:
|
|
# Should only happen on test failure
|
|
# Avoid using logger, as that may have already been shutdown when
|
|
# this destructor is called.
|
|
print(self._node_msg("Cleaning up leftover process"))
|
|
self.process.kill()
|
|
|
|
def __getattr__(self, name):
|
|
"""Dispatches any unrecognised messages to the RPC connection or a CLI instance."""
|
|
if self.use_cli:
|
|
return getattr(self.cli, name)
|
|
else:
|
|
assert self.rpc_connected and self.rpc is not None, self._node_msg("Error: no RPC connection")
|
|
return getattr(self.rpc, name)
|
|
|
|
def start(self, extra_args=None, stderr=None, *args, **kwargs):
|
|
"""Start the node."""
|
|
if extra_args is None:
|
|
extra_args = self.extra_args
|
|
if stderr is None:
|
|
stderr = self.stderr
|
|
all_args = self.args + self.extra_args_from_options + extra_args
|
|
if self.mocktime != 0:
|
|
all_args = all_args + ["-mocktime=%d" % self.mocktime]
|
|
# Delete any existing cookie file -- if such a file exists (eg due to
|
|
# unclean shutdown), it will get overwritten anyway by dashd, and
|
|
# potentially interfere with our attempt to authenticate
|
|
delete_cookie_file(self.datadir, self.chain)
|
|
self.process = subprocess.Popen(all_args, stderr=stderr, *args, **kwargs)
|
|
self.running = True
|
|
self.log.debug("dashd started, waiting for RPC to come up")
|
|
|
|
def wait_for_rpc_connection(self):
|
|
"""Sets up an RPC connection to the dashd process. Returns False if unable to connect."""
|
|
# Poll at a rate of four times per second
|
|
poll_per_s = 4
|
|
for _ in range(poll_per_s * self.rpc_timeout):
|
|
if self.process.poll() is not None:
|
|
raise FailedToStartError(self._node_msg(
|
|
'dashd exited with status {} during initialization'.format(self.process.returncode)))
|
|
try:
|
|
self.rpc = get_rpc_proxy(rpc_url(self.datadir, self.index, self.chain, self.rpchost), self.index, timeout=self.rpc_timeout, coveragedir=self.coverage_dir)
|
|
self.rpc.getblockcount()
|
|
# If the call to getblockcount() succeeds then the RPC connection is up
|
|
self.rpc_connected = True
|
|
self.url = self.rpc.url
|
|
self.log.debug("RPC successfully started")
|
|
return
|
|
except IOError as e:
|
|
if e.errno != errno.ECONNREFUSED: # Port not yet open?
|
|
raise # unknown IO error
|
|
except JSONRPCException as e: # Initialization phase
|
|
# -28 RPC in warmup
|
|
# -342 Service unavailable, RPC server started but is shutting down due to error
|
|
if e.error['code'] != -28 and e.error['code'] != -342:
|
|
raise # unknown JSON RPC exception
|
|
except ValueError as e: # cookie file not found and no rpcuser or rpcassword. dashd still starting
|
|
if "No RPC credentials" not in str(e):
|
|
raise
|
|
time.sleep(1.0 / poll_per_s)
|
|
self._raise_assertion_error("Unable to connect to dashd")
|
|
|
|
def get_wallet_rpc(self, wallet_name):
|
|
if self.use_cli:
|
|
return self.cli("-rpcwallet={}".format(wallet_name))
|
|
else:
|
|
assert self.rpc_connected and self.rpc, self._node_msg("RPC not connected")
|
|
wallet_path = "wallet/{}".format(urllib.parse.quote(wallet_name))
|
|
return self.rpc / wallet_path
|
|
|
|
def stop_node(self, wait=0):
|
|
"""Stop the node."""
|
|
if not self.running:
|
|
return
|
|
self.log.debug("Stopping node")
|
|
try:
|
|
self.stop(wait=wait)
|
|
except http.client.CannotSendRequest:
|
|
self.log.exception("Unable to stop node.")
|
|
del self.p2ps[:]
|
|
|
|
def is_node_stopped(self):
|
|
"""Checks whether the node has stopped.
|
|
|
|
Returns True if the node has stopped. False otherwise.
|
|
This method is responsible for freeing resources (self.process)."""
|
|
if not self.running:
|
|
return True
|
|
return_code = self.process.poll()
|
|
if return_code is None:
|
|
return False
|
|
|
|
# process has stopped. Assert that it didn't return an error code.
|
|
assert return_code == 0, self._node_msg(
|
|
"Node returned non-zero exit code (%d) when stopping" % return_code)
|
|
self.running = False
|
|
self.process = None
|
|
self.rpc_connected = False
|
|
self.rpc = None
|
|
self.log.debug("Node stopped")
|
|
return True
|
|
|
|
def wait_until_stopped(self, timeout=BITCOIND_PROC_WAIT_TIMEOUT):
|
|
wait_until(self.is_node_stopped, timeout=timeout)
|
|
|
|
@contextlib.contextmanager
|
|
def assert_debug_log(self, expected_msgs):
|
|
chain = get_chain_folder(self.datadir, self.chain)
|
|
debug_log = os.path.join(self.datadir, chain, 'debug.log')
|
|
with open(debug_log, encoding='utf-8') as dl:
|
|
dl.seek(0, 2)
|
|
prev_size = dl.tell()
|
|
try:
|
|
yield
|
|
finally:
|
|
with open(debug_log, encoding='utf-8') as dl:
|
|
dl.seek(prev_size)
|
|
log = dl.read()
|
|
print_log = " - " + "\n - ".join(log.splitlines())
|
|
for expected_msg in expected_msgs:
|
|
if re.search(re.escape(expected_msg), log, flags=re.MULTILINE) is None:
|
|
self._raise_assertion_error('Expected message "{}" does not partially match log:\n\n{}\n\n'.format(expected_msg, print_log))
|
|
|
|
def assert_start_raises_init_error(self, extra_args=None, expected_msg=None, partial_match=False, *args, **kwargs):
|
|
"""Attempt to start the node and expect it to raise an error.
|
|
|
|
extra_args: extra arguments to pass through to dashd
|
|
expected_msg: regex that stderr should match when dashd fails
|
|
|
|
Will throw if dashd starts without an error.
|
|
Will throw if an expected_msg is provided and it does not match dashd's stdout."""
|
|
with tempfile.SpooledTemporaryFile(max_size=2**16) as log_stderr:
|
|
try:
|
|
self.start(extra_args, stderr=log_stderr, *args, **kwargs)
|
|
self.wait_for_rpc_connection()
|
|
self.stop_node()
|
|
self.wait_until_stopped()
|
|
except FailedToStartError as e:
|
|
self.log.debug('dashd failed to start: %s', e)
|
|
self.running = False
|
|
self.process = None
|
|
# Check stderr for expected message
|
|
if expected_msg is not None:
|
|
log_stderr.seek(0)
|
|
stderr = log_stderr.read().decode('utf-8').strip()
|
|
if partial_match:
|
|
if re.search(expected_msg, stderr, flags=re.MULTILINE) is None:
|
|
self._raise_assertion_error(
|
|
'Expected message "{}" does not partially match stderr:\n"{}"'.format(expected_msg, stderr))
|
|
else:
|
|
if re.fullmatch(expected_msg, stderr) is None:
|
|
self._raise_assertion_error(
|
|
'Expected message "{}" does not fully match stderr:\n"{}"'.format(expected_msg, stderr))
|
|
else:
|
|
if expected_msg is None:
|
|
assert_msg = "dashd should have exited with an error"
|
|
else:
|
|
assert_msg = "dashd should have exited with expected error " + expected_msg
|
|
self._raise_assertion_error(assert_msg)
|
|
|
|
def node_encrypt_wallet(self, passphrase):
|
|
""""Encrypts the wallet.
|
|
|
|
This causes dashd to shutdown, so this method takes
|
|
care of cleaning up resources."""
|
|
self.encryptwallet(passphrase)
|
|
self.wait_until_stopped()
|
|
|
|
def add_p2p_connection(self, p2p_conn, *args, **kwargs):
|
|
"""Add a p2p connection to the node.
|
|
|
|
This method adds the p2p connection to the self.p2ps list and also
|
|
returns the connection to the caller."""
|
|
if 'dstport' not in kwargs:
|
|
kwargs['dstport'] = p2p_port(self.index)
|
|
if 'dstaddr' not in kwargs:
|
|
kwargs['dstaddr'] = '127.0.0.1'
|
|
|
|
p2p_conn.peer_connect(*args, **kwargs, net=self.chain)
|
|
self.p2ps.append(p2p_conn)
|
|
|
|
return p2p_conn
|
|
|
|
@property
|
|
def p2p(self):
|
|
"""Return the first p2p connection
|
|
|
|
Convenience property - most tests only use a single p2p connection to each
|
|
node, so this saves having to write node.p2ps[0] many times."""
|
|
assert self.p2ps, self._node_msg("No p2p connection")
|
|
return self.p2ps[0]
|
|
|
|
def disconnect_p2ps(self):
|
|
"""Close all p2p connections to the node."""
|
|
for p in self.p2ps:
|
|
p.peer_disconnect()
|
|
del self.p2ps[:]
|
|
|
|
# wait for p2p connections to disappear from getpeerinfo()
|
|
def check_peers():
|
|
for p in self.getpeerinfo():
|
|
if p['subver'] == MY_SUBVERSION.decode():
|
|
return False
|
|
return True
|
|
wait_until(check_peers, timeout=5)
|
|
|
|
class TestNodeCLIAttr:
|
|
def __init__(self, cli, command):
|
|
self.cli = cli
|
|
self.command = command
|
|
|
|
def __call__(self, *args, **kwargs):
|
|
return self.cli.send_cli(self.command, *args, **kwargs)
|
|
|
|
def get_request(self, *args, **kwargs):
|
|
return lambda: self(*args, **kwargs)
|
|
|
|
class TestNodeCLI():
|
|
"""Interface to dash-cli for an individual node"""
|
|
|
|
def __init__(self, binary, datadir):
|
|
self.options = []
|
|
self.binary = binary
|
|
self.datadir = datadir
|
|
self.input = None
|
|
self.log = logging.getLogger('TestFramework.dashcli')
|
|
|
|
def __call__(self, *options, input=None):
|
|
# TestNodeCLI is callable with dash-cli command-line options
|
|
cli = TestNodeCLI(self.binary, self.datadir)
|
|
cli.options = [str(o) for o in options]
|
|
cli.input = input
|
|
return cli
|
|
|
|
def __getattr__(self, command):
|
|
return TestNodeCLIAttr(self, command)
|
|
|
|
def batch(self, requests):
|
|
results = []
|
|
for request in requests:
|
|
try:
|
|
results.append(dict(result=request()))
|
|
except JSONRPCException as e:
|
|
results.append(dict(error=e))
|
|
return results
|
|
|
|
def send_cli(self, command=None, *args, **kwargs):
|
|
"""Run dash-cli command. Deserializes returned string as python object."""
|
|
|
|
pos_args = [str(arg) for arg in args]
|
|
named_args = [str(key) + "=" + str(value) for (key, value) in kwargs.items()]
|
|
assert not (pos_args and named_args), "Cannot use positional arguments and named arguments in the same dash-cli call"
|
|
p_args = [self.binary, "-datadir=" + self.datadir] + self.options
|
|
if named_args:
|
|
p_args += ["-named"]
|
|
if command is not None:
|
|
p_args += [command]
|
|
p_args += pos_args + named_args
|
|
self.log.debug("Running dash-cli command: %s" % command)
|
|
process = subprocess.Popen(p_args, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=subprocess.PIPE, universal_newlines=True)
|
|
cli_stdout, cli_stderr = process.communicate(input=self.input)
|
|
returncode = process.poll()
|
|
if returncode:
|
|
match = re.match(r'error code: ([-0-9]+)\nerror message:\n(.*)', cli_stderr)
|
|
if match:
|
|
code, message = match.groups()
|
|
raise JSONRPCException(dict(code=int(code), message=message))
|
|
# Ignore cli_stdout, raise with cli_stderr
|
|
raise subprocess.CalledProcessError(returncode, self.binary, output=cli_stderr)
|
|
try:
|
|
return json.loads(cli_stdout, parse_float=decimal.Decimal)
|
|
except JSONDecodeError:
|
|
return cli_stdout.rstrip("\n")
|