mirror of
https://github.com/dashpay/dash.git
synced 2024-12-27 13:03:17 +01:00
f293c046f4
223588b1bbc63dc57098bbd0baa48635e0cc0b82 Add a --descriptors option to various tests (Andrew Chow) 869f7ab30aeb4d7fbd563c535b55467a8a0430cf tests: Add RPCOverloadWrapper which overloads some disabled RPCs (Andrew Chow) cf060628590fab87d73f278e744d70ef2d5d81db Correctly check for default wallet (Andrew Chow) 886e0d75f5fea2421190aa4812777d89f68962cc Implement CWallet::IsSpentKey for non-LegacySPKMans (Andrew Chow) 3c19fdd2a2fd5394fcfa75b2ba84ab2277cbdabf Return error when no ScriptPubKeyMan is available for specified type (Andrew Chow) 388ba94231f2f10a0be751c562cdd4650510a90a Change wallet_encryption.py to use signmessage instead of dumpprivkey (Andrew Chow) 1346e14831489f9c8f53a08f9dfed61d55d53c6f Functional tests for descriptor wallets (Andrew Chow) f193ea889ddb53d9a5c47647966681d525e38368 add importdescriptors RPC and tests for native descriptor wallets (Hugo Nguyen) ce24a944940019185efebcc5d85eac458ed26016 Add IsLegacy to CWallet so that the GUI knows whether to show watchonly (Andrew Chow) 1cb42b22b11c27e64462afc25a94b2fc50bfa113 Generate new descriptors when encrypting (Andrew Chow) 82ae02b1656819f4bd5023b8955447e1d4ea8692 Be able to create new wallets with DescriptorScriptPubKeyMans as backing (Andrew Chow) b713baa75a62335ab9c0eed9ef76a95bfec30668 Implement GetMetadata in DescriptorScriptPubKeyMan (Andrew Chow) 8b9603bd0b443e2f7984eb72bf2e21cf02af0bcb Change GetMetadata to use unique_ptr<CKeyMetadata> (Andrew Chow) 72a9540df96ffdb94f039b9c14eaacdc7d961196 Implement FillPSBT in DescriptorScriptPubKeyMan (Andrew Chow) 84b4978c02102171775c77a45f6ec198930f0a88 Implement SignMessage for descriptor wallets (Andrew Chow) bde7c9fa38775a81d53ac0484fa9c98076a0c7d1 Implement SignTransaction in DescriptorScriptPubKeyMan (Andrew Chow) d50c8ddd4190f20bf0debd410348b73408ec3143 Implement GetSolvingProvider for DescriptorScriptPubKeyMan (Andrew Chow) f1ca5feb4ad668a3e1ae543d0addd5f483f1a88f Implement GetKeypoolOldestTime and only display it if greater than 0 (Andrew Chow) 586b57a9a6b4b12a78f792785b63a5a1743bce0c Implement ReturnDestination in DescriptorScriptPubKeyMan (Andrew Chow) f866957979c23cefd41efa9dae9e53b9177818dc Implement GetReservedDestination in DescriptorScriptPubKeyMan (Andrew Chow) a775f7c7fd0b9094fcbeee6ba92206d5bbb19164 Implement Unlock and Encrypt in DescriptorScriptPubKeyMan (Andrew Chow) bfdd0734869a22217c15858d7a76d0dacc2ebc86 Implement GetNewDestination for DescriptorScriptPubKeyMan (Andrew Chow) 58c7651821b0eeff0a99dc61d78d2e9e07986580 Implement TopUp in DescriptorScriptPubKeyMan (Andrew Chow) e014886a342508f7c8d80323eee9a5f314eaf94c Implement SetupGeneration for DescriptorScriptPubKeyMan (Andrew Chow) 46dfb99768e7d03a3cf552812d5b41ceaebc06be Implement writing descriptorkeys, descriptorckeys, and descriptors to wallet file (Andrew Chow) 4cb9b69be031e1dc65d8964794781b347fd948f5 Implement several simple functions in DescriptorScriptPubKeyMan (Andrew Chow) d1ec3e4f19487b4b100f80ad02eac063c571777d Add IsSingleType to Descriptors (Andrew Chow) 953feb3d2724f5398dd48990c4957a19313d2c8c Implement loading of keys for DescriptorScriptPubKeyMan (Andrew Chow) 2363e9fcaa41b68bf11153f591b95f2d41ff9a1a Load the descriptor cache from the wallet file (Andrew Chow) 46c46aebb7943e1e2e96755e94dc6c197920bf75 Implement GetID for DescriptorScriptPubKeyMan (Andrew Chow) ec2f9e1178c8e38c0a5ca063fe81adac8f916348 Implement IsHDEnabled in DescriptorScriptPubKeyMan (Andrew Chow) 741122d4c1a62ced3e96d16d67f4eeb3a6522d99 Implement MarkUnusedAddresses in DescriptorScriptPubKeyMan (Andrew Chow) 2db7ca765c8fb2c71dd6f7c4f29ad70e68ff1720 Implement IsMine for DescriptorScriptPubKeyMan (Andrew Chow) db7177af8c159abbcc209f2caafcd45d54c181c5 Add LoadDescriptorScriptPubKeyMan and SetActiveScriptPubKeyMan to CWallet (Andrew Chow) 78f8a92910d34247fa5d04368338c598d9908267 Implement SetType in DescriptorScriptPubKeyMan (Andrew Chow) 834de0300cde57ca3f662fb7aa5b1bdaed68bc8f Store WalletDescriptor in DescriptorScriptPubKeyMan (Andrew Chow) d8132669e10c1db9ae0c2ea0d3f822d7d2f01345 Add a lock cs_desc_man for DescriptorScriptPubKeyMan (Andrew Chow) 3194a7f88ac1a32997b390b4f188c4b6a4af04a5 Introduce WalletDescriptor class (Andrew Chow) 6b13cd3fa854dfaeb9e269bff3d67cacc0e5b5dc Create LegacyScriptPubKeyMan when not a descriptor wallet (Andrew Chow) aeac157c9dc141546b45e06ba9c2e641ad86083f Return nullptr from GetLegacyScriptPubKeyMan if descriptor wallet (Andrew Chow) 96accc73f067c7c95946e9932645dd821ef67f63 Add WALLET_FLAG_DESCRIPTORS (Andrew Chow) 6b8119af53ee2fdb4c4b5b24b4e650c0dc3bd27c Introduce DescriptorScriptPubKeyMan as a dummy class (Andrew Chow) 06620302c713cae65ee8e4ff9302e4c88e2a1285 Introduce SetType function to tell ScriptPubKeyMans the type and internal-ness of it (Andrew Chow) Pull request description: Introducing the wallet of the glorious future (again): native descriptor wallets. With native descriptor wallets, addresses are generated from descriptors. Instead of generating keys and deriving addresses from keys, addresses come from the scriptPubKeys produced by a descriptor. Native descriptor wallets will be optional for now and can only be created by using `createwallet`. Descriptor wallets will store descriptors, master keys from the descriptor, and descriptor cache entries. Keys are derived from descriptors on the fly. In order to allow choosing different address types, 6 descriptors are needed for normal use. There is a pair of primary and change descriptors for each of the 3 address types. With the default keypool size of 1000, each descriptor has 1000 scriptPubKeys and descriptor cache entries pregenerated. This has a side effect of making wallets large since 6000 pubkeys are written to the wallet by default, instead of the current 2000. scriptPubKeys are kept only in memory and are generated every time a descriptor is loaded. By default, we use the standard BIP 44, 49, 84 derivation paths with an external and internal derivation chain for each. Descriptors can also be imported with a new `importdescriptors` RPC. Native descriptor wallets use the `ScriptPubKeyMan` interface introduced in #16341 to add a `DescriptorScriptPubKeyMan`. This defines a different IsMine which uses the simpler model of "does this scriptPubKey exist in this wallet". Furthermore, `DescriptorScriptPubKeyMan` does not have watchonly, so with native descriptor wallets, it is not possible to have a wallet with both watchonly and non-watchonly things. Rather a wallet with `disable_private_keys` needs to be used for watchonly things. A `--descriptor` option was added to some tests (`wallet_basic.py`, `wallet_encryption.py`, `wallet_keypool.py`, `wallet_keypool_topup.py`, and `wallet_labels.py`) to allow for these tests to use descriptor wallets. Additionally, several RPCs are disabled for descriptor wallets (`importprivkey`, `importpubkey`, `importaddress`, `importmulti`, `addmultisigaddress`, `dumpprivkey`, `dumpwallet`, `importwallet`, and `sethdseed`). ACKs for top commit: Sjors: utACK 223588b1bbc63dc57098bbd0baa48635e0cc0b82 (rebased, nits addressed) jonatack: Code review re-ACK 223588b1bbc63dc57098bbd0baa48635e0cc0b82. fjahr: re-ACK 223588b1bbc63dc57098bbd0baa48635e0cc0b82 instagibbs: light re-ACK 223588b meshcollider: Code review ACK 223588b1bbc63dc57098bbd0baa48635e0cc0b82 Tree-SHA512: 59bc52aeddbb769ed5f420d5d240d8137847ac821b588eb616b34461253510c1717d6a70bab8765631738747336ae06f45ba39603ccd17f483843e5ed9a90986 Introduce SetType function to tell ScriptPubKeyMans the type and internal-ness of it Introduce DescriptorScriptPubKeyMan as a dummy class Add WALLET_FLAG_DESCRIPTORS Return nullptr from GetLegacyScriptPubKeyMan if descriptor wallet Create LegacyScriptPubKeyMan when not a descriptor wallet Introduce WalletDescriptor class WalletDescriptor is a Descriptor with other wallet metadata Add a lock cs_desc_man for DescriptorScriptPubKeyMan Store WalletDescriptor in DescriptorScriptPubKeyMan Implement SetType in DescriptorScriptPubKeyMan Add LoadDescriptorScriptPubKeyMan and SetActiveScriptPubKeyMan to CWallet Implement IsMine for DescriptorScriptPubKeyMan Adds a set of scriptPubKeys that DescriptorScriptPubKeyMan tracks. If the given script is in that set, it is considered ISMINE_SPENDABLE Implement MarkUnusedAddresses in DescriptorScriptPubKeyMan Implement IsHDEnabled in DescriptorScriptPubKeyMan Implement GetID for DescriptorScriptPubKeyMan Load the descriptor cache from the wallet file Implement loading of keys for DescriptorScriptPubKeyMan Add IsSingleType to Descriptors IsSingleType will return whether the descriptor will give one or multiple scriptPubKeys Implement several simple functions in DescriptorScriptPubKeyMan Implements a bunch of one liners: UpgradeKeyMetadata, IsFirstRun, HavePrivateKeys, KeypoolCountExternalKeys, GetKeypoolSize, GetTimeFirstKey, CanGetAddresses, RewriteDB Implement writing descriptorkeys, descriptorckeys, and descriptors to wallet file Implement SetupGeneration for DescriptorScriptPubKeyMan Implement TopUp in DescriptorScriptPubKeyMan Implement GetNewDestination for DescriptorScriptPubKeyMan Implement Unlock and Encrypt in DescriptorScriptPubKeyMan Implement GetReservedDestination in DescriptorScriptPubKeyMan Implement ReturnDestination in DescriptorScriptPubKeyMan Implement GetKeypoolOldestTime and only display it if greater than 0 Implement GetSolvingProvider for DescriptorScriptPubKeyMan Internally, a GetSigningProvider function is introduced which allows for some private keys to be optionally included. This can be called with a script as the argument (i.e. a scriptPubKey from our wallet when we are signing) or with a pubkey. In order to know what index to expand the private keys for that pubkey, we need to also cache all of the pubkeys involved when we expand the descriptor. So SetCache and TopUp are updated to do this too. Implement SignTransaction in DescriptorScriptPubKeyMan Implement SignMessage for descriptor wallets Implement FillPSBT in DescriptorScriptPubKeyMan FillPSBT will add our own scripts to the PSBT if those inputs are ours. If an input also lists pubkeys that we happen to know the private keys for, we will sign those inputs too. Change GetMetadata to use unique_ptr<CKeyMetadata> Implement GetMetadata in DescriptorScriptPubKeyMan Be able to create new wallets with DescriptorScriptPubKeyMans as backing Generate new descriptors when encrypting Add IsLegacy to CWallet so that the GUI knows whether to show watchonly add importdescriptors RPC and tests for native descriptor wallets Co-authored-by: Andrew Chow <achow101-github@achow101.com> Functional tests for descriptor wallets Change wallet_encryption.py to use signmessage instead of dumpprivkey Return error when no ScriptPubKeyMan is available for specified type When a CWallet doesn't have a ScriptPubKeyMan for the requested type in GetNewDestination, give a meaningful error. Also handle this in Qt which did not do anything with errors. Implement CWallet::IsSpentKey for non-LegacySPKMans tests: Add RPCOverloadWrapper which overloads some disabled RPCs RPCOverloadWrapper overloads some deprecated or disabled RPCs with an implementation using other RPCs to avoid having a ton of code churn around replacing those RPCs. Add a --descriptors option to various tests Adds a --descriptors option globally to the test framework. This will make the test create and use descriptor wallets. However some tests may not work with this. Some tests are modified to work with --descriptors and run with that option in test_runer: * wallet_basic.py * wallet_encryption.py * wallet_keypool.py <---- wallet_keypool_hd.py actually * wallet_keypool_topup.py * wallet_labels.py * wallet_avoidreuse.py
826 lines
33 KiB
Python
Executable File
826 lines
33 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# Copyright (c) 2014-2020 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
"""Run regression test suite.
|
|
|
|
This module calls down into individual test cases via subprocess. It will
|
|
forward all unrecognized arguments onto the individual test scripts.
|
|
|
|
For a description of arguments recognized by test scripts, see
|
|
`test/functional/test_framework/test_framework.py:BitcoinTestFramework.main`.
|
|
|
|
"""
|
|
|
|
import argparse
|
|
from collections import deque
|
|
import configparser
|
|
import datetime
|
|
import os
|
|
import time
|
|
import shutil
|
|
import signal
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
import re
|
|
import logging
|
|
import unittest
|
|
|
|
# Formatting. Default colors to empty strings.
|
|
DEFAULT, BOLD, GREEN, RED = ("", ""), ("", ""), ("", ""), ("", "")
|
|
try:
|
|
# Make sure python thinks it can write unicode to its stdout
|
|
"\u2713".encode("utf_8").decode(sys.stdout.encoding)
|
|
TICK = "✓ "
|
|
CROSS = "✖ "
|
|
CIRCLE = "○ "
|
|
except UnicodeDecodeError:
|
|
TICK = "P "
|
|
CROSS = "x "
|
|
CIRCLE = "o "
|
|
|
|
if os.name != 'nt' or sys.getwindowsversion() >= (10, 0, 14393):
|
|
if os.name == 'nt':
|
|
import ctypes
|
|
kernel32 = ctypes.windll.kernel32 # type: ignore
|
|
ENABLE_VIRTUAL_TERMINAL_PROCESSING = 4
|
|
STD_OUTPUT_HANDLE = -11
|
|
STD_ERROR_HANDLE = -12
|
|
# Enable ascii color control to stdout
|
|
stdout = kernel32.GetStdHandle(STD_OUTPUT_HANDLE)
|
|
stdout_mode = ctypes.c_int32()
|
|
kernel32.GetConsoleMode(stdout, ctypes.byref(stdout_mode))
|
|
kernel32.SetConsoleMode(stdout, stdout_mode.value | ENABLE_VIRTUAL_TERMINAL_PROCESSING)
|
|
# Enable ascii color control to stderr
|
|
stderr = kernel32.GetStdHandle(STD_ERROR_HANDLE)
|
|
stderr_mode = ctypes.c_int32()
|
|
kernel32.GetConsoleMode(stderr, ctypes.byref(stderr_mode))
|
|
kernel32.SetConsoleMode(stderr, stderr_mode.value | ENABLE_VIRTUAL_TERMINAL_PROCESSING)
|
|
# primitive formatting on supported
|
|
# terminal via ANSI escape sequences:
|
|
DEFAULT = ('\033[0m', '\033[0m')
|
|
BOLD = ('\033[0m', '\033[1m')
|
|
GREEN = ('\033[0m', '\033[0;32m')
|
|
RED = ('\033[0m', '\033[0;31m')
|
|
|
|
TEST_EXIT_PASSED = 0
|
|
TEST_EXIT_SKIPPED = 77
|
|
|
|
# List of framework modules containing unit tests. Should be kept in sync with
|
|
# the output of `git grep unittest.TestCase ./test/functional/test_framework`
|
|
TEST_FRAMEWORK_MODULES = [
|
|
"address",
|
|
"crypto.bip324_cipher",
|
|
"blocktools",
|
|
"crypto.chacha20",
|
|
"crypto.ellswift",
|
|
"key",
|
|
"crypto.muhash",
|
|
"crypto.poly1305",
|
|
"crypto.ripemd160",
|
|
"script",
|
|
]
|
|
|
|
EXTENDED_SCRIPTS = [
|
|
# These tests are not run by default.
|
|
# Longest test should go first, to favor running tests in parallel
|
|
'feature_pruning.py', # NOTE: Prune mode is incompatible with -txindex, should work with governance validation disabled though.
|
|
'feature_dbcrash.py',
|
|
]
|
|
|
|
BASE_SCRIPTS = [
|
|
# Scripts that are run by default.
|
|
# Longest test should go first, to favor running tests in parallel
|
|
'feature_dip3_deterministicmns.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_data_recovery.py',
|
|
'wallet_hd.py',
|
|
'wallet_hd.py --descriptors',
|
|
'wallet_backup.py',
|
|
# vv Tests less than 5m vv
|
|
'mining_getblocktemplate_longpoll.py', # FIXME: "socket.error: [Errno 54] Connection reset by peer" on my Mac, same as https://github.com/bitcoin/bitcoin/issues/6651
|
|
'feature_maxuploadtarget.py',
|
|
'feature_block.py', # NOTE: needs dash_hash to pass
|
|
'rpc_fundrawtransaction.py',
|
|
'rpc_fundrawtransaction.py --nohd',
|
|
'wallet_multiwallet.py --usecli',
|
|
'p2p_quorum_data.py',
|
|
# vv Tests less than 2m vv
|
|
'p2p_instantsend.py',
|
|
'wallet_basic.py',
|
|
'wallet_basic.py --descriptors',
|
|
'wallet_labels.py',
|
|
'wallet_labels.py --descriptors',
|
|
'p2p_timeouts.py',
|
|
'feature_bip68_sequence.py',
|
|
'mempool_updatefromblock.py',
|
|
'p2p_tx_download.py',
|
|
'wallet_dump.py',
|
|
'wallet_listtransactions.py',
|
|
'feature_multikeysporks.py',
|
|
'feature_dip3_v19.py',
|
|
'feature_llmq_signing.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_signing.py --spork21', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_chainlocks.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_rotation.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_connections.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_evo.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_simplepose.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_is_cl_conflicts.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_is_retroactive.py', # NOTE: needs dash_hash to pass
|
|
'feature_llmq_dkgerrors.py', # NOTE: needs dash_hash to pass
|
|
'feature_dip4_coinbasemerkleroots.py', # NOTE: needs dash_hash to pass
|
|
'feature_asset_locks.py', # NOTE: needs dash_hash to pass
|
|
'feature_mnehf.py', # NOTE: needs dash_hash to pass
|
|
# vv Tests less than 60s vv
|
|
'p2p_sendheaders.py', # NOTE: needs dash_hash to pass
|
|
'p2p_sendheaders_compressed.py', # NOTE: needs dash_hash to pass
|
|
'wallet_importmulti.py',
|
|
'mempool_limit.py',
|
|
'rpc_txoutproof.py',
|
|
'wallet_listreceivedby.py',
|
|
'wallet_abandonconflict.py',
|
|
'feature_csv_activation.py',
|
|
'rpc_rawtransaction.py',
|
|
'feature_reindex.py',
|
|
'feature_abortnode.py',
|
|
# vv Tests less than 30s vv
|
|
'rpc_quorum.py',
|
|
'wallet_keypool_topup.py',
|
|
'wallet_keypool_topup.py --descriptors',
|
|
'feature_fee_estimation.py',
|
|
'interface_zmq_dash.py',
|
|
'interface_zmq.py',
|
|
'rpc_invalid_address_message.py',
|
|
'interface_bitcoin_cli.py',
|
|
'mempool_resurrect.py',
|
|
'wallet_txn_doublespend.py --mineblock',
|
|
'tool_wallet.py',
|
|
'wallet_txn_clone.py',
|
|
'rpc_getchaintips.py',
|
|
'rpc_misc.py',
|
|
'interface_rest.py',
|
|
'mempool_spend_coinbase.py',
|
|
'wallet_avoidreuse.py',
|
|
'wallet_avoidreuse.py --descriptors',
|
|
'mempool_reorg.py',
|
|
'mempool_persist.py',
|
|
'wallet_multiwallet.py',
|
|
'wallet_createwallet.py',
|
|
'wallet_createwallet.py --usecli',
|
|
'wallet_reorgsrestore.py',
|
|
'wallet_watchonly.py',
|
|
'wallet_watchonly.py --usecli',
|
|
'interface_http.py',
|
|
'interface_rpc.py',
|
|
'rpc_psbt.py',
|
|
'rpc_psbt.py --descriptors',
|
|
'rpc_users.py',
|
|
'rpc_whitelist.py',
|
|
'feature_proxy.py',
|
|
'rpc_signrawtransaction.py',
|
|
'p2p_addrv2_relay.py',
|
|
'wallet_groups.py',
|
|
'p2p_disconnect_ban.py',
|
|
'feature_addressindex.py',
|
|
'feature_timestampindex.py',
|
|
'feature_spentindex.py',
|
|
'rpc_decodescript.py',
|
|
'rpc_blockchain.py',
|
|
'rpc_deprecated.py',
|
|
'wallet_disable.py',
|
|
'p2p_addr_relay.py',
|
|
'p2p_getaddr_caching.py',
|
|
'p2p_getdata.py',
|
|
'rpc_net.py',
|
|
'wallet_keypool.py',
|
|
'wallet_keypool_hd.py',
|
|
'wallet_keypool_hd.py --descriptors',
|
|
'wallet_descriptor.py',
|
|
'p2p_nobloomfilter_messages.py',
|
|
'p2p_filter.py',
|
|
'p2p_blocksonly.py',
|
|
'rpc_setban.py',
|
|
'mining_prioritisetransaction.py',
|
|
'p2p_invalid_locator.py',
|
|
'p2p_invalid_block.py',
|
|
'p2p_invalid_messages.py',
|
|
'p2p_invalid_tx.py',
|
|
'feature_assumevalid.py',
|
|
'example_test.py',
|
|
'wallet_txn_doublespend.py',
|
|
'feature_backwards_compatibility.py',
|
|
'wallet_txn_clone.py --mineblock',
|
|
'feature_notifications.py',
|
|
'rpc_getblockfilter.py',
|
|
'rpc_invalidateblock.py',
|
|
'feature_txindex.py',
|
|
'feature_utxo_set_hash.py',
|
|
'mempool_packages.py',
|
|
'mempool_package_onemore.py',
|
|
'rpc_createmultisig.py',
|
|
'rpc_createmultisig.py --descriptors',
|
|
'rpc_packages.py',
|
|
'feature_versionbits_warning.py',
|
|
'rpc_preciousblock.py',
|
|
'wallet_importprunedfunds.py',
|
|
'p2p_leak_tx.py',
|
|
'p2p_eviction.py',
|
|
'rpc_signmessage.py',
|
|
'rpc_generateblock.py',
|
|
'wallet_balance.py',
|
|
'feature_nulldummy.py',
|
|
'mempool_accept.py',
|
|
'mempool_expiry.py',
|
|
'wallet_import_rescan.py',
|
|
'wallet_import_with_label.py',
|
|
'wallet_upgradewallet.py',
|
|
'wallet_importdescriptors.py',
|
|
'wallet_mnemonicbits.py',
|
|
'rpc_bind.py --ipv4',
|
|
'rpc_bind.py --ipv6',
|
|
'rpc_bind.py --nonloopback',
|
|
'mining_basic.py',
|
|
'rpc_named_arguments.py',
|
|
'wallet_listsinceblock.py',
|
|
'p2p_leak.py',
|
|
'p2p_compactblocks.py',
|
|
'p2p_connect_to_devnet.py',
|
|
'feature_sporks.py',
|
|
'rpc_getblockstats.py',
|
|
'wallet_encryption.py',
|
|
'wallet_encryption.py --descriptors',
|
|
'wallet_upgradetohd.py',
|
|
'feature_dersig.py',
|
|
'feature_cltv.py',
|
|
'feature_new_quorum_type_activation.py',
|
|
'feature_governance_objects.py',
|
|
'feature_governance.py',
|
|
'rpc_uptime.py',
|
|
'wallet_resendwallettransactions.py',
|
|
'wallet_fallbackfee.py',
|
|
'rpc_dumptxoutset.py',
|
|
'feature_minchainwork.py',
|
|
'rpc_estimatefee.py',
|
|
'p2p_unrequested_blocks.py', # NOTE: needs dash_hash to pass
|
|
'feature_shutdown.py',
|
|
'rpc_coinjoin.py',
|
|
'rpc_masternode.py',
|
|
'rpc_mnauth.py',
|
|
'rpc_verifyislock.py',
|
|
'rpc_verifychainlock.py',
|
|
'wallet_create_tx.py',
|
|
'p2p_fingerprint.py',
|
|
'rpc_platform_filter.py',
|
|
'rpc_wipewallettxes.py',
|
|
'feature_dip0020_activation.py',
|
|
'feature_uacomment.py',
|
|
'wallet_coinbase_category.py',
|
|
'feature_filelock.py',
|
|
'feature_loadblock.py',
|
|
'p2p_blockfilters.py',
|
|
'p2p_message_capture.py',
|
|
'feature_asmap.py',
|
|
'feature_includeconf.py',
|
|
'mempool_unbroadcast.py',
|
|
'mempool_compatibility.py',
|
|
'rpc_deriveaddresses.py',
|
|
'rpc_deriveaddresses.py --usecli',
|
|
'p2p_ping.py',
|
|
'rpc_scantxoutset.py',
|
|
'feature_logging.py',
|
|
'feature_coinstatsindex.py',
|
|
'wallet_orphanedreward.py',
|
|
'p2p_node_network_limited.py',
|
|
'p2p_permissions.py',
|
|
'feature_blocksdir.py',
|
|
'wallet_startup.py',
|
|
'p2p_i2p_ports.py',
|
|
'feature_config_args.py',
|
|
'feature_settings.py',
|
|
'rpc_getdescriptorinfo.py',
|
|
'rpc_getaddressinfo_labels_purpose_deprecation.py',
|
|
'rpc_getaddressinfo_label_deprecation.py',
|
|
'rpc_help.py',
|
|
'feature_help.py',
|
|
'feature_blockfilterindex_prune.py'
|
|
# Don't append tests at the end to avoid merge conflicts
|
|
# Put them in a random line within the section that fits their approximate run-time
|
|
]
|
|
|
|
# Place EXTENDED_SCRIPTS first since it has the 3 longest running tests
|
|
ALL_SCRIPTS = EXTENDED_SCRIPTS + BASE_SCRIPTS
|
|
|
|
NON_SCRIPTS = [
|
|
# These are python files that live in the functional tests directory, but are not test scripts.
|
|
"combine_logs.py",
|
|
"create_cache.py",
|
|
"test_runner.py",
|
|
]
|
|
|
|
def main():
|
|
# Parse arguments and pass through unrecognised args
|
|
parser = argparse.ArgumentParser(add_help=False,
|
|
usage='%(prog)s [test_runner.py options] [script options] [scripts]',
|
|
description=__doc__,
|
|
epilog='''
|
|
Help text and arguments for individual test script:''',
|
|
formatter_class=argparse.RawTextHelpFormatter)
|
|
parser.add_argument('--ansi', action='store_true', default=sys.stdout.isatty(), help="Use ANSI colors and dots in output (enabled by default when standard output is a TTY)")
|
|
parser.add_argument('--attempts', '-a', type=int, default=1, help='how many attempts should be allowed for the non-deterministic test suite. Default=1.')
|
|
parser.add_argument('--combinedlogslen', '-c', type=int, default=0, metavar='n', help='On failure, print a log (of length n lines) to the console, combined from the test framework and all test nodes.')
|
|
parser.add_argument('--coverage', action='store_true', help='generate a basic coverage report for the RPC interface')
|
|
parser.add_argument('--ci', action='store_true', help='Run checks and code that are usually only enabled in a continuous integration environment')
|
|
parser.add_argument('--exclude', '-x', help='specify a comma-separated-list of scripts to exclude.')
|
|
parser.add_argument('--extended', action='store_true', help='run the extended test suite in addition to the basic tests')
|
|
parser.add_argument('--help', '-h', '-?', action='store_true', help='print help text and exit')
|
|
parser.add_argument('--jobs', '-j', type=int, default=4, help='how many test scripts to run in parallel. Default=4.')
|
|
parser.add_argument('--keepcache', '-k', action='store_true', help='the default behavior is to flush the cache directory on startup. --keepcache retains the cache from the previous testrun.')
|
|
parser.add_argument('--quiet', '-q', action='store_true', help='only print dots, results summary and failure logs')
|
|
parser.add_argument('--tmpdirprefix', '-t', default=tempfile.gettempdir(), help="Root directory for datadirs")
|
|
parser.add_argument('--failfast', '-F', action='store_true', help='stop execution after the first test failure')
|
|
parser.add_argument('--filter', help='filter scripts to run by regular expression')
|
|
|
|
args, unknown_args = parser.parse_known_args()
|
|
if not args.ansi:
|
|
global DEFAULT, BOLD, GREEN, RED
|
|
DEFAULT = ("", "")
|
|
BOLD = ("", "")
|
|
GREEN = ("", "")
|
|
RED = ("", "")
|
|
|
|
# args to be passed on always start with two dashes; tests are the remaining unknown args
|
|
tests = [arg for arg in unknown_args if arg[:2] != "--"]
|
|
passon_args = [arg for arg in unknown_args if arg[:2] == "--"]
|
|
|
|
# Read config generated by configure.
|
|
config = configparser.ConfigParser()
|
|
configfile = os.path.abspath(os.path.dirname(__file__)) + "/../config.ini"
|
|
config.read_file(open(configfile, encoding="utf8"))
|
|
|
|
passon_args.append("--configfile=%s" % configfile)
|
|
|
|
# Set up logging
|
|
logging_level = logging.INFO if args.quiet else logging.DEBUG
|
|
logging.basicConfig(format='%(message)s', level=logging_level)
|
|
|
|
# Create base test directory
|
|
tmpdir = "%s/test_runner_∋_🏃_%s" % (args.tmpdirprefix, datetime.datetime.now().strftime("%Y%m%d_%H%M%S"))
|
|
|
|
os.makedirs(tmpdir)
|
|
|
|
logging.debug("Temporary test directory at %s" % tmpdir)
|
|
|
|
enable_bitcoind = config["components"].getboolean("ENABLE_BITCOIND")
|
|
|
|
if not enable_bitcoind:
|
|
print("No functional tests to run.")
|
|
print("Rerun ./configure with --with-daemon and then make")
|
|
sys.exit(0)
|
|
|
|
# Build list of tests
|
|
test_list = []
|
|
if tests:
|
|
# Individual tests have been specified. Run specified tests that exist
|
|
# in the ALL_SCRIPTS list. Accept names with or without a .py extension.
|
|
# Specified tests can contain wildcards, but in that case the supplied
|
|
# paths should be coherent, e.g. the same path as that provided to call
|
|
# test_runner.py. Examples:
|
|
# `test/functional/test_runner.py test/functional/wallet*`
|
|
# `test/functional/test_runner.py ./test/functional/wallet*`
|
|
# `test_runner.py wallet*`
|
|
# but not:
|
|
# `test/functional/test_runner.py wallet*`
|
|
# Multiple wildcards can be passed:
|
|
# `test_runner.py tool* mempool*`
|
|
for test in tests:
|
|
script = test.split("/")[-1]
|
|
script = script + ".py" if ".py" not in script else script
|
|
matching_scripts = [s for s in ALL_SCRIPTS if s.startswith(script)]
|
|
if matching_scripts:
|
|
test_list.extend(matching_scripts)
|
|
else:
|
|
print("{}WARNING!{} Test '{}' not found in full test list.".format(BOLD[1], BOLD[0], test))
|
|
elif args.extended:
|
|
# Include extended tests
|
|
test_list += ALL_SCRIPTS
|
|
else:
|
|
# Run base tests only
|
|
test_list += BASE_SCRIPTS
|
|
|
|
# Remove the test cases that the user has explicitly asked to exclude.
|
|
if args.exclude:
|
|
exclude_tests = [test.split('.py')[0] for test in args.exclude.split(',')]
|
|
for exclude_test in exclude_tests:
|
|
# Remove <test_name>.py and <test_name>.py --arg from the test list
|
|
exclude_list = [test for test in test_list if test.split('.py')[0] == exclude_test]
|
|
for exclude_item in exclude_list:
|
|
test_list.remove(exclude_item)
|
|
if not exclude_list:
|
|
print("{}WARNING!{} Test '{}' not found in current test list.".format(BOLD[1], BOLD[0], exclude_test))
|
|
|
|
if args.filter:
|
|
test_list = list(filter(re.compile(args.filter).search, test_list))
|
|
|
|
if not test_list:
|
|
print("No valid test scripts specified. Check that your test is in one "
|
|
"of the test lists in test_runner.py, or run test_runner.py with no arguments to run all tests")
|
|
sys.exit(0)
|
|
|
|
if args.help:
|
|
# Print help for test_runner.py, then print help of the first script (with args removed) and exit.
|
|
parser.print_help()
|
|
subprocess.check_call([sys.executable, os.path.join(config["environment"]["SRCDIR"], 'test', 'functional', test_list[0].split()[0]), '-h'])
|
|
sys.exit(0)
|
|
|
|
check_script_list(src_dir=config["environment"]["SRCDIR"], fail_on_warn=args.ci)
|
|
check_script_prefixes()
|
|
|
|
if not args.keepcache:
|
|
shutil.rmtree("%s/test/cache" % config["environment"]["BUILDDIR"], ignore_errors=True)
|
|
|
|
run_tests(
|
|
test_list=test_list,
|
|
src_dir=config["environment"]["SRCDIR"],
|
|
build_dir=config["environment"]["BUILDDIR"],
|
|
tmpdir=tmpdir,
|
|
jobs=args.jobs,
|
|
attempts=args.attempts,
|
|
enable_coverage=args.coverage,
|
|
args=passon_args,
|
|
combined_logs_len=args.combinedlogslen,
|
|
failfast=args.failfast,
|
|
use_term_control=args.ansi,
|
|
)
|
|
|
|
def run_tests(*, test_list, src_dir, build_dir, tmpdir, jobs=1, attempts=1, enable_coverage=False, args=None, combined_logs_len=0,failfast=False, use_term_control):
|
|
args = args or []
|
|
|
|
# Warn if dashd is already running
|
|
try:
|
|
# pgrep exits with code zero when one or more matching processes found
|
|
if subprocess.run(["pgrep", "-x", "dashd"], stdout=subprocess.DEVNULL).returncode == 0:
|
|
print("%sWARNING!%s There is already a dashd process running on this system. Tests may fail unexpectedly due to resource contention!" % (BOLD[1], BOLD[0]))
|
|
except OSError:
|
|
# pgrep not supported
|
|
pass
|
|
|
|
# Warn if there is a cache directory
|
|
cache_dir = "%s/test/cache" % build_dir
|
|
if os.path.isdir(cache_dir):
|
|
print("%sWARNING!%s There is a cache directory here: %s. If tests fail unexpectedly, try deleting the cache directory." % (BOLD[1], BOLD[0], cache_dir))
|
|
|
|
# Test Framework Tests
|
|
print("Running Unit Tests for Test Framework Modules")
|
|
test_framework_tests = unittest.TestSuite()
|
|
for module in TEST_FRAMEWORK_MODULES:
|
|
test_framework_tests.addTest(unittest.TestLoader().loadTestsFromName("test_framework.{}".format(module)))
|
|
result = unittest.TextTestRunner(verbosity=1, failfast=True).run(test_framework_tests)
|
|
if not result.wasSuccessful():
|
|
logging.debug("Early exiting after failure in TestFramework unit tests")
|
|
sys.exit(False)
|
|
|
|
tests_dir = src_dir + '/test/functional/'
|
|
|
|
flags = ['--cachedir={}'.format(cache_dir)] + args
|
|
|
|
if enable_coverage:
|
|
coverage = RPCCoverage()
|
|
flags.append(coverage.flag)
|
|
logging.debug("Initializing coverage directory at %s" % coverage.dir)
|
|
else:
|
|
coverage = None
|
|
|
|
if len(test_list) > 1 and jobs > 1:
|
|
# Populate cache
|
|
try:
|
|
subprocess.check_output([sys.executable, tests_dir + 'create_cache.py'] + flags + ["--tmpdir=%s/cache" % tmpdir])
|
|
except subprocess.CalledProcessError as e:
|
|
sys.stdout.buffer.write(e.output)
|
|
raise
|
|
|
|
#Run Tests
|
|
job_queue = TestHandler(
|
|
num_tests_parallel=jobs,
|
|
tests_dir=tests_dir,
|
|
tmpdir=tmpdir,
|
|
test_list=test_list,
|
|
flags=flags,
|
|
use_term_control=use_term_control,
|
|
attempts=attempts,
|
|
)
|
|
start_time = time.time()
|
|
test_results = []
|
|
|
|
max_len_name = len(max(test_list, key=len))
|
|
test_count = len(test_list)
|
|
for i in range(test_count):
|
|
test_result, testdir, stdout, stderr = job_queue.get_next()
|
|
test_results.append(test_result)
|
|
done_str = "{}/{} - {}{}{}".format(i + 1, test_count, BOLD[1], test_result.name, BOLD[0])
|
|
if test_result.status == "Passed":
|
|
logging.debug("%s passed, Duration: %s s" % (done_str, test_result.time))
|
|
elif test_result.status == "Skipped":
|
|
logging.debug("%s skipped" % (done_str))
|
|
else:
|
|
print("%s failed, Duration: %s s\n" % (done_str, test_result.time))
|
|
print(BOLD[1] + 'stdout:\n' + BOLD[0] + stdout + '\n')
|
|
print(BOLD[1] + 'stderr:\n' + BOLD[0] + stderr + '\n')
|
|
if combined_logs_len and os.path.isdir(testdir):
|
|
# Print the final `combinedlogslen` lines of the combined logs
|
|
print('{}Combine the logs and print the last {} lines ...{}'.format(BOLD[1], combined_logs_len, BOLD[0]))
|
|
print('\n============')
|
|
print('{}Combined log for {}:{}'.format(BOLD[1], testdir, BOLD[0]))
|
|
print('============\n')
|
|
combined_logs_args = [sys.executable, os.path.join(tests_dir, 'combine_logs.py'), testdir]
|
|
if BOLD[0]:
|
|
combined_logs_args += ['--color']
|
|
combined_logs, _ = subprocess.Popen(combined_logs_args, universal_newlines=True, stdout=subprocess.PIPE).communicate()
|
|
print("\n".join(deque(combined_logs.splitlines(), combined_logs_len)))
|
|
|
|
if failfast:
|
|
logging.debug("Early exiting after test failure")
|
|
break
|
|
|
|
print_results(test_results, max_len_name, (int(time.time() - start_time)))
|
|
|
|
if coverage:
|
|
coverage_passed = coverage.report_rpc_coverage()
|
|
|
|
logging.debug("Cleaning up coverage data")
|
|
coverage.cleanup()
|
|
else:
|
|
coverage_passed = True
|
|
|
|
# Clear up the temp directory if all subdirectories are gone
|
|
if not os.listdir(tmpdir):
|
|
os.rmdir(tmpdir)
|
|
|
|
all_passed = all(map(lambda test_result: test_result.was_successful, test_results)) and coverage_passed
|
|
|
|
# Clean up dangling processes if any. This may only happen with --failfast option.
|
|
# Killing the process group will also terminate the current process but that is
|
|
# not an issue
|
|
if not os.getenv("CI_FAILFAST_TEST_LEAVE_DANGLING") and len(job_queue.jobs):
|
|
os.killpg(os.getpgid(0), signal.SIGKILL)
|
|
|
|
sys.exit(not all_passed)
|
|
|
|
|
|
def print_results(test_results, max_len_name, runtime):
|
|
results = "\n" + BOLD[1] + "%s | %s | %s\n\n" % ("TEST".ljust(max_len_name), "STATUS ", "DURATION") + BOLD[0]
|
|
|
|
test_results.sort(key=TestResult.sort_key)
|
|
all_passed = True
|
|
time_sum = 0
|
|
|
|
for test_result in test_results:
|
|
all_passed = all_passed and test_result.was_successful
|
|
time_sum += test_result.time
|
|
test_result.padding = max_len_name
|
|
results += str(test_result)
|
|
|
|
status = TICK + "Passed" if all_passed else CROSS + "Failed"
|
|
if not all_passed:
|
|
results += RED[1]
|
|
results += BOLD[1] + "\n%s | %s | %s s (accumulated) \n" % ("ALL".ljust(max_len_name), status.ljust(9), time_sum) + BOLD[0]
|
|
if not all_passed:
|
|
results += RED[0]
|
|
results += "Runtime: %s s\n" % (runtime)
|
|
print(results)
|
|
|
|
class TestHandler:
|
|
"""
|
|
Trigger the test scripts passed in via the list.
|
|
"""
|
|
|
|
def __init__(self, *, num_tests_parallel, tests_dir, tmpdir, test_list, flags, use_term_control, attempts):
|
|
assert num_tests_parallel >= 1
|
|
self.num_jobs = num_tests_parallel
|
|
self.tests_dir = tests_dir
|
|
self.tmpdir = tmpdir
|
|
self.test_list = test_list
|
|
self.flags = flags
|
|
self.num_running = 0
|
|
self.jobs = []
|
|
self.use_term_control = use_term_control
|
|
self.attempts = attempts
|
|
|
|
def get_next(self):
|
|
while self.num_running < self.num_jobs and self.test_list:
|
|
# Add tests
|
|
self.num_running += 1
|
|
test = self.test_list.pop(0)
|
|
portseed = len(self.test_list)
|
|
portseed_arg = ["--portseed={}".format(portseed)]
|
|
log_stdout = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
log_stderr = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
test_argv = test.split()
|
|
testdir = "{}/{}_{}".format(self.tmpdir, re.sub(".py$", "", test_argv[0]), portseed)
|
|
tmpdir_arg = ["--tmpdir={}".format(testdir)]
|
|
self.jobs.append((test,
|
|
time.time(),
|
|
subprocess.Popen([sys.executable, self.tests_dir + test_argv[0]] + test_argv[1:] + self.flags + portseed_arg + tmpdir_arg,
|
|
universal_newlines=True,
|
|
stdout=log_stdout,
|
|
stderr=log_stderr),
|
|
testdir,
|
|
log_stdout,
|
|
log_stderr,
|
|
portseed,
|
|
1)) # attempt
|
|
if not self.jobs:
|
|
raise IndexError('pop from empty list')
|
|
|
|
# Print remaining running jobs when all jobs have been started.
|
|
if not self.test_list:
|
|
print("Remaining jobs: [{}]".format(", ".join(j[0] for j in self.jobs)))
|
|
|
|
dot_count = 0
|
|
while True:
|
|
# Return first proc that finishes
|
|
time.sleep(.5)
|
|
for job in self.jobs:
|
|
(name, start_time, proc, testdir, log_out, log_err, portseed, attempt) = job
|
|
if proc.poll() is not None:
|
|
log_out.seek(0), log_err.seek(0)
|
|
[stdout, stderr] = [log_file.read().decode('utf-8') for log_file in (log_out, log_err)]
|
|
log_out.close(), log_err.close()
|
|
if proc.returncode == TEST_EXIT_PASSED and stderr == "":
|
|
status = "Passed"
|
|
elif proc.returncode == TEST_EXIT_SKIPPED:
|
|
status = "Skipped"
|
|
elif attempt < self.attempts:
|
|
# cleanup
|
|
if self.use_term_control:
|
|
clearline = '\r' + (' ' * dot_count) + '\r'
|
|
print(clearline, end='', flush=True)
|
|
dot_count = 0
|
|
shutil.rmtree(testdir, ignore_errors=True)
|
|
self.jobs.remove(job)
|
|
print(f"{name} failed at attempt {attempt}/{self.attempts}, Duration: {int(time.time() - start_time)} s")
|
|
# start over
|
|
portseed_arg = ["--portseed={}".format(portseed)]
|
|
log_stdout = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
log_stderr = tempfile.SpooledTemporaryFile(max_size=2**16)
|
|
test_argv = name.split()
|
|
tmpdir_arg = ["--tmpdir={}".format(testdir)]
|
|
self.jobs.append((name,
|
|
time.time(),
|
|
subprocess.Popen([sys.executable, self.tests_dir + test_argv[0]] + test_argv[1:] + self.flags + portseed_arg + tmpdir_arg,
|
|
universal_newlines=True,
|
|
stdout=log_stdout,
|
|
stderr=log_stderr),
|
|
testdir,
|
|
log_stdout,
|
|
log_stderr,
|
|
portseed,
|
|
attempt + 1)) # attempt
|
|
# no results for now, move to the next job
|
|
continue
|
|
else:
|
|
status = "Failed"
|
|
self.num_running -= 1
|
|
self.jobs.remove(job)
|
|
if self.use_term_control:
|
|
clearline = '\r' + (' ' * dot_count) + '\r'
|
|
print(clearline, end='', flush=True)
|
|
dot_count = 0
|
|
return TestResult(name, status, int(time.time() - start_time)), testdir, stdout, stderr
|
|
if self.use_term_control:
|
|
print('.', end='', flush=True)
|
|
dot_count += 1
|
|
|
|
|
|
class TestResult():
|
|
def __init__(self, name, status, time):
|
|
self.name = name
|
|
self.status = status
|
|
self.time = time
|
|
self.padding = 0
|
|
|
|
def sort_key(self):
|
|
if self.status == "Passed":
|
|
return 0, self.name.lower()
|
|
elif self.status == "Failed":
|
|
return 2, self.name.lower()
|
|
elif self.status == "Skipped":
|
|
return 1, self.name.lower()
|
|
|
|
def __repr__(self):
|
|
if self.status == "Passed":
|
|
color = GREEN
|
|
glyph = TICK
|
|
elif self.status == "Failed":
|
|
color = RED
|
|
glyph = CROSS
|
|
elif self.status == "Skipped":
|
|
color = DEFAULT
|
|
glyph = CIRCLE
|
|
|
|
return color[1] + "%s | %s%s | %s s\n" % (self.name.ljust(self.padding), glyph, self.status.ljust(7), self.time) + color[0]
|
|
|
|
@property
|
|
def was_successful(self):
|
|
return self.status != "Failed"
|
|
|
|
|
|
def check_script_prefixes():
|
|
"""Check that test scripts start with one of the allowed name prefixes."""
|
|
|
|
good_prefixes_re = re.compile("^(example|feature|interface|mempool|mining|p2p|rpc|wallet|tool)_")
|
|
bad_script_names = [script for script in ALL_SCRIPTS if good_prefixes_re.match(script) is None]
|
|
|
|
if bad_script_names:
|
|
print("%sERROR:%s %d tests not meeting naming conventions:" % (BOLD[1], BOLD[0], len(bad_script_names)))
|
|
print(" %s" % ("\n ".join(sorted(bad_script_names))))
|
|
raise AssertionError("Some tests are not following naming convention!")
|
|
|
|
|
|
def check_script_list(*, src_dir, fail_on_warn):
|
|
"""Check scripts directory.
|
|
|
|
Check that there are no scripts in the functional tests directory which are
|
|
not being run by pull-tester.py."""
|
|
script_dir = src_dir + '/test/functional/'
|
|
python_files = set([test_file for test_file in os.listdir(script_dir) if test_file.endswith(".py")])
|
|
missed_tests = list(python_files - set(map(lambda x: x.split()[0], ALL_SCRIPTS + NON_SCRIPTS)))
|
|
if len(missed_tests) != 0:
|
|
print("%sWARNING!%s The following scripts are not being run: %s. Check the test lists in test_runner.py." % (BOLD[1], BOLD[0], str(missed_tests)))
|
|
if fail_on_warn:
|
|
# On CI this warning is an error to prevent merging incomplete commits into master
|
|
sys.exit(1)
|
|
|
|
|
|
class RPCCoverage():
|
|
"""
|
|
Coverage reporting utilities for test_runner.
|
|
|
|
Coverage calculation works by having each test script subprocess write
|
|
coverage files into a particular directory. These files contain the RPC
|
|
commands invoked during testing, as well as a complete listing of RPC
|
|
commands per `dash-cli help` (`rpc_interface.txt`).
|
|
|
|
After all tests complete, the commands run are combined and diff'd against
|
|
the complete list to calculate uncovered RPC commands.
|
|
|
|
See also: test/functional/test_framework/coverage.py
|
|
|
|
"""
|
|
def __init__(self):
|
|
self.dir = tempfile.mkdtemp(prefix="coverage")
|
|
self.flag = '--coveragedir=%s' % self.dir
|
|
|
|
def report_rpc_coverage(self):
|
|
"""
|
|
Print out RPC commands that were unexercised by tests.
|
|
|
|
"""
|
|
uncovered = self._get_uncovered_rpc_commands()
|
|
|
|
if uncovered:
|
|
print("Uncovered RPC commands:")
|
|
print("".join((" - %s\n" % command) for command in sorted(uncovered)))
|
|
return False
|
|
else:
|
|
print("All RPC commands covered.")
|
|
return True
|
|
|
|
def cleanup(self):
|
|
return shutil.rmtree(self.dir)
|
|
|
|
def _get_uncovered_rpc_commands(self):
|
|
"""
|
|
Return a set of currently untested RPC commands.
|
|
|
|
"""
|
|
# This is shared from `test/functional/test-framework/coverage.py`
|
|
reference_filename = 'rpc_interface.txt'
|
|
coverage_file_prefix = 'coverage.'
|
|
|
|
coverage_ref_filename = os.path.join(self.dir, reference_filename)
|
|
coverage_filenames = set()
|
|
all_cmds = set()
|
|
covered_cmds = set()
|
|
|
|
if not os.path.isfile(coverage_ref_filename):
|
|
raise RuntimeError("No coverage reference found")
|
|
|
|
with open(coverage_ref_filename, 'r', encoding="utf8") as coverage_ref_file:
|
|
all_cmds.update([line.strip() for line in coverage_ref_file.readlines()])
|
|
|
|
for root, dirs, files in os.walk(self.dir):
|
|
for filename in files:
|
|
if filename.startswith(coverage_file_prefix):
|
|
coverage_filenames.add(os.path.join(root, filename))
|
|
|
|
for filename in coverage_filenames:
|
|
with open(filename, 'r', encoding="utf8") as coverage_file:
|
|
covered_cmds.update([line.strip() for line in coverage_file.readlines()])
|
|
|
|
return all_cmds - covered_cmds
|
|
|
|
|
|
if __name__ == '__main__':
|
|
main()
|