mirror of
https://github.com/dashpay/dash.git
synced 2024-12-26 20:42:59 +01:00
9d33b30a87
dac7a111bdd3b0233d94cf68dae7a8bfc6ac9c64 refactor: test: use _ variable for unused loop counters (Sebastian Falbesoner)
Pull request description:
This tiny PR substitutes Python loops in the form of `for x in range(N): ...` by `for _ in range(N): ...` where applicable. The idea is indicating to the reader that a block (or statement, in list comprehensions) is just repeated N times, and that the loop counter is not used in the body, hence using the throwaway variable. This is already done quite often in the current tests (see e.g. `$ git grep "for _ in range("`). Another alternative would be using `itertools.repeat` (according to Python core developer Raymond Hettinger it's [even faster](https://twitter.com/raymondh/status/1144527183341375488)), but that doesn't seem to be widespread in use and I'm not sure about a readability increase.
The only drawback I see is that whenever one wants to debug loop iterations, one would need to introduce a loop variable again. Reviewing this is basically a no-brainer, since tests would fail immediately if a a substitution has taken place on a loop where the variable is used.
Instances to replace were found by `$ git grep "for.*in range("` and manually checked.
ACKs for top commit:
darosior:
ACK dac7a111bdd3b0233d94cf68dae7a8bfc6ac9c64
instagibbs:
manual inspection ACK dac7a111bd
practicalswift:
ACK dac7a111bdd3b0233d94cf68dae7a8bfc6ac9c64 -- the updated code is easier to reason about since the throwaway nature of a variable is expressed explicitly (using the Pythonic `_` idiom) instead of implicitly. Explicit is better than implicit was we all know by now :)
Tree-SHA512: 5f43ded9ce14e5e00b3876ec445b90acda1842f813149ae7bafa93f3ac3d510bb778e2c701187fd2c73585e6b87797bb2d2987139bd1a9ba7d58775a59392406
283 lines
11 KiB
Python
Executable File
283 lines
11 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
|
|
# Copyright (c) 2023 The Dash Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
|
|
import struct
|
|
from io import BytesIO
|
|
|
|
from test_framework.authproxy import JSONRPCException
|
|
from test_framework.key import ECKey
|
|
from test_framework.messages import (
|
|
CMnEhf,
|
|
CTransaction,
|
|
hash256,
|
|
ser_string,
|
|
)
|
|
|
|
from test_framework.test_framework import DashTestFramework
|
|
from test_framework.util import (
|
|
assert_equal,
|
|
assert_greater_than,
|
|
get_bip9_details,
|
|
)
|
|
|
|
class MnehfTest(DashTestFramework):
|
|
def set_test_params(self):
|
|
extra_args = [["-vbparams=testdummy:0:999999999999:12:12:12:5:1", "-persistmempool=0"] for _ in range(4)]
|
|
self.set_dash_test_params(4, 3, fast_dip3_enforcement=True, extra_args=extra_args)
|
|
|
|
def skip_test_if_missing_module(self):
|
|
self.skip_if_no_wallet()
|
|
|
|
def restart_all_nodes(self, params=None):
|
|
for inode in range(self.num_nodes):
|
|
self.log.info(f"Restart node {inode} with {self.extra_args[inode]}")
|
|
if params is not None:
|
|
self.extra_args[inode][0] = f"-vbparams=testdummy:{params[0]}:{params[1]}:12:12:12:5:1"
|
|
self.log.info(f"Actual restart options: {self.extra_args[inode]}")
|
|
|
|
self.restart_node(0)
|
|
for mn in self.mninfo:
|
|
index = mn.node.index
|
|
self.stop_node(index)
|
|
self.start_masternode(mn)
|
|
for i in range(1, self.num_nodes):
|
|
self.connect_nodes(i, 0)
|
|
|
|
def slowly_generate_batch(self, amount):
|
|
self.log.info(f"Slowly generate {amount} blocks")
|
|
while amount > 0:
|
|
self.log.info(f"Generating batch of blocks {amount} left")
|
|
next = min(10, amount)
|
|
amount -= next
|
|
self.bump_mocktime(next)
|
|
self.nodes[1].generate(next)
|
|
self.sync_all()
|
|
|
|
def create_mnehf(self, versionBit, pubkey=None):
|
|
# request ID = sha256("mnhf", versionBit)
|
|
request_id_buf = ser_string(b"mnhf") + struct.pack("<Q", versionBit)
|
|
request_id = hash256(request_id_buf)[::-1].hex()
|
|
|
|
quorumHash = self.mninfo[0].node.quorum("selectquorum", 100, request_id)["quorumHash"]
|
|
mnehf_payload = CMnEhf(
|
|
version = 1,
|
|
versionBit = versionBit,
|
|
quorumHash = int(quorumHash, 16),
|
|
quorumSig = b'\00' * 96)
|
|
|
|
mnehf_tx = CTransaction()
|
|
mnehf_tx.vin = []
|
|
mnehf_tx.vout = []
|
|
mnehf_tx.nVersion = 3
|
|
mnehf_tx.nType = 7 # mnehf signal
|
|
mnehf_tx.vExtraPayload = mnehf_payload.serialize()
|
|
|
|
mnehf_tx.calc_sha256()
|
|
msgHash = format(mnehf_tx.sha256, '064x')
|
|
|
|
self.log.info(f"Signing request_id: {request_id} msgHash: {msgHash} quorum: {quorumHash}")
|
|
recsig = self.get_recovered_sig(request_id, msgHash)
|
|
|
|
mnehf_payload.quorumSig = bytearray.fromhex(recsig["sig"])
|
|
mnehf_tx.vExtraPayload = mnehf_payload.serialize()
|
|
return mnehf_tx
|
|
|
|
|
|
def set_sporks(self):
|
|
spork_enabled = 0
|
|
spork_disabled = 4070908800
|
|
|
|
self.nodes[0].sporkupdate("SPORK_17_QUORUM_DKG_ENABLED", spork_enabled)
|
|
self.nodes[0].sporkupdate("SPORK_19_CHAINLOCKS_ENABLED", spork_disabled)
|
|
self.nodes[0].sporkupdate("SPORK_3_INSTANTSEND_BLOCK_FILTERING", spork_disabled)
|
|
self.nodes[0].sporkupdate("SPORK_2_INSTANTSEND_ENABLED", spork_disabled)
|
|
self.wait_for_sporks_same()
|
|
|
|
def check_fork(self, expected):
|
|
status = get_bip9_details(self.nodes[0], 'testdummy')['status']
|
|
self.log.info(f"height: {self.nodes[0].getblockcount()} status: {status}")
|
|
assert_equal(status, expected)
|
|
|
|
def ensure_tx_is_not_mined(self, tx_id):
|
|
try:
|
|
assert_equal(self.nodes[0].getrawtransaction(tx_id, 1)['height'], -1)
|
|
raise AssertionError("Transaction should not be mined")
|
|
except KeyError:
|
|
# KeyError is expected
|
|
pass
|
|
|
|
def send_tx(self, tx, expected_error = None, reason = None):
|
|
try:
|
|
self.log.info(f"Send tx with expected_error:'{expected_error}'...")
|
|
tx = self.nodes[0].sendrawtransaction(hexstring=tx.serialize().hex(), maxfeerate=0)
|
|
if expected_error is None:
|
|
return tx
|
|
|
|
# failure didn't happen, but expected:
|
|
message = "Transaction should not be accepted"
|
|
if reason is not None:
|
|
message += ": " + reason
|
|
|
|
raise AssertionError(message)
|
|
except JSONRPCException as e:
|
|
self.log.info(f"Send tx triggered an error: {e.error}")
|
|
assert expected_error in e.error['message']
|
|
|
|
|
|
def run_test(self):
|
|
node = self.nodes[0]
|
|
|
|
self.set_sporks()
|
|
self.activate_v19()
|
|
self.log.info(f"After v19 activation should be plenty of blocks: {node.getblockcount()}")
|
|
assert_greater_than(node.getblockcount(), 900)
|
|
assert_equal(get_bip9_details(node, 'testdummy')['status'], 'defined')
|
|
|
|
self.log.info("Mine a quorum...")
|
|
self.mine_quorum()
|
|
assert_equal(get_bip9_details(node, 'testdummy')['status'], 'defined')
|
|
|
|
key = ECKey()
|
|
key.generate()
|
|
pubkey = key.get_pubkey().get_bytes()
|
|
ehf_tx = self.create_mnehf(28, pubkey)
|
|
ehf_unknown_tx = self.create_mnehf(27, pubkey)
|
|
ehf_invalid_tx = self.create_mnehf(9, pubkey) # deployment that is known as non-EHF
|
|
|
|
self.log.info("Checking deserialization of CMnEhf by python's code")
|
|
mnehf_payload = CMnEhf()
|
|
mnehf_payload.deserialize(BytesIO(ehf_tx.vExtraPayload))
|
|
assert_equal(mnehf_payload.version, 1)
|
|
assert_equal(mnehf_payload.versionBit, 28)
|
|
self.log.info("Checking correctness of requestId and quorumHash")
|
|
assert_equal(mnehf_payload.quorumHash, int(self.mninfo[0].node.quorum("selectquorum", 100, 'a0eee872d7d3170dd20d5c5e8380c92b3aa887da5f63d8033289fafa35a90691')["quorumHash"], 16))
|
|
|
|
self.send_tx(ehf_tx, expected_error='mnhf-before-v20')
|
|
|
|
assert_equal(get_bip9_details(node, 'testdummy')['status'], 'defined')
|
|
self.activate_v20()
|
|
assert_equal(get_bip9_details(node, 'testdummy')['status'], 'defined')
|
|
assert_equal(get_bip9_details(node, 'mn_rr')['status'], 'defined')
|
|
|
|
ehf_tx_sent = self.send_tx(ehf_tx)
|
|
self.log.info(f"ehf tx: {ehf_tx_sent}")
|
|
ehf_unknown_tx_sent = self.send_tx(ehf_unknown_tx)
|
|
self.log.info(f"unknown ehf tx: {ehf_unknown_tx_sent}")
|
|
self.send_tx(ehf_invalid_tx, expected_error='bad-mnhf-non-ehf')
|
|
self.sync_all()
|
|
ehf_blockhash = self.nodes[1].generate(1)[0]
|
|
self.sync_blocks()
|
|
self.sync_all()
|
|
|
|
self.log.info(f"Check MnEhfTx {ehf_tx_sent} was mined in {ehf_blockhash}")
|
|
assert ehf_tx_sent in node.getblock(ehf_blockhash)['tx']
|
|
assert ehf_unknown_tx_sent in node.getblock(ehf_blockhash)['tx']
|
|
|
|
self.log.info(f"MnEhf tx: '{ehf_tx}' is sent: {ehf_tx_sent}")
|
|
self.log.info(f"MnEhf 'unknown' tx: '{ehf_unknown_tx}' is sent: {ehf_unknown_tx_sent}")
|
|
self.log.info(f"mempool: {node.getmempoolinfo()}")
|
|
assert_equal(node.getmempoolinfo()['size'], 0)
|
|
|
|
while (node.getblockcount() + 1) % 12 != 0:
|
|
self.check_fork('defined')
|
|
node.generate(1)
|
|
self.sync_all()
|
|
|
|
|
|
self.restart_all_nodes()
|
|
|
|
for _ in range(12):
|
|
self.check_fork('started')
|
|
node.generate(1)
|
|
self.sync_all()
|
|
|
|
|
|
for i in range(12):
|
|
self.check_fork('locked_in')
|
|
node.generate(1)
|
|
self.sync_all()
|
|
if i == 7:
|
|
self.restart_all_nodes()
|
|
|
|
self.check_fork('active')
|
|
|
|
fork_active_blockhash = node.getbestblockhash()
|
|
self.log.info(f"Invalidate block: {ehf_blockhash} with tip {fork_active_blockhash}")
|
|
for inode in self.nodes:
|
|
inode.invalidateblock(ehf_blockhash)
|
|
|
|
self.log.info("Expecting for fork to be defined in next blocks because no MnEHF tx here")
|
|
for _ in range(12):
|
|
self.check_fork('defined')
|
|
node.generate(1)
|
|
self.sync_all()
|
|
|
|
|
|
self.log.info("Re-sending MnEHF for new fork")
|
|
tx_sent_2 = self.send_tx(ehf_tx)
|
|
ehf_blockhash_2 = node.generate(1)[0]
|
|
self.sync_all()
|
|
|
|
self.log.info(f"Check MnEhfTx again {tx_sent_2} was mined in {ehf_blockhash_2}")
|
|
assert tx_sent_2 in node.getblock(ehf_blockhash_2)['tx']
|
|
|
|
self.log.info(f"Generate some more block to jump to `started` status")
|
|
for _ in range(12):
|
|
node.generate(1)
|
|
self.check_fork('started')
|
|
self.restart_all_nodes()
|
|
self.check_fork('started')
|
|
|
|
|
|
self.log.info(f"Re-consider block {ehf_blockhash} to the old MnEHF and forget new fork")
|
|
for inode in self.nodes:
|
|
inode.reconsiderblock(ehf_blockhash)
|
|
assert_equal(node.getbestblockhash(), fork_active_blockhash)
|
|
|
|
self.check_fork('active')
|
|
self.restart_all_nodes()
|
|
self.check_fork('active')
|
|
|
|
self.log.info("Testing duplicate EHF signal with same bit")
|
|
ehf_tx_duplicate = self.send_tx(self.create_mnehf(28, pubkey))
|
|
tip_blockhash = node.generate(1)[0]
|
|
self.sync_blocks()
|
|
block = node.getblock(tip_blockhash)
|
|
assert ehf_tx_duplicate in node.getrawmempool() and ehf_tx_duplicate not in block['tx']
|
|
|
|
self.log.info("Testing EHF signal with same bit but with newer start time")
|
|
self.bump_mocktime(int(60 * 60 * 24 * 14))
|
|
node.generate(1)
|
|
self.sync_blocks()
|
|
self.restart_all_nodes(params=[self.mocktime, self.mocktime + 1000000])
|
|
self.mine_quorum()
|
|
|
|
ehf_tx_new_start = self.create_mnehf(28, pubkey)
|
|
|
|
self.log.info("activate MN_RR also by enabling spork 24")
|
|
assert_equal(get_bip9_details(node, 'mn_rr')['status'], 'defined')
|
|
self.nodes[0].sporkupdate("SPORK_24_TEST_EHF", 0)
|
|
self.wait_for_sporks_same()
|
|
|
|
self.check_fork('defined')
|
|
|
|
self.log.info("Mine one block and ensure EHF tx for the new deployment is mined")
|
|
ehf_tx_sent = self.send_tx(ehf_tx_new_start)
|
|
tip_blockhash = node.generate(1)[0]
|
|
self.sync_all()
|
|
block = node.getblock(tip_blockhash)
|
|
assert ehf_tx_sent in block['tx']
|
|
|
|
self.check_fork('defined')
|
|
self.slowly_generate_batch(12 * 4)
|
|
self.check_fork('active')
|
|
self.log.info(f"bip9: {get_bip9_details(node, 'mn_rr')}")
|
|
assert_equal(get_bip9_details(node, 'mn_rr')['status'], 'active')
|
|
|
|
|
|
if __name__ == '__main__':
|
|
MnehfTest().main()
|