mirror of
https://github.com/dashpay/dash.git
synced 2024-12-27 21:12:48 +01:00
d6c6174958
* llmq: Implement automated DKG recovery threads * llmq: Implement quorum verification vector sync * init: Validiate quorum data recovery related command line parameter * test: Add quorum_data_request_timeout_seconds in DashTestFramework * test: Test quorum data recovery in feature_llmq_data_recovery.py * test: Add feature_llmq_data_recovery.py to BASE_SCRIPTS * test: Fix quorum_data_request_expiration_timeout in wait_for_quorum_data * test: Always test the existence of secretKeyShare in test_mn_quorum_data With this change it also validates that "secretKeyShare" is not in `quorum_info` if its not expected to be in there. Before this was basically just not tested. * llmq|test: Use bool as argument type for -llmq-data-recovery * llmq: Always set nTimeLastSuccess to 0 * test: Set -llmq-data-recovery=0 in p2p_quorum_data.py * test: Simplify test_mns Co-authored-by: UdjinM6 <UdjinM6@users.noreply.github.com> * refactor: pass CQuorumCPtr to StartQuorumDataRecoveryThread * test: Fix thread name in comment Co-authored-by: UdjinM6 <UdjinM6@users.noreply.github.com>
462 lines
21 KiB
Python
Executable File
462 lines
21 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# Copyright (c) 2021 The Dash Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
|
|
import time
|
|
|
|
from test_framework.messages import msg_qgetdata, msg_qwatch
|
|
from test_framework.mininode import (
|
|
mininode_lock,
|
|
network_thread_start,
|
|
network_thread_join,
|
|
P2PInterface,
|
|
)
|
|
from test_framework.test_framework import DashTestFramework
|
|
from test_framework.util import (
|
|
assert_equal,
|
|
assert_raises_rpc_error,
|
|
connect_nodes,
|
|
force_finish_mnsync,
|
|
wait_until,
|
|
)
|
|
|
|
'''
|
|
p2p_quorum_data.py
|
|
|
|
Tests QGETDATA/QDATA functionality
|
|
'''
|
|
|
|
# Possible error values of QDATA
|
|
QUORUM_TYPE_INVALID = 1
|
|
QUORUM_BLOCK_NOT_FOUND = 2
|
|
QUORUM_NOT_FOUND = 3
|
|
MASTERNODE_IS_NO_MEMBER = 4
|
|
QUORUM_VERIFICATION_VECTOR_MISSING = 5
|
|
ENCRYPTED_CONTRIBUTIONS_MISSING = 6
|
|
|
|
# Used to overwrite MNAUTH for mininode connections
|
|
fake_mnauth_1 = ["cecf37bf0ec05d2d22cb8227f88074bb882b94cd2081ba318a5a444b1b15b9fd",
|
|
"087ba00bf61135f3860c4944a0debabe186ef82628fbe4ceaed1ad51d672c58dde14ea4b321efe0b89257a40322bc972"]
|
|
fake_mnauth_2 = ["6ad7ed7a2d6c2c1db30fc364114602b36b2730a9aa96d8f11f1871a9cee37378",
|
|
"122463411a86362966a5161805f24cf6a0eef08a586b8e00c4f0ad0b084c5bb3f5c9a60ee5ffc78db2313897e3ab2223"]
|
|
|
|
# Used to distinguish mininode connections
|
|
uacomment_m3_1 = "MN3_1"
|
|
uacomment_m3_2 = "MN3_2"
|
|
|
|
|
|
def assert_qdata(qdata, qgetdata, error, len_vvec=0, len_contributions=0):
|
|
assert qdata is not None and qgetdata is not None
|
|
assert_equal(qdata.quorum_type, qgetdata.quorum_type)
|
|
assert_equal(qdata.quorum_hash, qgetdata.quorum_hash)
|
|
assert_equal(qdata.data_mask, qgetdata.data_mask)
|
|
assert_equal(qdata.protx_hash, qgetdata.protx_hash)
|
|
assert_equal(qdata.error, error)
|
|
assert_equal(len(qdata.quorum_vvec), len_vvec)
|
|
assert_equal(len(qdata.enc_contributions), len_contributions)
|
|
|
|
|
|
def wait_for_banscore(node, peer_id, expected_score):
|
|
def get_score():
|
|
for peer in node.getpeerinfo():
|
|
if peer["id"] == peer_id:
|
|
return peer["banscore"]
|
|
return None
|
|
wait_until(lambda: get_score() == expected_score, timeout=6)
|
|
|
|
|
|
def p2p_connection(node, uacomment=None):
|
|
return node.add_p2p_connection(QuorumDataInterface(), uacomment=uacomment)
|
|
|
|
|
|
def get_mininode_id(node, uacomment=None):
|
|
def get_id():
|
|
for p in node.getpeerinfo():
|
|
for p2p in node.p2ps:
|
|
if uacomment is not None and p2p.uacomment != uacomment:
|
|
continue
|
|
if p["subver"] == p2p.strSubVer.decode():
|
|
return p["id"]
|
|
return None
|
|
wait_until(lambda: get_id() is not None, timeout=10)
|
|
return get_id()
|
|
|
|
|
|
def mnauth(node, node_id, protx_hash, operator_pubkey):
|
|
assert node.mnauth(node_id, protx_hash, operator_pubkey)
|
|
mnauth_peer_id = None
|
|
for peer in node.getpeerinfo():
|
|
if "verified_proregtx_hash" in peer and peer["verified_proregtx_hash"] == protx_hash:
|
|
assert_equal(mnauth_peer_id, None)
|
|
mnauth_peer_id = peer["id"]
|
|
assert_equal(mnauth_peer_id, node_id)
|
|
|
|
|
|
class QuorumDataInterface(P2PInterface):
|
|
def __init__(self):
|
|
super().__init__()
|
|
|
|
def test_qgetdata(self, qgetdata, expected_error=0, len_vvec=0, len_contributions=0, response_expected=True):
|
|
self.send_message(qgetdata)
|
|
self.wait_for_qdata(message_expected=response_expected)
|
|
if response_expected:
|
|
assert_qdata(self.get_qdata(), qgetdata, expected_error, len_vvec, len_contributions)
|
|
|
|
def wait_for_qgetdata(self, timeout=3, message_expected=True):
|
|
def test_function():
|
|
return self.message_count["qgetdata"]
|
|
wait_until(test_function, timeout=timeout, lock=mininode_lock, do_assert=message_expected)
|
|
self.message_count["qgetdata"] = 0
|
|
if not message_expected:
|
|
assert not self.message_count["qgetdata"]
|
|
|
|
def get_qdata(self):
|
|
return self.last_message["qdata"]
|
|
|
|
def wait_for_qdata(self, timeout=10, message_expected=True):
|
|
def test_function():
|
|
return self.message_count["qdata"]
|
|
wait_until(test_function, timeout=timeout, lock=mininode_lock, do_assert=message_expected)
|
|
self.message_count["qdata"] = 0
|
|
if not message_expected:
|
|
assert not self.message_count["qdata"]
|
|
|
|
|
|
class QuorumDataMessagesTest(DashTestFramework):
|
|
def set_test_params(self):
|
|
extra_args = [["-llmq-data-recovery=0"]] * 4
|
|
self.set_dash_test_params(4, 3, fast_dip3_enforcement=True, extra_args=extra_args)
|
|
|
|
def restart_mn(self, mn, reindex=False):
|
|
args = self.extra_args[mn.nodeIdx] + ['-masternodeblsprivkey=%s' % mn.keyOperator]
|
|
if reindex:
|
|
args.append('-reindex')
|
|
self.restart_node(mn.nodeIdx, args)
|
|
force_finish_mnsync(mn.node)
|
|
connect_nodes(mn.node, 0)
|
|
self.sync_blocks()
|
|
|
|
def run_test(self):
|
|
|
|
def force_request_expire(bump_seconds=self.quorum_data_request_expiration_timeout + 1):
|
|
self.bump_mocktime(bump_seconds)
|
|
# Test with/without expired request cleanup
|
|
if node0.getblockcount() % 2:
|
|
node0.generate(1)
|
|
self.sync_blocks()
|
|
|
|
def test_basics():
|
|
self.log.info("Testing basics of QGETDATA/QDATA")
|
|
p2p_node0 = p2p_connection(node0)
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
network_thread_start()
|
|
p2p_node0.wait_for_verack()
|
|
p2p_mn1.wait_for_verack()
|
|
id_p2p_node0 = get_mininode_id(node0)
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
|
|
# Ensure that both nodes start with zero ban score
|
|
wait_for_banscore(node0, id_p2p_node0, 0)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 0)
|
|
|
|
self.log.info("Check that normal node doesn't respond to qgetdata "
|
|
"and does bump our score")
|
|
p2p_node0.test_qgetdata(qgetdata_all, response_expected=False)
|
|
wait_for_banscore(node0, id_p2p_node0, 10)
|
|
# The masternode should not respond to qgetdata for non-masternode connections
|
|
self.log.info("Check that masternode doesn't respond to "
|
|
"non-masternode connection. Doesn't bump score.")
|
|
p2p_mn1.test_qgetdata(qgetdata_all, response_expected=False)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 10)
|
|
# Open a fake MNAUTH authenticated P2P connection to the masternode to allow qgetdata
|
|
node0.disconnect_p2ps()
|
|
mn1.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
network_thread_start()
|
|
p2p_mn1.wait_for_verack()
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
mnauth(mn1.node, id_p2p_mn1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
# The masternode should now respond to qgetdata requests
|
|
self.log.info("Request verification vector")
|
|
p2p_mn1.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 0)
|
|
# Note: our banscore is bumped as we are requesting too rapidly,
|
|
# however the node still returns the data
|
|
self.log.info("Request encrypted contributions")
|
|
p2p_mn1.test_qgetdata(qgetdata_contributions, 0, 0, self.llmq_size)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 25)
|
|
# Request both
|
|
# Note: our banscore is bumped as we are requesting too rapidly,
|
|
# however the node still returns the data
|
|
self.log.info("Request both")
|
|
p2p_mn1.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 50)
|
|
mn1.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
self.log.info("Test ban score increase for invalid / unexpected QDATA")
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
p2p_mn2 = p2p_connection(mn2.node)
|
|
network_thread_start()
|
|
p2p_mn1.wait_for_verack()
|
|
p2p_mn2.wait_for_verack()
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
id_p2p_mn2 = get_mininode_id(mn2.node)
|
|
mnauth(mn1.node, id_p2p_mn1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
mnauth(mn2.node, id_p2p_mn2, fake_mnauth_2[0], fake_mnauth_2[1])
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 0)
|
|
p2p_mn2.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
qdata_valid = p2p_mn2.get_qdata()
|
|
# - Not requested
|
|
p2p_mn1.send_message(qdata_valid)
|
|
time.sleep(1)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 10)
|
|
# - Already received
|
|
force_request_expire()
|
|
assert mn1.node.quorum("getdata", id_p2p_mn1, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
p2p_mn1.wait_for_qgetdata()
|
|
p2p_mn1.send_message(qdata_valid)
|
|
time.sleep(1)
|
|
p2p_mn1.send_message(qdata_valid)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 20)
|
|
# - Not like requested
|
|
force_request_expire()
|
|
assert mn1.node.quorum("getdata", id_p2p_mn1, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
p2p_mn1.wait_for_qgetdata()
|
|
qdata_invalid_request = qdata_valid
|
|
qdata_invalid_request.data_mask = 2
|
|
p2p_mn1.send_message(qdata_invalid_request)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 30)
|
|
# - Invalid verification vector
|
|
force_request_expire()
|
|
assert mn1.node.quorum("getdata", id_p2p_mn1, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
p2p_mn1.wait_for_qgetdata()
|
|
qdata_invalid_vvec = qdata_valid
|
|
qdata_invalid_vvec.quorum_vvec.pop()
|
|
p2p_mn1.send_message(qdata_invalid_vvec)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 40)
|
|
# - Invalid contributions
|
|
force_request_expire()
|
|
assert mn1.node.quorum("getdata", id_p2p_mn1, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
p2p_mn1.wait_for_qgetdata()
|
|
qdata_invalid_contribution = qdata_valid
|
|
qdata_invalid_contribution.enc_contributions.pop()
|
|
p2p_mn1.send_message(qdata_invalid_contribution)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 50)
|
|
mn1.node.disconnect_p2ps()
|
|
mn2.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
self.log.info("Test all available error codes")
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
network_thread_start()
|
|
p2p_mn1.wait_for_verack()
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
mnauth(mn1.node, id_p2p_mn1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
qgetdata_invalid_type = msg_qgetdata(quorum_hash_int, 103, 0x01, protx_hash_int)
|
|
qgetdata_invalid_block = msg_qgetdata(protx_hash_int, 100, 0x01, protx_hash_int)
|
|
qgetdata_invalid_quorum = msg_qgetdata(int(mn1.node.getblockhash(0), 16), 100, 0x01, protx_hash_int)
|
|
qgetdata_invalid_no_member = msg_qgetdata(quorum_hash_int, 100, 0x02, quorum_hash_int)
|
|
p2p_mn1.test_qgetdata(qgetdata_invalid_type, QUORUM_TYPE_INVALID)
|
|
p2p_mn1.test_qgetdata(qgetdata_invalid_block, QUORUM_BLOCK_NOT_FOUND)
|
|
p2p_mn1.test_qgetdata(qgetdata_invalid_quorum, QUORUM_NOT_FOUND)
|
|
p2p_mn1.test_qgetdata(qgetdata_invalid_no_member, MASTERNODE_IS_NO_MEMBER)
|
|
# The last two error case require the node to miss its DKG data so we just reindex the node.
|
|
mn1.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
self.restart_mn(mn1, reindex=True)
|
|
# Re-connect to the masternode
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
p2p_mn2 = p2p_connection(mn2.node)
|
|
network_thread_start()
|
|
p2p_mn1.wait_for_verack()
|
|
p2p_mn2.wait_for_verack()
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
id_p2p_mn2 = get_mininode_id(mn2.node)
|
|
assert id_p2p_mn1 is not None
|
|
assert id_p2p_mn2 is not None
|
|
mnauth(mn1.node, id_p2p_mn1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
mnauth(mn2.node, id_p2p_mn2, fake_mnauth_2[0], fake_mnauth_2[1])
|
|
# Validate the DKG data is missing
|
|
p2p_mn1.test_qgetdata(qgetdata_vvec, QUORUM_VERIFICATION_VECTOR_MISSING)
|
|
p2p_mn1.test_qgetdata(qgetdata_contributions, ENCRYPTED_CONTRIBUTIONS_MISSING)
|
|
self.log.info("Test DKG data recovery with QDATA")
|
|
# Now that mn1 is missing its DKG data try to recover it by querying the data from mn2 and then sending it
|
|
# to mn1 with a direct QDATA message.
|
|
#
|
|
# mininode - QGETDATA -> mn2 - QDATA -> mininode - QDATA -> mn1
|
|
#
|
|
# However, mn1 only accepts self requested QDATA messages, that's why we trigger mn1 - QGETDATA -> mininode
|
|
# via the RPC command "quorum getdata".
|
|
#
|
|
# Get the required DKG data for mn1
|
|
p2p_mn2.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
# Trigger mn1 - QGETDATA -> p2p_mn1
|
|
assert mn1.node.quorum("getdata", id_p2p_mn1, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
# Wait until mn1 sent the QGETDATA to p2p_mn1
|
|
p2p_mn1.wait_for_qgetdata()
|
|
# Send the QDATA received from mn2 to mn1
|
|
p2p_mn1.send_message(p2p_mn2.get_qdata())
|
|
# Now mn1 should have its data back!
|
|
self.wait_for_quorum_data([mn1], 100, quorum_hash, recover=False)
|
|
# Restart one more time and make sure data gets saved to db
|
|
mn1.node.disconnect_p2ps()
|
|
mn2.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
self.restart_mn(mn1)
|
|
self.wait_for_quorum_data([mn1], 100, quorum_hash, recover=False)
|
|
|
|
# Test request limiting / banscore increase
|
|
def test_request_limit():
|
|
|
|
def test_send_from_two_to_one(send_1, expected_score_1, send_2, expected_score_2, clear_requests=False):
|
|
if clear_requests:
|
|
force_request_expire()
|
|
if send_1:
|
|
p2p_mn3_1.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
if send_2:
|
|
p2p_mn3_2.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_1, expected_score_1)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_2, expected_score_2)
|
|
|
|
self.log.info("Test request limiting / banscore increases")
|
|
|
|
p2p_mn1 = p2p_connection(mn1.node)
|
|
network_thread_start()
|
|
p2p_mn1.wait_for_verack()
|
|
id_p2p_mn1 = get_mininode_id(mn1.node)
|
|
mnauth(mn1.node, id_p2p_mn1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
p2p_mn1.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 0)
|
|
force_request_expire(299) # This shouldn't clear requests, next request should bump score
|
|
p2p_mn1.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 25)
|
|
force_request_expire(1) # This should clear the requests now, next request should not bump score
|
|
p2p_mn1.test_qgetdata(qgetdata_vvec, 0, self.llmq_threshold, 0)
|
|
wait_for_banscore(mn1.node, id_p2p_mn1, 25)
|
|
mn1.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
# Requesting one QDATA with mn1 and mn2 from mn3 should not result
|
|
# in banscore increase for either of both.
|
|
p2p_mn3_1 = p2p_connection(mn3.node, uacomment_m3_1)
|
|
p2p_mn3_2 = p2p_connection(mn3.node, uacomment_m3_2)
|
|
network_thread_start()
|
|
p2p_mn3_1.wait_for_verack()
|
|
p2p_mn3_2.wait_for_verack()
|
|
id_p2p_mn3_1 = get_mininode_id(mn3.node, uacomment_m3_1)
|
|
id_p2p_mn3_2 = get_mininode_id(mn3.node, uacomment_m3_2)
|
|
assert id_p2p_mn3_1 != id_p2p_mn3_2
|
|
mnauth(mn3.node, id_p2p_mn3_1, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
mnauth(mn3.node, id_p2p_mn3_2, fake_mnauth_2[0], fake_mnauth_2[1])
|
|
# Now try some {mn1, mn2} - QGETDATA -> mn3 combinations to make
|
|
# sure request limit works connection based
|
|
test_send_from_two_to_one(False, 0, True, 0, True)
|
|
test_send_from_two_to_one(True, 0, True, 25)
|
|
test_send_from_two_to_one(True, 25, False, 25)
|
|
test_send_from_two_to_one(False, 25, True, 25, True)
|
|
test_send_from_two_to_one(True, 25, True, 50)
|
|
test_send_from_two_to_one(True, 50, True, 75)
|
|
test_send_from_two_to_one(True, 50, True, 75, True)
|
|
test_send_from_two_to_one(True, 75, False, 75)
|
|
test_send_from_two_to_one(False, 75, True, None)
|
|
# mn1 should still have a score of 75
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_1, 75)
|
|
# mn2 should be "banned" now
|
|
wait_until(lambda: not p2p_mn3_2.is_connected, timeout=10)
|
|
mn3.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
|
|
# Test that QWATCH connections are also allowed to query data but all
|
|
# QWATCH connections share one request limit slot
|
|
def test_qwatch_connections():
|
|
self.log.info("Test QWATCH connections")
|
|
force_request_expire()
|
|
p2p_mn3_1 = p2p_connection(mn3.node, uacomment_m3_1)
|
|
p2p_mn3_2 = p2p_connection(mn3.node, uacomment_m3_2)
|
|
network_thread_start()
|
|
p2p_mn3_1.wait_for_verack()
|
|
p2p_mn3_2.wait_for_verack()
|
|
id_p2p_mn3_1 = get_mininode_id(mn3.node, uacomment_m3_1)
|
|
id_p2p_mn3_2 = get_mininode_id(mn3.node, uacomment_m3_2)
|
|
assert id_p2p_mn3_1 != id_p2p_mn3_2
|
|
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_1, 0)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_2, 0)
|
|
|
|
# Send QWATCH for both connections
|
|
p2p_mn3_1.send_message(msg_qwatch())
|
|
p2p_mn3_2.send_message(msg_qwatch())
|
|
|
|
# Now send alternating and make sure they share the same request limit
|
|
p2p_mn3_1.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_1, 0)
|
|
p2p_mn3_2.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_2, 25)
|
|
p2p_mn3_1.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
wait_for_banscore(mn3.node, id_p2p_mn3_1, 25)
|
|
mn3.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
|
|
def test_watchquorums():
|
|
self.log.info("Test -watchquorums support")
|
|
for extra_args in [[], ["-watchquorums"]]:
|
|
self.restart_node(0, self.extra_args[0] + extra_args)
|
|
for i in range(self.num_nodes - 1):
|
|
connect_nodes(node0, i + 1)
|
|
p2p_node0 = p2p_connection(node0)
|
|
p2p_mn2 = p2p_connection(mn2.node)
|
|
network_thread_start()
|
|
p2p_node0.wait_for_verack()
|
|
p2p_mn2.wait_for_verack()
|
|
id_p2p_node0 = get_mininode_id(node0)
|
|
id_p2p_mn2 = get_mininode_id(mn2.node)
|
|
mnauth(node0, id_p2p_node0, fake_mnauth_1[0], fake_mnauth_1[1])
|
|
mnauth(mn2.node, id_p2p_mn2, fake_mnauth_2[0], fake_mnauth_2[1])
|
|
p2p_mn2.test_qgetdata(qgetdata_all, 0, self.llmq_threshold, self.llmq_size)
|
|
assert node0.quorum("getdata", id_p2p_node0, 100, quorum_hash, 0x03, mn1.proTxHash)
|
|
p2p_node0.wait_for_qgetdata()
|
|
p2p_node0.send_message(p2p_mn2.get_qdata())
|
|
wait_for_banscore(node0, id_p2p_node0, (1 - len(extra_args)) * 10)
|
|
node0.disconnect_p2ps()
|
|
mn2.node.disconnect_p2ps()
|
|
network_thread_join()
|
|
|
|
def test_rpc_quorum_getdata_protx_hash():
|
|
self.log.info("Test optional proTxHash of `quorum getdata`")
|
|
assert_raises_rpc_error(-8, "proTxHash missing",
|
|
mn1.node.quorum, "getdata", 0, 100, quorum_hash, 0x02)
|
|
assert_raises_rpc_error(-8, "proTxHash invalid",
|
|
mn1.node.quorum, "getdata", 0, 100, quorum_hash, 0x03,
|
|
"0000000000000000000000000000000000000000000000000000000000000000")
|
|
|
|
# Enable DKG and disable ChainLocks
|
|
self.nodes[0].spork("SPORK_17_QUORUM_DKG_ENABLED", 0)
|
|
self.nodes[0].spork("SPORK_19_CHAINLOCKS_ENABLED", 4070908800)
|
|
|
|
self.wait_for_sporks_same()
|
|
quorum_hash = self.mine_quorum()
|
|
|
|
node0 = self.nodes[0]
|
|
mn1 = self.mninfo[0]
|
|
mn2 = self.mninfo[1]
|
|
mn3 = self.mninfo[2]
|
|
|
|
# Convert the hex values into integer values
|
|
quorum_hash_int = int(quorum_hash, 16)
|
|
protx_hash_int = int(mn1.proTxHash, 16)
|
|
|
|
# Valid requests
|
|
qgetdata_vvec = msg_qgetdata(quorum_hash_int, 100, 0x01, protx_hash_int)
|
|
qgetdata_contributions = msg_qgetdata(quorum_hash_int, 100, 0x02, protx_hash_int)
|
|
qgetdata_all = msg_qgetdata(quorum_hash_int, 100, 0x03, protx_hash_int)
|
|
|
|
test_basics()
|
|
test_request_limit()
|
|
test_qwatch_connections()
|
|
test_watchquorums()
|
|
test_rpc_quorum_getdata_protx_hash()
|
|
|
|
|
|
if __name__ == '__main__':
|
|
QuorumDataMessagesTest().main()
|