2016-05-06 11:23:48 +02:00
|
|
|
#!/usr/bin/env python3
|
|
|
|
# Copyright (c) 2015-2016 The Bitcoin Core developers
|
|
|
|
# Distributed under the MIT software license, see the accompanying
|
2015-09-18 21:59:55 +02:00
|
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
2019-01-07 10:55:35 +01:00
|
|
|
"""Test behavior of -maxuploadtarget.
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
* Verify that getdata requests for old blocks (>1week) are dropped
|
|
|
|
if uploadtarget has been reached.
|
|
|
|
* Verify that getdata requests for recent blocks are respecteved even
|
|
|
|
if uploadtarget has been reached.
|
|
|
|
* Verify that the upload counters are reset after 24 hours.
|
2019-01-07 10:55:35 +01:00
|
|
|
"""
|
2017-05-02 19:10:23 +02:00
|
|
|
from collections import defaultdict
|
|
|
|
import time
|
2019-01-07 10:55:35 +01:00
|
|
|
|
|
|
|
from test_framework.mininode import *
|
|
|
|
from test_framework.test_framework import BitcoinTestFramework
|
|
|
|
from test_framework.util import *
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
class TestNode(NodeConnCB):
|
|
|
|
def __init__(self):
|
2017-03-30 09:22:59 +02:00
|
|
|
super().__init__()
|
2017-05-02 19:10:23 +02:00
|
|
|
self.block_receive_map = defaultdict(int)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
def on_inv(self, conn, message):
|
|
|
|
pass
|
|
|
|
|
|
|
|
def on_block(self, conn, message):
|
|
|
|
message.block.calc_sha256()
|
2017-05-02 19:10:23 +02:00
|
|
|
self.block_receive_map[message.block.sha256] += 1
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
class MaxUploadTest(BitcoinTestFramework):
|
|
|
|
|
2017-09-01 18:47:13 +02:00
|
|
|
def set_test_params(self):
|
2016-05-20 15:16:51 +02:00
|
|
|
self.setup_clean_chain = True
|
|
|
|
self.num_nodes = 1
|
2017-05-02 20:02:55 +02:00
|
|
|
self.extra_args = [["-maxuploadtarget=200", "-blockmaxsize=999000", "-maxtipage="+str(2*60*60*24*7)]]
|
2016-05-20 15:16:51 +02:00
|
|
|
|
2016-12-06 12:05:31 +01:00
|
|
|
# Cache for utxos, as the listunspent may take a long time later in the test
|
|
|
|
self.utxo_cache = []
|
|
|
|
|
2015-09-18 21:59:55 +02:00
|
|
|
def run_test(self):
|
2018-04-18 13:48:59 +02:00
|
|
|
# Advance all nodes 2 weeks in the future
|
2019-08-09 01:14:11 +02:00
|
|
|
old_mocktime = self.mocktime
|
2018-04-18 13:48:59 +02:00
|
|
|
current_mocktime = old_mocktime + 2*60*60*24*7
|
2019-08-12 19:10:56 +02:00
|
|
|
self.mocktime = current_mocktime
|
2018-04-18 13:48:59 +02:00
|
|
|
set_node_times(self.nodes, current_mocktime)
|
|
|
|
|
2015-09-18 21:59:55 +02:00
|
|
|
# Before we connect anything, we first set the time on the node
|
|
|
|
# to be in the past, otherwise things break because the CNode
|
|
|
|
# time counters can't be reset backward after initialization
|
2018-04-18 13:48:59 +02:00
|
|
|
self.nodes[0].setmocktime(old_mocktime)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# Generate some old blocks
|
|
|
|
self.nodes[0].generate(130)
|
|
|
|
|
2017-11-08 19:10:43 +01:00
|
|
|
# p2p_conns[0] will only request old blocks
|
|
|
|
# p2p_conns[1] will only request new blocks
|
|
|
|
# p2p_conns[2] will test resetting the counters
|
|
|
|
p2p_conns = []
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-11-14 08:56:04 +01:00
|
|
|
for _ in range(3):
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns.append(self.nodes[0].add_p2p_connection(TestNode()))
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-12-12 12:52:33 +01:00
|
|
|
network_thread_start()
|
2017-11-08 19:10:43 +01:00
|
|
|
for p2pc in p2p_conns:
|
|
|
|
p2pc.wait_for_verack()
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# Test logic begins here
|
|
|
|
|
|
|
|
# Now mine a big block
|
2016-12-06 12:05:31 +01:00
|
|
|
mine_large_block(self.nodes[0], self.utxo_cache)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# Store the hash; we'll request this later
|
|
|
|
big_old_block = self.nodes[0].getbestblockhash()
|
|
|
|
old_block_size = self.nodes[0].getblock(big_old_block, True)['size']
|
|
|
|
big_old_block = int(big_old_block, 16)
|
|
|
|
|
|
|
|
# Advance to two days ago
|
2018-04-18 13:48:59 +02:00
|
|
|
self.nodes[0].setmocktime(current_mocktime - 2*60*60*24)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# Mine one more block, so that the prior block looks old
|
2016-12-06 12:05:31 +01:00
|
|
|
mine_large_block(self.nodes[0], self.utxo_cache)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# We'll be requesting this new block too
|
|
|
|
big_new_block = self.nodes[0].getbestblockhash()
|
|
|
|
big_new_block = int(big_new_block, 16)
|
|
|
|
|
2017-11-08 19:10:43 +01:00
|
|
|
# p2p_conns[0] will test what happens if we just keep requesting the
|
2015-09-18 21:59:55 +02:00
|
|
|
# the same big old block too many times (expect: disconnect)
|
|
|
|
|
|
|
|
getdata_request = msg_getdata()
|
|
|
|
getdata_request.inv.append(CInv(2, big_old_block))
|
|
|
|
|
|
|
|
max_bytes_per_day = 200*1024*1024
|
2016-01-19 17:47:55 +01:00
|
|
|
daily_buffer = 144 * MAX_BLOCK_SIZE
|
2015-11-06 00:05:06 +01:00
|
|
|
max_bytes_available = max_bytes_per_day - daily_buffer
|
2016-03-19 21:36:32 +01:00
|
|
|
success_count = max_bytes_available // old_block_size
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2015-09-18 21:59:55 +02:00
|
|
|
# 144MB will be reserved for relaying new blocks, so expect this to
|
|
|
|
# succeed for ~70 tries.
|
2016-05-06 11:23:48 +02:00
|
|
|
for i in range(success_count):
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns[0].send_message(getdata_request)
|
|
|
|
p2p_conns[0].sync_with_ping()
|
|
|
|
assert_equal(p2p_conns[0].block_receive_map[big_old_block], i+1)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 3)
|
|
|
|
# At most a couple more tries should succeed (depending on how long
|
|
|
|
# the test has been running so far).
|
2016-05-06 11:23:48 +02:00
|
|
|
for i in range(3):
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns[0].send_message(getdata_request)
|
|
|
|
p2p_conns[0].wait_for_disconnect()
|
2015-09-18 21:59:55 +02:00
|
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 2)
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Peer 0 disconnected after downloading old block too many times")
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-11-08 19:10:43 +01:00
|
|
|
# Requesting the current block on p2p_conns[1] should succeed indefinitely,
|
2015-09-18 21:59:55 +02:00
|
|
|
# even when over the max upload target.
|
|
|
|
# We'll try 200 times
|
|
|
|
getdata_request.inv = [CInv(2, big_new_block)]
|
2016-05-06 11:23:48 +02:00
|
|
|
for i in range(200):
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns[1].send_message(getdata_request)
|
|
|
|
p2p_conns[1].sync_with_ping()
|
|
|
|
assert_equal(p2p_conns[1].block_receive_map[big_new_block], i+1)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Peer 1 able to repeatedly download new block")
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-11-08 19:10:43 +01:00
|
|
|
# But if p2p_conns[1] tries for an old block, it gets disconnected too.
|
2015-09-18 21:59:55 +02:00
|
|
|
getdata_request.inv = [CInv(2, big_old_block)]
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns[1].send_message(getdata_request)
|
|
|
|
p2p_conns[1].wait_for_disconnect()
|
2015-09-18 21:59:55 +02:00
|
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 1)
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Peer 1 disconnected after trying to download old block")
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Advancing system time on node to clear counters...")
|
2015-09-18 21:59:55 +02:00
|
|
|
|
|
|
|
# If we advance the time by 24 hours, then the counters should reset,
|
2017-11-08 19:10:43 +01:00
|
|
|
# and p2p_conns[2] should be able to retrieve the old block.
|
2018-04-18 13:48:59 +02:00
|
|
|
self.nodes[0].setmocktime(current_mocktime)
|
2017-11-08 19:10:43 +01:00
|
|
|
p2p_conns[2].sync_with_ping()
|
|
|
|
p2p_conns[2].send_message(getdata_request)
|
|
|
|
p2p_conns[2].sync_with_ping()
|
|
|
|
assert_equal(p2p_conns[2].block_receive_map[big_old_block], 1)
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Peer 2 able to download old block")
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2017-11-14 08:56:04 +01:00
|
|
|
self.nodes[0].disconnect_p2ps()
|
2015-09-18 21:59:55 +02:00
|
|
|
|
2015-11-11 10:10:48 +01:00
|
|
|
#stop and start node 0 with 1MB maxuploadtarget, whitelist 127.0.0.1
|
2017-03-09 21:16:20 +01:00
|
|
|
self.log.info("Restarting nodes with -whitelist=127.0.0.1")
|
2017-06-02 12:08:48 +02:00
|
|
|
self.stop_node(0)
|
2019-09-24 16:23:31 +02:00
|
|
|
self.start_node(0, ["-whitelist=127.0.0.1", "-maxuploadtarget=1", "-blockmaxsize=999000", "-maxtipage="+str(2*60*60*24*7), "-mocktime="+str(current_mocktime)])
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2017-11-08 19:10:43 +01:00
|
|
|
# Reconnect to self.nodes[0]
|
|
|
|
self.nodes[0].add_p2p_connection(TestNode())
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2017-12-12 12:52:33 +01:00
|
|
|
network_thread_start()
|
2017-11-08 19:10:43 +01:00
|
|
|
self.nodes[0].p2p.wait_for_verack()
|
2015-11-11 10:10:48 +01:00
|
|
|
|
|
|
|
#retrieve 20 blocks which should be enough to break the 1MB limit
|
|
|
|
getdata_request.inv = [CInv(2, big_new_block)]
|
2016-05-06 11:23:48 +02:00
|
|
|
for i in range(20):
|
2017-11-08 19:10:43 +01:00
|
|
|
self.nodes[0].p2p.send_message(getdata_request)
|
|
|
|
self.nodes[0].p2p.sync_with_ping()
|
|
|
|
assert_equal(self.nodes[0].p2p.block_receive_map[big_new_block], i+1)
|
2015-11-11 10:10:48 +01:00
|
|
|
|
|
|
|
getdata_request.inv = [CInv(2, big_old_block)]
|
2017-11-08 19:10:43 +01:00
|
|
|
self.nodes[0].p2p.send_and_ping(getdata_request)
|
2017-05-02 19:10:23 +02:00
|
|
|
assert_equal(len(self.nodes[0].getpeerinfo()), 1) #node is still connected because of the whitelist
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2017-05-02 19:10:23 +02:00
|
|
|
self.log.info("Peer still connected after trying to download old block (whitelisted)")
|
2015-11-11 10:10:48 +01:00
|
|
|
|
2015-09-18 21:59:55 +02:00
|
|
|
if __name__ == '__main__':
|
|
|
|
MaxUploadTest().main()
|