mirror of
https://github.com/dashpay/dash.git
synced 2024-12-27 04:52:59 +01:00
6ad9bdf722
e78aaf41f43d0e2ad78fa6d8dad61032c8ef73d0 [docs] Add release notes for burying bip 9 soft fork deployments (John Newbery) 8319e738f9f118025b332e4fa804d4c31e4113f4 [tests] Add coverage for the content of getblockchaininfo.softforks (James O'Beirne) 0328dcdcfcb56dc8918697716d7686be048ad0b3 [Consensus] Bury segwit deployment (John Newbery) 1c93b9b31c2ab7358f9d55f52dd46340397c906d [Consensus] Bury CSV deployment height (John Newbery) 3862e473f0cb71a762c0306b171b591341d58142 [rpc] Tidy up reporting of buried and ongoing softforks (John Newbery) Pull request description: This hardcodes CSV and segwit activation heights, similar to the BIP 90 buried deployments for BIPs 34, 65 and 66. CSV and segwit have been active for over 18 months. Hardcoding the activation height is a code simplification, makes it easier to understand segwit activation status, and reduces technical debt. This was originally attempted by jl2012 in #11398 and again by me in #12360. ACKs for top commit: ajtowns: ACK e78aaf41f43d0e2ad78fa6d8dad61032c8ef73d0 ; checked diff to previous acked commit, checked tests still work ariard: ACK e78aaf4, check diff, run the tests again and successfully activated csv/segwit heights on mainnet as expected. MarcoFalke: ACK e78aaf41f43d0e2ad78fa6d8dad61032c8ef73d0 (still didn't check if the mainnet block heights are correct, but the code looks good now) Tree-SHA512: 7e951829106e21a81725f7d3e236eddbb59349189740907bb47e33f5dbf95c43753ac1231f47ae7bee85c8c81b2146afcdfdc11deb1503947f23093a9c399912
487 lines
25 KiB
Python
Executable File
487 lines
25 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# Copyright (c) 2015 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
"""Test CSV soft fork activation.
|
|
|
|
This soft fork will activate the following BIPS:
|
|
BIP 68 - nSequence relative lock times
|
|
BIP 112 - CHECKSEQUENCEVERIFY
|
|
BIP 113 - MedianTimePast semantics for nLockTime
|
|
|
|
mine 82 blocks whose coinbases will be used to generate inputs for our tests
|
|
mine 345 blocks and seed block chain with the 82 inputs will use for our tests at height 427
|
|
mine 2 blocks and verify soft fork not yet activated
|
|
mine 1 block and test that soft fork is activated (rules enforced for next block)
|
|
Test BIP 113 is enforced
|
|
Mine 4 blocks so next height is 580 and test BIP 68 is enforced for time and height
|
|
Mine 1 block so next height is 581 and test BIP 68 now passes time but not height
|
|
Mine 1 block so next height is 582 and test BIP 68 now passes time and height
|
|
Test that BIP 112 is enforced
|
|
|
|
Various transactions will be used to test that the BIPs rules are not enforced before the soft fork activates
|
|
And that after the soft fork activates transactions pass and fail as they should according to the rules.
|
|
For each BIP, transactions of versions 1 and 2 will be tested.
|
|
----------------
|
|
BIP 113:
|
|
bip113tx - modify the nLocktime variable
|
|
|
|
BIP 68:
|
|
bip68txs - 16 txs with nSequence relative locktime of 10 with various bits set as per the relative_locktimes below
|
|
|
|
BIP 112:
|
|
bip112txs_vary_nSequence - 16 txs with nSequence relative_locktimes of 10 evaluated against 10 OP_CSV OP_DROP
|
|
bip112txs_vary_nSequence_9 - 16 txs with nSequence relative_locktimes of 9 evaluated against 10 OP_CSV OP_DROP
|
|
bip112txs_vary_OP_CSV - 16 txs with nSequence = 10 evaluated against varying {relative_locktimes of 10} OP_CSV OP_DROP
|
|
bip112txs_vary_OP_CSV_9 - 16 txs with nSequence = 9 evaluated against varying {relative_locktimes of 10} OP_CSV OP_DROP
|
|
bip112tx_special - test negative argument to OP_CSV
|
|
bip112tx_emptystack - test empty stack (= no argument) OP_CSV
|
|
"""
|
|
from decimal import Decimal
|
|
from itertools import product
|
|
from io import BytesIO
|
|
|
|
from test_framework.blocktools import create_coinbase, create_block, create_transaction, TIME_GENESIS_BLOCK
|
|
from test_framework.messages import ToHex, CTransaction
|
|
from test_framework.mininode import P2PDataStore
|
|
from test_framework.script import (
|
|
CScript,
|
|
OP_CHECKSEQUENCEVERIFY,
|
|
OP_DROP,
|
|
)
|
|
from test_framework.test_framework import BitcoinTestFramework
|
|
from test_framework.util import (
|
|
assert_equal,
|
|
hex_str_to_bytes,
|
|
softfork_active,
|
|
)
|
|
|
|
TESTING_TX_COUNT = 83 # Number of testing transactions: 1 BIP113 tx, 16 BIP68 txs, 66 BIP112 txs (see comments above)
|
|
COINBASE_BLOCK_COUNT = TESTING_TX_COUNT # Number of coinbase blocks we need to generate as inputs for our txs
|
|
BASE_RELATIVE_LOCKTIME = 10
|
|
CSV_ACTIVATION_HEIGHT = 432
|
|
SEQ_DISABLE_FLAG = 1 << 31
|
|
SEQ_RANDOM_HIGH_BIT = 1 << 25
|
|
SEQ_TYPE_FLAG = 1 << 22
|
|
SEQ_RANDOM_LOW_BIT = 1 << 18
|
|
|
|
def relative_locktime(sdf, srhb, stf, srlb):
|
|
"""Returns a locktime with certain bits set."""
|
|
|
|
locktime = BASE_RELATIVE_LOCKTIME
|
|
if sdf:
|
|
locktime |= SEQ_DISABLE_FLAG
|
|
if srhb:
|
|
locktime |= SEQ_RANDOM_HIGH_BIT
|
|
if stf:
|
|
locktime |= SEQ_TYPE_FLAG
|
|
if srlb:
|
|
locktime |= SEQ_RANDOM_LOW_BIT
|
|
return locktime
|
|
|
|
def all_rlt_txs(txs):
|
|
return [tx['tx'] for tx in txs]
|
|
|
|
def sign_transaction(node, unsignedtx):
|
|
rawtx = ToHex(unsignedtx)
|
|
signresult = node.signrawtransactionwithwallet(rawtx)
|
|
tx = CTransaction()
|
|
f = BytesIO(hex_str_to_bytes(signresult['hex']))
|
|
tx.deserialize(f)
|
|
return tx
|
|
|
|
def create_bip112special(node, input, txversion, address):
|
|
tx = create_transaction(node, input, address, amount=Decimal("499.98"))
|
|
tx.nVersion = txversion
|
|
signtx = sign_transaction(node, tx)
|
|
signtx.vin[0].scriptSig = CScript([-1, OP_CHECKSEQUENCEVERIFY, OP_DROP] + list(CScript(signtx.vin[0].scriptSig)))
|
|
return signtx
|
|
|
|
def create_bip112emptystack(node, input, txversion, address):
|
|
tx = create_transaction(node, input, address, amount=Decimal("499.98"))
|
|
tx.nVersion = txversion
|
|
signtx = sign_transaction(node, tx)
|
|
signtx.vin[0].scriptSig = CScript([OP_CHECKSEQUENCEVERIFY] + list(CScript(signtx.vin[0].scriptSig)))
|
|
return signtx
|
|
|
|
def send_generic_input_tx(node, coinbases, address):
|
|
return node.sendrawtransaction(ToHex(sign_transaction(node, create_transaction(node, node.getblock(coinbases.pop())['tx'][0], address, amount=Decimal("499.99")))))
|
|
|
|
def create_bip68txs(node, bip68inputs, txversion, address, locktime_delta=0):
|
|
"""Returns a list of bip68 transactions with different bits set."""
|
|
txs = []
|
|
assert len(bip68inputs) >= 16
|
|
for i, (sdf, srhb, stf, srlb) in enumerate(product(*[[True, False]] * 4)):
|
|
locktime = relative_locktime(sdf, srhb, stf, srlb)
|
|
tx = create_transaction(node, bip68inputs[i], address, amount=Decimal("499.98"))
|
|
tx.nVersion = txversion
|
|
tx.vin[0].nSequence = locktime + locktime_delta
|
|
tx = sign_transaction(node, tx)
|
|
tx.rehash()
|
|
txs.append({'tx': tx, 'sdf': sdf, 'stf': stf})
|
|
|
|
return txs
|
|
|
|
def create_bip112txs(node, bip112inputs, varyOP_CSV, txversion, address, locktime_delta=0):
|
|
"""Returns a list of bip68 transactions with different bits set."""
|
|
txs = []
|
|
assert len(bip112inputs) >= 16
|
|
for i, (sdf, srhb, stf, srlb) in enumerate(product(*[[True, False]] * 4)):
|
|
locktime = relative_locktime(sdf, srhb, stf, srlb)
|
|
tx = create_transaction(node, bip112inputs[i], address, amount=Decimal("499.98"))
|
|
if (varyOP_CSV): # if varying OP_CSV, nSequence is fixed
|
|
tx.vin[0].nSequence = BASE_RELATIVE_LOCKTIME + locktime_delta
|
|
else: # vary nSequence instead, OP_CSV is fixed
|
|
tx.vin[0].nSequence = locktime + locktime_delta
|
|
tx.nVersion = txversion
|
|
signtx = sign_transaction(node, tx)
|
|
if (varyOP_CSV):
|
|
signtx.vin[0].scriptSig = CScript([locktime, OP_CHECKSEQUENCEVERIFY, OP_DROP] + list(CScript(signtx.vin[0].scriptSig)))
|
|
else:
|
|
signtx.vin[0].scriptSig = CScript([BASE_RELATIVE_LOCKTIME, OP_CHECKSEQUENCEVERIFY, OP_DROP] + list(CScript(signtx.vin[0].scriptSig)))
|
|
tx.rehash()
|
|
txs.append({'tx': signtx, 'sdf': sdf, 'stf': stf})
|
|
return txs
|
|
|
|
class BIP68_112_113Test(BitcoinTestFramework):
|
|
def set_test_params(self):
|
|
self.num_nodes = 1
|
|
self.setup_clean_chain = True
|
|
# Must also set '-maxtipage=600100' to allow syncing from very old blocks
|
|
# and '-dip3params=2000:2000' to create pre-dip3 blocks only
|
|
self.extra_args = [[
|
|
'-whitelist=noban@127.0.0.1',
|
|
'-blockversion=4',
|
|
'-maxtipage=600100', '-dip3params=2000:2000',
|
|
'-par=1', # Use only one script thread to get the exact reject reason for testing
|
|
]]
|
|
self.supports_cli = False
|
|
|
|
def setup_network(self):
|
|
self.setup_nodes()
|
|
|
|
def skip_test_if_missing_module(self):
|
|
self.skip_if_no_wallet()
|
|
|
|
def generate_blocks(self, number):
|
|
test_blocks = []
|
|
for i in range(number):
|
|
block = self.create_test_block([])
|
|
test_blocks.append(block)
|
|
self.last_block_time += 600
|
|
self.tip = block.sha256
|
|
self.tipheight += 1
|
|
return test_blocks
|
|
|
|
def create_test_block(self, txs):
|
|
block = create_block(self.tip, create_coinbase(self.tipheight + 1), self.last_block_time + 600)
|
|
block.nVersion = 4
|
|
block.vtx.extend(txs)
|
|
block.hashMerkleRoot = block.calc_merkle_root()
|
|
block.rehash()
|
|
block.solve()
|
|
return block
|
|
|
|
def send_blocks(self, blocks, success=True, reject_reason=None):
|
|
"""Sends blocks to test node. Syncs and verifies that tip has advanced to most recent block.
|
|
|
|
Call with success = False if the tip shouldn't advance to the most recent block."""
|
|
self.nodes[0].p2p.send_blocks_and_test(blocks, self.nodes[0], success=success, reject_reason=reject_reason)
|
|
|
|
def run_test(self):
|
|
self.nodes[0].add_p2p_connection(P2PDataStore())
|
|
|
|
self.log.info("Generate blocks in the past for coinbase outputs.")
|
|
self.coinbase_blocks = self.nodes[0].generate(COINBASE_BLOCK_COUNT) # blocks generated for inputs
|
|
# set time so that there was enough time to build up to 1000 blocks 10 minutes apart on top of the last one
|
|
# without worrying about getting into the future
|
|
self.nodes[0].setmocktime(TIME_GENESIS_BLOCK + 600 * 1000 + 100)
|
|
self.tipheight = COINBASE_BLOCK_COUNT # height of the next block to build
|
|
self.last_block_time = TIME_GENESIS_BLOCK
|
|
self.tip = int(self.nodes[0].getbestblockhash(), 16)
|
|
self.nodeaddress = self.nodes[0].getnewaddress()
|
|
|
|
# Activation height is hardcoded
|
|
test_blocks = self.generate_blocks(CSV_ACTIVATION_HEIGHT-5 - COINBASE_BLOCK_COUNT)
|
|
#test_blocks = self.generate_blocks(345)
|
|
self.send_blocks(test_blocks)
|
|
assert not softfork_active(self.nodes[0], 'csv')
|
|
|
|
# Inputs at height = 431
|
|
#
|
|
# Put inputs for all tests in the chain at height 431 (tip now = 430) (time increases by 600s per block)
|
|
# Note we reuse inputs for v1 and v2 txs so must test these separately
|
|
# 16 normal inputs
|
|
bip68inputs = []
|
|
for i in range(16):
|
|
bip68inputs.append(send_generic_input_tx(self.nodes[0], self.coinbase_blocks, self.nodeaddress))
|
|
|
|
# 2 sets of 16 inputs with 10 OP_CSV OP_DROP (actually will be prepended to spending scriptSig)
|
|
bip112basicinputs = []
|
|
for j in range(2):
|
|
inputs = []
|
|
for i in range(16):
|
|
inputs.append(send_generic_input_tx(self.nodes[0], self.coinbase_blocks, self.nodeaddress))
|
|
bip112basicinputs.append(inputs)
|
|
|
|
# 2 sets of 16 varied inputs with (relative_lock_time) OP_CSV OP_DROP (actually will be prepended to spending scriptSig)
|
|
bip112diverseinputs = []
|
|
for j in range(2):
|
|
inputs = []
|
|
for i in range(16):
|
|
inputs.append(send_generic_input_tx(self.nodes[0], self.coinbase_blocks, self.nodeaddress))
|
|
bip112diverseinputs.append(inputs)
|
|
|
|
# 1 special input with -1 OP_CSV OP_DROP (actually will be prepended to spending scriptSig)
|
|
bip112specialinput = send_generic_input_tx(self.nodes[0], self.coinbase_blocks, self.nodeaddress)
|
|
# 1 special input with (empty stack) OP_CSV (actually will be prepended to spending scriptSig)
|
|
bip112emptystackinput = send_generic_input_tx(self.nodes[0],self.coinbase_blocks, self.nodeaddress)
|
|
|
|
# 1 normal input
|
|
bip113input = send_generic_input_tx(self.nodes[0], self.coinbase_blocks, self.nodeaddress)
|
|
|
|
self.nodes[0].setmocktime(self.last_block_time + 600)
|
|
inputblockhash = self.nodes[0].generate(1)[0] # 1 block generated for inputs to be in chain at height 431
|
|
self.nodes[0].setmocktime(TIME_GENESIS_BLOCK + 600 * 1000 + 100)
|
|
self.tip = int(inputblockhash, 16)
|
|
self.tipheight += 1
|
|
self.last_block_time += 600
|
|
assert_equal(len(self.nodes[0].getblock(inputblockhash, True)["tx"]), TESTING_TX_COUNT + 1)
|
|
|
|
# 2 more version 4 blocks
|
|
test_blocks = self.generate_blocks(2)
|
|
self.send_blocks(test_blocks)
|
|
|
|
assert_equal(self.tipheight, CSV_ACTIVATION_HEIGHT - 2)
|
|
self.log.info("Height = {}, CSV not yet active (will activate for block {}, not {})".format(self.tipheight, CSV_ACTIVATION_HEIGHT, CSV_ACTIVATION_HEIGHT - 1))
|
|
assert not softfork_active(self.nodes[0], 'csv')
|
|
|
|
# Test both version 1 and version 2 transactions for all tests
|
|
# BIP113 test transaction will be modified before each use to put in appropriate block time
|
|
bip113tx_v1 = create_transaction(self.nodes[0], bip113input, self.nodeaddress, amount=Decimal("499.98"))
|
|
bip113tx_v1.vin[0].nSequence = 0xFFFFFFFE
|
|
bip113tx_v1.nVersion = 1
|
|
bip113tx_v2 = create_transaction(self.nodes[0], bip113input, self.nodeaddress, amount=Decimal("499.98"))
|
|
bip113tx_v2.vin[0].nSequence = 0xFFFFFFFE
|
|
bip113tx_v2.nVersion = 2
|
|
|
|
# For BIP68 test all 16 relative sequence locktimes
|
|
bip68txs_v1 = create_bip68txs(self.nodes[0], bip68inputs, 1, self.nodeaddress)
|
|
bip68txs_v2 = create_bip68txs(self.nodes[0], bip68inputs, 2, self.nodeaddress)
|
|
|
|
# For BIP112 test:
|
|
# 16 relative sequence locktimes of 10 against 10 OP_CSV OP_DROP inputs
|
|
bip112txs_vary_nSequence_v1 = create_bip112txs(self.nodes[0], bip112basicinputs[0], False, 1, self.nodeaddress)
|
|
bip112txs_vary_nSequence_v2 = create_bip112txs(self.nodes[0], bip112basicinputs[0], False, 2, self.nodeaddress)
|
|
# 16 relative sequence locktimes of 9 against 10 OP_CSV OP_DROP inputs
|
|
bip112txs_vary_nSequence_9_v1 = create_bip112txs(self.nodes[0], bip112basicinputs[1], False, 1, self.nodeaddress, -1)
|
|
bip112txs_vary_nSequence_9_v2 = create_bip112txs(self.nodes[0], bip112basicinputs[1], False, 2, self.nodeaddress, -1)
|
|
# sequence lock time of 10 against 16 (relative_lock_time) OP_CSV OP_DROP inputs
|
|
bip112txs_vary_OP_CSV_v1 = create_bip112txs(self.nodes[0], bip112diverseinputs[0], True, 1, self.nodeaddress)
|
|
bip112txs_vary_OP_CSV_v2 = create_bip112txs(self.nodes[0], bip112diverseinputs[0], True, 2, self.nodeaddress)
|
|
# sequence lock time of 9 against 16 (relative_lock_time) OP_CSV OP_DROP inputs
|
|
bip112txs_vary_OP_CSV_9_v1 = create_bip112txs(self.nodes[0], bip112diverseinputs[1], True, 1, self.nodeaddress, -1)
|
|
bip112txs_vary_OP_CSV_9_v2 = create_bip112txs(self.nodes[0], bip112diverseinputs[1], True, 2, self.nodeaddress, -1)
|
|
# -1 OP_CSV OP_DROP input
|
|
bip112tx_special_v1 = create_bip112special(self.nodes[0], bip112specialinput, 1, self.nodeaddress)
|
|
bip112tx_special_v2 = create_bip112special(self.nodes[0], bip112specialinput, 2, self.nodeaddress)
|
|
# (empty stack) OP_CSV input
|
|
bip112tx_emptystack_v1 = create_bip112emptystack(self.nodes[0], bip112emptystackinput, 1, self.nodeaddress)
|
|
bip112tx_emptystack_v2 = create_bip112emptystack(self.nodes[0], bip112emptystackinput, 2, self.nodeaddress)
|
|
|
|
self.log.info("TESTING")
|
|
|
|
self.log.info("Pre-Soft Fork Tests. All txs should pass.")
|
|
self.log.info("Test version 1 txs")
|
|
|
|
success_txs = []
|
|
# BIP113 tx, -1 CSV tx and empty stack CSV tx should succeed
|
|
bip113tx_v1.nLockTime = self.last_block_time - 600 * 5 # = MTP of prior block (not <) but < time put on current block
|
|
bip113signed1 = sign_transaction(self.nodes[0], bip113tx_v1)
|
|
success_txs.append(bip113signed1)
|
|
success_txs.append(bip112tx_special_v1)
|
|
success_txs.append(bip112tx_emptystack_v1)
|
|
# add BIP 68 txs
|
|
success_txs.extend(all_rlt_txs(bip68txs_v1))
|
|
# add BIP 112 with seq=10 txs
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_nSequence_v1))
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_OP_CSV_v1))
|
|
# try BIP 112 with seq=9 txs
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_nSequence_9_v1))
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_OP_CSV_9_v1))
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
self.log.info("Test version 2 txs")
|
|
|
|
success_txs = []
|
|
# BIP113 tx, -1 CSV tx and empty stack CSV tx should succeed
|
|
bip113tx_v2.nLockTime = self.last_block_time - 600 * 5 # = MTP of prior block (not <) but < time put on current block
|
|
bip113signed2 = sign_transaction(self.nodes[0], bip113tx_v2)
|
|
success_txs.append(bip113signed2)
|
|
success_txs.append(bip112tx_special_v2)
|
|
success_txs.append(bip112tx_emptystack_v2)
|
|
# add BIP 68 txs
|
|
success_txs.extend(all_rlt_txs(bip68txs_v2))
|
|
# add BIP 112 with seq=10 txs
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_nSequence_v2))
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_OP_CSV_v2))
|
|
# try BIP 112 with seq=9 txs
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_nSequence_9_v2))
|
|
success_txs.extend(all_rlt_txs(bip112txs_vary_OP_CSV_9_v2))
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# 1 more version 4 block to get us to height 432 so the fork should now be active for the next block
|
|
assert not softfork_active(self.nodes[0], 'csv')
|
|
test_blocks = self.generate_blocks(1)
|
|
self.send_blocks(test_blocks)
|
|
assert softfork_active(self.nodes[0], 'csv')
|
|
|
|
self.log.info("Post-Soft Fork Tests.")
|
|
|
|
self.log.info("BIP 113 tests")
|
|
# BIP 113 tests should now fail regardless of version number if nLockTime isn't satisfied by new rules
|
|
bip113tx_v1.nLockTime = self.last_block_time - 600 * 5 # = MTP of prior block (not <) but < time put on current block
|
|
bip113signed1 = sign_transaction(self.nodes[0], bip113tx_v1)
|
|
bip113tx_v2.nLockTime = self.last_block_time - 600 * 5 # = MTP of prior block (not <) but < time put on current block
|
|
bip113signed2 = sign_transaction(self.nodes[0], bip113tx_v2)
|
|
for bip113tx in [bip113signed1, bip113signed2]:
|
|
self.send_blocks([self.create_test_block([bip113tx])], success=False, reject_reason='bad-txns-nonfinal')
|
|
# BIP 113 tests should now pass if the locktime is < MTP
|
|
bip113tx_v1.nLockTime = self.last_block_time - 600 * 5 - 1 # < MTP of prior block
|
|
bip113signed1 = sign_transaction(self.nodes[0], bip113tx_v1)
|
|
bip113tx_v2.nLockTime = self.last_block_time - 600 * 5 - 1 # < MTP of prior block
|
|
bip113signed2 = sign_transaction(self.nodes[0], bip113tx_v2)
|
|
for bip113tx in [bip113signed1, bip113signed2]:
|
|
self.send_blocks([self.create_test_block([bip113tx])])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# Next block height = 437 after 4 blocks of random version
|
|
test_blocks = self.generate_blocks(4)
|
|
self.send_blocks(test_blocks)
|
|
|
|
self.log.info("BIP 68 tests")
|
|
self.log.info("Test version 1 txs - all should still pass")
|
|
|
|
success_txs = []
|
|
success_txs.extend(all_rlt_txs(bip68txs_v1))
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
self.log.info("Test version 2 txs")
|
|
|
|
# All txs with SEQUENCE_LOCKTIME_DISABLE_FLAG set pass
|
|
bip68success_txs = [tx['tx'] for tx in bip68txs_v2 if tx['sdf']]
|
|
self.send_blocks([self.create_test_block(bip68success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# All txs without flag fail as we are at delta height = 8 < 10 and delta time = 8 * 600 < 10 * 512
|
|
bip68timetxs = [tx['tx'] for tx in bip68txs_v2 if not tx['sdf'] and tx['stf']]
|
|
for tx in bip68timetxs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False, reject_reason='bad-txns-nonfinal')
|
|
|
|
bip68heighttxs = [tx['tx'] for tx in bip68txs_v2 if not tx['sdf'] and not tx['stf']]
|
|
for tx in bip68heighttxs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False, reject_reason='bad-txns-nonfinal')
|
|
|
|
# Advance one block to 438
|
|
test_blocks = self.generate_blocks(1)
|
|
self.send_blocks(test_blocks)
|
|
|
|
# Height txs should fail and time txs should now pass 9 * 600 > 10 * 512
|
|
bip68success_txs.extend(bip68timetxs)
|
|
self.send_blocks([self.create_test_block(bip68success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
for tx in bip68heighttxs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False, reject_reason='bad-txns-nonfinal')
|
|
|
|
# Advance one block to 439
|
|
test_blocks = self.generate_blocks(1)
|
|
self.send_blocks(test_blocks)
|
|
|
|
# All BIP 68 txs should pass
|
|
bip68success_txs.extend(bip68heighttxs)
|
|
self.send_blocks([self.create_test_block(bip68success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
self.log.info("BIP 112 tests")
|
|
self.log.info("Test version 1 txs")
|
|
|
|
# -1 OP_CSV tx and (empty stack) OP_CSV tx should fail
|
|
self.send_blocks([self.create_test_block([bip112tx_special_v1])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Negative locktime)')
|
|
self.send_blocks([self.create_test_block([bip112tx_emptystack_v1])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Operation not valid with the current stack size)')
|
|
# If SEQUENCE_LOCKTIME_DISABLE_FLAG is set in argument to OP_CSV, version 1 txs should still pass
|
|
|
|
success_txs = [tx['tx'] for tx in bip112txs_vary_OP_CSV_v1 if tx['sdf']]
|
|
success_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_9_v1 if tx['sdf']]
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# If SEQUENCE_LOCKTIME_DISABLE_FLAG is unset in argument to OP_CSV, version 1 txs should now fail
|
|
fail_txs = all_rlt_txs(bip112txs_vary_nSequence_v1)
|
|
fail_txs += all_rlt_txs(bip112txs_vary_nSequence_9_v1)
|
|
fail_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_v1 if not tx['sdf']]
|
|
fail_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_9_v1 if not tx['sdf']]
|
|
for tx in fail_txs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Locktime requirement not satisfied)')
|
|
|
|
self.log.info("Test version 2 txs")
|
|
|
|
# -1 OP_CSV tx and (empty stack) OP_CSV tx should fail
|
|
self.send_blocks([self.create_test_block([bip112tx_special_v2])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Negative locktime)')
|
|
self.send_blocks([self.create_test_block([bip112tx_emptystack_v2])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Operation not valid with the current stack size)')
|
|
|
|
# If SEQUENCE_LOCKTIME_DISABLE_FLAG is set in argument to OP_CSV, version 2 txs should pass (all sequence locks are met)
|
|
success_txs = [tx['tx'] for tx in bip112txs_vary_OP_CSV_v2 if tx['sdf']]
|
|
success_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_9_v2 if tx['sdf']]
|
|
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# SEQUENCE_LOCKTIME_DISABLE_FLAG is unset in argument to OP_CSV for all remaining txs ##
|
|
|
|
# All txs with nSequence 9 should fail either due to earlier mismatch or failing the CSV check
|
|
fail_txs = all_rlt_txs(bip112txs_vary_nSequence_9_v2)
|
|
fail_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_9_v2 if not tx['sdf']]
|
|
for tx in fail_txs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Locktime requirement not satisfied)')
|
|
|
|
# If SEQUENCE_LOCKTIME_DISABLE_FLAG is set in nSequence, tx should fail
|
|
fail_txs = [tx['tx'] for tx in bip112txs_vary_nSequence_v2 if tx['sdf']]
|
|
for tx in fail_txs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Locktime requirement not satisfied)')
|
|
|
|
# If sequencelock types mismatch, tx should fail
|
|
fail_txs = [tx['tx'] for tx in bip112txs_vary_nSequence_v2 if not tx['sdf'] and tx['stf']]
|
|
fail_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_v2 if not tx['sdf'] and tx['stf']]
|
|
for tx in fail_txs:
|
|
self.send_blocks([self.create_test_block([tx])], success=False,
|
|
reject_reason='non-mandatory-script-verify-flag (Locktime requirement not satisfied)')
|
|
|
|
# Remaining txs should pass, just test masking works properly
|
|
success_txs = [tx['tx'] for tx in bip112txs_vary_nSequence_v2 if not tx['sdf'] and not tx['stf']]
|
|
success_txs += [tx['tx'] for tx in bip112txs_vary_OP_CSV_v2 if not tx['sdf'] and not tx['stf']]
|
|
self.send_blocks([self.create_test_block(success_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
# Additional test, of checking that comparison of two time types works properly
|
|
time_txs = []
|
|
for tx in [tx['tx'] for tx in bip112txs_vary_OP_CSV_v2 if not tx['sdf'] and tx['stf']]:
|
|
tx.vin[0].nSequence = BASE_RELATIVE_LOCKTIME | SEQ_TYPE_FLAG
|
|
signtx = sign_transaction(self.nodes[0], tx)
|
|
time_txs.append(signtx)
|
|
|
|
self.send_blocks([self.create_test_block(time_txs)])
|
|
self.nodes[0].invalidateblock(self.nodes[0].getbestblockhash())
|
|
|
|
if __name__ == '__main__':
|
|
BIP68_112_113Test().main()
|