mirror of
https://github.com/dashpay/dash.git
synced 2024-12-25 20:12:57 +01:00
4aa197dbdb
fa4632c41714dfaa699bacc6a947d72668a4deef test: Move boost/stdlib includes last (MarcoFalke) fa488f131fd4f5bab0d01376c5a5013306f1abcd scripted-diff: Bump copyright headers (MarcoFalke) fac5c373006a9e4bcbb56843bb85f1aca4d87599 scripted-diff: Sort test includes (MarcoFalke) Pull request description: When writing tests, often includes need to be added or removed. Currently the list of includes is not sorted, so developers that write tests and have `clang-format` installed will either have an unrelated change (sorting) included in their commit or they will have to manually undo the sort. This pull preempts both issues by just sorting all includes in one commit. Please be aware that this is **NOT** a change to policy to enforce clang-format or any other developer guideline or process. Developers are free to use whatever tool they want, see also #18651. Edit: Also includes a commit to bump the copyright headers, so that the touched files don't need to be touched again for that. ACKs for top commit: practicalswift: ACK fa4632c41714dfaa699bacc6a947d72668a4deef jonatack: ACK fa4632c41714dfaa, light review and sanity checks with gcc build and clang fuzz build Tree-SHA512: 130a8d073a379ba556b1e64104d37c46b671425c0aef0ed725fd60156a95e8dc83fb6f0b5330b2f8152cf5daaf3983b4aca5e75812598f2626c39fd12b88b180
176 lines
5.8 KiB
Python
Executable File
176 lines
5.8 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
# Copyright (c) 2013-2020 The Bitcoin Core developers
|
|
# Distributed under the MIT software license, see the accompanying
|
|
# file COPYING or http://www.opensource.org/licenses/mit-license.php.
|
|
#
|
|
# Generate seeds.txt from "protx list valid 1"
|
|
#
|
|
|
|
import re
|
|
import sys
|
|
import dns.resolver
|
|
import collections
|
|
import json
|
|
import multiprocessing
|
|
|
|
NSEEDS=512
|
|
|
|
MAX_SEEDS_PER_ASN=4
|
|
|
|
# These are hosts that have been observed to be behaving strangely (e.g.
|
|
# aggressively connecting to every node).
|
|
with open("suspicious_hosts.txt", mode="r", encoding="utf-8") as f:
|
|
SUSPICIOUS_HOSTS = {s.strip() for s in f if s.strip()}
|
|
|
|
|
|
PATTERN_IPV4 = re.compile(r"^((\d{1,3})\.(\d{1,3})\.(\d{1,3})\.(\d{1,3})):(\d+)$")
|
|
PATTERN_IPV6 = re.compile(r"^\[([0-9a-z:]+)\]:(\d+)$")
|
|
PATTERN_ONION = re.compile(r"^([abcdefghijklmnopqrstuvwxyz234567]{16}\.onion):(\d+)$")
|
|
|
|
def parseip(ip):
|
|
m = PATTERN_IPV4.match(ip)
|
|
sortkey = None
|
|
ip = None
|
|
if m is None:
|
|
m = PATTERN_IPV6.match(ip)
|
|
if m is None:
|
|
m = PATTERN_ONION.match(ip)
|
|
if m is None:
|
|
return None
|
|
else:
|
|
net = 'onion'
|
|
ipstr = sortkey = m.group(1)
|
|
port = int(m.group(2))
|
|
else:
|
|
net = 'ipv6'
|
|
if m.group(1) in ['::']: # Not interested in localhost
|
|
return None
|
|
ipstr = m.group(1)
|
|
sortkey = ipstr # XXX parse IPv6 into number, could use name_to_ipv6 from generate-seeds
|
|
port = int(m.group(2))
|
|
else:
|
|
# Do IPv4 sanity check
|
|
ip = 0
|
|
for i in range(0,4):
|
|
if int(m.group(i+2)) < 0 or int(m.group(i+2)) > 255:
|
|
return None
|
|
ip = ip + (int(m.group(i+2)) << (8*(3-i)))
|
|
if ip == 0:
|
|
return None
|
|
net = 'ipv4'
|
|
sortkey = ip
|
|
ipstr = m.group(1)
|
|
port = int(m.group(6))
|
|
|
|
return {
|
|
"net": net,
|
|
"ip": ipstr,
|
|
"port": port,
|
|
"ipnum": ip,
|
|
"sortkey": sortkey
|
|
}
|
|
|
|
def filtermulticollateralhash(mns):
|
|
'''Filter out MNs sharing the same collateral hash'''
|
|
hist = collections.defaultdict(list)
|
|
for mn in mns:
|
|
hist[mn['collateralHash']].append(mn)
|
|
return [mn for mn in mns if len(hist[mn['collateralHash']]) == 1]
|
|
|
|
def filtermulticollateraladdress(mns):
|
|
'''Filter out MNs sharing the same collateral address'''
|
|
hist = collections.defaultdict(list)
|
|
for mn in mns:
|
|
hist[mn['collateralAddress']].append(mn)
|
|
return [mn for mn in mns if len(hist[mn['collateralAddress']]) == 1]
|
|
|
|
def filtermultipayoutaddress(mns):
|
|
'''Filter out MNs sharing the same payout address'''
|
|
hist = collections.defaultdict(list)
|
|
for mn in mns:
|
|
hist[mn['state']['payoutAddress']].append(mn)
|
|
return [mn for mn in mns if len(hist[mn['state']['payoutAddress']]) == 1]
|
|
|
|
def resolveasn(resolver, ip):
|
|
if ip['net'] == 'ipv4':
|
|
ipaddr = ip['ip']
|
|
prefix = '.origin'
|
|
else: # http://www.team-cymru.com/IP-ASN-mapping.html
|
|
res = str() # 2001:4860:b002:23::68
|
|
for nb in ip['ip'].split(':')[:4]: # pick the first 4 nibbles
|
|
for c in nb.zfill(4): # right padded with '0'
|
|
res += c + '.' # 2001 4860 b002 0023
|
|
ipaddr = res.rstrip('.') # 2.0.0.1.4.8.6.0.b.0.0.2.0.0.2.3
|
|
prefix = '.origin6'
|
|
asn = int([x.to_text() for x in resolver.resolve('.'.join(reversed(ipaddr.split('.'))) + prefix + '.asn.cymru.com', 'TXT').response.answer][0].split('\"')[1].split(' ')[0])
|
|
return asn
|
|
|
|
# Based on Greg Maxwell's seed_filter.py
|
|
def filterbyasn(ips, max_per_asn, max_total):
|
|
# Sift out ips by type
|
|
ips_ipv46 = [ip for ip in ips if ip['net'] in ['ipv4', 'ipv6']]
|
|
ips_onion = [ip for ip in ips if ip['net'] == 'onion']
|
|
|
|
my_resolver = dns.resolver.Resolver()
|
|
|
|
pool = multiprocessing.Pool(processes=16)
|
|
|
|
# OpenDNS servers
|
|
my_resolver.nameservers = ['208.67.222.222', '208.67.220.220']
|
|
|
|
# Resolve ASNs in parallel
|
|
asns = [pool.apply_async(resolveasn, args=(my_resolver, ip)) for ip in ips_ipv46]
|
|
|
|
# Filter IPv46 by ASN
|
|
result = []
|
|
asn_count = {}
|
|
for i, ip in enumerate(ips_ipv46):
|
|
if len(result) == max_total:
|
|
break
|
|
try:
|
|
asn = asns[i].get()
|
|
if asn not in asn_count:
|
|
asn_count[asn] = 0
|
|
if asn_count[asn] == max_per_asn:
|
|
continue
|
|
asn_count[asn] += 1
|
|
result.append(ip)
|
|
except:
|
|
sys.stderr.write('ERR: Could not resolve ASN for "' + ip['ip'] + '"\n')
|
|
|
|
# Add back Onions
|
|
result.extend(ips_onion)
|
|
return result
|
|
|
|
def main():
|
|
# This expects a json as outputted by "protx list valid 1"
|
|
if len(sys.argv) > 1:
|
|
with open(sys.argv[1], 'r', encoding="utf8") as f:
|
|
mns = json.load(f)
|
|
else:
|
|
mns = json.load(sys.stdin)
|
|
|
|
# Skip PoSe banned MNs
|
|
mns = [mn for mn in mns if mn['state']['PoSeBanHeight'] == -1]
|
|
# Skip MNs with < 10000 confirmations
|
|
mns = [mn for mn in mns if mn['confirmations'] >= 10000]
|
|
# Filter out MNs which are definitely from the same person/operator
|
|
mns = filtermulticollateralhash(mns)
|
|
mns = filtermulticollateraladdress(mns)
|
|
mns = filtermultipayoutaddress(mns)
|
|
# Extract IPs
|
|
ips = [parseip(mn['state']['service']) for mn in mns]
|
|
# Look up ASNs and limit results, both per ASN and globally.
|
|
ips = filterbyasn(ips, MAX_SEEDS_PER_ASN, NSEEDS)
|
|
# Sort the results by IP address (for deterministic output).
|
|
ips.sort(key=lambda x: (x['net'], x['sortkey']))
|
|
|
|
for ip in ips:
|
|
if ip['net'] == 'ipv6':
|
|
print('[%s]:%i' % (ip['ip'], ip['port']))
|
|
else:
|
|
print('%s:%i' % (ip['ip'], ip['port']))
|
|
|
|
if __name__ == '__main__':
|
|
main()
|