mirror of
https://github.com/dashpay/dash.git
synced 2024-12-25 12:02:48 +01:00
Merge fba3a04ac2
into ad7a373529
This commit is contained in:
commit
c4abdb73fc
@ -787,7 +787,7 @@ private:
|
|||||||
|
|
||||||
/** Protects m_peer_map. This mutex must not be locked while holding a lock
|
/** Protects m_peer_map. This mutex must not be locked while holding a lock
|
||||||
* on any of the mutexes inside a Peer object. */
|
* on any of the mutexes inside a Peer object. */
|
||||||
mutable Mutex m_peer_mutex;
|
mutable SharedMutex m_peer_mutex;
|
||||||
/**
|
/**
|
||||||
* Map of all Peer objects, keyed by peer id. This map is protected
|
* Map of all Peer objects, keyed by peer id. This map is protected
|
||||||
* by the m_peer_mutex. Once a shared pointer reference is
|
* by the m_peer_mutex. Once a shared pointer reference is
|
||||||
@ -1695,7 +1695,7 @@ void PeerManagerImpl::FinalizeNode(const CNode& node) {
|
|||||||
|
|
||||||
PeerRef PeerManagerImpl::GetPeerRef(NodeId id) const
|
PeerRef PeerManagerImpl::GetPeerRef(NodeId id) const
|
||||||
{
|
{
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
auto it = m_peer_map.find(id);
|
auto it = m_peer_map.find(id);
|
||||||
return it != m_peer_map.end() ? it->second : nullptr;
|
return it != m_peer_map.end() ? it->second : nullptr;
|
||||||
}
|
}
|
||||||
@ -2265,7 +2265,7 @@ bool PeerManagerImpl::AlreadyHaveBlock(const uint256& block_hash)
|
|||||||
|
|
||||||
void PeerManagerImpl::SendPings()
|
void PeerManagerImpl::SendPings()
|
||||||
{
|
{
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
for(auto& it : m_peer_map) it.second->m_ping_queued = true;
|
for(auto& it : m_peer_map) it.second->m_ping_queued = true;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -2275,7 +2275,7 @@ void PeerManagerImpl::AskPeersForTransaction(const uint256& txid, bool is_master
|
|||||||
peersToAsk.reserve(4);
|
peersToAsk.reserve(4);
|
||||||
|
|
||||||
{
|
{
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
// TODO consider prioritizing MNs again, once that flag is moved into Peer
|
// TODO consider prioritizing MNs again, once that flag is moved into Peer
|
||||||
for (const auto& [_, peer] : m_peer_map) {
|
for (const auto& [_, peer] : m_peer_map) {
|
||||||
if (peersToAsk.size() >= 4) {
|
if (peersToAsk.size() >= 4) {
|
||||||
@ -2340,7 +2340,7 @@ void PeerManagerImpl::RelayInv(CInv &inv, const int minProtoVersion)
|
|||||||
|
|
||||||
void PeerManagerImpl::RelayInv(CInv &inv)
|
void PeerManagerImpl::RelayInv(CInv &inv)
|
||||||
{
|
{
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
for (const auto& [_, peer] : m_peer_map) {
|
for (const auto& [_, peer] : m_peer_map) {
|
||||||
if (!peer->GetInvRelay()) continue;
|
if (!peer->GetInvRelay()) continue;
|
||||||
PushInv(*peer, inv);
|
PushInv(*peer, inv);
|
||||||
@ -2352,7 +2352,7 @@ void PeerManagerImpl::RelayDSQ(const CCoinJoinQueue& queue)
|
|||||||
CInv inv{MSG_DSQ, queue.GetHash()};
|
CInv inv{MSG_DSQ, queue.GetHash()};
|
||||||
std::vector<NodeId> nodes_send_all;
|
std::vector<NodeId> nodes_send_all;
|
||||||
{
|
{
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
for (const auto& [nodeid, peer] : m_peer_map) {
|
for (const auto& [nodeid, peer] : m_peer_map) {
|
||||||
switch (peer->m_wants_dsq) {
|
switch (peer->m_wants_dsq) {
|
||||||
case Peer::WantsDSQ::NONE:
|
case Peer::WantsDSQ::NONE:
|
||||||
@ -2428,7 +2428,7 @@ void PeerManagerImpl::RelayInvFiltered(CInv &inv, const uint256& relatedTxHash,
|
|||||||
void PeerManagerImpl::RelayTransaction(const uint256& txid)
|
void PeerManagerImpl::RelayTransaction(const uint256& txid)
|
||||||
{
|
{
|
||||||
const CInv inv{m_cj_ctx->dstxman->GetDSTX(txid) ? MSG_DSTX : MSG_TX, txid};
|
const CInv inv{m_cj_ctx->dstxman->GetDSTX(txid) ? MSG_DSTX : MSG_TX, txid};
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
for(auto& it : m_peer_map) {
|
for(auto& it : m_peer_map) {
|
||||||
Peer& peer = *it.second;
|
Peer& peer = *it.second;
|
||||||
auto tx_relay = peer.GetTxRelay();
|
auto tx_relay = peer.GetTxRelay();
|
||||||
@ -2441,7 +2441,7 @@ void PeerManagerImpl::RelayTransaction(const uint256& txid)
|
|||||||
void PeerManagerImpl::RelayRecoveredSig(const uint256& sigHash)
|
void PeerManagerImpl::RelayRecoveredSig(const uint256& sigHash)
|
||||||
{
|
{
|
||||||
const CInv inv{MSG_QUORUM_RECOVERED_SIG, sigHash};
|
const CInv inv{MSG_QUORUM_RECOVERED_SIG, sigHash};
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
for (const auto& [_, peer] : m_peer_map) {
|
for (const auto& [_, peer] : m_peer_map) {
|
||||||
if (peer->m_wants_recsigs) {
|
if (peer->m_wants_recsigs) {
|
||||||
PushInv(*peer, inv);
|
PushInv(*peer, inv);
|
||||||
@ -2477,7 +2477,7 @@ void PeerManagerImpl::RelayAddress(NodeId originator,
|
|||||||
std::array<std::pair<uint64_t, Peer*>, 2> best{{{0, nullptr}, {0, nullptr}}};
|
std::array<std::pair<uint64_t, Peer*>, 2> best{{{0, nullptr}, {0, nullptr}}};
|
||||||
assert(nRelayNodes <= best.size());
|
assert(nRelayNodes <= best.size());
|
||||||
|
|
||||||
LOCK(m_peer_mutex);
|
READ_LOCK(m_peer_mutex);
|
||||||
|
|
||||||
for (auto& [id, peer] : m_peer_map) {
|
for (auto& [id, peer] : m_peer_map) {
|
||||||
if (peer->m_addr_relay_enabled && id != originator && IsAddrCompatible(*peer, addr)) {
|
if (peer->m_addr_relay_enabled && id != originator && IsAddrCompatible(*peer, addr)) {
|
||||||
|
Loading…
Reference in New Issue
Block a user