mirror of
https://github.com/zerotier/ZeroTierOne.git
synced 2025-06-06 20:43:44 +02:00
Fix TRACE and CLUSTER builds.
This commit is contained in:
parent
e4896b257f
commit
91c9f4cb20
4 changed files with 12 additions and 12 deletions
|
@ -346,7 +346,7 @@ void Cluster::handleIncomingStateMessage(const void *msg,unsigned int len)
|
||||||
Mutex::Lock _l(_remotePeers_m);
|
Mutex::Lock _l(_remotePeers_m);
|
||||||
_RemotePeer &rp = _remotePeers[std::pair<Address,unsigned int>(id.address(),(unsigned int)fromMemberId)];
|
_RemotePeer &rp = _remotePeers[std::pair<Address,unsigned int>(id.address(),(unsigned int)fromMemberId)];
|
||||||
if (!rp.lastHavePeerReceived) {
|
if (!rp.lastHavePeerReceived) {
|
||||||
RR->topology->saveIdentity(id);
|
RR->topology->saveIdentity((void *)0,id);
|
||||||
RR->identity.agree(id,rp.key,ZT_PEER_SECRET_KEY_LENGTH);
|
RR->identity.agree(id,rp.key,ZT_PEER_SECRET_KEY_LENGTH);
|
||||||
}
|
}
|
||||||
rp.lastHavePeerReceived = RR->node->now();
|
rp.lastHavePeerReceived = RR->node->now();
|
||||||
|
@ -459,7 +459,7 @@ void Cluster::handleIncomingStateMessage(const void *msg,unsigned int len)
|
||||||
Mutex::Lock _l2(_members[fromMemberId].lock);
|
Mutex::Lock _l2(_members[fromMemberId].lock);
|
||||||
_send(fromMemberId,CLUSTER_MESSAGE_PROXY_SEND,rendezvousForRemote.data(),rendezvousForRemote.size());
|
_send(fromMemberId,CLUSTER_MESSAGE_PROXY_SEND,rendezvousForRemote.data(),rendezvousForRemote.size());
|
||||||
}
|
}
|
||||||
RR->sw->send(rendezvousForLocal,true);
|
RR->sw->send((void *)0,rendezvousForLocal,true);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
} break;
|
} break;
|
||||||
|
@ -470,7 +470,7 @@ void Cluster::handleIncomingStateMessage(const void *msg,unsigned int len)
|
||||||
const unsigned int len = dmsg.at<uint16_t>(ptr); ptr += 2;
|
const unsigned int len = dmsg.at<uint16_t>(ptr); ptr += 2;
|
||||||
Packet outp(rcpt,RR->identity.address(),verb);
|
Packet outp(rcpt,RR->identity.address(),verb);
|
||||||
outp.append(dmsg.field(ptr,len),len); ptr += len;
|
outp.append(dmsg.field(ptr,len),len); ptr += len;
|
||||||
RR->sw->send(outp,true);
|
RR->sw->send((void *)0,outp,true);
|
||||||
//TRACE("[%u] proxy send %s to %s length %u",(unsigned int)fromMemberId,Packet::verbString(verb),rcpt.toString().c_str(),len);
|
//TRACE("[%u] proxy send %s to %s length %u",(unsigned int)fromMemberId,Packet::verbString(verb),rcpt.toString().c_str(),len);
|
||||||
} break;
|
} break;
|
||||||
|
|
||||||
|
@ -479,7 +479,7 @@ void Cluster::handleIncomingStateMessage(const void *msg,unsigned int len)
|
||||||
if (network) {
|
if (network) {
|
||||||
// Copy into a Packet just to conform to Network API. Eventually
|
// Copy into a Packet just to conform to Network API. Eventually
|
||||||
// will want to refactor.
|
// will want to refactor.
|
||||||
network->handleConfigChunk(0,Address(),Buffer<ZT_PROTO_MAX_PACKET_LENGTH>(dmsg),ptr);
|
network->handleConfigChunk((void *)0,0,Address(),Buffer<ZT_PROTO_MAX_PACKET_LENGTH>(dmsg),ptr);
|
||||||
}
|
}
|
||||||
} break;
|
} break;
|
||||||
}
|
}
|
||||||
|
@ -576,7 +576,7 @@ bool Cluster::sendViaCluster(int mostRecentMemberId,const Address &toPeerAddress
|
||||||
for(std::vector<InetAddress>::const_iterator i2(_members[mostRecentMemberId].zeroTierPhysicalEndpoints.begin());i2!=_members[mostRecentMemberId].zeroTierPhysicalEndpoints.end();++i2) {
|
for(std::vector<InetAddress>::const_iterator i2(_members[mostRecentMemberId].zeroTierPhysicalEndpoints.begin());i2!=_members[mostRecentMemberId].zeroTierPhysicalEndpoints.end();++i2) {
|
||||||
if (i1->ss_family == i2->ss_family) {
|
if (i1->ss_family == i2->ss_family) {
|
||||||
TRACE("sendViaCluster sending %u bytes to %s by way of %u (%s->%s)",len,toPeerAddress.toString().c_str(),(unsigned int)mostRecentMemberId,i1->toString().c_str(),i2->toString().c_str());
|
TRACE("sendViaCluster sending %u bytes to %s by way of %u (%s->%s)",len,toPeerAddress.toString().c_str(),(unsigned int)mostRecentMemberId,i1->toString().c_str(),i2->toString().c_str());
|
||||||
RR->node->putPacket(*i1,*i2,data,len);
|
RR->node->putPacket((void *)0,*i1,*i2,data,len);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -679,7 +679,7 @@ void Cluster::relayViaCluster(const Address &fromPeerAddress,const Address &toPe
|
||||||
for(std::vector<InetAddress>::const_iterator i2(_members[mostRecentMemberId].zeroTierPhysicalEndpoints.begin());i2!=_members[mostRecentMemberId].zeroTierPhysicalEndpoints.end();++i2) {
|
for(std::vector<InetAddress>::const_iterator i2(_members[mostRecentMemberId].zeroTierPhysicalEndpoints.begin());i2!=_members[mostRecentMemberId].zeroTierPhysicalEndpoints.end();++i2) {
|
||||||
if (i1->ss_family == i2->ss_family) {
|
if (i1->ss_family == i2->ss_family) {
|
||||||
TRACE("relayViaCluster relaying %u bytes from %s to %s by way of %u (%s->%s)",len,fromPeerAddress.toString().c_str(),toPeerAddress.toString().c_str(),(unsigned int)mostRecentMemberId,i1->toString().c_str(),i2->toString().c_str());
|
TRACE("relayViaCluster relaying %u bytes from %s to %s by way of %u (%s->%s)",len,fromPeerAddress.toString().c_str(),toPeerAddress.toString().c_str(),(unsigned int)mostRecentMemberId,i1->toString().c_str(),i2->toString().c_str());
|
||||||
RR->node->putPacket(*i1,*i2,data,len);
|
RR->node->putPacket((void *)0,*i1,*i2,data,len);
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -981,7 +981,7 @@ void Cluster::_flush(uint16_t memberId)
|
||||||
void Cluster::_doREMOTE_WHOIS(uint64_t fromMemberId,const Packet &remotep)
|
void Cluster::_doREMOTE_WHOIS(uint64_t fromMemberId,const Packet &remotep)
|
||||||
{
|
{
|
||||||
if (remotep.payloadLength() >= ZT_ADDRESS_LENGTH) {
|
if (remotep.payloadLength() >= ZT_ADDRESS_LENGTH) {
|
||||||
Identity queried(RR->topology->getIdentity(Address(remotep.payload(),ZT_ADDRESS_LENGTH)));
|
Identity queried(RR->topology->getIdentity((void *)0,Address(remotep.payload(),ZT_ADDRESS_LENGTH)));
|
||||||
if (queried) {
|
if (queried) {
|
||||||
Buffer<1024> routp;
|
Buffer<1024> routp;
|
||||||
remotep.source().appendTo(routp);
|
remotep.source().appendTo(routp);
|
||||||
|
|
|
@ -285,7 +285,7 @@ ZT_ResultCode Node::processBackgroundTasks(void *tptr,uint64_t now,volatile uint
|
||||||
#ifdef ZT_ENABLE_CLUSTER
|
#ifdef ZT_ENABLE_CLUSTER
|
||||||
// If clustering is enabled we have to call cluster->doPeriodicTasks() very often, so we override normal timer deadline behavior
|
// If clustering is enabled we have to call cluster->doPeriodicTasks() very often, so we override normal timer deadline behavior
|
||||||
if (RR->cluster) {
|
if (RR->cluster) {
|
||||||
RR->sw->doTimerTasks(now);
|
RR->sw->doTimerTasks(tptr,now);
|
||||||
RR->cluster->doPeriodicTasks();
|
RR->cluster->doPeriodicTasks();
|
||||||
*nextBackgroundTaskDeadline = now + ZT_CLUSTER_PERIODIC_TASK_PERIOD; // this is really short so just tick at this rate
|
*nextBackgroundTaskDeadline = now + ZT_CLUSTER_PERIODIC_TASK_PERIOD; // this is really short so just tick at this rate
|
||||||
} else {
|
} else {
|
||||||
|
@ -686,7 +686,7 @@ void Node::postTrace(const char *module,unsigned int line,const char *fmt,...)
|
||||||
tmp2[sizeof(tmp2)-1] = (char)0;
|
tmp2[sizeof(tmp2)-1] = (char)0;
|
||||||
|
|
||||||
Utils::snprintf(tmp1,sizeof(tmp1),"[%s] %s:%u %s",nowstr,module,line,tmp2);
|
Utils::snprintf(tmp1,sizeof(tmp1),"[%s] %s:%u %s",nowstr,module,line,tmp2);
|
||||||
postEvent(ZT_EVENT_TRACE,tmp1);
|
postEvent((void *)0,ZT_EVENT_TRACE,tmp1);
|
||||||
}
|
}
|
||||||
#endif // ZT_TRACE
|
#endif // ZT_TRACE
|
||||||
|
|
||||||
|
|
|
@ -103,7 +103,7 @@ void Peer::received(
|
||||||
}
|
}
|
||||||
outp.append((uint16_t)redirectTo.port());
|
outp.append((uint16_t)redirectTo.port());
|
||||||
outp.armor(_key,true,path->nextOutgoingCounter());
|
outp.armor(_key,true,path->nextOutgoingCounter());
|
||||||
path->send(RR,outp.data(),outp.size(),now);
|
path->send(RR,tPtr,outp.data(),outp.size(),now);
|
||||||
} else {
|
} else {
|
||||||
// For older peers we use RENDEZVOUS to coax them into contacting us elsewhere.
|
// For older peers we use RENDEZVOUS to coax them into contacting us elsewhere.
|
||||||
Packet outp(_id.address(),RR->identity.address(),Packet::VERB_RENDEZVOUS);
|
Packet outp(_id.address(),RR->identity.address(),Packet::VERB_RENDEZVOUS);
|
||||||
|
@ -118,7 +118,7 @@ void Peer::received(
|
||||||
outp.append(redirectTo.rawIpData(),16);
|
outp.append(redirectTo.rawIpData(),16);
|
||||||
}
|
}
|
||||||
outp.armor(_key,true,path->nextOutgoingCounter());
|
outp.armor(_key,true,path->nextOutgoingCounter());
|
||||||
path->send(RR,outp.data(),outp.size(),now);
|
path->send(RR,tPtr,outp.data(),outp.size(),now);
|
||||||
}
|
}
|
||||||
suboptimalPath = true;
|
suboptimalPath = true;
|
||||||
}
|
}
|
||||||
|
|
|
@ -863,7 +863,7 @@ bool Switch::_trySend(void *tPtr,Packet &packet,bool encrypt)
|
||||||
Packet::Fragment frag(packet,fragStart,chunkSize,fno,totalFragments);
|
Packet::Fragment frag(packet,fragStart,chunkSize,fno,totalFragments);
|
||||||
#ifdef ZT_ENABLE_CLUSTER
|
#ifdef ZT_ENABLE_CLUSTER
|
||||||
if (viaPath)
|
if (viaPath)
|
||||||
viaPath->send(RR,frag.data(),frag.size(),now);
|
viaPath->send(RR,tPtr,frag.data(),frag.size(),now);
|
||||||
else if (clusterMostRecentMemberId >= 0)
|
else if (clusterMostRecentMemberId >= 0)
|
||||||
RR->cluster->sendViaCluster(clusterMostRecentMemberId,destination,frag.data(),frag.size());
|
RR->cluster->sendViaCluster(clusterMostRecentMemberId,destination,frag.data(),frag.size());
|
||||||
#else
|
#else
|
||||||
|
|
Loading…
Add table
Reference in a new issue