mirror of
https://github.com/zerotier/ZeroTierOne.git
synced 2025-06-05 03:53:44 +02:00
Merge branch 'dev' of github.com:zerotier/ZeroTierOne into dev
This commit is contained in:
commit
fa58909d44
6 changed files with 229 additions and 146 deletions
220
node/Bond.cpp
220
node/Bond.cpp
|
@ -16,6 +16,7 @@
|
||||||
|
|
||||||
#include <cmath>
|
#include <cmath>
|
||||||
#include <string>
|
#include <string>
|
||||||
|
#include <cstdio>
|
||||||
|
|
||||||
namespace ZeroTier {
|
namespace ZeroTier {
|
||||||
|
|
||||||
|
@ -95,7 +96,6 @@ SharedPtr<Bond> Bond::createTransportTriggeredBond(const RuntimeEnvironment* ren
|
||||||
int64_t identity = peer->identity().address().toInt();
|
int64_t identity = peer->identity().address().toInt();
|
||||||
Bond* bond = nullptr;
|
Bond* bond = nullptr;
|
||||||
if (! _bonds.count(identity)) {
|
if (! _bonds.count(identity)) {
|
||||||
std::string policyAlias;
|
|
||||||
if (! _policyTemplateAssignments.count(identity)) {
|
if (! _policyTemplateAssignments.count(identity)) {
|
||||||
if (_defaultPolicy) {
|
if (_defaultPolicy) {
|
||||||
bond = new Bond(renv, _defaultPolicy, peer);
|
bond = new Bond(renv, _defaultPolicy, peer);
|
||||||
|
@ -103,7 +103,7 @@ SharedPtr<Bond> Bond::createTransportTriggeredBond(const RuntimeEnvironment* ren
|
||||||
}
|
}
|
||||||
if (! _defaultPolicy && _defaultPolicyStr.length()) {
|
if (! _defaultPolicy && _defaultPolicyStr.length()) {
|
||||||
bond = new Bond(renv, _bondPolicyTemplates[_defaultPolicyStr].ptr(), peer);
|
bond = new Bond(renv, _bondPolicyTemplates[_defaultPolicyStr].ptr(), peer);
|
||||||
bond->log("new default custom bond");
|
bond->log("new default custom bond (based on %s)", bond->getPolicyStrByCode(bond->policy()).c_str());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
|
@ -148,14 +148,12 @@ SharedPtr<Bond> Bond::createTransportTriggeredBond(const RuntimeEnvironment* ren
|
||||||
SharedPtr<Link> Bond::getLinkBySocket(const std::string& policyAlias, uint64_t localSocket)
|
SharedPtr<Link> Bond::getLinkBySocket(const std::string& policyAlias, uint64_t localSocket)
|
||||||
{
|
{
|
||||||
Mutex::Lock _l(_links_m);
|
Mutex::Lock _l(_links_m);
|
||||||
char ifname[32] = { 0 }; // 256 because interfaces on Windows can potentially be that long
|
char ifname[64] = { 0 };
|
||||||
_phy->getIfName((PhySocket*)((uintptr_t)localSocket), ifname, sizeof(ifname) - 1);
|
_phy->getIfName((PhySocket*)((uintptr_t)localSocket), ifname, sizeof(ifname) - 1);
|
||||||
// fprintf(stderr, "ifname %s\n",ifname);
|
|
||||||
std::string ifnameStr(ifname);
|
std::string ifnameStr(ifname);
|
||||||
auto search = _interfaceToLinkMap[policyAlias].find(ifnameStr);
|
auto search = _interfaceToLinkMap[policyAlias].find(ifnameStr);
|
||||||
if (search == _interfaceToLinkMap[policyAlias].end()) {
|
if (search == _interfaceToLinkMap[policyAlias].end()) {
|
||||||
// If the link wasn't already known, add a new entry
|
// If the link wasn't already known, add a new entry
|
||||||
// fprintf(stderr, "adding new link?? %s\n", ifnameStr.c_str());
|
|
||||||
SharedPtr<Link> s = new Link(ifnameStr, 0, 0, true, ZT_BOND_SLAVE_MODE_SPARE, "", 0.0);
|
SharedPtr<Link> s = new Link(ifnameStr, 0, 0, true, ZT_BOND_SLAVE_MODE_SPARE, "", 0.0);
|
||||||
_interfaceToLinkMap[policyAlias].insert(std::pair<std::string, SharedPtr<Link> >(ifnameStr, s));
|
_interfaceToLinkMap[policyAlias].insert(std::pair<std::string, SharedPtr<Link> >(ifnameStr, s));
|
||||||
return s;
|
return s;
|
||||||
|
@ -183,7 +181,6 @@ void Bond::processBackgroundTasks(void* tPtr, const int64_t now)
|
||||||
while (bondItr != _bonds.end()) {
|
while (bondItr != _bonds.end()) {
|
||||||
// Update Bond Controller's background processing timer
|
// Update Bond Controller's background processing timer
|
||||||
_currMinReqMonitorInterval = std::min(_currMinReqMonitorInterval, (unsigned long)(bondItr->second->monitorInterval()));
|
_currMinReqMonitorInterval = std::min(_currMinReqMonitorInterval, (unsigned long)(bondItr->second->monitorInterval()));
|
||||||
// Process bond tasks
|
|
||||||
bondItr->second->processBackgroundBondTasks(tPtr, now);
|
bondItr->second->processBackgroundBondTasks(tPtr, now);
|
||||||
++bondItr;
|
++bondItr;
|
||||||
}
|
}
|
||||||
|
@ -216,8 +213,6 @@ Bond::Bond(const RuntimeEnvironment* renv, SharedPtr<Bond> originalBond, const S
|
||||||
|
|
||||||
void Bond::nominatePathToBond(const SharedPtr<Path>& path, int64_t now)
|
void Bond::nominatePathToBond(const SharedPtr<Path>& path, int64_t now)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
path->address().toString(pathStr);
|
|
||||||
Mutex::Lock _l(_paths_m);
|
Mutex::Lock _l(_paths_m);
|
||||||
/**
|
/**
|
||||||
* Ensure the link is allowed and the path is not already present
|
* Ensure the link is allowed and the path is not already present
|
||||||
|
@ -261,7 +256,7 @@ void Bond::nominatePathToBond(const SharedPtr<Path>& path, int64_t now)
|
||||||
_paths[i].enabled = sl->enabled();
|
_paths[i].enabled = sl->enabled();
|
||||||
_paths[i].onlyPathOnLink = ! bFoundCommonLink;
|
_paths[i].onlyPathOnLink = ! bFoundCommonLink;
|
||||||
}
|
}
|
||||||
log("nominate link %s/%s (now in trial period)", getLink(path)->ifname().c_str(), pathStr);
|
log("nominate link %s", pathToStr(path).c_str());
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -365,9 +360,6 @@ SharedPtr<Path> Bond::getAppropriatePath(int64_t now, int32_t flowId)
|
||||||
|
|
||||||
void Bond::recordIncomingInvalidPacket(const SharedPtr<Path>& path)
|
void Bond::recordIncomingInvalidPacket(const SharedPtr<Path>& path)
|
||||||
{
|
{
|
||||||
// char pathStr[64] = { 0 }; path->address().toString(pathStr);
|
|
||||||
// log("%s (qos) Invalid packet on link %s/%s from peer %llx",
|
|
||||||
// getLink(path)->ifname().c_str(), pathStr);
|
|
||||||
Mutex::Lock _l(_paths_m);
|
Mutex::Lock _l(_paths_m);
|
||||||
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
||||||
if (_paths[i].p == path) {
|
if (_paths[i].p == path) {
|
||||||
|
@ -380,13 +372,6 @@ void Bond::recordOutgoingPacket(const SharedPtr<Path>& path, uint64_t packetId,
|
||||||
{
|
{
|
||||||
_freeRandomByte += (unsigned char)(packetId >> 8); // Grab entropy to use in path selection logic
|
_freeRandomByte += (unsigned char)(packetId >> 8); // Grab entropy to use in path selection logic
|
||||||
bool isFrame = (verb == Packet::Packet::VERB_ECHO || verb == Packet::VERB_FRAME || verb == Packet::VERB_EXT_FRAME);
|
bool isFrame = (verb == Packet::Packet::VERB_ECHO || verb == Packet::VERB_FRAME || verb == Packet::VERB_EXT_FRAME);
|
||||||
if (isFrame) {
|
|
||||||
// char pathStr[64] = { 0 };
|
|
||||||
// path->address().toString(pathStr);
|
|
||||||
// int pathIdx = getNominatedPathIdx(path);
|
|
||||||
// log("outgoing packet via [%d]", pathIdx);
|
|
||||||
// log("outgoing packet via %s/%s", getLink(path)->ifname().c_str(), pathStr);
|
|
||||||
}
|
|
||||||
bool shouldRecord = (packetId & (ZT_QOS_ACK_DIVISOR - 1) && (verb != Packet::VERB_ACK) && (verb != Packet::VERB_QOS_MEASUREMENT));
|
bool shouldRecord = (packetId & (ZT_QOS_ACK_DIVISOR - 1) && (verb != Packet::VERB_ACK) && (verb != Packet::VERB_QOS_MEASUREMENT));
|
||||||
if (isFrame || shouldRecord) {
|
if (isFrame || shouldRecord) {
|
||||||
Mutex::Lock _l(_paths_m);
|
Mutex::Lock _l(_paths_m);
|
||||||
|
@ -417,12 +402,6 @@ void Bond::recordOutgoingPacket(const SharedPtr<Path>& path, uint64_t packetId,
|
||||||
void Bond::recordIncomingPacket(const SharedPtr<Path>& path, uint64_t packetId, uint16_t payloadLength, Packet::Verb verb, int32_t flowId, int64_t now)
|
void Bond::recordIncomingPacket(const SharedPtr<Path>& path, uint64_t packetId, uint16_t payloadLength, Packet::Verb verb, int32_t flowId, int64_t now)
|
||||||
{
|
{
|
||||||
bool isFrame = (verb == Packet::Packet::VERB_ECHO || verb == Packet::VERB_FRAME || verb == Packet::VERB_EXT_FRAME);
|
bool isFrame = (verb == Packet::Packet::VERB_ECHO || verb == Packet::VERB_FRAME || verb == Packet::VERB_EXT_FRAME);
|
||||||
if (isFrame) {
|
|
||||||
// char pathStr[64] = { 0 }; path->address().toString(pathStr);
|
|
||||||
// int pathIdx = getNominatedPathIdx(path);
|
|
||||||
// log("incoming packet via [%d] [id=%llx, len=%d, verb=%d, flowId=%x]", pathIdx, packetId, payloadLength, verb, flowId);
|
|
||||||
// log("incoming packet via %s/%s (ls=%llx) [id=%llx, len=%d, verb=%d, flowId=%x]", getLink(path)->ifname().c_str(), pathStr, path->localSocket(), packetId, payloadLength, verb, flowId);
|
|
||||||
}
|
|
||||||
bool shouldRecord = (packetId & (ZT_QOS_ACK_DIVISOR - 1) && (verb != Packet::VERB_ACK) && (verb != Packet::VERB_QOS_MEASUREMENT));
|
bool shouldRecord = (packetId & (ZT_QOS_ACK_DIVISOR - 1) && (verb != Packet::VERB_ACK) && (verb != Packet::VERB_QOS_MEASUREMENT));
|
||||||
Mutex::Lock _l(_paths_m);
|
Mutex::Lock _l(_paths_m);
|
||||||
int pathIdx = getNominatedPathIdx(path);
|
int pathIdx = getNominatedPathIdx(path);
|
||||||
|
@ -474,8 +453,7 @@ void Bond::receivedQoS(const SharedPtr<Path>& path, int64_t now, int count, uint
|
||||||
if (pathIdx == ZT_MAX_PEER_NETWORK_PATHS) {
|
if (pathIdx == ZT_MAX_PEER_NETWORK_PATHS) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
// char pathStr[64] = { 0 }; path->address().toString(pathStr);
|
// log("received QoS packet (sampling %d frames) via %s", count, pathToStr(path).c_str());
|
||||||
// log("received QoS packet (sampling %d frames) via %s/%s", count, getLink(path)->ifname().c_str(), pathStr);
|
|
||||||
// Look up egress times and compute latency values for each record
|
// Look up egress times and compute latency values for each record
|
||||||
std::map<uint64_t, uint64_t>::iterator it;
|
std::map<uint64_t, uint64_t>::iterator it;
|
||||||
for (int j = 0; j < count; j++) {
|
for (int j = 0; j < count; j++) {
|
||||||
|
@ -510,12 +488,9 @@ int32_t Bond::generateQoSPacket(int pathIdx, int64_t now, char* qosBuffer)
|
||||||
|
|
||||||
bool Bond::assignFlowToBondedPath(SharedPtr<Flow>& flow, int64_t now)
|
bool Bond::assignFlowToBondedPath(SharedPtr<Flow>& flow, int64_t now)
|
||||||
{
|
{
|
||||||
char curPathStr[64] = { 0 };
|
|
||||||
unsigned int idx = ZT_MAX_PEER_NETWORK_PATHS;
|
unsigned int idx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
if (_policy == ZT_BOND_POLICY_BALANCE_XOR) {
|
if (_policy == ZT_BOND_POLICY_BALANCE_XOR) {
|
||||||
idx = abs((int)(flow->id % (_numBondedPaths)));
|
idx = abs((int)(flow->id % (_numBondedPaths)));
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[_bondIdxMap[idx]].p->localSocket());
|
|
||||||
_paths[_bondIdxMap[idx]].p->address().toString(curPathStr);
|
|
||||||
flow->assignPath(_bondIdxMap[idx], now);
|
flow->assignPath(_bondIdxMap[idx], now);
|
||||||
++(_paths[_bondIdxMap[idx]].assignedFlowCount);
|
++(_paths[_bondIdxMap[idx]].assignedFlowCount);
|
||||||
}
|
}
|
||||||
|
@ -542,8 +517,6 @@ bool Bond::assignFlowToBondedPath(SharedPtr<Flow>& flow, int64_t now)
|
||||||
entropy %= totalIncompleteAllocation;
|
entropy %= totalIncompleteAllocation;
|
||||||
for (unsigned int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
for (unsigned int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
||||||
if (_paths[i].p && _paths[i].bonded) {
|
if (_paths[i].p && _paths[i].bonded) {
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
|
||||||
_paths[i].p->address().toString(curPathStr);
|
|
||||||
uint8_t probabilitySegment = (_totalBondUnderload > 0) ? _paths[i].affinity : _paths[i].allocation;
|
uint8_t probabilitySegment = (_totalBondUnderload > 0) ? _paths[i].affinity : _paths[i].allocation;
|
||||||
if (entropy <= probabilitySegment) {
|
if (entropy <= probabilitySegment) {
|
||||||
idx = i;
|
idx = i;
|
||||||
|
@ -567,15 +540,13 @@ bool Bond::assignFlowToBondedPath(SharedPtr<Flow>& flow, int64_t now)
|
||||||
}
|
}
|
||||||
flow->assignPath(_abPathIdx, now);
|
flow->assignPath(_abPathIdx, now);
|
||||||
}
|
}
|
||||||
_paths[flow->assignedPath].p->address().toString(curPathStr);
|
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[flow->assignedPath].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[flow->assignedPath].p->localSocket());
|
||||||
log("assign out-flow %x to link %s/%s (%lu / %lu flows)", flow->id, link->ifname().c_str(), curPathStr, _paths[flow->assignedPath].assignedFlowCount, (unsigned long)_flows.size());
|
log("assign out-flow %04x to link %s (%lu / %lu flows)", flow->id, pathToStr(_paths[flow->assignedPath].p).c_str(), _paths[flow->assignedPath].assignedFlowCount, (unsigned long)_flows.size());
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
SharedPtr<Bond::Flow> Bond::createFlow(int pathIdx, int32_t flowId, unsigned char entropy, int64_t now)
|
SharedPtr<Bond::Flow> Bond::createFlow(int pathIdx, int32_t flowId, unsigned char entropy, int64_t now)
|
||||||
{
|
{
|
||||||
char curPathStr[64] = { 0 };
|
|
||||||
if (! _numBondedPaths) {
|
if (! _numBondedPaths) {
|
||||||
log("unable to assign flow %x (bond has no links)\n", flowId);
|
log("unable to assign flow %x (bond has no links)\n", flowId);
|
||||||
return SharedPtr<Flow>();
|
return SharedPtr<Flow>();
|
||||||
|
@ -593,10 +564,8 @@ SharedPtr<Bond::Flow> Bond::createFlow(int pathIdx, int32_t flowId, unsigned cha
|
||||||
*/
|
*/
|
||||||
if (pathIdx != ZT_MAX_PEER_NETWORK_PATHS) {
|
if (pathIdx != ZT_MAX_PEER_NETWORK_PATHS) {
|
||||||
flow->assignPath(pathIdx, now);
|
flow->assignPath(pathIdx, now);
|
||||||
_paths[pathIdx].p->address().toString(curPathStr);
|
|
||||||
_paths[pathIdx].assignedFlowCount++;
|
_paths[pathIdx].assignedFlowCount++;
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[flow->assignedPath].p->localSocket());
|
log("assign in-flow %x to link %s (%lu / %lu)", flow->id, pathToStr(_paths[pathIdx].p).c_str(), _paths[pathIdx].assignedFlowCount, (unsigned long)_flows.size());
|
||||||
log("assign in-flow %x to link %s/%s (%lu / %lu)", flow->id, link->ifname().c_str(), curPathStr, _paths[pathIdx].assignedFlowCount, (unsigned long)_flows.size());
|
|
||||||
}
|
}
|
||||||
/**
|
/**
|
||||||
* Add a flow when no path was provided. This means that it is an outgoing packet
|
* Add a flow when no path was provided. This means that it is an outgoing packet
|
||||||
|
@ -680,7 +649,6 @@ void Bond::processIncomingPathNegotiationRequest(uint64_t now, SharedPtr<Path>&
|
||||||
|
|
||||||
void Bond::pathNegotiationCheck(void* tPtr, int64_t now)
|
void Bond::pathNegotiationCheck(void* tPtr, int64_t now)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
int maxInPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
int maxInPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
int maxOutPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
int maxOutPathIdx = ZT_MAX_PEER_NETWORK_PATHS;
|
||||||
uint64_t maxInCount = 0;
|
uint64_t maxInCount = 0;
|
||||||
|
@ -718,8 +686,6 @@ void Bond::pathNegotiationCheck(void* tPtr, int64_t now)
|
||||||
sendPATH_NEGOTIATION_REQUEST(tPtr, _paths[maxOutPathIdx].p);
|
sendPATH_NEGOTIATION_REQUEST(tPtr, _paths[maxOutPathIdx].p);
|
||||||
++_numSentPathNegotiationRequests;
|
++_numSentPathNegotiationRequests;
|
||||||
_lastSentPathNegotiationRequest = now;
|
_lastSentPathNegotiationRequest = now;
|
||||||
_paths[maxOutPathIdx].p->address().toString(pathStr);
|
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[maxOutPathIdx].p->localSocket());
|
|
||||||
// fprintf(stderr, "sending request to use %s on %s, ls=%llx, utility=%d\n", pathStr, link->ifname().c_str(), _paths[maxOutPathIdx].p->localSocket(), _localUtility);
|
// fprintf(stderr, "sending request to use %s on %s, ls=%llx, utility=%d\n", pathStr, link->ifname().c_str(), _paths[maxOutPathIdx].p->localSocket(), _localUtility);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -738,9 +704,7 @@ void Bond::pathNegotiationCheck(void* tPtr, int64_t now)
|
||||||
|
|
||||||
void Bond::sendPATH_NEGOTIATION_REQUEST(void* tPtr, int pathIdx)
|
void Bond::sendPATH_NEGOTIATION_REQUEST(void* tPtr, int pathIdx)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
log("send link negotiation request to peer via link %s, local utility is %d", pathToStr(_paths[pathIdx].p).c_str(), _localUtility);
|
||||||
_paths[pathIdx].p->address().toString(pathStr);
|
|
||||||
log("send link negotiation request to peer via link %s/%s, local utility is %d", getLink(_paths[pathIdx].p)->ifname().c_str(), pathStr, _localUtility);
|
|
||||||
if (_abLinkSelectMethod != ZT_BOND_RESELECTION_POLICY_OPTIMIZE) {
|
if (_abLinkSelectMethod != ZT_BOND_RESELECTION_POLICY_OPTIMIZE) {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
@ -754,8 +718,6 @@ void Bond::sendPATH_NEGOTIATION_REQUEST(void* tPtr, int pathIdx)
|
||||||
|
|
||||||
void Bond::sendQOS_MEASUREMENT(void* tPtr, int pathIdx, int64_t localSocket, const InetAddress& atAddress, int64_t now)
|
void Bond::sendQOS_MEASUREMENT(void* tPtr, int pathIdx, int64_t localSocket, const InetAddress& atAddress, int64_t now)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
_paths[pathIdx].p->address().toString(pathStr);
|
|
||||||
int64_t _now = RR->node->now();
|
int64_t _now = RR->node->now();
|
||||||
Packet outp(_peer->_id.address(), RR->identity.address(), Packet::VERB_QOS_MEASUREMENT);
|
Packet outp(_peer->_id.address(), RR->identity.address(), Packet::VERB_QOS_MEASUREMENT);
|
||||||
char qosData[ZT_QOS_MAX_PACKET_SIZE];
|
char qosData[ZT_QOS_MAX_PACKET_SIZE];
|
||||||
|
@ -773,7 +735,7 @@ void Bond::sendQOS_MEASUREMENT(void* tPtr, int pathIdx, int64_t localSocket, con
|
||||||
_paths[pathIdx].packetsReceivedSinceLastQoS = 0;
|
_paths[pathIdx].packetsReceivedSinceLastQoS = 0;
|
||||||
_paths[pathIdx].lastQoSMeasurement = now;
|
_paths[pathIdx].lastQoSMeasurement = now;
|
||||||
}
|
}
|
||||||
// log("send QOS via link %s/%s (len=%d)", getLink(_paths[pathIdx].p)->ifname().c_str(), pathStr, len);
|
// log("send QOS via link %s (len=%d)", pathToStr(_paths[pathIdx].p).c_str(), len);
|
||||||
}
|
}
|
||||||
|
|
||||||
void Bond::processBackgroundBondTasks(void* tPtr, int64_t now)
|
void Bond::processBackgroundBondTasks(void* tPtr, int64_t now)
|
||||||
|
@ -794,17 +756,15 @@ void Bond::processBackgroundBondTasks(void* tPtr, int64_t now)
|
||||||
// Send ambient monitoring traffic
|
// Send ambient monitoring traffic
|
||||||
for (unsigned int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
for (unsigned int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
||||||
if (_paths[i].p && _paths[i].allowed()) {
|
if (_paths[i].p && _paths[i].allowed()) {
|
||||||
// ECHO (this is our bond's heartbeat)
|
|
||||||
if ((_monitorInterval > 0) && ((now - _paths[i].p->_lastOut) >= _monitorInterval)) {
|
if ((_monitorInterval > 0) && ((now - _paths[i].p->_lastOut) >= _monitorInterval)) {
|
||||||
if ((_peer->remoteVersionProtocol() >= 5) && (! ((_peer->remoteVersionMajor() == 1) && (_peer->remoteVersionMinor() == 1) && (_peer->remoteVersionRevision() == 0)))) {
|
if ((_peer->remoteVersionProtocol() >= 5) && (! ((_peer->remoteVersionMajor() == 1) && (_peer->remoteVersionMinor() == 1) && (_peer->remoteVersionRevision() == 0)))) {
|
||||||
Packet outp(_peer->address(), RR->identity.address(), Packet::VERB_ECHO);
|
Packet outp(_peer->address(), RR->identity.address(), Packet::VERB_ECHO); // ECHO (this is our bond's heartbeat)
|
||||||
outp.armor(_peer->key(), true, _peer->aesKeysIfSupported());
|
outp.armor(_peer->key(), true, _peer->aesKeysIfSupported());
|
||||||
RR->node->expectReplyTo(outp.packetId());
|
RR->node->expectReplyTo(outp.packetId());
|
||||||
RR->node->putPacket(tPtr, _paths[i].p->localSocket(), _paths[i].p->address(), outp.data(), outp.size());
|
RR->node->putPacket(tPtr, _paths[i].p->localSocket(), _paths[i].p->address(), outp.data(), outp.size());
|
||||||
|
_paths[i].p->_lastOut = now;
|
||||||
_overheadBytes += outp.size();
|
_overheadBytes += outp.size();
|
||||||
char pathStr[64] = { 0 };
|
log("sent ECHO via link %s", pathToStr(_paths[i].p).c_str());
|
||||||
_paths[i].p->address().toString(pathStr);
|
|
||||||
// log("send HELLO via link %s/%s (len=%d)", getLink(_paths[i].p)->ifname().c_str(), pathStr, outp.size());
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// QOS
|
// QOS
|
||||||
|
@ -837,7 +797,6 @@ void Bond::processBackgroundBondTasks(void* tPtr, int64_t now)
|
||||||
|
|
||||||
void Bond::curateBond(int64_t now, bool rebuildBond)
|
void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
uint8_t tmpNumAliveLinks = 0;
|
uint8_t tmpNumAliveLinks = 0;
|
||||||
uint8_t tmpNumTotalLinks = 0;
|
uint8_t tmpNumTotalLinks = 0;
|
||||||
/**
|
/**
|
||||||
|
@ -855,7 +814,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Determine alive-ness
|
* Determine aliveness
|
||||||
*/
|
*/
|
||||||
_paths[i].alive = (now - _paths[i].p->_lastIn) < _failoverInterval;
|
_paths[i].alive = (now - _paths[i].p->_lastIn) < _failoverInterval;
|
||||||
|
|
||||||
|
@ -880,12 +839,11 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
* Note eligibility state change (if any) and take appropriate action
|
* Note eligibility state change (if any) and take appropriate action
|
||||||
*/
|
*/
|
||||||
if (currEligibility != _paths[i].eligible) {
|
if (currEligibility != _paths[i].eligible) {
|
||||||
_paths[i].p->address().toString(pathStr);
|
|
||||||
if (currEligibility == 0) {
|
if (currEligibility == 0) {
|
||||||
log("link %s/%s is no longer eligible", getLink(_paths[i].p)->ifname().c_str(), pathStr);
|
log("link %s is no longer eligible", pathToStr(_paths[i].p).c_str());
|
||||||
}
|
}
|
||||||
if (currEligibility == 1) {
|
if (currEligibility == 1) {
|
||||||
log("link %s/%s is eligible", getLink(_paths[i].p)->ifname().c_str(), pathStr);
|
log("link %s is eligible", pathToStr(_paths[i].p).c_str());
|
||||||
}
|
}
|
||||||
dumpPathStatus(now, i);
|
dumpPathStatus(now, i);
|
||||||
if (currEligibility) {
|
if (currEligibility) {
|
||||||
|
@ -895,8 +853,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
_paths[i].adjustRefractoryPeriod(now, _defaultPathRefractoryPeriod, ! currEligibility);
|
_paths[i].adjustRefractoryPeriod(now, _defaultPathRefractoryPeriod, ! currEligibility);
|
||||||
if (_paths[i].bonded) {
|
if (_paths[i].bonded) {
|
||||||
if (_allowFlowHashing) {
|
if (_allowFlowHashing) {
|
||||||
_paths[i].p->address().toString(pathStr);
|
log("link %s was bonded, flow reallocation will occur soon", pathToStr(_paths[i].p).c_str());
|
||||||
log("link %s/%s was bonded, flow reallocation will occur soon", getLink(_paths[i].p)->ifname().c_str(), pathStr);
|
|
||||||
rebuildBond = true;
|
rebuildBond = true;
|
||||||
_paths[i].shouldReallocateFlows = _paths[i].bonded;
|
_paths[i].shouldReallocateFlows = _paths[i].bonded;
|
||||||
}
|
}
|
||||||
|
@ -917,19 +874,13 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
_numTotalLinks = tmpNumTotalLinks;
|
_numTotalLinks = tmpNumTotalLinks;
|
||||||
bool tmpHealthStatus = true;
|
bool tmpHealthStatus = true;
|
||||||
|
|
||||||
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
|
||||||
if (_numAliveLinks < 2) {
|
|
||||||
// Considered healthy if there is at least one backup link
|
|
||||||
tmpHealthStatus = false;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if (_policy == ZT_BOND_POLICY_BROADCAST) {
|
if (_policy == ZT_BOND_POLICY_BROADCAST) {
|
||||||
if (_numAliveLinks < 1) {
|
if (_numAliveLinks < 1) {
|
||||||
// Considered healthy if we're able to send frames at all
|
// Considered healthy if we're able to send frames at all
|
||||||
tmpHealthStatus = false;
|
tmpHealthStatus = false;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if ((_policy == ZT_BOND_POLICY_BALANCE_RR) || (_policy == ZT_BOND_POLICY_BALANCE_XOR) || (_policy == ZT_BOND_POLICY_BALANCE_AWARE)) {
|
if ((_policy == ZT_BOND_POLICY_BALANCE_RR) || (_policy == ZT_BOND_POLICY_BALANCE_XOR) || (_policy == ZT_BOND_POLICY_BALANCE_AWARE || (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP))) {
|
||||||
if (_numAliveLinks < _numTotalLinks) {
|
if (_numAliveLinks < _numTotalLinks) {
|
||||||
tmpHealthStatus = false;
|
tmpHealthStatus = false;
|
||||||
}
|
}
|
||||||
|
@ -942,7 +893,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
else {
|
else {
|
||||||
healthStatusStr = "DEGRADED";
|
healthStatusStr = "DEGRADED";
|
||||||
}
|
}
|
||||||
log("bond is in a %s state (links: %d/%d)", healthStatusStr.c_str(), _numAliveLinks, _numTotalLinks);
|
log("bond is %s (%d/%d links)", healthStatusStr.c_str(), _numAliveLinks, _numTotalLinks);
|
||||||
dumpInfo(now, true);
|
dumpInfo(now, true);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -959,7 +910,6 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
}
|
}
|
||||||
if (rebuildBond) {
|
if (rebuildBond) {
|
||||||
log("rebuilding bond");
|
log("rebuilding bond");
|
||||||
// TODO: Obey blacklisting
|
|
||||||
int updatedBondedPathCount = 0;
|
int updatedBondedPathCount = 0;
|
||||||
// Build map associating paths with local physical links. Will be selected from in next step
|
// Build map associating paths with local physical links. Will be selected from in next step
|
||||||
std::map<SharedPtr<Link>, std::vector<int> > linkMap;
|
std::map<SharedPtr<Link>, std::vector<int> > linkMap;
|
||||||
|
@ -984,8 +934,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
}
|
}
|
||||||
addPathToBond(idx, updatedBondedPathCount);
|
addPathToBond(idx, updatedBondedPathCount);
|
||||||
++updatedBondedPathCount;
|
++updatedBondedPathCount;
|
||||||
_paths[idx].p->address().toString(pathStr);
|
log("add %s (no user addr preference)", pathToStr(_paths[idx].p).c_str());
|
||||||
log("add %s/%s (no user addr preference)", link->ifname().c_str(), pathStr);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// If the user prefers to only use one address type (IPv4 or IPv6)
|
// If the user prefers to only use one address type (IPv4 or IPv6)
|
||||||
|
@ -996,14 +945,12 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (! _paths[idx].allowed()) {
|
if (! _paths[idx].allowed()) {
|
||||||
_paths[idx].p->address().toString(pathStr);
|
log("did not add %s (user addr preference %d)", pathToStr(_paths[idx].p).c_str(), ipvPref);
|
||||||
log("did not add %s/%s (user addr preference %d)", link->ifname().c_str(), pathStr, ipvPref);
|
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
addPathToBond(idx, updatedBondedPathCount);
|
addPathToBond(idx, updatedBondedPathCount);
|
||||||
++updatedBondedPathCount;
|
++updatedBondedPathCount;
|
||||||
_paths[idx].p->address().toString(pathStr);
|
log("add path %s (user addr preference %d)", pathToStr(_paths[idx].p).c_str(), ipvPref);
|
||||||
log("add path %s/%s (user addr preference %d)", link->ifname().c_str(), pathStr, ipvPref);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// If the users prefers one address type to another, try to find at least
|
// If the users prefers one address type to another, try to find at least
|
||||||
|
@ -1019,8 +966,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
if (_paths[idx].preferred()) {
|
if (_paths[idx].preferred()) {
|
||||||
addPathToBond(idx, updatedBondedPathCount);
|
addPathToBond(idx, updatedBondedPathCount);
|
||||||
++updatedBondedPathCount;
|
++updatedBondedPathCount;
|
||||||
_paths[idx].p->address().toString(pathStr);
|
log("add %s (user addr preference %d)", pathToStr(_paths[idx].p).c_str(), ipvPref);
|
||||||
log("add %s/%s (user addr preference %d)", link->ifname().c_str(), pathStr, ipvPref);
|
|
||||||
foundPreferredPath = true;
|
foundPreferredPath = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1034,8 +980,7 @@ void Bond::curateBond(int64_t now, bool rebuildBond)
|
||||||
}
|
}
|
||||||
addPathToBond(idx, updatedBondedPathCount);
|
addPathToBond(idx, updatedBondedPathCount);
|
||||||
++updatedBondedPathCount;
|
++updatedBondedPathCount;
|
||||||
_paths[idx].p->address().toString(pathStr);
|
log("add %s (user addr preference %d)", pathToStr(_paths[idx].p).c_str(), ipvPref);
|
||||||
log("add %s/%s (user addr preference %d)", link->ifname().c_str(), pathStr, ipvPref);
|
|
||||||
foundPreferredPath = true;
|
foundPreferredPath = true;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1151,8 +1096,6 @@ void Bond::estimatePathQuality(int64_t now)
|
||||||
|
|
||||||
void Bond::processBalanceTasks(int64_t now)
|
void Bond::processBalanceTasks(int64_t now)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
|
|
||||||
if (_allowFlowHashing) {
|
if (_allowFlowHashing) {
|
||||||
/**
|
/**
|
||||||
* Clean up and reset flows if necessary
|
* Clean up and reset flows if necessary
|
||||||
|
@ -1177,8 +1120,7 @@ void Bond::processBalanceTasks(int64_t now)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (! _paths[i].eligible && _paths[i].shouldReallocateFlows) {
|
if (! _paths[i].eligible && _paths[i].shouldReallocateFlows) {
|
||||||
_paths[i].p->address().toString(pathStr);
|
log("reallocate flows from dead link %s", pathToStr(_paths[i].p).c_str());
|
||||||
log("reallocate flows from dead link %s/%s", getLink(_paths[i].p)->ifname().c_str(), pathStr);
|
|
||||||
std::map<int32_t, SharedPtr<Flow> >::iterator flow_it = _flows.begin();
|
std::map<int32_t, SharedPtr<Flow> >::iterator flow_it = _flows.begin();
|
||||||
while (flow_it != _flows.end()) {
|
while (flow_it != _flows.end()) {
|
||||||
if (_paths[flow_it->second->assignedPath].p == _paths[i].p) {
|
if (_paths[flow_it->second->assignedPath].p == _paths[i].p) {
|
||||||
|
@ -1214,8 +1156,7 @@ void Bond::processBalanceTasks(int64_t now)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
if (_paths[i].p && _paths[i].bonded && _paths[i].eligible && (_paths[i].allocation < minimumAllocationValue) && _paths[i].assignedFlowCount) {
|
if (_paths[i].p && _paths[i].bonded && _paths[i].eligible && (_paths[i].allocation < minimumAllocationValue) && _paths[i].assignedFlowCount) {
|
||||||
_paths[i].p->address().toString(pathStr);
|
log("reallocate flows from under-performing link %s\n", pathToStr(_paths[i].p).c_str());
|
||||||
log("reallocate flows from under-performing link %s/%s\n", getLink(_paths[i].p)->ifname().c_str(), pathStr);
|
|
||||||
std::map<int32_t, SharedPtr<Flow> >::iterator flow_it = _flows.begin();
|
std::map<int32_t, SharedPtr<Flow> >::iterator flow_it = _flows.begin();
|
||||||
while (flow_it != _flows.end()) {
|
while (flow_it != _flows.end()) {
|
||||||
if (flow_it->second->assignedPath == _paths[i].p) {
|
if (flow_it->second->assignedPath == _paths[i].p) {
|
||||||
|
@ -1249,14 +1190,10 @@ void Bond::dequeueNextActiveBackupPath(uint64_t now)
|
||||||
|
|
||||||
bool Bond::abForciblyRotateLink()
|
bool Bond::abForciblyRotateLink()
|
||||||
{
|
{
|
||||||
char prevPathStr[64];
|
|
||||||
char curPathStr[64];
|
|
||||||
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
if (_policy == ZT_BOND_POLICY_ACTIVE_BACKUP) {
|
||||||
int prevPathIdx = _abPathIdx;
|
int prevPathIdx = _abPathIdx;
|
||||||
_paths[_abPathIdx].p->address().toString(prevPathStr);
|
|
||||||
dequeueNextActiveBackupPath(RR->node->now());
|
dequeueNextActiveBackupPath(RR->node->now());
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("active link rotated from %s to %s", pathToStr(_paths[prevPathIdx].p).c_str(), pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("forcibly rotate link from %s/%s to %s/%s", getLink(_paths[prevPathIdx].p)->ifname().c_str(), prevPathStr, getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
return false;
|
return false;
|
||||||
|
@ -1264,9 +1201,6 @@ bool Bond::abForciblyRotateLink()
|
||||||
|
|
||||||
void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
|
||||||
char prevPathStr[64];
|
|
||||||
char curPathStr[64];
|
|
||||||
int prevActiveBackupPathIdx = _abPathIdx;
|
int prevActiveBackupPathIdx = _abPathIdx;
|
||||||
int nonPreferredPathIdx;
|
int nonPreferredPathIdx;
|
||||||
bool bFoundPrimaryLink = false;
|
bool bFoundPrimaryLink = false;
|
||||||
|
@ -1280,11 +1214,10 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
log("no active link");
|
log("no active link");
|
||||||
}
|
}
|
||||||
else if (_paths[_abPathIdx].p) {
|
else if (_paths[_abPathIdx].p) {
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("active link is %s, failover queue size is %zu", pathToStr(_paths[_abPathIdx].p).c_str(), _abFailoverQueue.size());
|
||||||
log("active link is %s/%s, failover queue size is %zu", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr, _abFailoverQueue.size());
|
|
||||||
}
|
}
|
||||||
if (_abFailoverQueue.empty()) {
|
if (_abFailoverQueue.empty()) {
|
||||||
log("failover queue is empty, no longer fault-tolerant");
|
log("failover queue is empty, bond is no longer fault-tolerant");
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1304,10 +1237,9 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
log("no user-specified links");
|
log("no user-specified links");
|
||||||
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
||||||
if (_paths[i].p && _paths[i].eligible) {
|
if (_paths[i].p && _paths[i].eligible) {
|
||||||
_paths[i].p->address().toString(curPathStr);
|
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
||||||
if (link) {
|
if (link) {
|
||||||
log("found eligible link %s/%s", getLink(_paths[i].p)->ifname().c_str(), curPathStr);
|
log("found eligible link %s", pathToStr(_paths[i].p).c_str());
|
||||||
_abPathIdx = i;
|
_abPathIdx = i;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
@ -1327,18 +1259,16 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
||||||
if (_paths[i].eligible && link->primary()) {
|
if (_paths[i].eligible && link->primary()) {
|
||||||
if (! _paths[i].preferred()) {
|
if (! _paths[i].preferred()) {
|
||||||
_paths[i].p->address().toString(curPathStr);
|
|
||||||
// Found path on primary link, take note in case we don't find a preferred path
|
// Found path on primary link, take note in case we don't find a preferred path
|
||||||
nonPreferredPathIdx = i;
|
nonPreferredPathIdx = i;
|
||||||
bFoundPrimaryLink = true;
|
bFoundPrimaryLink = true;
|
||||||
}
|
}
|
||||||
if (_paths[i].preferred()) {
|
if (_paths[i].preferred()) {
|
||||||
_abPathIdx = i;
|
_abPathIdx = i;
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
|
||||||
bFoundPrimaryLink = true;
|
bFoundPrimaryLink = true;
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
||||||
if (link) {
|
if (link) {
|
||||||
log("found preferred primary link %s/%s", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
log("found preferred primary link %s", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
}
|
}
|
||||||
break; // Found preferred path on primary link
|
break; // Found preferred path on primary link
|
||||||
}
|
}
|
||||||
|
@ -1364,8 +1294,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS) {
|
if (_abPathIdx != ZT_MAX_PEER_NETWORK_PATHS) {
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[_abPathIdx].p->localSocket());
|
||||||
if (link) {
|
if (link) {
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("select non-primary link %s", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("select non-primary link %s/%s", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1380,11 +1309,10 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
// Remove ineligible paths from the failover link queue
|
// Remove ineligible paths from the failover link queue
|
||||||
for (std::deque<int>::iterator it(_abFailoverQueue.begin()); it != _abFailoverQueue.end();) {
|
for (std::deque<int>::iterator it(_abFailoverQueue.begin()); it != _abFailoverQueue.end();) {
|
||||||
if (_paths[(*it)].p && ! _paths[(*it)].eligible) {
|
if (_paths[(*it)].p && ! _paths[(*it)].eligible) {
|
||||||
_paths[(*it)].p->address().toString(curPathStr);
|
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[(*it)].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[(*it)].p->localSocket());
|
||||||
it = _abFailoverQueue.erase(it);
|
it = _abFailoverQueue.erase(it);
|
||||||
if (link) {
|
if (link) {
|
||||||
log("link %s/%s is now ineligible, removing from failover queue (%zu links in queue)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr, _abFailoverQueue.size());
|
log("link %s is ineligible, removing from failover queue (%zu links in queue)", pathToStr(_paths[_abPathIdx].p).c_str(), _abFailoverQueue.size());
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
|
@ -1410,7 +1338,6 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
SharedPtr<Link> link = RR->bc->getLinkBySocket(_policyAlias, _paths[i].p->localSocket());
|
||||||
_paths[i].p->address().toString(pathStr);
|
|
||||||
|
|
||||||
int failoverScoreHandicap = _paths[i].failoverScore;
|
int failoverScoreHandicap = _paths[i].failoverScore;
|
||||||
if (_paths[i].preferred()) {
|
if (_paths[i].preferred()) {
|
||||||
|
@ -1432,7 +1359,6 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
if (failoverLink) {
|
if (failoverLink) {
|
||||||
for (int j = 0; j < ZT_MAX_PEER_NETWORK_PATHS; j++) {
|
for (int j = 0; j < ZT_MAX_PEER_NETWORK_PATHS; j++) {
|
||||||
if (_paths[j].p && getLink(_paths[j].p) == failoverLink.ptr()) {
|
if (_paths[j].p && getLink(_paths[j].p) == failoverLink.ptr()) {
|
||||||
_paths[j].p->address().toString(pathStr);
|
|
||||||
int inheritedHandicap = failoverScoreHandicap - 10;
|
int inheritedHandicap = failoverScoreHandicap - 10;
|
||||||
int newHandicap = _paths[j].failoverScore > inheritedHandicap ? _paths[j].failoverScore : inheritedHandicap;
|
int newHandicap = _paths[j].failoverScore > inheritedHandicap ? _paths[j].failoverScore : inheritedHandicap;
|
||||||
if (! _paths[j].preferred()) {
|
if (! _paths[j].preferred()) {
|
||||||
|
@ -1451,8 +1377,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
}
|
}
|
||||||
if (! bFoundPathInQueue) {
|
if (! bFoundPathInQueue) {
|
||||||
_abFailoverQueue.push_front(i);
|
_abFailoverQueue.push_front(i);
|
||||||
_paths[i].p->address().toString(curPathStr);
|
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
||||||
log("add link %s/%s to failover queue (%zu links in queue)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr, _abFailoverQueue.size());
|
|
||||||
addPathToBond(0, i);
|
addPathToBond(0, i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1497,8 +1422,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
}
|
}
|
||||||
if (! bFoundPathInQueue) {
|
if (! bFoundPathInQueue) {
|
||||||
_abFailoverQueue.push_front(i);
|
_abFailoverQueue.push_front(i);
|
||||||
_paths[i].p->address().toString(curPathStr);
|
log("add link %s to failover queue (%zu links in queue)", pathToStr(_paths[i].p).c_str(), _abFailoverQueue.size());
|
||||||
log("add link %s/%s to failover queue (%zu links in queue)", getLink(_paths[i].p)->ifname().c_str(), curPathStr, _abFailoverQueue.size());
|
|
||||||
addPathToBond(0, i);
|
addPathToBond(0, i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1528,12 +1452,10 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
* Fulfill primary re-select obligations
|
* Fulfill primary re-select obligations
|
||||||
*/
|
*/
|
||||||
if (_paths[_abPathIdx].p && ! _paths[_abPathIdx].eligible) { // Implicit ZT_BOND_RESELECTION_POLICY_FAILURE
|
if (_paths[_abPathIdx].p && ! _paths[_abPathIdx].eligible) { // Implicit ZT_BOND_RESELECTION_POLICY_FAILURE
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("link %s has failed, select link from failover queue (%zu links in queue)", pathToStr(_paths[_abPathIdx].p).c_str(), _abFailoverQueue.size());
|
||||||
log("link %s/%s has failed, select link from failover queue (%zu links in queue)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr, _abFailoverQueue.size());
|
|
||||||
if (! _abFailoverQueue.empty()) {
|
if (! _abFailoverQueue.empty()) {
|
||||||
dequeueNextActiveBackupPath(now);
|
dequeueNextActiveBackupPath(now);
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("active link switched to %s", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("active link switched to %s/%s", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
log("failover queue is empty, no links to choose from");
|
log("failover queue is empty, no links to choose from");
|
||||||
|
@ -1548,8 +1470,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
if (_abLinkSelectMethod == ZT_BOND_RESELECTION_POLICY_ALWAYS) {
|
if (_abLinkSelectMethod == ZT_BOND_RESELECTION_POLICY_ALWAYS) {
|
||||||
if (_paths[_abPathIdx].p && ! getLink(_paths[_abPathIdx].p)->primary() && getLink(_paths[_abFailoverQueue.front()].p)->primary()) {
|
if (_paths[_abPathIdx].p && ! getLink(_paths[_abPathIdx].p)->primary() && getLink(_paths[_abFailoverQueue.front()].p)->primary()) {
|
||||||
dequeueNextActiveBackupPath(now);
|
dequeueNextActiveBackupPath(now);
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("switch back to available primary link %s (select mode: always)", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("switch back to available primary link %s/%s (select: always)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (_abLinkSelectMethod == ZT_BOND_RESELECTION_POLICY_BETTER) {
|
if (_abLinkSelectMethod == ZT_BOND_RESELECTION_POLICY_BETTER) {
|
||||||
|
@ -1557,8 +1478,7 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
// Active backup has switched to "better" primary link according to re-select policy.
|
// Active backup has switched to "better" primary link according to re-select policy.
|
||||||
if (getLink(_paths[_abFailoverQueue.front()].p)->primary() && (_paths[_abFailoverQueue.front()].failoverScore > _paths[_abPathIdx].failoverScore)) {
|
if (getLink(_paths[_abFailoverQueue.front()].p)->primary() && (_paths[_abFailoverQueue.front()].failoverScore > _paths[_abPathIdx].failoverScore)) {
|
||||||
dequeueNextActiveBackupPath(now);
|
dequeueNextActiveBackupPath(now);
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("switch back to user-defined primary link %s (select mode: better)", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("switch back to user-defined primary link %s/%s (select: better)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1568,10 +1488,8 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
*/
|
*/
|
||||||
if (_paths[_abFailoverQueue.front()].negotiated) {
|
if (_paths[_abFailoverQueue.front()].negotiated) {
|
||||||
dequeueNextActiveBackupPath(now);
|
dequeueNextActiveBackupPath(now);
|
||||||
_paths[_abPathIdx].p->address().toString(prevPathStr);
|
|
||||||
_lastPathNegotiationCheck = now;
|
_lastPathNegotiationCheck = now;
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("switch negotiated link %s (select mode: optimize)", pathToStr(_paths[_abPathIdx].p).c_str());
|
||||||
log("switch negotiated link %s/%s (select: optimize)", getLink(_paths[_abPathIdx].p)->ifname().c_str(), curPathStr);
|
|
||||||
}
|
}
|
||||||
else {
|
else {
|
||||||
// Try to find a better path and automatically switch to it -- not too often, though.
|
// Try to find a better path and automatically switch to it -- not too often, though.
|
||||||
|
@ -1584,18 +1502,12 @@ void Bond::processActiveBackupTasks(void* tPtr, int64_t now)
|
||||||
int thresholdQuantity = (int)(ZT_BOND_ACTIVE_BACKUP_OPTIMIZE_MIN_THRESHOLD * (float)_paths[_abPathIdx].allocation);
|
int thresholdQuantity = (int)(ZT_BOND_ACTIVE_BACKUP_OPTIMIZE_MIN_THRESHOLD * (float)_paths[_abPathIdx].allocation);
|
||||||
if ((failoverScoreDifference > 0) && (failoverScoreDifference > thresholdQuantity)) {
|
if ((failoverScoreDifference > 0) && (failoverScoreDifference > thresholdQuantity)) {
|
||||||
SharedPtr<Path> oldPath = _paths[_abPathIdx].p;
|
SharedPtr<Path> oldPath = _paths[_abPathIdx].p;
|
||||||
_paths[_abPathIdx].p->address().toString(prevPathStr);
|
|
||||||
dequeueNextActiveBackupPath(now);
|
dequeueNextActiveBackupPath(now);
|
||||||
_paths[_abPathIdx].p->address().toString(curPathStr);
|
log("switch from %s (score: %d) to better link %s (score: %d) (select mode: optimize)",
|
||||||
log("ab",
|
pathToStr(oldPath).c_str(),
|
||||||
"switch from %s/%s (score: %d) to better link %s/%s (score: %d) for peer %llx (select: optimize)",
|
|
||||||
getLink(oldPath)->ifname().c_str(),
|
|
||||||
prevPathStr,
|
|
||||||
prevFScore,
|
prevFScore,
|
||||||
getLink(_paths[_abPathIdx].p)->ifname().c_str(),
|
pathToStr(_paths[_abPathIdx].p).c_str(),
|
||||||
curPathStr,
|
newFScore);
|
||||||
newFScore,
|
|
||||||
_peerId);
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -1717,6 +1629,7 @@ void Bond::setBondParameters(int policy, SharedPtr<Bond> templateBond, bool useT
|
||||||
/* If a user has specified custom parameters for this bonding policy, overlay them onto the defaults */
|
/* If a user has specified custom parameters for this bonding policy, overlay them onto the defaults */
|
||||||
if (useTemplate) {
|
if (useTemplate) {
|
||||||
_policyAlias = templateBond->_policyAlias;
|
_policyAlias = templateBond->_policyAlias;
|
||||||
|
_policy = templateBond->policy();
|
||||||
_failoverInterval = templateBond->_failoverInterval >= ZT_BOND_FAILOVER_MIN_INTERVAL ? templateBond->_failoverInterval : ZT_BOND_FAILOVER_MIN_INTERVAL;
|
_failoverInterval = templateBond->_failoverInterval >= ZT_BOND_FAILOVER_MIN_INTERVAL ? templateBond->_failoverInterval : ZT_BOND_FAILOVER_MIN_INTERVAL;
|
||||||
_downDelay = templateBond->_downDelay;
|
_downDelay = templateBond->_downDelay;
|
||||||
_upDelay = templateBond->_upDelay;
|
_upDelay = templateBond->_upDelay;
|
||||||
|
@ -1751,28 +1664,44 @@ SharedPtr<Link> Bond::getLink(const SharedPtr<Path>& path)
|
||||||
return RR->bc->getLinkBySocket(_policyAlias, path->localSocket());
|
return RR->bc->getLinkBySocket(_policyAlias, path->localSocket());
|
||||||
}
|
}
|
||||||
|
|
||||||
|
std::string Bond::pathToStr(const SharedPtr<Path>& path)
|
||||||
|
{
|
||||||
|
#ifdef ZT_TRACE
|
||||||
|
char pathStr[64] = { 0 };
|
||||||
|
char fullPathStr[256] = { 0 };
|
||||||
|
path->address().toString(pathStr);
|
||||||
|
snprintf(fullPathStr, 256, "%.16llx-%s/%s", (unsigned long long)(path->localSocket()), getLink(path)->ifname().c_str(), pathStr);
|
||||||
|
return std::string(fullPathStr);
|
||||||
|
#else
|
||||||
|
return "";
|
||||||
|
#endif
|
||||||
|
}
|
||||||
|
|
||||||
void Bond::dumpPathStatus(int64_t now, int pathIdx)
|
void Bond::dumpPathStatus(int64_t now, int pathIdx)
|
||||||
{
|
{
|
||||||
char pathStr[64] = { 0 };
|
#ifdef ZT_TRACE
|
||||||
_paths[pathIdx].p->address().toString(pathStr);
|
std::string aliveOrDead = _paths[pathIdx].alive ? std::string("alive") : std::string("dead");
|
||||||
log("path status: [%2d] alive:%d, eli:%d, bonded:%d, flows:%6d, lat:%10.3f, jitter:%10.3f, error:%6.4f, loss:%6.4f, age:%6d alloc:%d--- (%s/%s)",
|
std::string eligibleOrNot = _paths[pathIdx].eligible ? std::string("eligible") : std::string("ineligible");
|
||||||
|
std::string bondedOrNot = _paths[pathIdx].bonded ? std::string("bonded") : std::string("unbonded");
|
||||||
|
log("path[%2d] --- %5s (%7d), %10s, %8s, flows=%-6d lat=%-8.3f pdv=%-7.3f err=%-6.4f loss=%-6.4f alloc=%-3d --- (%s)",
|
||||||
pathIdx,
|
pathIdx,
|
||||||
_paths[pathIdx].alive,
|
aliveOrDead.c_str(),
|
||||||
_paths[pathIdx].eligible,
|
_paths[pathIdx].p->age(now),
|
||||||
_paths[pathIdx].bonded,
|
eligibleOrNot.c_str(),
|
||||||
|
bondedOrNot.c_str(),
|
||||||
_paths[pathIdx].assignedFlowCount,
|
_paths[pathIdx].assignedFlowCount,
|
||||||
_paths[pathIdx].latencyMean,
|
_paths[pathIdx].latencyMean,
|
||||||
_paths[pathIdx].latencyVariance,
|
_paths[pathIdx].latencyVariance,
|
||||||
_paths[pathIdx].packetErrorRatio,
|
_paths[pathIdx].packetErrorRatio,
|
||||||
_paths[pathIdx].packetLossRatio,
|
_paths[pathIdx].packetLossRatio,
|
||||||
_paths[pathIdx].p->age(now),
|
|
||||||
_paths[pathIdx].allocation,
|
_paths[pathIdx].allocation,
|
||||||
getLink(_paths[pathIdx].p)->ifname().c_str(),
|
pathToStr(_paths[pathIdx].p).c_str());
|
||||||
pathStr);
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
void Bond::dumpInfo(int64_t now, bool force)
|
void Bond::dumpInfo(int64_t now, bool force)
|
||||||
{
|
{
|
||||||
|
#ifdef ZT_TRACE
|
||||||
uint64_t timeSinceLastDump = now - _lastSummaryDump;
|
uint64_t timeSinceLastDump = now - _lastSummaryDump;
|
||||||
if (! force && timeSinceLastDump < ZT_BOND_STATUS_INTERVAL) {
|
if (! force && timeSinceLastDump < ZT_BOND_STATUS_INTERVAL) {
|
||||||
return;
|
return;
|
||||||
|
@ -1780,12 +1709,21 @@ void Bond::dumpInfo(int64_t now, bool force)
|
||||||
_lastSummaryDump = now;
|
_lastSummaryDump = now;
|
||||||
float overhead = (_overheadBytes / (timeSinceLastDump / 1000.0f) / 1000.0f);
|
float overhead = (_overheadBytes / (timeSinceLastDump / 1000.0f) / 1000.0f);
|
||||||
_overheadBytes = 0;
|
_overheadBytes = 0;
|
||||||
log("bond status: bp: %d, fi: %d, mi: %d, ud: %d, dd: %d, flows: %lu, ambient: %f KB/s", _policy, _failoverInterval, _monitorInterval, _upDelay, _downDelay, (unsigned long)_flows.size(), overhead);
|
log("bond: bp=%d, fi=%d, mi=%d, ud=%d, dd=%d, flows=%lu, overhead=%f KB/s",
|
||||||
|
_policy,
|
||||||
|
_failoverInterval,
|
||||||
|
_monitorInterval,
|
||||||
|
_upDelay,
|
||||||
|
_downDelay,
|
||||||
|
(unsigned long)_flows.size(),
|
||||||
|
overhead);
|
||||||
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
for (int i = 0; i < ZT_MAX_PEER_NETWORK_PATHS; ++i) {
|
||||||
if (_paths[i].p) {
|
if (_paths[i].p) {
|
||||||
dumpPathStatus(now, i);
|
dumpPathStatus(now, i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
log("");
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
} // namespace ZeroTier
|
} // namespace ZeroTier
|
||||||
|
|
|
@ -518,6 +518,7 @@ class Bond {
|
||||||
|
|
||||||
public:
|
public:
|
||||||
void dumpInfo(int64_t now, bool force);
|
void dumpInfo(int64_t now, bool force);
|
||||||
|
std::string pathToStr(const SharedPtr<Path>& path);
|
||||||
void dumpPathStatus(int64_t now, int pathIdx);
|
void dumpPathStatus(int64_t now, int pathIdx);
|
||||||
|
|
||||||
SharedPtr<Link> getLink(const SharedPtr<Path>& path);
|
SharedPtr<Link> getLink(const SharedPtr<Path>& path);
|
||||||
|
@ -1000,12 +1001,12 @@ class Bond {
|
||||||
/**
|
/**
|
||||||
* @param policy Bonding policy for this bond
|
* @param policy Bonding policy for this bond
|
||||||
*/
|
*/
|
||||||
/*
|
|
||||||
inline void setPolicy(uint8_t policy)
|
inline void setPolicy(uint8_t policy)
|
||||||
{
|
{
|
||||||
_policy = policy;
|
_policy = policy;
|
||||||
}
|
}
|
||||||
*/
|
|
||||||
/**
|
/**
|
||||||
* @return the current bonding policy
|
* @return the current bonding policy
|
||||||
*/
|
*/
|
||||||
|
@ -1138,6 +1139,7 @@ class Bond {
|
||||||
*/
|
*/
|
||||||
void log(const char* fmt, ...)
|
void log(const char* fmt, ...)
|
||||||
{
|
{
|
||||||
|
#ifdef ZT_TRACE
|
||||||
time_t rawtime;
|
time_t rawtime;
|
||||||
struct tm* timeinfo;
|
struct tm* timeinfo;
|
||||||
char timestamp[80];
|
char timestamp[80];
|
||||||
|
@ -1157,6 +1159,7 @@ class Bond {
|
||||||
va_end(args);
|
va_end(args);
|
||||||
RR->t->bondStateMessage(NULL, traceMsg);
|
RR->t->bondStateMessage(NULL, traceMsg);
|
||||||
#undef MAX_MSG_LEN
|
#undef MAX_MSG_LEN
|
||||||
|
#endif
|
||||||
}
|
}
|
||||||
|
|
||||||
private:
|
private:
|
||||||
|
|
|
@ -3,6 +3,7 @@
|
||||||
|
|
||||||
#include <vector>
|
#include <vector>
|
||||||
#include "../node/InetAddress.hpp"
|
#include "../node/InetAddress.hpp"
|
||||||
|
#include "../node/MAC.hpp"
|
||||||
|
|
||||||
namespace ZeroTier {
|
namespace ZeroTier {
|
||||||
|
|
||||||
|
@ -11,6 +12,8 @@ class MacDNSHelper
|
||||||
public:
|
public:
|
||||||
static void setDNS(uint64_t nwid, const char *domain, const std::vector<InetAddress> &servers);
|
static void setDNS(uint64_t nwid, const char *domain, const std::vector<InetAddress> &servers);
|
||||||
static void removeDNS(uint64_t nwid);
|
static void removeDNS(uint64_t nwid);
|
||||||
|
static bool addIps(uint64_t nwid, const MAC mac, const char *dev, const std::vector<InetAddress> &addrs);
|
||||||
|
static bool removeIps(uint64_t nwid);
|
||||||
};
|
};
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
|
@ -6,6 +6,11 @@
|
||||||
|
|
||||||
namespace ZeroTier {
|
namespace ZeroTier {
|
||||||
|
|
||||||
|
static void printKeys (const void* key, const void* value, void* context) {
|
||||||
|
CFShow(key);
|
||||||
|
CFShow(value);
|
||||||
|
}
|
||||||
|
|
||||||
void MacDNSHelper::setDNS(uint64_t nwid, const char *domain, const std::vector<InetAddress> &servers)
|
void MacDNSHelper::setDNS(uint64_t nwid, const char *domain, const std::vector<InetAddress> &servers)
|
||||||
{
|
{
|
||||||
SCDynamicStoreRef ds = SCDynamicStoreCreate(NULL, CFSTR("zerotier"), NULL, NULL);
|
SCDynamicStoreRef ds = SCDynamicStoreCreate(NULL, CFSTR("zerotier"), NULL, NULL);
|
||||||
|
@ -85,4 +90,130 @@ void MacDNSHelper::removeDNS(uint64_t nwid)
|
||||||
CFRelease(ds);
|
CFRelease(ds);
|
||||||
}
|
}
|
||||||
|
|
||||||
}
|
// Make macOS believe we do in fact have ipv6 connectivity and that it should resolve dns names
|
||||||
|
// over ipv6 if we ask for them.
|
||||||
|
// Originally I planned to put all the v6 ip addresses from the network into the config.
|
||||||
|
// But only the link local address is necessary and sufficient. Added other v6 addresses
|
||||||
|
// doesn't do anything.
|
||||||
|
bool MacDNSHelper::addIps(uint64_t nwid, const MAC mac, const char *dev, const std::vector<InetAddress>& addrs)
|
||||||
|
{
|
||||||
|
|
||||||
|
bool hasV6 = false;
|
||||||
|
for (unsigned int i = 0; i < addrs.size(); ++i) {
|
||||||
|
if (addrs[i].isV6()) {
|
||||||
|
hasV6 = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!hasV6) {
|
||||||
|
MacDNSHelper::removeIps(nwid);
|
||||||
|
return true;
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
SCDynamicStoreRef ds = SCDynamicStoreCreate(NULL, CFSTR("zerotier"), NULL, NULL);
|
||||||
|
char buf[256] = { 0 };
|
||||||
|
sprintf(buf, "State:/Network/Service/%.16llx/IPv6", nwid);
|
||||||
|
|
||||||
|
InetAddress ll = InetAddress::makeIpv6LinkLocal(mac);
|
||||||
|
char buf2[256] = {0};
|
||||||
|
const char* llStr = ll.toIpString(buf2);
|
||||||
|
|
||||||
|
|
||||||
|
CFStringRef key = CFStringCreateWithCString(NULL, buf, kCFStringEncodingUTF8);
|
||||||
|
CFStringRef* cfaddrs = new CFStringRef[1];
|
||||||
|
CFStringRef* cfprefixes = new CFStringRef[1];
|
||||||
|
CFStringRef* cfdestaddrs = new CFStringRef[1];
|
||||||
|
CFStringRef* cfflags = new CFStringRef[1];
|
||||||
|
|
||||||
|
|
||||||
|
cfaddrs[0] = CFStringCreateWithCString(NULL, llStr, kCFStringEncodingUTF8);
|
||||||
|
cfprefixes[0] = CFStringCreateWithCString(NULL, "64", kCFStringEncodingUTF8);
|
||||||
|
cfdestaddrs[0] = CFStringCreateWithCString(NULL, "::ffff:ffff:ffff:ffff:0:0", kCFStringEncodingUTF8);
|
||||||
|
cfflags[0] = CFStringCreateWithCString(NULL, "0", kCFStringEncodingUTF8);
|
||||||
|
|
||||||
|
CFArrayRef addrArray = CFArrayCreate(NULL, (const void**)cfaddrs, 1, &kCFTypeArrayCallBacks);
|
||||||
|
CFArrayRef prefixArray = CFArrayCreate(NULL, (const void**)cfprefixes, 1, &kCFTypeArrayCallBacks);
|
||||||
|
CFArrayRef destArray = CFArrayCreate(NULL, (const void**)cfdestaddrs, 1, &kCFTypeArrayCallBacks);
|
||||||
|
CFArrayRef flagsArray = CFArrayCreate(NULL, (const void**)cfflags, 1, &kCFTypeArrayCallBacks);
|
||||||
|
CFStringRef cfdev = CFStringCreateWithCString(NULL, dev, kCFStringEncodingUTF8);
|
||||||
|
|
||||||
|
const int SIZE = 5;
|
||||||
|
CFStringRef keys[SIZE];
|
||||||
|
keys[0] = CFSTR("Addresses");
|
||||||
|
keys[1] = CFSTR("DestAddresses");
|
||||||
|
keys[2] = CFSTR("Flags");
|
||||||
|
keys[3] = CFSTR("InterfaceName");
|
||||||
|
keys[4] = CFSTR("PrefixLength");
|
||||||
|
|
||||||
|
CFTypeRef values[SIZE];
|
||||||
|
values[0] = addrArray;
|
||||||
|
values[1] = destArray;
|
||||||
|
values[2] = flagsArray;
|
||||||
|
// values[3] = devArray;
|
||||||
|
values[3] = cfdev;
|
||||||
|
values[4] = prefixArray;
|
||||||
|
|
||||||
|
|
||||||
|
CFDictionaryRef dict = CFDictionaryCreate(NULL,
|
||||||
|
(const void**)keys, (const void**)values, SIZE, &kCFCopyStringDictionaryKeyCallBacks,
|
||||||
|
&kCFTypeDictionaryValueCallBacks);
|
||||||
|
|
||||||
|
// CFDictionaryApplyFunction(dict, printKeys, NULL);
|
||||||
|
|
||||||
|
CFArrayRef list = SCDynamicStoreCopyKeyList(ds, key);
|
||||||
|
CFIndex i = 0, j = CFArrayGetCount(list);
|
||||||
|
bool addrsChanged = true;
|
||||||
|
CFPropertyListRef oldAddrs = NULL;
|
||||||
|
|
||||||
|
bool ret = TRUE;
|
||||||
|
if (j > 0) {
|
||||||
|
oldAddrs = SCDynamicStoreCopyValue(ds, (CFStringRef)CFArrayGetValueAtIndex(list, i));
|
||||||
|
addrsChanged = !CFEqual(oldAddrs,dict);
|
||||||
|
}
|
||||||
|
if (addrsChanged) {
|
||||||
|
if (j <= 0) {
|
||||||
|
ret &= SCDynamicStoreAddValue(ds, key, dict);
|
||||||
|
} else {
|
||||||
|
ret &= SCDynamicStoreSetValue(ds, (CFStringRef)CFArrayGetValueAtIndex(list, i), dict);
|
||||||
|
}
|
||||||
|
if (!ret) {
|
||||||
|
fprintf(stderr, "Error writing IPv6 configuration\n");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
CFRelease(addrArray);
|
||||||
|
CFRelease(prefixArray);
|
||||||
|
CFRelease(destArray);
|
||||||
|
CFRelease(flagsArray);
|
||||||
|
CFRelease(cfdev);
|
||||||
|
|
||||||
|
CFRelease(list);
|
||||||
|
CFRelease(dict);
|
||||||
|
|
||||||
|
CFRelease(ds);
|
||||||
|
CFRelease(key);
|
||||||
|
|
||||||
|
delete[] cfaddrs;
|
||||||
|
delete[] cfprefixes;
|
||||||
|
delete[] cfdestaddrs;
|
||||||
|
delete[] cfflags;
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
bool MacDNSHelper::removeIps(uint64_t nwid)
|
||||||
|
{
|
||||||
|
SCDynamicStoreRef ds = SCDynamicStoreCreate(NULL, CFSTR("zerotier"), NULL, NULL);
|
||||||
|
|
||||||
|
char buf[256] = {0};
|
||||||
|
sprintf(buf, "State:/Network/Service/%.16llx/IPv6", nwid);
|
||||||
|
CFStringRef key = CFStringCreateWithCString(NULL, buf, kCFStringEncodingUTF8);
|
||||||
|
bool res = SCDynamicStoreRemoveValue(ds, key);
|
||||||
|
CFRelease(key);
|
||||||
|
CFRelease(ds);
|
||||||
|
|
||||||
|
return res;
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
|
@ -244,6 +244,7 @@ MacEthernetTap::~MacEthernetTap()
|
||||||
pid_t pid0,pid1;
|
pid_t pid0,pid1;
|
||||||
|
|
||||||
MacDNSHelper::removeDNS(_nwid);
|
MacDNSHelper::removeDNS(_nwid);
|
||||||
|
MacDNSHelper::removeIps(_nwid);
|
||||||
|
|
||||||
Mutex::Lock _gl(globalTapCreateLock);
|
Mutex::Lock _gl(globalTapCreateLock);
|
||||||
::write(_shutdownSignalPipe[1],"\0",1); // causes thread to exit
|
::write(_shutdownSignalPipe[1],"\0",1); // causes thread to exit
|
||||||
|
|
|
@ -1769,7 +1769,8 @@ public:
|
||||||
if (basePolicyStr.empty()) {
|
if (basePolicyStr.empty()) {
|
||||||
fprintf(stderr, "error: no base policy was specified for custom policy (%s)\n", customPolicyStr.c_str());
|
fprintf(stderr, "error: no base policy was specified for custom policy (%s)\n", customPolicyStr.c_str());
|
||||||
}
|
}
|
||||||
if (_node->bondController()->getPolicyCodeByStr(basePolicyStr) == ZT_BOND_POLICY_NONE) {
|
int basePolicyCode = _node->bondController()->getPolicyCodeByStr(basePolicyStr);
|
||||||
|
if (basePolicyCode == ZT_BOND_POLICY_NONE) {
|
||||||
fprintf(stderr, "error: custom policy (%s) is invalid, unknown base policy (%s).\n",
|
fprintf(stderr, "error: custom policy (%s) is invalid, unknown base policy (%s).\n",
|
||||||
customPolicyStr.c_str(), basePolicyStr.c_str());
|
customPolicyStr.c_str(), basePolicyStr.c_str());
|
||||||
continue;
|
continue;
|
||||||
|
@ -1781,6 +1782,7 @@ public:
|
||||||
// New bond, used as a copy template for new instances
|
// New bond, used as a copy template for new instances
|
||||||
SharedPtr<Bond> newTemplateBond = new Bond(NULL, basePolicyStr, customPolicyStr, SharedPtr<Peer>());
|
SharedPtr<Bond> newTemplateBond = new Bond(NULL, basePolicyStr, customPolicyStr, SharedPtr<Peer>());
|
||||||
// Acceptable ranges
|
// Acceptable ranges
|
||||||
|
newTemplateBond->setPolicy(basePolicyCode);
|
||||||
newTemplateBond->setMaxAcceptableLatency(OSUtils::jsonInt(customPolicy["maxAcceptableLatency"],-1));
|
newTemplateBond->setMaxAcceptableLatency(OSUtils::jsonInt(customPolicy["maxAcceptableLatency"],-1));
|
||||||
newTemplateBond->setMaxAcceptableMeanLatency(OSUtils::jsonInt(customPolicy["maxAcceptableMeanLatency"],-1));
|
newTemplateBond->setMaxAcceptableMeanLatency(OSUtils::jsonInt(customPolicy["maxAcceptableMeanLatency"],-1));
|
||||||
newTemplateBond->setMaxAcceptablePacketDelayVariance(OSUtils::jsonInt(customPolicy["maxAcceptablePacketDelayVariance"],-1));
|
newTemplateBond->setMaxAcceptablePacketDelayVariance(OSUtils::jsonInt(customPolicy["maxAcceptablePacketDelayVariance"],-1));
|
||||||
|
@ -1805,7 +1807,7 @@ public:
|
||||||
newTemplateBond->setUpDelay(OSUtils::jsonInt(customPolicy["upDelay"],-1));
|
newTemplateBond->setUpDelay(OSUtils::jsonInt(customPolicy["upDelay"],-1));
|
||||||
newTemplateBond->setDownDelay(OSUtils::jsonInt(customPolicy["downDelay"],-1));
|
newTemplateBond->setDownDelay(OSUtils::jsonInt(customPolicy["downDelay"],-1));
|
||||||
newTemplateBond->setFlowRebalanceStrategy(OSUtils::jsonInt(customPolicy["flowRebalanceStrategy"],(uint64_t)0));
|
newTemplateBond->setFlowRebalanceStrategy(OSUtils::jsonInt(customPolicy["flowRebalanceStrategy"],(uint64_t)0));
|
||||||
newTemplateBond->setFailoverInterval(OSUtils::jsonInt(customPolicy["failoverInterval"],(uint64_t)0));
|
newTemplateBond->setFailoverInterval(OSUtils::jsonInt(customPolicy["failoverInterval"],ZT_BOND_FAILOVER_DEFAULT_INTERVAL));
|
||||||
newTemplateBond->setPacketsPerLink(OSUtils::jsonInt(customPolicy["packetsPerLink"],-1));
|
newTemplateBond->setPacketsPerLink(OSUtils::jsonInt(customPolicy["packetsPerLink"],-1));
|
||||||
|
|
||||||
// Policy-Specific link set
|
// Policy-Specific link set
|
||||||
|
@ -2041,6 +2043,11 @@ public:
|
||||||
fprintf(stderr,"ERROR: unable to add ip address %s" ZT_EOL_S, ip->toString(ipbuf));
|
fprintf(stderr,"ERROR: unable to add ip address %s" ZT_EOL_S, ip->toString(ipbuf));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#ifdef __APPLE__
|
||||||
|
if (!MacDNSHelper::addIps(n.config.nwid, n.config.mac, n.tap->deviceName().c_str(), newManagedIps))
|
||||||
|
fprintf(stderr, "ERROR: unable to add v6 addresses to system configuration" ZT_EOL_S);
|
||||||
|
#endif
|
||||||
#endif
|
#endif
|
||||||
n.managedIps.swap(newManagedIps);
|
n.managedIps.swap(newManagedIps);
|
||||||
}
|
}
|
||||||
|
|
Loading…
Add table
Reference in a new issue