5
0
mirror of https://github.com/cwinfo/yggdrasil-go.git synced 2024-11-16 00:50:33 +00:00
yggdrasil-go/src/yggdrasil/switch.go

509 lines
18 KiB
Go
Raw Normal View History

2017-12-29 04:16:20 +00:00
package yggdrasil
// This part constructs a spanning tree of the network
// It routes packets based on distance on the spanning tree
// In general, this is *not* equivalent to routing on the tree
// It falls back to the tree in the worst case, but it can take shortcuts too
// This is the part that makse routing reasonably efficient on scale-free graphs
// TODO document/comment everything in a lot more detail
// TODO? use a pre-computed lookup table (python version had this)
2018-06-07 19:24:02 +00:00
// A little annoying to do with constant changes from backpressure
2017-12-29 04:16:20 +00:00
2018-06-12 22:50:08 +00:00
import (
"sort"
"sync"
"sync/atomic"
"time"
)
2017-12-29 04:16:20 +00:00
const switch_timeout = time.Minute
const switch_updateInterval = switch_timeout / 2
const switch_throttle = switch_updateInterval / 2
2017-12-29 04:16:20 +00:00
// The switch locator represents the topology and network state dependent info about a node, minus the signatures that go with it.
// Nodes will pick the best root they see, provided that the root continues to push out updates with new timestamps.
// The coords represent a path from the root to a node.
// This path is generally part of a spanning tree, except possibly the last hop (it can loop when sending coords to your parent, but they see this and know not to use a looping path).
2017-12-29 04:16:20 +00:00
type switchLocator struct {
2018-01-04 22:37:51 +00:00
root sigPubKey
tstamp int64
coords []switchPort
2017-12-29 04:16:20 +00:00
}
// Returns true if the first sigPubKey has a higher TreeID.
2017-12-29 04:16:20 +00:00
func firstIsBetter(first, second *sigPubKey) bool {
2018-01-04 22:37:51 +00:00
// Higher TreeID is better
ftid := getTreeID(first)
stid := getTreeID(second)
for idx := 0; idx < len(ftid); idx++ {
if ftid[idx] == stid[idx] {
continue
}
return ftid[idx] > stid[idx]
}
// Edge case, when comparing identical IDs
return false
2017-12-29 04:16:20 +00:00
}
// Returns a copy of the locator which can safely be mutated.
2017-12-29 04:16:20 +00:00
func (l *switchLocator) clone() switchLocator {
2018-01-04 22:37:51 +00:00
// Used to create a deep copy for use in messages
// Copy required because we need to mutate coords before sending
// (By appending the port from us to the destination)
loc := *l
loc.coords = make([]switchPort, len(l.coords), len(l.coords)+1)
copy(loc.coords, l.coords)
return loc
2017-12-29 04:16:20 +00:00
}
// Gets the distance a locator is from the provided destination coords, with the coords provided in []byte format (used to compress integers sent over the wire).
2017-12-29 04:16:20 +00:00
func (l *switchLocator) dist(dest []byte) int {
2018-01-04 22:37:51 +00:00
// Returns distance (on the tree) from these coords
offset := 0
fdc := 0
for {
if fdc >= len(l.coords) {
break
}
coord, length := wire_decode_uint64(dest[offset:])
if length == 0 {
break
}
if l.coords[fdc] != switchPort(coord) {
break
}
fdc++
offset += length
}
dist := len(l.coords[fdc:])
for {
_, length := wire_decode_uint64(dest[offset:])
if length == 0 {
break
}
dist++
offset += length
}
return dist
2017-12-29 04:16:20 +00:00
}
// Gets coords in wire encoded format, with *no* length prefix.
2017-12-29 04:16:20 +00:00
func (l *switchLocator) getCoords() []byte {
2018-01-04 22:37:51 +00:00
bs := make([]byte, 0, len(l.coords))
for _, coord := range l.coords {
c := wire_encode_uint64(uint64(coord))
bs = append(bs, c...)
}
return bs
2017-12-29 04:16:20 +00:00
}
// Returns true if the this locator represents an ancestor of the locator given as an argument.
// Ancestor means that it's the parent node, or the parent of parent, and so on...
2017-12-29 04:16:20 +00:00
func (x *switchLocator) isAncestorOf(y *switchLocator) bool {
2018-01-04 22:37:51 +00:00
if x.root != y.root {
return false
}
if len(x.coords) > len(y.coords) {
return false
}
for idx := range x.coords {
if x.coords[idx] != y.coords[idx] {
return false
}
}
return true
2017-12-29 04:16:20 +00:00
}
// Information about a peer, used by the switch to build the tree and eventually make routing decisions.
2017-12-29 04:16:20 +00:00
type peerInfo struct {
2018-01-04 22:37:51 +00:00
key sigPubKey // ID of this peer
locator switchLocator // Should be able to respond with signatures upon request
degree uint64 // Self-reported degree
time time.Time // Time this node was last seen
firstSeen time.Time
port switchPort // Interface number of this peer
msg switchMsg // The wire switchMsg used
2017-12-29 04:16:20 +00:00
}
// This is just a uint64 with a named type for clarity reasons.
2017-12-29 04:16:20 +00:00
type switchPort uint64
// This is the subset of the information about a peer needed to make routing decisions, and it stored separately in an atomically accessed table, which gets hammered in the "hot loop" of the routing logic (see: peer.handleTraffic in peers.go).
2017-12-29 04:16:20 +00:00
type tableElem struct {
2018-05-16 22:48:53 +00:00
port switchPort
locator switchLocator
2017-12-29 04:16:20 +00:00
}
// This is the subset of the information about all peers needed to make routing decisions, and it stored separately in an atomically accessed table, which gets hammered in the "hot loop" of the routing logic (see: peer.handleTraffic in peers.go).
2017-12-29 04:16:20 +00:00
type lookupTable struct {
2018-01-04 22:37:51 +00:00
self switchLocator
2018-01-19 00:48:34 +00:00
elems []tableElem
2017-12-29 04:16:20 +00:00
}
// This is switch information which is mutable and needs to be modified by other goroutines, but is not accessed atomically.
// Use the switchTable functions to access it safely using the RWMutex for synchronization.
2017-12-29 04:16:20 +00:00
type switchData struct {
2018-01-04 22:37:51 +00:00
// All data that's mutable and used by exported Table methods
// To be read/written with atomic.Value Store/Load calls
locator switchLocator
seq uint64 // Sequence number, reported to peers, so they know about changes
peers map[switchPort]peerInfo
msg *switchMsg
2017-12-29 04:16:20 +00:00
}
// All the information stored by the switch.
2017-12-29 04:16:20 +00:00
type switchTable struct {
2018-01-04 22:37:51 +00:00
core *Core
key sigPubKey // Our own key
time time.Time // Time when locator.tstamp was last updated
parent switchPort // Port of whatever peer is our parent, or self if we're root
drop map[sigPubKey]int64 // Tstamp associated with a dropped root
mutex sync.RWMutex // Lock for reads/writes of switchData
data switchData
updater atomic.Value //*sync.Once
table atomic.Value //lookupTable
2017-12-29 04:16:20 +00:00
}
// Initializes the switchTable struct.
2017-12-29 04:16:20 +00:00
func (t *switchTable) init(core *Core, key sigPubKey) {
2018-01-04 22:37:51 +00:00
now := time.Now()
t.core = core
t.key = key
locator := switchLocator{root: key, tstamp: now.Unix()}
peers := make(map[switchPort]peerInfo)
t.data = switchData{locator: locator, peers: peers}
t.updater.Store(&sync.Once{})
2018-01-19 00:48:34 +00:00
t.table.Store(lookupTable{})
2018-01-04 22:37:51 +00:00
t.drop = make(map[sigPubKey]int64)
2018-05-27 21:13:37 +00:00
}
// Safely gets a copy of this node's locator.
2017-12-29 04:16:20 +00:00
func (t *switchTable) getLocator() switchLocator {
2018-01-04 22:37:51 +00:00
t.mutex.RLock()
defer t.mutex.RUnlock()
return t.data.locator.clone()
2017-12-29 04:16:20 +00:00
}
// Regular maintenance to possibly timeout/reset the root and similar.
func (t *switchTable) doMaintenance() {
2018-01-04 22:37:51 +00:00
// Periodic maintenance work to keep things internally consistent
t.mutex.Lock() // Write lock
defer t.mutex.Unlock() // Release lock when we're done
t.cleanRoot()
t.cleanDropped()
2017-12-29 04:16:20 +00:00
}
// Updates the root periodically if it is ourself, or promotes ourself to root if we're better than the current root or if the current root has timed out.
2017-12-29 04:16:20 +00:00
func (t *switchTable) cleanRoot() {
2018-01-04 22:37:51 +00:00
// TODO rethink how this is done?...
// Get rid of the root if it looks like its timed out
now := time.Now()
doUpdate := false
if now.Sub(t.time) > switch_timeout {
dropped := t.data.peers[t.parent]
dropped.time = t.time
t.drop[t.data.locator.root] = t.data.locator.tstamp
doUpdate = true
}
// Or, if we're better than our root, root ourself
if firstIsBetter(&t.key, &t.data.locator.root) {
doUpdate = true
}
// Or, if we are the root, possibly update our timestamp
if t.data.locator.root == t.key &&
now.Sub(t.time) > switch_updateInterval {
2018-01-04 22:37:51 +00:00
doUpdate = true
}
if doUpdate {
t.parent = switchPort(0)
t.time = now
if t.data.locator.root != t.key {
t.data.seq++
t.updater.Store(&sync.Once{})
select {
case t.core.router.reset <- struct{}{}:
default:
}
}
t.data.locator = switchLocator{root: t.key, tstamp: now.Unix()}
t.core.peers.sendSwitchMsgs()
2018-01-04 22:37:51 +00:00
}
2017-12-29 04:16:20 +00:00
}
// Removes a peer.
// Must be called by the router mainLoop goroutine, e.g. call router.doAdmin with a lambda that calls this.
// If the removed peer was this node's parent, it immediately tries to find a new parent.
func (t *switchTable) unlockedRemovePeer(port switchPort) {
delete(t.data.peers, port)
t.updater.Store(&sync.Once{})
if port != t.parent {
return
}
2018-06-07 21:49:51 +00:00
for _, info := range t.data.peers {
t.unlockedHandleMsg(&info.msg, info.port)
}
2017-12-29 04:16:20 +00:00
}
// Dropped is a list of roots that are better than the current root, but stopped sending new timestamps.
// If we switch to a new root, and that root is better than an old root that previously timed out, then we can clean up the old dropped root infos.
// This function is called periodically to do that cleanup.
2017-12-29 04:16:20 +00:00
func (t *switchTable) cleanDropped() {
// TODO? only call this after root changes, not periodically
2018-01-04 22:37:51 +00:00
for root := range t.drop {
if !firstIsBetter(&root, &t.data.locator.root) {
delete(t.drop, root)
}
}
2017-12-29 04:16:20 +00:00
}
// A switchMsg contains the root node's sig key, timestamp, and signed per-hop information about a path from the root node to some other node in the network.
// This is exchanged with peers to construct the spanning tree.
// A subset of this information, excluding the signatures, is used to construct locators that are used elsewhere in the code.
type switchMsg struct {
Root sigPubKey
TStamp int64
Hops []switchMsgHop
}
// This represents the signed information about the path leading from the root the Next node, via the Port specified here.
type switchMsgHop struct {
Port switchPort
Next sigPubKey
Sig sigBytes
}
// This returns a *switchMsg to a copy of this node's current switchMsg, which can safely have additional information appended to Hops and sent to a peer.
func (t *switchTable) getMsg() *switchMsg {
t.mutex.RLock()
defer t.mutex.RUnlock()
if t.parent == 0 {
return &switchMsg{Root: t.key, TStamp: t.data.locator.tstamp}
} else if parent, isIn := t.data.peers[t.parent]; isIn {
msg := parent.msg
msg.Hops = append([]switchMsgHop(nil), msg.Hops...)
return &msg
} else {
return nil
}
}
// This function checks that the root information in a switchMsg is OK.
// In particular, that the root is better, or else the same as the current root but with a good timestamp, and that this root+timestamp haven't been dropped due to timeout.
func (t *switchTable) checkRoot(msg *switchMsg) bool {
// returns false if it's a dropped root, not a better root, or has an older timestamp
// returns true otherwise
// used elsewhere to keep inserting peers into the dht only if root info is OK
t.mutex.RLock()
defer t.mutex.RUnlock()
dropTstamp, isIn := t.drop[msg.Root]
switch {
case isIn && dropTstamp >= msg.TStamp:
return false
case firstIsBetter(&msg.Root, &t.data.locator.root):
return true
case t.data.locator.root != msg.Root:
return false
case t.data.locator.tstamp > msg.TStamp:
return false
default:
return true
}
}
// This is a mutexed wrapper to unlockedHandleMsg, and is called by the peer structs in peers.go to pass a switchMsg for that peer into the switch.
func (t *switchTable) handleMsg(msg *switchMsg, fromPort switchPort) {
2018-01-04 22:37:51 +00:00
t.mutex.Lock()
defer t.mutex.Unlock()
2018-06-07 21:49:51 +00:00
t.unlockedHandleMsg(msg, fromPort)
}
// This updates the switch with information about a peer.
// Then the tricky part, it decides if it should update our own locator as a result.
// That happens if this node is already our parent, or is advertising a better root, or is advertising a better path to the same root, etc...
// There are a lot of very delicate order sensitive checks here, so its' best to just read the code if you need to understand what it's doing.
// It's very important to not change the order of the statements in the case function unless you're absolutely sure that it's safe, including safe if used along side nodes that used the previous order.
2018-06-07 21:49:51 +00:00
func (t *switchTable) unlockedHandleMsg(msg *switchMsg, fromPort switchPort) {
// TODO directly use a switchMsg instead of switchMessage + sigs
2018-01-04 22:37:51 +00:00
now := time.Now()
// Set up the sender peerInfo
var sender peerInfo
sender.locator.root = msg.Root
sender.locator.tstamp = msg.TStamp
prevKey := msg.Root
for _, hop := range msg.Hops {
// Build locator
sender.locator.coords = append(sender.locator.coords, hop.Port)
sender.key = prevKey
prevKey = hop.Next
}
sender.msg = *msg
2018-01-04 22:37:51 +00:00
oldSender, isIn := t.data.peers[fromPort]
if !isIn {
oldSender.firstSeen = now
}
sender.firstSeen = oldSender.firstSeen
sender.port = fromPort
sender.time = now
// Decide what to do
2018-01-04 22:37:51 +00:00
equiv := func(x *switchLocator, y *switchLocator) bool {
if x.root != y.root {
return false
}
if len(x.coords) != len(y.coords) {
return false
}
for idx := range x.coords {
if x.coords[idx] != y.coords[idx] {
return false
}
}
return true
}
doUpdate := false
if !equiv(&sender.locator, &oldSender.locator) {
2018-01-04 22:37:51 +00:00
doUpdate = true
sender.firstSeen = now
2018-01-04 22:37:51 +00:00
}
t.data.peers[fromPort] = sender
updateRoot := false
oldParent, isIn := t.data.peers[t.parent]
noParent := !isIn
noLoop := func() bool {
for idx := 0; idx < len(msg.Hops)-1; idx++ {
if msg.Hops[idx].Next == t.core.sigPub {
2018-01-04 22:37:51 +00:00
return false
}
}
if sender.locator.root == t.core.sigPub {
2018-01-04 22:37:51 +00:00
return false
}
return true
}()
sTime := now.Sub(sender.firstSeen)
pTime := oldParent.time.Sub(oldParent.firstSeen) + switch_timeout
// Really want to compare sLen/sTime and pLen/pTime
// Cross multiplied to avoid divide-by-zero
cost := len(sender.locator.coords) * int(pTime.Seconds())
2018-01-04 22:37:51 +00:00
pCost := len(t.data.locator.coords) * int(sTime.Seconds())
dropTstamp, isIn := t.drop[sender.locator.root]
2018-01-04 22:37:51 +00:00
// Here be dragons
switch {
case !noLoop: // do nothing
case isIn && dropTstamp >= sender.locator.tstamp: // do nothing
case firstIsBetter(&sender.locator.root, &t.data.locator.root):
2018-01-04 22:37:51 +00:00
updateRoot = true
case t.data.locator.root != sender.locator.root: // do nothing
case t.data.locator.tstamp > sender.locator.tstamp: // do nothing
2018-01-04 22:37:51 +00:00
case noParent:
updateRoot = true
case cost < pCost:
updateRoot = true
case sender.port != t.parent: // do nothing
case !equiv(&sender.locator, &t.data.locator):
// Special case
// If coords changed, then this may now be a worse parent than before
// Re-parent the node (de-parent and reprocess the message)
// Then reprocess *all* messages to look for a better parent
// This is so we don't keep using this node as our parent if there's something better
t.parent = 0
t.unlockedHandleMsg(msg, fromPort)
for _, info := range t.data.peers {
t.unlockedHandleMsg(&info.msg, info.port)
}
case now.Sub(t.time) < switch_throttle: // do nothing
case sender.locator.tstamp > t.data.locator.tstamp:
2018-01-04 22:37:51 +00:00
updateRoot = true
}
if updateRoot {
if !equiv(&sender.locator, &t.data.locator) {
2018-01-04 22:37:51 +00:00
doUpdate = true
t.data.seq++
select {
case t.core.router.reset <- struct{}{}:
default:
}
}
if t.data.locator.tstamp != sender.locator.tstamp {
2018-01-04 22:37:51 +00:00
t.time = now
}
t.data.locator = sender.locator
2018-01-04 22:37:51 +00:00
t.parent = sender.port
t.core.peers.sendSwitchMsgs()
2018-01-04 22:37:51 +00:00
}
if doUpdate {
t.updater.Store(&sync.Once{})
}
return
2017-12-29 04:16:20 +00:00
}
// This is called via a sync.Once to update the atomically readable subset of switch information that gets used for routing decisions.
2017-12-29 04:16:20 +00:00
func (t *switchTable) updateTable() {
2018-01-04 22:37:51 +00:00
// WARNING this should only be called from within t.data.updater.Do()
// It relies on the sync.Once for synchronization with messages and lookups
// TODO use a pre-computed faster lookup table
// Instead of checking distance for every destination every time
// Array of structs, indexed by first coord that differs from self
// Each struct has stores the best port to forward to, and a next coord map
// Move to struct, then iterate over coord maps until you dead end
// The last port before the dead end should be the closest
t.mutex.RLock()
defer t.mutex.RUnlock()
newTable := lookupTable{
self: t.data.locator.clone(),
2018-01-19 00:48:34 +00:00
elems: make([]tableElem, 0, len(t.data.peers)),
2018-01-04 22:37:51 +00:00
}
for _, pinfo := range t.data.peers {
//if !pinfo.forward { continue }
if pinfo.locator.root != newTable.self.root {
continue
}
2018-01-04 22:37:51 +00:00
loc := pinfo.locator.clone()
loc.coords = loc.coords[:len(loc.coords)-1] // Remove the them->self link
2018-01-19 00:48:34 +00:00
newTable.elems = append(newTable.elems, tableElem{
2018-01-04 22:37:51 +00:00
locator: loc,
2018-05-16 22:48:53 +00:00
port: pinfo.port,
2018-01-19 00:48:34 +00:00
})
2018-01-04 22:37:51 +00:00
}
sort.SliceStable(newTable.elems, func(i, j int) bool {
return t.data.peers[newTable.elems[i].port].firstSeen.Before(t.data.peers[newTable.elems[j].port].firstSeen)
})
2018-01-04 22:37:51 +00:00
t.table.Store(newTable)
2017-12-29 04:16:20 +00:00
}
// This does the switch layer lookups that decide how to route traffic.
// Traffic uses greedy routing in a metric space, where the metric distance between nodes is equal to the distance between them on the tree.
// Traffic must be routed to a node that is closer to the destination via the metric space distance.
// In the event that two nodes are equally close, it gets routed to the one with the longest uptime (due to the order that things are iterated over).
// The size of the outgoing packet queue is added to a node's tree distance when the cost of forwarding to a node, subject to the constraint that the real tree distance puts them closer to the destination than ourself.
// Doing so adds a limited form of backpressure routing, based on local information, which allows us to forward traffic around *local* bottlenecks, provided that another greedy path exists.
2018-06-08 01:29:22 +00:00
func (t *switchTable) lookup(dest []byte) switchPort {
2018-01-04 22:37:51 +00:00
t.updater.Load().(*sync.Once).Do(t.updateTable)
table := t.table.Load().(lookupTable)
2018-06-08 01:29:22 +00:00
myDist := table.self.dist(dest)
2018-06-08 01:18:13 +00:00
if myDist == 0 {
2018-06-08 01:29:22 +00:00
return 0
2018-01-04 22:37:51 +00:00
}
// cost is in units of (expected distance) + (expected queue size), where expected distance is used as an approximation of the minimum backpressure gradient needed for packets to flow
ports := t.core.peers.getPorts()
var best switchPort
bestCost := int64(^uint64(0) >> 1)
2018-01-19 00:48:34 +00:00
for _, info := range table.elems {
2018-06-08 01:29:22 +00:00
dist := info.locator.dist(dest)
2018-01-04 22:37:51 +00:00
if !(dist < myDist) {
continue
}
p, isIn := ports[info.port]
if !isIn {
continue
}
cost := int64(dist) + p.getQueueSize()
if cost < bestCost {
2018-01-19 00:48:34 +00:00
best = info.port
bestCost = cost
2018-01-04 22:37:51 +00:00
}
}
2018-06-08 01:29:22 +00:00
return best
2017-12-29 04:16:20 +00:00
}