1272 lines
		
	
	
		
			35 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			1272 lines
		
	
	
		
			35 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2016 The go-ethereum Authors
 | ||
| // This file is part of the go-ethereum library.
 | ||
| //
 | ||
| // The go-ethereum library is free software: you can redistribute it and/or modify
 | ||
| // it under the terms of the GNU Lesser General Public License as published by
 | ||
| // the Free Software Foundation, either version 3 of the License, or
 | ||
| // (at your option) any later version.
 | ||
| //
 | ||
| // The go-ethereum library is distributed in the hope that it will be useful,
 | ||
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | ||
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | ||
| // GNU Lesser General Public License for more details.
 | ||
| //
 | ||
| // You should have received a copy of the GNU Lesser General Public License
 | ||
| // along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
 | ||
| 
 | ||
| package discv5
 | ||
| 
 | ||
| import (
 | ||
| 	"bytes"
 | ||
| 	"crypto/ecdsa"
 | ||
| 	"errors"
 | ||
| 	"fmt"
 | ||
| 	"net"
 | ||
| 	"time"
 | ||
| 
 | ||
| 	"github.com/ethereum/go-ethereum/common"
 | ||
| 	"github.com/ethereum/go-ethereum/common/mclock"
 | ||
| 	"github.com/ethereum/go-ethereum/crypto"
 | ||
| 	"github.com/ethereum/go-ethereum/log"
 | ||
| 	"github.com/ethereum/go-ethereum/p2p/netutil"
 | ||
| 	"github.com/ethereum/go-ethereum/rlp"
 | ||
| 	"golang.org/x/crypto/sha3"
 | ||
| )
 | ||
| 
 | ||
| var (
 | ||
| 	errInvalidEvent = errors.New("invalid in current state")
 | ||
| 	errNoQuery      = errors.New("no pending query")
 | ||
| )
 | ||
| 
 | ||
| const (
 | ||
| 	autoRefreshInterval   = 1 * time.Hour
 | ||
| 	bucketRefreshInterval = 1 * time.Minute
 | ||
| 	seedCount             = 30
 | ||
| 	seedMaxAge            = 5 * 24 * time.Hour
 | ||
| 	lowPort               = 1024
 | ||
| )
 | ||
| 
 | ||
| const testTopic = "foo"
 | ||
| 
 | ||
| const (
 | ||
| 	printTestImgLogs = false
 | ||
| )
 | ||
| 
 | ||
| // Network manages the table and all protocol interaction.
 | ||
| type Network struct {
 | ||
| 	db          *nodeDB // database of known nodes
 | ||
| 	conn        transport
 | ||
| 	netrestrict *netutil.Netlist
 | ||
| 
 | ||
| 	closed           chan struct{}          // closed when loop is done
 | ||
| 	closeReq         chan struct{}          // 'request to close'
 | ||
| 	refreshReq       chan []*Node           // lookups ask for refresh on this channel
 | ||
| 	refreshResp      chan (<-chan struct{}) // ...and get the channel to block on from this one
 | ||
| 	read             chan ingressPacket     // ingress packets arrive here
 | ||
| 	timeout          chan timeoutEvent
 | ||
| 	queryReq         chan *findnodeQuery // lookups submit findnode queries on this channel
 | ||
| 	tableOpReq       chan func()
 | ||
| 	tableOpResp      chan struct{}
 | ||
| 	topicRegisterReq chan topicRegisterReq
 | ||
| 	topicSearchReq   chan topicSearchReq
 | ||
| 
 | ||
| 	// State of the main loop.
 | ||
| 	tab           *Table
 | ||
| 	topictab      *topicTable
 | ||
| 	ticketStore   *ticketStore
 | ||
| 	nursery       []*Node
 | ||
| 	nodes         map[NodeID]*Node // tracks active nodes with state != known
 | ||
| 	timeoutTimers map[timeoutEvent]*time.Timer
 | ||
| 
 | ||
| 	// Revalidation queues.
 | ||
| 	// Nodes put on these queues will be pinged eventually.
 | ||
| 	slowRevalidateQueue []*Node
 | ||
| 	fastRevalidateQueue []*Node
 | ||
| 
 | ||
| 	// Buffers for state transition.
 | ||
| 	sendBuf []*ingressPacket
 | ||
| }
 | ||
| 
 | ||
| // transport is implemented by the UDP transport.
 | ||
| // it is an interface so we can test without opening lots of UDP
 | ||
| // sockets and without generating a private key.
 | ||
| type transport interface {
 | ||
| 	sendPing(remote *Node, remoteAddr *net.UDPAddr, topics []Topic) (hash []byte)
 | ||
| 	sendNeighbours(remote *Node, nodes []*Node)
 | ||
| 	sendFindnodeHash(remote *Node, target common.Hash)
 | ||
| 	sendTopicRegister(remote *Node, topics []Topic, topicIdx int, pong []byte)
 | ||
| 	sendTopicNodes(remote *Node, queryHash common.Hash, nodes []*Node)
 | ||
| 
 | ||
| 	send(remote *Node, ptype nodeEvent, p interface{}) (hash []byte)
 | ||
| 
 | ||
| 	localAddr() *net.UDPAddr
 | ||
| 	Close()
 | ||
| }
 | ||
| 
 | ||
| type findnodeQuery struct {
 | ||
| 	remote   *Node
 | ||
| 	target   common.Hash
 | ||
| 	reply    chan<- []*Node
 | ||
| 	nresults int // counter for received nodes
 | ||
| }
 | ||
| 
 | ||
| type topicRegisterReq struct {
 | ||
| 	add   bool
 | ||
| 	topic Topic
 | ||
| }
 | ||
| 
 | ||
| type topicSearchReq struct {
 | ||
| 	topic  Topic
 | ||
| 	found  chan<- *Node
 | ||
| 	lookup chan<- bool
 | ||
| 	delay  time.Duration
 | ||
| }
 | ||
| 
 | ||
| type topicSearchResult struct {
 | ||
| 	target lookupInfo
 | ||
| 	nodes  []*Node
 | ||
| }
 | ||
| 
 | ||
| type timeoutEvent struct {
 | ||
| 	ev   nodeEvent
 | ||
| 	node *Node
 | ||
| }
 | ||
| 
 | ||
| func newNetwork(conn transport, ourPubkey ecdsa.PublicKey, dbPath string, netrestrict *netutil.Netlist) (*Network, error) {
 | ||
| 	ourID := PubkeyID(&ourPubkey)
 | ||
| 
 | ||
| 	var db *nodeDB
 | ||
| 	if dbPath != "<no database>" {
 | ||
| 		var err error
 | ||
| 		if db, err = newNodeDB(dbPath, Version, ourID); err != nil {
 | ||
| 			return nil, err
 | ||
| 		}
 | ||
| 	}
 | ||
| 
 | ||
| 	tab := newTable(ourID, conn.localAddr())
 | ||
| 	net := &Network{
 | ||
| 		db:               db,
 | ||
| 		conn:             conn,
 | ||
| 		netrestrict:      netrestrict,
 | ||
| 		tab:              tab,
 | ||
| 		topictab:         newTopicTable(db, tab.self),
 | ||
| 		ticketStore:      newTicketStore(),
 | ||
| 		refreshReq:       make(chan []*Node),
 | ||
| 		refreshResp:      make(chan (<-chan struct{})),
 | ||
| 		closed:           make(chan struct{}),
 | ||
| 		closeReq:         make(chan struct{}),
 | ||
| 		read:             make(chan ingressPacket, 100),
 | ||
| 		timeout:          make(chan timeoutEvent),
 | ||
| 		timeoutTimers:    make(map[timeoutEvent]*time.Timer),
 | ||
| 		tableOpReq:       make(chan func()),
 | ||
| 		tableOpResp:      make(chan struct{}),
 | ||
| 		queryReq:         make(chan *findnodeQuery),
 | ||
| 		topicRegisterReq: make(chan topicRegisterReq),
 | ||
| 		topicSearchReq:   make(chan topicSearchReq),
 | ||
| 		nodes:            make(map[NodeID]*Node),
 | ||
| 	}
 | ||
| 	go net.loop()
 | ||
| 	return net, nil
 | ||
| }
 | ||
| 
 | ||
| // Close terminates the network listener and flushes the node database.
 | ||
| func (net *Network) Close() {
 | ||
| 	net.conn.Close()
 | ||
| 	select {
 | ||
| 	case <-net.closed:
 | ||
| 	case net.closeReq <- struct{}{}:
 | ||
| 		<-net.closed
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| // Self returns the local node.
 | ||
| // The returned node should not be modified by the caller.
 | ||
| func (net *Network) Self() *Node {
 | ||
| 	return net.tab.self
 | ||
| }
 | ||
| 
 | ||
| // ReadRandomNodes fills the given slice with random nodes from the
 | ||
| // table. It will not write the same node more than once. The nodes in
 | ||
| // the slice are copies and can be modified by the caller.
 | ||
| func (net *Network) ReadRandomNodes(buf []*Node) (n int) {
 | ||
| 	net.reqTableOp(func() { n = net.tab.readRandomNodes(buf) })
 | ||
| 	return n
 | ||
| }
 | ||
| 
 | ||
| // SetFallbackNodes sets the initial points of contact. These nodes
 | ||
| // are used to connect to the network if the table is empty and there
 | ||
| // are no known nodes in the database.
 | ||
| func (net *Network) SetFallbackNodes(nodes []*Node) error {
 | ||
| 	nursery := make([]*Node, 0, len(nodes))
 | ||
| 	for _, n := range nodes {
 | ||
| 		if err := n.validateComplete(); err != nil {
 | ||
| 			return fmt.Errorf("bad bootstrap/fallback node %q (%v)", n, err)
 | ||
| 		}
 | ||
| 		// Recompute cpy.sha because the node might not have been
 | ||
| 		// created by NewNode or ParseNode.
 | ||
| 		cpy := *n
 | ||
| 		cpy.sha = crypto.Keccak256Hash(n.ID[:])
 | ||
| 		nursery = append(nursery, &cpy)
 | ||
| 	}
 | ||
| 	net.reqRefresh(nursery)
 | ||
| 	return nil
 | ||
| }
 | ||
| 
 | ||
| // Resolve searches for a specific node with the given ID.
 | ||
| // It returns nil if the node could not be found.
 | ||
| func (net *Network) Resolve(targetID NodeID) *Node {
 | ||
| 	result := net.lookup(crypto.Keccak256Hash(targetID[:]), true)
 | ||
| 	for _, n := range result {
 | ||
| 		if n.ID == targetID {
 | ||
| 			return n
 | ||
| 		}
 | ||
| 	}
 | ||
| 	return nil
 | ||
| }
 | ||
| 
 | ||
| // Lookup performs a network search for nodes close
 | ||
| // to the given target. It approaches the target by querying
 | ||
| // nodes that are closer to it on each iteration.
 | ||
| // The given target does not need to be an actual node
 | ||
| // identifier.
 | ||
| //
 | ||
| // The local node may be included in the result.
 | ||
| func (net *Network) Lookup(targetID NodeID) []*Node {
 | ||
| 	return net.lookup(crypto.Keccak256Hash(targetID[:]), false)
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) lookup(target common.Hash, stopOnMatch bool) []*Node {
 | ||
| 	var (
 | ||
| 		asked          = make(map[NodeID]bool)
 | ||
| 		seen           = make(map[NodeID]bool)
 | ||
| 		reply          = make(chan []*Node, alpha)
 | ||
| 		result         = nodesByDistance{target: target}
 | ||
| 		pendingQueries = 0
 | ||
| 	)
 | ||
| 	// Get initial answers from the local node.
 | ||
| 	result.push(net.tab.self, bucketSize)
 | ||
| 	for {
 | ||
| 		// Ask the α closest nodes that we haven't asked yet.
 | ||
| 		for i := 0; i < len(result.entries) && pendingQueries < alpha; i++ {
 | ||
| 			n := result.entries[i]
 | ||
| 			if !asked[n.ID] {
 | ||
| 				asked[n.ID] = true
 | ||
| 				pendingQueries++
 | ||
| 				net.reqQueryFindnode(n, target, reply)
 | ||
| 			}
 | ||
| 		}
 | ||
| 		if pendingQueries == 0 {
 | ||
| 			// We have asked all closest nodes, stop the search.
 | ||
| 			break
 | ||
| 		}
 | ||
| 		// Wait for the next reply.
 | ||
| 		select {
 | ||
| 		case nodes := <-reply:
 | ||
| 			for _, n := range nodes {
 | ||
| 				if n != nil && !seen[n.ID] {
 | ||
| 					seen[n.ID] = true
 | ||
| 					result.push(n, bucketSize)
 | ||
| 					if stopOnMatch && n.sha == target {
 | ||
| 						return result.entries
 | ||
| 					}
 | ||
| 				}
 | ||
| 			}
 | ||
| 			pendingQueries--
 | ||
| 		case <-time.After(respTimeout):
 | ||
| 			// forget all pending requests, start new ones
 | ||
| 			pendingQueries = 0
 | ||
| 			reply = make(chan []*Node, alpha)
 | ||
| 		}
 | ||
| 	}
 | ||
| 	return result.entries
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) RegisterTopic(topic Topic, stop <-chan struct{}) {
 | ||
| 	select {
 | ||
| 	case net.topicRegisterReq <- topicRegisterReq{true, topic}:
 | ||
| 	case <-net.closed:
 | ||
| 		return
 | ||
| 	}
 | ||
| 	select {
 | ||
| 	case <-net.closed:
 | ||
| 	case <-stop:
 | ||
| 		select {
 | ||
| 		case net.topicRegisterReq <- topicRegisterReq{false, topic}:
 | ||
| 		case <-net.closed:
 | ||
| 		}
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) SearchTopic(topic Topic, setPeriod <-chan time.Duration, found chan<- *Node, lookup chan<- bool) {
 | ||
| 	for {
 | ||
| 		select {
 | ||
| 		case <-net.closed:
 | ||
| 			return
 | ||
| 		case delay, ok := <-setPeriod:
 | ||
| 			select {
 | ||
| 			case net.topicSearchReq <- topicSearchReq{topic: topic, found: found, lookup: lookup, delay: delay}:
 | ||
| 			case <-net.closed:
 | ||
| 				return
 | ||
| 			}
 | ||
| 			if !ok {
 | ||
| 				return
 | ||
| 			}
 | ||
| 		}
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) reqRefresh(nursery []*Node) <-chan struct{} {
 | ||
| 	select {
 | ||
| 	case net.refreshReq <- nursery:
 | ||
| 		return <-net.refreshResp
 | ||
| 	case <-net.closed:
 | ||
| 		return net.closed
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) reqQueryFindnode(n *Node, target common.Hash, reply chan []*Node) bool {
 | ||
| 	q := &findnodeQuery{remote: n, target: target, reply: reply}
 | ||
| 	select {
 | ||
| 	case net.queryReq <- q:
 | ||
| 		return true
 | ||
| 	case <-net.closed:
 | ||
| 		return false
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) reqReadPacket(pkt ingressPacket) {
 | ||
| 	select {
 | ||
| 	case net.read <- pkt:
 | ||
| 	case <-net.closed:
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) reqTableOp(f func()) (called bool) {
 | ||
| 	select {
 | ||
| 	case net.tableOpReq <- f:
 | ||
| 		<-net.tableOpResp
 | ||
| 		return true
 | ||
| 	case <-net.closed:
 | ||
| 		return false
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| // TODO: external address handling.
 | ||
| 
 | ||
| type topicSearchInfo struct {
 | ||
| 	lookupChn chan<- bool
 | ||
| 	period    time.Duration
 | ||
| }
 | ||
| 
 | ||
| const maxSearchCount = 5
 | ||
| 
 | ||
| func (net *Network) loop() {
 | ||
| 	var (
 | ||
| 		refreshTimer       = time.NewTicker(autoRefreshInterval)
 | ||
| 		bucketRefreshTimer = time.NewTimer(bucketRefreshInterval)
 | ||
| 		refreshDone        chan struct{} // closed when the 'refresh' lookup has ended
 | ||
| 	)
 | ||
| 
 | ||
| 	// Tracking the next ticket to register.
 | ||
| 	var (
 | ||
| 		nextTicket        *ticketRef
 | ||
| 		nextRegisterTimer *time.Timer
 | ||
| 		nextRegisterTime  <-chan time.Time
 | ||
| 	)
 | ||
| 	defer func() {
 | ||
| 		if nextRegisterTimer != nil {
 | ||
| 			nextRegisterTimer.Stop()
 | ||
| 		}
 | ||
| 	}()
 | ||
| 	resetNextTicket := func() {
 | ||
| 		ticket, timeout := net.ticketStore.nextFilteredTicket()
 | ||
| 		if nextTicket != ticket {
 | ||
| 			nextTicket = ticket
 | ||
| 			if nextRegisterTimer != nil {
 | ||
| 				nextRegisterTimer.Stop()
 | ||
| 				nextRegisterTime = nil
 | ||
| 			}
 | ||
| 			if ticket != nil {
 | ||
| 				nextRegisterTimer = time.NewTimer(timeout)
 | ||
| 				nextRegisterTime = nextRegisterTimer.C
 | ||
| 			}
 | ||
| 		}
 | ||
| 	}
 | ||
| 
 | ||
| 	// Tracking registration and search lookups.
 | ||
| 	var (
 | ||
| 		topicRegisterLookupTarget lookupInfo
 | ||
| 		topicRegisterLookupDone   chan []*Node
 | ||
| 		topicRegisterLookupTick   = time.NewTimer(0)
 | ||
| 		searchReqWhenRefreshDone  []topicSearchReq
 | ||
| 		searchInfo                = make(map[Topic]topicSearchInfo)
 | ||
| 		activeSearchCount         int
 | ||
| 	)
 | ||
| 	topicSearchLookupDone := make(chan topicSearchResult, 100)
 | ||
| 	topicSearch := make(chan Topic, 100)
 | ||
| 	<-topicRegisterLookupTick.C
 | ||
| 
 | ||
| 	statsDump := time.NewTicker(10 * time.Second)
 | ||
| 
 | ||
| loop:
 | ||
| 	for {
 | ||
| 		resetNextTicket()
 | ||
| 
 | ||
| 		select {
 | ||
| 		case <-net.closeReq:
 | ||
| 			log.Trace("<-net.closeReq")
 | ||
| 			break loop
 | ||
| 
 | ||
| 		// Ingress packet handling.
 | ||
| 		case pkt := <-net.read:
 | ||
| 			//fmt.Println("read", pkt.ev)
 | ||
| 			log.Trace("<-net.read")
 | ||
| 			n := net.internNode(&pkt)
 | ||
| 			prestate := n.state
 | ||
| 			status := "ok"
 | ||
| 			if err := net.handle(n, pkt.ev, &pkt); err != nil {
 | ||
| 				status = err.Error()
 | ||
| 			}
 | ||
| 			log.Trace("", "msg", log.Lazy{Fn: func() string {
 | ||
| 				return fmt.Sprintf("<<< (%d) %v from %x@%v: %v -> %v (%v)",
 | ||
| 					net.tab.count, pkt.ev, pkt.remoteID[:8], pkt.remoteAddr, prestate, n.state, status)
 | ||
| 			}})
 | ||
| 			// TODO: persist state if n.state goes >= known, delete if it goes <= known
 | ||
| 
 | ||
| 		// State transition timeouts.
 | ||
| 		case timeout := <-net.timeout:
 | ||
| 			log.Trace("<-net.timeout")
 | ||
| 			if net.timeoutTimers[timeout] == nil {
 | ||
| 				// Stale timer (was aborted).
 | ||
| 				continue
 | ||
| 			}
 | ||
| 			delete(net.timeoutTimers, timeout)
 | ||
| 			prestate := timeout.node.state
 | ||
| 			status := "ok"
 | ||
| 			if err := net.handle(timeout.node, timeout.ev, nil); err != nil {
 | ||
| 				status = err.Error()
 | ||
| 			}
 | ||
| 			log.Trace("", "msg", log.Lazy{Fn: func() string {
 | ||
| 				return fmt.Sprintf("--- (%d) %v for %x@%v: %v -> %v (%v)",
 | ||
| 					net.tab.count, timeout.ev, timeout.node.ID[:8], timeout.node.addr(), prestate, timeout.node.state, status)
 | ||
| 			}})
 | ||
| 
 | ||
| 		// Querying.
 | ||
| 		case q := <-net.queryReq:
 | ||
| 			log.Trace("<-net.queryReq")
 | ||
| 			if !q.start(net) {
 | ||
| 				q.remote.deferQuery(q)
 | ||
| 			}
 | ||
| 
 | ||
| 		// Interacting with the table.
 | ||
| 		case f := <-net.tableOpReq:
 | ||
| 			log.Trace("<-net.tableOpReq")
 | ||
| 			f()
 | ||
| 			net.tableOpResp <- struct{}{}
 | ||
| 
 | ||
| 		// Topic registration stuff.
 | ||
| 		case req := <-net.topicRegisterReq:
 | ||
| 			log.Trace("<-net.topicRegisterReq")
 | ||
| 			if !req.add {
 | ||
| 				net.ticketStore.removeRegisterTopic(req.topic)
 | ||
| 				continue
 | ||
| 			}
 | ||
| 			net.ticketStore.addTopic(req.topic, true)
 | ||
| 			// If we're currently waiting idle (nothing to look up), give the ticket store a
 | ||
| 			// chance to start it sooner. This should speed up convergence of the radius
 | ||
| 			// determination for new topics.
 | ||
| 			// if topicRegisterLookupDone == nil {
 | ||
| 			if topicRegisterLookupTarget.target == (common.Hash{}) {
 | ||
| 				log.Trace("topicRegisterLookupTarget == null")
 | ||
| 				if topicRegisterLookupTick.Stop() {
 | ||
| 					<-topicRegisterLookupTick.C
 | ||
| 				}
 | ||
| 				target, delay := net.ticketStore.nextRegisterLookup()
 | ||
| 				topicRegisterLookupTarget = target
 | ||
| 				topicRegisterLookupTick.Reset(delay)
 | ||
| 			}
 | ||
| 
 | ||
| 		case nodes := <-topicRegisterLookupDone:
 | ||
| 			log.Trace("<-topicRegisterLookupDone")
 | ||
| 			net.ticketStore.registerLookupDone(topicRegisterLookupTarget, nodes, func(n *Node) []byte {
 | ||
| 				net.ping(n, n.addr())
 | ||
| 				return n.pingEcho
 | ||
| 			})
 | ||
| 			target, delay := net.ticketStore.nextRegisterLookup()
 | ||
| 			topicRegisterLookupTarget = target
 | ||
| 			topicRegisterLookupTick.Reset(delay)
 | ||
| 			topicRegisterLookupDone = nil
 | ||
| 
 | ||
| 		case <-topicRegisterLookupTick.C:
 | ||
| 			log.Trace("<-topicRegisterLookupTick")
 | ||
| 			if (topicRegisterLookupTarget.target == common.Hash{}) {
 | ||
| 				target, delay := net.ticketStore.nextRegisterLookup()
 | ||
| 				topicRegisterLookupTarget = target
 | ||
| 				topicRegisterLookupTick.Reset(delay)
 | ||
| 				topicRegisterLookupDone = nil
 | ||
| 			} else {
 | ||
| 				topicRegisterLookupDone = make(chan []*Node)
 | ||
| 				target := topicRegisterLookupTarget.target
 | ||
| 				go func() { topicRegisterLookupDone <- net.lookup(target, false) }()
 | ||
| 			}
 | ||
| 
 | ||
| 		case <-nextRegisterTime:
 | ||
| 			log.Trace("<-nextRegisterTime")
 | ||
| 			net.ticketStore.ticketRegistered(*nextTicket)
 | ||
| 			//fmt.Println("sendTopicRegister", nextTicket.t.node.addr().String(), nextTicket.t.topics, nextTicket.idx, nextTicket.t.pong)
 | ||
| 			net.conn.sendTopicRegister(nextTicket.t.node, nextTicket.t.topics, nextTicket.idx, nextTicket.t.pong)
 | ||
| 
 | ||
| 		case req := <-net.topicSearchReq:
 | ||
| 			if refreshDone == nil {
 | ||
| 				log.Trace("<-net.topicSearchReq")
 | ||
| 				info, ok := searchInfo[req.topic]
 | ||
| 				if ok {
 | ||
| 					if req.delay == time.Duration(0) {
 | ||
| 						delete(searchInfo, req.topic)
 | ||
| 						net.ticketStore.removeSearchTopic(req.topic)
 | ||
| 					} else {
 | ||
| 						info.period = req.delay
 | ||
| 						searchInfo[req.topic] = info
 | ||
| 					}
 | ||
| 					continue
 | ||
| 				}
 | ||
| 				if req.delay != time.Duration(0) {
 | ||
| 					var info topicSearchInfo
 | ||
| 					info.period = req.delay
 | ||
| 					info.lookupChn = req.lookup
 | ||
| 					searchInfo[req.topic] = info
 | ||
| 					net.ticketStore.addSearchTopic(req.topic, req.found)
 | ||
| 					topicSearch <- req.topic
 | ||
| 				}
 | ||
| 			} else {
 | ||
| 				searchReqWhenRefreshDone = append(searchReqWhenRefreshDone, req)
 | ||
| 			}
 | ||
| 
 | ||
| 		case topic := <-topicSearch:
 | ||
| 			if activeSearchCount < maxSearchCount {
 | ||
| 				activeSearchCount++
 | ||
| 				target := net.ticketStore.nextSearchLookup(topic)
 | ||
| 				go func() {
 | ||
| 					nodes := net.lookup(target.target, false)
 | ||
| 					topicSearchLookupDone <- topicSearchResult{target: target, nodes: nodes}
 | ||
| 				}()
 | ||
| 			}
 | ||
| 			period := searchInfo[topic].period
 | ||
| 			if period != time.Duration(0) {
 | ||
| 				go func() {
 | ||
| 					time.Sleep(period)
 | ||
| 					topicSearch <- topic
 | ||
| 				}()
 | ||
| 			}
 | ||
| 
 | ||
| 		case res := <-topicSearchLookupDone:
 | ||
| 			activeSearchCount--
 | ||
| 			if lookupChn := searchInfo[res.target.topic].lookupChn; lookupChn != nil {
 | ||
| 				lookupChn <- net.ticketStore.radius[res.target.topic].converged
 | ||
| 			}
 | ||
| 			net.ticketStore.searchLookupDone(res.target, res.nodes, func(n *Node, topic Topic) []byte {
 | ||
| 				if n.state != nil && n.state.canQuery {
 | ||
| 					return net.conn.send(n, topicQueryPacket, topicQuery{Topic: topic}) // TODO: set expiration
 | ||
| 				}
 | ||
| 				if n.state == unknown {
 | ||
| 					net.ping(n, n.addr())
 | ||
| 				}
 | ||
| 				return nil
 | ||
| 			})
 | ||
| 
 | ||
| 		case <-statsDump.C:
 | ||
| 			log.Trace("<-statsDump.C")
 | ||
| 			/*r, ok := net.ticketStore.radius[testTopic]
 | ||
| 			if !ok {
 | ||
| 				fmt.Printf("(%x) no radius @ %v\n", net.tab.self.ID[:8], time.Now())
 | ||
| 			} else {
 | ||
| 				topics := len(net.ticketStore.tickets)
 | ||
| 				tickets := len(net.ticketStore.nodes)
 | ||
| 				rad := r.radius / (maxRadius/10000+1)
 | ||
| 				fmt.Printf("(%x) topics:%d radius:%d tickets:%d @ %v\n", net.tab.self.ID[:8], topics, rad, tickets, time.Now())
 | ||
| 			}*/
 | ||
| 
 | ||
| 			tm := mclock.Now()
 | ||
| 			for topic, r := range net.ticketStore.radius {
 | ||
| 				if printTestImgLogs {
 | ||
| 					rad := r.radius / (maxRadius/1000000 + 1)
 | ||
| 					minrad := r.minRadius / (maxRadius/1000000 + 1)
 | ||
| 					fmt.Printf("*R %d %v %016x %v\n", tm/1000000, topic, net.tab.self.sha[:8], rad)
 | ||
| 					fmt.Printf("*MR %d %v %016x %v\n", tm/1000000, topic, net.tab.self.sha[:8], minrad)
 | ||
| 				}
 | ||
| 			}
 | ||
| 			for topic, t := range net.topictab.topics {
 | ||
| 				wp := t.wcl.nextWaitPeriod(tm)
 | ||
| 				if printTestImgLogs {
 | ||
| 					fmt.Printf("*W %d %v %016x %d\n", tm/1000000, topic, net.tab.self.sha[:8], wp/1000000)
 | ||
| 				}
 | ||
| 			}
 | ||
| 
 | ||
| 		// Periodic / lookup-initiated bucket refresh.
 | ||
| 		case <-refreshTimer.C:
 | ||
| 			log.Trace("<-refreshTimer.C")
 | ||
| 			// TODO: ideally we would start the refresh timer after
 | ||
| 			// fallback nodes have been set for the first time.
 | ||
| 			if refreshDone == nil {
 | ||
| 				refreshDone = make(chan struct{})
 | ||
| 				net.refresh(refreshDone)
 | ||
| 			}
 | ||
| 		case <-bucketRefreshTimer.C:
 | ||
| 			target := net.tab.chooseBucketRefreshTarget()
 | ||
| 			go func() {
 | ||
| 				net.lookup(target, false)
 | ||
| 				bucketRefreshTimer.Reset(bucketRefreshInterval)
 | ||
| 			}()
 | ||
| 		case newNursery := <-net.refreshReq:
 | ||
| 			log.Trace("<-net.refreshReq")
 | ||
| 			if newNursery != nil {
 | ||
| 				net.nursery = newNursery
 | ||
| 			}
 | ||
| 			if refreshDone == nil {
 | ||
| 				refreshDone = make(chan struct{})
 | ||
| 				net.refresh(refreshDone)
 | ||
| 			}
 | ||
| 			net.refreshResp <- refreshDone
 | ||
| 		case <-refreshDone:
 | ||
| 			log.Trace("<-net.refreshDone", "table size", net.tab.count)
 | ||
| 			if net.tab.count != 0 {
 | ||
| 				refreshDone = nil
 | ||
| 				list := searchReqWhenRefreshDone
 | ||
| 				searchReqWhenRefreshDone = nil
 | ||
| 				go func() {
 | ||
| 					for _, req := range list {
 | ||
| 						net.topicSearchReq <- req
 | ||
| 					}
 | ||
| 				}()
 | ||
| 			} else {
 | ||
| 				refreshDone = make(chan struct{})
 | ||
| 				net.refresh(refreshDone)
 | ||
| 			}
 | ||
| 		}
 | ||
| 	}
 | ||
| 	log.Trace("loop stopped")
 | ||
| 
 | ||
| 	log.Debug(fmt.Sprintf("shutting down"))
 | ||
| 	if net.conn != nil {
 | ||
| 		net.conn.Close()
 | ||
| 	}
 | ||
| 	if refreshDone != nil {
 | ||
| 		// TODO: wait for pending refresh.
 | ||
| 		//<-refreshResults
 | ||
| 	}
 | ||
| 	// Cancel all pending timeouts.
 | ||
| 	for _, timer := range net.timeoutTimers {
 | ||
| 		timer.Stop()
 | ||
| 	}
 | ||
| 	if net.db != nil {
 | ||
| 		net.db.close()
 | ||
| 	}
 | ||
| 	close(net.closed)
 | ||
| }
 | ||
| 
 | ||
| // Everything below runs on the Network.loop goroutine
 | ||
| // and can modify Node, Table and Network at any time without locking.
 | ||
| 
 | ||
| func (net *Network) refresh(done chan<- struct{}) {
 | ||
| 	var seeds []*Node
 | ||
| 	if net.db != nil {
 | ||
| 		seeds = net.db.querySeeds(seedCount, seedMaxAge)
 | ||
| 	}
 | ||
| 	if len(seeds) == 0 {
 | ||
| 		seeds = net.nursery
 | ||
| 	}
 | ||
| 	if len(seeds) == 0 {
 | ||
| 		log.Trace("no seed nodes found")
 | ||
| 		time.AfterFunc(time.Second*10, func() { close(done) })
 | ||
| 		return
 | ||
| 	}
 | ||
| 	for _, n := range seeds {
 | ||
| 		log.Debug("", "msg", log.Lazy{Fn: func() string {
 | ||
| 			var age string
 | ||
| 			if net.db != nil {
 | ||
| 				age = time.Since(net.db.lastPong(n.ID)).String()
 | ||
| 			} else {
 | ||
| 				age = "unknown"
 | ||
| 			}
 | ||
| 			return fmt.Sprintf("seed node (age %s): %v", age, n)
 | ||
| 		}})
 | ||
| 		n = net.internNodeFromDB(n)
 | ||
| 		if n.state == unknown {
 | ||
| 			net.transition(n, verifyinit)
 | ||
| 		}
 | ||
| 		// Force-add the seed node so Lookup does something.
 | ||
| 		// It will be deleted again if verification fails.
 | ||
| 		net.tab.add(n)
 | ||
| 	}
 | ||
| 	// Start self lookup to fill up the buckets.
 | ||
| 	go func() {
 | ||
| 		net.Lookup(net.tab.self.ID)
 | ||
| 		close(done)
 | ||
| 	}()
 | ||
| }
 | ||
| 
 | ||
| // Node Interning.
 | ||
| 
 | ||
| func (net *Network) internNode(pkt *ingressPacket) *Node {
 | ||
| 	if n := net.nodes[pkt.remoteID]; n != nil {
 | ||
| 		n.IP = pkt.remoteAddr.IP
 | ||
| 		n.UDP = uint16(pkt.remoteAddr.Port)
 | ||
| 		n.TCP = uint16(pkt.remoteAddr.Port)
 | ||
| 		return n
 | ||
| 	}
 | ||
| 	n := NewNode(pkt.remoteID, pkt.remoteAddr.IP, uint16(pkt.remoteAddr.Port), uint16(pkt.remoteAddr.Port))
 | ||
| 	n.state = unknown
 | ||
| 	net.nodes[pkt.remoteID] = n
 | ||
| 	return n
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) internNodeFromDB(dbn *Node) *Node {
 | ||
| 	if n := net.nodes[dbn.ID]; n != nil {
 | ||
| 		return n
 | ||
| 	}
 | ||
| 	n := NewNode(dbn.ID, dbn.IP, dbn.UDP, dbn.TCP)
 | ||
| 	n.state = unknown
 | ||
| 	net.nodes[n.ID] = n
 | ||
| 	return n
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) internNodeFromNeighbours(sender *net.UDPAddr, rn rpcNode) (n *Node, err error) {
 | ||
| 	if rn.ID == net.tab.self.ID {
 | ||
| 		return nil, errors.New("is self")
 | ||
| 	}
 | ||
| 	if rn.UDP <= lowPort {
 | ||
| 		return nil, errors.New("low port")
 | ||
| 	}
 | ||
| 	n = net.nodes[rn.ID]
 | ||
| 	if n == nil {
 | ||
| 		// We haven't seen this node before.
 | ||
| 		n, err = nodeFromRPC(sender, rn)
 | ||
| 		if net.netrestrict != nil && !net.netrestrict.Contains(n.IP) {
 | ||
| 			return n, errors.New("not contained in netrestrict whitelist")
 | ||
| 		}
 | ||
| 		if err == nil {
 | ||
| 			n.state = unknown
 | ||
| 			net.nodes[n.ID] = n
 | ||
| 		}
 | ||
| 		return n, err
 | ||
| 	}
 | ||
| 	if !n.IP.Equal(rn.IP) || n.UDP != rn.UDP || n.TCP != rn.TCP {
 | ||
| 		if n.state == known {
 | ||
| 			// reject address change if node is known by us
 | ||
| 			err = fmt.Errorf("metadata mismatch: got %v, want %v", rn, n)
 | ||
| 		} else {
 | ||
| 			// accept otherwise; this will be handled nicer with signed ENRs
 | ||
| 			n.IP = rn.IP
 | ||
| 			n.UDP = rn.UDP
 | ||
| 			n.TCP = rn.TCP
 | ||
| 		}
 | ||
| 	}
 | ||
| 	return n, err
 | ||
| }
 | ||
| 
 | ||
| // nodeNetGuts is embedded in Node and contains fields.
 | ||
| type nodeNetGuts struct {
 | ||
| 	// This is a cached copy of sha3(ID) which is used for node
 | ||
| 	// distance calculations. This is part of Node in order to make it
 | ||
| 	// possible to write tests that need a node at a certain distance.
 | ||
| 	// In those tests, the content of sha will not actually correspond
 | ||
| 	// with ID.
 | ||
| 	sha common.Hash
 | ||
| 
 | ||
| 	// State machine fields. Access to these fields
 | ||
| 	// is restricted to the Network.loop goroutine.
 | ||
| 	state             *nodeState
 | ||
| 	pingEcho          []byte           // hash of last ping sent by us
 | ||
| 	pingTopics        []Topic          // topic set sent by us in last ping
 | ||
| 	deferredQueries   []*findnodeQuery // queries that can't be sent yet
 | ||
| 	pendingNeighbours *findnodeQuery   // current query, waiting for reply
 | ||
| 	queryTimeouts     int
 | ||
| }
 | ||
| 
 | ||
| func (n *nodeNetGuts) deferQuery(q *findnodeQuery) {
 | ||
| 	n.deferredQueries = append(n.deferredQueries, q)
 | ||
| }
 | ||
| 
 | ||
| func (n *nodeNetGuts) startNextQuery(net *Network) {
 | ||
| 	if len(n.deferredQueries) == 0 {
 | ||
| 		return
 | ||
| 	}
 | ||
| 	nextq := n.deferredQueries[0]
 | ||
| 	if nextq.start(net) {
 | ||
| 		n.deferredQueries = append(n.deferredQueries[:0], n.deferredQueries[1:]...)
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (q *findnodeQuery) start(net *Network) bool {
 | ||
| 	// Satisfy queries against the local node directly.
 | ||
| 	if q.remote == net.tab.self {
 | ||
| 		closest := net.tab.closest(q.target, bucketSize)
 | ||
| 		q.reply <- closest.entries
 | ||
| 		return true
 | ||
| 	}
 | ||
| 	if q.remote.state.canQuery && q.remote.pendingNeighbours == nil {
 | ||
| 		net.conn.sendFindnodeHash(q.remote, q.target)
 | ||
| 		net.timedEvent(respTimeout, q.remote, neighboursTimeout)
 | ||
| 		q.remote.pendingNeighbours = q
 | ||
| 		return true
 | ||
| 	}
 | ||
| 	// If the node is not known yet, it won't accept queries.
 | ||
| 	// Initiate the transition to known.
 | ||
| 	// The request will be sent later when the node reaches known state.
 | ||
| 	if q.remote.state == unknown {
 | ||
| 		net.transition(q.remote, verifyinit)
 | ||
| 	}
 | ||
| 	return false
 | ||
| }
 | ||
| 
 | ||
| // Node Events (the input to the state machine).
 | ||
| 
 | ||
| type nodeEvent uint
 | ||
| 
 | ||
| //go:generate stringer -type=nodeEvent
 | ||
| 
 | ||
| const (
 | ||
| 
 | ||
| 	// Packet type events.
 | ||
| 	// These correspond to packet types in the UDP protocol.
 | ||
| 	pingPacket = iota + 1
 | ||
| 	pongPacket
 | ||
| 	findnodePacket
 | ||
| 	neighborsPacket
 | ||
| 	findnodeHashPacket
 | ||
| 	topicRegisterPacket
 | ||
| 	topicQueryPacket
 | ||
| 	topicNodesPacket
 | ||
| 
 | ||
| 	// Non-packet events.
 | ||
| 	// Event values in this category are allocated outside
 | ||
| 	// the packet type range (packet types are encoded as a single byte).
 | ||
| 	pongTimeout nodeEvent = iota + 256
 | ||
| 	pingTimeout
 | ||
| 	neighboursTimeout
 | ||
| )
 | ||
| 
 | ||
| // Node State Machine.
 | ||
| 
 | ||
| type nodeState struct {
 | ||
| 	name     string
 | ||
| 	handle   func(*Network, *Node, nodeEvent, *ingressPacket) (next *nodeState, err error)
 | ||
| 	enter    func(*Network, *Node)
 | ||
| 	canQuery bool
 | ||
| }
 | ||
| 
 | ||
| func (s *nodeState) String() string {
 | ||
| 	return s.name
 | ||
| }
 | ||
| 
 | ||
| var (
 | ||
| 	unknown          *nodeState
 | ||
| 	verifyinit       *nodeState
 | ||
| 	verifywait       *nodeState
 | ||
| 	remoteverifywait *nodeState
 | ||
| 	known            *nodeState
 | ||
| 	contested        *nodeState
 | ||
| 	unresponsive     *nodeState
 | ||
| )
 | ||
| 
 | ||
| func init() {
 | ||
| 	unknown = &nodeState{
 | ||
| 		name: "unknown",
 | ||
| 		enter: func(net *Network, n *Node) {
 | ||
| 			net.tab.delete(n)
 | ||
| 			n.pingEcho = nil
 | ||
| 			// Abort active queries.
 | ||
| 			for _, q := range n.deferredQueries {
 | ||
| 				q.reply <- nil
 | ||
| 			}
 | ||
| 			n.deferredQueries = nil
 | ||
| 			if n.pendingNeighbours != nil {
 | ||
| 				n.pendingNeighbours.reply <- nil
 | ||
| 				n.pendingNeighbours = nil
 | ||
| 			}
 | ||
| 			n.queryTimeouts = 0
 | ||
| 		},
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				net.ping(n, pkt.remoteAddr)
 | ||
| 				return verifywait, nil
 | ||
| 			default:
 | ||
| 				return unknown, errInvalidEvent
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	verifyinit = &nodeState{
 | ||
| 		name: "verifyinit",
 | ||
| 		enter: func(net *Network, n *Node) {
 | ||
| 			net.ping(n, n.addr())
 | ||
| 		},
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return verifywait, nil
 | ||
| 			case pongPacket:
 | ||
| 				err := net.handleKnownPong(n, pkt)
 | ||
| 				return remoteverifywait, err
 | ||
| 			case pongTimeout:
 | ||
| 				return unknown, nil
 | ||
| 			default:
 | ||
| 				return verifyinit, errInvalidEvent
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	verifywait = &nodeState{
 | ||
| 		name: "verifywait",
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return verifywait, nil
 | ||
| 			case pongPacket:
 | ||
| 				err := net.handleKnownPong(n, pkt)
 | ||
| 				return known, err
 | ||
| 			case pongTimeout:
 | ||
| 				return unknown, nil
 | ||
| 			default:
 | ||
| 				return verifywait, errInvalidEvent
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	remoteverifywait = &nodeState{
 | ||
| 		name: "remoteverifywait",
 | ||
| 		enter: func(net *Network, n *Node) {
 | ||
| 			net.timedEvent(respTimeout, n, pingTimeout)
 | ||
| 		},
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return remoteverifywait, nil
 | ||
| 			case pingTimeout:
 | ||
| 				return known, nil
 | ||
| 			default:
 | ||
| 				return remoteverifywait, errInvalidEvent
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	known = &nodeState{
 | ||
| 		name:     "known",
 | ||
| 		canQuery: true,
 | ||
| 		enter: func(net *Network, n *Node) {
 | ||
| 			n.queryTimeouts = 0
 | ||
| 			n.startNextQuery(net)
 | ||
| 			// Insert into the table and start revalidation of the last node
 | ||
| 			// in the bucket if it is full.
 | ||
| 			last := net.tab.add(n)
 | ||
| 			if last != nil && last.state == known {
 | ||
| 				// TODO: do this asynchronously
 | ||
| 				net.transition(last, contested)
 | ||
| 			}
 | ||
| 		},
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return known, nil
 | ||
| 			case pongPacket:
 | ||
| 				err := net.handleKnownPong(n, pkt)
 | ||
| 				return known, err
 | ||
| 			default:
 | ||
| 				return net.handleQueryEvent(n, ev, pkt)
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	contested = &nodeState{
 | ||
| 		name:     "contested",
 | ||
| 		canQuery: true,
 | ||
| 		enter: func(net *Network, n *Node) {
 | ||
| 			net.ping(n, n.addr())
 | ||
| 		},
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pongPacket:
 | ||
| 				// Node is still alive.
 | ||
| 				err := net.handleKnownPong(n, pkt)
 | ||
| 				return known, err
 | ||
| 			case pongTimeout:
 | ||
| 				net.tab.deleteReplace(n)
 | ||
| 				return unresponsive, nil
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return contested, nil
 | ||
| 			default:
 | ||
| 				return net.handleQueryEvent(n, ev, pkt)
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| 
 | ||
| 	unresponsive = &nodeState{
 | ||
| 		name:     "unresponsive",
 | ||
| 		canQuery: true,
 | ||
| 		handle: func(net *Network, n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 			switch ev {
 | ||
| 			case pingPacket:
 | ||
| 				net.handlePing(n, pkt)
 | ||
| 				return known, nil
 | ||
| 			case pongPacket:
 | ||
| 				err := net.handleKnownPong(n, pkt)
 | ||
| 				return known, err
 | ||
| 			default:
 | ||
| 				return net.handleQueryEvent(n, ev, pkt)
 | ||
| 			}
 | ||
| 		},
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| // handle processes packets sent by n and events related to n.
 | ||
| func (net *Network) handle(n *Node, ev nodeEvent, pkt *ingressPacket) error {
 | ||
| 	//fmt.Println("handle", n.addr().String(), n.state, ev)
 | ||
| 	if pkt != nil {
 | ||
| 		if err := net.checkPacket(n, ev, pkt); err != nil {
 | ||
| 			//fmt.Println("check err:", err)
 | ||
| 			return err
 | ||
| 		}
 | ||
| 		// Start the background expiration goroutine after the first
 | ||
| 		// successful communication. Subsequent calls have no effect if it
 | ||
| 		// is already running. We do this here instead of somewhere else
 | ||
| 		// so that the search for seed nodes also considers older nodes
 | ||
| 		// that would otherwise be removed by the expirer.
 | ||
| 		if net.db != nil {
 | ||
| 			net.db.ensureExpirer()
 | ||
| 		}
 | ||
| 	}
 | ||
| 	if n.state == nil {
 | ||
| 		n.state = unknown //???
 | ||
| 	}
 | ||
| 	next, err := n.state.handle(net, n, ev, pkt)
 | ||
| 	net.transition(n, next)
 | ||
| 	//fmt.Println("new state:", n.state)
 | ||
| 	return err
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) checkPacket(n *Node, ev nodeEvent, pkt *ingressPacket) error {
 | ||
| 	// Replay prevention checks.
 | ||
| 	switch ev {
 | ||
| 	case pingPacket, findnodeHashPacket, neighborsPacket:
 | ||
| 		// TODO: check date is > last date seen
 | ||
| 		// TODO: check ping version
 | ||
| 	case pongPacket:
 | ||
| 		if !bytes.Equal(pkt.data.(*pong).ReplyTok, n.pingEcho) {
 | ||
| 			// fmt.Println("pong reply token mismatch")
 | ||
| 			return fmt.Errorf("pong reply token mismatch")
 | ||
| 		}
 | ||
| 		n.pingEcho = nil
 | ||
| 	}
 | ||
| 	// Address validation.
 | ||
| 	// TODO: Ideally we would do the following:
 | ||
| 	//  - reject all packets with wrong address except ping.
 | ||
| 	//  - for ping with new address, transition to verifywait but keep the
 | ||
| 	//    previous node (with old address) around. if the new one reaches known,
 | ||
| 	//    swap it out.
 | ||
| 	return nil
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) transition(n *Node, next *nodeState) {
 | ||
| 	if n.state != next {
 | ||
| 		n.state = next
 | ||
| 		if next.enter != nil {
 | ||
| 			next.enter(net, n)
 | ||
| 		}
 | ||
| 	}
 | ||
| 
 | ||
| 	// TODO: persist/unpersist node
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) timedEvent(d time.Duration, n *Node, ev nodeEvent) {
 | ||
| 	timeout := timeoutEvent{ev, n}
 | ||
| 	net.timeoutTimers[timeout] = time.AfterFunc(d, func() {
 | ||
| 		select {
 | ||
| 		case net.timeout <- timeout:
 | ||
| 		case <-net.closed:
 | ||
| 		}
 | ||
| 	})
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) abortTimedEvent(n *Node, ev nodeEvent) {
 | ||
| 	timer := net.timeoutTimers[timeoutEvent{ev, n}]
 | ||
| 	if timer != nil {
 | ||
| 		timer.Stop()
 | ||
| 		delete(net.timeoutTimers, timeoutEvent{ev, n})
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) ping(n *Node, addr *net.UDPAddr) {
 | ||
| 	//fmt.Println("ping", n.addr().String(), n.ID.String(), n.sha.Hex())
 | ||
| 	if n.pingEcho != nil || n.ID == net.tab.self.ID {
 | ||
| 		//fmt.Println(" not sent")
 | ||
| 		return
 | ||
| 	}
 | ||
| 	log.Trace("Pinging remote node", "node", n.ID)
 | ||
| 	n.pingTopics = net.ticketStore.regTopicSet()
 | ||
| 	n.pingEcho = net.conn.sendPing(n, addr, n.pingTopics)
 | ||
| 	net.timedEvent(respTimeout, n, pongTimeout)
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) handlePing(n *Node, pkt *ingressPacket) {
 | ||
| 	log.Trace("Handling remote ping", "node", n.ID)
 | ||
| 	ping := pkt.data.(*ping)
 | ||
| 	n.TCP = ping.From.TCP
 | ||
| 	t := net.topictab.getTicket(n, ping.Topics)
 | ||
| 
 | ||
| 	pong := &pong{
 | ||
| 		To:         makeEndpoint(n.addr(), n.TCP), // TODO: maybe use known TCP port from DB
 | ||
| 		ReplyTok:   pkt.hash,
 | ||
| 		Expiration: uint64(time.Now().Add(expiration).Unix()),
 | ||
| 	}
 | ||
| 	ticketToPong(t, pong)
 | ||
| 	net.conn.send(n, pongPacket, pong)
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) handleKnownPong(n *Node, pkt *ingressPacket) error {
 | ||
| 	log.Trace("Handling known pong", "node", n.ID)
 | ||
| 	net.abortTimedEvent(n, pongTimeout)
 | ||
| 	now := mclock.Now()
 | ||
| 	ticket, err := pongToTicket(now, n.pingTopics, n, pkt)
 | ||
| 	if err == nil {
 | ||
| 		// fmt.Printf("(%x) ticket: %+v\n", net.tab.self.ID[:8], pkt.data)
 | ||
| 		net.ticketStore.addTicket(now, pkt.data.(*pong).ReplyTok, ticket)
 | ||
| 	} else {
 | ||
| 		log.Trace("Failed to convert pong to ticket", "err", err)
 | ||
| 	}
 | ||
| 	n.pingEcho = nil
 | ||
| 	n.pingTopics = nil
 | ||
| 	return err
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) handleQueryEvent(n *Node, ev nodeEvent, pkt *ingressPacket) (*nodeState, error) {
 | ||
| 	switch ev {
 | ||
| 	case findnodePacket:
 | ||
| 		target := crypto.Keccak256Hash(pkt.data.(*findnode).Target[:])
 | ||
| 		results := net.tab.closest(target, bucketSize).entries
 | ||
| 		net.conn.sendNeighbours(n, results)
 | ||
| 		return n.state, nil
 | ||
| 	case neighborsPacket:
 | ||
| 		err := net.handleNeighboursPacket(n, pkt)
 | ||
| 		return n.state, err
 | ||
| 	case neighboursTimeout:
 | ||
| 		if n.pendingNeighbours != nil {
 | ||
| 			n.pendingNeighbours.reply <- nil
 | ||
| 			n.pendingNeighbours = nil
 | ||
| 		}
 | ||
| 		n.queryTimeouts++
 | ||
| 		if n.queryTimeouts > maxFindnodeFailures && n.state == known {
 | ||
| 			return contested, errors.New("too many timeouts")
 | ||
| 		}
 | ||
| 		return n.state, nil
 | ||
| 
 | ||
| 	// v5
 | ||
| 
 | ||
| 	case findnodeHashPacket:
 | ||
| 		results := net.tab.closest(pkt.data.(*findnodeHash).Target, bucketSize).entries
 | ||
| 		net.conn.sendNeighbours(n, results)
 | ||
| 		return n.state, nil
 | ||
| 	case topicRegisterPacket:
 | ||
| 		//fmt.Println("got topicRegisterPacket")
 | ||
| 		regdata := pkt.data.(*topicRegister)
 | ||
| 		pong, err := net.checkTopicRegister(regdata)
 | ||
| 		if err != nil {
 | ||
| 			//fmt.Println(err)
 | ||
| 			return n.state, fmt.Errorf("bad waiting ticket: %v", err)
 | ||
| 		}
 | ||
| 		net.topictab.useTicket(n, pong.TicketSerial, regdata.Topics, int(regdata.Idx), pong.Expiration, pong.WaitPeriods)
 | ||
| 		return n.state, nil
 | ||
| 	case topicQueryPacket:
 | ||
| 		// TODO: handle expiration
 | ||
| 		topic := pkt.data.(*topicQuery).Topic
 | ||
| 		results := net.topictab.getEntries(topic)
 | ||
| 		if _, ok := net.ticketStore.tickets[topic]; ok {
 | ||
| 			results = append(results, net.tab.self) // we're not registering in our own table but if we're advertising, return ourselves too
 | ||
| 		}
 | ||
| 		if len(results) > 10 {
 | ||
| 			results = results[:10]
 | ||
| 		}
 | ||
| 		var hash common.Hash
 | ||
| 		copy(hash[:], pkt.hash)
 | ||
| 		net.conn.sendTopicNodes(n, hash, results)
 | ||
| 		return n.state, nil
 | ||
| 	case topicNodesPacket:
 | ||
| 		p := pkt.data.(*topicNodes)
 | ||
| 		if net.ticketStore.gotTopicNodes(n, p.Echo, p.Nodes) {
 | ||
| 			n.queryTimeouts++
 | ||
| 			if n.queryTimeouts > maxFindnodeFailures && n.state == known {
 | ||
| 				return contested, errors.New("too many timeouts")
 | ||
| 			}
 | ||
| 		}
 | ||
| 		return n.state, nil
 | ||
| 
 | ||
| 	default:
 | ||
| 		return n.state, errInvalidEvent
 | ||
| 	}
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) checkTopicRegister(data *topicRegister) (*pong, error) {
 | ||
| 	var pongpkt ingressPacket
 | ||
| 	if err := decodePacket(data.Pong, &pongpkt); err != nil {
 | ||
| 		return nil, err
 | ||
| 	}
 | ||
| 	if pongpkt.ev != pongPacket {
 | ||
| 		return nil, errors.New("is not pong packet")
 | ||
| 	}
 | ||
| 	if pongpkt.remoteID != net.tab.self.ID {
 | ||
| 		return nil, errors.New("not signed by us")
 | ||
| 	}
 | ||
| 	// check that we previously authorised all topics
 | ||
| 	// that the other side is trying to register.
 | ||
| 	if rlpHash(data.Topics) != pongpkt.data.(*pong).TopicHash {
 | ||
| 		return nil, errors.New("topic hash mismatch")
 | ||
| 	}
 | ||
| 	if data.Idx >= uint(len(data.Topics)) {
 | ||
| 		return nil, errors.New("topic index out of range")
 | ||
| 	}
 | ||
| 	return pongpkt.data.(*pong), nil
 | ||
| }
 | ||
| 
 | ||
| func rlpHash(x interface{}) (h common.Hash) {
 | ||
| 	hw := sha3.NewLegacyKeccak256()
 | ||
| 	rlp.Encode(hw, x)
 | ||
| 	hw.Sum(h[:0])
 | ||
| 	return h
 | ||
| }
 | ||
| 
 | ||
| func (net *Network) handleNeighboursPacket(n *Node, pkt *ingressPacket) error {
 | ||
| 	if n.pendingNeighbours == nil {
 | ||
| 		return errNoQuery
 | ||
| 	}
 | ||
| 	net.abortTimedEvent(n, neighboursTimeout)
 | ||
| 
 | ||
| 	req := pkt.data.(*neighbors)
 | ||
| 	nodes := make([]*Node, len(req.Nodes))
 | ||
| 	for i, rn := range req.Nodes {
 | ||
| 		nn, err := net.internNodeFromNeighbours(pkt.remoteAddr, rn)
 | ||
| 		if err != nil {
 | ||
| 			log.Debug(fmt.Sprintf("invalid neighbour (%v) from %x@%v: %v", rn.IP, n.ID[:8], pkt.remoteAddr, err))
 | ||
| 			continue
 | ||
| 		}
 | ||
| 		nodes[i] = nn
 | ||
| 		// Start validation of query results immediately.
 | ||
| 		// This fills the table quickly.
 | ||
| 		// TODO: generates way too many packets, maybe do it via queue.
 | ||
| 		if nn.state == unknown {
 | ||
| 			net.transition(nn, verifyinit)
 | ||
| 		}
 | ||
| 	}
 | ||
| 	// TODO: don't ignore second packet
 | ||
| 	n.pendingNeighbours.reply <- nodes
 | ||
| 	n.pendingNeighbours = nil
 | ||
| 	// Now that this query is done, start the next one.
 | ||
| 	n.startNextQuery(net)
 | ||
| 	return nil
 | ||
| }
 |