* swarm/network: DRY out repeated giga comment I not necessarily agree with the way we wait for event propagation. But I truly disagree with having duplicated giga comments. * p2p/simulations: encapsulate Node.Up field so we avoid data races The Node.Up field was accessed concurrently without "proper" locking. There was a lock on Network and that was used sometimes to access the field. Other times the locking was missed and we had a data race. For example: https://github.com/ethereum/go-ethereum/pull/18464 The case above was solved, but there were still intermittent/hard to reproduce races. So let's solve the issue permanently. resolves: ethersphere/go-ethereum#1146 * p2p/simulations: fix unmarshal of simulations.Node Making Node.Up field private in13292ee897broke TestHTTPNetwork and TestHTTPSnapshot. Because the default UnmarshalJSON does not handle unexported fields. Important: The fix is partial and not proper to my taste. But I cut scope as I think the fix may require a change to the current serialization format. New ticket: https://github.com/ethersphere/go-ethereum/issues/1177 * p2p/simulations: Add a sanity test case for Node.Config UnmarshalJSON * p2p/simulations: revert back to defer Unlock() pattern for Network It's a good patten to call `defer Unlock()` right after `Lock()` so (new) error cases won't miss to unlock. Let's get back to that pattern. The patten was abandoned in85a79b3ad3, while fixing a data race. That data race does not exist anymore, since the Node.Up field got hidden behind its own lock. * p2p/simulations: consistent naming for test providers Node.UnmarshalJSON * p2p/simulations: remove JSON annotation from private fields of Node As unexported fields are not serialized. * p2p/simulations: fix deadlock in Network.GetRandomDownNode() Problem: GetRandomDownNode() locks -> getDownNodeIDs() -> GetNodes() tries to lock -> deadlock On Network type, unexported functions must assume that `net.lock` is already acquired and should not call exported functions which might try to lock again. * p2p/simulations: ensure method conformity for Network Connect* methods were moved to p2p/simulations.Network from swarm/network/simulation. However these new methods did not follow the pattern of Network methods, i.e., all exported method locks the whole Network either for read or write. * p2p/simulations: fix deadlock during network shutdown `TestDiscoveryPersistenceSimulationSimAdapter` often got into deadlock. The execution was stuck on two locks, i.e, `Kademlia.lock` and `p2p/simulations.Network.lock`. Usually the test got stuck once in each 20 executions with high confidence. `Kademlia` was stuck in `Kademlia.EachAddr()` and `Network` in `Network.Stop()`. Solution: in `Network.Stop()` `net.lock` must be released before calling `node.Stop()` as stopping a node (somehow - I did not find the exact code path) causes `Network.InitConn()` to be called from `Kademlia.SuggestPeer()` and that blocks on `net.lock`. Related ticket: https://github.com/ethersphere/go-ethereum/issues/1223 * swarm/state: simplify if statement in DBStore.Put() * p2p/simulations: remove faulty godoc from private function The comment started with the wrong method name. The method is simple and self explanatory. Also, it's private. => Let's just remove the comment.
		
			
				
	
	
		
			195 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			195 lines
		
	
	
		
			5.3 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
// Copyright 2018 The go-ethereum Authors
 | 
						|
// This file is part of the go-ethereum library.
 | 
						|
//
 | 
						|
// The go-ethereum library is free software: you can redistribute it and/or modify
 | 
						|
// it under the terms of the GNU Lesser General Public License as published by
 | 
						|
// the Free Software Foundation, either version 3 of the License, or
 | 
						|
// (at your option) any later version.
 | 
						|
//
 | 
						|
// The go-ethereum library is distributed in the hope that it will be useful,
 | 
						|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
 | 
						|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | 
						|
// GNU Lesser General Public License for more details.
 | 
						|
//
 | 
						|
// You should have received a copy of the GNU Lesser General Public License
 | 
						|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
 | 
						|
package main
 | 
						|
 | 
						|
import (
 | 
						|
	"context"
 | 
						|
	"encoding/json"
 | 
						|
	"fmt"
 | 
						|
	"io/ioutil"
 | 
						|
	"net/http"
 | 
						|
	"net/http/httptest"
 | 
						|
	"net/url"
 | 
						|
	"testing"
 | 
						|
	"time"
 | 
						|
 | 
						|
	"github.com/ethereum/go-ethereum/p2p/enode"
 | 
						|
	"github.com/ethereum/go-ethereum/p2p/simulations"
 | 
						|
	"github.com/ethereum/go-ethereum/swarm/log"
 | 
						|
)
 | 
						|
 | 
						|
var (
 | 
						|
	nodeCount = 10
 | 
						|
)
 | 
						|
 | 
						|
//This test is used to test the overlay simulation.
 | 
						|
//As the simulation is executed via a main, it is easily missed on changes
 | 
						|
//An automated test will prevent that
 | 
						|
//The test just connects to the simulations, starts the network,
 | 
						|
//starts the mocker, gets the number of nodes, and stops it again.
 | 
						|
//It also provides a documentation on the steps needed by frontends
 | 
						|
//to use the simulations
 | 
						|
func TestOverlaySim(t *testing.T) {
 | 
						|
	//start the simulation
 | 
						|
	log.Info("Start simulation backend")
 | 
						|
	//get the simulation networ; needed to subscribe for up events
 | 
						|
	net := newSimulationNetwork()
 | 
						|
	//create the overlay simulation
 | 
						|
	sim := newOverlaySim(net)
 | 
						|
	//create a http test server with it
 | 
						|
	srv := httptest.NewServer(sim)
 | 
						|
	defer srv.Close()
 | 
						|
 | 
						|
	log.Debug("Http simulation server started. Start simulation network")
 | 
						|
	//start the simulation network (initialization of simulation)
 | 
						|
	resp, err := http.Post(srv.URL+"/start", "application/json", nil)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	defer resp.Body.Close()
 | 
						|
	if resp.StatusCode != http.StatusOK {
 | 
						|
		t.Fatalf("Expected Status Code %d, got %d", http.StatusOK, resp.StatusCode)
 | 
						|
	}
 | 
						|
 | 
						|
	log.Debug("Start mocker")
 | 
						|
	//start the mocker, needs a node count and an ID
 | 
						|
	resp, err = http.PostForm(srv.URL+"/mocker/start",
 | 
						|
		url.Values{
 | 
						|
			"node-count":  {fmt.Sprintf("%d", nodeCount)},
 | 
						|
			"mocker-type": {simulations.GetMockerList()[0]},
 | 
						|
		})
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	defer resp.Body.Close()
 | 
						|
	if resp.StatusCode != http.StatusOK {
 | 
						|
		reason, err := ioutil.ReadAll(resp.Body)
 | 
						|
		if err != nil {
 | 
						|
			t.Fatal(err)
 | 
						|
		}
 | 
						|
		t.Fatalf("Expected Status Code %d, got %d, response body %s", http.StatusOK, resp.StatusCode, string(reason))
 | 
						|
	}
 | 
						|
 | 
						|
	//variables needed to wait for nodes being up
 | 
						|
	var upCount int
 | 
						|
	trigger := make(chan enode.ID)
 | 
						|
 | 
						|
	//wait for all nodes to be up
 | 
						|
	ctx, cancel := context.WithTimeout(context.Background(), 10*time.Second)
 | 
						|
	defer cancel()
 | 
						|
 | 
						|
	//start watching node up events...
 | 
						|
	go watchSimEvents(net, ctx, trigger)
 | 
						|
 | 
						|
	//...and wait until all expected up events (nodeCount) have been received
 | 
						|
LOOP:
 | 
						|
	for {
 | 
						|
		select {
 | 
						|
		case <-trigger:
 | 
						|
			//new node up event received, increase counter
 | 
						|
			upCount++
 | 
						|
			//all expected node up events received
 | 
						|
			if upCount == nodeCount {
 | 
						|
				break LOOP
 | 
						|
			}
 | 
						|
		case <-ctx.Done():
 | 
						|
			t.Fatalf("Timed out waiting for up events")
 | 
						|
		}
 | 
						|
 | 
						|
	}
 | 
						|
 | 
						|
	//at this point we can query the server
 | 
						|
	log.Info("Get number of nodes")
 | 
						|
	//get the number of nodes
 | 
						|
	resp, err = http.Get(srv.URL + "/nodes")
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	defer resp.Body.Close()
 | 
						|
	if resp.StatusCode != http.StatusOK {
 | 
						|
		t.Fatalf("err %s", resp.Status)
 | 
						|
	}
 | 
						|
	b, err := ioutil.ReadAll(resp.Body)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	//unmarshal number of nodes from JSON response
 | 
						|
	var nodesArr []simulations.Node
 | 
						|
	err = json.Unmarshal(b, &nodesArr)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
 | 
						|
	//check if number of nodes received is same as sent
 | 
						|
	if len(nodesArr) != nodeCount {
 | 
						|
		t.Fatal(fmt.Errorf("Expected %d number of nodes, got %d", nodeCount, len(nodesArr)))
 | 
						|
	}
 | 
						|
 | 
						|
	//need to let it run for a little while, otherwise stopping it immediately can crash due running nodes
 | 
						|
	//wanting to connect to already stopped nodes
 | 
						|
	time.Sleep(1 * time.Second)
 | 
						|
 | 
						|
	log.Info("Stop the network")
 | 
						|
	//stop the network
 | 
						|
	resp, err = http.Post(srv.URL+"/stop", "application/json", nil)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	defer resp.Body.Close()
 | 
						|
	if resp.StatusCode != http.StatusOK {
 | 
						|
		t.Fatalf("err %s", resp.Status)
 | 
						|
	}
 | 
						|
 | 
						|
	log.Info("Reset the network")
 | 
						|
	//reset the network (removes all nodes and connections)
 | 
						|
	resp, err = http.Post(srv.URL+"/reset", "application/json", nil)
 | 
						|
	if err != nil {
 | 
						|
		t.Fatal(err)
 | 
						|
	}
 | 
						|
	defer resp.Body.Close()
 | 
						|
	if resp.StatusCode != http.StatusOK {
 | 
						|
		t.Fatalf("err %s", resp.Status)
 | 
						|
	}
 | 
						|
}
 | 
						|
 | 
						|
//watch for events so we know when all nodes are up
 | 
						|
func watchSimEvents(net *simulations.Network, ctx context.Context, trigger chan enode.ID) {
 | 
						|
	events := make(chan *simulations.Event)
 | 
						|
	sub := net.Events().Subscribe(events)
 | 
						|
	defer sub.Unsubscribe()
 | 
						|
 | 
						|
	for {
 | 
						|
		select {
 | 
						|
		case ev := <-events:
 | 
						|
			//only catch node up events
 | 
						|
			if ev.Type == simulations.EventTypeNode {
 | 
						|
				if ev.Node.Up() {
 | 
						|
					log.Debug("got node up event", "event", ev, "node", ev.Node.Config.ID)
 | 
						|
					select {
 | 
						|
					case trigger <- ev.Node.Config.ID:
 | 
						|
					case <-ctx.Done():
 | 
						|
						return
 | 
						|
					}
 | 
						|
				}
 | 
						|
			}
 | 
						|
		case <-ctx.Done():
 | 
						|
			return
 | 
						|
		}
 | 
						|
	}
 | 
						|
}
 |