481 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
			
		
		
	
	
			481 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Go
		
	
	
	
	
	
| // Copyright 2017 The go-ethereum Authors
 | |
| // This file is part of go-ethereum.
 | |
| //
 | |
| // go-ethereum is free software: you can redistribute it and/or modify
 | |
| // it under the terms of the GNU General Public License as published by
 | |
| // the Free Software Foundation, either version 3 of the License, or
 | |
| // (at your option) any later version.
 | |
| //
 | |
| // go-ethereum is distributed in the hope that it will be useful,
 | |
| // but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
| // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | |
| // GNU General Public License for more details.
 | |
| //
 | |
| // You should have received a copy of the GNU General Public License
 | |
| // along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
 | |
| 
 | |
| package main
 | |
| 
 | |
| import (
 | |
| 	"context"
 | |
| 	"crypto/ecdsa"
 | |
| 	"flag"
 | |
| 	"fmt"
 | |
| 	"io/ioutil"
 | |
| 	"net"
 | |
| 	"os"
 | |
| 	"path"
 | |
| 	"path/filepath"
 | |
| 	"runtime"
 | |
| 	"sync"
 | |
| 	"syscall"
 | |
| 	"testing"
 | |
| 	"time"
 | |
| 
 | |
| 	"github.com/docker/docker/pkg/reexec"
 | |
| 	"github.com/ethereum/go-ethereum/accounts"
 | |
| 	"github.com/ethereum/go-ethereum/accounts/keystore"
 | |
| 	"github.com/ethereum/go-ethereum/internal/cmdtest"
 | |
| 	"github.com/ethereum/go-ethereum/node"
 | |
| 	"github.com/ethereum/go-ethereum/p2p"
 | |
| 	"github.com/ethereum/go-ethereum/rpc"
 | |
| 	"github.com/ethereum/go-ethereum/swarm"
 | |
| 	"github.com/ethereum/go-ethereum/swarm/api"
 | |
| 	swarmhttp "github.com/ethereum/go-ethereum/swarm/api/http"
 | |
| )
 | |
| 
 | |
| var loglevel = flag.Int("loglevel", 3, "verbosity of logs")
 | |
| 
 | |
| func init() {
 | |
| 	// Run the app if we've been exec'd as "swarm-test" in runSwarm.
 | |
| 	reexec.Register("swarm-test", func() {
 | |
| 		if err := app.Run(os.Args); err != nil {
 | |
| 			fmt.Fprintln(os.Stderr, err)
 | |
| 			os.Exit(1)
 | |
| 		}
 | |
| 		os.Exit(0)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| const clusterSize = 3
 | |
| 
 | |
| var clusteronce sync.Once
 | |
| var cluster *testCluster
 | |
| 
 | |
| func initCluster(t *testing.T) {
 | |
| 	clusteronce.Do(func() {
 | |
| 		cluster = newTestCluster(t, clusterSize)
 | |
| 	})
 | |
| }
 | |
| 
 | |
| func serverFunc(api *api.API) swarmhttp.TestServer {
 | |
| 	return swarmhttp.NewServer(api, "")
 | |
| }
 | |
| func TestMain(m *testing.M) {
 | |
| 	// check if we have been reexec'd
 | |
| 	if reexec.Init() {
 | |
| 		return
 | |
| 	}
 | |
| 	os.Exit(m.Run())
 | |
| }
 | |
| 
 | |
| func runSwarm(t *testing.T, args ...string) *cmdtest.TestCmd {
 | |
| 	tt := cmdtest.NewTestCmd(t, nil)
 | |
| 
 | |
| 	// Boot "swarm". This actually runs the test binary but the TestMain
 | |
| 	// function will prevent any tests from running.
 | |
| 	tt.Run("swarm-test", args...)
 | |
| 
 | |
| 	return tt
 | |
| }
 | |
| 
 | |
| type testCluster struct {
 | |
| 	Nodes  []*testNode
 | |
| 	TmpDir string
 | |
| }
 | |
| 
 | |
| // newTestCluster starts a test swarm cluster of the given size.
 | |
| //
 | |
| // A temporary directory is created and each node gets a data directory inside
 | |
| // it.
 | |
| //
 | |
| // Each node listens on 127.0.0.1 with random ports for both the HTTP and p2p
 | |
| // ports (assigned by first listening on 127.0.0.1:0 and then passing the ports
 | |
| // as flags).
 | |
| //
 | |
| // When starting more than one node, they are connected together using the
 | |
| // admin SetPeer RPC method.
 | |
| 
 | |
| func newTestCluster(t *testing.T, size int) *testCluster {
 | |
| 	cluster := &testCluster{}
 | |
| 	defer func() {
 | |
| 		if t.Failed() {
 | |
| 			cluster.Shutdown()
 | |
| 		}
 | |
| 	}()
 | |
| 
 | |
| 	tmpdir, err := ioutil.TempDir("", "swarm-test")
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	cluster.TmpDir = tmpdir
 | |
| 
 | |
| 	// start the nodes
 | |
| 	cluster.StartNewNodes(t, size)
 | |
| 
 | |
| 	if size == 1 {
 | |
| 		return cluster
 | |
| 	}
 | |
| 
 | |
| 	// connect the nodes together
 | |
| 	for _, node := range cluster.Nodes {
 | |
| 		if err := node.Client.Call(nil, "admin_addPeer", cluster.Nodes[0].Enode); err != nil {
 | |
| 			t.Fatal(err)
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	// wait until all nodes have the correct number of peers
 | |
| outer:
 | |
| 	for _, node := range cluster.Nodes {
 | |
| 		var peers []*p2p.PeerInfo
 | |
| 		for start := time.Now(); time.Since(start) < time.Minute; time.Sleep(50 * time.Millisecond) {
 | |
| 			if err := node.Client.Call(&peers, "admin_peers"); err != nil {
 | |
| 				t.Fatal(err)
 | |
| 			}
 | |
| 			if len(peers) == len(cluster.Nodes)-1 {
 | |
| 				continue outer
 | |
| 			}
 | |
| 		}
 | |
| 		t.Fatalf("%s only has %d / %d peers", node.Name, len(peers), len(cluster.Nodes)-1)
 | |
| 	}
 | |
| 
 | |
| 	return cluster
 | |
| }
 | |
| 
 | |
| func (c *testCluster) Shutdown() {
 | |
| 	for _, node := range c.Nodes {
 | |
| 		node.Shutdown()
 | |
| 	}
 | |
| 	os.RemoveAll(c.TmpDir)
 | |
| }
 | |
| 
 | |
| func (c *testCluster) Stop() {
 | |
| 	for _, node := range c.Nodes {
 | |
| 		node.Shutdown()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (c *testCluster) StartNewNodes(t *testing.T, size int) {
 | |
| 	c.Nodes = make([]*testNode, 0, size)
 | |
| 	for i := 0; i < size; i++ {
 | |
| 		dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
 | |
| 		if err := os.Mkdir(dir, 0700); err != nil {
 | |
| 			t.Fatal(err)
 | |
| 		}
 | |
| 
 | |
| 		node := newTestNode(t, dir)
 | |
| 		node.Name = fmt.Sprintf("swarm%02d", i)
 | |
| 
 | |
| 		c.Nodes = append(c.Nodes, node)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (c *testCluster) StartExistingNodes(t *testing.T, size int, bzzaccount string) {
 | |
| 	c.Nodes = make([]*testNode, 0, size)
 | |
| 	for i := 0; i < size; i++ {
 | |
| 		dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
 | |
| 		node := existingTestNode(t, dir, bzzaccount)
 | |
| 		node.Name = fmt.Sprintf("swarm%02d", i)
 | |
| 
 | |
| 		c.Nodes = append(c.Nodes, node)
 | |
| 	}
 | |
| }
 | |
| 
 | |
| func (c *testCluster) Cleanup() {
 | |
| 	os.RemoveAll(c.TmpDir)
 | |
| }
 | |
| 
 | |
| type testNode struct {
 | |
| 	Name       string
 | |
| 	Addr       string
 | |
| 	URL        string
 | |
| 	Enode      string
 | |
| 	Dir        string
 | |
| 	IpcPath    string
 | |
| 	PrivateKey *ecdsa.PrivateKey
 | |
| 	Client     *rpc.Client
 | |
| 	Cmd        *cmdtest.TestCmd
 | |
| }
 | |
| 
 | |
| const testPassphrase = "swarm-test-passphrase"
 | |
| 
 | |
| func getTestAccount(t *testing.T, dir string) (conf *node.Config, account accounts.Account) {
 | |
| 	// create key
 | |
| 	conf = &node.Config{
 | |
| 		DataDir: dir,
 | |
| 		IPCPath: "bzzd.ipc",
 | |
| 		NoUSB:   true,
 | |
| 	}
 | |
| 	n, err := node.New(conf)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	account, err = n.AccountManager().Backends(keystore.KeyStoreType)[0].(*keystore.KeyStore).NewAccount(testPassphrase)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// use a unique IPCPath when running tests on Windows
 | |
| 	if runtime.GOOS == "windows" {
 | |
| 		conf.IPCPath = fmt.Sprintf("bzzd-%s.ipc", account.Address.String())
 | |
| 	}
 | |
| 
 | |
| 	return conf, account
 | |
| }
 | |
| 
 | |
| func existingTestNode(t *testing.T, dir string, bzzaccount string) *testNode {
 | |
| 	conf, _ := getTestAccount(t, dir)
 | |
| 	node := &testNode{Dir: dir}
 | |
| 
 | |
| 	// use a unique IPCPath when running tests on Windows
 | |
| 	if runtime.GOOS == "windows" {
 | |
| 		conf.IPCPath = fmt.Sprintf("bzzd-%s.ipc", bzzaccount)
 | |
| 	}
 | |
| 
 | |
| 	// assign ports
 | |
| 	ports, err := getAvailableTCPPorts(2)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	p2pPort := ports[0]
 | |
| 	httpPort := ports[1]
 | |
| 
 | |
| 	// start the node
 | |
| 	node.Cmd = runSwarm(t,
 | |
| 		"--port", p2pPort,
 | |
| 		"--nat", "extip:127.0.0.1",
 | |
| 		"--datadir", dir,
 | |
| 		"--ipcpath", conf.IPCPath,
 | |
| 		"--ens-api", "",
 | |
| 		"--bzzaccount", bzzaccount,
 | |
| 		"--bzznetworkid", "321",
 | |
| 		"--bzzport", httpPort,
 | |
| 		"--verbosity", fmt.Sprint(*loglevel),
 | |
| 	)
 | |
| 	node.Cmd.InputLine(testPassphrase)
 | |
| 	defer func() {
 | |
| 		if t.Failed() {
 | |
| 			node.Shutdown()
 | |
| 		}
 | |
| 	}()
 | |
| 
 | |
| 	ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
 | |
| 	defer cancel()
 | |
| 
 | |
| 	// ensure that all ports have active listeners
 | |
| 	// so that the next node will not get the same
 | |
| 	// when calling getAvailableTCPPorts
 | |
| 	err = waitTCPPorts(ctx, ports...)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// wait for the node to start
 | |
| 	for start := time.Now(); time.Since(start) < 10*time.Second; time.Sleep(50 * time.Millisecond) {
 | |
| 		node.Client, err = rpc.Dial(conf.IPCEndpoint())
 | |
| 		if err == nil {
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 	if node.Client == nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// load info
 | |
| 	var info swarm.Info
 | |
| 	if err := node.Client.Call(&info, "bzz_info"); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	node.Addr = net.JoinHostPort("127.0.0.1", info.Port)
 | |
| 	node.URL = "http://" + node.Addr
 | |
| 
 | |
| 	var nodeInfo p2p.NodeInfo
 | |
| 	if err := node.Client.Call(&nodeInfo, "admin_nodeInfo"); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	node.Enode = nodeInfo.Enode
 | |
| 	node.IpcPath = conf.IPCPath
 | |
| 	return node
 | |
| }
 | |
| 
 | |
| func newTestNode(t *testing.T, dir string) *testNode {
 | |
| 
 | |
| 	conf, account := getTestAccount(t, dir)
 | |
| 	ks := keystore.NewKeyStore(path.Join(dir, "keystore"), 1<<18, 1)
 | |
| 
 | |
| 	pk := decryptStoreAccount(ks, account.Address.Hex(), []string{testPassphrase})
 | |
| 
 | |
| 	node := &testNode{Dir: dir, PrivateKey: pk}
 | |
| 
 | |
| 	// assign ports
 | |
| 	ports, err := getAvailableTCPPorts(2)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	p2pPort := ports[0]
 | |
| 	httpPort := ports[1]
 | |
| 
 | |
| 	// start the node
 | |
| 	node.Cmd = runSwarm(t,
 | |
| 		"--port", p2pPort,
 | |
| 		"--nat", "extip:127.0.0.1",
 | |
| 		"--datadir", dir,
 | |
| 		"--ipcpath", conf.IPCPath,
 | |
| 		"--ens-api", "",
 | |
| 		"--bzzaccount", account.Address.String(),
 | |
| 		"--bzznetworkid", "321",
 | |
| 		"--bzzport", httpPort,
 | |
| 		"--verbosity", fmt.Sprint(*loglevel),
 | |
| 	)
 | |
| 	node.Cmd.InputLine(testPassphrase)
 | |
| 	defer func() {
 | |
| 		if t.Failed() {
 | |
| 			node.Shutdown()
 | |
| 		}
 | |
| 	}()
 | |
| 
 | |
| 	ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
 | |
| 	defer cancel()
 | |
| 
 | |
| 	// ensure that all ports have active listeners
 | |
| 	// so that the next node will not get the same
 | |
| 	// when calling getAvailableTCPPorts
 | |
| 	err = waitTCPPorts(ctx, ports...)
 | |
| 	if err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// wait for the node to start
 | |
| 	for start := time.Now(); time.Since(start) < 10*time.Second; time.Sleep(50 * time.Millisecond) {
 | |
| 		node.Client, err = rpc.Dial(conf.IPCEndpoint())
 | |
| 		if err == nil {
 | |
| 			break
 | |
| 		}
 | |
| 	}
 | |
| 	if node.Client == nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 
 | |
| 	// load info
 | |
| 	var info swarm.Info
 | |
| 	if err := node.Client.Call(&info, "bzz_info"); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	node.Addr = net.JoinHostPort("127.0.0.1", info.Port)
 | |
| 	node.URL = "http://" + node.Addr
 | |
| 
 | |
| 	var nodeInfo p2p.NodeInfo
 | |
| 	if err := node.Client.Call(&nodeInfo, "admin_nodeInfo"); err != nil {
 | |
| 		t.Fatal(err)
 | |
| 	}
 | |
| 	node.Enode = nodeInfo.Enode
 | |
| 	node.IpcPath = conf.IPCPath
 | |
| 	return node
 | |
| }
 | |
| 
 | |
| func (n *testNode) Shutdown() {
 | |
| 	if n.Cmd != nil {
 | |
| 		n.Cmd.Kill()
 | |
| 	}
 | |
| }
 | |
| 
 | |
| // getAvailableTCPPorts returns a set of ports that
 | |
| // nothing is listening on at the time.
 | |
| //
 | |
| // Function assignTCPPort cannot be called in sequence
 | |
| // and guardantee that the same port will be returned in
 | |
| // different calls as the listener is closed within the function,
 | |
| // not after all listeners are started and selected unique
 | |
| // available ports.
 | |
| func getAvailableTCPPorts(count int) (ports []string, err error) {
 | |
| 	for i := 0; i < count; i++ {
 | |
| 		l, err := net.Listen("tcp", "127.0.0.1:0")
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		// defer close in the loop to be sure the same port will not
 | |
| 		// be selected in the next iteration
 | |
| 		defer l.Close()
 | |
| 
 | |
| 		_, port, err := net.SplitHostPort(l.Addr().String())
 | |
| 		if err != nil {
 | |
| 			return nil, err
 | |
| 		}
 | |
| 		ports = append(ports, port)
 | |
| 	}
 | |
| 	return ports, nil
 | |
| }
 | |
| 
 | |
| // waitTCPPorts blocks until tcp connections can be
 | |
| // established on all provided ports. It runs all
 | |
| // ports dialers in parallel, and returns the first
 | |
| // encountered error.
 | |
| // See waitTCPPort also.
 | |
| func waitTCPPorts(ctx context.Context, ports ...string) error {
 | |
| 	var err error
 | |
| 	// mu locks err variable that is assigned in
 | |
| 	// other goroutines
 | |
| 	var mu sync.Mutex
 | |
| 
 | |
| 	// cancel is canceling all goroutines
 | |
| 	// when the firs error is returned
 | |
| 	// to prevent unnecessary waiting
 | |
| 	ctx, cancel := context.WithCancel(ctx)
 | |
| 	defer cancel()
 | |
| 
 | |
| 	var wg sync.WaitGroup
 | |
| 	for _, port := range ports {
 | |
| 		wg.Add(1)
 | |
| 		go func(port string) {
 | |
| 			defer wg.Done()
 | |
| 
 | |
| 			e := waitTCPPort(ctx, port)
 | |
| 
 | |
| 			mu.Lock()
 | |
| 			defer mu.Unlock()
 | |
| 			if e != nil && err == nil {
 | |
| 				err = e
 | |
| 				cancel()
 | |
| 			}
 | |
| 		}(port)
 | |
| 	}
 | |
| 	wg.Wait()
 | |
| 
 | |
| 	return err
 | |
| }
 | |
| 
 | |
| // waitTCPPort blocks until tcp connection can be established
 | |
| // ona provided port. It has a 3 minute timeout as maximum,
 | |
| // to prevent long waiting, but it can be shortened with
 | |
| // a provided context instance. Dialer has a 10 second timeout
 | |
| // in every iteration, and connection refused error will be
 | |
| // retried in 100 milliseconds periods.
 | |
| func waitTCPPort(ctx context.Context, port string) error {
 | |
| 	ctx, cancel := context.WithTimeout(ctx, 3*time.Minute)
 | |
| 	defer cancel()
 | |
| 
 | |
| 	for {
 | |
| 		c, err := (&net.Dialer{Timeout: 10 * time.Second}).DialContext(ctx, "tcp", "127.0.0.1:"+port)
 | |
| 		if err != nil {
 | |
| 			if operr, ok := err.(*net.OpError); ok {
 | |
| 				if syserr, ok := operr.Err.(*os.SyscallError); ok && syserr.Err == syscall.ECONNREFUSED {
 | |
| 					time.Sleep(100 * time.Millisecond)
 | |
| 					continue
 | |
| 				}
 | |
| 			}
 | |
| 			return err
 | |
| 		}
 | |
| 		return c.Close()
 | |
| 	}
 | |
| }
 |