Compare commits
273 Commits
Author | SHA1 | Date | |
---|---|---|---|
|
225171a4bf | ||
|
702b8a7aec | ||
|
5d7e18539e | ||
|
c4a1d4fecf | ||
|
fb9f7261ec | ||
|
d927cbb638 | ||
|
2fbc454355 | ||
|
d6efa69187 | ||
|
3ea8ac6a9a | ||
|
8a9c31a307 | ||
|
6380c06c65 | ||
|
54d1111965 | ||
|
f00d0daf33 | ||
|
2cffd4ff3c | ||
|
7b1aa64220 | ||
|
8f4c4fea20 | ||
|
d42ce0f2c1 | ||
|
273c7a9dc4 | ||
|
a5d5609e38 | ||
|
93c0f1715d | ||
|
d9575e92fc | ||
|
11a402f747 | ||
|
021d6fbbbb | ||
|
feed8069a6 | ||
|
514022bde6 | ||
|
ff22ec31b6 | ||
|
0598707129 | ||
|
a511f6b515 | ||
|
8b1e14b7f8 | ||
|
6c412e313c | ||
|
6b232ce325 | ||
|
7abedf9bbb | ||
|
27a278e6e3 | ||
|
a0bcb16875 | ||
|
bc0a43191e | ||
|
1064b9e691 | ||
|
10780e8a00 | ||
|
2433349c80 | ||
|
58243b4d3e | ||
|
cab1cff11c | ||
|
2909f6d7a2 | ||
|
d96ba77113 | ||
|
62467e4405 | ||
|
d0082bb7ec | ||
|
21c059b67e | ||
|
9e24491c65 | ||
|
49f63deb24 | ||
|
b536460f8e | ||
|
afd8b84706 | ||
|
f6206efe5b | ||
|
ae674a3660 | ||
|
894022a3d5 | ||
|
68da9aa716 | ||
|
14bdcdeab4 | ||
|
fe6a9473dc | ||
|
427316a707 | ||
|
0647c4de7b | ||
|
7ddc2c9e95 | ||
|
dcaaa3c804 | ||
|
f5b128a5b3 | ||
|
fd982d3f3b | ||
|
526abe2736 | ||
|
8997efe31f | ||
|
4ea2d707f9 | ||
|
ee8877509a | ||
|
763e64cad8 | ||
|
040dd5bd5d | ||
|
dcdd57df62 | ||
|
323428865f | ||
|
65c91ad5e7 | ||
|
5d30be412b | ||
|
eb7f901289 | ||
|
db5e403afe | ||
|
96116758d2 | ||
|
65cebb7730 | ||
|
2e0391ea84 | ||
|
d483da766f | ||
|
7c9314f231 | ||
|
e1f1d3085c | ||
|
96339daf40 | ||
|
f7d3678c28 | ||
|
facf1bc9d6 | ||
|
e8824f6e74 | ||
|
a9835c1816 | ||
|
2eedbe799f | ||
|
b3711af051 | ||
|
30bdf817a0 | ||
|
fbeb4f20f9 | ||
|
0b20b1a050 | ||
|
4dbefc1f25 | ||
|
dbae1dc7b3 | ||
|
3b07451564 | ||
|
37685930d9 | ||
|
51df1c1f20 | ||
|
f524ec4326 | ||
|
eb794af833 | ||
|
67a7857124 | ||
|
c73b654fd1 | ||
|
9da128db70 | ||
|
4e5d1f1c39 | ||
|
d57e85ecc9 | ||
|
1990c9e621 | ||
|
319098cc1c | ||
|
223d943481 | ||
|
8974e2e5e0 | ||
|
a4a2343cdc | ||
|
fdfd6d3c39 | ||
|
b5537c5601 | ||
|
e916f9786d | ||
|
909e968ebb | ||
|
598f786aab | ||
|
461291882e | ||
|
1f0f6f0272 | ||
|
0a22ae5572 | ||
|
b0cfd9c786 | ||
|
6d8a1bfb08 | ||
|
4895665670 | ||
|
eaff89291c | ||
|
e187711c65 | ||
|
d926bf2c7e | ||
|
8db8d074e2 | ||
|
1a70338734 | ||
|
61a5976368 | ||
|
88c42ab4e7 | ||
|
4210dd1500 | ||
|
c971ab617d | ||
|
f1986f86f2 | ||
|
28aca90716 | ||
|
9b1536b26a | ||
|
3e57c33147 | ||
|
baa7eb901e | ||
|
574378edb5 | ||
|
c95e4a80d1 | ||
|
897ea01d5f | ||
|
ec192f18b4 | ||
|
aa34173f13 | ||
|
c343f75c26 | ||
|
52b1d09457 | ||
|
9402f96597 | ||
|
d0fd8d6fc2 | ||
|
cfde0b5f52 | ||
|
de06185fc3 | ||
|
3fb5f3ae11 | ||
|
e75d0a6e4c | ||
|
947e0afeb3 | ||
|
1836366ac1 | ||
|
591cef17d4 | ||
|
e33a5de454 | ||
|
f04c0e341e | ||
|
ea89f40f0d | ||
|
1fc54d92ec | ||
|
8c4a7fa8d3 | ||
|
423d4254f5 | ||
|
dea1ce052a | ||
|
25982375a8 | ||
|
049f5b3572 | ||
|
0255951587 | ||
|
85cd64df0e | ||
|
9608ccf106 | ||
|
3f06da7b5f | ||
|
546d42179e | ||
|
90829a04bf | ||
|
f991995918 | ||
|
aab7ab04b0 | ||
|
43b940ec5a | ||
|
b487bdf0ba | ||
|
a3267ed929 | ||
|
9f7592c802 | ||
|
99483e85b9 | ||
|
1d666cf27e | ||
|
eac16f9824 | ||
|
69c52bde3f | ||
|
2977538ac0 | ||
|
7f0726f706 | ||
|
13af276418 | ||
|
ea06da0892 | ||
|
feb6620c34 | ||
|
90b22773e9 | ||
|
9e4f96a1a6 | ||
|
01a7e267dc | ||
|
e8ea5aa0d5 | ||
|
5bee5d69d7 | ||
|
cbfb40b0aa | ||
|
4cf2b4110e | ||
|
0029a869f0 | ||
|
2ab24a2a8f | ||
|
400332b99d | ||
|
a5237a27ea | ||
|
7a22e89080 | ||
|
e3a993d774 | ||
|
ed40767355 | ||
|
a20cc75b4a | ||
|
b659718fd0 | ||
|
be2aec092d | ||
|
17f80cc2e2 | ||
|
143c4341d8 | ||
|
3f33a7c8ce | ||
|
c8dcb9584e | ||
|
af28d12847 | ||
|
0ad32d3be7 | ||
|
68b0d30d4a | ||
|
eae63c511c | ||
|
ca34e8230e | ||
|
342ec83d67 | ||
|
38c7eb0f26 | ||
|
d51faee240 | ||
|
426f62f1a8 | ||
|
7677ec1f34 | ||
|
d258eee211 | ||
|
84f8c0cc1f | ||
|
998f6564b2 | ||
|
40a2c52397 | ||
|
a9c6ef6905 | ||
|
ccc0debb63 | ||
|
ff9b14617e | ||
|
d6ed2f67a8 | ||
|
54294b45b1 | ||
|
d31802312a | ||
|
55b579e02c | ||
|
be22ee8dda | ||
|
56de337e57 | ||
|
c934c06cc1 | ||
|
fbf57d53e2 | ||
|
6ce21a4744 | ||
|
9af364e42b | ||
|
09d44247f7 | ||
|
0fe47e98c4 | ||
|
415969f534 | ||
|
d9cee2c172 | ||
|
ab6bdbd9b0 | ||
|
953b5ac015 | ||
|
f2fdb75dd9 | ||
|
f9c456e02d | ||
|
579bd0f9fb | ||
|
49719e21bc | ||
|
a2e43d28d0 | ||
|
6286c255f1 | ||
|
f6bc65fc68 | ||
|
ff8a033f18 | ||
|
247b5f0369 | ||
|
49ec4f0cd1 | ||
|
2688dab48c | ||
|
595b47e535 | ||
|
784aa83942 | ||
|
fcc18f4c80 | ||
|
53a18d2e27 | ||
|
7beccb29be | ||
|
5dbd8b42a9 | ||
|
4e7dc34ff1 | ||
|
4747aad160 | ||
|
4ea493e7eb | ||
|
c60f6f6214 | ||
|
ba975dc093 | ||
|
eab6e5a317 | ||
|
c4a4613d95 | ||
|
fedae95015 | ||
|
864e80a48f | ||
|
a42be3b78d | ||
|
6cf0ab38bd | ||
|
5463ed9996 | ||
|
d7be5c6619 | ||
|
d2fe83dc5c | ||
|
16f3c31773 | ||
|
5b3af4c3d1 | ||
|
60b433ab84 | ||
|
fd3da7c69d | ||
|
cd9a1d5b37 | ||
|
2ad511ce09 | ||
|
541f299fbb | ||
|
7c02933275 | ||
|
f2447bd4c3 | ||
|
ea1724de1a | ||
|
577d375a0d |
38
.github/CODEOWNERS
vendored
38
.github/CODEOWNERS
vendored
@@ -1,12 +1,32 @@
|
|||||||
# Lines starting with '#' are comments.
|
# Lines starting with '#' are comments.
|
||||||
# Each line is a file pattern followed by one or more owners.
|
# Each line is a file pattern followed by one or more owners.
|
||||||
|
|
||||||
accounts/usbwallet @karalabe
|
accounts/usbwallet @karalabe
|
||||||
consensus @karalabe
|
consensus @karalabe
|
||||||
core/ @karalabe @holiman
|
core/ @karalabe @holiman
|
||||||
eth/ @karalabe
|
eth/ @karalabe
|
||||||
les/ @zsfelfoldi
|
les/ @zsfelfoldi
|
||||||
light/ @zsfelfoldi
|
light/ @zsfelfoldi
|
||||||
mobile/ @karalabe
|
mobile/ @karalabe
|
||||||
p2p/ @fjl @zsfelfoldi
|
p2p/ @fjl @zsfelfoldi
|
||||||
whisper/ @gballet @gluk256
|
swarm/bmt @zelig
|
||||||
|
swarm/dev @lmars
|
||||||
|
swarm/fuse @jmozah @holisticode
|
||||||
|
swarm/grafana_dashboards @nonsense
|
||||||
|
swarm/metrics @nonsense @holisticode
|
||||||
|
swarm/multihash @nolash
|
||||||
|
swarm/network/bitvector @zelig @janos @gbalint
|
||||||
|
swarm/network/priorityqueue @zelig @janos @gbalint
|
||||||
|
swarm/network/simulations @zelig
|
||||||
|
swarm/network/stream @janos @zelig @gbalint @holisticode @justelad
|
||||||
|
swarm/network/stream/intervals @janos
|
||||||
|
swarm/network/stream/testing @zelig
|
||||||
|
swarm/pot @zelig
|
||||||
|
swarm/pss @nolash @zelig @nonsense
|
||||||
|
swarm/services @zelig
|
||||||
|
swarm/state @justelad
|
||||||
|
swarm/storage/encryption @gbalint @zelig @nagydani
|
||||||
|
swarm/storage/mock @janos
|
||||||
|
swarm/storage/mru @nolash
|
||||||
|
swarm/testutil @lmars
|
||||||
|
whisper/ @gballet @gluk256
|
||||||
|
12
.travis.yml
12
.travis.yml
@@ -126,7 +126,7 @@ matrix:
|
|||||||
|
|
||||||
# This builder does the Android Maven and Azure uploads
|
# This builder does the Android Maven and Azure uploads
|
||||||
- os: linux
|
- os: linux
|
||||||
dist: precise # Needed for the android tools
|
dist: trusty
|
||||||
addons:
|
addons:
|
||||||
apt:
|
apt:
|
||||||
packages:
|
packages:
|
||||||
@@ -146,16 +146,16 @@ matrix:
|
|||||||
git:
|
git:
|
||||||
submodules: false # avoid cloning ethereum/tests
|
submodules: false # avoid cloning ethereum/tests
|
||||||
before_install:
|
before_install:
|
||||||
- curl https://storage.googleapis.com/golang/go1.10.1.linux-amd64.tar.gz | tar -xz
|
- curl https://storage.googleapis.com/golang/go1.10.3.linux-amd64.tar.gz | tar -xz
|
||||||
- export PATH=`pwd`/go/bin:$PATH
|
- export PATH=`pwd`/go/bin:$PATH
|
||||||
- export GOROOT=`pwd`/go
|
- export GOROOT=`pwd`/go
|
||||||
- export GOPATH=$HOME/go
|
- export GOPATH=$HOME/go
|
||||||
script:
|
script:
|
||||||
# Build the Android archive and upload it to Maven Central and Azure
|
# Build the Android archive and upload it to Maven Central and Azure
|
||||||
- curl https://dl.google.com/android/repository/android-ndk-r15c-linux-x86_64.zip -o android-ndk-r15c.zip
|
- curl https://dl.google.com/android/repository/android-ndk-r17b-linux-x86_64.zip -o android-ndk-r17b.zip
|
||||||
- unzip -q android-ndk-r15c.zip && rm android-ndk-r15c.zip
|
- unzip -q android-ndk-r17b.zip && rm android-ndk-r17b.zip
|
||||||
- mv android-ndk-r15c $HOME
|
- mv android-ndk-r17b $HOME
|
||||||
- export ANDROID_NDK=$HOME/android-ndk-r15c
|
- export ANDROID_NDK=$HOME/android-ndk-r17b
|
||||||
|
|
||||||
- mkdir -p $GOPATH/src/github.com/ethereum
|
- mkdir -p $GOPATH/src/github.com/ethereum
|
||||||
- ln -s `pwd` $GOPATH/src/github.com/ethereum
|
- ln -s `pwd` $GOPATH/src/github.com/ethereum
|
||||||
|
1
AUTHORS
1
AUTHORS
@@ -171,3 +171,4 @@ xiekeyang <xiekeyang@users.noreply.github.com>
|
|||||||
yoza <yoza.is12s@gmail.com>
|
yoza <yoza.is12s@gmail.com>
|
||||||
ΞTHΞЯSPHΞЯΞ <{viktor.tron,nagydani,zsfelfoldi}@gmail.com>
|
ΞTHΞЯSPHΞЯΞ <{viktor.tron,nagydani,zsfelfoldi}@gmail.com>
|
||||||
Максим Чусовлянов <mchusovlianov@gmail.com>
|
Максим Чусовлянов <mchusovlianov@gmail.com>
|
||||||
|
Ralph Caraveo <deckarep@gmail.com>
|
||||||
|
1
Makefile
1
Makefile
@@ -41,6 +41,7 @@ lint: ## Run linters.
|
|||||||
build/env.sh go run build/ci.go lint
|
build/env.sh go run build/ci.go lint
|
||||||
|
|
||||||
clean:
|
clean:
|
||||||
|
./build/clean_go_build_cache.sh
|
||||||
rm -fr build/_workspace/pkg/ $(GOBIN)/*
|
rm -fr build/_workspace/pkg/ $(GOBIN)/*
|
||||||
|
|
||||||
# The devtools target installs tools required for 'go generate'.
|
# The devtools target installs tools required for 'go generate'.
|
||||||
|
@@ -40,7 +40,7 @@ The go-ethereum project comes with several wrappers/executables found in the `cm
|
|||||||
| `evm` | Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode. Its purpose is to allow isolated, fine-grained debugging of EVM opcodes (e.g. `evm --code 60ff60ff --debug`). |
|
| `evm` | Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode. Its purpose is to allow isolated, fine-grained debugging of EVM opcodes (e.g. `evm --code 60ff60ff --debug`). |
|
||||||
| `gethrpctest` | Developer utility tool to support our [ethereum/rpc-test](https://github.com/ethereum/rpc-tests) test suite which validates baseline conformity to the [Ethereum JSON RPC](https://github.com/ethereum/wiki/wiki/JSON-RPC) specs. Please see the [test suite's readme](https://github.com/ethereum/rpc-tests/blob/master/README.md) for details. |
|
| `gethrpctest` | Developer utility tool to support our [ethereum/rpc-test](https://github.com/ethereum/rpc-tests) test suite which validates baseline conformity to the [Ethereum JSON RPC](https://github.com/ethereum/wiki/wiki/JSON-RPC) specs. Please see the [test suite's readme](https://github.com/ethereum/rpc-tests/blob/master/README.md) for details. |
|
||||||
| `rlpdump` | Developer utility tool to convert binary RLP ([Recursive Length Prefix](https://github.com/ethereum/wiki/wiki/RLP)) dumps (data encoding used by the Ethereum protocol both network as well as consensus wise) to user friendlier hierarchical representation (e.g. `rlpdump --hex CE0183FFFFFFC4C304050583616263`). |
|
| `rlpdump` | Developer utility tool to convert binary RLP ([Recursive Length Prefix](https://github.com/ethereum/wiki/wiki/RLP)) dumps (data encoding used by the Ethereum protocol both network as well as consensus wise) to user friendlier hierarchical representation (e.g. `rlpdump --hex CE0183FFFFFFC4C304050583616263`). |
|
||||||
| `swarm` | swarm daemon and tools. This is the entrypoint for the swarm network. `swarm --help` for command line options and subcommands. See https://swarm-guide.readthedocs.io for swarm documentation. |
|
| `swarm` | Swarm daemon and tools. This is the entrypoint for the Swarm network. `swarm --help` for command line options and subcommands. See [Swarm README](https://github.com/ethereum/go-ethereum/tree/master/swarm) for more information. |
|
||||||
| `puppeth` | a CLI wizard that aids in creating a new Ethereum network. |
|
| `puppeth` | a CLI wizard that aids in creating a new Ethereum network. |
|
||||||
|
|
||||||
## Running geth
|
## Running geth
|
||||||
|
@@ -111,9 +111,14 @@ func (arguments Arguments) unpackTuple(v interface{}, marshalledValues []interfa
|
|||||||
if err := requireUnpackKind(value, typ, kind, arguments); err != nil {
|
if err := requireUnpackKind(value, typ, kind, arguments); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
// If the output interface is a struct, make sure names don't collide
|
|
||||||
|
// If the interface is a struct, get of abi->struct_field mapping
|
||||||
|
|
||||||
|
var abi2struct map[string]string
|
||||||
if kind == reflect.Struct {
|
if kind == reflect.Struct {
|
||||||
if err := requireUniqueStructFieldNames(arguments); err != nil {
|
var err error
|
||||||
|
abi2struct, err = mapAbiToStructFields(arguments, value)
|
||||||
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -123,9 +128,10 @@ func (arguments Arguments) unpackTuple(v interface{}, marshalledValues []interfa
|
|||||||
|
|
||||||
switch kind {
|
switch kind {
|
||||||
case reflect.Struct:
|
case reflect.Struct:
|
||||||
err := unpackStruct(value, reflectValue, arg)
|
if structField, ok := abi2struct[arg.Name]; ok {
|
||||||
if err != nil {
|
if err := set(value.FieldByName(structField), reflectValue, arg); err != nil {
|
||||||
return err
|
return err
|
||||||
|
}
|
||||||
}
|
}
|
||||||
case reflect.Slice, reflect.Array:
|
case reflect.Slice, reflect.Array:
|
||||||
if value.Len() < i {
|
if value.Len() < i {
|
||||||
@@ -151,17 +157,22 @@ func (arguments Arguments) unpackAtomic(v interface{}, marshalledValues []interf
|
|||||||
if len(marshalledValues) != 1 {
|
if len(marshalledValues) != 1 {
|
||||||
return fmt.Errorf("abi: wrong length, expected single value, got %d", len(marshalledValues))
|
return fmt.Errorf("abi: wrong length, expected single value, got %d", len(marshalledValues))
|
||||||
}
|
}
|
||||||
|
|
||||||
elem := reflect.ValueOf(v).Elem()
|
elem := reflect.ValueOf(v).Elem()
|
||||||
kind := elem.Kind()
|
kind := elem.Kind()
|
||||||
reflectValue := reflect.ValueOf(marshalledValues[0])
|
reflectValue := reflect.ValueOf(marshalledValues[0])
|
||||||
|
|
||||||
|
var abi2struct map[string]string
|
||||||
if kind == reflect.Struct {
|
if kind == reflect.Struct {
|
||||||
//make sure names don't collide
|
var err error
|
||||||
if err := requireUniqueStructFieldNames(arguments); err != nil {
|
if abi2struct, err = mapAbiToStructFields(arguments, elem); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
arg := arguments.NonIndexed()[0]
|
||||||
return unpackStruct(elem, reflectValue, arguments[0])
|
if structField, ok := abi2struct[arg.Name]; ok {
|
||||||
|
return set(elem.FieldByName(structField), reflectValue, arg)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
return set(elem, reflectValue, arguments.NonIndexed()[0])
|
return set(elem, reflectValue, arguments.NonIndexed()[0])
|
||||||
@@ -277,18 +288,3 @@ func capitalise(input string) string {
|
|||||||
}
|
}
|
||||||
return strings.ToUpper(input[:1]) + input[1:]
|
return strings.ToUpper(input[:1]) + input[1:]
|
||||||
}
|
}
|
||||||
|
|
||||||
//unpackStruct extracts each argument into its corresponding struct field
|
|
||||||
func unpackStruct(value, reflectValue reflect.Value, arg Argument) error {
|
|
||||||
name := capitalise(arg.Name)
|
|
||||||
typ := value.Type()
|
|
||||||
for j := 0; j < typ.NumField(); j++ {
|
|
||||||
// TODO read tags: `abi:"fieldName"`
|
|
||||||
if typ.Field(j).Name == name {
|
|
||||||
if err := set(value.Field(j), reflectValue, arg); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
@@ -31,6 +31,7 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/consensus/ethash"
|
"github.com/ethereum/go-ethereum/consensus/ethash"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core"
|
||||||
"github.com/ethereum/go-ethereum/core/bloombits"
|
"github.com/ethereum/go-ethereum/core/bloombits"
|
||||||
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
"github.com/ethereum/go-ethereum/core/state"
|
"github.com/ethereum/go-ethereum/core/state"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/core/vm"
|
"github.com/ethereum/go-ethereum/core/vm"
|
||||||
@@ -65,7 +66,7 @@ type SimulatedBackend struct {
|
|||||||
// NewSimulatedBackend creates a new binding backend using a simulated blockchain
|
// NewSimulatedBackend creates a new binding backend using a simulated blockchain
|
||||||
// for testing purposes.
|
// for testing purposes.
|
||||||
func NewSimulatedBackend(alloc core.GenesisAlloc) *SimulatedBackend {
|
func NewSimulatedBackend(alloc core.GenesisAlloc) *SimulatedBackend {
|
||||||
database, _ := ethdb.NewMemDatabase()
|
database := ethdb.NewMemDatabase()
|
||||||
genesis := core.Genesis{Config: params.AllEthashProtocolChanges, Alloc: alloc}
|
genesis := core.Genesis{Config: params.AllEthashProtocolChanges, Alloc: alloc}
|
||||||
genesis.MustCommit(database)
|
genesis.MustCommit(database)
|
||||||
blockchain, _ := core.NewBlockChain(database, nil, genesis.Config, ethash.NewFaker(), vm.Config{})
|
blockchain, _ := core.NewBlockChain(database, nil, genesis.Config, ethash.NewFaker(), vm.Config{})
|
||||||
@@ -159,7 +160,7 @@ func (b *SimulatedBackend) StorageAt(ctx context.Context, contract common.Addres
|
|||||||
|
|
||||||
// TransactionReceipt returns the receipt of a transaction.
|
// TransactionReceipt returns the receipt of a transaction.
|
||||||
func (b *SimulatedBackend) TransactionReceipt(ctx context.Context, txHash common.Hash) (*types.Receipt, error) {
|
func (b *SimulatedBackend) TransactionReceipt(ctx context.Context, txHash common.Hash) (*types.Receipt, error) {
|
||||||
receipt, _, _, _ := core.GetReceipt(b.database, txHash)
|
receipt, _, _, _ := rawdb.ReadReceipt(b.database, txHash)
|
||||||
return receipt, nil
|
return receipt, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -323,18 +324,24 @@ func (b *SimulatedBackend) SendTransaction(ctx context.Context, tx *types.Transa
|
|||||||
//
|
//
|
||||||
// TODO(karalabe): Deprecate when the subscription one can return past data too.
|
// TODO(karalabe): Deprecate when the subscription one can return past data too.
|
||||||
func (b *SimulatedBackend) FilterLogs(ctx context.Context, query ethereum.FilterQuery) ([]types.Log, error) {
|
func (b *SimulatedBackend) FilterLogs(ctx context.Context, query ethereum.FilterQuery) ([]types.Log, error) {
|
||||||
// Initialize unset filter boundaried to run from genesis to chain head
|
var filter *filters.Filter
|
||||||
from := int64(0)
|
if query.BlockHash != nil {
|
||||||
if query.FromBlock != nil {
|
// Block filter requested, construct a single-shot filter
|
||||||
from = query.FromBlock.Int64()
|
filter = filters.NewBlockFilter(&filterBackend{b.database, b.blockchain}, *query.BlockHash, query.Addresses, query.Topics)
|
||||||
|
} else {
|
||||||
|
// Initialize unset filter boundaried to run from genesis to chain head
|
||||||
|
from := int64(0)
|
||||||
|
if query.FromBlock != nil {
|
||||||
|
from = query.FromBlock.Int64()
|
||||||
|
}
|
||||||
|
to := int64(-1)
|
||||||
|
if query.ToBlock != nil {
|
||||||
|
to = query.ToBlock.Int64()
|
||||||
|
}
|
||||||
|
// Construct the range filter
|
||||||
|
filter = filters.NewRangeFilter(&filterBackend{b.database, b.blockchain}, from, to, query.Addresses, query.Topics)
|
||||||
}
|
}
|
||||||
to := int64(-1)
|
// Run the filter and return all the logs
|
||||||
if query.ToBlock != nil {
|
|
||||||
to = query.ToBlock.Int64()
|
|
||||||
}
|
|
||||||
// Construct and execute the filter
|
|
||||||
filter := filters.New(&filterBackend{b.database, b.blockchain}, from, to, query.Addresses, query.Topics)
|
|
||||||
|
|
||||||
logs, err := filter.Logs(ctx)
|
logs, err := filter.Logs(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@@ -429,12 +436,24 @@ func (fb *filterBackend) HeaderByNumber(ctx context.Context, block rpc.BlockNumb
|
|||||||
return fb.bc.GetHeaderByNumber(uint64(block.Int64())), nil
|
return fb.bc.GetHeaderByNumber(uint64(block.Int64())), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (fb *filterBackend) HeaderByHash(ctx context.Context, hash common.Hash) (*types.Header, error) {
|
||||||
|
return fb.bc.GetHeaderByHash(hash), nil
|
||||||
|
}
|
||||||
|
|
||||||
func (fb *filterBackend) GetReceipts(ctx context.Context, hash common.Hash) (types.Receipts, error) {
|
func (fb *filterBackend) GetReceipts(ctx context.Context, hash common.Hash) (types.Receipts, error) {
|
||||||
return core.GetBlockReceipts(fb.db, hash, core.GetBlockNumber(fb.db, hash)), nil
|
number := rawdb.ReadHeaderNumber(fb.db, hash)
|
||||||
|
if number == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
return rawdb.ReadReceipts(fb.db, hash, *number), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (fb *filterBackend) GetLogs(ctx context.Context, hash common.Hash) ([][]*types.Log, error) {
|
func (fb *filterBackend) GetLogs(ctx context.Context, hash common.Hash) ([][]*types.Log, error) {
|
||||||
receipts := core.GetBlockReceipts(fb.db, hash, core.GetBlockNumber(fb.db, hash))
|
number := rawdb.ReadHeaderNumber(fb.db, hash)
|
||||||
|
if number == nil {
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
receipts := rawdb.ReadReceipts(fb.db, hash, *number)
|
||||||
if receipts == nil {
|
if receipts == nil {
|
||||||
return nil, nil
|
return nil, nil
|
||||||
}
|
}
|
||||||
@@ -445,7 +464,7 @@ func (fb *filterBackend) GetLogs(ctx context.Context, hash common.Hash) ([][]*ty
|
|||||||
return logs, nil
|
return logs, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (fb *filterBackend) SubscribeTxPreEvent(ch chan<- core.TxPreEvent) event.Subscription {
|
func (fb *filterBackend) SubscribeNewTxsEvent(ch chan<- core.NewTxsEvent) event.Subscription {
|
||||||
return event.NewSubscription(func(quit <-chan struct{}) error {
|
return event.NewSubscription(func(quit <-chan struct{}) error {
|
||||||
<-quit
|
<-quit
|
||||||
return nil
|
return nil
|
||||||
|
@@ -58,12 +58,28 @@ var jsonEventPledge = []byte(`{
|
|||||||
"type": "event"
|
"type": "event"
|
||||||
}`)
|
}`)
|
||||||
|
|
||||||
|
var jsonEventMixedCase = []byte(`{
|
||||||
|
"anonymous": false,
|
||||||
|
"inputs": [{
|
||||||
|
"indexed": false, "name": "value", "type": "uint256"
|
||||||
|
}, {
|
||||||
|
"indexed": false, "name": "_value", "type": "uint256"
|
||||||
|
}, {
|
||||||
|
"indexed": false, "name": "Value", "type": "uint256"
|
||||||
|
}],
|
||||||
|
"name": "MixedCase",
|
||||||
|
"type": "event"
|
||||||
|
}`)
|
||||||
|
|
||||||
// 1000000
|
// 1000000
|
||||||
var transferData1 = "00000000000000000000000000000000000000000000000000000000000f4240"
|
var transferData1 = "00000000000000000000000000000000000000000000000000000000000f4240"
|
||||||
|
|
||||||
// "0x00Ce0d46d924CC8437c806721496599FC3FFA268", 2218516807680, "usd"
|
// "0x00Ce0d46d924CC8437c806721496599FC3FFA268", 2218516807680, "usd"
|
||||||
var pledgeData1 = "00000000000000000000000000ce0d46d924cc8437c806721496599fc3ffa2680000000000000000000000000000000000000000000000000000020489e800007573640000000000000000000000000000000000000000000000000000000000"
|
var pledgeData1 = "00000000000000000000000000ce0d46d924cc8437c806721496599fc3ffa2680000000000000000000000000000000000000000000000000000020489e800007573640000000000000000000000000000000000000000000000000000000000"
|
||||||
|
|
||||||
|
// 1000000,2218516807680,1000001
|
||||||
|
var mixedCaseData1 = "00000000000000000000000000000000000000000000000000000000000f42400000000000000000000000000000000000000000000000000000020489e8000000000000000000000000000000000000000000000000000000000000000f4241"
|
||||||
|
|
||||||
func TestEventId(t *testing.T) {
|
func TestEventId(t *testing.T) {
|
||||||
var table = []struct {
|
var table = []struct {
|
||||||
definition string
|
definition string
|
||||||
@@ -121,6 +137,27 @@ func TestEventTupleUnpack(t *testing.T) {
|
|||||||
Value *big.Int
|
Value *big.Int
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type EventTransferWithTag struct {
|
||||||
|
// this is valid because `value` is not exportable,
|
||||||
|
// so value is only unmarshalled into `Value1`.
|
||||||
|
value *big.Int
|
||||||
|
Value1 *big.Int `abi:"value"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BadEventTransferWithSameFieldAndTag struct {
|
||||||
|
Value *big.Int
|
||||||
|
Value1 *big.Int `abi:"value"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BadEventTransferWithDuplicatedTag struct {
|
||||||
|
Value1 *big.Int `abi:"value"`
|
||||||
|
Value2 *big.Int `abi:"value"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type BadEventTransferWithEmptyTag struct {
|
||||||
|
Value *big.Int `abi:""`
|
||||||
|
}
|
||||||
|
|
||||||
type EventPledge struct {
|
type EventPledge struct {
|
||||||
Who common.Address
|
Who common.Address
|
||||||
Wad *big.Int
|
Wad *big.Int
|
||||||
@@ -133,9 +170,16 @@ func TestEventTupleUnpack(t *testing.T) {
|
|||||||
Currency [3]byte
|
Currency [3]byte
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type EventMixedCase struct {
|
||||||
|
Value1 *big.Int `abi:"value"`
|
||||||
|
Value2 *big.Int `abi:"_value"`
|
||||||
|
Value3 *big.Int `abi:"Value"`
|
||||||
|
}
|
||||||
|
|
||||||
bigint := new(big.Int)
|
bigint := new(big.Int)
|
||||||
bigintExpected := big.NewInt(1000000)
|
bigintExpected := big.NewInt(1000000)
|
||||||
bigintExpected2 := big.NewInt(2218516807680)
|
bigintExpected2 := big.NewInt(2218516807680)
|
||||||
|
bigintExpected3 := big.NewInt(1000001)
|
||||||
addr := common.HexToAddress("0x00Ce0d46d924CC8437c806721496599FC3FFA268")
|
addr := common.HexToAddress("0x00Ce0d46d924CC8437c806721496599FC3FFA268")
|
||||||
var testCases = []struct {
|
var testCases = []struct {
|
||||||
data string
|
data string
|
||||||
@@ -158,6 +202,34 @@ func TestEventTupleUnpack(t *testing.T) {
|
|||||||
jsonEventTransfer,
|
jsonEventTransfer,
|
||||||
"",
|
"",
|
||||||
"Can unpack ERC20 Transfer event into slice",
|
"Can unpack ERC20 Transfer event into slice",
|
||||||
|
}, {
|
||||||
|
transferData1,
|
||||||
|
&EventTransferWithTag{},
|
||||||
|
&EventTransferWithTag{Value1: bigintExpected},
|
||||||
|
jsonEventTransfer,
|
||||||
|
"",
|
||||||
|
"Can unpack ERC20 Transfer event into structure with abi: tag",
|
||||||
|
}, {
|
||||||
|
transferData1,
|
||||||
|
&BadEventTransferWithDuplicatedTag{},
|
||||||
|
&BadEventTransferWithDuplicatedTag{},
|
||||||
|
jsonEventTransfer,
|
||||||
|
"struct: abi tag in 'Value2' already mapped",
|
||||||
|
"Can not unpack ERC20 Transfer event with duplicated abi tag",
|
||||||
|
}, {
|
||||||
|
transferData1,
|
||||||
|
&BadEventTransferWithSameFieldAndTag{},
|
||||||
|
&BadEventTransferWithSameFieldAndTag{},
|
||||||
|
jsonEventTransfer,
|
||||||
|
"abi: multiple variables maps to the same abi field 'value'",
|
||||||
|
"Can not unpack ERC20 Transfer event with a field and a tag mapping to the same abi variable",
|
||||||
|
}, {
|
||||||
|
transferData1,
|
||||||
|
&BadEventTransferWithEmptyTag{},
|
||||||
|
&BadEventTransferWithEmptyTag{},
|
||||||
|
jsonEventTransfer,
|
||||||
|
"struct: abi tag in 'Value' is empty",
|
||||||
|
"Can not unpack ERC20 Transfer event with an empty tag",
|
||||||
}, {
|
}, {
|
||||||
pledgeData1,
|
pledgeData1,
|
||||||
&EventPledge{},
|
&EventPledge{},
|
||||||
@@ -216,6 +288,13 @@ func TestEventTupleUnpack(t *testing.T) {
|
|||||||
jsonEventPledge,
|
jsonEventPledge,
|
||||||
"abi: cannot unmarshal tuple into map[string]interface {}",
|
"abi: cannot unmarshal tuple into map[string]interface {}",
|
||||||
"Can not unpack Pledge event into map",
|
"Can not unpack Pledge event into map",
|
||||||
|
}, {
|
||||||
|
mixedCaseData1,
|
||||||
|
&EventMixedCase{},
|
||||||
|
&EventMixedCase{Value1: bigintExpected, Value2: bigintExpected2, Value3: bigintExpected3},
|
||||||
|
jsonEventMixedCase,
|
||||||
|
"",
|
||||||
|
"Can unpack abi variables with mixed case",
|
||||||
}}
|
}}
|
||||||
|
|
||||||
for _, tc := range testCases {
|
for _, tc := range testCases {
|
||||||
@@ -227,7 +306,7 @@ func TestEventTupleUnpack(t *testing.T) {
|
|||||||
assert.Nil(err, "Should be able to unpack event data.")
|
assert.Nil(err, "Should be able to unpack event data.")
|
||||||
assert.Equal(tc.expected, tc.dest, tc.name)
|
assert.Equal(tc.expected, tc.dest, tc.name)
|
||||||
} else {
|
} else {
|
||||||
assert.EqualError(err, tc.error)
|
assert.EqualError(err, tc.error, tc.name)
|
||||||
}
|
}
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
@@ -47,10 +47,8 @@ type Method struct {
|
|||||||
// Please note that "int" is substitute for its canonical representation "int256"
|
// Please note that "int" is substitute for its canonical representation "int256"
|
||||||
func (method Method) Sig() string {
|
func (method Method) Sig() string {
|
||||||
types := make([]string, len(method.Inputs))
|
types := make([]string, len(method.Inputs))
|
||||||
i := 0
|
for i, input := range method.Inputs {
|
||||||
for _, input := range method.Inputs {
|
|
||||||
types[i] = input.Type.String()
|
types[i] = input.Type.String()
|
||||||
i++
|
|
||||||
}
|
}
|
||||||
return fmt.Sprintf("%v(%v)", method.Name, strings.Join(types, ","))
|
return fmt.Sprintf("%v(%v)", method.Name, strings.Join(types, ","))
|
||||||
}
|
}
|
||||||
|
@@ -31,29 +31,14 @@ var (
|
|||||||
uint16T = reflect.TypeOf(uint16(0))
|
uint16T = reflect.TypeOf(uint16(0))
|
||||||
uint32T = reflect.TypeOf(uint32(0))
|
uint32T = reflect.TypeOf(uint32(0))
|
||||||
uint64T = reflect.TypeOf(uint64(0))
|
uint64T = reflect.TypeOf(uint64(0))
|
||||||
intT = reflect.TypeOf(int(0))
|
|
||||||
int8T = reflect.TypeOf(int8(0))
|
int8T = reflect.TypeOf(int8(0))
|
||||||
int16T = reflect.TypeOf(int16(0))
|
int16T = reflect.TypeOf(int16(0))
|
||||||
int32T = reflect.TypeOf(int32(0))
|
int32T = reflect.TypeOf(int32(0))
|
||||||
int64T = reflect.TypeOf(int64(0))
|
int64T = reflect.TypeOf(int64(0))
|
||||||
addressT = reflect.TypeOf(common.Address{})
|
addressT = reflect.TypeOf(common.Address{})
|
||||||
intTS = reflect.TypeOf([]int(nil))
|
|
||||||
int8TS = reflect.TypeOf([]int8(nil))
|
|
||||||
int16TS = reflect.TypeOf([]int16(nil))
|
|
||||||
int32TS = reflect.TypeOf([]int32(nil))
|
|
||||||
int64TS = reflect.TypeOf([]int64(nil))
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// U256 converts a big Int into a 256bit EVM number.
|
// U256 converts a big Int into a 256bit EVM number.
|
||||||
func U256(n *big.Int) []byte {
|
func U256(n *big.Int) []byte {
|
||||||
return math.PaddedBigBytes(math.U256(n), 32)
|
return math.PaddedBigBytes(math.U256(n), 32)
|
||||||
}
|
}
|
||||||
|
|
||||||
// checks whether the given reflect value is signed. This also works for slices with a number type
|
|
||||||
func isSigned(v reflect.Value) bool {
|
|
||||||
switch v.Type() {
|
|
||||||
case intTS, int8TS, int16TS, int32TS, int64TS, intT, int8T, int16T, int32T, int64T:
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
return false
|
|
||||||
}
|
|
||||||
|
@@ -19,7 +19,6 @@ package abi
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"math/big"
|
"math/big"
|
||||||
"reflect"
|
|
||||||
"testing"
|
"testing"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -32,13 +31,3 @@ func TestNumberTypes(t *testing.T) {
|
|||||||
t.Errorf("expected %x got %x", ubytes, unsigned)
|
t.Errorf("expected %x got %x", ubytes, unsigned)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSigned(t *testing.T) {
|
|
||||||
if isSigned(reflect.ValueOf(uint(10))) {
|
|
||||||
t.Error("signed")
|
|
||||||
}
|
|
||||||
|
|
||||||
if !isSigned(reflect.ValueOf(int(10))) {
|
|
||||||
t.Error("not signed")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
@@ -19,6 +19,7 @@ package abi
|
|||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
"reflect"
|
"reflect"
|
||||||
|
"strings"
|
||||||
)
|
)
|
||||||
|
|
||||||
// indirect recursively dereferences the value until it either gets the value
|
// indirect recursively dereferences the value until it either gets the value
|
||||||
@@ -111,18 +112,101 @@ func requireUnpackKind(v reflect.Value, t reflect.Type, k reflect.Kind,
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// requireUniqueStructFieldNames makes sure field names don't collide
|
// mapAbiToStringField maps abi to struct fields.
|
||||||
func requireUniqueStructFieldNames(args Arguments) error {
|
// first round: for each Exportable field that contains a `abi:""` tag
|
||||||
exists := make(map[string]bool)
|
// and this field name exists in the arguments, pair them together.
|
||||||
for _, arg := range args {
|
// second round: for each argument field that has not been already linked,
|
||||||
field := capitalise(arg.Name)
|
// find what variable is expected to be mapped into, if it exists and has not been
|
||||||
if field == "" {
|
// used, pair them.
|
||||||
return fmt.Errorf("abi: purely underscored output cannot unpack to struct")
|
func mapAbiToStructFields(args Arguments, value reflect.Value) (map[string]string, error) {
|
||||||
|
|
||||||
|
typ := value.Type()
|
||||||
|
|
||||||
|
abi2struct := make(map[string]string)
|
||||||
|
struct2abi := make(map[string]string)
|
||||||
|
|
||||||
|
// first round ~~~
|
||||||
|
for i := 0; i < typ.NumField(); i++ {
|
||||||
|
structFieldName := typ.Field(i).Name
|
||||||
|
|
||||||
|
// skip private struct fields.
|
||||||
|
if structFieldName[:1] != strings.ToUpper(structFieldName[:1]) {
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
if exists[field] {
|
|
||||||
return fmt.Errorf("abi: multiple outputs mapping to the same struct field '%s'", field)
|
// skip fields that have no abi:"" tag.
|
||||||
|
var ok bool
|
||||||
|
var tagName string
|
||||||
|
if tagName, ok = typ.Field(i).Tag.Lookup("abi"); !ok {
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
exists[field] = true
|
|
||||||
|
// check if tag is empty.
|
||||||
|
if tagName == "" {
|
||||||
|
return nil, fmt.Errorf("struct: abi tag in '%s' is empty", structFieldName)
|
||||||
|
}
|
||||||
|
|
||||||
|
// check which argument field matches with the abi tag.
|
||||||
|
found := false
|
||||||
|
for _, abiField := range args.NonIndexed() {
|
||||||
|
if abiField.Name == tagName {
|
||||||
|
if abi2struct[abiField.Name] != "" {
|
||||||
|
return nil, fmt.Errorf("struct: abi tag in '%s' already mapped", structFieldName)
|
||||||
|
}
|
||||||
|
// pair them
|
||||||
|
abi2struct[abiField.Name] = structFieldName
|
||||||
|
struct2abi[structFieldName] = abiField.Name
|
||||||
|
found = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// check if this tag has been mapped.
|
||||||
|
if !found {
|
||||||
|
return nil, fmt.Errorf("struct: abi tag '%s' defined but not found in abi", tagName)
|
||||||
|
}
|
||||||
|
|
||||||
}
|
}
|
||||||
return nil
|
|
||||||
|
// second round ~~~
|
||||||
|
for _, arg := range args {
|
||||||
|
|
||||||
|
abiFieldName := arg.Name
|
||||||
|
structFieldName := capitalise(abiFieldName)
|
||||||
|
|
||||||
|
if structFieldName == "" {
|
||||||
|
return nil, fmt.Errorf("abi: purely underscored output cannot unpack to struct")
|
||||||
|
}
|
||||||
|
|
||||||
|
// this abi has already been paired, skip it... unless there exists another, yet unassigned
|
||||||
|
// struct field with the same field name. If so, raise an error:
|
||||||
|
// abi: [ { "name": "value" } ]
|
||||||
|
// struct { Value *big.Int , Value1 *big.Int `abi:"value"`}
|
||||||
|
if abi2struct[abiFieldName] != "" {
|
||||||
|
if abi2struct[abiFieldName] != structFieldName &&
|
||||||
|
struct2abi[structFieldName] == "" &&
|
||||||
|
value.FieldByName(structFieldName).IsValid() {
|
||||||
|
return nil, fmt.Errorf("abi: multiple variables maps to the same abi field '%s'", abiFieldName)
|
||||||
|
}
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
// return an error if this struct field has already been paired.
|
||||||
|
if struct2abi[structFieldName] != "" {
|
||||||
|
return nil, fmt.Errorf("abi: multiple outputs mapping to the same struct field '%s'", structFieldName)
|
||||||
|
}
|
||||||
|
|
||||||
|
if value.FieldByName(structFieldName).IsValid() {
|
||||||
|
// pair them
|
||||||
|
abi2struct[abiFieldName] = structFieldName
|
||||||
|
struct2abi[structFieldName] = abiFieldName
|
||||||
|
} else {
|
||||||
|
// not paired, but annotate as used, to detect cases like
|
||||||
|
// abi : [ { "name": "value" }, { "name": "_value" } ]
|
||||||
|
// struct { Value *big.Int }
|
||||||
|
struct2abi[structFieldName] = abiFieldName
|
||||||
|
}
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
return abi2struct, nil
|
||||||
}
|
}
|
||||||
|
@@ -27,10 +27,10 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
mapset "github.com/deckarep/golang-set"
|
||||||
"github.com/ethereum/go-ethereum/accounts"
|
"github.com/ethereum/go-ethereum/accounts"
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"gopkg.in/fatih/set.v0"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// Minimum amount of time between cache reloads. This limit applies if the platform does
|
// Minimum amount of time between cache reloads. This limit applies if the platform does
|
||||||
@@ -79,7 +79,7 @@ func newAccountCache(keydir string) (*accountCache, chan struct{}) {
|
|||||||
keydir: keydir,
|
keydir: keydir,
|
||||||
byAddr: make(map[common.Address][]accounts.Account),
|
byAddr: make(map[common.Address][]accounts.Account),
|
||||||
notify: make(chan struct{}, 1),
|
notify: make(chan struct{}, 1),
|
||||||
fileC: fileCache{all: set.NewNonTS()},
|
fileC: fileCache{all: mapset.NewThreadUnsafeSet()},
|
||||||
}
|
}
|
||||||
ac.watcher = newWatcher(ac)
|
ac.watcher = newWatcher(ac)
|
||||||
return ac, ac.notify
|
return ac, ac.notify
|
||||||
@@ -237,7 +237,7 @@ func (ac *accountCache) scanAccounts() error {
|
|||||||
log.Debug("Failed to reload keystore contents", "err", err)
|
log.Debug("Failed to reload keystore contents", "err", err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
if creates.Size() == 0 && deletes.Size() == 0 && updates.Size() == 0 {
|
if creates.Cardinality() == 0 && deletes.Cardinality() == 0 && updates.Cardinality() == 0 {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
// Create a helper method to scan the contents of the key files
|
// Create a helper method to scan the contents of the key files
|
||||||
@@ -272,15 +272,15 @@ func (ac *accountCache) scanAccounts() error {
|
|||||||
// Process all the file diffs
|
// Process all the file diffs
|
||||||
start := time.Now()
|
start := time.Now()
|
||||||
|
|
||||||
for _, p := range creates.List() {
|
for _, p := range creates.ToSlice() {
|
||||||
if a := readAccount(p.(string)); a != nil {
|
if a := readAccount(p.(string)); a != nil {
|
||||||
ac.add(*a)
|
ac.add(*a)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
for _, p := range deletes.List() {
|
for _, p := range deletes.ToSlice() {
|
||||||
ac.deleteByFile(p.(string))
|
ac.deleteByFile(p.(string))
|
||||||
}
|
}
|
||||||
for _, p := range updates.List() {
|
for _, p := range updates.ToSlice() {
|
||||||
path := p.(string)
|
path := p.(string)
|
||||||
ac.deleteByFile(path)
|
ac.deleteByFile(path)
|
||||||
if a := readAccount(path); a != nil {
|
if a := readAccount(path); a != nil {
|
||||||
|
@@ -24,20 +24,20 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
mapset "github.com/deckarep/golang-set"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
set "gopkg.in/fatih/set.v0"
|
|
||||||
)
|
)
|
||||||
|
|
||||||
// fileCache is a cache of files seen during scan of keystore.
|
// fileCache is a cache of files seen during scan of keystore.
|
||||||
type fileCache struct {
|
type fileCache struct {
|
||||||
all *set.SetNonTS // Set of all files from the keystore folder
|
all mapset.Set // Set of all files from the keystore folder
|
||||||
lastMod time.Time // Last time instance when a file was modified
|
lastMod time.Time // Last time instance when a file was modified
|
||||||
mu sync.RWMutex
|
mu sync.RWMutex
|
||||||
}
|
}
|
||||||
|
|
||||||
// scan performs a new scan on the given directory, compares against the already
|
// scan performs a new scan on the given directory, compares against the already
|
||||||
// cached filenames, and returns file sets: creates, deletes, updates.
|
// cached filenames, and returns file sets: creates, deletes, updates.
|
||||||
func (fc *fileCache) scan(keyDir string) (set.Interface, set.Interface, set.Interface, error) {
|
func (fc *fileCache) scan(keyDir string) (mapset.Set, mapset.Set, mapset.Set, error) {
|
||||||
t0 := time.Now()
|
t0 := time.Now()
|
||||||
|
|
||||||
// List all the failes from the keystore folder
|
// List all the failes from the keystore folder
|
||||||
@@ -51,8 +51,8 @@ func (fc *fileCache) scan(keyDir string) (set.Interface, set.Interface, set.Inte
|
|||||||
defer fc.mu.Unlock()
|
defer fc.mu.Unlock()
|
||||||
|
|
||||||
// Iterate all the files and gather their metadata
|
// Iterate all the files and gather their metadata
|
||||||
all := set.NewNonTS()
|
all := mapset.NewThreadUnsafeSet()
|
||||||
mods := set.NewNonTS()
|
mods := mapset.NewThreadUnsafeSet()
|
||||||
|
|
||||||
var newLastMod time.Time
|
var newLastMod time.Time
|
||||||
for _, fi := range files {
|
for _, fi := range files {
|
||||||
@@ -76,9 +76,9 @@ func (fc *fileCache) scan(keyDir string) (set.Interface, set.Interface, set.Inte
|
|||||||
t2 := time.Now()
|
t2 := time.Now()
|
||||||
|
|
||||||
// Update the tracked files and return the three sets
|
// Update the tracked files and return the three sets
|
||||||
deletes := set.Difference(fc.all, all) // Deletes = previous - current
|
deletes := fc.all.Difference(all) // Deletes = previous - current
|
||||||
creates := set.Difference(all, fc.all) // Creates = current - previous
|
creates := all.Difference(fc.all) // Creates = current - previous
|
||||||
updates := set.Difference(mods, creates) // Updates = modified - creates
|
updates := mods.Difference(creates) // Updates = modified - creates
|
||||||
|
|
||||||
fc.all, fc.lastMod = all, newLastMod
|
fc.all, fc.lastMod = all, newLastMod
|
||||||
t3 := time.Now()
|
t3 := time.Now()
|
||||||
|
@@ -50,7 +50,7 @@ var (
|
|||||||
var KeyStoreType = reflect.TypeOf(&KeyStore{})
|
var KeyStoreType = reflect.TypeOf(&KeyStore{})
|
||||||
|
|
||||||
// KeyStoreScheme is the protocol scheme prefixing account and wallet URLs.
|
// KeyStoreScheme is the protocol scheme prefixing account and wallet URLs.
|
||||||
var KeyStoreScheme = "keystore"
|
const KeyStoreScheme = "keystore"
|
||||||
|
|
||||||
// Maximum time between wallet refreshes (if filesystem notifications don't work).
|
// Maximum time between wallet refreshes (if filesystem notifications don't work).
|
||||||
const walletRefreshCycle = 3 * time.Second
|
const walletRefreshCycle = 3 * time.Second
|
||||||
|
@@ -28,18 +28,18 @@ package keystore
|
|||||||
import (
|
import (
|
||||||
"bytes"
|
"bytes"
|
||||||
"crypto/aes"
|
"crypto/aes"
|
||||||
crand "crypto/rand"
|
"crypto/rand"
|
||||||
"crypto/sha256"
|
"crypto/sha256"
|
||||||
"encoding/hex"
|
"encoding/hex"
|
||||||
"encoding/json"
|
"encoding/json"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/common/math"
|
"github.com/ethereum/go-ethereum/common/math"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/crypto/randentropy"
|
|
||||||
"github.com/pborman/uuid"
|
"github.com/pborman/uuid"
|
||||||
"golang.org/x/crypto/pbkdf2"
|
"golang.org/x/crypto/pbkdf2"
|
||||||
"golang.org/x/crypto/scrypt"
|
"golang.org/x/crypto/scrypt"
|
||||||
@@ -93,7 +93,7 @@ func (ks keyStorePassphrase) GetKey(addr common.Address, filename, auth string)
|
|||||||
|
|
||||||
// StoreKey generates a key, encrypts with 'auth' and stores in the given directory
|
// StoreKey generates a key, encrypts with 'auth' and stores in the given directory
|
||||||
func StoreKey(dir, auth string, scryptN, scryptP int) (common.Address, error) {
|
func StoreKey(dir, auth string, scryptN, scryptP int) (common.Address, error) {
|
||||||
_, a, err := storeNewKey(&keyStorePassphrase{dir, scryptN, scryptP}, crand.Reader, auth)
|
_, a, err := storeNewKey(&keyStorePassphrase{dir, scryptN, scryptP}, rand.Reader, auth)
|
||||||
return a.Address, err
|
return a.Address, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -108,16 +108,19 @@ func (ks keyStorePassphrase) StoreKey(filename string, key *Key, auth string) er
|
|||||||
func (ks keyStorePassphrase) JoinPath(filename string) string {
|
func (ks keyStorePassphrase) JoinPath(filename string) string {
|
||||||
if filepath.IsAbs(filename) {
|
if filepath.IsAbs(filename) {
|
||||||
return filename
|
return filename
|
||||||
} else {
|
|
||||||
return filepath.Join(ks.keysDirPath, filename)
|
|
||||||
}
|
}
|
||||||
|
return filepath.Join(ks.keysDirPath, filename)
|
||||||
}
|
}
|
||||||
|
|
||||||
// EncryptKey encrypts a key using the specified scrypt parameters into a json
|
// EncryptKey encrypts a key using the specified scrypt parameters into a json
|
||||||
// blob that can be decrypted later on.
|
// blob that can be decrypted later on.
|
||||||
func EncryptKey(key *Key, auth string, scryptN, scryptP int) ([]byte, error) {
|
func EncryptKey(key *Key, auth string, scryptN, scryptP int) ([]byte, error) {
|
||||||
authArray := []byte(auth)
|
authArray := []byte(auth)
|
||||||
salt := randentropy.GetEntropyCSPRNG(32)
|
|
||||||
|
salt := make([]byte, 32)
|
||||||
|
if _, err := io.ReadFull(rand.Reader, salt); err != nil {
|
||||||
|
panic("reading from crypto/rand failed: " + err.Error())
|
||||||
|
}
|
||||||
derivedKey, err := scrypt.Key(authArray, salt, scryptN, scryptR, scryptP, scryptDKLen)
|
derivedKey, err := scrypt.Key(authArray, salt, scryptN, scryptR, scryptP, scryptDKLen)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
@@ -125,7 +128,10 @@ func EncryptKey(key *Key, auth string, scryptN, scryptP int) ([]byte, error) {
|
|||||||
encryptKey := derivedKey[:16]
|
encryptKey := derivedKey[:16]
|
||||||
keyBytes := math.PaddedBigBytes(key.PrivateKey.D, 32)
|
keyBytes := math.PaddedBigBytes(key.PrivateKey.D, 32)
|
||||||
|
|
||||||
iv := randentropy.GetEntropyCSPRNG(aes.BlockSize) // 16
|
iv := make([]byte, aes.BlockSize) // 16
|
||||||
|
if _, err := io.ReadFull(rand.Reader, iv); err != nil {
|
||||||
|
panic("reading from crypto/rand failed: " + err.Error())
|
||||||
|
}
|
||||||
cipherText, err := aesCTRXOR(encryptKey, keyBytes, iv)
|
cipherText, err := aesCTRXOR(encryptKey, keyBytes, iv)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
|
@@ -56,7 +56,6 @@ func (ks keyStorePlain) StoreKey(filename string, key *Key, auth string) error {
|
|||||||
func (ks keyStorePlain) JoinPath(filename string) string {
|
func (ks keyStorePlain) JoinPath(filename string) string {
|
||||||
if filepath.IsAbs(filename) {
|
if filepath.IsAbs(filename) {
|
||||||
return filename
|
return filename
|
||||||
} else {
|
|
||||||
return filepath.Join(ks.keysDirPath, filename)
|
|
||||||
}
|
}
|
||||||
|
return filepath.Join(ks.keysDirPath, filename)
|
||||||
}
|
}
|
||||||
|
@@ -76,12 +76,12 @@ func (u URL) MarshalJSON() ([]byte, error) {
|
|||||||
|
|
||||||
// UnmarshalJSON parses url.
|
// UnmarshalJSON parses url.
|
||||||
func (u *URL) UnmarshalJSON(input []byte) error {
|
func (u *URL) UnmarshalJSON(input []byte) error {
|
||||||
var textUrl string
|
var textURL string
|
||||||
err := json.Unmarshal(input, &textUrl)
|
err := json.Unmarshal(input, &textURL)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
url, err := parseURL(textUrl)
|
url, err := parseURL(textURL)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
96
accounts/url_test.go
Normal file
96
accounts/url_test.go
Normal file
@@ -0,0 +1,96 @@
|
|||||||
|
// Copyright 2017 The go-ethereum Authors
|
||||||
|
// This file is part of the go-ethereum library.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU Lesser General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// The go-ethereum library is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU Lesser General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU Lesser General Public License
|
||||||
|
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package accounts
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestURLParsing(t *testing.T) {
|
||||||
|
url, err := parseURL("https://ethereum.org")
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("unexpected error: %v", err)
|
||||||
|
}
|
||||||
|
if url.Scheme != "https" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "https", url.Scheme)
|
||||||
|
}
|
||||||
|
if url.Path != "ethereum.org" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "ethereum.org", url.Path)
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err = parseURL("ethereum.org")
|
||||||
|
if err == nil {
|
||||||
|
t.Error("expected err, got: nil")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestURLString(t *testing.T) {
|
||||||
|
url := URL{Scheme: "https", Path: "ethereum.org"}
|
||||||
|
if url.String() != "https://ethereum.org" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "https://ethereum.org", url.String())
|
||||||
|
}
|
||||||
|
|
||||||
|
url = URL{Scheme: "", Path: "ethereum.org"}
|
||||||
|
if url.String() != "ethereum.org" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "ethereum.org", url.String())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestURLMarshalJSON(t *testing.T) {
|
||||||
|
url := URL{Scheme: "https", Path: "ethereum.org"}
|
||||||
|
json, err := url.MarshalJSON()
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("unexpcted error: %v", err)
|
||||||
|
}
|
||||||
|
if string(json) != "\"https://ethereum.org\"" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "\"https://ethereum.org\"", string(json))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestURLUnmarshalJSON(t *testing.T) {
|
||||||
|
url := &URL{}
|
||||||
|
err := url.UnmarshalJSON([]byte("\"https://ethereum.org\""))
|
||||||
|
if err != nil {
|
||||||
|
t.Errorf("unexpcted error: %v", err)
|
||||||
|
}
|
||||||
|
if url.Scheme != "https" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "https", url.Scheme)
|
||||||
|
}
|
||||||
|
if url.Path != "ethereum.org" {
|
||||||
|
t.Errorf("expected: %v, got: %v", "https", url.Path)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestURLComparison(t *testing.T) {
|
||||||
|
tests := []struct {
|
||||||
|
urlA URL
|
||||||
|
urlB URL
|
||||||
|
expect int
|
||||||
|
}{
|
||||||
|
{URL{"https", "ethereum.org"}, URL{"https", "ethereum.org"}, 0},
|
||||||
|
{URL{"http", "ethereum.org"}, URL{"https", "ethereum.org"}, -1},
|
||||||
|
{URL{"https", "ethereum.org/a"}, URL{"https", "ethereum.org"}, 1},
|
||||||
|
{URL{"https", "abc.org"}, URL{"https", "ethereum.org"}, -1},
|
||||||
|
}
|
||||||
|
|
||||||
|
for i, tt := range tests {
|
||||||
|
result := tt.urlA.Cmp(tt.urlB)
|
||||||
|
if result != tt.expect {
|
||||||
|
t.Errorf("test %d: cmp mismatch: expected: %d, got: %d", i, tt.expect, result)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
@@ -36,7 +36,7 @@ func Type(msg proto.Message) uint16 {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Name returns the friendly message type name of a specific protocol buffer
|
// Name returns the friendly message type name of a specific protocol buffer
|
||||||
// type numbers.
|
// type number.
|
||||||
func Name(kind uint16) string {
|
func Name(kind uint16) string {
|
||||||
name := MessageType_name[int32(kind)]
|
name := MessageType_name[int32(kind)]
|
||||||
if len(name) < 12 {
|
if len(name) < 12 {
|
||||||
|
@@ -302,7 +302,7 @@ func (w *ledgerDriver) ledgerSign(derivationPath []uint32, tx *types.Transaction
|
|||||||
for i, component := range derivationPath {
|
for i, component := range derivationPath {
|
||||||
binary.BigEndian.PutUint32(path[1+4*i:], component)
|
binary.BigEndian.PutUint32(path[1+4*i:], component)
|
||||||
}
|
}
|
||||||
// Create the transaction RLP based on whether legacy or EIP155 signing was requeste
|
// Create the transaction RLP based on whether legacy or EIP155 signing was requested
|
||||||
var (
|
var (
|
||||||
txrlp []byte
|
txrlp []byte
|
||||||
err error
|
err error
|
||||||
|
@@ -23,8 +23,8 @@ environment:
|
|||||||
install:
|
install:
|
||||||
- git submodule update --init
|
- git submodule update --init
|
||||||
- rmdir C:\go /s /q
|
- rmdir C:\go /s /q
|
||||||
- appveyor DownloadFile https://storage.googleapis.com/golang/go1.10.1.windows-%GETH_ARCH%.zip
|
- appveyor DownloadFile https://storage.googleapis.com/golang/go1.10.3.windows-%GETH_ARCH%.zip
|
||||||
- 7z x go1.10.1.windows-%GETH_ARCH%.zip -y -oC:\ > NUL
|
- 7z x go1.10.3.windows-%GETH_ARCH%.zip -y -oC:\ > NUL
|
||||||
- go version
|
- go version
|
||||||
- gcc --version
|
- gcc --version
|
||||||
|
|
||||||
|
560
bmt/bmt.go
560
bmt/bmt.go
@@ -1,560 +0,0 @@
|
|||||||
// Copyright 2017 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
// Package bmt provides a binary merkle tree implementation
|
|
||||||
package bmt
|
|
||||||
|
|
||||||
import (
|
|
||||||
"fmt"
|
|
||||||
"hash"
|
|
||||||
"io"
|
|
||||||
"strings"
|
|
||||||
"sync"
|
|
||||||
"sync/atomic"
|
|
||||||
)
|
|
||||||
|
|
||||||
/*
|
|
||||||
Binary Merkle Tree Hash is a hash function over arbitrary datachunks of limited size
|
|
||||||
It is defined as the root hash of the binary merkle tree built over fixed size segments
|
|
||||||
of the underlying chunk using any base hash function (e.g keccak 256 SHA3)
|
|
||||||
|
|
||||||
It is used as the chunk hash function in swarm which in turn is the basis for the
|
|
||||||
128 branching swarm hash http://swarm-guide.readthedocs.io/en/latest/architecture.html#swarm-hash
|
|
||||||
|
|
||||||
The BMT is optimal for providing compact inclusion proofs, i.e. prove that a
|
|
||||||
segment is a substring of a chunk starting at a particular offset
|
|
||||||
The size of the underlying segments is fixed at 32 bytes (called the resolution
|
|
||||||
of the BMT hash), the EVM word size to optimize for on-chain BMT verification
|
|
||||||
as well as the hash size optimal for inclusion proofs in the merkle tree of the swarm hash.
|
|
||||||
|
|
||||||
Two implementations are provided:
|
|
||||||
|
|
||||||
* RefHasher is optimized for code simplicity and meant as a reference implementation
|
|
||||||
* Hasher is optimized for speed taking advantage of concurrency with minimalistic
|
|
||||||
control structure to coordinate the concurrent routines
|
|
||||||
It implements the ChunkHash interface as well as the go standard hash.Hash interface
|
|
||||||
|
|
||||||
*/
|
|
||||||
|
|
||||||
const (
|
|
||||||
// DefaultSegmentCount is the maximum number of segments of the underlying chunk
|
|
||||||
DefaultSegmentCount = 128 // Should be equal to storage.DefaultBranches
|
|
||||||
// DefaultPoolSize is the maximum number of bmt trees used by the hashers, i.e,
|
|
||||||
// the maximum number of concurrent BMT hashing operations performed by the same hasher
|
|
||||||
DefaultPoolSize = 8
|
|
||||||
)
|
|
||||||
|
|
||||||
// BaseHasher is a hash.Hash constructor function used for the base hash of the BMT.
|
|
||||||
type BaseHasher func() hash.Hash
|
|
||||||
|
|
||||||
// Hasher a reusable hasher for fixed maximum size chunks representing a BMT
|
|
||||||
// implements the hash.Hash interface
|
|
||||||
// reuse pool of Tree-s for amortised memory allocation and resource control
|
|
||||||
// supports order-agnostic concurrent segment writes
|
|
||||||
// as well as sequential read and write
|
|
||||||
// can not be called concurrently on more than one chunk
|
|
||||||
// can be further appended after Sum
|
|
||||||
// Reset gives back the Tree to the pool and guaranteed to leave
|
|
||||||
// the tree and itself in a state reusable for hashing a new chunk
|
|
||||||
type Hasher struct {
|
|
||||||
pool *TreePool // BMT resource pool
|
|
||||||
bmt *Tree // prebuilt BMT resource for flowcontrol and proofs
|
|
||||||
blocksize int // segment size (size of hash) also for hash.Hash
|
|
||||||
count int // segment count
|
|
||||||
size int // for hash.Hash same as hashsize
|
|
||||||
cur int // cursor position for rightmost currently open chunk
|
|
||||||
segment []byte // the rightmost open segment (not complete)
|
|
||||||
depth int // index of last level
|
|
||||||
result chan []byte // result channel
|
|
||||||
hash []byte // to record the result
|
|
||||||
max int32 // max segments for SegmentWriter interface
|
|
||||||
blockLength []byte // The block length that needes to be added in Sum
|
|
||||||
}
|
|
||||||
|
|
||||||
// New creates a reusable Hasher
|
|
||||||
// implements the hash.Hash interface
|
|
||||||
// pulls a new Tree from a resource pool for hashing each chunk
|
|
||||||
func New(p *TreePool) *Hasher {
|
|
||||||
return &Hasher{
|
|
||||||
pool: p,
|
|
||||||
depth: depth(p.SegmentCount),
|
|
||||||
size: p.SegmentSize,
|
|
||||||
blocksize: p.SegmentSize,
|
|
||||||
count: p.SegmentCount,
|
|
||||||
result: make(chan []byte),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Node is a reuseable segment hasher representing a node in a BMT
|
|
||||||
// it allows for continued writes after a Sum
|
|
||||||
// and is left in completely reusable state after Reset
|
|
||||||
type Node struct {
|
|
||||||
level, index int // position of node for information/logging only
|
|
||||||
initial bool // first and last node
|
|
||||||
root bool // whether the node is root to a smaller BMT
|
|
||||||
isLeft bool // whether it is left side of the parent double segment
|
|
||||||
unbalanced bool // indicates if a node has only the left segment
|
|
||||||
parent *Node // BMT connections
|
|
||||||
state int32 // atomic increment impl concurrent boolean toggle
|
|
||||||
left, right []byte
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewNode constructor for segment hasher nodes in the BMT
|
|
||||||
func NewNode(level, index int, parent *Node) *Node {
|
|
||||||
return &Node{
|
|
||||||
parent: parent,
|
|
||||||
level: level,
|
|
||||||
index: index,
|
|
||||||
initial: index == 0,
|
|
||||||
isLeft: index%2 == 0,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// TreePool provides a pool of Trees used as resources by Hasher
|
|
||||||
// a Tree popped from the pool is guaranteed to have clean state
|
|
||||||
// for hashing a new chunk
|
|
||||||
// Hasher Reset releases the Tree to the pool
|
|
||||||
type TreePool struct {
|
|
||||||
lock sync.Mutex
|
|
||||||
c chan *Tree
|
|
||||||
hasher BaseHasher
|
|
||||||
SegmentSize int
|
|
||||||
SegmentCount int
|
|
||||||
Capacity int
|
|
||||||
count int
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewTreePool creates a Tree pool with hasher, segment size, segment count and capacity
|
|
||||||
// on GetTree it reuses free Trees or creates a new one if size is not reached
|
|
||||||
func NewTreePool(hasher BaseHasher, segmentCount, capacity int) *TreePool {
|
|
||||||
return &TreePool{
|
|
||||||
c: make(chan *Tree, capacity),
|
|
||||||
hasher: hasher,
|
|
||||||
SegmentSize: hasher().Size(),
|
|
||||||
SegmentCount: segmentCount,
|
|
||||||
Capacity: capacity,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Drain drains the pool until it has no more than n resources
|
|
||||||
func (self *TreePool) Drain(n int) {
|
|
||||||
self.lock.Lock()
|
|
||||||
defer self.lock.Unlock()
|
|
||||||
for len(self.c) > n {
|
|
||||||
<-self.c
|
|
||||||
self.count--
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reserve is blocking until it returns an available Tree
|
|
||||||
// it reuses free Trees or creates a new one if size is not reached
|
|
||||||
func (self *TreePool) Reserve() *Tree {
|
|
||||||
self.lock.Lock()
|
|
||||||
defer self.lock.Unlock()
|
|
||||||
var t *Tree
|
|
||||||
if self.count == self.Capacity {
|
|
||||||
return <-self.c
|
|
||||||
}
|
|
||||||
select {
|
|
||||||
case t = <-self.c:
|
|
||||||
default:
|
|
||||||
t = NewTree(self.hasher, self.SegmentSize, self.SegmentCount)
|
|
||||||
self.count++
|
|
||||||
}
|
|
||||||
return t
|
|
||||||
}
|
|
||||||
|
|
||||||
// Release gives back a Tree to the pool.
|
|
||||||
// This Tree is guaranteed to be in reusable state
|
|
||||||
// does not need locking
|
|
||||||
func (self *TreePool) Release(t *Tree) {
|
|
||||||
self.c <- t // can never fail but...
|
|
||||||
}
|
|
||||||
|
|
||||||
// Tree is a reusable control structure representing a BMT
|
|
||||||
// organised in a binary tree
|
|
||||||
// Hasher uses a TreePool to pick one for each chunk hash
|
|
||||||
// the Tree is 'locked' while not in the pool
|
|
||||||
type Tree struct {
|
|
||||||
leaves []*Node
|
|
||||||
}
|
|
||||||
|
|
||||||
// Draw draws the BMT (badly)
|
|
||||||
func (self *Tree) Draw(hash []byte, d int) string {
|
|
||||||
var left, right []string
|
|
||||||
var anc []*Node
|
|
||||||
for i, n := range self.leaves {
|
|
||||||
left = append(left, fmt.Sprintf("%v", hashstr(n.left)))
|
|
||||||
if i%2 == 0 {
|
|
||||||
anc = append(anc, n.parent)
|
|
||||||
}
|
|
||||||
right = append(right, fmt.Sprintf("%v", hashstr(n.right)))
|
|
||||||
}
|
|
||||||
anc = self.leaves
|
|
||||||
var hashes [][]string
|
|
||||||
for l := 0; len(anc) > 0; l++ {
|
|
||||||
var nodes []*Node
|
|
||||||
hash := []string{""}
|
|
||||||
for i, n := range anc {
|
|
||||||
hash = append(hash, fmt.Sprintf("%v|%v", hashstr(n.left), hashstr(n.right)))
|
|
||||||
if i%2 == 0 && n.parent != nil {
|
|
||||||
nodes = append(nodes, n.parent)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
hash = append(hash, "")
|
|
||||||
hashes = append(hashes, hash)
|
|
||||||
anc = nodes
|
|
||||||
}
|
|
||||||
hashes = append(hashes, []string{"", fmt.Sprintf("%v", hashstr(hash)), ""})
|
|
||||||
total := 60
|
|
||||||
del := " "
|
|
||||||
var rows []string
|
|
||||||
for i := len(hashes) - 1; i >= 0; i-- {
|
|
||||||
var textlen int
|
|
||||||
hash := hashes[i]
|
|
||||||
for _, s := range hash {
|
|
||||||
textlen += len(s)
|
|
||||||
}
|
|
||||||
if total < textlen {
|
|
||||||
total = textlen + len(hash)
|
|
||||||
}
|
|
||||||
delsize := (total - textlen) / (len(hash) - 1)
|
|
||||||
if delsize > len(del) {
|
|
||||||
delsize = len(del)
|
|
||||||
}
|
|
||||||
row := fmt.Sprintf("%v: %v", len(hashes)-i-1, strings.Join(hash, del[:delsize]))
|
|
||||||
rows = append(rows, row)
|
|
||||||
|
|
||||||
}
|
|
||||||
rows = append(rows, strings.Join(left, " "))
|
|
||||||
rows = append(rows, strings.Join(right, " "))
|
|
||||||
return strings.Join(rows, "\n") + "\n"
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewTree initialises the Tree by building up the nodes of a BMT
|
|
||||||
// segment size is stipulated to be the size of the hash
|
|
||||||
// segmentCount needs to be positive integer and does not need to be
|
|
||||||
// a power of two and can even be an odd number
|
|
||||||
// segmentSize * segmentCount determines the maximum chunk size
|
|
||||||
// hashed using the tree
|
|
||||||
func NewTree(hasher BaseHasher, segmentSize, segmentCount int) *Tree {
|
|
||||||
n := NewNode(0, 0, nil)
|
|
||||||
n.root = true
|
|
||||||
prevlevel := []*Node{n}
|
|
||||||
// iterate over levels and creates 2^level nodes
|
|
||||||
level := 1
|
|
||||||
count := 2
|
|
||||||
for d := 1; d <= depth(segmentCount); d++ {
|
|
||||||
nodes := make([]*Node, count)
|
|
||||||
for i := 0; i < len(nodes); i++ {
|
|
||||||
parent := prevlevel[i/2]
|
|
||||||
t := NewNode(level, i, parent)
|
|
||||||
nodes[i] = t
|
|
||||||
}
|
|
||||||
prevlevel = nodes
|
|
||||||
level++
|
|
||||||
count *= 2
|
|
||||||
}
|
|
||||||
// the datanode level is the nodes on the last level where
|
|
||||||
return &Tree{
|
|
||||||
leaves: prevlevel,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// methods needed by hash.Hash
|
|
||||||
|
|
||||||
// Size returns the size
|
|
||||||
func (self *Hasher) Size() int {
|
|
||||||
return self.size
|
|
||||||
}
|
|
||||||
|
|
||||||
// BlockSize returns the block size
|
|
||||||
func (self *Hasher) BlockSize() int {
|
|
||||||
return self.blocksize
|
|
||||||
}
|
|
||||||
|
|
||||||
// Sum returns the hash of the buffer
|
|
||||||
// hash.Hash interface Sum method appends the byte slice to the underlying
|
|
||||||
// data before it calculates and returns the hash of the chunk
|
|
||||||
func (self *Hasher) Sum(b []byte) (r []byte) {
|
|
||||||
t := self.bmt
|
|
||||||
i := self.cur
|
|
||||||
n := t.leaves[i]
|
|
||||||
j := i
|
|
||||||
// must run strictly before all nodes calculate
|
|
||||||
// datanodes are guaranteed to have a parent
|
|
||||||
if len(self.segment) > self.size && i > 0 && n.parent != nil {
|
|
||||||
n = n.parent
|
|
||||||
} else {
|
|
||||||
i *= 2
|
|
||||||
}
|
|
||||||
d := self.finalise(n, i)
|
|
||||||
self.writeSegment(j, self.segment, d)
|
|
||||||
c := <-self.result
|
|
||||||
self.releaseTree()
|
|
||||||
|
|
||||||
// sha3(length + BMT(pure_chunk))
|
|
||||||
if self.blockLength == nil {
|
|
||||||
return c
|
|
||||||
}
|
|
||||||
res := self.pool.hasher()
|
|
||||||
res.Reset()
|
|
||||||
res.Write(self.blockLength)
|
|
||||||
res.Write(c)
|
|
||||||
return res.Sum(nil)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hasher implements the SwarmHash interface
|
|
||||||
|
|
||||||
// Hash waits for the hasher result and returns it
|
|
||||||
// caller must call this on a BMT Hasher being written to
|
|
||||||
func (self *Hasher) Hash() []byte {
|
|
||||||
return <-self.result
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hasher implements the io.Writer interface
|
|
||||||
|
|
||||||
// Write fills the buffer to hash
|
|
||||||
// with every full segment complete launches a hasher go routine
|
|
||||||
// that shoots up the BMT
|
|
||||||
func (self *Hasher) Write(b []byte) (int, error) {
|
|
||||||
l := len(b)
|
|
||||||
if l <= 0 {
|
|
||||||
return 0, nil
|
|
||||||
}
|
|
||||||
s := self.segment
|
|
||||||
i := self.cur
|
|
||||||
count := (self.count + 1) / 2
|
|
||||||
need := self.count*self.size - self.cur*2*self.size
|
|
||||||
size := self.size
|
|
||||||
if need > size {
|
|
||||||
size *= 2
|
|
||||||
}
|
|
||||||
if l < need {
|
|
||||||
need = l
|
|
||||||
}
|
|
||||||
// calculate missing bit to complete current open segment
|
|
||||||
rest := size - len(s)
|
|
||||||
if need < rest {
|
|
||||||
rest = need
|
|
||||||
}
|
|
||||||
s = append(s, b[:rest]...)
|
|
||||||
need -= rest
|
|
||||||
// read full segments and the last possibly partial segment
|
|
||||||
for need > 0 && i < count-1 {
|
|
||||||
// push all finished chunks we read
|
|
||||||
self.writeSegment(i, s, self.depth)
|
|
||||||
need -= size
|
|
||||||
if need < 0 {
|
|
||||||
size += need
|
|
||||||
}
|
|
||||||
s = b[rest : rest+size]
|
|
||||||
rest += size
|
|
||||||
i++
|
|
||||||
}
|
|
||||||
self.segment = s
|
|
||||||
self.cur = i
|
|
||||||
// otherwise, we can assume len(s) == 0, so all buffer is read and chunk is not yet full
|
|
||||||
return l, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hasher implements the io.ReaderFrom interface
|
|
||||||
|
|
||||||
// ReadFrom reads from io.Reader and appends to the data to hash using Write
|
|
||||||
// it reads so that chunk to hash is maximum length or reader reaches EOF
|
|
||||||
// caller must Reset the hasher prior to call
|
|
||||||
func (self *Hasher) ReadFrom(r io.Reader) (m int64, err error) {
|
|
||||||
bufsize := self.size*self.count - self.size*self.cur - len(self.segment)
|
|
||||||
buf := make([]byte, bufsize)
|
|
||||||
var read int
|
|
||||||
for {
|
|
||||||
var n int
|
|
||||||
n, err = r.Read(buf)
|
|
||||||
read += n
|
|
||||||
if err == io.EOF || read == len(buf) {
|
|
||||||
hash := self.Sum(buf[:n])
|
|
||||||
if read == len(buf) {
|
|
||||||
err = NewEOC(hash)
|
|
||||||
}
|
|
||||||
break
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
n, err = self.Write(buf[:n])
|
|
||||||
if err != nil {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return int64(read), err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Reset needs to be called before writing to the hasher
|
|
||||||
func (self *Hasher) Reset() {
|
|
||||||
self.getTree()
|
|
||||||
self.blockLength = nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hasher implements the SwarmHash interface
|
|
||||||
|
|
||||||
// ResetWithLength needs to be called before writing to the hasher
|
|
||||||
// the argument is supposed to be the byte slice binary representation of
|
|
||||||
// the length of the data subsumed under the hash
|
|
||||||
func (self *Hasher) ResetWithLength(l []byte) {
|
|
||||||
self.Reset()
|
|
||||||
self.blockLength = l
|
|
||||||
}
|
|
||||||
|
|
||||||
// Release gives back the Tree to the pool whereby it unlocks
|
|
||||||
// it resets tree, segment and index
|
|
||||||
func (self *Hasher) releaseTree() {
|
|
||||||
if self.bmt != nil {
|
|
||||||
n := self.bmt.leaves[self.cur]
|
|
||||||
for ; n != nil; n = n.parent {
|
|
||||||
n.unbalanced = false
|
|
||||||
if n.parent != nil {
|
|
||||||
n.root = false
|
|
||||||
}
|
|
||||||
}
|
|
||||||
self.pool.Release(self.bmt)
|
|
||||||
self.bmt = nil
|
|
||||||
|
|
||||||
}
|
|
||||||
self.cur = 0
|
|
||||||
self.segment = nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (self *Hasher) writeSegment(i int, s []byte, d int) {
|
|
||||||
h := self.pool.hasher()
|
|
||||||
n := self.bmt.leaves[i]
|
|
||||||
|
|
||||||
if len(s) > self.size && n.parent != nil {
|
|
||||||
go func() {
|
|
||||||
h.Reset()
|
|
||||||
h.Write(s)
|
|
||||||
s = h.Sum(nil)
|
|
||||||
|
|
||||||
if n.root {
|
|
||||||
self.result <- s
|
|
||||||
return
|
|
||||||
}
|
|
||||||
self.run(n.parent, h, d, n.index, s)
|
|
||||||
}()
|
|
||||||
return
|
|
||||||
}
|
|
||||||
go self.run(n, h, d, i*2, s)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (self *Hasher) run(n *Node, h hash.Hash, d int, i int, s []byte) {
|
|
||||||
isLeft := i%2 == 0
|
|
||||||
for {
|
|
||||||
if isLeft {
|
|
||||||
n.left = s
|
|
||||||
} else {
|
|
||||||
n.right = s
|
|
||||||
}
|
|
||||||
if !n.unbalanced && n.toggle() {
|
|
||||||
return
|
|
||||||
}
|
|
||||||
if !n.unbalanced || !isLeft || i == 0 && d == 0 {
|
|
||||||
h.Reset()
|
|
||||||
h.Write(n.left)
|
|
||||||
h.Write(n.right)
|
|
||||||
s = h.Sum(nil)
|
|
||||||
|
|
||||||
} else {
|
|
||||||
s = append(n.left, n.right...)
|
|
||||||
}
|
|
||||||
|
|
||||||
self.hash = s
|
|
||||||
if n.root {
|
|
||||||
self.result <- s
|
|
||||||
return
|
|
||||||
}
|
|
||||||
|
|
||||||
isLeft = n.isLeft
|
|
||||||
n = n.parent
|
|
||||||
i++
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// getTree obtains a BMT resource by reserving one from the pool
|
|
||||||
func (self *Hasher) getTree() *Tree {
|
|
||||||
if self.bmt != nil {
|
|
||||||
return self.bmt
|
|
||||||
}
|
|
||||||
t := self.pool.Reserve()
|
|
||||||
self.bmt = t
|
|
||||||
return t
|
|
||||||
}
|
|
||||||
|
|
||||||
// atomic bool toggle implementing a concurrent reusable 2-state object
|
|
||||||
// atomic addint with %2 implements atomic bool toggle
|
|
||||||
// it returns true if the toggler just put it in the active/waiting state
|
|
||||||
func (self *Node) toggle() bool {
|
|
||||||
return atomic.AddInt32(&self.state, 1)%2 == 1
|
|
||||||
}
|
|
||||||
|
|
||||||
func hashstr(b []byte) string {
|
|
||||||
end := len(b)
|
|
||||||
if end > 4 {
|
|
||||||
end = 4
|
|
||||||
}
|
|
||||||
return fmt.Sprintf("%x", b[:end])
|
|
||||||
}
|
|
||||||
|
|
||||||
func depth(n int) (d int) {
|
|
||||||
for l := (n - 1) / 2; l > 0; l /= 2 {
|
|
||||||
d++
|
|
||||||
}
|
|
||||||
return d
|
|
||||||
}
|
|
||||||
|
|
||||||
// finalise is following the zigzags on the tree belonging
|
|
||||||
// to the final datasegment
|
|
||||||
func (self *Hasher) finalise(n *Node, i int) (d int) {
|
|
||||||
isLeft := i%2 == 0
|
|
||||||
for {
|
|
||||||
// when the final segment's path is going via left segments
|
|
||||||
// the incoming data is pushed to the parent upon pulling the left
|
|
||||||
// we do not need toggle the state since this condition is
|
|
||||||
// detectable
|
|
||||||
n.unbalanced = isLeft
|
|
||||||
n.right = nil
|
|
||||||
if n.initial {
|
|
||||||
n.root = true
|
|
||||||
return d
|
|
||||||
}
|
|
||||||
isLeft = n.isLeft
|
|
||||||
n = n.parent
|
|
||||||
d++
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// EOC (end of chunk) implements the error interface
|
|
||||||
type EOC struct {
|
|
||||||
Hash []byte // read the hash of the chunk off the error
|
|
||||||
}
|
|
||||||
|
|
||||||
// Error returns the error string
|
|
||||||
func (self *EOC) Error() string {
|
|
||||||
return fmt.Sprintf("hasher limit reached, chunk hash: %x", self.Hash)
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewEOC creates new end of chunk error with the hash
|
|
||||||
func NewEOC(hash []byte) *EOC {
|
|
||||||
return &EOC{hash}
|
|
||||||
}
|
|
85
bmt/bmt_r.go
85
bmt/bmt_r.go
@@ -1,85 +0,0 @@
|
|||||||
// Copyright 2017 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
// simple nonconcurrent reference implementation for hashsize segment based
|
|
||||||
// Binary Merkle tree hash on arbitrary but fixed maximum chunksize
|
|
||||||
//
|
|
||||||
// This implementation does not take advantage of any paralellisms and uses
|
|
||||||
// far more memory than necessary, but it is easy to see that it is correct.
|
|
||||||
// It can be used for generating test cases for optimized implementations.
|
|
||||||
// see testBMTHasherCorrectness function in bmt_test.go
|
|
||||||
package bmt
|
|
||||||
|
|
||||||
import (
|
|
||||||
"hash"
|
|
||||||
)
|
|
||||||
|
|
||||||
// RefHasher is the non-optimized easy to read reference implementation of BMT
|
|
||||||
type RefHasher struct {
|
|
||||||
span int
|
|
||||||
section int
|
|
||||||
cap int
|
|
||||||
h hash.Hash
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewRefHasher returns a new RefHasher
|
|
||||||
func NewRefHasher(hasher BaseHasher, count int) *RefHasher {
|
|
||||||
h := hasher()
|
|
||||||
hashsize := h.Size()
|
|
||||||
maxsize := hashsize * count
|
|
||||||
c := 2
|
|
||||||
for ; c < count; c *= 2 {
|
|
||||||
}
|
|
||||||
if c > 2 {
|
|
||||||
c /= 2
|
|
||||||
}
|
|
||||||
return &RefHasher{
|
|
||||||
section: 2 * hashsize,
|
|
||||||
span: c * hashsize,
|
|
||||||
cap: maxsize,
|
|
||||||
h: h,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Hash returns the BMT hash of the byte slice
|
|
||||||
// implements the SwarmHash interface
|
|
||||||
func (rh *RefHasher) Hash(d []byte) []byte {
|
|
||||||
if len(d) > rh.cap {
|
|
||||||
d = d[:rh.cap]
|
|
||||||
}
|
|
||||||
|
|
||||||
return rh.hash(d, rh.span)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (rh *RefHasher) hash(d []byte, s int) []byte {
|
|
||||||
l := len(d)
|
|
||||||
left := d
|
|
||||||
var right []byte
|
|
||||||
if l > rh.section {
|
|
||||||
for ; s >= l; s /= 2 {
|
|
||||||
}
|
|
||||||
left = rh.hash(d[:s], s)
|
|
||||||
right = d[s:]
|
|
||||||
if l-s > rh.section/2 {
|
|
||||||
right = rh.hash(right, s)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
defer rh.h.Reset()
|
|
||||||
rh.h.Write(left)
|
|
||||||
rh.h.Write(right)
|
|
||||||
h := rh.h.Sum(nil)
|
|
||||||
return h
|
|
||||||
}
|
|
481
bmt/bmt_test.go
481
bmt/bmt_test.go
@@ -1,481 +0,0 @@
|
|||||||
// Copyright 2017 The go-ethereum Authors
|
|
||||||
// This file is part of the go-ethereum library.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is free software: you can redistribute it and/or modify
|
|
||||||
// it under the terms of the GNU Lesser General Public License as published by
|
|
||||||
// the Free Software Foundation, either version 3 of the License, or
|
|
||||||
// (at your option) any later version.
|
|
||||||
//
|
|
||||||
// The go-ethereum library is distributed in the hope that it will be useful,
|
|
||||||
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
||||||
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
||||||
// GNU Lesser General Public License for more details.
|
|
||||||
//
|
|
||||||
// You should have received a copy of the GNU Lesser General Public License
|
|
||||||
// along with the go-ethereum library. If not, see <http://www.gnu.org/licenses/>.
|
|
||||||
|
|
||||||
package bmt
|
|
||||||
|
|
||||||
import (
|
|
||||||
"bytes"
|
|
||||||
crand "crypto/rand"
|
|
||||||
"fmt"
|
|
||||||
"hash"
|
|
||||||
"io"
|
|
||||||
"math/rand"
|
|
||||||
"sync"
|
|
||||||
"sync/atomic"
|
|
||||||
"testing"
|
|
||||||
"time"
|
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/crypto/sha3"
|
|
||||||
)
|
|
||||||
|
|
||||||
const (
|
|
||||||
maxproccnt = 8
|
|
||||||
)
|
|
||||||
|
|
||||||
// TestRefHasher tests that the RefHasher computes the expected BMT hash for
|
|
||||||
// all data lengths between 0 and 256 bytes
|
|
||||||
func TestRefHasher(t *testing.T) {
|
|
||||||
hashFunc := sha3.NewKeccak256
|
|
||||||
|
|
||||||
sha3 := func(data ...[]byte) []byte {
|
|
||||||
h := hashFunc()
|
|
||||||
for _, v := range data {
|
|
||||||
h.Write(v)
|
|
||||||
}
|
|
||||||
return h.Sum(nil)
|
|
||||||
}
|
|
||||||
|
|
||||||
// the test struct is used to specify the expected BMT hash for data
|
|
||||||
// lengths between "from" and "to"
|
|
||||||
type test struct {
|
|
||||||
from int64
|
|
||||||
to int64
|
|
||||||
expected func([]byte) []byte
|
|
||||||
}
|
|
||||||
|
|
||||||
var tests []*test
|
|
||||||
|
|
||||||
// all lengths in [0,64] should be:
|
|
||||||
//
|
|
||||||
// sha3(data)
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 0,
|
|
||||||
to: 64,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(data)
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [65,96] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// data[64:]
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 65,
|
|
||||||
to: 96,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(data[:64]), data[64:])
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [97,128] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// sha3(data[64:])
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 97,
|
|
||||||
to: 128,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(data[:64]), sha3(data[64:]))
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [129,160] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// sha3(data[64:128])
|
|
||||||
// )
|
|
||||||
// data[128:]
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 129,
|
|
||||||
to: 160,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(sha3(data[:64]), sha3(data[64:128])), data[128:])
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [161,192] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// sha3(data[64:128])
|
|
||||||
// )
|
|
||||||
// sha3(data[128:])
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 161,
|
|
||||||
to: 192,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(sha3(data[:64]), sha3(data[64:128])), sha3(data[128:]))
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [193,224] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// sha3(data[64:128])
|
|
||||||
// )
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[128:192])
|
|
||||||
// data[192:]
|
|
||||||
// )
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 193,
|
|
||||||
to: 224,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(sha3(data[:64]), sha3(data[64:128])), sha3(sha3(data[128:192]), data[192:]))
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// all lengths in [225,256] should be:
|
|
||||||
//
|
|
||||||
// sha3(
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[:64])
|
|
||||||
// sha3(data[64:128])
|
|
||||||
// )
|
|
||||||
// sha3(
|
|
||||||
// sha3(data[128:192])
|
|
||||||
// sha3(data[192:])
|
|
||||||
// )
|
|
||||||
// )
|
|
||||||
//
|
|
||||||
tests = append(tests, &test{
|
|
||||||
from: 225,
|
|
||||||
to: 256,
|
|
||||||
expected: func(data []byte) []byte {
|
|
||||||
return sha3(sha3(sha3(data[:64]), sha3(data[64:128])), sha3(sha3(data[128:192]), sha3(data[192:])))
|
|
||||||
},
|
|
||||||
})
|
|
||||||
|
|
||||||
// run the tests
|
|
||||||
for _, x := range tests {
|
|
||||||
for length := x.from; length <= x.to; length++ {
|
|
||||||
t.Run(fmt.Sprintf("%d_bytes", length), func(t *testing.T) {
|
|
||||||
data := make([]byte, length)
|
|
||||||
if _, err := io.ReadFull(crand.Reader, data); err != nil && err != io.EOF {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
expected := x.expected(data)
|
|
||||||
actual := NewRefHasher(hashFunc, 128).Hash(data)
|
|
||||||
if !bytes.Equal(actual, expected) {
|
|
||||||
t.Fatalf("expected %x, got %x", expected, actual)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func testDataReader(l int) (r io.Reader) {
|
|
||||||
return io.LimitReader(crand.Reader, int64(l))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHasherCorrectness(t *testing.T) {
|
|
||||||
err := testHasher(testBaseHasher)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func testHasher(f func(BaseHasher, []byte, int, int) error) error {
|
|
||||||
tdata := testDataReader(4128)
|
|
||||||
data := make([]byte, 4128)
|
|
||||||
tdata.Read(data)
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
size := hasher().Size()
|
|
||||||
counts := []int{1, 2, 3, 4, 5, 8, 16, 32, 64, 128}
|
|
||||||
|
|
||||||
var err error
|
|
||||||
for _, count := range counts {
|
|
||||||
max := count * size
|
|
||||||
incr := 1
|
|
||||||
for n := 0; n <= max+incr; n += incr {
|
|
||||||
err = f(hasher, data, n, count)
|
|
||||||
if err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHasherReuseWithoutRelease(t *testing.T) {
|
|
||||||
testHasherReuse(1, t)
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHasherReuseWithRelease(t *testing.T) {
|
|
||||||
testHasherReuse(maxproccnt, t)
|
|
||||||
}
|
|
||||||
|
|
||||||
func testHasherReuse(i int, t *testing.T) {
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
pool := NewTreePool(hasher, 128, i)
|
|
||||||
defer pool.Drain(0)
|
|
||||||
bmt := New(pool)
|
|
||||||
|
|
||||||
for i := 0; i < 500; i++ {
|
|
||||||
n := rand.Intn(4096)
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata.Read(data)
|
|
||||||
|
|
||||||
err := testHasherCorrectness(bmt, hasher, data, n, 128)
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestHasherConcurrency(t *testing.T) {
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
pool := NewTreePool(hasher, 128, maxproccnt)
|
|
||||||
defer pool.Drain(0)
|
|
||||||
wg := sync.WaitGroup{}
|
|
||||||
cycles := 100
|
|
||||||
wg.Add(maxproccnt * cycles)
|
|
||||||
errc := make(chan error)
|
|
||||||
|
|
||||||
for p := 0; p < maxproccnt; p++ {
|
|
||||||
for i := 0; i < cycles; i++ {
|
|
||||||
go func() {
|
|
||||||
bmt := New(pool)
|
|
||||||
n := rand.Intn(4096)
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata.Read(data)
|
|
||||||
err := testHasherCorrectness(bmt, hasher, data, n, 128)
|
|
||||||
wg.Done()
|
|
||||||
if err != nil {
|
|
||||||
errc <- err
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
go func() {
|
|
||||||
wg.Wait()
|
|
||||||
close(errc)
|
|
||||||
}()
|
|
||||||
var err error
|
|
||||||
select {
|
|
||||||
case <-time.NewTimer(5 * time.Second).C:
|
|
||||||
err = fmt.Errorf("timed out")
|
|
||||||
case err = <-errc:
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func testBaseHasher(hasher BaseHasher, d []byte, n, count int) error {
|
|
||||||
pool := NewTreePool(hasher, count, 1)
|
|
||||||
defer pool.Drain(0)
|
|
||||||
bmt := New(pool)
|
|
||||||
return testHasherCorrectness(bmt, hasher, d, n, count)
|
|
||||||
}
|
|
||||||
|
|
||||||
func testHasherCorrectness(bmt hash.Hash, hasher BaseHasher, d []byte, n, count int) (err error) {
|
|
||||||
data := d[:n]
|
|
||||||
rbmt := NewRefHasher(hasher, count)
|
|
||||||
exp := rbmt.Hash(data)
|
|
||||||
timeout := time.NewTimer(time.Second)
|
|
||||||
c := make(chan error)
|
|
||||||
|
|
||||||
go func() {
|
|
||||||
bmt.Reset()
|
|
||||||
bmt.Write(data)
|
|
||||||
got := bmt.Sum(nil)
|
|
||||||
if !bytes.Equal(got, exp) {
|
|
||||||
c <- fmt.Errorf("wrong hash: expected %x, got %x", exp, got)
|
|
||||||
}
|
|
||||||
close(c)
|
|
||||||
}()
|
|
||||||
select {
|
|
||||||
case <-timeout.C:
|
|
||||||
err = fmt.Errorf("BMT hash calculation timed out")
|
|
||||||
case err = <-c:
|
|
||||||
}
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
func BenchmarkSHA3_4k(t *testing.B) { benchmarkSHA3(4096, t) }
|
|
||||||
func BenchmarkSHA3_2k(t *testing.B) { benchmarkSHA3(4096/2, t) }
|
|
||||||
func BenchmarkSHA3_1k(t *testing.B) { benchmarkSHA3(4096/4, t) }
|
|
||||||
func BenchmarkSHA3_512b(t *testing.B) { benchmarkSHA3(4096/8, t) }
|
|
||||||
func BenchmarkSHA3_256b(t *testing.B) { benchmarkSHA3(4096/16, t) }
|
|
||||||
func BenchmarkSHA3_128b(t *testing.B) { benchmarkSHA3(4096/32, t) }
|
|
||||||
|
|
||||||
func BenchmarkBMTBaseline_4k(t *testing.B) { benchmarkBMTBaseline(4096, t) }
|
|
||||||
func BenchmarkBMTBaseline_2k(t *testing.B) { benchmarkBMTBaseline(4096/2, t) }
|
|
||||||
func BenchmarkBMTBaseline_1k(t *testing.B) { benchmarkBMTBaseline(4096/4, t) }
|
|
||||||
func BenchmarkBMTBaseline_512b(t *testing.B) { benchmarkBMTBaseline(4096/8, t) }
|
|
||||||
func BenchmarkBMTBaseline_256b(t *testing.B) { benchmarkBMTBaseline(4096/16, t) }
|
|
||||||
func BenchmarkBMTBaseline_128b(t *testing.B) { benchmarkBMTBaseline(4096/32, t) }
|
|
||||||
|
|
||||||
func BenchmarkRefHasher_4k(t *testing.B) { benchmarkRefHasher(4096, t) }
|
|
||||||
func BenchmarkRefHasher_2k(t *testing.B) { benchmarkRefHasher(4096/2, t) }
|
|
||||||
func BenchmarkRefHasher_1k(t *testing.B) { benchmarkRefHasher(4096/4, t) }
|
|
||||||
func BenchmarkRefHasher_512b(t *testing.B) { benchmarkRefHasher(4096/8, t) }
|
|
||||||
func BenchmarkRefHasher_256b(t *testing.B) { benchmarkRefHasher(4096/16, t) }
|
|
||||||
func BenchmarkRefHasher_128b(t *testing.B) { benchmarkRefHasher(4096/32, t) }
|
|
||||||
|
|
||||||
func BenchmarkHasher_4k(t *testing.B) { benchmarkHasher(4096, t) }
|
|
||||||
func BenchmarkHasher_2k(t *testing.B) { benchmarkHasher(4096/2, t) }
|
|
||||||
func BenchmarkHasher_1k(t *testing.B) { benchmarkHasher(4096/4, t) }
|
|
||||||
func BenchmarkHasher_512b(t *testing.B) { benchmarkHasher(4096/8, t) }
|
|
||||||
func BenchmarkHasher_256b(t *testing.B) { benchmarkHasher(4096/16, t) }
|
|
||||||
func BenchmarkHasher_128b(t *testing.B) { benchmarkHasher(4096/32, t) }
|
|
||||||
|
|
||||||
func BenchmarkHasherNoReuse_4k(t *testing.B) { benchmarkHasherReuse(1, 4096, t) }
|
|
||||||
func BenchmarkHasherNoReuse_2k(t *testing.B) { benchmarkHasherReuse(1, 4096/2, t) }
|
|
||||||
func BenchmarkHasherNoReuse_1k(t *testing.B) { benchmarkHasherReuse(1, 4096/4, t) }
|
|
||||||
func BenchmarkHasherNoReuse_512b(t *testing.B) { benchmarkHasherReuse(1, 4096/8, t) }
|
|
||||||
func BenchmarkHasherNoReuse_256b(t *testing.B) { benchmarkHasherReuse(1, 4096/16, t) }
|
|
||||||
func BenchmarkHasherNoReuse_128b(t *testing.B) { benchmarkHasherReuse(1, 4096/32, t) }
|
|
||||||
|
|
||||||
func BenchmarkHasherReuse_4k(t *testing.B) { benchmarkHasherReuse(16, 4096, t) }
|
|
||||||
func BenchmarkHasherReuse_2k(t *testing.B) { benchmarkHasherReuse(16, 4096/2, t) }
|
|
||||||
func BenchmarkHasherReuse_1k(t *testing.B) { benchmarkHasherReuse(16, 4096/4, t) }
|
|
||||||
func BenchmarkHasherReuse_512b(t *testing.B) { benchmarkHasherReuse(16, 4096/8, t) }
|
|
||||||
func BenchmarkHasherReuse_256b(t *testing.B) { benchmarkHasherReuse(16, 4096/16, t) }
|
|
||||||
func BenchmarkHasherReuse_128b(t *testing.B) { benchmarkHasherReuse(16, 4096/32, t) }
|
|
||||||
|
|
||||||
// benchmarks the minimum hashing time for a balanced (for simplicity) BMT
|
|
||||||
// by doing count/segmentsize parallel hashings of 2*segmentsize bytes
|
|
||||||
// doing it on n maxproccnt each reusing the base hasher
|
|
||||||
// the premise is that this is the minimum computation needed for a BMT
|
|
||||||
// therefore this serves as a theoretical optimum for concurrent implementations
|
|
||||||
func benchmarkBMTBaseline(n int, t *testing.B) {
|
|
||||||
tdata := testDataReader(64)
|
|
||||||
data := make([]byte, 64)
|
|
||||||
tdata.Read(data)
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
|
|
||||||
t.ReportAllocs()
|
|
||||||
t.ResetTimer()
|
|
||||||
for i := 0; i < t.N; i++ {
|
|
||||||
count := int32((n-1)/hasher().Size() + 1)
|
|
||||||
wg := sync.WaitGroup{}
|
|
||||||
wg.Add(maxproccnt)
|
|
||||||
var i int32
|
|
||||||
for j := 0; j < maxproccnt; j++ {
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
h := hasher()
|
|
||||||
for atomic.AddInt32(&i, 1) < count {
|
|
||||||
h.Reset()
|
|
||||||
h.Write(data)
|
|
||||||
h.Sum(nil)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func benchmarkHasher(n int, t *testing.B) {
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata.Read(data)
|
|
||||||
|
|
||||||
size := 1
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
segmentCount := 128
|
|
||||||
pool := NewTreePool(hasher, segmentCount, size)
|
|
||||||
bmt := New(pool)
|
|
||||||
|
|
||||||
t.ReportAllocs()
|
|
||||||
t.ResetTimer()
|
|
||||||
for i := 0; i < t.N; i++ {
|
|
||||||
bmt.Reset()
|
|
||||||
bmt.Write(data)
|
|
||||||
bmt.Sum(nil)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func benchmarkHasherReuse(poolsize, n int, t *testing.B) {
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata.Read(data)
|
|
||||||
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
segmentCount := 128
|
|
||||||
pool := NewTreePool(hasher, segmentCount, poolsize)
|
|
||||||
cycles := 200
|
|
||||||
|
|
||||||
t.ReportAllocs()
|
|
||||||
t.ResetTimer()
|
|
||||||
for i := 0; i < t.N; i++ {
|
|
||||||
wg := sync.WaitGroup{}
|
|
||||||
wg.Add(cycles)
|
|
||||||
for j := 0; j < cycles; j++ {
|
|
||||||
bmt := New(pool)
|
|
||||||
go func() {
|
|
||||||
defer wg.Done()
|
|
||||||
bmt.Reset()
|
|
||||||
bmt.Write(data)
|
|
||||||
bmt.Sum(nil)
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
wg.Wait()
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func benchmarkSHA3(n int, t *testing.B) {
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
tdata.Read(data)
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
h := hasher()
|
|
||||||
|
|
||||||
t.ReportAllocs()
|
|
||||||
t.ResetTimer()
|
|
||||||
for i := 0; i < t.N; i++ {
|
|
||||||
h.Reset()
|
|
||||||
h.Write(data)
|
|
||||||
h.Sum(nil)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func benchmarkRefHasher(n int, t *testing.B) {
|
|
||||||
data := make([]byte, n)
|
|
||||||
tdata := testDataReader(n)
|
|
||||||
tdata.Read(data)
|
|
||||||
hasher := sha3.NewKeccak256
|
|
||||||
rbmt := NewRefHasher(hasher, 128)
|
|
||||||
|
|
||||||
t.ReportAllocs()
|
|
||||||
t.ResetTimer()
|
|
||||||
for i := 0; i < t.N; i++ {
|
|
||||||
rbmt.Hash(data)
|
|
||||||
}
|
|
||||||
}
|
|
232
build/ci.go
232
build/ci.go
@@ -26,7 +26,7 @@ Available commands are:
|
|||||||
install [ -arch architecture ] [ -cc compiler ] [ packages... ] -- builds packages and executables
|
install [ -arch architecture ] [ -cc compiler ] [ packages... ] -- builds packages and executables
|
||||||
test [ -coverage ] [ packages... ] -- runs the tests
|
test [ -coverage ] [ packages... ] -- runs the tests
|
||||||
lint -- runs certain pre-selected linters
|
lint -- runs certain pre-selected linters
|
||||||
archive [ -arch architecture ] [ -type zip|tar ] [ -signer key-envvar ] [ -upload dest ] -- archives build artefacts
|
archive [ -arch architecture ] [ -type zip|tar ] [ -signer key-envvar ] [ -upload dest ] -- archives build artifacts
|
||||||
importkeys -- imports signing keys from env
|
importkeys -- imports signing keys from env
|
||||||
debsrc [ -signer key-id ] [ -upload dest ] -- creates a debian source package
|
debsrc [ -signer key-id ] [ -upload dest ] -- creates a debian source package
|
||||||
nsis -- creates a Windows NSIS installer
|
nsis -- creates a Windows NSIS installer
|
||||||
@@ -59,6 +59,8 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/internal/build"
|
"github.com/ethereum/go-ethereum/internal/build"
|
||||||
|
"github.com/ethereum/go-ethereum/params"
|
||||||
|
sv "github.com/ethereum/go-ethereum/swarm/version"
|
||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
@@ -77,52 +79,81 @@ var (
|
|||||||
executablePath("geth"),
|
executablePath("geth"),
|
||||||
executablePath("puppeth"),
|
executablePath("puppeth"),
|
||||||
executablePath("rlpdump"),
|
executablePath("rlpdump"),
|
||||||
executablePath("swarm"),
|
|
||||||
executablePath("wnode"),
|
executablePath("wnode"),
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Files that end up in the swarm*.zip archive.
|
||||||
|
swarmArchiveFiles = []string{
|
||||||
|
"COPYING",
|
||||||
|
executablePath("swarm"),
|
||||||
|
}
|
||||||
|
|
||||||
// A debian package is created for all executables listed here.
|
// A debian package is created for all executables listed here.
|
||||||
debExecutables = []debExecutable{
|
debExecutables = []debExecutable{
|
||||||
{
|
{
|
||||||
Name: "abigen",
|
BinaryName: "abigen",
|
||||||
Description: "Source code generator to convert Ethereum contract definitions into easy to use, compile-time type-safe Go packages.",
|
Description: "Source code generator to convert Ethereum contract definitions into easy to use, compile-time type-safe Go packages.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "bootnode",
|
BinaryName: "bootnode",
|
||||||
Description: "Ethereum bootnode.",
|
Description: "Ethereum bootnode.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "evm",
|
BinaryName: "evm",
|
||||||
Description: "Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode.",
|
Description: "Developer utility version of the EVM (Ethereum Virtual Machine) that is capable of running bytecode snippets within a configurable environment and execution mode.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "geth",
|
BinaryName: "geth",
|
||||||
Description: "Ethereum CLI client.",
|
Description: "Ethereum CLI client.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "puppeth",
|
BinaryName: "puppeth",
|
||||||
Description: "Ethereum private network manager.",
|
Description: "Ethereum private network manager.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "rlpdump",
|
BinaryName: "rlpdump",
|
||||||
Description: "Developer utility tool that prints RLP structures.",
|
Description: "Developer utility tool that prints RLP structures.",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "swarm",
|
BinaryName: "wnode",
|
||||||
Description: "Ethereum Swarm daemon and tools",
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "wnode",
|
|
||||||
Description: "Ethereum Whisper diagnostic tool",
|
Description: "Ethereum Whisper diagnostic tool",
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// A debian package is created for all executables listed here.
|
||||||
|
debSwarmExecutables = []debExecutable{
|
||||||
|
{
|
||||||
|
BinaryName: "swarm",
|
||||||
|
PackageName: "ethereum-swarm",
|
||||||
|
Description: "Ethereum Swarm daemon and tools",
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
debEthereum = debPackage{
|
||||||
|
Name: "ethereum",
|
||||||
|
Version: params.Version,
|
||||||
|
Executables: debExecutables,
|
||||||
|
}
|
||||||
|
|
||||||
|
debSwarm = debPackage{
|
||||||
|
Name: "ethereum-swarm",
|
||||||
|
Version: sv.Version,
|
||||||
|
Executables: debSwarmExecutables,
|
||||||
|
}
|
||||||
|
|
||||||
|
// Debian meta packages to build and push to Ubuntu PPA
|
||||||
|
debPackages = []debPackage{
|
||||||
|
debSwarm,
|
||||||
|
debEthereum,
|
||||||
|
}
|
||||||
|
|
||||||
// Distros for which packages are created.
|
// Distros for which packages are created.
|
||||||
// Note: vivid is unsupported because there is no golang-1.6 package for it.
|
// Note: vivid is unsupported because there is no golang-1.6 package for it.
|
||||||
// Note: wily is unsupported because it was officially deprecated on lanchpad.
|
// Note: wily is unsupported because it was officially deprecated on lanchpad.
|
||||||
// Note: yakkety is unsupported because it was officially deprecated on lanchpad.
|
// Note: yakkety is unsupported because it was officially deprecated on lanchpad.
|
||||||
// Note: zesty is unsupported because it was officially deprecated on lanchpad.
|
// Note: zesty is unsupported because it was officially deprecated on lanchpad.
|
||||||
debDistros = []string{"trusty", "xenial", "artful", "bionic"}
|
// Note: artful is unsupported because it was officially deprecated on lanchpad.
|
||||||
|
debDistros = []string{"trusty", "xenial", "bionic", "cosmic"}
|
||||||
)
|
)
|
||||||
|
|
||||||
var GOBIN, _ = filepath.Abs(filepath.Join("build", "bin"))
|
var GOBIN, _ = filepath.Abs(filepath.Join("build", "bin"))
|
||||||
@@ -330,6 +361,7 @@ func doLint(cmdline []string) {
|
|||||||
configs := []string{
|
configs := []string{
|
||||||
"--vendor",
|
"--vendor",
|
||||||
"--tests",
|
"--tests",
|
||||||
|
"--deadline=2m",
|
||||||
"--disable-all",
|
"--disable-all",
|
||||||
"--enable=goimports",
|
"--enable=goimports",
|
||||||
"--enable=varcheck",
|
"--enable=varcheck",
|
||||||
@@ -349,7 +381,6 @@ func doLint(cmdline []string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Release Packaging
|
// Release Packaging
|
||||||
|
|
||||||
func doArchive(cmdline []string) {
|
func doArchive(cmdline []string) {
|
||||||
var (
|
var (
|
||||||
arch = flag.String("arch", runtime.GOARCH, "Architecture cross packaging")
|
arch = flag.String("arch", runtime.GOARCH, "Architecture cross packaging")
|
||||||
@@ -369,10 +400,14 @@ func doArchive(cmdline []string) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
var (
|
var (
|
||||||
env = build.Env()
|
env = build.Env()
|
||||||
base = archiveBasename(*arch, env)
|
|
||||||
geth = "geth-" + base + ext
|
basegeth = archiveBasename(*arch, params.ArchiveVersion(env.Commit))
|
||||||
alltools = "geth-alltools-" + base + ext
|
geth = "geth-" + basegeth + ext
|
||||||
|
alltools = "geth-alltools-" + basegeth + ext
|
||||||
|
|
||||||
|
baseswarm = archiveBasename(*arch, sv.ArchiveVersion(env.Commit))
|
||||||
|
swarm = "swarm-" + baseswarm + ext
|
||||||
)
|
)
|
||||||
maybeSkipArchive(env)
|
maybeSkipArchive(env)
|
||||||
if err := build.WriteArchive(geth, gethArchiveFiles); err != nil {
|
if err := build.WriteArchive(geth, gethArchiveFiles); err != nil {
|
||||||
@@ -381,14 +416,17 @@ func doArchive(cmdline []string) {
|
|||||||
if err := build.WriteArchive(alltools, allToolsArchiveFiles); err != nil {
|
if err := build.WriteArchive(alltools, allToolsArchiveFiles); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
for _, archive := range []string{geth, alltools} {
|
if err := build.WriteArchive(swarm, swarmArchiveFiles); err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
for _, archive := range []string{geth, alltools, swarm} {
|
||||||
if err := archiveUpload(archive, *upload, *signer); err != nil {
|
if err := archiveUpload(archive, *upload, *signer); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func archiveBasename(arch string, env build.Environment) string {
|
func archiveBasename(arch string, archiveVersion string) string {
|
||||||
platform := runtime.GOOS + "-" + arch
|
platform := runtime.GOOS + "-" + arch
|
||||||
if arch == "arm" {
|
if arch == "arm" {
|
||||||
platform += os.Getenv("GOARM")
|
platform += os.Getenv("GOARM")
|
||||||
@@ -399,18 +437,7 @@ func archiveBasename(arch string, env build.Environment) string {
|
|||||||
if arch == "ios" {
|
if arch == "ios" {
|
||||||
platform = "ios-all"
|
platform = "ios-all"
|
||||||
}
|
}
|
||||||
return platform + "-" + archiveVersion(env)
|
return platform + "-" + archiveVersion
|
||||||
}
|
|
||||||
|
|
||||||
func archiveVersion(env build.Environment) string {
|
|
||||||
version := build.VERSION()
|
|
||||||
if isUnstableBuild(env) {
|
|
||||||
version += "-unstable"
|
|
||||||
}
|
|
||||||
if env.Commit != "" {
|
|
||||||
version += "-" + env.Commit[:8]
|
|
||||||
}
|
|
||||||
return version
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func archiveUpload(archive string, blobstore string, signer string) error {
|
func archiveUpload(archive string, blobstore string, signer string) error {
|
||||||
@@ -460,7 +487,6 @@ func maybeSkipArchive(env build.Environment) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Debian Packaging
|
// Debian Packaging
|
||||||
|
|
||||||
func doDebianSource(cmdline []string) {
|
func doDebianSource(cmdline []string) {
|
||||||
var (
|
var (
|
||||||
signer = flag.String("signer", "", `Signing key name, also used as package author`)
|
signer = flag.String("signer", "", `Signing key name, also used as package author`)
|
||||||
@@ -484,21 +510,23 @@ func doDebianSource(cmdline []string) {
|
|||||||
build.MustRun(gpg)
|
build.MustRun(gpg)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Create the packages.
|
// Create Debian packages and upload them
|
||||||
for _, distro := range debDistros {
|
for _, pkg := range debPackages {
|
||||||
meta := newDebMetadata(distro, *signer, env, now)
|
for _, distro := range debDistros {
|
||||||
pkgdir := stageDebianSource(*workdir, meta)
|
meta := newDebMetadata(distro, *signer, env, now, pkg.Name, pkg.Version, pkg.Executables)
|
||||||
debuild := exec.Command("debuild", "-S", "-sa", "-us", "-uc")
|
pkgdir := stageDebianSource(*workdir, meta)
|
||||||
debuild.Dir = pkgdir
|
debuild := exec.Command("debuild", "-S", "-sa", "-us", "-uc")
|
||||||
build.MustRun(debuild)
|
debuild.Dir = pkgdir
|
||||||
|
build.MustRun(debuild)
|
||||||
|
|
||||||
changes := fmt.Sprintf("%s_%s_source.changes", meta.Name(), meta.VersionString())
|
changes := fmt.Sprintf("%s_%s_source.changes", meta.Name(), meta.VersionString())
|
||||||
changes = filepath.Join(*workdir, changes)
|
changes = filepath.Join(*workdir, changes)
|
||||||
if *signer != "" {
|
if *signer != "" {
|
||||||
build.MustRunCommand("debsign", changes)
|
build.MustRunCommand("debsign", changes)
|
||||||
}
|
}
|
||||||
if *upload != "" {
|
if *upload != "" {
|
||||||
build.MustRunCommand("dput", *upload, changes)
|
build.MustRunCommand("dput", *upload, changes)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -523,9 +551,17 @@ func isUnstableBuild(env build.Environment) bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
type debPackage struct {
|
||||||
|
Name string // the name of the Debian package to produce, e.g. "ethereum", or "ethereum-swarm"
|
||||||
|
Version string // the clean version of the debPackage, e.g. 1.8.12 or 0.3.0, without any metadata
|
||||||
|
Executables []debExecutable // executables to be included in the package
|
||||||
|
}
|
||||||
|
|
||||||
type debMetadata struct {
|
type debMetadata struct {
|
||||||
Env build.Environment
|
Env build.Environment
|
||||||
|
|
||||||
|
PackageName string
|
||||||
|
|
||||||
// go-ethereum version being built. Note that this
|
// go-ethereum version being built. Note that this
|
||||||
// is not the debian package version. The package version
|
// is not the debian package version. The package version
|
||||||
// is constructed by VersionString.
|
// is constructed by VersionString.
|
||||||
@@ -537,21 +573,33 @@ type debMetadata struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type debExecutable struct {
|
type debExecutable struct {
|
||||||
Name, Description string
|
PackageName string
|
||||||
|
BinaryName string
|
||||||
|
Description string
|
||||||
}
|
}
|
||||||
|
|
||||||
func newDebMetadata(distro, author string, env build.Environment, t time.Time) debMetadata {
|
// Package returns the name of the package if present, or
|
||||||
|
// fallbacks to BinaryName
|
||||||
|
func (d debExecutable) Package() string {
|
||||||
|
if d.PackageName != "" {
|
||||||
|
return d.PackageName
|
||||||
|
}
|
||||||
|
return d.BinaryName
|
||||||
|
}
|
||||||
|
|
||||||
|
func newDebMetadata(distro, author string, env build.Environment, t time.Time, name string, version string, exes []debExecutable) debMetadata {
|
||||||
if author == "" {
|
if author == "" {
|
||||||
// No signing key, use default author.
|
// No signing key, use default author.
|
||||||
author = "Ethereum Builds <fjl@ethereum.org>"
|
author = "Ethereum Builds <fjl@ethereum.org>"
|
||||||
}
|
}
|
||||||
return debMetadata{
|
return debMetadata{
|
||||||
|
PackageName: name,
|
||||||
Env: env,
|
Env: env,
|
||||||
Author: author,
|
Author: author,
|
||||||
Distro: distro,
|
Distro: distro,
|
||||||
Version: build.VERSION(),
|
Version: version,
|
||||||
Time: t.Format(time.RFC1123Z),
|
Time: t.Format(time.RFC1123Z),
|
||||||
Executables: debExecutables,
|
Executables: exes,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -559,9 +607,9 @@ func newDebMetadata(distro, author string, env build.Environment, t time.Time) d
|
|||||||
// on all executable packages.
|
// on all executable packages.
|
||||||
func (meta debMetadata) Name() string {
|
func (meta debMetadata) Name() string {
|
||||||
if isUnstableBuild(meta.Env) {
|
if isUnstableBuild(meta.Env) {
|
||||||
return "ethereum-unstable"
|
return meta.PackageName + "-unstable"
|
||||||
}
|
}
|
||||||
return "ethereum"
|
return meta.PackageName
|
||||||
}
|
}
|
||||||
|
|
||||||
// VersionString returns the debian version of the packages.
|
// VersionString returns the debian version of the packages.
|
||||||
@@ -588,9 +636,20 @@ func (meta debMetadata) ExeList() string {
|
|||||||
// ExeName returns the package name of an executable package.
|
// ExeName returns the package name of an executable package.
|
||||||
func (meta debMetadata) ExeName(exe debExecutable) string {
|
func (meta debMetadata) ExeName(exe debExecutable) string {
|
||||||
if isUnstableBuild(meta.Env) {
|
if isUnstableBuild(meta.Env) {
|
||||||
return exe.Name + "-unstable"
|
return exe.Package() + "-unstable"
|
||||||
}
|
}
|
||||||
return exe.Name
|
return exe.Package()
|
||||||
|
}
|
||||||
|
|
||||||
|
// EthereumSwarmPackageName returns the name of the swarm package based on
|
||||||
|
// environment, e.g. "ethereum-swarm-unstable", or "ethereum-swarm".
|
||||||
|
// This is needed so that we make sure that "ethereum" package,
|
||||||
|
// depends on and installs "ethereum-swarm"
|
||||||
|
func (meta debMetadata) EthereumSwarmPackageName() string {
|
||||||
|
if isUnstableBuild(meta.Env) {
|
||||||
|
return debSwarm.Name + "-unstable"
|
||||||
|
}
|
||||||
|
return debSwarm.Name
|
||||||
}
|
}
|
||||||
|
|
||||||
// ExeConflicts returns the content of the Conflicts field
|
// ExeConflicts returns the content of the Conflicts field
|
||||||
@@ -605,7 +664,7 @@ func (meta debMetadata) ExeConflicts(exe debExecutable) string {
|
|||||||
// be preferred and the conflicting files should be handled via
|
// be preferred and the conflicting files should be handled via
|
||||||
// alternates. We might do this eventually but using a conflict is
|
// alternates. We might do this eventually but using a conflict is
|
||||||
// easier now.
|
// easier now.
|
||||||
return "ethereum, " + exe.Name
|
return "ethereum, " + exe.Package()
|
||||||
}
|
}
|
||||||
return ""
|
return ""
|
||||||
}
|
}
|
||||||
@@ -622,24 +681,23 @@ func stageDebianSource(tmpdir string, meta debMetadata) (pkgdir string) {
|
|||||||
|
|
||||||
// Put the debian build files in place.
|
// Put the debian build files in place.
|
||||||
debian := filepath.Join(pkgdir, "debian")
|
debian := filepath.Join(pkgdir, "debian")
|
||||||
build.Render("build/deb.rules", filepath.Join(debian, "rules"), 0755, meta)
|
build.Render("build/deb/"+meta.PackageName+"/deb.rules", filepath.Join(debian, "rules"), 0755, meta)
|
||||||
build.Render("build/deb.changelog", filepath.Join(debian, "changelog"), 0644, meta)
|
build.Render("build/deb/"+meta.PackageName+"/deb.changelog", filepath.Join(debian, "changelog"), 0644, meta)
|
||||||
build.Render("build/deb.control", filepath.Join(debian, "control"), 0644, meta)
|
build.Render("build/deb/"+meta.PackageName+"/deb.control", filepath.Join(debian, "control"), 0644, meta)
|
||||||
build.Render("build/deb.copyright", filepath.Join(debian, "copyright"), 0644, meta)
|
build.Render("build/deb/"+meta.PackageName+"/deb.copyright", filepath.Join(debian, "copyright"), 0644, meta)
|
||||||
build.RenderString("8\n", filepath.Join(debian, "compat"), 0644, meta)
|
build.RenderString("8\n", filepath.Join(debian, "compat"), 0644, meta)
|
||||||
build.RenderString("3.0 (native)\n", filepath.Join(debian, "source/format"), 0644, meta)
|
build.RenderString("3.0 (native)\n", filepath.Join(debian, "source/format"), 0644, meta)
|
||||||
for _, exe := range meta.Executables {
|
for _, exe := range meta.Executables {
|
||||||
install := filepath.Join(debian, meta.ExeName(exe)+".install")
|
install := filepath.Join(debian, meta.ExeName(exe)+".install")
|
||||||
docs := filepath.Join(debian, meta.ExeName(exe)+".docs")
|
docs := filepath.Join(debian, meta.ExeName(exe)+".docs")
|
||||||
build.Render("build/deb.install", install, 0644, exe)
|
build.Render("build/deb/"+meta.PackageName+"/deb.install", install, 0644, exe)
|
||||||
build.Render("build/deb.docs", docs, 0644, exe)
|
build.Render("build/deb/"+meta.PackageName+"/deb.docs", docs, 0644, exe)
|
||||||
}
|
}
|
||||||
|
|
||||||
return pkgdir
|
return pkgdir
|
||||||
}
|
}
|
||||||
|
|
||||||
// Windows installer
|
// Windows installer
|
||||||
|
|
||||||
func doWindowsInstaller(cmdline []string) {
|
func doWindowsInstaller(cmdline []string) {
|
||||||
// Parse the flags and make skip installer generation on PRs
|
// Parse the flags and make skip installer generation on PRs
|
||||||
var (
|
var (
|
||||||
@@ -689,11 +747,11 @@ func doWindowsInstaller(cmdline []string) {
|
|||||||
// Build the installer. This assumes that all the needed files have been previously
|
// Build the installer. This assumes that all the needed files have been previously
|
||||||
// built (don't mix building and packaging to keep cross compilation complexity to a
|
// built (don't mix building and packaging to keep cross compilation complexity to a
|
||||||
// minimum).
|
// minimum).
|
||||||
version := strings.Split(build.VERSION(), ".")
|
version := strings.Split(params.Version, ".")
|
||||||
if env.Commit != "" {
|
if env.Commit != "" {
|
||||||
version[2] += "-" + env.Commit[:8]
|
version[2] += "-" + env.Commit[:8]
|
||||||
}
|
}
|
||||||
installer, _ := filepath.Abs("geth-" + archiveBasename(*arch, env) + ".exe")
|
installer, _ := filepath.Abs("geth-" + archiveBasename(*arch, params.ArchiveVersion(env.Commit)) + ".exe")
|
||||||
build.MustRunCommand("makensis.exe",
|
build.MustRunCommand("makensis.exe",
|
||||||
"/DOUTPUTFILE="+installer,
|
"/DOUTPUTFILE="+installer,
|
||||||
"/DMAJORVERSION="+version[0],
|
"/DMAJORVERSION="+version[0],
|
||||||
@@ -745,7 +803,7 @@ func doAndroidArchive(cmdline []string) {
|
|||||||
maybeSkipArchive(env)
|
maybeSkipArchive(env)
|
||||||
|
|
||||||
// Sign and upload the archive to Azure
|
// Sign and upload the archive to Azure
|
||||||
archive := "geth-" + archiveBasename("android", env) + ".aar"
|
archive := "geth-" + archiveBasename("android", params.ArchiveVersion(env.Commit)) + ".aar"
|
||||||
os.Rename("geth.aar", archive)
|
os.Rename("geth.aar", archive)
|
||||||
|
|
||||||
if err := archiveUpload(archive, *upload, *signer); err != nil {
|
if err := archiveUpload(archive, *upload, *signer); err != nil {
|
||||||
@@ -755,14 +813,18 @@ func doAndroidArchive(cmdline []string) {
|
|||||||
os.Rename(archive, meta.Package+".aar")
|
os.Rename(archive, meta.Package+".aar")
|
||||||
if *signer != "" && *deploy != "" {
|
if *signer != "" && *deploy != "" {
|
||||||
// Import the signing key into the local GPG instance
|
// Import the signing key into the local GPG instance
|
||||||
if b64key := os.Getenv(*signer); b64key != "" {
|
b64key := os.Getenv(*signer)
|
||||||
key, err := base64.StdEncoding.DecodeString(b64key)
|
key, err := base64.StdEncoding.DecodeString(b64key)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Fatalf("invalid base64 %s", *signer)
|
log.Fatalf("invalid base64 %s", *signer)
|
||||||
}
|
}
|
||||||
gpg := exec.Command("gpg", "--import")
|
gpg := exec.Command("gpg", "--import")
|
||||||
gpg.Stdin = bytes.NewReader(key)
|
gpg.Stdin = bytes.NewReader(key)
|
||||||
build.MustRun(gpg)
|
build.MustRun(gpg)
|
||||||
|
|
||||||
|
keyID, err := build.PGPKeyID(string(key))
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
// Upload the artifacts to Sonatype and/or Maven Central
|
// Upload the artifacts to Sonatype and/or Maven Central
|
||||||
repo := *deploy + "/service/local/staging/deploy/maven2"
|
repo := *deploy + "/service/local/staging/deploy/maven2"
|
||||||
@@ -771,6 +833,7 @@ func doAndroidArchive(cmdline []string) {
|
|||||||
}
|
}
|
||||||
build.MustRunCommand("mvn", "gpg:sign-and-deploy-file", "-e", "-X",
|
build.MustRunCommand("mvn", "gpg:sign-and-deploy-file", "-e", "-X",
|
||||||
"-settings=build/mvn.settings", "-Durl="+repo, "-DrepositoryId=ossrh",
|
"-settings=build/mvn.settings", "-Durl="+repo, "-DrepositoryId=ossrh",
|
||||||
|
"-Dgpg.keyname="+keyID,
|
||||||
"-DpomFile="+meta.Package+".pom", "-Dfile="+meta.Package+".aar")
|
"-DpomFile="+meta.Package+".pom", "-Dfile="+meta.Package+".aar")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -825,7 +888,7 @@ func newMavenMetadata(env build.Environment) mavenMetadata {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Render the version and package strings
|
// Render the version and package strings
|
||||||
version := build.VERSION()
|
version := params.Version
|
||||||
if isUnstableBuild(env) {
|
if isUnstableBuild(env) {
|
||||||
version += "-SNAPSHOT"
|
version += "-SNAPSHOT"
|
||||||
}
|
}
|
||||||
@@ -860,7 +923,7 @@ func doXCodeFramework(cmdline []string) {
|
|||||||
build.MustRun(bind)
|
build.MustRun(bind)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
archive := "geth-" + archiveBasename("ios", env)
|
archive := "geth-" + archiveBasename("ios", params.ArchiveVersion(env.Commit))
|
||||||
if err := os.Mkdir(archive, os.ModePerm); err != nil {
|
if err := os.Mkdir(archive, os.ModePerm); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
@@ -916,7 +979,7 @@ func newPodMetadata(env build.Environment, archive string) podMetadata {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
version := build.VERSION()
|
version := params.Version
|
||||||
if isUnstableBuild(env) {
|
if isUnstableBuild(env) {
|
||||||
version += "-unstable." + env.Buildnum
|
version += "-unstable." + env.Buildnum
|
||||||
}
|
}
|
||||||
@@ -1012,23 +1075,14 @@ func doPurge(cmdline []string) {
|
|||||||
}
|
}
|
||||||
for i := 0; i < len(blobs); i++ {
|
for i := 0; i < len(blobs); i++ {
|
||||||
for j := i + 1; j < len(blobs); j++ {
|
for j := i + 1; j < len(blobs); j++ {
|
||||||
iTime, err := time.Parse(time.RFC1123, blobs[i].Properties.LastModified)
|
if blobs[i].Properties.LastModified.After(blobs[j].Properties.LastModified) {
|
||||||
if err != nil {
|
|
||||||
log.Fatal(err)
|
|
||||||
}
|
|
||||||
jTime, err := time.Parse(time.RFC1123, blobs[j].Properties.LastModified)
|
|
||||||
if err != nil {
|
|
||||||
log.Fatal(err)
|
|
||||||
}
|
|
||||||
if iTime.After(jTime) {
|
|
||||||
blobs[i], blobs[j] = blobs[j], blobs[i]
|
blobs[i], blobs[j] = blobs[j], blobs[i]
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Filter out all archives more recent that the given threshold
|
// Filter out all archives more recent that the given threshold
|
||||||
for i, blob := range blobs {
|
for i, blob := range blobs {
|
||||||
timestamp, _ := time.Parse(time.RFC1123, blob.Properties.LastModified)
|
if time.Since(blob.Properties.LastModified) < time.Duration(*limit)*24*time.Hour {
|
||||||
if time.Since(timestamp) < time.Duration(*limit)*24*time.Hour {
|
|
||||||
blobs = blobs[:i]
|
blobs = blobs[:i]
|
||||||
break
|
break
|
||||||
}
|
}
|
||||||
|
19
build/clean_go_build_cache.sh
Executable file
19
build/clean_go_build_cache.sh
Executable file
@@ -0,0 +1,19 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
# Cleaning the Go cache only makes sense if we actually have Go installed... or
|
||||||
|
# if Go is actually callable. This does not hold true during deb packaging, so
|
||||||
|
# we need an explicit check to avoid build failures.
|
||||||
|
if ! command -v go > /dev/null; then
|
||||||
|
exit
|
||||||
|
fi
|
||||||
|
|
||||||
|
version_gt() {
|
||||||
|
test "$(printf '%s\n' "$@" | sort -V | head -n 1)" != "$1"
|
||||||
|
}
|
||||||
|
|
||||||
|
golang_version=$(go version |cut -d' ' -f3 |sed 's/go//')
|
||||||
|
|
||||||
|
# Clean go build cache when go version is greater than or equal to 1.10
|
||||||
|
if !(version_gt 1.10 $golang_version); then
|
||||||
|
go clean -cache
|
||||||
|
fi
|
@@ -1 +0,0 @@
|
|||||||
build/bin/{{.Name}} usr/bin
|
|
19
build/deb/ethereum-swarm/deb.control
Normal file
19
build/deb/ethereum-swarm/deb.control
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
Source: {{.Name}}
|
||||||
|
Section: science
|
||||||
|
Priority: extra
|
||||||
|
Maintainer: {{.Author}}
|
||||||
|
Build-Depends: debhelper (>= 8.0.0), golang-1.10
|
||||||
|
Standards-Version: 3.9.5
|
||||||
|
Homepage: https://ethereum.org
|
||||||
|
Vcs-Git: git://github.com/ethereum/go-ethereum.git
|
||||||
|
Vcs-Browser: https://github.com/ethereum/go-ethereum
|
||||||
|
|
||||||
|
{{range .Executables}}
|
||||||
|
Package: {{$.ExeName .}}
|
||||||
|
Conflicts: {{$.ExeConflicts .}}
|
||||||
|
Architecture: any
|
||||||
|
Depends: ${shlibs:Depends}, ${misc:Depends}
|
||||||
|
Built-Using: ${misc:Built-Using}
|
||||||
|
Description: {{.Description}}
|
||||||
|
{{.Description}}
|
||||||
|
{{end}}
|
@@ -1,4 +1,4 @@
|
|||||||
Copyright 2016 The go-ethereum Authors
|
Copyright 2018 The go-ethereum Authors
|
||||||
|
|
||||||
go-ethereum is free software: you can redistribute it and/or modify
|
go-ethereum is free software: you can redistribute it and/or modify
|
||||||
it under the terms of the GNU General Public License as published by
|
it under the terms of the GNU General Public License as published by
|
1
build/deb/ethereum-swarm/deb.install
Normal file
1
build/deb/ethereum-swarm/deb.install
Normal file
@@ -0,0 +1 @@
|
|||||||
|
build/bin/{{.BinaryName}} usr/bin
|
5
build/deb/ethereum/deb.changelog
Normal file
5
build/deb/ethereum/deb.changelog
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
{{.Name}} ({{.VersionString}}) {{.Distro}}; urgency=low
|
||||||
|
|
||||||
|
* git build of {{.Env.Commit}}
|
||||||
|
|
||||||
|
-- {{.Author}} {{.Time}}
|
@@ -10,9 +10,9 @@ Vcs-Browser: https://github.com/ethereum/go-ethereum
|
|||||||
|
|
||||||
Package: {{.Name}}
|
Package: {{.Name}}
|
||||||
Architecture: any
|
Architecture: any
|
||||||
Depends: ${misc:Depends}, {{.ExeList}}
|
Depends: ${misc:Depends}, {{.EthereumSwarmPackageName}}, {{.ExeList}}
|
||||||
Description: Meta-package to install geth and other tools
|
Description: Meta-package to install geth, swarm, and other tools
|
||||||
Meta-package to install geth and other tools
|
Meta-package to install geth, swarm and other tools
|
||||||
|
|
||||||
{{range .Executables}}
|
{{range .Executables}}
|
||||||
Package: {{$.ExeName .}}
|
Package: {{$.ExeName .}}
|
14
build/deb/ethereum/deb.copyright
Normal file
14
build/deb/ethereum/deb.copyright
Normal file
@@ -0,0 +1,14 @@
|
|||||||
|
Copyright 2018 The go-ethereum Authors
|
||||||
|
|
||||||
|
go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
it under the terms of the GNU General Public License as published by
|
||||||
|
the Free Software Foundation, either version 3 of the License, or
|
||||||
|
(at your option) any later version.
|
||||||
|
|
||||||
|
go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
GNU General Public License for more details.
|
||||||
|
|
||||||
|
You should have received a copy of the GNU General Public License
|
||||||
|
along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
1
build/deb/ethereum/deb.docs
Normal file
1
build/deb/ethereum/deb.docs
Normal file
@@ -0,0 +1 @@
|
|||||||
|
AUTHORS
|
1
build/deb/ethereum/deb.install
Normal file
1
build/deb/ethereum/deb.install
Normal file
@@ -0,0 +1 @@
|
|||||||
|
build/bin/{{.BinaryName}} usr/bin
|
13
build/deb/ethereum/deb.rules
Normal file
13
build/deb/ethereum/deb.rules
Normal file
@@ -0,0 +1,13 @@
|
|||||||
|
#!/usr/bin/make -f
|
||||||
|
# -*- makefile -*-
|
||||||
|
|
||||||
|
# Uncomment this to turn on verbose mode.
|
||||||
|
#export DH_VERBOSE=1
|
||||||
|
|
||||||
|
override_dh_auto_build:
|
||||||
|
build/env.sh /usr/lib/go-1.10/bin/go run build/ci.go install -git-commit={{.Env.Commit}} -git-branch={{.Env.Branch}} -git-tag={{.Env.Tag}} -buildnum={{.Env.Buildnum}} -pull-request={{.Env.IsPullRequest}}
|
||||||
|
|
||||||
|
override_dh_auto_test:
|
||||||
|
|
||||||
|
%:
|
||||||
|
dh $@
|
@@ -1,18 +1,18 @@
|
|||||||
#!/usr/bin/env bash
|
#!/bin/sh
|
||||||
|
|
||||||
find_files() {
|
find_files() {
|
||||||
find . -not \( \
|
find . ! \( \
|
||||||
\( \
|
\( \
|
||||||
-wholename '.github' \
|
-path '.github' \
|
||||||
-o -wholename './build/_workspace' \
|
-o -path './build/_workspace' \
|
||||||
-o -wholename './build/bin' \
|
-o -path './build/bin' \
|
||||||
-o -wholename './crypto/bn256' \
|
-o -path './crypto/bn256' \
|
||||||
-o -wholename '*/vendor/*' \
|
-o -path '*/vendor/*' \
|
||||||
\) -prune \
|
\) -prune \
|
||||||
\) -name '*.go'
|
\) -name '*.go'
|
||||||
}
|
}
|
||||||
|
|
||||||
GOFMT="gofmt -s -w";
|
GOFMT="gofmt -s -w"
|
||||||
GOIMPORTS="goimports -w";
|
GOIMPORTS="goimports -w"
|
||||||
find_files | xargs $GOFMT;
|
find_files | xargs $GOFMT
|
||||||
find_files | xargs $GOIMPORTS;
|
find_files | xargs $GOIMPORTS
|
||||||
|
@@ -29,7 +29,7 @@ import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
var (
|
var (
|
||||||
abiFlag = flag.String("abi", "", "Path to the Ethereum contract ABI json to bind")
|
abiFlag = flag.String("abi", "", "Path to the Ethereum contract ABI json to bind, - for STDIN")
|
||||||
binFlag = flag.String("bin", "", "Path to the Ethereum contract bytecode (generate deploy method)")
|
binFlag = flag.String("bin", "", "Path to the Ethereum contract bytecode (generate deploy method)")
|
||||||
typFlag = flag.String("type", "", "Struct name for the binding (default = package name)")
|
typFlag = flag.String("type", "", "Struct name for the binding (default = package name)")
|
||||||
|
|
||||||
@@ -75,16 +75,27 @@ func main() {
|
|||||||
bins []string
|
bins []string
|
||||||
types []string
|
types []string
|
||||||
)
|
)
|
||||||
if *solFlag != "" {
|
if *solFlag != "" || *abiFlag == "-" {
|
||||||
// Generate the list of types to exclude from binding
|
// Generate the list of types to exclude from binding
|
||||||
exclude := make(map[string]bool)
|
exclude := make(map[string]bool)
|
||||||
for _, kind := range strings.Split(*excFlag, ",") {
|
for _, kind := range strings.Split(*excFlag, ",") {
|
||||||
exclude[strings.ToLower(kind)] = true
|
exclude[strings.ToLower(kind)] = true
|
||||||
}
|
}
|
||||||
contracts, err := compiler.CompileSolidity(*solcFlag, *solFlag)
|
|
||||||
if err != nil {
|
var contracts map[string]*compiler.Contract
|
||||||
fmt.Printf("Failed to build Solidity contract: %v\n", err)
|
var err error
|
||||||
os.Exit(-1)
|
if *solFlag != "" {
|
||||||
|
contracts, err = compiler.CompileSolidity(*solcFlag, *solFlag)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Failed to build Solidity contract: %v\n", err)
|
||||||
|
os.Exit(-1)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
contracts, err = contractsFromStdin()
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Failed to read input ABIs from STDIN: %v\n", err)
|
||||||
|
os.Exit(-1)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
// Gather all non-excluded contract for binding
|
// Gather all non-excluded contract for binding
|
||||||
for name, contract := range contracts {
|
for name, contract := range contracts {
|
||||||
@@ -138,3 +149,12 @@ func main() {
|
|||||||
os.Exit(-1)
|
os.Exit(-1)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func contractsFromStdin() (map[string]*compiler.Contract, error) {
|
||||||
|
bytes, err := ioutil.ReadAll(os.Stdin)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
return compiler.ParseCombinedJSON(bytes, "", "", "", "")
|
||||||
|
}
|
||||||
|
@@ -47,11 +47,11 @@ import (
|
|||||||
"gopkg.in/urfave/cli.v1"
|
"gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
// ExternalApiVersion -- see extapi_changelog.md
|
// ExternalAPIVersion -- see extapi_changelog.md
|
||||||
const ExternalApiVersion = "2.0.0"
|
const ExternalAPIVersion = "2.0.0"
|
||||||
|
|
||||||
// InternalApiVersion -- see intapi_changelog.md
|
// InternalAPIVersion -- see intapi_changelog.md
|
||||||
const InternalApiVersion = "2.0.0"
|
const InternalAPIVersion = "2.0.0"
|
||||||
|
|
||||||
const legalWarning = `
|
const legalWarning = `
|
||||||
WARNING!
|
WARNING!
|
||||||
@@ -398,10 +398,10 @@ func signer(c *cli.Context) error {
|
|||||||
}
|
}
|
||||||
// register signer API with server
|
// register signer API with server
|
||||||
var (
|
var (
|
||||||
extapiUrl = "n/a"
|
extapiURL = "n/a"
|
||||||
ipcApiUrl = "n/a"
|
ipcapiURL = "n/a"
|
||||||
)
|
)
|
||||||
rpcApi := []rpc.API{
|
rpcAPI := []rpc.API{
|
||||||
{
|
{
|
||||||
Namespace: "account",
|
Namespace: "account",
|
||||||
Public: true,
|
Public: true,
|
||||||
@@ -415,12 +415,12 @@ func signer(c *cli.Context) error {
|
|||||||
|
|
||||||
// start http server
|
// start http server
|
||||||
httpEndpoint := fmt.Sprintf("%s:%d", c.String(utils.RPCListenAddrFlag.Name), c.Int(rpcPortFlag.Name))
|
httpEndpoint := fmt.Sprintf("%s:%d", c.String(utils.RPCListenAddrFlag.Name), c.Int(rpcPortFlag.Name))
|
||||||
listener, _, err := rpc.StartHTTPEndpoint(httpEndpoint, rpcApi, []string{"account"}, cors, vhosts)
|
listener, _, err := rpc.StartHTTPEndpoint(httpEndpoint, rpcAPI, []string{"account"}, cors, vhosts, rpc.DefaultHTTPTimeouts)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Could not start RPC api: %v", err)
|
utils.Fatalf("Could not start RPC api: %v", err)
|
||||||
}
|
}
|
||||||
extapiUrl = fmt.Sprintf("http://%s", httpEndpoint)
|
extapiURL = fmt.Sprintf("http://%s", httpEndpoint)
|
||||||
log.Info("HTTP endpoint opened", "url", extapiUrl)
|
log.Info("HTTP endpoint opened", "url", extapiURL)
|
||||||
|
|
||||||
defer func() {
|
defer func() {
|
||||||
listener.Close()
|
listener.Close()
|
||||||
@@ -430,19 +430,19 @@ func signer(c *cli.Context) error {
|
|||||||
}
|
}
|
||||||
if !c.Bool(utils.IPCDisabledFlag.Name) {
|
if !c.Bool(utils.IPCDisabledFlag.Name) {
|
||||||
if c.IsSet(utils.IPCPathFlag.Name) {
|
if c.IsSet(utils.IPCPathFlag.Name) {
|
||||||
ipcApiUrl = c.String(utils.IPCPathFlag.Name)
|
ipcapiURL = c.String(utils.IPCPathFlag.Name)
|
||||||
} else {
|
} else {
|
||||||
ipcApiUrl = filepath.Join(configDir, "clef.ipc")
|
ipcapiURL = filepath.Join(configDir, "clef.ipc")
|
||||||
}
|
}
|
||||||
|
|
||||||
listener, _, err := rpc.StartIPCEndpoint(ipcApiUrl, rpcApi)
|
listener, _, err := rpc.StartIPCEndpoint(ipcapiURL, rpcAPI)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Could not start IPC api: %v", err)
|
utils.Fatalf("Could not start IPC api: %v", err)
|
||||||
}
|
}
|
||||||
log.Info("IPC endpoint opened", "url", ipcApiUrl)
|
log.Info("IPC endpoint opened", "url", ipcapiURL)
|
||||||
defer func() {
|
defer func() {
|
||||||
listener.Close()
|
listener.Close()
|
||||||
log.Info("IPC endpoint closed", "url", ipcApiUrl)
|
log.Info("IPC endpoint closed", "url", ipcapiURL)
|
||||||
}()
|
}()
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -453,10 +453,10 @@ func signer(c *cli.Context) error {
|
|||||||
}
|
}
|
||||||
ui.OnSignerStartup(core.StartupInfo{
|
ui.OnSignerStartup(core.StartupInfo{
|
||||||
Info: map[string]interface{}{
|
Info: map[string]interface{}{
|
||||||
"extapi_version": ExternalApiVersion,
|
"extapi_version": ExternalAPIVersion,
|
||||||
"intapi_version": InternalApiVersion,
|
"intapi_version": InternalAPIVersion,
|
||||||
"extapi_http": extapiUrl,
|
"extapi_http": extapiURL,
|
||||||
"extapi_ipc": ipcApiUrl,
|
"extapi_ipc": ipcapiURL,
|
||||||
},
|
},
|
||||||
})
|
})
|
||||||
|
|
||||||
|
72
cmd/ethkey/changepassphrase.go
Normal file
72
cmd/ethkey/changepassphrase.go
Normal file
@@ -0,0 +1,72 @@
|
|||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/accounts/keystore"
|
||||||
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
"gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
var newPassphraseFlag = cli.StringFlag{
|
||||||
|
Name: "newpasswordfile",
|
||||||
|
Usage: "the file that contains the new passphrase for the keyfile",
|
||||||
|
}
|
||||||
|
|
||||||
|
var commandChangePassphrase = cli.Command{
|
||||||
|
Name: "changepassphrase",
|
||||||
|
Usage: "change the passphrase on a keyfile",
|
||||||
|
ArgsUsage: "<keyfile>",
|
||||||
|
Description: `
|
||||||
|
Change the passphrase of a keyfile.`,
|
||||||
|
Flags: []cli.Flag{
|
||||||
|
passphraseFlag,
|
||||||
|
newPassphraseFlag,
|
||||||
|
},
|
||||||
|
Action: func(ctx *cli.Context) error {
|
||||||
|
keyfilepath := ctx.Args().First()
|
||||||
|
|
||||||
|
// Read key from file.
|
||||||
|
keyjson, err := ioutil.ReadFile(keyfilepath)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Failed to read the keyfile at '%s': %v", keyfilepath, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Decrypt key with passphrase.
|
||||||
|
passphrase := getPassphrase(ctx)
|
||||||
|
key, err := keystore.DecryptKey(keyjson, passphrase)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error decrypting key: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get a new passphrase.
|
||||||
|
fmt.Println("Please provide a new passphrase")
|
||||||
|
var newPhrase string
|
||||||
|
if passFile := ctx.String(newPassphraseFlag.Name); passFile != "" {
|
||||||
|
content, err := ioutil.ReadFile(passFile)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Failed to read new passphrase file '%s': %v", passFile, err)
|
||||||
|
}
|
||||||
|
newPhrase = strings.TrimRight(string(content), "\r\n")
|
||||||
|
} else {
|
||||||
|
newPhrase = promptPassphrase(true)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Encrypt the key with the new passphrase.
|
||||||
|
newJson, err := keystore.EncryptKey(key, newPhrase, keystore.StandardScryptN, keystore.StandardScryptP)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error encrypting with new passphrase: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Then write the new keyfile in place of the old one.
|
||||||
|
if err := ioutil.WriteFile(keyfilepath, newJson, 600); err != nil {
|
||||||
|
utils.Fatalf("Error writing new keyfile to disk: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Don't print anything. Just return successfully,
|
||||||
|
// producing a positive exit code.
|
||||||
|
return nil
|
||||||
|
},
|
||||||
|
}
|
@@ -90,7 +90,7 @@ If you want to encrypt an existing private key, it can be specified by setting
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Encrypt key with passphrase.
|
// Encrypt key with passphrase.
|
||||||
passphrase := getPassPhrase(ctx, true)
|
passphrase := promptPassphrase(true)
|
||||||
keyjson, err := keystore.EncryptKey(key, passphrase, keystore.StandardScryptN, keystore.StandardScryptP)
|
keyjson, err := keystore.EncryptKey(key, passphrase, keystore.StandardScryptN, keystore.StandardScryptP)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Error encrypting key: %v", err)
|
utils.Fatalf("Error encrypting key: %v", err)
|
||||||
|
@@ -60,7 +60,7 @@ make sure to use this feature with great caution!`,
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Decrypt key with passphrase.
|
// Decrypt key with passphrase.
|
||||||
passphrase := getPassPhrase(ctx, false)
|
passphrase := getPassphrase(ctx)
|
||||||
key, err := keystore.DecryptKey(keyjson, passphrase)
|
key, err := keystore.DecryptKey(keyjson, passphrase)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Error decrypting key: %v", err)
|
utils.Fatalf("Error decrypting key: %v", err)
|
||||||
|
@@ -38,6 +38,7 @@ func init() {
|
|||||||
app.Commands = []cli.Command{
|
app.Commands = []cli.Command{
|
||||||
commandGenerate,
|
commandGenerate,
|
||||||
commandInspect,
|
commandInspect,
|
||||||
|
commandChangePassphrase,
|
||||||
commandSignMessage,
|
commandSignMessage,
|
||||||
commandVerifyMessage,
|
commandVerifyMessage,
|
||||||
}
|
}
|
||||||
|
@@ -62,7 +62,7 @@ To sign a message contained in a file, use the --msgfile flag.
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Decrypt key with passphrase.
|
// Decrypt key with passphrase.
|
||||||
passphrase := getPassPhrase(ctx, false)
|
passphrase := getPassphrase(ctx)
|
||||||
key, err := keystore.DecryptKey(keyjson, passphrase)
|
key, err := keystore.DecryptKey(keyjson, passphrase)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Error decrypting key: %v", err)
|
utils.Fatalf("Error decrypting key: %v", err)
|
||||||
|
@@ -28,11 +28,32 @@ import (
|
|||||||
"gopkg.in/urfave/cli.v1"
|
"gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
// getPassPhrase obtains a passphrase given by the user. It first checks the
|
// promptPassphrase prompts the user for a passphrase. Set confirmation to true
|
||||||
// --passphrase command line flag and ultimately prompts the user for a
|
// to require the user to confirm the passphrase.
|
||||||
|
func promptPassphrase(confirmation bool) string {
|
||||||
|
passphrase, err := console.Stdin.PromptPassword("Passphrase: ")
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Failed to read passphrase: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if confirmation {
|
||||||
|
confirm, err := console.Stdin.PromptPassword("Repeat passphrase: ")
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Failed to read passphrase confirmation: %v", err)
|
||||||
|
}
|
||||||
|
if passphrase != confirm {
|
||||||
|
utils.Fatalf("Passphrases do not match")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return passphrase
|
||||||
|
}
|
||||||
|
|
||||||
|
// getPassphrase obtains a passphrase given by the user. It first checks the
|
||||||
|
// --passfile command line flag and ultimately prompts the user for a
|
||||||
// passphrase.
|
// passphrase.
|
||||||
func getPassPhrase(ctx *cli.Context, confirmation bool) string {
|
func getPassphrase(ctx *cli.Context) string {
|
||||||
// Look for the --passphrase flag.
|
// Look for the --passwordfile flag.
|
||||||
passphraseFile := ctx.String(passphraseFlag.Name)
|
passphraseFile := ctx.String(passphraseFlag.Name)
|
||||||
if passphraseFile != "" {
|
if passphraseFile != "" {
|
||||||
content, err := ioutil.ReadFile(passphraseFile)
|
content, err := ioutil.ReadFile(passphraseFile)
|
||||||
@@ -44,20 +65,7 @@ func getPassPhrase(ctx *cli.Context, confirmation bool) string {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// Otherwise prompt the user for the passphrase.
|
// Otherwise prompt the user for the passphrase.
|
||||||
passphrase, err := console.Stdin.PromptPassword("Passphrase: ")
|
return promptPassphrase(false)
|
||||||
if err != nil {
|
|
||||||
utils.Fatalf("Failed to read passphrase: %v", err)
|
|
||||||
}
|
|
||||||
if confirmation {
|
|
||||||
confirm, err := console.Stdin.PromptPassword("Repeat passphrase: ")
|
|
||||||
if err != nil {
|
|
||||||
utils.Fatalf("Failed to read passphrase confirmation: %v", err)
|
|
||||||
}
|
|
||||||
if passphrase != confirm {
|
|
||||||
utils.Fatalf("Passphrases do not match")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
return passphrase
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// signHash is a helper function that calculates a hash for the given message
|
// signHash is a helper function that calculates a hash for the given message
|
||||||
|
@@ -50,6 +50,6 @@ func compileCmd(ctx *cli.Context) error {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
fmt.Println(bin)
|
fmt.Fprintln(ctx.App.Writer, bin)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@@ -45,6 +45,6 @@ func disasmCmd(ctx *cli.Context) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
code := strings.TrimSpace(string(in[:]))
|
code := strings.TrimSpace(string(in[:]))
|
||||||
fmt.Printf("%v\n", code)
|
fmt.Fprintf(ctx.App.Writer, "%v\n", code)
|
||||||
return asm.PrintDisassembled(code)
|
return asm.PrintDisassembled(code)
|
||||||
}
|
}
|
||||||
|
@@ -30,10 +30,11 @@ func Compile(fn string, src []byte, debug bool) (string, error) {
|
|||||||
bin, compileErrors := compiler.Compile()
|
bin, compileErrors := compiler.Compile()
|
||||||
if len(compileErrors) > 0 {
|
if len(compileErrors) > 0 {
|
||||||
// report errors
|
// report errors
|
||||||
|
errs := ""
|
||||||
for _, err := range compileErrors {
|
for _, err := range compileErrors {
|
||||||
fmt.Printf("%s:%v\n", fn, err)
|
errs += fmt.Sprintf("%s:%v\n", fn, err)
|
||||||
}
|
}
|
||||||
return "", errors.New("compiling failed")
|
return "", errors.New(errs + "compiling failed\n")
|
||||||
}
|
}
|
||||||
return bin, nil
|
return bin, nil
|
||||||
}
|
}
|
||||||
|
@@ -32,6 +32,8 @@ type JSONLogger struct {
|
|||||||
cfg *vm.LogConfig
|
cfg *vm.LogConfig
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// NewJSONLogger creates a new EVM tracer that prints execution steps as JSON objects
|
||||||
|
// into the provided stream.
|
||||||
func NewJSONLogger(cfg *vm.LogConfig, writer io.Writer) *JSONLogger {
|
func NewJSONLogger(cfg *vm.LogConfig, writer io.Writer) *JSONLogger {
|
||||||
return &JSONLogger{json.NewEncoder(writer), cfg}
|
return &JSONLogger{json.NewEncoder(writer), cfg}
|
||||||
}
|
}
|
||||||
|
@@ -98,14 +98,13 @@ func runCmd(ctx *cli.Context) error {
|
|||||||
}
|
}
|
||||||
if ctx.GlobalString(GenesisFlag.Name) != "" {
|
if ctx.GlobalString(GenesisFlag.Name) != "" {
|
||||||
gen := readGenesis(ctx.GlobalString(GenesisFlag.Name))
|
gen := readGenesis(ctx.GlobalString(GenesisFlag.Name))
|
||||||
db, _ := ethdb.NewMemDatabase()
|
db := ethdb.NewMemDatabase()
|
||||||
genesis := gen.ToBlock(db)
|
genesis := gen.ToBlock(db)
|
||||||
statedb, _ = state.New(genesis.Root(), state.NewDatabase(db))
|
statedb, _ = state.New(genesis.Root(), state.NewDatabase(db))
|
||||||
chainConfig = gen.Config
|
chainConfig = gen.Config
|
||||||
blockNumber = gen.Number
|
blockNumber = gen.Number
|
||||||
} else {
|
} else {
|
||||||
db, _ := ethdb.NewMemDatabase()
|
statedb, _ = state.New(common.Hash{}, state.NewDatabase(ethdb.NewMemDatabase()))
|
||||||
statedb, _ = state.New(common.Hash{}, state.NewDatabase(db))
|
|
||||||
}
|
}
|
||||||
if ctx.GlobalString(SenderFlag.Name) != "" {
|
if ctx.GlobalString(SenderFlag.Name) != "" {
|
||||||
sender = common.HexToAddress(ctx.GlobalString(SenderFlag.Name))
|
sender = common.HexToAddress(ctx.GlobalString(SenderFlag.Name))
|
||||||
@@ -129,13 +128,13 @@ func runCmd(ctx *cli.Context) error {
|
|||||||
if ctx.GlobalString(CodeFileFlag.Name) == "-" {
|
if ctx.GlobalString(CodeFileFlag.Name) == "-" {
|
||||||
//Try reading from stdin
|
//Try reading from stdin
|
||||||
if hexcode, err = ioutil.ReadAll(os.Stdin); err != nil {
|
if hexcode, err = ioutil.ReadAll(os.Stdin); err != nil {
|
||||||
fmt.Printf("Could not load code from stdin: %v\n", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "Could not load code from stdin: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// Codefile with hex assembly
|
// Codefile with hex assembly
|
||||||
if hexcode, err = ioutil.ReadFile(ctx.GlobalString(CodeFileFlag.Name)); err != nil {
|
if hexcode, err = ioutil.ReadFile(ctx.GlobalString(CodeFileFlag.Name)); err != nil {
|
||||||
fmt.Printf("Could not load code from file: %v\n", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "Could not load code from file: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -173,11 +172,11 @@ func runCmd(ctx *cli.Context) error {
|
|||||||
if cpuProfilePath := ctx.GlobalString(CPUProfileFlag.Name); cpuProfilePath != "" {
|
if cpuProfilePath := ctx.GlobalString(CPUProfileFlag.Name); cpuProfilePath != "" {
|
||||||
f, err := os.Create(cpuProfilePath)
|
f, err := os.Create(cpuProfilePath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Println("could not create CPU profile: ", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "could not create CPU profile: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
if err := pprof.StartCPUProfile(f); err != nil {
|
if err := pprof.StartCPUProfile(f); err != nil {
|
||||||
fmt.Println("could not start CPU profile: ", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "could not start CPU profile: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
defer pprof.StopCPUProfile()
|
defer pprof.StopCPUProfile()
|
||||||
@@ -201,17 +200,17 @@ func runCmd(ctx *cli.Context) error {
|
|||||||
|
|
||||||
if ctx.GlobalBool(DumpFlag.Name) {
|
if ctx.GlobalBool(DumpFlag.Name) {
|
||||||
statedb.IntermediateRoot(true)
|
statedb.IntermediateRoot(true)
|
||||||
fmt.Println(string(statedb.Dump()))
|
fmt.Fprintln(ctx.App.Writer, string(statedb.Dump()))
|
||||||
}
|
}
|
||||||
|
|
||||||
if memProfilePath := ctx.GlobalString(MemProfileFlag.Name); memProfilePath != "" {
|
if memProfilePath := ctx.GlobalString(MemProfileFlag.Name); memProfilePath != "" {
|
||||||
f, err := os.Create(memProfilePath)
|
f, err := os.Create(memProfilePath)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Println("could not create memory profile: ", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "could not create memory profile: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
if err := pprof.WriteHeapProfile(f); err != nil {
|
if err := pprof.WriteHeapProfile(f); err != nil {
|
||||||
fmt.Println("could not write memory profile: ", err)
|
fmt.Fprintf(ctx.App.ErrWriter, "could not create memory profile: %v\n", err)
|
||||||
os.Exit(1)
|
os.Exit(1)
|
||||||
}
|
}
|
||||||
f.Close()
|
f.Close()
|
||||||
@@ -219,17 +218,17 @@ func runCmd(ctx *cli.Context) error {
|
|||||||
|
|
||||||
if ctx.GlobalBool(DebugFlag.Name) {
|
if ctx.GlobalBool(DebugFlag.Name) {
|
||||||
if debugLogger != nil {
|
if debugLogger != nil {
|
||||||
fmt.Fprintln(os.Stderr, "#### TRACE ####")
|
fmt.Fprintln(ctx.App.ErrWriter, "#### TRACE ####")
|
||||||
vm.WriteTrace(os.Stderr, debugLogger.StructLogs())
|
vm.WriteTrace(ctx.App.ErrWriter, debugLogger.StructLogs())
|
||||||
}
|
}
|
||||||
fmt.Fprintln(os.Stderr, "#### LOGS ####")
|
fmt.Fprintln(ctx.App.ErrWriter, "#### LOGS ####")
|
||||||
vm.WriteLogs(os.Stderr, statedb.Logs())
|
vm.WriteLogs(ctx.App.ErrWriter, statedb.Logs())
|
||||||
}
|
}
|
||||||
|
|
||||||
if ctx.GlobalBool(StatDumpFlag.Name) {
|
if ctx.GlobalBool(StatDumpFlag.Name) {
|
||||||
var mem goruntime.MemStats
|
var mem goruntime.MemStats
|
||||||
goruntime.ReadMemStats(&mem)
|
goruntime.ReadMemStats(&mem)
|
||||||
fmt.Fprintf(os.Stderr, `evm execution time: %v
|
fmt.Fprintf(ctx.App.ErrWriter, `evm execution time: %v
|
||||||
heap objects: %d
|
heap objects: %d
|
||||||
allocations: %d
|
allocations: %d
|
||||||
total allocations: %d
|
total allocations: %d
|
||||||
@@ -239,9 +238,9 @@ Gas used: %d
|
|||||||
`, execTime, mem.HeapObjects, mem.Alloc, mem.TotalAlloc, mem.NumGC, initialGas-leftOverGas)
|
`, execTime, mem.HeapObjects, mem.Alloc, mem.TotalAlloc, mem.NumGC, initialGas-leftOverGas)
|
||||||
}
|
}
|
||||||
if tracer == nil {
|
if tracer == nil {
|
||||||
fmt.Printf("0x%x\n", ret)
|
fmt.Fprintf(ctx.App.Writer, "0x%x\n", ret)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
fmt.Printf(" error: %v\n", err)
|
fmt.Fprintf(ctx.App.ErrWriter, " error: %v\n", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -38,6 +38,8 @@ var stateTestCommand = cli.Command{
|
|||||||
ArgsUsage: "<file>",
|
ArgsUsage: "<file>",
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// StatetestResult contains the execution status after running a state test, any
|
||||||
|
// error that might have occurred and a dump of the final state if requested.
|
||||||
type StatetestResult struct {
|
type StatetestResult struct {
|
||||||
Name string `json:"name"`
|
Name string `json:"name"`
|
||||||
Pass bool `json:"pass"`
|
Pass bool `json:"pass"`
|
||||||
@@ -105,7 +107,7 @@ func stateTestCmd(ctx *cli.Context) error {
|
|||||||
}
|
}
|
||||||
// print state root for evmlab tracing (already committed above, so no need to delete objects again
|
// print state root for evmlab tracing (already committed above, so no need to delete objects again
|
||||||
if ctx.GlobalBool(MachineFlag.Name) && state != nil {
|
if ctx.GlobalBool(MachineFlag.Name) && state != nil {
|
||||||
fmt.Fprintf(os.Stderr, "{\"stateRoot\": \"%x\"}\n", state.IntermediateRoot(false))
|
fmt.Fprintf(ctx.App.ErrWriter, "{\"stateRoot\": \"%x\"}\n", state.IntermediateRoot(false))
|
||||||
}
|
}
|
||||||
|
|
||||||
results = append(results, *result)
|
results = append(results, *result)
|
||||||
@@ -113,13 +115,13 @@ func stateTestCmd(ctx *cli.Context) error {
|
|||||||
// Print any structured logs collected
|
// Print any structured logs collected
|
||||||
if ctx.GlobalBool(DebugFlag.Name) {
|
if ctx.GlobalBool(DebugFlag.Name) {
|
||||||
if debugger != nil {
|
if debugger != nil {
|
||||||
fmt.Fprintln(os.Stderr, "#### TRACE ####")
|
fmt.Fprintln(ctx.App.ErrWriter, "#### TRACE ####")
|
||||||
vm.WriteTrace(os.Stderr, debugger.StructLogs())
|
vm.WriteTrace(ctx.App.ErrWriter, debugger.StructLogs())
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
out, _ := json.MarshalIndent(results, "", " ")
|
out, _ := json.MarshalIndent(results, "", " ")
|
||||||
fmt.Println(string(out))
|
fmt.Fprintln(ctx.App.Writer, string(out))
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@@ -77,9 +77,6 @@ var (
|
|||||||
accJSONFlag = flag.String("account.json", "", "Key json file to fund user requests with")
|
accJSONFlag = flag.String("account.json", "", "Key json file to fund user requests with")
|
||||||
accPassFlag = flag.String("account.pass", "", "Decryption password to access faucet funds")
|
accPassFlag = flag.String("account.pass", "", "Decryption password to access faucet funds")
|
||||||
|
|
||||||
githubUser = flag.String("github.user", "", "GitHub user to authenticate with for Gist access")
|
|
||||||
githubToken = flag.String("github.token", "", "GitHub personal token to access Gists with")
|
|
||||||
|
|
||||||
captchaToken = flag.String("captcha.token", "", "Recaptcha site key to authenticate client side")
|
captchaToken = flag.String("captcha.token", "", "Recaptcha site key to authenticate client side")
|
||||||
captchaSecret = flag.String("captcha.secret", "", "Recaptcha secret key to authenticate server side")
|
captchaSecret = flag.String("captcha.secret", "", "Recaptcha secret key to authenticate server side")
|
||||||
|
|
||||||
@@ -216,7 +213,7 @@ func newFaucet(genesis *core.Genesis, port int, enodes []*discv5.Node, network u
|
|||||||
// Assemble the raw devp2p protocol stack
|
// Assemble the raw devp2p protocol stack
|
||||||
stack, err := node.New(&node.Config{
|
stack, err := node.New(&node.Config{
|
||||||
Name: "geth",
|
Name: "geth",
|
||||||
Version: params.Version,
|
Version: params.VersionWithMeta,
|
||||||
DataDir: filepath.Join(os.Getenv("HOME"), ".faucet"),
|
DataDir: filepath.Join(os.Getenv("HOME"), ".faucet"),
|
||||||
P2P: p2p.Config{
|
P2P: p2p.Config{
|
||||||
NAT: nat.Any(),
|
NAT: nat.Any(),
|
||||||
@@ -474,7 +471,7 @@ func (f *faucet) apiHandler(conn *websocket.Conn) {
|
|||||||
amount = new(big.Int).Div(amount, new(big.Int).Exp(big.NewInt(2), big.NewInt(int64(msg.Tier)), nil))
|
amount = new(big.Int).Div(amount, new(big.Int).Exp(big.NewInt(2), big.NewInt(int64(msg.Tier)), nil))
|
||||||
|
|
||||||
tx := types.NewTransaction(f.nonce+uint64(len(f.reqs)), address, amount, 21000, f.price, nil)
|
tx := types.NewTransaction(f.nonce+uint64(len(f.reqs)), address, amount, 21000, f.price, nil)
|
||||||
signed, err := f.keystore.SignTx(f.account, tx, f.config.ChainId)
|
signed, err := f.keystore.SignTx(f.account, tx, f.config.ChainID)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
f.lock.Unlock()
|
f.lock.Unlock()
|
||||||
if err = sendError(conn, err); err != nil {
|
if err = sendError(conn, err); err != nil {
|
||||||
@@ -638,59 +635,6 @@ func sendSuccess(conn *websocket.Conn, msg string) error {
|
|||||||
return send(conn, map[string]string{"success": msg}, time.Second)
|
return send(conn, map[string]string{"success": msg}, time.Second)
|
||||||
}
|
}
|
||||||
|
|
||||||
// authGitHub tries to authenticate a faucet request using GitHub gists, returning
|
|
||||||
// the username, avatar URL and Ethereum address to fund on success.
|
|
||||||
func authGitHub(url string) (string, string, common.Address, error) {
|
|
||||||
// Retrieve the gist from the GitHub Gist APIs
|
|
||||||
parts := strings.Split(url, "/")
|
|
||||||
req, _ := http.NewRequest("GET", "https://api.github.com/gists/"+parts[len(parts)-1], nil)
|
|
||||||
if *githubUser != "" {
|
|
||||||
req.SetBasicAuth(*githubUser, *githubToken)
|
|
||||||
}
|
|
||||||
res, err := http.DefaultClient.Do(req)
|
|
||||||
if err != nil {
|
|
||||||
return "", "", common.Address{}, err
|
|
||||||
}
|
|
||||||
var gist struct {
|
|
||||||
Owner struct {
|
|
||||||
Login string `json:"login"`
|
|
||||||
} `json:"owner"`
|
|
||||||
Files map[string]struct {
|
|
||||||
Content string `json:"content"`
|
|
||||||
} `json:"files"`
|
|
||||||
}
|
|
||||||
err = json.NewDecoder(res.Body).Decode(&gist)
|
|
||||||
res.Body.Close()
|
|
||||||
if err != nil {
|
|
||||||
return "", "", common.Address{}, err
|
|
||||||
}
|
|
||||||
if gist.Owner.Login == "" {
|
|
||||||
return "", "", common.Address{}, errors.New("Anonymous Gists not allowed")
|
|
||||||
}
|
|
||||||
// Iterate over all the files and look for Ethereum addresses
|
|
||||||
var address common.Address
|
|
||||||
for _, file := range gist.Files {
|
|
||||||
content := strings.TrimSpace(file.Content)
|
|
||||||
if len(content) == 2+common.AddressLength*2 {
|
|
||||||
address = common.HexToAddress(content)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if address == (common.Address{}) {
|
|
||||||
return "", "", common.Address{}, errors.New("No Ethereum address found to fund")
|
|
||||||
}
|
|
||||||
// Validate the user's existence since the API is unhelpful here
|
|
||||||
if res, err = http.Head("https://github.com/" + gist.Owner.Login); err != nil {
|
|
||||||
return "", "", common.Address{}, err
|
|
||||||
}
|
|
||||||
res.Body.Close()
|
|
||||||
|
|
||||||
if res.StatusCode != 200 {
|
|
||||||
return "", "", common.Address{}, errors.New("Invalid user... boom!")
|
|
||||||
}
|
|
||||||
// Everything passed validation, return the gathered infos
|
|
||||||
return gist.Owner.Login + "@github", fmt.Sprintf("https://github.com/%s.png?size=64", gist.Owner.Login), address, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// authTwitter tries to authenticate a faucet request using Twitter posts, returning
|
// authTwitter tries to authenticate a faucet request using Twitter posts, returning
|
||||||
// the username, avatar URL and Ethereum address to fund on success.
|
// the username, avatar URL and Ethereum address to fund on success.
|
||||||
func authTwitter(url string) (string, string, common.Address, error) {
|
func authTwitter(url string) (string, string, common.Address, error) {
|
||||||
|
@@ -340,7 +340,7 @@ func importWallet(ctx *cli.Context) error {
|
|||||||
if len(keyfile) == 0 {
|
if len(keyfile) == 0 {
|
||||||
utils.Fatalf("keyfile must be given as argument")
|
utils.Fatalf("keyfile must be given as argument")
|
||||||
}
|
}
|
||||||
keyJson, err := ioutil.ReadFile(keyfile)
|
keyJSON, err := ioutil.ReadFile(keyfile)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Could not read wallet file: %v", err)
|
utils.Fatalf("Could not read wallet file: %v", err)
|
||||||
}
|
}
|
||||||
@@ -349,7 +349,7 @@ func importWallet(ctx *cli.Context) error {
|
|||||||
passphrase := getPassPhrase("", false, 0, utils.MakePasswordList(ctx))
|
passphrase := getPassPhrase("", false, 0, utils.MakePasswordList(ctx))
|
||||||
|
|
||||||
ks := stack.AccountManager().Backends(keystore.KeyStoreType)[0].(*keystore.KeyStore)
|
ks := stack.AccountManager().Backends(keystore.KeyStoreType)[0].(*keystore.KeyStore)
|
||||||
acct, err := ks.ImportPreSaleKey(keyJson, passphrase)
|
acct, err := ks.ImportPreSaleKey(keyJSON, passphrase)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("%v", err)
|
utils.Fatalf("%v", err)
|
||||||
}
|
}
|
||||||
|
@@ -41,7 +41,7 @@ var bugCommand = cli.Command{
|
|||||||
Category: "MISCELLANEOUS COMMANDS",
|
Category: "MISCELLANEOUS COMMANDS",
|
||||||
}
|
}
|
||||||
|
|
||||||
const issueUrl = "https://github.com/ethereum/go-ethereum/issues/new"
|
const issueURL = "https://github.com/ethereum/go-ethereum/issues/new"
|
||||||
|
|
||||||
// reportBug reports a bug by opening a new URL to the go-ethereum GH issue
|
// reportBug reports a bug by opening a new URL to the go-ethereum GH issue
|
||||||
// tracker and setting default values as the issue body.
|
// tracker and setting default values as the issue body.
|
||||||
@@ -51,15 +51,15 @@ func reportBug(ctx *cli.Context) error {
|
|||||||
|
|
||||||
fmt.Fprintln(&buff, "#### System information")
|
fmt.Fprintln(&buff, "#### System information")
|
||||||
fmt.Fprintln(&buff)
|
fmt.Fprintln(&buff)
|
||||||
fmt.Fprintln(&buff, "Version:", params.Version)
|
fmt.Fprintln(&buff, "Version:", params.VersionWithMeta)
|
||||||
fmt.Fprintln(&buff, "Go Version:", runtime.Version())
|
fmt.Fprintln(&buff, "Go Version:", runtime.Version())
|
||||||
fmt.Fprintln(&buff, "OS:", runtime.GOOS)
|
fmt.Fprintln(&buff, "OS:", runtime.GOOS)
|
||||||
printOSDetails(&buff)
|
printOSDetails(&buff)
|
||||||
fmt.Fprintln(&buff, header)
|
fmt.Fprintln(&buff, header)
|
||||||
|
|
||||||
// open a new GH issue
|
// open a new GH issue
|
||||||
if !browser.Open(issueUrl + "?body=" + url.QueryEscape(buff.String())) {
|
if !browser.Open(issueURL + "?body=" + url.QueryEscape(buff.String())) {
|
||||||
fmt.Printf("Please file a new issue at %s using this template:\n\n%s", issueUrl, buff.String())
|
fmt.Printf("Please file a new issue at %s using this template:\n\n%s", issueURL, buff.String())
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@@ -94,7 +94,8 @@ processing will proceed even if an individual RLP-file import failure occurs.`,
|
|||||||
Requires a first argument of the file to write to.
|
Requires a first argument of the file to write to.
|
||||||
Optional second and third arguments control the first and
|
Optional second and third arguments control the first and
|
||||||
last block to write. In this mode, the file will be appended
|
last block to write. In this mode, the file will be appended
|
||||||
if already existing.`,
|
if already existing. If the file ends with .gz, the output will
|
||||||
|
be gzipped.`,
|
||||||
}
|
}
|
||||||
importPreimagesCommand = cli.Command{
|
importPreimagesCommand = cli.Command{
|
||||||
Action: utils.MigrateFlags(importPreimages),
|
Action: utils.MigrateFlags(importPreimages),
|
||||||
|
@@ -50,7 +50,7 @@ func TestConsoleWelcome(t *testing.T) {
|
|||||||
geth.SetTemplateFunc("goos", func() string { return runtime.GOOS })
|
geth.SetTemplateFunc("goos", func() string { return runtime.GOOS })
|
||||||
geth.SetTemplateFunc("goarch", func() string { return runtime.GOARCH })
|
geth.SetTemplateFunc("goarch", func() string { return runtime.GOARCH })
|
||||||
geth.SetTemplateFunc("gover", runtime.Version)
|
geth.SetTemplateFunc("gover", runtime.Version)
|
||||||
geth.SetTemplateFunc("gethver", func() string { return params.Version })
|
geth.SetTemplateFunc("gethver", func() string { return params.VersionWithMeta })
|
||||||
geth.SetTemplateFunc("niltime", func() string { return time.Unix(0, 0).Format(time.RFC1123) })
|
geth.SetTemplateFunc("niltime", func() string { return time.Unix(0, 0).Format(time.RFC1123) })
|
||||||
geth.SetTemplateFunc("apis", func() string { return ipcAPIs })
|
geth.SetTemplateFunc("apis", func() string { return ipcAPIs })
|
||||||
|
|
||||||
@@ -133,7 +133,7 @@ func testAttachWelcome(t *testing.T, geth *testgeth, endpoint, apis string) {
|
|||||||
attach.SetTemplateFunc("goos", func() string { return runtime.GOOS })
|
attach.SetTemplateFunc("goos", func() string { return runtime.GOOS })
|
||||||
attach.SetTemplateFunc("goarch", func() string { return runtime.GOARCH })
|
attach.SetTemplateFunc("goarch", func() string { return runtime.GOARCH })
|
||||||
attach.SetTemplateFunc("gover", runtime.Version)
|
attach.SetTemplateFunc("gover", runtime.Version)
|
||||||
attach.SetTemplateFunc("gethver", func() string { return params.Version })
|
attach.SetTemplateFunc("gethver", func() string { return params.VersionWithMeta })
|
||||||
attach.SetTemplateFunc("etherbase", func() string { return geth.Etherbase })
|
attach.SetTemplateFunc("etherbase", func() string { return geth.Etherbase })
|
||||||
attach.SetTemplateFunc("niltime", func() string { return time.Unix(0, 0).Format(time.RFC1123) })
|
attach.SetTemplateFunc("niltime", func() string { return time.Unix(0, 0).Format(time.RFC1123) })
|
||||||
attach.SetTemplateFunc("ipc", func() bool { return strings.HasPrefix(endpoint, "ipc") })
|
attach.SetTemplateFunc("ipc", func() bool { return strings.HasPrefix(endpoint, "ipc") })
|
||||||
|
@@ -24,7 +24,7 @@ import (
|
|||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
"github.com/ethereum/go-ethereum/params"
|
"github.com/ethereum/go-ethereum/params"
|
||||||
)
|
)
|
||||||
@@ -131,8 +131,8 @@ func testDAOForkBlockNewChain(t *testing.T, test int, genesis string, expectBloc
|
|||||||
if genesis != "" {
|
if genesis != "" {
|
||||||
genesisHash = daoGenesisHash
|
genesisHash = daoGenesisHash
|
||||||
}
|
}
|
||||||
config, err := core.GetChainConfig(db, genesisHash)
|
config := rawdb.ReadChainConfig(db, genesisHash)
|
||||||
if err != nil {
|
if config == nil {
|
||||||
t.Errorf("test %d: failed to retrieve chain config: %v", test, err)
|
t.Errorf("test %d: failed to retrieve chain config: %v", test, err)
|
||||||
return // we want to return here, the other checks can't make it past this point (nil panic).
|
return // we want to return here, the other checks can't make it past this point (nil panic).
|
||||||
}
|
}
|
||||||
|
@@ -77,7 +77,7 @@ var customGenesisTests = []struct {
|
|||||||
"homesteadBlock" : 314,
|
"homesteadBlock" : 314,
|
||||||
"daoForkBlock" : 141,
|
"daoForkBlock" : 141,
|
||||||
"daoForkSupport" : true
|
"daoForkSupport" : true
|
||||||
},
|
}
|
||||||
}`,
|
}`,
|
||||||
query: "eth.getBlock(0).nonce",
|
query: "eth.getBlock(0).nonce",
|
||||||
result: "0x0000000000000042",
|
result: "0x0000000000000042",
|
||||||
|
@@ -19,12 +19,16 @@ package main
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"math"
|
||||||
"os"
|
"os"
|
||||||
"runtime"
|
"runtime"
|
||||||
|
godebug "runtime/debug"
|
||||||
"sort"
|
"sort"
|
||||||
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
|
"github.com/elastic/gosigar"
|
||||||
"github.com/ethereum/go-ethereum/accounts"
|
"github.com/ethereum/go-ethereum/accounts"
|
||||||
"github.com/ethereum/go-ethereum/accounts/keystore"
|
"github.com/ethereum/go-ethereum/accounts/keystore"
|
||||||
"github.com/ethereum/go-ethereum/cmd/utils"
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
@@ -140,6 +144,15 @@ var (
|
|||||||
utils.WhisperMaxMessageSizeFlag,
|
utils.WhisperMaxMessageSizeFlag,
|
||||||
utils.WhisperMinPOWFlag,
|
utils.WhisperMinPOWFlag,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
metricsFlags = []cli.Flag{
|
||||||
|
utils.MetricsEnableInfluxDBFlag,
|
||||||
|
utils.MetricsInfluxDBEndpointFlag,
|
||||||
|
utils.MetricsInfluxDBDatabaseFlag,
|
||||||
|
utils.MetricsInfluxDBUsernameFlag,
|
||||||
|
utils.MetricsInfluxDBPasswordFlag,
|
||||||
|
utils.MetricsInfluxDBHostTagFlag,
|
||||||
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
@@ -182,12 +195,37 @@ func init() {
|
|||||||
app.Flags = append(app.Flags, consoleFlags...)
|
app.Flags = append(app.Flags, consoleFlags...)
|
||||||
app.Flags = append(app.Flags, debug.Flags...)
|
app.Flags = append(app.Flags, debug.Flags...)
|
||||||
app.Flags = append(app.Flags, whisperFlags...)
|
app.Flags = append(app.Flags, whisperFlags...)
|
||||||
|
app.Flags = append(app.Flags, metricsFlags...)
|
||||||
|
|
||||||
app.Before = func(ctx *cli.Context) error {
|
app.Before = func(ctx *cli.Context) error {
|
||||||
runtime.GOMAXPROCS(runtime.NumCPU())
|
runtime.GOMAXPROCS(runtime.NumCPU())
|
||||||
if err := debug.Setup(ctx); err != nil {
|
|
||||||
|
logdir := ""
|
||||||
|
if ctx.GlobalBool(utils.DashboardEnabledFlag.Name) {
|
||||||
|
logdir = (&node.Config{DataDir: utils.MakeDataDir(ctx)}).ResolvePath("logs")
|
||||||
|
}
|
||||||
|
if err := debug.Setup(ctx, logdir); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
// Cap the cache allowance and tune the garbage collector
|
||||||
|
var mem gosigar.Mem
|
||||||
|
if err := mem.Get(); err == nil {
|
||||||
|
allowance := int(mem.Total / 1024 / 1024 / 3)
|
||||||
|
if cache := ctx.GlobalInt(utils.CacheFlag.Name); cache > allowance {
|
||||||
|
log.Warn("Sanitizing cache to Go's GC limits", "provided", cache, "updated", allowance)
|
||||||
|
ctx.GlobalSet(utils.CacheFlag.Name, strconv.Itoa(allowance))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Ensure Go's GC ignores the database cache for trigger percentage
|
||||||
|
cache := ctx.GlobalInt(utils.CacheFlag.Name)
|
||||||
|
gogc := math.Max(20, math.Min(100, 100/(float64(cache)/1024)))
|
||||||
|
|
||||||
|
log.Debug("Sanitizing Go's GC trigger", "percent", int(gogc))
|
||||||
|
godebug.SetGCPercent(int(gogc))
|
||||||
|
|
||||||
|
// Start metrics export if enabled
|
||||||
|
utils.SetupMetrics(ctx)
|
||||||
|
|
||||||
// Start system runtime metrics collection
|
// Start system runtime metrics collection
|
||||||
go metrics.CollectProcessMetrics(3 * time.Second)
|
go metrics.CollectProcessMetrics(3 * time.Second)
|
||||||
|
|
||||||
@@ -213,6 +251,9 @@ func main() {
|
|||||||
// It creates a default node based on the command line arguments and runs it in
|
// It creates a default node based on the command line arguments and runs it in
|
||||||
// blocking mode, waiting for it to be shut down.
|
// blocking mode, waiting for it to be shut down.
|
||||||
func geth(ctx *cli.Context) error {
|
func geth(ctx *cli.Context) error {
|
||||||
|
if args := ctx.Args(); len(args) > 0 {
|
||||||
|
return fmt.Errorf("invalid command: %q", args[0])
|
||||||
|
}
|
||||||
node := makeFullNode(ctx)
|
node := makeFullNode(ctx)
|
||||||
startNode(ctx, node)
|
startNode(ctx, node)
|
||||||
node.Wait()
|
node.Wait()
|
||||||
@@ -267,11 +308,11 @@ func startNode(ctx *cli.Context, stack *node.Node) {
|
|||||||
status, _ := event.Wallet.Status()
|
status, _ := event.Wallet.Status()
|
||||||
log.Info("New wallet appeared", "url", event.Wallet.URL(), "status", status)
|
log.Info("New wallet appeared", "url", event.Wallet.URL(), "status", status)
|
||||||
|
|
||||||
|
derivationPath := accounts.DefaultBaseDerivationPath
|
||||||
if event.Wallet.URL().Scheme == "ledger" {
|
if event.Wallet.URL().Scheme == "ledger" {
|
||||||
event.Wallet.SelfDerive(accounts.DefaultLedgerBaseDerivationPath, stateReader)
|
derivationPath = accounts.DefaultLedgerBaseDerivationPath
|
||||||
} else {
|
|
||||||
event.Wallet.SelfDerive(accounts.DefaultBaseDerivationPath, stateReader)
|
|
||||||
}
|
}
|
||||||
|
event.Wallet.SelfDerive(derivationPath, stateReader)
|
||||||
|
|
||||||
case accounts.WalletDropped:
|
case accounts.WalletDropped:
|
||||||
log.Info("Old wallet dropped", "url", event.Wallet.URL())
|
log.Info("Old wallet dropped", "url", event.Wallet.URL())
|
||||||
|
@@ -108,7 +108,7 @@ func makedag(ctx *cli.Context) error {
|
|||||||
|
|
||||||
func version(ctx *cli.Context) error {
|
func version(ctx *cli.Context) error {
|
||||||
fmt.Println(strings.Title(clientIdentifier))
|
fmt.Println(strings.Title(clientIdentifier))
|
||||||
fmt.Println("Version:", params.Version)
|
fmt.Println("Version:", params.VersionWithMeta)
|
||||||
if gitCommit != "" {
|
if gitCommit != "" {
|
||||||
fmt.Println("Git Commit:", gitCommit)
|
fmt.Println("Git Commit:", gitCommit)
|
||||||
}
|
}
|
||||||
|
@@ -185,12 +185,12 @@ func resolveMetric(metrics map[string]interface{}, pattern string, path string)
|
|||||||
parts := strings.SplitN(pattern, "/", 2)
|
parts := strings.SplitN(pattern, "/", 2)
|
||||||
if len(parts) > 1 {
|
if len(parts) > 1 {
|
||||||
for _, variation := range strings.Split(parts[0], ",") {
|
for _, variation := range strings.Split(parts[0], ",") {
|
||||||
if submetrics, ok := metrics[variation].(map[string]interface{}); !ok {
|
submetrics, ok := metrics[variation].(map[string]interface{})
|
||||||
|
if !ok {
|
||||||
utils.Fatalf("Failed to retrieve system metrics: %s", path+variation)
|
utils.Fatalf("Failed to retrieve system metrics: %s", path+variation)
|
||||||
return nil
|
return nil
|
||||||
} else {
|
|
||||||
results = append(results, resolveMetric(submetrics, parts[1], path+variation+"/")...)
|
|
||||||
}
|
}
|
||||||
|
results = append(results, resolveMetric(submetrics, parts[1], path+variation+"/")...)
|
||||||
}
|
}
|
||||||
return results
|
return results
|
||||||
}
|
}
|
||||||
|
@@ -83,7 +83,8 @@ var AppHelpFlagGroups = []flagGroup{
|
|||||||
utils.LightKDFFlag,
|
utils.LightKDFFlag,
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{Name: "DEVELOPER CHAIN",
|
{
|
||||||
|
Name: "DEVELOPER CHAIN",
|
||||||
Flags: []cli.Flag{
|
Flags: []cli.Flag{
|
||||||
utils.DeveloperFlag,
|
utils.DeveloperFlag,
|
||||||
utils.DeveloperPeriodFlag,
|
utils.DeveloperPeriodFlag,
|
||||||
@@ -206,11 +207,22 @@ var AppHelpFlagGroups = []flagGroup{
|
|||||||
{
|
{
|
||||||
Name: "LOGGING AND DEBUGGING",
|
Name: "LOGGING AND DEBUGGING",
|
||||||
Flags: append([]cli.Flag{
|
Flags: append([]cli.Flag{
|
||||||
utils.MetricsEnabledFlag,
|
|
||||||
utils.FakePoWFlag,
|
utils.FakePoWFlag,
|
||||||
utils.NoCompactionFlag,
|
utils.NoCompactionFlag,
|
||||||
}, debug.Flags...),
|
}, debug.Flags...),
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
Name: "METRICS AND STATS",
|
||||||
|
Flags: []cli.Flag{
|
||||||
|
utils.MetricsEnabledFlag,
|
||||||
|
utils.MetricsEnableInfluxDBFlag,
|
||||||
|
utils.MetricsInfluxDBEndpointFlag,
|
||||||
|
utils.MetricsInfluxDBDatabaseFlag,
|
||||||
|
utils.MetricsInfluxDBUsernameFlag,
|
||||||
|
utils.MetricsInfluxDBPasswordFlag,
|
||||||
|
utils.MetricsInfluxDBHostTagFlag,
|
||||||
|
},
|
||||||
|
},
|
||||||
{
|
{
|
||||||
Name: "WHISPER (EXPERIMENTAL)",
|
Name: "WHISPER (EXPERIMENTAL)",
|
||||||
Flags: whisperFlags,
|
Flags: whisperFlags,
|
||||||
|
@@ -180,7 +180,10 @@ func main() {
|
|||||||
},
|
},
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
app.Run(os.Args)
|
if err := app.Run(os.Args); err != nil {
|
||||||
|
fmt.Fprintln(os.Stderr, err)
|
||||||
|
os.Exit(1)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func showNetwork(ctx *cli.Context) error {
|
func showNetwork(ctx *cli.Context) error {
|
||||||
@@ -275,9 +278,8 @@ func createNode(ctx *cli.Context) error {
|
|||||||
if len(ctx.Args()) != 0 {
|
if len(ctx.Args()) != 0 {
|
||||||
return cli.ShowCommandHelp(ctx, ctx.Command.Name)
|
return cli.ShowCommandHelp(ctx, ctx.Command.Name)
|
||||||
}
|
}
|
||||||
config := &adapters.NodeConfig{
|
config := adapters.RandomNodeConfig()
|
||||||
Name: ctx.String("name"),
|
config.Name = ctx.String("name")
|
||||||
}
|
|
||||||
if key := ctx.String("key"); key != "" {
|
if key := ctx.String("key"); key != "" {
|
||||||
privKey, err := crypto.HexToECDSA(key)
|
privKey, err := crypto.HexToECDSA(key)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
|
@@ -103,8 +103,8 @@ func newCppEthereumGenesisSpec(network string, genesis *core.Genesis) (*cppEther
|
|||||||
spec.Params.ByzantiumForkBlock = (hexutil.Uint64)(genesis.Config.ByzantiumBlock.Uint64())
|
spec.Params.ByzantiumForkBlock = (hexutil.Uint64)(genesis.Config.ByzantiumBlock.Uint64())
|
||||||
spec.Params.ConstantinopleForkBlock = (hexutil.Uint64)(math.MaxUint64)
|
spec.Params.ConstantinopleForkBlock = (hexutil.Uint64)(math.MaxUint64)
|
||||||
|
|
||||||
spec.Params.NetworkID = (hexutil.Uint64)(genesis.Config.ChainId.Uint64())
|
spec.Params.NetworkID = (hexutil.Uint64)(genesis.Config.ChainID.Uint64())
|
||||||
spec.Params.ChainID = (hexutil.Uint64)(genesis.Config.ChainId.Uint64())
|
spec.Params.ChainID = (hexutil.Uint64)(genesis.Config.ChainID.Uint64())
|
||||||
|
|
||||||
spec.Params.MaximumExtraDataSize = (hexutil.Uint64)(params.MaximumExtraDataSize)
|
spec.Params.MaximumExtraDataSize = (hexutil.Uint64)(params.MaximumExtraDataSize)
|
||||||
spec.Params.MinGasLimit = (hexutil.Uint64)(params.MinGasLimit)
|
spec.Params.MinGasLimit = (hexutil.Uint64)(params.MinGasLimit)
|
||||||
@@ -284,7 +284,7 @@ func newParityChainSpec(network string, genesis *core.Genesis, bootnodes []strin
|
|||||||
spec.Params.MaximumExtraDataSize = (hexutil.Uint64)(params.MaximumExtraDataSize)
|
spec.Params.MaximumExtraDataSize = (hexutil.Uint64)(params.MaximumExtraDataSize)
|
||||||
spec.Params.MinGasLimit = (hexutil.Uint64)(params.MinGasLimit)
|
spec.Params.MinGasLimit = (hexutil.Uint64)(params.MinGasLimit)
|
||||||
spec.Params.GasLimitBoundDivisor = (hexutil.Uint64)(params.GasLimitBoundDivisor)
|
spec.Params.GasLimitBoundDivisor = (hexutil.Uint64)(params.GasLimitBoundDivisor)
|
||||||
spec.Params.NetworkID = (hexutil.Uint64)(genesis.Config.ChainId.Uint64())
|
spec.Params.NetworkID = (hexutil.Uint64)(genesis.Config.ChainID.Uint64())
|
||||||
spec.Params.MaxCodeSize = params.MaxCodeSize
|
spec.Params.MaxCodeSize = params.MaxCodeSize
|
||||||
spec.Params.EIP155Transition = genesis.Config.EIP155Block.Uint64()
|
spec.Params.EIP155Transition = genesis.Config.EIP155Block.Uint64()
|
||||||
spec.Params.EIP98Transition = math.MaxUint64
|
spec.Params.EIP98Transition = math.MaxUint64
|
||||||
|
@@ -609,7 +609,7 @@ func deployDashboard(client *sshClient, network string, conf *config, config *da
|
|||||||
}
|
}
|
||||||
template.Must(template.New("").Parse(dashboardContent)).Execute(indexfile, map[string]interface{}{
|
template.Must(template.New("").Parse(dashboardContent)).Execute(indexfile, map[string]interface{}{
|
||||||
"Network": network,
|
"Network": network,
|
||||||
"NetworkID": conf.Genesis.Config.ChainId,
|
"NetworkID": conf.Genesis.Config.ChainID,
|
||||||
"NetworkTitle": strings.Title(network),
|
"NetworkTitle": strings.Title(network),
|
||||||
"EthstatsPage": config.ethstats,
|
"EthstatsPage": config.ethstats,
|
||||||
"ExplorerPage": config.explorer,
|
"ExplorerPage": config.explorer,
|
||||||
|
@@ -122,7 +122,7 @@ func (info *ethstatsInfos) Report() map[string]string {
|
|||||||
"Website address": info.host,
|
"Website address": info.host,
|
||||||
"Website listener port": strconv.Itoa(info.port),
|
"Website listener port": strconv.Itoa(info.port),
|
||||||
"Login secret": info.secret,
|
"Login secret": info.secret,
|
||||||
"Banned addresses": fmt.Sprintf("%v", info.banned),
|
"Banned addresses": strings.Join(info.banned, "\n"),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@@ -221,7 +221,7 @@ func checkNode(client *sshClient, network string, boot bool) (*nodeInfos, error)
|
|||||||
|
|
||||||
// Container available, retrieve its node ID and its genesis json
|
// Container available, retrieve its node ID and its genesis json
|
||||||
var out []byte
|
var out []byte
|
||||||
if out, err = client.Run(fmt.Sprintf("docker exec %s_%s_1 geth --exec admin.nodeInfo.id attach", network, kind)); err != nil {
|
if out, err = client.Run(fmt.Sprintf("docker exec %s_%s_1 geth --exec admin.nodeInfo.id --cache=16 attach", network, kind)); err != nil {
|
||||||
return nil, ErrServiceUnreachable
|
return nil, ErrServiceUnreachable
|
||||||
}
|
}
|
||||||
id := bytes.Trim(bytes.TrimSpace(out), "\"")
|
id := bytes.Trim(bytes.TrimSpace(out), "\"")
|
||||||
|
@@ -49,7 +49,7 @@ func (w *wizard) deployFaucet() {
|
|||||||
existed := err == nil
|
existed := err == nil
|
||||||
|
|
||||||
infos.node.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
infos.node.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
||||||
infos.node.network = w.conf.Genesis.Config.ChainId.Int64()
|
infos.node.network = w.conf.Genesis.Config.ChainID.Int64()
|
||||||
|
|
||||||
// Figure out which port to listen on
|
// Figure out which port to listen on
|
||||||
fmt.Println()
|
fmt.Println()
|
||||||
|
@@ -121,7 +121,7 @@ func (w *wizard) makeGenesis() {
|
|||||||
// Query the user for some custom extras
|
// Query the user for some custom extras
|
||||||
fmt.Println()
|
fmt.Println()
|
||||||
fmt.Println("Specify your chain/network ID if you want an explicit one (default = random)")
|
fmt.Println("Specify your chain/network ID if you want an explicit one (default = random)")
|
||||||
genesis.Config.ChainId = new(big.Int).SetUint64(uint64(w.readDefaultInt(rand.Intn(65536))))
|
genesis.Config.ChainID = new(big.Int).SetUint64(uint64(w.readDefaultInt(rand.Intn(65536))))
|
||||||
|
|
||||||
// All done, store the genesis and flush to disk
|
// All done, store the genesis and flush to disk
|
||||||
log.Info("Configured new genesis block")
|
log.Info("Configured new genesis block")
|
||||||
|
@@ -203,7 +203,7 @@ func (stats serverStats) render() {
|
|||||||
|
|
||||||
table.SetHeader([]string{"Server", "Address", "Service", "Config", "Value"})
|
table.SetHeader([]string{"Server", "Address", "Service", "Config", "Value"})
|
||||||
table.SetAlignment(tablewriter.ALIGN_LEFT)
|
table.SetAlignment(tablewriter.ALIGN_LEFT)
|
||||||
table.SetColWidth(100)
|
table.SetColWidth(40)
|
||||||
|
|
||||||
// Find the longest lines for all columns for the hacked separator
|
// Find the longest lines for all columns for the hacked separator
|
||||||
separator := make([]string, 5)
|
separator := make([]string, 5)
|
||||||
@@ -222,8 +222,10 @@ func (stats serverStats) render() {
|
|||||||
if len(config) > len(separator[3]) {
|
if len(config) > len(separator[3]) {
|
||||||
separator[3] = strings.Repeat("-", len(config))
|
separator[3] = strings.Repeat("-", len(config))
|
||||||
}
|
}
|
||||||
if len(value) > len(separator[4]) {
|
for _, val := range strings.Split(value, "\n") {
|
||||||
separator[4] = strings.Repeat("-", len(value))
|
if len(val) > len(separator[4]) {
|
||||||
|
separator[4] = strings.Repeat("-", len(val))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -263,26 +265,20 @@ func (stats serverStats) render() {
|
|||||||
sort.Strings(configs)
|
sort.Strings(configs)
|
||||||
|
|
||||||
for k, config := range configs {
|
for k, config := range configs {
|
||||||
switch {
|
for l, value := range strings.Split(stats[server].services[service][config], "\n") {
|
||||||
case j == 0 && k == 0:
|
switch {
|
||||||
table.Append([]string{server, stats[server].address, service, config, stats[server].services[service][config]})
|
case j == 0 && k == 0 && l == 0:
|
||||||
case k == 0:
|
table.Append([]string{server, stats[server].address, service, config, value})
|
||||||
table.Append([]string{"", "", service, config, stats[server].services[service][config]})
|
case k == 0 && l == 0:
|
||||||
default:
|
table.Append([]string{"", "", service, config, value})
|
||||||
table.Append([]string{"", "", "", config, stats[server].services[service][config]})
|
case l == 0:
|
||||||
|
table.Append([]string{"", "", "", config, value})
|
||||||
|
default:
|
||||||
|
table.Append([]string{"", "", "", "", value})
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
table.Render()
|
table.Render()
|
||||||
}
|
}
|
||||||
|
|
||||||
// protips contains a collection of network infos to report pro-tips
|
|
||||||
// based on.
|
|
||||||
type protips struct {
|
|
||||||
genesis string
|
|
||||||
network int64
|
|
||||||
bootFull []string
|
|
||||||
bootLight []string
|
|
||||||
ethstats string
|
|
||||||
}
|
|
||||||
|
@@ -56,7 +56,7 @@ func (w *wizard) deployNode(boot bool) {
|
|||||||
existed := err == nil
|
existed := err == nil
|
||||||
|
|
||||||
infos.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
infos.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
||||||
infos.network = w.conf.Genesis.Config.ChainId.Int64()
|
infos.network = w.conf.Genesis.Config.ChainID.Int64()
|
||||||
|
|
||||||
// Figure out where the user wants to store the persistent data
|
// Figure out where the user wants to store the persistent data
|
||||||
fmt.Println()
|
fmt.Println()
|
||||||
@@ -107,7 +107,7 @@ func (w *wizard) deployNode(boot bool) {
|
|||||||
// Ethash based miners only need an etherbase to mine against
|
// Ethash based miners only need an etherbase to mine against
|
||||||
fmt.Println()
|
fmt.Println()
|
||||||
if infos.etherbase == "" {
|
if infos.etherbase == "" {
|
||||||
fmt.Printf("What address should the miner user?\n")
|
fmt.Printf("What address should the miner use?\n")
|
||||||
for {
|
for {
|
||||||
if address := w.readAddress(); address != nil {
|
if address := w.readAddress(); address != nil {
|
||||||
infos.etherbase = address.Hex()
|
infos.etherbase = address.Hex()
|
||||||
@@ -115,7 +115,7 @@ func (w *wizard) deployNode(boot bool) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
fmt.Printf("What address should the miner user? (default = %s)\n", infos.etherbase)
|
fmt.Printf("What address should the miner use? (default = %s)\n", infos.etherbase)
|
||||||
infos.etherbase = w.readDefaultAddress(common.HexToAddress(infos.etherbase)).Hex()
|
infos.etherbase = w.readDefaultAddress(common.HexToAddress(infos.etherbase)).Hex()
|
||||||
}
|
}
|
||||||
} else if w.conf.Genesis.Config.Clique != nil {
|
} else if w.conf.Genesis.Config.Clique != nil {
|
||||||
|
@@ -52,7 +52,7 @@ func (w *wizard) deployWallet() {
|
|||||||
existed := err == nil
|
existed := err == nil
|
||||||
|
|
||||||
infos.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
infos.genesis, _ = json.MarshalIndent(w.conf.Genesis, "", " ")
|
||||||
infos.network = w.conf.Genesis.Config.ChainId.Int64()
|
infos.network = w.conf.Genesis.Config.ChainID.Int64()
|
||||||
|
|
||||||
// Figure out which port to listen on
|
// Figure out which port to listen on
|
||||||
fmt.Println()
|
fmt.Println()
|
||||||
|
@@ -24,6 +24,7 @@ import (
|
|||||||
"reflect"
|
"reflect"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
"unicode"
|
"unicode"
|
||||||
|
|
||||||
cli "gopkg.in/urfave/cli.v1"
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
@@ -58,19 +59,25 @@ var (
|
|||||||
|
|
||||||
//constants for environment variables
|
//constants for environment variables
|
||||||
const (
|
const (
|
||||||
SWARM_ENV_CHEQUEBOOK_ADDR = "SWARM_CHEQUEBOOK_ADDR"
|
SWARM_ENV_CHEQUEBOOK_ADDR = "SWARM_CHEQUEBOOK_ADDR"
|
||||||
SWARM_ENV_ACCOUNT = "SWARM_ACCOUNT"
|
SWARM_ENV_ACCOUNT = "SWARM_ACCOUNT"
|
||||||
SWARM_ENV_LISTEN_ADDR = "SWARM_LISTEN_ADDR"
|
SWARM_ENV_LISTEN_ADDR = "SWARM_LISTEN_ADDR"
|
||||||
SWARM_ENV_PORT = "SWARM_PORT"
|
SWARM_ENV_PORT = "SWARM_PORT"
|
||||||
SWARM_ENV_NETWORK_ID = "SWARM_NETWORK_ID"
|
SWARM_ENV_NETWORK_ID = "SWARM_NETWORK_ID"
|
||||||
SWARM_ENV_SWAP_ENABLE = "SWARM_SWAP_ENABLE"
|
SWARM_ENV_SWAP_ENABLE = "SWARM_SWAP_ENABLE"
|
||||||
SWARM_ENV_SWAP_API = "SWARM_SWAP_API"
|
SWARM_ENV_SWAP_API = "SWARM_SWAP_API"
|
||||||
SWARM_ENV_SYNC_ENABLE = "SWARM_SYNC_ENABLE"
|
SWARM_ENV_SYNC_DISABLE = "SWARM_SYNC_DISABLE"
|
||||||
SWARM_ENV_ENS_API = "SWARM_ENS_API"
|
SWARM_ENV_SYNC_UPDATE_DELAY = "SWARM_ENV_SYNC_UPDATE_DELAY"
|
||||||
SWARM_ENV_ENS_ADDR = "SWARM_ENS_ADDR"
|
SWARM_ENV_DELIVERY_SKIP_CHECK = "SWARM_DELIVERY_SKIP_CHECK"
|
||||||
SWARM_ENV_CORS = "SWARM_CORS"
|
SWARM_ENV_ENS_API = "SWARM_ENS_API"
|
||||||
SWARM_ENV_BOOTNODES = "SWARM_BOOTNODES"
|
SWARM_ENV_ENS_ADDR = "SWARM_ENS_ADDR"
|
||||||
GETH_ENV_DATADIR = "GETH_DATADIR"
|
SWARM_ENV_CORS = "SWARM_CORS"
|
||||||
|
SWARM_ENV_BOOTNODES = "SWARM_BOOTNODES"
|
||||||
|
SWARM_ENV_PSS_ENABLE = "SWARM_PSS_ENABLE"
|
||||||
|
SWARM_ENV_STORE_PATH = "SWARM_STORE_PATH"
|
||||||
|
SWARM_ENV_STORE_CAPACITY = "SWARM_STORE_CAPACITY"
|
||||||
|
SWARM_ENV_STORE_CACHE_CAPACITY = "SWARM_STORE_CACHE_CAPACITY"
|
||||||
|
GETH_ENV_DATADIR = "GETH_DATADIR"
|
||||||
)
|
)
|
||||||
|
|
||||||
// These settings ensure that TOML keys use the same names as Go struct fields.
|
// These settings ensure that TOML keys use the same names as Go struct fields.
|
||||||
@@ -92,10 +99,8 @@ var tomlSettings = toml.Config{
|
|||||||
|
|
||||||
//before booting the swarm node, build the configuration
|
//before booting the swarm node, build the configuration
|
||||||
func buildConfig(ctx *cli.Context) (config *bzzapi.Config, err error) {
|
func buildConfig(ctx *cli.Context) (config *bzzapi.Config, err error) {
|
||||||
//check for deprecated flags
|
|
||||||
checkDeprecated(ctx)
|
|
||||||
//start by creating a default config
|
//start by creating a default config
|
||||||
config = bzzapi.NewDefaultConfig()
|
config = bzzapi.NewConfig()
|
||||||
//first load settings from config file (if provided)
|
//first load settings from config file (if provided)
|
||||||
config, err = configFileOverride(config, ctx)
|
config, err = configFileOverride(config, ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -168,7 +173,7 @@ func cmdLineOverride(currentConfig *bzzapi.Config, ctx *cli.Context) *bzzapi.Con
|
|||||||
|
|
||||||
if networkid := ctx.GlobalString(SwarmNetworkIdFlag.Name); networkid != "" {
|
if networkid := ctx.GlobalString(SwarmNetworkIdFlag.Name); networkid != "" {
|
||||||
if id, _ := strconv.Atoi(networkid); id != 0 {
|
if id, _ := strconv.Atoi(networkid); id != 0 {
|
||||||
currentConfig.NetworkId = uint64(id)
|
currentConfig.NetworkID = uint64(id)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -191,12 +196,20 @@ func cmdLineOverride(currentConfig *bzzapi.Config, ctx *cli.Context) *bzzapi.Con
|
|||||||
currentConfig.SwapEnabled = true
|
currentConfig.SwapEnabled = true
|
||||||
}
|
}
|
||||||
|
|
||||||
if ctx.GlobalIsSet(SwarmSyncEnabledFlag.Name) {
|
if ctx.GlobalIsSet(SwarmSyncDisabledFlag.Name) {
|
||||||
currentConfig.SyncEnabled = true
|
currentConfig.SyncEnabled = false
|
||||||
}
|
}
|
||||||
|
|
||||||
currentConfig.SwapApi = ctx.GlobalString(SwarmSwapAPIFlag.Name)
|
if d := ctx.GlobalDuration(SwarmSyncUpdateDelay.Name); d > 0 {
|
||||||
if currentConfig.SwapEnabled && currentConfig.SwapApi == "" {
|
currentConfig.SyncUpdateDelay = d
|
||||||
|
}
|
||||||
|
|
||||||
|
if ctx.GlobalIsSet(SwarmDeliverySkipCheckFlag.Name) {
|
||||||
|
currentConfig.DeliverySkipCheck = true
|
||||||
|
}
|
||||||
|
|
||||||
|
currentConfig.SwapAPI = ctx.GlobalString(SwarmSwapAPIFlag.Name)
|
||||||
|
if currentConfig.SwapEnabled && currentConfig.SwapAPI == "" {
|
||||||
utils.Fatalf(SWARM_ERR_SWAP_SET_NO_API)
|
utils.Fatalf(SWARM_ERR_SWAP_SET_NO_API)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -209,10 +222,6 @@ func cmdLineOverride(currentConfig *bzzapi.Config, ctx *cli.Context) *bzzapi.Con
|
|||||||
currentConfig.EnsAPIs = ensAPIs
|
currentConfig.EnsAPIs = ensAPIs
|
||||||
}
|
}
|
||||||
|
|
||||||
if ensaddr := ctx.GlobalString(DeprecatedEnsAddrFlag.Name); ensaddr != "" {
|
|
||||||
currentConfig.EnsRoot = common.HexToAddress(ensaddr)
|
|
||||||
}
|
|
||||||
|
|
||||||
if cors := ctx.GlobalString(CorsStringFlag.Name); cors != "" {
|
if cors := ctx.GlobalString(CorsStringFlag.Name); cors != "" {
|
||||||
currentConfig.Cors = cors
|
currentConfig.Cors = cors
|
||||||
}
|
}
|
||||||
@@ -221,6 +230,18 @@ func cmdLineOverride(currentConfig *bzzapi.Config, ctx *cli.Context) *bzzapi.Con
|
|||||||
currentConfig.BootNodes = ctx.GlobalString(utils.BootnodesFlag.Name)
|
currentConfig.BootNodes = ctx.GlobalString(utils.BootnodesFlag.Name)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if storePath := ctx.GlobalString(SwarmStorePath.Name); storePath != "" {
|
||||||
|
currentConfig.LocalStoreParams.ChunkDbPath = storePath
|
||||||
|
}
|
||||||
|
|
||||||
|
if storeCapacity := ctx.GlobalUint64(SwarmStoreCapacity.Name); storeCapacity != 0 {
|
||||||
|
currentConfig.LocalStoreParams.DbCapacity = storeCapacity
|
||||||
|
}
|
||||||
|
|
||||||
|
if storeCacheCapacity := ctx.GlobalUint(SwarmStoreCacheCapacity.Name); storeCacheCapacity != 0 {
|
||||||
|
currentConfig.LocalStoreParams.CacheCapacity = storeCacheCapacity
|
||||||
|
}
|
||||||
|
|
||||||
return currentConfig
|
return currentConfig
|
||||||
|
|
||||||
}
|
}
|
||||||
@@ -239,7 +260,7 @@ func envVarsOverride(currentConfig *bzzapi.Config) (config *bzzapi.Config) {
|
|||||||
|
|
||||||
if networkid := os.Getenv(SWARM_ENV_NETWORK_ID); networkid != "" {
|
if networkid := os.Getenv(SWARM_ENV_NETWORK_ID); networkid != "" {
|
||||||
if id, _ := strconv.Atoi(networkid); id != 0 {
|
if id, _ := strconv.Atoi(networkid); id != 0 {
|
||||||
currentConfig.NetworkId = uint64(id)
|
currentConfig.NetworkID = uint64(id)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -262,17 +283,29 @@ func envVarsOverride(currentConfig *bzzapi.Config) (config *bzzapi.Config) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if syncenable := os.Getenv(SWARM_ENV_SYNC_ENABLE); syncenable != "" {
|
if syncdisable := os.Getenv(SWARM_ENV_SYNC_DISABLE); syncdisable != "" {
|
||||||
if sync, err := strconv.ParseBool(syncenable); err != nil {
|
if sync, err := strconv.ParseBool(syncdisable); err != nil {
|
||||||
currentConfig.SyncEnabled = sync
|
currentConfig.SyncEnabled = !sync
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if v := os.Getenv(SWARM_ENV_DELIVERY_SKIP_CHECK); v != "" {
|
||||||
|
if skipCheck, err := strconv.ParseBool(v); err != nil {
|
||||||
|
currentConfig.DeliverySkipCheck = skipCheck
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if v := os.Getenv(SWARM_ENV_SYNC_UPDATE_DELAY); v != "" {
|
||||||
|
if d, err := time.ParseDuration(v); err != nil {
|
||||||
|
currentConfig.SyncUpdateDelay = d
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if swapapi := os.Getenv(SWARM_ENV_SWAP_API); swapapi != "" {
|
if swapapi := os.Getenv(SWARM_ENV_SWAP_API); swapapi != "" {
|
||||||
currentConfig.SwapApi = swapapi
|
currentConfig.SwapAPI = swapapi
|
||||||
}
|
}
|
||||||
|
|
||||||
if currentConfig.SwapEnabled && currentConfig.SwapApi == "" {
|
if currentConfig.SwapEnabled && currentConfig.SwapAPI == "" {
|
||||||
utils.Fatalf(SWARM_ERR_SWAP_SET_NO_API)
|
utils.Fatalf(SWARM_ERR_SWAP_SET_NO_API)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -312,18 +345,6 @@ func dumpConfig(ctx *cli.Context) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
//deprecated flags checked here
|
|
||||||
func checkDeprecated(ctx *cli.Context) {
|
|
||||||
// exit if the deprecated --ethapi flag is set
|
|
||||||
if ctx.GlobalString(DeprecatedEthAPIFlag.Name) != "" {
|
|
||||||
utils.Fatalf("--ethapi is no longer a valid command line flag, please use --ens-api and/or --swap-api.")
|
|
||||||
}
|
|
||||||
// warn if --ens-api flag is set
|
|
||||||
if ctx.GlobalString(DeprecatedEnsAddrFlag.Name) != "" {
|
|
||||||
log.Warn("--ens-addr is no longer a valid command line flag, please use --ens-api to specify contract address.")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//validate configuration parameters
|
//validate configuration parameters
|
||||||
func validateConfig(cfg *bzzapi.Config) (err error) {
|
func validateConfig(cfg *bzzapi.Config) (err error) {
|
||||||
for _, ensAPI := range cfg.EnsAPIs {
|
for _, ensAPI := range cfg.EnsAPIs {
|
||||||
|
@@ -34,7 +34,7 @@ import (
|
|||||||
|
|
||||||
func TestDumpConfig(t *testing.T) {
|
func TestDumpConfig(t *testing.T) {
|
||||||
swarm := runSwarm(t, "dumpconfig")
|
swarm := runSwarm(t, "dumpconfig")
|
||||||
defaultConf := api.NewDefaultConfig()
|
defaultConf := api.NewConfig()
|
||||||
out, err := tomlSettings.Marshal(&defaultConf)
|
out, err := tomlSettings.Marshal(&defaultConf)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@@ -43,7 +43,7 @@ func TestDumpConfig(t *testing.T) {
|
|||||||
swarm.ExpectExit()
|
swarm.ExpectExit()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFailsSwapEnabledNoSwapApi(t *testing.T) {
|
func TestConfigFailsSwapEnabledNoSwapApi(t *testing.T) {
|
||||||
flags := []string{
|
flags := []string{
|
||||||
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
||||||
fmt.Sprintf("--%s", SwarmPortFlag.Name), "54545",
|
fmt.Sprintf("--%s", SwarmPortFlag.Name), "54545",
|
||||||
@@ -55,7 +55,7 @@ func TestFailsSwapEnabledNoSwapApi(t *testing.T) {
|
|||||||
swarm.ExpectExit()
|
swarm.ExpectExit()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFailsNoBzzAccount(t *testing.T) {
|
func TestConfigFailsNoBzzAccount(t *testing.T) {
|
||||||
flags := []string{
|
flags := []string{
|
||||||
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
||||||
fmt.Sprintf("--%s", SwarmPortFlag.Name), "54545",
|
fmt.Sprintf("--%s", SwarmPortFlag.Name), "54545",
|
||||||
@@ -66,7 +66,7 @@ func TestFailsNoBzzAccount(t *testing.T) {
|
|||||||
swarm.ExpectExit()
|
swarm.ExpectExit()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCmdLineOverrides(t *testing.T) {
|
func TestConfigCmdLineOverrides(t *testing.T) {
|
||||||
dir, err := ioutil.TempDir("", "bzztest")
|
dir, err := ioutil.TempDir("", "bzztest")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
@@ -85,9 +85,10 @@ func TestCmdLineOverrides(t *testing.T) {
|
|||||||
flags := []string{
|
flags := []string{
|
||||||
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "42",
|
||||||
fmt.Sprintf("--%s", SwarmPortFlag.Name), httpPort,
|
fmt.Sprintf("--%s", SwarmPortFlag.Name), httpPort,
|
||||||
fmt.Sprintf("--%s", SwarmSyncEnabledFlag.Name),
|
fmt.Sprintf("--%s", SwarmSyncDisabledFlag.Name),
|
||||||
fmt.Sprintf("--%s", CorsStringFlag.Name), "*",
|
fmt.Sprintf("--%s", CorsStringFlag.Name), "*",
|
||||||
fmt.Sprintf("--%s", SwarmAccountFlag.Name), account.Address.String(),
|
fmt.Sprintf("--%s", SwarmAccountFlag.Name), account.Address.String(),
|
||||||
|
fmt.Sprintf("--%s", SwarmDeliverySkipCheckFlag.Name),
|
||||||
fmt.Sprintf("--%s", EnsAPIFlag.Name), "",
|
fmt.Sprintf("--%s", EnsAPIFlag.Name), "",
|
||||||
"--datadir", dir,
|
"--datadir", dir,
|
||||||
"--ipcpath", conf.IPCPath,
|
"--ipcpath", conf.IPCPath,
|
||||||
@@ -120,12 +121,16 @@ func TestCmdLineOverrides(t *testing.T) {
|
|||||||
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.NetworkId != 42 {
|
if info.NetworkID != 42 {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", 42, info.NetworkId)
|
t.Fatalf("Expected network ID to be %d, got %d", 42, info.NetworkID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if !info.SyncEnabled {
|
if info.SyncEnabled {
|
||||||
t.Fatal("Expected Sync to be enabled, but is false")
|
t.Fatal("Expected Sync to be disabled, but is true")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !info.DeliverySkipCheck {
|
||||||
|
t.Fatal("Expected DeliverySkipCheck to be enabled, but it is not")
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.Cors != "*" {
|
if info.Cors != "*" {
|
||||||
@@ -135,7 +140,7 @@ func TestCmdLineOverrides(t *testing.T) {
|
|||||||
node.Shutdown()
|
node.Shutdown()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFileOverrides(t *testing.T) {
|
func TestConfigFileOverrides(t *testing.T) {
|
||||||
|
|
||||||
// assign ports
|
// assign ports
|
||||||
httpPort, err := assignTCPPort()
|
httpPort, err := assignTCPPort()
|
||||||
@@ -145,16 +150,16 @@ func TestFileOverrides(t *testing.T) {
|
|||||||
|
|
||||||
//create a config file
|
//create a config file
|
||||||
//first, create a default conf
|
//first, create a default conf
|
||||||
defaultConf := api.NewDefaultConfig()
|
defaultConf := api.NewConfig()
|
||||||
//change some values in order to test if they have been loaded
|
//change some values in order to test if they have been loaded
|
||||||
defaultConf.SyncEnabled = true
|
defaultConf.SyncEnabled = false
|
||||||
defaultConf.NetworkId = 54
|
defaultConf.DeliverySkipCheck = true
|
||||||
|
defaultConf.NetworkID = 54
|
||||||
defaultConf.Port = httpPort
|
defaultConf.Port = httpPort
|
||||||
defaultConf.StoreParams.DbCapacity = 9000000
|
defaultConf.DbCapacity = 9000000
|
||||||
defaultConf.ChunkerParams.Branches = 64
|
defaultConf.HiveParams.KeepAliveInterval = 6000000000
|
||||||
defaultConf.HiveParams.CallInterval = 6000000000
|
|
||||||
defaultConf.Swap.Params.Strategy.AutoCashInterval = 600 * time.Second
|
defaultConf.Swap.Params.Strategy.AutoCashInterval = 600 * time.Second
|
||||||
defaultConf.SyncParams.KeyBufferSize = 512
|
//defaultConf.SyncParams.KeyBufferSize = 512
|
||||||
//create a TOML string
|
//create a TOML string
|
||||||
out, err := tomlSettings.Marshal(&defaultConf)
|
out, err := tomlSettings.Marshal(&defaultConf)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -215,38 +220,38 @@ func TestFileOverrides(t *testing.T) {
|
|||||||
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.NetworkId != 54 {
|
if info.NetworkID != 54 {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", 54, info.NetworkId)
|
t.Fatalf("Expected network ID to be %d, got %d", 54, info.NetworkID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if !info.SyncEnabled {
|
if info.SyncEnabled {
|
||||||
t.Fatal("Expected Sync to be enabled, but is false")
|
t.Fatal("Expected Sync to be disabled, but is true")
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.StoreParams.DbCapacity != 9000000 {
|
if !info.DeliverySkipCheck {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", 54, info.NetworkId)
|
t.Fatal("Expected DeliverySkipCheck to be enabled, but it is not")
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.ChunkerParams.Branches != 64 {
|
if info.DbCapacity != 9000000 {
|
||||||
t.Fatalf("Expected chunker params branches to be %d, got %d", 64, info.ChunkerParams.Branches)
|
t.Fatalf("Expected network ID to be %d, got %d", 54, info.NetworkID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.HiveParams.CallInterval != 6000000000 {
|
if info.HiveParams.KeepAliveInterval != 6000000000 {
|
||||||
t.Fatalf("Expected HiveParams CallInterval to be %d, got %d", uint64(6000000000), uint64(info.HiveParams.CallInterval))
|
t.Fatalf("Expected HiveParams KeepAliveInterval to be %d, got %d", uint64(6000000000), uint64(info.HiveParams.KeepAliveInterval))
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.Swap.Params.Strategy.AutoCashInterval != 600*time.Second {
|
if info.Swap.Params.Strategy.AutoCashInterval != 600*time.Second {
|
||||||
t.Fatalf("Expected SwapParams AutoCashInterval to be %ds, got %d", 600, info.Swap.Params.Strategy.AutoCashInterval)
|
t.Fatalf("Expected SwapParams AutoCashInterval to be %ds, got %d", 600, info.Swap.Params.Strategy.AutoCashInterval)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.SyncParams.KeyBufferSize != 512 {
|
// if info.SyncParams.KeyBufferSize != 512 {
|
||||||
t.Fatalf("Expected info.SyncParams.KeyBufferSize to be %d, got %d", 512, info.SyncParams.KeyBufferSize)
|
// t.Fatalf("Expected info.SyncParams.KeyBufferSize to be %d, got %d", 512, info.SyncParams.KeyBufferSize)
|
||||||
}
|
// }
|
||||||
|
|
||||||
node.Shutdown()
|
node.Shutdown()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestEnvVars(t *testing.T) {
|
func TestConfigEnvVars(t *testing.T) {
|
||||||
// assign ports
|
// assign ports
|
||||||
httpPort, err := assignTCPPort()
|
httpPort, err := assignTCPPort()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -257,7 +262,8 @@ func TestEnvVars(t *testing.T) {
|
|||||||
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmPortFlag.EnvVar, httpPort))
|
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmPortFlag.EnvVar, httpPort))
|
||||||
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmNetworkIdFlag.EnvVar, "999"))
|
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmNetworkIdFlag.EnvVar, "999"))
|
||||||
envVars = append(envVars, fmt.Sprintf("%s=%s", CorsStringFlag.EnvVar, "*"))
|
envVars = append(envVars, fmt.Sprintf("%s=%s", CorsStringFlag.EnvVar, "*"))
|
||||||
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmSyncEnabledFlag.EnvVar, "true"))
|
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmSyncDisabledFlag.EnvVar, "true"))
|
||||||
|
envVars = append(envVars, fmt.Sprintf("%s=%s", SwarmDeliverySkipCheckFlag.EnvVar, "true"))
|
||||||
|
|
||||||
dir, err := ioutil.TempDir("", "bzztest")
|
dir, err := ioutil.TempDir("", "bzztest")
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -326,23 +332,27 @@ func TestEnvVars(t *testing.T) {
|
|||||||
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.NetworkId != 999 {
|
if info.NetworkID != 999 {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", 999, info.NetworkId)
|
t.Fatalf("Expected network ID to be %d, got %d", 999, info.NetworkID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.Cors != "*" {
|
if info.Cors != "*" {
|
||||||
t.Fatalf("Expected Cors flag to be set to %s, got %s", "*", info.Cors)
|
t.Fatalf("Expected Cors flag to be set to %s, got %s", "*", info.Cors)
|
||||||
}
|
}
|
||||||
|
|
||||||
if !info.SyncEnabled {
|
if info.SyncEnabled {
|
||||||
t.Fatal("Expected Sync to be enabled, but is false")
|
t.Fatal("Expected Sync to be disabled, but is true")
|
||||||
|
}
|
||||||
|
|
||||||
|
if !info.DeliverySkipCheck {
|
||||||
|
t.Fatal("Expected DeliverySkipCheck to be enabled, but it is not")
|
||||||
}
|
}
|
||||||
|
|
||||||
node.Shutdown()
|
node.Shutdown()
|
||||||
cmd.Process.Kill()
|
cmd.Process.Kill()
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestCmdLineOverridesFile(t *testing.T) {
|
func TestConfigCmdLineOverridesFile(t *testing.T) {
|
||||||
|
|
||||||
// assign ports
|
// assign ports
|
||||||
httpPort, err := assignTCPPort()
|
httpPort, err := assignTCPPort()
|
||||||
@@ -352,26 +362,27 @@ func TestCmdLineOverridesFile(t *testing.T) {
|
|||||||
|
|
||||||
//create a config file
|
//create a config file
|
||||||
//first, create a default conf
|
//first, create a default conf
|
||||||
defaultConf := api.NewDefaultConfig()
|
defaultConf := api.NewConfig()
|
||||||
//change some values in order to test if they have been loaded
|
//change some values in order to test if they have been loaded
|
||||||
defaultConf.SyncEnabled = false
|
defaultConf.SyncEnabled = true
|
||||||
defaultConf.NetworkId = 54
|
defaultConf.NetworkID = 54
|
||||||
defaultConf.Port = "8588"
|
defaultConf.Port = "8588"
|
||||||
defaultConf.StoreParams.DbCapacity = 9000000
|
defaultConf.DbCapacity = 9000000
|
||||||
defaultConf.ChunkerParams.Branches = 64
|
defaultConf.HiveParams.KeepAliveInterval = 6000000000
|
||||||
defaultConf.HiveParams.CallInterval = 6000000000
|
|
||||||
defaultConf.Swap.Params.Strategy.AutoCashInterval = 600 * time.Second
|
defaultConf.Swap.Params.Strategy.AutoCashInterval = 600 * time.Second
|
||||||
defaultConf.SyncParams.KeyBufferSize = 512
|
//defaultConf.SyncParams.KeyBufferSize = 512
|
||||||
//create a TOML file
|
//create a TOML file
|
||||||
out, err := tomlSettings.Marshal(&defaultConf)
|
out, err := tomlSettings.Marshal(&defaultConf)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Error creating TOML file in TestFileOverride: %v", err)
|
t.Fatalf("Error creating TOML file in TestFileOverride: %v", err)
|
||||||
}
|
}
|
||||||
//write file
|
//write file
|
||||||
f, err := ioutil.TempFile("", "testconfig.toml")
|
fname := "testconfig.toml"
|
||||||
|
f, err := ioutil.TempFile("", fname)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatalf("Error writing TOML file in TestFileOverride: %v", err)
|
t.Fatalf("Error writing TOML file in TestFileOverride: %v", err)
|
||||||
}
|
}
|
||||||
|
defer os.Remove(fname)
|
||||||
//write file
|
//write file
|
||||||
_, err = f.WriteString(string(out))
|
_, err = f.WriteString(string(out))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -392,7 +403,7 @@ func TestCmdLineOverridesFile(t *testing.T) {
|
|||||||
flags := []string{
|
flags := []string{
|
||||||
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "77",
|
fmt.Sprintf("--%s", SwarmNetworkIdFlag.Name), "77",
|
||||||
fmt.Sprintf("--%s", SwarmPortFlag.Name), httpPort,
|
fmt.Sprintf("--%s", SwarmPortFlag.Name), httpPort,
|
||||||
fmt.Sprintf("--%s", SwarmSyncEnabledFlag.Name),
|
fmt.Sprintf("--%s", SwarmSyncDisabledFlag.Name),
|
||||||
fmt.Sprintf("--%s", SwarmTomlConfigPathFlag.Name), f.Name(),
|
fmt.Sprintf("--%s", SwarmTomlConfigPathFlag.Name), f.Name(),
|
||||||
fmt.Sprintf("--%s", SwarmAccountFlag.Name), account.Address.String(),
|
fmt.Sprintf("--%s", SwarmAccountFlag.Name), account.Address.String(),
|
||||||
"--ens-api", "",
|
"--ens-api", "",
|
||||||
@@ -427,33 +438,29 @@ func TestCmdLineOverridesFile(t *testing.T) {
|
|||||||
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
t.Fatalf("Expected port to be %s, got %s", httpPort, info.Port)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.NetworkId != expectNetworkId {
|
if info.NetworkID != expectNetworkId {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", expectNetworkId, info.NetworkId)
|
t.Fatalf("Expected network ID to be %d, got %d", expectNetworkId, info.NetworkID)
|
||||||
}
|
}
|
||||||
|
|
||||||
if !info.SyncEnabled {
|
if info.SyncEnabled {
|
||||||
t.Fatal("Expected Sync to be enabled, but is false")
|
t.Fatal("Expected Sync to be disabled, but is true")
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.StoreParams.DbCapacity != 9000000 {
|
if info.LocalStoreParams.DbCapacity != 9000000 {
|
||||||
t.Fatalf("Expected network ID to be %d, got %d", 54, info.NetworkId)
|
t.Fatalf("Expected Capacity to be %d, got %d", 9000000, info.LocalStoreParams.DbCapacity)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.ChunkerParams.Branches != 64 {
|
if info.HiveParams.KeepAliveInterval != 6000000000 {
|
||||||
t.Fatalf("Expected chunker params branches to be %d, got %d", 64, info.ChunkerParams.Branches)
|
t.Fatalf("Expected HiveParams KeepAliveInterval to be %d, got %d", uint64(6000000000), uint64(info.HiveParams.KeepAliveInterval))
|
||||||
}
|
|
||||||
|
|
||||||
if info.HiveParams.CallInterval != 6000000000 {
|
|
||||||
t.Fatalf("Expected HiveParams CallInterval to be %d, got %d", uint64(6000000000), uint64(info.HiveParams.CallInterval))
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.Swap.Params.Strategy.AutoCashInterval != 600*time.Second {
|
if info.Swap.Params.Strategy.AutoCashInterval != 600*time.Second {
|
||||||
t.Fatalf("Expected SwapParams AutoCashInterval to be %ds, got %d", 600, info.Swap.Params.Strategy.AutoCashInterval)
|
t.Fatalf("Expected SwapParams AutoCashInterval to be %ds, got %d", 600, info.Swap.Params.Strategy.AutoCashInterval)
|
||||||
}
|
}
|
||||||
|
|
||||||
if info.SyncParams.KeyBufferSize != 512 {
|
// if info.SyncParams.KeyBufferSize != 512 {
|
||||||
t.Fatalf("Expected info.SyncParams.KeyBufferSize to be %d, got %d", 512, info.SyncParams.KeyBufferSize)
|
// t.Fatalf("Expected info.SyncParams.KeyBufferSize to be %d, got %d", 512, info.SyncParams.KeyBufferSize)
|
||||||
}
|
// }
|
||||||
|
|
||||||
node.Shutdown()
|
node.Shutdown()
|
||||||
}
|
}
|
||||||
|
@@ -23,6 +23,7 @@ import (
|
|||||||
"path/filepath"
|
"path/filepath"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/cmd/utils"
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/swarm/storage"
|
"github.com/ethereum/go-ethereum/swarm/storage"
|
||||||
"gopkg.in/urfave/cli.v1"
|
"gopkg.in/urfave/cli.v1"
|
||||||
@@ -30,11 +31,11 @@ import (
|
|||||||
|
|
||||||
func dbExport(ctx *cli.Context) {
|
func dbExport(ctx *cli.Context) {
|
||||||
args := ctx.Args()
|
args := ctx.Args()
|
||||||
if len(args) != 2 {
|
if len(args) != 3 {
|
||||||
utils.Fatalf("invalid arguments, please specify both <chunkdb> (path to a local chunk database) and <file> (path to write the tar archive to, - for stdout)")
|
utils.Fatalf("invalid arguments, please specify both <chunkdb> (path to a local chunk database), <file> (path to write the tar archive to, - for stdout) and the base key")
|
||||||
}
|
}
|
||||||
|
|
||||||
store, err := openDbStore(args[0])
|
store, err := openLDBStore(args[0], common.Hex2Bytes(args[2]))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("error opening local chunk database: %s", err)
|
utils.Fatalf("error opening local chunk database: %s", err)
|
||||||
}
|
}
|
||||||
@@ -62,11 +63,11 @@ func dbExport(ctx *cli.Context) {
|
|||||||
|
|
||||||
func dbImport(ctx *cli.Context) {
|
func dbImport(ctx *cli.Context) {
|
||||||
args := ctx.Args()
|
args := ctx.Args()
|
||||||
if len(args) != 2 {
|
if len(args) != 3 {
|
||||||
utils.Fatalf("invalid arguments, please specify both <chunkdb> (path to a local chunk database) and <file> (path to read the tar archive from, - for stdin)")
|
utils.Fatalf("invalid arguments, please specify both <chunkdb> (path to a local chunk database), <file> (path to read the tar archive from, - for stdin) and the base key")
|
||||||
}
|
}
|
||||||
|
|
||||||
store, err := openDbStore(args[0])
|
store, err := openLDBStore(args[0], common.Hex2Bytes(args[2]))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("error opening local chunk database: %s", err)
|
utils.Fatalf("error opening local chunk database: %s", err)
|
||||||
}
|
}
|
||||||
@@ -94,11 +95,11 @@ func dbImport(ctx *cli.Context) {
|
|||||||
|
|
||||||
func dbClean(ctx *cli.Context) {
|
func dbClean(ctx *cli.Context) {
|
||||||
args := ctx.Args()
|
args := ctx.Args()
|
||||||
if len(args) != 1 {
|
if len(args) != 2 {
|
||||||
utils.Fatalf("invalid arguments, please specify <chunkdb> (path to a local chunk database)")
|
utils.Fatalf("invalid arguments, please specify <chunkdb> (path to a local chunk database) and the base key")
|
||||||
}
|
}
|
||||||
|
|
||||||
store, err := openDbStore(args[0])
|
store, err := openLDBStore(args[0], common.Hex2Bytes(args[1]))
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("error opening local chunk database: %s", err)
|
utils.Fatalf("error opening local chunk database: %s", err)
|
||||||
}
|
}
|
||||||
@@ -107,10 +108,13 @@ func dbClean(ctx *cli.Context) {
|
|||||||
store.Cleanup()
|
store.Cleanup()
|
||||||
}
|
}
|
||||||
|
|
||||||
func openDbStore(path string) (*storage.DbStore, error) {
|
func openLDBStore(path string, basekey []byte) (*storage.LDBStore, error) {
|
||||||
if _, err := os.Stat(filepath.Join(path, "CURRENT")); err != nil {
|
if _, err := os.Stat(filepath.Join(path, "CURRENT")); err != nil {
|
||||||
return nil, fmt.Errorf("invalid chunkdb path: %s", err)
|
return nil, fmt.Errorf("invalid chunkdb path: %s", err)
|
||||||
}
|
}
|
||||||
hash := storage.MakeHashFunc("SHA3")
|
|
||||||
return storage.NewDbStore(path, hash, 10000000, 0)
|
storeparams := storage.NewDefaultStoreParams()
|
||||||
|
ldbparams := storage.NewLDBStoreParams(storeparams, path)
|
||||||
|
ldbparams.BaseKey = basekey
|
||||||
|
return storage.NewLDBStore(ldbparams)
|
||||||
}
|
}
|
||||||
|
85
cmd/swarm/download.go
Normal file
85
cmd/swarm/download.go
Normal file
@@ -0,0 +1,85 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/api"
|
||||||
|
swarm "github.com/ethereum/go-ethereum/swarm/api/client"
|
||||||
|
"gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
func download(ctx *cli.Context) {
|
||||||
|
log.Debug("downloading content using swarm down")
|
||||||
|
args := ctx.Args()
|
||||||
|
dest := "."
|
||||||
|
|
||||||
|
switch len(args) {
|
||||||
|
case 0:
|
||||||
|
utils.Fatalf("Usage: swarm down [options] <bzz locator> [<destination path>]")
|
||||||
|
case 1:
|
||||||
|
log.Trace(fmt.Sprintf("swarm down: no destination path - assuming working dir"))
|
||||||
|
default:
|
||||||
|
log.Trace(fmt.Sprintf("destination path arg: %s", args[1]))
|
||||||
|
if absDest, err := filepath.Abs(args[1]); err == nil {
|
||||||
|
dest = absDest
|
||||||
|
} else {
|
||||||
|
utils.Fatalf("could not get download path: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var (
|
||||||
|
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
||||||
|
isRecursive = ctx.Bool(SwarmRecursiveFlag.Name)
|
||||||
|
client = swarm.NewClient(bzzapi)
|
||||||
|
)
|
||||||
|
|
||||||
|
if fi, err := os.Stat(dest); err == nil {
|
||||||
|
if isRecursive && !fi.Mode().IsDir() {
|
||||||
|
utils.Fatalf("destination path is not a directory!")
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
if !os.IsNotExist(err) {
|
||||||
|
utils.Fatalf("could not stat path: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
uri, err := api.Parse(args[0])
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("could not parse uri argument: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// assume behaviour according to --recursive switch
|
||||||
|
if isRecursive {
|
||||||
|
if err := client.DownloadDirectory(uri.Addr, uri.Path, dest); err != nil {
|
||||||
|
utils.Fatalf("encoutered an error while downloading directory: %v", err)
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
// we are downloading a file
|
||||||
|
log.Debug(fmt.Sprintf("downloading file/path from a manifest. hash: %s, path:%s", uri.Addr, uri.Path))
|
||||||
|
|
||||||
|
err := client.DownloadFile(uri.Addr, uri.Path, dest)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("could not download %s from given address: %s. error: %v", uri.Path, uri.Addr, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
139
cmd/swarm/export_test.go
Normal file
139
cmd/swarm/export_test.go
Normal file
@@ -0,0 +1,139 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"crypto/md5"
|
||||||
|
"crypto/rand"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"net/http"
|
||||||
|
"os"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/swarm"
|
||||||
|
)
|
||||||
|
|
||||||
|
// TestCLISwarmExportImport perform the following test:
|
||||||
|
// 1. runs swarm node
|
||||||
|
// 2. uploads a random file
|
||||||
|
// 3. runs an export of the local datastore
|
||||||
|
// 4. runs a second swarm node
|
||||||
|
// 5. imports the exported datastore
|
||||||
|
// 6. fetches the uploaded random file from the second node
|
||||||
|
func TestCLISwarmExportImport(t *testing.T) {
|
||||||
|
cluster := newTestCluster(t, 1)
|
||||||
|
|
||||||
|
// generate random 10mb file
|
||||||
|
f, cleanup := generateRandomFile(t, 10000000)
|
||||||
|
defer cleanup()
|
||||||
|
|
||||||
|
// upload the file with 'swarm up' and expect a hash
|
||||||
|
up := runSwarm(t, "--bzzapi", cluster.Nodes[0].URL, "up", f.Name())
|
||||||
|
_, matches := up.ExpectRegexp(`[a-f\d]{64}`)
|
||||||
|
up.ExpectExit()
|
||||||
|
hash := matches[0]
|
||||||
|
|
||||||
|
var info swarm.Info
|
||||||
|
if err := cluster.Nodes[0].Client.Call(&info, "bzz_info"); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
cluster.Stop()
|
||||||
|
defer cluster.Cleanup()
|
||||||
|
|
||||||
|
// generate an export.tar
|
||||||
|
exportCmd := runSwarm(t, "db", "export", info.Path+"/chunks", info.Path+"/export.tar", strings.TrimPrefix(info.BzzKey, "0x"))
|
||||||
|
exportCmd.ExpectExit()
|
||||||
|
|
||||||
|
// start second cluster
|
||||||
|
cluster2 := newTestCluster(t, 1)
|
||||||
|
|
||||||
|
var info2 swarm.Info
|
||||||
|
if err := cluster2.Nodes[0].Client.Call(&info2, "bzz_info"); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// stop second cluster, so that we close LevelDB
|
||||||
|
cluster2.Stop()
|
||||||
|
defer cluster2.Cleanup()
|
||||||
|
|
||||||
|
// import the export.tar
|
||||||
|
importCmd := runSwarm(t, "db", "import", info2.Path+"/chunks", info.Path+"/export.tar", strings.TrimPrefix(info2.BzzKey, "0x"))
|
||||||
|
importCmd.ExpectExit()
|
||||||
|
|
||||||
|
// spin second cluster back up
|
||||||
|
cluster2.StartExistingNodes(t, 1, strings.TrimPrefix(info2.BzzAccount, "0x"))
|
||||||
|
|
||||||
|
// try to fetch imported file
|
||||||
|
res, err := http.Get(cluster2.Nodes[0].URL + "/bzz:/" + hash)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if res.StatusCode != 200 {
|
||||||
|
t.Fatalf("expected HTTP status %d, got %s", 200, res.Status)
|
||||||
|
}
|
||||||
|
|
||||||
|
// compare downloaded file with the generated random file
|
||||||
|
mustEqualFiles(t, f, res.Body)
|
||||||
|
}
|
||||||
|
|
||||||
|
func mustEqualFiles(t *testing.T, up io.Reader, down io.Reader) {
|
||||||
|
h := md5.New()
|
||||||
|
upLen, err := io.Copy(h, up)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
upHash := h.Sum(nil)
|
||||||
|
h.Reset()
|
||||||
|
downLen, err := io.Copy(h, down)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
downHash := h.Sum(nil)
|
||||||
|
|
||||||
|
if !bytes.Equal(upHash, downHash) || upLen != downLen {
|
||||||
|
t.Fatalf("downloaded imported file md5=%x (length %v) is not the same as the generated one mp5=%x (length %v)", downHash, downLen, upHash, upLen)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func generateRandomFile(t *testing.T, size int) (f *os.File, teardown func()) {
|
||||||
|
// create a tmp file
|
||||||
|
tmp, err := ioutil.TempFile("", "swarm-test")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// callback for tmp file cleanup
|
||||||
|
teardown = func() {
|
||||||
|
tmp.Close()
|
||||||
|
os.Remove(tmp.Name())
|
||||||
|
}
|
||||||
|
|
||||||
|
// write 10mb random data to file
|
||||||
|
buf := make([]byte, 10000000)
|
||||||
|
_, err = rand.Read(buf)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
ioutil.WriteFile(tmp.Name(), buf, 0755)
|
||||||
|
|
||||||
|
return tmp, teardown
|
||||||
|
}
|
127
cmd/swarm/fs.go
Normal file
127
cmd/swarm/fs.go
Normal file
@@ -0,0 +1,127 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
"github.com/ethereum/go-ethereum/node"
|
||||||
|
"github.com/ethereum/go-ethereum/rpc"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/fuse"
|
||||||
|
"gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
func mount(cliContext *cli.Context) {
|
||||||
|
args := cliContext.Args()
|
||||||
|
if len(args) < 2 {
|
||||||
|
utils.Fatalf("Usage: swarm fs mount --ipcpath <path to bzzd.ipc> <manifestHash> <file name>")
|
||||||
|
}
|
||||||
|
|
||||||
|
client, err := dialRPC(cliContext)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("had an error dailing to RPC endpoint: %v", err)
|
||||||
|
}
|
||||||
|
defer client.Close()
|
||||||
|
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
mf := &fuse.MountInfo{}
|
||||||
|
mountPoint, err := filepath.Abs(filepath.Clean(args[1]))
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("error expanding path for mount point: %v", err)
|
||||||
|
}
|
||||||
|
err = client.CallContext(ctx, mf, "swarmfs_mount", args[0], mountPoint)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("had an error calling the RPC endpoint while mounting: %v", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func unmount(cliContext *cli.Context) {
|
||||||
|
args := cliContext.Args()
|
||||||
|
|
||||||
|
if len(args) < 1 {
|
||||||
|
utils.Fatalf("Usage: swarm fs unmount --ipcpath <path to bzzd.ipc> <mount path>")
|
||||||
|
}
|
||||||
|
client, err := dialRPC(cliContext)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("had an error dailing to RPC endpoint: %v", err)
|
||||||
|
}
|
||||||
|
defer client.Close()
|
||||||
|
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
mf := fuse.MountInfo{}
|
||||||
|
err = client.CallContext(ctx, &mf, "swarmfs_unmount", args[0])
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("encountered an error calling the RPC endpoint while unmounting: %v", err)
|
||||||
|
}
|
||||||
|
fmt.Printf("%s\n", mf.LatestManifest) //print the latest manifest hash for user reference
|
||||||
|
}
|
||||||
|
|
||||||
|
func listMounts(cliContext *cli.Context) {
|
||||||
|
client, err := dialRPC(cliContext)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("had an error dailing to RPC endpoint: %v", err)
|
||||||
|
}
|
||||||
|
defer client.Close()
|
||||||
|
|
||||||
|
ctx, cancel := context.WithTimeout(context.Background(), 30*time.Second)
|
||||||
|
defer cancel()
|
||||||
|
|
||||||
|
mf := []fuse.MountInfo{}
|
||||||
|
err = client.CallContext(ctx, &mf, "swarmfs_listmounts")
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("encountered an error calling the RPC endpoint while unmounting: %v", err)
|
||||||
|
}
|
||||||
|
if len(mf) == 0 {
|
||||||
|
fmt.Print("Could not found any swarmfs mounts. Please make sure you've specified the correct RPC endpoint\n")
|
||||||
|
} else {
|
||||||
|
fmt.Printf("Found %d swarmfs mount(s):\n", len(mf))
|
||||||
|
for i, mountInfo := range mf {
|
||||||
|
fmt.Printf("%d:\n", i)
|
||||||
|
fmt.Printf("\tMount point: %s\n", mountInfo.MountPoint)
|
||||||
|
fmt.Printf("\tLatest Manifest: %s\n", mountInfo.LatestManifest)
|
||||||
|
fmt.Printf("\tStart Manifest: %s\n", mountInfo.StartManifest)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func dialRPC(ctx *cli.Context) (*rpc.Client, error) {
|
||||||
|
var endpoint string
|
||||||
|
|
||||||
|
if ctx.IsSet(utils.IPCPathFlag.Name) {
|
||||||
|
endpoint = ctx.String(utils.IPCPathFlag.Name)
|
||||||
|
} else {
|
||||||
|
utils.Fatalf("swarm ipc endpoint not specified")
|
||||||
|
}
|
||||||
|
|
||||||
|
if endpoint == "" {
|
||||||
|
endpoint = node.DefaultIPCEndpoint(clientIdentifier)
|
||||||
|
} else if strings.HasPrefix(endpoint, "rpc:") || strings.HasPrefix(endpoint, "ipc:") {
|
||||||
|
// Backwards compatibility with geth < 1.5 which required
|
||||||
|
// these prefixes.
|
||||||
|
endpoint = endpoint[4:]
|
||||||
|
}
|
||||||
|
return rpc.Dial(endpoint)
|
||||||
|
}
|
236
cmd/swarm/fs_test.go
Normal file
236
cmd/swarm/fs_test.go
Normal file
@@ -0,0 +1,236 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
// +build linux darwin freebsd
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"os"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
colorable "github.com/mattn/go-colorable"
|
||||||
|
)
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
log.PrintOrigins(true)
|
||||||
|
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(*loglevel), log.StreamHandler(colorable.NewColorableStderr(), log.TerminalFormat(true))))
|
||||||
|
}
|
||||||
|
|
||||||
|
type testFile struct {
|
||||||
|
filePath string
|
||||||
|
content string
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestCLISwarmFs is a high-level test of swarmfs
|
||||||
|
func TestCLISwarmFs(t *testing.T) {
|
||||||
|
cluster := newTestCluster(t, 3)
|
||||||
|
defer cluster.Shutdown()
|
||||||
|
|
||||||
|
// create a tmp dir
|
||||||
|
mountPoint, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
log.Debug("swarmfs cli test", "1st mount", mountPoint)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(mountPoint)
|
||||||
|
|
||||||
|
handlingNode := cluster.Nodes[0]
|
||||||
|
mhash := doUploadEmptyDir(t, handlingNode)
|
||||||
|
log.Debug("swarmfs cli test: mounting first run", "ipc path", filepath.Join(handlingNode.Dir, handlingNode.IpcPath))
|
||||||
|
|
||||||
|
mount := runSwarm(t, []string{
|
||||||
|
"fs",
|
||||||
|
"mount",
|
||||||
|
"--ipcpath", filepath.Join(handlingNode.Dir, handlingNode.IpcPath),
|
||||||
|
mhash,
|
||||||
|
mountPoint,
|
||||||
|
}...)
|
||||||
|
mount.ExpectExit()
|
||||||
|
|
||||||
|
filesToAssert := []*testFile{}
|
||||||
|
|
||||||
|
dirPath, err := createDirInDir(mountPoint, "testSubDir")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
dirPath2, err := createDirInDir(dirPath, "AnotherTestSubDir")
|
||||||
|
|
||||||
|
dummyContent := "somerandomtestcontentthatshouldbeasserted"
|
||||||
|
dirs := []string{
|
||||||
|
mountPoint,
|
||||||
|
dirPath,
|
||||||
|
dirPath2,
|
||||||
|
}
|
||||||
|
files := []string{"f1.tmp", "f2.tmp"}
|
||||||
|
for _, d := range dirs {
|
||||||
|
for _, entry := range files {
|
||||||
|
tFile, err := createTestFileInPath(d, entry, dummyContent)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
filesToAssert = append(filesToAssert, tFile)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if len(filesToAssert) != len(dirs)*len(files) {
|
||||||
|
t.Fatalf("should have %d files to assert now, got %d", len(dirs)*len(files), len(filesToAssert))
|
||||||
|
}
|
||||||
|
hashRegexp := `[a-f\d]{64}`
|
||||||
|
log.Debug("swarmfs cli test: unmounting first run...", "ipc path", filepath.Join(handlingNode.Dir, handlingNode.IpcPath))
|
||||||
|
|
||||||
|
unmount := runSwarm(t, []string{
|
||||||
|
"fs",
|
||||||
|
"unmount",
|
||||||
|
"--ipcpath", filepath.Join(handlingNode.Dir, handlingNode.IpcPath),
|
||||||
|
mountPoint,
|
||||||
|
}...)
|
||||||
|
_, matches := unmount.ExpectRegexp(hashRegexp)
|
||||||
|
unmount.ExpectExit()
|
||||||
|
|
||||||
|
hash := matches[0]
|
||||||
|
if hash == mhash {
|
||||||
|
t.Fatal("this should not be equal")
|
||||||
|
}
|
||||||
|
log.Debug("swarmfs cli test: asserting no files in mount point")
|
||||||
|
|
||||||
|
//check that there's nothing in the mount folder
|
||||||
|
filesInDir, err := ioutil.ReadDir(mountPoint)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("had an error reading the directory: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(filesInDir) != 0 {
|
||||||
|
t.Fatal("there shouldn't be anything here")
|
||||||
|
}
|
||||||
|
|
||||||
|
secondMountPoint, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
log.Debug("swarmfs cli test", "2nd mount point at", secondMountPoint)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(secondMountPoint)
|
||||||
|
|
||||||
|
log.Debug("swarmfs cli test: remounting at second mount point", "ipc path", filepath.Join(handlingNode.Dir, handlingNode.IpcPath))
|
||||||
|
|
||||||
|
//remount, check files
|
||||||
|
newMount := runSwarm(t, []string{
|
||||||
|
"fs",
|
||||||
|
"mount",
|
||||||
|
"--ipcpath", filepath.Join(handlingNode.Dir, handlingNode.IpcPath),
|
||||||
|
hash, // the latest hash
|
||||||
|
secondMountPoint,
|
||||||
|
}...)
|
||||||
|
|
||||||
|
newMount.ExpectExit()
|
||||||
|
time.Sleep(1 * time.Second)
|
||||||
|
|
||||||
|
filesInDir, err = ioutil.ReadDir(secondMountPoint)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
if len(filesInDir) == 0 {
|
||||||
|
t.Fatal("there should be something here")
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Debug("swarmfs cli test: traversing file tree to see it matches previous mount")
|
||||||
|
|
||||||
|
for _, file := range filesToAssert {
|
||||||
|
file.filePath = strings.Replace(file.filePath, mountPoint, secondMountPoint, -1)
|
||||||
|
fileBytes, err := ioutil.ReadFile(file.filePath)
|
||||||
|
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
if !bytes.Equal(fileBytes, bytes.NewBufferString(file.content).Bytes()) {
|
||||||
|
t.Fatal("this should be equal")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Debug("swarmfs cli test: unmounting second run", "ipc path", filepath.Join(handlingNode.Dir, handlingNode.IpcPath))
|
||||||
|
|
||||||
|
unmountSec := runSwarm(t, []string{
|
||||||
|
"fs",
|
||||||
|
"unmount",
|
||||||
|
"--ipcpath", filepath.Join(handlingNode.Dir, handlingNode.IpcPath),
|
||||||
|
secondMountPoint,
|
||||||
|
}...)
|
||||||
|
|
||||||
|
_, matches = unmountSec.ExpectRegexp(hashRegexp)
|
||||||
|
unmountSec.ExpectExit()
|
||||||
|
|
||||||
|
if matches[0] != hash {
|
||||||
|
t.Fatal("these should be equal - no changes made")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func doUploadEmptyDir(t *testing.T, node *testNode) string {
|
||||||
|
// create a tmp dir
|
||||||
|
tmpDir, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(tmpDir)
|
||||||
|
|
||||||
|
hashRegexp := `[a-f\d]{64}`
|
||||||
|
|
||||||
|
flags := []string{
|
||||||
|
"--bzzapi", node.URL,
|
||||||
|
"--recursive",
|
||||||
|
"up",
|
||||||
|
tmpDir}
|
||||||
|
|
||||||
|
log.Info("swarmfs cli test: uploading dir with 'swarm up'")
|
||||||
|
up := runSwarm(t, flags...)
|
||||||
|
_, matches := up.ExpectRegexp(hashRegexp)
|
||||||
|
up.ExpectExit()
|
||||||
|
hash := matches[0]
|
||||||
|
log.Info("swarmfs cli test: dir uploaded", "hash", hash)
|
||||||
|
return hash
|
||||||
|
}
|
||||||
|
|
||||||
|
func createDirInDir(createInDir string, dirToCreate string) (string, error) {
|
||||||
|
fullpath := filepath.Join(createInDir, dirToCreate)
|
||||||
|
err := os.MkdirAll(fullpath, 0777)
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
return fullpath, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func createTestFileInPath(dir, filename, content string) (*testFile, error) {
|
||||||
|
tFile := &testFile{}
|
||||||
|
filePath := filepath.Join(dir, filename)
|
||||||
|
if file, err := os.Create(filePath); err == nil {
|
||||||
|
tFile.content = content
|
||||||
|
tFile.filePath = filePath
|
||||||
|
|
||||||
|
_, err = io.WriteString(file, content)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
file.Close()
|
||||||
|
}
|
||||||
|
|
||||||
|
return tFile, nil
|
||||||
|
}
|
@@ -18,6 +18,7 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"context"
|
||||||
"fmt"
|
"fmt"
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
@@ -38,11 +39,11 @@ func hash(ctx *cli.Context) {
|
|||||||
defer f.Close()
|
defer f.Close()
|
||||||
|
|
||||||
stat, _ := f.Stat()
|
stat, _ := f.Stat()
|
||||||
chunker := storage.NewTreeChunker(storage.NewChunkerParams())
|
fileStore := storage.NewFileStore(storage.NewMapChunkStore(), storage.NewFileStoreParams())
|
||||||
key, err := chunker.Split(f, stat.Size(), nil, nil, nil)
|
addr, _, err := fileStore.Store(context.TODO(), f, stat.Size(), false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("%v\n", err)
|
utils.Fatalf("%v\n", err)
|
||||||
} else {
|
} else {
|
||||||
fmt.Printf("%v\n", key)
|
fmt.Printf("%v\n", addr)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -34,21 +34,37 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/console"
|
"github.com/ethereum/go-ethereum/console"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/ethclient"
|
|
||||||
"github.com/ethereum/go-ethereum/internal/debug"
|
"github.com/ethereum/go-ethereum/internal/debug"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/node"
|
"github.com/ethereum/go-ethereum/node"
|
||||||
"github.com/ethereum/go-ethereum/p2p"
|
"github.com/ethereum/go-ethereum/p2p"
|
||||||
"github.com/ethereum/go-ethereum/p2p/discover"
|
"github.com/ethereum/go-ethereum/p2p/discover"
|
||||||
"github.com/ethereum/go-ethereum/params"
|
|
||||||
"github.com/ethereum/go-ethereum/swarm"
|
"github.com/ethereum/go-ethereum/swarm"
|
||||||
bzzapi "github.com/ethereum/go-ethereum/swarm/api"
|
bzzapi "github.com/ethereum/go-ethereum/swarm/api"
|
||||||
swarmmetrics "github.com/ethereum/go-ethereum/swarm/metrics"
|
swarmmetrics "github.com/ethereum/go-ethereum/swarm/metrics"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/tracing"
|
||||||
|
sv "github.com/ethereum/go-ethereum/swarm/version"
|
||||||
|
|
||||||
"gopkg.in/urfave/cli.v1"
|
"gopkg.in/urfave/cli.v1"
|
||||||
)
|
)
|
||||||
|
|
||||||
const clientIdentifier = "swarm"
|
const clientIdentifier = "swarm"
|
||||||
|
const helpTemplate = `NAME:
|
||||||
|
{{.HelpName}} - {{.Usage}}
|
||||||
|
|
||||||
|
USAGE:
|
||||||
|
{{if .UsageText}}{{.UsageText}}{{else}}{{.HelpName}}{{if .VisibleFlags}} [command options]{{end}} {{if .ArgsUsage}}{{.ArgsUsage}}{{else}}[arguments...]{{end}}{{end}}{{if .Category}}
|
||||||
|
|
||||||
|
CATEGORY:
|
||||||
|
{{.Category}}{{end}}{{if .Description}}
|
||||||
|
|
||||||
|
DESCRIPTION:
|
||||||
|
{{.Description}}{{end}}{{if .VisibleFlags}}
|
||||||
|
|
||||||
|
OPTIONS:
|
||||||
|
{{range .VisibleFlags}}{{.}}
|
||||||
|
{{end}}{{end}}
|
||||||
|
`
|
||||||
|
|
||||||
var (
|
var (
|
||||||
gitCommit string // Git SHA1 commit hash of the release (set via linker flags)
|
gitCommit string // Git SHA1 commit hash of the release (set via linker flags)
|
||||||
@@ -87,10 +103,6 @@ var (
|
|||||||
Usage: "Network identifier (integer, default 3=swarm testnet)",
|
Usage: "Network identifier (integer, default 3=swarm testnet)",
|
||||||
EnvVar: SWARM_ENV_NETWORK_ID,
|
EnvVar: SWARM_ENV_NETWORK_ID,
|
||||||
}
|
}
|
||||||
SwarmConfigPathFlag = cli.StringFlag{
|
|
||||||
Name: "bzzconfig",
|
|
||||||
Usage: "DEPRECATED: please use --config path/to/TOML-file",
|
|
||||||
}
|
|
||||||
SwarmSwapEnabledFlag = cli.BoolFlag{
|
SwarmSwapEnabledFlag = cli.BoolFlag{
|
||||||
Name: "swap",
|
Name: "swap",
|
||||||
Usage: "Swarm SWAP enabled (default false)",
|
Usage: "Swarm SWAP enabled (default false)",
|
||||||
@@ -101,10 +113,20 @@ var (
|
|||||||
Usage: "URL of the Ethereum API provider to use to settle SWAP payments",
|
Usage: "URL of the Ethereum API provider to use to settle SWAP payments",
|
||||||
EnvVar: SWARM_ENV_SWAP_API,
|
EnvVar: SWARM_ENV_SWAP_API,
|
||||||
}
|
}
|
||||||
SwarmSyncEnabledFlag = cli.BoolTFlag{
|
SwarmSyncDisabledFlag = cli.BoolTFlag{
|
||||||
Name: "sync",
|
Name: "nosync",
|
||||||
Usage: "Swarm Syncing enabled (default true)",
|
Usage: "Disable swarm syncing",
|
||||||
EnvVar: SWARM_ENV_SYNC_ENABLE,
|
EnvVar: SWARM_ENV_SYNC_DISABLE,
|
||||||
|
}
|
||||||
|
SwarmSyncUpdateDelay = cli.DurationFlag{
|
||||||
|
Name: "sync-update-delay",
|
||||||
|
Usage: "Duration for sync subscriptions update after no new peers are added (default 15s)",
|
||||||
|
EnvVar: SWARM_ENV_SYNC_UPDATE_DELAY,
|
||||||
|
}
|
||||||
|
SwarmDeliverySkipCheckFlag = cli.BoolFlag{
|
||||||
|
Name: "delivery-skip-check",
|
||||||
|
Usage: "Skip chunk delivery check (default false)",
|
||||||
|
EnvVar: SWARM_ENV_DELIVERY_SKIP_CHECK,
|
||||||
}
|
}
|
||||||
EnsAPIFlag = cli.StringSliceFlag{
|
EnsAPIFlag = cli.StringSliceFlag{
|
||||||
Name: "ens-api",
|
Name: "ens-api",
|
||||||
@@ -116,13 +138,13 @@ var (
|
|||||||
Usage: "Swarm HTTP endpoint",
|
Usage: "Swarm HTTP endpoint",
|
||||||
Value: "http://127.0.0.1:8500",
|
Value: "http://127.0.0.1:8500",
|
||||||
}
|
}
|
||||||
SwarmRecursiveUploadFlag = cli.BoolFlag{
|
SwarmRecursiveFlag = cli.BoolFlag{
|
||||||
Name: "recursive",
|
Name: "recursive",
|
||||||
Usage: "Upload directories recursively",
|
Usage: "Upload directories recursively",
|
||||||
}
|
}
|
||||||
SwarmWantManifestFlag = cli.BoolTFlag{
|
SwarmWantManifestFlag = cli.BoolTFlag{
|
||||||
Name: "manifest",
|
Name: "manifest",
|
||||||
Usage: "Automatic manifest upload",
|
Usage: "Automatic manifest upload (default true)",
|
||||||
}
|
}
|
||||||
SwarmUploadDefaultPath = cli.StringFlag{
|
SwarmUploadDefaultPath = cli.StringFlag{
|
||||||
Name: "defaultpath",
|
Name: "defaultpath",
|
||||||
@@ -134,22 +156,43 @@ var (
|
|||||||
}
|
}
|
||||||
SwarmUploadMimeType = cli.StringFlag{
|
SwarmUploadMimeType = cli.StringFlag{
|
||||||
Name: "mime",
|
Name: "mime",
|
||||||
Usage: "force mime type",
|
Usage: "Manually specify MIME type",
|
||||||
|
}
|
||||||
|
SwarmEncryptedFlag = cli.BoolFlag{
|
||||||
|
Name: "encrypt",
|
||||||
|
Usage: "use encrypted upload",
|
||||||
}
|
}
|
||||||
CorsStringFlag = cli.StringFlag{
|
CorsStringFlag = cli.StringFlag{
|
||||||
Name: "corsdomain",
|
Name: "corsdomain",
|
||||||
Usage: "Domain on which to send Access-Control-Allow-Origin header (multiple domains can be supplied separated by a ',')",
|
Usage: "Domain on which to send Access-Control-Allow-Origin header (multiple domains can be supplied separated by a ',')",
|
||||||
EnvVar: SWARM_ENV_CORS,
|
EnvVar: SWARM_ENV_CORS,
|
||||||
}
|
}
|
||||||
|
SwarmStorePath = cli.StringFlag{
|
||||||
// the following flags are deprecated and should be removed in the future
|
Name: "store.path",
|
||||||
DeprecatedEthAPIFlag = cli.StringFlag{
|
Usage: "Path to leveldb chunk DB (default <$GETH_ENV_DIR>/swarm/bzz-<$BZZ_KEY>/chunks)",
|
||||||
Name: "ethapi",
|
EnvVar: SWARM_ENV_STORE_PATH,
|
||||||
Usage: "DEPRECATED: please use --ens-api and --swap-api",
|
|
||||||
}
|
}
|
||||||
DeprecatedEnsAddrFlag = cli.StringFlag{
|
SwarmStoreCapacity = cli.Uint64Flag{
|
||||||
Name: "ens-addr",
|
Name: "store.size",
|
||||||
Usage: "DEPRECATED: ENS contract address, please use --ens-api with contract address according to its format",
|
Usage: "Number of chunks (5M is roughly 20-25GB) (default 5000000)",
|
||||||
|
EnvVar: SWARM_ENV_STORE_CAPACITY,
|
||||||
|
}
|
||||||
|
SwarmStoreCacheCapacity = cli.UintFlag{
|
||||||
|
Name: "store.cache.size",
|
||||||
|
Usage: "Number of recent chunks cached in memory (default 5000)",
|
||||||
|
EnvVar: SWARM_ENV_STORE_CACHE_CAPACITY,
|
||||||
|
}
|
||||||
|
SwarmResourceMultihashFlag = cli.BoolFlag{
|
||||||
|
Name: "multihash",
|
||||||
|
Usage: "Determines how to interpret data for a resource update. If not present, data will be interpreted as raw, literal data that will be included in the resource",
|
||||||
|
}
|
||||||
|
SwarmResourceNameFlag = cli.StringFlag{
|
||||||
|
Name: "name",
|
||||||
|
Usage: "User-defined name for the new resource",
|
||||||
|
}
|
||||||
|
SwarmResourceDataOnCreateFlag = cli.StringFlag{
|
||||||
|
Name: "data",
|
||||||
|
Usage: "Initializes the resource with the given hex-encoded data. Data must be prefixed by 0x",
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -159,12 +202,21 @@ var (
|
|||||||
SWARM_ERR_SWAP_SET_NO_API = "SWAP is enabled but --swap-api is not set"
|
SWARM_ERR_SWAP_SET_NO_API = "SWAP is enabled but --swap-api is not set"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// this help command gets added to any subcommand that does not define it explicitly
|
||||||
|
var defaultSubcommandHelp = cli.Command{
|
||||||
|
Action: func(ctx *cli.Context) { cli.ShowCommandHelpAndExit(ctx, "", 1) },
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "help",
|
||||||
|
Usage: "shows this help",
|
||||||
|
Hidden: true,
|
||||||
|
}
|
||||||
|
|
||||||
var defaultNodeConfig = node.DefaultConfig
|
var defaultNodeConfig = node.DefaultConfig
|
||||||
|
|
||||||
// This init function sets defaults so cmd/swarm can run alongside geth.
|
// This init function sets defaults so cmd/swarm can run alongside geth.
|
||||||
func init() {
|
func init() {
|
||||||
defaultNodeConfig.Name = clientIdentifier
|
defaultNodeConfig.Name = clientIdentifier
|
||||||
defaultNodeConfig.Version = params.VersionWithCommit(gitCommit)
|
defaultNodeConfig.Version = sv.VersionWithCommit(gitCommit)
|
||||||
defaultNodeConfig.P2P.ListenAddr = ":30399"
|
defaultNodeConfig.P2P.ListenAddr = ":30399"
|
||||||
defaultNodeConfig.IPCPath = "bzzd.ipc"
|
defaultNodeConfig.IPCPath = "bzzd.ipc"
|
||||||
// Set flag defaults for --help display.
|
// Set flag defaults for --help display.
|
||||||
@@ -180,91 +232,165 @@ func init() {
|
|||||||
app.Copyright = "Copyright 2013-2016 The go-ethereum Authors"
|
app.Copyright = "Copyright 2013-2016 The go-ethereum Authors"
|
||||||
app.Commands = []cli.Command{
|
app.Commands = []cli.Command{
|
||||||
{
|
{
|
||||||
Action: version,
|
Action: version,
|
||||||
Name: "version",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "Print version numbers",
|
Name: "version",
|
||||||
ArgsUsage: " ",
|
Usage: "Print version numbers",
|
||||||
Description: `
|
Description: "The output of this command is supposed to be machine-readable",
|
||||||
The output of this command is supposed to be machine-readable.
|
|
||||||
`,
|
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: upload,
|
Action: upload,
|
||||||
Name: "up",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "upload a file or directory to swarm using the HTTP API",
|
Name: "up",
|
||||||
ArgsUsage: " <file>",
|
Usage: "uploads a file or directory to swarm using the HTTP API",
|
||||||
Description: `
|
ArgsUsage: "<file>",
|
||||||
"upload a file or directory to swarm using the HTTP API and prints the root hash",
|
Flags: []cli.Flag{SwarmEncryptedFlag},
|
||||||
`,
|
Description: "uploads a file or directory to swarm using the HTTP API and prints the root hash",
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: list,
|
CustomHelpTemplate: helpTemplate,
|
||||||
Name: "ls",
|
Name: "resource",
|
||||||
Usage: "list files and directories contained in a manifest",
|
Usage: "(Advanced) Create and update Mutable Resources",
|
||||||
ArgsUsage: " <manifest> [<prefix>]",
|
ArgsUsage: "<create|update|info>",
|
||||||
Description: `
|
Description: "Works with Mutable Resource Updates",
|
||||||
Lists files and directories contained in a manifest.
|
|
||||||
`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Action: hash,
|
|
||||||
Name: "hash",
|
|
||||||
Usage: "print the swarm hash of a file or directory",
|
|
||||||
ArgsUsage: " <file>",
|
|
||||||
Description: `
|
|
||||||
Prints the swarm hash of file or directory.
|
|
||||||
`,
|
|
||||||
},
|
|
||||||
{
|
|
||||||
Name: "manifest",
|
|
||||||
Usage: "update a MANIFEST",
|
|
||||||
ArgsUsage: "manifest COMMAND",
|
|
||||||
Description: `
|
|
||||||
Updates a MANIFEST by adding/removing/updating the hash of a path.
|
|
||||||
`,
|
|
||||||
Subcommands: []cli.Command{
|
Subcommands: []cli.Command{
|
||||||
{
|
{
|
||||||
Action: add,
|
Action: resourceCreate,
|
||||||
Name: "add",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "add a new path to the manifest",
|
Name: "create",
|
||||||
ArgsUsage: "<MANIFEST> <path> <hash> [<content-type>]",
|
Usage: "creates a new Mutable Resource",
|
||||||
Description: `
|
ArgsUsage: "<frequency>",
|
||||||
Adds a new path to the manifest
|
Description: "creates a new Mutable Resource",
|
||||||
`,
|
Flags: []cli.Flag{SwarmResourceNameFlag, SwarmResourceDataOnCreateFlag, SwarmResourceMultihashFlag},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: update,
|
Action: resourceUpdate,
|
||||||
Name: "update",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "update the hash for an already existing path in the manifest",
|
Name: "update",
|
||||||
ArgsUsage: "<MANIFEST> <path> <newhash> [<newcontent-type>]",
|
Usage: "updates the content of an existing Mutable Resource",
|
||||||
Description: `
|
ArgsUsage: "<Manifest Address or ENS domain> <0x Hex data>",
|
||||||
Update the hash for an already existing path in the manifest
|
Description: "updates the content of an existing Mutable Resource",
|
||||||
`,
|
Flags: []cli.Flag{SwarmResourceMultihashFlag},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: remove,
|
Action: resourceInfo,
|
||||||
Name: "remove",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "removes a path from the manifest",
|
Name: "info",
|
||||||
ArgsUsage: "<MANIFEST> <path>",
|
Usage: "obtains information about an existing Mutable Resource",
|
||||||
Description: `
|
ArgsUsage: "<Manifest Address or ENS domain>",
|
||||||
Removes a path from the manifest
|
Description: "obtains information about an existing Mutable Resource",
|
||||||
`,
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Name: "db",
|
Action: list,
|
||||||
Usage: "manage the local chunk database",
|
CustomHelpTemplate: helpTemplate,
|
||||||
ArgsUsage: "db COMMAND",
|
Name: "ls",
|
||||||
|
Usage: "list files and directories contained in a manifest",
|
||||||
|
ArgsUsage: "<manifest> [<prefix>]",
|
||||||
|
Description: "Lists files and directories contained in a manifest",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: hash,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "hash",
|
||||||
|
Usage: "print the swarm hash of a file or directory",
|
||||||
|
ArgsUsage: "<file>",
|
||||||
|
Description: "Prints the swarm hash of file or directory",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: download,
|
||||||
|
Name: "down",
|
||||||
|
Flags: []cli.Flag{SwarmRecursiveFlag},
|
||||||
|
Usage: "downloads a swarm manifest or a file inside a manifest",
|
||||||
|
ArgsUsage: " <uri> [<dir>]",
|
||||||
Description: `
|
Description: `
|
||||||
Manage the local chunk database.
|
Downloads a swarm bzz uri to the given dir. When no dir is provided, working directory is assumed. --recursive flag is expected when downloading a manifest with multiple entries.
|
||||||
`,
|
`,
|
||||||
|
},
|
||||||
|
|
||||||
|
{
|
||||||
|
Name: "manifest",
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Usage: "perform operations on swarm manifests",
|
||||||
|
ArgsUsage: "COMMAND",
|
||||||
|
Description: "Updates a MANIFEST by adding/removing/updating the hash of a path.\nCOMMAND could be: add, update, remove",
|
||||||
Subcommands: []cli.Command{
|
Subcommands: []cli.Command{
|
||||||
{
|
{
|
||||||
Action: dbExport,
|
Action: add,
|
||||||
Name: "export",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "export a local chunk database as a tar archive (use - to send to stdout)",
|
Name: "add",
|
||||||
ArgsUsage: "<chunkdb> <file>",
|
Usage: "add a new path to the manifest",
|
||||||
|
ArgsUsage: "<MANIFEST> <path> <hash> [<content-type>]",
|
||||||
|
Description: "Adds a new path to the manifest",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: update,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "update",
|
||||||
|
Usage: "update the hash for an already existing path in the manifest",
|
||||||
|
ArgsUsage: "<MANIFEST> <path> <newhash> [<newcontent-type>]",
|
||||||
|
Description: "Update the hash for an already existing path in the manifest",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: remove,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "remove",
|
||||||
|
Usage: "removes a path from the manifest",
|
||||||
|
ArgsUsage: "<MANIFEST> <path>",
|
||||||
|
Description: "Removes a path from the manifest",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "fs",
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Usage: "perform FUSE operations",
|
||||||
|
ArgsUsage: "fs COMMAND",
|
||||||
|
Description: "Performs FUSE operations by mounting/unmounting/listing mount points. This assumes you already have a Swarm node running locally. For all operation you must reference the correct path to bzzd.ipc in order to communicate with the node",
|
||||||
|
Subcommands: []cli.Command{
|
||||||
|
{
|
||||||
|
Action: mount,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "mount",
|
||||||
|
Flags: []cli.Flag{utils.IPCPathFlag},
|
||||||
|
Usage: "mount a swarm hash to a mount point",
|
||||||
|
ArgsUsage: "swarm fs mount --ipcpath <path to bzzd.ipc> <manifest hash> <mount point>",
|
||||||
|
Description: "Mounts a Swarm manifest hash to a given mount point. This assumes you already have a Swarm node running locally. You must reference the correct path to your bzzd.ipc file",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: unmount,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "unmount",
|
||||||
|
Flags: []cli.Flag{utils.IPCPathFlag},
|
||||||
|
Usage: "unmount a swarmfs mount",
|
||||||
|
ArgsUsage: "swarm fs unmount --ipcpath <path to bzzd.ipc> <mount point>",
|
||||||
|
Description: "Unmounts a swarmfs mount residing at <mount point>. This assumes you already have a Swarm node running locally. You must reference the correct path to your bzzd.ipc file",
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Action: listMounts,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "list",
|
||||||
|
Flags: []cli.Flag{utils.IPCPathFlag},
|
||||||
|
Usage: "list swarmfs mounts",
|
||||||
|
ArgsUsage: "swarm fs list --ipcpath <path to bzzd.ipc>",
|
||||||
|
Description: "Lists all mounted swarmfs volumes. This assumes you already have a Swarm node running locally. You must reference the correct path to your bzzd.ipc file",
|
||||||
|
},
|
||||||
|
},
|
||||||
|
},
|
||||||
|
{
|
||||||
|
Name: "db",
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Usage: "manage the local chunk database",
|
||||||
|
ArgsUsage: "db COMMAND",
|
||||||
|
Description: "Manage the local chunk database",
|
||||||
|
Subcommands: []cli.Command{
|
||||||
|
{
|
||||||
|
Action: dbExport,
|
||||||
|
CustomHelpTemplate: helpTemplate,
|
||||||
|
Name: "export",
|
||||||
|
Usage: "export a local chunk database as a tar archive (use - to send to stdout)",
|
||||||
|
ArgsUsage: "<chunkdb> <file>",
|
||||||
Description: `
|
Description: `
|
||||||
Export a local chunk database as a tar archive (use - to send to stdout).
|
Export a local chunk database as a tar archive (use - to send to stdout).
|
||||||
|
|
||||||
@@ -277,10 +403,11 @@ pv(1) tool to get a progress bar:
|
|||||||
`,
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: dbImport,
|
Action: dbImport,
|
||||||
Name: "import",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "import chunks from a tar archive into a local chunk database (use - to read from stdin)",
|
Name: "import",
|
||||||
ArgsUsage: "<chunkdb> <file>",
|
Usage: "import chunks from a tar archive into a local chunk database (use - to read from stdin)",
|
||||||
|
ArgsUsage: "<chunkdb> <file>",
|
||||||
Description: `
|
Description: `
|
||||||
Import chunks from a tar archive into a local chunk database (use - to read from stdin).
|
Import chunks from a tar archive into a local chunk database (use - to read from stdin).
|
||||||
|
|
||||||
@@ -293,30 +420,24 @@ pv(1) tool to get a progress bar:
|
|||||||
`,
|
`,
|
||||||
},
|
},
|
||||||
{
|
{
|
||||||
Action: dbClean,
|
Action: dbClean,
|
||||||
Name: "clean",
|
CustomHelpTemplate: helpTemplate,
|
||||||
Usage: "remove corrupt entries from a local chunk database",
|
Name: "clean",
|
||||||
ArgsUsage: "<chunkdb>",
|
Usage: "remove corrupt entries from a local chunk database",
|
||||||
Description: `
|
ArgsUsage: "<chunkdb>",
|
||||||
Remove corrupt entries from a local chunk database.
|
Description: "Remove corrupt entries from a local chunk database",
|
||||||
`,
|
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
},
|
},
|
||||||
{
|
|
||||||
Action: func(ctx *cli.Context) {
|
|
||||||
utils.Fatalf("ERROR: 'swarm cleandb' has been removed, please use 'swarm db clean'.")
|
|
||||||
},
|
|
||||||
Name: "cleandb",
|
|
||||||
Usage: "DEPRECATED: use 'swarm db clean'",
|
|
||||||
ArgsUsage: " ",
|
|
||||||
Description: `
|
|
||||||
DEPRECATED: use 'swarm db clean'.
|
|
||||||
`,
|
|
||||||
},
|
|
||||||
// See config.go
|
// See config.go
|
||||||
DumpConfigCommand,
|
DumpConfigCommand,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// append a hidden help subcommand to all commands that have subcommands
|
||||||
|
// if a help command was already defined above, that one will take precedence.
|
||||||
|
addDefaultHelpSubcommands(app.Commands)
|
||||||
|
|
||||||
sort.Sort(cli.CommandsByName(app.Commands))
|
sort.Sort(cli.CommandsByName(app.Commands))
|
||||||
|
|
||||||
app.Flags = []cli.Flag{
|
app.Flags = []cli.Flag{
|
||||||
@@ -339,10 +460,11 @@ DEPRECATED: use 'swarm db clean'.
|
|||||||
CorsStringFlag,
|
CorsStringFlag,
|
||||||
EnsAPIFlag,
|
EnsAPIFlag,
|
||||||
SwarmTomlConfigPathFlag,
|
SwarmTomlConfigPathFlag,
|
||||||
SwarmConfigPathFlag,
|
|
||||||
SwarmSwapEnabledFlag,
|
SwarmSwapEnabledFlag,
|
||||||
SwarmSwapAPIFlag,
|
SwarmSwapAPIFlag,
|
||||||
SwarmSyncEnabledFlag,
|
SwarmSyncDisabledFlag,
|
||||||
|
SwarmSyncUpdateDelay,
|
||||||
|
SwarmDeliverySkipCheckFlag,
|
||||||
SwarmListenAddrFlag,
|
SwarmListenAddrFlag,
|
||||||
SwarmPortFlag,
|
SwarmPortFlag,
|
||||||
SwarmAccountFlag,
|
SwarmAccountFlag,
|
||||||
@@ -350,23 +472,34 @@ DEPRECATED: use 'swarm db clean'.
|
|||||||
ChequebookAddrFlag,
|
ChequebookAddrFlag,
|
||||||
// upload flags
|
// upload flags
|
||||||
SwarmApiFlag,
|
SwarmApiFlag,
|
||||||
SwarmRecursiveUploadFlag,
|
SwarmRecursiveFlag,
|
||||||
SwarmWantManifestFlag,
|
SwarmWantManifestFlag,
|
||||||
SwarmUploadDefaultPath,
|
SwarmUploadDefaultPath,
|
||||||
SwarmUpFromStdinFlag,
|
SwarmUpFromStdinFlag,
|
||||||
SwarmUploadMimeType,
|
SwarmUploadMimeType,
|
||||||
//deprecated flags
|
// storage flags
|
||||||
DeprecatedEthAPIFlag,
|
SwarmStorePath,
|
||||||
DeprecatedEnsAddrFlag,
|
SwarmStoreCapacity,
|
||||||
|
SwarmStoreCacheCapacity,
|
||||||
}
|
}
|
||||||
|
rpcFlags := []cli.Flag{
|
||||||
|
utils.WSEnabledFlag,
|
||||||
|
utils.WSListenAddrFlag,
|
||||||
|
utils.WSPortFlag,
|
||||||
|
utils.WSApiFlag,
|
||||||
|
utils.WSAllowedOriginsFlag,
|
||||||
|
}
|
||||||
|
app.Flags = append(app.Flags, rpcFlags...)
|
||||||
app.Flags = append(app.Flags, debug.Flags...)
|
app.Flags = append(app.Flags, debug.Flags...)
|
||||||
app.Flags = append(app.Flags, swarmmetrics.Flags...)
|
app.Flags = append(app.Flags, swarmmetrics.Flags...)
|
||||||
|
app.Flags = append(app.Flags, tracing.Flags...)
|
||||||
app.Before = func(ctx *cli.Context) error {
|
app.Before = func(ctx *cli.Context) error {
|
||||||
runtime.GOMAXPROCS(runtime.NumCPU())
|
runtime.GOMAXPROCS(runtime.NumCPU())
|
||||||
if err := debug.Setup(ctx); err != nil {
|
if err := debug.Setup(ctx, ""); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
swarmmetrics.Setup(ctx)
|
swarmmetrics.Setup(ctx)
|
||||||
|
tracing.Setup(ctx)
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
app.After = func(ctx *cli.Context) error {
|
app.After = func(ctx *cli.Context) error {
|
||||||
@@ -384,15 +517,12 @@ func main() {
|
|||||||
|
|
||||||
func version(ctx *cli.Context) error {
|
func version(ctx *cli.Context) error {
|
||||||
fmt.Println(strings.Title(clientIdentifier))
|
fmt.Println(strings.Title(clientIdentifier))
|
||||||
fmt.Println("Version:", params.Version)
|
fmt.Println("Version:", sv.VersionWithMeta)
|
||||||
if gitCommit != "" {
|
if gitCommit != "" {
|
||||||
fmt.Println("Git Commit:", gitCommit)
|
fmt.Println("Git Commit:", gitCommit)
|
||||||
}
|
}
|
||||||
fmt.Println("Network Id:", ctx.GlobalInt(utils.NetworkIdFlag.Name))
|
|
||||||
fmt.Println("Go Version:", runtime.Version())
|
fmt.Println("Go Version:", runtime.Version())
|
||||||
fmt.Println("OS:", runtime.GOOS)
|
fmt.Println("OS:", runtime.GOOS)
|
||||||
fmt.Printf("GOPATH=%s\n", os.Getenv("GOPATH"))
|
|
||||||
fmt.Printf("GOROOT=%s\n", runtime.GOROOT())
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -405,6 +535,10 @@ func bzzd(ctx *cli.Context) error {
|
|||||||
}
|
}
|
||||||
|
|
||||||
cfg := defaultNodeConfig
|
cfg := defaultNodeConfig
|
||||||
|
|
||||||
|
//pss operates on ws
|
||||||
|
cfg.WSModules = append(cfg.WSModules, "pss")
|
||||||
|
|
||||||
//geth only supports --datadir via command line
|
//geth only supports --datadir via command line
|
||||||
//in order to be consistent within swarm, if we pass --datadir via environment variable
|
//in order to be consistent within swarm, if we pass --datadir via environment variable
|
||||||
//or via config file, we get the same directory for geth and swarm
|
//or via config file, we get the same directory for geth and swarm
|
||||||
@@ -421,7 +555,7 @@ func bzzd(ctx *cli.Context) error {
|
|||||||
//due to overriding behavior
|
//due to overriding behavior
|
||||||
initSwarmNode(bzzconfig, stack, ctx)
|
initSwarmNode(bzzconfig, stack, ctx)
|
||||||
//register BZZ as node.Service in the ethereum node
|
//register BZZ as node.Service in the ethereum node
|
||||||
registerBzzService(bzzconfig, ctx, stack)
|
registerBzzService(bzzconfig, stack)
|
||||||
//start the node
|
//start the node
|
||||||
utils.StartNode(stack)
|
utils.StartNode(stack)
|
||||||
|
|
||||||
@@ -439,7 +573,7 @@ func bzzd(ctx *cli.Context) error {
|
|||||||
bootnodes := strings.Split(bzzconfig.BootNodes, ",")
|
bootnodes := strings.Split(bzzconfig.BootNodes, ",")
|
||||||
injectBootnodes(stack.Server(), bootnodes)
|
injectBootnodes(stack.Server(), bootnodes)
|
||||||
} else {
|
} else {
|
||||||
if bzzconfig.NetworkId == 3 {
|
if bzzconfig.NetworkID == 3 {
|
||||||
injectBootnodes(stack.Server(), testbetBootNodes)
|
injectBootnodes(stack.Server(), testbetBootNodes)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -448,21 +582,11 @@ func bzzd(ctx *cli.Context) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func registerBzzService(bzzconfig *bzzapi.Config, ctx *cli.Context, stack *node.Node) {
|
func registerBzzService(bzzconfig *bzzapi.Config, stack *node.Node) {
|
||||||
|
|
||||||
//define the swarm service boot function
|
//define the swarm service boot function
|
||||||
boot := func(ctx *node.ServiceContext) (node.Service, error) {
|
boot := func(_ *node.ServiceContext) (node.Service, error) {
|
||||||
var swapClient *ethclient.Client
|
// In production, mockStore must be always nil.
|
||||||
var err error
|
return swarm.NewSwarm(bzzconfig, nil)
|
||||||
if bzzconfig.SwapApi != "" {
|
|
||||||
log.Info("connecting to SWAP API", "url", bzzconfig.SwapApi)
|
|
||||||
swapClient, err = ethclient.Dial(bzzconfig.SwapApi)
|
|
||||||
if err != nil {
|
|
||||||
return nil, fmt.Errorf("error connecting to SWAP API %s: %s", bzzconfig.SwapApi, err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return swarm.NewSwarm(ctx, swapClient, bzzconfig)
|
|
||||||
}
|
}
|
||||||
//register within the ethereum node
|
//register within the ethereum node
|
||||||
if err := stack.Register(boot); err != nil {
|
if err := stack.Register(boot); err != nil {
|
||||||
@@ -487,6 +611,26 @@ func getAccount(bzzaccount string, ctx *cli.Context, stack *node.Node) *ecdsa.Pr
|
|||||||
return decryptStoreAccount(ks, bzzaccount, utils.MakePasswordList(ctx))
|
return decryptStoreAccount(ks, bzzaccount, utils.MakePasswordList(ctx))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// getPrivKey returns the private key of the specified bzzaccount
|
||||||
|
// Used only by client commands, such as `resource`
|
||||||
|
func getPrivKey(ctx *cli.Context) *ecdsa.PrivateKey {
|
||||||
|
// booting up the swarm node just as we do in bzzd action
|
||||||
|
bzzconfig, err := buildConfig(ctx)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("unable to configure swarm: %v", err)
|
||||||
|
}
|
||||||
|
cfg := defaultNodeConfig
|
||||||
|
if _, err := os.Stat(bzzconfig.Path); err == nil {
|
||||||
|
cfg.DataDir = bzzconfig.Path
|
||||||
|
}
|
||||||
|
utils.SetNodeConfig(ctx, &cfg)
|
||||||
|
stack, err := node.New(&cfg)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("can't create node: %v", err)
|
||||||
|
}
|
||||||
|
return getAccount(bzzconfig.BzzAccount, ctx, stack)
|
||||||
|
}
|
||||||
|
|
||||||
func decryptStoreAccount(ks *keystore.KeyStore, account string, passwords []string) *ecdsa.PrivateKey {
|
func decryptStoreAccount(ks *keystore.KeyStore, account string, passwords []string) *ecdsa.PrivateKey {
|
||||||
var a accounts.Account
|
var a accounts.Account
|
||||||
var err error
|
var err error
|
||||||
@@ -551,3 +695,16 @@ func injectBootnodes(srv *p2p.Server, nodes []string) {
|
|||||||
srv.AddPeer(n)
|
srv.AddPeer(n)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// addDefaultHelpSubcommand scans through defined CLI commands and adds
|
||||||
|
// a basic help subcommand to each
|
||||||
|
// if a help command is already defined, it will take precedence over the default.
|
||||||
|
func addDefaultHelpSubcommands(commands []cli.Command) {
|
||||||
|
for i := range commands {
|
||||||
|
cmd := &commands[i]
|
||||||
|
if cmd.Subcommands != nil {
|
||||||
|
cmd.Subcommands = append(cmd.Subcommands, defaultSubcommandHelp)
|
||||||
|
addDefaultHelpSubcommands(cmd.Subcommands)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
@@ -131,13 +131,13 @@ func addEntryToManifest(ctx *cli.Context, mhash, path, hash, ctype string) strin
|
|||||||
longestPathEntry = api.ManifestEntry{}
|
longestPathEntry = api.ManifestEntry{}
|
||||||
)
|
)
|
||||||
|
|
||||||
mroot, err := client.DownloadManifest(mhash)
|
mroot, isEncrypted, err := client.DownloadManifest(mhash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest download failed: %v", err)
|
utils.Fatalf("Manifest download failed: %v", err)
|
||||||
}
|
}
|
||||||
|
|
||||||
//TODO: check if the "hash" to add is valid and present in swarm
|
//TODO: check if the "hash" to add is valid and present in swarm
|
||||||
_, err = client.DownloadManifest(hash)
|
_, _, err = client.DownloadManifest(hash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Hash to add is not present: %v", err)
|
utils.Fatalf("Hash to add is not present: %v", err)
|
||||||
}
|
}
|
||||||
@@ -180,7 +180,7 @@ func addEntryToManifest(ctx *cli.Context, mhash, path, hash, ctype string) strin
|
|||||||
mroot.Entries = append(mroot.Entries, newEntry)
|
mroot.Entries = append(mroot.Entries, newEntry)
|
||||||
}
|
}
|
||||||
|
|
||||||
newManifestHash, err := client.UploadManifest(mroot)
|
newManifestHash, err := client.UploadManifest(mroot, isEncrypted)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest upload failed: %v", err)
|
utils.Fatalf("Manifest upload failed: %v", err)
|
||||||
}
|
}
|
||||||
@@ -197,7 +197,7 @@ func updateEntryInManifest(ctx *cli.Context, mhash, path, hash, ctype string) st
|
|||||||
longestPathEntry = api.ManifestEntry{}
|
longestPathEntry = api.ManifestEntry{}
|
||||||
)
|
)
|
||||||
|
|
||||||
mroot, err := client.DownloadManifest(mhash)
|
mroot, isEncrypted, err := client.DownloadManifest(mhash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest download failed: %v", err)
|
utils.Fatalf("Manifest download failed: %v", err)
|
||||||
}
|
}
|
||||||
@@ -257,7 +257,7 @@ func updateEntryInManifest(ctx *cli.Context, mhash, path, hash, ctype string) st
|
|||||||
mroot = newMRoot
|
mroot = newMRoot
|
||||||
}
|
}
|
||||||
|
|
||||||
newManifestHash, err := client.UploadManifest(mroot)
|
newManifestHash, err := client.UploadManifest(mroot, isEncrypted)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest upload failed: %v", err)
|
utils.Fatalf("Manifest upload failed: %v", err)
|
||||||
}
|
}
|
||||||
@@ -273,7 +273,7 @@ func removeEntryFromManifest(ctx *cli.Context, mhash, path string) string {
|
|||||||
longestPathEntry = api.ManifestEntry{}
|
longestPathEntry = api.ManifestEntry{}
|
||||||
)
|
)
|
||||||
|
|
||||||
mroot, err := client.DownloadManifest(mhash)
|
mroot, isEncrypted, err := client.DownloadManifest(mhash)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest download failed: %v", err)
|
utils.Fatalf("Manifest download failed: %v", err)
|
||||||
}
|
}
|
||||||
@@ -323,7 +323,7 @@ func removeEntryFromManifest(ctx *cli.Context, mhash, path string) string {
|
|||||||
mroot = newMRoot
|
mroot = newMRoot
|
||||||
}
|
}
|
||||||
|
|
||||||
newManifestHash, err := client.UploadManifest(mroot)
|
newManifestHash, err := client.UploadManifest(mroot, isEncrypted)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Manifest upload failed: %v", err)
|
utils.Fatalf("Manifest upload failed: %v", err)
|
||||||
}
|
}
|
||||||
|
169
cmd/swarm/mru.go
Normal file
169
cmd/swarm/mru.go
Normal file
@@ -0,0 +1,169 @@
|
|||||||
|
// Copyright 2016 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
// Command resource allows the user to create and update signed mutable resource updates
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"strconv"
|
||||||
|
"strings"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/common/hexutil"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/cmd/utils"
|
||||||
|
swarm "github.com/ethereum/go-ethereum/swarm/api/client"
|
||||||
|
"github.com/ethereum/go-ethereum/swarm/storage/mru"
|
||||||
|
"gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
func NewGenericSigner(ctx *cli.Context) mru.Signer {
|
||||||
|
return mru.NewGenericSigner(getPrivKey(ctx))
|
||||||
|
}
|
||||||
|
|
||||||
|
// swarm resource create <frequency> [--name <name>] [--data <0x Hexdata> [--multihash=false]]
|
||||||
|
// swarm resource update <Manifest Address or ENS domain> <0x Hexdata> [--multihash=false]
|
||||||
|
// swarm resource info <Manifest Address or ENS domain>
|
||||||
|
|
||||||
|
func resourceCreate(ctx *cli.Context) {
|
||||||
|
args := ctx.Args()
|
||||||
|
|
||||||
|
var (
|
||||||
|
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
||||||
|
client = swarm.NewClient(bzzapi)
|
||||||
|
multihash = ctx.Bool(SwarmResourceMultihashFlag.Name)
|
||||||
|
initialData = ctx.String(SwarmResourceDataOnCreateFlag.Name)
|
||||||
|
name = ctx.String(SwarmResourceNameFlag.Name)
|
||||||
|
)
|
||||||
|
|
||||||
|
if len(args) < 1 {
|
||||||
|
fmt.Println("Incorrect number of arguments")
|
||||||
|
cli.ShowCommandHelpAndExit(ctx, "create", 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
signer := NewGenericSigner(ctx)
|
||||||
|
frequency, err := strconv.ParseUint(args[0], 10, 64)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Frequency formatting error: %s\n", err.Error())
|
||||||
|
cli.ShowCommandHelpAndExit(ctx, "create", 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
metadata := mru.ResourceMetadata{
|
||||||
|
Name: name,
|
||||||
|
Frequency: frequency,
|
||||||
|
Owner: signer.Address(),
|
||||||
|
}
|
||||||
|
|
||||||
|
var newResourceRequest *mru.Request
|
||||||
|
if initialData != "" {
|
||||||
|
initialDataBytes, err := hexutil.Decode(initialData)
|
||||||
|
if err != nil {
|
||||||
|
fmt.Printf("Error parsing data: %s\n", err.Error())
|
||||||
|
cli.ShowCommandHelpAndExit(ctx, "create", 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
newResourceRequest, err = mru.NewCreateUpdateRequest(&metadata)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error creating new resource request: %s", err)
|
||||||
|
}
|
||||||
|
newResourceRequest.SetData(initialDataBytes, multihash)
|
||||||
|
if err = newResourceRequest.Sign(signer); err != nil {
|
||||||
|
utils.Fatalf("Error signing resource update: %s", err.Error())
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
newResourceRequest, err = mru.NewCreateRequest(&metadata)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error creating new resource request: %s", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
manifestAddress, err := client.CreateResource(newResourceRequest)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error creating resource: %s", err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
fmt.Println(manifestAddress) // output manifest address to the user in a single line (useful for other commands to pick up)
|
||||||
|
|
||||||
|
}
|
||||||
|
|
||||||
|
func resourceUpdate(ctx *cli.Context) {
|
||||||
|
args := ctx.Args()
|
||||||
|
|
||||||
|
var (
|
||||||
|
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
||||||
|
client = swarm.NewClient(bzzapi)
|
||||||
|
multihash = ctx.Bool(SwarmResourceMultihashFlag.Name)
|
||||||
|
)
|
||||||
|
|
||||||
|
if len(args) < 2 {
|
||||||
|
fmt.Println("Incorrect number of arguments")
|
||||||
|
cli.ShowCommandHelpAndExit(ctx, "update", 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
signer := NewGenericSigner(ctx)
|
||||||
|
manifestAddressOrDomain := args[0]
|
||||||
|
data, err := hexutil.Decode(args[1])
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error parsing data: %s", err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
|
||||||
|
// Retrieve resource status and metadata out of the manifest
|
||||||
|
updateRequest, err := client.GetResourceMetadata(manifestAddressOrDomain)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error retrieving resource status: %s", err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
// set the new data
|
||||||
|
updateRequest.SetData(data, multihash)
|
||||||
|
|
||||||
|
// sign update
|
||||||
|
if err = updateRequest.Sign(signer); err != nil {
|
||||||
|
utils.Fatalf("Error signing resource update: %s", err.Error())
|
||||||
|
}
|
||||||
|
|
||||||
|
// post update
|
||||||
|
err = client.UpdateResource(updateRequest)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error updating resource: %s", err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func resourceInfo(ctx *cli.Context) {
|
||||||
|
var (
|
||||||
|
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
||||||
|
client = swarm.NewClient(bzzapi)
|
||||||
|
)
|
||||||
|
args := ctx.Args()
|
||||||
|
if len(args) < 1 {
|
||||||
|
fmt.Println("Incorrect number of arguments.")
|
||||||
|
cli.ShowCommandHelpAndExit(ctx, "info", 1)
|
||||||
|
return
|
||||||
|
}
|
||||||
|
manifestAddressOrDomain := args[0]
|
||||||
|
metadata, err := client.GetResourceMetadata(manifestAddressOrDomain)
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error retrieving resource metadata: %s", err.Error())
|
||||||
|
return
|
||||||
|
}
|
||||||
|
encodedMetadata, err := metadata.MarshalJSON()
|
||||||
|
if err != nil {
|
||||||
|
utils.Fatalf("Error encoding metadata to JSON for display:%s", err)
|
||||||
|
}
|
||||||
|
fmt.Println(string(encodedMetadata))
|
||||||
|
}
|
@@ -81,6 +81,7 @@ type testCluster struct {
|
|||||||
//
|
//
|
||||||
// When starting more than one node, they are connected together using the
|
// When starting more than one node, they are connected together using the
|
||||||
// admin SetPeer RPC method.
|
// admin SetPeer RPC method.
|
||||||
|
|
||||||
func newTestCluster(t *testing.T, size int) *testCluster {
|
func newTestCluster(t *testing.T, size int) *testCluster {
|
||||||
cluster := &testCluster{}
|
cluster := &testCluster{}
|
||||||
defer func() {
|
defer func() {
|
||||||
@@ -96,18 +97,7 @@ func newTestCluster(t *testing.T, size int) *testCluster {
|
|||||||
cluster.TmpDir = tmpdir
|
cluster.TmpDir = tmpdir
|
||||||
|
|
||||||
// start the nodes
|
// start the nodes
|
||||||
cluster.Nodes = make([]*testNode, 0, size)
|
cluster.StartNewNodes(t, size)
|
||||||
for i := 0; i < size; i++ {
|
|
||||||
dir := filepath.Join(cluster.TmpDir, fmt.Sprintf("swarm%02d", i))
|
|
||||||
if err := os.Mkdir(dir, 0700); err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
node := newTestNode(t, dir)
|
|
||||||
node.Name = fmt.Sprintf("swarm%02d", i)
|
|
||||||
|
|
||||||
cluster.Nodes = append(cluster.Nodes, node)
|
|
||||||
}
|
|
||||||
|
|
||||||
if size == 1 {
|
if size == 1 {
|
||||||
return cluster
|
return cluster
|
||||||
@@ -145,14 +135,51 @@ func (c *testCluster) Shutdown() {
|
|||||||
os.RemoveAll(c.TmpDir)
|
os.RemoveAll(c.TmpDir)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *testCluster) Stop() {
|
||||||
|
for _, node := range c.Nodes {
|
||||||
|
node.Shutdown()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *testCluster) StartNewNodes(t *testing.T, size int) {
|
||||||
|
c.Nodes = make([]*testNode, 0, size)
|
||||||
|
for i := 0; i < size; i++ {
|
||||||
|
dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
|
||||||
|
if err := os.Mkdir(dir, 0700); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
node := newTestNode(t, dir)
|
||||||
|
node.Name = fmt.Sprintf("swarm%02d", i)
|
||||||
|
|
||||||
|
c.Nodes = append(c.Nodes, node)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *testCluster) StartExistingNodes(t *testing.T, size int, bzzaccount string) {
|
||||||
|
c.Nodes = make([]*testNode, 0, size)
|
||||||
|
for i := 0; i < size; i++ {
|
||||||
|
dir := filepath.Join(c.TmpDir, fmt.Sprintf("swarm%02d", i))
|
||||||
|
node := existingTestNode(t, dir, bzzaccount)
|
||||||
|
node.Name = fmt.Sprintf("swarm%02d", i)
|
||||||
|
|
||||||
|
c.Nodes = append(c.Nodes, node)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *testCluster) Cleanup() {
|
||||||
|
os.RemoveAll(c.TmpDir)
|
||||||
|
}
|
||||||
|
|
||||||
type testNode struct {
|
type testNode struct {
|
||||||
Name string
|
Name string
|
||||||
Addr string
|
Addr string
|
||||||
URL string
|
URL string
|
||||||
Enode string
|
Enode string
|
||||||
Dir string
|
Dir string
|
||||||
Client *rpc.Client
|
IpcPath string
|
||||||
Cmd *cmdtest.TestCmd
|
Client *rpc.Client
|
||||||
|
Cmd *cmdtest.TestCmd
|
||||||
}
|
}
|
||||||
|
|
||||||
const testPassphrase = "swarm-test-passphrase"
|
const testPassphrase = "swarm-test-passphrase"
|
||||||
@@ -181,6 +208,72 @@ func getTestAccount(t *testing.T, dir string) (conf *node.Config, account accoun
|
|||||||
return conf, account
|
return conf, account
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func existingTestNode(t *testing.T, dir string, bzzaccount string) *testNode {
|
||||||
|
conf, _ := getTestAccount(t, dir)
|
||||||
|
node := &testNode{Dir: dir}
|
||||||
|
|
||||||
|
// use a unique IPCPath when running tests on Windows
|
||||||
|
if runtime.GOOS == "windows" {
|
||||||
|
conf.IPCPath = fmt.Sprintf("bzzd-%s.ipc", bzzaccount)
|
||||||
|
}
|
||||||
|
|
||||||
|
// assign ports
|
||||||
|
httpPort, err := assignTCPPort()
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
p2pPort, err := assignTCPPort()
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// start the node
|
||||||
|
node.Cmd = runSwarm(t,
|
||||||
|
"--port", p2pPort,
|
||||||
|
"--nodiscover",
|
||||||
|
"--datadir", dir,
|
||||||
|
"--ipcpath", conf.IPCPath,
|
||||||
|
"--ens-api", "",
|
||||||
|
"--bzzaccount", bzzaccount,
|
||||||
|
"--bzznetworkid", "321",
|
||||||
|
"--bzzport", httpPort,
|
||||||
|
"--verbosity", "6",
|
||||||
|
)
|
||||||
|
node.Cmd.InputLine(testPassphrase)
|
||||||
|
defer func() {
|
||||||
|
if t.Failed() {
|
||||||
|
node.Shutdown()
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
// wait for the node to start
|
||||||
|
for start := time.Now(); time.Since(start) < 10*time.Second; time.Sleep(50 * time.Millisecond) {
|
||||||
|
node.Client, err = rpc.Dial(conf.IPCEndpoint())
|
||||||
|
if err == nil {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if node.Client == nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// load info
|
||||||
|
var info swarm.Info
|
||||||
|
if err := node.Client.Call(&info, "bzz_info"); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
node.Addr = net.JoinHostPort("127.0.0.1", info.Port)
|
||||||
|
node.URL = "http://" + node.Addr
|
||||||
|
|
||||||
|
var nodeInfo p2p.NodeInfo
|
||||||
|
if err := node.Client.Call(&nodeInfo, "admin_nodeInfo"); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
node.Enode = fmt.Sprintf("enode://%s@127.0.0.1:%s", nodeInfo.ID, p2pPort)
|
||||||
|
|
||||||
|
return node
|
||||||
|
}
|
||||||
|
|
||||||
func newTestNode(t *testing.T, dir string) *testNode {
|
func newTestNode(t *testing.T, dir string) *testNode {
|
||||||
|
|
||||||
conf, account := getTestAccount(t, dir)
|
conf, account := getTestAccount(t, dir)
|
||||||
@@ -239,6 +332,7 @@ func newTestNode(t *testing.T, dir string) *testNode {
|
|||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
node.Enode = fmt.Sprintf("enode://%s@127.0.0.1:%s", nodeInfo.ID, p2pPort)
|
node.Enode = fmt.Sprintf("enode://%s@127.0.0.1:%s", nodeInfo.ID, p2pPort)
|
||||||
|
node.IpcPath = conf.IPCPath
|
||||||
|
|
||||||
return node
|
return node
|
||||||
}
|
}
|
||||||
|
101
cmd/swarm/swarm-smoke/main.go
Normal file
101
cmd/swarm/swarm-smoke/main.go
Normal file
@@ -0,0 +1,101 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"os"
|
||||||
|
"sort"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
colorable "github.com/mattn/go-colorable"
|
||||||
|
|
||||||
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
endpoints []string
|
||||||
|
includeLocalhost bool
|
||||||
|
cluster string
|
||||||
|
scheme string
|
||||||
|
filesize int
|
||||||
|
from int
|
||||||
|
to int
|
||||||
|
)
|
||||||
|
|
||||||
|
func main() {
|
||||||
|
log.PrintOrigins(true)
|
||||||
|
log.Root().SetHandler(log.LvlFilterHandler(log.LvlTrace, log.StreamHandler(colorable.NewColorableStderr(), log.TerminalFormat(true))))
|
||||||
|
|
||||||
|
app := cli.NewApp()
|
||||||
|
app.Name = "smoke-test"
|
||||||
|
app.Usage = ""
|
||||||
|
|
||||||
|
app.Flags = []cli.Flag{
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "cluster-endpoint",
|
||||||
|
Value: "testing",
|
||||||
|
Usage: "cluster to point to (open, or testing)",
|
||||||
|
Destination: &cluster,
|
||||||
|
},
|
||||||
|
cli.IntFlag{
|
||||||
|
Name: "cluster-from",
|
||||||
|
Value: 8501,
|
||||||
|
Usage: "swarm node (from)",
|
||||||
|
Destination: &from,
|
||||||
|
},
|
||||||
|
cli.IntFlag{
|
||||||
|
Name: "cluster-to",
|
||||||
|
Value: 8512,
|
||||||
|
Usage: "swarm node (to)",
|
||||||
|
Destination: &to,
|
||||||
|
},
|
||||||
|
cli.StringFlag{
|
||||||
|
Name: "cluster-scheme",
|
||||||
|
Value: "http",
|
||||||
|
Usage: "http or https",
|
||||||
|
Destination: &scheme,
|
||||||
|
},
|
||||||
|
cli.BoolFlag{
|
||||||
|
Name: "include-localhost",
|
||||||
|
Usage: "whether to include localhost:8500 as an endpoint",
|
||||||
|
Destination: &includeLocalhost,
|
||||||
|
},
|
||||||
|
cli.IntFlag{
|
||||||
|
Name: "filesize",
|
||||||
|
Value: 1,
|
||||||
|
Usage: "file size for generated random file in MB",
|
||||||
|
Destination: &filesize,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
app.Commands = []cli.Command{
|
||||||
|
{
|
||||||
|
Name: "upload_and_sync",
|
||||||
|
Aliases: []string{"c"},
|
||||||
|
Usage: "upload and sync",
|
||||||
|
Action: cliUploadAndSync,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
sort.Sort(cli.FlagsByName(app.Flags))
|
||||||
|
sort.Sort(cli.CommandsByName(app.Commands))
|
||||||
|
|
||||||
|
err := app.Run(os.Args)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
}
|
||||||
|
}
|
191
cmd/swarm/swarm-smoke/upload_and_sync.go
Normal file
191
cmd/swarm/swarm-smoke/upload_and_sync.go
Normal file
@@ -0,0 +1,191 @@
|
|||||||
|
// Copyright 2018 The go-ethereum Authors
|
||||||
|
// This file is part of go-ethereum.
|
||||||
|
//
|
||||||
|
// go-ethereum is free software: you can redistribute it and/or modify
|
||||||
|
// it under the terms of the GNU General Public License as published by
|
||||||
|
// the Free Software Foundation, either version 3 of the License, or
|
||||||
|
// (at your option) any later version.
|
||||||
|
//
|
||||||
|
// go-ethereum is distributed in the hope that it will be useful,
|
||||||
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
// GNU General Public License for more details.
|
||||||
|
//
|
||||||
|
// You should have received a copy of the GNU General Public License
|
||||||
|
// along with go-ethereum. If not, see <http://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
package main
|
||||||
|
|
||||||
|
import (
|
||||||
|
"bytes"
|
||||||
|
"crypto/md5"
|
||||||
|
"crypto/rand"
|
||||||
|
"fmt"
|
||||||
|
"io"
|
||||||
|
"io/ioutil"
|
||||||
|
"net/http"
|
||||||
|
"os"
|
||||||
|
"os/exec"
|
||||||
|
"strings"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
"github.com/pborman/uuid"
|
||||||
|
|
||||||
|
cli "gopkg.in/urfave/cli.v1"
|
||||||
|
)
|
||||||
|
|
||||||
|
func generateEndpoints(scheme string, cluster string, from int, to int) {
|
||||||
|
if cluster == "prod" {
|
||||||
|
cluster = ""
|
||||||
|
} else {
|
||||||
|
cluster = cluster + "."
|
||||||
|
}
|
||||||
|
|
||||||
|
for port := from; port <= to; port++ {
|
||||||
|
endpoints = append(endpoints, fmt.Sprintf("%s://%v.%sswarm-gateways.net", scheme, port, cluster))
|
||||||
|
}
|
||||||
|
|
||||||
|
if includeLocalhost {
|
||||||
|
endpoints = append(endpoints, "http://localhost:8500")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func cliUploadAndSync(c *cli.Context) error {
|
||||||
|
defer func(now time.Time) { log.Info("total time", "time", time.Since(now), "size", filesize) }(time.Now())
|
||||||
|
|
||||||
|
generateEndpoints(scheme, cluster, from, to)
|
||||||
|
|
||||||
|
log.Info("uploading to " + endpoints[0] + " and syncing")
|
||||||
|
|
||||||
|
f, cleanup := generateRandomFile(filesize * 1000000)
|
||||||
|
defer cleanup()
|
||||||
|
|
||||||
|
hash, err := upload(f, endpoints[0])
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
fhash, err := digest(f)
|
||||||
|
if err != nil {
|
||||||
|
log.Error(err.Error())
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Info("uploaded successfully", "hash", hash, "digest", fmt.Sprintf("%x", fhash))
|
||||||
|
|
||||||
|
if filesize < 10 {
|
||||||
|
time.Sleep(35 * time.Second)
|
||||||
|
} else {
|
||||||
|
time.Sleep(15 * time.Second)
|
||||||
|
time.Sleep(2 * time.Duration(filesize) * time.Second)
|
||||||
|
}
|
||||||
|
|
||||||
|
wg := sync.WaitGroup{}
|
||||||
|
for _, endpoint := range endpoints {
|
||||||
|
endpoint := endpoint
|
||||||
|
ruid := uuid.New()[:8]
|
||||||
|
wg.Add(1)
|
||||||
|
go func(endpoint string, ruid string) {
|
||||||
|
for {
|
||||||
|
err := fetch(hash, endpoint, fhash, ruid)
|
||||||
|
if err != nil {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
|
||||||
|
wg.Done()
|
||||||
|
return
|
||||||
|
}
|
||||||
|
}(endpoint, ruid)
|
||||||
|
}
|
||||||
|
wg.Wait()
|
||||||
|
log.Info("all endpoints synced random file successfully")
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// fetch is getting the requested `hash` from the `endpoint` and compares it with the `original` file
|
||||||
|
func fetch(hash string, endpoint string, original []byte, ruid string) error {
|
||||||
|
log.Trace("sleeping", "ruid", ruid)
|
||||||
|
time.Sleep(5 * time.Second)
|
||||||
|
|
||||||
|
log.Trace("http get request", "ruid", ruid, "api", endpoint, "hash", hash)
|
||||||
|
res, err := http.Get(endpoint + "/bzz:/" + hash + "/")
|
||||||
|
if err != nil {
|
||||||
|
log.Warn(err.Error(), "ruid", ruid)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
log.Trace("http get response", "ruid", ruid, "api", endpoint, "hash", hash, "code", res.StatusCode, "len", res.ContentLength)
|
||||||
|
|
||||||
|
if res.StatusCode != 200 {
|
||||||
|
err := fmt.Errorf("expected status code %d, got %v", 200, res.StatusCode)
|
||||||
|
log.Warn(err.Error(), "ruid", ruid)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
defer res.Body.Close()
|
||||||
|
|
||||||
|
rdigest, err := digest(res.Body)
|
||||||
|
if err != nil {
|
||||||
|
log.Warn(err.Error(), "ruid", ruid)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
if !bytes.Equal(rdigest, original) {
|
||||||
|
err := fmt.Errorf("downloaded imported file md5=%x is not the same as the generated one=%x", rdigest, original)
|
||||||
|
log.Warn(err.Error(), "ruid", ruid)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
log.Trace("downloaded file matches random file", "ruid", ruid, "len", res.ContentLength)
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// upload is uploading a file `f` to `endpoint` via the `swarm up` cmd
|
||||||
|
func upload(f *os.File, endpoint string) (string, error) {
|
||||||
|
var out bytes.Buffer
|
||||||
|
cmd := exec.Command("swarm", "--bzzapi", endpoint, "up", f.Name())
|
||||||
|
cmd.Stdout = &out
|
||||||
|
err := cmd.Run()
|
||||||
|
if err != nil {
|
||||||
|
return "", err
|
||||||
|
}
|
||||||
|
hash := strings.TrimRight(out.String(), "\r\n")
|
||||||
|
return hash, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func digest(r io.Reader) ([]byte, error) {
|
||||||
|
h := md5.New()
|
||||||
|
_, err := io.Copy(h, r)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
return h.Sum(nil), nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// generateRandomFile is creating a temporary file with the requested byte size
|
||||||
|
func generateRandomFile(size int) (f *os.File, teardown func()) {
|
||||||
|
// create a tmp file
|
||||||
|
tmp, err := ioutil.TempFile("", "swarm-test")
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// callback for tmp file cleanup
|
||||||
|
teardown = func() {
|
||||||
|
tmp.Close()
|
||||||
|
os.Remove(tmp.Name())
|
||||||
|
}
|
||||||
|
|
||||||
|
buf := make([]byte, size)
|
||||||
|
_, err = rand.Read(buf)
|
||||||
|
if err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
ioutil.WriteFile(tmp.Name(), buf, 0755)
|
||||||
|
|
||||||
|
return tmp, teardown
|
||||||
|
}
|
@@ -40,12 +40,13 @@ func upload(ctx *cli.Context) {
|
|||||||
args := ctx.Args()
|
args := ctx.Args()
|
||||||
var (
|
var (
|
||||||
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
bzzapi = strings.TrimRight(ctx.GlobalString(SwarmApiFlag.Name), "/")
|
||||||
recursive = ctx.GlobalBool(SwarmRecursiveUploadFlag.Name)
|
recursive = ctx.GlobalBool(SwarmRecursiveFlag.Name)
|
||||||
wantManifest = ctx.GlobalBoolT(SwarmWantManifestFlag.Name)
|
wantManifest = ctx.GlobalBoolT(SwarmWantManifestFlag.Name)
|
||||||
defaultPath = ctx.GlobalString(SwarmUploadDefaultPath.Name)
|
defaultPath = ctx.GlobalString(SwarmUploadDefaultPath.Name)
|
||||||
fromStdin = ctx.GlobalBool(SwarmUpFromStdinFlag.Name)
|
fromStdin = ctx.GlobalBool(SwarmUpFromStdinFlag.Name)
|
||||||
mimeType = ctx.GlobalString(SwarmUploadMimeType.Name)
|
mimeType = ctx.GlobalString(SwarmUploadMimeType.Name)
|
||||||
client = swarm.NewClient(bzzapi)
|
client = swarm.NewClient(bzzapi)
|
||||||
|
toEncrypt = ctx.Bool(SwarmEncryptedFlag.Name)
|
||||||
file string
|
file string
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -76,7 +77,7 @@ func upload(ctx *cli.Context) {
|
|||||||
utils.Fatalf("Error opening file: %s", err)
|
utils.Fatalf("Error opening file: %s", err)
|
||||||
}
|
}
|
||||||
defer f.Close()
|
defer f.Close()
|
||||||
hash, err := client.UploadRaw(f, f.Size)
|
hash, err := client.UploadRaw(f, f.Size, toEncrypt)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
utils.Fatalf("Upload failed: %s", err)
|
utils.Fatalf("Upload failed: %s", err)
|
||||||
}
|
}
|
||||||
@@ -97,7 +98,7 @@ func upload(ctx *cli.Context) {
|
|||||||
if !recursive {
|
if !recursive {
|
||||||
return "", errors.New("Argument is a directory and recursive upload is disabled")
|
return "", errors.New("Argument is a directory and recursive upload is disabled")
|
||||||
}
|
}
|
||||||
return client.UploadDirectory(file, defaultPath, "")
|
return client.UploadDirectory(file, defaultPath, "", toEncrypt)
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
doUpload = func() (string, error) {
|
doUpload = func() (string, error) {
|
||||||
@@ -110,7 +111,7 @@ func upload(ctx *cli.Context) {
|
|||||||
mimeType = detectMimeType(file)
|
mimeType = detectMimeType(file)
|
||||||
}
|
}
|
||||||
f.ContentType = mimeType
|
f.ContentType = mimeType
|
||||||
return client.Upload(f, "")
|
return client.Upload(f, "", toEncrypt)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
hash, err := doUpload()
|
hash, err := doUpload()
|
||||||
|
@@ -17,60 +17,259 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"bytes"
|
||||||
|
"flag"
|
||||||
|
"fmt"
|
||||||
"io"
|
"io"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
|
"path"
|
||||||
|
"path/filepath"
|
||||||
|
"strings"
|
||||||
"testing"
|
"testing"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ethereum/go-ethereum/log"
|
||||||
|
swarm "github.com/ethereum/go-ethereum/swarm/api/client"
|
||||||
|
colorable "github.com/mattn/go-colorable"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var loglevel = flag.Int("loglevel", 3, "verbosity of logs")
|
||||||
|
|
||||||
|
func init() {
|
||||||
|
log.PrintOrigins(true)
|
||||||
|
log.Root().SetHandler(log.LvlFilterHandler(log.Lvl(*loglevel), log.StreamHandler(colorable.NewColorableStderr(), log.TerminalFormat(true))))
|
||||||
|
}
|
||||||
|
|
||||||
// TestCLISwarmUp tests that running 'swarm up' makes the resulting file
|
// TestCLISwarmUp tests that running 'swarm up' makes the resulting file
|
||||||
// available from all nodes via the HTTP API
|
// available from all nodes via the HTTP API
|
||||||
func TestCLISwarmUp(t *testing.T) {
|
func TestCLISwarmUp(t *testing.T) {
|
||||||
// start 3 node cluster
|
testCLISwarmUp(false, t)
|
||||||
t.Log("starting 3 node cluster")
|
}
|
||||||
|
func TestCLISwarmUpRecursive(t *testing.T) {
|
||||||
|
testCLISwarmUpRecursive(false, t)
|
||||||
|
}
|
||||||
|
|
||||||
|
// TestCLISwarmUpEncrypted tests that running 'swarm encrypted-up' makes the resulting file
|
||||||
|
// available from all nodes via the HTTP API
|
||||||
|
func TestCLISwarmUpEncrypted(t *testing.T) {
|
||||||
|
testCLISwarmUp(true, t)
|
||||||
|
}
|
||||||
|
func TestCLISwarmUpEncryptedRecursive(t *testing.T) {
|
||||||
|
testCLISwarmUpRecursive(true, t)
|
||||||
|
}
|
||||||
|
|
||||||
|
func testCLISwarmUp(toEncrypt bool, t *testing.T) {
|
||||||
|
log.Info("starting 3 node cluster")
|
||||||
cluster := newTestCluster(t, 3)
|
cluster := newTestCluster(t, 3)
|
||||||
defer cluster.Shutdown()
|
defer cluster.Shutdown()
|
||||||
|
|
||||||
// create a tmp file
|
// create a tmp file
|
||||||
tmp, err := ioutil.TempFile("", "swarm-test")
|
tmp, err := ioutil.TempFile("", "swarm-test")
|
||||||
assertNil(t, err)
|
|
||||||
defer tmp.Close()
|
|
||||||
defer os.Remove(tmp.Name())
|
|
||||||
_, err = io.WriteString(tmp, "data")
|
|
||||||
assertNil(t, err)
|
|
||||||
|
|
||||||
// upload the file with 'swarm up' and expect a hash
|
|
||||||
t.Log("uploading file with 'swarm up'")
|
|
||||||
up := runSwarm(t, "--bzzapi", cluster.Nodes[0].URL, "up", tmp.Name())
|
|
||||||
_, matches := up.ExpectRegexp(`[a-f\d]{64}`)
|
|
||||||
up.ExpectExit()
|
|
||||||
hash := matches[0]
|
|
||||||
t.Logf("file uploaded with hash %s", hash)
|
|
||||||
|
|
||||||
// get the file from the HTTP API of each node
|
|
||||||
for _, node := range cluster.Nodes {
|
|
||||||
t.Logf("getting file from %s", node.Name)
|
|
||||||
res, err := http.Get(node.URL + "/bzz:/" + hash)
|
|
||||||
assertNil(t, err)
|
|
||||||
assertHTTPResponse(t, res, http.StatusOK, "data")
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func assertNil(t *testing.T, err error) {
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
t.Fatal(err)
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
|
defer tmp.Close()
|
||||||
|
defer os.Remove(tmp.Name())
|
||||||
|
|
||||||
|
// write data to file
|
||||||
|
data := "notsorandomdata"
|
||||||
|
_, err = io.WriteString(tmp, data)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
hashRegexp := `[a-f\d]{64}`
|
||||||
|
flags := []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"up",
|
||||||
|
tmp.Name()}
|
||||||
|
if toEncrypt {
|
||||||
|
hashRegexp = `[a-f\d]{128}`
|
||||||
|
flags = []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"up",
|
||||||
|
"--encrypt",
|
||||||
|
tmp.Name()}
|
||||||
|
}
|
||||||
|
// upload the file with 'swarm up' and expect a hash
|
||||||
|
log.Info(fmt.Sprintf("uploading file with 'swarm up'"))
|
||||||
|
up := runSwarm(t, flags...)
|
||||||
|
_, matches := up.ExpectRegexp(hashRegexp)
|
||||||
|
up.ExpectExit()
|
||||||
|
hash := matches[0]
|
||||||
|
log.Info("file uploaded", "hash", hash)
|
||||||
|
|
||||||
|
// get the file from the HTTP API of each node
|
||||||
|
for _, node := range cluster.Nodes {
|
||||||
|
log.Info("getting file from node", "node", node.Name)
|
||||||
|
|
||||||
|
res, err := http.Get(node.URL + "/bzz:/" + hash)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer res.Body.Close()
|
||||||
|
|
||||||
|
reply, err := ioutil.ReadAll(res.Body)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
if res.StatusCode != 200 {
|
||||||
|
t.Fatalf("expected HTTP status 200, got %s", res.Status)
|
||||||
|
}
|
||||||
|
if string(reply) != data {
|
||||||
|
t.Fatalf("expected HTTP body %q, got %q", data, reply)
|
||||||
|
}
|
||||||
|
log.Debug("verifying uploaded file using `swarm down`")
|
||||||
|
//try to get the content with `swarm down`
|
||||||
|
tmpDownload, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
tmpDownload = path.Join(tmpDownload, "tmpfile.tmp")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(tmpDownload)
|
||||||
|
|
||||||
|
bzzLocator := "bzz:/" + hash
|
||||||
|
flags = []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"down",
|
||||||
|
bzzLocator,
|
||||||
|
tmpDownload,
|
||||||
|
}
|
||||||
|
|
||||||
|
down := runSwarm(t, flags...)
|
||||||
|
down.ExpectExit()
|
||||||
|
|
||||||
|
fi, err := os.Stat(tmpDownload)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("could not stat path: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
switch mode := fi.Mode(); {
|
||||||
|
case mode.IsRegular():
|
||||||
|
downloadedBytes, err := ioutil.ReadFile(tmpDownload)
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("had an error reading the downloaded file: %v", err)
|
||||||
|
}
|
||||||
|
if !bytes.Equal(downloadedBytes, bytes.NewBufferString(data).Bytes()) {
|
||||||
|
t.Fatalf("retrieved data and posted data not equal!")
|
||||||
|
}
|
||||||
|
|
||||||
|
default:
|
||||||
|
t.Fatalf("expected to download regular file, got %s", fi.Mode())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
timeout := time.Duration(2 * time.Second)
|
||||||
|
httpClient := http.Client{
|
||||||
|
Timeout: timeout,
|
||||||
|
}
|
||||||
|
|
||||||
|
// try to squeeze a timeout by getting an non-existent hash from each node
|
||||||
|
for _, node := range cluster.Nodes {
|
||||||
|
_, err := httpClient.Get(node.URL + "/bzz:/1023e8bae0f70be7d7b5f74343088ba408a218254391490c85ae16278e230340")
|
||||||
|
// we're speeding up the timeout here since netstore has a 60 seconds timeout on a request
|
||||||
|
if err != nil && !strings.Contains(err.Error(), "Client.Timeout exceeded while awaiting headers") {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
// this is disabled since it takes 60s due to netstore timeout
|
||||||
|
// if res.StatusCode != 404 {
|
||||||
|
// t.Fatalf("expected HTTP status 404, got %s", res.Status)
|
||||||
|
// }
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
func assertHTTPResponse(t *testing.T, res *http.Response, expectedStatus int, expectedBody string) {
|
func testCLISwarmUpRecursive(toEncrypt bool, t *testing.T) {
|
||||||
defer res.Body.Close()
|
fmt.Println("starting 3 node cluster")
|
||||||
if res.StatusCode != expectedStatus {
|
cluster := newTestCluster(t, 3)
|
||||||
t.Fatalf("expected HTTP status %d, got %s", expectedStatus, res.Status)
|
defer cluster.Shutdown()
|
||||||
|
|
||||||
|
tmpUploadDir, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
}
|
}
|
||||||
data, err := ioutil.ReadAll(res.Body)
|
defer os.RemoveAll(tmpUploadDir)
|
||||||
assertNil(t, err)
|
// create tmp files
|
||||||
if string(data) != expectedBody {
|
data := "notsorandomdata"
|
||||||
t.Fatalf("expected HTTP body %q, got %q", expectedBody, data)
|
for _, path := range []string{"tmp1", "tmp2"} {
|
||||||
|
if err := ioutil.WriteFile(filepath.Join(tmpUploadDir, path), bytes.NewBufferString(data).Bytes(), 0644); err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
hashRegexp := `[a-f\d]{64}`
|
||||||
|
flags := []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"--recursive",
|
||||||
|
"up",
|
||||||
|
tmpUploadDir}
|
||||||
|
if toEncrypt {
|
||||||
|
hashRegexp = `[a-f\d]{128}`
|
||||||
|
flags = []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"--recursive",
|
||||||
|
"up",
|
||||||
|
"--encrypt",
|
||||||
|
tmpUploadDir}
|
||||||
|
}
|
||||||
|
// upload the file with 'swarm up' and expect a hash
|
||||||
|
log.Info(fmt.Sprintf("uploading file with 'swarm up'"))
|
||||||
|
up := runSwarm(t, flags...)
|
||||||
|
_, matches := up.ExpectRegexp(hashRegexp)
|
||||||
|
up.ExpectExit()
|
||||||
|
hash := matches[0]
|
||||||
|
log.Info("dir uploaded", "hash", hash)
|
||||||
|
|
||||||
|
// get the file from the HTTP API of each node
|
||||||
|
for _, node := range cluster.Nodes {
|
||||||
|
log.Info("getting file from node", "node", node.Name)
|
||||||
|
//try to get the content with `swarm down`
|
||||||
|
tmpDownload, err := ioutil.TempDir("", "swarm-test")
|
||||||
|
if err != nil {
|
||||||
|
t.Fatal(err)
|
||||||
|
}
|
||||||
|
defer os.RemoveAll(tmpDownload)
|
||||||
|
bzzLocator := "bzz:/" + hash
|
||||||
|
flagss := []string{}
|
||||||
|
flagss = []string{
|
||||||
|
"--bzzapi", cluster.Nodes[0].URL,
|
||||||
|
"down",
|
||||||
|
"--recursive",
|
||||||
|
bzzLocator,
|
||||||
|
tmpDownload,
|
||||||
|
}
|
||||||
|
|
||||||
|
fmt.Println("downloading from swarm with recursive")
|
||||||
|
down := runSwarm(t, flagss...)
|
||||||
|
down.ExpectExit()
|
||||||
|
|
||||||
|
files, err := ioutil.ReadDir(tmpDownload)
|
||||||
|
for _, v := range files {
|
||||||
|
fi, err := os.Stat(path.Join(tmpDownload, v.Name()))
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("got an error: %v", err)
|
||||||
|
}
|
||||||
|
|
||||||
|
switch mode := fi.Mode(); {
|
||||||
|
case mode.IsRegular():
|
||||||
|
if file, err := swarm.Open(path.Join(tmpDownload, v.Name())); err != nil {
|
||||||
|
t.Fatalf("encountered an error opening the file returned from the CLI: %v", err)
|
||||||
|
} else {
|
||||||
|
ff := make([]byte, len(data))
|
||||||
|
io.ReadFull(file, ff)
|
||||||
|
buf := bytes.NewBufferString(data)
|
||||||
|
|
||||||
|
if !bytes.Equal(ff, buf.Bytes()) {
|
||||||
|
t.Fatalf("retrieved data and posted data not equal!")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
default:
|
||||||
|
t.Fatalf("this shouldnt happen")
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
t.Fatalf("could not list files at: %v", files)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -29,6 +29,7 @@ import (
|
|||||||
|
|
||||||
"github.com/ethereum/go-ethereum/common"
|
"github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
"github.com/ethereum/go-ethereum/core"
|
||||||
|
"github.com/ethereum/go-ethereum/core/rawdb"
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/ethereum/go-ethereum/ethdb"
|
"github.com/ethereum/go-ethereum/ethdb"
|
||||||
@@ -271,15 +272,13 @@ func ImportPreimages(db *ethdb.LDBDatabase, fn string) error {
|
|||||||
// Accumulate the preimages and flush when enough ws gathered
|
// Accumulate the preimages and flush when enough ws gathered
|
||||||
preimages[crypto.Keccak256Hash(blob)] = common.CopyBytes(blob)
|
preimages[crypto.Keccak256Hash(blob)] = common.CopyBytes(blob)
|
||||||
if len(preimages) > 1024 {
|
if len(preimages) > 1024 {
|
||||||
if err := core.WritePreimages(db, 0, preimages); err != nil {
|
rawdb.WritePreimages(db, 0, preimages)
|
||||||
return err
|
|
||||||
}
|
|
||||||
preimages = make(map[common.Hash][]byte)
|
preimages = make(map[common.Hash][]byte)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
// Flush the last batch preimage data
|
// Flush the last batch preimage data
|
||||||
if len(preimages) > 0 {
|
if len(preimages) > 0 {
|
||||||
return core.WritePreimages(db, 0, preimages)
|
rawdb.WritePreimages(db, 0, preimages)
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
@@ -27,6 +27,7 @@ import (
|
|||||||
"runtime"
|
"runtime"
|
||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/accounts"
|
"github.com/ethereum/go-ethereum/accounts"
|
||||||
"github.com/ethereum/go-ethereum/accounts/keystore"
|
"github.com/ethereum/go-ethereum/accounts/keystore"
|
||||||
@@ -48,6 +49,7 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/les"
|
"github.com/ethereum/go-ethereum/les"
|
||||||
"github.com/ethereum/go-ethereum/log"
|
"github.com/ethereum/go-ethereum/log"
|
||||||
"github.com/ethereum/go-ethereum/metrics"
|
"github.com/ethereum/go-ethereum/metrics"
|
||||||
|
"github.com/ethereum/go-ethereum/metrics/influxdb"
|
||||||
"github.com/ethereum/go-ethereum/node"
|
"github.com/ethereum/go-ethereum/node"
|
||||||
"github.com/ethereum/go-ethereum/p2p"
|
"github.com/ethereum/go-ethereum/p2p"
|
||||||
"github.com/ethereum/go-ethereum/p2p/discover"
|
"github.com/ethereum/go-ethereum/p2p/discover"
|
||||||
@@ -96,7 +98,7 @@ func NewApp(gitCommit, usage string) *cli.App {
|
|||||||
app.Author = ""
|
app.Author = ""
|
||||||
//app.Authors = nil
|
//app.Authors = nil
|
||||||
app.Email = ""
|
app.Email = ""
|
||||||
app.Version = params.Version
|
app.Version = params.VersionWithMeta
|
||||||
if len(gitCommit) >= 8 {
|
if len(gitCommit) >= 8 {
|
||||||
app.Version += "-" + gitCommit[:8]
|
app.Version += "-" + gitCommit[:8]
|
||||||
}
|
}
|
||||||
@@ -191,7 +193,7 @@ var (
|
|||||||
}
|
}
|
||||||
// Dashboard settings
|
// Dashboard settings
|
||||||
DashboardEnabledFlag = cli.BoolFlag{
|
DashboardEnabledFlag = cli.BoolFlag{
|
||||||
Name: "dashboard",
|
Name: metrics.DashboardEnabledFlag,
|
||||||
Usage: "Enable the dashboard",
|
Usage: "Enable the dashboard",
|
||||||
}
|
}
|
||||||
DashboardAddrFlag = cli.StringFlag{
|
DashboardAddrFlag = cli.StringFlag{
|
||||||
@@ -360,10 +362,6 @@ var (
|
|||||||
Name: "ethstats",
|
Name: "ethstats",
|
||||||
Usage: "Reporting URL of a ethstats service (nodename:secret@host:port)",
|
Usage: "Reporting URL of a ethstats service (nodename:secret@host:port)",
|
||||||
}
|
}
|
||||||
MetricsEnabledFlag = cli.BoolFlag{
|
|
||||||
Name: metrics.MetricsEnabledFlag,
|
|
||||||
Usage: "Enable metrics collection and reporting",
|
|
||||||
}
|
|
||||||
FakePoWFlag = cli.BoolFlag{
|
FakePoWFlag = cli.BoolFlag{
|
||||||
Name: "fakepow",
|
Name: "fakepow",
|
||||||
Usage: "Disables proof-of-work verification",
|
Usage: "Disables proof-of-work verification",
|
||||||
@@ -532,6 +530,45 @@ var (
|
|||||||
Usage: "Minimum POW accepted",
|
Usage: "Minimum POW accepted",
|
||||||
Value: whisper.DefaultMinimumPoW,
|
Value: whisper.DefaultMinimumPoW,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Metrics flags
|
||||||
|
MetricsEnabledFlag = cli.BoolFlag{
|
||||||
|
Name: metrics.MetricsEnabledFlag,
|
||||||
|
Usage: "Enable metrics collection and reporting",
|
||||||
|
}
|
||||||
|
MetricsEnableInfluxDBFlag = cli.BoolFlag{
|
||||||
|
Name: "metrics.influxdb",
|
||||||
|
Usage: "Enable metrics export/push to an external InfluxDB database",
|
||||||
|
}
|
||||||
|
MetricsInfluxDBEndpointFlag = cli.StringFlag{
|
||||||
|
Name: "metrics.influxdb.endpoint",
|
||||||
|
Usage: "InfluxDB API endpoint to report metrics to",
|
||||||
|
Value: "http://localhost:8086",
|
||||||
|
}
|
||||||
|
MetricsInfluxDBDatabaseFlag = cli.StringFlag{
|
||||||
|
Name: "metrics.influxdb.database",
|
||||||
|
Usage: "InfluxDB database name to push reported metrics to",
|
||||||
|
Value: "geth",
|
||||||
|
}
|
||||||
|
MetricsInfluxDBUsernameFlag = cli.StringFlag{
|
||||||
|
Name: "metrics.influxdb.username",
|
||||||
|
Usage: "Username to authorize access to the database",
|
||||||
|
Value: "test",
|
||||||
|
}
|
||||||
|
MetricsInfluxDBPasswordFlag = cli.StringFlag{
|
||||||
|
Name: "metrics.influxdb.password",
|
||||||
|
Usage: "Password to authorize access to the database",
|
||||||
|
Value: "test",
|
||||||
|
}
|
||||||
|
// The `host` tag is part of every measurement sent to InfluxDB. Queries on tags are faster in InfluxDB.
|
||||||
|
// It is used so that we can group all nodes and average a measurement across all of them, but also so
|
||||||
|
// that we can select a specific node and inspect its measurements.
|
||||||
|
// https://docs.influxdata.com/influxdb/v1.4/concepts/key_concepts/#tag-key
|
||||||
|
MetricsInfluxDBHostTagFlag = cli.StringFlag{
|
||||||
|
Name: "metrics.influxdb.host.tag",
|
||||||
|
Usage: "InfluxDB `host` tag attached to all measurements",
|
||||||
|
Value: "localhost",
|
||||||
|
}
|
||||||
)
|
)
|
||||||
|
|
||||||
// MakeDataDir retrieves the currently requested data directory, terminating
|
// MakeDataDir retrieves the currently requested data directory, terminating
|
||||||
@@ -607,8 +644,7 @@ func setBootstrapNodes(ctx *cli.Context, cfg *p2p.Config) {
|
|||||||
for _, url := range urls {
|
for _, url := range urls {
|
||||||
node, err := discover.ParseNode(url)
|
node, err := discover.ParseNode(url)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Error("Bootstrap URL invalid", "enode", url, "err", err)
|
log.Crit("Bootstrap URL invalid", "enode", url, "err", err)
|
||||||
continue
|
|
||||||
}
|
}
|
||||||
cfg.BootstrapNodes = append(cfg.BootstrapNodes, node)
|
cfg.BootstrapNodes = append(cfg.BootstrapNodes, node)
|
||||||
}
|
}
|
||||||
@@ -961,7 +997,7 @@ func setEthash(ctx *cli.Context, cfg *eth.Config) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// checkExclusive verifies that only a single isntance of the provided flags was
|
// checkExclusive verifies that only a single instance of the provided flags was
|
||||||
// set by the user. Each flag might optionally be followed by a string type to
|
// set by the user. Each flag might optionally be followed by a string type to
|
||||||
// specialize it further.
|
// specialize it further.
|
||||||
func checkExclusive(ctx *cli.Context, args ...interface{}) {
|
func checkExclusive(ctx *cli.Context, args ...interface{}) {
|
||||||
@@ -1084,6 +1120,9 @@ func SetEthConfig(ctx *cli.Context, stack *node.Node, cfg *eth.Config) {
|
|||||||
}
|
}
|
||||||
cfg.Genesis = core.DefaultRinkebyGenesisBlock()
|
cfg.Genesis = core.DefaultRinkebyGenesisBlock()
|
||||||
case ctx.GlobalBool(DeveloperFlag.Name):
|
case ctx.GlobalBool(DeveloperFlag.Name):
|
||||||
|
if !ctx.GlobalIsSet(NetworkIdFlag.Name) {
|
||||||
|
cfg.NetworkId = 1337
|
||||||
|
}
|
||||||
// Create new developer account or reuse existing one
|
// Create new developer account or reuse existing one
|
||||||
var (
|
var (
|
||||||
developer accounts.Account
|
developer accounts.Account
|
||||||
@@ -1145,7 +1184,7 @@ func RegisterEthService(stack *node.Node, cfg *eth.Config) {
|
|||||||
// RegisterDashboardService adds a dashboard to the stack.
|
// RegisterDashboardService adds a dashboard to the stack.
|
||||||
func RegisterDashboardService(stack *node.Node, cfg *dashboard.Config, commit string) {
|
func RegisterDashboardService(stack *node.Node, cfg *dashboard.Config, commit string) {
|
||||||
stack.Register(func(ctx *node.ServiceContext) (node.Service, error) {
|
stack.Register(func(ctx *node.ServiceContext) (node.Service, error) {
|
||||||
return dashboard.New(cfg, commit)
|
return dashboard.New(cfg, commit, ctx.ResolvePath("logs")), nil
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1159,7 +1198,7 @@ func RegisterShhService(stack *node.Node, cfg *whisper.Config) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// RegisterEthStatsService configures the Ethereum Stats daemon and adds it to
|
// RegisterEthStatsService configures the Ethereum Stats daemon and adds it to
|
||||||
// th egiven node.
|
// the given node.
|
||||||
func RegisterEthStatsService(stack *node.Node, url string) {
|
func RegisterEthStatsService(stack *node.Node, url string) {
|
||||||
if err := stack.Register(func(ctx *node.ServiceContext) (node.Service, error) {
|
if err := stack.Register(func(ctx *node.ServiceContext) (node.Service, error) {
|
||||||
// Retrieve both eth and les services
|
// Retrieve both eth and les services
|
||||||
@@ -1181,6 +1220,27 @@ func SetupNetwork(ctx *cli.Context) {
|
|||||||
params.TargetGasLimit = ctx.GlobalUint64(TargetGasLimitFlag.Name)
|
params.TargetGasLimit = ctx.GlobalUint64(TargetGasLimitFlag.Name)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func SetupMetrics(ctx *cli.Context) {
|
||||||
|
if metrics.Enabled {
|
||||||
|
log.Info("Enabling metrics collection")
|
||||||
|
var (
|
||||||
|
enableExport = ctx.GlobalBool(MetricsEnableInfluxDBFlag.Name)
|
||||||
|
endpoint = ctx.GlobalString(MetricsInfluxDBEndpointFlag.Name)
|
||||||
|
database = ctx.GlobalString(MetricsInfluxDBDatabaseFlag.Name)
|
||||||
|
username = ctx.GlobalString(MetricsInfluxDBUsernameFlag.Name)
|
||||||
|
password = ctx.GlobalString(MetricsInfluxDBPasswordFlag.Name)
|
||||||
|
hosttag = ctx.GlobalString(MetricsInfluxDBHostTagFlag.Name)
|
||||||
|
)
|
||||||
|
|
||||||
|
if enableExport {
|
||||||
|
log.Info("Enabling metrics export to InfluxDB")
|
||||||
|
go influxdb.InfluxDBWithTags(metrics.DefaultRegistry, 10*time.Second, endpoint, database, username, password, "geth.", map[string]string{
|
||||||
|
"host": hosttag,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// MakeChainDatabase open an LevelDB using the flags passed to the client and will hard crash if it fails.
|
// MakeChainDatabase open an LevelDB using the flags passed to the client and will hard crash if it fails.
|
||||||
func MakeChainDatabase(ctx *cli.Context, stack *node.Node) ethdb.Database {
|
func MakeChainDatabase(ctx *cli.Context, stack *node.Node) ethdb.Database {
|
||||||
var (
|
var (
|
||||||
|
@@ -140,8 +140,8 @@ func processArgs() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if *asymmetricMode && len(*argPub) > 0 {
|
if *asymmetricMode && len(*argPub) > 0 {
|
||||||
pub = crypto.ToECDSAPub(common.FromHex(*argPub))
|
var err error
|
||||||
if !isKeyValid(pub) {
|
if pub, err = crypto.UnmarshalPubkey(common.FromHex(*argPub)); err != nil {
|
||||||
utils.Fatalf("invalid public key")
|
utils.Fatalf("invalid public key")
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -271,7 +271,9 @@ func initialize() {
|
|||||||
|
|
||||||
if *mailServerMode {
|
if *mailServerMode {
|
||||||
shh.RegisterServer(&mailServer)
|
shh.RegisterServer(&mailServer)
|
||||||
mailServer.Init(shh, *argDBPath, msPassword, *argServerPoW)
|
if err := mailServer.Init(shh, *argDBPath, msPassword, *argServerPoW); err != nil {
|
||||||
|
utils.Fatalf("Failed to init MailServer: %s", err)
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
server = &p2p.Server{
|
server = &p2p.Server{
|
||||||
@@ -319,10 +321,6 @@ func startServer() error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func isKeyValid(k *ecdsa.PublicKey) bool {
|
|
||||||
return k.X != nil && k.Y != nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func configureNode() {
|
func configureNode() {
|
||||||
var err error
|
var err error
|
||||||
var p2pAccept bool
|
var p2pAccept bool
|
||||||
@@ -338,9 +336,8 @@ func configureNode() {
|
|||||||
if b == nil {
|
if b == nil {
|
||||||
utils.Fatalf("Error: can not convert hexadecimal string")
|
utils.Fatalf("Error: can not convert hexadecimal string")
|
||||||
}
|
}
|
||||||
pub = crypto.ToECDSAPub(b)
|
if pub, err = crypto.UnmarshalPubkey(b); err != nil {
|
||||||
if !isKeyValid(pub) {
|
utils.Fatalf("Error: invalid peer public key")
|
||||||
utils.Fatalf("Error: invalid public key")
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -19,15 +19,20 @@ package common
|
|||||||
|
|
||||||
import "encoding/hex"
|
import "encoding/hex"
|
||||||
|
|
||||||
|
// ToHex returns the hex representation of b, prefixed with '0x'.
|
||||||
|
// For empty slices, the return value is "0x0".
|
||||||
|
//
|
||||||
|
// Deprecated: use hexutil.Encode instead.
|
||||||
func ToHex(b []byte) string {
|
func ToHex(b []byte) string {
|
||||||
hex := Bytes2Hex(b)
|
hex := Bytes2Hex(b)
|
||||||
// Prefer output of "0x0" instead of "0x"
|
|
||||||
if len(hex) == 0 {
|
if len(hex) == 0 {
|
||||||
hex = "0"
|
hex = "0"
|
||||||
}
|
}
|
||||||
return "0x" + hex
|
return "0x" + hex
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// FromHex returns the bytes represented by the hexadecimal string s.
|
||||||
|
// s may be prefixed with "0x".
|
||||||
func FromHex(s string) []byte {
|
func FromHex(s string) []byte {
|
||||||
if len(s) > 1 {
|
if len(s) > 1 {
|
||||||
if s[0:2] == "0x" || s[0:2] == "0X" {
|
if s[0:2] == "0x" || s[0:2] == "0X" {
|
||||||
@@ -40,9 +45,7 @@ func FromHex(s string) []byte {
|
|||||||
return Hex2Bytes(s)
|
return Hex2Bytes(s)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Copy bytes
|
// CopyBytes returns an exact copy of the provided bytes.
|
||||||
//
|
|
||||||
// Returns an exact copy of the provided bytes
|
|
||||||
func CopyBytes(b []byte) (copiedBytes []byte) {
|
func CopyBytes(b []byte) (copiedBytes []byte) {
|
||||||
if b == nil {
|
if b == nil {
|
||||||
return nil
|
return nil
|
||||||
@@ -53,14 +56,17 @@ func CopyBytes(b []byte) (copiedBytes []byte) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// hasHexPrefix validates str begins with '0x' or '0X'.
|
||||||
func hasHexPrefix(str string) bool {
|
func hasHexPrefix(str string) bool {
|
||||||
return len(str) >= 2 && str[0] == '0' && (str[1] == 'x' || str[1] == 'X')
|
return len(str) >= 2 && str[0] == '0' && (str[1] == 'x' || str[1] == 'X')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isHexCharacter returns bool of c being a valid hexadecimal.
|
||||||
func isHexCharacter(c byte) bool {
|
func isHexCharacter(c byte) bool {
|
||||||
return ('0' <= c && c <= '9') || ('a' <= c && c <= 'f') || ('A' <= c && c <= 'F')
|
return ('0' <= c && c <= '9') || ('a' <= c && c <= 'f') || ('A' <= c && c <= 'F')
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// isHex validates whether each byte is valid hexadecimal string.
|
||||||
func isHex(str string) bool {
|
func isHex(str string) bool {
|
||||||
if len(str)%2 != 0 {
|
if len(str)%2 != 0 {
|
||||||
return false
|
return false
|
||||||
@@ -73,31 +79,32 @@ func isHex(str string) bool {
|
|||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Bytes2Hex returns the hexadecimal encoding of d.
|
||||||
func Bytes2Hex(d []byte) string {
|
func Bytes2Hex(d []byte) string {
|
||||||
return hex.EncodeToString(d)
|
return hex.EncodeToString(d)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Hex2Bytes returns the bytes represented by the hexadecimal string str.
|
||||||
func Hex2Bytes(str string) []byte {
|
func Hex2Bytes(str string) []byte {
|
||||||
h, _ := hex.DecodeString(str)
|
h, _ := hex.DecodeString(str)
|
||||||
|
|
||||||
return h
|
return h
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Hex2BytesFixed returns bytes of a specified fixed length flen.
|
||||||
func Hex2BytesFixed(str string, flen int) []byte {
|
func Hex2BytesFixed(str string, flen int) []byte {
|
||||||
h, _ := hex.DecodeString(str)
|
h, _ := hex.DecodeString(str)
|
||||||
if len(h) == flen {
|
if len(h) == flen {
|
||||||
return h
|
return h
|
||||||
} else {
|
|
||||||
if len(h) > flen {
|
|
||||||
return h[len(h)-flen:]
|
|
||||||
} else {
|
|
||||||
hh := make([]byte, flen)
|
|
||||||
copy(hh[flen-len(h):flen], h[:])
|
|
||||||
return hh
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
if len(h) > flen {
|
||||||
|
return h[len(h)-flen:]
|
||||||
|
}
|
||||||
|
hh := make([]byte, flen)
|
||||||
|
copy(hh[flen-len(h):flen], h[:])
|
||||||
|
return hh
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// RightPadBytes zero-pads slice to the right up to length l.
|
||||||
func RightPadBytes(slice []byte, l int) []byte {
|
func RightPadBytes(slice []byte, l int) []byte {
|
||||||
if l <= len(slice) {
|
if l <= len(slice) {
|
||||||
return slice
|
return slice
|
||||||
@@ -109,6 +116,7 @@ func RightPadBytes(slice []byte, l int) []byte {
|
|||||||
return padded
|
return padded
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// LeftPadBytes zero-pads slice to the left up to length l.
|
||||||
func LeftPadBytes(slice []byte, l int) []byte {
|
func LeftPadBytes(slice []byte, l int) []byte {
|
||||||
if l <= len(slice) {
|
if l <= len(slice) {
|
||||||
return slice
|
return slice
|
||||||
|
@@ -31,11 +31,17 @@ import (
|
|||||||
|
|
||||||
var versionRegexp = regexp.MustCompile(`([0-9]+)\.([0-9]+)\.([0-9]+)`)
|
var versionRegexp = regexp.MustCompile(`([0-9]+)\.([0-9]+)\.([0-9]+)`)
|
||||||
|
|
||||||
|
// Contract contains information about a compiled contract, alongside its code.
|
||||||
type Contract struct {
|
type Contract struct {
|
||||||
Code string `json:"code"`
|
Code string `json:"code"`
|
||||||
Info ContractInfo `json:"info"`
|
Info ContractInfo `json:"info"`
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// ContractInfo contains information about a compiled contract, including access
|
||||||
|
// to the ABI definition, user and developer docs, and metadata.
|
||||||
|
//
|
||||||
|
// Depending on the source, language version, compiler version, and compiler
|
||||||
|
// options will provide information about how the contract was compiled.
|
||||||
type ContractInfo struct {
|
type ContractInfo struct {
|
||||||
Source string `json:"source"`
|
Source string `json:"source"`
|
||||||
Language string `json:"language"`
|
Language string `json:"language"`
|
||||||
@@ -142,8 +148,22 @@ func (s *Solidity) run(cmd *exec.Cmd, source string) (map[string]*Contract, erro
|
|||||||
if err := cmd.Run(); err != nil {
|
if err := cmd.Run(); err != nil {
|
||||||
return nil, fmt.Errorf("solc: %v\n%s", err, stderr.Bytes())
|
return nil, fmt.Errorf("solc: %v\n%s", err, stderr.Bytes())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
return ParseCombinedJSON(stdout.Bytes(), source, s.Version, s.Version, strings.Join(s.makeArgs(), " "))
|
||||||
|
}
|
||||||
|
|
||||||
|
// ParseCombinedJSON takes the direct output of a solc --combined-output run and
|
||||||
|
// parses it into a map of string contract name to Contract structs. The
|
||||||
|
// provided source, language and compiler version, and compiler options are all
|
||||||
|
// passed through into the Contract structs.
|
||||||
|
//
|
||||||
|
// The solc output is expected to contain ABI, user docs, and dev docs.
|
||||||
|
//
|
||||||
|
// Returns an error if the JSON is malformed or missing data, or if the JSON
|
||||||
|
// embedded within the JSON is malformed.
|
||||||
|
func ParseCombinedJSON(combinedJSON []byte, source string, languageVersion string, compilerVersion string, compilerOptions string) (map[string]*Contract, error) {
|
||||||
var output solcOutput
|
var output solcOutput
|
||||||
if err := json.Unmarshal(stdout.Bytes(), &output); err != nil {
|
if err := json.Unmarshal(combinedJSON, &output); err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -168,9 +188,9 @@ func (s *Solidity) run(cmd *exec.Cmd, source string) (map[string]*Contract, erro
|
|||||||
Info: ContractInfo{
|
Info: ContractInfo{
|
||||||
Source: source,
|
Source: source,
|
||||||
Language: "Solidity",
|
Language: "Solidity",
|
||||||
LanguageVersion: s.Version,
|
LanguageVersion: languageVersion,
|
||||||
CompilerVersion: s.Version,
|
CompilerVersion: compilerVersion,
|
||||||
CompilerOptions: strings.Join(s.makeArgs(), " "),
|
CompilerOptions: compilerOptions,
|
||||||
AbiDefinition: abi,
|
AbiDefinition: abi,
|
||||||
UserDoc: userdoc,
|
UserDoc: userdoc,
|
||||||
DeveloperDoc: devdoc,
|
DeveloperDoc: devdoc,
|
||||||
|
@@ -39,6 +39,7 @@ import (
|
|||||||
|
|
||||||
const uintBits = 32 << (uint64(^uint(0)) >> 63)
|
const uintBits = 32 << (uint64(^uint(0)) >> 63)
|
||||||
|
|
||||||
|
// Errors
|
||||||
var (
|
var (
|
||||||
ErrEmptyString = &decError{"empty hex string"}
|
ErrEmptyString = &decError{"empty hex string"}
|
||||||
ErrSyntax = &decError{"invalid hex string"}
|
ErrSyntax = &decError{"invalid hex string"}
|
||||||
|
@@ -22,12 +22,13 @@ import (
|
|||||||
"math/big"
|
"math/big"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
// Various big integer limit values.
|
||||||
var (
|
var (
|
||||||
tt255 = BigPow(2, 255)
|
tt255 = BigPow(2, 255)
|
||||||
tt256 = BigPow(2, 256)
|
tt256 = BigPow(2, 256)
|
||||||
tt256m1 = new(big.Int).Sub(tt256, big.NewInt(1))
|
tt256m1 = new(big.Int).Sub(tt256, big.NewInt(1))
|
||||||
MaxBig256 = new(big.Int).Set(tt256m1)
|
|
||||||
tt63 = BigPow(2, 63)
|
tt63 = BigPow(2, 63)
|
||||||
|
MaxBig256 = new(big.Int).Set(tt256m1)
|
||||||
MaxBig63 = new(big.Int).Sub(tt63, big.NewInt(1))
|
MaxBig63 = new(big.Int).Sub(tt63, big.NewInt(1))
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -78,7 +79,7 @@ func ParseBig256(s string) (*big.Int, bool) {
|
|||||||
return bigint, ok
|
return bigint, ok
|
||||||
}
|
}
|
||||||
|
|
||||||
// MustParseBig parses s as a 256 bit big integer and panics if the string is invalid.
|
// MustParseBig256 parses s as a 256 bit big integer and panics if the string is invalid.
|
||||||
func MustParseBig256(s string) *big.Int {
|
func MustParseBig256(s string) *big.Int {
|
||||||
v, ok := ParseBig256(s)
|
v, ok := ParseBig256(s)
|
||||||
if !ok {
|
if !ok {
|
||||||
@@ -186,9 +187,8 @@ func U256(x *big.Int) *big.Int {
|
|||||||
func S256(x *big.Int) *big.Int {
|
func S256(x *big.Int) *big.Int {
|
||||||
if x.Cmp(tt255) < 0 {
|
if x.Cmp(tt255) < 0 {
|
||||||
return x
|
return x
|
||||||
} else {
|
|
||||||
return new(big.Int).Sub(x, tt256)
|
|
||||||
}
|
}
|
||||||
|
return new(big.Int).Sub(x, tt256)
|
||||||
}
|
}
|
||||||
|
|
||||||
// Exp implements exponentiation by squaring.
|
// Exp implements exponentiation by squaring.
|
||||||
|
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user