mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 11:26:44 +01:00
Compare commits
13 Commits
refactore/
...
feature/se
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
a5ef822c64 | ||
|
|
5501f30062 | ||
|
|
d4f6926311 | ||
|
|
bfba1ba2d2 | ||
|
|
eed635539f | ||
|
|
87610f6188 | ||
|
|
4b596072d2 | ||
|
|
95c4019cb2 | ||
|
|
c4d5e8a7ab | ||
|
|
c1375d9c5f | ||
|
|
26e2bbc262 | ||
|
|
bb4c464200 | ||
|
|
982899efed |
1
.gitignore
vendored
1
.gitignore
vendored
@@ -1 +0,0 @@
|
||||
bin/
|
||||
135
Makefile
135
Makefile
@@ -1,135 +0,0 @@
|
||||
#! /usr/bin/make -f
|
||||
|
||||
# Project variables.
|
||||
PACKAGE := github.com/hermeznetwork/hermez-node
|
||||
VERSION := $(shell git describe --tags --always)
|
||||
BUILD := $(shell git rev-parse --short HEAD)
|
||||
BUILD_DATE := $(shell date +%Y-%m-%dT%H:%M:%S%z)
|
||||
PROJECT_NAME := $(shell basename "$(PWD)")
|
||||
|
||||
# Go related variables.
|
||||
GO_FILES ?= $$(find . -name '*.go' | grep -v vendor)
|
||||
GOBASE := $(shell pwd)
|
||||
GOBIN := $(GOBASE)/bin
|
||||
GOPKG := $(.)
|
||||
GOENVVARS := GOBIN=$(GOBIN)
|
||||
GOCMD := $(GOBASE)/cli/node
|
||||
GOPROOF := $(GOBASE)/test/proofserver/cli
|
||||
GOBINARY := node
|
||||
|
||||
# Project configs.
|
||||
MODE ?= sync
|
||||
CONFIG ?= $(GOBASE)/cli/node/cfg.buidler.toml
|
||||
POSTGRES_PASS ?= yourpasswordhere
|
||||
|
||||
# Use linker flags to provide version/build settings.
|
||||
LDFLAGS=-ldflags "-X=main.Version=$(VERSION) -X=main.Build=$(BUILD) -X=main.Date=$(BUILD_DATE)"
|
||||
|
||||
# PID file will keep the process id of the server.
|
||||
PID_PROOF_MOCK := /tmp/.$(PROJECT_NAME).proof.pid
|
||||
|
||||
# Make is verbose in Linux. Make it silent.
|
||||
MAKEFLAGS += --silent
|
||||
|
||||
.PHONY: help
|
||||
help: Makefile
|
||||
@echo
|
||||
@echo " Choose a command run in "$(PROJECT_NAME)":"
|
||||
@echo
|
||||
@sed -n 's/^##//p' $< | column -t -s ':' | sed -e 's/^/ /'
|
||||
@echo
|
||||
|
||||
## test: Run the application check and all tests.
|
||||
test: govet gocilint test-unit
|
||||
|
||||
## test-unit: Run all unit tests.
|
||||
test-unit:
|
||||
@echo " > Running unit tests"
|
||||
$(GOENVVARS) go test -race -p 1 -failfast -timeout 300s -v ./...
|
||||
|
||||
## test-api-server: Run the API server using the Go tests.
|
||||
test-api-server:
|
||||
@echo " > Running unit tests"
|
||||
$(GOENVVARS) FAKE_SERVER=yes go test -timeout 0 ./api -p 1 -count 1 -v
|
||||
|
||||
## gofmt: Run `go fmt` for all go files.
|
||||
gofmt:
|
||||
@echo " > Format all go files"
|
||||
$(GOENVVARS) gofmt -w ${GO_FILES}
|
||||
|
||||
## govet: Run go vet.
|
||||
govet:
|
||||
@echo " > Running go vet"
|
||||
$(GOENVVARS) go vet ./...
|
||||
|
||||
## golint: Run default golint.
|
||||
golint:
|
||||
@echo " > Running golint"
|
||||
$(GOENVVARS) golint -set_exit_status ./...
|
||||
|
||||
## gocilint: Run Golang CI Lint.
|
||||
gocilint:
|
||||
@echo " > Running Golang CI Lint"
|
||||
$-golangci-lint run --timeout=5m -E whitespace -E gosec -E gci -E misspell -E gomnd -E gofmt -E goimports -E golint --exclude-use-default=false --max-same-issues 0
|
||||
|
||||
## exec: Run given command. e.g; make exec run="go test ./..."
|
||||
exec:
|
||||
GOBIN=$(GOBIN) $(run)
|
||||
|
||||
## clean: Clean build files. Runs `go clean` internally.
|
||||
clean:
|
||||
@-rm $(GOBIN)/ 2> /dev/null
|
||||
@echo " > Cleaning build cache"
|
||||
$(GOENVVARS) go clean
|
||||
|
||||
## build: Build the project.
|
||||
build: install
|
||||
@echo " > Building Hermez binary..."
|
||||
@bash -c "$(MAKE) migration-pack"
|
||||
$(GOENVVARS) go build $(LDFLAGS) -o $(GOBIN)/$(GOBINARY) $(GOCMD)
|
||||
@bash -c "$(MAKE) migration-clean"
|
||||
|
||||
## install: Install missing dependencies. Runs `go get` internally. e.g; make install get=github.com/foo/bar
|
||||
install:
|
||||
@echo " > Checking if there is any missing dependencies..."
|
||||
$(GOENVVARS) go get $(GOCMD)/... $(get)
|
||||
|
||||
## run: Run Hermez node.
|
||||
run:
|
||||
@bash -c "$(MAKE) clean build"
|
||||
@echo " > Running $(PROJECT_NAME)"
|
||||
@$(GOBIN)/$(GOBINARY) --mode $(MODE) --cfg $(CONFIG) run
|
||||
|
||||
## run-proof-mock: Run proof server mock API.
|
||||
run-proof-mock: stop-proof-mock
|
||||
@echo " > Running Proof Server Mock"
|
||||
$(GOENVVARS) go build -o $(GOBIN)/proof $(GOPROOF)
|
||||
@$(GOBIN)/proof 2>&1 & echo $$! > $(PID_PROOF_MOCK)
|
||||
@cat $(PID_PROOF_MOCK) | sed "/^/s/^/ \> Proof Server Mock PID: /"
|
||||
|
||||
## stop-proof-mock: Stop proof server mock API.
|
||||
stop-proof-mock:
|
||||
@-touch $(PID_PROOF_MOCK)
|
||||
@-kill -s INT `cat $(PID_PROOF_MOCK)` 2> /dev/null || true
|
||||
@-rm $(PID_PROOF_MOCK) $(GOBIN)/proof 2> /dev/null || true
|
||||
|
||||
## migration-pack: Pack the database migrations into the binary.
|
||||
migration-pack:
|
||||
@echo " > Packing the migrations..."
|
||||
@cd /tmp && go get -u github.com/gobuffalo/packr/v2/packr2 && cd -
|
||||
@cd $(GOBASE)/db && packr2 && cd -
|
||||
|
||||
## migration-clean: Clean the database migrations pack.
|
||||
migration-clean:
|
||||
@echo " > Cleaning the migrations..."
|
||||
@cd $(GOBASE)/db && packr2 clean && cd -
|
||||
|
||||
## run-database-container: Run the Postgres container
|
||||
run-database-container:
|
||||
@echo " > Running the postgreSQL DB..."
|
||||
@-docker run --rm --name hermez-db -p 5432:5432 -e POSTGRES_DB=hermez -e POSTGRES_USER=hermez -e POSTGRES_PASSWORD="$(POSTGRES_PASS)" -d postgres
|
||||
|
||||
## stop-database-container: Stop the Postgres container
|
||||
stop-database-container:
|
||||
@echo " > Stopping the postgreSQL DB..."
|
||||
@-docker stop hermez-db
|
||||
83
README.md
83
README.md
@@ -8,75 +8,42 @@ Go implementation of the Hermez node.
|
||||
|
||||
The `hermez-node` has been tested with go version 1.14
|
||||
|
||||
### Build
|
||||
|
||||
Build the binary and check the current version:
|
||||
|
||||
```shell
|
||||
$ make build
|
||||
$ bin/node version
|
||||
```
|
||||
|
||||
### Run
|
||||
|
||||
First you must edit the default/template config file into [cli/node/cfg.buidler.toml](cli/node/cfg.buidler.toml),
|
||||
there are more information about the config file into [cli/node/README.md](cli/node/README.md)
|
||||
|
||||
After setting the config, you can build and run the Hermez Node as a synchronizer:
|
||||
|
||||
```shell
|
||||
$ make run
|
||||
```
|
||||
|
||||
Or build and run as a coordinator, and also passing the config file from other location:
|
||||
|
||||
```shell
|
||||
$ MODE=sync CONFIG=cli/node/cfg.buidler.toml make run
|
||||
```
|
||||
|
||||
To check the useful make commands:
|
||||
|
||||
```shell
|
||||
$ make help
|
||||
```
|
||||
|
||||
### Unit testing
|
||||
|
||||
Running the unit tests requires a connection to a PostgreSQL database. You can
|
||||
run PostgreSQL with docker easily this way (where `yourpasswordhere` should
|
||||
start PostgreSQL with docker easily this way (where `yourpasswordhere` should
|
||||
be your password):
|
||||
|
||||
```shell
|
||||
$ POSTGRES_PASS="yourpasswordhere" make run-database-container
|
||||
```
|
||||
POSTGRES_PASS=yourpasswordhere; sudo docker run --rm --name hermez-db-test -p 5432:5432 -e POSTGRES_DB=hermez -e POSTGRES_USER=hermez -e POSTGRES_PASSWORD="$POSTGRES_PASS" -d postgres
|
||||
```
|
||||
|
||||
Afterward, run the tests with the password as env var:
|
||||
Afterwards, run the tests with the password as env var:
|
||||
|
||||
```shell
|
||||
$ POSTGRES_PASS="yourpasswordhere" make test
|
||||
```
|
||||
POSTGRES_PASS=yourpasswordhere go test -p 1 ./...
|
||||
```
|
||||
|
||||
NOTE: `-p 1` forces execution of package test in serial. Otherwise, they may be
|
||||
executed in parallel, and the test may find unexpected entries in the SQL database
|
||||
NOTE: `-p 1` forces execution of package test in serial. Otherwise they may be
|
||||
executed in paralel and the test may find unexpected entries in the SQL databse
|
||||
because it's shared among all tests.
|
||||
|
||||
There is an extra temporary option that allows you to run the API server using the
|
||||
Go tests. It will be removed once the API can be properly initialized with data
|
||||
from the synchronizer. To use this, run:
|
||||
There is an extra temporary option that allows you to run the API server using
|
||||
the Go tests. This will be removed once the API can be properly initialized,
|
||||
with data from the synchronizer and so on. To use this, run:
|
||||
|
||||
```shell
|
||||
$ POSTGRES_PASS="yourpasswordhere" make test-api-server
|
||||
```
|
||||
FAKE_SERVER=yes POSTGRES_PASS=yourpasswordhere go test -timeout 0 ./api -p 1 -count 1 -v`
|
||||
```
|
||||
|
||||
### Lint
|
||||
|
||||
All Pull Requests need to pass the configured linter.
|
||||
|
||||
To run the linter locally, first, install [golangci-lint](https://golangci-lint.run).
|
||||
Afterward, you can check the lints with this command:
|
||||
To run the linter locally, first install [golangci-lint](https://golangci-lint.run). Afterwards you can check the lints with this command:
|
||||
|
||||
```shell
|
||||
$ make gocilint
|
||||
```
|
||||
golangci-lint run --timeout=5m -E whitespace -E gosec -E gci -E misspell -E gomnd -E gofmt -E goimports -E golint --exclude-use-default=false --max-same-issues 0
|
||||
```
|
||||
|
||||
## Usage
|
||||
@@ -87,13 +54,13 @@ See [cli/node/README.md](cli/node/README.md)
|
||||
|
||||
### Proof Server
|
||||
|
||||
The node in mode coordinator requires a proof server (a server capable of
|
||||
calculating proofs from the zkInputs). There is a mock proof server CLI
|
||||
at `test/proofserver/cli` for testing purposes.
|
||||
The node in mode coordinator requires a proof server (a server that is capable
|
||||
of calculating proofs from the zkInputs). For testing purposes there is a mock
|
||||
proof server cli at `test/proofserver/cli`.
|
||||
|
||||
Usage of `test/proofserver/cli`:
|
||||
|
||||
```shell
|
||||
```
|
||||
USAGE:
|
||||
go run ./test/proofserver/cli OPTIONS
|
||||
|
||||
@@ -104,19 +71,11 @@ OPTIONS:
|
||||
proving time duration (default 2s)
|
||||
```
|
||||
|
||||
Also, the Makefile commands can be used to run and stop the proof server
|
||||
in the background:
|
||||
|
||||
```shell
|
||||
$ make run-proof-mock
|
||||
$ make stop-proof-mock
|
||||
```
|
||||
|
||||
### `/tmp` as tmpfs
|
||||
|
||||
For every processed batch, the node builds a temporary exit tree in a key-value
|
||||
DB stored in `/tmp`. It is highly recommended that `/tmp` is mounted as a RAM
|
||||
file system in production to avoid unnecessary reads a writes to disk. This
|
||||
file system in production to avoid unecessary reads an writes to disk. This
|
||||
can be done by mounting `/tmp` as tmpfs; for example, by having this line in
|
||||
`/etc/fstab`:
|
||||
```
|
||||
|
||||
@@ -44,7 +44,7 @@ func (a *API) getAccounts(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type accountResponse struct {
|
||||
Accounts []historydb.AccountAPI `json:"accounts"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -47,7 +47,7 @@ func (a *API) getAccountCreationAuth(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
c.JSON(http.StatusOK, auth)
|
||||
}
|
||||
|
||||
|
||||
12
api/api.go
12
api/api.go
@@ -34,20 +34,20 @@ func NewAPI(
|
||||
if explorerEndpoints && hdb == nil {
|
||||
return nil, tracerr.Wrap(errors.New("cannot serve Explorer endpoints without HistoryDB"))
|
||||
}
|
||||
consts, err := hdb.GetConstants()
|
||||
ni, err := hdb.GetNodeInfo()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
a := &API{
|
||||
h: hdb,
|
||||
cg: &configAPI{
|
||||
RollupConstants: *newRollupConstants(consts.Rollup),
|
||||
AuctionConstants: consts.Auction,
|
||||
WDelayerConstants: consts.WDelayer,
|
||||
RollupConstants: *newRollupConstants(ni.Constants.RollupConstants),
|
||||
AuctionConstants: ni.Constants.AuctionConstants,
|
||||
WDelayerConstants: ni.Constants.WDelayerConstants,
|
||||
},
|
||||
l2: l2db,
|
||||
chainID: consts.ChainID,
|
||||
hermezAddress: consts.HermezAddress,
|
||||
chainID: ni.Constants.ChainID,
|
||||
hermezAddress: ni.Constants.HermezAddress,
|
||||
}
|
||||
|
||||
// Add coordinator endpoints
|
||||
|
||||
@@ -24,7 +24,6 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/stateapiupdater"
|
||||
"github.com/hermeznetwork/hermez-node/test"
|
||||
"github.com/hermeznetwork/hermez-node/test/til"
|
||||
"github.com/hermeznetwork/hermez-node/test/txsets"
|
||||
@@ -187,7 +186,6 @@ type testCommon struct {
|
||||
var tc testCommon
|
||||
var config configAPI
|
||||
var api *API
|
||||
var stateAPIUpdater *stateapiupdater.Updater
|
||||
|
||||
// TestMain initializes the API server, and fill HistoryDB and StateDB with fake data,
|
||||
// emulating the task of the synchronizer in order to have data to be returned
|
||||
@@ -203,7 +201,7 @@ func TestMain(m *testing.M) {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
apiConnCon := db.NewAPIConnectionController(1, time.Second)
|
||||
apiConnCon := db.NewAPICnnectionController(1, time.Second)
|
||||
hdb := historydb.NewHistoryDB(database, database, apiConnCon)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
@@ -224,27 +222,15 @@ func TestMain(m *testing.M) {
|
||||
apiGin := gin.Default()
|
||||
// Reset DB
|
||||
test.WipeDB(hdb.DB())
|
||||
|
||||
constants := &historydb.Constants{
|
||||
SCConsts: common.SCConsts{
|
||||
Rollup: _config.RollupConstants,
|
||||
Auction: _config.AuctionConstants,
|
||||
WDelayer: _config.WDelayerConstants,
|
||||
},
|
||||
ChainID: chainID,
|
||||
HermezAddress: _config.HermezAddress,
|
||||
}
|
||||
if err := hdb.SetConstants(constants); err != nil {
|
||||
if err := hdb.SetInitialNodeInfo(10, 0.0, &historydb.Constants{
|
||||
RollupConstants: _config.RollupConstants,
|
||||
AuctionConstants: _config.AuctionConstants,
|
||||
WDelayerConstants: _config.WDelayerConstants,
|
||||
ChainID: chainID,
|
||||
HermezAddress: _config.HermezAddress,
|
||||
}); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
nodeConfig := &historydb.NodeConfig{
|
||||
MaxPoolTxs: 10,
|
||||
MinFeeUSD: 0,
|
||||
}
|
||||
if err := hdb.SetNodeConfig(nodeConfig); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
|
||||
api, err = NewAPI(
|
||||
true,
|
||||
true,
|
||||
@@ -269,7 +255,7 @@ func TestMain(m *testing.M) {
|
||||
}
|
||||
}()
|
||||
|
||||
// Generate blockchain data with til
|
||||
// Genratre blockchain data with til
|
||||
tcc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||
tilCfgExtra := til.ConfigExtra{
|
||||
BootCoordAddr: ethCommon.HexToAddress("0xE39fEc6224708f0772D2A74fd3f9055A90E0A9f2"),
|
||||
@@ -315,7 +301,7 @@ func TestMain(m *testing.M) {
|
||||
USD: ðUSD,
|
||||
USDUpdate: ðNow,
|
||||
})
|
||||
err = api.h.UpdateTokenValue(common.EmptyAddr, ethUSD)
|
||||
err = api.h.UpdateTokenValue(test.EthToken.Symbol, ethUSD)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
@@ -342,7 +328,7 @@ func TestMain(m *testing.M) {
|
||||
token.USD = &value
|
||||
token.USDUpdate = &now
|
||||
// Set value in DB
|
||||
err = api.h.UpdateTokenValue(token.EthAddr, value)
|
||||
err = api.h.UpdateTokenValue(token.Symbol, value)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
@@ -499,14 +485,12 @@ func TestMain(m *testing.M) {
|
||||
// Slot 7
|
||||
nextForgers[6].Coordinator = nonBootForger
|
||||
|
||||
buckets := make([]common.BucketParams, 5)
|
||||
var buckets [common.RollupConstNumBuckets]common.BucketParams
|
||||
for i := range buckets {
|
||||
buckets[i].CeilUSD = big.NewInt(int64(i) * 10)
|
||||
buckets[i].BlockStamp = big.NewInt(int64(i) * 100)
|
||||
buckets[i].Withdrawals = big.NewInt(int64(i) * 1000)
|
||||
buckets[i].RateBlocks = big.NewInt(int64(i) * 10000)
|
||||
buckets[i].RateWithdrawals = big.NewInt(int64(i) * 100000)
|
||||
buckets[i].MaxWithdrawals = big.NewInt(int64(i) * 1000000)
|
||||
buckets[i].Withdrawals = big.NewInt(int64(i) * 100)
|
||||
buckets[i].BlockWithdrawalRate = big.NewInt(int64(i) * 1000)
|
||||
buckets[i].MaxWithdrawals = big.NewInt(int64(i) * 10000)
|
||||
}
|
||||
|
||||
// Generate SC vars and add them to HistoryDB (if needed)
|
||||
@@ -523,12 +507,6 @@ func TestMain(m *testing.M) {
|
||||
WithdrawalDelay: uint64(3000),
|
||||
}
|
||||
|
||||
stateAPIUpdater = stateapiupdater.NewUpdater(hdb, nodeConfig, &common.SCVariables{
|
||||
Rollup: rollupVars,
|
||||
Auction: auctionVars,
|
||||
WDelayer: wdelayerVars,
|
||||
}, constants)
|
||||
|
||||
// Generate test data, as expected to be received/sended from/to the API
|
||||
testCoords := genTestCoordinators(commonCoords)
|
||||
testBids := genTestBids(commonBlocks, testCoords, bids)
|
||||
@@ -613,7 +591,7 @@ func TestTimeout(t *testing.T) {
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
databaseTO, err := db.ConnectSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
apiConnConTO := db.NewAPIConnectionController(1, 100*time.Millisecond)
|
||||
apiConnConTO := db.NewAPICnnectionController(1, 100*time.Millisecond)
|
||||
hdbTO := historydb.NewHistoryDB(databaseTO, databaseTO, apiConnConTO)
|
||||
require.NoError(t, err)
|
||||
// L2DB
|
||||
|
||||
@@ -52,7 +52,7 @@ func (a *API) getBatches(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type batchesResponse struct {
|
||||
Batches []historydb.BatchAPI `json:"batches"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -7,12 +7,10 @@ import (
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/mitchellh/copystructure"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
type testBatch struct {
|
||||
@@ -22,7 +20,7 @@ type testBatch struct {
|
||||
EthBlockHash ethCommon.Hash `json:"ethereumBlockHash"`
|
||||
Timestamp time.Time `json:"timestamp"`
|
||||
ForgerAddr ethCommon.Address `json:"forgerAddr"`
|
||||
CollectedFees apitypes.CollectedFeesAPI `json:"collectedFees"`
|
||||
CollectedFees map[common.TokenID]string `json:"collectedFees"`
|
||||
TotalFeesUSD *float64 `json:"historicTotalCollectedFeesUSD"`
|
||||
StateRoot string `json:"stateRoot"`
|
||||
NumAccounts int `json:"numAccounts"`
|
||||
@@ -75,9 +73,9 @@ func genTestBatches(
|
||||
if !found {
|
||||
panic("block not found")
|
||||
}
|
||||
collectedFees := apitypes.CollectedFeesAPI(make(map[common.TokenID]apitypes.BigIntStr))
|
||||
collectedFees := make(map[common.TokenID]string)
|
||||
for k, v := range cBatches[i].CollectedFees {
|
||||
collectedFees[k] = *apitypes.NewBigIntStr(v)
|
||||
collectedFees[k] = v.String()
|
||||
}
|
||||
forgedTxs := 0
|
||||
for _, tx := range txs {
|
||||
@@ -134,7 +132,7 @@ func TestGetBatches(t *testing.T) {
|
||||
limit := 3
|
||||
path := fmt.Sprintf("%s?limit=%d", endpoint, limit)
|
||||
err := doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
assertBatches(t, tc.batches, fetchedBatches)
|
||||
|
||||
// minBatchNum
|
||||
@@ -143,7 +141,7 @@ func TestGetBatches(t *testing.T) {
|
||||
minBatchNum := tc.batches[len(tc.batches)/2].BatchNum
|
||||
path = fmt.Sprintf("%s?minBatchNum=%d&limit=%d", endpoint, minBatchNum, limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
minBatchNumBatches := []testBatch{}
|
||||
for i := 0; i < len(tc.batches); i++ {
|
||||
if tc.batches[i].BatchNum > minBatchNum {
|
||||
@@ -158,7 +156,7 @@ func TestGetBatches(t *testing.T) {
|
||||
maxBatchNum := tc.batches[len(tc.batches)/2].BatchNum
|
||||
path = fmt.Sprintf("%s?maxBatchNum=%d&limit=%d", endpoint, maxBatchNum, limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
maxBatchNumBatches := []testBatch{}
|
||||
for i := 0; i < len(tc.batches); i++ {
|
||||
if tc.batches[i].BatchNum < maxBatchNum {
|
||||
@@ -173,7 +171,7 @@ func TestGetBatches(t *testing.T) {
|
||||
slotNum := tc.batches[len(tc.batches)/2].SlotNum
|
||||
path = fmt.Sprintf("%s?slotNum=%d&limit=%d", endpoint, slotNum, limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
slotNumBatches := []testBatch{}
|
||||
for i := 0; i < len(tc.batches); i++ {
|
||||
if tc.batches[i].SlotNum == slotNum {
|
||||
@@ -188,7 +186,7 @@ func TestGetBatches(t *testing.T) {
|
||||
forgerAddr := tc.batches[len(tc.batches)/2].ForgerAddr
|
||||
path = fmt.Sprintf("%s?forgerAddr=%s&limit=%d", endpoint, forgerAddr.String(), limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
forgerAddrBatches := []testBatch{}
|
||||
for i := 0; i < len(tc.batches); i++ {
|
||||
if tc.batches[i].ForgerAddr == forgerAddr {
|
||||
@@ -202,7 +200,7 @@ func TestGetBatches(t *testing.T) {
|
||||
limit = 6
|
||||
path = fmt.Sprintf("%s?limit=%d", endpoint, limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderDesc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
flippedBatches := []testBatch{}
|
||||
for i := len(tc.batches) - 1; i >= 0; i-- {
|
||||
flippedBatches = append(flippedBatches, tc.batches[i])
|
||||
@@ -216,7 +214,7 @@ func TestGetBatches(t *testing.T) {
|
||||
minBatchNum = tc.batches[len(tc.batches)/4].BatchNum
|
||||
path = fmt.Sprintf("%s?minBatchNum=%d&maxBatchNum=%d&limit=%d", endpoint, minBatchNum, maxBatchNum, limit)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
minMaxBatchNumBatches := []testBatch{}
|
||||
for i := 0; i < len(tc.batches); i++ {
|
||||
if tc.batches[i].BatchNum < maxBatchNum && tc.batches[i].BatchNum > minBatchNum {
|
||||
@@ -229,25 +227,25 @@ func TestGetBatches(t *testing.T) {
|
||||
fetchedBatches = []testBatch{}
|
||||
path = fmt.Sprintf("%s?slotNum=%d&minBatchNum=%d", endpoint, 1, 25)
|
||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testBatchesResponse{}, appendIter)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
assertBatches(t, []testBatch{}, fetchedBatches)
|
||||
|
||||
// 400
|
||||
// Invalid minBatchNum
|
||||
path = fmt.Sprintf("%s?minBatchNum=%d", endpoint, -2)
|
||||
err = doBadReq("GET", path, nil, 400)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
// Invalid forgerAddr
|
||||
path = fmt.Sprintf("%s?forgerAddr=%s", endpoint, "0xG0000001")
|
||||
err = doBadReq("GET", path, nil, 400)
|
||||
require.NoError(t, err)
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
func TestGetBatch(t *testing.T) {
|
||||
endpoint := apiURL + "batches/"
|
||||
for _, batch := range tc.batches {
|
||||
fetchedBatch := testBatch{}
|
||||
require.NoError(
|
||||
assert.NoError(
|
||||
t, doGoodReq(
|
||||
"GET",
|
||||
endpoint+strconv.Itoa(int(batch.BatchNum)),
|
||||
@@ -257,16 +255,16 @@ func TestGetBatch(t *testing.T) {
|
||||
assertBatch(t, batch, fetchedBatch)
|
||||
}
|
||||
// 400
|
||||
require.NoError(t, doBadReq("GET", endpoint+"foo", nil, 400))
|
||||
assert.NoError(t, doBadReq("GET", endpoint+"foo", nil, 400))
|
||||
// 404
|
||||
require.NoError(t, doBadReq("GET", endpoint+"99999", nil, 404))
|
||||
assert.NoError(t, doBadReq("GET", endpoint+"99999", nil, 404))
|
||||
}
|
||||
|
||||
func TestGetFullBatch(t *testing.T) {
|
||||
endpoint := apiURL + "full-batches/"
|
||||
for _, fullBatch := range tc.fullBatches {
|
||||
fetchedFullBatch := testFullBatch{}
|
||||
require.NoError(
|
||||
assert.NoError(
|
||||
t, doGoodReq(
|
||||
"GET",
|
||||
endpoint+strconv.Itoa(int(fullBatch.Batch.BatchNum)),
|
||||
@@ -277,9 +275,9 @@ func TestGetFullBatch(t *testing.T) {
|
||||
assertTxs(t, fullBatch.Txs, fetchedFullBatch.Txs)
|
||||
}
|
||||
// 400
|
||||
require.NoError(t, doBadReq("GET", endpoint+"foo", nil, 400))
|
||||
assert.NoError(t, doBadReq("GET", endpoint+"foo", nil, 400))
|
||||
// 404
|
||||
require.NoError(t, doBadReq("GET", endpoint+"99999", nil, 404))
|
||||
assert.NoError(t, doBadReq("GET", endpoint+"99999", nil, 404))
|
||||
}
|
||||
|
||||
func assertBatches(t *testing.T, expected, actual []testBatch) {
|
||||
|
||||
@@ -34,7 +34,7 @@ func (a *API) getBids(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type bidsResponse struct {
|
||||
Bids []historydb.BidAPI `json:"bids"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -41,6 +41,7 @@ func newRollupConstants(publicConstants common.RollupConstants) *rollupConstants
|
||||
MaxL1UserTx: common.RollupConstMaxL1UserTx,
|
||||
MaxL1Tx: common.RollupConstMaxL1Tx,
|
||||
InputSHAConstantBytes: common.RollupConstInputSHAConstantBytes,
|
||||
NumBuckets: common.RollupConstNumBuckets,
|
||||
MaxWithdrawalDelay: common.RollupConstMaxWithdrawalDelay,
|
||||
ExchangeMultiplier: common.RollupConstExchangeMultiplier,
|
||||
}
|
||||
|
||||
@@ -32,7 +32,7 @@ func (a *API) getCoordinators(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type coordinatorsResponse struct {
|
||||
Coordinators []historydb.CoordinatorAPI `json:"coordinators"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -43,7 +43,7 @@ func (a *API) getExits(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type exitsResponse struct {
|
||||
Exits []historydb.ExitAPI `json:"exits"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
@@ -72,6 +72,6 @@ func (a *API) getExit(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
c.JSON(http.StatusOK, exit)
|
||||
}
|
||||
|
||||
@@ -14,7 +14,7 @@ import (
|
||||
)
|
||||
|
||||
const (
|
||||
// maxLimit is the max permitted items to be returned in paginated responses
|
||||
// maxLimit is the max permited items to be returned in paginated responses
|
||||
maxLimit uint = 2049
|
||||
|
||||
// dfltOrder indicates how paginated endpoints are ordered if not specified
|
||||
@@ -40,8 +40,8 @@ const (
|
||||
)
|
||||
|
||||
var (
|
||||
// ErrNilBidderAddr is used when a nil bidderAddr is received in the getCoordinator method
|
||||
ErrNilBidderAddr = errors.New("biderAddr can not be nil")
|
||||
// ErrNillBidderAddr is used when a nil bidderAddr is received in the getCoordinator method
|
||||
ErrNillBidderAddr = errors.New("biderAddr can not be nil")
|
||||
)
|
||||
|
||||
func retSQLErr(err error, c *gin.Context) {
|
||||
|
||||
@@ -50,19 +50,19 @@ func parsePagination(c querier) (fromItem *uint, order string, limit *uint, err
|
||||
return fromItem, order, limit, nil
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwritten
|
||||
// nolint reason: res may be not overwriten
|
||||
func parseQueryUint(name string, dflt *uint, min, max uint, c querier) (*uint, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
return stringToUint(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwritten
|
||||
// nolint reason: res may be not overwriten
|
||||
func parseQueryInt64(name string, dflt *int64, min, max int64, c querier) (*int64, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
return stringToInt64(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwritten
|
||||
// nolint reason: res may be not overwriten
|
||||
func parseQueryBool(name string, dflt *bool, c querier) (*bool, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
if str == "" {
|
||||
@@ -295,13 +295,13 @@ func parseParamIdx(c paramer) (*common.Idx, error) {
|
||||
return stringToIdx(idxStr, name)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwritten
|
||||
// nolint reason: res may be not overwriten
|
||||
func parseParamUint(name string, dflt *uint, min, max uint, c paramer) (*uint, error) { //nolint:SA4009
|
||||
str := c.Param(name)
|
||||
return stringToUint(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwritten
|
||||
// nolint reason: res may be not overwriten
|
||||
func parseParamInt64(name string, dflt *int64, min, max int64, c paramer) (*int64, error) { //nolint:SA4009
|
||||
str := c.Param(name)
|
||||
return stringToInt64(str, name, dflt, min, max)
|
||||
|
||||
@@ -11,7 +11,7 @@ import (
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
// SlotAPI is a representation of a slot information
|
||||
// SlotAPI is a repesentation of a slot information
|
||||
type SlotAPI struct {
|
||||
ItemID uint64 `json:"itemId"`
|
||||
SlotNum int64 `json:"slotNum"`
|
||||
@@ -316,7 +316,7 @@ func (a *API) getSlots(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type slotsResponse struct {
|
||||
Slots []SlotAPI `json:"slots"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -99,14 +99,14 @@ func TestGetSlot(t *testing.T) {
|
||||
nil, &fetchedSlot,
|
||||
),
|
||||
)
|
||||
// ni, err := api.h.GetNodeInfoAPI()
|
||||
// assert.NoError(t, err)
|
||||
emptySlot := api.getEmptyTestSlot(slotNum, 0, tc.auctionVars)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
assert.NoError(t, err)
|
||||
emptySlot := api.getEmptyTestSlot(slotNum, ni.APIState.Network.LastSyncBlock, tc.auctionVars)
|
||||
assertSlot(t, emptySlot, fetchedSlot)
|
||||
|
||||
// Invalid slotNum
|
||||
path := endpoint + strconv.Itoa(-2)
|
||||
err := doBadReq("GET", path, nil, 400)
|
||||
err = doBadReq("GET", path, nil, 400)
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
@@ -129,10 +129,10 @@ func TestGetSlots(t *testing.T) {
|
||||
err := doGoodReqPaginated(path, historydb.OrderAsc, &testSlotsResponse{}, appendIter)
|
||||
assert.NoError(t, err)
|
||||
allSlots := tc.slots
|
||||
// ni, err := api.h.GetNodeInfoAPI()
|
||||
// assert.NoError(t, err)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
assert.NoError(t, err)
|
||||
for i := tc.slots[len(tc.slots)-1].SlotNum; i < maxSlotNum; i++ {
|
||||
emptySlot := api.getEmptyTestSlot(i+1, 0, tc.auctionVars)
|
||||
emptySlot := api.getEmptyTestSlot(i+1, ni.APIState.Network.LastSyncBlock, tc.auctionVars)
|
||||
allSlots = append(allSlots, emptySlot)
|
||||
}
|
||||
assertSlots(t, allSlots, fetchedSlots)
|
||||
|
||||
107
api/state.go
107
api/state.go
@@ -1,9 +1,13 @@
|
||||
package api
|
||||
|
||||
import (
|
||||
"database/sql"
|
||||
"net/http"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
func (a *API) getState(c *gin.Context) {
|
||||
@@ -14,3 +18,106 @@ func (a *API) getState(c *gin.Context) {
|
||||
}
|
||||
c.JSON(http.StatusOK, stateAPI)
|
||||
}
|
||||
|
||||
type APIStateUpdater struct {
|
||||
hdb *historydb.HistoryDB
|
||||
state historydb.StateAPI
|
||||
config historydb.NodeConfig
|
||||
vars common.SCVariablesPtr
|
||||
consts historydb.Constants
|
||||
}
|
||||
|
||||
func NewAPIStateUpdater(hdb *historydb.HistoryDB, config *historydb.NodeConfig, vars *common.SCVariables,
|
||||
consts *historydb.Constants) *APIStateUpdater {
|
||||
u := APIStateUpdater{
|
||||
hdb: hdb,
|
||||
config: *config,
|
||||
consts: *consts,
|
||||
}
|
||||
u.SetSCVars(&common.SCVariablesPtr{&vars.Rollup, &vars.Auction, &vars.WDelayer})
|
||||
return &u
|
||||
}
|
||||
|
||||
func (u *APIStateUpdater) Store() error {
|
||||
return tracerr.Wrap(u.hdb.SetAPIState(&u.state))
|
||||
}
|
||||
|
||||
func (u *APIStateUpdater) SetSCVars(vars *common.SCVariablesPtr) {
|
||||
if vars.Rollup != nil {
|
||||
u.vars.Rollup = vars.Rollup
|
||||
rollupVars := historydb.NewRollupVariablesAPI(u.vars.Rollup)
|
||||
u.state.Rollup = *rollupVars
|
||||
}
|
||||
if vars.Auction != nil {
|
||||
u.vars.Auction = vars.Auction
|
||||
auctionVars := historydb.NewAuctionVariablesAPI(u.vars.Auction)
|
||||
u.state.Auction = *auctionVars
|
||||
}
|
||||
if vars.WDelayer != nil {
|
||||
u.vars.WDelayer = vars.WDelayer
|
||||
u.state.WithdrawalDelayer = *u.vars.WDelayer
|
||||
}
|
||||
}
|
||||
|
||||
func (u *APIStateUpdater) UpdateMetrics() error {
|
||||
if u.state.Network.LastBatch == nil {
|
||||
return nil
|
||||
}
|
||||
lastBatchNum := u.state.Network.LastBatch.BatchNum
|
||||
metrics, err := u.hdb.GetMetricsInternalAPI(lastBatchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
u.state.Metrics = *metrics
|
||||
return nil
|
||||
}
|
||||
|
||||
func (u *APIStateUpdater) UpdateNetworkInfoBlock(lastEthBlock, lastSyncBlock common.Block) {
|
||||
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
||||
u.state.Network.LastEthBlock = lastEthBlock.Num
|
||||
}
|
||||
|
||||
func (u *APIStateUpdater) UpdateNetworkInfo(
|
||||
lastEthBlock, lastSyncBlock common.Block,
|
||||
lastBatchNum common.BatchNum, currentSlot int64,
|
||||
) error {
|
||||
// Get last batch in API format
|
||||
lastBatch, err := u.hdb.GetBatchInternalAPI(lastBatchNum)
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
lastBatch = nil
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Get next forgers
|
||||
lastClosedSlot := currentSlot + int64(u.state.Auction.ClosedAuctionSlots)
|
||||
nextForgers, err := u.hdb.GetNextForgersInternalAPI(u.vars.Auction, &u.consts.Auction,
|
||||
lastSyncBlock, currentSlot, lastClosedSlot)
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
nextForgers = nil
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
bucketUpdates, err := u.hdb.GetBucketUpdatesInternalAPI()
|
||||
if err == sql.ErrNoRows {
|
||||
bucketUpdates = nil
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Update NodeInfo struct
|
||||
for i, bucketParams := range u.state.Rollup.Buckets {
|
||||
for _, bucketUpdate := range bucketUpdates {
|
||||
if bucketUpdate.NumBucket == i {
|
||||
bucketParams.Withdrawals = bucketUpdate.Withdrawals
|
||||
u.state.Rollup.Buckets[i] = bucketParams
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
||||
u.state.Network.LastEthBlock = lastEthBlock.Num
|
||||
u.state.Network.LastBatch = lastBatch
|
||||
u.state.Network.CurrentSlot = currentSlot
|
||||
u.state.Network.NextForgers = nextForgers
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -29,11 +29,10 @@ type testNetwork struct {
|
||||
}
|
||||
|
||||
func TestSetRollupVariables(t *testing.T) {
|
||||
stateAPIUpdater.SetSCVars(&common.SCVariablesPtr{Rollup: &tc.rollupVars})
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
api.h.SetRollupVariables(&tc.rollupVars)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assertEqualRollupVariables(t, tc.rollupVars, ni.StateAPI.Rollup, true)
|
||||
assert.NoError(t, err)
|
||||
assertEqualRollupVariables(t, tc.rollupVars, ni.APIState.Rollup, true)
|
||||
}
|
||||
|
||||
func assertEqualRollupVariables(t *testing.T, rollupVariables common.RollupVariables, apiVariables historydb.RollupVariablesAPI, checkBuckets bool) {
|
||||
@@ -43,9 +42,7 @@ func assertEqualRollupVariables(t *testing.T, rollupVariables common.RollupVaria
|
||||
assert.Equal(t, rollupVariables.SafeMode, apiVariables.SafeMode)
|
||||
if checkBuckets {
|
||||
for i, bucket := range rollupVariables.Buckets {
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.BlockStamp), apiVariables.Buckets[i].BlockStamp)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.RateBlocks), apiVariables.Buckets[i].RateBlocks)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.RateWithdrawals), apiVariables.Buckets[i].RateWithdrawals)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.BlockWithdrawalRate), apiVariables.Buckets[i].BlockWithdrawalRate)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.CeilUSD), apiVariables.Buckets[i].CeilUSD)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.MaxWithdrawals), apiVariables.Buckets[i].MaxWithdrawals)
|
||||
assert.Equal(t, apitypes.NewBigIntStr(bucket.Withdrawals), apiVariables.Buckets[i].Withdrawals)
|
||||
@@ -54,19 +51,17 @@ func assertEqualRollupVariables(t *testing.T, rollupVariables common.RollupVaria
|
||||
}
|
||||
|
||||
func TestSetWDelayerVariables(t *testing.T) {
|
||||
stateAPIUpdater.SetSCVars(&common.SCVariablesPtr{WDelayer: &tc.wdelayerVars})
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
api.h.SetWDelayerVariables(&tc.wdelayerVars)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tc.wdelayerVars, ni.StateAPI.WithdrawalDelayer)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, tc.wdelayerVars, ni.APIState.WithdrawalDelayer)
|
||||
}
|
||||
|
||||
func TestSetAuctionVariables(t *testing.T) {
|
||||
stateAPIUpdater.SetSCVars(&common.SCVariablesPtr{Auction: &tc.auctionVars})
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
api.h.SetAuctionVariables(&tc.auctionVars)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assertEqualAuctionVariables(t, tc.auctionVars, ni.StateAPI.Auction)
|
||||
assert.NoError(t, err)
|
||||
assertEqualAuctionVariables(t, tc.auctionVars, ni.APIState.Auction)
|
||||
}
|
||||
|
||||
func assertEqualAuctionVariables(t *testing.T, auctionVariables common.AuctionVariables, apiVariables historydb.AuctionVariablesAPI) {
|
||||
@@ -118,17 +113,16 @@ func TestUpdateNetworkInfo(t *testing.T) {
|
||||
err := api.h.AddBucketUpdatesTest(api.h.DB(), bucketUpdates)
|
||||
require.NoError(t, err)
|
||||
|
||||
err = stateAPIUpdater.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
err = api.h.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
assert.NoError(t, err)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, lastBlock.Num, ni.StateAPI.Network.LastSyncBlock)
|
||||
assert.Equal(t, lastBatchNum, ni.StateAPI.Network.LastBatch.BatchNum)
|
||||
assert.Equal(t, currentSlotNum, ni.StateAPI.Network.CurrentSlot)
|
||||
assert.Equal(t, int(ni.StateAPI.Auction.ClosedAuctionSlots)+1, len(ni.StateAPI.Network.NextForgers))
|
||||
assert.Equal(t, ni.StateAPI.Rollup.Buckets[0].Withdrawals, apitypes.NewBigIntStr(big.NewInt(123)))
|
||||
assert.Equal(t, ni.StateAPI.Rollup.Buckets[2].Withdrawals, apitypes.NewBigIntStr(big.NewInt(43)))
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, lastBlock.Num, ni.APIState.Network.LastSyncBlock)
|
||||
assert.Equal(t, lastBatchNum, ni.APIState.Network.LastBatch.BatchNum)
|
||||
assert.Equal(t, currentSlotNum, ni.APIState.Network.CurrentSlot)
|
||||
assert.Equal(t, int(ni.APIState.Auction.ClosedAuctionSlots)+1, len(ni.APIState.Network.NextForgers))
|
||||
assert.Equal(t, ni.APIState.Rollup.Buckets[0].Withdrawals, apitypes.NewBigIntStr(big.NewInt(123)))
|
||||
assert.Equal(t, ni.APIState.Rollup.Buckets[2].Withdrawals, apitypes.NewBigIntStr(big.NewInt(43)))
|
||||
}
|
||||
|
||||
func TestUpdateMetrics(t *testing.T) {
|
||||
@@ -136,62 +130,55 @@ func TestUpdateMetrics(t *testing.T) {
|
||||
lastBlock := tc.blocks[3]
|
||||
lastBatchNum := common.BatchNum(12)
|
||||
currentSlotNum := int64(1)
|
||||
err := stateAPIUpdater.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
require.NoError(t, err)
|
||||
err := api.h.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
assert.NoError(t, err)
|
||||
|
||||
err = stateAPIUpdater.UpdateMetrics()
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
err = api.h.UpdateMetrics()
|
||||
assert.NoError(t, err)
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assert.Greater(t, ni.StateAPI.Metrics.TransactionsPerBatch, float64(0))
|
||||
assert.Greater(t, ni.StateAPI.Metrics.BatchFrequency, float64(0))
|
||||
assert.Greater(t, ni.StateAPI.Metrics.TransactionsPerSecond, float64(0))
|
||||
assert.Greater(t, ni.StateAPI.Metrics.TotalAccounts, int64(0))
|
||||
assert.Greater(t, ni.StateAPI.Metrics.TotalBJJs, int64(0))
|
||||
assert.Greater(t, ni.StateAPI.Metrics.AvgTransactionFee, float64(0))
|
||||
assert.NoError(t, err)
|
||||
assert.Greater(t, ni.APIState.Metrics.TransactionsPerBatch, float64(0))
|
||||
assert.Greater(t, ni.APIState.Metrics.BatchFrequency, float64(0))
|
||||
assert.Greater(t, ni.APIState.Metrics.TransactionsPerSecond, float64(0))
|
||||
assert.Greater(t, ni.APIState.Metrics.TotalAccounts, int64(0))
|
||||
assert.Greater(t, ni.APIState.Metrics.TotalBJJs, int64(0))
|
||||
assert.Greater(t, ni.APIState.Metrics.AvgTransactionFee, float64(0))
|
||||
}
|
||||
|
||||
func TestUpdateRecommendedFee(t *testing.T) {
|
||||
err := stateAPIUpdater.UpdateRecommendedFee()
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
err := api.h.UpdateRecommendedFee()
|
||||
assert.NoError(t, err)
|
||||
var minFeeUSD float64
|
||||
if api.l2 != nil {
|
||||
minFeeUSD = api.l2.MinFeeUSD()
|
||||
}
|
||||
ni, err := api.h.GetNodeInfoAPI()
|
||||
require.NoError(t, err)
|
||||
assert.Greater(t, ni.StateAPI.RecommendedFee.ExistingAccount, minFeeUSD)
|
||||
assert.Equal(t, ni.StateAPI.RecommendedFee.CreatesAccount,
|
||||
ni.StateAPI.RecommendedFee.ExistingAccount*
|
||||
historydb.CreateAccountExtraFeePercentage)
|
||||
assert.Equal(t, ni.StateAPI.RecommendedFee.CreatesAccountInternal,
|
||||
ni.StateAPI.RecommendedFee.ExistingAccount*
|
||||
historydb.CreateAccountInternalExtraFeePercentage)
|
||||
assert.NoError(t, err)
|
||||
assert.Greater(t, ni.APIState.RecommendedFee.ExistingAccount, minFeeUSD)
|
||||
// assert.Equal(t, ni.StateAPI.RecommendedFee.CreatesAccount,
|
||||
// ni.StateAPI.RecommendedFee.ExistingAccount*createAccountExtraFeePercentage)
|
||||
// assert.Equal(t, ni.StateAPI.RecommendedFee.CreatesAccountAndRegister,
|
||||
// ni.StateAPI.RecommendedFee.ExistingAccount*createAccountInternalExtraFeePercentage)
|
||||
}
|
||||
|
||||
func TestGetState(t *testing.T) {
|
||||
lastBlock := tc.blocks[3]
|
||||
lastBatchNum := common.BatchNum(12)
|
||||
currentSlotNum := int64(1)
|
||||
stateAPIUpdater.SetSCVars(&common.SCVariablesPtr{
|
||||
Rollup: &tc.rollupVars,
|
||||
Auction: &tc.auctionVars,
|
||||
WDelayer: &tc.wdelayerVars,
|
||||
})
|
||||
err := stateAPIUpdater.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
require.NoError(t, err)
|
||||
err = stateAPIUpdater.UpdateMetrics()
|
||||
require.NoError(t, err)
|
||||
err = stateAPIUpdater.UpdateRecommendedFee()
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, stateAPIUpdater.Store())
|
||||
api.h.SetRollupVariables(&tc.rollupVars)
|
||||
api.h.SetWDelayerVariables(&tc.wdelayerVars)
|
||||
api.h.SetAuctionVariables(&tc.auctionVars)
|
||||
err := api.h.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
assert.NoError(t, err)
|
||||
err = api.h.UpdateMetrics()
|
||||
assert.NoError(t, err)
|
||||
err = api.h.UpdateRecommendedFee()
|
||||
assert.NoError(t, err)
|
||||
|
||||
endpoint := apiURL + "state"
|
||||
var status testStatus
|
||||
|
||||
require.NoError(t, doGoodReq("GET", endpoint, nil, &status))
|
||||
assert.NoError(t, doGoodReq("GET", endpoint, nil, &status))
|
||||
|
||||
// SC vars
|
||||
// UpdateNetworkInfo will overwrite buckets withdrawal values
|
||||
@@ -218,12 +205,10 @@ func TestGetState(t *testing.T) {
|
||||
// Recommended fee
|
||||
// TODO: perform real asserts (not just greater than 0)
|
||||
assert.Greater(t, status.RecommendedFee.ExistingAccount, float64(0))
|
||||
assert.Equal(t, status.RecommendedFee.CreatesAccount,
|
||||
status.RecommendedFee.ExistingAccount*
|
||||
historydb.CreateAccountExtraFeePercentage)
|
||||
assert.Equal(t, status.RecommendedFee.CreatesAccountInternal,
|
||||
status.RecommendedFee.ExistingAccount*
|
||||
historydb.CreateAccountInternalExtraFeePercentage)
|
||||
// assert.Equal(t, status.RecommendedFee.CreatesAccount,
|
||||
// status.RecommendedFee.ExistingAccount*createAccountExtraFeePercentage)
|
||||
// assert.Equal(t, status.RecommendedFee.CreatesAccountAndRegister,
|
||||
// status.RecommendedFee.ExistingAccount*createAccountInternalExtraFeePercentage)
|
||||
}
|
||||
|
||||
func assertNextForgers(t *testing.T, expected, actual []historydb.NextForgerAPI) {
|
||||
|
||||
@@ -2729,24 +2729,16 @@ components:
|
||||
properties:
|
||||
ceilUSD:
|
||||
type: string
|
||||
description: Max USD value that bucket holds
|
||||
description: Max USD value
|
||||
example: "1000"
|
||||
blockStamp:
|
||||
type: string
|
||||
description: Block number of the last bucket update
|
||||
example: "1"
|
||||
withdrawals:
|
||||
type: string
|
||||
description: Available withdrawals of the bucket
|
||||
example: "4"
|
||||
rateBlocks:
|
||||
blockWithdrawalRate:
|
||||
type: string
|
||||
description: rateBlocks every `rateBlocks` blocks add `rateWithdrawals` withdrawal
|
||||
example: "2"
|
||||
rateWithdrawals:
|
||||
type: string
|
||||
description: add `rateWithdrawals` every `rateBlocks`
|
||||
example: "3"
|
||||
description: Every `blockWithdrawalRate` blocks add 1 withdrawal
|
||||
example: "8"
|
||||
maxWithdrawals:
|
||||
type: string
|
||||
description: Max withdrawals the bucket can hold
|
||||
@@ -2754,10 +2746,8 @@ components:
|
||||
additionalProperties: false
|
||||
required:
|
||||
- ceilUSD
|
||||
- blockStamp
|
||||
- withdrawals
|
||||
- rateBlocks
|
||||
- rateWithdrawals
|
||||
- blockWithdrawalRate
|
||||
- maxWithdrawals
|
||||
additionalProperties: false
|
||||
required:
|
||||
|
||||
@@ -53,7 +53,7 @@ func (a *API) getTokens(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type tokensResponse struct {
|
||||
Tokens []historydb.TokenWithUSD `json:"tokens"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -42,7 +42,7 @@ func (a *API) getHistoryTxs(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
type txsResponse struct {
|
||||
Txs []historydb.TxAPI `json:"transactions"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
@@ -66,6 +66,6 @@ func (a *API) getHistoryTx(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
c.JSON(http.StatusOK, tx)
|
||||
}
|
||||
|
||||
@@ -455,7 +455,7 @@ func TestGetHistoryTx(t *testing.T) {
|
||||
// 400, due invalid TxID
|
||||
err := doBadReq("GET", endpoint+"0x001", nil, 400)
|
||||
assert.NoError(t, err)
|
||||
// 404, due nonexistent TxID in DB
|
||||
// 404, due inexistent TxID in DB
|
||||
err = doBadReq("GET", endpoint+"0x00eb5e95e1ce5e9f6c4ed402d415e8d0bdd7664769cfd2064d28da04a2c76be432", nil, 404)
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
@@ -51,7 +51,7 @@ func (a *API) getPoolTx(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build successful response
|
||||
// Build succesfull response
|
||||
c.JSON(http.StatusOK, tx)
|
||||
}
|
||||
|
||||
@@ -179,7 +179,7 @@ func (a *API) verifyPoolL2TxWrite(txw *l2db.PoolL2TxWrite) error {
|
||||
// Get public key
|
||||
account, err := a.h.GetCommonAccountAPI(poolTx.FromIdx)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("Error getting from account: %w", err))
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Validate TokenID
|
||||
if poolTx.TokenID != account.TokenID {
|
||||
|
||||
@@ -2,15 +2,10 @@ package api
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"crypto/ecdsa"
|
||||
"encoding/binary"
|
||||
"encoding/hex"
|
||||
"encoding/json"
|
||||
"math/big"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
ethCrypto "github.com/ethereum/go-ethereum/crypto"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
@@ -240,7 +235,7 @@ func TestPoolTxs(t *testing.T) {
|
||||
// 400, due invalid TxID
|
||||
err = doBadReq("GET", endpoint+"0xG2241b6f2b1dd772dba391f4a1a3407c7c21f598d86e2585a14e616fb4a255f823", nil, 400)
|
||||
require.NoError(t, err)
|
||||
// 404, due nonexistent TxID in DB
|
||||
// 404, due inexistent TxID in DB
|
||||
err = doBadReq("GET", endpoint+"0x02241b6f2b1dd772dba391f4a1a3407c7c21f598d86e2585a14e616fb4a255f823", nil, 404)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
@@ -262,73 +257,3 @@ func assertPoolTx(t *testing.T, expected, actual testPoolTxReceive) {
|
||||
}
|
||||
assert.Equal(t, expected, actual)
|
||||
}
|
||||
|
||||
// TestAllTosNull test that the API doesn't accept txs with all the TOs set to null (to eth, to bjj, to idx)
|
||||
func TestAllTosNull(t *testing.T) {
|
||||
// Generate account:
|
||||
// Ethereum private key
|
||||
var key ecdsa.PrivateKey
|
||||
key.D = big.NewInt(int64(4444)) // only for testing
|
||||
key.PublicKey.X, key.PublicKey.Y = ethCrypto.S256().ScalarBaseMult(key.D.Bytes())
|
||||
key.Curve = ethCrypto.S256()
|
||||
addr := ethCrypto.PubkeyToAddress(key.PublicKey)
|
||||
// BJJ private key
|
||||
var sk babyjub.PrivateKey
|
||||
var iBytes [8]byte
|
||||
binary.LittleEndian.PutUint64(iBytes[:], 4444)
|
||||
copy(sk[:], iBytes[:]) // only for testing
|
||||
account := common.Account{
|
||||
Idx: 4444,
|
||||
TokenID: 0,
|
||||
BatchNum: 1,
|
||||
BJJ: sk.Public().Compress(),
|
||||
EthAddr: addr,
|
||||
Nonce: 0,
|
||||
Balance: big.NewInt(1000000),
|
||||
}
|
||||
// Add account to history DB (required to verify signature)
|
||||
err := api.h.AddAccounts([]common.Account{account})
|
||||
assert.NoError(t, err)
|
||||
// Genrate tx with all tos set to nil (to eth, to bjj, to idx)
|
||||
tx := common.PoolL2Tx{
|
||||
FromIdx: account.Idx,
|
||||
TokenID: account.TokenID,
|
||||
Amount: big.NewInt(1000),
|
||||
Fee: 200,
|
||||
Nonce: 0,
|
||||
}
|
||||
// Set idx and type manually, and check that the function doesn't allow it
|
||||
_, err = common.NewPoolL2Tx(&tx)
|
||||
assert.Error(t, err)
|
||||
tx.Type = common.TxTypeTransfer
|
||||
var txID common.TxID
|
||||
txIDRaw, err := hex.DecodeString("02e66e24f7f25272906647c8fd1d7fe8acf3cf3e9b38ffc9f94bbb5090dc275073")
|
||||
assert.NoError(t, err)
|
||||
copy(txID[:], txIDRaw)
|
||||
tx.TxID = txID
|
||||
// Sign tx
|
||||
toSign, err := tx.HashToSign(0)
|
||||
assert.NoError(t, err)
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
tx.Signature = sig.Compress()
|
||||
// Transform common.PoolL2Tx ==> testPoolTxSend
|
||||
txToSend := testPoolTxSend{
|
||||
TxID: tx.TxID,
|
||||
Type: tx.Type,
|
||||
TokenID: tx.TokenID,
|
||||
FromIdx: idxToHez(tx.FromIdx, "ETH"),
|
||||
Amount: tx.Amount.String(),
|
||||
Fee: tx.Fee,
|
||||
Nonce: tx.Nonce,
|
||||
Signature: tx.Signature,
|
||||
}
|
||||
// Send tx to the API
|
||||
jsonTxBytes, err := json.Marshal(txToSend)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader := bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", apiURL+"transactions-pool", jsonTxReader, 400)
|
||||
require.NoError(t, err)
|
||||
// Clean historyDB: the added account shouldn't be there for other tests
|
||||
_, err = api.h.DB().DB.Exec("delete from account where idx = 4444")
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
@@ -4,6 +4,7 @@ import (
|
||||
"database/sql/driver"
|
||||
"encoding/base64"
|
||||
"encoding/hex"
|
||||
"encoding/json"
|
||||
"errors"
|
||||
"fmt"
|
||||
"math/big"
|
||||
@@ -18,10 +19,7 @@ import (
|
||||
|
||||
// BigIntStr is used to scan/value *big.Int directly into strings from/to sql DBs.
|
||||
// It assumes that *big.Int are inserted/fetched to/from the DB using the BigIntMeddler meddler
|
||||
// defined at github.com/hermeznetwork/hermez-node/db. Since *big.Int is
|
||||
// stored as DECIMAL in SQL, there's no need to implement Scan()/Value()
|
||||
// because DECIMALS are encoded/decoded as strings by the sql driver, and
|
||||
// BigIntStr is already a string.
|
||||
// defined at github.com/hermeznetwork/hermez-node/db
|
||||
type BigIntStr string
|
||||
|
||||
// NewBigIntStr creates a *BigIntStr from a *big.Int.
|
||||
@@ -34,6 +32,34 @@ func NewBigIntStr(bigInt *big.Int) *BigIntStr {
|
||||
return &bigIntStr
|
||||
}
|
||||
|
||||
// Scan implements Scanner for database/sql
|
||||
func (b *BigIntStr) Scan(src interface{}) error {
|
||||
srcBytes, ok := src.([]byte)
|
||||
if !ok {
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan %T into apitypes.BigIntStr", src))
|
||||
}
|
||||
// bytes to *big.Int
|
||||
bigInt := new(big.Int).SetBytes(srcBytes)
|
||||
// *big.Int to BigIntStr
|
||||
bigIntStr := NewBigIntStr(bigInt)
|
||||
if bigIntStr == nil {
|
||||
return nil
|
||||
}
|
||||
*b = *bigIntStr
|
||||
return nil
|
||||
}
|
||||
|
||||
// Value implements valuer for database/sql
|
||||
func (b BigIntStr) Value() (driver.Value, error) {
|
||||
// string to *big.Int
|
||||
bigInt, ok := new(big.Int).SetString(string(b), 10)
|
||||
if !ok || bigInt == nil {
|
||||
return nil, tracerr.Wrap(errors.New("invalid representation of a *big.Int"))
|
||||
}
|
||||
// *big.Int to bytes
|
||||
return bigInt.Bytes(), nil
|
||||
}
|
||||
|
||||
// StrBigInt is used to unmarshal BigIntStr directly into an alias of big.Int
|
||||
type StrBigInt big.Int
|
||||
|
||||
@@ -47,19 +73,25 @@ func (s *StrBigInt) UnmarshalText(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// CollectedFeesAPI is send common.batch.CollectedFee through the API
|
||||
type CollectedFeesAPI map[common.TokenID]BigIntStr
|
||||
// CollectedFees is used to retrieve common.batch.CollectedFee from the DB
|
||||
type CollectedFees map[common.TokenID]BigIntStr
|
||||
|
||||
// NewCollectedFeesAPI creates a new CollectedFeesAPI from a *big.Int map
|
||||
func NewCollectedFeesAPI(m map[common.TokenID]*big.Int) CollectedFeesAPI {
|
||||
c := CollectedFeesAPI(make(map[common.TokenID]BigIntStr))
|
||||
for k, v := range m {
|
||||
c[k] = *NewBigIntStr(v)
|
||||
// UnmarshalJSON unmarshals a json representation of map[common.TokenID]*big.Int
|
||||
func (c *CollectedFees) UnmarshalJSON(text []byte) error {
|
||||
bigIntMap := make(map[common.TokenID]*big.Int)
|
||||
if err := json.Unmarshal(text, &bigIntMap); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return c
|
||||
*c = CollectedFees(make(map[common.TokenID]BigIntStr))
|
||||
for k, v := range bigIntMap {
|
||||
bStr := NewBigIntStr(v)
|
||||
(CollectedFees(*c)[k]) = *bStr
|
||||
}
|
||||
// *c = CollectedFees(bStrMap)
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezEthAddr is used to scan/value Ethereum Address directly into strings that follow the Ethereum address hez format (^hez:0x[a-fA-F0-9]{40}$) from/to sql DBs.
|
||||
// HezEthAddr is used to scan/value Ethereum Address directly into strings that follow the Ethereum address hez fotmat (^hez:0x[a-fA-F0-9]{40}$) from/to sql DBs.
|
||||
// It assumes that Ethereum Address are inserted/fetched to/from the DB using the default Scan/Value interface
|
||||
type HezEthAddr string
|
||||
|
||||
@@ -111,7 +143,7 @@ func (s *StrHezEthAddr) UnmarshalText(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezBJJ is used to scan/value *babyjub.PublicKeyComp directly into strings that follow the BJJ public key hez format (^hez:[A-Za-z0-9_-]{44}$) from/to sql DBs.
|
||||
// HezBJJ is used to scan/value *babyjub.PublicKeyComp directly into strings that follow the BJJ public key hez fotmat (^hez:[A-Za-z0-9_-]{44}$) from/to sql DBs.
|
||||
// It assumes that *babyjub.PublicKeyComp are inserted/fetched to/from the DB using the default Scan/Value interface
|
||||
type HezBJJ string
|
||||
|
||||
@@ -184,7 +216,7 @@ func (b HezBJJ) Value() (driver.Value, error) {
|
||||
// StrHezBJJ is used to unmarshal HezBJJ directly into an alias of babyjub.PublicKeyComp
|
||||
type StrHezBJJ babyjub.PublicKeyComp
|
||||
|
||||
// UnmarshalText unmarshalls a StrHezBJJ
|
||||
// UnmarshalText unmarshals a StrHezBJJ
|
||||
func (s *StrHezBJJ) UnmarshalText(text []byte) error {
|
||||
bjj, err := hezStrToBJJ(string(text))
|
||||
if err != nil {
|
||||
@@ -194,8 +226,8 @@ func (s *StrHezBJJ) UnmarshalText(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezIdx is used to value common.Idx directly into strings that follow the Idx key hez format (hez:tokenSymbol:idx) to sql DBs.
|
||||
// Note that this can only be used to insert to DB since there is no way to automatically read from the DB since it needs the tokenSymbol
|
||||
// HezIdx is used to value common.Idx directly into strings that follow the Idx key hez fotmat (hez:tokenSymbol:idx) to sql DBs.
|
||||
// Note that this can only be used to insert to DB since there is no way to automaticaly read from the DB since it needs the tokenSymbol
|
||||
type HezIdx string
|
||||
|
||||
// StrHezIdx is used to unmarshal HezIdx directly into an alias of common.Idx
|
||||
|
||||
@@ -28,8 +28,7 @@ type ConfigBatch struct {
|
||||
|
||||
// NewBatchBuilder constructs a new BatchBuilder, and executes the bb.Reset
|
||||
// method
|
||||
func NewBatchBuilder(dbpath string, synchronizerStateDB *statedb.StateDB, batchNum common.BatchNum,
|
||||
nLevels uint64) (*BatchBuilder, error) {
|
||||
func NewBatchBuilder(dbpath string, synchronizerStateDB *statedb.StateDB, batchNum common.BatchNum, nLevels uint64) (*BatchBuilder, error) {
|
||||
localStateDB, err := statedb.NewLocalStateDB(
|
||||
statedb.Config{
|
||||
Path: dbpath,
|
||||
|
||||
@@ -15,8 +15,7 @@ func TestBatchBuilder(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
defer assert.Nil(t, os.RemoveAll(dir))
|
||||
|
||||
synchDB, err := statedb.NewStateDB(statedb.Config{Path: dir, Keep: 128,
|
||||
Type: statedb.TypeBatchBuilder, NLevels: 0})
|
||||
synchDB, err := statedb.NewStateDB(statedb.Config{Path: dir, Keep: 128, Type: statedb.TypeBatchBuilder, NLevels: 0})
|
||||
assert.Nil(t, err)
|
||||
|
||||
bbDir, err := ioutil.TempDir("", "tmpBatchBuilderDB")
|
||||
|
||||
@@ -8,7 +8,7 @@ The `hermez-node` has been tested with go version 1.14
|
||||
|
||||
## Usage
|
||||
|
||||
```shell
|
||||
```
|
||||
NAME:
|
||||
hermez-node - A new cli application
|
||||
|
||||
@@ -16,18 +16,18 @@ USAGE:
|
||||
node [global options] command [command options] [arguments...]
|
||||
|
||||
VERSION:
|
||||
v0.1.0-6-gd8a50c5
|
||||
0.1.0-alpha
|
||||
|
||||
COMMANDS:
|
||||
version Show the application version
|
||||
importkey Import ethereum private key
|
||||
genbjj Generate a new BabyJubJub key
|
||||
wipesql Wipe the SQL DB (HistoryDB and L2DB) and the StateDBs, leaving the DB in a clean state
|
||||
wipesql Wipe the SQL DB (HistoryDB and L2DB), leaving the DB in a clean state
|
||||
run Run the hermez-node in the indicated mode
|
||||
discard Discard blocks up to a specified block number
|
||||
help, h Shows a list of commands or help for one command
|
||||
|
||||
GLOBAL OPTIONS:
|
||||
--mode MODE Set node MODE (can be "sync" or "coord")
|
||||
--cfg FILE Node configuration FILE
|
||||
--help, -h show help (default: false)
|
||||
--version, -v print the version (default: false)
|
||||
```
|
||||
@@ -75,7 +75,7 @@ when running the coordinator in sync mode
|
||||
|
||||
Building the node requires using the packr utility to bundle the database
|
||||
migrations inside the resulting binary. Install the packr utility with:
|
||||
```shell
|
||||
```
|
||||
cd /tmp && go get -u github.com/gobuffalo/packr/v2/packr2 && cd -
|
||||
```
|
||||
|
||||
@@ -83,7 +83,7 @@ Make sure your `$PATH` contains `$GOPATH/bin`, otherwise the packr utility will
|
||||
not be found.
|
||||
|
||||
Now build the node executable:
|
||||
```shell
|
||||
```
|
||||
cd ../../db && packr2 && cd -
|
||||
go build .
|
||||
cd ../../db && packr2 clean && cd -
|
||||
@@ -98,40 +98,35 @@ run the following examples by replacing `./node` with `go run .` and executing
|
||||
them in the `cli/node` directory to build from source and run at the same time.
|
||||
|
||||
Run the node in mode synchronizer:
|
||||
```shell
|
||||
./node run --mode sync --cfg cfg.buidler.toml
|
||||
```
|
||||
./node --mode sync --cfg cfg.buidler.toml run
|
||||
```
|
||||
|
||||
Run the node in mode coordinator:
|
||||
```shell
|
||||
./node run --mode coord --cfg cfg.buidler.toml
|
||||
```
|
||||
./node --mode coord --cfg cfg.buidler.toml run
|
||||
```
|
||||
|
||||
Import an ethereum private key into the keystore:
|
||||
```shell
|
||||
./node importkey --mode coord --cfg cfg.buidler.toml --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
||||
```
|
||||
./node --mode coord --cfg cfg.buidler.toml importkey --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
||||
```
|
||||
|
||||
Generate a new BabyJubJub key pair:
|
||||
```shell
|
||||
./node genbjj
|
||||
```
|
||||
|
||||
Check the binary version:
|
||||
```shell
|
||||
./node version
|
||||
./node --mode coord --cfg cfg.buidler.toml genbjj
|
||||
```
|
||||
|
||||
Wipe the entier SQL database (this will destroy all synchronized and pool
|
||||
data):
|
||||
```shell
|
||||
./node wipesql --mode coord --cfg cfg.buidler.toml
|
||||
```
|
||||
./node --mode coord --cfg cfg.buidler.toml wipesql
|
||||
```
|
||||
|
||||
Discard all synchronized blocks and associated state up to a given block
|
||||
number. This command is useful in case the synchronizer reaches an invalid
|
||||
state and you want to roll back a few blocks and try again (maybe with some
|
||||
fixes in the code).
|
||||
```shell
|
||||
./node discard --mode coord --cfg cfg.buidler.toml --block 8061330
|
||||
```
|
||||
./node --mode coord --cfg cfg.buidler.toml discard --block 8061330
|
||||
```
|
||||
|
||||
@@ -1,24 +0,0 @@
|
||||
[API]
|
||||
Address = "localhost:8386"
|
||||
Explorer = true
|
||||
MaxSQLConnections = 10
|
||||
SQLConnectionTimeout = "2s"
|
||||
|
||||
[PostgreSQL]
|
||||
PortWrite = 5432
|
||||
HostWrite = "localhost"
|
||||
UserWrite = "hermez"
|
||||
PasswordWrite = "yourpasswordhere"
|
||||
NameWrite = "hermez"
|
||||
|
||||
[Coordinator.L2DB]
|
||||
SafetyPeriod = 10
|
||||
MaxTxs = 512
|
||||
TTL = "24h"
|
||||
PurgeBatchDelay = 10
|
||||
InvalidateBatchDelay = 20
|
||||
PurgeBlockDelay = 10
|
||||
InvalidateBlockDelay = 20
|
||||
|
||||
[Coordinator.API]
|
||||
Coordinator = true
|
||||
@@ -8,31 +8,8 @@ SQLConnectionTimeout = "2s"
|
||||
|
||||
[PriceUpdater]
|
||||
Interval = "10s"
|
||||
URLBitfinexV2 = "https://api-pub.bitfinex.com/v2/"
|
||||
URLCoinGeckoV3 = "https://api.coingecko.com/api/v3/"
|
||||
# Available update methods:
|
||||
# - coingeckoV3 (recommended): get price by SC addr using coingecko API
|
||||
# - bitfinexV2: get price by token symbol using bitfinex API
|
||||
# - static (recommended for blacklisting tokens): use the given StaticValue to set the price (if not provided 0 will be used)
|
||||
# - ignore: don't update the price leave it as it is on the DB
|
||||
DefaultUpdateMethod = "coingeckoV3" # Update method used for all the tokens registered on the network, and not listed in [[PriceUpdater.TokensConfig]]
|
||||
[[PriceUpdater.TokensConfig]]
|
||||
UpdateMethod = "bitfinexV2"
|
||||
Symbol = "USDT"
|
||||
Addr = "0xdac17f958d2ee523a2206206994597c13d831ec7"
|
||||
[[PriceUpdater.TokensConfig]]
|
||||
UpdateMethod = "coingeckoV3"
|
||||
Symbol = "ETH"
|
||||
Addr = "0x0000000000000000000000000000000000000000"
|
||||
[[PriceUpdater.TokensConfig]]
|
||||
UpdateMethod = "static"
|
||||
Symbol = "UNI"
|
||||
Addr = "0x1f9840a85d5af5bf1d1762f925bdaddc4201f984"
|
||||
StaticValue = 30.12
|
||||
[[PriceUpdater.TokensConfig]]
|
||||
UpdateMethod = "ignore"
|
||||
Symbol = "SUSHI"
|
||||
Addr = "0x6b3595068778dd592e39a122f4f5a5cf09c90fe2"
|
||||
URL = "https://api-pub.bitfinex.com/v2/"
|
||||
Type = "bitfinexV2"
|
||||
|
||||
[Debug]
|
||||
APIAddress = "localhost:12345"
|
||||
@@ -74,7 +51,7 @@ ForgerAddress = "0x05c23b938a85ab26A36E6314a0D02080E9ca6BeD" # Non-Boot Coordina
|
||||
# ForgerAddressPrivateKey = "0x30f5fddb34cd4166adb2c6003fa6b18f380fd2341376be42cf1c7937004ac7a3"
|
||||
# ForgerAddress = "0xb4124ceb3451635dacedd11767f004d8a28c6ee7" # Boot Coordinator
|
||||
# ForgerAddressPrivateKey = "0xa8a54b2d8197bc0b19bb8a084031be71835580a01e70a45a13babd16c9bc1563"
|
||||
MinimumForgeAddressBalance = "0"
|
||||
MinimumForgeAddressBalance = 0
|
||||
ConfirmBlocks = 10
|
||||
L1BatchTimeoutPerc = 0.6
|
||||
StartSlotBlocksDelay = 2
|
||||
@@ -86,8 +63,6 @@ SyncRetryInterval = "1s"
|
||||
ForgeDelay = "10s"
|
||||
ForgeNoTxsDelay = "0s"
|
||||
PurgeByExtDelInterval = "1m"
|
||||
MustForgeAtSlotDeadline = true
|
||||
IgnoreSlotCommitment = false
|
||||
|
||||
[Coordinator.FeeAccount]
|
||||
Address = "0x56232B1c5B10038125Bc7345664B4AFD745bcF8E"
|
||||
@@ -132,10 +107,10 @@ Path = "/tmp/iden3-test/hermez/ethkeystore"
|
||||
Password = "yourpasswordhere"
|
||||
|
||||
[Coordinator.EthClient.ForgeBatchGasCost]
|
||||
Fixed = 600000
|
||||
L1UserTx = 15000
|
||||
L1CoordTx = 8000
|
||||
L2Tx = 250
|
||||
Fixed = 500000
|
||||
L1UserTx = 8000
|
||||
L1CoordTx = 9000
|
||||
L2Tx = 1
|
||||
|
||||
[Coordinator.API]
|
||||
Coordinator = true
|
||||
|
||||
247
cli/node/main.go
247
cli/node/main.go
@@ -5,16 +5,13 @@ import (
|
||||
"fmt"
|
||||
"os"
|
||||
"os/signal"
|
||||
"path"
|
||||
"strings"
|
||||
|
||||
ethKeystore "github.com/ethereum/go-ethereum/accounts/keystore"
|
||||
"github.com/ethereum/go-ethereum/crypto"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/config"
|
||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/kvdb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/node"
|
||||
@@ -34,22 +31,6 @@ const (
|
||||
modeCoord = "coord"
|
||||
)
|
||||
|
||||
var (
|
||||
// Version represents the program based on the git tag
|
||||
Version = "v0.1.0"
|
||||
// Build represents the program based on the git commit
|
||||
Build = "dev"
|
||||
// Date represents the date of application was built
|
||||
Date = ""
|
||||
)
|
||||
|
||||
func cmdVersion(c *cli.Context) error {
|
||||
fmt.Printf("Version = \"%v\"\n", Version)
|
||||
fmt.Printf("Build = \"%v\"\n", Build)
|
||||
fmt.Printf("Date = \"%v\"\n", Date)
|
||||
return nil
|
||||
}
|
||||
|
||||
func cmdGenBJJ(c *cli.Context) error {
|
||||
sk := babyjub.NewRandPrivKey()
|
||||
skBuf := [32]byte(sk)
|
||||
@@ -91,86 +72,6 @@ func cmdImportKey(c *cli.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func resetStateDBs(cfg *Config, batchNum common.BatchNum) error {
|
||||
log.Infof("Reset Synchronizer StateDB to batchNum %v...", batchNum)
|
||||
|
||||
// Manually make a checkpoint from batchNum to current to force current
|
||||
// to be a valid checkpoint. This is useful because in case of a
|
||||
// crash, current can be corrupted and the first thing that
|
||||
// `kvdb.NewKVDB` does is read the current checkpoint, which wouldn't
|
||||
// succeed in case of corruption.
|
||||
dbPath := cfg.node.StateDB.Path
|
||||
source := path.Join(dbPath, fmt.Sprintf("%s%d", kvdb.PathBatchNum, batchNum))
|
||||
current := path.Join(dbPath, kvdb.PathCurrent)
|
||||
last := path.Join(dbPath, kvdb.PathLast)
|
||||
if err := os.RemoveAll(last); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
if batchNum == 0 {
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
} else {
|
||||
if err := kvdb.PebbleMakeCheckpoint(source, current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.PebbleMakeCheckpoint: %w", err))
|
||||
}
|
||||
}
|
||||
db, err := kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(batchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
|
||||
if cfg.mode == node.ModeCoordinator {
|
||||
log.Infof("Wipe Coordinator StateDBs...")
|
||||
|
||||
// We wipe the Coordinator StateDBs entirely (by deleting
|
||||
// current and resetting to batchNum 0) because the Coordinator
|
||||
// StateDBs are always reset from Synchronizer when the
|
||||
// coordinator pipeline starts.
|
||||
dbPath := cfg.node.Coordinator.TxSelector.Path
|
||||
current := path.Join(dbPath, kvdb.PathCurrent)
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
db, err := kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
|
||||
dbPath = cfg.node.Coordinator.BatchBuilder.Path
|
||||
current = path.Join(dbPath, kvdb.PathCurrent)
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
db, err = kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("statedb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func cmdWipeSQL(c *cli.Context) error {
|
||||
_cfg, err := parseCli(c)
|
||||
if err != nil {
|
||||
@@ -179,8 +80,7 @@ func cmdWipeSQL(c *cli.Context) error {
|
||||
cfg := _cfg.node
|
||||
yes := c.Bool(flagYes)
|
||||
if !yes {
|
||||
fmt.Print("*WARNING* Are you sure you want to delete " +
|
||||
"the SQL DB and StateDBs? [y/N]: ")
|
||||
fmt.Print("*WARNING* Are you sure you want to delete the SQL DB? [y/N]: ")
|
||||
var input string
|
||||
if _, err := fmt.Scanln(&input); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
@@ -202,17 +102,22 @@ func cmdWipeSQL(c *cli.Context) error {
|
||||
}
|
||||
log.Info("Wiping SQL DB...")
|
||||
if err := dbUtils.MigrationsDown(db.DB); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("dbUtils.MigrationsDown: %w", err))
|
||||
}
|
||||
|
||||
log.Info("Wiping StateDBs...")
|
||||
if err := resetStateDBs(_cfg, 0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("resetStateDBs: %w", err))
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func waitSigInt() {
|
||||
func cmdRun(c *cli.Context) error {
|
||||
cfg, err := parseCli(c)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error parsing flags and config: %w", err))
|
||||
}
|
||||
node, err := node.NewNode(cfg.mode, cfg.node)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error starting node: %w", err))
|
||||
}
|
||||
node.Start()
|
||||
|
||||
stopCh := make(chan interface{})
|
||||
|
||||
// catch ^C to send the stop signal
|
||||
@@ -233,36 +138,48 @@ func waitSigInt() {
|
||||
}
|
||||
}()
|
||||
<-stopCh
|
||||
}
|
||||
|
||||
func cmdRun(c *cli.Context) error {
|
||||
cfg, err := parseCli(c)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error parsing flags and config: %w", err))
|
||||
}
|
||||
node, err := node.NewNode(cfg.mode, cfg.node)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error starting node: %w", err))
|
||||
}
|
||||
node.Start()
|
||||
waitSigInt()
|
||||
node.Stop()
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func cmdServeAPI(c *cli.Context) error {
|
||||
cfg, err := parseCliAPIServer(c)
|
||||
cfgPath := c.String(flagCfg)
|
||||
cfg, err := config.LoadAPIServer(cfgPath)
|
||||
if err != nil {
|
||||
if err := cli.ShowAppHelp(c); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
return tracerr.Wrap(fmt.Errorf("error parsing flags and config: %w", err))
|
||||
}
|
||||
srv, err := node.NewAPIServer(cfg.mode, cfg.server)
|
||||
|
||||
node, err := node.NewNode(cfg.mode, cfg.node)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error starting api server: %w", err))
|
||||
return tracerr.Wrap(fmt.Errorf("error starting node: %w", err))
|
||||
}
|
||||
srv.Start()
|
||||
waitSigInt()
|
||||
srv.Stop()
|
||||
node.Start()
|
||||
|
||||
stopCh := make(chan interface{})
|
||||
|
||||
// catch ^C to send the stop signal
|
||||
ossig := make(chan os.Signal, 1)
|
||||
signal.Notify(ossig, os.Interrupt)
|
||||
const forceStopCount = 3
|
||||
go func() {
|
||||
n := 0
|
||||
for sig := range ossig {
|
||||
if sig == os.Interrupt {
|
||||
log.Info("Received Interrupt Signal")
|
||||
stopCh <- nil
|
||||
n++
|
||||
if n == forceStopCount {
|
||||
log.Fatalf("Received %v Interrupt Signals", forceStopCount)
|
||||
}
|
||||
}
|
||||
}
|
||||
}()
|
||||
<-stopCh
|
||||
node.Stop()
|
||||
|
||||
return nil
|
||||
}
|
||||
@@ -325,11 +242,6 @@ func cmdDiscard(c *cli.Context) error {
|
||||
return tracerr.Wrap(fmt.Errorf("l2DB.Reorg: %w", err))
|
||||
}
|
||||
|
||||
log.Info("Resetting StateDBs...")
|
||||
if err := resetStateDBs(_cfg, batchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("resetStateDBs: %w", err))
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -358,55 +270,13 @@ func getConfig(c *cli.Context) (*Config, error) {
|
||||
switch mode {
|
||||
case modeSync:
|
||||
cfg.mode = node.ModeSynchronizer
|
||||
cfg.node, err = config.LoadNode(nodeCfgPath, false)
|
||||
cfg.node, err = config.LoadNode(nodeCfgPath)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
case modeCoord:
|
||||
cfg.mode = node.ModeCoordinator
|
||||
cfg.node, err = config.LoadNode(nodeCfgPath, true)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
default:
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid mode \"%v\"", mode))
|
||||
}
|
||||
|
||||
return &cfg, nil
|
||||
}
|
||||
|
||||
// ConfigAPIServer is the configuration of the api server execution
|
||||
type ConfigAPIServer struct {
|
||||
mode node.Mode
|
||||
server *config.APIServer
|
||||
}
|
||||
|
||||
func parseCliAPIServer(c *cli.Context) (*ConfigAPIServer, error) {
|
||||
cfg, err := getConfigAPIServer(c)
|
||||
if err != nil {
|
||||
if err := cli.ShowAppHelp(c); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return cfg, nil
|
||||
}
|
||||
|
||||
func getConfigAPIServer(c *cli.Context) (*ConfigAPIServer, error) {
|
||||
var cfg ConfigAPIServer
|
||||
mode := c.String(flagMode)
|
||||
nodeCfgPath := c.String(flagCfg)
|
||||
var err error
|
||||
switch mode {
|
||||
case modeSync:
|
||||
cfg.mode = node.ModeSynchronizer
|
||||
cfg.server, err = config.LoadAPIServer(nodeCfgPath, false)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
case modeCoord:
|
||||
cfg.mode = node.ModeCoordinator
|
||||
cfg.server, err = config.LoadAPIServer(nodeCfgPath, true)
|
||||
cfg.node, err = config.LoadCoordinator(nodeCfgPath)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -420,8 +290,8 @@ func getConfigAPIServer(c *cli.Context) (*ConfigAPIServer, error) {
|
||||
func main() {
|
||||
app := cli.NewApp()
|
||||
app.Name = "hermez-node"
|
||||
app.Version = Version
|
||||
flags := []cli.Flag{
|
||||
app.Version = "0.1.0-alpha"
|
||||
app.Flags = []cli.Flag{
|
||||
&cli.StringFlag{
|
||||
Name: flagMode,
|
||||
Usage: fmt.Sprintf("Set node `MODE` (can be \"%v\" or \"%v\")", modeSync, modeCoord),
|
||||
@@ -435,23 +305,17 @@ func main() {
|
||||
}
|
||||
|
||||
app.Commands = []*cli.Command{
|
||||
{
|
||||
Name: "version",
|
||||
Aliases: []string{},
|
||||
Usage: "Show the application version and build",
|
||||
Action: cmdVersion,
|
||||
},
|
||||
{
|
||||
Name: "importkey",
|
||||
Aliases: []string{},
|
||||
Usage: "Import ethereum private key",
|
||||
Action: cmdImportKey,
|
||||
Flags: append(flags,
|
||||
Flags: []cli.Flag{
|
||||
&cli.StringFlag{
|
||||
Name: flagSK,
|
||||
Usage: "ethereum `PRIVATE_KEY` in hex",
|
||||
Required: true,
|
||||
}),
|
||||
}},
|
||||
},
|
||||
{
|
||||
Name: "genbjj",
|
||||
@@ -462,22 +326,21 @@ func main() {
|
||||
{
|
||||
Name: "wipesql",
|
||||
Aliases: []string{},
|
||||
Usage: "Wipe the SQL DB (HistoryDB and L2DB) and the StateDBs, " +
|
||||
Usage: "Wipe the SQL DB (HistoryDB and L2DB), " +
|
||||
"leaving the DB in a clean state",
|
||||
Action: cmdWipeSQL,
|
||||
Flags: append(flags,
|
||||
Flags: []cli.Flag{
|
||||
&cli.BoolFlag{
|
||||
Name: flagYes,
|
||||
Usage: "automatic yes to the prompt",
|
||||
Required: false,
|
||||
}),
|
||||
}},
|
||||
},
|
||||
{
|
||||
Name: "run",
|
||||
Aliases: []string{},
|
||||
Usage: "Run the hermez-node in the indicated mode",
|
||||
Action: cmdRun,
|
||||
Flags: flags,
|
||||
},
|
||||
{
|
||||
Name: "serveapi",
|
||||
@@ -490,12 +353,12 @@ func main() {
|
||||
Aliases: []string{},
|
||||
Usage: "Discard blocks up to a specified block number",
|
||||
Action: cmdDiscard,
|
||||
Flags: append(flags,
|
||||
Flags: []cli.Flag{
|
||||
&cli.Int64Flag{
|
||||
Name: flagBlock,
|
||||
Usage: "last block number to keep",
|
||||
Required: false,
|
||||
}),
|
||||
}},
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
@@ -72,8 +72,7 @@ func (idx Idx) BigInt() *big.Int {
|
||||
// IdxFromBytes returns Idx from a byte array
|
||||
func IdxFromBytes(b []byte) (Idx, error) {
|
||||
if len(b) != IdxBytesLen {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse Idx, bytes len %d, expected %d",
|
||||
len(b), IdxBytesLen))
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse Idx, bytes len %d, expected %d", len(b), IdxBytesLen))
|
||||
}
|
||||
var idxBytes [8]byte
|
||||
copy(idxBytes[2:], b[:])
|
||||
@@ -195,8 +194,7 @@ func (a *Account) BigInts() ([NLeafElems]*big.Int, error) {
|
||||
return e, nil
|
||||
}
|
||||
|
||||
// HashValue returns the value of the Account, which is the Poseidon hash of its
|
||||
// *big.Int representation
|
||||
// HashValue returns the value of the Account, which is the Poseidon hash of its *big.Int representation
|
||||
func (a *Account) HashValue() (*big.Int, error) {
|
||||
bi, err := a.BigInts()
|
||||
if err != nil {
|
||||
|
||||
@@ -76,8 +76,7 @@ func TestNonceParser(t *testing.T) {
|
||||
|
||||
func TestAccount(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -116,8 +115,7 @@ func TestAccountLoop(t *testing.T) {
|
||||
// check that for different deterministic BabyJubJub keys & random Address there is no problem
|
||||
for i := 0; i < 256; i++ {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -201,8 +199,7 @@ func bigFromStr(h string, u int) *big.Int {
|
||||
|
||||
func TestAccountHashValue(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -215,16 +212,13 @@ func TestAccountHashValue(t *testing.T) {
|
||||
}
|
||||
v, err := account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"447675324273474410516096114710387312413478475468606444107594732044698919451",
|
||||
v.String())
|
||||
assert.Equal(t, "16297758255249203915951182296472515138555043617458222397753168518282206850764", v.String())
|
||||
}
|
||||
|
||||
func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
// values from js test vectors
|
||||
ay := new(big.Int).Sub(new(big.Int).Exp(big.NewInt(2), big.NewInt(253), nil), big.NewInt(1))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff",
|
||||
(hex.EncodeToString(ay.Bytes())))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff", (hex.EncodeToString(ay.Bytes())))
|
||||
bjjPoint, err := babyjub.PointFromSignAndY(true, ay)
|
||||
require.NoError(t, err)
|
||||
bjj := babyjub.PublicKey(*bjjPoint)
|
||||
@@ -242,22 +236,16 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "9444732965739290427391", e[0].String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895", e[1].String())
|
||||
assert.Equal(t,
|
||||
"14474011154664524427946373126085988481658748083205070504932198000989141204991",
|
||||
e[2].String())
|
||||
assert.Equal(t, "14474011154664524427946373126085988481658748083205070504932198000989141204991", e[2].String())
|
||||
assert.Equal(t, "1461501637330902918203684832716283019655932542975", e[3].String())
|
||||
|
||||
h, err := poseidon.Hash(e[:])
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"13265203488631320682117942952393454767418777767637549409684833552016769103047",
|
||||
h.String())
|
||||
assert.Equal(t, "4550823210217540218403400309533329186487982452461145263910122718498735057257", h.String())
|
||||
|
||||
v, err := account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"13265203488631320682117942952393454767418777767637549409684833552016769103047",
|
||||
v.String())
|
||||
assert.Equal(t, "4550823210217540218403400309533329186487982452461145263910122718498735057257", v.String())
|
||||
|
||||
// second account
|
||||
ay = big.NewInt(0)
|
||||
@@ -273,9 +261,7 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
}
|
||||
v, err = account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"2351654555892372227640888372176282444150254868378439619268573230312091195718",
|
||||
v.String())
|
||||
assert.Equal(t, "7750253361301235345986002241352365187241910378619330147114280396816709365657", v.String())
|
||||
|
||||
// third account
|
||||
ay = bigFromStr("21b0a1688b37f77b1d1d5539ec3b826db5ac78b2513f574a04c50a7d4f8246d7", 16)
|
||||
@@ -293,15 +279,11 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "554050781187", e[0].String())
|
||||
assert.Equal(t, "42000000000000000000", e[1].String())
|
||||
assert.Equal(t,
|
||||
"15238403086306505038849621710779816852318505119327426213168494964113886299863",
|
||||
e[2].String())
|
||||
assert.Equal(t, "15238403086306505038849621710779816852318505119327426213168494964113886299863", e[2].String())
|
||||
assert.Equal(t, "935037732739828347587684875151694054123613453305", e[3].String())
|
||||
v, err = account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"15036148928138382129196903417666258171042923749783835283230591475172197254845",
|
||||
v.String())
|
||||
assert.Equal(t, "10565754214047872850889045989683221123564392137456000481397520902594455245517", v.String())
|
||||
}
|
||||
|
||||
func TestAccountErrNotInFF(t *testing.T) {
|
||||
@@ -330,8 +312,7 @@ func TestAccountErrNotInFF(t *testing.T) {
|
||||
|
||||
func TestAccountErrNumOverflowNonce(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -358,8 +339,7 @@ func TestAccountErrNumOverflowNonce(t *testing.T) {
|
||||
|
||||
func TestAccountErrNumOverflowBalance(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -371,16 +351,14 @@ func TestAccountErrNumOverflowBalance(t *testing.T) {
|
||||
BJJ: pk.Compress(),
|
||||
EthAddr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
||||
}
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895",
|
||||
account.Balance.String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895", account.Balance.String())
|
||||
|
||||
_, err = account.Bytes()
|
||||
assert.NoError(t, err)
|
||||
|
||||
// force value overflow
|
||||
account.Balance = new(big.Int).Exp(big.NewInt(2), big.NewInt(192), nil)
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512896",
|
||||
account.Balance.String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512896", account.Balance.String())
|
||||
b, err := account.Bytes()
|
||||
assert.NotNil(t, err)
|
||||
assert.Equal(t, fmt.Errorf("%s Balance", ErrNumOverflow), tracerr.Unwrap(err))
|
||||
|
||||
@@ -11,15 +11,15 @@ import (
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
)
|
||||
|
||||
const (
|
||||
// AccountCreationAuthMsg is the message that is signed to authorize a
|
||||
// Hermez account creation
|
||||
AccountCreationAuthMsg = "Account creation"
|
||||
// EIP712Version is the used version of the EIP-712
|
||||
EIP712Version = "1"
|
||||
// EIP712Provider defines the Provider for the EIP-712
|
||||
EIP712Provider = "Hermez Network"
|
||||
)
|
||||
// AccountCreationAuthMsg is the message that is signed to authorize a Hermez
|
||||
// account creation
|
||||
const AccountCreationAuthMsg = "Account creation"
|
||||
|
||||
// EIP712Version is the used version of the EIP-712
|
||||
const EIP712Version = "1"
|
||||
|
||||
// EIP712Provider defines the Provider for the EIP-712
|
||||
const EIP712Provider = "Hermez Network"
|
||||
|
||||
var (
|
||||
// EmptyEthSignature is an ethereum signature of all zeroes
|
||||
@@ -84,7 +84,7 @@ func (a *AccountCreationAuth) toHash(chainID uint16,
|
||||
return rawData, nil
|
||||
}
|
||||
|
||||
// HashToSign returns the hash to be signed by the Ethereum address to authorize
|
||||
// HashToSign returns the hash to be signed by the Etherum address to authorize
|
||||
// the account creation, which follows the EIP-712 encoding
|
||||
func (a *AccountCreationAuth) HashToSign(chainID uint16,
|
||||
hermezContractAddr ethCommon.Address) ([]byte, error) {
|
||||
@@ -96,9 +96,9 @@ func (a *AccountCreationAuth) HashToSign(chainID uint16,
|
||||
}
|
||||
|
||||
// Sign signs the account creation authorization message using the provided
|
||||
// `signHash` function, and stores the signature in `a.Signature`. `signHash`
|
||||
// `signHash` function, and stores the signaure in `a.Signature`. `signHash`
|
||||
// should do an ethereum signature using the account corresponding to
|
||||
// `a.EthAddr`. The `signHash` function is used to make signing flexible: in
|
||||
// `a.EthAddr`. The `signHash` function is used to make signig flexible: in
|
||||
// tests we sign directly using the private key, outside tests we sign using
|
||||
// the keystore (which never exposes the private key). Sign follows the EIP-712
|
||||
// encoding.
|
||||
|
||||
@@ -13,8 +13,7 @@ import (
|
||||
|
||||
func TestAccountCreationAuthSignVerify(t *testing.T) {
|
||||
// Ethereum key
|
||||
ethSk, err :=
|
||||
ethCrypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
ethSk, err := ethCrypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
require.NoError(t, err)
|
||||
ethAddr := ethCrypto.PubkeyToAddress(ethSk.PublicKey)
|
||||
|
||||
@@ -70,7 +69,6 @@ func TestAccountCreationAuthJSComp(t *testing.T) {
|
||||
sigExpected string
|
||||
}
|
||||
var tvs []testVector
|
||||
//nolint:lll
|
||||
tv0 := testVector{
|
||||
ethSk: "0000000000000000000000000000000000000000000000000000000000000001",
|
||||
expectedAddress: "0x7E5F4552091A69125d5DfCb7b8C2659029395Bdf",
|
||||
@@ -81,7 +79,6 @@ func TestAccountCreationAuthJSComp(t *testing.T) {
|
||||
hashExpected: "c56eba41e511df100c804c5c09288f35887efea4f033be956481af335df3bea2",
|
||||
sigExpected: "dbedcc5ce02db8f48afbdb2feba9a3a31848eaa8fca5f312ce37b01db45d2199208335330d4445bd2f51d1db68dbc0d0bf3585c4a07504b4efbe46a69eaae5a21b",
|
||||
}
|
||||
//nolint:lll
|
||||
tv1 := testVector{
|
||||
ethSk: "0000000000000000000000000000000000000000000000000000000000000002",
|
||||
expectedAddress: "0x2B5AD5c4795c026514f8317c7a215E218DcCD6cF",
|
||||
@@ -92,7 +89,6 @@ func TestAccountCreationAuthJSComp(t *testing.T) {
|
||||
hashExpected: "deb9afa479282cf27b442ce8ba86b19448aa87eacef691521a33db5d0feb9959",
|
||||
sigExpected: "6a0da90ba2d2b1be679a28ebe54ee03082d44b836087391cd7d2607c1e4dafe04476e6e88dccb8707c68312512f16c947524b35c80f26c642d23953e9bb84c701c",
|
||||
}
|
||||
//nolint:lll
|
||||
tv2 := testVector{
|
||||
ethSk: "c5e8f61d1ab959b397eecc0a37a6517b8e67a0e7cf1f4bce5591f3ed80199122",
|
||||
expectedAddress: "0xc783df8a850f42e7F7e57013759C285caa701eB6",
|
||||
|
||||
@@ -13,9 +13,8 @@ const batchNumBytesLen = 8
|
||||
|
||||
// Batch is a struct that represents Hermez network batch
|
||||
type Batch struct {
|
||||
BatchNum BatchNum `meddler:"batch_num"`
|
||||
// Ethereum block in which the batch is forged
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
BatchNum BatchNum `meddler:"batch_num"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum block in which the batch is forged
|
||||
ForgerAddr ethCommon.Address `meddler:"forger_addr"`
|
||||
CollectedFees map[TokenID]*big.Int `meddler:"fees_collected,json"`
|
||||
FeeIdxsCoordinator []Idx `meddler:"fee_idxs_coordinator,json"`
|
||||
@@ -23,11 +22,9 @@ type Batch struct {
|
||||
NumAccounts int `meddler:"num_accounts"`
|
||||
LastIdx int64 `meddler:"last_idx"`
|
||||
ExitRoot *big.Int `meddler:"exit_root,bigint"`
|
||||
// ForgeL1TxsNum is optional, Only when the batch forges L1 txs. Identifier that corresponds
|
||||
// to the group of L1 txs forged in the current batch.
|
||||
ForgeL1TxsNum *int64 `meddler:"forge_l1_txs_num"`
|
||||
SlotNum int64 `meddler:"slot_num"` // Slot in which the batch is forged
|
||||
TotalFeesUSD *float64 `meddler:"total_fees_usd"`
|
||||
ForgeL1TxsNum *int64 `meddler:"forge_l1_txs_num"` // optional, Only when the batch forges L1 txs. Identifier that corresponds to the group of L1 txs forged in the current batch.
|
||||
SlotNum int64 `meddler:"slot_num"` // Slot in which the batch is forged
|
||||
TotalFeesUSD *float64 `meddler:"total_fees_usd"`
|
||||
}
|
||||
|
||||
// NewEmptyBatch creates a new empty batch
|
||||
@@ -66,9 +63,7 @@ func (bn BatchNum) BigInt() *big.Int {
|
||||
// BatchNumFromBytes returns BatchNum from a []byte
|
||||
func BatchNumFromBytes(b []byte) (BatchNum, error) {
|
||||
if len(b) != batchNumBytesLen {
|
||||
return 0,
|
||||
tracerr.Wrap(fmt.Errorf("can not parse BatchNumFromBytes, bytes len %d, expected %d",
|
||||
len(b), batchNumBytesLen))
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse BatchNumFromBytes, bytes len %d, expected %d", len(b), batchNumBytesLen))
|
||||
}
|
||||
batchNum := binary.BigEndian.Uint64(b[:batchNumBytesLen])
|
||||
return BatchNum(batchNum), nil
|
||||
|
||||
@@ -34,7 +34,7 @@ type Slot struct {
|
||||
// BatchesLen int
|
||||
BidValue *big.Int
|
||||
BootCoord bool
|
||||
// Bidder, Forger and URL correspond to the winner of the slot (which is
|
||||
// Bidder, Forer and URL correspond to the winner of the slot (which is
|
||||
// not always the highest bidder). These are the values of the
|
||||
// coordinator that is able to forge exclusively before the deadline.
|
||||
Bidder ethCommon.Address
|
||||
|
||||
@@ -5,15 +5,10 @@ import (
|
||||
)
|
||||
|
||||
// Coordinator represents a Hermez network coordinator who wins an auction for an specific slot
|
||||
// WARNING: this is strongly based on the previous implementation, once the new spec is done, this
|
||||
// may change a lot.
|
||||
// WARNING: this is strongly based on the previous implementation, once the new spec is done, this may change a lot.
|
||||
type Coordinator struct {
|
||||
// Bidder is the address of the bidder
|
||||
Bidder ethCommon.Address `meddler:"bidder_addr"`
|
||||
// Forger is the address of the forger
|
||||
Forger ethCommon.Address `meddler:"forger_addr"`
|
||||
// EthBlockNum is the block in which the coordinator was registered
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
// URL of the coordinators API
|
||||
URL string `meddler:"url"`
|
||||
Bidder ethCommon.Address `meddler:"bidder_addr"` // address of the bidder
|
||||
Forger ethCommon.Address `meddler:"forger_addr"` // address of the forger
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // block in which the coordinator was registered
|
||||
URL string `meddler:"url"` // URL of the coordinators API
|
||||
}
|
||||
|
||||
@@ -7,8 +7,6 @@ type SCVariables struct {
|
||||
WDelayer WDelayerVariables `validate:"required"`
|
||||
}
|
||||
|
||||
// AsPtr returns the SCVariables as a SCVariablesPtr using pointers to the
|
||||
// original SCVariables
|
||||
func (v *SCVariables) AsPtr() *SCVariablesPtr {
|
||||
return &SCVariablesPtr{
|
||||
Rollup: &v.Rollup,
|
||||
|
||||
@@ -68,13 +68,11 @@ type AuctionVariables struct {
|
||||
ClosedAuctionSlots uint16 `meddler:"closed_auction_slots" validate:"required"`
|
||||
// Distance (#slots) to the farthest slot to which you can bid (30 days = 4320 slots )
|
||||
OpenAuctionSlots uint16 `meddler:"open_auction_slots" validate:"required"`
|
||||
// How the HEZ tokens deposited by the slot winner are distributed (Burn: 40% - Donation:
|
||||
// 40% - HGT: 20%)
|
||||
// How the HEZ tokens deposited by the slot winner are distributed (Burn: 40% - Donation: 40% - HGT: 20%)
|
||||
AllocationRatio [3]uint16 `meddler:"allocation_ratio,json" validate:"required"`
|
||||
// Minimum outbid (percentage) over the previous one to consider it valid
|
||||
Outbidding uint16 `meddler:"outbidding" validate:"required"`
|
||||
// Number of blocks at the end of a slot in which any coordinator can forge if the winner
|
||||
// has not forged one before
|
||||
// Number of blocks at the end of a slot in which any coordinator can forge if the winner has not forged one before
|
||||
SlotDeadline uint8 `meddler:"slot_deadline" validate:"required"`
|
||||
}
|
||||
|
||||
|
||||
@@ -20,23 +20,22 @@ const (
|
||||
// RollupConstExitIDx IDX 1 is reserved for exits
|
||||
RollupConstExitIDx = 1
|
||||
// RollupConstLimitTokens Max number of tokens allowed to be registered inside the rollup
|
||||
RollupConstLimitTokens = (1 << 32) //nolint:gomnd
|
||||
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes]
|
||||
// compressedSignature
|
||||
RollupConstLimitTokens = (1 << 32)
|
||||
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes] compressedSignature
|
||||
RollupConstL1CoordinatorTotalBytes = 101
|
||||
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6
|
||||
// bytes] fromIdx + [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes]
|
||||
// tokenId + [6 bytes] toIdx
|
||||
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx +
|
||||
// [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes] tokenId + [6 bytes] toIdx
|
||||
RollupConstL1UserTotalBytes = 78
|
||||
// RollupConstMaxL1UserTx Maximum L1-user transactions allowed to be queued in a batch
|
||||
RollupConstMaxL1UserTx = 128
|
||||
// RollupConstMaxL1Tx Maximum L1 transactions allowed to be queued in a batch
|
||||
RollupConstMaxL1Tx = 256
|
||||
// RollupConstInputSHAConstantBytes [6 bytes] lastIdx + [6 bytes] newLastIdx + [32 bytes]
|
||||
// stateRoot + [32 bytes] newStRoot + [32 bytes] newExitRoot + [_MAX_L1_TX *
|
||||
// _L1_USER_TOTALBYTES bytes] l1TxsData + totalL2TxsDataLength + feeIdxCoordinatorLength +
|
||||
// [2 bytes] chainID = 18542 bytes + totalL2TxsDataLength + feeIdxCoordinatorLength
|
||||
// RollupConstInputSHAConstantBytes [6 bytes] lastIdx + [6 bytes] newLastIdx + [32 bytes] stateRoot + [32 bytes] newStRoot + [32 bytes] newExitRoot +
|
||||
// [_MAX_L1_TX * _L1_USER_TOTALBYTES bytes] l1TxsData + totalL2TxsDataLength + feeIdxCoordinatorLength + [2 bytes] chainID =
|
||||
// 18542 bytes + totalL2TxsDataLength + feeIdxCoordinatorLength
|
||||
RollupConstInputSHAConstantBytes = 18546
|
||||
// RollupConstNumBuckets Number of buckets
|
||||
RollupConstNumBuckets = 5
|
||||
// RollupConstMaxWithdrawalDelay max withdrawal delay in seconds
|
||||
RollupConstMaxWithdrawalDelay = 2 * 7 * 24 * 60 * 60
|
||||
// RollupConstExchangeMultiplier exchange multiplier
|
||||
@@ -45,18 +44,14 @@ const (
|
||||
|
||||
var (
|
||||
// RollupConstLimitDepositAmount Max deposit amount allowed (depositAmount: L1 --> L2)
|
||||
RollupConstLimitDepositAmount, _ = new(big.Int).SetString(
|
||||
"340282366920938463463374607431768211456", 10)
|
||||
RollupConstLimitDepositAmount, _ = new(big.Int).SetString("340282366920938463463374607431768211456", 10)
|
||||
// RollupConstLimitL2TransferAmount Max amount allowed (amount L2 --> L2)
|
||||
RollupConstLimitL2TransferAmount, _ = new(big.Int).SetString(
|
||||
"6277101735386680763835789423207666416102355444464034512896", 10)
|
||||
RollupConstLimitL2TransferAmount, _ = new(big.Int).SetString("6277101735386680763835789423207666416102355444464034512896", 10)
|
||||
|
||||
// RollupConstEthAddressInternalOnly This ethereum address is used internally for rollup
|
||||
// accounts that don't have ethereum address, only Babyjubjub.
|
||||
// This non-ethereum accounts can be created by the coordinator and allow users to have a
|
||||
// rollup account without needing an ethereum address
|
||||
RollupConstEthAddressInternalOnly = ethCommon.HexToAddress(
|
||||
"0xFFfFfFffFFfffFFfFFfFFFFFffFFFffffFfFFFfF")
|
||||
// RollupConstEthAddressInternalOnly This ethereum address is used internally for rollup accounts that don't have ethereum address, only Babyjubjub
|
||||
// This non-ethereum accounts can be created by the coordinator and allow users to have a rollup
|
||||
// account without needing an ethereum address
|
||||
RollupConstEthAddressInternalOnly = ethCommon.HexToAddress("0xFFfFfFffFFfffFFfFFfFFFFFffFFFffffFfFFFfF")
|
||||
// RollupConstRfield Modulus zkSNARK
|
||||
RollupConstRfield, _ = new(big.Int).SetString(
|
||||
"21888242871839275222246405745257275088548364400416034343698204186575808495617", 10)
|
||||
@@ -68,32 +63,24 @@ var (
|
||||
|
||||
// RollupConstRecipientInterfaceHash ERC777 recipient interface hash
|
||||
RollupConstRecipientInterfaceHash = crypto.Keccak256([]byte("ERC777TokensRecipient"))
|
||||
// RollupConstPerformL1UserTxSignature the signature of the function that can be called thru
|
||||
// an ERC777 `send`
|
||||
RollupConstPerformL1UserTxSignature = crypto.Keccak256([]byte(
|
||||
"addL1Transaction(uint256,uint48,uint16,uint16,uint32,uint48)"))
|
||||
// RollupConstAddTokenSignature the signature of the function that can be called thru an
|
||||
// ERC777 `send`
|
||||
// RollupConstPerformL1UserTxSignature the signature of the function that can be called thru an ERC777 `send`
|
||||
RollupConstPerformL1UserTxSignature = crypto.Keccak256([]byte("addL1Transaction(uint256,uint48,uint16,uint16,uint32,uint48)"))
|
||||
// RollupConstAddTokenSignature the signature of the function that can be called thru an ERC777 `send`
|
||||
RollupConstAddTokenSignature = crypto.Keccak256([]byte("addToken(address)"))
|
||||
// RollupConstSendSignature ERC777 Signature
|
||||
RollupConstSendSignature = crypto.Keccak256([]byte("send(address,uint256,bytes)"))
|
||||
// RollupConstERC777Granularity ERC777 Signature
|
||||
RollupConstERC777Granularity = crypto.Keccak256([]byte("granularity()"))
|
||||
// RollupConstWithdrawalDelayerDeposit This constant are used to deposit tokens from ERC77
|
||||
// tokens into withdrawal delayer
|
||||
// RollupConstWithdrawalDelayerDeposit This constant are used to deposit tokens from ERC77 tokens into withdrawal delayer
|
||||
RollupConstWithdrawalDelayerDeposit = crypto.Keccak256([]byte("deposit(address,address,uint192)"))
|
||||
|
||||
// ERC20 signature
|
||||
|
||||
// RollupConstTransferSignature This constant is used in the _safeTransfer internal method
|
||||
// in order to safe GAS.
|
||||
// RollupConstTransferSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
RollupConstTransferSignature = crypto.Keccak256([]byte("transfer(address,uint256)"))
|
||||
// RollupConstTransferFromSignature This constant is used in the _safeTransfer internal
|
||||
// method in order to safe GAS.
|
||||
RollupConstTransferFromSignature = crypto.Keccak256([]byte(
|
||||
"transferFrom(address,address,uint256)"))
|
||||
// RollupConstApproveSignature This constant is used in the _safeTransfer internal method in
|
||||
// order to safe GAS.
|
||||
// RollupConstTransferFromSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
RollupConstTransferFromSignature = crypto.Keccak256([]byte("transferFrom(address,address,uint256)"))
|
||||
// RollupConstApproveSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
RollupConstApproveSignature = crypto.Keccak256([]byte("approve(address,uint256)"))
|
||||
// RollupConstERC20Signature ERC20 decimals signature
|
||||
RollupConstERC20Signature = crypto.Keccak256([]byte("decimals()"))
|
||||
@@ -130,12 +117,10 @@ func (c *RollupConstants) FindVerifierIdx(MaxTx, NLevels int64) (int, error) {
|
||||
// BucketParams are the parameter variables of each Bucket of Rollup Smart
|
||||
// Contract
|
||||
type BucketParams struct {
|
||||
CeilUSD *big.Int
|
||||
BlockStamp *big.Int
|
||||
Withdrawals *big.Int
|
||||
RateBlocks *big.Int
|
||||
RateWithdrawals *big.Int
|
||||
MaxWithdrawals *big.Int
|
||||
CeilUSD *big.Int
|
||||
Withdrawals *big.Int
|
||||
BlockWithdrawalRate *big.Int
|
||||
MaxWithdrawals *big.Int
|
||||
}
|
||||
|
||||
// BucketUpdate are the bucket updates (tracking the withdrawals value changes)
|
||||
@@ -156,14 +141,13 @@ type TokenExchange struct {
|
||||
}
|
||||
|
||||
// RollupVariables are the variables of the Rollup Smart Contract
|
||||
//nolint:lll
|
||||
type RollupVariables struct {
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
FeeAddToken *big.Int `meddler:"fee_add_token,bigint" validate:"required"`
|
||||
ForgeL1L2BatchTimeout int64 `meddler:"forge_l1_timeout" validate:"required"`
|
||||
WithdrawalDelay uint64 `meddler:"withdrawal_delay" validate:"required"`
|
||||
Buckets []BucketParams `meddler:"buckets,json"`
|
||||
SafeMode bool `meddler:"safe_mode"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
FeeAddToken *big.Int `meddler:"fee_add_token,bigint" validate:"required"`
|
||||
ForgeL1L2BatchTimeout int64 `meddler:"forge_l1_timeout" validate:"required"`
|
||||
WithdrawalDelay uint64 `meddler:"withdrawal_delay" validate:"required"`
|
||||
Buckets [RollupConstNumBuckets]BucketParams `meddler:"buckets,json"`
|
||||
SafeMode bool `meddler:"safe_mode"`
|
||||
}
|
||||
|
||||
// Copy returns a deep copy of the Variables
|
||||
|
||||
@@ -27,7 +27,6 @@ type WDelayerEscapeHatchWithdrawal struct {
|
||||
}
|
||||
|
||||
// WDelayerVariables are the variables of the Withdrawal Delayer Smart Contract
|
||||
//nolint:lll
|
||||
type WDelayerVariables struct {
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
// HermezRollupAddress ethCommon.Address `json:"hermezRollupAddress" meddler:"rollup_address"`
|
||||
|
||||
@@ -22,9 +22,9 @@ var FeeFactorLsh60 [256]*big.Int
|
||||
// the coordinator according to the tx type (if the tx requires to create an
|
||||
// account and register, only register or he account already esists)
|
||||
type RecommendedFee struct {
|
||||
ExistingAccount float64 `json:"existingAccount"`
|
||||
CreatesAccount float64 `json:"createAccount"`
|
||||
CreatesAccountInternal float64 `json:"createAccountInternal"`
|
||||
ExistingAccount float64 `json:"existingAccount"`
|
||||
CreatesAccount float64 `json:"createAccount"`
|
||||
CreatesAccountAndRegister float64 `json:"createAccountInternal"`
|
||||
}
|
||||
|
||||
// FeeSelector is used to select a percentage from the FeePlan.
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
// Package common float40.go provides methods to work with Hermez custom half
|
||||
// Package common Float40 provides methods to work with Hermez custom half
|
||||
// float precision, 40 bits, codification internally called Float40 has been
|
||||
// adopted to encode large integers. This is done in order to save bits when L2
|
||||
// transactions are published.
|
||||
@@ -32,8 +32,6 @@ var (
|
||||
// ErrFloat40NotEnoughPrecission is used when the given *big.Int can
|
||||
// not be represented as Float40 due not enough precission
|
||||
ErrFloat40NotEnoughPrecission = errors.New("Float40 error, not enough precission")
|
||||
|
||||
thres = big.NewInt(0x08_00_00_00_00)
|
||||
)
|
||||
|
||||
// Float40 represents a float in a 64 bit format
|
||||
@@ -70,7 +68,7 @@ func (f40 Float40) BigInt() (*big.Int, error) {
|
||||
var f40Uint64 uint64 = uint64(f40) & 0x00_00_00_FF_FF_FF_FF_FF
|
||||
f40Bytes, err := f40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, err
|
||||
}
|
||||
|
||||
e := f40Bytes[0] & 0xF8 >> 3 // take first 5 bits
|
||||
@@ -88,41 +86,18 @@ func NewFloat40(f *big.Int) (Float40, error) {
|
||||
e := big.NewInt(0)
|
||||
zero := big.NewInt(0)
|
||||
ten := big.NewInt(10)
|
||||
thres := big.NewInt(0x08_00_00_00_00)
|
||||
for new(big.Int).Mod(m, ten).Cmp(zero) == 0 && m.Cmp(thres) >= 0 {
|
||||
m = new(big.Int).Div(m, ten)
|
||||
e = new(big.Int).Add(e, big.NewInt(1))
|
||||
}
|
||||
if e.Int64() > 31 {
|
||||
return 0, tracerr.Wrap(ErrFloat40E31)
|
||||
return 0, ErrFloat40E31
|
||||
}
|
||||
if m.Cmp(thres) >= 0 {
|
||||
return 0, tracerr.Wrap(ErrFloat40NotEnoughPrecission)
|
||||
return 0, ErrFloat40NotEnoughPrecission
|
||||
}
|
||||
r := new(big.Int).Add(m,
|
||||
new(big.Int).Mul(e, thres))
|
||||
return Float40(r.Uint64()), nil
|
||||
}
|
||||
|
||||
// NewFloat40Floor encodes a *big.Int integer as a Float40, rounding down in
|
||||
// case of loss during the encoding. It returns an error in case that the number
|
||||
// is too big (e>31). Warning: this method should not be used inside the
|
||||
// hermez-node, it's a helper for external usage to generate valid Float40
|
||||
// values.
|
||||
func NewFloat40Floor(f *big.Int) (Float40, error) {
|
||||
m := f
|
||||
e := big.NewInt(0)
|
||||
// zero := big.NewInt(0)
|
||||
ten := big.NewInt(10)
|
||||
for m.Cmp(thres) >= 0 {
|
||||
m = new(big.Int).Div(m, ten)
|
||||
e = new(big.Int).Add(e, big.NewInt(1))
|
||||
}
|
||||
if e.Int64() > 31 {
|
||||
return 0, tracerr.Wrap(ErrFloat40E31)
|
||||
}
|
||||
|
||||
r := new(big.Int).Add(m,
|
||||
new(big.Int).Mul(e, thres))
|
||||
|
||||
return Float40(r.Uint64()), nil
|
||||
}
|
||||
|
||||
@@ -1,11 +1,9 @@
|
||||
package common
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"math/big"
|
||||
"testing"
|
||||
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
@@ -57,56 +55,7 @@ func TestExpectError(t *testing.T) {
|
||||
bi, ok := new(big.Int).SetString(test, 10)
|
||||
require.True(t, ok)
|
||||
_, err := NewFloat40(bi)
|
||||
assert.Equal(t, testVector[test], tracerr.Unwrap(err))
|
||||
}
|
||||
}
|
||||
|
||||
func TestNewFloat40Floor(t *testing.T) {
|
||||
testVector := map[string][]string{
|
||||
// []int contains [Float40 value, Flot40 Floor value], when
|
||||
// Float40 value is expected to be 0, is because is expected to
|
||||
// be an error
|
||||
"9922334455000000000000000000000000000000": {
|
||||
"1040714485495", "1040714485495", "9922334455000000000000000000000000000000"},
|
||||
"9922334455000000000000000000000000000001": { // Floor [2] will be same as prev line
|
||||
"0", "1040714485495", "9922334455000000000000000000000000000000"},
|
||||
"9922334454999999999999999999999999999999": {
|
||||
"0", "1040714485494", "9922334454000000000000000000000000000000"},
|
||||
"42949672950000000000000000000000000000000": {
|
||||
"1069446856703", "1069446856703", "42949672950000000000000000000000000000000"},
|
||||
"99223344556573838487575": {
|
||||
"0", "456598933239", "99223344550000000000000"},
|
||||
"992233445500000000000000000000000000000000": {
|
||||
"0", "0", "0"}, // e>31, returns 0, err
|
||||
"343597383670000000000000000000000000000000": {
|
||||
"1099511627775", "1099511627775", "343597383670000000000000000000000000000000"},
|
||||
"343597383680000000000000000000000000000000": {
|
||||
"0", "0", "0"}, // e>31, returns 0, err
|
||||
"1157073197879933027": {
|
||||
"0", "286448638922", "1157073197800000000"},
|
||||
}
|
||||
for test := range testVector {
|
||||
bi, ok := new(big.Int).SetString(test, 10)
|
||||
require.True(t, ok)
|
||||
f40, err := NewFloat40(bi)
|
||||
if f40 == 0 {
|
||||
assert.Error(t, err)
|
||||
} else {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
assert.Equal(t, testVector[test][0], fmt.Sprint(uint64(f40)))
|
||||
|
||||
f40, err = NewFloat40Floor(bi)
|
||||
if f40 == 0 {
|
||||
assert.Equal(t, ErrFloat40E31, tracerr.Unwrap(err))
|
||||
} else {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
assert.Equal(t, testVector[test][1], fmt.Sprint(uint64(f40)))
|
||||
|
||||
bi2, err := f40.BigInt()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, fmt.Sprint(testVector[test][2]), bi2.String())
|
||||
assert.Equal(t, testVector[test], err)
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -21,33 +21,25 @@ type L1Tx struct {
|
||||
// where type:
|
||||
// - L1UserTx: 0
|
||||
// - L1CoordinatorTx: 1
|
||||
TxID TxID `meddler:"id"`
|
||||
// ToForgeL1TxsNum indicates in which the tx was forged / will be forged
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"`
|
||||
Position int `meddler:"position"`
|
||||
// UserOrigin is set to true if the tx was originated by a user, false if it was
|
||||
// aoriginated by a coordinator. Note that this differ from the spec for implementation
|
||||
// simplification purpposes
|
||||
UserOrigin bool `meddler:"user_origin"`
|
||||
// FromIdx is used by L1Tx/Deposit to indicate the Idx receiver of the L1Tx.DepositAmount
|
||||
// (deposit)
|
||||
FromIdx Idx `meddler:"from_idx,zeroisnull"`
|
||||
TxID TxID `meddler:"id"`
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"` // toForgeL1TxsNum in which the tx was forged / will be forged
|
||||
Position int `meddler:"position"`
|
||||
UserOrigin bool `meddler:"user_origin"` // true if the tx was originated by a user, false if it was aoriginated by a coordinator. Note that this differ from the spec for implementation simplification purpposes
|
||||
FromIdx Idx `meddler:"from_idx,zeroisnull"` // FromIdx is used by L1Tx/Deposit to indicate the Idx receiver of the L1Tx.DepositAmount (deposit)
|
||||
EffectiveFromIdx Idx `meddler:"effective_from_idx,zeroisnull"`
|
||||
FromEthAddr ethCommon.Address `meddler:"from_eth_addr,zeroisnull"`
|
||||
FromBJJ babyjub.PublicKeyComp `meddler:"from_bjj,zeroisnull"`
|
||||
// ToIdx is ignored in L1Tx/Deposit, but used in the L1Tx/DepositAndTransfer
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
ToIdx Idx `meddler:"to_idx"` // ToIdx is ignored in L1Tx/Deposit, but used in the L1Tx/DepositAndTransfer
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
// EffectiveAmount only applies to L1UserTx.
|
||||
EffectiveAmount *big.Int `meddler:"effective_amount,bigintnull"`
|
||||
DepositAmount *big.Int `meddler:"deposit_amount,bigint"`
|
||||
// EffectiveDepositAmount only applies to L1UserTx.
|
||||
EffectiveDepositAmount *big.Int `meddler:"effective_deposit_amount,bigintnull"`
|
||||
// Ethereum Block Number in which this L1Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
Type TxType `meddler:"type"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
EffectiveDepositAmount *big.Int `meddler:"effective_deposit_amount,bigintnull"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum Block Number in which this L1Tx was added to the queue
|
||||
Type TxType `meddler:"type"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
}
|
||||
|
||||
// NewL1Tx returns the given L1Tx with the TxId & Type parameters calculated
|
||||
@@ -259,7 +251,7 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
||||
}
|
||||
l1tx.ToIdx = toIdx
|
||||
l1tx.EffectiveAmount, err = Float40FromBytes(amountBytes).BigInt()
|
||||
return &l1tx, tracerr.Wrap(err)
|
||||
return &l1tx, err
|
||||
}
|
||||
|
||||
// BytesGeneric returns the generic representation of a L1Tx. This method is
|
||||
@@ -339,9 +331,7 @@ func (tx *L1Tx) BytesCoordinatorTx(compressedSignatureBytes []byte) ([]byte, err
|
||||
// L1UserTxFromBytes decodes a L1Tx from []byte
|
||||
func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||
if len(b) != RollupConstL1UserTotalBytes {
|
||||
return nil,
|
||||
tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d",
|
||||
68, len(b)))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d", 68, len(b)))
|
||||
}
|
||||
|
||||
tx := &L1Tx{
|
||||
@@ -379,12 +369,9 @@ func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||
}
|
||||
|
||||
// L1CoordinatorTxFromBytes decodes a L1Tx from []byte
|
||||
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx,
|
||||
error) {
|
||||
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx, error) {
|
||||
if len(b) != RollupConstL1CoordinatorTotalBytes {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d",
|
||||
101, len(b)))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d", 101, len(b)))
|
||||
}
|
||||
|
||||
tx := &L1Tx{
|
||||
|
||||
@@ -29,8 +29,7 @@ func TestNewL1UserTx(t *testing.T) {
|
||||
}
|
||||
l1Tx, err := NewL1Tx(l1Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x00a6cbae3b8661fb75b0919ca6605a02cfb04d9c6dd16870fa0fcdf01befa32768",
|
||||
l1Tx.TxID.String())
|
||||
assert.Equal(t, "0x00a6cbae3b8661fb75b0919ca6605a02cfb04d9c6dd16870fa0fcdf01befa32768", l1Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestNewL1CoordinatorTx(t *testing.T) {
|
||||
@@ -47,8 +46,7 @@ func TestNewL1CoordinatorTx(t *testing.T) {
|
||||
}
|
||||
l1Tx, err := NewL1Tx(l1Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x01274482d73df4dab34a1b6740adfca347a462513aa14e82f27b12f818d1b68c84",
|
||||
l1Tx.TxID.String())
|
||||
assert.Equal(t, "0x01274482d73df4dab34a1b6740adfca347a462513aa14e82f27b12f818d1b68c84", l1Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestL1TxCompressedData(t *testing.T) {
|
||||
@@ -201,8 +199,7 @@ func TestL1userTxByteParsers(t *testing.T) {
|
||||
func TestL1TxByteParsersCompatibility(t *testing.T) {
|
||||
// Data from compatibility test
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompB, err :=
|
||||
hex.DecodeString("0dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a")
|
||||
pkCompB, err := hex.DecodeString("0dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a")
|
||||
require.NoError(t, err)
|
||||
pkCompL := SwapEndianness(pkCompB)
|
||||
err = pkComp.UnmarshalText([]byte(hex.EncodeToString(pkCompL)))
|
||||
@@ -223,8 +220,7 @@ func TestL1TxByteParsersCompatibility(t *testing.T) {
|
||||
|
||||
encodedData, err := l1Tx.BytesUser()
|
||||
require.NoError(t, err)
|
||||
expected := "85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79e" +
|
||||
"e1f41a7def613e73a20c32eece5a000001c638db52540be400459682f0000020039c0000053cb88d"
|
||||
expected := "85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a000001c638db52540be400459682f0000020039c0000053cb88d"
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
}
|
||||
|
||||
@@ -232,8 +228,7 @@ func TestL1CoordinatorTxByteParsers(t *testing.T) {
|
||||
hermezAddress := ethCommon.HexToAddress("0xD6C850aeBFDC46D7F4c207e445cC0d6B0919BDBe")
|
||||
chainID := big.NewInt(1337)
|
||||
|
||||
privateKey, err :=
|
||||
crypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
privateKey, err := crypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
require.NoError(t, err)
|
||||
|
||||
publicKey := privateKey.Public()
|
||||
@@ -305,8 +300,7 @@ func TestL1CoordinatorTxByteParsersCompatibility(t *testing.T) {
|
||||
signature = append(signature, v[:]...)
|
||||
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompB, err :=
|
||||
hex.DecodeString("a2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c")
|
||||
pkCompB, err := hex.DecodeString("a2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c")
|
||||
require.NoError(t, err)
|
||||
pkCompL := SwapEndianness(pkCompB)
|
||||
err = pkComp.UnmarshalText([]byte(hex.EncodeToString(pkCompL)))
|
||||
@@ -321,9 +315,7 @@ func TestL1CoordinatorTxByteParsersCompatibility(t *testing.T) {
|
||||
encodeData, err := l1Tx.BytesCoordinatorTx(signature)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected, err := utils.HexDecode("1b186d7122ff7f654cfed3156719774898d573900c86599a885a706" +
|
||||
"dbdffe5ea8cda71e5eb097e115405d84d1e7b464009b434b32c014a2df502d1f065ced8bc3ba2c28" +
|
||||
"07ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c000000e7")
|
||||
expected, err := utils.HexDecode("1b186d7122ff7f654cfed3156719774898d573900c86599a885a706dbdffe5ea8cda71e5eb097e115405d84d1e7b464009b434b32c014a2df502d1f065ced8bc3ba2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c000000e7")
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, expected, encodeData)
|
||||
|
||||
@@ -10,7 +10,7 @@ import (
|
||||
|
||||
// L2Tx is a struct that represents an already forged L2 tx
|
||||
type L2Tx struct {
|
||||
// Stored in DB: mandatory fields
|
||||
// Stored in DB: mandatory fileds
|
||||
TxID TxID `meddler:"id"`
|
||||
BatchNum BatchNum `meddler:"batch_num"` // batchNum in which this tx was forged.
|
||||
Position int `meddler:"position"`
|
||||
@@ -21,10 +21,9 @@ type L2Tx struct {
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
Fee FeeSelector `meddler:"fee"`
|
||||
// Nonce is filled by the TxProcessor
|
||||
Nonce Nonce `meddler:"nonce"`
|
||||
Type TxType `meddler:"type"`
|
||||
// EthBlockNum in which this L2Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
Nonce Nonce `meddler:"nonce"`
|
||||
Type TxType `meddler:"type"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // EthereumBlockNumber in which this L2Tx was added to the queue
|
||||
}
|
||||
|
||||
// NewL2Tx returns the given L2Tx with the TxId & Type parameters calculated
|
||||
|
||||
@@ -19,8 +19,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err := NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -31,8 +30,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x029e7499a830f8f5eb17c07da48cf91415710f1bcbe0169d363ff91e81faf92fc2",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x029e7499a830f8f5eb17c07da48cf91415710f1bcbe0169d363ff91e81faf92fc2", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -44,8 +42,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x0255c70ed20e1b8935232e1b9c5884dbcc88a6e1a3454d24f2d77252eb2bb0b64e",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0255c70ed20e1b8935232e1b9c5884dbcc88a6e1a3454d24f2d77252eb2bb0b64e", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -57,8 +54,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x0206b372f967061d1148bbcff679de38120e075141a80a07326d0f514c2efc6ca9",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0206b372f967061d1148bbcff679de38120e075141a80a07326d0f514c2efc6ca9", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 1,
|
||||
@@ -70,8 +66,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x0236f7ea5bccf78ba60baf56c058d235a844f9b09259fd0efa4f5f72a7d4a26618",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0236f7ea5bccf78ba60baf56c058d235a844f9b09259fd0efa4f5f72a7d4a26618", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 999,
|
||||
@@ -83,8 +78,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02ac122f5b709ce190129fecbbe35bfd30c70e6433dbd85a8eb743d110906a1dc1",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x02ac122f5b709ce190129fecbbe35bfd30c70e6433dbd85a8eb743d110906a1dc1", l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 4444,
|
||||
@@ -96,8 +90,7 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02c674951a81881b7bc50db3b9e5efd97ac88550c7426ac548720e5057cfba515a",
|
||||
l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x02c674951a81881b7bc50db3b9e5efd97ac88550c7426ac548720e5057cfba515a", l2Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestL2TxByteParsers(t *testing.T) {
|
||||
|
||||
@@ -16,8 +16,7 @@ import (
|
||||
// EmptyBJJComp contains the 32 byte array of a empty BabyJubJub PublicKey
|
||||
// Compressed. It is a valid point in the BabyJubJub curve, so does not give
|
||||
// errors when being decompressed.
|
||||
var EmptyBJJComp = babyjub.PublicKeyComp([32]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
|
||||
var EmptyBJJComp = babyjub.PublicKeyComp([32]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
|
||||
|
||||
// PoolL2Tx is a struct that represents a L2Tx sent by an account to the
|
||||
// coordinator that is waiting to be forged
|
||||
@@ -101,8 +100,6 @@ func (tx *PoolL2Tx) SetType() error {
|
||||
tx.Type = TxTypeTransferToBJJ
|
||||
} else if tx.ToEthAddr != FFAddr && tx.ToEthAddr != EmptyAddr {
|
||||
tx.Type = TxTypeTransferToEthAddr
|
||||
} else {
|
||||
return tracerr.Wrap(errors.New("malformed transaction"))
|
||||
}
|
||||
} else {
|
||||
return tracerr.Wrap(errors.New("malformed transaction"))
|
||||
@@ -306,8 +303,10 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(e1B[0:5], amountFloat40Bytes)
|
||||
copy(e1B[5:25], tx.ToEthAddr[:])
|
||||
toEthAddr := EthAddrToBigInt(tx.ToEthAddr)
|
||||
copy(e1B[5:25], toEthAddr.Bytes())
|
||||
e1 := new(big.Int).SetBytes(e1B[:])
|
||||
|
||||
rqToEthAddr := EthAddrToBigInt(tx.RqToEthAddr)
|
||||
|
||||
_, toBJJY := babyjub.UnpackSignY(tx.ToBJJ)
|
||||
@@ -319,8 +318,7 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
||||
|
||||
_, rqToBJJY := babyjub.UnpackSignY(tx.RqToBJJ)
|
||||
|
||||
return poseidon.Hash([]*big.Int{toCompressedData, e1, toBJJY, rqTxCompressedDataV2,
|
||||
rqToEthAddr, rqToBJJY})
|
||||
return poseidon.Hash([]*big.Int{toCompressedData, e1, toBJJY, rqTxCompressedDataV2, rqToEthAddr, rqToBJJY})
|
||||
}
|
||||
|
||||
// VerifySignature returns true if the signature verification is correct for the given PublicKeyComp
|
||||
|
||||
@@ -21,20 +21,17 @@ func TestNewPoolL2Tx(t *testing.T) {
|
||||
}
|
||||
poolL2Tx, err := NewPoolL2Tx(poolL2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e",
|
||||
poolL2Tx.TxID.String())
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e", poolL2Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestTxCompressedDataAndTxCompressedDataV2JSVectors(t *testing.T) {
|
||||
// test vectors values generated from javascript implementation
|
||||
var skPositive babyjub.PrivateKey // 'Positive' refers to the sign
|
||||
_, err := hex.Decode(skPositive[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(skPositive[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
|
||||
var skNegative babyjub.PrivateKey // 'Negative' refers to the sign
|
||||
_, err = hex.Decode(skNegative[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090002"))
|
||||
_, err = hex.Decode(skNegative[:], []byte("0001020304050607080900010203040506070809000102030405060708090002"))
|
||||
assert.NoError(t, err)
|
||||
|
||||
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||
@@ -126,8 +123,7 @@ func TestTxCompressedDataAndTxCompressedDataV2JSVectors(t *testing.T) {
|
||||
|
||||
func TestRqTxCompressedDataV2(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
RqFromIdx: 7,
|
||||
@@ -146,8 +142,7 @@ func TestRqTxCompressedDataV2(t *testing.T) {
|
||||
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
||||
assert.True(t, ok)
|
||||
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
||||
assert.Equal(t, "010c000000000b0000000a0000000009000000000008000000000007",
|
||||
hex.EncodeToString(txCompressedData.Bytes()))
|
||||
assert.Equal(t, "010c000000000b0000000a0000000009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
}
|
||||
|
||||
func TestHashToSign(t *testing.T) {
|
||||
@@ -162,15 +157,13 @@ func TestHashToSign(t *testing.T) {
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0b8abaf6b7933464e4450df2514da8b72606c02bf7f89bf6e54816fbda9d9d57",
|
||||
hex.EncodeToString(toSign.Bytes()))
|
||||
assert.Equal(t, "2d49ce1d4136e06f64e3eb1f79a346e6ee3e93ceeac909a57806a8d87005c263", hex.EncodeToString(toSign.Bytes()))
|
||||
}
|
||||
|
||||
func TestVerifyTxSignature(t *testing.T) {
|
||||
chainID := uint16(0)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 2,
|
||||
@@ -184,49 +177,18 @@ func TestVerifyTxSignature(t *testing.T) {
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"3144939470626721092564692894890580265754250231349521601298746071096761507003",
|
||||
toSign.String())
|
||||
assert.Equal(t, "1571327027383224465388301747239444557034990637650927918405777653988509342917", toSign.String())
|
||||
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
tx.Signature = sig.Compress()
|
||||
assert.True(t, tx.VerifySignature(chainID, sk.Public().Compress()))
|
||||
}
|
||||
|
||||
func TestVerifyTxSignatureEthAddrWith0(t *testing.T) {
|
||||
chainID := uint16(5)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("02f0b4f87065af3797aaaf934e8b5c31563c17f2272fa71bd0146535bfbb4184"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 10659,
|
||||
ToIdx: 0,
|
||||
ToEthAddr: ethCommon.HexToAddress("0x0004308BD15Ead4F1173624dC289DBdcC806a309"),
|
||||
Amount: big.NewInt(5000),
|
||||
TokenID: 0,
|
||||
Nonce: 946,
|
||||
Fee: 231,
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
assert.Equal(t,
|
||||
"f208b8298d5f37148ac3c0c03703272ea47b9f836851bcf8dd5f7e4e3b336ca1d2f6e92ad85dc25f174daf7a0abfd5f71dead3f059b783f4c4b2f56a18a47000",
|
||||
sig.Compress().String(),
|
||||
)
|
||||
tx.Signature = sig.Compress()
|
||||
assert.True(t, tx.VerifySignature(chainID, sk.Public().Compress()))
|
||||
}
|
||||
|
||||
func TestDecompressEmptyBJJComp(t *testing.T) {
|
||||
pkComp := EmptyBJJComp
|
||||
pk, err := pkComp.Decompress()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t,
|
||||
"2957874849018779266517920829765869116077630550401372566248359756137677864698",
|
||||
pk.X.String())
|
||||
assert.Equal(t, "2957874849018779266517920829765869116077630550401372566248359756137677864698", pk.X.String())
|
||||
assert.Equal(t, "0", pk.Y.String())
|
||||
}
|
||||
|
||||
|
||||
@@ -15,9 +15,8 @@ const tokenIDBytesLen = 4
|
||||
|
||||
// Token is a struct that represents an Ethereum token that is supported in Hermez network
|
||||
type Token struct {
|
||||
TokenID TokenID `json:"id" meddler:"token_id"`
|
||||
// EthBlockNum indicates the Ethereum block number in which this token was registered
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
TokenID TokenID `json:"id" meddler:"token_id"`
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"` // Ethereum block number in which this token was registered
|
||||
EthAddr ethCommon.Address `json:"ethereumAddress" meddler:"eth_addr"`
|
||||
Name string `json:"name" meddler:"name"`
|
||||
Symbol string `json:"symbol" meddler:"symbol"`
|
||||
@@ -49,8 +48,7 @@ func (t TokenID) BigInt() *big.Int {
|
||||
// TokenIDFromBytes returns TokenID from a byte array
|
||||
func TokenIDFromBytes(b []byte) (TokenID, error) {
|
||||
if len(b) != tokenIDBytesLen {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse TokenID, bytes len %d, expected 4",
|
||||
len(b)))
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse TokenID, bytes len %d, expected 4", len(b)))
|
||||
}
|
||||
tid := binary.BigEndian.Uint32(b[:4])
|
||||
return TokenID(tid), nil
|
||||
|
||||
54
common/tx.go
54
common/tx.go
@@ -15,12 +15,12 @@ import (
|
||||
)
|
||||
|
||||
const (
|
||||
// TxIDPrefixL1UserTx is the prefix that determines that the TxID is for
|
||||
// a L1UserTx
|
||||
// TXIDPrefixL1UserTx is the prefix that determines that the TxID is
|
||||
// for a L1UserTx
|
||||
//nolinter:gomnd
|
||||
TxIDPrefixL1UserTx = byte(0)
|
||||
|
||||
// TxIDPrefixL1CoordTx is the prefix that determines that the TxID is
|
||||
// TXIDPrefixL1CoordTx is the prefix that determines that the TxID is
|
||||
// for a L1CoordinatorTx
|
||||
//nolinter:gomnd
|
||||
TxIDPrefixL1CoordTx = byte(1)
|
||||
@@ -51,8 +51,7 @@ func (txid *TxID) Scan(src interface{}) error {
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan %T into TxID", src))
|
||||
}
|
||||
if len(srcB) != TxIDLen {
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan []byte of len %d into TxID, need %d",
|
||||
len(srcB), TxIDLen))
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan []byte of len %d into TxID, need %d", len(srcB), TxIDLen))
|
||||
}
|
||||
copy(txid[:], srcB)
|
||||
return nil
|
||||
@@ -88,7 +87,7 @@ func (txid TxID) MarshalText() ([]byte, error) {
|
||||
return []byte(txid.String()), nil
|
||||
}
|
||||
|
||||
// UnmarshalText unmarshalls a TxID
|
||||
// UnmarshalText unmarshals a TxID
|
||||
func (txid *TxID) UnmarshalText(data []byte) error {
|
||||
idStr := string(data)
|
||||
id, err := NewTxIDFromString(idStr)
|
||||
@@ -103,15 +102,13 @@ func (txid *TxID) UnmarshalText(data []byte) error {
|
||||
type TxType string
|
||||
|
||||
const (
|
||||
// TxTypeExit represents L2->L1 token transfer. A leaf for this account appears in the exit
|
||||
// tree of the block
|
||||
// TxTypeExit represents L2->L1 token transfer. A leaf for this account appears in the exit tree of the block
|
||||
TxTypeExit TxType = "Exit"
|
||||
// TxTypeTransfer represents L2->L2 token transfer
|
||||
TxTypeTransfer TxType = "Transfer"
|
||||
// TxTypeDeposit represents L1->L2 transfer
|
||||
TxTypeDeposit TxType = "Deposit"
|
||||
// TxTypeCreateAccountDeposit represents creation of a new leaf in the state tree
|
||||
// (newAcconut) + L1->L2 transfer
|
||||
// TxTypeCreateAccountDeposit represents creation of a new leaf in the state tree (newAcconut) + L1->L2 transfer
|
||||
TxTypeCreateAccountDeposit TxType = "CreateAccountDeposit"
|
||||
// TxTypeCreateAccountDepositTransfer represents L1->L2 transfer + L2->L2 transfer
|
||||
TxTypeCreateAccountDepositTransfer TxType = "CreateAccountDepositTransfer"
|
||||
@@ -127,31 +124,24 @@ const (
|
||||
TxTypeTransferToBJJ TxType = "TransferToBJJ"
|
||||
)
|
||||
|
||||
// Tx is a struct used by the TxSelector & BatchBuilder as a generic type generated from L1Tx &
|
||||
// PoolL2Tx
|
||||
// Tx is a struct used by the TxSelector & BatchBuilder as a generic type generated from L1Tx & PoolL2Tx
|
||||
type Tx struct {
|
||||
// Generic
|
||||
IsL1 bool `meddler:"is_l1"`
|
||||
TxID TxID `meddler:"id"`
|
||||
Type TxType `meddler:"type"`
|
||||
Position int `meddler:"position"`
|
||||
FromIdx Idx `meddler:"from_idx"`
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
AmountFloat float64 `meddler:"amount_f"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
USD *float64 `meddler:"amount_usd"`
|
||||
// BatchNum in which this tx was forged. If the tx is L2, this must be != 0
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
// Ethereum Block Number in which this L1Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
IsL1 bool `meddler:"is_l1"`
|
||||
TxID TxID `meddler:"id"`
|
||||
Type TxType `meddler:"type"`
|
||||
Position int `meddler:"position"`
|
||||
FromIdx Idx `meddler:"from_idx"`
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
AmountFloat float64 `meddler:"amount_f"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
USD *float64 `meddler:"amount_usd"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"` // batchNum in which this tx was forged. If the tx is L2, this must be != 0
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum Block Number in which this L1Tx was added to the queue
|
||||
// L1
|
||||
// ToForgeL1TxsNum in which the tx was forged / will be forged
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"`
|
||||
// UserOrigin is set to true if the tx was originated by a user, false if it was aoriginated
|
||||
// by a coordinator. Note that this differ from the spec for implementation simplification
|
||||
// purpposes
|
||||
UserOrigin *bool `meddler:"user_origin"`
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"` // toForgeL1TxsNum in which the tx was forged / will be forged
|
||||
UserOrigin *bool `meddler:"user_origin"` // true if the tx was originated by a user, false if it was aoriginated by a coordinator. Note that this differ from the spec for implementation simplification purpposes
|
||||
FromEthAddr ethCommon.Address `meddler:"from_eth_addr"`
|
||||
FromBJJ babyjub.PublicKeyComp `meddler:"from_bjj"`
|
||||
DepositAmount *big.Int `meddler:"deposit_amount,bigintnull"`
|
||||
|
||||
@@ -21,10 +21,8 @@ func TestSignatureConstant(t *testing.T) {
|
||||
func TestTxIDScannerValue(t *testing.T) {
|
||||
txid0 := &TxID{}
|
||||
txid1 := &TxID{}
|
||||
txid0B := [TxIDLen]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2,
|
||||
3, 4, 5, 6, 7, 8, 9, 0, 1, 2}
|
||||
txid1B := [TxIDLen]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
||||
txid0B := [TxIDLen]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2}
|
||||
txid1B := [TxIDLen]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
||||
|
||||
copy(txid0[:], txid0B[:])
|
||||
copy(txid1[:], txid1B[:])
|
||||
|
||||
@@ -21,23 +21,16 @@ func TestBJJFromStringWithChecksum(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
|
||||
// expected values computed with js implementation
|
||||
assert.Equal(t,
|
||||
"2492816973395423007340226948038371729989170225696553239457870892535792679622",
|
||||
pk.X.String())
|
||||
assert.Equal(t,
|
||||
"15238403086306505038849621710779816852318505119327426213168494964113886299863",
|
||||
pk.Y.String())
|
||||
assert.Equal(t, "2492816973395423007340226948038371729989170225696553239457870892535792679622", pk.X.String())
|
||||
assert.Equal(t, "15238403086306505038849621710779816852318505119327426213168494964113886299863", pk.Y.String())
|
||||
}
|
||||
|
||||
func TestRmEndingZeroes(t *testing.T) {
|
||||
s0, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000000")
|
||||
s0, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000000")
|
||||
require.NoError(t, err)
|
||||
s1, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000001")
|
||||
s1, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000001")
|
||||
require.NoError(t, err)
|
||||
s2, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000002")
|
||||
s2, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000002")
|
||||
require.NoError(t, err)
|
||||
|
||||
// expect cropped last zeroes
|
||||
|
||||
20
common/zk.go
20
common/zk.go
@@ -1,4 +1,4 @@
|
||||
// Package common zk.go contains all the common data structures used at the
|
||||
// Package common contains all the common data structures used at the
|
||||
// hermez-node, zk.go contains the zkSnark inputs used to generate the proof
|
||||
package common
|
||||
|
||||
@@ -67,7 +67,7 @@ type ZKInputs struct {
|
||||
|
||||
// accumulate fees
|
||||
// FeePlanTokens contains all the tokenIDs for which the fees are being
|
||||
// accumulated and those fees accumulated will be paid to the FeeIdxs
|
||||
// accumulated and those fees accoumulated will be paid to the FeeIdxs
|
||||
// array. The order of FeeIdxs & FeePlanTokens & State3 must match.
|
||||
// Coordinator fees are processed correlated such as:
|
||||
// [FeePlanTokens[i], FeeIdxs[i]]
|
||||
@@ -130,8 +130,8 @@ type ZKInputs struct {
|
||||
RqOffset []*big.Int `json:"rqOffset"` // uint8 (max 3 bits), len: [maxTx]
|
||||
|
||||
// transaction L2 request data
|
||||
// RqTxCompressedDataV2 big.Int (max 251 bits), len: [maxTx]
|
||||
RqTxCompressedDataV2 []*big.Int `json:"rqTxCompressedDataV2"`
|
||||
// RqTxCompressedDataV2
|
||||
RqTxCompressedDataV2 []*big.Int `json:"rqTxCompressedDataV2"` // big.Int (max 251 bits), len: [maxTx]
|
||||
// RqToEthAddr
|
||||
RqToEthAddr []*big.Int `json:"rqToEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||
// RqToBJJAy
|
||||
@@ -301,8 +301,7 @@ func (z ZKInputs) MarshalJSON() ([]byte, error) {
|
||||
}
|
||||
|
||||
// NewZKInputs returns a pointer to an initialized struct of ZKInputs
|
||||
func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32,
|
||||
currentNumBatch *big.Int) *ZKInputs {
|
||||
func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32, currentNumBatch *big.Int) *ZKInputs {
|
||||
zki := &ZKInputs{}
|
||||
zki.Metadata.MaxFeeIdxs = maxFeeIdxs
|
||||
zki.Metadata.MaxLevels = uint32(48) //nolint:gomnd
|
||||
@@ -481,7 +480,7 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
||||
b = append(b, newExitRoot...)
|
||||
|
||||
// [MAX_L1_TX * (2 * MAX_NLEVELS + 528) bits] L1TxsData
|
||||
l1TxDataLen := (2*z.Metadata.MaxLevels + 528) //nolint:gomnd
|
||||
l1TxDataLen := (2*z.Metadata.MaxLevels + 528)
|
||||
l1TxsDataLen := (z.Metadata.MaxL1Tx * l1TxDataLen)
|
||||
l1TxsData := make([]byte, l1TxsDataLen/8) //nolint:gomnd
|
||||
for i := 0; i < len(z.Metadata.L1TxsData); i++ {
|
||||
@@ -507,14 +506,11 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
||||
l2TxsData = append(l2TxsData, z.Metadata.L2TxsData[i]...)
|
||||
}
|
||||
if len(l2TxsData) > int(expectedL2TxsDataLen) {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("len(l2TxsData): %d, expected: %d",
|
||||
len(l2TxsData), expectedL2TxsDataLen))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("len(l2TxsData): %d, expected: %d", len(l2TxsData), expectedL2TxsDataLen))
|
||||
}
|
||||
|
||||
b = append(b, l2TxsData...)
|
||||
l2TxsPadding := make([]byte,
|
||||
(int(z.Metadata.MaxTx)-len(z.Metadata.L1TxsDataAvailability)-
|
||||
len(z.Metadata.L2TxsData))*int(l2TxDataLen)/8) //nolint:gomnd
|
||||
l2TxsPadding := make([]byte, (int(z.Metadata.MaxTx)-len(z.Metadata.L1TxsDataAvailability)-len(z.Metadata.L2TxsData))*int(l2TxDataLen)/8) //nolint:gomnd
|
||||
b = append(b, l2TxsPadding...)
|
||||
|
||||
// [NLevels * MAX_TOKENS_FEE bits] feeTxsData
|
||||
|
||||
151
config/config.go
151
config/config.go
@@ -9,7 +9,6 @@ import (
|
||||
"github.com/BurntSushi/toml"
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/priceupdater"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
"gopkg.in/go-playground/validator.v9"
|
||||
@@ -81,7 +80,7 @@ type Coordinator struct {
|
||||
// checking the next block), used to decide when to stop scheduling new
|
||||
// batches (by stopping the pipeline).
|
||||
// For example, if we are at block 10 and ScheduleBatchBlocksAheadCheck
|
||||
// is 5, even though at block 11 we canForge, the pipeline will be
|
||||
// is 5, eventhough at block 11 we canForge, the pipeline will be
|
||||
// stopped if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// scheduling a batch and having it mined.
|
||||
@@ -91,7 +90,7 @@ type Coordinator struct {
|
||||
// from the next block; used to decide when to stop sending batches to
|
||||
// the smart contract.
|
||||
// For example, if we are at block 10 and SendBatchBlocksMarginCheck is
|
||||
// 5, even though at block 11 we canForge, the batch will be discarded
|
||||
// 5, eventhough at block 11 we canForge, the batch will be discarded
|
||||
// if we can't forge at block 15.
|
||||
SendBatchBlocksMarginCheck int64
|
||||
// ProofServerPollInterval is the waiting interval between polling the
|
||||
@@ -109,20 +108,6 @@ type Coordinator struct {
|
||||
// to 0s, the coordinator will continuously forge even if the batches
|
||||
// are empty.
|
||||
ForgeNoTxsDelay Duration `validate:"-"`
|
||||
// MustForgeAtSlotDeadline enables the coordinator to forge slots if
|
||||
// the empty slots reach the slot deadline.
|
||||
MustForgeAtSlotDeadline bool
|
||||
// IgnoreSlotCommitment disables forcing the coordinator to forge a
|
||||
// slot immediately when the slot is not committed. If set to false,
|
||||
// the coordinator will immediately forge a batch at the beginning of a
|
||||
// slot if it's the slot winner.
|
||||
IgnoreSlotCommitment bool
|
||||
// ForgeOncePerSlotIfTxs will make the coordinator forge at most one
|
||||
// batch per slot, only if there are included txs in that batch, or
|
||||
// pending l1UserTxs in the smart contract. Setting this parameter
|
||||
// overrides `ForgeDelay`, `ForgeNoTxsDelay`, `MustForgeAtSlotDeadline`
|
||||
// and `IgnoreSlotCommitment`.
|
||||
ForgeOncePerSlotIfTxs bool
|
||||
// SyncRetryInterval is the waiting interval between calls to the main
|
||||
// handler of a synced block after an error
|
||||
SyncRetryInterval Duration `validate:"required"`
|
||||
@@ -148,7 +133,7 @@ type Coordinator struct {
|
||||
// L2Txs is reached, L2Txs older than TTL will be deleted.
|
||||
TTL Duration `validate:"required"`
|
||||
// PurgeBatchDelay is the delay between batches to purge
|
||||
// outdated transactions. Outdated L2Txs are those that have
|
||||
// outdated transactions. Oudated L2Txs are those that have
|
||||
// been forged or marked as invalid for longer than the
|
||||
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||
// for longer than TTL once there are MaxTxs.
|
||||
@@ -158,7 +143,7 @@ type Coordinator struct {
|
||||
// nonce.
|
||||
InvalidateBatchDelay int64 `validate:"required"`
|
||||
// PurgeBlockDelay is the delay between blocks to purge
|
||||
// outdated transactions. Outdated L2Txs are those that have
|
||||
// outdated transactions. Oudated L2Txs are those that have
|
||||
// been forged or marked as invalid for longer than the
|
||||
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||
// for longer than TTL once there are MaxTxs.
|
||||
@@ -190,7 +175,7 @@ type Coordinator struct {
|
||||
MaxGasPrice *big.Int `validate:"required"`
|
||||
// GasPriceIncPerc is the percentage increase of gas price set
|
||||
// in an ethereum transaction from the suggested gas price by
|
||||
// the ethereum node
|
||||
// the ehtereum node
|
||||
GasPriceIncPerc int64
|
||||
// CheckLoopInterval is the waiting interval between receipt
|
||||
// checks of ethereum transactions in the TxManager
|
||||
@@ -234,9 +219,28 @@ type Coordinator struct {
|
||||
}
|
||||
}
|
||||
|
||||
// PostgreSQL is the postgreSQL configuration parameters. It's possible to use
|
||||
// diferentiated SQL connections for read/write. If the read configuration is
|
||||
// not provided, the write one it's going to be used for both reads and writes
|
||||
// NodeAPI specifies the configuration parameters of the API
|
||||
type NodeAPI struct {
|
||||
// Address where the API will listen if set
|
||||
Address string
|
||||
// Explorer enables the Explorer API endpoints
|
||||
Explorer bool
|
||||
// UpdateMetricsInterval is the interval between updates of the
|
||||
// API metrics
|
||||
UpdateMetricsInterval Duration
|
||||
// UpdateRecommendedFeeInterval is the interval between updates of the
|
||||
// recommended fees
|
||||
UpdateRecommendedFeeInterval Duration
|
||||
// Maximum concurrent connections allowed between API and SQL
|
||||
MaxSQLConnections int `validate:"required"`
|
||||
// SQLConnectionTimeout is the maximum amount of time that an API request
|
||||
// can wait to stablish a SQL connection
|
||||
SQLConnectionTimeout Duration
|
||||
}
|
||||
|
||||
// It's possible to use diferentiated SQL connections for read/write.
|
||||
// If the read configuration is not provided, the write one it's going to be used
|
||||
// for both reads and writes
|
||||
type PostgreSQL struct {
|
||||
// Port of the PostgreSQL write server
|
||||
PortWrite int `validate:"required"`
|
||||
@@ -277,15 +281,11 @@ type NodeDebug struct {
|
||||
type Node struct {
|
||||
PriceUpdater struct {
|
||||
// Interval between price updater calls
|
||||
Interval Duration `validate:"required"`
|
||||
// URLBitfinexV2 is the URL of bitfinex V2 API
|
||||
URLBitfinexV2 string `validate:"required"`
|
||||
// URLCoinGeckoV3 is the URL of coingecko V3 API
|
||||
URLCoinGeckoV3 string `validate:"required"`
|
||||
// DefaultUpdateMethod to get token prices
|
||||
DefaultUpdateMethod priceupdater.UpdateMethodType `validate:"required"`
|
||||
// TokensConfig to specify how each token get it's price updated
|
||||
TokensConfig []priceupdater.TokenConfig
|
||||
Interval Duration `valudate:"required"`
|
||||
// URL of the token prices provider
|
||||
URL string `valudate:"required"`
|
||||
// Type of the API of the token prices provider
|
||||
Type string `valudate:"required"`
|
||||
} `validate:"required"`
|
||||
StateDB struct {
|
||||
// Path where the synchronizer StateDB is stored
|
||||
@@ -324,60 +324,31 @@ type Node struct {
|
||||
// TokenHEZ address
|
||||
TokenHEZName string `validate:"required"`
|
||||
} `validate:"required"`
|
||||
// API specifies the configuration parameters of the API
|
||||
API struct {
|
||||
// Address where the API will listen if set
|
||||
Address string
|
||||
// Explorer enables the Explorer API endpoints
|
||||
Explorer bool
|
||||
// UpdateMetricsInterval is the interval between updates of the
|
||||
// API metrics
|
||||
UpdateMetricsInterval Duration
|
||||
// UpdateRecommendedFeeInterval is the interval between updates of the
|
||||
// recommended fees
|
||||
UpdateRecommendedFeeInterval Duration
|
||||
// Maximum concurrent connections allowed between API and SQL
|
||||
MaxSQLConnections int `validate:"required"`
|
||||
// SQLConnectionTimeout is the maximum amount of time that an API request
|
||||
// can wait to stablish a SQL connection
|
||||
SQLConnectionTimeout Duration
|
||||
} `validate:"required"`
|
||||
API NodeAPI `validate:"required"`
|
||||
Debug NodeDebug `validate:"required"`
|
||||
Coordinator Coordinator `validate:"-"`
|
||||
}
|
||||
|
||||
// APIServer is the api server configuration parameters
|
||||
type APIServer struct {
|
||||
// NodeAPI specifies the configuration parameters of the API
|
||||
API struct {
|
||||
// Address where the API will listen if set
|
||||
Address string `validate:"required"`
|
||||
// Explorer enables the Explorer API endpoints
|
||||
Explorer bool
|
||||
// Maximum concurrent connections allowed between API and SQL
|
||||
MaxSQLConnections int `validate:"required"`
|
||||
// SQLConnectionTimeout is the maximum amount of time that an API request
|
||||
// can wait to stablish a SQL connection
|
||||
SQLConnectionTimeout Duration
|
||||
} `validate:"required"`
|
||||
API NodeAPI `validate:"required"`
|
||||
PostgreSQL PostgreSQL `validate:"required"`
|
||||
Coordinator struct {
|
||||
API struct {
|
||||
// Coordinator enables the coordinator API endpoints
|
||||
Coordinator bool
|
||||
} `validate:"required"`
|
||||
L2DB struct {
|
||||
// MaxTxs is the maximum number of pending L2Txs that can be
|
||||
// stored in the pool. Once this number of pending L2Txs is
|
||||
// reached, inserts to the pool will be denied until some of
|
||||
// the pending txs are forged.
|
||||
MaxTxs uint32 `validate:"required"`
|
||||
// MinFeeUSD is the minimum fee in USD that a tx must pay in
|
||||
// order to be accepted into the pool. Txs with lower than
|
||||
// minimum fee will be rejected at the API level.
|
||||
MinFeeUSD float64
|
||||
} `validate:"required"`
|
||||
}
|
||||
} `validate:"required"`
|
||||
L2DB struct {
|
||||
// MaxTxs is the maximum number of pending L2Txs that can be
|
||||
// stored in the pool. Once this number of pending L2Txs is
|
||||
// reached, inserts to the pool will be denied until some of
|
||||
// the pending txs are forged.
|
||||
MaxTxs uint32 `validate:"required"`
|
||||
// MinFeeUSD is the minimum fee in USD that a tx must pay in
|
||||
// order to be accepted into the pool. Txs with lower than
|
||||
// minimum fee will be rejected at the API level.
|
||||
MinFeeUSD float64
|
||||
} `validate:"required"`
|
||||
Debug NodeDebug `validate:"required"`
|
||||
}
|
||||
|
||||
@@ -394,8 +365,8 @@ func Load(path string, cfg interface{}) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// LoadNode loads the Node configuration from path.
|
||||
func LoadNode(path string, coordinator bool) (*Node, error) {
|
||||
// LoadCoordinator loads the Coordinator configuration from path.
|
||||
func LoadCoordinator(path string) (*Node, error) {
|
||||
var cfg Node
|
||||
if err := Load(path, &cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
|
||||
@@ -404,16 +375,27 @@ func LoadNode(path string, coordinator bool) (*Node, error) {
|
||||
if err := validate.Struct(cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
if coordinator {
|
||||
if err := validate.Struct(cfg.Coordinator); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
if err := validate.Struct(cfg.Coordinator); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
return &cfg, nil
|
||||
}
|
||||
|
||||
// LoadNode loads the Node configuration from path.
|
||||
func LoadNode(path string) (*Node, error) {
|
||||
var cfg Node
|
||||
if err := Load(path, &cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
|
||||
}
|
||||
validate := validator.New()
|
||||
if err := validate.Struct(cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
return &cfg, nil
|
||||
}
|
||||
|
||||
// LoadAPIServer loads the APIServer configuration from path.
|
||||
func LoadAPIServer(path string, coordinator bool) (*APIServer, error) {
|
||||
func LoadAPIServer(path string) (*APIServer, error) {
|
||||
var cfg APIServer
|
||||
if err := Load(path, &cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error loading apiServer configuration file: %w", err))
|
||||
@@ -422,10 +404,5 @@ func LoadAPIServer(path string, coordinator bool) (*APIServer, error) {
|
||||
if err := validate.Struct(cfg); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
if coordinator {
|
||||
if err := validate.Struct(cfg.Coordinator); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
|
||||
}
|
||||
}
|
||||
return &cfg, nil
|
||||
}
|
||||
|
||||
@@ -8,7 +8,6 @@ import (
|
||||
"path"
|
||||
"time"
|
||||
|
||||
"github.com/ethereum/go-ethereum/accounts/abi/bind"
|
||||
"github.com/ethereum/go-ethereum/core/types"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
@@ -85,15 +84,15 @@ type BatchInfo struct {
|
||||
PublicInputs []*big.Int
|
||||
L1Batch bool
|
||||
VerifierIdx uint8
|
||||
L1UserTxs []common.L1Tx
|
||||
L1UserTxsExtra []common.L1Tx
|
||||
L1CoordTxs []common.L1Tx
|
||||
L1CoordinatorTxsAuths [][]byte
|
||||
L2Txs []common.L2Tx
|
||||
CoordIdxs []common.Idx
|
||||
ForgeBatchArgs *eth.RollupForgeBatchArgs
|
||||
Auth *bind.TransactOpts `json:"-"`
|
||||
EthTx *types.Transaction
|
||||
EthTxErr error
|
||||
// FeesInfo
|
||||
EthTx *types.Transaction
|
||||
EthTxErr error
|
||||
// SendTimestamp the time of batch sent to ethereum
|
||||
SendTimestamp time.Time
|
||||
Receipt *types.Receipt
|
||||
|
||||
@@ -24,8 +24,9 @@ import (
|
||||
)
|
||||
|
||||
var (
|
||||
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
||||
errSkipBatchByPolicy = fmt.Errorf("skip batch by policy")
|
||||
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
||||
errForgeNoTxsBeforeDelay = fmt.Errorf("no txs to forge and we haven't reached the forge no txs delay")
|
||||
errForgeBeforeDelay = fmt.Errorf("we haven't reached the forge delay")
|
||||
)
|
||||
|
||||
const (
|
||||
@@ -52,7 +53,7 @@ type Config struct {
|
||||
// checking the next block), used to decide when to stop scheduling new
|
||||
// batches (by stopping the pipeline).
|
||||
// For example, if we are at block 10 and ScheduleBatchBlocksAheadCheck
|
||||
// is 5, even though at block 11 we canForge, the pipeline will be
|
||||
// is 5, eventhough at block 11 we canForge, the pipeline will be
|
||||
// stopped if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// scheduling a batch and having it mined.
|
||||
@@ -62,7 +63,7 @@ type Config struct {
|
||||
// from the next block; used to decide when to stop sending batches to
|
||||
// the smart contract.
|
||||
// For example, if we are at block 10 and SendBatchBlocksMarginCheck is
|
||||
// 5, even though at block 11 we canForge, the batch will be discarded
|
||||
// 5, eventhough at block 11 we canForge, the batch will be discarded
|
||||
// if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// sending a batch and having it mined.
|
||||
@@ -82,20 +83,6 @@ type Config struct {
|
||||
// to 0s, the coordinator will continuously forge even if the batches
|
||||
// are empty.
|
||||
ForgeNoTxsDelay time.Duration
|
||||
// MustForgeAtSlotDeadline enables the coordinator to forge slots if
|
||||
// the empty slots reach the slot deadline.
|
||||
MustForgeAtSlotDeadline bool
|
||||
// IgnoreSlotCommitment disables forcing the coordinator to forge a
|
||||
// slot immediately when the slot is not committed. If set to false,
|
||||
// the coordinator will immediately forge a batch at the beginning of
|
||||
// a slot if it's the slot winner.
|
||||
IgnoreSlotCommitment bool
|
||||
// ForgeOncePerSlotIfTxs will make the coordinator forge at most one
|
||||
// batch per slot, only if there are included txs in that batch, or
|
||||
// pending l1UserTxs in the smart contract. Setting this parameter
|
||||
// overrides `ForgeDelay`, `ForgeNoTxsDelay`, `MustForgeAtSlotDeadline`
|
||||
// and `IgnoreSlotCommitment`.
|
||||
ForgeOncePerSlotIfTxs bool
|
||||
// SyncRetryInterval is the waiting interval between calls to the main
|
||||
// handler of a synced block after an error
|
||||
SyncRetryInterval time.Duration
|
||||
@@ -157,8 +144,8 @@ type Coordinator struct {
|
||||
pipelineNum int // Pipeline sequential number. The first pipeline is 1
|
||||
pipelineFromBatch fromBatch // batch from which we started the pipeline
|
||||
provers []prover.Client
|
||||
consts common.SCConsts
|
||||
vars common.SCVariables
|
||||
consts synchronizer.SCConsts
|
||||
vars synchronizer.SCVariables
|
||||
stats synchronizer.Stats
|
||||
started bool
|
||||
|
||||
@@ -288,13 +275,13 @@ type MsgSyncBlock struct {
|
||||
Batches []common.BatchData
|
||||
// Vars contains each Smart Contract variables if they are updated, or
|
||||
// nil if they haven't changed.
|
||||
Vars common.SCVariablesPtr
|
||||
Vars synchronizer.SCVariablesPtr
|
||||
}
|
||||
|
||||
// MsgSyncReorg indicates a reorg
|
||||
type MsgSyncReorg struct {
|
||||
Stats synchronizer.Stats
|
||||
Vars common.SCVariablesPtr
|
||||
Vars synchronizer.SCVariablesPtr
|
||||
}
|
||||
|
||||
// MsgStopPipeline indicates a signal to reset the pipeline
|
||||
@@ -313,7 +300,7 @@ func (c *Coordinator) SendMsg(ctx context.Context, msg interface{}) {
|
||||
}
|
||||
}
|
||||
|
||||
func updateSCVars(vars *common.SCVariables, update common.SCVariablesPtr) {
|
||||
func updateSCVars(vars *synchronizer.SCVariables, update synchronizer.SCVariablesPtr) {
|
||||
if update.Rollup != nil {
|
||||
vars.Rollup = *update.Rollup
|
||||
}
|
||||
@@ -325,13 +312,12 @@ func updateSCVars(vars *common.SCVariables, update common.SCVariablesPtr) {
|
||||
}
|
||||
}
|
||||
|
||||
func (c *Coordinator) syncSCVars(vars common.SCVariablesPtr) {
|
||||
func (c *Coordinator) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
updateSCVars(&c.vars, vars)
|
||||
}
|
||||
|
||||
func canForge(auctionConstants *common.AuctionConstants, auctionVars *common.AuctionVariables,
|
||||
currentSlot *common.Slot, nextSlot *common.Slot, addr ethCommon.Address, blockNum int64,
|
||||
mustForgeAtDeadline bool) bool {
|
||||
currentSlot *common.Slot, nextSlot *common.Slot, addr ethCommon.Address, blockNum int64) bool {
|
||||
if blockNum < auctionConstants.GenesisBlockNum {
|
||||
log.Infow("canForge: requested blockNum is < genesis", "blockNum", blockNum,
|
||||
"genesis", auctionConstants.GenesisBlockNum)
|
||||
@@ -356,7 +342,7 @@ func canForge(auctionConstants *common.AuctionConstants, auctionVars *common.Auc
|
||||
"block", blockNum)
|
||||
anyoneForge = true
|
||||
}
|
||||
if slot.Forger == addr || (anyoneForge && mustForgeAtDeadline) {
|
||||
if slot.Forger == addr || anyoneForge {
|
||||
return true
|
||||
}
|
||||
log.Debugw("canForge: can't forge", "slot.Forger", slot.Forger)
|
||||
@@ -366,14 +352,14 @@ func canForge(auctionConstants *common.AuctionConstants, auctionVars *common.Auc
|
||||
func (c *Coordinator) canForgeAt(blockNum int64) bool {
|
||||
return canForge(&c.consts.Auction, &c.vars.Auction,
|
||||
&c.stats.Sync.Auction.CurrentSlot, &c.stats.Sync.Auction.NextSlot,
|
||||
c.cfg.ForgerAddress, blockNum, c.cfg.MustForgeAtSlotDeadline)
|
||||
c.cfg.ForgerAddress, blockNum)
|
||||
}
|
||||
|
||||
func (c *Coordinator) canForge() bool {
|
||||
blockNum := c.stats.Eth.LastBlock.Num + 1
|
||||
return canForge(&c.consts.Auction, &c.vars.Auction,
|
||||
&c.stats.Sync.Auction.CurrentSlot, &c.stats.Sync.Auction.NextSlot,
|
||||
c.cfg.ForgerAddress, blockNum, c.cfg.MustForgeAtSlotDeadline)
|
||||
c.cfg.ForgerAddress, blockNum)
|
||||
}
|
||||
|
||||
func (c *Coordinator) syncStats(ctx context.Context, stats *synchronizer.Stats) error {
|
||||
@@ -487,8 +473,7 @@ func (c *Coordinator) handleReorg(ctx context.Context, msg *MsgSyncReorg) error
|
||||
// handleStopPipeline handles stopping the pipeline. If failedBatchNum is 0,
|
||||
// the next pipeline will start from the last state of the synchronizer,
|
||||
// otherwise, it will state from failedBatchNum-1.
|
||||
func (c *Coordinator) handleStopPipeline(ctx context.Context, reason string,
|
||||
failedBatchNum common.BatchNum) error {
|
||||
func (c *Coordinator) handleStopPipeline(ctx context.Context, reason string, failedBatchNum common.BatchNum) error {
|
||||
batchNum := c.stats.Sync.LastBatch.BatchNum
|
||||
if failedBatchNum != 0 {
|
||||
batchNum = failedBatchNum - 1
|
||||
|
||||
@@ -126,8 +126,7 @@ func newTestModules(t *testing.T) modules {
|
||||
batchBuilderDBPath, err = ioutil.TempDir("", "tmpBatchBuilderDB")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, batchBuilderDBPath)
|
||||
batchBuilder, err := batchbuilder.NewBatchBuilder(batchBuilderDBPath, syncStateDB, 0,
|
||||
uint64(nLevels))
|
||||
batchBuilder, err := batchbuilder.NewBatchBuilder(batchBuilderDBPath, syncStateDB, 0, uint64(nLevels))
|
||||
assert.NoError(t, err)
|
||||
|
||||
return modules{
|
||||
@@ -159,15 +158,14 @@ func newTestCoordinator(t *testing.T, forgerAddr ethCommon.Address, ethClient *t
|
||||
deleteme = append(deleteme, debugBatchPath)
|
||||
|
||||
conf := Config{
|
||||
ForgerAddress: forgerAddr,
|
||||
ConfirmBlocks: 5,
|
||||
L1BatchTimeoutPerc: 0.5,
|
||||
EthClientAttempts: 5,
|
||||
SyncRetryInterval: 400 * time.Microsecond,
|
||||
EthClientAttemptsDelay: 100 * time.Millisecond,
|
||||
TxManagerCheckInterval: 300 * time.Millisecond,
|
||||
DebugBatchPath: debugBatchPath,
|
||||
MustForgeAtSlotDeadline: true,
|
||||
ForgerAddress: forgerAddr,
|
||||
ConfirmBlocks: 5,
|
||||
L1BatchTimeoutPerc: 0.5,
|
||||
EthClientAttempts: 5,
|
||||
SyncRetryInterval: 400 * time.Microsecond,
|
||||
EthClientAttemptsDelay: 100 * time.Millisecond,
|
||||
TxManagerCheckInterval: 300 * time.Millisecond,
|
||||
DebugBatchPath: debugBatchPath,
|
||||
Purger: PurgerCfg{
|
||||
PurgeBatchDelay: 10,
|
||||
PurgeBlockDelay: 10,
|
||||
@@ -189,12 +187,12 @@ func newTestCoordinator(t *testing.T, forgerAddr ethCommon.Address, ethClient *t
|
||||
&prover.MockClient{Delay: 400 * time.Millisecond},
|
||||
}
|
||||
|
||||
scConsts := &common.SCConsts{
|
||||
scConsts := &synchronizer.SCConsts{
|
||||
Rollup: *ethClientSetup.RollupConstants,
|
||||
Auction: *ethClientSetup.AuctionConstants,
|
||||
WDelayer: *ethClientSetup.WDelayerConstants,
|
||||
}
|
||||
initSCVars := &common.SCVariables{
|
||||
initSCVars := &synchronizer.SCVariables{
|
||||
Rollup: *ethClientSetup.RollupVariables,
|
||||
Auction: *ethClientSetup.AuctionVariables,
|
||||
WDelayer: *ethClientSetup.WDelayerVariables,
|
||||
@@ -207,7 +205,7 @@ func newTestCoordinator(t *testing.T, forgerAddr ethCommon.Address, ethClient *t
|
||||
|
||||
func newTestSynchronizer(t *testing.T, ethClient *test.Client, ethClientSetup *test.ClientSetup,
|
||||
modules modules) *synchronizer.Synchronizer {
|
||||
sync, err := synchronizer.NewSynchronizer(ethClient, modules.historyDB, modules.l2DB, modules.stateDB,
|
||||
sync, err := synchronizer.NewSynchronizer(ethClient, modules.historyDB, modules.stateDB,
|
||||
synchronizer.Config{
|
||||
StatsRefreshPeriod: 0 * time.Second,
|
||||
})
|
||||
@@ -392,10 +390,6 @@ func TestCoordCanForge(t *testing.T) {
|
||||
assert.Equal(t, true, coord.canForge())
|
||||
assert.Equal(t, true, bootCoord.canForge())
|
||||
|
||||
// Anyone can forge but the node MustForgeAtSlotDeadline as set as false
|
||||
coord.cfg.MustForgeAtSlotDeadline = false
|
||||
assert.Equal(t, false, coord.canForge())
|
||||
|
||||
// Slot 3. coordinator bid, so the winner is the coordinator
|
||||
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||
3*int64(ethClientSetup.AuctionConstants.BlocksPerSlot)
|
||||
@@ -534,7 +528,7 @@ func TestCoordinatorStress(t *testing.T) {
|
||||
coord.SendMsg(ctx, MsgSyncBlock{
|
||||
Stats: *stats,
|
||||
Batches: blockData.Rollup.Batches,
|
||||
Vars: common.SCVariablesPtr{
|
||||
Vars: synchronizer.SCVariablesPtr{
|
||||
Rollup: blockData.Rollup.Vars,
|
||||
Auction: blockData.Auction.Vars,
|
||||
WDelayer: blockData.WDelayer.Vars,
|
||||
|
||||
@@ -22,7 +22,7 @@ import (
|
||||
|
||||
type statsVars struct {
|
||||
Stats synchronizer.Stats
|
||||
Vars common.SCVariablesPtr
|
||||
Vars synchronizer.SCVariablesPtr
|
||||
}
|
||||
|
||||
type state struct {
|
||||
@@ -36,7 +36,7 @@ type state struct {
|
||||
type Pipeline struct {
|
||||
num int
|
||||
cfg Config
|
||||
consts common.SCConsts
|
||||
consts synchronizer.SCConsts
|
||||
|
||||
// state
|
||||
state state
|
||||
@@ -57,7 +57,7 @@ type Pipeline struct {
|
||||
purger *Purger
|
||||
|
||||
stats synchronizer.Stats
|
||||
vars common.SCVariables
|
||||
vars synchronizer.SCVariables
|
||||
statsVarsCh chan statsVars
|
||||
|
||||
ctx context.Context
|
||||
@@ -90,7 +90,7 @@ func NewPipeline(ctx context.Context,
|
||||
coord *Coordinator,
|
||||
txManager *TxManager,
|
||||
provers []prover.Client,
|
||||
scConsts *common.SCConsts,
|
||||
scConsts *synchronizer.SCConsts,
|
||||
) (*Pipeline, error) {
|
||||
proversPool := NewProversPool(len(provers))
|
||||
proversPoolSize := 0
|
||||
@@ -124,8 +124,7 @@ func NewPipeline(ctx context.Context,
|
||||
}
|
||||
|
||||
// SetSyncStatsVars is a thread safe method to sets the synchronizer Stats
|
||||
func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars *common.SCVariablesPtr) {
|
||||
func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats, vars *synchronizer.SCVariablesPtr) {
|
||||
select {
|
||||
case p.statsVarsCh <- statsVars{Stats: *stats, Vars: *vars}:
|
||||
case <-ctx.Done():
|
||||
@@ -134,7 +133,7 @@ func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Sta
|
||||
|
||||
// reset pipeline state
|
||||
func (p *Pipeline) reset(batchNum common.BatchNum,
|
||||
stats *synchronizer.Stats, vars *common.SCVariables) error {
|
||||
stats *synchronizer.Stats, vars *synchronizer.SCVariables) error {
|
||||
p.state = state{
|
||||
batchNum: batchNum,
|
||||
lastForgeL1TxsNum: stats.Sync.LastForgeL1TxsNum,
|
||||
@@ -195,7 +194,7 @@ func (p *Pipeline) reset(batchNum common.BatchNum,
|
||||
return nil
|
||||
}
|
||||
|
||||
func (p *Pipeline) syncSCVars(vars common.SCVariablesPtr) {
|
||||
func (p *Pipeline) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
updateSCVars(&p.vars, vars)
|
||||
}
|
||||
|
||||
@@ -210,7 +209,7 @@ func (p *Pipeline) handleForgeBatch(ctx context.Context,
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
log.Errorw("proversPool.Get", "err", err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, err
|
||||
}
|
||||
defer func() {
|
||||
// If we encounter any error (notice that this function returns
|
||||
@@ -224,9 +223,8 @@ func (p *Pipeline) handleForgeBatch(ctx context.Context,
|
||||
|
||||
// 2. Forge the batch internally (make a selection of txs and prepare
|
||||
// all the smart contract arguments)
|
||||
var skipReason *string
|
||||
p.mutexL2DBUpdateDelete.Lock()
|
||||
batchInfo, skipReason, err = p.forgeBatch(batchNum)
|
||||
batchInfo, err = p.forgeBatch(batchNum)
|
||||
p.mutexL2DBUpdateDelete.Unlock()
|
||||
if ctx.Err() != nil {
|
||||
return nil, ctx.Err()
|
||||
@@ -235,13 +233,13 @@ func (p *Pipeline) handleForgeBatch(ctx context.Context,
|
||||
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err,
|
||||
"lastForgeL1TxsNum", p.state.lastForgeL1TxsNum,
|
||||
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum)
|
||||
} else if tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||
// no log
|
||||
} else {
|
||||
log.Errorw("forgeBatch", "err", err)
|
||||
}
|
||||
return nil, tracerr.Wrap(err)
|
||||
} else if skipReason != nil {
|
||||
log.Debugw("skipping batch", "batch", batchNum, "reason", *skipReason)
|
||||
return nil, tracerr.Wrap(errSkipBatchByPolicy)
|
||||
return nil, err
|
||||
}
|
||||
|
||||
// 3. Send the ZKInputs to the proof server
|
||||
@@ -250,14 +248,14 @@ func (p *Pipeline) handleForgeBatch(ctx context.Context,
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
log.Errorw("sendServerProof", "err", err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, err
|
||||
}
|
||||
return batchInfo, nil
|
||||
}
|
||||
|
||||
// Start the forging pipeline
|
||||
func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||
stats *synchronizer.Stats, vars *common.SCVariables) error {
|
||||
stats *synchronizer.Stats, vars *synchronizer.SCVariables) error {
|
||||
if p.started {
|
||||
log.Fatal("Pipeline already started")
|
||||
}
|
||||
@@ -296,7 +294,8 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||
if p.ctx.Err() != nil {
|
||||
continue
|
||||
} else if tracerr.Unwrap(err) == errLastL1BatchNotSynced ||
|
||||
tracerr.Unwrap(err) == errSkipBatchByPolicy {
|
||||
tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||
continue
|
||||
} else if err != nil {
|
||||
p.setErrAtBatchNum(batchNum)
|
||||
@@ -389,109 +388,17 @@ func (p *Pipeline) sendServerProof(ctx context.Context, batchInfo *BatchInfo) er
|
||||
return nil
|
||||
}
|
||||
|
||||
// slotCommitted returns true if the current slot has already been committed
|
||||
func (p *Pipeline) slotCommitted() bool {
|
||||
// Synchronizer has synchronized a batch in the current slot (setting
|
||||
// CurrentSlot.ForgerCommitment) or the pipeline has already
|
||||
// internally-forged a batch in the current slot
|
||||
return p.stats.Sync.Auction.CurrentSlot.ForgerCommitment ||
|
||||
p.stats.Sync.Auction.CurrentSlot.SlotNum == p.state.lastSlotForged
|
||||
}
|
||||
|
||||
// forgePolicySkipPreSelection is called before doing a tx selection in a batch to
|
||||
// determine by policy if we should forge the batch or not. Returns true and
|
||||
// the reason when the forging of the batch must be skipped.
|
||||
func (p *Pipeline) forgePolicySkipPreSelection(now time.Time) (bool, string) {
|
||||
// Check if the slot is not yet fulfilled
|
||||
slotCommitted := p.slotCommitted()
|
||||
if p.cfg.ForgeOncePerSlotIfTxs {
|
||||
if slotCommitted {
|
||||
return true, "cfg.ForgeOncePerSlotIfTxs = true and slot already committed"
|
||||
}
|
||||
return false, ""
|
||||
}
|
||||
// Determine if we must commit the slot
|
||||
if !p.cfg.IgnoreSlotCommitment && !slotCommitted {
|
||||
return false, ""
|
||||
}
|
||||
|
||||
// If we haven't reached the ForgeDelay, skip forging the batch
|
||||
if now.Sub(p.lastForgeTime) < p.cfg.ForgeDelay {
|
||||
return true, "we haven't reached the forge delay"
|
||||
}
|
||||
return false, ""
|
||||
}
|
||||
|
||||
// forgePolicySkipPostSelection is called after doing a tx selection in a batch to
|
||||
// determine by policy if we should forge the batch or not. Returns true and
|
||||
// the reason when the forging of the batch must be skipped.
|
||||
func (p *Pipeline) forgePolicySkipPostSelection(now time.Time, l1UserTxsExtra, l1CoordTxs []common.L1Tx,
|
||||
poolL2Txs []common.PoolL2Tx, batchInfo *BatchInfo) (bool, string, error) {
|
||||
// Check if the slot is not yet fulfilled
|
||||
slotCommitted := p.slotCommitted()
|
||||
|
||||
pendingTxs := true
|
||||
if len(l1UserTxsExtra) == 0 && len(l1CoordTxs) == 0 && len(poolL2Txs) == 0 {
|
||||
if batchInfo.L1Batch {
|
||||
// Query the number of unforged L1UserTxs
|
||||
// (either in a open queue or in a frozen
|
||||
// not-yet-forged queue).
|
||||
count, err := p.historyDB.GetUnforgedL1UserTxsCount()
|
||||
if err != nil {
|
||||
return false, "", err
|
||||
}
|
||||
// If there are future L1UserTxs, we forge a
|
||||
// batch to advance the queues to be able to
|
||||
// forge the L1UserTxs in the future.
|
||||
// Otherwise, skip.
|
||||
if count == 0 {
|
||||
pendingTxs = false
|
||||
}
|
||||
} else {
|
||||
pendingTxs = false
|
||||
}
|
||||
}
|
||||
|
||||
if p.cfg.ForgeOncePerSlotIfTxs {
|
||||
if slotCommitted {
|
||||
return true, "cfg.ForgeOncePerSlotIfTxs = true and slot already committed",
|
||||
nil
|
||||
}
|
||||
if pendingTxs {
|
||||
return false, "", nil
|
||||
}
|
||||
return true, "cfg.ForgeOncePerSlotIfTxs = true and no pending txs",
|
||||
nil
|
||||
}
|
||||
|
||||
// Determine if we must commit the slot
|
||||
if !p.cfg.IgnoreSlotCommitment && !slotCommitted {
|
||||
return false, "", nil
|
||||
}
|
||||
|
||||
// check if there is no txs to forge, no l1UserTxs in the open queue to
|
||||
// freeze and we haven't reached the ForgeNoTxsDelay
|
||||
if now.Sub(p.lastForgeTime) < p.cfg.ForgeNoTxsDelay {
|
||||
if !pendingTxs {
|
||||
return true, "no txs to forge and we haven't reached the forge no txs delay",
|
||||
nil
|
||||
}
|
||||
}
|
||||
return false, "", nil
|
||||
}
|
||||
|
||||
// forgeBatch forges the batchNum batch.
|
||||
func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo,
|
||||
skipReason *string, err error) {
|
||||
func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, err error) {
|
||||
// remove transactions from the pool that have been there for too long
|
||||
_, err = p.purger.InvalidateMaybe(p.l2DB, p.txSelector.LocalAccountsDB(),
|
||||
p.stats.Sync.LastBlock.Num, int64(batchNum))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
_, err = p.purger.PurgeMaybe(p.l2DB, p.stats.Sync.LastBlock.Num, int64(batchNum))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
// Structure to accumulate data and metadata of the batch
|
||||
now := time.Now()
|
||||
@@ -499,50 +406,85 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo,
|
||||
batchInfo.Debug.StartTimestamp = now
|
||||
batchInfo.Debug.StartBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||
|
||||
selectionCfg := &txselector.SelectionConfig{
|
||||
MaxL1UserTxs: common.RollupConstMaxL1UserTx,
|
||||
TxProcessorConfig: p.cfg.TxProcessorConfig,
|
||||
}
|
||||
|
||||
var poolL2Txs []common.PoolL2Tx
|
||||
var discardedL2Txs []common.PoolL2Tx
|
||||
var l1UserTxs, l1CoordTxs []common.L1Tx
|
||||
var l1UserTxsExtra, l1CoordTxs []common.L1Tx
|
||||
var auths [][]byte
|
||||
var coordIdxs []common.Idx
|
||||
|
||||
if skip, reason := p.forgePolicySkipPreSelection(now); skip {
|
||||
return nil, &reason, nil
|
||||
// Check if the slot is not yet fulfilled
|
||||
slotCommitted := false
|
||||
if p.stats.Sync.Auction.CurrentSlot.ForgerCommitment ||
|
||||
p.stats.Sync.Auction.CurrentSlot.SlotNum == p.state.lastSlotForged {
|
||||
slotCommitted = true
|
||||
}
|
||||
|
||||
// If we haven't reached the ForgeDelay, skip forging the batch
|
||||
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeDelay {
|
||||
return nil, errForgeBeforeDelay
|
||||
}
|
||||
|
||||
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||
if p.shouldL1L2Batch(batchInfo) {
|
||||
batchInfo.L1Batch = true
|
||||
if p.state.lastForgeL1TxsNum != p.stats.Sync.LastForgeL1TxsNum {
|
||||
return nil, nil, tracerr.Wrap(errLastL1BatchNotSynced)
|
||||
return nil, tracerr.Wrap(errLastL1BatchNotSynced)
|
||||
}
|
||||
// 2a: L1+L2 txs
|
||||
_l1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(p.state.lastForgeL1TxsNum + 1)
|
||||
l1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(p.state.lastForgeL1TxsNum + 1)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
coordIdxs, auths, l1UserTxs, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||
p.txSelector.GetL1L2TxSelection(p.cfg.TxProcessorConfig, _l1UserTxs)
|
||||
coordIdxs, auths, l1UserTxsExtra, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||
p.txSelector.GetL1L2TxSelection(selectionCfg, l1UserTxs)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
} else {
|
||||
// 2b: only L2 txs
|
||||
coordIdxs, auths, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||
p.txSelector.GetL2TxSelection(p.cfg.TxProcessorConfig)
|
||||
p.txSelector.GetL2TxSelection(selectionCfg)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
l1UserTxs = nil
|
||||
l1UserTxsExtra = nil
|
||||
}
|
||||
|
||||
if skip, reason, err := p.forgePolicySkipPostSelection(now,
|
||||
l1UserTxs, l1CoordTxs, poolL2Txs, batchInfo); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
} else if skip {
|
||||
if err := p.txSelector.Reset(batchInfo.BatchNum-1, false); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
// If there are no txs to forge, no l1UserTxs in the open queue to
|
||||
// freeze, and we haven't reached the ForgeNoTxsDelay, skip forging the
|
||||
// batch.
|
||||
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeNoTxsDelay {
|
||||
noTxs := false
|
||||
if len(l1UserTxsExtra) == 0 && len(l1CoordTxs) == 0 && len(poolL2Txs) == 0 {
|
||||
if batchInfo.L1Batch {
|
||||
// Query the L1UserTxs in the queue following
|
||||
// the one we are trying to forge.
|
||||
nextL1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(
|
||||
p.state.lastForgeL1TxsNum + 1)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
// If there are future L1UserTxs, we forge a
|
||||
// batch to advance the queues and forge the
|
||||
// L1UserTxs in the future. Otherwise, skip.
|
||||
if len(nextL1UserTxs) == 0 {
|
||||
noTxs = true
|
||||
}
|
||||
} else {
|
||||
noTxs = true
|
||||
}
|
||||
}
|
||||
if noTxs {
|
||||
if err := p.txSelector.Reset(batchInfo.BatchNum-1, false); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return nil, errForgeNoTxsBeforeDelay
|
||||
}
|
||||
return nil, &reason, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
if batchInfo.L1Batch {
|
||||
@@ -551,41 +493,40 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo,
|
||||
}
|
||||
|
||||
// 3. Save metadata from TxSelector output for BatchNum
|
||||
batchInfo.L1UserTxs = l1UserTxs
|
||||
batchInfo.L1UserTxsExtra = l1UserTxsExtra
|
||||
batchInfo.L1CoordTxs = l1CoordTxs
|
||||
batchInfo.L1CoordinatorTxsAuths = auths
|
||||
batchInfo.CoordIdxs = coordIdxs
|
||||
batchInfo.VerifierIdx = p.cfg.VerifierIdx
|
||||
|
||||
if err := p.l2DB.StartForging(common.TxIDsFromPoolL2Txs(poolL2Txs),
|
||||
batchInfo.BatchNum); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := p.l2DB.StartForging(common.TxIDsFromPoolL2Txs(poolL2Txs), batchInfo.BatchNum); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if err := p.l2DB.UpdateTxsInfo(discardedL2Txs); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// Invalidate transactions that become invalid because of
|
||||
// Invalidate transactions that become invalid beause of
|
||||
// the poolL2Txs selected. Will mark as invalid the txs that have a
|
||||
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||
// all the nonces smaller than the current one)
|
||||
err = p.l2DB.InvalidateOldNonces(idxsNonceFromPoolL2Txs(poolL2Txs), batchInfo.BatchNum)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// 4. Call BatchBuilder with TxSelector output
|
||||
configBatch := &batchbuilder.ConfigBatch{
|
||||
TxProcessorConfig: p.cfg.TxProcessorConfig,
|
||||
}
|
||||
zkInputs, err := p.batchBuilder.BuildBatch(coordIdxs, configBatch, l1UserTxs,
|
||||
zkInputs, err := p.batchBuilder.BuildBatch(coordIdxs, configBatch, l1UserTxsExtra,
|
||||
l1CoordTxs, poolL2Txs)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
l2Txs, err := common.PoolL2TxsToL2Txs(poolL2Txs) // NOTE: This is a big uggly, find a better way
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
batchInfo.L2Txs = l2Txs
|
||||
|
||||
@@ -597,13 +538,12 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo,
|
||||
|
||||
p.state.lastSlotForged = p.stats.Sync.Auction.CurrentSlot.SlotNum
|
||||
|
||||
return batchInfo, nil, nil
|
||||
return batchInfo, nil
|
||||
}
|
||||
|
||||
// waitServerProof gets the generated zkProof & sends it to the SmartContract
|
||||
func (p *Pipeline) waitServerProof(ctx context.Context, batchInfo *BatchInfo) error {
|
||||
proof, pubInputs, err := batchInfo.ServerProof.GetProof(ctx) // blocking call,
|
||||
// until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
proof, pubInputs, err := batchInfo.ServerProof.GetProof(ctx) // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -642,7 +582,7 @@ func prepareForgeBatchArgs(batchInfo *BatchInfo) *eth.RollupForgeBatchArgs {
|
||||
NewLastIdx: int64(zki.Metadata.NewLastIdxRaw),
|
||||
NewStRoot: zki.Metadata.NewStateRootRaw.BigInt(),
|
||||
NewExitRoot: zki.Metadata.NewExitRootRaw.BigInt(),
|
||||
L1UserTxs: batchInfo.L1UserTxs,
|
||||
L1UserTxs: batchInfo.L1UserTxsExtra,
|
||||
L1CoordinatorTxs: batchInfo.L1CoordTxs,
|
||||
L1CoordinatorTxsAuths: batchInfo.L1CoordinatorTxsAuths,
|
||||
L2TxsData: batchInfo.L2Txs,
|
||||
|
||||
@@ -140,7 +140,7 @@ func preloadSync(t *testing.T, ethClient *test.Client, sync *synchronizer.Synchr
|
||||
blocks[0].Rollup.Batches[0].Batch.StateRoot =
|
||||
newBigInt("0")
|
||||
blocks[0].Rollup.Batches[1].Batch.StateRoot =
|
||||
newBigInt("6860514559199319426609623120853503165917774887908204288119245630904770452486")
|
||||
newBigInt("10941365282189107056349764238909072001483688090878331371699519307087372995595")
|
||||
|
||||
ethAddTokens(blocks, ethClient)
|
||||
err = ethClient.CtlAddBlocks(blocks)
|
||||
@@ -206,7 +206,11 @@ PoolTransfer(0) User2-User3: 300 (126)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
err = pipeline.reset(batchNum, syncStats, syncSCVars)
|
||||
err = pipeline.reset(batchNum, syncStats, &synchronizer.SCVariables{
|
||||
Rollup: *syncSCVars.Rollup,
|
||||
Auction: *syncSCVars.Auction,
|
||||
WDelayer: *syncSCVars.WDelayer,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
// Sanity check
|
||||
sdbAccounts, err := pipeline.txSelector.LocalAccountsDB().TestGetAccounts()
|
||||
@@ -224,12 +228,12 @@ PoolTransfer(0) User2-User3: 300 (126)
|
||||
|
||||
batchNum++
|
||||
|
||||
batchInfo, _, err := pipeline.forgeBatch(batchNum)
|
||||
batchInfo, err := pipeline.forgeBatch(batchNum)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, 3, len(batchInfo.L2Txs))
|
||||
|
||||
batchNum++
|
||||
batchInfo, _, err = pipeline.forgeBatch(batchNum)
|
||||
batchInfo, err = pipeline.forgeBatch(batchNum)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, 0, len(batchInfo.L2Txs))
|
||||
}
|
||||
|
||||
@@ -14,7 +14,7 @@ import (
|
||||
// PurgerCfg is the purger configuration
|
||||
type PurgerCfg struct {
|
||||
// PurgeBatchDelay is the delay between batches to purge outdated
|
||||
// transactions. Outdated L2Txs are those that have been forged or
|
||||
// transactions. Oudated L2Txs are those that have been forged or
|
||||
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||
// that have been in the pool for longer than TTL once there are
|
||||
// MaxTxs.
|
||||
@@ -23,7 +23,7 @@ type PurgerCfg struct {
|
||||
// transactions due to nonce lower than the account nonce.
|
||||
InvalidateBatchDelay int64
|
||||
// PurgeBlockDelay is the delay between blocks to purge outdated
|
||||
// transactions. Outdated L2Txs are those that have been forged or
|
||||
// transactions. Oudated L2Txs are those that have been forged or
|
||||
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||
// that have been in the pool for longer than TTL once there are
|
||||
// MaxTxs.
|
||||
|
||||
@@ -31,10 +31,10 @@ type TxManager struct {
|
||||
batchCh chan *BatchInfo
|
||||
chainID *big.Int
|
||||
account accounts.Account
|
||||
consts common.SCConsts
|
||||
consts synchronizer.SCConsts
|
||||
|
||||
stats synchronizer.Stats
|
||||
vars common.SCVariables
|
||||
vars synchronizer.SCVariables
|
||||
statsVarsCh chan statsVars
|
||||
|
||||
discardPipelineCh chan int // int refers to the pipelineNum
|
||||
@@ -55,8 +55,7 @@ type TxManager struct {
|
||||
|
||||
// NewTxManager creates a new TxManager
|
||||
func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterface, l2DB *l2db.L2DB,
|
||||
coord *Coordinator, scConsts *common.SCConsts, initSCVars *common.SCVariables) (
|
||||
*TxManager, error) {
|
||||
coord *Coordinator, scConsts *synchronizer.SCConsts, initSCVars *synchronizer.SCVariables) (*TxManager, error) {
|
||||
chainID, err := ethClient.EthChainID()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -67,7 +66,7 @@ func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterfac
|
||||
}
|
||||
accNonce, err := ethClient.EthNonceAt(ctx, *address, nil)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, err
|
||||
}
|
||||
log.Infow("TxManager started", "nonce", accNonce)
|
||||
return &TxManager{
|
||||
@@ -103,8 +102,7 @@ func (t *TxManager) AddBatch(ctx context.Context, batchInfo *BatchInfo) {
|
||||
}
|
||||
|
||||
// SetSyncStatsVars is a thread safe method to sets the synchronizer Stats
|
||||
func (t *TxManager) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars *common.SCVariablesPtr) {
|
||||
func (t *TxManager) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats, vars *synchronizer.SCVariablesPtr) {
|
||||
select {
|
||||
case t.statsVarsCh <- statsVars{Stats: *stats, Vars: *vars}:
|
||||
case <-ctx.Done():
|
||||
@@ -120,7 +118,7 @@ func (t *TxManager) DiscardPipeline(ctx context.Context, pipelineNum int) {
|
||||
}
|
||||
}
|
||||
|
||||
func (t *TxManager) syncSCVars(vars common.SCVariablesPtr) {
|
||||
func (t *TxManager) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
updateSCVars(&t.vars, vars)
|
||||
}
|
||||
|
||||
@@ -147,7 +145,7 @@ func (t *TxManager) NewAuth(ctx context.Context, batchInfo *BatchInfo) (*bind.Tr
|
||||
auth.Value = big.NewInt(0) // in wei
|
||||
|
||||
gasLimit := t.cfg.ForgeBatchGasCost.Fixed +
|
||||
uint64(len(batchInfo.L1UserTxs))*t.cfg.ForgeBatchGasCost.L1UserTx +
|
||||
uint64(len(batchInfo.L1UserTxsExtra))*t.cfg.ForgeBatchGasCost.L1UserTx +
|
||||
uint64(len(batchInfo.L1CoordTxs))*t.cfg.ForgeBatchGasCost.L1CoordTx +
|
||||
uint64(len(batchInfo.L2Txs))*t.cfg.ForgeBatchGasCost.L2Tx
|
||||
auth.GasLimit = gasLimit
|
||||
@@ -184,30 +182,19 @@ func addPerc(v *big.Int, p int64) *big.Int {
|
||||
r.Mul(r, big.NewInt(p))
|
||||
// nolint reason: to calculate percentages we divide by 100
|
||||
r.Div(r, big.NewInt(100)) //nolit:gomnd
|
||||
// If the increase is 0, force it to be 1 so that a gas increase
|
||||
// doesn't result in the same value, making the transaction to be equal
|
||||
// than before.
|
||||
if r.Cmp(big.NewInt(0)) == 0 {
|
||||
r = big.NewInt(1)
|
||||
}
|
||||
return r.Add(v, r)
|
||||
}
|
||||
|
||||
func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchInfo,
|
||||
resend bool) error {
|
||||
func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchInfo, resend bool) error {
|
||||
var ethTx *types.Transaction
|
||||
var err error
|
||||
var auth *bind.TransactOpts
|
||||
auth, err := t.NewAuth(ctx, batchInfo)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
auth.Nonce = big.NewInt(int64(t.accNextNonce))
|
||||
if resend {
|
||||
auth = batchInfo.Auth
|
||||
auth.GasPrice = addPerc(auth.GasPrice, 10)
|
||||
} else {
|
||||
auth, err = t.NewAuth(ctx, batchInfo)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
batchInfo.Auth = auth
|
||||
auth.Nonce = big.NewInt(int64(t.accNextNonce))
|
||||
auth.Nonce = big.NewInt(int64(batchInfo.EthTx.Nonce()))
|
||||
}
|
||||
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ {
|
||||
if auth.GasPrice.Cmp(t.cfg.MaxGasPrice) > 0 {
|
||||
@@ -278,8 +265,7 @@ func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchIn
|
||||
t.lastSentL1BatchBlockNum = t.stats.Eth.LastBlock.Num + 1
|
||||
}
|
||||
}
|
||||
if err := t.l2DB.DoneForging(common.TxIDsFromL2Txs(batchInfo.L2Txs),
|
||||
batchInfo.BatchNum); err != nil {
|
||||
if err := t.l2DB.DoneForging(common.TxIDsFromL2Txs(batchInfo.L2Txs), batchInfo.BatchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
@@ -311,9 +297,7 @@ func (t *TxManager) checkEthTransactionReceipt(ctx context.Context, batchInfo *B
|
||||
}
|
||||
}
|
||||
if err != nil {
|
||||
return tracerr.Wrap(
|
||||
fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w",
|
||||
err))
|
||||
return tracerr.Wrap(fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w", err))
|
||||
}
|
||||
batchInfo.Receipt = receipt
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
@@ -503,7 +487,7 @@ func (t *TxManager) Run(ctx context.Context) {
|
||||
// Our ethNode is giving an error different
|
||||
// than "not found" when getting the receipt
|
||||
// for the transaction, so we can't figure out
|
||||
// if it was not mined, mined and successful or
|
||||
// if it was not mined, mined and succesfull or
|
||||
// mined and failed. This could be due to the
|
||||
// ethNode failure.
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
@@ -568,7 +552,7 @@ func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
||||
// Our ethNode is giving an error different
|
||||
// than "not found" when getting the receipt
|
||||
// for the transaction, so we can't figure out
|
||||
// if it was not mined, mined and successful or
|
||||
// if it was not mined, mined and succesfull or
|
||||
// mined and failed. This could be due to the
|
||||
// ethNode failure.
|
||||
next++
|
||||
@@ -608,7 +592,7 @@ func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
||||
func (t *TxManager) canForgeAt(blockNum int64) bool {
|
||||
return canForge(&t.consts.Auction, &t.vars.Auction,
|
||||
&t.stats.Sync.Auction.CurrentSlot, &t.stats.Sync.Auction.NextSlot,
|
||||
t.cfg.ForgerAddress, blockNum, t.cfg.MustForgeAtSlotDeadline)
|
||||
t.cfg.ForgerAddress, blockNum)
|
||||
}
|
||||
|
||||
func (t *TxManager) mustL1L2Batch(blockNum int64) bool {
|
||||
|
||||
@@ -8,7 +8,6 @@ import (
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
@@ -39,14 +38,14 @@ func (hdb *HistoryDB) GetBatchAPI(batchNum common.BatchNum) (*BatchAPI, error) {
|
||||
return hdb.getBatchAPI(hdb.dbRead, batchNum)
|
||||
}
|
||||
|
||||
// GetBatchInternalAPI return the batch with the given batchNum
|
||||
// GetBatchAPI return the batch with the given batchNum
|
||||
func (hdb *HistoryDB) GetBatchInternalAPI(batchNum common.BatchNum) (*BatchAPI, error) {
|
||||
return hdb.getBatchAPI(hdb.dbRead, batchNum)
|
||||
}
|
||||
|
||||
func (hdb *HistoryDB) getBatchAPI(d meddler.DB, batchNum common.BatchNum) (*BatchAPI, error) {
|
||||
batch := &BatchAPI{}
|
||||
if err := meddler.QueryRow(
|
||||
return batch, tracerr.Wrap(meddler.QueryRow(
|
||||
d, batch,
|
||||
`SELECT batch.item_id, batch.batch_num, batch.eth_block_num,
|
||||
batch.forger_addr, batch.fees_collected, batch.total_fees_usd, batch.state_root,
|
||||
@@ -55,11 +54,7 @@ func (hdb *HistoryDB) getBatchAPI(d meddler.DB, batchNum common.BatchNum) (*Batc
|
||||
COALESCE ((SELECT COUNT(*) FROM tx WHERE batch_num = batch.batch_num), 0) AS forged_txs
|
||||
FROM batch INNER JOIN block ON batch.eth_block_num = block.eth_block_num
|
||||
WHERE batch_num = $1;`, batchNum,
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
batch.CollectedFeesAPI = apitypes.NewCollectedFeesAPI(batch.CollectedFeesDB)
|
||||
return batch, nil
|
||||
))
|
||||
}
|
||||
|
||||
// GetBatchesAPI return the batches applying the given filters
|
||||
@@ -160,9 +155,6 @@ func (hdb *HistoryDB) GetBatchesAPI(
|
||||
if len(batches) == 0 {
|
||||
return batches, 0, nil
|
||||
}
|
||||
for i := range batches {
|
||||
batches[i].CollectedFeesAPI = apitypes.NewCollectedFeesAPI(batches[i].CollectedFeesDB)
|
||||
}
|
||||
return batches, batches[0].TotalItems - uint64(len(batches)), nil
|
||||
}
|
||||
|
||||
@@ -937,8 +929,7 @@ func (hdb *HistoryDB) GetCommonAccountAPI(idx common.Idx) (*common.Account, erro
|
||||
defer hdb.apiConnCon.Release()
|
||||
account := &common.Account{}
|
||||
err = meddler.QueryRow(
|
||||
hdb.dbRead, account, `SELECT idx, token_id, batch_num, bjj, eth_addr
|
||||
FROM account WHERE idx = $1;`, idx,
|
||||
hdb.dbRead, account, `SELECT * FROM account WHERE idx = $1;`, idx,
|
||||
)
|
||||
return account, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -953,7 +944,6 @@ func (hdb *HistoryDB) GetCoordinatorAPI(bidderAddr ethCommon.Address) (*Coordina
|
||||
defer hdb.apiConnCon.Release()
|
||||
return hdb.getCoordinatorAPI(hdb.dbRead, bidderAddr)
|
||||
}
|
||||
|
||||
func (hdb *HistoryDB) getCoordinatorAPI(d meddler.DB, bidderAddr ethCommon.Address) (*CoordinatorAPI, error) {
|
||||
coordinator := &CoordinatorAPI{}
|
||||
err := meddler.QueryRow(
|
||||
@@ -964,7 +954,6 @@ func (hdb *HistoryDB) getCoordinatorAPI(d meddler.DB, bidderAddr ethCommon.Addre
|
||||
return coordinator, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetNodeInfoAPI retusnt he NodeInfo
|
||||
func (hdb *HistoryDB) GetNodeInfoAPI() (*NodeInfo, error) {
|
||||
cancel, err := hdb.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
@@ -975,9 +964,9 @@ func (hdb *HistoryDB) GetNodeInfoAPI() (*NodeInfo, error) {
|
||||
return hdb.GetNodeInfo()
|
||||
}
|
||||
|
||||
// GetBucketUpdatesInternalAPI returns the latest bucket updates
|
||||
func (hdb *HistoryDB) GetBucketUpdatesInternalAPI() ([]BucketUpdateAPI, error) {
|
||||
var bucketUpdates []*BucketUpdateAPI
|
||||
// var bucketUpdates []*common.BucketUpdate
|
||||
err := meddler.QueryAll(
|
||||
hdb.dbRead, &bucketUpdates,
|
||||
`SELECT num_bucket, withdrawals FROM bucket_update
|
||||
@@ -988,7 +977,7 @@ func (hdb *HistoryDB) GetBucketUpdatesInternalAPI() ([]BucketUpdateAPI, error) {
|
||||
return db.SlicePtrsToSlice(bucketUpdates).([]BucketUpdateAPI), tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetNextForgersInternalAPI returns next forgers
|
||||
// getNextForgers returns next forgers
|
||||
func (hdb *HistoryDB) GetNextForgersInternalAPI(auctionVars *common.AuctionVariables,
|
||||
auctionConsts *common.AuctionConstants,
|
||||
lastBlock common.Block, currentSlot, lastClosedSlot int64) ([]NextForgerAPI, error) {
|
||||
@@ -1082,7 +1071,7 @@ func (hdb *HistoryDB) GetNextForgersInternalAPI(auctionVars *common.AuctionVaria
|
||||
return nextForgers, nil
|
||||
}
|
||||
|
||||
// GetMetricsInternalAPI returns the MetricsAPI
|
||||
// UpdateMetrics update Status.Metrics information
|
||||
func (hdb *HistoryDB) GetMetricsInternalAPI(lastBatchNum common.BatchNum) (*MetricsAPI, error) {
|
||||
var metrics MetricsAPI
|
||||
// Get the first and last batch of the last 24h and their timestamps
|
||||
@@ -1129,6 +1118,7 @@ func (hdb *HistoryDB) GetMetricsInternalAPI(lastBatchNum common.BatchNum) (*Metr
|
||||
nBatches++
|
||||
}
|
||||
if (p.ToBatchNum - p.FromBatchNum) > 0 {
|
||||
fmt.Printf("DBG ntxs: %v, nBatches: %v\n", nTxs, nBatches)
|
||||
metrics.TransactionsPerBatch = float64(nTxs) /
|
||||
float64(nBatches)
|
||||
} else {
|
||||
@@ -1145,17 +1135,8 @@ func (hdb *HistoryDB) GetMetricsInternalAPI(lastBatchNum common.BatchNum) (*Metr
|
||||
}
|
||||
// Set batch frequency
|
||||
metrics.BatchFrequency = seconds / float64(nBatches)
|
||||
// Set avg transaction fee (only L2 txs have fee)
|
||||
row = hdb.dbRead.QueryRow(
|
||||
`SELECT COUNT(*) as total_txs FROM tx WHERE tx.batch_num between $1 AND $2 AND NOT is_l1;`,
|
||||
p.FromBatchNum, p.ToBatchNum,
|
||||
)
|
||||
var nL2Txs int
|
||||
if err := row.Scan(&nL2Txs); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if nL2Txs > 0 {
|
||||
metrics.AvgTransactionFee = totalFee / float64(nL2Txs)
|
||||
if nTxs > 0 {
|
||||
metrics.AvgTransactionFee = totalFee / float64(nTxs)
|
||||
} else {
|
||||
metrics.AvgTransactionFee = 0
|
||||
}
|
||||
@@ -1190,7 +1171,6 @@ func (hdb *HistoryDB) GetMetricsInternalAPI(lastBatchNum common.BatchNum) (*Metr
|
||||
return &metrics, nil
|
||||
}
|
||||
|
||||
// GetStateAPI returns the StateAPI
|
||||
func (hdb *HistoryDB) GetStateAPI() (*StateAPI, error) {
|
||||
cancel, err := hdb.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
|
||||
@@ -179,7 +179,7 @@ func (hdb *HistoryDB) GetBatch(batchNum common.BatchNum) (*common.Batch, error)
|
||||
batch.slot_num, batch.total_fees_usd FROM batch WHERE batch_num = $1;`,
|
||||
batchNum,
|
||||
)
|
||||
return &batch, tracerr.Wrap(err)
|
||||
return &batch, err
|
||||
}
|
||||
|
||||
// GetAllBatches retrieve all batches from the DB
|
||||
@@ -235,7 +235,7 @@ func (hdb *HistoryDB) GetLastBatch() (*common.Batch, error) {
|
||||
batch.num_accounts, batch.last_idx, batch.exit_root, batch.forge_l1_txs_num,
|
||||
batch.slot_num, batch.total_fees_usd FROM batch ORDER BY batch_num DESC LIMIT 1;`,
|
||||
)
|
||||
return &batch, tracerr.Wrap(err)
|
||||
return &batch, err
|
||||
}
|
||||
|
||||
// GetLastL1BatchBlockNum returns the blockNum of the latest forged l1Batch
|
||||
@@ -456,10 +456,13 @@ func (hdb *HistoryDB) addTokens(d meddler.DB, tokens []common.Token) error {
|
||||
|
||||
// UpdateTokenValue updates the USD value of a token. Value is the price in
|
||||
// USD of a normalized token (1 token = 10^decimals units)
|
||||
func (hdb *HistoryDB) UpdateTokenValue(tokenAddr ethCommon.Address, value float64) error {
|
||||
func (hdb *HistoryDB) UpdateTokenValue(tokenSymbol string, value float64) error {
|
||||
// Sanitize symbol
|
||||
tokenSymbol = strings.ToValidUTF8(tokenSymbol, " ")
|
||||
|
||||
_, err := hdb.dbWrite.Exec(
|
||||
"UPDATE token SET usd = $1 WHERE eth_addr = $2;",
|
||||
value, tokenAddr,
|
||||
"UPDATE token SET usd = $1 WHERE symbol = $2;",
|
||||
value, tokenSymbol,
|
||||
)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -483,14 +486,23 @@ func (hdb *HistoryDB) GetAllTokens() ([]TokenWithUSD, error) {
|
||||
return db.SlicePtrsToSlice(tokens).([]TokenWithUSD), tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetTokenSymbolsAndAddrs returns all the token symbols and addresses from the DB
|
||||
func (hdb *HistoryDB) GetTokenSymbolsAndAddrs() ([]TokenSymbolAndAddr, error) {
|
||||
var tokens []*TokenSymbolAndAddr
|
||||
err := meddler.QueryAll(
|
||||
hdb.dbRead, &tokens,
|
||||
"SELECT symbol, eth_addr FROM token;",
|
||||
)
|
||||
return db.SlicePtrsToSlice(tokens).([]TokenSymbolAndAddr), tracerr.Wrap(err)
|
||||
// GetTokenSymbols returns all the token symbols from the DB
|
||||
func (hdb *HistoryDB) GetTokenSymbols() ([]string, error) {
|
||||
var tokenSymbols []string
|
||||
rows, err := hdb.dbRead.Query("SELECT symbol FROM token;")
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
defer db.RowsClose(rows)
|
||||
sym := new(string)
|
||||
for rows.Next() {
|
||||
err = rows.Scan(sym)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tokenSymbols = append(tokenSymbols, *sym)
|
||||
}
|
||||
return tokenSymbols, nil
|
||||
}
|
||||
|
||||
// AddAccounts insert accounts into the DB
|
||||
@@ -693,11 +705,11 @@ func (hdb *HistoryDB) GetAllExits() ([]common.ExitInfo, error) {
|
||||
func (hdb *HistoryDB) GetAllL1UserTxs() ([]common.L1Tx, error) {
|
||||
var txs []*common.L1Tx
|
||||
err := meddler.QueryAll(
|
||||
hdb.dbRead, &txs,
|
||||
hdb.dbRead, &txs, // Note that '\x' gets parsed as a big.Int with value = 0
|
||||
`SELECT tx.id, tx.to_forge_l1_txs_num, tx.position, tx.user_origin,
|
||||
tx.from_idx, tx.effective_from_idx, tx.from_eth_addr, tx.from_bjj, tx.to_idx, tx.token_id,
|
||||
tx.amount, (CASE WHEN tx.batch_num IS NULL THEN NULL WHEN tx.amount_success THEN tx.amount ELSE 0 END) AS effective_amount,
|
||||
tx.deposit_amount, (CASE WHEN tx.batch_num IS NULL THEN NULL WHEN tx.deposit_amount_success THEN tx.deposit_amount ELSE 0 END) AS effective_deposit_amount,
|
||||
tx.amount, (CASE WHEN tx.batch_num IS NULL THEN NULL WHEN tx.amount_success THEN tx.amount ELSE '\x' END) AS effective_amount,
|
||||
tx.deposit_amount, (CASE WHEN tx.batch_num IS NULL THEN NULL WHEN tx.deposit_amount_success THEN tx.deposit_amount ELSE '\x' END) AS effective_deposit_amount,
|
||||
tx.eth_block_num, tx.type, tx.batch_num
|
||||
FROM tx WHERE is_l1 = TRUE AND user_origin = TRUE ORDER BY item_id;`,
|
||||
)
|
||||
@@ -751,16 +763,6 @@ func (hdb *HistoryDB) GetUnforgedL1UserTxs(toForgeL1TxsNum int64) ([]common.L1Tx
|
||||
return db.SlicePtrsToSlice(txs).([]common.L1Tx), tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetUnforgedL1UserTxsCount returns the count of unforged L1Txs (either in
|
||||
// open or frozen queues that are not yet forged)
|
||||
func (hdb *HistoryDB) GetUnforgedL1UserTxsCount() (int, error) {
|
||||
row := hdb.dbRead.QueryRow(
|
||||
`SELECT COUNT(*) FROM tx WHERE batch_num IS NULL;`,
|
||||
)
|
||||
var count int
|
||||
return count, tracerr.Wrap(row.Scan(&count))
|
||||
}
|
||||
|
||||
// TODO: Think about chaning all the queries that return a last value, to queries that return the next valid value.
|
||||
|
||||
// GetLastTxsPosition for a given to_forge_l1_txs_num
|
||||
@@ -1159,7 +1161,7 @@ func (hdb *HistoryDB) GetTokensTest() ([]TokenWithUSD, error) {
|
||||
tokens := []*TokenWithUSD{}
|
||||
if err := meddler.QueryAll(
|
||||
hdb.dbRead, &tokens,
|
||||
"SELECT * FROM token ORDER BY token_id ASC",
|
||||
"SELECT * FROM TOKEN",
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -1169,18 +1171,7 @@ func (hdb *HistoryDB) GetTokensTest() ([]TokenWithUSD, error) {
|
||||
return db.SlicePtrsToSlice(tokens).([]TokenWithUSD), nil
|
||||
}
|
||||
|
||||
const (
|
||||
// CreateAccountExtraFeePercentage is the multiplication factor over
|
||||
// the average fee for CreateAccount that is applied to obtain the
|
||||
// recommended fee for CreateAccount
|
||||
CreateAccountExtraFeePercentage float64 = 2.5
|
||||
// CreateAccountInternalExtraFeePercentage is the multiplication factor
|
||||
// over the average fee for CreateAccountInternal that is applied to
|
||||
// obtain the recommended fee for CreateAccountInternal
|
||||
CreateAccountInternalExtraFeePercentage float64 = 2.0
|
||||
)
|
||||
|
||||
// GetRecommendedFee returns the RecommendedFee information
|
||||
// UpdateRecommendedFee update Status.RecommendedFee information
|
||||
func (hdb *HistoryDB) GetRecommendedFee(minFeeUSD float64) (*common.RecommendedFee, error) {
|
||||
var recommendedFee common.RecommendedFee
|
||||
// Get total txs and the batch of the first selected tx of the last hour
|
||||
@@ -1220,8 +1211,8 @@ func (hdb *HistoryDB) GetRecommendedFee(minFeeUSD float64) (*common.RecommendedF
|
||||
recommendedFee.ExistingAccount =
|
||||
math.Max(avgTransactionFee, minFeeUSD)
|
||||
recommendedFee.CreatesAccount =
|
||||
math.Max(CreateAccountExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||
recommendedFee.CreatesAccountInternal =
|
||||
math.Max(CreateAccountInternalExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||
math.Max(createAccountExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||
recommendedFee.CreatesAccountAndRegister =
|
||||
math.Max(createAccountInternalExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||
return &recommendedFee, nil
|
||||
}
|
||||
|
||||
@@ -11,7 +11,6 @@ import (
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
@@ -44,7 +43,7 @@ func TestMain(m *testing.M) {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
apiConnCon := dbUtils.NewAPIConnectionController(1, time.Second)
|
||||
apiConnCon := dbUtils.NewAPICnnectionController(1, time.Second)
|
||||
historyDBWithACC = NewHistoryDB(db, db, apiConnCon)
|
||||
// Run tests
|
||||
result := m.Run()
|
||||
@@ -167,7 +166,7 @@ func TestBatches(t *testing.T) {
|
||||
if i%2 != 0 {
|
||||
// Set value to the token
|
||||
value := (float64(i) + 5) * 5.389329
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.EthAddr, value))
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.Symbol, value))
|
||||
tokensValue[token.TokenID] = value / math.Pow(10, float64(token.Decimals))
|
||||
}
|
||||
}
|
||||
@@ -277,7 +276,7 @@ func TestTokens(t *testing.T) {
|
||||
// Update token value
|
||||
for i, token := range tokens {
|
||||
value := 1.01 * float64(i)
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.EthAddr, value))
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.Symbol, value))
|
||||
}
|
||||
// Fetch tokens
|
||||
fetchedTokens, err = historyDB.GetTokensTest()
|
||||
@@ -303,7 +302,7 @@ func TestTokensUTF8(t *testing.T) {
|
||||
// Generate fake tokens
|
||||
const nTokens = 5
|
||||
tokens, ethToken := test.GenTokens(nTokens, blocks)
|
||||
nonUTFTokens := make([]common.Token, len(tokens))
|
||||
nonUTFTokens := make([]common.Token, len(tokens)+1)
|
||||
// Force token.name and token.symbol to be non UTF-8 Strings
|
||||
for i, token := range tokens {
|
||||
token.Name = fmt.Sprint("NON-UTF8-NAME-\xc5-", i)
|
||||
@@ -333,7 +332,7 @@ func TestTokensUTF8(t *testing.T) {
|
||||
// Update token value
|
||||
for i, token := range nonUTFTokens {
|
||||
value := 1.01 * float64(i)
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.EthAddr, value))
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.Symbol, value))
|
||||
}
|
||||
// Fetch tokens
|
||||
fetchedTokens, err = historyDB.GetTokensTest()
|
||||
@@ -721,10 +720,6 @@ func TestGetUnforgedL1UserTxs(t *testing.T) {
|
||||
assert.Equal(t, 5, len(l1UserTxs))
|
||||
assert.Equal(t, blocks[0].Rollup.L1UserTxs, l1UserTxs)
|
||||
|
||||
count, err := historyDB.GetUnforgedL1UserTxsCount()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, 5, count)
|
||||
|
||||
// No l1UserTxs for this toForgeL1TxsNum
|
||||
l1UserTxs, err = historyDB.GetUnforgedL1UserTxs(2)
|
||||
require.NoError(t, err)
|
||||
@@ -738,7 +733,7 @@ func exampleInitSCVars() (*common.RollupVariables, *common.AuctionVariables, *co
|
||||
big.NewInt(10),
|
||||
12,
|
||||
13,
|
||||
[]common.BucketParams{},
|
||||
[5]common.BucketParams{},
|
||||
false,
|
||||
}
|
||||
//nolint:govet
|
||||
@@ -1177,6 +1172,15 @@ func TestGetMetricsAPI(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
// clientSetupExample := test.NewClientSetupExample()
|
||||
// apiStateUpdater := NewAPIStateUpdater(historyDB, &NodeConfig{1000, 0.5},
|
||||
// &Constants{
|
||||
// RollupConstants: *clientSetupExample.RollupConstants,
|
||||
// AuctionConstants: *clientSetupExample.AuctionConstants,
|
||||
// WDelayerConstants: *clientSetupExample.WDelayerConstants,
|
||||
// ChainID: uint16(clientSetupExample.ChainID.Int64()),
|
||||
// HermezAddress: clientSetupExample.AuctionConstants.HermezRollup,
|
||||
// })
|
||||
res, err := historyDB.GetMetricsInternalAPI(common.BatchNum(numBatches))
|
||||
assert.NoError(t, err)
|
||||
|
||||
@@ -1463,7 +1467,7 @@ func setTestBlocks(from, to int64) []common.Block {
|
||||
func TestNodeInfo(t *testing.T) {
|
||||
test.WipeDB(historyDB.DB())
|
||||
|
||||
err := historyDB.SetStateInternalAPI(&StateAPI{})
|
||||
err := historyDB.SetAPIState(&StateAPI{})
|
||||
require.NoError(t, err)
|
||||
|
||||
clientSetup := test.NewClientSetupExample()
|
||||
@@ -1480,10 +1484,6 @@ func TestNodeInfo(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
// Test parameters
|
||||
var f64 float64 = 1.2
|
||||
var i64 int64 = 8888
|
||||
addr := ethCommon.HexToAddress("0x1234")
|
||||
hash := ethCommon.HexToHash("0x5678")
|
||||
stateAPI := &StateAPI{
|
||||
NodePublicConfig: NodePublicConfig{
|
||||
ForgeDelay: 3.1,
|
||||
@@ -1491,56 +1491,6 @@ func TestNodeInfo(t *testing.T) {
|
||||
Network: NetworkAPI{
|
||||
LastEthBlock: 12,
|
||||
LastSyncBlock: 34,
|
||||
LastBatch: &BatchAPI{
|
||||
ItemID: 123,
|
||||
BatchNum: 456,
|
||||
EthBlockNum: 789,
|
||||
EthBlockHash: hash,
|
||||
Timestamp: time.Now(),
|
||||
ForgerAddr: addr,
|
||||
// CollectedFeesDB: map[common.TokenID]*big.Int{
|
||||
// 0: big.NewInt(11111),
|
||||
// 1: big.NewInt(21111),
|
||||
// 2: big.NewInt(31111),
|
||||
// },
|
||||
CollectedFeesAPI: apitypes.CollectedFeesAPI(map[common.TokenID]apitypes.BigIntStr{
|
||||
0: apitypes.BigIntStr("11111"),
|
||||
1: apitypes.BigIntStr("21111"),
|
||||
2: apitypes.BigIntStr("31111"),
|
||||
}),
|
||||
TotalFeesUSD: &f64,
|
||||
StateRoot: apitypes.BigIntStr("1234"),
|
||||
NumAccounts: 11,
|
||||
ExitRoot: apitypes.BigIntStr("5678"),
|
||||
ForgeL1TxsNum: &i64,
|
||||
SlotNum: 44,
|
||||
ForgedTxs: 23,
|
||||
TotalItems: 0,
|
||||
FirstItem: 0,
|
||||
LastItem: 0,
|
||||
},
|
||||
CurrentSlot: 22,
|
||||
NextForgers: []NextForgerAPI{
|
||||
{
|
||||
Coordinator: CoordinatorAPI{
|
||||
ItemID: 111,
|
||||
Bidder: addr,
|
||||
Forger: addr,
|
||||
EthBlockNum: 566,
|
||||
URL: "asd",
|
||||
TotalItems: 0,
|
||||
FirstItem: 0,
|
||||
LastItem: 0,
|
||||
},
|
||||
Period: Period{
|
||||
SlotNum: 33,
|
||||
FromBlock: 55,
|
||||
ToBlock: 66,
|
||||
FromTimestamp: time.Now(),
|
||||
ToTimestamp: time.Now(),
|
||||
},
|
||||
},
|
||||
},
|
||||
},
|
||||
Metrics: MetricsAPI{
|
||||
TransactionsPerBatch: 1.1,
|
||||
@@ -1553,7 +1503,7 @@ func TestNodeInfo(t *testing.T) {
|
||||
ExistingAccount: 0.15,
|
||||
},
|
||||
}
|
||||
err = historyDB.SetStateInternalAPI(stateAPI)
|
||||
err = historyDB.SetAPIState(stateAPI)
|
||||
require.NoError(t, err)
|
||||
|
||||
nodeConfig := &NodeConfig{
|
||||
@@ -1571,16 +1521,7 @@ func TestNodeInfo(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, nodeConfig, dbNodeConfig)
|
||||
|
||||
dbStateAPI, err := historyDB.getStateAPI(historyDB.dbRead)
|
||||
dbStateAPI, err := historyDB.GetStateAPI()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, stateAPI.Network.LastBatch.Timestamp.Unix(),
|
||||
dbStateAPI.Network.LastBatch.Timestamp.Unix())
|
||||
dbStateAPI.Network.LastBatch.Timestamp = stateAPI.Network.LastBatch.Timestamp
|
||||
assert.Equal(t, stateAPI.Network.NextForgers[0].Period.FromTimestamp.Unix(),
|
||||
dbStateAPI.Network.NextForgers[0].Period.FromTimestamp.Unix())
|
||||
dbStateAPI.Network.NextForgers[0].Period.FromTimestamp = stateAPI.Network.NextForgers[0].Period.FromTimestamp
|
||||
assert.Equal(t, stateAPI.Network.NextForgers[0].Period.ToTimestamp.Unix(),
|
||||
dbStateAPI.Network.NextForgers[0].Period.ToTimestamp.Unix())
|
||||
dbStateAPI.Network.NextForgers[0].Period.ToTimestamp = stateAPI.Network.NextForgers[0].Period.ToTimestamp
|
||||
assert.Equal(t, stateAPI, dbStateAPI)
|
||||
}
|
||||
|
||||
@@ -4,13 +4,16 @@ import (
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/russross/meddler"
|
||||
)
|
||||
|
||||
// Period represents a time period in ethereum
|
||||
const (
|
||||
createAccountExtraFeePercentage float64 = 2
|
||||
createAccountInternalExtraFeePercentage float64 = 2.5
|
||||
)
|
||||
|
||||
type Period struct {
|
||||
SlotNum int64 `json:"slotNum"`
|
||||
FromBlock int64 `json:"fromBlock"`
|
||||
@@ -19,13 +22,11 @@ type Period struct {
|
||||
ToTimestamp time.Time `json:"toTimestamp"`
|
||||
}
|
||||
|
||||
// NextForgerAPI represents the next forger exposed via the API
|
||||
type NextForgerAPI struct {
|
||||
Coordinator CoordinatorAPI `json:"coordinator"`
|
||||
Period Period `json:"period"`
|
||||
}
|
||||
|
||||
// NetworkAPI is the network state exposed via the API
|
||||
type NetworkAPI struct {
|
||||
LastEthBlock int64 `json:"lastEthereumBlock"`
|
||||
LastSyncBlock int64 `json:"lastSynchedBlock"`
|
||||
@@ -40,7 +41,6 @@ type NodePublicConfig struct {
|
||||
ForgeDelay float64 `json:"forgeDelay"`
|
||||
}
|
||||
|
||||
// StateAPI is an object representing the node and network state exposed via the API
|
||||
type StateAPI struct {
|
||||
// NodePublicConfig is the configuration of the node that is exposed via API
|
||||
NodePublicConfig NodePublicConfig `json:"nodeConfig"`
|
||||
@@ -52,29 +52,27 @@ type StateAPI struct {
|
||||
RecommendedFee common.RecommendedFee `json:"recommendedFee"`
|
||||
}
|
||||
|
||||
// Constants contains network constants
|
||||
type Constants struct {
|
||||
// RollupConstants common.RollupConstants
|
||||
// AuctionConstants common.AuctionConstants
|
||||
// WDelayerConstants common.WDelayerConstants
|
||||
common.SCConsts
|
||||
ChainID uint16
|
||||
HermezAddress ethCommon.Address
|
||||
}
|
||||
|
||||
// NodeConfig contains the node config exposed in the API
|
||||
type NodeConfig struct {
|
||||
MaxPoolTxs uint32
|
||||
MinFeeUSD float64
|
||||
ForgeDelay float64
|
||||
MaxPoolTxs uint32 `meddler:"max_pool_txs"`
|
||||
MinFeeUSD float64 `meddler:"min_fee"`
|
||||
}
|
||||
|
||||
// NodeInfo contains information about he node used when serving the API
|
||||
type NodeInfo struct {
|
||||
ItemID int `meddler:"item_id,pk"`
|
||||
StateAPI *StateAPI `meddler:"state,json"`
|
||||
APIState *StateAPI `meddler:"state,json"`
|
||||
NodeConfig *NodeConfig `meddler:"config,json"`
|
||||
Constants *Constants `meddler:"constants,json"`
|
||||
}
|
||||
|
||||
// GetNodeInfo returns the NodeInfo
|
||||
func (hdb *HistoryDB) GetNodeInfo() (*NodeInfo, error) {
|
||||
ni := &NodeInfo{}
|
||||
err := meddler.QueryRow(
|
||||
@@ -83,7 +81,6 @@ func (hdb *HistoryDB) GetNodeInfo() (*NodeInfo, error) {
|
||||
return ni, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetConstants returns the Constats
|
||||
func (hdb *HistoryDB) GetConstants() (*Constants, error) {
|
||||
var nodeInfo NodeInfo
|
||||
err := meddler.QueryRow(
|
||||
@@ -93,7 +90,6 @@ func (hdb *HistoryDB) GetConstants() (*Constants, error) {
|
||||
return nodeInfo.Constants, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// SetConstants sets the Constants
|
||||
func (hdb *HistoryDB) SetConstants(constants *Constants) error {
|
||||
_constants := struct {
|
||||
Constants *Constants `meddler:"constants,json"`
|
||||
@@ -109,7 +105,6 @@ func (hdb *HistoryDB) SetConstants(constants *Constants) error {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetStateInternalAPI returns the StateAPI
|
||||
func (hdb *HistoryDB) GetStateInternalAPI() (*StateAPI, error) {
|
||||
return hdb.getStateAPI(hdb.dbRead)
|
||||
}
|
||||
@@ -120,19 +115,14 @@ func (hdb *HistoryDB) getStateAPI(d meddler.DB) (*StateAPI, error) {
|
||||
d, &nodeInfo,
|
||||
"SELECT state FROM node_info WHERE item_id = 1;",
|
||||
)
|
||||
return nodeInfo.StateAPI, tracerr.Wrap(err)
|
||||
return nodeInfo.APIState, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// SetStateInternalAPI sets the StateAPI
|
||||
func (hdb *HistoryDB) SetStateInternalAPI(stateAPI *StateAPI) error {
|
||||
if stateAPI.Network.LastBatch != nil {
|
||||
stateAPI.Network.LastBatch.CollectedFeesAPI =
|
||||
apitypes.NewCollectedFeesAPI(stateAPI.Network.LastBatch.CollectedFeesDB)
|
||||
}
|
||||
_stateAPI := struct {
|
||||
StateAPI *StateAPI `meddler:"state,json"`
|
||||
}{stateAPI}
|
||||
values, err := meddler.Default.Values(&_stateAPI, false)
|
||||
func (hdb *HistoryDB) SetAPIState(apiState *StateAPI) error {
|
||||
_apiState := struct {
|
||||
APIState *StateAPI `meddler:"state,json"`
|
||||
}{apiState}
|
||||
values, err := meddler.Default.Values(&_apiState, false)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -143,7 +133,6 @@ func (hdb *HistoryDB) SetStateInternalAPI(stateAPI *StateAPI) error {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetNodeConfig returns the NodeConfig
|
||||
func (hdb *HistoryDB) GetNodeConfig() (*NodeConfig, error) {
|
||||
var nodeInfo NodeInfo
|
||||
err := meddler.QueryRow(
|
||||
@@ -153,7 +142,6 @@ func (hdb *HistoryDB) GetNodeConfig() (*NodeConfig, error) {
|
||||
return nodeInfo.NodeConfig, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// SetNodeConfig sets the NodeConfig
|
||||
func (hdb *HistoryDB) SetNodeConfig(nodeConfig *NodeConfig) error {
|
||||
_nodeConfig := struct {
|
||||
NodeConfig *NodeConfig `meddler:"config,json"`
|
||||
@@ -163,8 +151,65 @@ func (hdb *HistoryDB) SetNodeConfig(nodeConfig *NodeConfig) error {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
_, err = hdb.dbWrite.Exec(
|
||||
"UPDATE node_info SET config = $1 WHERE item_id = 1;",
|
||||
"UPDATE config SET state = $1 WHERE item_id = 1;",
|
||||
values[0],
|
||||
)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// func (hdb *HistoryDB) SetInitialNodeInfo(maxPoolTxs uint32, minFeeUSD float64, constants *Constants) error {
|
||||
// ni := &NodeInfo{
|
||||
// MaxPoolTxs: &maxPoolTxs,
|
||||
// MinFeeUSD: &minFeeUSD,
|
||||
// Constants: constants,
|
||||
// }
|
||||
// return tracerr.Wrap(meddler.Insert(hdb.dbWrite, "node_info", ni))
|
||||
// }
|
||||
|
||||
// apiSlotToBigInts converts from [6]*apitypes.BigIntStr to [6]*big.Int
|
||||
// func apiSlotToBigInts(defaultSlotSetBid [6]*apitypes.BigIntStr) ([6]*big.Int, error) {
|
||||
// var slots [6]*big.Int
|
||||
//
|
||||
// for i, slot := range defaultSlotSetBid {
|
||||
// bigInt, ok := new(big.Int).SetString(string(*slot), 10)
|
||||
// if !ok {
|
||||
// return slots, tracerr.Wrap(fmt.Errorf("can't convert %T into big.Int", slot))
|
||||
// }
|
||||
// slots[i] = bigInt
|
||||
// }
|
||||
//
|
||||
// return slots, nil
|
||||
// }
|
||||
|
||||
// func (hdb *HistoryDB) updateNodeInfo(setUpdatedNodeInfo func(*sqlx.Tx, *NodeInfo) error) error {
|
||||
// // Create a SQL transaction or read and update atomicaly
|
||||
// txn, err := hdb.dbWrite.Beginx()
|
||||
// if err != nil {
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
// defer func() {
|
||||
// if err != nil {
|
||||
// db.Rollback(txn)
|
||||
// }
|
||||
// }()
|
||||
// // Read current node info
|
||||
// ni := &NodeInfo{}
|
||||
// if err := meddler.QueryRow(
|
||||
// txn, ni, "SELECT * FROM node_info;",
|
||||
// ); err != nil {
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
// // Update NodeInfo struct
|
||||
// if err := setUpdatedNodeInfo(txn, ni); err != nil {
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
// // Update NodeInfo at DB
|
||||
// if _, err := txn.Exec("DELETE FROM node_info;"); err != nil {
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
// if err := meddler.Insert(txn, "node_info", ni); err != nil {
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
// // Commit NodeInfo update
|
||||
// return tracerr.Wrap(txn.Commit())
|
||||
// }
|
||||
|
||||
@@ -147,12 +147,6 @@ type txWrite struct {
|
||||
Nonce *common.Nonce `meddler:"nonce"`
|
||||
}
|
||||
|
||||
// TokenSymbolAndAddr token representation with only Eth addr and symbol
|
||||
type TokenSymbolAndAddr struct {
|
||||
Symbol string `meddler:"symbol"`
|
||||
Addr ethCommon.Address `meddler:"eth_addr"`
|
||||
}
|
||||
|
||||
// TokenWithUSD add USD info to common.Token
|
||||
type TokenWithUSD struct {
|
||||
ItemID uint64 `json:"itemId" meddler:"item_id"`
|
||||
@@ -289,24 +283,23 @@ func (account AccountAPI) MarshalJSON() ([]byte, error) {
|
||||
// BatchAPI is a representation of a batch with additional information
|
||||
// required by the API, and extracted by joining block table
|
||||
type BatchAPI struct {
|
||||
ItemID uint64 `json:"itemId" meddler:"item_id"`
|
||||
BatchNum common.BatchNum `json:"batchNum" meddler:"batch_num"`
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
EthBlockHash ethCommon.Hash `json:"ethereumBlockHash" meddler:"hash"`
|
||||
Timestamp time.Time `json:"timestamp" meddler:"timestamp,utctime"`
|
||||
ForgerAddr ethCommon.Address `json:"forgerAddr" meddler:"forger_addr"`
|
||||
CollectedFeesDB map[common.TokenID]*big.Int `json:"-" meddler:"fees_collected,json"`
|
||||
CollectedFeesAPI apitypes.CollectedFeesAPI `json:"collectedFees" meddler:"-"`
|
||||
TotalFeesUSD *float64 `json:"historicTotalCollectedFeesUSD" meddler:"total_fees_usd"`
|
||||
StateRoot apitypes.BigIntStr `json:"stateRoot" meddler:"state_root"`
|
||||
NumAccounts int `json:"numAccounts" meddler:"num_accounts"`
|
||||
ExitRoot apitypes.BigIntStr `json:"exitRoot" meddler:"exit_root"`
|
||||
ForgeL1TxsNum *int64 `json:"forgeL1TransactionsNum" meddler:"forge_l1_txs_num"`
|
||||
SlotNum int64 `json:"slotNum" meddler:"slot_num"`
|
||||
ForgedTxs int `json:"forgedTransactions" meddler:"forged_txs"`
|
||||
TotalItems uint64 `json:"-" meddler:"total_items"`
|
||||
FirstItem uint64 `json:"-" meddler:"first_item"`
|
||||
LastItem uint64 `json:"-" meddler:"last_item"`
|
||||
ItemID uint64 `json:"itemId" meddler:"item_id"`
|
||||
BatchNum common.BatchNum `json:"batchNum" meddler:"batch_num"`
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
EthBlockHash ethCommon.Hash `json:"ethereumBlockHash" meddler:"hash"`
|
||||
Timestamp time.Time `json:"timestamp" meddler:"timestamp,utctime"`
|
||||
ForgerAddr ethCommon.Address `json:"forgerAddr" meddler:"forger_addr"`
|
||||
CollectedFees apitypes.CollectedFees `json:"collectedFees" meddler:"fees_collected,json"`
|
||||
TotalFeesUSD *float64 `json:"historicTotalCollectedFeesUSD" meddler:"total_fees_usd"`
|
||||
StateRoot apitypes.BigIntStr `json:"stateRoot" meddler:"state_root"`
|
||||
NumAccounts int `json:"numAccounts" meddler:"num_accounts"`
|
||||
ExitRoot apitypes.BigIntStr `json:"exitRoot" meddler:"exit_root"`
|
||||
ForgeL1TxsNum *int64 `json:"forgeL1TransactionsNum" meddler:"forge_l1_txs_num"`
|
||||
SlotNum int64 `json:"slotNum" meddler:"slot_num"`
|
||||
ForgedTxs int `json:"forgedTransactions" meddler:"forged_txs"`
|
||||
TotalItems uint64 `json:"-" meddler:"total_items"`
|
||||
FirstItem uint64 `json:"-" meddler:"first_item"`
|
||||
LastItem uint64 `json:"-" meddler:"last_item"`
|
||||
}
|
||||
|
||||
// MetricsAPI define metrics of the network
|
||||
@@ -317,7 +310,7 @@ type MetricsAPI struct {
|
||||
TotalAccounts int64 `json:"totalAccounts" meddler:"total_accounts"`
|
||||
TotalBJJs int64 `json:"totalBJJs" meddler:"total_bjjs"`
|
||||
AvgTransactionFee float64 `json:"avgTransactionFee"`
|
||||
EstimatedTimeToForgeL1 float64 `json:"estimatedTimeToForgeL1" meddler:"estimated_time_to_forge_l1"`
|
||||
EstimatedTimeToForgeL1 float64 `json:"estimatedTimeToForgeL1" meddler:"estimatedTimeToForgeL1"`
|
||||
}
|
||||
|
||||
// BidAPI is a representation of a bid with additional information
|
||||
@@ -354,43 +347,38 @@ type BucketUpdateAPI struct {
|
||||
// BucketParamsAPI are the parameter variables of each Bucket of Rollup Smart
|
||||
// Contract
|
||||
type BucketParamsAPI struct {
|
||||
CeilUSD *apitypes.BigIntStr `json:"ceilUSD"`
|
||||
BlockStamp *apitypes.BigIntStr `json:"blockStamp"`
|
||||
Withdrawals *apitypes.BigIntStr `json:"withdrawals"`
|
||||
RateBlocks *apitypes.BigIntStr `json:"rateBlocks"`
|
||||
RateWithdrawals *apitypes.BigIntStr `json:"rateWithdrawals"`
|
||||
MaxWithdrawals *apitypes.BigIntStr `json:"maxWithdrawals"`
|
||||
CeilUSD *apitypes.BigIntStr `json:"ceilUSD"`
|
||||
Withdrawals *apitypes.BigIntStr `json:"withdrawals"`
|
||||
BlockWithdrawalRate *apitypes.BigIntStr `json:"blockWithdrawalRate"`
|
||||
MaxWithdrawals *apitypes.BigIntStr `json:"maxWithdrawals"`
|
||||
}
|
||||
|
||||
// RollupVariablesAPI are the variables of the Rollup Smart Contract
|
||||
type RollupVariablesAPI struct {
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
FeeAddToken *apitypes.BigIntStr `json:"feeAddToken" meddler:"fee_add_token" validate:"required"`
|
||||
ForgeL1L2BatchTimeout int64 `json:"forgeL1L2BatchTimeout" meddler:"forge_l1_timeout" validate:"required"`
|
||||
WithdrawalDelay uint64 `json:"withdrawalDelay" meddler:"withdrawal_delay" validate:"required"`
|
||||
Buckets []BucketParamsAPI `json:"buckets" meddler:"buckets,json"`
|
||||
SafeMode bool `json:"safeMode" meddler:"safe_mode"`
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
FeeAddToken *apitypes.BigIntStr `json:"feeAddToken" meddler:"fee_add_token" validate:"required"`
|
||||
ForgeL1L2BatchTimeout int64 `json:"forgeL1L2BatchTimeout" meddler:"forge_l1_timeout" validate:"required"`
|
||||
WithdrawalDelay uint64 `json:"withdrawalDelay" meddler:"withdrawal_delay" validate:"required"`
|
||||
Buckets [common.RollupConstNumBuckets]BucketParamsAPI `json:"buckets" meddler:"buckets,json"`
|
||||
SafeMode bool `json:"safeMode" meddler:"safe_mode"`
|
||||
}
|
||||
|
||||
// NewRollupVariablesAPI creates a RollupVariablesAPI from common.RollupVariables
|
||||
func NewRollupVariablesAPI(rollupVariables *common.RollupVariables) *RollupVariablesAPI {
|
||||
buckets := make([]BucketParamsAPI, len(rollupVariables.Buckets))
|
||||
rollupVars := RollupVariablesAPI{
|
||||
EthBlockNum: rollupVariables.EthBlockNum,
|
||||
FeeAddToken: apitypes.NewBigIntStr(rollupVariables.FeeAddToken),
|
||||
ForgeL1L2BatchTimeout: rollupVariables.ForgeL1L2BatchTimeout,
|
||||
WithdrawalDelay: rollupVariables.WithdrawalDelay,
|
||||
SafeMode: rollupVariables.SafeMode,
|
||||
Buckets: buckets,
|
||||
}
|
||||
|
||||
for i, bucket := range rollupVariables.Buckets {
|
||||
rollupVars.Buckets[i] = BucketParamsAPI{
|
||||
CeilUSD: apitypes.NewBigIntStr(bucket.CeilUSD),
|
||||
BlockStamp: apitypes.NewBigIntStr(bucket.BlockStamp),
|
||||
Withdrawals: apitypes.NewBigIntStr(bucket.Withdrawals),
|
||||
RateBlocks: apitypes.NewBigIntStr(bucket.RateBlocks),
|
||||
RateWithdrawals: apitypes.NewBigIntStr(bucket.RateWithdrawals),
|
||||
MaxWithdrawals: apitypes.NewBigIntStr(bucket.MaxWithdrawals),
|
||||
CeilUSD: apitypes.NewBigIntStr(bucket.CeilUSD),
|
||||
Withdrawals: apitypes.NewBigIntStr(bucket.Withdrawals),
|
||||
BlockWithdrawalRate: apitypes.NewBigIntStr(bucket.BlockWithdrawalRate),
|
||||
MaxWithdrawals: apitypes.NewBigIntStr(bucket.MaxWithdrawals),
|
||||
}
|
||||
}
|
||||
return &rollupVars
|
||||
|
||||
@@ -316,7 +316,7 @@ func (k *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKVDB
|
||||
|
||||
checkpointPath := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
|
||||
// copy synchronizer 'BatchNumX' to 'BatchNumX'
|
||||
// copy synchronizer'BatchNumX' to 'BatchNumX'
|
||||
if err := synchronizerKVDB.MakeCheckpointFromTo(batchNum, checkpointPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -458,7 +458,7 @@ func (k *KVDB) CheckpointExists(batchNum common.BatchNum) (bool, error) {
|
||||
if _, err := os.Stat(source); os.IsNotExist(err) {
|
||||
return false, nil
|
||||
} else if err != nil {
|
||||
return false, tracerr.Wrap(err)
|
||||
return false, err
|
||||
}
|
||||
return true, nil
|
||||
}
|
||||
@@ -544,12 +544,10 @@ func (k *KVDB) MakeCheckpointFromTo(fromBatchNum common.BatchNum, dest string) e
|
||||
// synchronizer to the same batchNum
|
||||
k.m.Lock()
|
||||
defer k.m.Unlock()
|
||||
return PebbleMakeCheckpoint(source, dest)
|
||||
return pebbleMakeCheckpoint(source, dest)
|
||||
}
|
||||
|
||||
// PebbleMakeCheckpoint is a hepler function to make a pebble checkpoint from
|
||||
// source to dest.
|
||||
func PebbleMakeCheckpoint(source, dest string) error {
|
||||
func pebbleMakeCheckpoint(source, dest string) error {
|
||||
// Remove dest folder (if it exists) before doing the checkpoint
|
||||
if _, err := os.Stat(dest); os.IsNotExist(err) {
|
||||
} else if err != nil {
|
||||
|
||||
@@ -80,7 +80,7 @@ func (l2db *L2DB) AddTxAPI(tx *PoolL2TxWrite) error {
|
||||
q := fmt.Sprintf(
|
||||
`INSERT INTO tx_pool (%s)
|
||||
SELECT %s
|
||||
WHERE (SELECT COUNT(*) FROM tx_pool WHERE state = $%v AND NOT external_delete) < $%v;`,
|
||||
WHERE (SELECT COUNT(*) FROM tx_pool WHERE state = $%v) < $%v;`,
|
||||
namesPart, valuesPart,
|
||||
len(values)+1, len(values)+2) //nolint:gomnd
|
||||
values = append(values, common.PoolL2TxStatePending, l2db.maxTxs)
|
||||
|
||||
@@ -74,16 +74,6 @@ func (l2db *L2DB) AddAccountCreationAuth(auth *common.AccountCreationAuth) error
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// AddManyAccountCreationAuth inserts a batch of accounts creation authorization
|
||||
// if not exist into the DB
|
||||
func (l2db *L2DB) AddManyAccountCreationAuth(auths []common.AccountCreationAuth) error {
|
||||
_, err := sqlx.NamedExec(l2db.dbWrite,
|
||||
`INSERT INTO account_creation_auth (eth_addr, bjj, signature)
|
||||
VALUES (:ethaddr, :bjj, :signature)
|
||||
ON CONFLICT (eth_addr) DO NOTHING`, auths)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetAccountCreationAuth returns an account creation authorization from the DB
|
||||
func (l2db *L2DB) GetAccountCreationAuth(addr ethCommon.Address) (*common.AccountCreationAuth, error) {
|
||||
auth := new(common.AccountCreationAuth)
|
||||
@@ -204,7 +194,7 @@ func (l2db *L2DB) GetPendingTxs() ([]common.PoolL2Tx, error) {
|
||||
var txs []*common.PoolL2Tx
|
||||
err := meddler.QueryAll(
|
||||
l2db.dbRead, &txs,
|
||||
selectPoolTxCommon+"WHERE state = $1 AND NOT external_delete;",
|
||||
selectPoolTxCommon+"WHERE state = $1",
|
||||
common.PoolL2TxStatePending,
|
||||
)
|
||||
return db.SlicePtrsToSlice(txs).([]common.PoolL2Tx), tracerr.Wrap(err)
|
||||
@@ -320,7 +310,7 @@ func (l2db *L2DB) InvalidateOldNonces(updatedAccounts []common.IdxNonce, batchNu
|
||||
return nil
|
||||
}
|
||||
// Fill the batch_num in the query with Sprintf because we are using a
|
||||
// named query which works with slices, and doesn't handle an extra
|
||||
// named query which works with slices, and doens't handle an extra
|
||||
// individual argument.
|
||||
query := fmt.Sprintf(invalidateOldNoncesQuery, batchNum)
|
||||
if _, err := sqlx.NamedExec(l2db.dbWrite, query, updatedAccounts); err != nil {
|
||||
|
||||
@@ -38,7 +38,7 @@ func TestMain(m *testing.M) {
|
||||
panic(err)
|
||||
}
|
||||
l2DB = NewL2DB(db, db, 10, 1000, 0.0, 24*time.Hour, nil)
|
||||
apiConnCon := dbUtils.NewAPIConnectionController(1, time.Second)
|
||||
apiConnCon := dbUtils.NewAPICnnectionController(1, time.Second)
|
||||
l2DBWithACC = NewL2DB(db, db, 10, 1000, 0.0, 24*time.Hour, apiConnCon)
|
||||
test.WipeDB(l2DB.DB())
|
||||
historyDB = historydb.NewHistoryDB(db, db, nil)
|
||||
@@ -121,7 +121,7 @@ func prepareHistoryDB(historyDB *historydb.HistoryDB) error {
|
||||
}
|
||||
tokens[token.TokenID] = readToken
|
||||
// Set value to the tokens
|
||||
err := historyDB.UpdateTokenValue(readToken.EthAddr, *readToken.USD)
|
||||
err := historyDB.UpdateTokenValue(readToken.Symbol, *readToken.USD)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -725,43 +725,6 @@ func TestAuth(t *testing.T) {
|
||||
}
|
||||
}
|
||||
|
||||
func TestManyAuth(t *testing.T) {
|
||||
test.WipeDB(l2DB.DB())
|
||||
const nAuths = 5
|
||||
chainID := uint16(0)
|
||||
hermezContractAddr := ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
||||
// Generate authorizations
|
||||
genAuths := test.GenAuths(nAuths, chainID, hermezContractAddr)
|
||||
auths := make([]common.AccountCreationAuth, len(genAuths))
|
||||
// Convert to a non-pointer slice
|
||||
for i := 0; i < len(genAuths); i++ {
|
||||
auths[i] = *genAuths[i]
|
||||
}
|
||||
|
||||
// Add a duplicate one to check the not exist condition
|
||||
err := l2DB.AddAccountCreationAuth(genAuths[0])
|
||||
require.NoError(t, err)
|
||||
|
||||
// Add to the DB
|
||||
err = l2DB.AddManyAccountCreationAuth(auths)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Assert the result
|
||||
for i := 0; i < len(auths); i++ {
|
||||
// Fetch from DB
|
||||
auth, err := l2DB.GetAccountCreationAuth(auths[i].EthAddr)
|
||||
require.NoError(t, err)
|
||||
// Check fetched vs generated
|
||||
assert.Equal(t, auths[i].EthAddr, auth.EthAddr)
|
||||
assert.Equal(t, auths[i].BJJ, auth.BJJ)
|
||||
assert.Equal(t, auths[i].Signature, auth.Signature)
|
||||
assert.Equal(t, auths[i].Timestamp.Unix(), auths[i].Timestamp.Unix())
|
||||
nameZone, offset := auths[i].Timestamp.Zone()
|
||||
assert.Equal(t, "UTC", nameZone)
|
||||
assert.Equal(t, 0, offset)
|
||||
}
|
||||
}
|
||||
|
||||
func TestAddGet(t *testing.T) {
|
||||
err := prepareHistoryDB(historyDB)
|
||||
if err != nil {
|
||||
|
||||
@@ -1,11 +1,5 @@
|
||||
-- +migrate Up
|
||||
|
||||
-- NOTE: We use "DECIMAL(78,0)" to encode go *big.Int types. All the *big.Int
|
||||
-- that we deal with represent a value in the SNARK field, which is an integer
|
||||
-- of 256 bits. `log(2**256, 10) = 77.06`: that is, a 256 bit number can have
|
||||
-- at most 78 digits, so we use this value to specify the precision in the
|
||||
-- PostgreSQL DECIMAL guaranteeing that we will never lose precision.
|
||||
|
||||
-- History
|
||||
CREATE TABLE block (
|
||||
eth_block_num BIGINT PRIMARY KEY,
|
||||
@@ -28,10 +22,10 @@ CREATE TABLE batch (
|
||||
forger_addr BYTEA NOT NULL, -- fake foreign key for coordinator
|
||||
fees_collected BYTEA NOT NULL,
|
||||
fee_idxs_coordinator BYTEA NOT NULL,
|
||||
state_root DECIMAL(78,0) NOT NULL,
|
||||
state_root BYTEA NOT NULL,
|
||||
num_accounts BIGINT NOT NULL,
|
||||
last_idx BIGINT NOT NULL,
|
||||
exit_root DECIMAL(78,0) NOT NULL,
|
||||
exit_root BYTEA NOT NULL,
|
||||
forge_l1_txs_num BIGINT,
|
||||
slot_num BIGINT NOT NULL,
|
||||
total_fees_usd NUMERIC
|
||||
@@ -40,7 +34,7 @@ CREATE TABLE batch (
|
||||
CREATE TABLE bid (
|
||||
item_id SERIAL PRIMARY KEY,
|
||||
slot_num BIGINT NOT NULL,
|
||||
bid_value DECIMAL(78,0) NOT NULL,
|
||||
bid_value BYTEA NOT NULL,
|
||||
eth_block_num BIGINT NOT NULL REFERENCES block (eth_block_num) ON DELETE CASCADE,
|
||||
bidder_addr BYTEA NOT NULL -- fake foreign key for coordinator
|
||||
);
|
||||
@@ -112,7 +106,7 @@ CREATE TABLE account_update (
|
||||
batch_num BIGINT NOT NULL REFERENCES batch (batch_num) ON DELETE CASCADE,
|
||||
idx BIGINT NOT NULL REFERENCES account (idx) ON DELETE CASCADE,
|
||||
nonce BIGINT NOT NULL,
|
||||
balance DECIMAL(78,0) NOT NULL
|
||||
balance BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE exit_tree (
|
||||
@@ -120,7 +114,7 @@ CREATE TABLE exit_tree (
|
||||
batch_num BIGINT REFERENCES batch (batch_num) ON DELETE CASCADE,
|
||||
account_idx BIGINT REFERENCES account (idx) ON DELETE CASCADE,
|
||||
merkle_proof BYTEA NOT NULL,
|
||||
balance DECIMAL(78,0) NOT NULL,
|
||||
balance BYTEA NOT NULL,
|
||||
instant_withdrawn BIGINT REFERENCES block (eth_block_num) ON DELETE SET NULL,
|
||||
delayed_withdraw_request BIGINT REFERENCES block (eth_block_num) ON DELETE SET NULL,
|
||||
owner BYTEA,
|
||||
@@ -170,7 +164,7 @@ CREATE TABLE tx (
|
||||
to_idx BIGINT NOT NULL,
|
||||
to_eth_addr BYTEA,
|
||||
to_bjj BYTEA,
|
||||
amount DECIMAL(78,0) NOT NULL,
|
||||
amount BYTEA NOT NULL,
|
||||
amount_success BOOLEAN NOT NULL DEFAULT true,
|
||||
amount_f NUMERIC NOT NULL,
|
||||
token_id INT NOT NULL REFERENCES token (token_id),
|
||||
@@ -180,7 +174,7 @@ CREATE TABLE tx (
|
||||
-- L1
|
||||
to_forge_l1_txs_num BIGINT,
|
||||
user_origin BOOLEAN,
|
||||
deposit_amount DECIMAL(78,0),
|
||||
deposit_amount BYTEA,
|
||||
deposit_amount_success BOOLEAN NOT NULL DEFAULT true,
|
||||
deposit_amount_f NUMERIC,
|
||||
deposit_amount_usd NUMERIC,
|
||||
@@ -550,7 +544,7 @@ FOR EACH ROW EXECUTE PROCEDURE forge_l1_user_txs();
|
||||
|
||||
CREATE TABLE rollup_vars (
|
||||
eth_block_num BIGINT PRIMARY KEY REFERENCES block (eth_block_num) ON DELETE CASCADE,
|
||||
fee_add_token DECIMAL(78,0) NOT NULL,
|
||||
fee_add_token BYTEA NOT NULL,
|
||||
forge_l1_timeout BIGINT NOT NULL,
|
||||
withdrawal_delay BIGINT NOT NULL,
|
||||
buckets BYTEA NOT NULL,
|
||||
@@ -562,7 +556,7 @@ CREATE TABLE bucket_update (
|
||||
eth_block_num BIGINT NOT NULL REFERENCES block (eth_block_num) ON DELETE CASCADE,
|
||||
num_bucket BIGINT NOT NULL,
|
||||
block_stamp BIGINT NOT NULL,
|
||||
withdrawals DECIMAL(78,0) NOT NULL
|
||||
withdrawals BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE token_exchange (
|
||||
@@ -578,7 +572,7 @@ CREATE TABLE escape_hatch_withdrawal (
|
||||
who_addr BYTEA NOT NULL,
|
||||
to_addr BYTEA NOT NULL,
|
||||
token_addr BYTEA NOT NULL,
|
||||
amount DECIMAL(78,0) NOT NULL
|
||||
amount BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE auction_vars (
|
||||
@@ -616,7 +610,7 @@ CREATE TABLE tx_pool (
|
||||
effective_to_eth_addr BYTEA,
|
||||
effective_to_bjj BYTEA,
|
||||
token_id INT NOT NULL REFERENCES token (token_id) ON DELETE CASCADE,
|
||||
amount DECIMAL(78,0) NOT NULL,
|
||||
amount BYTEA NOT NULL,
|
||||
amount_f NUMERIC NOT NULL,
|
||||
fee SMALLINT NOT NULL,
|
||||
nonce BIGINT NOT NULL,
|
||||
@@ -630,7 +624,7 @@ CREATE TABLE tx_pool (
|
||||
rq_to_eth_addr BYTEA,
|
||||
rq_to_bjj BYTEA,
|
||||
rq_token_id INT,
|
||||
rq_amount DECIMAL(78,0),
|
||||
rq_amount BYTEA,
|
||||
rq_fee SMALLINT,
|
||||
rq_nonce BIGINT,
|
||||
tx_type VARCHAR(40) NOT NULL,
|
||||
|
||||
@@ -17,8 +17,7 @@ import (
|
||||
var (
|
||||
// ErrStateDBWithoutMT is used when a method that requires a MerkleTree
|
||||
// is called in a StateDB that does not have a MerkleTree defined
|
||||
ErrStateDBWithoutMT = errors.New(
|
||||
"Can not call method to use MerkleTree in a StateDB without MerkleTree")
|
||||
ErrStateDBWithoutMT = errors.New("Can not call method to use MerkleTree in a StateDB without MerkleTree")
|
||||
|
||||
// ErrAccountAlreadyExists is used when CreateAccount is called and the
|
||||
// Account already exists
|
||||
@@ -29,8 +28,7 @@ var (
|
||||
ErrIdxNotFound = errors.New("Idx can not be found")
|
||||
// ErrGetIdxNoCase is used when trying to get the Idx from EthAddr &
|
||||
// BJJ with not compatible combination
|
||||
ErrGetIdxNoCase = errors.New(
|
||||
"Can not get Idx due unexpected combination of ethereum Address & BabyJubJub PublicKey")
|
||||
ErrGetIdxNoCase = errors.New("Can not get Idx due unexpected combination of ethereum Address & BabyJubJub PublicKey")
|
||||
|
||||
// PrefixKeyIdx is the key prefix for idx in the db
|
||||
PrefixKeyIdx = []byte("i:")
|
||||
@@ -146,8 +144,7 @@ func NewStateDB(cfg Config) (*StateDB, error) {
|
||||
}
|
||||
}
|
||||
if cfg.Type == TypeTxSelector && cfg.NLevels != 0 {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("invalid StateDB parameters: StateDB type==TypeStateDB can not have nLevels!=0"))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid StateDB parameters: StateDB type==TypeStateDB can not have nLevels!=0"))
|
||||
}
|
||||
|
||||
return &StateDB{
|
||||
@@ -350,8 +347,7 @@ func GetAccountInTreeDB(sto db.Storage, idx common.Idx) (*common.Account, error)
|
||||
// CreateAccount creates a new Account in the StateDB for the given Idx. If
|
||||
// StateDB.MT==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (
|
||||
*merkletree.CircomProcessorProof, error) {
|
||||
func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
cpp, err := CreateAccountInTreeDB(s.db.DB(), s.MT, idx, account)
|
||||
if err != nil {
|
||||
return cpp, tracerr.Wrap(err)
|
||||
@@ -365,8 +361,7 @@ func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (
|
||||
// from ExitTree. Creates a new Account in the StateDB for the given Idx. If
|
||||
// StateDB.MT==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx,
|
||||
account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
// store at the DB the key: v, and value: leaf.Bytes()
|
||||
v, err := account.HashValue()
|
||||
if err != nil {
|
||||
@@ -415,8 +410,7 @@ func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common
|
||||
// UpdateAccount updates the Account in the StateDB for the given Idx. If
|
||||
// StateDB.mt==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (
|
||||
*merkletree.CircomProcessorProof, error) {
|
||||
func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
return UpdateAccountInTreeDB(s.db.DB(), s.MT, idx, account)
|
||||
}
|
||||
|
||||
@@ -424,8 +418,7 @@ func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (
|
||||
// from ExitTree. Updates the Account in the StateDB for the given Idx. If
|
||||
// StateDB.mt==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func UpdateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx,
|
||||
account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func UpdateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
// store at the DB the key: v, and value: account.Bytes()
|
||||
v, err := account.HashValue()
|
||||
if err != nil {
|
||||
@@ -510,7 +503,7 @@ func (l *LocalStateDB) CheckpointExists(batchNum common.BatchNum) (bool, error)
|
||||
return l.db.CheckpointExists(batchNum)
|
||||
}
|
||||
|
||||
// Reset performs a reset in the LocalStateDB. If fromSynchronizer is true, it
|
||||
// Reset performs a reset in the LocaStateDB. If fromSynchronizer is true, it
|
||||
// gets the state from LocalStateDB.synchronizerStateDB for the given batchNum.
|
||||
// If fromSynchronizer is false, get the state from LocalStateDB checkpoints.
|
||||
func (l *LocalStateDB) Reset(batchNum common.BatchNum, fromSynchronizer bool) error {
|
||||
|
||||
@@ -22,8 +22,7 @@ import (
|
||||
|
||||
func newAccount(t *testing.T, i int) *common.Account {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
require.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -372,8 +371,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal, err := ioutil.TempDir("", "ldb")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal))
|
||||
ldb, err := NewLocalStateDB(Config{Path: dirLocal, Keep: 128, Type: TypeBatchBuilder,
|
||||
NLevels: 32}, sdb)
|
||||
ldb, err := NewLocalStateDB(Config{Path: dirLocal, Keep: 128, Type: TypeBatchBuilder, NLevels: 32}, sdb)
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
@@ -394,8 +392,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal2, err := ioutil.TempDir("", "ldb2")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal2))
|
||||
ldb2, err := NewLocalStateDB(Config{Path: dirLocal2, Keep: 128, Type: TypeBatchBuilder,
|
||||
NLevels: 32}, sdb)
|
||||
ldb2, err := NewLocalStateDB(Config{Path: dirLocal2, Keep: 128, Type: TypeBatchBuilder, NLevels: 32}, sdb)
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
@@ -474,8 +471,7 @@ func TestCheckAccountsTreeTestVectors(t *testing.T) {
|
||||
|
||||
ay0 := new(big.Int).Sub(new(big.Int).Exp(big.NewInt(2), big.NewInt(253), nil), big.NewInt(1))
|
||||
// test value from js version (compatibility-canary)
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff",
|
||||
(hex.EncodeToString(ay0.Bytes())))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff", (hex.EncodeToString(ay0.Bytes())))
|
||||
bjjPoint0Comp := babyjub.PackSignY(true, ay0)
|
||||
bjj0 := babyjub.PublicKeyComp(bjjPoint0Comp)
|
||||
|
||||
@@ -534,9 +530,7 @@ func TestCheckAccountsTreeTestVectors(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
}
|
||||
// root value generated by js version:
|
||||
assert.Equal(t,
|
||||
"13174362770971232417413036794215823584762073355951212910715422236001731746065",
|
||||
sdb.MT.Root().BigInt().String())
|
||||
assert.Equal(t, "17298264051379321456969039521810887093935433569451713402227686942080129181291", sdb.MT.Root().BigInt().String())
|
||||
}
|
||||
|
||||
// TestListCheckpoints performs almost the same test than kvdb/kvdb_test.go
|
||||
|
||||
@@ -18,8 +18,7 @@ func concatEthAddrTokenID(addr ethCommon.Address, tokenID common.TokenID) []byte
|
||||
b = append(b[:], tokenID.Bytes()[:]...)
|
||||
return b
|
||||
}
|
||||
func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp,
|
||||
tokenID common.TokenID) []byte {
|
||||
func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp, tokenID common.TokenID) []byte {
|
||||
pkComp := pk
|
||||
var b []byte
|
||||
b = append(b, addr.Bytes()...)
|
||||
@@ -33,8 +32,7 @@ func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp,
|
||||
// - key: EthAddr & BabyJubJub PublicKey Compressed, value: idx
|
||||
// If Idx already exist for the given EthAddr & BJJ, the remaining Idx will be
|
||||
// always the smallest one.
|
||||
func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address,
|
||||
pk babyjub.PublicKeyComp, tokenID common.TokenID) error {
|
||||
func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address, pk babyjub.PublicKeyComp, tokenID common.TokenID) error {
|
||||
oldIdx, err := s.GetIdxByEthAddrBJJ(addr, pk, tokenID)
|
||||
if err == nil {
|
||||
// EthAddr & BJJ already have an Idx
|
||||
@@ -42,8 +40,7 @@ func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address,
|
||||
// if new idx is smaller, store the new one
|
||||
// if new idx is bigger, don't store and return, as the used one will be the old
|
||||
if idx >= oldIdx {
|
||||
log.Debug("StateDB.setIdxByEthAddrBJJ: Idx not stored because there " +
|
||||
"already exist a smaller Idx for the given EthAddr & BJJ")
|
||||
log.Debug("StateDB.setIdxByEthAddrBJJ: Idx not stored because there already exist a smaller Idx for the given EthAddr & BJJ")
|
||||
return nil
|
||||
}
|
||||
}
|
||||
@@ -83,8 +80,7 @@ func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address,
|
||||
// GetIdxByEthAddr returns the smallest Idx in the StateDB for the given
|
||||
// Ethereum Address. Will return common.Idx(0) and error in case that Idx is
|
||||
// not found in the StateDB.
|
||||
func (s *StateDB) GetIdxByEthAddr(addr ethCommon.Address, tokenID common.TokenID) (common.Idx,
|
||||
error) {
|
||||
func (s *StateDB) GetIdxByEthAddr(addr ethCommon.Address, tokenID common.TokenID) (common.Idx, error) {
|
||||
k := concatEthAddrTokenID(addr, tokenID)
|
||||
b, err := s.db.DB().Get(append(PrefixKeyAddr, k...))
|
||||
if err != nil {
|
||||
@@ -120,22 +116,18 @@ func (s *StateDB) GetIdxByEthAddrBJJ(addr ethCommon.Address, pk babyjub.PublicKe
|
||||
return common.Idx(0), tracerr.Wrap(ErrIdxNotFound)
|
||||
} else if err != nil {
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
ErrIdxNotFound, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrIdxNotFound, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
idx, err := common.IdxFromBytes(b)
|
||||
if err != nil {
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
err, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", err, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
return idx, nil
|
||||
}
|
||||
// rest of cases (included case ToEthAddr==0) are not possible
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(
|
||||
fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
|
||||
// GetTokenIDsFromIdxs returns a map containing the common.TokenID with its
|
||||
@@ -145,9 +137,7 @@ func (s *StateDB) GetTokenIDsFromIdxs(idxs []common.Idx) (map[common.TokenID]com
|
||||
for i := 0; i < len(idxs); i++ {
|
||||
a, err := s.GetAccount(idxs[i])
|
||||
if err != nil {
|
||||
return nil,
|
||||
tracerr.Wrap(fmt.Errorf("GetTokenIDsFromIdxs error on GetAccount with Idx==%d: %s",
|
||||
idxs[i], err.Error()))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("GetTokenIDsFromIdxs error on GetAccount with Idx==%d: %s", idxs[i], err.Error()))
|
||||
}
|
||||
m[a.TokenID] = idxs[i]
|
||||
}
|
||||
|
||||
24
db/utils.go
24
db/utils.go
@@ -13,9 +13,6 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/jmoiron/sqlx"
|
||||
|
||||
//nolint:errcheck // driver for postgres DB
|
||||
_ "github.com/lib/pq"
|
||||
migrate "github.com/rubenv/sql-migrate"
|
||||
"github.com/russross/meddler"
|
||||
"golang.org/x/sync/semaphore"
|
||||
@@ -96,8 +93,8 @@ type APIConnectionController struct {
|
||||
timeout time.Duration
|
||||
}
|
||||
|
||||
// NewAPIConnectionController initialize APIConnectionController
|
||||
func NewAPIConnectionController(maxConnections int, timeout time.Duration) *APIConnectionController {
|
||||
// NewAPICnnectionController initialize APIConnectionController
|
||||
func NewAPICnnectionController(maxConnections int, timeout time.Duration) *APIConnectionController {
|
||||
return &APIConnectionController{
|
||||
smphr: semaphore.NewWeighted(int64(maxConnections)),
|
||||
timeout: timeout,
|
||||
@@ -168,11 +165,7 @@ func (b BigIntMeddler) PostRead(fieldPtr, scanTarget interface{}) error {
|
||||
return tracerr.Wrap(fmt.Errorf("BigIntMeddler.PostRead: nil pointer"))
|
||||
}
|
||||
field := fieldPtr.(**big.Int)
|
||||
var ok bool
|
||||
*field, ok = new(big.Int).SetString(*ptr, 10)
|
||||
if !ok {
|
||||
return tracerr.Wrap(fmt.Errorf("big.Int.SetString failed on \"%v\"", *ptr))
|
||||
}
|
||||
*field = new(big.Int).SetBytes([]byte(*ptr))
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -180,7 +173,7 @@ func (b BigIntMeddler) PostRead(fieldPtr, scanTarget interface{}) error {
|
||||
func (b BigIntMeddler) PreWrite(fieldPtr interface{}) (saveValue interface{}, err error) {
|
||||
field := fieldPtr.(*big.Int)
|
||||
|
||||
return field.String(), nil
|
||||
return field.Bytes(), nil
|
||||
}
|
||||
|
||||
// BigIntNullMeddler encodes or decodes the field value to or from JSON
|
||||
@@ -205,12 +198,7 @@ func (b BigIntNullMeddler) PostRead(fieldPtr, scanTarget interface{}) error {
|
||||
if ptr == nil {
|
||||
return tracerr.Wrap(fmt.Errorf("BigIntMeddler.PostRead: nil pointer"))
|
||||
}
|
||||
var ok bool
|
||||
*field, ok = new(big.Int).SetString(string(ptr), 10)
|
||||
if !ok {
|
||||
return tracerr.Wrap(fmt.Errorf("big.Int.SetString failed on \"%v\"", string(ptr)))
|
||||
}
|
||||
|
||||
*field = new(big.Int).SetBytes(ptr)
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -220,7 +208,7 @@ func (b BigIntNullMeddler) PreWrite(fieldPtr interface{}) (saveValue interface{}
|
||||
if field == nil {
|
||||
return nil, nil
|
||||
}
|
||||
return field.String(), nil
|
||||
return field.Bytes(), nil
|
||||
}
|
||||
|
||||
// SliceToSlicePtrs converts any []Foo to []*Foo
|
||||
|
||||
@@ -1,13 +1,9 @@
|
||||
package db
|
||||
|
||||
import (
|
||||
"math/big"
|
||||
"os"
|
||||
"testing"
|
||||
|
||||
"github.com/russross/meddler"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
type foo struct {
|
||||
@@ -37,42 +33,3 @@ func TestSlicePtrsToSlice(t *testing.T) {
|
||||
assert.Equal(t, *a[i], b[i])
|
||||
}
|
||||
}
|
||||
|
||||
func TestBigInt(t *testing.T) {
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
db, err := InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
defer func() {
|
||||
_, err := db.Exec("DROP TABLE IF EXISTS test_big_int;")
|
||||
require.NoError(t, err)
|
||||
err = db.Close()
|
||||
require.NoError(t, err)
|
||||
}()
|
||||
|
||||
_, err = db.Exec("DROP TABLE IF EXISTS test_big_int;")
|
||||
require.NoError(t, err)
|
||||
|
||||
_, err = db.Exec(`CREATE TABLE test_big_int (
|
||||
item_id SERIAL PRIMARY KEY,
|
||||
value1 DECIMAL(78, 0) NOT NULL,
|
||||
value2 DECIMAL(78, 0),
|
||||
value3 DECIMAL(78, 0)
|
||||
);`)
|
||||
require.NoError(t, err)
|
||||
|
||||
type Entry struct {
|
||||
ItemID int `meddler:"item_id"`
|
||||
Value1 *big.Int `meddler:"value1,bigint"`
|
||||
Value2 *big.Int `meddler:"value2,bigintnull"`
|
||||
Value3 *big.Int `meddler:"value3,bigintnull"`
|
||||
}
|
||||
|
||||
entry := Entry{ItemID: 1, Value1: big.NewInt(1234567890), Value2: big.NewInt(9876543210), Value3: nil}
|
||||
err = meddler.Insert(db, "test_big_int", &entry)
|
||||
require.NoError(t, err)
|
||||
|
||||
var dbEntry Entry
|
||||
err = meddler.QueryRow(db, &dbEntry, "SELECT * FROM test_big_int WHERE item_id = 1;")
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, entry, dbEntry)
|
||||
}
|
||||
|
||||
@@ -8,7 +8,7 @@ The first step is to clone the github repository where the contracts are located
|
||||
|
||||
While the prepared deployment is not found to master, branch in repository must be changed:
|
||||
|
||||
`git checkout feature/newDeploymentScript-eth-mik` (tested with commit `0eccd237dda102a4ad788a6e11f98361d39d0d9c`)
|
||||
`git checkout feature/newDeploymentScript-eth-edu` (tested with commit `e6c5b7db8da2de1b9cc55e281c8d1dfa524b06f0`)
|
||||
|
||||
Now, install the dependencies:
|
||||
|
||||
|
||||
186
eth/auction.go
186
eth/auction.go
@@ -70,8 +70,7 @@ type AuctionEventInitialize struct {
|
||||
}
|
||||
|
||||
// AuctionVariables returns the AuctionVariables from the initialize event
|
||||
func (ei *AuctionEventInitialize) AuctionVariables(
|
||||
InitialMinimalBidding *big.Int) *common.AuctionVariables {
|
||||
func (ei *AuctionEventInitialize) AuctionVariables(InitialMinimalBidding *big.Int) *common.AuctionVariables {
|
||||
return &common.AuctionVariables{
|
||||
EthBlockNum: 0,
|
||||
DonationAddress: ei.DonationAddress,
|
||||
@@ -223,15 +222,12 @@ type AuctionInterface interface {
|
||||
AuctionGetAllocationRatio() ([3]uint16, error)
|
||||
AuctionSetDonationAddress(newDonationAddress ethCommon.Address) (*types.Transaction, error)
|
||||
AuctionGetDonationAddress() (*ethCommon.Address, error)
|
||||
AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address,
|
||||
newBootCoordinatorURL string) (*types.Transaction, error)
|
||||
AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address, newBootCoordinatorURL string) (*types.Transaction, error)
|
||||
AuctionGetBootCoordinator() (*ethCommon.Address, error)
|
||||
AuctionChangeDefaultSlotSetBid(slotSet int64,
|
||||
newInitialMinBid *big.Int) (*types.Transaction, error)
|
||||
AuctionChangeDefaultSlotSetBid(slotSet int64, newInitialMinBid *big.Int) (*types.Transaction, error)
|
||||
|
||||
// Coordinator Management
|
||||
AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (*types.Transaction,
|
||||
error)
|
||||
AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (*types.Transaction, error)
|
||||
|
||||
// Slot Info
|
||||
AuctionGetSlotNumber(blockNum int64) (int64, error)
|
||||
@@ -241,8 +237,7 @@ type AuctionInterface interface {
|
||||
AuctionGetSlotSet(slot int64) (*big.Int, error)
|
||||
|
||||
// Bidding
|
||||
AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (
|
||||
tx *types.Transaction, err error)
|
||||
AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64, slotSets [6]bool,
|
||||
maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
|
||||
@@ -280,10 +275,8 @@ type AuctionClient struct {
|
||||
}
|
||||
|
||||
// NewAuctionClient creates a new AuctionClient. `tokenAddress` is the address of the HEZ tokens.
|
||||
func NewAuctionClient(client *EthereumClient, address ethCommon.Address,
|
||||
tokenHEZCfg TokenConfig) (*AuctionClient, error) {
|
||||
contractAbi, err :=
|
||||
abi.JSON(strings.NewReader(string(HermezAuctionProtocol.HermezAuctionProtocolABI)))
|
||||
func NewAuctionClient(client *EthereumClient, address ethCommon.Address, tokenHEZCfg TokenConfig) (*AuctionClient, error) {
|
||||
contractAbi, err := abi.JSON(strings.NewReader(string(HermezAuctionProtocol.HermezAuctionProtocolABI)))
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -338,8 +331,7 @@ func (c *AuctionClient) AuctionGetSlotDeadline() (slotDeadline uint8, err error)
|
||||
}
|
||||
|
||||
// AuctionSetOpenAuctionSlots is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetOpenAuctionSlots(
|
||||
newOpenAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetOpenAuctionSlots(newOpenAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -363,8 +355,7 @@ func (c *AuctionClient) AuctionGetOpenAuctionSlots() (openAuctionSlots uint16, e
|
||||
}
|
||||
|
||||
// AuctionSetClosedAuctionSlots is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetClosedAuctionSlots(
|
||||
newClosedAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetClosedAuctionSlots(newClosedAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -388,8 +379,7 @@ func (c *AuctionClient) AuctionGetClosedAuctionSlots() (closedAuctionSlots uint1
|
||||
}
|
||||
|
||||
// AuctionSetOutbidding is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetOutbidding(newOutbidding uint16) (tx *types.Transaction,
|
||||
err error) {
|
||||
func (c *AuctionClient) AuctionSetOutbidding(newOutbidding uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
12500000, //nolint:gomnd
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -413,8 +403,7 @@ func (c *AuctionClient) AuctionGetOutbidding() (outbidding uint16, err error) {
|
||||
}
|
||||
|
||||
// AuctionSetAllocationRatio is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetAllocationRatio(
|
||||
newAllocationRatio [3]uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetAllocationRatio(newAllocationRatio [3]uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -438,8 +427,7 @@ func (c *AuctionClient) AuctionGetAllocationRatio() (allocationRation [3]uint16,
|
||||
}
|
||||
|
||||
// AuctionSetDonationAddress is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetDonationAddress(
|
||||
newDonationAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetDonationAddress(newDonationAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -452,8 +440,7 @@ func (c *AuctionClient) AuctionSetDonationAddress(
|
||||
}
|
||||
|
||||
// AuctionGetDonationAddress is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.Address,
|
||||
err error) {
|
||||
func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.Address, err error) {
|
||||
var _donationAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_donationAddress, err = c.auction.GetDonationAddress(c.opts)
|
||||
@@ -465,13 +452,11 @@ func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.
|
||||
}
|
||||
|
||||
// AuctionSetBootCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address,
|
||||
newBootCoordinatorURL string) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address, newBootCoordinatorURL string) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
return c.auction.SetBootCoordinator(auth, newBootCoordinator,
|
||||
newBootCoordinatorURL)
|
||||
return c.auction.SetBootCoordinator(auth, newBootCoordinator, newBootCoordinatorURL)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed setting bootCoordinator: %w", err))
|
||||
@@ -480,8 +465,7 @@ func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.A
|
||||
}
|
||||
|
||||
// AuctionGetBootCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.Address,
|
||||
err error) {
|
||||
func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.Address, err error) {
|
||||
var _bootCoordinator ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_bootCoordinator, err = c.auction.GetBootCoordinator(c.opts)
|
||||
@@ -493,8 +477,7 @@ func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.
|
||||
}
|
||||
|
||||
// AuctionChangeDefaultSlotSetBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64,
|
||||
newInitialMinBid *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64, newInitialMinBid *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -508,8 +491,7 @@ func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64,
|
||||
}
|
||||
|
||||
// AuctionGetClaimableHEZ is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetClaimableHEZ(
|
||||
claimAddress ethCommon.Address) (claimableHEZ *big.Int, err error) {
|
||||
func (c *AuctionClient) AuctionGetClaimableHEZ(claimAddress ethCommon.Address) (claimableHEZ *big.Int, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
claimableHEZ, err = c.auction.GetClaimableHEZ(c.opts, claimAddress)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -520,8 +502,7 @@ func (c *AuctionClient) AuctionGetClaimableHEZ(
|
||||
}
|
||||
|
||||
// AuctionSetCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetCoordinator(forger ethCommon.Address,
|
||||
coordinatorURL string) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -570,8 +551,7 @@ func (c *AuctionClient) AuctionGetSlotSet(slot int64) (slotSet *big.Int, err err
|
||||
}
|
||||
|
||||
// AuctionGetDefaultSlotSetBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetDefaultSlotSetBid(slotSet uint8) (minBidSlotSet *big.Int,
|
||||
err error) {
|
||||
func (c *AuctionClient) AuctionGetDefaultSlotSetBid(slotSet uint8) (minBidSlotSet *big.Int, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
minBidSlotSet, err = c.auction.GetDefaultSlotSetBid(c.opts, slotSet)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -594,8 +574,7 @@ func (c *AuctionClient) AuctionGetSlotNumber(blockNum int64) (slot int64, err er
|
||||
}
|
||||
|
||||
// AuctionBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -607,8 +586,7 @@ func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.I
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
_slot := big.NewInt(slot)
|
||||
@@ -621,8 +599,8 @@ func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.I
|
||||
}
|
||||
|
||||
// AuctionMultiBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64,
|
||||
slotSets [6]bool, maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64, slotSets [6]bool,
|
||||
maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
1000000, //nolint:gomnd
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -635,14 +613,12 @@ func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlo
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
_startingSlot := big.NewInt(startingSlot)
|
||||
_endingSlot := big.NewInt(endingSlot)
|
||||
return c.auction.ProcessMultiBid(auth, amount, _startingSlot, _endingSlot,
|
||||
slotSets, maxBid, minBid, permit)
|
||||
return c.auction.ProcessMultiBid(auth, amount, _startingSlot, _endingSlot, slotSets, maxBid, minBid, permit)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed multibid: %w", err))
|
||||
@@ -651,8 +627,7 @@ func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlo
|
||||
}
|
||||
|
||||
// AuctionCanForge is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionCanForge(forger ethCommon.Address, blockNum int64) (canForge bool,
|
||||
err error) {
|
||||
func (c *AuctionClient) AuctionCanForge(forger ethCommon.Address, blockNum int64) (canForge bool, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
canForge, err = c.auction.CanForge(c.opts, forger, big.NewInt(blockNum))
|
||||
return tracerr.Wrap(err)
|
||||
@@ -705,8 +680,7 @@ func (c *AuctionClient) AuctionConstants() (auctionConstants *common.AuctionCons
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
auctionConstants.InitialMinimalBidding, err =
|
||||
c.auction.INITIALMINIMALBIDDING(c.opts)
|
||||
auctionConstants.InitialMinimalBidding, err = c.auction.INITIALMINIMALBIDDING(c.opts)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -777,35 +751,21 @@ func (c *AuctionClient) AuctionVariables() (auctionVariables *common.AuctionVari
|
||||
}
|
||||
|
||||
var (
|
||||
logAuctionNewBid = crypto.Keccak256Hash([]byte(
|
||||
"NewBid(uint128,uint128,address)"))
|
||||
logAuctionNewSlotDeadline = crypto.Keccak256Hash([]byte(
|
||||
"NewSlotDeadline(uint8)"))
|
||||
logAuctionNewClosedAuctionSlots = crypto.Keccak256Hash([]byte(
|
||||
"NewClosedAuctionSlots(uint16)"))
|
||||
logAuctionNewOutbidding = crypto.Keccak256Hash([]byte(
|
||||
"NewOutbidding(uint16)"))
|
||||
logAuctionNewDonationAddress = crypto.Keccak256Hash([]byte(
|
||||
"NewDonationAddress(address)"))
|
||||
logAuctionNewBootCoordinator = crypto.Keccak256Hash([]byte(
|
||||
"NewBootCoordinator(address,string)"))
|
||||
logAuctionNewOpenAuctionSlots = crypto.Keccak256Hash([]byte(
|
||||
"NewOpenAuctionSlots(uint16)"))
|
||||
logAuctionNewAllocationRatio = crypto.Keccak256Hash([]byte(
|
||||
"NewAllocationRatio(uint16[3])"))
|
||||
logAuctionSetCoordinator = crypto.Keccak256Hash([]byte(
|
||||
"SetCoordinator(address,address,string)"))
|
||||
logAuctionNewForgeAllocated = crypto.Keccak256Hash([]byte(
|
||||
"NewForgeAllocated(address,address,uint128,uint128,uint128,uint128)"))
|
||||
logAuctionNewDefaultSlotSetBid = crypto.Keccak256Hash([]byte(
|
||||
"NewDefaultSlotSetBid(uint128,uint128)"))
|
||||
logAuctionNewForge = crypto.Keccak256Hash([]byte(
|
||||
"NewForge(address,uint128)"))
|
||||
logAuctionHEZClaimed = crypto.Keccak256Hash([]byte(
|
||||
"HEZClaimed(address,uint128)"))
|
||||
logAuctionInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezAuctionProtocolEvent(address,address,string," +
|
||||
"uint16,uint8,uint16,uint16,uint16[3])"))
|
||||
logAuctionNewBid = crypto.Keccak256Hash([]byte("NewBid(uint128,uint128,address)"))
|
||||
logAuctionNewSlotDeadline = crypto.Keccak256Hash([]byte("NewSlotDeadline(uint8)"))
|
||||
logAuctionNewClosedAuctionSlots = crypto.Keccak256Hash([]byte("NewClosedAuctionSlots(uint16)"))
|
||||
logAuctionNewOutbidding = crypto.Keccak256Hash([]byte("NewOutbidding(uint16)"))
|
||||
logAuctionNewDonationAddress = crypto.Keccak256Hash([]byte("NewDonationAddress(address)"))
|
||||
logAuctionNewBootCoordinator = crypto.Keccak256Hash([]byte("NewBootCoordinator(address,string)"))
|
||||
logAuctionNewOpenAuctionSlots = crypto.Keccak256Hash([]byte("NewOpenAuctionSlots(uint16)"))
|
||||
logAuctionNewAllocationRatio = crypto.Keccak256Hash([]byte("NewAllocationRatio(uint16[3])"))
|
||||
logAuctionSetCoordinator = crypto.Keccak256Hash([]byte("SetCoordinator(address,address,string)"))
|
||||
logAuctionNewForgeAllocated = crypto.Keccak256Hash([]byte("NewForgeAllocated(address,address,uint128,uint128,uint128,uint128)"))
|
||||
logAuctionNewDefaultSlotSetBid = crypto.Keccak256Hash([]byte("NewDefaultSlotSetBid(uint128,uint128)"))
|
||||
logAuctionNewForge = crypto.Keccak256Hash([]byte("NewForge(address,uint128)"))
|
||||
logAuctionHEZClaimed = crypto.Keccak256Hash([]byte("HEZClaimed(address,uint128)"))
|
||||
logAuctionInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezAuctionProtocolEvent(address,address,string,uint16,uint8,uint16,uint16,uint16[3])"))
|
||||
)
|
||||
|
||||
// AuctionEventInit returns the initialize event with its corresponding block number
|
||||
@@ -821,8 +781,7 @@ func (c *AuctionClient) AuctionEventInit() (*AuctionEventInitialize, int64, erro
|
||||
return nil, 0, tracerr.Wrap(err)
|
||||
}
|
||||
if len(logs) != 1 {
|
||||
return nil, 0,
|
||||
tracerr.Wrap(fmt.Errorf("no event of type InitializeHermezAuctionProtocolEvent found"))
|
||||
return nil, 0, tracerr.Wrap(fmt.Errorf("no event of type InitializeHermezAuctionProtocolEvent found"))
|
||||
}
|
||||
vLog := logs[0]
|
||||
if vLog.Topics[0] != logAuctionInitialize {
|
||||
@@ -870,8 +829,7 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
|
||||
for _, vLog := range logs {
|
||||
if blockHash != nil && vLog.BlockHash != *blockHash {
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got",
|
||||
vLog.BlockHash.String())
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got", vLog.BlockHash.String())
|
||||
return nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
}
|
||||
switch vLog.Topics[0] {
|
||||
@@ -882,8 +840,7 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
Address ethCommon.Address
|
||||
}
|
||||
var newBid AuctionEventNewBid
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewBid, "NewBid",
|
||||
vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewBid, "NewBid", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newBid.BidAmount = auxNewBid.BidAmount
|
||||
@@ -892,60 +849,48 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
auctionEvents.NewBid = append(auctionEvents.NewBid, newBid)
|
||||
case logAuctionNewSlotDeadline:
|
||||
var newSlotDeadline AuctionEventNewSlotDeadline
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newSlotDeadline,
|
||||
"NewSlotDeadline", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newSlotDeadline, "NewSlotDeadline", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewSlotDeadline = append(auctionEvents.NewSlotDeadline, newSlotDeadline)
|
||||
case logAuctionNewClosedAuctionSlots:
|
||||
var newClosedAuctionSlots AuctionEventNewClosedAuctionSlots
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newClosedAuctionSlots,
|
||||
"NewClosedAuctionSlots", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newClosedAuctionSlots, "NewClosedAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewClosedAuctionSlots =
|
||||
append(auctionEvents.NewClosedAuctionSlots, newClosedAuctionSlots)
|
||||
auctionEvents.NewClosedAuctionSlots = append(auctionEvents.NewClosedAuctionSlots, newClosedAuctionSlots)
|
||||
case logAuctionNewOutbidding:
|
||||
var newOutbidding AuctionEventNewOutbidding
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOutbidding, "NewOutbidding",
|
||||
vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOutbidding, "NewOutbidding", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewOutbidding = append(auctionEvents.NewOutbidding, newOutbidding)
|
||||
case logAuctionNewDonationAddress:
|
||||
var newDonationAddress AuctionEventNewDonationAddress
|
||||
newDonationAddress.NewDonationAddress = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
auctionEvents.NewDonationAddress = append(auctionEvents.NewDonationAddress,
|
||||
newDonationAddress)
|
||||
auctionEvents.NewDonationAddress = append(auctionEvents.NewDonationAddress, newDonationAddress)
|
||||
case logAuctionNewBootCoordinator:
|
||||
var newBootCoordinator AuctionEventNewBootCoordinator
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newBootCoordinator,
|
||||
"NewBootCoordinator", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newBootCoordinator, "NewBootCoordinator", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newBootCoordinator.NewBootCoordinator = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
auctionEvents.NewBootCoordinator = append(auctionEvents.NewBootCoordinator,
|
||||
newBootCoordinator)
|
||||
auctionEvents.NewBootCoordinator = append(auctionEvents.NewBootCoordinator, newBootCoordinator)
|
||||
case logAuctionNewOpenAuctionSlots:
|
||||
var newOpenAuctionSlots AuctionEventNewOpenAuctionSlots
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOpenAuctionSlots,
|
||||
"NewOpenAuctionSlots", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOpenAuctionSlots, "NewOpenAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewOpenAuctionSlots =
|
||||
append(auctionEvents.NewOpenAuctionSlots, newOpenAuctionSlots)
|
||||
auctionEvents.NewOpenAuctionSlots = append(auctionEvents.NewOpenAuctionSlots, newOpenAuctionSlots)
|
||||
case logAuctionNewAllocationRatio:
|
||||
var newAllocationRatio AuctionEventNewAllocationRatio
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newAllocationRatio,
|
||||
"NewAllocationRatio", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newAllocationRatio, "NewAllocationRatio", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewAllocationRatio = append(auctionEvents.NewAllocationRatio,
|
||||
newAllocationRatio)
|
||||
auctionEvents.NewAllocationRatio = append(auctionEvents.NewAllocationRatio, newAllocationRatio)
|
||||
case logAuctionSetCoordinator:
|
||||
var setCoordinator AuctionEventSetCoordinator
|
||||
if err := c.contractAbi.UnpackIntoInterface(&setCoordinator,
|
||||
"SetCoordinator", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&setCoordinator, "SetCoordinator", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
setCoordinator.BidderAddress = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
@@ -953,29 +898,25 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
auctionEvents.SetCoordinator = append(auctionEvents.SetCoordinator, setCoordinator)
|
||||
case logAuctionNewForgeAllocated:
|
||||
var newForgeAllocated AuctionEventNewForgeAllocated
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newForgeAllocated,
|
||||
"NewForgeAllocated", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newForgeAllocated, "NewForgeAllocated", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newForgeAllocated.Bidder = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
newForgeAllocated.Forger = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
newForgeAllocated.SlotToForge = new(big.Int).SetBytes(vLog.Topics[3][:]).Int64()
|
||||
auctionEvents.NewForgeAllocated = append(auctionEvents.NewForgeAllocated,
|
||||
newForgeAllocated)
|
||||
auctionEvents.NewForgeAllocated = append(auctionEvents.NewForgeAllocated, newForgeAllocated)
|
||||
case logAuctionNewDefaultSlotSetBid:
|
||||
var auxNewDefaultSlotSetBid struct {
|
||||
SlotSet *big.Int
|
||||
NewInitialMinBid *big.Int
|
||||
}
|
||||
var newDefaultSlotSetBid AuctionEventNewDefaultSlotSetBid
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewDefaultSlotSetBid,
|
||||
"NewDefaultSlotSetBid", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewDefaultSlotSetBid, "NewDefaultSlotSetBid", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newDefaultSlotSetBid.NewInitialMinBid = auxNewDefaultSlotSetBid.NewInitialMinBid
|
||||
newDefaultSlotSetBid.SlotSet = auxNewDefaultSlotSetBid.SlotSet.Int64()
|
||||
auctionEvents.NewDefaultSlotSetBid =
|
||||
append(auctionEvents.NewDefaultSlotSetBid, newDefaultSlotSetBid)
|
||||
auctionEvents.NewDefaultSlotSetBid = append(auctionEvents.NewDefaultSlotSetBid, newDefaultSlotSetBid)
|
||||
case logAuctionNewForge:
|
||||
var newForge AuctionEventNewForge
|
||||
newForge.Forger = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
@@ -983,8 +924,7 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
auctionEvents.NewForge = append(auctionEvents.NewForge, newForge)
|
||||
case logAuctionHEZClaimed:
|
||||
var HEZClaimed AuctionEventHEZClaimed
|
||||
if err := c.contractAbi.UnpackIntoInterface(&HEZClaimed, "HEZClaimed",
|
||||
vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&HEZClaimed, "HEZClaimed", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
HEZClaimed.Owner = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
|
||||
@@ -58,8 +58,7 @@ func TestAuctionConstants(t *testing.T) {
|
||||
func TestAuctionVariables(t *testing.T) {
|
||||
INITMINBID := new(big.Int)
|
||||
INITMINBID.SetString(minBidStr, 10)
|
||||
defaultSlotSetBid := [6]*big.Int{INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID,
|
||||
INITMINBID}
|
||||
defaultSlotSetBid := [6]*big.Int{INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID}
|
||||
|
||||
auctionVariables, err := auctionClientTest.AuctionVariables()
|
||||
require.Nil(t, err)
|
||||
@@ -133,8 +132,7 @@ func TestAuctionSetClosedAuctionSlots(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newClosedAuctionSlots,
|
||||
auctionEvents.NewClosedAuctionSlots[0].NewClosedAuctionSlots)
|
||||
assert.Equal(t, newClosedAuctionSlots, auctionEvents.NewClosedAuctionSlots[0].NewClosedAuctionSlots)
|
||||
_, err = auctionClientTest.AuctionSetClosedAuctionSlots(closedAuctionSlots)
|
||||
require.Nil(t, err)
|
||||
}
|
||||
@@ -230,8 +228,7 @@ func TestAuctionSetBootCoordinator(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newBootCoordinator, auctionEvents.NewBootCoordinator[0].NewBootCoordinator)
|
||||
assert.Equal(t, newBootCoordinatorURL, auctionEvents.NewBootCoordinator[0].NewBootCoordinatorURL)
|
||||
_, err = auctionClientTest.AuctionSetBootCoordinator(bootCoordinatorAddressConst,
|
||||
bootCoordinatorURL)
|
||||
_, err = auctionClientTest.AuctionSetBootCoordinator(bootCoordinatorAddressConst, bootCoordinatorURL)
|
||||
require.Nil(t, err)
|
||||
}
|
||||
|
||||
@@ -345,8 +342,7 @@ func TestAuctionMultiBid(t *testing.T) {
|
||||
budget := new(big.Int)
|
||||
budget.SetString("45200000000000000000", 10)
|
||||
bidderAddress := governanceAddressConst
|
||||
_, err = auctionClientTest.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet,
|
||||
maxBid, minBid, deadline)
|
||||
_, err = auctionClientTest.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet, maxBid, minBid, deadline)
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
@@ -387,8 +383,7 @@ func TestAuctionClaimHEZ(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestAuctionForge(t *testing.T) {
|
||||
auctionClientTestHermez, err := NewAuctionClient(ethereumClientHermez,
|
||||
auctionTestAddressConst, tokenHEZ)
|
||||
auctionClientTestHermez, err := NewAuctionClient(ethereumClientHermez, auctionTestAddressConst, tokenHEZ)
|
||||
require.Nil(t, err)
|
||||
slotConst := 4
|
||||
blockNum := int64(int(blocksPerSlot)*slotConst + int(genesisBlock))
|
||||
|
||||
@@ -64,19 +64,16 @@ type ClientConfig struct {
|
||||
}
|
||||
|
||||
// NewClient creates a new Client to interact with Ethereum and the Hermez smart contracts.
|
||||
func NewClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore,
|
||||
cfg *ClientConfig) (*Client, error) {
|
||||
func NewClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore, cfg *ClientConfig) (*Client, error) {
|
||||
ethereumClient, err := NewEthereumClient(client, account, ks, &cfg.Ethereum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionClient, err := NewAuctionClient(ethereumClient, cfg.Auction.Address,
|
||||
cfg.Auction.TokenHEZ)
|
||||
auctionClient, err := NewAuctionClient(ethereumClient, cfg.Auction.Address, cfg.Auction.TokenHEZ)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupClient, err := NewRollupClient(ethereumClient, cfg.Rollup.Address,
|
||||
cfg.Auction.TokenHEZ)
|
||||
rollupClient, err := NewRollupClient(ethereumClient, cfg.Rollup.Address, cfg.Auction.TokenHEZ)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -309,12 +309,9 @@ func (_HermezAuctionProtocol *HermezAuctionProtocolCaller) Coordinators(opts *bi
|
||||
Forger common.Address
|
||||
CoordinatorURL string
|
||||
})
|
||||
if err != nil {
|
||||
return *outstruct, err
|
||||
}
|
||||
|
||||
outstruct.Forger = *abi.ConvertType(out[0], new(common.Address)).(*common.Address)
|
||||
outstruct.CoordinatorURL = *abi.ConvertType(out[1], new(string)).(*string)
|
||||
outstruct.Forger = out[0].(common.Address)
|
||||
outstruct.CoordinatorURL = out[1].(string)
|
||||
|
||||
return *outstruct, err
|
||||
|
||||
@@ -887,15 +884,12 @@ func (_HermezAuctionProtocol *HermezAuctionProtocolCaller) Slots(opts *bind.Call
|
||||
BidAmount *big.Int
|
||||
ClosedMinBid *big.Int
|
||||
})
|
||||
if err != nil {
|
||||
return *outstruct, err
|
||||
}
|
||||
|
||||
outstruct.Bidder = *abi.ConvertType(out[0], new(common.Address)).(*common.Address)
|
||||
outstruct.Fulfilled = *abi.ConvertType(out[1], new(bool)).(*bool)
|
||||
outstruct.ForgerCommitment = *abi.ConvertType(out[2], new(bool)).(*bool)
|
||||
outstruct.BidAmount = *abi.ConvertType(out[3], new(*big.Int)).(**big.Int)
|
||||
outstruct.ClosedMinBid = *abi.ConvertType(out[4], new(*big.Int)).(**big.Int)
|
||||
outstruct.Bidder = out[0].(common.Address)
|
||||
outstruct.Fulfilled = out[1].(bool)
|
||||
outstruct.ForgerCommitment = out[2].(bool)
|
||||
outstruct.BidAmount = out[3].(*big.Int)
|
||||
outstruct.ClosedMinBid = out[4].(*big.Int)
|
||||
|
||||
return *outstruct, err
|
||||
|
||||
|
||||
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
@@ -64,8 +64,7 @@ type EthereumConfig struct {
|
||||
GasPriceDiv uint64
|
||||
}
|
||||
|
||||
// EthereumClient is an ethereum client to call Smart Contract methods and check blockchain
|
||||
// information.
|
||||
// EthereumClient is an ethereum client to call Smart Contract methods and check blockchain information.
|
||||
type EthereumClient struct {
|
||||
client *ethclient.Client
|
||||
chainID *big.Int
|
||||
@@ -77,8 +76,7 @@ type EthereumClient struct {
|
||||
|
||||
// NewEthereumClient creates a EthereumClient instance. The account is not mandatory (it can
|
||||
// be nil). If the account is nil, CallAuth will fail with ErrAccountNil.
|
||||
func NewEthereumClient(client *ethclient.Client, account *accounts.Account,
|
||||
ks *ethKeystore.KeyStore, config *EthereumConfig) (*EthereumClient, error) {
|
||||
func NewEthereumClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore, config *EthereumConfig) (*EthereumClient, error) {
|
||||
if config == nil {
|
||||
config = &EthereumConfig{
|
||||
CallGasLimit: defaultCallGasLimit,
|
||||
@@ -168,8 +166,7 @@ func (c *EthereumClient) NewAuth() (*bind.TransactOpts, error) {
|
||||
// This call requires a valid account with Ether that can be spend during the
|
||||
// call.
|
||||
func (c *EthereumClient) CallAuth(gasLimit uint64,
|
||||
fn func(*ethclient.Client, *bind.TransactOpts) (*types.Transaction, error)) (*types.Transaction,
|
||||
error) {
|
||||
fn func(*ethclient.Client, *bind.TransactOpts) (*types.Transaction, error)) (*types.Transaction, error) {
|
||||
if c.account == nil {
|
||||
return nil, tracerr.Wrap(ErrAccountNil)
|
||||
}
|
||||
@@ -215,8 +212,7 @@ func (c *EthereumClient) Call(fn func(*ethclient.Client) error) error {
|
||||
}
|
||||
|
||||
// EthTransactionReceipt returns the transaction receipt of the given txHash
|
||||
func (c *EthereumClient) EthTransactionReceipt(ctx context.Context,
|
||||
txHash ethCommon.Hash) (*types.Receipt, error) {
|
||||
func (c *EthereumClient) EthTransactionReceipt(ctx context.Context, txHash ethCommon.Hash) (*types.Receipt, error) {
|
||||
return c.client.TransactionReceipt(ctx, txHash)
|
||||
}
|
||||
|
||||
@@ -232,15 +228,13 @@ func (c *EthereumClient) EthLastBlock() (int64, error) {
|
||||
}
|
||||
|
||||
// EthHeaderByNumber internally calls ethclient.Client HeaderByNumber
|
||||
// func (c *EthereumClient) EthHeaderByNumber(ctx context.Context, number *big.Int) (*types.Header,
|
||||
// error) {
|
||||
// func (c *EthereumClient) EthHeaderByNumber(ctx context.Context, number *big.Int) (*types.Header, error) {
|
||||
// return c.client.HeaderByNumber(ctx, number)
|
||||
// }
|
||||
|
||||
// EthBlockByNumber internally calls ethclient.Client BlockByNumber and returns
|
||||
// *common.Block. If number == -1, the latests known block is returned.
|
||||
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block,
|
||||
error) {
|
||||
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block, error) {
|
||||
blockNum := big.NewInt(number)
|
||||
if number == -1 {
|
||||
blockNum = nil
|
||||
|
||||
@@ -14,8 +14,7 @@ import (
|
||||
func addBlock(url string) {
|
||||
method := "POST"
|
||||
|
||||
payload := strings.NewReader(
|
||||
"{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_mine\",\n \"params\":[],\n \"id\":1\n}")
|
||||
payload := strings.NewReader("{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_mine\",\n \"params\":[],\n \"id\":1\n}")
|
||||
|
||||
client := &http.Client{}
|
||||
req, err := http.NewRequest(method, url, payload)
|
||||
@@ -46,9 +45,7 @@ func addTime(seconds float64, url string) {
|
||||
secondsStr := strconv.FormatFloat(seconds, 'E', -1, 32)
|
||||
|
||||
method := "POST"
|
||||
payload := strings.NewReader(
|
||||
"{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_increaseTime\",\n \"params\":[" +
|
||||
secondsStr + "],\n \"id\":1\n}")
|
||||
payload := strings.NewReader("{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_increaseTime\",\n \"params\":[" + secondsStr + "],\n \"id\":1\n}")
|
||||
|
||||
client := &http.Client{}
|
||||
req, err := http.NewRequest(method, url, payload)
|
||||
@@ -69,16 +66,13 @@ func addTime(seconds float64, url string) {
|
||||
}()
|
||||
}
|
||||
|
||||
func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, value, nonce,
|
||||
deadline *big.Int, tokenName string) ([]byte, error) {
|
||||
func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, value, nonce, deadline *big.Int, tokenName string) ([]byte, error) {
|
||||
// NOTE: We ignore hash.Write errors because we are writing to a memory
|
||||
// buffer and don't expect any errors to occur.
|
||||
abiPermit :=
|
||||
[]byte("Permit(address owner,address spender,uint256 value,uint256 nonce,uint256 deadline)")
|
||||
abiPermit := []byte("Permit(address owner,address spender,uint256 value,uint256 nonce,uint256 deadline)")
|
||||
hashPermit := sha3.NewLegacyKeccak256()
|
||||
hashPermit.Write(abiPermit) //nolint:errcheck,gosec
|
||||
abiEIP712Domain :=
|
||||
[]byte("EIP712Domain(string name,string version,uint256 chainId,address verifyingContract)")
|
||||
abiEIP712Domain := []byte("EIP712Domain(string name,string version,uint256 chainId,address verifyingContract)")
|
||||
hashEIP712Domain := sha3.NewLegacyKeccak256()
|
||||
hashEIP712Domain.Write(abiEIP712Domain) //nolint:errcheck,gosec
|
||||
var encodeBytes []byte
|
||||
@@ -130,8 +124,7 @@ func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, va
|
||||
return hashBytes2.Sum(nil), nil
|
||||
}
|
||||
|
||||
func createPermit(owner, spender ethCommon.Address, amount, deadline *big.Int, digest,
|
||||
signature []byte) []byte {
|
||||
func createPermit(owner, spender ethCommon.Address, amount, deadline *big.Int, digest, signature []byte) []byte {
|
||||
r := signature[0:32]
|
||||
s := signature[32:64]
|
||||
v := signature[64] + byte(27) //nolint:gomnd
|
||||
|
||||
@@ -26,8 +26,7 @@ var (
|
||||
mnemonic = "explain tackle mirror kit van hammer degree position ginger unfair soup bonus"
|
||||
)
|
||||
|
||||
func genAcc(w *hdwallet.Wallet, ks *keystore.KeyStore, i int) (*accounts.Account,
|
||||
ethCommon.Address) {
|
||||
func genAcc(w *hdwallet.Wallet, ks *keystore.KeyStore, i int) (*accounts.Account, ethCommon.Address) {
|
||||
path := hdwallet.MustParseDerivationPath(fmt.Sprintf("m/44'/60'/0'/0/%d", i))
|
||||
account, err := w.Derive(path, false)
|
||||
if err != nil {
|
||||
@@ -112,9 +111,7 @@ func getEnvVariables() {
|
||||
if err != nil {
|
||||
log.Fatal(errEnvVar)
|
||||
}
|
||||
if auctionAddressStr == "" || auctionTestAddressStr == "" || tokenHEZAddressStr == "" ||
|
||||
hermezRollupAddressStr == "" || wdelayerAddressStr == "" || wdelayerTestAddressStr == "" ||
|
||||
genesisBlockEnv == "" {
|
||||
if auctionAddressStr == "" || auctionTestAddressStr == "" || tokenHEZAddressStr == "" || hermezRollupAddressStr == "" || wdelayerAddressStr == "" || wdelayerTestAddressStr == "" || genesisBlockEnv == "" {
|
||||
log.Fatal(errEnvVar)
|
||||
}
|
||||
|
||||
@@ -192,8 +189,7 @@ func TestMain(m *testing.M) {
|
||||
log.Fatal(err)
|
||||
}
|
||||
|
||||
ethereumClientEmergencyCouncil, err = NewEthereumClient(ethClient,
|
||||
emergencyCouncilAccount, ks, nil)
|
||||
ethereumClientEmergencyCouncil, err = NewEthereumClient(ethClient, emergencyCouncilAccount, ks, nil)
|
||||
if err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
|
||||
236
eth/rollup.go
236
eth/rollup.go
@@ -4,6 +4,7 @@ import (
|
||||
"context"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"strconv"
|
||||
"strings"
|
||||
|
||||
"github.com/ethereum/go-ethereum"
|
||||
@@ -60,12 +61,21 @@ type RollupEventInitialize struct {
|
||||
|
||||
// RollupVariables returns the RollupVariables from the initialize event
|
||||
func (ei *RollupEventInitialize) RollupVariables() *common.RollupVariables {
|
||||
var buckets [common.RollupConstNumBuckets]common.BucketParams
|
||||
for i := range buckets {
|
||||
buckets[i] = common.BucketParams{
|
||||
CeilUSD: big.NewInt(0),
|
||||
Withdrawals: big.NewInt(0),
|
||||
BlockWithdrawalRate: big.NewInt(0),
|
||||
MaxWithdrawals: big.NewInt(0),
|
||||
}
|
||||
}
|
||||
return &common.RollupVariables{
|
||||
EthBlockNum: 0,
|
||||
FeeAddToken: ei.FeeAddToken,
|
||||
ForgeL1L2BatchTimeout: int64(ei.ForgeL1L2BatchTimeout),
|
||||
WithdrawalDelay: ei.WithdrawalDelay,
|
||||
Buckets: []common.BucketParams{},
|
||||
Buckets: buckets,
|
||||
SafeMode: false,
|
||||
}
|
||||
}
|
||||
@@ -136,21 +146,20 @@ type RollupEventUpdateWithdrawalDelay struct {
|
||||
|
||||
// RollupUpdateBucketsParameters are the bucket parameters used in an update
|
||||
type RollupUpdateBucketsParameters struct {
|
||||
CeilUSD *big.Int
|
||||
BlockStamp *big.Int
|
||||
Withdrawals *big.Int
|
||||
RateBlocks *big.Int
|
||||
RateWithdrawals *big.Int
|
||||
MaxWithdrawals *big.Int
|
||||
CeilUSD *big.Int
|
||||
Withdrawals *big.Int
|
||||
BlockWithdrawalRate *big.Int
|
||||
MaxWithdrawals *big.Int
|
||||
}
|
||||
|
||||
type rollupEventUpdateBucketsParametersAux struct {
|
||||
ArrayBuckets []*big.Int
|
||||
ArrayBuckets [common.RollupConstNumBuckets][4]*big.Int
|
||||
}
|
||||
|
||||
// RollupEventUpdateBucketsParameters is an event of the Rollup Smart Contract
|
||||
type RollupEventUpdateBucketsParameters struct {
|
||||
ArrayBuckets []RollupUpdateBucketsParameters
|
||||
// ArrayBuckets [common.RollupConstNumBuckets][4]*big.Int
|
||||
ArrayBuckets [common.RollupConstNumBuckets]RollupUpdateBucketsParameters
|
||||
SafeMode bool
|
||||
}
|
||||
|
||||
@@ -234,20 +243,13 @@ type RollupInterface interface {
|
||||
// Public Functions
|
||||
|
||||
RollupForgeBatch(*RollupForgeBatchArgs, *bind.TransactOpts) (*types.Transaction, error)
|
||||
RollupAddToken(tokenAddress ethCommon.Address, feeAddToken,
|
||||
deadline *big.Int) (*types.Transaction, error)
|
||||
RollupAddToken(tokenAddress ethCommon.Address, feeAddToken, deadline *big.Int) (*types.Transaction, error)
|
||||
|
||||
RollupWithdrawMerkleProof(babyPubKey babyjub.PublicKeyComp, tokenID uint32, numExitRoot,
|
||||
idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (*types.Transaction,
|
||||
error)
|
||||
RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32,
|
||||
numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
RollupWithdrawMerkleProof(babyPubKey babyjub.PublicKeyComp, tokenID uint32, numExitRoot, idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
|
||||
RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int,
|
||||
amount *big.Int, tokenID uint32, toIdx int64) (*types.Transaction, error)
|
||||
RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64,
|
||||
deadline *big.Int) (tx *types.Transaction, err error)
|
||||
RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (*types.Transaction, error)
|
||||
RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
|
||||
// Governance Public Functions
|
||||
RollupUpdateForgeL1L2BatchTimeout(newForgeL1L2BatchTimeout int64) (*types.Transaction, error)
|
||||
@@ -285,8 +287,7 @@ type RollupClient struct {
|
||||
}
|
||||
|
||||
// NewRollupClient creates a new RollupClient
|
||||
func NewRollupClient(client *EthereumClient, address ethCommon.Address,
|
||||
tokenHEZCfg TokenConfig) (*RollupClient, error) {
|
||||
func NewRollupClient(client *EthereumClient, address ethCommon.Address, tokenHEZCfg TokenConfig) (*RollupClient, error) {
|
||||
contractAbi, err := abi.JSON(strings.NewReader(string(Hermez.HermezABI)))
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -322,8 +323,7 @@ func NewRollupClient(client *EthereumClient, address ethCommon.Address,
|
||||
}
|
||||
|
||||
// RollupForgeBatch is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs,
|
||||
auth *bind.TransactOpts) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs, auth *bind.TransactOpts) (tx *types.Transaction, err error) {
|
||||
if auth == nil {
|
||||
auth, err = c.client.NewAuth()
|
||||
if err != nil {
|
||||
@@ -401,8 +401,7 @@ func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs,
|
||||
// RollupAddToken is the interface to call the smart contract function.
|
||||
// `feeAddToken` is the amount of HEZ tokens that will be paid to add the
|
||||
// token. `feeAddToken` must match the public value of the smart contract.
|
||||
func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToken,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToken, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -414,11 +413,9 @@ func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToke
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
feeAddToken, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, feeAddToken, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, feeAddToken, deadline, digest,
|
||||
signature)
|
||||
permit := createPermit(owner, spender, feeAddToken, deadline, digest, signature)
|
||||
|
||||
return c.hermez.AddToken(auth, tokenAddress, permit)
|
||||
},
|
||||
@@ -429,9 +426,7 @@ func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToke
|
||||
}
|
||||
|
||||
// RollupWithdrawMerkleProof is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp, tokenID uint32,
|
||||
numExitRoot, idx int64, amount *big.Int, siblings []*big.Int,
|
||||
instantWithdraw bool) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp, tokenID uint32, numExitRoot, idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -439,8 +434,7 @@ func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp,
|
||||
babyPubKey := new(big.Int).SetBytes(pkCompB)
|
||||
numExitRootB := uint32(numExitRoot)
|
||||
idxBig := big.NewInt(idx)
|
||||
return c.hermez.WithdrawMerkleProof(auth, tokenID, amount, babyPubKey,
|
||||
numExitRootB, siblings, idxBig, instantWithdraw)
|
||||
return c.hermez.WithdrawMerkleProof(auth, tokenID, amount, babyPubKey, numExitRootB, siblings, idxBig, instantWithdraw)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed update WithdrawMerkleProof: %w", err))
|
||||
@@ -449,17 +443,13 @@ func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp,
|
||||
}
|
||||
|
||||
// RollupWithdrawCircuit is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int,
|
||||
tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction,
|
||||
error) {
|
||||
func (c *RollupClient) RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error) {
|
||||
log.Error("TODO")
|
||||
return nil, tracerr.Wrap(errTODO)
|
||||
}
|
||||
|
||||
// RollupL1UserTxERC20ETH is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (tx *types.Transaction,
|
||||
err error) {
|
||||
func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -484,8 +474,8 @@ func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fro
|
||||
auth.Value = depositAmount
|
||||
}
|
||||
var permit []byte
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig, big.NewInt(int64(depositAmountF)),
|
||||
big.NewInt(int64(amountF)), tokenID, toIdxBig, permit)
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig, uint16(depositAmountF),
|
||||
uint16(amountF), tokenID, toIdxBig, permit)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed add L1 Tx ERC20/ETH: %w", err))
|
||||
@@ -494,9 +484,7 @@ func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fro
|
||||
}
|
||||
|
||||
// RollupL1UserTxERC20Permit is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -528,12 +516,11 @@ func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp,
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig,
|
||||
big.NewInt(int64(depositAmountF)), big.NewInt(int64(amountF)), tokenID, toIdxBig, permit)
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig, uint16(depositAmountF),
|
||||
uint16(amountF), tokenID, toIdxBig, permit)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed add L1 Tx ERC20Permit: %w", err))
|
||||
@@ -565,13 +552,11 @@ func (c *RollupClient) RollupLastForgedBatch() (lastForgedBatch int64, err error
|
||||
}
|
||||
|
||||
// RollupUpdateForgeL1L2BatchTimeout is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(
|
||||
newForgeL1L2BatchTimeout int64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(newForgeL1L2BatchTimeout int64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
return c.hermez.UpdateForgeL1L2BatchTimeout(auth,
|
||||
uint8(newForgeL1L2BatchTimeout))
|
||||
return c.hermez.UpdateForgeL1L2BatchTimeout(auth, uint8(newForgeL1L2BatchTimeout))
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed update ForgeL1L2BatchTimeout: %w", err))
|
||||
@@ -580,8 +565,7 @@ func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(
|
||||
}
|
||||
|
||||
// RollupUpdateFeeAddToken is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateFeeAddToken(newFeeAddToken *big.Int) (tx *types.Transaction,
|
||||
err error) {
|
||||
func (c *RollupClient) RollupUpdateFeeAddToken(newFeeAddToken *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -595,20 +579,18 @@ func (c *RollupClient) RollupUpdateFeeAddToken(newFeeAddToken *big.Int) (tx *typ
|
||||
|
||||
// RollupUpdateBucketsParameters is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateBucketsParameters(
|
||||
arrayBuckets []RollupUpdateBucketsParameters,
|
||||
arrayBuckets [common.RollupConstNumBuckets]RollupUpdateBucketsParameters,
|
||||
) (tx *types.Transaction, err error) {
|
||||
params := [common.RollupConstNumBuckets][4]*big.Int{}
|
||||
for i, bucket := range arrayBuckets {
|
||||
params[i][0] = bucket.CeilUSD
|
||||
params[i][1] = bucket.Withdrawals
|
||||
params[i][2] = bucket.BlockWithdrawalRate
|
||||
params[i][3] = bucket.MaxWithdrawals
|
||||
}
|
||||
if tx, err = c.client.CallAuth(
|
||||
12500000, //nolint:gomnd
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
params := make([]*big.Int, len(arrayBuckets))
|
||||
for i, bucket := range arrayBuckets {
|
||||
params[i], err = c.hermez.PackBucket(c.opts,
|
||||
bucket.CeilUSD, bucket.BlockStamp, bucket.Withdrawals,
|
||||
bucket.RateBlocks, bucket.RateWithdrawals, bucket.MaxWithdrawals)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("failed to pack bucket: %w", err))
|
||||
}
|
||||
}
|
||||
return c.hermez.UpdateBucketsParameters(auth, params)
|
||||
},
|
||||
); err != nil {
|
||||
@@ -618,8 +600,7 @@ func (c *RollupClient) RollupUpdateBucketsParameters(
|
||||
}
|
||||
|
||||
// RollupUpdateTokenExchange is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Address,
|
||||
valueArray []uint64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Address, valueArray []uint64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -632,8 +613,7 @@ func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Addres
|
||||
}
|
||||
|
||||
// RollupUpdateWithdrawalDelay is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateWithdrawalDelay(newWithdrawalDelay int64) (tx *types.Transaction,
|
||||
err error) {
|
||||
func (c *RollupClient) RollupUpdateWithdrawalDelay(newWithdrawalDelay int64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -659,8 +639,7 @@ func (c *RollupClient) RollupSafeMode() (tx *types.Transaction, err error) {
|
||||
}
|
||||
|
||||
// RollupInstantWithdrawalViewer is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupInstantWithdrawalViewer(tokenAddress ethCommon.Address,
|
||||
amount *big.Int) (instantAllowed bool, err error) {
|
||||
func (c *RollupClient) RollupInstantWithdrawalViewer(tokenAddress ethCommon.Address, amount *big.Int) (instantAllowed bool, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
instantAllowed, err = c.hermez.InstantWithdrawalViewer(c.opts, tokenAddress, amount)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -695,8 +674,7 @@ func (c *RollupClient) RollupConstants() (rollupConstants *common.RollupConstant
|
||||
}
|
||||
newRollupVerifier.MaxTx = rollupVerifier.MaxTx.Int64()
|
||||
newRollupVerifier.NLevels = rollupVerifier.NLevels.Int64()
|
||||
rollupConstants.Verifiers = append(rollupConstants.Verifiers,
|
||||
newRollupVerifier)
|
||||
rollupConstants.Verifiers = append(rollupConstants.Verifiers, newRollupVerifier)
|
||||
}
|
||||
rollupConstants.HermezAuctionContract, err = c.hermez.HermezAuctionContract(c.opts)
|
||||
if err != nil {
|
||||
@@ -715,30 +693,19 @@ func (c *RollupClient) RollupConstants() (rollupConstants *common.RollupConstant
|
||||
}
|
||||
|
||||
var (
|
||||
logHermezL1UserTxEvent = crypto.Keccak256Hash([]byte(
|
||||
"L1UserTxEvent(uint32,uint8,bytes)"))
|
||||
logHermezAddToken = crypto.Keccak256Hash([]byte(
|
||||
"AddToken(address,uint32)"))
|
||||
logHermezForgeBatch = crypto.Keccak256Hash([]byte(
|
||||
"ForgeBatch(uint32,uint16)"))
|
||||
logHermezUpdateForgeL1L2BatchTimeout = crypto.Keccak256Hash([]byte(
|
||||
"UpdateForgeL1L2BatchTimeout(uint8)"))
|
||||
logHermezUpdateFeeAddToken = crypto.Keccak256Hash([]byte(
|
||||
"UpdateFeeAddToken(uint256)"))
|
||||
logHermezWithdrawEvent = crypto.Keccak256Hash([]byte(
|
||||
"WithdrawEvent(uint48,uint32,bool)"))
|
||||
logHermezUpdateBucketWithdraw = crypto.Keccak256Hash([]byte(
|
||||
"UpdateBucketWithdraw(uint8,uint256,uint256)"))
|
||||
logHermezUpdateWithdrawalDelay = crypto.Keccak256Hash([]byte(
|
||||
"UpdateWithdrawalDelay(uint64)"))
|
||||
logHermezUpdateBucketsParameters = crypto.Keccak256Hash([]byte(
|
||||
"UpdateBucketsParameters(uint256[])"))
|
||||
logHermezUpdateTokenExchange = crypto.Keccak256Hash([]byte(
|
||||
"UpdateTokenExchange(address[],uint64[])"))
|
||||
logHermezSafeMode = crypto.Keccak256Hash([]byte(
|
||||
"SafeMode()"))
|
||||
logHermezInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezEvent(uint8,uint256,uint64)"))
|
||||
logHermezL1UserTxEvent = crypto.Keccak256Hash([]byte("L1UserTxEvent(uint32,uint8,bytes)"))
|
||||
logHermezAddToken = crypto.Keccak256Hash([]byte("AddToken(address,uint32)"))
|
||||
logHermezForgeBatch = crypto.Keccak256Hash([]byte("ForgeBatch(uint32,uint16)"))
|
||||
logHermezUpdateForgeL1L2BatchTimeout = crypto.Keccak256Hash([]byte("UpdateForgeL1L2BatchTimeout(uint8)"))
|
||||
logHermezUpdateFeeAddToken = crypto.Keccak256Hash([]byte("UpdateFeeAddToken(uint256)"))
|
||||
logHermezWithdrawEvent = crypto.Keccak256Hash([]byte("WithdrawEvent(uint48,uint32,bool)"))
|
||||
logHermezUpdateBucketWithdraw = crypto.Keccak256Hash([]byte("UpdateBucketWithdraw(uint8,uint256,uint256)"))
|
||||
logHermezUpdateWithdrawalDelay = crypto.Keccak256Hash([]byte("UpdateWithdrawalDelay(uint64)"))
|
||||
logHermezUpdateBucketsParameters = crypto.Keccak256Hash([]byte("UpdateBucketsParameters(uint256[4][" +
|
||||
strconv.Itoa(common.RollupConstNumBuckets) + "])"))
|
||||
logHermezUpdateTokenExchange = crypto.Keccak256Hash([]byte("UpdateTokenExchange(address[],uint64[])"))
|
||||
logHermezSafeMode = crypto.Keccak256Hash([]byte("SafeMode()"))
|
||||
logHermezInitialize = crypto.Keccak256Hash([]byte("InitializeHermezEvent(uint8,uint256,uint64)"))
|
||||
)
|
||||
|
||||
// RollupEventInit returns the initialize event with its corresponding block number
|
||||
@@ -762,8 +729,7 @@ func (c *RollupClient) RollupEventInit() (*RollupEventInitialize, int64, error)
|
||||
}
|
||||
|
||||
var rollupInit RollupEventInitialize
|
||||
if err := c.contractAbi.UnpackIntoInterface(&rollupInit, "InitializeHermezEvent",
|
||||
vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&rollupInit, "InitializeHermezEvent", vLog.Data); err != nil {
|
||||
return nil, 0, tracerr.Wrap(err)
|
||||
}
|
||||
return &rollupInit, int64(vLog.BlockNumber), tracerr.Wrap(err)
|
||||
@@ -844,8 +810,7 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
var updateForgeL1L2BatchTimeout struct {
|
||||
NewForgeL1L2BatchTimeout uint8
|
||||
}
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateForgeL1L2BatchTimeout,
|
||||
"UpdateForgeL1L2BatchTimeout", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateForgeL1L2BatchTimeout, "UpdateForgeL1L2BatchTimeout", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -873,16 +838,14 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
case logHermezUpdateBucketWithdraw:
|
||||
var updateBucketWithdrawAux rollupEventUpdateBucketWithdrawAux
|
||||
var updateBucketWithdraw RollupEventUpdateBucketWithdraw
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateBucketWithdrawAux,
|
||||
"UpdateBucketWithdraw", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateBucketWithdrawAux, "UpdateBucketWithdraw", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
updateBucketWithdraw.Withdrawals = updateBucketWithdrawAux.Withdrawals
|
||||
updateBucketWithdraw.NumBucket = int(new(big.Int).SetBytes(vLog.Topics[1][:]).Int64())
|
||||
updateBucketWithdraw.BlockStamp = new(big.Int).SetBytes(vLog.Topics[2][:]).Int64()
|
||||
rollupEvents.UpdateBucketWithdraw =
|
||||
append(rollupEvents.UpdateBucketWithdraw, updateBucketWithdraw)
|
||||
rollupEvents.UpdateBucketWithdraw = append(rollupEvents.UpdateBucketWithdraw, updateBucketWithdraw)
|
||||
|
||||
case logHermezUpdateWithdrawalDelay:
|
||||
var withdrawalDelay RollupEventUpdateWithdrawalDelay
|
||||
@@ -894,26 +857,17 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
case logHermezUpdateBucketsParameters:
|
||||
var bucketsParametersAux rollupEventUpdateBucketsParametersAux
|
||||
var bucketsParameters RollupEventUpdateBucketsParameters
|
||||
err := c.contractAbi.UnpackIntoInterface(&bucketsParametersAux,
|
||||
"UpdateBucketsParameters", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&bucketsParametersAux, "UpdateBucketsParameters", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
bucketsParameters.ArrayBuckets = make([]RollupUpdateBucketsParameters, len(bucketsParametersAux.ArrayBuckets))
|
||||
for i, bucket := range bucketsParametersAux.ArrayBuckets {
|
||||
bucket, err := c.hermez.UnpackBucket(c.opts, bucket)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
bucketsParameters.ArrayBuckets[i].CeilUSD = bucket.CeilUSD
|
||||
bucketsParameters.ArrayBuckets[i].BlockStamp = bucket.BlockStamp
|
||||
bucketsParameters.ArrayBuckets[i].Withdrawals = bucket.Withdrawals
|
||||
bucketsParameters.ArrayBuckets[i].RateBlocks = bucket.RateBlocks
|
||||
bucketsParameters.ArrayBuckets[i].RateWithdrawals = bucket.RateWithdrawals
|
||||
bucketsParameters.ArrayBuckets[i].MaxWithdrawals = bucket.MaxWithdrawals
|
||||
bucketsParameters.ArrayBuckets[i].CeilUSD = bucket[0]
|
||||
bucketsParameters.ArrayBuckets[i].Withdrawals = bucket[1]
|
||||
bucketsParameters.ArrayBuckets[i].BlockWithdrawalRate = bucket[2]
|
||||
bucketsParameters.ArrayBuckets[i].MaxWithdrawals = bucket[3]
|
||||
}
|
||||
rollupEvents.UpdateBucketsParameters =
|
||||
append(rollupEvents.UpdateBucketsParameters, bucketsParameters)
|
||||
rollupEvents.UpdateBucketsParameters = append(rollupEvents.UpdateBucketsParameters, bucketsParameters)
|
||||
case logHermezUpdateTokenExchange:
|
||||
var tokensExchange RollupEventUpdateTokenExchange
|
||||
err := c.contractAbi.UnpackIntoInterface(&tokensExchange, "UpdateTokenExchange", vLog.Data)
|
||||
@@ -932,10 +886,8 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
}
|
||||
for i := range bucketsParameters.ArrayBuckets {
|
||||
bucketsParameters.ArrayBuckets[i].CeilUSD = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].BlockStamp = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].Withdrawals = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].RateBlocks = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].RateWithdrawals = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].BlockWithdrawalRate = big.NewInt(0)
|
||||
bucketsParameters.ArrayBuckets[i].MaxWithdrawals = big.NewInt(0)
|
||||
}
|
||||
rollupEvents.UpdateBucketsParameters = append(rollupEvents.UpdateBucketsParameters,
|
||||
@@ -947,8 +899,7 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
|
||||
// RollupForgeBatchArgs returns the arguments used in a ForgeBatch call in the
|
||||
// Rollup Smart Contract in the given transaction, and the sender address.
|
||||
func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
l1UserTxsLen uint16) (*RollupForgeBatchArgs, *ethCommon.Address, error) {
|
||||
func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsLen uint16) (*RollupForgeBatchArgs, *ethCommon.Address, error) {
|
||||
tx, _, err := c.client.client.TransactionByHash(context.Background(), ethTxHash)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("TransactionByHash: %w", err))
|
||||
@@ -963,8 +914,7 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
sender, err := c.client.client.TransactionSender(context.Background(), tx,
|
||||
receipt.Logs[0].BlockHash, receipt.Logs[0].Index)
|
||||
sender, err := c.client.client.TransactionSender(context.Background(), tx, receipt.Logs[0].BlockHash, receipt.Logs[0].Index)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -989,7 +939,7 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
FeeIdxCoordinator: []common.Idx{},
|
||||
}
|
||||
nLevels := c.consts.Verifiers[rollupForgeBatchArgs.VerifierIdx].NLevels
|
||||
lenL1L2TxsBytes := int((nLevels/8)*2 + common.Float40BytesLength + 1) //nolint:gomnd
|
||||
lenL1L2TxsBytes := int((nLevels/8)*2 + common.Float40BytesLength + 1)
|
||||
numBytesL1TxUser := int(l1UserTxsLen) * lenL1L2TxsBytes
|
||||
numTxsL1Coord := len(aux.EncodedL1CoordinatorTx) / common.RollupConstL1CoordinatorTotalBytes
|
||||
numBytesL1TxCoord := numTxsL1Coord * lenL1L2TxsBytes
|
||||
@@ -999,9 +949,7 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
l1UserTxsData = aux.L1L2TxsData[:numBytesL1TxUser]
|
||||
}
|
||||
for i := 0; i < int(l1UserTxsLen); i++ {
|
||||
l1Tx, err :=
|
||||
common.L1TxFromDataAvailability(l1UserTxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes],
|
||||
uint32(nLevels))
|
||||
l1Tx, err := common.L1TxFromDataAvailability(l1UserTxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes], uint32(nLevels))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -1013,17 +961,14 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
}
|
||||
numTxsL2 := len(l2TxsData) / lenL1L2TxsBytes
|
||||
for i := 0; i < numTxsL2; i++ {
|
||||
l2Tx, err :=
|
||||
common.L2TxFromBytesDataAvailability(l2TxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes],
|
||||
int(nLevels))
|
||||
l2Tx, err := common.L2TxFromBytesDataAvailability(l2TxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes], int(nLevels))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupForgeBatchArgs.L2TxsData = append(rollupForgeBatchArgs.L2TxsData, *l2Tx)
|
||||
}
|
||||
for i := 0; i < numTxsL1Coord; i++ {
|
||||
bytesL1Coordinator :=
|
||||
aux.EncodedL1CoordinatorTx[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes] //nolint:lll
|
||||
bytesL1Coordinator := aux.EncodedL1CoordinatorTx[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes]
|
||||
var signature []byte
|
||||
v := bytesL1Coordinator[0]
|
||||
s := bytesL1Coordinator[1:33]
|
||||
@@ -1036,29 +981,24 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupForgeBatchArgs.L1CoordinatorTxs = append(rollupForgeBatchArgs.L1CoordinatorTxs, *l1Tx)
|
||||
rollupForgeBatchArgs.L1CoordinatorTxsAuths =
|
||||
append(rollupForgeBatchArgs.L1CoordinatorTxsAuths, signature)
|
||||
rollupForgeBatchArgs.L1CoordinatorTxsAuths = append(rollupForgeBatchArgs.L1CoordinatorTxsAuths, signature)
|
||||
}
|
||||
lenFeeIdxCoordinatorBytes := int(nLevels / 8) //nolint:gomnd
|
||||
numFeeIdxCoordinator := len(aux.FeeIdxCoordinator) / lenFeeIdxCoordinatorBytes
|
||||
for i := 0; i < numFeeIdxCoordinator; i++ {
|
||||
var paddedFeeIdx [6]byte
|
||||
// TODO: This check is not necessary: the first case will always work. Test it
|
||||
// before removing the if.
|
||||
// TODO: This check is not necessary: the first case will always work. Test it before removing the if.
|
||||
if lenFeeIdxCoordinatorBytes < common.IdxBytesLen {
|
||||
copy(paddedFeeIdx[6-lenFeeIdxCoordinatorBytes:],
|
||||
aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
copy(paddedFeeIdx[6-lenFeeIdxCoordinatorBytes:], aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
} else {
|
||||
copy(paddedFeeIdx[:],
|
||||
aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
copy(paddedFeeIdx[:], aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
}
|
||||
feeIdxCoordinator, err := common.IdxFromBytes(paddedFeeIdx[:])
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
if feeIdxCoordinator != common.Idx(0) {
|
||||
rollupForgeBatchArgs.FeeIdxCoordinator =
|
||||
append(rollupForgeBatchArgs.FeeIdxCoordinator, feeIdxCoordinator)
|
||||
rollupForgeBatchArgs.FeeIdxCoordinator = append(rollupForgeBatchArgs.FeeIdxCoordinator, feeIdxCoordinator)
|
||||
}
|
||||
}
|
||||
return &rollupForgeBatchArgs, &sender, nil
|
||||
|
||||
@@ -116,8 +116,7 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
minBid.SetString("11000000000000000000", 10)
|
||||
budget := new(big.Int)
|
||||
budget.SetString("45200000000000000000", 10)
|
||||
_, err = auctionClient.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet,
|
||||
maxBid, minBid, deadline)
|
||||
_, err = auctionClient.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet, maxBid, minBid, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Add Blocks
|
||||
@@ -129,18 +128,12 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
|
||||
// Forge Batch 1
|
||||
args := new(RollupForgeBatchArgs)
|
||||
// When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
args.FeeIdxCoordinator = []common.Idx{}
|
||||
l1CoordinatorBytes, err := hex.DecodeString(
|
||||
"1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf" +
|
||||
"42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230" +
|
||||
"de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
||||
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
l1CoordinatorBytes, err := hex.DecodeString("1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
||||
require.NoError(t, err)
|
||||
numTxsL1 := len(l1CoordinatorBytes) / common.RollupConstL1CoordinatorTotalBytes
|
||||
for i := 0; i < numTxsL1; i++ {
|
||||
bytesL1Coordinator :=
|
||||
l1CoordinatorBytes[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*
|
||||
common.RollupConstL1CoordinatorTotalBytes]
|
||||
bytesL1Coordinator := l1CoordinatorBytes[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes]
|
||||
var signature []byte
|
||||
v := bytesL1Coordinator[0]
|
||||
s := bytesL1Coordinator[1:33]
|
||||
@@ -156,12 +149,9 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
args.L1UserTxs = []common.L1Tx{}
|
||||
args.L2TxsData = []common.L2Tx{}
|
||||
newStateRoot := new(big.Int)
|
||||
newStateRoot.SetString(
|
||||
"18317824016047294649053625209337295956588174734569560016974612130063629505228",
|
||||
10)
|
||||
newStateRoot.SetString("18317824016047294649053625209337295956588174734569560016974612130063629505228", 10)
|
||||
newExitRoot := new(big.Int)
|
||||
bytesNumExitRoot, err := hex.DecodeString(
|
||||
"10a89d5fe8d488eda1ba371d633515739933c706c210c604f5bd209180daa43b")
|
||||
bytesNumExitRoot, err := hex.DecodeString("10a89d5fe8d488eda1ba371d633515739933c706c210c604f5bd209180daa43b")
|
||||
require.NoError(t, err)
|
||||
newExitRoot.SetBytes(bytesNumExitRoot)
|
||||
args.NewLastIdx = int64(300)
|
||||
@@ -216,8 +206,7 @@ func TestRollupUpdateForgeL1L2BatchTimeout(t *testing.T) {
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, newForgeL1L2BatchTimeout,
|
||||
rollupEvents.UpdateForgeL1L2BatchTimeout[0].NewForgeL1L2BatchTimeout)
|
||||
assert.Equal(t, newForgeL1L2BatchTimeout, rollupEvents.UpdateForgeL1L2BatchTimeout[0].NewForgeL1L2BatchTimeout)
|
||||
}
|
||||
|
||||
func TestRollupUpdateFeeAddToken(t *testing.T) {
|
||||
@@ -234,15 +223,12 @@ func TestRollupUpdateFeeAddToken(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestRollupUpdateBucketsParameters(t *testing.T) {
|
||||
bucketsParameters := make([]RollupUpdateBucketsParameters, 5)
|
||||
ceilUSD, _ := new(big.Int).SetString("10000000", 10)
|
||||
var bucketsParameters [common.RollupConstNumBuckets]RollupUpdateBucketsParameters
|
||||
for i := range bucketsParameters {
|
||||
bucketsParameters[i].CeilUSD = big.NewInt(0).Mul(ceilUSD, big.NewInt(int64(i+1)))
|
||||
bucketsParameters[i].BlockStamp = big.NewInt(int64(0))
|
||||
bucketsParameters[i].CeilUSD = big.NewInt(int64((i + 1) * 100))
|
||||
bucketsParameters[i].Withdrawals = big.NewInt(int64(i + 1))
|
||||
bucketsParameters[i].RateBlocks = big.NewInt(int64(i+1) * 4)
|
||||
bucketsParameters[i].RateWithdrawals = big.NewInt(int64(3))
|
||||
bucketsParameters[i].MaxWithdrawals = big.NewInt(int64(1215752192))
|
||||
bucketsParameters[i].BlockWithdrawalRate = big.NewInt(int64(i+1) * 100)
|
||||
bucketsParameters[i].MaxWithdrawals = big.NewInt(int64(100000000000))
|
||||
}
|
||||
_, err := rollupClient.RollupUpdateBucketsParameters(bucketsParameters)
|
||||
require.NoError(t, err)
|
||||
@@ -251,14 +237,7 @@ func TestRollupUpdateBucketsParameters(t *testing.T) {
|
||||
blockStampBucket = currentBlockNum
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
for i := range bucketsParameters {
|
||||
assert.Equal(t, 0, bucketsParameters[i].CeilUSD.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].CeilUSD))
|
||||
assert.Equal(t, 0, bucketsParameters[i].BlockStamp.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].BlockStamp))
|
||||
assert.Equal(t, 0, bucketsParameters[i].Withdrawals.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].Withdrawals))
|
||||
assert.Equal(t, 0, bucketsParameters[i].RateBlocks.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].RateBlocks))
|
||||
assert.Equal(t, 0, bucketsParameters[i].RateWithdrawals.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].RateWithdrawals))
|
||||
assert.Equal(t, 0, bucketsParameters[i].MaxWithdrawals.Cmp(rollupEvents.UpdateBucketsParameters[0].ArrayBuckets[i].MaxWithdrawals))
|
||||
}
|
||||
assert.Equal(t, bucketsParameters, rollupEvents.UpdateBucketsParameters[0].ArrayBuckets)
|
||||
}
|
||||
|
||||
func TestRollupUpdateWithdrawalDelay(t *testing.T) {
|
||||
@@ -269,8 +248,7 @@ func TestRollupUpdateWithdrawalDelay(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, newWithdrawalDelay,
|
||||
int64(rollupEvents.UpdateWithdrawalDelay[0].NewWithdrawalDelay))
|
||||
assert.Equal(t, newWithdrawalDelay, int64(rollupEvents.UpdateWithdrawalDelay[0].NewWithdrawalDelay))
|
||||
}
|
||||
|
||||
func TestRollupUpdateTokenExchange(t *testing.T) {
|
||||
@@ -309,8 +287,7 @@ func TestRollupL1UserTxETHCreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -322,13 +299,11 @@ func TestRollupL1UserTxETHCreateAccountDeposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20CreateAccountDeposit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
key := genKeysBjj(1)
|
||||
fromIdxInt64 := int64(0)
|
||||
@@ -344,8 +319,7 @@ func TestRollupL1UserTxERC20CreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -357,13 +331,11 @@ func TestRollupL1UserTxERC20CreateAccountDeposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitCreateAccountDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
key := genKeysBjj(3)
|
||||
fromIdxInt64 := int64(0)
|
||||
@@ -379,8 +351,7 @@ func TestRollupL1UserTxERC20PermitCreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -392,13 +363,11 @@ func TestRollupL1UserTxERC20PermitCreateAccountDeposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -414,8 +383,7 @@ func TestRollupL1UserTxETHDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -426,13 +394,11 @@ func TestRollupL1UserTxETHDeposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20Deposit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -447,8 +413,7 @@ func TestRollupL1UserTxERC20Deposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -459,13 +424,11 @@ func TestRollupL1UserTxERC20Deposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -479,8 +442,7 @@ func TestRollupL1UserTxERC20PermitDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -491,13 +453,11 @@ func TestRollupL1UserTxERC20PermitDeposit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -513,8 +473,7 @@ func TestRollupL1UserTxETHDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -525,13 +484,11 @@ func TestRollupL1UserTxETHDepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20DepositTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -546,8 +503,7 @@ func TestRollupL1UserTxERC20DepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -558,13 +514,11 @@ func TestRollupL1UserTxERC20DepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(259)
|
||||
@@ -579,8 +533,7 @@ func TestRollupL1UserTxERC20PermitDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -591,13 +544,11 @@ func TestRollupL1UserTxERC20PermitDepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHCreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -613,8 +564,7 @@ func TestRollupL1UserTxETHCreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -625,13 +575,11 @@ func TestRollupL1UserTxETHCreateAccountDepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20CreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -646,8 +594,7 @@ func TestRollupL1UserTxERC20CreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -658,13 +605,11 @@ func TestRollupL1UserTxERC20CreateAccountDepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitCreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(259)
|
||||
@@ -679,8 +624,7 @@ func TestRollupL1UserTxERC20PermitCreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -691,13 +635,11 @@ func TestRollupL1UserTxERC20PermitCreateAccountDepositTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHForceTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -712,8 +654,7 @@ func TestRollupL1UserTxETHForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -724,13 +665,11 @@ func TestRollupL1UserTxETHForceTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20ForceTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -744,8 +683,7 @@ func TestRollupL1UserTxERC20ForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -756,13 +694,11 @@ func TestRollupL1UserTxERC20ForceTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitForceTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(259)
|
||||
toIdxInt64 := int64(260)
|
||||
@@ -776,8 +712,7 @@ func TestRollupL1UserTxERC20PermitForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -788,13 +723,11 @@ func TestRollupL1UserTxERC20PermitForceTransfer(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHForceExit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -809,8 +742,7 @@ func TestRollupL1UserTxETHForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -821,13 +753,11 @@ func TestRollupL1UserTxETHForceExit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20ForceExit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -841,8 +771,7 @@ func TestRollupL1UserTxERC20ForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -853,13 +782,11 @@ func TestRollupL1UserTxERC20ForceExit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitForceExit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -875,8 +802,7 @@ func TestRollupL1UserTxERC20PermitForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -887,8 +813,7 @@ func TestRollupL1UserTxERC20PermitForceExit(t *testing.T) {
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupForgeBatch2(t *testing.T) {
|
||||
@@ -904,8 +829,7 @@ func TestRollupForgeBatch2(t *testing.T) {
|
||||
|
||||
// Forge Batch 3
|
||||
args := new(RollupForgeBatchArgs)
|
||||
// When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
args.FeeIdxCoordinator = []common.Idx{}
|
||||
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
args.L1CoordinatorTxs = argsForge.L1CoordinatorTxs
|
||||
args.L1CoordinatorTxsAuths = argsForge.L1CoordinatorTxsAuths
|
||||
for i := 0; i < len(L1UserTxs); i++ {
|
||||
@@ -913,19 +837,14 @@ func TestRollupForgeBatch2(t *testing.T) {
|
||||
l1UserTx.EffectiveAmount = l1UserTx.Amount
|
||||
l1Bytes, err := l1UserTx.BytesDataAvailability(uint32(nLevels))
|
||||
require.NoError(t, err)
|
||||
l1UserTxDataAvailability, err := common.L1TxFromDataAvailability(l1Bytes,
|
||||
uint32(nLevels))
|
||||
l1UserTxDataAvailability, err := common.L1TxFromDataAvailability(l1Bytes, uint32(nLevels))
|
||||
require.NoError(t, err)
|
||||
args.L1UserTxs = append(args.L1UserTxs, *l1UserTxDataAvailability)
|
||||
}
|
||||
newStateRoot := new(big.Int)
|
||||
newStateRoot.SetString(
|
||||
"18317824016047294649053625209337295956588174734569560016974612130063629505228",
|
||||
10)
|
||||
newStateRoot.SetString("18317824016047294649053625209337295956588174734569560016974612130063629505228", 10)
|
||||
newExitRoot := new(big.Int)
|
||||
newExitRoot.SetString(
|
||||
"1114281409737474688393837964161044726766678436313681099613347372031079422302",
|
||||
10)
|
||||
newExitRoot.SetString("1114281409737474688393837964161044726766678436313681099613347372031079422302", 10)
|
||||
amount := new(big.Int)
|
||||
amount.SetString("79000000", 10)
|
||||
l2Tx := common.L2Tx{
|
||||
@@ -985,8 +904,7 @@ func TestRollupWithdrawMerkleProof(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompBE, err :=
|
||||
hex.DecodeString("adc3b754f8da621967b073a787bef8eec7052f2ba712b23af57d98f65beea8b2")
|
||||
pkCompBE, err := hex.DecodeString("adc3b754f8da621967b073a787bef8eec7052f2ba712b23af57d98f65beea8b2")
|
||||
require.NoError(t, err)
|
||||
pkCompLE := common.SwapEndianness(pkCompBE)
|
||||
copy(pkComp[:], pkCompLE)
|
||||
@@ -996,20 +914,16 @@ func TestRollupWithdrawMerkleProof(t *testing.T) {
|
||||
numExitRoot := int64(3)
|
||||
fromIdx := int64(256)
|
||||
amount, _ := new(big.Int).SetString("20000000000000000000", 10)
|
||||
// siblingBytes0, err := new(big.Int).SetString(
|
||||
// "19508838618377323910556678335932426220272947530531646682154552299216398748115",
|
||||
// 10)
|
||||
// siblingBytes0, err := new(big.Int).SetString("19508838618377323910556678335932426220272947530531646682154552299216398748115", 10)
|
||||
// require.NoError(t, err)
|
||||
// siblingBytes1, err := new(big.Int).SetString(
|
||||
// "15198806719713909654457742294233381653226080862567104272457668857208564789571", 10)
|
||||
// siblingBytes1, err := new(big.Int).SetString("15198806719713909654457742294233381653226080862567104272457668857208564789571", 10)
|
||||
// require.NoError(t, err)
|
||||
var siblings []*big.Int
|
||||
// siblings = append(siblings, siblingBytes0)
|
||||
// siblings = append(siblings, siblingBytes1)
|
||||
instantWithdraw := true
|
||||
|
||||
_, err = rollupClientAux.RollupWithdrawMerkleProof(pkComp, tokenID, numExitRoot, fromIdx,
|
||||
amount, siblings, instantWithdraw)
|
||||
_, err = rollupClientAux.RollupWithdrawMerkleProof(pkComp, tokenID, numExitRoot, fromIdx, amount, siblings, instantWithdraw)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
@@ -1026,9 +940,9 @@ func TestRollupWithdrawMerkleProof(t *testing.T) {
|
||||
// Bucket 1
|
||||
// Bucket[0].withdrawals = 1, Bucket[1].withdrawals = 2, ...
|
||||
// Bucket[1].withdrawals - 1 = 1
|
||||
assert.Equal(t, 0, rollupEvents.UpdateBucketWithdraw[0].NumBucket)
|
||||
assert.Equal(t, int64(442), rollupEvents.UpdateBucketWithdraw[0].BlockStamp)
|
||||
assert.Equal(t, big.NewInt(15), rollupEvents.UpdateBucketWithdraw[0].Withdrawals)
|
||||
assert.Equal(t, 1, rollupEvents.UpdateBucketWithdraw[0].NumBucket)
|
||||
assert.Equal(t, blockStampBucket, rollupEvents.UpdateBucketWithdraw[0].BlockStamp)
|
||||
assert.Equal(t, big.NewInt(1), rollupEvents.UpdateBucketWithdraw[0].Withdrawals)
|
||||
}
|
||||
|
||||
func TestRollupSafeMode(t *testing.T) {
|
||||
|
||||
@@ -132,8 +132,7 @@ type WDelayerInterface interface {
|
||||
WDelayerDepositInfo(owner, token ethCommon.Address) (depositInfo DepositState, err error)
|
||||
WDelayerDeposit(onwer, token ethCommon.Address, amount *big.Int) (*types.Transaction, error)
|
||||
WDelayerWithdrawal(owner, token ethCommon.Address) (*types.Transaction, error)
|
||||
WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address,
|
||||
amount *big.Int) (*types.Transaction, error)
|
||||
WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address, amount *big.Int) (*types.Transaction, error)
|
||||
|
||||
WDelayerEventsByBlock(blockNum int64, blockHash *ethCommon.Hash) (*WDelayerEvents, error)
|
||||
WDelayerConstants() (*common.WDelayerConstants, error)
|
||||
@@ -144,8 +143,7 @@ type WDelayerInterface interface {
|
||||
// Implementation
|
||||
//
|
||||
|
||||
// WDelayerClient is the implementation of the interface to the WithdrawDelayer
|
||||
// Smart Contract in ethereum.
|
||||
// WDelayerClient is the implementation of the interface to the WithdrawDelayer Smart Contract in ethereum.
|
||||
type WDelayerClient struct {
|
||||
client *EthereumClient
|
||||
address ethCommon.Address
|
||||
@@ -174,8 +172,7 @@ func NewWDelayerClient(client *EthereumClient, address ethCommon.Address) (*WDel
|
||||
}
|
||||
|
||||
// WDelayerGetHermezGovernanceAddress is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (
|
||||
hermezGovernanceAddress *ethCommon.Address, err error) {
|
||||
func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (hermezGovernanceAddress *ethCommon.Address, err error) {
|
||||
var _hermezGovernanceAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_hermezGovernanceAddress, err = c.wdelayer.GetHermezGovernanceAddress(c.opts)
|
||||
@@ -187,8 +184,7 @@ func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (
|
||||
}
|
||||
|
||||
// WDelayerTransferGovernance is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerTransferGovernance(newAddress ethCommon.Address) (
|
||||
tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerTransferGovernance(newAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -214,8 +210,7 @@ func (c *WDelayerClient) WDelayerClaimGovernance() (tx *types.Transaction, err e
|
||||
}
|
||||
|
||||
// WDelayerGetEmergencyCouncil is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress *ethCommon.Address,
|
||||
err error) {
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress *ethCommon.Address, err error) {
|
||||
var _emergencyCouncilAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_emergencyCouncilAddress, err = c.wdelayer.GetEmergencyCouncil(c.opts)
|
||||
@@ -227,8 +222,7 @@ func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress
|
||||
}
|
||||
|
||||
// WDelayerTransferEmergencyCouncil is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerTransferEmergencyCouncil(newAddress ethCommon.Address) (
|
||||
tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerTransferEmergencyCouncil(newAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -277,8 +271,7 @@ func (c *WDelayerClient) WDelayerGetWithdrawalDelay() (withdrawalDelay int64, er
|
||||
}
|
||||
|
||||
// WDelayerGetEmergencyModeStartingTime is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyModeStartingTime() (emergencyModeStartingTime int64,
|
||||
err error) {
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyModeStartingTime() (emergencyModeStartingTime int64, err error) {
|
||||
var _emergencyModeStartingTime uint64
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_emergencyModeStartingTime, err = c.wdelayer.GetEmergencyModeStartingTime(c.opts)
|
||||
@@ -303,8 +296,7 @@ func (c *WDelayerClient) WDelayerEnableEmergencyMode() (tx *types.Transaction, e
|
||||
}
|
||||
|
||||
// WDelayerChangeWithdrawalDelay is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64) (
|
||||
tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -317,8 +309,7 @@ func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64
|
||||
}
|
||||
|
||||
// WDelayerDepositInfo is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (
|
||||
depositInfo DepositState, err error) {
|
||||
func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (depositInfo DepositState, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
amount, depositTimestamp, err := c.wdelayer.DepositInfo(c.opts, owner, token)
|
||||
depositInfo.Amount = amount
|
||||
@@ -331,8 +322,7 @@ func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (
|
||||
}
|
||||
|
||||
// WDelayerDeposit is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount *big.Int) (
|
||||
tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -345,8 +335,7 @@ func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount
|
||||
}
|
||||
|
||||
// WDelayerWithdrawal is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx *types.Transaction,
|
||||
err error) {
|
||||
func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -359,8 +348,7 @@ func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx
|
||||
}
|
||||
|
||||
// WDelayerEscapeHatchWithdrawal is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address,
|
||||
amount *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address, amount *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -396,21 +384,14 @@ func (c *WDelayerClient) WDelayerConstants() (constants *common.WDelayerConstant
|
||||
}
|
||||
|
||||
var (
|
||||
logWDelayerDeposit = crypto.Keccak256Hash([]byte(
|
||||
"Deposit(address,address,uint192,uint64)"))
|
||||
logWDelayerWithdraw = crypto.Keccak256Hash([]byte(
|
||||
"Withdraw(address,address,uint192)"))
|
||||
logWDelayerEmergencyModeEnabled = crypto.Keccak256Hash([]byte(
|
||||
"EmergencyModeEnabled()"))
|
||||
logWDelayerNewWithdrawalDelay = crypto.Keccak256Hash([]byte(
|
||||
"NewWithdrawalDelay(uint64)"))
|
||||
logWDelayerEscapeHatchWithdrawal = crypto.Keccak256Hash([]byte(
|
||||
"EscapeHatchWithdrawal(address,address,address,uint256)"))
|
||||
logWDelayerNewEmergencyCouncil = crypto.Keccak256Hash([]byte(
|
||||
"NewEmergencyCouncil(address)"))
|
||||
logWDelayerNewHermezGovernanceAddress = crypto.Keccak256Hash([]byte(
|
||||
"NewHermezGovernanceAddress(address)"))
|
||||
logWDelayerInitialize = crypto.Keccak256Hash([]byte(
|
||||
logWDelayerDeposit = crypto.Keccak256Hash([]byte("Deposit(address,address,uint192,uint64)"))
|
||||
logWDelayerWithdraw = crypto.Keccak256Hash([]byte("Withdraw(address,address,uint192)"))
|
||||
logWDelayerEmergencyModeEnabled = crypto.Keccak256Hash([]byte("EmergencyModeEnabled()"))
|
||||
logWDelayerNewWithdrawalDelay = crypto.Keccak256Hash([]byte("NewWithdrawalDelay(uint64)"))
|
||||
logWDelayerEscapeHatchWithdrawal = crypto.Keccak256Hash([]byte("EscapeHatchWithdrawal(address,address,address,uint256)"))
|
||||
logWDelayerNewEmergencyCouncil = crypto.Keccak256Hash([]byte("NewEmergencyCouncil(address)"))
|
||||
logWDelayerNewHermezGovernanceAddress = crypto.Keccak256Hash([]byte("NewHermezGovernanceAddress(address)"))
|
||||
logWDelayerInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeWithdrawalDelayerEvent(uint64,address,address)"))
|
||||
)
|
||||
|
||||
@@ -502,51 +483,42 @@ func (c *WDelayerClient) WDelayerEventsByBlock(blockNum int64,
|
||||
|
||||
case logWDelayerEmergencyModeEnabled:
|
||||
var emergencyModeEnabled WDelayerEventEmergencyModeEnabled
|
||||
wdelayerEvents.EmergencyModeEnabled =
|
||||
append(wdelayerEvents.EmergencyModeEnabled, emergencyModeEnabled)
|
||||
wdelayerEvents.EmergencyModeEnabled = append(wdelayerEvents.EmergencyModeEnabled, emergencyModeEnabled)
|
||||
|
||||
case logWDelayerNewWithdrawalDelay:
|
||||
var withdrawalDelay WDelayerEventNewWithdrawalDelay
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdrawalDelay,
|
||||
"NewWithdrawalDelay", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdrawalDelay, "NewWithdrawalDelay", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewWithdrawalDelay =
|
||||
append(wdelayerEvents.NewWithdrawalDelay, withdrawalDelay)
|
||||
wdelayerEvents.NewWithdrawalDelay = append(wdelayerEvents.NewWithdrawalDelay, withdrawalDelay)
|
||||
|
||||
case logWDelayerEscapeHatchWithdrawal:
|
||||
var escapeHatchWithdrawal WDelayerEventEscapeHatchWithdrawal
|
||||
err := c.contractAbi.UnpackIntoInterface(&escapeHatchWithdrawal,
|
||||
"EscapeHatchWithdrawal", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&escapeHatchWithdrawal, "EscapeHatchWithdrawal", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
escapeHatchWithdrawal.Who = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
escapeHatchWithdrawal.To = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
escapeHatchWithdrawal.Token = ethCommon.BytesToAddress(vLog.Topics[3].Bytes())
|
||||
wdelayerEvents.EscapeHatchWithdrawal =
|
||||
append(wdelayerEvents.EscapeHatchWithdrawal, escapeHatchWithdrawal)
|
||||
wdelayerEvents.EscapeHatchWithdrawal = append(wdelayerEvents.EscapeHatchWithdrawal, escapeHatchWithdrawal)
|
||||
|
||||
case logWDelayerNewEmergencyCouncil:
|
||||
var emergencyCouncil WDelayerEventNewEmergencyCouncil
|
||||
err := c.contractAbi.UnpackIntoInterface(&emergencyCouncil,
|
||||
"NewEmergencyCouncil", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&emergencyCouncil, "NewEmergencyCouncil", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewEmergencyCouncil =
|
||||
append(wdelayerEvents.NewEmergencyCouncil, emergencyCouncil)
|
||||
wdelayerEvents.NewEmergencyCouncil = append(wdelayerEvents.NewEmergencyCouncil, emergencyCouncil)
|
||||
|
||||
case logWDelayerNewHermezGovernanceAddress:
|
||||
var governanceAddress WDelayerEventNewHermezGovernanceAddress
|
||||
err := c.contractAbi.UnpackIntoInterface(&governanceAddress,
|
||||
"NewHermezGovernanceAddress", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&governanceAddress, "NewHermezGovernanceAddress", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewHermezGovernanceAddress =
|
||||
append(wdelayerEvents.NewHermezGovernanceAddress, governanceAddress)
|
||||
wdelayerEvents.NewHermezGovernanceAddress = append(wdelayerEvents.NewHermezGovernanceAddress, governanceAddress)
|
||||
}
|
||||
}
|
||||
return &wdelayerEvents, nil
|
||||
|
||||
@@ -54,8 +54,7 @@ func TestWDelayerSetHermezGovernanceAddress(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, auxAddressConst,
|
||||
wdelayerEvents.NewHermezGovernanceAddress[0].NewHermezGovernanceAddress)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.NewHermezGovernanceAddress[0].NewHermezGovernanceAddress)
|
||||
_, err = wdelayerClientAux.WDelayerTransferGovernance(governanceAddressConst)
|
||||
require.Nil(t, err)
|
||||
_, err = wdelayerClientTest.WDelayerClaimGovernance()
|
||||
@@ -69,8 +68,7 @@ func TestWDelayerGetEmergencyCouncil(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestWDelayerSetEmergencyCouncil(t *testing.T) {
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil,
|
||||
wdelayerTestAddressConst)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil, wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
wdelayerClientAux, err := NewWDelayerClient(ethereumClientAux, wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
@@ -202,18 +200,13 @@ func TestWDelayerGetEmergencyModeStartingTime(t *testing.T) {
|
||||
func TestWDelayerEscapeHatchWithdrawal(t *testing.T) {
|
||||
amount := new(big.Int)
|
||||
amount.SetString("10000000000000000", 10)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil,
|
||||
wdelayerTestAddressConst)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil, wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
_, err =
|
||||
wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst,
|
||||
tokenHEZAddressConst, amount)
|
||||
_, err = wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst, tokenHEZAddressConst, amount)
|
||||
require.Contains(t, err.Error(), "NO_MAX_EMERGENCY_MODE_TIME")
|
||||
seconds := maxEmergencyModeTime.Seconds()
|
||||
addTime(seconds, ethClientDialURL)
|
||||
_, err =
|
||||
wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst,
|
||||
tokenHEZAddressConst, amount)
|
||||
_, err = wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst, tokenHEZAddressConst, amount)
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
|
||||
4
go.mod
4
go.mod
@@ -11,8 +11,8 @@ require (
|
||||
github.com/gin-gonic/gin v1.5.0
|
||||
github.com/gobuffalo/packr/v2 v2.8.1
|
||||
github.com/hermeznetwork/tracerr v0.3.1-0.20210120162744-5da60b576169
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12
|
||||
github.com/jinzhu/copier v0.0.0-20190924061706-b57f9002281a
|
||||
github.com/jmoiron/sqlx v1.2.1-0.20200615141059-0794cb1f47ee
|
||||
github.com/joho/godotenv v1.3.0
|
||||
|
||||
19
go.sum
19
go.sum
@@ -24,6 +24,8 @@ github.com/DataDog/zstd v1.3.6-0.20190409195224-796139022798/go.mod h1:1jcaCB/uf
|
||||
github.com/Joker/hpp v1.0.0/go.mod h1:8x5n+M1Hp5hC0g8okX3sR3vFQwynaX/UgSOM9MeBKzY=
|
||||
github.com/Joker/jade v1.0.1-0.20190614124447-d475f43051e7/go.mod h1:6E6s8o2AE4KhCrqr6GRJjdC/gNfTdxkIXvuGZZda2VM=
|
||||
github.com/Knetic/govaluate v3.0.1-0.20171022003610-9aa49832a739+incompatible/go.mod h1:r7JcOSlj0wfOMncg0iLm8Leh48TZaKVeNIfJntJ2wa0=
|
||||
github.com/KyleBanks/depth v1.2.1 h1:5h8fQADFrWtarTdtDudMmGsC7GPbOAu6RVB3ffsVFHc=
|
||||
github.com/KyleBanks/depth v1.2.1/go.mod h1:jzSb9d0L43HxTQfT+oSA1EEp2q+ne2uh6XgeJcm8brE=
|
||||
github.com/OneOfOne/xxhash v1.2.2/go.mod h1:HSdplMjZKSmBqAxg5vPj2TmRDmfkzw+cTzAElWljhcU=
|
||||
github.com/OneOfOne/xxhash v1.2.5/go.mod h1:eZbhyaAYD41SGSSsnmcpxVoRiQ/MPUTjUdIIOT9Um7Q=
|
||||
github.com/Shopify/goreferrer v0.0.0-20181106222321-ec9c9a553398/go.mod h1:a1uqRtAwp2Xwc6WNPJEufxJ7fx3npB4UV/JOLmbu5I0=
|
||||
@@ -85,6 +87,8 @@ github.com/cenkalti/backoff v2.2.1+incompatible/go.mod h1:90ReRw6GdpyfrHakVjL/QH
|
||||
github.com/census-instrumentation/opencensus-proto v0.2.1/go.mod h1:f6KPmirojxKA12rnyqOA5BBL4O983OfeGPqjHWSTneU=
|
||||
github.com/cespare/cp v0.1.0 h1:SE+dxFebS7Iik5LK0tsi1k9ZCxEaFX4AjQmoyA+1dJk=
|
||||
github.com/cespare/cp v0.1.0/go.mod h1:SOGHArjBr4JWaSDEVpWpo/hNg6RoKrls6Oh40hiwW+s=
|
||||
github.com/cespare/cp v1.1.1 h1:nCb6ZLdB7NRaqsm91JtQTAme2SKJzXVsdPIPkyJr1MU=
|
||||
github.com/cespare/cp v1.1.1/go.mod h1:SOGHArjBr4JWaSDEVpWpo/hNg6RoKrls6Oh40hiwW+s=
|
||||
github.com/cespare/xxhash v1.1.0 h1:a6HrQnmkObjyL+Gs60czilIUGqrzKutQD6XZog3p+ko=
|
||||
github.com/cespare/xxhash v1.1.0/go.mod h1:XrSqR1VqqWfGrhpAt58auRo0WTKS1nRRg3ghfAqPWnc=
|
||||
github.com/cespare/xxhash/v2 v2.0.1-0.20190104013014-3767db7a7e18/go.mod h1:HD5P3vAIAh+Y2GAxg0PrPN1P8WkepXGpjbUPDHJqqKM=
|
||||
@@ -170,6 +174,8 @@ github.com/fatih/color v1.7.0/go.mod h1:Zm6kSWBoL9eyXnKyktHP6abPY2pDugNf5Kwzbycv
|
||||
github.com/fatih/structs v1.1.0/go.mod h1:9NiDSp5zOcgEDl+j00MP/WkGVPOlPRLejGD8Ga6PJ7M=
|
||||
github.com/fjl/memsize v0.0.0-20180418122429-ca190fb6ffbc h1:jtW8jbpkO4YirRSyepBOH8E+2HEw6/hKkBvFPwhUN8c=
|
||||
github.com/fjl/memsize v0.0.0-20180418122429-ca190fb6ffbc/go.mod h1:VvhXpOYNQvB+uIk2RvXzuaQtkQJzzIx6lSBe1xv7hi0=
|
||||
github.com/fjl/memsize v0.0.0-20190710130421-bcb5799ab5e5 h1:FtmdgXiUlNeRsoNMFlKLDt+S+6hbjVMEW6RGQ7aUf7c=
|
||||
github.com/fjl/memsize v0.0.0-20190710130421-bcb5799ab5e5/go.mod h1:VvhXpOYNQvB+uIk2RvXzuaQtkQJzzIx6lSBe1xv7hi0=
|
||||
github.com/flosch/pongo2 v0.0.0-20190707114632-bbf5a6c351f4/go.mod h1:T9YF2M40nIgbVgp3rreNmTged+9HrbNTIQf1PsaIiTA=
|
||||
github.com/franela/goblin v0.0.0-20200105215937-c9ffbefa60db/go.mod h1:7dvUGVsVBjqR7JHJk0brhHOZYGmfBYOrK0ZhYMEtBr4=
|
||||
github.com/franela/goreq v0.0.0-20171204163338-bcd34c9993f8/go.mod h1:ZhphrRTfi2rbfLwlschooIH4+wKKDR4Pdxhh+TRoA20=
|
||||
@@ -330,10 +336,11 @@ github.com/huin/goupnp v1.0.0 h1:wg75sLpL6DZqwHQN6E1Cfk6mtfzS45z8OV+ic+DtHRo=
|
||||
github.com/huin/goupnp v1.0.0/go.mod h1:n9v9KO1tAxYH82qOn+UTIFQDmx5n1Zxd/ClZDMX7Bnc=
|
||||
github.com/huin/goutil v0.0.0-20170803182201-1ca381bf3150/go.mod h1:PpLOETDnJ0o3iZrZfqZzyLl6l7F3c6L1oWn7OICBi6o=
|
||||
github.com/hydrogen18/memlistener v0.0.0-20141126152155-54553eb933fb/go.mod h1:qEIFzExnS6016fRpRfxrExeVn2gbClQA99gQhnIcdhE=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef h1:72PG9b2eDlLqKszJVLrsoJbpt4CtgJLhKOjH1MJqCVY=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189 h1:hoarWk/SwNwnMXE0kiskcZULW0XBLQIUGYK4C39ozfs=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189/go.mod h1:56abMeBKD4BIFe346rk+yuJ4MQgfMHe28sRx4o2gOpk=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201218111145-a2015adb2f1b/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c h1:D2u8FFYey6iFXLsqqJZ8R7ch8gZum+/b98whvoSDbyg=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12 h1:DXWT0BLCSm7cJmTMQy7+iOlxkA1/5ADglufhLK52e10=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12/go.mod h1:FdUFTW2qJiwHyy5R70uErwq7Kaq1uskyFdTfodcUJqA=
|
||||
github.com/imkira/go-interpol v1.1.0/go.mod h1:z0h2/2T3XF8kyEPpRgJ3kmNv+C43p+I/CoI+jC3w2iA=
|
||||
github.com/inconshreveable/mousetrap v1.0.0/go.mod h1:PxqpIevigyE2G7u3NXJIT2ANytuPF1OarO4DADm73n8=
|
||||
github.com/influxdata/influxdb v1.2.3-0.20180221223340-01288bdb0883/go.mod h1:qZna6X/4elxqT3yI9iZYdZrWWdeFOOprn86kgg4+IzY=
|
||||
@@ -601,6 +608,8 @@ github.com/shurcooL/sanitized_anchor_name v1.0.0/go.mod h1:1NzhyTcUVG4SuEtjjoZeV
|
||||
github.com/sirupsen/logrus v1.2.0/go.mod h1:LxeOpSwHxABJmUn/MG1IvRgCAasNZTLOkJPxbbu5VWo=
|
||||
github.com/sirupsen/logrus v1.4.2 h1:SPIRibHv4MatM3XXNO2BJeFLZwZ2LvZgfQ5+UNI2im4=
|
||||
github.com/sirupsen/logrus v1.4.2/go.mod h1:tLMulIdttU9McNUspp0xgXVQah82FyeX6MwdIuYE2rE=
|
||||
github.com/sirupsen/logrus v1.5.0 h1:1N5EYkVAPEywqZRJd7cwnRtCb6xJx7NH3T3WUTF980Q=
|
||||
github.com/sirupsen/logrus v1.5.0/go.mod h1:+F7Ogzej0PZc/94MaYx/nvG9jOFMD2osvC3s+Squfpo=
|
||||
github.com/smartystreets/assertions v0.0.0-20180927180507-b2de0cb4f26d/go.mod h1:OnSkiWE9lh6wB0YB77sQom3nweQdgAjqCqsofrRNTgc=
|
||||
github.com/smartystreets/goconvey v1.6.4/go.mod h1:syvi0/a8iFYH4r/RixwvyeAJjdLS9QV7WQ/tjFTllLA=
|
||||
github.com/soheilhy/cmux v0.1.4/go.mod h1:IM3LyeVVIOuxMH7sFAkER9+bJ4dT7Ms6E4xg4kGIyLM=
|
||||
@@ -619,6 +628,8 @@ github.com/spf13/viper v1.3.2/go.mod h1:ZiWeW+zYFKm7srdB9IoDzzZXaJaI5eL9QjNiN/DM
|
||||
github.com/spf13/viper v1.4.0/go.mod h1:PTJ7Z/lr49W6bUbkmS1V3by4uWynFiR9p7+dSq/yZzE=
|
||||
github.com/status-im/keycard-go v0.0.0-20190316090335-8537d3370df4 h1:Gb2Tyox57NRNuZ2d3rmvB3pcmbu7O1RS3m8WRx7ilrg=
|
||||
github.com/status-im/keycard-go v0.0.0-20190316090335-8537d3370df4/go.mod h1:RZLeN1LMWmRsyYjvAu+I6Dm9QmlDaIIt+Y+4Kd7Tp+Q=
|
||||
github.com/status-im/keycard-go v0.0.0-20190424133014-d95853db0f48 h1:ju5UTwk5Odtm4trrY+4Ca4RMj5OyXbmVeDAVad2T0Jw=
|
||||
github.com/status-im/keycard-go v0.0.0-20190424133014-d95853db0f48/go.mod h1:RZLeN1LMWmRsyYjvAu+I6Dm9QmlDaIIt+Y+4Kd7Tp+Q=
|
||||
github.com/steakknife/bloomfilter v0.0.0-20180922174646-6819c0d2a570 h1:gIlAHnH1vJb5vwEjIp5kBj/eu99p/bl0Ay2goiPe5xE=
|
||||
github.com/steakknife/bloomfilter v0.0.0-20180922174646-6819c0d2a570/go.mod h1:8OR4w3TdeIHIh1g6EMY5p0gVNOovcWC+1vpc7naMuAw=
|
||||
github.com/steakknife/hamming v0.0.0-20180906055917-c99c65617cd3 h1:njlZPzLwU639dk2kqnCPPv+wNjq7Xb6EfUxe/oX0/NM=
|
||||
|
||||
255
node/node.go
255
node/node.go
@@ -27,7 +27,6 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/priceupdater"
|
||||
"github.com/hermeznetwork/hermez-node/prover"
|
||||
"github.com/hermeznetwork/hermez-node/stateapiupdater"
|
||||
"github.com/hermeznetwork/hermez-node/synchronizer"
|
||||
"github.com/hermeznetwork/hermez-node/test/debugapi"
|
||||
"github.com/hermeznetwork/hermez-node/txprocessor"
|
||||
@@ -55,7 +54,7 @@ const (
|
||||
// Node is the Hermez Node
|
||||
type Node struct {
|
||||
nodeAPI *NodeAPI
|
||||
stateAPIUpdater *stateapiupdater.Updater
|
||||
apiStateUpdater *api.APIStateUpdater
|
||||
debugAPI *debugapi.DebugAPI
|
||||
priceUpdater *priceupdater.PriceUpdater
|
||||
// Coordinator
|
||||
@@ -110,7 +109,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
}
|
||||
var apiConnCon *dbUtils.APIConnectionController
|
||||
if cfg.API.Explorer || mode == ModeCoordinator {
|
||||
apiConnCon = dbUtils.NewAPIConnectionController(
|
||||
apiConnCon = dbUtils.NewAPICnnectionController(
|
||||
cfg.API.MaxSQLConnections,
|
||||
cfg.API.SQLConnectionTimeout.Duration,
|
||||
)
|
||||
@@ -149,12 +148,12 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
if minForgeBalance != nil && balance.Cmp(minForgeBalance) == -1 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf(
|
||||
"forger account balance is less than cfg.Coordinator.MinimumForgeAddressBalance: %v < %v",
|
||||
balance, minForgeBalance))
|
||||
balance.Int64(), minForgeBalance))
|
||||
}
|
||||
log.Infow("forger ethereum account balance",
|
||||
"addr", cfg.Coordinator.ForgerAddress,
|
||||
"balance", balance,
|
||||
"minForgeBalance", minForgeBalance,
|
||||
"balance", balance.Int64(),
|
||||
"minForgeBalance", minForgeBalance.Int64(),
|
||||
)
|
||||
|
||||
// Unlock Coordinator ForgerAddr in the keystore to make calls
|
||||
@@ -223,19 +222,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var l2DB *l2db.L2DB
|
||||
if mode == ModeCoordinator {
|
||||
l2DB = l2db.NewL2DB(
|
||||
dbRead, dbWrite,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
apiConnCon,
|
||||
)
|
||||
}
|
||||
|
||||
sync, err := synchronizer.NewSynchronizer(client, historyDB, l2DB, stateDB, synchronizer.Config{
|
||||
sync, err := synchronizer.NewSynchronizer(client, historyDB, stateDB, synchronizer.Config{
|
||||
StatsRefreshPeriod: cfg.Synchronizer.StatsRefreshPeriod.Duration,
|
||||
ChainID: chainIDU16,
|
||||
})
|
||||
@@ -253,7 +240,6 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
hdbNodeCfg := historydb.NodeConfig{
|
||||
MaxPoolTxs: cfg.Coordinator.L2DB.MaxTxs,
|
||||
MinFeeUSD: cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
ForgeDelay: cfg.Coordinator.ForgeDelay.Duration.Seconds(),
|
||||
}
|
||||
if err := historyDB.SetNodeConfig(&hdbNodeCfg); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -271,10 +257,20 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
stateAPIUpdater := stateapiupdater.NewUpdater(historyDB, &hdbNodeCfg, initSCVars, &hdbConsts)
|
||||
apiStateUpdater := api.NewAPIStateUpdater(historyDB, &hdbNodeCfg, initSCVars, &hdbConsts)
|
||||
|
||||
var coord *coordinator.Coordinator
|
||||
var l2DB *l2db.L2DB
|
||||
if mode == ModeCoordinator {
|
||||
l2DB = l2db.NewL2DB(
|
||||
dbRead, dbWrite,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
apiConnCon,
|
||||
)
|
||||
|
||||
// Unlock FeeAccount EthAddr in the keystore to generate the
|
||||
// account creation authorization
|
||||
if !keyStore.HasAddress(cfg.Coordinator.FeeAccount.Address) {
|
||||
@@ -296,15 +292,14 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
if err := auth.Sign(func(msg []byte) ([]byte, error) {
|
||||
return keyStore.SignHash(feeAccount, msg)
|
||||
}, chainIDU16, cfg.SmartContracts.Rollup); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, err
|
||||
}
|
||||
coordAccount := &txselector.CoordAccount{
|
||||
Addr: cfg.Coordinator.FeeAccount.Address,
|
||||
BJJ: cfg.Coordinator.FeeAccount.BJJ,
|
||||
AccountCreationAuth: auth.Signature,
|
||||
}
|
||||
txSelector, err := txselector.NewTxSelector(coordAccount,
|
||||
cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
||||
txSelector, err := txselector.NewTxSelector(coordAccount, cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -360,25 +355,22 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
|
||||
coord, err = coordinator.NewCoordinator(
|
||||
coordinator.Config{
|
||||
ForgerAddress: cfg.Coordinator.ForgerAddress,
|
||||
ConfirmBlocks: cfg.Coordinator.ConfirmBlocks,
|
||||
L1BatchTimeoutPerc: cfg.Coordinator.L1BatchTimeoutPerc,
|
||||
ForgeRetryInterval: cfg.Coordinator.ForgeRetryInterval.Duration,
|
||||
ForgeDelay: cfg.Coordinator.ForgeDelay.Duration,
|
||||
MustForgeAtSlotDeadline: cfg.Coordinator.MustForgeAtSlotDeadline,
|
||||
IgnoreSlotCommitment: cfg.Coordinator.IgnoreSlotCommitment,
|
||||
ForgeOncePerSlotIfTxs: cfg.Coordinator.ForgeOncePerSlotIfTxs,
|
||||
ForgeNoTxsDelay: cfg.Coordinator.ForgeNoTxsDelay.Duration,
|
||||
SyncRetryInterval: cfg.Coordinator.SyncRetryInterval.Duration,
|
||||
PurgeByExtDelInterval: cfg.Coordinator.PurgeByExtDelInterval.Duration,
|
||||
EthClientAttempts: cfg.Coordinator.EthClient.Attempts,
|
||||
EthClientAttemptsDelay: cfg.Coordinator.EthClient.AttemptsDelay.Duration,
|
||||
EthNoReuseNonce: cfg.Coordinator.EthClient.NoReuseNonce,
|
||||
EthTxResendTimeout: cfg.Coordinator.EthClient.TxResendTimeout.Duration,
|
||||
MaxGasPrice: cfg.Coordinator.EthClient.MaxGasPrice,
|
||||
GasPriceIncPerc: cfg.Coordinator.EthClient.GasPriceIncPerc,
|
||||
TxManagerCheckInterval: cfg.Coordinator.EthClient.CheckLoopInterval.Duration,
|
||||
DebugBatchPath: cfg.Coordinator.Debug.BatchPath,
|
||||
ForgerAddress: cfg.Coordinator.ForgerAddress,
|
||||
ConfirmBlocks: cfg.Coordinator.ConfirmBlocks,
|
||||
L1BatchTimeoutPerc: cfg.Coordinator.L1BatchTimeoutPerc,
|
||||
ForgeRetryInterval: cfg.Coordinator.ForgeRetryInterval.Duration,
|
||||
ForgeDelay: cfg.Coordinator.ForgeDelay.Duration,
|
||||
ForgeNoTxsDelay: cfg.Coordinator.ForgeNoTxsDelay.Duration,
|
||||
SyncRetryInterval: cfg.Coordinator.SyncRetryInterval.Duration,
|
||||
PurgeByExtDelInterval: cfg.Coordinator.PurgeByExtDelInterval.Duration,
|
||||
EthClientAttempts: cfg.Coordinator.EthClient.Attempts,
|
||||
EthClientAttemptsDelay: cfg.Coordinator.EthClient.AttemptsDelay.Duration,
|
||||
EthNoReuseNonce: cfg.Coordinator.EthClient.NoReuseNonce,
|
||||
EthTxResendTimeout: cfg.Coordinator.EthClient.TxResendTimeout.Duration,
|
||||
MaxGasPrice: cfg.Coordinator.EthClient.MaxGasPrice,
|
||||
GasPriceIncPerc: cfg.Coordinator.EthClient.GasPriceIncPerc,
|
||||
TxManagerCheckInterval: cfg.Coordinator.EthClient.CheckLoopInterval.Duration,
|
||||
DebugBatchPath: cfg.Coordinator.Debug.BatchPath,
|
||||
Purger: coordinator.PurgerCfg{
|
||||
PurgeBatchDelay: cfg.Coordinator.L2DB.PurgeBatchDelay,
|
||||
InvalidateBatchDelay: cfg.Coordinator.L2DB.InvalidateBatchDelay,
|
||||
@@ -438,19 +430,14 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
if cfg.Debug.APIAddress != "" {
|
||||
debugAPI = debugapi.NewDebugAPI(cfg.Debug.APIAddress, stateDB, sync)
|
||||
}
|
||||
priceUpdater, err := priceupdater.NewPriceUpdater(
|
||||
cfg.PriceUpdater.DefaultUpdateMethod,
|
||||
cfg.PriceUpdater.TokensConfig,
|
||||
historyDB,
|
||||
cfg.PriceUpdater.URLBitfinexV2,
|
||||
cfg.PriceUpdater.URLCoinGeckoV3,
|
||||
)
|
||||
priceUpdater, err := priceupdater.NewPriceUpdater(cfg.PriceUpdater.URL,
|
||||
priceupdater.APIType(cfg.PriceUpdater.Type), historyDB)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
return &Node{
|
||||
stateAPIUpdater: stateAPIUpdater,
|
||||
apiStateUpdater: apiStateUpdater,
|
||||
nodeAPI: nodeAPI,
|
||||
debugAPI: debugAPI,
|
||||
priceUpdater: priceUpdater,
|
||||
@@ -469,14 +456,11 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
// APIServer is a server that only runs the API
|
||||
type APIServer struct {
|
||||
nodeAPI *NodeAPI
|
||||
mode Mode
|
||||
ctx context.Context
|
||||
wg sync.WaitGroup
|
||||
cancel context.CancelFunc
|
||||
}
|
||||
|
||||
// NewAPIServer creates a new APIServer
|
||||
func NewAPIServer(mode Mode, cfg *config.APIServer) (*APIServer, error) {
|
||||
// NOTE: I just copied some parts of NewNode related to starting the
|
||||
// API, but it still cotains many parameters that are not available
|
||||
meddler.Debug = cfg.Debug.MeddlerLogs
|
||||
// Stablish DB connection
|
||||
dbWrite, err := dbUtils.InitSQLDB(
|
||||
@@ -508,10 +492,13 @@ func NewAPIServer(mode Mode, cfg *config.APIServer) (*APIServer, error) {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||
}
|
||||
}
|
||||
apiConnCon := dbUtils.NewAPIConnectionController(
|
||||
cfg.API.MaxSQLConnections,
|
||||
cfg.API.SQLConnectionTimeout.Duration,
|
||||
)
|
||||
var apiConnCon *dbUtils.APIConnectionController
|
||||
if cfg.API.Explorer || mode == ModeCoordinator {
|
||||
apiConnCon = dbUtils.NewAPICnnectionController(
|
||||
cfg.API.MaxSQLConnections,
|
||||
cfg.API.SQLConnectionTimeout.Duration,
|
||||
)
|
||||
}
|
||||
|
||||
historyDB := historydb.NewHistoryDB(dbRead, dbWrite, apiConnCon)
|
||||
|
||||
@@ -519,67 +506,47 @@ func NewAPIServer(mode Mode, cfg *config.APIServer) (*APIServer, error) {
|
||||
if mode == ModeCoordinator {
|
||||
l2DB = l2db.NewL2DB(
|
||||
dbRead, dbWrite,
|
||||
0,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
0,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
apiConnCon,
|
||||
)
|
||||
}
|
||||
|
||||
if cfg.Debug.GinDebugMode {
|
||||
gin.SetMode(gin.DebugMode)
|
||||
} else {
|
||||
gin.SetMode(gin.ReleaseMode)
|
||||
}
|
||||
server := gin.Default()
|
||||
coord := false
|
||||
if mode == ModeCoordinator {
|
||||
coord = cfg.Coordinator.API.Coordinator
|
||||
}
|
||||
nodeAPI, err := NewNodeAPI(
|
||||
cfg.API.Address,
|
||||
coord, cfg.API.Explorer,
|
||||
server,
|
||||
historyDB,
|
||||
l2DB,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
return &APIServer{
|
||||
nodeAPI: nodeAPI,
|
||||
mode: mode,
|
||||
ctx: ctx,
|
||||
cancel: cancel,
|
||||
}, nil
|
||||
}
|
||||
|
||||
// Start the APIServer
|
||||
func (s *APIServer) Start() {
|
||||
log.Infow("Starting api server...", "mode", s.mode)
|
||||
log.Info("Starting NodeAPI...")
|
||||
s.wg.Add(1)
|
||||
go func() {
|
||||
defer func() {
|
||||
log.Info("NodeAPI routine stopped")
|
||||
s.wg.Done()
|
||||
}()
|
||||
if err := s.nodeAPI.Run(s.ctx); err != nil {
|
||||
if s.ctx.Err() != nil {
|
||||
return
|
||||
}
|
||||
log.Fatalw("NodeAPI.Run", "err", err)
|
||||
var nodeAPI *NodeAPI
|
||||
if cfg.API.Address != "" {
|
||||
if cfg.Debug.GinDebugMode {
|
||||
gin.SetMode(gin.DebugMode)
|
||||
} else {
|
||||
gin.SetMode(gin.ReleaseMode)
|
||||
}
|
||||
}()
|
||||
}
|
||||
|
||||
// Stop the APIServer
|
||||
func (s *APIServer) Stop() {
|
||||
log.Infow("Stopping NodeAPI...")
|
||||
s.cancel()
|
||||
s.wg.Wait()
|
||||
if cfg.API.UpdateMetricsInterval.Duration == 0 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid cfg.API.UpdateMetricsInterval: %v",
|
||||
cfg.API.UpdateMetricsInterval.Duration))
|
||||
}
|
||||
if cfg.API.UpdateRecommendedFeeInterval.Duration == 0 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid cfg.API.UpdateRecommendedFeeInterval: %v",
|
||||
cfg.API.UpdateRecommendedFeeInterval.Duration))
|
||||
}
|
||||
server := gin.Default()
|
||||
coord := false
|
||||
if mode == ModeCoordinator {
|
||||
coord = cfg.Coordinator.API.Coordinator
|
||||
}
|
||||
var err error
|
||||
nodeAPI, err = NewNodeAPI(
|
||||
cfg.API.Address,
|
||||
coord, cfg.API.Explorer,
|
||||
server,
|
||||
historyDB,
|
||||
l2DB,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
// ETC...
|
||||
}
|
||||
|
||||
// NodeAPI holds the node http API
|
||||
@@ -622,8 +589,8 @@ func NewNodeAPI(
|
||||
}, nil
|
||||
}
|
||||
|
||||
// Run starts the http server of the NodeAPI. To stop it, pass a context
|
||||
// with cancellation.
|
||||
// Run starts the http server of the NodeAPI. To stop it, pass a context with
|
||||
// cancelation.
|
||||
func (a *NodeAPI) Run(ctx context.Context) error {
|
||||
server := &http.Server{
|
||||
Handler: a.engine,
|
||||
@@ -655,18 +622,18 @@ func (a *NodeAPI) Run(ctx context.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars *common.SCVariablesPtr, batches []common.BatchData) error {
|
||||
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, vars *common.SCVariablesPtr,
|
||||
batches []common.BatchData) error {
|
||||
if n.mode == ModeCoordinator {
|
||||
n.coord.SendMsg(ctx, coordinator.MsgSyncBlock{
|
||||
Stats: *stats,
|
||||
Vars: *vars,
|
||||
Vars: vars,
|
||||
Batches: batches,
|
||||
})
|
||||
}
|
||||
n.stateAPIUpdater.SetSCVars(vars)
|
||||
n.apiStateUpdater.SetSCVars(vars)
|
||||
if stats.Synced() {
|
||||
if err := n.stateAPIUpdater.UpdateNetworkInfo(
|
||||
if err := n.apiStateUpdater.UpdateNetworkInfo(
|
||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||
common.BatchNum(stats.Eth.LastBatchNum),
|
||||
stats.Sync.Auction.CurrentSlot.SlotNum,
|
||||
@@ -674,11 +641,11 @@ func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats,
|
||||
log.Errorw("ApiStateUpdater.UpdateNetworkInfo", "err", err)
|
||||
}
|
||||
} else {
|
||||
n.stateAPIUpdater.UpdateNetworkInfoBlock(
|
||||
n.apiStateUpdater.UpdateNetworkInfoBlock(
|
||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||
)
|
||||
}
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
if err := n.apiStateUpdater.Store(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
@@ -689,14 +656,14 @@ func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats,
|
||||
if n.mode == ModeCoordinator {
|
||||
n.coord.SendMsg(ctx, coordinator.MsgSyncReorg{
|
||||
Stats: *stats,
|
||||
Vars: *vars.AsPtr(),
|
||||
Vars: vars,
|
||||
})
|
||||
}
|
||||
n.stateAPIUpdater.SetSCVars(vars.AsPtr())
|
||||
n.stateAPIUpdater.UpdateNetworkInfoBlock(
|
||||
n.apiStateUpdater.SetSCVars(vars.AsPtr())
|
||||
n.apiStateUpdater.UpdateNetworkInfoBlock(
|
||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||
)
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
if err := n.apiStateUpdater.Store(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
@@ -704,8 +671,7 @@ func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats,
|
||||
|
||||
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
||||
// don't have to pass it around.
|
||||
func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common.Block,
|
||||
time.Duration, error) {
|
||||
func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common.Block, time.Duration, error) {
|
||||
blockData, discarded, err := n.sync.Sync(ctx, lastBlock)
|
||||
stats := n.sync.Stats()
|
||||
if err != nil {
|
||||
@@ -747,9 +713,7 @@ func (n *Node) StartSynchronizer() {
|
||||
// the last synced one) is synchronized
|
||||
stats := n.sync.Stats()
|
||||
vars := n.sync.SCVars()
|
||||
if err := n.handleNewBlock(n.ctx, stats, vars.AsPtr(), []common.BatchData{}); err != nil {
|
||||
log.Fatalw("Node.handleNewBlock", "err", err)
|
||||
}
|
||||
n.handleNewBlock(n.ctx, stats, vars.AsPtr(), []common.BatchData{})
|
||||
|
||||
n.wg.Add(1)
|
||||
go func() {
|
||||
@@ -836,24 +800,24 @@ func (n *Node) StartNodeAPI() {
|
||||
n.wg.Add(1)
|
||||
go func() {
|
||||
// Do an initial update on startup
|
||||
if err := n.stateAPIUpdater.UpdateMetrics(); err != nil {
|
||||
if err := n.apiStateUpdater.UpdateMetrics(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.UpdateMetrics", "err", err)
|
||||
}
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
if err := n.apiStateUpdater.Store(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||
}
|
||||
for {
|
||||
select {
|
||||
case <-n.ctx.Done():
|
||||
log.Info("ApiStateUpdater.UpdateMetrics loop done")
|
||||
log.Info("API.UpdateMetrics loop done")
|
||||
n.wg.Done()
|
||||
return
|
||||
case <-time.After(n.cfg.API.UpdateMetricsInterval.Duration):
|
||||
if err := n.stateAPIUpdater.UpdateMetrics(); err != nil {
|
||||
if err := n.apiStateUpdater.UpdateMetrics(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.UpdateMetrics", "err", err)
|
||||
continue
|
||||
}
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
if err := n.apiStateUpdater.Store(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||
}
|
||||
}
|
||||
@@ -863,25 +827,18 @@ func (n *Node) StartNodeAPI() {
|
||||
n.wg.Add(1)
|
||||
go func() {
|
||||
// Do an initial update on startup
|
||||
if err := n.stateAPIUpdater.UpdateRecommendedFee(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.UpdateRecommendedFee", "err", err)
|
||||
}
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||
if err := n.historyDB.UpdateRecommendedFee(); err != nil {
|
||||
log.Errorw("API.UpdateRecommendedFee", "err", err)
|
||||
}
|
||||
for {
|
||||
select {
|
||||
case <-n.ctx.Done():
|
||||
log.Info("ApiStateUpdaterAPI.UpdateRecommendedFee loop done")
|
||||
log.Info("API.UpdateRecommendedFee loop done")
|
||||
n.wg.Done()
|
||||
return
|
||||
case <-time.After(n.cfg.API.UpdateRecommendedFeeInterval.Duration):
|
||||
if err := n.stateAPIUpdater.UpdateRecommendedFee(); err != nil {
|
||||
log.Errorw("ApiStateUpdaterAPI.UpdateRecommendedFee", "err", err)
|
||||
continue
|
||||
}
|
||||
if err := n.stateAPIUpdater.Store(); err != nil {
|
||||
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||
if err := n.historyDB.UpdateRecommendedFee(); err != nil {
|
||||
log.Errorw("API.UpdateRecommendedFee", "err", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -4,12 +4,9 @@ import (
|
||||
"context"
|
||||
"fmt"
|
||||
"net/http"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/dghubble/sling"
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
@@ -20,107 +17,53 @@ const (
|
||||
defaultIdleConnTimeout = 2 * time.Second
|
||||
)
|
||||
|
||||
// UpdateMethodType defines the token price update mechanism
|
||||
type UpdateMethodType string
|
||||
// APIType defines the token exchange API
|
||||
type APIType string
|
||||
|
||||
const (
|
||||
// UpdateMethodTypeBitFinexV2 is the http API used by bitfinex V2
|
||||
UpdateMethodTypeBitFinexV2 UpdateMethodType = "bitfinexV2"
|
||||
// UpdateMethodTypeCoingeckoV3 is the http API used by copingecko V3
|
||||
UpdateMethodTypeCoingeckoV3 UpdateMethodType = "coingeckoV3"
|
||||
// UpdateMethodTypeStatic is the value given by the configuration
|
||||
UpdateMethodTypeStatic UpdateMethodType = "static"
|
||||
// UpdateMethodTypeIgnore indicates to not update the value, to set value 0
|
||||
// it's better to use UpdateMethodTypeStatic
|
||||
UpdateMethodTypeIgnore UpdateMethodType = "ignore"
|
||||
// APITypeBitFinexV2 is the http API used by bitfinex V2
|
||||
APITypeBitFinexV2 APIType = "bitfinexV2"
|
||||
)
|
||||
|
||||
func (t *UpdateMethodType) valid() bool {
|
||||
func (t *APIType) valid() bool {
|
||||
switch *t {
|
||||
case UpdateMethodTypeBitFinexV2:
|
||||
return true
|
||||
case UpdateMethodTypeCoingeckoV3:
|
||||
return true
|
||||
case UpdateMethodTypeStatic:
|
||||
return true
|
||||
case UpdateMethodTypeIgnore:
|
||||
case APITypeBitFinexV2:
|
||||
return true
|
||||
default:
|
||||
return false
|
||||
}
|
||||
}
|
||||
|
||||
// TokenConfig specifies how a single token get its price updated
|
||||
type TokenConfig struct {
|
||||
UpdateMethod UpdateMethodType
|
||||
StaticValue float64 // required by UpdateMethodTypeStatic
|
||||
Symbol string
|
||||
Addr ethCommon.Address
|
||||
}
|
||||
|
||||
func (t *TokenConfig) valid() bool {
|
||||
if (t.Addr == common.EmptyAddr && t.Symbol != "ETH") ||
|
||||
(t.Symbol == "" && t.UpdateMethod == UpdateMethodTypeBitFinexV2) {
|
||||
return false
|
||||
}
|
||||
return t.UpdateMethod.valid()
|
||||
}
|
||||
|
||||
// PriceUpdater definition
|
||||
type PriceUpdater struct {
|
||||
db *historydb.HistoryDB
|
||||
defaultUpdateMethod UpdateMethodType
|
||||
tokensList []historydb.TokenSymbolAndAddr
|
||||
tokensConfig map[ethCommon.Address]TokenConfig
|
||||
clientCoingeckoV3 *sling.Sling
|
||||
clientBitfinexV2 *sling.Sling
|
||||
db *historydb.HistoryDB
|
||||
apiURL string
|
||||
apiType APIType
|
||||
tokenSymbols []string
|
||||
}
|
||||
|
||||
// NewPriceUpdater is the constructor for the updater
|
||||
func NewPriceUpdater(
|
||||
defaultUpdateMethodType UpdateMethodType,
|
||||
tokensConfig []TokenConfig,
|
||||
db *historydb.HistoryDB,
|
||||
bitfinexV2URL, coingeckoV3URL string,
|
||||
) (*PriceUpdater, error) {
|
||||
// Validate params
|
||||
if !defaultUpdateMethodType.valid() || defaultUpdateMethodType == UpdateMethodTypeStatic {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("Invalid defaultUpdateMethodType: %v", defaultUpdateMethodType),
|
||||
)
|
||||
func NewPriceUpdater(apiURL string, apiType APIType, db *historydb.HistoryDB) (*PriceUpdater, error) {
|
||||
tokenSymbols := []string{}
|
||||
if !apiType.valid() {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Invalid apiType: %v", apiType))
|
||||
}
|
||||
tokensConfigMap := make(map[ethCommon.Address]TokenConfig)
|
||||
for _, t := range tokensConfig {
|
||||
if !t.valid() {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Invalid tokensConfig, wrong entry: %+v", t))
|
||||
}
|
||||
tokensConfigMap[t.Addr] = t
|
||||
}
|
||||
// Init
|
||||
tr := &http.Transport{
|
||||
MaxIdleConns: defaultMaxIdleConns,
|
||||
IdleConnTimeout: defaultIdleConnTimeout,
|
||||
DisableCompression: true,
|
||||
}
|
||||
httpClient := &http.Client{Transport: tr}
|
||||
return &PriceUpdater{
|
||||
db: db,
|
||||
defaultUpdateMethod: defaultUpdateMethodType,
|
||||
tokensList: []historydb.TokenSymbolAndAddr{},
|
||||
tokensConfig: tokensConfigMap,
|
||||
clientCoingeckoV3: sling.New().Base(coingeckoV3URL).Client(httpClient),
|
||||
clientBitfinexV2: sling.New().Base(bitfinexV2URL).Client(httpClient),
|
||||
db: db,
|
||||
apiURL: apiURL,
|
||||
apiType: apiType,
|
||||
tokenSymbols: tokenSymbols,
|
||||
}, nil
|
||||
}
|
||||
|
||||
func (p *PriceUpdater) getTokenPriceBitfinex(ctx context.Context, tokenSymbol string) (float64, error) {
|
||||
func getTokenPriceBitfinex(ctx context.Context, client *sling.Sling,
|
||||
tokenSymbol string) (float64, error) {
|
||||
state := [10]float64{}
|
||||
url := "ticker/t" + tokenSymbol + "USD"
|
||||
req, err := p.clientBitfinexV2.New().Get(url).Request()
|
||||
req, err := client.New().Get("ticker/t" + tokenSymbol + "USD").Request()
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
res, err := p.clientBitfinexV2.Do(req.WithContext(ctx), &state, nil)
|
||||
res, err := client.Do(req.WithContext(ctx), &state, nil)
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -130,83 +73,43 @@ func (p *PriceUpdater) getTokenPriceBitfinex(ctx context.Context, tokenSymbol st
|
||||
return state[6], nil
|
||||
}
|
||||
|
||||
func (p *PriceUpdater) getTokenPriceCoingecko(ctx context.Context, tokenAddr ethCommon.Address) (float64, error) {
|
||||
responseObject := make(map[string]map[string]float64)
|
||||
var url string
|
||||
var id string
|
||||
if tokenAddr == common.EmptyAddr { // Special case for Ether
|
||||
url = "simple/price?ids=ethereum&vs_currencies=usd"
|
||||
id = "ethereum"
|
||||
} else { // Common case (ERC20)
|
||||
id = strings.ToLower(tokenAddr.String())
|
||||
url = "simple/token_price/ethereum?contract_addresses=" +
|
||||
id + "&vs_currencies=usd"
|
||||
}
|
||||
req, err := p.clientCoingeckoV3.New().Get(url).Request()
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
res, err := p.clientCoingeckoV3.Do(req.WithContext(ctx), &responseObject, nil)
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
if res.StatusCode != http.StatusOK {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("http response is not is %v", res.StatusCode))
|
||||
}
|
||||
price := responseObject[id]["usd"]
|
||||
if price <= 0 {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("price not found for %v", id))
|
||||
}
|
||||
return price, nil
|
||||
}
|
||||
|
||||
// UpdatePrices is triggered by the Coordinator, and internally will update the
|
||||
// token prices in the db
|
||||
// UpdatePrices is triggered by the Coordinator, and internally will update the token prices in the db
|
||||
func (p *PriceUpdater) UpdatePrices(ctx context.Context) {
|
||||
for _, token := range p.tokensConfig {
|
||||
tr := &http.Transport{
|
||||
MaxIdleConns: defaultMaxIdleConns,
|
||||
IdleConnTimeout: defaultIdleConnTimeout,
|
||||
DisableCompression: true,
|
||||
}
|
||||
httpClient := &http.Client{Transport: tr}
|
||||
client := sling.New().Base(p.apiURL).Client(httpClient)
|
||||
|
||||
for _, tokenSymbol := range p.tokenSymbols {
|
||||
var tokenPrice float64
|
||||
var err error
|
||||
switch token.UpdateMethod {
|
||||
case UpdateMethodTypeBitFinexV2:
|
||||
tokenPrice, err = p.getTokenPriceBitfinex(ctx, token.Symbol)
|
||||
case UpdateMethodTypeCoingeckoV3:
|
||||
tokenPrice, err = p.getTokenPriceCoingecko(ctx, token.Addr)
|
||||
case UpdateMethodTypeStatic:
|
||||
tokenPrice = token.StaticValue
|
||||
case UpdateMethodTypeIgnore:
|
||||
continue
|
||||
switch p.apiType {
|
||||
case APITypeBitFinexV2:
|
||||
tokenPrice, err = getTokenPriceBitfinex(ctx, client, tokenSymbol)
|
||||
}
|
||||
if ctx.Err() != nil {
|
||||
return
|
||||
}
|
||||
if err != nil {
|
||||
log.Warnw("token price not updated (get error)",
|
||||
"err", err, "token", token.Symbol, "updateMethod", token.UpdateMethod)
|
||||
"err", err, "token", tokenSymbol, "apiType", p.apiType)
|
||||
}
|
||||
if err = p.db.UpdateTokenValue(token.Addr, tokenPrice); err != nil {
|
||||
if err = p.db.UpdateTokenValue(tokenSymbol, tokenPrice); err != nil {
|
||||
log.Errorw("token price not updated (db error)",
|
||||
"err", err, "token", token.Symbol, "updateMethod", token.UpdateMethod)
|
||||
"err", err, "token", tokenSymbol, "apiType", p.apiType)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// UpdateTokenList get the registered token symbols from HistoryDB
|
||||
func (p *PriceUpdater) UpdateTokenList() error {
|
||||
dbTokens, err := p.db.GetTokenSymbolsAndAddrs()
|
||||
tokenSymbols, err := p.db.GetTokenSymbols()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// For each token from the DB
|
||||
for _, dbToken := range dbTokens {
|
||||
// If the token doesn't exists in the config list,
|
||||
// add it with default update emthod
|
||||
if _, ok := p.tokensConfig[dbToken.Addr]; !ok {
|
||||
p.tokensConfig[dbToken.Addr] = TokenConfig{
|
||||
UpdateMethod: p.defaultUpdateMethod,
|
||||
Symbol: dbToken.Symbol,
|
||||
Addr: dbToken.Addr,
|
||||
}
|
||||
}
|
||||
}
|
||||
p.tokenSymbols = tokenSymbols
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -2,6 +2,7 @@ package priceupdater
|
||||
|
||||
import (
|
||||
"context"
|
||||
"math/big"
|
||||
"os"
|
||||
"testing"
|
||||
|
||||
@@ -14,130 +15,43 @@ import (
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
var historyDB *historydb.HistoryDB
|
||||
|
||||
const usdtAddr = "0xdac17f958d2ee523a2206206994597c13d831ec7"
|
||||
|
||||
func TestPriceUpdaterBitfinex(t *testing.T) {
|
||||
func TestPriceUpdater(t *testing.T) {
|
||||
// Init DB
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
historyDB = historydb.NewHistoryDB(db, db, nil)
|
||||
assert.NoError(t, err)
|
||||
historyDB := historydb.NewHistoryDB(db, db, nil)
|
||||
// Clean DB
|
||||
test.WipeDB(historyDB.DB())
|
||||
// Populate DB
|
||||
// Gen blocks and add them to DB
|
||||
blocks := test.GenBlocks(1, 2)
|
||||
require.NoError(t, historyDB.AddBlocks(blocks))
|
||||
assert.NoError(t, historyDB.AddBlocks(blocks))
|
||||
// Gen tokens and add them to DB
|
||||
tokens := []common.Token{
|
||||
{
|
||||
TokenID: 1,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.HexToAddress("0x1"),
|
||||
Name: "DAI",
|
||||
Symbol: "DAI",
|
||||
Decimals: 18,
|
||||
}, // Used to test get by SC addr
|
||||
{
|
||||
TokenID: 2,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.HexToAddress(usdtAddr),
|
||||
Name: "Tether",
|
||||
Symbol: "USDT",
|
||||
Decimals: 18,
|
||||
}, // Used to test get by token symbol
|
||||
{
|
||||
TokenID: 3,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.HexToAddress("0x2"),
|
||||
Name: "FOO",
|
||||
Symbol: "FOO",
|
||||
Decimals: 18,
|
||||
}, // Used to test ignore
|
||||
{
|
||||
TokenID: 4,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.HexToAddress("0x3"),
|
||||
Name: "BAR",
|
||||
Symbol: "BAR",
|
||||
Decimals: 18,
|
||||
}, // Used to test static
|
||||
{
|
||||
TokenID: 5,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.HexToAddress("0x1f9840a85d5af5bf1d1762f925bdaddc4201f984"),
|
||||
Name: "Uniswap",
|
||||
Symbol: "UNI",
|
||||
Decimals: 18,
|
||||
}, // Used to test default
|
||||
}
|
||||
require.NoError(t, historyDB.AddTokens(tokens)) // ETH token exist in DB by default
|
||||
// Update token price used to test ignore
|
||||
ignoreValue := 44.44
|
||||
require.NoError(t, historyDB.UpdateTokenValue(tokens[2].EthAddr, ignoreValue))
|
||||
|
||||
// Prepare token config
|
||||
staticValue := 0.12345
|
||||
tc := []TokenConfig{
|
||||
// ETH and UNI tokens use default method
|
||||
{ // DAI uses SC addr
|
||||
UpdateMethod: UpdateMethodTypeBitFinexV2,
|
||||
Addr: ethCommon.HexToAddress("0x1"),
|
||||
Symbol: "DAI",
|
||||
},
|
||||
{ // USDT uses symbol
|
||||
UpdateMethod: UpdateMethodTypeCoingeckoV3,
|
||||
Addr: ethCommon.HexToAddress(usdtAddr),
|
||||
},
|
||||
{ // FOO uses ignore
|
||||
UpdateMethod: UpdateMethodTypeIgnore,
|
||||
Addr: ethCommon.HexToAddress("0x2"),
|
||||
},
|
||||
{ // BAR uses static
|
||||
UpdateMethod: UpdateMethodTypeStatic,
|
||||
Addr: ethCommon.HexToAddress("0x3"),
|
||||
StaticValue: staticValue,
|
||||
},
|
||||
}
|
||||
|
||||
bitfinexV2URL := "https://api-pub.bitfinex.com/v2/"
|
||||
coingeckoV3URL := "https://api.coingecko.com/api/v3/"
|
||||
tokens := []common.Token{}
|
||||
tokens = append(tokens, common.Token{
|
||||
TokenID: 1,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.BigToAddress(big.NewInt(2)),
|
||||
Name: "DAI",
|
||||
Symbol: "DAI",
|
||||
Decimals: 18,
|
||||
})
|
||||
assert.NoError(t, historyDB.AddTokens(tokens))
|
||||
// Init price updater
|
||||
pu, err := NewPriceUpdater(
|
||||
UpdateMethodTypeCoingeckoV3,
|
||||
tc,
|
||||
historyDB,
|
||||
bitfinexV2URL,
|
||||
coingeckoV3URL,
|
||||
)
|
||||
pu, err := NewPriceUpdater("https://api-pub.bitfinex.com/v2/", APITypeBitFinexV2, historyDB)
|
||||
require.NoError(t, err)
|
||||
// Update token list
|
||||
require.NoError(t, pu.UpdateTokenList())
|
||||
assert.NoError(t, pu.UpdateTokenList())
|
||||
// Update prices
|
||||
pu.UpdatePrices(context.Background())
|
||||
|
||||
// Check results: get tokens from DB
|
||||
// Check that prices have been updated
|
||||
fetchedTokens, err := historyDB.GetTokensTest()
|
||||
require.NoError(t, err)
|
||||
// Check that tokens that are updated via API have value:
|
||||
// ETH
|
||||
require.NotNil(t, fetchedTokens[0].USDUpdate)
|
||||
assert.Greater(t, *fetchedTokens[0].USD, 0.0)
|
||||
// DAI
|
||||
require.NotNil(t, fetchedTokens[1].USDUpdate)
|
||||
assert.Greater(t, *fetchedTokens[1].USD, 0.0)
|
||||
// USDT
|
||||
require.NotNil(t, fetchedTokens[2].USDUpdate)
|
||||
assert.Greater(t, *fetchedTokens[2].USD, 0.0)
|
||||
// UNI
|
||||
require.NotNil(t, fetchedTokens[5].USDUpdate)
|
||||
assert.Greater(t, *fetchedTokens[5].USD, 0.0)
|
||||
// Check ignored token
|
||||
assert.Equal(t, ignoreValue, *fetchedTokens[3].USD)
|
||||
// Check static value
|
||||
assert.Equal(t, staticValue, *fetchedTokens[4].USD)
|
||||
// TokenID 0 (ETH) is always on the DB
|
||||
assert.Equal(t, 2, len(fetchedTokens))
|
||||
for _, token := range fetchedTokens {
|
||||
assert.NotNil(t, token.USD)
|
||||
assert.NotNil(t, token.USDUpdate)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -235,7 +235,7 @@ func (p *ProofServerClient) CalculateProof(ctx context.Context, zkInputs *common
|
||||
return tracerr.Wrap(p.apiInput(ctx, zkInputs))
|
||||
}
|
||||
|
||||
// GetProof retrieves the Proof and Public Data (public inputs) from the
|
||||
// GetProof retreives the Proof and Public Data (public inputs) from the
|
||||
// ServerProof, blocking until the proof is ready.
|
||||
func (p *ProofServerClient) GetProof(ctx context.Context) (*Proof, []*big.Int, error) {
|
||||
if err := p.WaitReady(ctx); err != nil {
|
||||
@@ -256,8 +256,7 @@ func (p *ProofServerClient) GetProof(ctx context.Context) (*Proof, []*big.Int, e
|
||||
}
|
||||
return &proof, pubInputs, nil
|
||||
}
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("status != %v, status = %v", StatusCodeSuccess,
|
||||
status.Status))
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("status != %v, status = %v", StatusCodeSuccess, status.Status))
|
||||
}
|
||||
|
||||
// Cancel cancels any current proof computation
|
||||
@@ -298,7 +297,7 @@ func (p *MockClient) CalculateProof(ctx context.Context, zkInputs *common.ZKInpu
|
||||
return nil
|
||||
}
|
||||
|
||||
// GetProof retrieves the Proof from the ServerProof
|
||||
// GetProof retreives the Proof from the ServerProof
|
||||
func (p *MockClient) GetProof(ctx context.Context) (*Proof, []*big.Int, error) {
|
||||
// Simulate a delay
|
||||
select {
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user