mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 11:26:44 +01:00
Compare commits
128 Commits
feature/sq
...
tmp/txsel-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
cde300077a | ||
|
|
b724348ddc | ||
|
|
0b6911052a | ||
|
|
a39d880fd4 | ||
|
|
efb23950db | ||
|
|
06b271fc47 | ||
|
|
ac66ede917 | ||
|
|
7305a8d7f1 | ||
|
|
81690039da | ||
|
|
661da0d53e | ||
|
|
d0ace0773d | ||
|
|
4fd757a03c | ||
|
|
1778b08043 | ||
|
|
dcd40f1943 | ||
|
|
7833cb11de | ||
|
|
9e96bc89f0 | ||
|
|
7bc83d7f70 | ||
|
|
6bf8584a82 | ||
|
|
18e9e437da | ||
|
|
e4343a73ad | ||
|
|
59558e5a4b | ||
|
|
dfd49164f4 | ||
|
|
1050da350a | ||
|
|
e460a7e58b | ||
|
|
97fdbfe037 | ||
|
|
2e2d5a3c29 | ||
|
|
f762fbdc4d | ||
|
|
4390fd8346 | ||
|
|
93b7866d3e | ||
|
|
059a3556fb | ||
|
|
8b88e1fbd0 | ||
|
|
1de0b08bac | ||
|
|
73c3a91746 | ||
|
|
74769da5ba | ||
|
|
65b007d59c | ||
|
|
6e4bd56c1a | ||
|
|
39b7882ef2 | ||
|
|
c7d0422c16 | ||
|
|
56ffea2190 | ||
|
|
4a75c35abf | ||
|
|
cf70111de5 | ||
|
|
f664a3a382 | ||
|
|
cd1df6ea8c | ||
|
|
6da827c751 | ||
|
|
60023e4574 | ||
|
|
9de3a4ec6a | ||
|
|
91c96eb429 | ||
|
|
70f874aaf1 | ||
|
|
54508b0ba6 | ||
|
|
0adcf1a2bc | ||
|
|
527cd9a2cc | ||
|
|
b269117a32 | ||
|
|
e14705c13b | ||
|
|
5ff0350f51 | ||
|
|
15fd3945dd | ||
|
|
83c256deda | ||
|
|
d50ae71710 | ||
|
|
ba108b1146 | ||
|
|
c771bdf94e | ||
|
|
3c68aa5014 | ||
|
|
4856251f01 | ||
|
|
4acfeb0ad9 | ||
|
|
400ab14f04 | ||
|
|
706e4c7a3d | ||
|
|
bf4acfad9f | ||
|
|
ba88b25425 | ||
|
|
df729b3b71 | ||
|
|
f8d01b5998 | ||
|
|
155e06484b | ||
|
|
5cdcae47b8 | ||
|
|
ed9bfdffa9 | ||
|
|
504e36ac47 | ||
|
|
ca53dc9b52 | ||
|
|
5b4cfac309 | ||
|
|
11a0707746 | ||
|
|
d16fba72a7 | ||
|
|
971b2c1d40 | ||
|
|
9d08ec6978 | ||
|
|
ffda9fa1ef | ||
|
|
5a11aa5c27 | ||
|
|
3e5e9bd633 | ||
|
|
c83047f527 | ||
|
|
bcd576480c | ||
|
|
35ea597ac4 | ||
|
|
8259aee884 | ||
|
|
72862147f3 | ||
|
|
3706ddb2fb | ||
|
|
df0cc32eed | ||
|
|
67b2b7da4b | ||
|
|
e23063380c | ||
|
|
ed4d39fcd1 | ||
|
|
d6ec1910da | ||
|
|
c829eb99dc | ||
|
|
6ecb8118bd | ||
|
|
4500820a03 | ||
|
|
b4e6104fd3 | ||
|
|
28f026f628 | ||
|
|
688d376ce0 | ||
|
|
2547d5dce7 | ||
|
|
bb8d81c3aa | ||
|
|
af6f114667 | ||
|
|
e2376980f8 | ||
|
|
264f01b572 | ||
|
|
a21793b2b0 | ||
|
|
7b01f6a288 | ||
|
|
f2e5800ebd | ||
|
|
f0e79f3d55 | ||
|
|
26fbeb5c68 | ||
|
|
05104b0565 | ||
|
|
53507edabb | ||
|
|
729966f854 | ||
|
|
1c10a01cf7 | ||
|
|
1d0abe438f | ||
|
|
6a01c0ac14 | ||
|
|
63151a285c | ||
|
|
52d4197330 | ||
|
|
ea63cba62a | ||
|
|
c7e6267189 | ||
|
|
2a77dac9c1 | ||
|
|
ac1fd9acf7 | ||
|
|
1bf29636db | ||
|
|
2bf3b843ed | ||
|
|
277a1bc321 | ||
|
|
3181c8738c | ||
|
|
62df063ccf | ||
|
|
48a538faa3 | ||
|
|
10a34c8801 | ||
|
|
6260dfedad |
@@ -4,6 +4,10 @@ Go implementation of the Hermez node.
|
||||
|
||||
## Developing
|
||||
|
||||
### Go version
|
||||
|
||||
The `hermez-node` has been tested with go version 1.14
|
||||
|
||||
### Unit testing
|
||||
|
||||
Running the unit tests requires a connection to a PostgreSQL database. You can
|
||||
@@ -11,7 +15,7 @@ start PostgreSQL with docker easily this way (where `yourpasswordhere` should
|
||||
be your password):
|
||||
|
||||
```
|
||||
POSTGRES_PASS=yourpasswordhere sudo docker run --rm --name hermez-db-test -p 5432:5432 -e POSTGRES_DB=hermez -e POSTGRES_USER=hermez -e POSTGRES_PASSWORD="$POSTGRES_PASS" -d postgres
|
||||
POSTGRES_PASS=yourpasswordhere; sudo docker run --rm --name hermez-db-test -p 5432:5432 -e POSTGRES_DB=hermez -e POSTGRES_USER=hermez -e POSTGRES_PASSWORD="$POSTGRES_PASS" -d postgres
|
||||
```
|
||||
|
||||
Afterwards, run the tests with the password as env var:
|
||||
|
||||
@@ -4,10 +4,7 @@ import (
|
||||
"net/http"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
func (a *API) getAccount(c *gin.Context) {
|
||||
@@ -23,16 +20,6 @@ func (a *API) getAccount(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Get balance from stateDB
|
||||
account, err := a.s.LastGetAccount(*idx)
|
||||
if err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
|
||||
apiAccount.Balance = apitypes.NewBigIntStr(account.Balance)
|
||||
apiAccount.Nonce = account.Nonce
|
||||
|
||||
c.JSON(http.StatusOK, apiAccount)
|
||||
}
|
||||
|
||||
@@ -57,27 +44,7 @@ func (a *API) getAccounts(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Get balances from stateDB
|
||||
if err := a.s.LastRead(func(sdb *statedb.Last) error {
|
||||
for x, apiAccount := range apiAccounts {
|
||||
idx, err := stringToIdx(string(apiAccount.Idx), "Account Idx")
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
account, err := sdb.GetAccount(*idx)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
apiAccounts[x].Balance = apitypes.NewBigIntStr(account.Balance)
|
||||
apiAccounts[x].Nonce = account.Nonce
|
||||
}
|
||||
return nil
|
||||
}); err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type accountResponse struct {
|
||||
Accounts []historydb.AccountAPI `json:"accounts"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -26,7 +26,7 @@ func (a *API) postAccountCreationAuth(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
// Insert to DB
|
||||
if err := a.l2.AddAccountCreationAuth(commonAuth); err != nil {
|
||||
if err := a.l2.AddAccountCreationAuthAPI(commonAuth); err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
@@ -47,7 +47,7 @@ func (a *API) getAccountCreationAuth(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
c.JSON(http.StatusOK, auth)
|
||||
}
|
||||
|
||||
|
||||
12
api/api.go
12
api/api.go
@@ -9,7 +9,6 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
@@ -22,6 +21,7 @@ const (
|
||||
// Status define status of the network
|
||||
type Status struct {
|
||||
sync.RWMutex
|
||||
NodeConfig NodeConfig `json:"nodeConfig"`
|
||||
Network Network `json:"network"`
|
||||
Metrics historydb.Metrics `json:"metrics"`
|
||||
Rollup historydb.RollupVariablesAPI `json:"rollup"`
|
||||
@@ -34,7 +34,6 @@ type Status struct {
|
||||
type API struct {
|
||||
h *historydb.HistoryDB
|
||||
cg *configAPI
|
||||
s *statedb.StateDB
|
||||
l2 *l2db.L2DB
|
||||
status Status
|
||||
chainID uint16
|
||||
@@ -46,9 +45,9 @@ func NewAPI(
|
||||
coordinatorEndpoints, explorerEndpoints bool,
|
||||
server *gin.Engine,
|
||||
hdb *historydb.HistoryDB,
|
||||
sdb *statedb.StateDB,
|
||||
l2db *l2db.L2DB,
|
||||
config *Config,
|
||||
nodeConfig *NodeConfig,
|
||||
) (*API, error) {
|
||||
// Check input
|
||||
// TODO: is stateDB only needed for explorer endpoints or for both?
|
||||
@@ -66,9 +65,10 @@ func NewAPI(
|
||||
AuctionConstants: config.AuctionConstants,
|
||||
WDelayerConstants: config.WDelayerConstants,
|
||||
},
|
||||
s: sdb,
|
||||
l2: l2db,
|
||||
status: Status{},
|
||||
l2: l2db,
|
||||
status: Status{
|
||||
NodeConfig: *nodeConfig,
|
||||
},
|
||||
chainID: config.ChainID,
|
||||
hermezAddress: config.HermezAddress,
|
||||
}
|
||||
|
||||
158
api/api_test.go
158
api/api_test.go
@@ -8,9 +8,11 @@ import (
|
||||
"io"
|
||||
"io/ioutil"
|
||||
"math/big"
|
||||
"net"
|
||||
"net/http"
|
||||
"os"
|
||||
"strconv"
|
||||
"sync"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
@@ -21,12 +23,12 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/test"
|
||||
"github.com/hermeznetwork/hermez-node/test/til"
|
||||
"github.com/hermeznetwork/hermez-node/test/txsets"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
// Pendinger is an interface that allows getting last returned item ID and PendingItems to be used for building fromItem
|
||||
@@ -37,8 +39,8 @@ type Pendinger interface {
|
||||
New() Pendinger
|
||||
}
|
||||
|
||||
const apiPort = ":4010"
|
||||
const apiURL = "http://localhost" + apiPort + "/"
|
||||
const apiAddr = ":4010"
|
||||
const apiURL = "http://localhost" + apiAddr + "/"
|
||||
|
||||
var SetBlockchain = `
|
||||
Type: Blockchain
|
||||
@@ -199,7 +201,8 @@ func TestMain(m *testing.M) {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
hdb := historydb.NewHistoryDB(database)
|
||||
apiConnCon := db.NewAPIConnectionController(1, time.Second)
|
||||
hdb := historydb.NewHistoryDB(database, database, apiConnCon)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
@@ -213,12 +216,8 @@ func TestMain(m *testing.M) {
|
||||
panic(err)
|
||||
}
|
||||
}()
|
||||
sdb, err := statedb.NewStateDB(dir, 128, statedb.TypeTxSelector, 0)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
// L2DB
|
||||
l2DB := l2db.NewL2DB(database, 10, 1000, 24*time.Hour)
|
||||
l2DB := l2db.NewL2DB(database, database, 10, 1000, 0.0, 24*time.Hour, apiConnCon)
|
||||
test.WipeDB(l2DB.DB()) // this will clean HistoryDB and L2DB
|
||||
// Config (smart contract constants)
|
||||
chainID := uint16(0)
|
||||
@@ -236,17 +235,24 @@ func TestMain(m *testing.M) {
|
||||
true,
|
||||
apiGin,
|
||||
hdb,
|
||||
sdb,
|
||||
l2DB,
|
||||
&_config,
|
||||
&NodeConfig{
|
||||
ForgeDelay: 180,
|
||||
},
|
||||
)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
// Start server
|
||||
server := &http.Server{Addr: apiPort, Handler: apiGin}
|
||||
listener, err := net.Listen("tcp", apiAddr) //nolint:gosec
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
server := &http.Server{Handler: apiGin}
|
||||
go func() {
|
||||
if err := server.ListenAndServe(); err != nil && tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
if err := server.Serve(listener); err != nil &&
|
||||
tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
panic(err)
|
||||
}
|
||||
}()
|
||||
@@ -254,7 +260,7 @@ func TestMain(m *testing.M) {
|
||||
// Reset DB
|
||||
test.WipeDB(api.h.DB())
|
||||
|
||||
// Genratre blockchain data with til
|
||||
// Generate blockchain data with til
|
||||
tcc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||
tilCfgExtra := til.ConfigExtra{
|
||||
BootCoordAddr: ethCommon.HexToAddress("0xE39fEc6224708f0772D2A74fd3f9055A90E0A9f2"),
|
||||
@@ -347,19 +353,6 @@ func TestMain(m *testing.M) {
|
||||
}
|
||||
}
|
||||
|
||||
// lastBlockNum2 := blocksData[len(blocksData)-1].Block.EthBlockNum
|
||||
|
||||
// Add accounts to StateDB
|
||||
for i := 0; i < len(commonAccounts); i++ {
|
||||
if _, err := api.s.CreateAccount(commonAccounts[i].Idx, &commonAccounts[i]); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
}
|
||||
// Make a checkpoint to make the accounts available in Last
|
||||
if err := api.s.MakeCheckpoint(); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
|
||||
// Generate Coordinators and add them to HistoryDB
|
||||
const nCoords = 10
|
||||
commonCoords := test.GenCoordinators(nCoords, commonBlocks)
|
||||
@@ -519,20 +512,48 @@ func TestMain(m *testing.M) {
|
||||
WithdrawalDelay: uint64(3000),
|
||||
}
|
||||
|
||||
// Generate test data, as expected to be received/sended from/to the API
|
||||
// Generate test data, as expected to be received/sent from/to the API
|
||||
testCoords := genTestCoordinators(commonCoords)
|
||||
testBids := genTestBids(commonBlocks, testCoords, bids)
|
||||
testExits := genTestExits(commonExitTree, testTokens, commonAccounts)
|
||||
testTxs := genTestTxs(commonL1Txs, commonL2Txs, commonAccounts, testTokens, commonBlocks)
|
||||
testBatches, testFullBatches := genTestBatches(commonBlocks, commonBatches, testTxs)
|
||||
poolTxsToSend, poolTxsToReceive := genTestPoolTxs(commonPoolTxs, testTokens, commonAccounts)
|
||||
// Add balance and nonce to historyDB
|
||||
accounts := genTestAccounts(commonAccounts, testTokens)
|
||||
accUpdates := []common.AccountUpdate{}
|
||||
for i := 0; i < len(accounts); i++ {
|
||||
balance := new(big.Int)
|
||||
balance.SetString(string(*accounts[i].Balance), 10)
|
||||
idx, err := stringToIdx(string(accounts[i].Idx), "foo")
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
accUpdates = append(accUpdates, common.AccountUpdate{
|
||||
EthBlockNum: 0,
|
||||
BatchNum: 1,
|
||||
Idx: *idx,
|
||||
Nonce: 0,
|
||||
Balance: balance,
|
||||
})
|
||||
accUpdates = append(accUpdates, common.AccountUpdate{
|
||||
EthBlockNum: 0,
|
||||
BatchNum: 1,
|
||||
Idx: *idx,
|
||||
Nonce: accounts[i].Nonce,
|
||||
Balance: balance,
|
||||
})
|
||||
}
|
||||
if err := api.h.AddAccountUpdates(accUpdates); err != nil {
|
||||
panic(err)
|
||||
}
|
||||
tc = testCommon{
|
||||
blocks: commonBlocks,
|
||||
tokens: testTokens,
|
||||
batches: testBatches,
|
||||
fullBatches: testFullBatches,
|
||||
coordinators: testCoords,
|
||||
accounts: genTestAccounts(commonAccounts, testTokens),
|
||||
accounts: accounts,
|
||||
txs: testTxs,
|
||||
exits: testExits,
|
||||
poolTxsToSend: poolTxsToSend,
|
||||
@@ -574,6 +595,87 @@ func TestMain(m *testing.M) {
|
||||
os.Exit(result)
|
||||
}
|
||||
|
||||
func TestTimeout(t *testing.T) {
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
databaseTO, err := db.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
apiConnConTO := db.NewAPIConnectionController(1, 100*time.Millisecond)
|
||||
hdbTO := historydb.NewHistoryDB(databaseTO, databaseTO, apiConnConTO)
|
||||
require.NoError(t, err)
|
||||
// L2DB
|
||||
l2DBTO := l2db.NewL2DB(databaseTO, databaseTO, 10, 1000, 1.0, 24*time.Hour, apiConnConTO)
|
||||
|
||||
// API
|
||||
apiGinTO := gin.Default()
|
||||
finishWait := make(chan interface{})
|
||||
startWait := make(chan interface{})
|
||||
apiGinTO.GET("/wait", func(c *gin.Context) {
|
||||
cancel, err := apiConnConTO.Acquire()
|
||||
defer cancel()
|
||||
require.NoError(t, err)
|
||||
defer apiConnConTO.Release()
|
||||
startWait <- nil
|
||||
<-finishWait
|
||||
})
|
||||
// Start server
|
||||
serverTO := &http.Server{Handler: apiGinTO}
|
||||
listener, err := net.Listen("tcp", ":4444") //nolint:gosec
|
||||
require.NoError(t, err)
|
||||
go func() {
|
||||
if err := serverTO.Serve(listener); err != nil &&
|
||||
tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
require.NoError(t, err)
|
||||
}
|
||||
}()
|
||||
_config := getConfigTest(0)
|
||||
_, err = NewAPI(
|
||||
true,
|
||||
true,
|
||||
apiGinTO,
|
||||
hdbTO,
|
||||
l2DBTO,
|
||||
&_config,
|
||||
&NodeConfig{
|
||||
ForgeDelay: 180,
|
||||
},
|
||||
)
|
||||
require.NoError(t, err)
|
||||
|
||||
client := &http.Client{}
|
||||
httpReq, err := http.NewRequest("GET", "http://localhost:4444/tokens", nil)
|
||||
require.NoError(t, err)
|
||||
httpReqWait, err := http.NewRequest("GET", "http://localhost:4444/wait", nil)
|
||||
require.NoError(t, err)
|
||||
// Request that will get timed out
|
||||
var wg sync.WaitGroup
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
// Request that will make the API busy
|
||||
_, err = client.Do(httpReqWait)
|
||||
require.NoError(t, err)
|
||||
wg.Done()
|
||||
}()
|
||||
<-startWait
|
||||
resp, err := client.Do(httpReq)
|
||||
require.NoError(t, err)
|
||||
require.Equal(t, http.StatusServiceUnavailable, resp.StatusCode)
|
||||
defer resp.Body.Close() //nolint
|
||||
body, err := ioutil.ReadAll(resp.Body)
|
||||
require.NoError(t, err)
|
||||
// Unmarshal body into return struct
|
||||
msg := &errorMsg{}
|
||||
err = json.Unmarshal(body, msg)
|
||||
require.NoError(t, err)
|
||||
// Check that the error was the expected down
|
||||
require.Equal(t, errSQLTimeout, msg.Message)
|
||||
finishWait <- nil
|
||||
|
||||
// Stop server
|
||||
wg.Wait()
|
||||
require.NoError(t, serverTO.Shutdown(context.Background()))
|
||||
require.NoError(t, databaseTO.Close())
|
||||
}
|
||||
|
||||
func doGoodReqPaginated(
|
||||
path, order string,
|
||||
iterStruct Pendinger,
|
||||
|
||||
@@ -52,7 +52,7 @@ func (a *API) getBatches(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type batchesResponse struct {
|
||||
Batches []historydb.BatchAPI `json:"batches"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
@@ -108,7 +108,7 @@ func (a *API) getFullBatch(c *gin.Context) {
|
||||
}
|
||||
// Fetch txs forged in the batch from historyDB
|
||||
maxTxsPerBatch := uint(2048) //nolint:gomnd
|
||||
txs, _, err := a.h.GetHistoryTxs(
|
||||
txs, _, err := a.h.GetTxsAPI(
|
||||
nil, nil, nil, nil, batchNum, nil, nil, &maxTxsPerBatch, historydb.OrderAsc,
|
||||
)
|
||||
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
||||
|
||||
@@ -34,7 +34,7 @@ func (a *API) getBids(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type bidsResponse struct {
|
||||
Bids []historydb.BidAPI `json:"bids"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -32,7 +32,7 @@ func (a *API) getCoordinators(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type coordinatorsResponse struct {
|
||||
Coordinators []historydb.CoordinatorAPI `json:"coordinators"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -43,7 +43,7 @@ func (a *API) getExits(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type exitsResponse struct {
|
||||
Exits []historydb.ExitAPI `json:"exits"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
@@ -72,6 +72,6 @@ func (a *API) getExit(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
c.JSON(http.StatusOK, exit)
|
||||
}
|
||||
|
||||
@@ -10,10 +10,11 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/lib/pq"
|
||||
"github.com/russross/meddler"
|
||||
)
|
||||
|
||||
const (
|
||||
// maxLimit is the max permited items to be returned in paginated responses
|
||||
// maxLimit is the max permitted items to be returned in paginated responses
|
||||
maxLimit uint = 2049
|
||||
|
||||
// dfltOrder indicates how paginated endpoints are ordered if not specified
|
||||
@@ -30,36 +31,55 @@ const (
|
||||
|
||||
// Error for duplicated key
|
||||
errDuplicatedKey = "Item already exists"
|
||||
|
||||
// Error for timeout due to SQL connection
|
||||
errSQLTimeout = "The node is under heavy preasure, please try again later"
|
||||
|
||||
// Error message returned when context reaches timeout
|
||||
errCtxTimeout = "context deadline exceeded"
|
||||
)
|
||||
|
||||
var (
|
||||
// ErrNillBidderAddr is used when a nil bidderAddr is received in the getCoordinator method
|
||||
ErrNillBidderAddr = errors.New("biderAddr can not be nil")
|
||||
// ErrNilBidderAddr is used when a nil bidderAddr is received in the getCoordinator method
|
||||
ErrNilBidderAddr = errors.New("biderAddr can not be nil")
|
||||
)
|
||||
|
||||
func retSQLErr(err error, c *gin.Context) {
|
||||
log.Warn("HTTP API SQL request error", "err", err)
|
||||
if sqlErr, ok := tracerr.Unwrap(err).(*pq.Error); ok {
|
||||
log.Warnw("HTTP API SQL request error", "err", err)
|
||||
errMsg := tracerr.Unwrap(err).Error()
|
||||
retDupKey := func(errCode pq.ErrorCode) {
|
||||
// https://www.postgresql.org/docs/current/errcodes-appendix.html
|
||||
if sqlErr.Code == "23505" {
|
||||
if errCode == "23505" {
|
||||
c.JSON(http.StatusInternalServerError, errorMsg{
|
||||
Message: errDuplicatedKey,
|
||||
})
|
||||
} else {
|
||||
c.JSON(http.StatusInternalServerError, errorMsg{
|
||||
Message: errMsg,
|
||||
})
|
||||
}
|
||||
}
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
if errMsg == errCtxTimeout {
|
||||
c.JSON(http.StatusServiceUnavailable, errorMsg{
|
||||
Message: errSQLTimeout,
|
||||
})
|
||||
} else if sqlErr, ok := tracerr.Unwrap(err).(*pq.Error); ok {
|
||||
retDupKey(sqlErr.Code)
|
||||
} else if sqlErr, ok := meddler.DriverErr(tracerr.Unwrap(err)); ok {
|
||||
retDupKey(sqlErr.(*pq.Error).Code)
|
||||
} else if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
c.JSON(http.StatusNotFound, errorMsg{
|
||||
Message: err.Error(),
|
||||
Message: errMsg,
|
||||
})
|
||||
} else {
|
||||
c.JSON(http.StatusInternalServerError, errorMsg{
|
||||
Message: err.Error(),
|
||||
Message: errMsg,
|
||||
})
|
||||
}
|
||||
}
|
||||
|
||||
func retBadReq(err error, c *gin.Context) {
|
||||
log.Warn("HTTP API Bad request error", "err", err)
|
||||
log.Warnw("HTTP API Bad request error", "err", err)
|
||||
c.JSON(http.StatusBadRequest, errorMsg{
|
||||
Message: err.Error(),
|
||||
})
|
||||
|
||||
@@ -50,19 +50,19 @@ func parsePagination(c querier) (fromItem *uint, order string, limit *uint, err
|
||||
return fromItem, order, limit, nil
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwriten
|
||||
// nolint reason: res may be not overwritten
|
||||
func parseQueryUint(name string, dflt *uint, min, max uint, c querier) (*uint, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
return stringToUint(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwriten
|
||||
// nolint reason: res may be not overwritten
|
||||
func parseQueryInt64(name string, dflt *int64, min, max int64, c querier) (*int64, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
return stringToInt64(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwriten
|
||||
// nolint reason: res may be not overwritten
|
||||
func parseQueryBool(name string, dflt *bool, c querier) (*bool, error) { //nolint:SA4009
|
||||
str := c.Query(name)
|
||||
if str == "" {
|
||||
@@ -295,13 +295,13 @@ func parseParamIdx(c paramer) (*common.Idx, error) {
|
||||
return stringToIdx(idxStr, name)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwriten
|
||||
// nolint reason: res may be not overwritten
|
||||
func parseParamUint(name string, dflt *uint, min, max uint, c paramer) (*uint, error) { //nolint:SA4009
|
||||
str := c.Param(name)
|
||||
return stringToUint(str, name, dflt, min, max)
|
||||
}
|
||||
|
||||
// nolint reason: res may be not overwriten
|
||||
// nolint reason: res may be not overwritten
|
||||
func parseParamInt64(name string, dflt *int64, min, max int64, c paramer) (*int64, error) { //nolint:SA4009
|
||||
str := c.Param(name)
|
||||
return stringToInt64(str, name, dflt, min, max)
|
||||
|
||||
16
api/slots.go
16
api/slots.go
@@ -11,7 +11,7 @@ import (
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
// SlotAPI is a repesentation of a slot information
|
||||
// SlotAPI is a representation of a slot information
|
||||
type SlotAPI struct {
|
||||
ItemID uint64 `json:"itemId"`
|
||||
SlotNum int64 `json:"slotNum"`
|
||||
@@ -97,12 +97,12 @@ func (a *API) getSlot(c *gin.Context) {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
}
|
||||
currentBlock, err := a.h.GetLastBlock()
|
||||
currentBlock, err := a.h.GetLastBlockAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
}
|
||||
auctionVars, err := a.h.GetAuctionVars()
|
||||
auctionVars, err := a.h.GetAuctionVarsAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
@@ -200,12 +200,12 @@ func (a *API) getSlots(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
currentBlock, err := a.h.GetLastBlock()
|
||||
currentBlock, err := a.h.GetLastBlockAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
}
|
||||
auctionVars, err := a.h.GetAuctionVars()
|
||||
auctionVars, err := a.h.GetAuctionVarsAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
@@ -220,13 +220,13 @@ func (a *API) getSlots(c *gin.Context) {
|
||||
retBadReq(errors.New("It is necessary to add maxSlotNum filter"), c)
|
||||
return
|
||||
} else if *finishedAuction {
|
||||
currentBlock, err := a.h.GetLastBlock()
|
||||
currentBlock, err := a.h.GetLastBlockAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
}
|
||||
currentSlot := a.getCurrentSlot(currentBlock.Num)
|
||||
auctionVars, err := a.h.GetAuctionVars()
|
||||
auctionVars, err := a.h.GetAuctionVarsAPI()
|
||||
if err != nil {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
@@ -316,7 +316,7 @@ func (a *API) getSlots(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type slotsResponse struct {
|
||||
Slots []SlotAPI `json:"slots"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
28
api/state.go
28
api/state.go
@@ -3,6 +3,7 @@ package api
|
||||
import (
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"math"
|
||||
"math/big"
|
||||
"net/http"
|
||||
"time"
|
||||
@@ -23,6 +24,12 @@ type Network struct {
|
||||
NextForgers []NextForger `json:"nextForgers"`
|
||||
}
|
||||
|
||||
// NodeConfig is the configuration of the node that is exposed via API
|
||||
type NodeConfig struct {
|
||||
// ForgeDelay in seconds
|
||||
ForgeDelay float64 `json:"forgeDelay"`
|
||||
}
|
||||
|
||||
// NextForger is a representation of the information of a coordinator and the period will forge
|
||||
type NextForger struct {
|
||||
Coordinator historydb.CoordinatorAPI `json:"coordinator"`
|
||||
@@ -141,7 +148,7 @@ func (a *API) UpdateNetworkInfo(
|
||||
a.status.Network.NextForgers = nextForgers
|
||||
|
||||
// Update buckets withdrawals
|
||||
bucketsUpdate, err := a.h.GetBucketUpdates()
|
||||
bucketsUpdate, err := a.h.GetBucketUpdatesAPI()
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
bucketsUpdate = nil
|
||||
} else if err != nil {
|
||||
@@ -201,7 +208,7 @@ func (a *API) getNextForgers(lastBlock common.Block, currentSlot, lastClosedSlot
|
||||
}}
|
||||
} else {
|
||||
// Get all the relevant updates from the DB
|
||||
minBidInfo, err = a.h.GetAuctionVarsUntilSetSlotNum(lastClosedSlot, int(lastClosedSlot-currentSlot)+1)
|
||||
minBidInfo, err = a.h.GetAuctionVarsUntilSetSlotNumAPI(lastClosedSlot, int(lastClosedSlot-currentSlot)+1)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -279,7 +286,7 @@ func (a *API) UpdateMetrics() error {
|
||||
}
|
||||
batchNum := a.status.Network.LastBatch.BatchNum
|
||||
a.status.RUnlock()
|
||||
metrics, err := a.h.GetMetrics(batchNum)
|
||||
metrics, err := a.h.GetMetricsAPI(batchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -293,14 +300,21 @@ func (a *API) UpdateMetrics() error {
|
||||
|
||||
// UpdateRecommendedFee update Status.RecommendedFee information
|
||||
func (a *API) UpdateRecommendedFee() error {
|
||||
feeExistingAccount, err := a.h.GetAvgTxFee()
|
||||
feeExistingAccount, err := a.h.GetAvgTxFeeAPI()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
var minFeeUSD float64
|
||||
if a.l2 != nil {
|
||||
minFeeUSD = a.l2.MinFeeUSD()
|
||||
}
|
||||
a.status.Lock()
|
||||
a.status.RecommendedFee.ExistingAccount = feeExistingAccount
|
||||
a.status.RecommendedFee.CreatesAccount = createAccountExtraFeePercentage * feeExistingAccount
|
||||
a.status.RecommendedFee.CreatesAccountAndRegister = createAccountInternalExtraFeePercentage * feeExistingAccount
|
||||
a.status.RecommendedFee.ExistingAccount =
|
||||
math.Max(feeExistingAccount, minFeeUSD)
|
||||
a.status.RecommendedFee.CreatesAccount =
|
||||
math.Max(createAccountExtraFeePercentage*feeExistingAccount, minFeeUSD)
|
||||
a.status.RecommendedFee.CreatesAccountAndRegister =
|
||||
math.Max(createAccountInternalExtraFeePercentage*feeExistingAccount, minFeeUSD)
|
||||
a.status.Unlock()
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -131,7 +131,7 @@ func TestUpdateNetworkInfo(t *testing.T) {
|
||||
func TestUpdateMetrics(t *testing.T) {
|
||||
// Update Metrics needs api.status.Network.LastBatch.BatchNum to be updated
|
||||
lastBlock := tc.blocks[3]
|
||||
lastBatchNum := common.BatchNum(3)
|
||||
lastBatchNum := common.BatchNum(12)
|
||||
currentSlotNum := int64(1)
|
||||
err := api.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||
assert.NoError(t, err)
|
||||
@@ -149,7 +149,11 @@ func TestUpdateMetrics(t *testing.T) {
|
||||
func TestUpdateRecommendedFee(t *testing.T) {
|
||||
err := api.UpdateRecommendedFee()
|
||||
assert.NoError(t, err)
|
||||
assert.Greater(t, api.status.RecommendedFee.ExistingAccount, float64(0))
|
||||
var minFeeUSD float64
|
||||
if api.l2 != nil {
|
||||
minFeeUSD = api.l2.MinFeeUSD()
|
||||
}
|
||||
assert.Greater(t, api.status.RecommendedFee.ExistingAccount, minFeeUSD)
|
||||
assert.Equal(t, api.status.RecommendedFee.CreatesAccount,
|
||||
api.status.RecommendedFee.ExistingAccount*createAccountExtraFeePercentage)
|
||||
assert.Equal(t, api.status.RecommendedFee.CreatesAccountAndRegister,
|
||||
@@ -158,7 +162,7 @@ func TestUpdateRecommendedFee(t *testing.T) {
|
||||
|
||||
func TestGetState(t *testing.T) {
|
||||
lastBlock := tc.blocks[3]
|
||||
lastBatchNum := common.BatchNum(3)
|
||||
lastBatchNum := common.BatchNum(12)
|
||||
currentSlotNum := int64(1)
|
||||
api.SetRollupVariables(tc.rollupVars)
|
||||
api.SetWDelayerVariables(tc.wdelayerVars)
|
||||
|
||||
@@ -1329,13 +1329,6 @@ components:
|
||||
type: string
|
||||
description: Moment in which the transaction was added to the pool.
|
||||
format: date-time
|
||||
batchNum:
|
||||
type: integer
|
||||
description: Identifier of a batch. Every new forged batch increases by one the batchNum, starting at 0.
|
||||
minimum: 0
|
||||
maximum: 4294967295
|
||||
nullable: true
|
||||
example: null
|
||||
requestFromAccountIndex:
|
||||
type: string
|
||||
description: >-
|
||||
@@ -1390,7 +1383,6 @@ components:
|
||||
$ref: '#/components/schemas/Token'
|
||||
example:
|
||||
amount: '100000000000000'
|
||||
batchNum:
|
||||
fee: 0
|
||||
fromAccountIndex: hez:SCC:256
|
||||
fromBJJ: hez:r_trOasVEk0zNaalOoS9aLedu6mO7jI5XTIPu_zGXoyn
|
||||
@@ -1438,7 +1430,6 @@ components:
|
||||
- info
|
||||
- signature
|
||||
- timestamp
|
||||
- batchNum
|
||||
- requestFromAccountIndex
|
||||
- requestToAccountIndex
|
||||
- requestToHezEthereumAddress
|
||||
@@ -2578,6 +2569,21 @@ components:
|
||||
description: List of next coordinators to forge.
|
||||
items:
|
||||
$ref: '#/components/schemas/NextForger'
|
||||
NodeConfig:
|
||||
type: object
|
||||
description: Configuration of the coordinator node. Note that this is specific for each coordinator.
|
||||
properties:
|
||||
forgeDelay:
|
||||
type: number
|
||||
description: |
|
||||
Delay in seconds after which a batch is forged if the slot is
|
||||
already committed. If set to 0s, the coordinator will continuously
|
||||
forge at the maximum rate. Note that this is a configuration parameter of a node,
|
||||
so each coordinator may have a different value.
|
||||
example: 193.4
|
||||
additionalProperties: false
|
||||
required:
|
||||
- forgeDelay
|
||||
State:
|
||||
type: object
|
||||
description: Gobal variables of the network
|
||||
@@ -2594,6 +2600,8 @@ components:
|
||||
$ref: '#/components/schemas/StateWithdrawDelayer'
|
||||
recommendedFee:
|
||||
$ref: '#/components/schemas/RecommendedFee'
|
||||
nodeConfig:
|
||||
$ref: '#/components/schemas/NodeConfig'
|
||||
additionalProperties: false
|
||||
required:
|
||||
- network
|
||||
@@ -2602,6 +2610,7 @@ components:
|
||||
- auction
|
||||
- withdrawalDelayer
|
||||
- recommendedFee
|
||||
- nodeConfig
|
||||
StateNetwork:
|
||||
type: object
|
||||
description: Gobal statistics of the network
|
||||
@@ -2812,6 +2821,10 @@ components:
|
||||
type: number
|
||||
description: Average fee percentage paid for L2 transactions in the last 24 hours.
|
||||
example: 1.54
|
||||
estimatedTimeToForgeL1:
|
||||
type: number
|
||||
description: Estimated time needed to forge a L1 transaction, from the time it's added on the smart contract, until it's actualy forged. In seconds.
|
||||
example: 193.4
|
||||
additionalProperties: false
|
||||
required:
|
||||
- transactionsPerBatch
|
||||
@@ -2820,6 +2833,7 @@ components:
|
||||
- totalAccounts
|
||||
- totalBJJs
|
||||
- avgTransactionFee
|
||||
- estimatedTimeToForgeL1
|
||||
PendingItems:
|
||||
type: integer
|
||||
description: Amount of items that will be returned in subsequent calls to the endpoint, as long as they are done with same filters. When the value is 0 it means that all items have been sent.
|
||||
@@ -2916,7 +2930,7 @@ components:
|
||||
example: 101
|
||||
l1UserTotalBytes:
|
||||
type: integer
|
||||
description: Number of bytes that a L1 user transaction has ([20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx + [2 bytes] depositAmountFloat16 + [2 bytes] amountFloat16 + [4 bytes] tokenId + [6 bytes] toIdx).
|
||||
description: Number of bytes that a L1 user transaction has ([20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx + [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes] tokenId + [6 bytes] toIdx).
|
||||
example: 72
|
||||
maxL1UserTx:
|
||||
type: integer
|
||||
|
||||
@@ -22,7 +22,7 @@ func (a *API) getToken(c *gin.Context) {
|
||||
}
|
||||
tokenID := common.TokenID(*tokenIDUint)
|
||||
// Fetch token from historyDB
|
||||
token, err := a.h.GetToken(tokenID)
|
||||
token, err := a.h.GetTokenAPI(tokenID)
|
||||
if err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
@@ -45,7 +45,7 @@ func (a *API) getTokens(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
// Fetch exits from historyDB
|
||||
tokens, pendingItems, err := a.h.GetTokens(
|
||||
tokens, pendingItems, err := a.h.GetTokensAPI(
|
||||
tokenIDs, symbols, name, fromItem, limit, order,
|
||||
)
|
||||
if err != nil {
|
||||
@@ -53,7 +53,7 @@ func (a *API) getTokens(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type tokensResponse struct {
|
||||
Tokens []historydb.TokenWithUSD `json:"tokens"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
|
||||
@@ -34,7 +34,7 @@ func (a *API) getHistoryTxs(c *gin.Context) {
|
||||
}
|
||||
|
||||
// Fetch txs from historyDB
|
||||
txs, pendingItems, err := a.h.GetHistoryTxs(
|
||||
txs, pendingItems, err := a.h.GetTxsAPI(
|
||||
addr, bjj, tokenID, idx, batchNum, txType, fromItem, limit, order,
|
||||
)
|
||||
if err != nil {
|
||||
@@ -42,7 +42,7 @@ func (a *API) getHistoryTxs(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
type txsResponse struct {
|
||||
Txs []historydb.TxAPI `json:"transactions"`
|
||||
PendingItems uint64 `json:"pendingItems"`
|
||||
@@ -61,11 +61,11 @@ func (a *API) getHistoryTx(c *gin.Context) {
|
||||
return
|
||||
}
|
||||
// Fetch tx from historyDB
|
||||
tx, err := a.h.GetHistoryTx(txID)
|
||||
tx, err := a.h.GetTxAPI(txID)
|
||||
if err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
c.JSON(http.StatusOK, tx)
|
||||
}
|
||||
|
||||
@@ -455,7 +455,7 @@ func TestGetHistoryTx(t *testing.T) {
|
||||
// 400, due invalid TxID
|
||||
err := doBadReq("GET", endpoint+"0x001", nil, 400)
|
||||
assert.NoError(t, err)
|
||||
// 404, due inexistent TxID in DB
|
||||
// 404, due nonexistent TxID in DB
|
||||
err = doBadReq("GET", endpoint+"0x00eb5e95e1ce5e9f6c4ed402d415e8d0bdd7664769cfd2064d28da04a2c76be432", nil, 404)
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
@@ -2,6 +2,7 @@ package api
|
||||
|
||||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"net/http"
|
||||
|
||||
@@ -27,8 +28,9 @@ func (a *API) postPoolTx(c *gin.Context) {
|
||||
retBadReq(err, c)
|
||||
return
|
||||
}
|
||||
writeTx.ClientIP = c.ClientIP()
|
||||
// Insert to DB
|
||||
if err := a.l2.AddTx(writeTx); err != nil {
|
||||
if err := a.l2.AddTxAPI(writeTx); err != nil {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
@@ -49,7 +51,7 @@ func (a *API) getPoolTx(c *gin.Context) {
|
||||
retSQLErr(err, c)
|
||||
return
|
||||
}
|
||||
// Build succesfull response
|
||||
// Build successful response
|
||||
c.JSON(http.StatusOK, tx)
|
||||
}
|
||||
|
||||
@@ -169,16 +171,21 @@ func (a *API) verifyPoolL2TxWrite(txw *l2db.PoolL2TxWrite) error {
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Get public key
|
||||
account, err := a.s.LastGetAccount(poolTx.FromIdx)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Validate feeAmount
|
||||
_, err = common.CalcFeeAmount(poolTx.Amount, poolTx.Fee)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Get public key
|
||||
account, err := a.h.GetCommonAccountAPI(poolTx.FromIdx)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// Validate TokenID
|
||||
if poolTx.TokenID != account.TokenID {
|
||||
return tracerr.Wrap(fmt.Errorf("tx.TokenID (%v) != account.TokenID (%v)",
|
||||
poolTx.TokenID, account.TokenID))
|
||||
}
|
||||
// Check signature
|
||||
if !poolTx.VerifySignature(a.chainID, account.BJJ) {
|
||||
return tracerr.Wrap(errors.New("wrong signature"))
|
||||
|
||||
@@ -2,14 +2,20 @@ package api
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"crypto/ecdsa"
|
||||
"encoding/binary"
|
||||
"encoding/hex"
|
||||
"encoding/json"
|
||||
"math/big"
|
||||
"testing"
|
||||
"time"
|
||||
|
||||
ethCrypto "github.com/ethereum/go-ethereum/crypto"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
// testPoolTxReceive is a struct to be used to assert the response
|
||||
@@ -170,9 +176,9 @@ func TestPoolTxs(t *testing.T) {
|
||||
fetchedTxID := common.TxID{}
|
||||
for _, tx := range tc.poolTxsToSend {
|
||||
jsonTxBytes, err := json.Marshal(tx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader := bytes.NewReader(jsonTxBytes)
|
||||
assert.NoError(
|
||||
require.NoError(
|
||||
t, doGoodReq(
|
||||
"POST",
|
||||
endpoint,
|
||||
@@ -187,42 +193,42 @@ func TestPoolTxs(t *testing.T) {
|
||||
badTx.Amount = "99950000000000000"
|
||||
badTx.Fee = 255
|
||||
jsonTxBytes, err := json.Marshal(badTx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader := bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", endpoint, jsonTxReader, 400)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Wrong signature
|
||||
badTx = tc.poolTxsToSend[0]
|
||||
badTx.FromIdx = "hez:foo:1000"
|
||||
jsonTxBytes, err = json.Marshal(badTx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader = bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", endpoint, jsonTxReader, 400)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Wrong to
|
||||
badTx = tc.poolTxsToSend[0]
|
||||
ethAddr := "hez:0xFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFFF"
|
||||
badTx.ToEthAddr = ðAddr
|
||||
badTx.ToIdx = nil
|
||||
jsonTxBytes, err = json.Marshal(badTx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader = bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", endpoint, jsonTxReader, 400)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Wrong rq
|
||||
badTx = tc.poolTxsToSend[0]
|
||||
rqFromIdx := "hez:foo:30"
|
||||
badTx.RqFromIdx = &rqFromIdx
|
||||
jsonTxBytes, err = json.Marshal(badTx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader = bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", endpoint, jsonTxReader, 400)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// GET
|
||||
endpoint += "/"
|
||||
for _, tx := range tc.poolTxsToReceive {
|
||||
fetchedTx := testPoolTxReceive{}
|
||||
assert.NoError(
|
||||
require.NoError(
|
||||
t, doGoodReq(
|
||||
"GET",
|
||||
endpoint+tx.TxID.String(),
|
||||
@@ -233,10 +239,10 @@ func TestPoolTxs(t *testing.T) {
|
||||
}
|
||||
// 400, due invalid TxID
|
||||
err = doBadReq("GET", endpoint+"0xG2241b6f2b1dd772dba391f4a1a3407c7c21f598d86e2585a14e616fb4a255f823", nil, 400)
|
||||
assert.NoError(t, err)
|
||||
// 404, due inexistent TxID in DB
|
||||
require.NoError(t, err)
|
||||
// 404, due nonexistent TxID in DB
|
||||
err = doBadReq("GET", endpoint+"0x02241b6f2b1dd772dba391f4a1a3407c7c21f598d86e2585a14e616fb4a255f823", nil, 404)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
func assertPoolTx(t *testing.T, expected, actual testPoolTxReceive) {
|
||||
@@ -256,3 +262,73 @@ func assertPoolTx(t *testing.T, expected, actual testPoolTxReceive) {
|
||||
}
|
||||
assert.Equal(t, expected, actual)
|
||||
}
|
||||
|
||||
// TestAllTosNull test that the API doesn't accept txs with all the TOs set to null (to eth, to bjj, to idx)
|
||||
func TestAllTosNull(t *testing.T) {
|
||||
// Generate account:
|
||||
// Ethereum private key
|
||||
var key ecdsa.PrivateKey
|
||||
key.D = big.NewInt(int64(4444)) // only for testing
|
||||
key.PublicKey.X, key.PublicKey.Y = ethCrypto.S256().ScalarBaseMult(key.D.Bytes())
|
||||
key.Curve = ethCrypto.S256()
|
||||
addr := ethCrypto.PubkeyToAddress(key.PublicKey)
|
||||
// BJJ private key
|
||||
var sk babyjub.PrivateKey
|
||||
var iBytes [8]byte
|
||||
binary.LittleEndian.PutUint64(iBytes[:], 4444)
|
||||
copy(sk[:], iBytes[:]) // only for testing
|
||||
account := common.Account{
|
||||
Idx: 4444,
|
||||
TokenID: 0,
|
||||
BatchNum: 1,
|
||||
BJJ: sk.Public().Compress(),
|
||||
EthAddr: addr,
|
||||
Nonce: 0,
|
||||
Balance: big.NewInt(1000000),
|
||||
}
|
||||
// Add account to history DB (required to verify signature)
|
||||
err := api.h.AddAccounts([]common.Account{account})
|
||||
assert.NoError(t, err)
|
||||
// Genrate tx with all tos set to nil (to eth, to bjj, to idx)
|
||||
tx := common.PoolL2Tx{
|
||||
FromIdx: account.Idx,
|
||||
TokenID: account.TokenID,
|
||||
Amount: big.NewInt(1000),
|
||||
Fee: 200,
|
||||
Nonce: 0,
|
||||
}
|
||||
// Set idx and type manually, and check that the function doesn't allow it
|
||||
_, err = common.NewPoolL2Tx(&tx)
|
||||
assert.Error(t, err)
|
||||
tx.Type = common.TxTypeTransfer
|
||||
var txID common.TxID
|
||||
txIDRaw, err := hex.DecodeString("02e66e24f7f25272906647c8fd1d7fe8acf3cf3e9b38ffc9f94bbb5090dc275073")
|
||||
assert.NoError(t, err)
|
||||
copy(txID[:], txIDRaw)
|
||||
tx.TxID = txID
|
||||
// Sign tx
|
||||
toSign, err := tx.HashToSign(0)
|
||||
assert.NoError(t, err)
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
tx.Signature = sig.Compress()
|
||||
// Transform common.PoolL2Tx ==> testPoolTxSend
|
||||
txToSend := testPoolTxSend{
|
||||
TxID: tx.TxID,
|
||||
Type: tx.Type,
|
||||
TokenID: tx.TokenID,
|
||||
FromIdx: idxToHez(tx.FromIdx, "ETH"),
|
||||
Amount: tx.Amount.String(),
|
||||
Fee: tx.Fee,
|
||||
Nonce: tx.Nonce,
|
||||
Signature: tx.Signature,
|
||||
}
|
||||
// Send tx to the API
|
||||
jsonTxBytes, err := json.Marshal(txToSend)
|
||||
require.NoError(t, err)
|
||||
jsonTxReader := bytes.NewReader(jsonTxBytes)
|
||||
err = doBadReq("POST", apiURL+"transactions-pool", jsonTxReader, 400)
|
||||
require.NoError(t, err)
|
||||
// Clean historyDB: the added account shouldn't be there for other tests
|
||||
_, err = api.h.DB().DB.Exec("delete from account where idx = 4444")
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
@@ -91,7 +91,7 @@ func (c *CollectedFees) UnmarshalJSON(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezEthAddr is used to scan/value Ethereum Address directly into strings that follow the Ethereum address hez fotmat (^hez:0x[a-fA-F0-9]{40}$) from/to sql DBs.
|
||||
// HezEthAddr is used to scan/value Ethereum Address directly into strings that follow the Ethereum address hez format (^hez:0x[a-fA-F0-9]{40}$) from/to sql DBs.
|
||||
// It assumes that Ethereum Address are inserted/fetched to/from the DB using the default Scan/Value interface
|
||||
type HezEthAddr string
|
||||
|
||||
@@ -143,7 +143,7 @@ func (s *StrHezEthAddr) UnmarshalText(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezBJJ is used to scan/value *babyjub.PublicKeyComp directly into strings that follow the BJJ public key hez fotmat (^hez:[A-Za-z0-9_-]{44}$) from/to sql DBs.
|
||||
// HezBJJ is used to scan/value *babyjub.PublicKeyComp directly into strings that follow the BJJ public key hez format (^hez:[A-Za-z0-9_-]{44}$) from/to sql DBs.
|
||||
// It assumes that *babyjub.PublicKeyComp are inserted/fetched to/from the DB using the default Scan/Value interface
|
||||
type HezBJJ string
|
||||
|
||||
@@ -216,7 +216,7 @@ func (b HezBJJ) Value() (driver.Value, error) {
|
||||
// StrHezBJJ is used to unmarshal HezBJJ directly into an alias of babyjub.PublicKeyComp
|
||||
type StrHezBJJ babyjub.PublicKeyComp
|
||||
|
||||
// UnmarshalText unmarshals a StrHezBJJ
|
||||
// UnmarshalText unmarshalls a StrHezBJJ
|
||||
func (s *StrHezBJJ) UnmarshalText(text []byte) error {
|
||||
bjj, err := hezStrToBJJ(string(text))
|
||||
if err != nil {
|
||||
@@ -226,8 +226,8 @@ func (s *StrHezBJJ) UnmarshalText(text []byte) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// HezIdx is used to value common.Idx directly into strings that follow the Idx key hez fotmat (hez:tokenSymbol:idx) to sql DBs.
|
||||
// Note that this can only be used to insert to DB since there is no way to automaticaly read from the DB since it needs the tokenSymbol
|
||||
// HezIdx is used to value common.Idx directly into strings that follow the Idx key hez format (hez:tokenSymbol:idx) to sql DBs.
|
||||
// Note that this can only be used to insert to DB since there is no way to automatically read from the DB since it needs the tokenSymbol
|
||||
type HezIdx string
|
||||
|
||||
// StrHezIdx is used to unmarshal HezIdx directly into an alias of common.Idx
|
||||
|
||||
@@ -2,6 +2,7 @@ package batchbuilder
|
||||
|
||||
import (
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/kvdb"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/hermez-node/txprocessor"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
@@ -27,9 +28,16 @@ type ConfigBatch struct {
|
||||
|
||||
// NewBatchBuilder constructs a new BatchBuilder, and executes the bb.Reset
|
||||
// method
|
||||
func NewBatchBuilder(dbpath string, synchronizerStateDB *statedb.StateDB, batchNum common.BatchNum, nLevels uint64) (*BatchBuilder, error) {
|
||||
localStateDB, err := statedb.NewLocalStateDB(dbpath, 128, synchronizerStateDB,
|
||||
statedb.TypeBatchBuilder, int(nLevels))
|
||||
func NewBatchBuilder(dbpath string, synchronizerStateDB *statedb.StateDB, batchNum common.BatchNum,
|
||||
nLevels uint64) (*BatchBuilder, error) {
|
||||
localStateDB, err := statedb.NewLocalStateDB(
|
||||
statedb.Config{
|
||||
Path: dbpath,
|
||||
Keep: kvdb.DefaultKeep,
|
||||
Type: statedb.TypeBatchBuilder,
|
||||
NLevels: int(nLevels),
|
||||
},
|
||||
synchronizerStateDB)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -47,7 +55,7 @@ func NewBatchBuilder(dbpath string, synchronizerStateDB *statedb.StateDB, batchN
|
||||
// copy of the rollup state from the Synchronizer at that `batchNum`, otherwise
|
||||
// it can just roll back the internal copy.
|
||||
func (bb *BatchBuilder) Reset(batchNum common.BatchNum, fromSynchronizer bool) error {
|
||||
return bb.localStateDB.Reset(batchNum, fromSynchronizer)
|
||||
return tracerr.Wrap(bb.localStateDB.Reset(batchNum, fromSynchronizer))
|
||||
}
|
||||
|
||||
// BuildBatch takes the transactions and returns the common.ZKInputs of the next batch
|
||||
@@ -57,7 +65,10 @@ func (bb *BatchBuilder) BuildBatch(coordIdxs []common.Idx, configBatch *ConfigBa
|
||||
tp := txprocessor.NewTxProcessor(bbStateDB, configBatch.TxProcessorConfig)
|
||||
|
||||
ptOut, err := tp.ProcessTxs(coordIdxs, l1usertxs, l1coordinatortxs, pooll2txs)
|
||||
return ptOut.ZKInputs, tracerr.Wrap(err)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return ptOut.ZKInputs, nil
|
||||
}
|
||||
|
||||
// LocalStateDB returns the underlying LocalStateDB
|
||||
|
||||
@@ -15,7 +15,8 @@ func TestBatchBuilder(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
defer assert.Nil(t, os.RemoveAll(dir))
|
||||
|
||||
synchDB, err := statedb.NewStateDB(dir, 128, statedb.TypeBatchBuilder, 0)
|
||||
synchDB, err := statedb.NewStateDB(statedb.Config{Path: dir, Keep: 128,
|
||||
Type: statedb.TypeBatchBuilder, NLevels: 0})
|
||||
assert.Nil(t, err)
|
||||
|
||||
bbDir, err := ioutil.TempDir("", "tmpBatchBuilderDB")
|
||||
|
||||
1
cli/node/.gitignore
vendored
1
cli/node/.gitignore
vendored
@@ -1,2 +1,3 @@
|
||||
cfg.example.secret.toml
|
||||
cfg.toml
|
||||
node
|
||||
|
||||
@@ -2,6 +2,10 @@
|
||||
|
||||
This is the main cli for the node
|
||||
|
||||
## Go version
|
||||
|
||||
The `hermez-node` has been tested with go version 1.14
|
||||
|
||||
## Usage
|
||||
|
||||
```
|
||||
@@ -65,29 +69,64 @@ when running the coordinator in sync mode
|
||||
- The node requires a PostgreSQL database. The parameters of the server and
|
||||
database must be set in the `PostgreSQL` section.
|
||||
|
||||
## Building
|
||||
|
||||
*All commands assume you are at the `cli/node` directory.*
|
||||
|
||||
Building the node requires using the packr utility to bundle the database
|
||||
migrations inside the resulting binary. Install the packr utility with:
|
||||
```
|
||||
cd /tmp && go get -u github.com/gobuffalo/packr/v2/packr2 && cd -
|
||||
```
|
||||
|
||||
Make sure your `$PATH` contains `$GOPATH/bin`, otherwise the packr utility will
|
||||
not be found.
|
||||
|
||||
Now build the node executable:
|
||||
```
|
||||
cd ../../db && packr2 && cd -
|
||||
go build .
|
||||
cd ../../db && packr2 clean && cd -
|
||||
```
|
||||
|
||||
The executable is `node`.
|
||||
|
||||
## Usage Examples
|
||||
|
||||
The following commands assume you have built the node previously. You can also
|
||||
run the following examples by replacing `./node` with `go run .` and executing
|
||||
them in the `cli/node` directory to build from source and run at the same time.
|
||||
|
||||
Run the node in mode synchronizer:
|
||||
```
|
||||
go run . --mode sync --cfg cfg.buidler.toml run
|
||||
./node --mode sync --cfg cfg.buidler.toml run
|
||||
```
|
||||
|
||||
Run the node in mode coordinator:
|
||||
```
|
||||
go run . --mode coord --cfg cfg.buidler.toml run
|
||||
./node --mode coord --cfg cfg.buidler.toml run
|
||||
```
|
||||
|
||||
Import an ethereum private key into the keystore:
|
||||
```
|
||||
go run . --mode coord --cfg cfg.buidler.toml importkey --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
||||
./node --mode coord --cfg cfg.buidler.toml importkey --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
||||
```
|
||||
|
||||
Generate a new BabyJubJub key pair:
|
||||
```
|
||||
go run . --mode coord --cfg cfg.buidler.toml genbjj
|
||||
./node --mode coord --cfg cfg.buidler.toml genbjj
|
||||
```
|
||||
|
||||
Wipe the entier SQL database (this will destroy all synchronized and pool data):
|
||||
Wipe the entier SQL database (this will destroy all synchronized and pool
|
||||
data):
|
||||
```
|
||||
go run . --mode coord --cfg cfg.buidler.toml wipesql
|
||||
./node --mode coord --cfg cfg.buidler.toml wipesql
|
||||
```
|
||||
|
||||
Discard all synchronized blocks and associated state up to a given block
|
||||
number. This command is useful in case the synchronizer reaches an invalid
|
||||
state and you want to roll back a few blocks and try again (maybe with some
|
||||
fixes in the code).
|
||||
```
|
||||
./node --mode coord --cfg cfg.buidler.toml discard --block 8061330
|
||||
```
|
||||
|
||||
@@ -3,26 +3,36 @@ Address = "localhost:8086"
|
||||
Explorer = true
|
||||
UpdateMetricsInterval = "10s"
|
||||
UpdateRecommendedFeeInterval = "10s"
|
||||
MaxSQLConnections = 100
|
||||
SQLConnectionTimeout = "2s"
|
||||
|
||||
[PriceUpdater]
|
||||
Interval = "10s"
|
||||
URL = "https://api-pub.bitfinex.com/v2/"
|
||||
Type = "bitfinexV2"
|
||||
# URL = "https://api.coingecko.com/api/v3/"
|
||||
# Type = "coingeckoV3"
|
||||
|
||||
[Debug]
|
||||
APIAddress = "localhost:12345"
|
||||
MeddlerLogs = true
|
||||
GinDebugMode = true
|
||||
|
||||
[StateDB]
|
||||
Path = "/tmp/iden3-test/hermez/statedb"
|
||||
Keep = 256
|
||||
|
||||
[PostgreSQL]
|
||||
Port = 5432
|
||||
Host = "localhost"
|
||||
User = "hermez"
|
||||
Password = "yourpasswordhere"
|
||||
Name = "hermez"
|
||||
PortWrite = 5432
|
||||
HostWrite = "localhost"
|
||||
UserWrite = "hermez"
|
||||
PasswordWrite = "yourpasswordhere"
|
||||
NameWrite = "hermez"
|
||||
# PortRead = 5432
|
||||
# HostRead = "localhost"
|
||||
# UserRead = "hermez"
|
||||
# PasswordRead = "yourpasswordhere"
|
||||
# NameRead = "hermez"
|
||||
|
||||
[Web3]
|
||||
URL = "http://localhost:8545"
|
||||
@@ -39,15 +49,22 @@ TokenHEZ = "0x5D94e3e7aeC542aB0F9129B9a7BAdeb5B3Ca0f77"
|
||||
TokenHEZName = "Hermez Network Token"
|
||||
|
||||
[Coordinator]
|
||||
# ForgerAddress = "0x05c23b938a85ab26A36E6314a0D02080E9ca6BeD" # Non-Boot Coordinator
|
||||
ForgerAddress = "0x05c23b938a85ab26A36E6314a0D02080E9ca6BeD" # Non-Boot Coordinator
|
||||
# ForgerAddressPrivateKey = "0x30f5fddb34cd4166adb2c6003fa6b18f380fd2341376be42cf1c7937004ac7a3"
|
||||
ForgerAddress = "0xb4124ceb3451635dacedd11767f004d8a28c6ee7" # Boot Coordinator
|
||||
# ForgerAddress = "0xb4124ceb3451635dacedd11767f004d8a28c6ee7" # Boot Coordinator
|
||||
# ForgerAddressPrivateKey = "0xa8a54b2d8197bc0b19bb8a084031be71835580a01e70a45a13babd16c9bc1563"
|
||||
MinimumForgeAddressBalance = "0"
|
||||
ConfirmBlocks = 10
|
||||
L1BatchTimeoutPerc = 0.999
|
||||
L1BatchTimeoutPerc = 0.6
|
||||
StartSlotBlocksDelay = 2
|
||||
ScheduleBatchBlocksAheadCheck = 3
|
||||
SendBatchBlocksMarginCheck = 1
|
||||
ProofServerPollInterval = "1s"
|
||||
ForgeRetryInterval = "500ms"
|
||||
SyncRetryInterval = "1s"
|
||||
ForgeDelay = "10s"
|
||||
ForgeNoTxsDelay = "0s"
|
||||
PurgeByExtDelInterval = "1m"
|
||||
|
||||
[Coordinator.FeeAccount]
|
||||
Address = "0x56232B1c5B10038125Bc7345664B4AFD745bcF8E"
|
||||
@@ -58,6 +75,7 @@ BJJ = "0x1b176232f78ba0d388ecc5f4896eca2d3b3d4f272092469f559247297f5c0c13"
|
||||
[Coordinator.L2DB]
|
||||
SafetyPeriod = 10
|
||||
MaxTxs = 512
|
||||
MinFeeUSD = 0.0
|
||||
TTL = "24h"
|
||||
PurgeBatchDelay = 10
|
||||
InvalidateBatchDelay = 20
|
||||
@@ -78,21 +96,28 @@ MaxTx = 512
|
||||
NLevels = 32
|
||||
|
||||
[Coordinator.EthClient]
|
||||
ReceiptTimeout = "60s"
|
||||
ReceiptLoopInterval = "500ms"
|
||||
CheckLoopInterval = "500ms"
|
||||
Attempts = 4
|
||||
AttemptsDelay = "500ms"
|
||||
CallGasLimit = 300000
|
||||
GasPriceDiv = 100
|
||||
TxResendTimeout = "2m"
|
||||
NoReuseNonce = false
|
||||
MaxGasPrice = "5000000000"
|
||||
GasPriceIncPerc = 10
|
||||
|
||||
[Coordinator.EthClient.Keystore]
|
||||
Path = "/tmp/iden3-test/hermez/ethkeystore"
|
||||
Password = "yourpasswordhere"
|
||||
|
||||
[Coordinator.EthClient.ForgeBatchGasCost]
|
||||
Fixed = 600000
|
||||
L1UserTx = 15000
|
||||
L1CoordTx = 8000
|
||||
L2Tx = 250
|
||||
|
||||
[Coordinator.API]
|
||||
Coordinator = true
|
||||
|
||||
[Coordinator.Debug]
|
||||
BatchPath = "/tmp/iden3-test/hermez/batchesdebug"
|
||||
LightScrypt = true
|
||||
# RollupVerifierIndex = 0
|
||||
|
||||
187
cli/node/main.go
187
cli/node/main.go
@@ -5,16 +5,22 @@ import (
|
||||
"fmt"
|
||||
"os"
|
||||
"os/signal"
|
||||
"path"
|
||||
"strings"
|
||||
|
||||
ethKeystore "github.com/ethereum/go-ethereum/accounts/keystore"
|
||||
"github.com/ethereum/go-ethereum/crypto"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/config"
|
||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/kvdb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/node"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
"github.com/jmoiron/sqlx"
|
||||
"github.com/urfave/cli/v2"
|
||||
)
|
||||
|
||||
@@ -23,6 +29,7 @@ const (
|
||||
flagMode = "mode"
|
||||
flagSK = "privatekey"
|
||||
flagYes = "yes"
|
||||
flagBlock = "block"
|
||||
modeSync = "sync"
|
||||
modeCoord = "coord"
|
||||
)
|
||||
@@ -68,6 +75,86 @@ func cmdImportKey(c *cli.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func resetStateDBs(cfg *Config, batchNum common.BatchNum) error {
|
||||
log.Infof("Reset Synchronizer StateDB to batchNum %v...", batchNum)
|
||||
|
||||
// Manually make a checkpoint from batchNum to current to force current
|
||||
// to be a valid checkpoint. This is useful because in case of a
|
||||
// crash, current can be corrupted and the first thing that
|
||||
// `kvdb.NewKVDB` does is read the current checkpoint, which wouldn't
|
||||
// succeed in case of corruption.
|
||||
dbPath := cfg.node.StateDB.Path
|
||||
source := path.Join(dbPath, fmt.Sprintf("%s%d", kvdb.PathBatchNum, batchNum))
|
||||
current := path.Join(dbPath, kvdb.PathCurrent)
|
||||
last := path.Join(dbPath, kvdb.PathLast)
|
||||
if err := os.RemoveAll(last); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
if batchNum == 0 {
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
} else {
|
||||
if err := kvdb.PebbleMakeCheckpoint(source, current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.PebbleMakeCheckpoint: %w", err))
|
||||
}
|
||||
}
|
||||
db, err := kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(batchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
|
||||
if cfg.mode == node.ModeCoordinator {
|
||||
log.Infof("Wipe Coordinator StateDBs...")
|
||||
|
||||
// We wipe the Coordinator StateDBs entirely (by deleting
|
||||
// current and resetting to batchNum 0) because the Coordinator
|
||||
// StateDBs are always reset from Synchronizer when the
|
||||
// coordinator pipeline starts.
|
||||
dbPath := cfg.node.Coordinator.TxSelector.Path
|
||||
current := path.Join(dbPath, kvdb.PathCurrent)
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
db, err := kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("kvdb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
|
||||
dbPath = cfg.node.Coordinator.BatchBuilder.Path
|
||||
current = path.Join(dbPath, kvdb.PathCurrent)
|
||||
if err := os.RemoveAll(current); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("os.RemoveAll: %w", err))
|
||||
}
|
||||
db, err = kvdb.NewKVDB(kvdb.Config{
|
||||
Path: dbPath,
|
||||
NoGapsCheck: true,
|
||||
NoLast: true,
|
||||
})
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("statedb.NewKVDB: %w", err))
|
||||
}
|
||||
if err := db.Reset(0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("db.Reset: %w", err))
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func cmdWipeSQL(c *cli.Context) error {
|
||||
_cfg, err := parseCli(c)
|
||||
if err != nil {
|
||||
@@ -76,7 +163,8 @@ func cmdWipeSQL(c *cli.Context) error {
|
||||
cfg := _cfg.node
|
||||
yes := c.Bool(flagYes)
|
||||
if !yes {
|
||||
fmt.Print("*WARNING* Are you sure you want to delete the SQL DB? [y/N]: ")
|
||||
fmt.Print("*WARNING* Are you sure you want to delete " +
|
||||
"the SQL DB and StateDBs? [y/N]: ")
|
||||
var input string
|
||||
if _, err := fmt.Scanln(&input); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
@@ -87,18 +175,23 @@ func cmdWipeSQL(c *cli.Context) error {
|
||||
}
|
||||
}
|
||||
db, err := dbUtils.ConnectSQLDB(
|
||||
cfg.PostgreSQL.Port,
|
||||
cfg.PostgreSQL.Host,
|
||||
cfg.PostgreSQL.User,
|
||||
cfg.PostgreSQL.Password,
|
||||
cfg.PostgreSQL.Name,
|
||||
cfg.PostgreSQL.PortWrite,
|
||||
cfg.PostgreSQL.HostWrite,
|
||||
cfg.PostgreSQL.UserWrite,
|
||||
cfg.PostgreSQL.PasswordWrite,
|
||||
cfg.PostgreSQL.NameWrite,
|
||||
)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
log.Info("Wiping SQL DB...")
|
||||
if err := dbUtils.MigrationsDown(db.DB); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
return tracerr.Wrap(fmt.Errorf("dbUtils.MigrationsDown: %w", err))
|
||||
}
|
||||
|
||||
log.Info("Wiping StateDBs...")
|
||||
if err := resetStateDBs(_cfg, 0); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("resetStateDBs: %w", err))
|
||||
}
|
||||
return nil
|
||||
}
|
||||
@@ -139,6 +232,72 @@ func cmdRun(c *cli.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func cmdDiscard(c *cli.Context) error {
|
||||
_cfg, err := parseCli(c)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("error parsing flags and config: %w", err))
|
||||
}
|
||||
cfg := _cfg.node
|
||||
blockNum := c.Int64(flagBlock)
|
||||
log.Infof("Discarding all blocks up to block %v...", blockNum)
|
||||
|
||||
dbWrite, err := dbUtils.InitSQLDB(
|
||||
cfg.PostgreSQL.PortWrite,
|
||||
cfg.PostgreSQL.HostWrite,
|
||||
cfg.PostgreSQL.UserWrite,
|
||||
cfg.PostgreSQL.PasswordWrite,
|
||||
cfg.PostgreSQL.NameWrite,
|
||||
)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||
}
|
||||
var dbRead *sqlx.DB
|
||||
if cfg.PostgreSQL.HostRead == "" {
|
||||
dbRead = dbWrite
|
||||
} else if cfg.PostgreSQL.HostRead == cfg.PostgreSQL.HostWrite {
|
||||
return tracerr.Wrap(fmt.Errorf(
|
||||
"PostgreSQL.HostRead and PostgreSQL.HostWrite must be different",
|
||||
))
|
||||
} else {
|
||||
dbRead, err = dbUtils.InitSQLDB(
|
||||
cfg.PostgreSQL.PortRead,
|
||||
cfg.PostgreSQL.HostRead,
|
||||
cfg.PostgreSQL.UserRead,
|
||||
cfg.PostgreSQL.PasswordRead,
|
||||
cfg.PostgreSQL.NameRead,
|
||||
)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||
}
|
||||
}
|
||||
historyDB := historydb.NewHistoryDB(dbRead, dbWrite, nil)
|
||||
if err := historyDB.Reorg(blockNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("historyDB.Reorg: %w", err))
|
||||
}
|
||||
batchNum, err := historyDB.GetLastBatchNum()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("historyDB.GetLastBatchNum: %w", err))
|
||||
}
|
||||
l2DB := l2db.NewL2DB(
|
||||
dbRead, dbWrite,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
nil,
|
||||
)
|
||||
if err := l2DB.Reorg(batchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("l2DB.Reorg: %w", err))
|
||||
}
|
||||
|
||||
log.Info("Resetting StateDBs...")
|
||||
if err := resetStateDBs(_cfg, batchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("resetStateDBs: %w", err))
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// Config is the configuration of the hermez node execution
|
||||
type Config struct {
|
||||
mode node.Mode
|
||||
@@ -223,7 +382,7 @@ func main() {
|
||||
{
|
||||
Name: "wipesql",
|
||||
Aliases: []string{},
|
||||
Usage: "Wipe the SQL DB (HistoryDB and L2DB), " +
|
||||
Usage: "Wipe the SQL DB (HistoryDB and L2DB) and the StateDBs, " +
|
||||
"leaving the DB in a clean state",
|
||||
Action: cmdWipeSQL,
|
||||
Flags: []cli.Flag{
|
||||
@@ -239,6 +398,18 @@ func main() {
|
||||
Usage: "Run the hermez-node in the indicated mode",
|
||||
Action: cmdRun,
|
||||
},
|
||||
{
|
||||
Name: "discard",
|
||||
Aliases: []string{},
|
||||
Usage: "Discard blocks up to a specified block number",
|
||||
Action: cmdDiscard,
|
||||
Flags: []cli.Flag{
|
||||
&cli.Int64Flag{
|
||||
Name: flagBlock,
|
||||
Usage: "last block number to keep",
|
||||
Required: false,
|
||||
}},
|
||||
},
|
||||
}
|
||||
|
||||
err := app.Run(os.Args)
|
||||
|
||||
@@ -72,7 +72,8 @@ func (idx Idx) BigInt() *big.Int {
|
||||
// IdxFromBytes returns Idx from a byte array
|
||||
func IdxFromBytes(b []byte) (Idx, error) {
|
||||
if len(b) != IdxBytesLen {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse Idx, bytes len %d, expected %d", len(b), IdxBytesLen))
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse Idx, bytes len %d, expected %d",
|
||||
len(b), IdxBytesLen))
|
||||
}
|
||||
var idxBytes [8]byte
|
||||
copy(idxBytes[2:], b[:])
|
||||
@@ -194,7 +195,8 @@ func (a *Account) BigInts() ([NLeafElems]*big.Int, error) {
|
||||
return e, nil
|
||||
}
|
||||
|
||||
// HashValue returns the value of the Account, which is the Poseidon hash of its *big.Int representation
|
||||
// HashValue returns the value of the Account, which is the Poseidon hash of its
|
||||
// *big.Int representation
|
||||
func (a *Account) HashValue() (*big.Int, error) {
|
||||
bi, err := a.BigInts()
|
||||
if err != nil {
|
||||
@@ -263,3 +265,13 @@ type IdxNonce struct {
|
||||
Idx Idx `db:"idx"`
|
||||
Nonce Nonce `db:"nonce"`
|
||||
}
|
||||
|
||||
// AccountUpdate represents an account balance and/or nonce update after a
|
||||
// processed batch
|
||||
type AccountUpdate struct {
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
BatchNum BatchNum `meddler:"batch_num"`
|
||||
Idx Idx `meddler:"idx"`
|
||||
Nonce Nonce `meddler:"nonce"`
|
||||
Balance *big.Int `meddler:"balance,bigint"`
|
||||
}
|
||||
|
||||
@@ -76,7 +76,8 @@ func TestNonceParser(t *testing.T) {
|
||||
|
||||
func TestAccount(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -115,7 +116,8 @@ func TestAccountLoop(t *testing.T) {
|
||||
// check that for different deterministic BabyJubJub keys & random Address there is no problem
|
||||
for i := 0; i < 256; i++ {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -199,7 +201,8 @@ func bigFromStr(h string, u int) *big.Int {
|
||||
|
||||
func TestAccountHashValue(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -212,13 +215,16 @@ func TestAccountHashValue(t *testing.T) {
|
||||
}
|
||||
v, err := account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "16297758255249203915951182296472515138555043617458222397753168518282206850764", v.String())
|
||||
assert.Equal(t,
|
||||
"447675324273474410516096114710387312413478475468606444107594732044698919451",
|
||||
v.String())
|
||||
}
|
||||
|
||||
func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
// values from js test vectors
|
||||
ay := new(big.Int).Sub(new(big.Int).Exp(big.NewInt(2), big.NewInt(253), nil), big.NewInt(1))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff", (hex.EncodeToString(ay.Bytes())))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff",
|
||||
(hex.EncodeToString(ay.Bytes())))
|
||||
bjjPoint, err := babyjub.PointFromSignAndY(true, ay)
|
||||
require.NoError(t, err)
|
||||
bjj := babyjub.PublicKey(*bjjPoint)
|
||||
@@ -236,16 +242,22 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "9444732965739290427391", e[0].String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895", e[1].String())
|
||||
assert.Equal(t, "14474011154664524427946373126085988481658748083205070504932198000989141204991", e[2].String())
|
||||
assert.Equal(t,
|
||||
"14474011154664524427946373126085988481658748083205070504932198000989141204991",
|
||||
e[2].String())
|
||||
assert.Equal(t, "1461501637330902918203684832716283019655932542975", e[3].String())
|
||||
|
||||
h, err := poseidon.Hash(e[:])
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "4550823210217540218403400309533329186487982452461145263910122718498735057257", h.String())
|
||||
assert.Equal(t,
|
||||
"13265203488631320682117942952393454767418777767637549409684833552016769103047",
|
||||
h.String())
|
||||
|
||||
v, err := account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "4550823210217540218403400309533329186487982452461145263910122718498735057257", v.String())
|
||||
assert.Equal(t,
|
||||
"13265203488631320682117942952393454767418777767637549409684833552016769103047",
|
||||
v.String())
|
||||
|
||||
// second account
|
||||
ay = big.NewInt(0)
|
||||
@@ -261,7 +273,9 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
}
|
||||
v, err = account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "7750253361301235345986002241352365187241910378619330147114280396816709365657", v.String())
|
||||
assert.Equal(t,
|
||||
"2351654555892372227640888372176282444150254868378439619268573230312091195718",
|
||||
v.String())
|
||||
|
||||
// third account
|
||||
ay = bigFromStr("21b0a1688b37f77b1d1d5539ec3b826db5ac78b2513f574a04c50a7d4f8246d7", 16)
|
||||
@@ -279,11 +293,15 @@ func TestAccountHashValueTestVectors(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "554050781187", e[0].String())
|
||||
assert.Equal(t, "42000000000000000000", e[1].String())
|
||||
assert.Equal(t, "15238403086306505038849621710779816852318505119327426213168494964113886299863", e[2].String())
|
||||
assert.Equal(t,
|
||||
"15238403086306505038849621710779816852318505119327426213168494964113886299863",
|
||||
e[2].String())
|
||||
assert.Equal(t, "935037732739828347587684875151694054123613453305", e[3].String())
|
||||
v, err = account.HashValue()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "10565754214047872850889045989683221123564392137456000481397520902594455245517", v.String())
|
||||
assert.Equal(t,
|
||||
"15036148928138382129196903417666258171042923749783835283230591475172197254845",
|
||||
v.String())
|
||||
}
|
||||
|
||||
func TestAccountErrNotInFF(t *testing.T) {
|
||||
@@ -312,7 +330,8 @@ func TestAccountErrNotInFF(t *testing.T) {
|
||||
|
||||
func TestAccountErrNumOverflowNonce(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -339,7 +358,8 @@ func TestAccountErrNumOverflowNonce(t *testing.T) {
|
||||
|
||||
func TestAccountErrNumOverflowBalance(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -351,14 +371,16 @@ func TestAccountErrNumOverflowBalance(t *testing.T) {
|
||||
BJJ: pk.Compress(),
|
||||
EthAddr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
||||
}
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895", account.Balance.String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512895",
|
||||
account.Balance.String())
|
||||
|
||||
_, err = account.Bytes()
|
||||
assert.NoError(t, err)
|
||||
|
||||
// force value overflow
|
||||
account.Balance = new(big.Int).Exp(big.NewInt(2), big.NewInt(192), nil)
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512896", account.Balance.String())
|
||||
assert.Equal(t, "6277101735386680763835789423207666416102355444464034512896",
|
||||
account.Balance.String())
|
||||
b, err := account.Bytes()
|
||||
assert.NotNil(t, err)
|
||||
assert.Equal(t, fmt.Errorf("%s Balance", ErrNumOverflow), tracerr.Unwrap(err))
|
||||
|
||||
@@ -1,21 +1,30 @@
|
||||
package common
|
||||
|
||||
import (
|
||||
"encoding/binary"
|
||||
"strconv"
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
ethMath "github.com/ethereum/go-ethereum/common/math"
|
||||
ethCrypto "github.com/ethereum/go-ethereum/crypto"
|
||||
ethSigner "github.com/ethereum/go-ethereum/signer/core"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
)
|
||||
|
||||
// AccountCreationAuthMsg is the message that is signed to authorize a Hermez
|
||||
// account creation
|
||||
const AccountCreationAuthMsg = "I authorize this babyjubjub key for hermez rollup account creation"
|
||||
const AccountCreationAuthMsg = "Account creation"
|
||||
|
||||
// EthMsgPrefix is the prefix for message signing at the Ethereum ecosystem
|
||||
const EthMsgPrefix = "\x19Ethereum Signed Message:\n"
|
||||
// EIP712Version is the used version of the EIP-712
|
||||
const EIP712Version = "1"
|
||||
|
||||
// EIP712Provider defines the Provider for the EIP-712
|
||||
const EIP712Provider = "Hermez Network"
|
||||
|
||||
var (
|
||||
// EmptyEthSignature is an ethereum signature of all zeroes
|
||||
EmptyEthSignature = make([]byte, 65)
|
||||
)
|
||||
|
||||
// AccountCreationAuth authorizations sent by users to the L2DB, to be used for
|
||||
// account creations when necessary
|
||||
@@ -26,44 +35,82 @@ type AccountCreationAuth struct {
|
||||
Timestamp time.Time `meddler:"timestamp,utctime"`
|
||||
}
|
||||
|
||||
// toHash returns a byte array to be hashed from the AccountCreationAuth, which
|
||||
// follows the EIP-712 encoding
|
||||
func (a *AccountCreationAuth) toHash(chainID uint16,
|
||||
hermezContractAddr ethCommon.Address) []byte {
|
||||
var chainIDBytes [2]byte
|
||||
binary.BigEndian.PutUint16(chainIDBytes[:], chainID)
|
||||
// [EthPrefix | AccountCreationAuthMsg | compressedBJJ | chainID | hermezContractAddr]
|
||||
var b []byte
|
||||
b = append(b, []byte(AccountCreationAuthMsg)...)
|
||||
b = append(b, SwapEndianness(a.BJJ[:])...) // for js implementation compatibility
|
||||
b = append(b, chainIDBytes[:]...)
|
||||
b = append(b, hermezContractAddr[:]...)
|
||||
hermezContractAddr ethCommon.Address) ([]byte, error) {
|
||||
chainIDFormatted := ethMath.NewHexOrDecimal256(int64(chainID))
|
||||
|
||||
ethPrefix := EthMsgPrefix + strconv.Itoa(len(b))
|
||||
return append([]byte(ethPrefix), b...)
|
||||
signerData := ethSigner.TypedData{
|
||||
Types: ethSigner.Types{
|
||||
"EIP712Domain": []ethSigner.Type{
|
||||
{Name: "name", Type: "string"},
|
||||
{Name: "version", Type: "string"},
|
||||
{Name: "chainId", Type: "uint256"},
|
||||
{Name: "verifyingContract", Type: "address"},
|
||||
},
|
||||
"Authorise": []ethSigner.Type{
|
||||
{Name: "Provider", Type: "string"},
|
||||
{Name: "Authorisation", Type: "string"},
|
||||
{Name: "BJJKey", Type: "bytes32"},
|
||||
},
|
||||
},
|
||||
PrimaryType: "Authorise",
|
||||
Domain: ethSigner.TypedDataDomain{
|
||||
Name: EIP712Provider,
|
||||
Version: EIP712Version,
|
||||
ChainId: chainIDFormatted,
|
||||
VerifyingContract: hermezContractAddr.Hex(),
|
||||
},
|
||||
Message: ethSigner.TypedDataMessage{
|
||||
"Provider": EIP712Provider,
|
||||
"Authorisation": AccountCreationAuthMsg,
|
||||
"BJJKey": SwapEndianness(a.BJJ[:]),
|
||||
},
|
||||
}
|
||||
|
||||
domainSeparator, err := signerData.HashStruct("EIP712Domain", signerData.Domain.Map())
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
typedDataHash, err := signerData.HashStruct(signerData.PrimaryType, signerData.Message)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
rawData := []byte{0x19, 0x01} // "\x19\x01"
|
||||
rawData = append(rawData, domainSeparator...)
|
||||
rawData = append(rawData, typedDataHash...)
|
||||
return rawData, nil
|
||||
}
|
||||
|
||||
// HashToSign returns the hash to be signed by the Etherum address to authorize
|
||||
// the account creation
|
||||
// HashToSign returns the hash to be signed by the Ethereum address to authorize
|
||||
// the account creation, which follows the EIP-712 encoding
|
||||
func (a *AccountCreationAuth) HashToSign(chainID uint16,
|
||||
hermezContractAddr ethCommon.Address) ([]byte, error) {
|
||||
b := a.toHash(chainID, hermezContractAddr)
|
||||
return ethCrypto.Keccak256Hash(b).Bytes(), nil
|
||||
b, err := a.toHash(chainID, hermezContractAddr)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return ethCrypto.Keccak256(b), nil
|
||||
}
|
||||
|
||||
// Sign signs the account creation authorization message using the provided
|
||||
// `signHash` function, and stores the signaure in `a.Signature`. `signHash`
|
||||
// `signHash` function, and stores the signature in `a.Signature`. `signHash`
|
||||
// should do an ethereum signature using the account corresponding to
|
||||
// `a.EthAddr`. The `signHash` function is used to make signig flexible: in
|
||||
// `a.EthAddr`. The `signHash` function is used to make signing flexible: in
|
||||
// tests we sign directly using the private key, outside tests we sign using
|
||||
// the keystore (which never exposes the private key).
|
||||
// the keystore (which never exposes the private key). Sign follows the EIP-712
|
||||
// encoding.
|
||||
func (a *AccountCreationAuth) Sign(signHash func(hash []byte) ([]byte, error),
|
||||
chainID uint16, hermezContractAddr ethCommon.Address) error {
|
||||
hash, err := a.HashToSign(chainID, hermezContractAddr)
|
||||
if err != nil {
|
||||
return err
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
sig, err := signHash(hash)
|
||||
if err != nil {
|
||||
return err
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
sig[64] += 27
|
||||
a.Signature = sig
|
||||
@@ -72,7 +119,8 @@ func (a *AccountCreationAuth) Sign(signHash func(hash []byte) ([]byte, error),
|
||||
}
|
||||
|
||||
// VerifySignature ensures that the Signature is done with the EthAddr, for the
|
||||
// chainID and hermezContractAddress passed by parameter
|
||||
// chainID and hermezContractAddress passed by parameter. VerifySignature
|
||||
// follows the EIP-712 encoding.
|
||||
func (a *AccountCreationAuth) VerifySignature(chainID uint16,
|
||||
hermezContractAddr ethCommon.Address) bool {
|
||||
// Calculate hash to be signed
|
||||
|
||||
@@ -13,7 +13,8 @@ import (
|
||||
|
||||
func TestAccountCreationAuthSignVerify(t *testing.T) {
|
||||
// Ethereum key
|
||||
ethSk, err := ethCrypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
ethSk, err :=
|
||||
ethCrypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
require.NoError(t, err)
|
||||
ethAddr := ethCrypto.PubkeyToAddress(ethSk.PublicKey)
|
||||
|
||||
@@ -39,7 +40,7 @@ func TestAccountCreationAuthSignVerify(t *testing.T) {
|
||||
// Hash and sign manually and compare the generated signature
|
||||
hash, err := a.HashToSign(chainID, hermezContractAddr)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, "4f8df75e96fdce1ac90bb2f8d81c42047600f85bfcef80ce3b91c2a2afc58c1e",
|
||||
assert.Equal(t, "9414667457e658dd31949b82996b75c65a055512244c3bbfd22ff56add02ba65",
|
||||
hex.EncodeToString(hash))
|
||||
sig, err := ethCrypto.Sign(hash, ethSk)
|
||||
require.NoError(t, err)
|
||||
@@ -69,35 +70,38 @@ func TestAccountCreationAuthJSComp(t *testing.T) {
|
||||
sigExpected string
|
||||
}
|
||||
var tvs []testVector
|
||||
//nolint:lll
|
||||
tv0 := testVector{
|
||||
ethSk: "0000000000000000000000000000000000000000000000000000000000000001",
|
||||
expectedAddress: "0x7E5F4552091A69125d5DfCb7b8C2659029395Bdf",
|
||||
pkCompStr: "21b0a1688b37f77b1d1d5539ec3b826db5ac78b2513f574a04c50a7d4f8246d7",
|
||||
chainID: uint16(4),
|
||||
hermezContractAddr: "0x7e5f4552091a69125d5dfcb7b8c2659029395bdf",
|
||||
toHashExpected: "19457468657265756d205369676e6564204d6573736167653a0a3132304920617574686f72697a65207468697320626162796a75626a7562206b657920666f72206865726d657a20726f6c6c7570206163636f756e74206372656174696f6e21b0a1688b37f77b1d1d5539ec3b826db5ac78b2513f574a04c50a7d4f8246d700047e5f4552091a69125d5dfcb7b8c2659029395bdf",
|
||||
hashExpected: "39afea52d843a4de905b6b5ebb0ee8c678141f711d96d9b429c4aec10ef9911f",
|
||||
sigExpected: "73d10d6ecf06ee8a5f60ac90f06b78bef9c650f414ba3ac73e176dc32e896159147457e9c86f0b4bd60fdaf2c0b2aec890a7df993d69a4805e242a6b845ebf231c",
|
||||
toHashExpected: "190189658bba487e11c7da602676ee32bc90b77d3f32a305b147e4f3c3b35f19672e5d84ccc38d0ab245c469b719549d837113465c2abf9972c49403ca6fd10ed3dc",
|
||||
hashExpected: "c56eba41e511df100c804c5c09288f35887efea4f033be956481af335df3bea2",
|
||||
sigExpected: "dbedcc5ce02db8f48afbdb2feba9a3a31848eaa8fca5f312ce37b01db45d2199208335330d4445bd2f51d1db68dbc0d0bf3585c4a07504b4efbe46a69eaae5a21b",
|
||||
}
|
||||
//nolint:lll
|
||||
tv1 := testVector{
|
||||
ethSk: "0000000000000000000000000000000000000000000000000000000000000002",
|
||||
expectedAddress: "0x2B5AD5c4795c026514f8317c7a215E218DcCD6cF",
|
||||
pkCompStr: "093985b1993d9f743f9d7d943ed56f38601cb8b196db025f79650c4007c3054d",
|
||||
chainID: uint16(0),
|
||||
hermezContractAddr: "0x2b5ad5c4795c026514f8317c7a215e218dccd6cf",
|
||||
toHashExpected: "19457468657265756d205369676e6564204d6573736167653a0a3132304920617574686f72697a65207468697320626162796a75626a7562206b657920666f72206865726d657a20726f6c6c7570206163636f756e74206372656174696f6e093985b1993d9f743f9d7d943ed56f38601cb8b196db025f79650c4007c3054d00002b5ad5c4795c026514f8317c7a215e218dccd6cf",
|
||||
hashExpected: "89a3895993a4736232212e59566294feb3da227af44375daf3307dcad5451d5d",
|
||||
sigExpected: "bb4156156c705494ad5f99030342c64657e51e2994750f92125717c40bf56ad632044aa6bd00979feea92c417b552401e65fe5f531f15010d9d1c278da8be1df1b",
|
||||
toHashExpected: "1901dafbc253dedf90d6421dc6e25d5d9efc6985133cb2a8d363d0a081a0e3eddddc65f603a88de36aaeabd3b4cf586538c7f3fd50c94780530a3707c8c14ad9fd11",
|
||||
hashExpected: "deb9afa479282cf27b442ce8ba86b19448aa87eacef691521a33db5d0feb9959",
|
||||
sigExpected: "6a0da90ba2d2b1be679a28ebe54ee03082d44b836087391cd7d2607c1e4dafe04476e6e88dccb8707c68312512f16c947524b35c80f26c642d23953e9bb84c701c",
|
||||
}
|
||||
//nolint:lll
|
||||
tv2 := testVector{
|
||||
ethSk: "c5e8f61d1ab959b397eecc0a37a6517b8e67a0e7cf1f4bce5591f3ed80199122",
|
||||
expectedAddress: "0xc783df8a850f42e7F7e57013759C285caa701eB6",
|
||||
pkCompStr: "22870c1bcc451396202d62f566026eab8e438c6c91decf8ddf63a6c162619b52",
|
||||
chainID: uint16(31337), // =0x7a69
|
||||
hermezContractAddr: "0xf4e77E5Da47AC3125140c470c71cBca77B5c638c",
|
||||
toHashExpected: "19457468657265756d205369676e6564204d6573736167653a0a3132304920617574686f72697a65207468697320626162796a75626a7562206b657920666f72206865726d657a20726f6c6c7570206163636f756e74206372656174696f6e22870c1bcc451396202d62f566026eab8e438c6c91decf8ddf63a6c162619b527a69f4e77e5da47ac3125140c470c71cbca77b5c638c",
|
||||
hashExpected: "4f6ead01278ba4597d4720e37482f585a713497cea994a95209f4c57a963b4a7",
|
||||
sigExpected: "43b5818802a137a72a190c1d8d767ca507f7a4804b1b69b5e055abf31f4f2b476c80bb1ba63260d95610f6f831420d32130e7f22fec5d76e16644ddfcedd0d441c",
|
||||
toHashExpected: "190167617949b934d7e01add4009cd3d47415a26727b7d6288e5dce33fb3721d5a1a9ce511b19b694c9aaf8183f4987ed752f24884c54c003d11daa2e98c7547a79e",
|
||||
hashExpected: "157b570c597e615b8356ce008ac39f43bc9b6d50080bc07d968031b9378acbbb",
|
||||
sigExpected: "a0766181102428b5672e523dc4b905c10ddf025c10dbd0b3534ef864632a14652737610041c670b302fc7dca28edd5d6eac42b72d69ce58da8ce21287b244e381b",
|
||||
}
|
||||
tvs = append(tvs, tv0)
|
||||
tvs = append(tvs, tv1)
|
||||
@@ -122,10 +126,10 @@ func TestAccountCreationAuthJSComp(t *testing.T) {
|
||||
BJJ: pkComp,
|
||||
}
|
||||
|
||||
toHash := a.toHash(chainID, hermezContractAddr)
|
||||
toHash, err := a.toHash(chainID, hermezContractAddr)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tv.toHashExpected,
|
||||
hex.EncodeToString(toHash))
|
||||
assert.Equal(t, 120+len(EthMsgPrefix)+len([]byte("120")), len(toHash))
|
||||
|
||||
msg, err := a.HashToSign(chainID, hermezContractAddr)
|
||||
require.NoError(t, err)
|
||||
|
||||
@@ -13,8 +13,9 @@ const batchNumBytesLen = 8
|
||||
|
||||
// Batch is a struct that represents Hermez network batch
|
||||
type Batch struct {
|
||||
BatchNum BatchNum `meddler:"batch_num"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum block in which the batch is forged
|
||||
BatchNum BatchNum `meddler:"batch_num"`
|
||||
// Ethereum block in which the batch is forged
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
ForgerAddr ethCommon.Address `meddler:"forger_addr"`
|
||||
CollectedFees map[TokenID]*big.Int `meddler:"fees_collected,json"`
|
||||
FeeIdxsCoordinator []Idx `meddler:"fee_idxs_coordinator,json"`
|
||||
@@ -22,9 +23,29 @@ type Batch struct {
|
||||
NumAccounts int `meddler:"num_accounts"`
|
||||
LastIdx int64 `meddler:"last_idx"`
|
||||
ExitRoot *big.Int `meddler:"exit_root,bigint"`
|
||||
ForgeL1TxsNum *int64 `meddler:"forge_l1_txs_num"` // optional, Only when the batch forges L1 txs. Identifier that corresponds to the group of L1 txs forged in the current batch.
|
||||
SlotNum int64 `meddler:"slot_num"` // Slot in which the batch is forged
|
||||
TotalFeesUSD *float64 `meddler:"total_fees_usd"`
|
||||
// ForgeL1TxsNum is optional, Only when the batch forges L1 txs. Identifier that corresponds
|
||||
// to the group of L1 txs forged in the current batch.
|
||||
ForgeL1TxsNum *int64 `meddler:"forge_l1_txs_num"`
|
||||
SlotNum int64 `meddler:"slot_num"` // Slot in which the batch is forged
|
||||
TotalFeesUSD *float64 `meddler:"total_fees_usd"`
|
||||
}
|
||||
|
||||
// NewEmptyBatch creates a new empty batch
|
||||
func NewEmptyBatch() *Batch {
|
||||
return &Batch{
|
||||
BatchNum: 0,
|
||||
EthBlockNum: 0,
|
||||
ForgerAddr: ethCommon.Address{},
|
||||
CollectedFees: make(map[TokenID]*big.Int),
|
||||
FeeIdxsCoordinator: make([]Idx, 0),
|
||||
StateRoot: big.NewInt(0),
|
||||
NumAccounts: 0,
|
||||
LastIdx: 0,
|
||||
ExitRoot: big.NewInt(0),
|
||||
ForgeL1TxsNum: nil,
|
||||
SlotNum: 0,
|
||||
TotalFeesUSD: nil,
|
||||
}
|
||||
}
|
||||
|
||||
// BatchNum identifies a batch
|
||||
@@ -45,7 +66,9 @@ func (bn BatchNum) BigInt() *big.Int {
|
||||
// BatchNumFromBytes returns BatchNum from a []byte
|
||||
func BatchNumFromBytes(b []byte) (BatchNum, error) {
|
||||
if len(b) != batchNumBytesLen {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse BatchNumFromBytes, bytes len %d, expected %d", len(b), batchNumBytesLen))
|
||||
return 0,
|
||||
tracerr.Wrap(fmt.Errorf("can not parse BatchNumFromBytes, bytes len %d, expected %d",
|
||||
len(b), batchNumBytesLen))
|
||||
}
|
||||
batchNum := binary.BigEndian.Uint64(b[:batchNumBytesLen])
|
||||
return BatchNum(batchNum), nil
|
||||
@@ -59,6 +82,7 @@ type BatchData struct {
|
||||
L1CoordinatorTxs []L1Tx
|
||||
L2Txs []L2Tx
|
||||
CreatedAccounts []Account
|
||||
UpdatedAccounts []AccountUpdate
|
||||
ExitTree []ExitInfo
|
||||
Batch Batch
|
||||
}
|
||||
|
||||
@@ -34,7 +34,7 @@ type Slot struct {
|
||||
// BatchesLen int
|
||||
BidValue *big.Int
|
||||
BootCoord bool
|
||||
// Bidder, Forer and URL correspond to the winner of the slot (which is
|
||||
// Bidder, Forger and URL correspond to the winner of the slot (which is
|
||||
// not always the highest bidder). These are the values of the
|
||||
// coordinator that is able to forge exclusively before the deadline.
|
||||
Bidder ethCommon.Address
|
||||
|
||||
@@ -5,10 +5,15 @@ import (
|
||||
)
|
||||
|
||||
// Coordinator represents a Hermez network coordinator who wins an auction for an specific slot
|
||||
// WARNING: this is strongly based on the previous implementation, once the new spec is done, this may change a lot.
|
||||
// WARNING: this is strongly based on the previous implementation, once the new spec is done, this
|
||||
// may change a lot.
|
||||
type Coordinator struct {
|
||||
Bidder ethCommon.Address `meddler:"bidder_addr"` // address of the bidder
|
||||
Forger ethCommon.Address `meddler:"forger_addr"` // address of the forger
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // block in which the coordinator was registered
|
||||
URL string `meddler:"url"` // URL of the coordinators API
|
||||
// Bidder is the address of the bidder
|
||||
Bidder ethCommon.Address `meddler:"bidder_addr"`
|
||||
// Forger is the address of the forger
|
||||
Forger ethCommon.Address `meddler:"forger_addr"`
|
||||
// EthBlockNum is the block in which the coordinator was registered
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
// URL of the coordinators API
|
||||
URL string `meddler:"url"`
|
||||
}
|
||||
|
||||
@@ -33,7 +33,8 @@ func (c *AuctionConstants) SlotNum(blockNum int64) int64 {
|
||||
if blockNum >= c.GenesisBlockNum {
|
||||
return (blockNum - c.GenesisBlockNum) / int64(c.BlocksPerSlot)
|
||||
}
|
||||
return -1
|
||||
// This result will be negative
|
||||
return (blockNum - c.GenesisBlockNum) / int64(c.BlocksPerSlot)
|
||||
}
|
||||
|
||||
// SlotBlocks returns the first and the last block numbers included in that slot
|
||||
@@ -67,11 +68,13 @@ type AuctionVariables struct {
|
||||
ClosedAuctionSlots uint16 `meddler:"closed_auction_slots" validate:"required"`
|
||||
// Distance (#slots) to the farthest slot to which you can bid (30 days = 4320 slots )
|
||||
OpenAuctionSlots uint16 `meddler:"open_auction_slots" validate:"required"`
|
||||
// How the HEZ tokens deposited by the slot winner are distributed (Burn: 40% - Donation: 40% - HGT: 20%)
|
||||
// How the HEZ tokens deposited by the slot winner are distributed (Burn: 40% - Donation:
|
||||
// 40% - HGT: 20%)
|
||||
AllocationRatio [3]uint16 `meddler:"allocation_ratio,json" validate:"required"`
|
||||
// Minimum outbid (percentage) over the previous one to consider it valid
|
||||
Outbidding uint16 `meddler:"outbidding" validate:"required"`
|
||||
// Number of blocks at the end of a slot in which any coordinator can forge if the winner has not forged one before
|
||||
// Number of blocks at the end of a slot in which any coordinator can forge if the winner
|
||||
// has not forged one before
|
||||
SlotDeadline uint8 `meddler:"slot_deadline" validate:"required"`
|
||||
}
|
||||
|
||||
|
||||
@@ -20,19 +20,22 @@ const (
|
||||
// RollupConstExitIDx IDX 1 is reserved for exits
|
||||
RollupConstExitIDx = 1
|
||||
// RollupConstLimitTokens Max number of tokens allowed to be registered inside the rollup
|
||||
RollupConstLimitTokens = (1 << 32)
|
||||
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes] compressedSignature
|
||||
RollupConstLimitTokens = (1 << 32) //nolint:gomnd
|
||||
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes]
|
||||
// compressedSignature
|
||||
RollupConstL1CoordinatorTotalBytes = 101
|
||||
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx +
|
||||
// [2 bytes] depositAmountFloat16 + [2 bytes] amountFloat16 + [4 bytes] tokenId + [6 bytes] toIdx
|
||||
RollupConstL1UserTotalBytes = 72
|
||||
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6
|
||||
// bytes] fromIdx + [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes]
|
||||
// tokenId + [6 bytes] toIdx
|
||||
RollupConstL1UserTotalBytes = 78
|
||||
// RollupConstMaxL1UserTx Maximum L1-user transactions allowed to be queued in a batch
|
||||
RollupConstMaxL1UserTx = 128
|
||||
// RollupConstMaxL1Tx Maximum L1 transactions allowed to be queued in a batch
|
||||
RollupConstMaxL1Tx = 256
|
||||
// RollupConstInputSHAConstantBytes [6 bytes] lastIdx + [6 bytes] newLastIdx + [32 bytes] stateRoot + [32 bytes] newStRoot + [32 bytes] newExitRoot +
|
||||
// [_MAX_L1_TX * _L1_USER_TOTALBYTES bytes] l1TxsData + totalL2TxsDataLength + feeIdxCoordinatorLength + [2 bytes] chainID =
|
||||
// 18542 bytes + totalL2TxsDataLength + feeIdxCoordinatorLength
|
||||
// RollupConstInputSHAConstantBytes [6 bytes] lastIdx + [6 bytes] newLastIdx + [32 bytes]
|
||||
// stateRoot + [32 bytes] newStRoot + [32 bytes] newExitRoot + [_MAX_L1_TX *
|
||||
// _L1_USER_TOTALBYTES bytes] l1TxsData + totalL2TxsDataLength + feeIdxCoordinatorLength +
|
||||
// [2 bytes] chainID = 18542 bytes + totalL2TxsDataLength + feeIdxCoordinatorLength
|
||||
RollupConstInputSHAConstantBytes = 18546
|
||||
// RollupConstNumBuckets Number of buckets
|
||||
RollupConstNumBuckets = 5
|
||||
@@ -44,14 +47,18 @@ const (
|
||||
|
||||
var (
|
||||
// RollupConstLimitDepositAmount Max deposit amount allowed (depositAmount: L1 --> L2)
|
||||
RollupConstLimitDepositAmount, _ = new(big.Int).SetString("340282366920938463463374607431768211456", 10)
|
||||
RollupConstLimitDepositAmount, _ = new(big.Int).SetString(
|
||||
"340282366920938463463374607431768211456", 10)
|
||||
// RollupConstLimitL2TransferAmount Max amount allowed (amount L2 --> L2)
|
||||
RollupConstLimitL2TransferAmount, _ = new(big.Int).SetString("6277101735386680763835789423207666416102355444464034512896", 10)
|
||||
RollupConstLimitL2TransferAmount, _ = new(big.Int).SetString(
|
||||
"6277101735386680763835789423207666416102355444464034512896", 10)
|
||||
|
||||
// RollupConstEthAddressInternalOnly This ethereum address is used internally for rollup accounts that don't have ethereum address, only Babyjubjub
|
||||
// This non-ethereum accounts can be created by the coordinator and allow users to have a rollup
|
||||
// account without needing an ethereum address
|
||||
RollupConstEthAddressInternalOnly = ethCommon.HexToAddress("0xFFfFfFffFFfffFFfFFfFFFFFffFFFffffFfFFFfF")
|
||||
// RollupConstEthAddressInternalOnly This ethereum address is used internally for rollup
|
||||
// accounts that don't have ethereum address, only Babyjubjub.
|
||||
// This non-ethereum accounts can be created by the coordinator and allow users to have a
|
||||
// rollup account without needing an ethereum address
|
||||
RollupConstEthAddressInternalOnly = ethCommon.HexToAddress(
|
||||
"0xFFfFfFffFFfffFFfFFfFFFFFffFFFffffFfFFFfF")
|
||||
// RollupConstRfield Modulus zkSNARK
|
||||
RollupConstRfield, _ = new(big.Int).SetString(
|
||||
"21888242871839275222246405745257275088548364400416034343698204186575808495617", 10)
|
||||
@@ -63,24 +70,32 @@ var (
|
||||
|
||||
// RollupConstRecipientInterfaceHash ERC777 recipient interface hash
|
||||
RollupConstRecipientInterfaceHash = crypto.Keccak256([]byte("ERC777TokensRecipient"))
|
||||
// RollupConstPerformL1UserTxSignature the signature of the function that can be called thru an ERC777 `send`
|
||||
RollupConstPerformL1UserTxSignature = crypto.Keccak256([]byte("addL1Transaction(uint256,uint48,uint16,uint16,uint32,uint48)"))
|
||||
// RollupConstAddTokenSignature the signature of the function that can be called thru an ERC777 `send`
|
||||
// RollupConstPerformL1UserTxSignature the signature of the function that can be called thru
|
||||
// an ERC777 `send`
|
||||
RollupConstPerformL1UserTxSignature = crypto.Keccak256([]byte(
|
||||
"addL1Transaction(uint256,uint48,uint16,uint16,uint32,uint48)"))
|
||||
// RollupConstAddTokenSignature the signature of the function that can be called thru an
|
||||
// ERC777 `send`
|
||||
RollupConstAddTokenSignature = crypto.Keccak256([]byte("addToken(address)"))
|
||||
// RollupConstSendSignature ERC777 Signature
|
||||
RollupConstSendSignature = crypto.Keccak256([]byte("send(address,uint256,bytes)"))
|
||||
// RollupConstERC777Granularity ERC777 Signature
|
||||
RollupConstERC777Granularity = crypto.Keccak256([]byte("granularity()"))
|
||||
// RollupConstWithdrawalDelayerDeposit This constant are used to deposit tokens from ERC77 tokens into withdrawal delayer
|
||||
// RollupConstWithdrawalDelayerDeposit This constant are used to deposit tokens from ERC77
|
||||
// tokens into withdrawal delayer
|
||||
RollupConstWithdrawalDelayerDeposit = crypto.Keccak256([]byte("deposit(address,address,uint192)"))
|
||||
|
||||
// ERC20 signature
|
||||
|
||||
// RollupConstTransferSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
// RollupConstTransferSignature This constant is used in the _safeTransfer internal method
|
||||
// in order to safe GAS.
|
||||
RollupConstTransferSignature = crypto.Keccak256([]byte("transfer(address,uint256)"))
|
||||
// RollupConstTransferFromSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
RollupConstTransferFromSignature = crypto.Keccak256([]byte("transferFrom(address,address,uint256)"))
|
||||
// RollupConstApproveSignature This constant is used in the _safeTransfer internal method in order to safe GAS.
|
||||
// RollupConstTransferFromSignature This constant is used in the _safeTransfer internal
|
||||
// method in order to safe GAS.
|
||||
RollupConstTransferFromSignature = crypto.Keccak256([]byte(
|
||||
"transferFrom(address,address,uint256)"))
|
||||
// RollupConstApproveSignature This constant is used in the _safeTransfer internal method in
|
||||
// order to safe GAS.
|
||||
RollupConstApproveSignature = crypto.Keccak256([]byte("approve(address,uint256)"))
|
||||
// RollupConstERC20Signature ERC20 decimals signature
|
||||
RollupConstERC20Signature = crypto.Keccak256([]byte("decimals()"))
|
||||
@@ -141,6 +156,7 @@ type TokenExchange struct {
|
||||
}
|
||||
|
||||
// RollupVariables are the variables of the Rollup Smart Contract
|
||||
//nolint:lll
|
||||
type RollupVariables struct {
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
FeeAddToken *big.Int `meddler:"fee_add_token,bigint" validate:"required"`
|
||||
|
||||
@@ -27,6 +27,7 @@ type WDelayerEscapeHatchWithdrawal struct {
|
||||
}
|
||||
|
||||
// WDelayerVariables are the variables of the Withdrawal Delayer Smart Contract
|
||||
//nolint:lll
|
||||
type WDelayerVariables struct {
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
// HermezRollupAddress ethCommon.Address `json:"hermezRollupAddress" meddler:"rollup_address"`
|
||||
|
||||
@@ -1,131 +0,0 @@
|
||||
// Package common Float16 provides methods to work with Hermez custom half float
|
||||
// precision, 16 bits, codification internally called Float16 has been adopted
|
||||
// to encode large integers. This is done in order to save bits when L2
|
||||
// transactions are published.
|
||||
//nolint:gomnd
|
||||
package common
|
||||
|
||||
import (
|
||||
"encoding/binary"
|
||||
"errors"
|
||||
"math/big"
|
||||
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
var (
|
||||
// ErrRoundingLoss is used when converted big.Int to Float16 causes rounding loss
|
||||
ErrRoundingLoss = errors.New("input value causes rounding loss")
|
||||
)
|
||||
|
||||
// Float16 represents a float in a 16 bit format
|
||||
type Float16 uint16
|
||||
|
||||
// Bytes return a byte array of length 2 with the Float16 value encoded in BigEndian
|
||||
func (f16 Float16) Bytes() []byte {
|
||||
var b [2]byte
|
||||
binary.BigEndian.PutUint16(b[:], uint16(f16))
|
||||
return b[:]
|
||||
}
|
||||
|
||||
// Float16FromBytes returns a Float16 from a byte array of 2 bytes.
|
||||
func Float16FromBytes(b []byte) *Float16 {
|
||||
f16 := Float16(binary.BigEndian.Uint16(b[:2]))
|
||||
return &f16
|
||||
}
|
||||
|
||||
// BigInt converts the Float16 to a *big.Int integer
|
||||
func (f16 *Float16) BigInt() *big.Int {
|
||||
fl := int64(*f16)
|
||||
|
||||
m := big.NewInt(fl & 0x3FF)
|
||||
e := big.NewInt(fl >> 11)
|
||||
e5 := (fl >> 10) & 0x01
|
||||
|
||||
exp := big.NewInt(0).Exp(big.NewInt(10), e, nil)
|
||||
res := m.Mul(m, exp)
|
||||
|
||||
if e5 != 0 && e.Cmp(big.NewInt(0)) != 0 {
|
||||
res.Add(res, exp.Div(exp, big.NewInt(2)))
|
||||
}
|
||||
return res
|
||||
}
|
||||
|
||||
// floorFix2Float converts a fix to a float, always rounding down
|
||||
func floorFix2Float(_f *big.Int) Float16 {
|
||||
zero := big.NewInt(0)
|
||||
ten := big.NewInt(10)
|
||||
e := int64(0)
|
||||
|
||||
m := big.NewInt(0)
|
||||
m.Set(_f)
|
||||
|
||||
if m.Cmp(zero) == 0 {
|
||||
return 0
|
||||
}
|
||||
|
||||
s := big.NewInt(0).Rsh(m, 10)
|
||||
|
||||
for s.Cmp(zero) != 0 {
|
||||
m.Div(m, ten)
|
||||
s.Rsh(m, 10)
|
||||
e++
|
||||
}
|
||||
|
||||
return Float16(m.Int64() | e<<11)
|
||||
}
|
||||
|
||||
// NewFloat16 encodes a *big.Int integer as a Float16, returning error in
|
||||
// case of loss during the encoding.
|
||||
func NewFloat16(f *big.Int) (Float16, error) {
|
||||
fl1 := floorFix2Float(f)
|
||||
fi1 := fl1.BigInt()
|
||||
fl2 := fl1 | 0x400
|
||||
fi2 := fl2.BigInt()
|
||||
|
||||
m3 := (fl1 & 0x3FF) + 1
|
||||
e3 := fl1 >> 11
|
||||
|
||||
if m3&0x400 == 0 {
|
||||
m3 = 0x66
|
||||
e3++
|
||||
}
|
||||
|
||||
fl3 := m3 + e3<<11
|
||||
fi3 := fl3.BigInt()
|
||||
|
||||
res := fl1
|
||||
|
||||
d := big.NewInt(0).Abs(fi1.Sub(fi1, f))
|
||||
d2 := big.NewInt(0).Abs(fi2.Sub(fi2, f))
|
||||
|
||||
if d.Cmp(d2) == 1 {
|
||||
res = fl2
|
||||
d = d2
|
||||
}
|
||||
|
||||
d3 := big.NewInt(0).Abs(fi3.Sub(fi3, f))
|
||||
|
||||
if d.Cmp(d3) == 1 {
|
||||
res = fl3
|
||||
}
|
||||
|
||||
// Do rounding check
|
||||
if res.BigInt().Cmp(f) == 0 {
|
||||
return res, nil
|
||||
}
|
||||
return res, tracerr.Wrap(ErrRoundingLoss)
|
||||
}
|
||||
|
||||
// NewFloat16Floor encodes a big.Int integer as a Float16, rounding down in
|
||||
// case of loss during the encoding.
|
||||
func NewFloat16Floor(f *big.Int) Float16 {
|
||||
fl1 := floorFix2Float(f)
|
||||
fl2 := fl1 | 0x400
|
||||
fi2 := fl2.BigInt()
|
||||
|
||||
if fi2.Cmp(f) < 1 {
|
||||
return fl2
|
||||
}
|
||||
return fl1
|
||||
}
|
||||
@@ -1,132 +0,0 @@
|
||||
package common
|
||||
|
||||
import (
|
||||
"math/big"
|
||||
"testing"
|
||||
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/stretchr/testify/assert"
|
||||
)
|
||||
|
||||
func TestConversions(t *testing.T) {
|
||||
testVector := map[Float16]string{
|
||||
0x307B: "123000000",
|
||||
0x1DC6: "454500",
|
||||
0xFFFF: "10235000000000000000000000000000000",
|
||||
0x0000: "0",
|
||||
0x0400: "0",
|
||||
0x0001: "1",
|
||||
0x0401: "1",
|
||||
0x0800: "0",
|
||||
0x0c00: "5",
|
||||
0x0801: "10",
|
||||
0x0c01: "15",
|
||||
}
|
||||
|
||||
for test := range testVector {
|
||||
fix := test.BigInt()
|
||||
|
||||
assert.Equal(t, fix.String(), testVector[test])
|
||||
|
||||
bi := big.NewInt(0)
|
||||
bi.SetString(testVector[test], 10)
|
||||
|
||||
fl, err := NewFloat16(bi)
|
||||
assert.Equal(t, nil, err)
|
||||
|
||||
fx2 := fl.BigInt()
|
||||
assert.Equal(t, fx2.String(), testVector[test])
|
||||
}
|
||||
}
|
||||
|
||||
func TestFloorFix2Float(t *testing.T) {
|
||||
testVector := map[string]Float16{
|
||||
"87999990000000000": 0x776f,
|
||||
"87950000000000001": 0x776f,
|
||||
"87950000000000000": 0x776f,
|
||||
"87949999999999999": 0x736f,
|
||||
}
|
||||
|
||||
for test := range testVector {
|
||||
bi := big.NewInt(0)
|
||||
bi.SetString(test, 10)
|
||||
|
||||
testFloat := NewFloat16Floor(bi)
|
||||
|
||||
assert.Equal(t, testFloat, testVector[test])
|
||||
}
|
||||
}
|
||||
|
||||
func TestConversionLosses(t *testing.T) {
|
||||
a := big.NewInt(1000)
|
||||
b, err := NewFloat16(a)
|
||||
assert.Equal(t, nil, err)
|
||||
c := b.BigInt()
|
||||
assert.Equal(t, c, a)
|
||||
|
||||
a = big.NewInt(1024)
|
||||
b, err = NewFloat16(a)
|
||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
||||
c = b.BigInt()
|
||||
assert.NotEqual(t, c, a)
|
||||
|
||||
a = big.NewInt(32767)
|
||||
b, err = NewFloat16(a)
|
||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
||||
c = b.BigInt()
|
||||
assert.NotEqual(t, c, a)
|
||||
|
||||
a = big.NewInt(32768)
|
||||
b, err = NewFloat16(a)
|
||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
||||
c = b.BigInt()
|
||||
assert.NotEqual(t, c, a)
|
||||
|
||||
a = big.NewInt(65536000)
|
||||
b, err = NewFloat16(a)
|
||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
||||
c = b.BigInt()
|
||||
assert.NotEqual(t, c, a)
|
||||
}
|
||||
|
||||
func BenchmarkFloat16(b *testing.B) {
|
||||
newBigInt := func(s string) *big.Int {
|
||||
bigInt, ok := new(big.Int).SetString(s, 10)
|
||||
if !ok {
|
||||
panic("Bad big int")
|
||||
}
|
||||
return bigInt
|
||||
}
|
||||
type pair struct {
|
||||
Float16 Float16
|
||||
BigInt *big.Int
|
||||
}
|
||||
testVector := []pair{
|
||||
{0x307B, newBigInt("123000000")},
|
||||
{0x1DC6, newBigInt("454500")},
|
||||
{0xFFFF, newBigInt("10235000000000000000000000000000000")},
|
||||
{0x0000, newBigInt("0")},
|
||||
{0x0400, newBigInt("0")},
|
||||
{0x0001, newBigInt("1")},
|
||||
{0x0401, newBigInt("1")},
|
||||
{0x0800, newBigInt("0")},
|
||||
{0x0c00, newBigInt("5")},
|
||||
{0x0801, newBigInt("10")},
|
||||
{0x0c01, newBigInt("15")},
|
||||
}
|
||||
b.Run("floorFix2Float()", func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
NewFloat16Floor(testVector[i%len(testVector)].BigInt)
|
||||
}
|
||||
})
|
||||
b.Run("NewFloat16()", func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
_, _ = NewFloat16(testVector[i%len(testVector)].BigInt)
|
||||
}
|
||||
})
|
||||
b.Run("Float16.BigInt()", func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
testVector[i%len(testVector)].Float16.BigInt()
|
||||
}
|
||||
})
|
||||
}
|
||||
128
common/float40.go
Normal file
128
common/float40.go
Normal file
@@ -0,0 +1,128 @@
|
||||
// Package common float40.go provides methods to work with Hermez custom half
|
||||
// float precision, 40 bits, codification internally called Float40 has been
|
||||
// adopted to encode large integers. This is done in order to save bits when L2
|
||||
// transactions are published.
|
||||
//nolint:gomnd
|
||||
package common
|
||||
|
||||
import (
|
||||
"encoding/binary"
|
||||
"errors"
|
||||
"math/big"
|
||||
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
const (
|
||||
// maxFloat40Value is the maximum value that the Float40 can have
|
||||
// (40 bits: maxFloat40Value=2**40-1)
|
||||
maxFloat40Value = 0xffffffffff
|
||||
// Float40BytesLength defines the length of the Float40 values
|
||||
// represented as byte arrays
|
||||
Float40BytesLength = 5
|
||||
)
|
||||
|
||||
var (
|
||||
// ErrFloat40Overflow is used when a given Float40 overflows the
|
||||
// maximum capacity of the Float40 (2**40-1)
|
||||
ErrFloat40Overflow = errors.New("Float40 overflow, max value: 2**40 -1")
|
||||
// ErrFloat40E31 is used when the e > 31 when trying to convert a
|
||||
// *big.Int to Float40
|
||||
ErrFloat40E31 = errors.New("Float40 error, e > 31")
|
||||
// ErrFloat40NotEnoughPrecission is used when the given *big.Int can
|
||||
// not be represented as Float40 due not enough precission
|
||||
ErrFloat40NotEnoughPrecission = errors.New("Float40 error, not enough precission")
|
||||
|
||||
thres = big.NewInt(0x08_00_00_00_00)
|
||||
)
|
||||
|
||||
// Float40 represents a float in a 64 bit format
|
||||
type Float40 uint64
|
||||
|
||||
// Bytes return a byte array of length 5 with the Float40 value encoded in
|
||||
// BigEndian
|
||||
func (f40 Float40) Bytes() ([]byte, error) {
|
||||
if f40 > maxFloat40Value {
|
||||
return []byte{}, tracerr.Wrap(ErrFloat40Overflow)
|
||||
}
|
||||
|
||||
var f40Bytes [8]byte
|
||||
binary.BigEndian.PutUint64(f40Bytes[:], uint64(f40))
|
||||
var b [5]byte
|
||||
copy(b[:], f40Bytes[3:])
|
||||
return b[:], nil
|
||||
}
|
||||
|
||||
// Float40FromBytes returns a Float40 from a byte array of 5 bytes in Bigendian
|
||||
// representation.
|
||||
func Float40FromBytes(b []byte) Float40 {
|
||||
var f40Bytes [8]byte
|
||||
copy(f40Bytes[3:], b[:])
|
||||
f40 := binary.BigEndian.Uint64(f40Bytes[:])
|
||||
return Float40(f40)
|
||||
}
|
||||
|
||||
// BigInt converts the Float40 to a *big.Int v, where v = m * 10^e, being:
|
||||
// [ e | m ]
|
||||
// [ 5 bits | 35 bits ]
|
||||
func (f40 Float40) BigInt() (*big.Int, error) {
|
||||
// take the 5 used bytes (FF * 5)
|
||||
var f40Uint64 uint64 = uint64(f40) & 0x00_00_00_FF_FF_FF_FF_FF
|
||||
f40Bytes, err := f40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
e := f40Bytes[0] & 0xF8 >> 3 // take first 5 bits
|
||||
m := f40Uint64 & 0x07_FF_FF_FF_FF // take the others 35 bits
|
||||
|
||||
exp := new(big.Int).Exp(big.NewInt(10), big.NewInt(int64(e)), nil)
|
||||
r := new(big.Int).Mul(big.NewInt(int64(m)), exp)
|
||||
return r, nil
|
||||
}
|
||||
|
||||
// NewFloat40 encodes a *big.Int integer as a Float40, returning error in case
|
||||
// of loss during the encoding.
|
||||
func NewFloat40(f *big.Int) (Float40, error) {
|
||||
m := f
|
||||
e := big.NewInt(0)
|
||||
zero := big.NewInt(0)
|
||||
ten := big.NewInt(10)
|
||||
for new(big.Int).Mod(m, ten).Cmp(zero) == 0 && m.Cmp(thres) >= 0 {
|
||||
m = new(big.Int).Div(m, ten)
|
||||
e = new(big.Int).Add(e, big.NewInt(1))
|
||||
}
|
||||
if e.Int64() > 31 {
|
||||
return 0, tracerr.Wrap(ErrFloat40E31)
|
||||
}
|
||||
if m.Cmp(thres) >= 0 {
|
||||
return 0, tracerr.Wrap(ErrFloat40NotEnoughPrecission)
|
||||
}
|
||||
r := new(big.Int).Add(m,
|
||||
new(big.Int).Mul(e, thres))
|
||||
return Float40(r.Uint64()), nil
|
||||
}
|
||||
|
||||
// NewFloat40Floor encodes a *big.Int integer as a Float40, rounding down in
|
||||
// case of loss during the encoding. It returns an error in case that the number
|
||||
// is too big (e>31). Warning: this method should not be used inside the
|
||||
// hermez-node, it's a helper for external usage to generate valid Float40
|
||||
// values.
|
||||
func NewFloat40Floor(f *big.Int) (Float40, error) {
|
||||
m := f
|
||||
e := big.NewInt(0)
|
||||
// zero := big.NewInt(0)
|
||||
ten := big.NewInt(10)
|
||||
for m.Cmp(thres) >= 0 {
|
||||
m = new(big.Int).Div(m, ten)
|
||||
e = new(big.Int).Add(e, big.NewInt(1))
|
||||
}
|
||||
if e.Int64() > 31 {
|
||||
return 0, tracerr.Wrap(ErrFloat40E31)
|
||||
}
|
||||
|
||||
r := new(big.Int).Add(m,
|
||||
new(big.Int).Mul(e, thres))
|
||||
|
||||
return Float40(r.Uint64()), nil
|
||||
}
|
||||
146
common/float40_test.go
Normal file
146
common/float40_test.go
Normal file
@@ -0,0 +1,146 @@
|
||||
package common
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"math/big"
|
||||
"testing"
|
||||
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/stretchr/testify/assert"
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func TestConversionsFloat40(t *testing.T) {
|
||||
testVector := map[Float40]string{
|
||||
6*0x800000000 + 123: "123000000",
|
||||
2*0x800000000 + 4545: "454500",
|
||||
30*0x800000000 + 10235: "10235000000000000000000000000000000",
|
||||
0x000000000: "0",
|
||||
0x800000000: "0",
|
||||
0x0001: "1",
|
||||
0x0401: "1025",
|
||||
0x800000000 + 1: "10",
|
||||
0xFFFFFFFFFF: "343597383670000000000000000000000000000000",
|
||||
}
|
||||
|
||||
for test := range testVector {
|
||||
fix, err := test.BigInt()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, fix.String(), testVector[test])
|
||||
|
||||
bi, ok := new(big.Int).SetString(testVector[test], 10)
|
||||
require.True(t, ok)
|
||||
|
||||
fl, err := NewFloat40(bi)
|
||||
assert.NoError(t, err)
|
||||
|
||||
fx2, err := fl.BigInt()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, fx2.String(), testVector[test])
|
||||
}
|
||||
}
|
||||
|
||||
func TestExpectError(t *testing.T) {
|
||||
testVector := map[string]error{
|
||||
"9922334455000000000000000000000000000000": nil,
|
||||
"9922334455000000000000000000000000000001": ErrFloat40NotEnoughPrecission,
|
||||
"9922334454999999999999999999999999999999": ErrFloat40NotEnoughPrecission,
|
||||
"42949672950000000000000000000000000000000": nil,
|
||||
"99223344556573838487575": ErrFloat40NotEnoughPrecission,
|
||||
"992233445500000000000000000000000000000000": ErrFloat40E31,
|
||||
"343597383670000000000000000000000000000000": nil,
|
||||
"343597383680000000000000000000000000000000": ErrFloat40NotEnoughPrecission,
|
||||
"343597383690000000000000000000000000000000": ErrFloat40NotEnoughPrecission,
|
||||
"343597383700000000000000000000000000000000": ErrFloat40E31,
|
||||
}
|
||||
for test := range testVector {
|
||||
bi, ok := new(big.Int).SetString(test, 10)
|
||||
require.True(t, ok)
|
||||
_, err := NewFloat40(bi)
|
||||
assert.Equal(t, testVector[test], tracerr.Unwrap(err))
|
||||
}
|
||||
}
|
||||
|
||||
func TestNewFloat40Floor(t *testing.T) {
|
||||
testVector := map[string][]string{
|
||||
// []int contains [Float40 value, Flot40 Floor value], when
|
||||
// Float40 value is expected to be 0, is because is expected to
|
||||
// be an error
|
||||
"9922334455000000000000000000000000000000": {
|
||||
"1040714485495", "1040714485495", "9922334455000000000000000000000000000000"},
|
||||
"9922334455000000000000000000000000000001": { // Floor [2] will be same as prev line
|
||||
"0", "1040714485495", "9922334455000000000000000000000000000000"},
|
||||
"9922334454999999999999999999999999999999": {
|
||||
"0", "1040714485494", "9922334454000000000000000000000000000000"},
|
||||
"42949672950000000000000000000000000000000": {
|
||||
"1069446856703", "1069446856703", "42949672950000000000000000000000000000000"},
|
||||
"99223344556573838487575": {
|
||||
"0", "456598933239", "99223344550000000000000"},
|
||||
"992233445500000000000000000000000000000000": {
|
||||
"0", "0", "0"}, // e>31, returns 0, err
|
||||
"343597383670000000000000000000000000000000": {
|
||||
"1099511627775", "1099511627775", "343597383670000000000000000000000000000000"},
|
||||
"343597383680000000000000000000000000000000": {
|
||||
"0", "0", "0"}, // e>31, returns 0, err
|
||||
"1157073197879933027": {
|
||||
"0", "286448638922", "1157073197800000000"},
|
||||
}
|
||||
for test := range testVector {
|
||||
bi, ok := new(big.Int).SetString(test, 10)
|
||||
require.True(t, ok)
|
||||
f40, err := NewFloat40(bi)
|
||||
if f40 == 0 {
|
||||
assert.Error(t, err)
|
||||
} else {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
assert.Equal(t, testVector[test][0], fmt.Sprint(uint64(f40)))
|
||||
|
||||
f40, err = NewFloat40Floor(bi)
|
||||
if f40 == 0 {
|
||||
assert.Equal(t, ErrFloat40E31, tracerr.Unwrap(err))
|
||||
} else {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
assert.Equal(t, testVector[test][1], fmt.Sprint(uint64(f40)))
|
||||
|
||||
bi2, err := f40.BigInt()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, fmt.Sprint(testVector[test][2]), bi2.String())
|
||||
}
|
||||
}
|
||||
|
||||
func BenchmarkFloat40(b *testing.B) {
|
||||
newBigInt := func(s string) *big.Int {
|
||||
bigInt, ok := new(big.Int).SetString(s, 10)
|
||||
if !ok {
|
||||
panic("Can not convert string to *big.Int")
|
||||
}
|
||||
return bigInt
|
||||
}
|
||||
type pair struct {
|
||||
Float40 Float40
|
||||
BigInt *big.Int
|
||||
}
|
||||
testVector := []pair{
|
||||
{6*0x800000000 + 123, newBigInt("123000000")},
|
||||
{2*0x800000000 + 4545, newBigInt("454500")},
|
||||
{30*0x800000000 + 10235, newBigInt("10235000000000000000000000000000000")},
|
||||
{0x000000000, newBigInt("0")},
|
||||
{0x800000000, newBigInt("0")},
|
||||
{0x0001, newBigInt("1")},
|
||||
{0x0401, newBigInt("1025")},
|
||||
{0x800000000 + 1, newBigInt("10")},
|
||||
{0xFFFFFFFFFF, newBigInt("343597383670000000000000000000000000000000")},
|
||||
}
|
||||
b.Run("NewFloat40()", func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
_, _ = NewFloat40(testVector[i%len(testVector)].BigInt)
|
||||
}
|
||||
})
|
||||
b.Run("Float40.BigInt()", func(b *testing.B) {
|
||||
for i := 0; i < b.N; i++ {
|
||||
_, _ = testVector[i%len(testVector)].Float40.BigInt()
|
||||
}
|
||||
})
|
||||
}
|
||||
161
common/l1tx.go
161
common/l1tx.go
@@ -11,42 +11,43 @@ import (
|
||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||
)
|
||||
|
||||
const (
|
||||
// L1UserTxBytesLen is the length of the byte array that represents the L1Tx
|
||||
L1UserTxBytesLen = 72
|
||||
// L1CoordinatorTxBytesLen is the length of the byte array that represents the L1CoordinatorTx
|
||||
L1CoordinatorTxBytesLen = 101
|
||||
)
|
||||
|
||||
// L1Tx is a struct that represents a L1 tx
|
||||
type L1Tx struct {
|
||||
// Stored in DB: mandatory fileds
|
||||
|
||||
// TxID (12 bytes) for L1Tx is:
|
||||
// TxID (32 bytes) for L1Tx is the Keccak256 (ethereum) hash of:
|
||||
// bytes: | 1 | 8 | 2 | 1 |
|
||||
// values: | type | ToForgeL1TxsNum | Position | 0 (padding) |
|
||||
// where type:
|
||||
// - L1UserTx: 0
|
||||
// - L1CoordinatorTx: 1
|
||||
TxID TxID `meddler:"id"`
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"` // toForgeL1TxsNum in which the tx was forged / will be forged
|
||||
Position int `meddler:"position"`
|
||||
UserOrigin bool `meddler:"user_origin"` // true if the tx was originated by a user, false if it was aoriginated by a coordinator. Note that this differ from the spec for implementation simplification purpposes
|
||||
FromIdx Idx `meddler:"from_idx,zeroisnull"` // FromIdx is used by L1Tx/Deposit to indicate the Idx receiver of the L1Tx.DepositAmount (deposit)
|
||||
TxID TxID `meddler:"id"`
|
||||
// ToForgeL1TxsNum indicates in which the tx was forged / will be forged
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"`
|
||||
Position int `meddler:"position"`
|
||||
// UserOrigin is set to true if the tx was originated by a user, false if it was
|
||||
// aoriginated by a coordinator. Note that this differ from the spec for implementation
|
||||
// simplification purpposes
|
||||
UserOrigin bool `meddler:"user_origin"`
|
||||
// FromIdx is used by L1Tx/Deposit to indicate the Idx receiver of the L1Tx.DepositAmount
|
||||
// (deposit)
|
||||
FromIdx Idx `meddler:"from_idx,zeroisnull"`
|
||||
EffectiveFromIdx Idx `meddler:"effective_from_idx,zeroisnull"`
|
||||
FromEthAddr ethCommon.Address `meddler:"from_eth_addr,zeroisnull"`
|
||||
FromBJJ babyjub.PublicKeyComp `meddler:"from_bjj,zeroisnull"`
|
||||
ToIdx Idx `meddler:"to_idx"` // ToIdx is ignored in L1Tx/Deposit, but used in the L1Tx/DepositAndTransfer
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
// ToIdx is ignored in L1Tx/Deposit, but used in the L1Tx/DepositAndTransfer
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
// EffectiveAmount only applies to L1UserTx.
|
||||
EffectiveAmount *big.Int `meddler:"effective_amount,bigintnull"`
|
||||
DepositAmount *big.Int `meddler:"deposit_amount,bigint"`
|
||||
// EffectiveDepositAmount only applies to L1UserTx.
|
||||
EffectiveDepositAmount *big.Int `meddler:"effective_deposit_amount,bigintnull"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum Block Number in which this L1Tx was added to the queue
|
||||
Type TxType `meddler:"type"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
EffectiveDepositAmount *big.Int `meddler:"effective_deposit_amount,bigintnull"`
|
||||
// Ethereum Block Number in which this L1Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
Type TxType `meddler:"type"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
}
|
||||
|
||||
// NewL1Tx returns the given L1Tx with the TxId & Type parameters calculated
|
||||
@@ -179,45 +180,38 @@ func (tx L1Tx) Tx() Tx {
|
||||
// [ 8 bits ] empty (userFee) // 1 byte
|
||||
// [ 40 bits ] empty (nonce) // 5 bytes
|
||||
// [ 32 bits ] tokenID // 4 bytes
|
||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
||||
// [ 48 bits ] toIdx // 6 bytes
|
||||
// [ 48 bits ] fromIdx // 6 bytes
|
||||
// [ 16 bits ] chainId // 2 bytes
|
||||
// [ 32 bits ] empty (signatureConstant) // 4 bytes
|
||||
// Total bits compressed data: 241 bits // 31 bytes in *big.Int representation
|
||||
// Total bits compressed data: 225 bits // 29 bytes in *big.Int representation
|
||||
func (tx L1Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var b [31]byte
|
||||
var b [29]byte
|
||||
// b[0:7] empty: no ToBJJSign, no fee, no nonce
|
||||
copy(b[7:11], tx.TokenID.Bytes())
|
||||
copy(b[11:13], amountFloat16.Bytes())
|
||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[13:19], toIdxBytes[:])
|
||||
copy(b[11:17], toIdxBytes[:])
|
||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[19:25], fromIdxBytes[:])
|
||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
||||
copy(b[27:31], SignatureConstantBytes[:])
|
||||
copy(b[17:23], fromIdxBytes[:])
|
||||
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||
copy(b[25:29], SignatureConstantBytes[:])
|
||||
|
||||
bi := new(big.Int).SetBytes(b[:])
|
||||
return bi, nil
|
||||
}
|
||||
|
||||
// BytesDataAvailability encodes a L1Tx into []byte for the Data Availability
|
||||
// [ fromIdx | toIdx | amountFloat16 | Fee ]
|
||||
// [ fromIdx | toIdx | amountFloat40 | Fee ]
|
||||
func (tx *L1Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||
idxLen := nLevels / 8 //nolint:gomnd
|
||||
|
||||
b := make([]byte, ((nLevels*2)+16+8)/8) //nolint:gomnd
|
||||
b := make([]byte, ((nLevels*2)+40+8)/8) //nolint:gomnd
|
||||
|
||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||
if err != nil {
|
||||
@@ -231,13 +225,17 @@ func (tx *L1Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
||||
|
||||
if tx.EffectiveAmount != nil {
|
||||
amountFloat16, err := NewFloat16(tx.EffectiveAmount)
|
||||
amountFloat40, err := NewFloat40(tx.EffectiveAmount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[idxLen*2:idxLen*2+2], amountFloat16.Bytes())
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[idxLen*2:idxLen*2+Float40BytesLength], amountFloat40Bytes)
|
||||
}
|
||||
// fee = 0 (as is L1Tx) b[10:11]
|
||||
// fee = 0 (as is L1Tx)
|
||||
return b[:], nil
|
||||
}
|
||||
|
||||
@@ -247,7 +245,7 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
||||
|
||||
fromIdxBytes := b[0:idxLen]
|
||||
toIdxBytes := b[idxLen : idxLen*2]
|
||||
amountBytes := b[idxLen*2 : idxLen*2+2]
|
||||
amountBytes := b[idxLen*2 : idxLen*2+Float40BytesLength]
|
||||
|
||||
l1tx := L1Tx{}
|
||||
fromIdx, err := IdxFromBytes(ethCommon.LeftPadBytes(fromIdxBytes, 6))
|
||||
@@ -260,8 +258,8 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
l1tx.ToIdx = toIdx
|
||||
l1tx.EffectiveAmount = Float16FromBytes(amountBytes).BigInt()
|
||||
return &l1tx, nil
|
||||
l1tx.EffectiveAmount, err = Float40FromBytes(amountBytes).BigInt()
|
||||
return &l1tx, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// BytesGeneric returns the generic representation of a L1Tx. This method is
|
||||
@@ -269,7 +267,7 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
||||
// the L1TxData for the ZKInputs (at the HashGlobalInputs), using this method
|
||||
// for L1CoordinatorTxs & L1UserTxs (for the ZKInputs case).
|
||||
func (tx *L1Tx) BytesGeneric() ([]byte, error) {
|
||||
var b [L1UserTxBytesLen]byte
|
||||
var b [RollupConstL1UserTotalBytes]byte
|
||||
copy(b[0:20], tx.FromEthAddr.Bytes())
|
||||
if tx.FromBJJ != EmptyBJJComp {
|
||||
pkCompL := tx.FromBJJ
|
||||
@@ -281,22 +279,33 @@ func (tx *L1Tx) BytesGeneric() ([]byte, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[52:58], fromIdxBytes[:])
|
||||
depositAmountFloat16, err := NewFloat16(tx.DepositAmount)
|
||||
|
||||
depositAmountFloat40, err := NewFloat40(tx.DepositAmount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[58:60], depositAmountFloat16.Bytes())
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
depositAmountFloat40Bytes, err := depositAmountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[60:62], amountFloat16.Bytes())
|
||||
copy(b[62:66], tx.TokenID.Bytes())
|
||||
copy(b[58:63], depositAmountFloat40Bytes)
|
||||
|
||||
amountFloat40, err := NewFloat40(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[63:68], amountFloat40Bytes)
|
||||
|
||||
copy(b[68:72], tx.TokenID.Bytes())
|
||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[66:72], toIdxBytes[:])
|
||||
copy(b[72:78], toIdxBytes[:])
|
||||
return b[:], nil
|
||||
}
|
||||
|
||||
@@ -313,7 +322,7 @@ func (tx *L1Tx) BytesCoordinatorTx(compressedSignatureBytes []byte) ([]byte, err
|
||||
if tx.UserOrigin {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Can not calculate BytesCoordinatorTx() for a L1UserTx"))
|
||||
}
|
||||
var b [L1CoordinatorTxBytesLen]byte
|
||||
var b [RollupConstL1CoordinatorTotalBytes]byte
|
||||
v := compressedSignatureBytes[64]
|
||||
s := compressedSignatureBytes[32:64]
|
||||
r := compressedSignatureBytes[0:32]
|
||||
@@ -329,8 +338,10 @@ func (tx *L1Tx) BytesCoordinatorTx(compressedSignatureBytes []byte) ([]byte, err
|
||||
|
||||
// L1UserTxFromBytes decodes a L1Tx from []byte
|
||||
func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||
if len(b) != L1UserTxBytesLen {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d", 68, len(b)))
|
||||
if len(b) != RollupConstL1UserTotalBytes {
|
||||
return nil,
|
||||
tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d",
|
||||
68, len(b)))
|
||||
}
|
||||
|
||||
tx := &L1Tx{
|
||||
@@ -347,13 +358,19 @@ func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tx.FromIdx = fromIdx
|
||||
tx.DepositAmount = Float16FromBytes(b[58:60]).BigInt()
|
||||
tx.Amount = Float16FromBytes(b[60:62]).BigInt()
|
||||
tx.TokenID, err = TokenIDFromBytes(b[62:66])
|
||||
tx.DepositAmount, err = Float40FromBytes(b[58:63]).BigInt()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tx.ToIdx, err = IdxFromBytes(b[66:72])
|
||||
tx.Amount, err = Float40FromBytes(b[63:68]).BigInt()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tx.TokenID, err = TokenIDFromBytes(b[68:72])
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tx.ToIdx, err = IdxFromBytes(b[72:78])
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -361,19 +378,15 @@ func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||
return tx, nil
|
||||
}
|
||||
|
||||
func signHash(data []byte) []byte {
|
||||
msg := fmt.Sprintf("\x19Ethereum Signed Message:\n%d%s", len(data), data)
|
||||
return ethCrypto.Keccak256([]byte(msg))
|
||||
}
|
||||
|
||||
// L1CoordinatorTxFromBytes decodes a L1Tx from []byte
|
||||
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx, error) {
|
||||
if len(b) != L1CoordinatorTxBytesLen {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d", 101, len(b)))
|
||||
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx,
|
||||
error) {
|
||||
if len(b) != RollupConstL1CoordinatorTotalBytes {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d",
|
||||
101, len(b)))
|
||||
}
|
||||
|
||||
bytesMessage := []byte("I authorize this babyjubjub key for hermez rollup account creation")
|
||||
|
||||
tx := &L1Tx{
|
||||
UserOrigin: false,
|
||||
}
|
||||
@@ -394,18 +407,20 @@ func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommo
|
||||
// L1CoordinatorTX ETH
|
||||
// Ethereum adds 27 to v
|
||||
v = b[0] - byte(27) //nolint:gomnd
|
||||
chainIDBytes := ethCommon.LeftPadBytes(chainID.Bytes(), 2)
|
||||
var data []byte
|
||||
data = append(data, bytesMessage...)
|
||||
data = append(data, pkCompB...)
|
||||
data = append(data, chainIDBytes[:]...)
|
||||
data = append(data, hermezAddress.Bytes()...)
|
||||
var signature []byte
|
||||
signature = append(signature, r[:]...)
|
||||
signature = append(signature, s[:]...)
|
||||
signature = append(signature, v)
|
||||
hash := signHash(data)
|
||||
pubKeyBytes, err := ethCrypto.Ecrecover(hash, signature)
|
||||
|
||||
accCreationAuth := AccountCreationAuth{
|
||||
BJJ: tx.FromBJJ,
|
||||
}
|
||||
h, err := accCreationAuth.HashToSign(uint16(chainID.Uint64()), hermezAddress)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
pubKeyBytes, err := ethCrypto.Ecrecover(h, signature)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -29,7 +29,8 @@ func TestNewL1UserTx(t *testing.T) {
|
||||
}
|
||||
l1Tx, err := NewL1Tx(l1Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x00a6cbae3b8661fb75b0919ca6605a02cfb04d9c6dd16870fa0fcdf01befa32768", l1Tx.TxID.String())
|
||||
assert.Equal(t, "0x00a6cbae3b8661fb75b0919ca6605a02cfb04d9c6dd16870fa0fcdf01befa32768",
|
||||
l1Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestNewL1CoordinatorTx(t *testing.T) {
|
||||
@@ -46,68 +47,115 @@ func TestNewL1CoordinatorTx(t *testing.T) {
|
||||
}
|
||||
l1Tx, err := NewL1Tx(l1Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x01274482d73df4dab34a1b6740adfca347a462513aa14e82f27b12f818d1b68c84", l1Tx.TxID.String())
|
||||
assert.Equal(t, "0x01274482d73df4dab34a1b6740adfca347a462513aa14e82f27b12f818d1b68c84",
|
||||
l1Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestL1TxCompressedData(t *testing.T) {
|
||||
// test vectors values generated from javascript implementation (using
|
||||
// PoolL2Tx values)
|
||||
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
tx := L1Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
Amount: big.NewInt(4),
|
||||
TokenID: 5,
|
||||
FromIdx: (1 << 48) - 1,
|
||||
ToIdx: (1 << 48) - 1,
|
||||
Amount: amount,
|
||||
TokenID: (1 << 32) - 1,
|
||||
}
|
||||
chainID := uint16(0)
|
||||
txCompressedData, err := tx.TxCompressedData(chainID)
|
||||
txCompressedData, err := tx.TxCompressedData(uint16((1 << 16) - 1))
|
||||
assert.NoError(t, err)
|
||||
expectedStr := "ffffffffffffffffffffffffffffffffffffc60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
// test vector value generated from javascript implementation
|
||||
expectedStr := "7307597389635308713748674793997299267459594577423"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
assert.Equal(t, "0500040000000000030000000000020000c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
tx = L1Tx{
|
||||
FromIdx: 0,
|
||||
ToIdx: 0,
|
||||
Amount: big.NewInt(0),
|
||||
TokenID: 0,
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||
assert.NoError(t, err)
|
||||
expectedStr = "c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
amount, ok = new(big.Int).SetString("63000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
tx = L1Tx{
|
||||
FromIdx: 324,
|
||||
ToIdx: 256,
|
||||
Amount: amount,
|
||||
TokenID: 123,
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(1))
|
||||
assert.NoError(t, err)
|
||||
expectedStr = "7b0000000001000000000001440001c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
tx = L1Tx{
|
||||
FromIdx: 1,
|
||||
ToIdx: 2,
|
||||
TokenID: 3,
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||
assert.NoError(t, err)
|
||||
expectedStr = "030000000000020000000000010000c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
}
|
||||
|
||||
func TestBytesDataAvailability(t *testing.T) {
|
||||
// test vectors values generated from javascript implementation
|
||||
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
tx := L1Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
Amount: big.NewInt(4),
|
||||
TokenID: 5,
|
||||
ToIdx: (1 << 16) - 1,
|
||||
FromIdx: (1 << 16) - 1,
|
||||
EffectiveAmount: amount,
|
||||
}
|
||||
txCompressedData, err := tx.BytesDataAvailability(32)
|
||||
txCompressedData, err := tx.BytesDataAvailability(16)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0000000200000003000000", hex.EncodeToString(txCompressedData))
|
||||
|
||||
tx = L1Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
EffectiveAmount: big.NewInt(4),
|
||||
TokenID: 5,
|
||||
}
|
||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0000000200000003000400", hex.EncodeToString(txCompressedData))
|
||||
}
|
||||
|
||||
func TestL1TxFromDataAvailability(t *testing.T) {
|
||||
tx := L1Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
Amount: big.NewInt(4),
|
||||
}
|
||||
txCompressedData, err := tx.BytesDataAvailability(32)
|
||||
assert.NoError(t, err)
|
||||
l1tx, err := L1TxFromDataAvailability(txCompressedData, 32)
|
||||
assert.Equal(t, "ffffffffffffffffff00", hex.EncodeToString(txCompressedData))
|
||||
l1tx, err := L1TxFromDataAvailability(txCompressedData, 16)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||
|
||||
tx = L1Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
EffectiveAmount: big.NewInt(4),
|
||||
ToIdx: (1 << 32) - 1,
|
||||
FromIdx: (1 << 32) - 1,
|
||||
EffectiveAmount: amount,
|
||||
}
|
||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "ffffffffffffffffffffffffff00", hex.EncodeToString(txCompressedData))
|
||||
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||
|
||||
tx = L1Tx{
|
||||
ToIdx: 0,
|
||||
FromIdx: 0,
|
||||
EffectiveAmount: big.NewInt(0),
|
||||
}
|
||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0000000000000000000000000000", hex.EncodeToString(txCompressedData))
|
||||
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||
|
||||
tx = L1Tx{
|
||||
ToIdx: 635,
|
||||
FromIdx: 296,
|
||||
EffectiveAmount: big.NewInt(1000000000000000000),
|
||||
}
|
||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "000001280000027b42540be40000", hex.EncodeToString(txCompressedData))
|
||||
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||
@@ -153,7 +201,8 @@ func TestL1userTxByteParsers(t *testing.T) {
|
||||
func TestL1TxByteParsersCompatibility(t *testing.T) {
|
||||
// Data from compatibility test
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompB, err := hex.DecodeString("0dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a")
|
||||
pkCompB, err :=
|
||||
hex.DecodeString("0dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a")
|
||||
require.NoError(t, err)
|
||||
pkCompL := SwapEndianness(pkCompB)
|
||||
err = pkComp.UnmarshalText([]byte(hex.EncodeToString(pkCompL)))
|
||||
@@ -172,20 +221,19 @@ func TestL1TxByteParsersCompatibility(t *testing.T) {
|
||||
UserOrigin: true,
|
||||
}
|
||||
|
||||
expected, err := utils.HexDecode("85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a000001c638db8be880f00020039c0000053cb88d")
|
||||
require.NoError(t, err)
|
||||
|
||||
encodedData, err := l1Tx.BytesUser()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, encodedData)
|
||||
expected := "85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79e" +
|
||||
"e1f41a7def613e73a20c32eece5a000001c638db52540be400459682f0000020039c0000053cb88d"
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
}
|
||||
|
||||
func TestL1CoordinatorTxByteParsers(t *testing.T) {
|
||||
hermezAddress := ethCommon.HexToAddress("0xD6C850aeBFDC46D7F4c207e445cC0d6B0919BDBe")
|
||||
chainID := big.NewInt(1337)
|
||||
chainIDBytes := ethCommon.LeftPadBytes(chainID.Bytes(), 2)
|
||||
|
||||
privateKey, err := crypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
privateKey, err :=
|
||||
crypto.HexToECDSA("fad9c8855b740a0b7ed4c221dbad0f33a83a49cad6b3fe8d5817ac83d38b6a19")
|
||||
require.NoError(t, err)
|
||||
|
||||
publicKey := privateKey.Public()
|
||||
@@ -201,18 +249,16 @@ func TestL1CoordinatorTxByteParsers(t *testing.T) {
|
||||
pkCompL := []byte("56ca90f80d7c374ae7485e9bcc47d4ac399460948da6aeeb899311097925a72c")
|
||||
err = pkComp.UnmarshalText(pkCompL)
|
||||
require.NoError(t, err)
|
||||
bytesMessage1 := []byte("\x19Ethereum Signed Message:\n120")
|
||||
bytesMessage2 := []byte("I authorize this babyjubjub key for hermez rollup account creation")
|
||||
|
||||
babyjubB := SwapEndianness(pkComp[:])
|
||||
var data []byte
|
||||
data = append(data, bytesMessage1...)
|
||||
data = append(data, bytesMessage2...)
|
||||
data = append(data, babyjubB[:]...)
|
||||
data = append(data, chainIDBytes...)
|
||||
data = append(data, hermezAddress.Bytes()...)
|
||||
hash := crypto.Keccak256Hash(data)
|
||||
signature, err := crypto.Sign(hash.Bytes(), privateKey)
|
||||
accCreationAuth := AccountCreationAuth{
|
||||
EthAddr: fromEthAddr,
|
||||
BJJ: pkComp,
|
||||
}
|
||||
|
||||
h, err := accCreationAuth.HashToSign(uint16(chainID.Uint64()), hermezAddress)
|
||||
require.NoError(t, err)
|
||||
|
||||
signature, err := crypto.Sign(h, privateKey)
|
||||
require.NoError(t, err)
|
||||
// Ethereum adds 27 to v
|
||||
v := int(signature[64])
|
||||
@@ -259,7 +305,8 @@ func TestL1CoordinatorTxByteParsersCompatibility(t *testing.T) {
|
||||
signature = append(signature, v[:]...)
|
||||
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompB, err := hex.DecodeString("a2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c")
|
||||
pkCompB, err :=
|
||||
hex.DecodeString("a2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c")
|
||||
require.NoError(t, err)
|
||||
pkCompL := SwapEndianness(pkCompB)
|
||||
err = pkComp.UnmarshalText([]byte(hex.EncodeToString(pkCompL)))
|
||||
@@ -274,7 +321,9 @@ func TestL1CoordinatorTxByteParsersCompatibility(t *testing.T) {
|
||||
encodeData, err := l1Tx.BytesCoordinatorTx(signature)
|
||||
require.NoError(t, err)
|
||||
|
||||
expected, err := utils.HexDecode("1b186d7122ff7f654cfed3156719774898d573900c86599a885a706dbdffe5ea8cda71e5eb097e115405d84d1e7b464009b434b32c014a2df502d1f065ced8bc3ba2c2807ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c000000e7")
|
||||
expected, err := utils.HexDecode("1b186d7122ff7f654cfed3156719774898d573900c86599a885a706" +
|
||||
"dbdffe5ea8cda71e5eb097e115405d84d1e7b464009b434b32c014a2df502d1f065ced8bc3ba2c28" +
|
||||
"07ee39c3b3378738cff85a46a9465bb8fcf44ea597c33da9719be7c259c000000e7")
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, expected, encodeData)
|
||||
|
||||
@@ -10,7 +10,7 @@ import (
|
||||
|
||||
// L2Tx is a struct that represents an already forged L2 tx
|
||||
type L2Tx struct {
|
||||
// Stored in DB: mandatory fileds
|
||||
// Stored in DB: mandatory fields
|
||||
TxID TxID `meddler:"id"`
|
||||
BatchNum BatchNum `meddler:"batch_num"` // batchNum in which this tx was forged.
|
||||
Position int `meddler:"position"`
|
||||
@@ -21,9 +21,10 @@ type L2Tx struct {
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
Fee FeeSelector `meddler:"fee"`
|
||||
// Nonce is filled by the TxProcessor
|
||||
Nonce Nonce `meddler:"nonce"`
|
||||
Type TxType `meddler:"type"`
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // EthereumBlockNumber in which this L2Tx was added to the queue
|
||||
Nonce Nonce `meddler:"nonce"`
|
||||
Type TxType `meddler:"type"`
|
||||
// EthBlockNum in which this L2Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
}
|
||||
|
||||
// NewL2Tx returns the given L2Tx with the TxId & Type parameters calculated
|
||||
@@ -89,11 +90,15 @@ func (tx L2Tx) CalculateTxID() ([TxIDLen]byte, error) {
|
||||
// TokenID
|
||||
b = append(b, tx.TokenID.Bytes()[:]...)
|
||||
// Amount
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
amountFloat40, err := NewFloat40(tx.Amount)
|
||||
if err != nil {
|
||||
return txID, tracerr.Wrap(fmt.Errorf("%s: %d", err, tx.Amount))
|
||||
}
|
||||
b = append(b, amountFloat16.Bytes()...)
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return txID, tracerr.Wrap(err)
|
||||
}
|
||||
b = append(b, amountFloat40Bytes...)
|
||||
// Nonce
|
||||
nonceBytes, err := tx.Nonce.Bytes()
|
||||
if err != nil {
|
||||
@@ -170,11 +175,11 @@ func TxIDsFromL2Txs(txs []L2Tx) []TxID {
|
||||
}
|
||||
|
||||
// BytesDataAvailability encodes a L2Tx into []byte for the Data Availability
|
||||
// [ fromIdx | toIdx | amountFloat16 | Fee ]
|
||||
// [ fromIdx | toIdx | amountFloat40 | Fee ]
|
||||
func (tx L2Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||
idxLen := nLevels / 8 //nolint:gomnd
|
||||
|
||||
b := make([]byte, ((nLevels*2)+16+8)/8) //nolint:gomnd
|
||||
b := make([]byte, ((nLevels*2)+40+8)/8) //nolint:gomnd
|
||||
|
||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||
if err != nil {
|
||||
@@ -188,13 +193,16 @@ func (tx L2Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||
}
|
||||
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
||||
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
amountFloat40, err := NewFloat40(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
copy(b[idxLen*2:idxLen*2+2], amountFloat16.Bytes())
|
||||
b[idxLen*2+2] = byte(tx.Fee)
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[idxLen*2:idxLen*2+Float40BytesLength], amountFloat40Bytes)
|
||||
b[idxLen*2+Float40BytesLength] = byte(tx.Fee)
|
||||
|
||||
return b[:], nil
|
||||
}
|
||||
@@ -219,7 +227,10 @@ func L2TxFromBytesDataAvailability(b []byte, nLevels int) (*L2Tx, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
tx.Amount = Float16FromBytes(b[idxLen*2 : idxLen*2+2]).BigInt()
|
||||
tx.Fee = FeeSelector(b[idxLen*2+2])
|
||||
tx.Amount, err = Float40FromBytes(b[idxLen*2 : idxLen*2+Float40BytesLength]).BigInt()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
tx.Fee = FeeSelector(b[idxLen*2+Float40BytesLength])
|
||||
return tx, nil
|
||||
}
|
||||
|
||||
@@ -19,7 +19,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err := NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02fb52b5d0b9ef2626c11701bb751b2720c76d59946b9a48146ac153bb6e63bf6a", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -30,7 +31,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x0276114a8f666fa1ff7dbf34b4a9da577808dc501e3b2760d01fe3ef5473f5737f", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x029e7499a830f8f5eb17c07da48cf91415710f1bcbe0169d363ff91e81faf92fc2",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -42,7 +44,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x025afb63126d3067f61f633d13e5a51da0551af3a4567a9af2db5321ed04214ff4", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0255c70ed20e1b8935232e1b9c5884dbcc88a6e1a3454d24f2d77252eb2bb0b64e",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 87654,
|
||||
@@ -54,7 +57,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02cf390157041c3b1b59f0aaed4da464f0d0d48f1d026e46fd89c7fe1e5aed7fcf", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0206b372f967061d1148bbcff679de38120e075141a80a07326d0f514c2efc6ca9",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 1,
|
||||
@@ -66,7 +70,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x020ec18eaae67fcd545998841a9c4be09ee3083e12db6ae5e5213a2ecaaa52d5cf", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x0236f7ea5bccf78ba60baf56c058d235a844f9b09259fd0efa4f5f72a7d4a26618",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 999,
|
||||
@@ -78,7 +83,8 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02f036223e79fac776de107f50822552cc964ee9fc4caa304613285f6976bcc940", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x02ac122f5b709ce190129fecbbe35bfd30c70e6433dbd85a8eb743d110906a1dc1",
|
||||
l2Tx.TxID.String())
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
FromIdx: 4444,
|
||||
@@ -90,25 +96,86 @@ func TestNewL2Tx(t *testing.T) {
|
||||
}
|
||||
l2Tx, err = NewL2Tx(l2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x029c8aef9ef24531e4cf84e78cbab1018ba1626a5a10afb6b7c356be1b5c28e92c", l2Tx.TxID.String())
|
||||
assert.Equal(t, "0x02c674951a81881b7bc50db3b9e5efd97ac88550c7426ac548720e5057cfba515a",
|
||||
l2Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestL2TxByteParsers(t *testing.T) {
|
||||
amount := new(big.Int)
|
||||
amount.SetString("79000000", 10)
|
||||
// test vectors values generated from javascript implementation
|
||||
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
l2Tx := &L2Tx{
|
||||
ToIdx: 256,
|
||||
ToIdx: (1 << 16) - 1,
|
||||
FromIdx: (1 << 16) - 1,
|
||||
Amount: amount,
|
||||
FromIdx: 257,
|
||||
Fee: 201,
|
||||
Fee: (1 << 8) - 1,
|
||||
}
|
||||
// Data from the compatibility test
|
||||
expected := "00000101000001002b16c9"
|
||||
encodedData, err := l2Tx.BytesDataAvailability(32)
|
||||
expected := "ffffffffffffffffffff"
|
||||
encodedData, err := l2Tx.BytesDataAvailability(16)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
|
||||
decodedData, err := L2TxFromBytesDataAvailability(encodedData, 32)
|
||||
decodedData, err := L2TxFromBytesDataAvailability(encodedData, 16)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l2Tx, decodedData)
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
ToIdx: (1 << 32) - 1,
|
||||
FromIdx: (1 << 32) - 1,
|
||||
Amount: amount,
|
||||
Fee: (1 << 8) - 1,
|
||||
}
|
||||
expected = "ffffffffffffffffffffffffffff"
|
||||
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
|
||||
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l2Tx, decodedData)
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
ToIdx: 0,
|
||||
FromIdx: 0,
|
||||
Amount: big.NewInt(0),
|
||||
Fee: 0,
|
||||
}
|
||||
expected = "0000000000000000000000000000"
|
||||
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
|
||||
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l2Tx, decodedData)
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
ToIdx: 0,
|
||||
FromIdx: 1061,
|
||||
Amount: big.NewInt(420000000000),
|
||||
Fee: 127,
|
||||
}
|
||||
expected = "000004250000000010fa56ea007f"
|
||||
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
|
||||
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l2Tx, decodedData)
|
||||
|
||||
l2Tx = &L2Tx{
|
||||
ToIdx: 256,
|
||||
FromIdx: 257,
|
||||
Amount: big.NewInt(79000000),
|
||||
Fee: 201,
|
||||
}
|
||||
expected = "00000101000001000004b571c0c9"
|
||||
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||
|
||||
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l2Tx, decodedData)
|
||||
}
|
||||
|
||||
@@ -16,7 +16,8 @@ import (
|
||||
// EmptyBJJComp contains the 32 byte array of a empty BabyJubJub PublicKey
|
||||
// Compressed. It is a valid point in the BabyJubJub curve, so does not give
|
||||
// errors when being decompressed.
|
||||
var EmptyBJJComp = babyjub.PublicKeyComp([32]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
|
||||
var EmptyBJJComp = babyjub.PublicKeyComp([32]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0})
|
||||
|
||||
// PoolL2Tx is a struct that represents a L2Tx sent by an account to the
|
||||
// coordinator that is waiting to be forged
|
||||
@@ -36,7 +37,7 @@ type PoolL2Tx struct {
|
||||
ToEthAddr ethCommon.Address `meddler:"to_eth_addr,zeroisnull"`
|
||||
ToBJJ babyjub.PublicKeyComp `meddler:"to_bjj,zeroisnull"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
Amount *big.Int `meddler:"amount,bigint"` // TODO: change to float16
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
Fee FeeSelector `meddler:"fee"`
|
||||
Nonce Nonce `meddler:"nonce"` // effective 40 bits used
|
||||
State PoolL2TxState `meddler:"state"`
|
||||
@@ -53,7 +54,7 @@ type PoolL2Tx struct {
|
||||
RqToEthAddr ethCommon.Address `meddler:"rq_to_eth_addr,zeroisnull"`
|
||||
RqToBJJ babyjub.PublicKeyComp `meddler:"rq_to_bjj,zeroisnull"`
|
||||
RqTokenID TokenID `meddler:"rq_token_id,zeroisnull"`
|
||||
RqAmount *big.Int `meddler:"rq_amount,bigintnull"` // TODO: change to float16
|
||||
RqAmount *big.Int `meddler:"rq_amount,bigintnull"`
|
||||
RqFee FeeSelector `meddler:"rq_fee,zeroisnull"`
|
||||
RqNonce Nonce `meddler:"rq_nonce,zeroisnull"` // effective 48 bits used
|
||||
AbsoluteFee float64 `meddler:"fee_usd,zeroisnull"`
|
||||
@@ -73,7 +74,7 @@ func NewPoolL2Tx(tx *PoolL2Tx) (*PoolL2Tx, error) {
|
||||
// If original Type doesn't match the correct one, return error
|
||||
if txTypeOld != "" && txTypeOld != tx.Type {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("L2Tx.Type: %s, should be: %s",
|
||||
tx.Type, txTypeOld))
|
||||
txTypeOld, tx.Type))
|
||||
}
|
||||
|
||||
txIDOld := tx.TxID
|
||||
@@ -83,7 +84,7 @@ func NewPoolL2Tx(tx *PoolL2Tx) (*PoolL2Tx, error) {
|
||||
// If original TxID doesn't match the correct one, return error
|
||||
if txIDOld != (TxID{}) && txIDOld != tx.TxID {
|
||||
return tx, tracerr.Wrap(fmt.Errorf("PoolL2Tx.TxID: %s, should be: %s",
|
||||
tx.TxID.String(), txIDOld.String()))
|
||||
txIDOld.String(), tx.TxID.String()))
|
||||
}
|
||||
|
||||
return tx, nil
|
||||
@@ -100,6 +101,8 @@ func (tx *PoolL2Tx) SetType() error {
|
||||
tx.Type = TxTypeTransferToBJJ
|
||||
} else if tx.ToEthAddr != FFAddr && tx.ToEthAddr != EmptyAddr {
|
||||
tx.Type = TxTypeTransferToEthAddr
|
||||
} else {
|
||||
return tracerr.Wrap(errors.New("malformed transaction"))
|
||||
}
|
||||
} else {
|
||||
return tracerr.Wrap(errors.New("malformed transaction"))
|
||||
@@ -122,18 +125,13 @@ func (tx *PoolL2Tx) SetID() error {
|
||||
// [ 8 bits ] userFee // 1 byte
|
||||
// [ 40 bits ] nonce // 5 bytes
|
||||
// [ 32 bits ] tokenID // 4 bytes
|
||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
||||
// [ 48 bits ] toIdx // 6 bytes
|
||||
// [ 48 bits ] fromIdx // 6 bytes
|
||||
// [ 16 bits ] chainId // 2 bytes
|
||||
// [ 32 bits ] signatureConstant // 4 bytes
|
||||
// Total bits compressed data: 241 bits // 31 bytes in *big.Int representation
|
||||
// Total bits compressed data: 225 bits // 29 bytes in *big.Int representation
|
||||
func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
var b [31]byte
|
||||
var b [29]byte
|
||||
|
||||
toBJJSign := byte(0)
|
||||
pkSign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
||||
@@ -149,19 +147,18 @@ func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||
}
|
||||
copy(b[2:7], nonceBytes[:])
|
||||
copy(b[7:11], tx.TokenID.Bytes())
|
||||
copy(b[11:13], amountFloat16.Bytes())
|
||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[13:19], toIdxBytes[:])
|
||||
copy(b[11:17], toIdxBytes[:])
|
||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[19:25], fromIdxBytes[:])
|
||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
||||
copy(b[27:31], SignatureConstantBytes[:])
|
||||
copy(b[17:23], fromIdxBytes[:])
|
||||
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||
copy(b[25:29], SignatureConstantBytes[:])
|
||||
|
||||
bi := new(big.Int).SetBytes(b[:])
|
||||
return bi, nil
|
||||
@@ -170,9 +167,9 @@ func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||
// TxCompressedDataEmpty calculates the TxCompressedData of an empty
|
||||
// transaction
|
||||
func TxCompressedDataEmpty(chainID uint16) *big.Int {
|
||||
var b [31]byte
|
||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
||||
copy(b[27:31], SignatureConstantBytes[:])
|
||||
var b [29]byte
|
||||
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||
copy(b[25:29], SignatureConstantBytes[:])
|
||||
bi := new(big.Int).SetBytes(b[:])
|
||||
return bi
|
||||
}
|
||||
@@ -182,19 +179,24 @@ func TxCompressedDataEmpty(chainID uint16) *big.Int {
|
||||
// [ 8 bits ] userFee // 1 byte
|
||||
// [ 40 bits ] nonce // 5 bytes
|
||||
// [ 32 bits ] tokenID // 4 bytes
|
||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
||||
// [ 40 bits ] amountFloat40 // 5 bytes
|
||||
// [ 48 bits ] toIdx // 6 bytes
|
||||
// [ 48 bits ] fromIdx // 6 bytes
|
||||
// Total bits compressed data: 193 bits // 25 bytes in *big.Int representation
|
||||
// Total bits compressed data: 217 bits // 28 bytes in *big.Int representation
|
||||
func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
||||
if tx.Amount == nil {
|
||||
tx.Amount = big.NewInt(0)
|
||||
}
|
||||
amountFloat16, err := NewFloat16(tx.Amount)
|
||||
amountFloat40, err := NewFloat40(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
var b [25]byte
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var b [28]byte
|
||||
toBJJSign := byte(0)
|
||||
if tx.ToBJJ != EmptyBJJComp {
|
||||
sign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
||||
@@ -210,17 +212,17 @@ func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
||||
}
|
||||
copy(b[2:7], nonceBytes[:])
|
||||
copy(b[7:11], tx.TokenID.Bytes())
|
||||
copy(b[11:13], amountFloat16.Bytes())
|
||||
copy(b[11:16], amountFloat40Bytes)
|
||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[13:19], toIdxBytes[:])
|
||||
copy(b[16:22], toIdxBytes[:])
|
||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[19:25], fromIdxBytes[:])
|
||||
copy(b[22:28], fromIdxBytes[:])
|
||||
|
||||
bi := new(big.Int).SetBytes(b[:])
|
||||
return bi, nil
|
||||
@@ -236,19 +238,24 @@ func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
||||
// [ 8 bits ] rqUserFee // 1 byte
|
||||
// [ 40 bits ] rqNonce // 5 bytes
|
||||
// [ 32 bits ] rqTokenID // 4 bytes
|
||||
// [ 16 bits ] rqAmountFloat16 // 2 bytes
|
||||
// [ 40 bits ] rqAmountFloat40 // 5 bytes
|
||||
// [ 48 bits ] rqToIdx // 6 bytes
|
||||
// [ 48 bits ] rqFromIdx // 6 bytes
|
||||
// Total bits compressed data: 193 bits // 25 bytes in *big.Int representation
|
||||
// Total bits compressed data: 217 bits // 28 bytes in *big.Int representation
|
||||
func (tx *PoolL2Tx) RqTxCompressedDataV2() (*big.Int, error) {
|
||||
if tx.RqAmount == nil {
|
||||
tx.RqAmount = big.NewInt(0)
|
||||
}
|
||||
amountFloat16, err := NewFloat16(tx.RqAmount)
|
||||
amountFloat40, err := NewFloat40(tx.RqAmount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
var b [25]byte
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var b [28]byte
|
||||
rqToBJJSign := byte(0)
|
||||
if tx.RqToBJJ != EmptyBJJComp {
|
||||
sign, _ := babyjub.UnpackSignY(tx.RqToBJJ)
|
||||
@@ -264,17 +271,17 @@ func (tx *PoolL2Tx) RqTxCompressedDataV2() (*big.Int, error) {
|
||||
}
|
||||
copy(b[2:7], nonceBytes[:])
|
||||
copy(b[7:11], tx.RqTokenID.Bytes())
|
||||
copy(b[11:13], amountFloat16.Bytes())
|
||||
copy(b[11:16], amountFloat40Bytes)
|
||||
toIdxBytes, err := tx.RqToIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[13:19], toIdxBytes[:])
|
||||
copy(b[16:22], toIdxBytes[:])
|
||||
fromIdxBytes, err := tx.RqFromIdx.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(b[19:25], fromIdxBytes[:])
|
||||
copy(b[22:28], fromIdxBytes[:])
|
||||
|
||||
bi := new(big.Int).SetBytes(b[:])
|
||||
return bi, nil
|
||||
@@ -287,7 +294,20 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
toEthAddr := EthAddrToBigInt(tx.ToEthAddr)
|
||||
|
||||
// e1: [5 bytes AmountFloat40 | 20 bytes ToEthAddr]
|
||||
var e1B [25]byte
|
||||
amountFloat40, err := NewFloat40(tx.Amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
copy(e1B[0:5], amountFloat40Bytes)
|
||||
copy(e1B[5:25], tx.ToEthAddr[:])
|
||||
e1 := new(big.Int).SetBytes(e1B[:])
|
||||
rqToEthAddr := EthAddrToBigInt(tx.RqToEthAddr)
|
||||
|
||||
_, toBJJY := babyjub.UnpackSignY(tx.ToBJJ)
|
||||
@@ -299,7 +319,8 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
||||
|
||||
_, rqToBJJY := babyjub.UnpackSignY(tx.RqToBJJ)
|
||||
|
||||
return poseidon.Hash([]*big.Int{toCompressedData, toEthAddr, toBJJY, rqTxCompressedDataV2, rqToEthAddr, rqToBJJY})
|
||||
return poseidon.Hash([]*big.Int{toCompressedData, e1, toBJJY, rqTxCompressedDataV2,
|
||||
rqToEthAddr, rqToBJJY})
|
||||
}
|
||||
|
||||
// VerifySignature returns true if the signature verification is correct for the given PublicKeyComp
|
||||
|
||||
@@ -21,85 +21,113 @@ func TestNewPoolL2Tx(t *testing.T) {
|
||||
}
|
||||
poolL2Tx, err := NewPoolL2Tx(poolL2Tx)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "0x02fb52b5d0b9ef2626c11701bb751b2720c76d59946b9a48146ac153bb6e63bf6a", poolL2Tx.TxID.String())
|
||||
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e",
|
||||
poolL2Tx.TxID.String())
|
||||
}
|
||||
|
||||
func TestTxCompressedData(t *testing.T) {
|
||||
chainID := uint16(0)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
func TestTxCompressedDataAndTxCompressedDataV2JSVectors(t *testing.T) {
|
||||
// test vectors values generated from javascript implementation
|
||||
var skPositive babyjub.PrivateKey // 'Positive' refers to the sign
|
||||
_, err := hex.Decode(skPositive[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
|
||||
var skNegative babyjub.PrivateKey // 'Negative' refers to the sign
|
||||
_, err = hex.Decode(skNegative[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090002"))
|
||||
assert.NoError(t, err)
|
||||
|
||||
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
Amount: big.NewInt(4),
|
||||
TokenID: 5,
|
||||
Nonce: 6,
|
||||
ToBJJ: sk.Public().Compress(),
|
||||
FromIdx: (1 << 48) - 1,
|
||||
ToIdx: (1 << 48) - 1,
|
||||
Amount: amount,
|
||||
TokenID: (1 << 32) - 1,
|
||||
Nonce: (1 << 40) - 1,
|
||||
Fee: (1 << 3) - 1,
|
||||
ToBJJ: skPositive.Public().Compress(),
|
||||
}
|
||||
txCompressedData, err := tx.TxCompressedData(chainID)
|
||||
assert.NoError(t, err)
|
||||
// test vector value generated from javascript implementation
|
||||
expectedStr := "1766847064778421992193717128424891165872736891548909569553540445094274575"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
assert.Equal(t, "010000000000060000000500040000000000030000000000020000c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
// using a different chainID
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(100))
|
||||
assert.NoError(t, err)
|
||||
expectedStr = "1766847064778421992193717128424891165872736891548909569553540874591004175"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
assert.Equal(t, "010000000000060000000500040000000000030000000000020064c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(65535))
|
||||
assert.NoError(t, err)
|
||||
expectedStr = "1766847064778421992193717128424891165872736891548909569553821915776017935"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
assert.Equal(t, "01000000000006000000050004000000000003000000000002ffffc60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
txCompressedData, err := tx.TxCompressedData(uint16((1 << 16) - 1))
|
||||
require.NoError(t, err)
|
||||
expectedStr := "0107ffffffffffffffffffffffffffffffffffffffffffffffc60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
txCompressedDataV2, err := tx.TxCompressedDataV2()
|
||||
require.NoError(t, err)
|
||||
expectedStr = "0107ffffffffffffffffffffffffffffffffffffffffffffffffffff"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedDataV2.Bytes()))
|
||||
|
||||
tx = PoolL2Tx{
|
||||
RqFromIdx: 7,
|
||||
RqToIdx: 8,
|
||||
RqAmount: big.NewInt(9),
|
||||
RqTokenID: 10,
|
||||
RqNonce: 11,
|
||||
RqFee: 12,
|
||||
RqToBJJ: sk.Public().Compress(),
|
||||
FromIdx: 0,
|
||||
ToIdx: 0,
|
||||
Amount: big.NewInt(0),
|
||||
TokenID: 0,
|
||||
Nonce: 0,
|
||||
Fee: 0,
|
||||
ToBJJ: skNegative.Public().Compress(),
|
||||
}
|
||||
rqTxCompressedData, err := tx.RqTxCompressedDataV2()
|
||||
assert.NoError(t, err)
|
||||
// test vector value generated from javascript implementation
|
||||
expectedStr = "6571340879233176732837827812956721483162819083004853354503"
|
||||
assert.Equal(t, expectedStr, rqTxCompressedData.String())
|
||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(rqTxCompressedData.Bytes()))
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||
require.NoError(t, err)
|
||||
expectedStr = "c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
func TestTxCompressedDataV2(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 7,
|
||||
ToIdx: 8,
|
||||
Amount: big.NewInt(9),
|
||||
TokenID: 10,
|
||||
Nonce: 11,
|
||||
Fee: 12,
|
||||
ToBJJ: sk.Public().Compress(),
|
||||
txCompressedDataV2, err = tx.TxCompressedDataV2()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, "0", txCompressedDataV2.String())
|
||||
|
||||
amount, ok = new(big.Int).SetString("63000000000000000", 10)
|
||||
require.True(t, ok)
|
||||
tx = PoolL2Tx{
|
||||
FromIdx: 324,
|
||||
ToIdx: 256,
|
||||
Amount: amount,
|
||||
TokenID: 123,
|
||||
Nonce: 76,
|
||||
Fee: 214,
|
||||
ToBJJ: skNegative.Public().Compress(),
|
||||
}
|
||||
txCompressedData, err := tx.TxCompressedDataV2()
|
||||
assert.NoError(t, err)
|
||||
// test vector value generated from javascript implementation
|
||||
expectedStr := "6571340879233176732837827812956721483162819083004853354503"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
||||
assert.True(t, ok)
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(1))
|
||||
require.NoError(t, err)
|
||||
expectedStr = "d6000000004c0000007b0000000001000000000001440001c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
txCompressedDataV2, err = tx.TxCompressedDataV2()
|
||||
require.NoError(t, err)
|
||||
expectedStr = "d6000000004c0000007b3977825f00000000000100000000000144"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedDataV2.Bytes()))
|
||||
|
||||
tx = PoolL2Tx{
|
||||
FromIdx: 1,
|
||||
ToIdx: 2,
|
||||
TokenID: 3,
|
||||
Nonce: 4,
|
||||
Fee: 5,
|
||||
ToBJJ: skNegative.Public().Compress(),
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||
require.NoError(t, err)
|
||||
expectedStr = "050000000004000000030000000000020000000000010000c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
|
||||
tx = PoolL2Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
TokenID: 4,
|
||||
Nonce: 5,
|
||||
Fee: 6,
|
||||
ToBJJ: skPositive.Public().Compress(),
|
||||
}
|
||||
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||
require.NoError(t, err)
|
||||
expectedStr = "01060000000005000000040000000000030000000000020000c60be60f"
|
||||
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||
}
|
||||
|
||||
func TestRqTxCompressedDataV2(t *testing.T) {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
RqFromIdx: 7,
|
||||
@@ -113,19 +141,17 @@ func TestRqTxCompressedDataV2(t *testing.T) {
|
||||
txCompressedData, err := tx.RqTxCompressedDataV2()
|
||||
assert.NoError(t, err)
|
||||
// test vector value generated from javascript implementation
|
||||
expectedStr := "6571340879233176732837827812956721483162819083004853354503"
|
||||
expectedStr := "110248805340524920412994530176819463725852160917809517418728390663"
|
||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
||||
assert.True(t, ok)
|
||||
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
||||
assert.Equal(t, "010c000000000b0000000a0000000009000000000008000000000007",
|
||||
hex.EncodeToString(txCompressedData.Bytes()))
|
||||
}
|
||||
|
||||
func TestHashToSign(t *testing.T) {
|
||||
chainID := uint16(0)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 2,
|
||||
ToIdx: 3,
|
||||
@@ -136,13 +162,15 @@ func TestHashToSign(t *testing.T) {
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "1469900657138253851938022936440971384682713995864967090251961124784132925291", toSign.String())
|
||||
assert.Equal(t, "0b8abaf6b7933464e4450df2514da8b72606c02bf7f89bf6e54816fbda9d9d57",
|
||||
hex.EncodeToString(toSign.Bytes()))
|
||||
}
|
||||
|
||||
func TestVerifyTxSignature(t *testing.T) {
|
||||
chainID := uint16(0)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 2,
|
||||
@@ -156,18 +184,49 @@ func TestVerifyTxSignature(t *testing.T) {
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, "18645218094210271622244722988708640202588315450486586312909439859037906375295", toSign.String())
|
||||
assert.Equal(t,
|
||||
"3144939470626721092564692894890580265754250231349521601298746071096761507003",
|
||||
toSign.String())
|
||||
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
tx.Signature = sig.Compress()
|
||||
assert.True(t, tx.VerifySignature(chainID, sk.Public().Compress()))
|
||||
}
|
||||
|
||||
func TestVerifyTxSignatureEthAddrWith0(t *testing.T) {
|
||||
chainID := uint16(5)
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("02f0b4f87065af3797aaaf934e8b5c31563c17f2272fa71bd0146535bfbb4184"))
|
||||
assert.NoError(t, err)
|
||||
tx := PoolL2Tx{
|
||||
FromIdx: 10659,
|
||||
ToIdx: 0,
|
||||
ToEthAddr: ethCommon.HexToAddress("0x0004308BD15Ead4F1173624dC289DBdcC806a309"),
|
||||
Amount: big.NewInt(5000),
|
||||
TokenID: 0,
|
||||
Nonce: 946,
|
||||
Fee: 231,
|
||||
}
|
||||
toSign, err := tx.HashToSign(chainID)
|
||||
assert.NoError(t, err)
|
||||
|
||||
sig := sk.SignPoseidon(toSign)
|
||||
assert.Equal(t,
|
||||
"f208b8298d5f37148ac3c0c03703272ea47b9f836851bcf8dd5f7e4e3b336ca1d2f6e92ad85dc25f174daf7a0abfd5f71dead3f059b783f4c4b2f56a18a47000",
|
||||
sig.Compress().String(),
|
||||
)
|
||||
tx.Signature = sig.Compress()
|
||||
assert.True(t, tx.VerifySignature(chainID, sk.Public().Compress()))
|
||||
}
|
||||
|
||||
func TestDecompressEmptyBJJComp(t *testing.T) {
|
||||
pkComp := EmptyBJJComp
|
||||
pk, err := pkComp.Decompress()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, "2957874849018779266517920829765869116077630550401372566248359756137677864698", pk.X.String())
|
||||
assert.Equal(t,
|
||||
"2957874849018779266517920829765869116077630550401372566248359756137677864698",
|
||||
pk.X.String())
|
||||
assert.Equal(t, "0", pk.Y.String())
|
||||
}
|
||||
|
||||
|
||||
@@ -15,8 +15,9 @@ const tokenIDBytesLen = 4
|
||||
|
||||
// Token is a struct that represents an Ethereum token that is supported in Hermez network
|
||||
type Token struct {
|
||||
TokenID TokenID `json:"id" meddler:"token_id"`
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"` // Ethereum block number in which this token was registered
|
||||
TokenID TokenID `json:"id" meddler:"token_id"`
|
||||
// EthBlockNum indicates the Ethereum block number in which this token was registered
|
||||
EthBlockNum int64 `json:"ethereumBlockNum" meddler:"eth_block_num"`
|
||||
EthAddr ethCommon.Address `json:"ethereumAddress" meddler:"eth_addr"`
|
||||
Name string `json:"name" meddler:"name"`
|
||||
Symbol string `json:"symbol" meddler:"symbol"`
|
||||
@@ -48,7 +49,8 @@ func (t TokenID) BigInt() *big.Int {
|
||||
// TokenIDFromBytes returns TokenID from a byte array
|
||||
func TokenIDFromBytes(b []byte) (TokenID, error) {
|
||||
if len(b) != tokenIDBytesLen {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse TokenID, bytes len %d, expected 4", len(b)))
|
||||
return 0, tracerr.Wrap(fmt.Errorf("can not parse TokenID, bytes len %d, expected 4",
|
||||
len(b)))
|
||||
}
|
||||
tid := binary.BigEndian.Uint32(b[:4])
|
||||
return TokenID(tid), nil
|
||||
|
||||
54
common/tx.go
54
common/tx.go
@@ -15,12 +15,12 @@ import (
|
||||
)
|
||||
|
||||
const (
|
||||
// TXIDPrefixL1UserTx is the prefix that determines that the TxID is
|
||||
// for a L1UserTx
|
||||
// TxIDPrefixL1UserTx is the prefix that determines that the TxID is for
|
||||
// a L1UserTx
|
||||
//nolinter:gomnd
|
||||
TxIDPrefixL1UserTx = byte(0)
|
||||
|
||||
// TXIDPrefixL1CoordTx is the prefix that determines that the TxID is
|
||||
// TxIDPrefixL1CoordTx is the prefix that determines that the TxID is
|
||||
// for a L1CoordinatorTx
|
||||
//nolinter:gomnd
|
||||
TxIDPrefixL1CoordTx = byte(1)
|
||||
@@ -51,7 +51,8 @@ func (txid *TxID) Scan(src interface{}) error {
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan %T into TxID", src))
|
||||
}
|
||||
if len(srcB) != TxIDLen {
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan []byte of len %d into TxID, need %d", len(srcB), TxIDLen))
|
||||
return tracerr.Wrap(fmt.Errorf("can't scan []byte of len %d into TxID, need %d",
|
||||
len(srcB), TxIDLen))
|
||||
}
|
||||
copy(txid[:], srcB)
|
||||
return nil
|
||||
@@ -87,7 +88,7 @@ func (txid TxID) MarshalText() ([]byte, error) {
|
||||
return []byte(txid.String()), nil
|
||||
}
|
||||
|
||||
// UnmarshalText unmarshals a TxID
|
||||
// UnmarshalText unmarshalls a TxID
|
||||
func (txid *TxID) UnmarshalText(data []byte) error {
|
||||
idStr := string(data)
|
||||
id, err := NewTxIDFromString(idStr)
|
||||
@@ -102,13 +103,15 @@ func (txid *TxID) UnmarshalText(data []byte) error {
|
||||
type TxType string
|
||||
|
||||
const (
|
||||
// TxTypeExit represents L2->L1 token transfer. A leaf for this account appears in the exit tree of the block
|
||||
// TxTypeExit represents L2->L1 token transfer. A leaf for this account appears in the exit
|
||||
// tree of the block
|
||||
TxTypeExit TxType = "Exit"
|
||||
// TxTypeTransfer represents L2->L2 token transfer
|
||||
TxTypeTransfer TxType = "Transfer"
|
||||
// TxTypeDeposit represents L1->L2 transfer
|
||||
TxTypeDeposit TxType = "Deposit"
|
||||
// TxTypeCreateAccountDeposit represents creation of a new leaf in the state tree (newAcconut) + L1->L2 transfer
|
||||
// TxTypeCreateAccountDeposit represents creation of a new leaf in the state tree
|
||||
// (newAcconut) + L1->L2 transfer
|
||||
TxTypeCreateAccountDeposit TxType = "CreateAccountDeposit"
|
||||
// TxTypeCreateAccountDepositTransfer represents L1->L2 transfer + L2->L2 transfer
|
||||
TxTypeCreateAccountDepositTransfer TxType = "CreateAccountDepositTransfer"
|
||||
@@ -124,24 +127,31 @@ const (
|
||||
TxTypeTransferToBJJ TxType = "TransferToBJJ"
|
||||
)
|
||||
|
||||
// Tx is a struct used by the TxSelector & BatchBuilder as a generic type generated from L1Tx & PoolL2Tx
|
||||
// Tx is a struct used by the TxSelector & BatchBuilder as a generic type generated from L1Tx &
|
||||
// PoolL2Tx
|
||||
type Tx struct {
|
||||
// Generic
|
||||
IsL1 bool `meddler:"is_l1"`
|
||||
TxID TxID `meddler:"id"`
|
||||
Type TxType `meddler:"type"`
|
||||
Position int `meddler:"position"`
|
||||
FromIdx Idx `meddler:"from_idx"`
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
AmountFloat float64 `meddler:"amount_f"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
USD *float64 `meddler:"amount_usd"`
|
||||
BatchNum *BatchNum `meddler:"batch_num"` // batchNum in which this tx was forged. If the tx is L2, this must be != 0
|
||||
EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum Block Number in which this L1Tx was added to the queue
|
||||
IsL1 bool `meddler:"is_l1"`
|
||||
TxID TxID `meddler:"id"`
|
||||
Type TxType `meddler:"type"`
|
||||
Position int `meddler:"position"`
|
||||
FromIdx Idx `meddler:"from_idx"`
|
||||
ToIdx Idx `meddler:"to_idx"`
|
||||
Amount *big.Int `meddler:"amount,bigint"`
|
||||
AmountFloat float64 `meddler:"amount_f"`
|
||||
TokenID TokenID `meddler:"token_id"`
|
||||
USD *float64 `meddler:"amount_usd"`
|
||||
// BatchNum in which this tx was forged. If the tx is L2, this must be != 0
|
||||
BatchNum *BatchNum `meddler:"batch_num"`
|
||||
// Ethereum Block Number in which this L1Tx was added to the queue
|
||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
||||
// L1
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"` // toForgeL1TxsNum in which the tx was forged / will be forged
|
||||
UserOrigin *bool `meddler:"user_origin"` // true if the tx was originated by a user, false if it was aoriginated by a coordinator. Note that this differ from the spec for implementation simplification purpposes
|
||||
// ToForgeL1TxsNum in which the tx was forged / will be forged
|
||||
ToForgeL1TxsNum *int64 `meddler:"to_forge_l1_txs_num"`
|
||||
// UserOrigin is set to true if the tx was originated by a user, false if it was aoriginated
|
||||
// by a coordinator. Note that this differ from the spec for implementation simplification
|
||||
// purpposes
|
||||
UserOrigin *bool `meddler:"user_origin"`
|
||||
FromEthAddr ethCommon.Address `meddler:"from_eth_addr"`
|
||||
FromBJJ babyjub.PublicKeyComp `meddler:"from_bjj"`
|
||||
DepositAmount *big.Int `meddler:"deposit_amount,bigintnull"`
|
||||
|
||||
@@ -21,8 +21,10 @@ func TestSignatureConstant(t *testing.T) {
|
||||
func TestTxIDScannerValue(t *testing.T) {
|
||||
txid0 := &TxID{}
|
||||
txid1 := &TxID{}
|
||||
txid0B := [TxIDLen]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2}
|
||||
txid1B := [TxIDLen]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
||||
txid0B := [TxIDLen]byte{0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 0, 1, 2,
|
||||
3, 4, 5, 6, 7, 8, 9, 0, 1, 2}
|
||||
txid1B := [TxIDLen]byte{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
|
||||
0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
||||
|
||||
copy(txid0[:], txid0B[:])
|
||||
copy(txid1[:], txid1B[:])
|
||||
|
||||
@@ -62,3 +62,17 @@ func RmEndingZeroes(siblings []*merkletree.Hash) []*merkletree.Hash {
|
||||
}
|
||||
return siblings[:pos]
|
||||
}
|
||||
|
||||
// TokensToUSD is a helper function to calculate the USD value of a certain
|
||||
// amount of tokens considering the normalized token price (which is the price
|
||||
// commonly reported by exhanges)
|
||||
func TokensToUSD(amount *big.Int, decimals uint64, valueUSD float64) float64 {
|
||||
amountF := new(big.Float).SetInt(amount)
|
||||
// Divide by 10^decimals to normalize the amount
|
||||
baseF := new(big.Float).SetInt(new(big.Int).Exp(
|
||||
big.NewInt(10), big.NewInt(int64(decimals)), nil)) //nolint:gomnd
|
||||
amountF.Mul(amountF, big.NewFloat(valueUSD))
|
||||
amountF.Quo(amountF, baseF)
|
||||
amountUSD, _ := amountF.Float64()
|
||||
return amountUSD
|
||||
}
|
||||
|
||||
@@ -21,16 +21,23 @@ func TestBJJFromStringWithChecksum(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
|
||||
// expected values computed with js implementation
|
||||
assert.Equal(t, "2492816973395423007340226948038371729989170225696553239457870892535792679622", pk.X.String())
|
||||
assert.Equal(t, "15238403086306505038849621710779816852318505119327426213168494964113886299863", pk.Y.String())
|
||||
assert.Equal(t,
|
||||
"2492816973395423007340226948038371729989170225696553239457870892535792679622",
|
||||
pk.X.String())
|
||||
assert.Equal(t,
|
||||
"15238403086306505038849621710779816852318505119327426213168494964113886299863",
|
||||
pk.Y.String())
|
||||
}
|
||||
|
||||
func TestRmEndingZeroes(t *testing.T) {
|
||||
s0, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000000")
|
||||
s0, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000000")
|
||||
require.NoError(t, err)
|
||||
s1, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000001")
|
||||
s1, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000001")
|
||||
require.NoError(t, err)
|
||||
s2, err := merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000002")
|
||||
s2, err :=
|
||||
merkletree.NewHashFromHex("0x0000000000000000000000000000000000000000000000000000000000000002")
|
||||
require.NoError(t, err)
|
||||
|
||||
// expect cropped last zeroes
|
||||
|
||||
33
common/zk.go
33
common/zk.go
@@ -1,4 +1,4 @@
|
||||
// Package common contains all the common data structures used at the
|
||||
// Package common zk.go contains all the common data structures used at the
|
||||
// hermez-node, zk.go contains the zkSnark inputs used to generate the proof
|
||||
package common
|
||||
|
||||
@@ -67,7 +67,7 @@ type ZKInputs struct {
|
||||
|
||||
// accumulate fees
|
||||
// FeePlanTokens contains all the tokenIDs for which the fees are being
|
||||
// accumulated and those fees accoumulated will be paid to the FeeIdxs
|
||||
// accumulated and those fees accumulated will be paid to the FeeIdxs
|
||||
// array. The order of FeeIdxs & FeePlanTokens & State3 must match.
|
||||
// Coordinator fees are processed correlated such as:
|
||||
// [FeePlanTokens[i], FeeIdxs[i]]
|
||||
@@ -102,6 +102,8 @@ type ZKInputs struct {
|
||||
ToBJJAy []*big.Int `json:"toBjjAy"` // big.Int, len: [maxTx]
|
||||
// ToEthAddr
|
||||
ToEthAddr []*big.Int `json:"toEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||
// AmountF encoded as float40
|
||||
AmountF []*big.Int `json:"amountF"` // uint40 len: [maxTx]
|
||||
|
||||
// OnChain determines if is L1 (1/true) or L2 (0/false)
|
||||
OnChain []*big.Int `json:"onChain"` // bool, len: [maxTx]
|
||||
@@ -112,8 +114,8 @@ type ZKInputs struct {
|
||||
// NewAccount boolean (0/1) flag set 'true' when L1 tx creates a new
|
||||
// account (fromIdx==0)
|
||||
NewAccount []*big.Int `json:"newAccount"` // bool, len: [maxTx]
|
||||
// DepositAmountF encoded as float16
|
||||
DepositAmountF []*big.Int `json:"loadAmountF"` // uint16, len: [maxTx]
|
||||
// DepositAmountF encoded as float40
|
||||
DepositAmountF []*big.Int `json:"loadAmountF"` // uint40, len: [maxTx]
|
||||
// FromEthAddr
|
||||
FromEthAddr []*big.Int `json:"fromEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||
// FromBJJCompressed boolean encoded where each value is a *big.Int
|
||||
@@ -128,8 +130,8 @@ type ZKInputs struct {
|
||||
RqOffset []*big.Int `json:"rqOffset"` // uint8 (max 3 bits), len: [maxTx]
|
||||
|
||||
// transaction L2 request data
|
||||
// RqTxCompressedDataV2
|
||||
RqTxCompressedDataV2 []*big.Int `json:"rqTxCompressedDataV2"` // big.Int (max 251 bits), len: [maxTx]
|
||||
// RqTxCompressedDataV2 big.Int (max 251 bits), len: [maxTx]
|
||||
RqTxCompressedDataV2 []*big.Int `json:"rqTxCompressedDataV2"`
|
||||
// RqToEthAddr
|
||||
RqToEthAddr []*big.Int `json:"rqToEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||
// RqToBJJAy
|
||||
@@ -299,7 +301,8 @@ func (z ZKInputs) MarshalJSON() ([]byte, error) {
|
||||
}
|
||||
|
||||
// NewZKInputs returns a pointer to an initialized struct of ZKInputs
|
||||
func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32, currentNumBatch *big.Int) *ZKInputs {
|
||||
func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32,
|
||||
currentNumBatch *big.Int) *ZKInputs {
|
||||
zki := &ZKInputs{}
|
||||
zki.Metadata.MaxFeeIdxs = maxFeeIdxs
|
||||
zki.Metadata.MaxLevels = uint32(48) //nolint:gomnd
|
||||
@@ -326,6 +329,7 @@ func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32, cur
|
||||
zki.AuxToIdx = newSlice(maxTx)
|
||||
zki.ToBJJAy = newSlice(maxTx)
|
||||
zki.ToEthAddr = newSlice(maxTx)
|
||||
zki.AmountF = newSlice(maxTx)
|
||||
zki.OnChain = newSlice(maxTx)
|
||||
zki.NewAccount = newSlice(maxTx)
|
||||
|
||||
@@ -476,8 +480,8 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
||||
copy(newExitRoot, z.Metadata.NewExitRootRaw.Bytes())
|
||||
b = append(b, newExitRoot...)
|
||||
|
||||
// [MAX_L1_TX * (2 * MAX_NLEVELS + 480) bits] L1TxsData
|
||||
l1TxDataLen := (2*z.Metadata.MaxLevels + 480)
|
||||
// [MAX_L1_TX * (2 * MAX_NLEVELS + 528) bits] L1TxsData
|
||||
l1TxDataLen := (2*z.Metadata.MaxLevels + 528) //nolint:gomnd
|
||||
l1TxsDataLen := (z.Metadata.MaxL1Tx * l1TxDataLen)
|
||||
l1TxsData := make([]byte, l1TxsDataLen/8) //nolint:gomnd
|
||||
for i := 0; i < len(z.Metadata.L1TxsData); i++ {
|
||||
@@ -494,20 +498,23 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
||||
}
|
||||
b = append(b, l1TxsDataAvailability...)
|
||||
|
||||
// [MAX_TX*(2*NLevels + 24) bits] L2TxsData
|
||||
// [MAX_TX*(2*NLevels + 48) bits] L2TxsData
|
||||
var l2TxsData []byte
|
||||
l2TxDataLen := 2*z.Metadata.NLevels + 24 //nolint:gomnd
|
||||
l2TxDataLen := 2*z.Metadata.NLevels + 48 //nolint:gomnd
|
||||
l2TxsDataLen := (z.Metadata.MaxTx * l2TxDataLen)
|
||||
expectedL2TxsDataLen := l2TxsDataLen / 8 //nolint:gomnd
|
||||
for i := 0; i < len(z.Metadata.L2TxsData); i++ {
|
||||
l2TxsData = append(l2TxsData, z.Metadata.L2TxsData[i]...)
|
||||
}
|
||||
if len(l2TxsData) > int(expectedL2TxsDataLen) {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("len(l2TxsData): %d, expected: %d", len(l2TxsData), expectedL2TxsDataLen))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("len(l2TxsData): %d, expected: %d",
|
||||
len(l2TxsData), expectedL2TxsDataLen))
|
||||
}
|
||||
|
||||
b = append(b, l2TxsData...)
|
||||
l2TxsPadding := make([]byte, (int(z.Metadata.MaxTx)-len(z.Metadata.L1TxsDataAvailability)-len(z.Metadata.L2TxsData))*int(l2TxDataLen)/8) //nolint:gomnd
|
||||
l2TxsPadding := make([]byte,
|
||||
(int(z.Metadata.MaxTx)-len(z.Metadata.L1TxsDataAvailability)-
|
||||
len(z.Metadata.L2TxsData))*int(l2TxDataLen)/8) //nolint:gomnd
|
||||
b = append(b, l2TxsPadding...)
|
||||
|
||||
// [NLevels * MAX_TOKENS_FEE bits] feeTxsData
|
||||
|
||||
149
config/config.go
149
config/config.go
@@ -3,6 +3,7 @@ package config
|
||||
import (
|
||||
"fmt"
|
||||
"io/ioutil"
|
||||
"math/big"
|
||||
"time"
|
||||
|
||||
"github.com/BurntSushi/toml"
|
||||
@@ -34,10 +35,23 @@ type ServerProof struct {
|
||||
URL string `validate:"required"`
|
||||
}
|
||||
|
||||
// ForgeBatchGasCost is the costs associated to a ForgeBatch transaction, split
|
||||
// into different parts to be used in a formula.
|
||||
type ForgeBatchGasCost struct {
|
||||
Fixed uint64 `validate:"required"`
|
||||
L1UserTx uint64 `validate:"required"`
|
||||
L1CoordTx uint64 `validate:"required"`
|
||||
L2Tx uint64 `validate:"required"`
|
||||
}
|
||||
|
||||
// Coordinator is the coordinator specific configuration.
|
||||
type Coordinator struct {
|
||||
// ForgerAddress is the address under which this coordinator is forging
|
||||
ForgerAddress ethCommon.Address `validate:"required"`
|
||||
// MinimumForgeAddressBalance is the minimum balance the forger address
|
||||
// needs to start the coordinator in wei. Of set to 0, the coordinator
|
||||
// will not check the balance before starting.
|
||||
MinimumForgeAddressBalance *big.Int
|
||||
// FeeAccount is the Hermez account that the coordinator uses to receive fees
|
||||
FeeAccount struct {
|
||||
// Address is the ethereum address of the account to receive fees
|
||||
@@ -51,33 +65,85 @@ type Coordinator struct {
|
||||
// L1BatchTimeoutPerc is the portion of the range before the L1Batch
|
||||
// timeout that will trigger a schedule to forge an L1Batch
|
||||
L1BatchTimeoutPerc float64 `validate:"required"`
|
||||
// StartSlotBlocksDelay is the number of blocks of delay to wait before
|
||||
// starting the pipeline when we reach a slot in which we can forge.
|
||||
StartSlotBlocksDelay int64
|
||||
// ScheduleBatchBlocksAheadCheck is the number of blocks ahead in which
|
||||
// the forger address is checked to be allowed to forge (apart from
|
||||
// checking the next block), used to decide when to stop scheduling new
|
||||
// batches (by stopping the pipeline).
|
||||
// For example, if we are at block 10 and ScheduleBatchBlocksAheadCheck
|
||||
// is 5, even though at block 11 we canForge, the pipeline will be
|
||||
// stopped if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// scheduling a batch and having it mined.
|
||||
ScheduleBatchBlocksAheadCheck int64
|
||||
// SendBatchBlocksMarginCheck is the number of margin blocks ahead in
|
||||
// which the coordinator is also checked to be allowed to forge, apart
|
||||
// from the next block; used to decide when to stop sending batches to
|
||||
// the smart contract.
|
||||
// For example, if we are at block 10 and SendBatchBlocksMarginCheck is
|
||||
// 5, even though at block 11 we canForge, the batch will be discarded
|
||||
// if we can't forge at block 15.
|
||||
SendBatchBlocksMarginCheck int64
|
||||
// ProofServerPollInterval is the waiting interval between polling the
|
||||
// ProofServer while waiting for a particular status
|
||||
ProofServerPollInterval Duration `validate:"required"`
|
||||
// ForgeRetryInterval is the waiting interval between calls forge a
|
||||
// batch after an error
|
||||
ForgeRetryInterval Duration `validate:"required"`
|
||||
// ForgeDelay is the delay after which a batch is forged if the slot is
|
||||
// already committed. If set to 0s, the coordinator will continuously
|
||||
// forge at the maximum rate.
|
||||
ForgeDelay Duration `validate:"-"`
|
||||
// ForgeNoTxsDelay is the delay after which a batch is forged even if
|
||||
// there are no txs to forge if the slot is already committed. If set
|
||||
// to 0s, the coordinator will continuously forge even if the batches
|
||||
// are empty.
|
||||
ForgeNoTxsDelay Duration `validate:"-"`
|
||||
// SyncRetryInterval is the waiting interval between calls to the main
|
||||
// handler of a synced block after an error
|
||||
SyncRetryInterval Duration `validate:"required"`
|
||||
// PurgeByExtDelInterval is the waiting interval between calls
|
||||
// to the PurgeByExternalDelete function of the l2db which deletes
|
||||
// pending txs externally marked by the column `external_delete`
|
||||
PurgeByExtDelInterval Duration `validate:"required"`
|
||||
// L2DB is the DB that holds the pool of L2Txs
|
||||
L2DB struct {
|
||||
// SafetyPeriod is the number of batches after which
|
||||
// non-pending L2Txs are deleted from the pool
|
||||
SafetyPeriod common.BatchNum `validate:"required"`
|
||||
// MaxTxs is the number of L2Txs that once reached triggers
|
||||
// deletion of old L2Txs
|
||||
// MaxTxs is the maximum number of pending L2Txs that can be
|
||||
// stored in the pool. Once this number of pending L2Txs is
|
||||
// reached, inserts to the pool will be denied until some of
|
||||
// the pending txs are forged.
|
||||
MaxTxs uint32 `validate:"required"`
|
||||
// MinFeeUSD is the minimum fee in USD that a tx must pay in
|
||||
// order to be accepted into the pool. Txs with lower than
|
||||
// minimum fee will be rejected at the API level.
|
||||
MinFeeUSD float64
|
||||
// TTL is the Time To Live for L2Txs in the pool. Once MaxTxs
|
||||
// L2Txs is reached, L2Txs older than TTL will be deleted.
|
||||
TTL Duration `validate:"required"`
|
||||
// PurgeBatchDelay is the delay between batches to purge outdated transactions
|
||||
// PurgeBatchDelay is the delay between batches to purge
|
||||
// outdated transactions. Outdated L2Txs are those that have
|
||||
// been forged or marked as invalid for longer than the
|
||||
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||
// for longer than TTL once there are MaxTxs.
|
||||
PurgeBatchDelay int64 `validate:"required"`
|
||||
// InvalidateBatchDelay is the delay between batches to mark invalid transactions
|
||||
// InvalidateBatchDelay is the delay between batches to mark
|
||||
// invalid transactions due to nonce lower than the account
|
||||
// nonce.
|
||||
InvalidateBatchDelay int64 `validate:"required"`
|
||||
// PurgeBlockDelay is the delay between blocks to purge outdated transactions
|
||||
// PurgeBlockDelay is the delay between blocks to purge
|
||||
// outdated transactions. Outdated L2Txs are those that have
|
||||
// been forged or marked as invalid for longer than the
|
||||
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||
// for longer than TTL once there are MaxTxs.
|
||||
PurgeBlockDelay int64 `validate:"required"`
|
||||
// InvalidateBlockDelay is the delay between blocks to mark invalid transactions
|
||||
// InvalidateBlockDelay is the delay between blocks to mark
|
||||
// invalid transactions due to nonce lower than the account
|
||||
// nonce.
|
||||
InvalidateBlockDelay int64 `validate:"required"`
|
||||
} `validate:"required"`
|
||||
TxSelector struct {
|
||||
@@ -90,7 +156,6 @@ type Coordinator struct {
|
||||
} `validate:"required"`
|
||||
ServerProofs []ServerProof `validate:"required"`
|
||||
Circuit struct {
|
||||
// VerifierIdx uint8 `validate:"required"`
|
||||
// MaxTx is the maximum number of txs supported by the circuit
|
||||
MaxTx int64 `validate:"required"`
|
||||
// NLevels is the maximum number of merkle tree levels
|
||||
@@ -98,12 +163,13 @@ type Coordinator struct {
|
||||
NLevels int64 `validate:"required"`
|
||||
} `validate:"required"`
|
||||
EthClient struct {
|
||||
// CallGasLimit is the default gas limit set for ethereum
|
||||
// calls, except for methods where a particular gas limit is
|
||||
// harcoded because it's known to be a big value
|
||||
CallGasLimit uint64 `validate:"required"`
|
||||
// GasPriceDiv is the gas price division
|
||||
GasPriceDiv uint64 `validate:"required"`
|
||||
// MaxGasPrice is the maximum gas price allowed for ethereum
|
||||
// transactions
|
||||
MaxGasPrice *big.Int `validate:"required"`
|
||||
// GasPriceIncPerc is the percentage increase of gas price set
|
||||
// in an ethereum transaction from the suggested gas price by
|
||||
// the ethereum node
|
||||
GasPriceIncPerc int64
|
||||
// CheckLoopInterval is the waiting interval between receipt
|
||||
// checks of ethereum transactions in the TxManager
|
||||
CheckLoopInterval Duration `validate:"required"`
|
||||
@@ -113,6 +179,13 @@ type Coordinator struct {
|
||||
// AttemptsDelay is delay between attempts do do an eth client
|
||||
// RPC call
|
||||
AttemptsDelay Duration `validate:"required"`
|
||||
// TxResendTimeout is the timeout after which a non-mined
|
||||
// ethereum transaction will be resent (reusing the nonce) with
|
||||
// a newly calculated gas price
|
||||
TxResendTimeout Duration `validate:"required"`
|
||||
// NoReuseNonce disables reusing nonces of pending transactions for
|
||||
// new replacement transactions
|
||||
NoReuseNonce bool
|
||||
// Keystore is the ethereum keystore where private keys are kept
|
||||
Keystore struct {
|
||||
// Path to the keystore
|
||||
@@ -120,6 +193,9 @@ type Coordinator struct {
|
||||
// Password used to decrypt the keys in the keystore
|
||||
Password string `validate:"required"`
|
||||
} `validate:"required"`
|
||||
// ForgeBatchGasCost contains the cost of each action in the
|
||||
// ForgeBatch transaction.
|
||||
ForgeBatchGasCost ForgeBatchGasCost `validate:"required"`
|
||||
} `validate:"required"`
|
||||
API struct {
|
||||
// Coordinator enables the coordinator API endpoints
|
||||
@@ -132,6 +208,10 @@ type Coordinator struct {
|
||||
// LightScrypt if set, uses light parameters for the ethereum
|
||||
// keystore encryption algorithm.
|
||||
LightScrypt bool
|
||||
// RollupVerifierIndex is the index of the verifier to use in
|
||||
// the Rollup smart contract. The verifier chosen by index
|
||||
// must match with the Circuit parameters.
|
||||
RollupVerifierIndex *int
|
||||
}
|
||||
}
|
||||
|
||||
@@ -151,17 +231,30 @@ type Node struct {
|
||||
// Keep is the number of checkpoints to keep
|
||||
Keep int `validate:"required"`
|
||||
} `validate:"required"`
|
||||
// It's possible to use diferentiated SQL connections for read/write.
|
||||
// If the read configuration is not provided, the write one it's going to be used
|
||||
// for both reads and writes
|
||||
PostgreSQL struct {
|
||||
// Port of the PostgreSQL server
|
||||
Port int `validate:"required"`
|
||||
// Host of the PostgreSQL server
|
||||
Host string `validate:"required"`
|
||||
// User of the PostgreSQL server
|
||||
User string `validate:"required"`
|
||||
// Password of the PostgreSQL server
|
||||
Password string `validate:"required"`
|
||||
// Name of the PostgreSQL server database
|
||||
Name string `validate:"required"`
|
||||
// Port of the PostgreSQL write server
|
||||
PortWrite int `validate:"required"`
|
||||
// Host of the PostgreSQL write server
|
||||
HostWrite string `validate:"required"`
|
||||
// User of the PostgreSQL write server
|
||||
UserWrite string `validate:"required"`
|
||||
// Password of the PostgreSQL write server
|
||||
PasswordWrite string `validate:"required"`
|
||||
// Name of the PostgreSQL write server database
|
||||
NameWrite string `validate:"required"`
|
||||
// Port of the PostgreSQL read server
|
||||
PortRead int
|
||||
// Host of the PostgreSQL read server
|
||||
HostRead string
|
||||
// User of the PostgreSQL read server
|
||||
UserRead string
|
||||
// Password of the PostgreSQL read server
|
||||
PasswordRead string
|
||||
// Name of the PostgreSQL read server database
|
||||
NameRead string
|
||||
} `validate:"required"`
|
||||
Web3 struct {
|
||||
// URL is the URL of the web3 ethereum-node RPC server
|
||||
@@ -201,9 +294,14 @@ type Node struct {
|
||||
// UpdateMetricsInterval is the interval between updates of the
|
||||
// API metrics
|
||||
UpdateMetricsInterval Duration
|
||||
// UpdateMetricsInterval is the interval between updates of the
|
||||
// UpdateRecommendedFeeInterval is the interval between updates of the
|
||||
// recommended fees
|
||||
UpdateRecommendedFeeInterval Duration
|
||||
// Maximum concurrent connections allowed between API and SQL
|
||||
MaxSQLConnections int `validate:"required"`
|
||||
// SQLConnectionTimeout is the maximum amount of time that an API request
|
||||
// can wait to stablish a SQL connection
|
||||
SQLConnectionTimeout Duration
|
||||
} `validate:"required"`
|
||||
Debug struct {
|
||||
// APIAddress is the address where the debugAPI will listen if
|
||||
@@ -212,6 +310,9 @@ type Node struct {
|
||||
// MeddlerLogs enables meddler debug mode, where unused columns and struct
|
||||
// fields will be logged
|
||||
MeddlerLogs bool
|
||||
// GinDebugMode sets Gin-Gonic (the web framework) to run in
|
||||
// debug mode
|
||||
GinDebugMode bool
|
||||
}
|
||||
Coordinator Coordinator `validate:"-"`
|
||||
}
|
||||
|
||||
@@ -8,6 +8,7 @@ import (
|
||||
"path"
|
||||
"time"
|
||||
|
||||
"github.com/ethereum/go-ethereum/accounts/abi/bind"
|
||||
"github.com/ethereum/go-ethereum/core/types"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
@@ -47,6 +48,8 @@ type Debug struct {
|
||||
MineBlockNum int64
|
||||
// SendBlockNum is the blockNum when the batch was sent to ethereum
|
||||
SendBlockNum int64
|
||||
// ResendNum is the number of times the tx has been resent
|
||||
ResendNum int
|
||||
// LastScheduledL1BatchBlockNum is the blockNum when the last L1Batch
|
||||
// was scheduled
|
||||
LastScheduledL1BatchBlockNum int64
|
||||
@@ -64,10 +67,17 @@ type Debug struct {
|
||||
// StartToSendDelay is the delay between starting a batch and sending
|
||||
// it to ethereum, in seconds
|
||||
StartToSendDelay float64
|
||||
// StartToMineDelay is the delay between starting a batch and having
|
||||
// it mined in seconds
|
||||
StartToMineDelay float64
|
||||
// SendToMineDelay is the delay between sending a batch tx and having
|
||||
// it mined in seconds
|
||||
SendToMineDelay float64
|
||||
}
|
||||
|
||||
// BatchInfo contans the Batch information
|
||||
type BatchInfo struct {
|
||||
PipelineNum int
|
||||
BatchNum common.BatchNum
|
||||
ServerProof prover.Client
|
||||
ZKInputs *common.ZKInputs
|
||||
@@ -81,10 +91,17 @@ type BatchInfo struct {
|
||||
L2Txs []common.L2Tx
|
||||
CoordIdxs []common.Idx
|
||||
ForgeBatchArgs *eth.RollupForgeBatchArgs
|
||||
// FeesInfo
|
||||
EthTx *types.Transaction
|
||||
Receipt *types.Receipt
|
||||
Debug Debug
|
||||
Auth *bind.TransactOpts `json:"-"`
|
||||
EthTx *types.Transaction
|
||||
EthTxErr error
|
||||
// SendTimestamp the time of batch sent to ethereum
|
||||
SendTimestamp time.Time
|
||||
Receipt *types.Receipt
|
||||
// Fail is true if:
|
||||
// - The receipt status is failed
|
||||
// - A previous parent batch is failed
|
||||
Fail bool
|
||||
Debug Debug
|
||||
}
|
||||
|
||||
// DebugStore is a debug function to store the BatchInfo as a json text file in
|
||||
|
||||
@@ -3,14 +3,15 @@ package coordinator
|
||||
import (
|
||||
"context"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"os"
|
||||
"strings"
|
||||
"sync"
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/config"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
@@ -23,7 +24,10 @@ import (
|
||||
)
|
||||
|
||||
var (
|
||||
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
||||
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
||||
errForgeNoTxsBeforeDelay = fmt.Errorf(
|
||||
"no txs to forge and we haven't reached the forge no txs delay")
|
||||
errForgeBeforeDelay = fmt.Errorf("we haven't reached the forge delay")
|
||||
)
|
||||
|
||||
const (
|
||||
@@ -42,26 +46,81 @@ type Config struct {
|
||||
// L1BatchTimeoutPerc is the portion of the range before the L1Batch
|
||||
// timeout that will trigger a schedule to forge an L1Batch
|
||||
L1BatchTimeoutPerc float64
|
||||
// StartSlotBlocksDelay is the number of blocks of delay to wait before
|
||||
// starting the pipeline when we reach a slot in which we can forge.
|
||||
StartSlotBlocksDelay int64
|
||||
// ScheduleBatchBlocksAheadCheck is the number of blocks ahead in which
|
||||
// the forger address is checked to be allowed to forge (apart from
|
||||
// checking the next block), used to decide when to stop scheduling new
|
||||
// batches (by stopping the pipeline).
|
||||
// For example, if we are at block 10 and ScheduleBatchBlocksAheadCheck
|
||||
// is 5, even though at block 11 we canForge, the pipeline will be
|
||||
// stopped if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// scheduling a batch and having it mined.
|
||||
ScheduleBatchBlocksAheadCheck int64
|
||||
// SendBatchBlocksMarginCheck is the number of margin blocks ahead in
|
||||
// which the coordinator is also checked to be allowed to forge, apart
|
||||
// from the next block; used to decide when to stop sending batches to
|
||||
// the smart contract.
|
||||
// For example, if we are at block 10 and SendBatchBlocksMarginCheck is
|
||||
// 5, even though at block 11 we canForge, the batch will be discarded
|
||||
// if we can't forge at block 15.
|
||||
// This value should be the expected number of blocks it takes between
|
||||
// sending a batch and having it mined.
|
||||
SendBatchBlocksMarginCheck int64
|
||||
// EthClientAttempts is the number of attempts to do an eth client RPC
|
||||
// call before giving up
|
||||
EthClientAttempts int
|
||||
// ForgeRetryInterval is the waiting interval between calls forge a
|
||||
// batch after an error
|
||||
ForgeRetryInterval time.Duration
|
||||
// ForgeDelay is the delay after which a batch is forged if the slot is
|
||||
// already committed. If set to 0s, the coordinator will continuously
|
||||
// forge at the maximum rate.
|
||||
ForgeDelay time.Duration
|
||||
// ForgeNoTxsDelay is the delay after which a batch is forged even if
|
||||
// there are no txs to forge if the slot is already committed. If set
|
||||
// to 0s, the coordinator will continuously forge even if the batches
|
||||
// are empty.
|
||||
ForgeNoTxsDelay time.Duration
|
||||
// SyncRetryInterval is the waiting interval between calls to the main
|
||||
// handler of a synced block after an error
|
||||
SyncRetryInterval time.Duration
|
||||
// PurgeByExtDelInterval is the waiting interval between calls
|
||||
// to the PurgeByExternalDelete function of the l2db which deletes
|
||||
// pending txs externally marked by the column `external_delete`
|
||||
PurgeByExtDelInterval time.Duration
|
||||
// EthClientAttemptsDelay is delay between attempts do do an eth client
|
||||
// RPC call
|
||||
EthClientAttemptsDelay time.Duration
|
||||
// EthTxResendTimeout is the timeout after which a non-mined ethereum
|
||||
// transaction will be resent (reusing the nonce) with a newly
|
||||
// calculated gas price
|
||||
EthTxResendTimeout time.Duration
|
||||
// EthNoReuseNonce disables reusing nonces of pending transactions for
|
||||
// new replacement transactions
|
||||
EthNoReuseNonce bool
|
||||
// MaxGasPrice is the maximum gas price allowed for ethereum
|
||||
// transactions
|
||||
MaxGasPrice *big.Int
|
||||
// GasPriceIncPerc is the percentage increase of gas price set in an
|
||||
// ethereum transaction from the suggested gas price by the ehtereum
|
||||
// node
|
||||
GasPriceIncPerc int64
|
||||
// TxManagerCheckInterval is the waiting interval between receipt
|
||||
// checks of ethereum transactions in the TxManager
|
||||
TxManagerCheckInterval time.Duration
|
||||
// DebugBatchPath if set, specifies the path where batchInfo is stored
|
||||
// in JSON in every step/update of the pipeline
|
||||
DebugBatchPath string
|
||||
Purger PurgerCfg
|
||||
VerifierIdx uint8
|
||||
DebugBatchPath string
|
||||
Purger PurgerCfg
|
||||
// VerifierIdx is the index of the verifier contract registered in the
|
||||
// smart contract
|
||||
VerifierIdx uint8
|
||||
// ForgeBatchGasCost contains the cost of each action in the
|
||||
// ForgeBatch transaction.
|
||||
ForgeBatchGasCost config.ForgeBatchGasCost
|
||||
TxProcessorConfig txprocessor.Config
|
||||
}
|
||||
|
||||
@@ -74,15 +133,22 @@ func (c *Config) debugBatchStore(batchInfo *BatchInfo) {
|
||||
}
|
||||
}
|
||||
|
||||
type fromBatch struct {
|
||||
BatchNum common.BatchNum
|
||||
ForgerAddr ethCommon.Address
|
||||
StateRoot *big.Int
|
||||
}
|
||||
|
||||
// Coordinator implements the Coordinator type
|
||||
type Coordinator struct {
|
||||
// State
|
||||
pipelineBatchNum common.BatchNum // batchNum from which we started the pipeline
|
||||
provers []prover.Client
|
||||
consts synchronizer.SCConsts
|
||||
vars synchronizer.SCVariables
|
||||
stats synchronizer.Stats
|
||||
started bool
|
||||
pipelineNum int // Pipeline sequential number. The first pipeline is 1
|
||||
pipelineFromBatch fromBatch // batch from which we started the pipeline
|
||||
provers []prover.Client
|
||||
consts synchronizer.SCConsts
|
||||
vars synchronizer.SCVariables
|
||||
stats synchronizer.Stats
|
||||
started bool
|
||||
|
||||
cfg Config
|
||||
|
||||
@@ -96,7 +162,17 @@ type Coordinator struct {
|
||||
wg sync.WaitGroup
|
||||
cancel context.CancelFunc
|
||||
|
||||
pipeline *Pipeline
|
||||
// mutexL2DBUpdateDelete protects updates to the L2DB so that
|
||||
// these two processes always happen exclusively:
|
||||
// - Pipeline taking pending txs, running through the TxProcessor and
|
||||
// marking selected txs as forging
|
||||
// - Coordinator deleting pending txs that have been marked with
|
||||
// `external_delete`.
|
||||
// Without this mutex, the coordinator could delete a pending txs that
|
||||
// has just been selected by the TxProcessor in the pipeline.
|
||||
mutexL2DBUpdateDelete sync.Mutex
|
||||
pipeline *Pipeline
|
||||
lastNonFailedBatchNum common.BatchNum
|
||||
|
||||
purger *Purger
|
||||
txManager *TxManager
|
||||
@@ -139,10 +215,15 @@ func NewCoordinator(cfg Config,
|
||||
|
||||
ctx, cancel := context.WithCancel(context.Background())
|
||||
c := Coordinator{
|
||||
pipelineBatchNum: -1,
|
||||
provers: serverProofs,
|
||||
consts: *scConsts,
|
||||
vars: *initSCVars,
|
||||
pipelineNum: 0,
|
||||
pipelineFromBatch: fromBatch{
|
||||
BatchNum: 0,
|
||||
ForgerAddr: ethCommon.Address{},
|
||||
StateRoot: big.NewInt(0),
|
||||
},
|
||||
provers: serverProofs,
|
||||
consts: *scConsts,
|
||||
vars: *initSCVars,
|
||||
|
||||
cfg: cfg,
|
||||
|
||||
@@ -183,8 +264,10 @@ func (c *Coordinator) BatchBuilder() *batchbuilder.BatchBuilder {
|
||||
}
|
||||
|
||||
func (c *Coordinator) newPipeline(ctx context.Context) (*Pipeline, error) {
|
||||
return NewPipeline(ctx, c.cfg, c.historyDB, c.l2DB, c.txSelector,
|
||||
c.batchBuilder, c.purger, c.txManager, c.provers, &c.consts)
|
||||
c.pipelineNum++
|
||||
return NewPipeline(ctx, c.cfg, c.pipelineNum, c.historyDB, c.l2DB, c.txSelector,
|
||||
c.batchBuilder, &c.mutexL2DBUpdateDelete, c.purger, c, c.txManager,
|
||||
c.provers, &c.consts)
|
||||
}
|
||||
|
||||
// MsgSyncBlock indicates an update to the Synchronizer stats
|
||||
@@ -205,6 +288,9 @@ type MsgSyncReorg struct {
|
||||
// MsgStopPipeline indicates a signal to reset the pipeline
|
||||
type MsgStopPipeline struct {
|
||||
Reason string
|
||||
// FailedBatchNum indicates the first batchNum that failed in the
|
||||
// pipeline. If FailedBatchNum is 0, it should be ignored.
|
||||
FailedBatchNum common.BatchNum
|
||||
}
|
||||
|
||||
// SendMsg is a thread safe method to pass a message to the Coordinator
|
||||
@@ -215,27 +301,36 @@ func (c *Coordinator) SendMsg(ctx context.Context, msg interface{}) {
|
||||
}
|
||||
}
|
||||
|
||||
func updateSCVars(vars *synchronizer.SCVariables, update synchronizer.SCVariablesPtr) {
|
||||
if update.Rollup != nil {
|
||||
vars.Rollup = *update.Rollup
|
||||
}
|
||||
if update.Auction != nil {
|
||||
vars.Auction = *update.Auction
|
||||
}
|
||||
if update.WDelayer != nil {
|
||||
vars.WDelayer = *update.WDelayer
|
||||
}
|
||||
}
|
||||
|
||||
func (c *Coordinator) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
if vars.Rollup != nil {
|
||||
c.vars.Rollup = *vars.Rollup
|
||||
}
|
||||
if vars.Auction != nil {
|
||||
c.vars.Auction = *vars.Auction
|
||||
}
|
||||
if vars.WDelayer != nil {
|
||||
c.vars.WDelayer = *vars.WDelayer
|
||||
}
|
||||
updateSCVars(&c.vars, vars)
|
||||
}
|
||||
|
||||
func canForge(auctionConstants *common.AuctionConstants, auctionVars *common.AuctionVariables,
|
||||
currentSlot *common.Slot, nextSlot *common.Slot, addr ethCommon.Address, blockNum int64) bool {
|
||||
if blockNum < auctionConstants.GenesisBlockNum {
|
||||
log.Infow("canForge: requested blockNum is < genesis", "blockNum", blockNum,
|
||||
"genesis", auctionConstants.GenesisBlockNum)
|
||||
return false
|
||||
}
|
||||
var slot *common.Slot
|
||||
if currentSlot.StartBlock <= blockNum && blockNum <= currentSlot.EndBlock {
|
||||
slot = currentSlot
|
||||
} else if nextSlot.StartBlock <= blockNum && blockNum <= nextSlot.EndBlock {
|
||||
slot = nextSlot
|
||||
} else {
|
||||
log.Warnw("Coordinator: requested blockNum for canForge is outside slot",
|
||||
log.Warnw("canForge: requested blockNum is outside current and next slot",
|
||||
"blockNum", blockNum, "currentSlot", currentSlot,
|
||||
"nextSlot", nextSlot,
|
||||
)
|
||||
@@ -244,16 +339,23 @@ func canForge(auctionConstants *common.AuctionConstants, auctionVars *common.Auc
|
||||
anyoneForge := false
|
||||
if !slot.ForgerCommitment &&
|
||||
auctionConstants.RelativeBlock(blockNum) >= int64(auctionVars.SlotDeadline) {
|
||||
log.Debugw("Coordinator: anyone can forge in the current slot (slotDeadline passed)",
|
||||
log.Debugw("canForge: anyone can forge in the current slot (slotDeadline passed)",
|
||||
"block", blockNum)
|
||||
anyoneForge = true
|
||||
}
|
||||
if slot.Forger == addr || anyoneForge {
|
||||
return true
|
||||
}
|
||||
log.Debugw("canForge: can't forge", "slot.Forger", slot.Forger)
|
||||
return false
|
||||
}
|
||||
|
||||
func (c *Coordinator) canForgeAt(blockNum int64) bool {
|
||||
return canForge(&c.consts.Auction, &c.vars.Auction,
|
||||
&c.stats.Sync.Auction.CurrentSlot, &c.stats.Sync.Auction.NextSlot,
|
||||
c.cfg.ForgerAddress, blockNum)
|
||||
}
|
||||
|
||||
func (c *Coordinator) canForge() bool {
|
||||
blockNum := c.stats.Eth.LastBlock.Num + 1
|
||||
return canForge(&c.consts.Auction, &c.vars.Auction,
|
||||
@@ -262,21 +364,51 @@ func (c *Coordinator) canForge() bool {
|
||||
}
|
||||
|
||||
func (c *Coordinator) syncStats(ctx context.Context, stats *synchronizer.Stats) error {
|
||||
canForge := c.canForge()
|
||||
nextBlock := c.stats.Eth.LastBlock.Num + 1
|
||||
canForge := c.canForgeAt(nextBlock)
|
||||
if c.cfg.ScheduleBatchBlocksAheadCheck != 0 && canForge {
|
||||
canForge = c.canForgeAt(nextBlock + c.cfg.ScheduleBatchBlocksAheadCheck)
|
||||
}
|
||||
if c.pipeline == nil {
|
||||
if canForge {
|
||||
relativeBlock := c.consts.Auction.RelativeBlock(nextBlock)
|
||||
if canForge && relativeBlock < c.cfg.StartSlotBlocksDelay {
|
||||
log.Debugf("Coordinator: delaying pipeline start due to "+
|
||||
"relativeBlock (%v) < cfg.StartSlotBlocksDelay (%v)",
|
||||
relativeBlock, c.cfg.StartSlotBlocksDelay)
|
||||
} else if canForge {
|
||||
log.Infow("Coordinator: forging state begin", "block",
|
||||
stats.Eth.LastBlock.Num+1, "batch", stats.Sync.LastBatch)
|
||||
batchNum := common.BatchNum(stats.Sync.LastBatch)
|
||||
stats.Eth.LastBlock.Num+1, "batch", stats.Sync.LastBatch.BatchNum)
|
||||
fromBatch := fromBatch{
|
||||
BatchNum: stats.Sync.LastBatch.BatchNum,
|
||||
ForgerAddr: stats.Sync.LastBatch.ForgerAddr,
|
||||
StateRoot: stats.Sync.LastBatch.StateRoot,
|
||||
}
|
||||
if c.lastNonFailedBatchNum > fromBatch.BatchNum {
|
||||
fromBatch.BatchNum = c.lastNonFailedBatchNum
|
||||
fromBatch.ForgerAddr = c.cfg.ForgerAddress
|
||||
fromBatch.StateRoot = big.NewInt(0)
|
||||
}
|
||||
// Before starting the pipeline make sure we reset any
|
||||
// l2tx from the pool that was forged in a batch that
|
||||
// didn't end up being mined. We are already doing
|
||||
// this in handleStopPipeline, but we do it again as a
|
||||
// failsafe in case the last synced batchnum is
|
||||
// different than in the previous call to l2DB.Reorg,
|
||||
// or in case the node was restarted when there was a
|
||||
// started batch that included l2txs but was not mined.
|
||||
if err := c.l2DB.Reorg(fromBatch.BatchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
var err error
|
||||
if c.pipeline, err = c.newPipeline(ctx); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if err := c.pipeline.Start(batchNum, stats, &c.vars); err != nil {
|
||||
c.pipelineFromBatch = fromBatch
|
||||
// Start the pipeline
|
||||
if err := c.pipeline.Start(fromBatch.BatchNum, stats, &c.vars); err != nil {
|
||||
c.pipeline = nil
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
c.pipelineBatchNum = batchNum
|
||||
}
|
||||
} else {
|
||||
if !canForge {
|
||||
@@ -286,25 +418,12 @@ func (c *Coordinator) syncStats(ctx context.Context, stats *synchronizer.Stats)
|
||||
}
|
||||
}
|
||||
if c.pipeline == nil {
|
||||
// Mark invalid in Pool due to forged L2Txs
|
||||
// for _, batch := range batches {
|
||||
// if err := c.l2DB.InvalidateOldNonces(
|
||||
// idxsNonceFromL2Txs(batch.L2Txs), batch.Batch.BatchNum); err != nil {
|
||||
// return err
|
||||
// }
|
||||
// }
|
||||
if c.purger.CanInvalidate(stats.Sync.LastBlock.Num, stats.Sync.LastBatch) {
|
||||
if err := c.txSelector.Reset(common.BatchNum(stats.Sync.LastBatch)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
_, err := c.purger.InvalidateMaybe(c.l2DB, c.txSelector.LocalAccountsDB(),
|
||||
stats.Sync.LastBlock.Num, stats.Sync.LastBatch)
|
||||
if err != nil {
|
||||
if _, err := c.purger.InvalidateMaybe(c.l2DB, c.txSelector.LocalAccountsDB(),
|
||||
stats.Sync.LastBlock.Num, int64(stats.Sync.LastBatch.BatchNum)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
_, err = c.purger.PurgeMaybe(c.l2DB, stats.Sync.LastBlock.Num, stats.Sync.LastBatch)
|
||||
if err != nil {
|
||||
if _, err := c.purger.PurgeMaybe(c.l2DB, stats.Sync.LastBlock.Num,
|
||||
int64(stats.Sync.LastBatch.BatchNum)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
@@ -331,33 +450,44 @@ func (c *Coordinator) handleReorg(ctx context.Context, msg *MsgSyncReorg) error
|
||||
if c.pipeline != nil {
|
||||
c.pipeline.SetSyncStatsVars(ctx, &msg.Stats, &msg.Vars)
|
||||
}
|
||||
if common.BatchNum(c.stats.Sync.LastBatch) < c.pipelineBatchNum {
|
||||
// There's been a reorg and the batch from which the pipeline
|
||||
// was started was in a block that was discarded. The batch
|
||||
// may not be in the main chain, so we stop the pipeline as a
|
||||
// precaution (it will be started again once the node is in
|
||||
// sync).
|
||||
log.Infow("Coordinator.handleReorg StopPipeline sync.LastBatch < c.pipelineBatchNum",
|
||||
"sync.LastBatch", c.stats.Sync.LastBatch,
|
||||
"c.pipelineBatchNum", c.pipelineBatchNum)
|
||||
if err := c.handleStopPipeline(ctx, "reorg"); err != nil {
|
||||
if c.stats.Sync.LastBatch.ForgerAddr != c.cfg.ForgerAddress &&
|
||||
(c.stats.Sync.LastBatch.StateRoot == nil || c.pipelineFromBatch.StateRoot == nil ||
|
||||
c.stats.Sync.LastBatch.StateRoot.Cmp(c.pipelineFromBatch.StateRoot) != 0) {
|
||||
// There's been a reorg and the batch state root from which the
|
||||
// pipeline was started has changed (probably because it was in
|
||||
// a block that was discarded), and it was sent by a different
|
||||
// coordinator than us. That batch may never be in the main
|
||||
// chain, so we stop the pipeline (it will be started again
|
||||
// once the node is in sync).
|
||||
log.Infow("Coordinator.handleReorg StopPipeline sync.LastBatch.ForgerAddr != cfg.ForgerAddr "+
|
||||
"& sync.LastBatch.StateRoot != pipelineFromBatch.StateRoot",
|
||||
"sync.LastBatch.StateRoot", c.stats.Sync.LastBatch.StateRoot,
|
||||
"pipelineFromBatch.StateRoot", c.pipelineFromBatch.StateRoot)
|
||||
c.txManager.DiscardPipeline(ctx, c.pipelineNum)
|
||||
if err := c.handleStopPipeline(ctx, "reorg", 0); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *Coordinator) handleStopPipeline(ctx context.Context, reason string) error {
|
||||
if err := c.l2DB.Reorg(common.BatchNum(c.stats.Sync.LastBatch)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
// handleStopPipeline handles stopping the pipeline. If failedBatchNum is 0,
|
||||
// the next pipeline will start from the last state of the synchronizer,
|
||||
// otherwise, it will state from failedBatchNum-1.
|
||||
func (c *Coordinator) handleStopPipeline(ctx context.Context, reason string,
|
||||
failedBatchNum common.BatchNum) error {
|
||||
batchNum := c.stats.Sync.LastBatch.BatchNum
|
||||
if failedBatchNum != 0 {
|
||||
batchNum = failedBatchNum - 1
|
||||
}
|
||||
if c.pipeline != nil {
|
||||
c.pipeline.Stop(c.ctx)
|
||||
c.pipeline = nil
|
||||
}
|
||||
if strings.Contains(reason, common.AuctionErrMsgCannotForge) { //nolint:staticcheck
|
||||
// TODO: Check that we are in a slot in which we can't forge
|
||||
if err := c.l2DB.Reorg(batchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
c.lastNonFailedBatchNum = batchNum
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -373,7 +503,7 @@ func (c *Coordinator) handleMsg(ctx context.Context, msg interface{}) error {
|
||||
}
|
||||
case MsgStopPipeline:
|
||||
log.Infow("Coordinator received MsgStopPipeline", "reason", msg.Reason)
|
||||
if err := c.handleStopPipeline(ctx, msg.Reason); err != nil {
|
||||
if err := c.handleStopPipeline(ctx, msg.Reason, msg.FailedBatchNum); err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("Coordinator.handleStopPipeline: %w", err))
|
||||
}
|
||||
default:
|
||||
@@ -396,7 +526,7 @@ func (c *Coordinator) Start() {
|
||||
|
||||
c.wg.Add(1)
|
||||
go func() {
|
||||
waitDuration := longWaitDuration
|
||||
timer := time.NewTimer(longWaitDuration)
|
||||
for {
|
||||
select {
|
||||
case <-c.ctx.Done():
|
||||
@@ -408,23 +538,45 @@ func (c *Coordinator) Start() {
|
||||
continue
|
||||
} else if err != nil {
|
||||
log.Errorw("Coordinator.handleMsg", "err", err)
|
||||
waitDuration = c.cfg.SyncRetryInterval
|
||||
if !timer.Stop() {
|
||||
<-timer.C
|
||||
}
|
||||
timer.Reset(c.cfg.SyncRetryInterval)
|
||||
continue
|
||||
}
|
||||
waitDuration = longWaitDuration
|
||||
case <-time.After(waitDuration):
|
||||
case <-timer.C:
|
||||
timer.Reset(longWaitDuration)
|
||||
if !c.stats.Synced() {
|
||||
waitDuration = longWaitDuration
|
||||
continue
|
||||
}
|
||||
if err := c.syncStats(c.ctx, &c.stats); c.ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
log.Errorw("Coordinator.syncStats", "err", err)
|
||||
waitDuration = c.cfg.SyncRetryInterval
|
||||
if !timer.Stop() {
|
||||
<-timer.C
|
||||
}
|
||||
timer.Reset(c.cfg.SyncRetryInterval)
|
||||
continue
|
||||
}
|
||||
waitDuration = longWaitDuration
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
c.wg.Add(1)
|
||||
go func() {
|
||||
for {
|
||||
select {
|
||||
case <-c.ctx.Done():
|
||||
log.Info("Coordinator L2DB.PurgeByExternalDelete loop done")
|
||||
c.wg.Done()
|
||||
return
|
||||
case <-time.After(c.cfg.PurgeByExtDelInterval):
|
||||
c.mutexL2DBUpdateDelete.Lock()
|
||||
if err := c.l2DB.PurgeByExternalDelete(); err != nil {
|
||||
log.Errorw("L2DB.PurgeByExternalDelete", "err", err)
|
||||
}
|
||||
c.mutexL2DBUpdateDelete.Unlock()
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
@@ -97,15 +97,16 @@ func newTestModules(t *testing.T) modules {
|
||||
syncDBPath, err = ioutil.TempDir("", "tmpSyncDB")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, syncDBPath)
|
||||
syncStateDB, err := statedb.NewStateDB(syncDBPath, 128, statedb.TypeSynchronizer, 48)
|
||||
syncStateDB, err := statedb.NewStateDB(statedb.Config{Path: syncDBPath, Keep: 128,
|
||||
Type: statedb.TypeSynchronizer, NLevels: 48})
|
||||
assert.NoError(t, err)
|
||||
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
test.WipeDB(db)
|
||||
l2DB := l2db.NewL2DB(db, 10, 100, 24*time.Hour)
|
||||
historyDB := historydb.NewHistoryDB(db)
|
||||
l2DB := l2db.NewL2DB(db, db, 10, 100, 0.0, 24*time.Hour, nil)
|
||||
historyDB := historydb.NewHistoryDB(db, db, nil)
|
||||
|
||||
txSelDBPath, err = ioutil.TempDir("", "tmpTxSelDB")
|
||||
require.NoError(t, err)
|
||||
@@ -125,7 +126,8 @@ func newTestModules(t *testing.T) modules {
|
||||
batchBuilderDBPath, err = ioutil.TempDir("", "tmpBatchBuilderDB")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, batchBuilderDBPath)
|
||||
batchBuilder, err := batchbuilder.NewBatchBuilder(batchBuilderDBPath, syncStateDB, 0, uint64(nLevels))
|
||||
batchBuilder, err := batchbuilder.NewBatchBuilder(batchBuilderDBPath, syncStateDB, 0,
|
||||
uint64(nLevels))
|
||||
assert.NoError(t, err)
|
||||
|
||||
return modules{
|
||||
@@ -204,7 +206,7 @@ func newTestCoordinator(t *testing.T, forgerAddr ethCommon.Address, ethClient *t
|
||||
|
||||
func newTestSynchronizer(t *testing.T, ethClient *test.Client, ethClientSetup *test.ClientSetup,
|
||||
modules modules) *synchronizer.Synchronizer {
|
||||
sync, err := synchronizer.NewSynchronizer(ethClient, modules.historyDB, modules.stateDB,
|
||||
sync, err := synchronizer.NewSynchronizer(ethClient, modules.historyDB, modules.l2DB, modules.stateDB,
|
||||
synchronizer.Config{
|
||||
StatsRefreshPeriod: 0 * time.Second,
|
||||
})
|
||||
@@ -260,8 +262,8 @@ func TestCoordinatorFlow(t *testing.T) {
|
||||
var stats synchronizer.Stats
|
||||
stats.Eth.LastBlock = *ethClient.CtlLastBlock()
|
||||
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||
stats.Eth.LastBatch = ethClient.CtlLastForgedBatch()
|
||||
stats.Sync.LastBatch = stats.Eth.LastBatch
|
||||
stats.Eth.LastBatchNum = ethClient.CtlLastForgedBatch()
|
||||
stats.Sync.LastBatch.BatchNum = common.BatchNum(stats.Eth.LastBatchNum)
|
||||
canForge, err := ethClient.AuctionCanForge(forger, blockNum+1)
|
||||
require.NoError(t, err)
|
||||
var slot common.Slot
|
||||
@@ -278,7 +280,7 @@ func TestCoordinatorFlow(t *testing.T) {
|
||||
// Copy stateDB to synchronizer if there was a new batch
|
||||
source := fmt.Sprintf("%v/BatchNum%v", batchBuilderDBPath, stats.Sync.LastBatch)
|
||||
dest := fmt.Sprintf("%v/BatchNum%v", syncDBPath, stats.Sync.LastBatch)
|
||||
if stats.Sync.LastBatch != 0 {
|
||||
if stats.Sync.LastBatch.BatchNum != 0 {
|
||||
if _, err := os.Stat(dest); os.IsNotExist(err) {
|
||||
log.Infow("Making pebble checkpoint for sync",
|
||||
"source", source, "dest", dest)
|
||||
@@ -516,7 +518,7 @@ func TestCoordinatorStress(t *testing.T) {
|
||||
wg.Add(1)
|
||||
go func() {
|
||||
for {
|
||||
blockData, _, err := syn.Sync2(ctx, nil)
|
||||
blockData, _, err := syn.Sync(ctx, nil)
|
||||
if ctx.Err() != nil {
|
||||
wg.Done()
|
||||
return
|
||||
|
||||
@@ -2,6 +2,7 @@ package coordinator
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"sync"
|
||||
@@ -24,25 +25,36 @@ type statsVars struct {
|
||||
Vars synchronizer.SCVariablesPtr
|
||||
}
|
||||
|
||||
type state struct {
|
||||
batchNum common.BatchNum
|
||||
lastScheduledL1BatchBlockNum int64
|
||||
lastForgeL1TxsNum int64
|
||||
lastSlotForged int64
|
||||
}
|
||||
|
||||
// Pipeline manages the forging of batches with parallel server proofs
|
||||
type Pipeline struct {
|
||||
num int
|
||||
cfg Config
|
||||
consts synchronizer.SCConsts
|
||||
|
||||
// state
|
||||
batchNum common.BatchNum
|
||||
lastScheduledL1BatchBlockNum int64
|
||||
lastForgeL1TxsNum int64
|
||||
started bool
|
||||
state state
|
||||
started bool
|
||||
rw sync.RWMutex
|
||||
errAtBatchNum common.BatchNum
|
||||
lastForgeTime time.Time
|
||||
|
||||
proversPool *ProversPool
|
||||
provers []prover.Client
|
||||
txManager *TxManager
|
||||
historyDB *historydb.HistoryDB
|
||||
l2DB *l2db.L2DB
|
||||
txSelector *txselector.TxSelector
|
||||
batchBuilder *batchbuilder.BatchBuilder
|
||||
purger *Purger
|
||||
proversPool *ProversPool
|
||||
provers []prover.Client
|
||||
coord *Coordinator
|
||||
txManager *TxManager
|
||||
historyDB *historydb.HistoryDB
|
||||
l2DB *l2db.L2DB
|
||||
txSelector *txselector.TxSelector
|
||||
batchBuilder *batchbuilder.BatchBuilder
|
||||
mutexL2DBUpdateDelete *sync.Mutex
|
||||
purger *Purger
|
||||
|
||||
stats synchronizer.Stats
|
||||
vars synchronizer.SCVariables
|
||||
@@ -53,14 +65,29 @@ type Pipeline struct {
|
||||
cancel context.CancelFunc
|
||||
}
|
||||
|
||||
func (p *Pipeline) setErrAtBatchNum(batchNum common.BatchNum) {
|
||||
p.rw.Lock()
|
||||
defer p.rw.Unlock()
|
||||
p.errAtBatchNum = batchNum
|
||||
}
|
||||
|
||||
func (p *Pipeline) getErrAtBatchNum() common.BatchNum {
|
||||
p.rw.RLock()
|
||||
defer p.rw.RUnlock()
|
||||
return p.errAtBatchNum
|
||||
}
|
||||
|
||||
// NewPipeline creates a new Pipeline
|
||||
func NewPipeline(ctx context.Context,
|
||||
cfg Config,
|
||||
num int, // Pipeline sequential number
|
||||
historyDB *historydb.HistoryDB,
|
||||
l2DB *l2db.L2DB,
|
||||
txSelector *txselector.TxSelector,
|
||||
batchBuilder *batchbuilder.BatchBuilder,
|
||||
mutexL2DBUpdateDelete *sync.Mutex,
|
||||
purger *Purger,
|
||||
coord *Coordinator,
|
||||
txManager *TxManager,
|
||||
provers []prover.Client,
|
||||
scConsts *synchronizer.SCConsts,
|
||||
@@ -79,22 +106,26 @@ func NewPipeline(ctx context.Context,
|
||||
return nil, tracerr.Wrap(fmt.Errorf("no provers in the pool"))
|
||||
}
|
||||
return &Pipeline{
|
||||
cfg: cfg,
|
||||
historyDB: historyDB,
|
||||
l2DB: l2DB,
|
||||
txSelector: txSelector,
|
||||
batchBuilder: batchBuilder,
|
||||
provers: provers,
|
||||
proversPool: proversPool,
|
||||
purger: purger,
|
||||
txManager: txManager,
|
||||
consts: *scConsts,
|
||||
statsVarsCh: make(chan statsVars, queueLen),
|
||||
num: num,
|
||||
cfg: cfg,
|
||||
historyDB: historyDB,
|
||||
l2DB: l2DB,
|
||||
txSelector: txSelector,
|
||||
batchBuilder: batchBuilder,
|
||||
provers: provers,
|
||||
proversPool: proversPool,
|
||||
mutexL2DBUpdateDelete: mutexL2DBUpdateDelete,
|
||||
purger: purger,
|
||||
coord: coord,
|
||||
txManager: txManager,
|
||||
consts: *scConsts,
|
||||
statsVarsCh: make(chan statsVars, queueLen),
|
||||
}, nil
|
||||
}
|
||||
|
||||
// SetSyncStatsVars is a thread safe method to sets the synchronizer Stats
|
||||
func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats, vars *synchronizer.SCVariablesPtr) {
|
||||
func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars *synchronizer.SCVariablesPtr) {
|
||||
select {
|
||||
case p.statsVarsCh <- statsVars{Stats: *stats, Vars: *vars}:
|
||||
case <-ctx.Done():
|
||||
@@ -104,68 +135,121 @@ func (p *Pipeline) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Sta
|
||||
// reset pipeline state
|
||||
func (p *Pipeline) reset(batchNum common.BatchNum,
|
||||
stats *synchronizer.Stats, vars *synchronizer.SCVariables) error {
|
||||
p.batchNum = batchNum
|
||||
p.lastForgeL1TxsNum = stats.Sync.LastForgeL1TxsNum
|
||||
p.state = state{
|
||||
batchNum: batchNum,
|
||||
lastForgeL1TxsNum: stats.Sync.LastForgeL1TxsNum,
|
||||
lastScheduledL1BatchBlockNum: 0,
|
||||
lastSlotForged: -1,
|
||||
}
|
||||
p.stats = *stats
|
||||
p.vars = *vars
|
||||
p.lastScheduledL1BatchBlockNum = 0
|
||||
|
||||
err := p.txSelector.Reset(p.batchNum)
|
||||
// Reset the StateDB in TxSelector and BatchBuilder from the
|
||||
// synchronizer only if the checkpoint we reset from either:
|
||||
// a. Doesn't exist in the TxSelector/BatchBuilder
|
||||
// b. The batch has already been synced by the synchronizer and has a
|
||||
// different MTRoot than the BatchBuilder
|
||||
// Otherwise, reset from the local checkpoint.
|
||||
|
||||
// First attempt to reset from local checkpoint if such checkpoint exists
|
||||
existsTxSelector, err := p.txSelector.LocalAccountsDB().CheckpointExists(p.state.batchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
err = p.batchBuilder.Reset(p.batchNum, true)
|
||||
fromSynchronizerTxSelector := !existsTxSelector
|
||||
if err := p.txSelector.Reset(p.state.batchNum, fromSynchronizerTxSelector); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
existsBatchBuilder, err := p.batchBuilder.LocalStateDB().CheckpointExists(p.state.batchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
fromSynchronizerBatchBuilder := !existsBatchBuilder
|
||||
if err := p.batchBuilder.Reset(p.state.batchNum, fromSynchronizerBatchBuilder); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// After reset, check that if the batch exists in the historyDB, the
|
||||
// stateRoot matches with the local one, if not, force a reset from
|
||||
// synchronizer
|
||||
batch, err := p.historyDB.GetBatch(p.state.batchNum)
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
// nothing to do
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
} else {
|
||||
localStateRoot := p.batchBuilder.LocalStateDB().MT.Root().BigInt()
|
||||
if batch.StateRoot.Cmp(localStateRoot) != 0 {
|
||||
log.Debugw("localStateRoot (%v) != historyDB stateRoot (%v). "+
|
||||
"Forcing reset from Synchronizer", localStateRoot, batch.StateRoot)
|
||||
// StateRoot from synchronizer doesn't match StateRoot
|
||||
// from batchBuilder, force a reset from synchronizer
|
||||
if err := p.txSelector.Reset(p.state.batchNum, true); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if err := p.batchBuilder.Reset(p.state.batchNum, true); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (p *Pipeline) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
if vars.Rollup != nil {
|
||||
p.vars.Rollup = *vars.Rollup
|
||||
}
|
||||
if vars.Auction != nil {
|
||||
p.vars.Auction = *vars.Auction
|
||||
}
|
||||
if vars.WDelayer != nil {
|
||||
p.vars.WDelayer = *vars.WDelayer
|
||||
}
|
||||
updateSCVars(&p.vars, vars)
|
||||
}
|
||||
|
||||
// handleForgeBatch calls p.forgeBatch to forge the batch and get the zkInputs,
|
||||
// and then waits for an available proof server and sends the zkInputs to it so
|
||||
// that the proof computation begins.
|
||||
func (p *Pipeline) handleForgeBatch(ctx context.Context, batchNum common.BatchNum) (*BatchInfo, error) {
|
||||
batchInfo, err := p.forgeBatch(batchNum)
|
||||
if ctx.Err() != nil {
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
if tracerr.Unwrap(err) == errLastL1BatchNotSynced {
|
||||
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err,
|
||||
"lastForgeL1TxsNum", p.lastForgeL1TxsNum,
|
||||
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum)
|
||||
} else {
|
||||
log.Errorw("forgeBatch", "err", err)
|
||||
}
|
||||
return nil, err
|
||||
}
|
||||
// 6. Wait for an available server proof (blocking call)
|
||||
// handleForgeBatch waits for an available proof server, calls p.forgeBatch to
|
||||
// forge the batch and get the zkInputs, and then sends the zkInputs to the
|
||||
// selected proof server so that the proof computation begins.
|
||||
func (p *Pipeline) handleForgeBatch(ctx context.Context,
|
||||
batchNum common.BatchNum) (batchInfo *BatchInfo, err error) {
|
||||
// 1. Wait for an available serverProof (blocking call)
|
||||
serverProof, err := p.proversPool.Get(ctx)
|
||||
if ctx.Err() != nil {
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
log.Errorw("proversPool.Get", "err", err)
|
||||
return nil, err
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
defer func() {
|
||||
// If we encounter any error (notice that this function returns
|
||||
// errors to notify that a batch is not forged not only because
|
||||
// of unexpected errors but also due to benign causes), add the
|
||||
// serverProof back to the pool
|
||||
if err != nil {
|
||||
p.proversPool.Add(ctx, serverProof)
|
||||
}
|
||||
}()
|
||||
|
||||
// 2. Forge the batch internally (make a selection of txs and prepare
|
||||
// all the smart contract arguments)
|
||||
p.mutexL2DBUpdateDelete.Lock()
|
||||
batchInfo, err = p.forgeBatch(batchNum)
|
||||
p.mutexL2DBUpdateDelete.Unlock()
|
||||
if ctx.Err() != nil {
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
if tracerr.Unwrap(err) == errLastL1BatchNotSynced {
|
||||
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err,
|
||||
"lastForgeL1TxsNum", p.state.lastForgeL1TxsNum,
|
||||
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum)
|
||||
} else if tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||
// no log
|
||||
} else {
|
||||
log.Errorw("forgeBatch", "err", err)
|
||||
}
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// 3. Send the ZKInputs to the proof server
|
||||
batchInfo.ServerProof = serverProof
|
||||
if err := p.sendServerProof(ctx, batchInfo); ctx.Err() != nil {
|
||||
return nil, ctx.Err()
|
||||
} else if err != nil {
|
||||
log.Errorw("sendServerProof", "err", err)
|
||||
batchInfo.ServerProof = nil
|
||||
p.proversPool.Add(ctx, serverProof)
|
||||
return nil, err
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return batchInfo, nil
|
||||
}
|
||||
@@ -188,7 +272,7 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||
|
||||
p.wg.Add(1)
|
||||
go func() {
|
||||
waitDuration := zeroDuration
|
||||
timer := time.NewTimer(zeroDuration)
|
||||
for {
|
||||
select {
|
||||
case <-p.ctx.Done():
|
||||
@@ -198,18 +282,42 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||
case statsVars := <-p.statsVarsCh:
|
||||
p.stats = statsVars.Stats
|
||||
p.syncSCVars(statsVars.Vars)
|
||||
case <-time.After(waitDuration):
|
||||
batchNum = p.batchNum + 1
|
||||
if batchInfo, err := p.handleForgeBatch(p.ctx, batchNum); err != nil {
|
||||
waitDuration = p.cfg.SyncRetryInterval
|
||||
case <-timer.C:
|
||||
timer.Reset(p.cfg.ForgeRetryInterval)
|
||||
// Once errAtBatchNum != 0, we stop forging
|
||||
// batches because there's been an error and we
|
||||
// wait for the pipeline to be stopped.
|
||||
if p.getErrAtBatchNum() != 0 {
|
||||
continue
|
||||
} else {
|
||||
p.batchNum = batchNum
|
||||
select {
|
||||
case batchChSentServerProof <- batchInfo:
|
||||
case <-p.ctx.Done():
|
||||
}
|
||||
}
|
||||
batchNum = p.state.batchNum + 1
|
||||
batchInfo, err := p.handleForgeBatch(p.ctx, batchNum)
|
||||
if p.ctx.Err() != nil {
|
||||
continue
|
||||
} else if tracerr.Unwrap(err) == errLastL1BatchNotSynced ||
|
||||
tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||
continue
|
||||
} else if err != nil {
|
||||
p.setErrAtBatchNum(batchNum)
|
||||
p.coord.SendMsg(p.ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf(
|
||||
"Pipeline.handleForgBatch: %v", err),
|
||||
FailedBatchNum: batchNum,
|
||||
})
|
||||
continue
|
||||
}
|
||||
p.lastForgeTime = time.Now()
|
||||
|
||||
p.state.batchNum = batchNum
|
||||
select {
|
||||
case batchChSentServerProof <- batchInfo:
|
||||
case <-p.ctx.Done():
|
||||
}
|
||||
if !timer.Stop() {
|
||||
<-timer.C
|
||||
}
|
||||
timer.Reset(zeroDuration)
|
||||
}
|
||||
}
|
||||
}()
|
||||
@@ -223,16 +331,27 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||
p.wg.Done()
|
||||
return
|
||||
case batchInfo := <-batchChSentServerProof:
|
||||
// Once errAtBatchNum != 0, we stop forging
|
||||
// batches because there's been an error and we
|
||||
// wait for the pipeline to be stopped.
|
||||
if p.getErrAtBatchNum() != 0 {
|
||||
continue
|
||||
}
|
||||
err := p.waitServerProof(p.ctx, batchInfo)
|
||||
// We are done with this serverProof, add it back to the pool
|
||||
p.proversPool.Add(p.ctx, batchInfo.ServerProof)
|
||||
batchInfo.ServerProof = nil
|
||||
if p.ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
log.Errorw("waitServerProof", "err", err)
|
||||
p.setErrAtBatchNum(batchInfo.BatchNum)
|
||||
p.coord.SendMsg(p.ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf(
|
||||
"Pipeline.waitServerProof: %v", err),
|
||||
FailedBatchNum: batchInfo.BatchNum,
|
||||
})
|
||||
continue
|
||||
}
|
||||
// We are done with this serverProof, add it back to the pool
|
||||
p.proversPool.Add(p.ctx, batchInfo.ServerProof)
|
||||
p.txManager.AddBatch(p.ctx, batchInfo)
|
||||
}
|
||||
}
|
||||
@@ -282,56 +401,94 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
batchInfo = &BatchInfo{BatchNum: batchNum} // to accumulate metadata of the batch
|
||||
batchInfo.Debug.StartTimestamp = time.Now()
|
||||
// Structure to accumulate data and metadata of the batch
|
||||
now := time.Now()
|
||||
batchInfo = &BatchInfo{PipelineNum: p.num, BatchNum: batchNum}
|
||||
batchInfo.Debug.StartTimestamp = now
|
||||
batchInfo.Debug.StartBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||
|
||||
selectionCfg := &txselector.SelectionConfig{
|
||||
MaxL1UserTxs: common.RollupConstMaxL1UserTx,
|
||||
TxProcessorConfig: p.cfg.TxProcessorConfig,
|
||||
}
|
||||
|
||||
var poolL2Txs []common.PoolL2Tx
|
||||
var discardedL2Txs []common.PoolL2Tx
|
||||
var l1UserTxsExtra, l1CoordTxs []common.L1Tx
|
||||
var auths [][]byte
|
||||
var coordIdxs []common.Idx
|
||||
|
||||
// Check if the slot is not yet fulfilled
|
||||
slotCommitted := false
|
||||
if p.stats.Sync.Auction.CurrentSlot.ForgerCommitment ||
|
||||
p.stats.Sync.Auction.CurrentSlot.SlotNum == p.state.lastSlotForged {
|
||||
slotCommitted = true
|
||||
}
|
||||
|
||||
// If we haven't reached the ForgeDelay, skip forging the batch
|
||||
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeDelay {
|
||||
return nil, tracerr.Wrap(errForgeBeforeDelay)
|
||||
}
|
||||
|
||||
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||
if p.shouldL1L2Batch(batchInfo) {
|
||||
batchInfo.L1Batch = true
|
||||
defer func() {
|
||||
// If there's no error, update the parameters related
|
||||
// to the last L1Batch forged
|
||||
if err == nil {
|
||||
p.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||
p.lastForgeL1TxsNum++
|
||||
}
|
||||
}()
|
||||
if p.lastForgeL1TxsNum != p.stats.Sync.LastForgeL1TxsNum {
|
||||
if p.state.lastForgeL1TxsNum != p.stats.Sync.LastForgeL1TxsNum {
|
||||
return nil, tracerr.Wrap(errLastL1BatchNotSynced)
|
||||
}
|
||||
// 2a: L1+L2 txs
|
||||
l1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(p.lastForgeL1TxsNum + 1)
|
||||
l1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(p.state.lastForgeL1TxsNum + 1)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
coordIdxs, auths, l1UserTxsExtra, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||
p.txSelector.GetL1L2TxSelection(selectionCfg, l1UserTxs)
|
||||
p.txSelector.GetL1L2TxSelection(p.cfg.TxProcessorConfig, l1UserTxs)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
} else {
|
||||
// 2b: only L2 txs
|
||||
coordIdxs, auths, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||
p.txSelector.GetL2TxSelection(selectionCfg)
|
||||
p.txSelector.GetL2TxSelection(p.cfg.TxProcessorConfig)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
l1UserTxsExtra = nil
|
||||
}
|
||||
|
||||
// If there are no txs to forge, no l1UserTxs in the open queue to
|
||||
// freeze, and we haven't reached the ForgeNoTxsDelay, skip forging the
|
||||
// batch.
|
||||
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeNoTxsDelay {
|
||||
noTxs := false
|
||||
if len(l1UserTxsExtra) == 0 && len(l1CoordTxs) == 0 && len(poolL2Txs) == 0 {
|
||||
if batchInfo.L1Batch {
|
||||
// Query the number of unforged L1UserTxs
|
||||
// (either in a open queue or in a frozen
|
||||
// not-yet-forged queue).
|
||||
count, err := p.historyDB.GetUnforgedL1UserTxsCount()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
// If there are future L1UserTxs, we forge a
|
||||
// batch to advance the queues to be able to
|
||||
// forge the L1UserTxs in the future.
|
||||
// Otherwise, skip.
|
||||
if count == 0 {
|
||||
noTxs = true
|
||||
}
|
||||
} else {
|
||||
noTxs = true
|
||||
}
|
||||
}
|
||||
if noTxs {
|
||||
if err := p.txSelector.Reset(batchInfo.BatchNum-1, false); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
return nil, tracerr.Wrap(errForgeNoTxsBeforeDelay)
|
||||
}
|
||||
}
|
||||
|
||||
if batchInfo.L1Batch {
|
||||
p.state.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||
p.state.lastForgeL1TxsNum++
|
||||
}
|
||||
|
||||
// 3. Save metadata from TxSelector output for BatchNum
|
||||
batchInfo.L1UserTxsExtra = l1UserTxsExtra
|
||||
batchInfo.L1CoordTxs = l1CoordTxs
|
||||
@@ -339,14 +496,15 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
||||
batchInfo.CoordIdxs = coordIdxs
|
||||
batchInfo.VerifierIdx = p.cfg.VerifierIdx
|
||||
|
||||
if err := p.l2DB.StartForging(common.TxIDsFromPoolL2Txs(poolL2Txs), batchInfo.BatchNum); err != nil {
|
||||
if err := p.l2DB.StartForging(common.TxIDsFromPoolL2Txs(poolL2Txs),
|
||||
batchInfo.BatchNum); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if err := p.l2DB.UpdateTxsInfo(discardedL2Txs); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// Invalidate transactions that become invalid beause of
|
||||
// Invalidate transactions that become invalid because of
|
||||
// the poolL2Txs selected. Will mark as invalid the txs that have a
|
||||
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||
// all the nonces smaller than the current one)
|
||||
@@ -376,12 +534,15 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
||||
p.cfg.debugBatchStore(batchInfo)
|
||||
log.Infow("Pipeline: batch forged internally", "batch", batchInfo.BatchNum)
|
||||
|
||||
p.state.lastSlotForged = p.stats.Sync.Auction.CurrentSlot.SlotNum
|
||||
|
||||
return batchInfo, nil
|
||||
}
|
||||
|
||||
// waitServerProof gets the generated zkProof & sends it to the SmartContract
|
||||
func (p *Pipeline) waitServerProof(ctx context.Context, batchInfo *BatchInfo) error {
|
||||
proof, pubInputs, err := batchInfo.ServerProof.GetProof(ctx) // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
proof, pubInputs, err := batchInfo.ServerProof.GetProof(ctx) // blocking call,
|
||||
// until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -397,12 +558,12 @@ func (p *Pipeline) waitServerProof(ctx context.Context, batchInfo *BatchInfo) er
|
||||
func (p *Pipeline) shouldL1L2Batch(batchInfo *BatchInfo) bool {
|
||||
// Take the lastL1BatchBlockNum as the biggest between the last
|
||||
// scheduled one, and the synchronized one.
|
||||
lastL1BatchBlockNum := p.lastScheduledL1BatchBlockNum
|
||||
lastL1BatchBlockNum := p.state.lastScheduledL1BatchBlockNum
|
||||
if p.stats.Sync.LastL1BatchBlock > lastL1BatchBlockNum {
|
||||
lastL1BatchBlockNum = p.stats.Sync.LastL1BatchBlock
|
||||
}
|
||||
// Set Debug information
|
||||
batchInfo.Debug.LastScheduledL1BatchBlockNum = p.lastScheduledL1BatchBlockNum
|
||||
batchInfo.Debug.LastScheduledL1BatchBlockNum = p.state.lastScheduledL1BatchBlockNum
|
||||
batchInfo.Debug.LastL1BatchBlock = p.stats.Sync.LastL1BatchBlock
|
||||
batchInfo.Debug.LastL1BatchBlockDelta = p.stats.Eth.LastBlock.Num + 1 - lastL1BatchBlockNum
|
||||
batchInfo.Debug.L1BatchBlockScheduleDeadline =
|
||||
|
||||
@@ -25,6 +25,14 @@ import (
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func newBigInt(s string) *big.Int {
|
||||
v, ok := new(big.Int).SetString(s, 10)
|
||||
if !ok {
|
||||
panic(fmt.Errorf("Can't set big.Int from %s", s))
|
||||
}
|
||||
return v
|
||||
}
|
||||
|
||||
func TestPipelineShouldL1L2Batch(t *testing.T) {
|
||||
ethClientSetup := test.NewClientSetupExample()
|
||||
ethClientSetup.ChainID = big.NewInt(int64(chainID))
|
||||
@@ -77,7 +85,7 @@ func TestPipelineShouldL1L2Batch(t *testing.T) {
|
||||
//
|
||||
// Scheduled L1Batch
|
||||
//
|
||||
pipeline.lastScheduledL1BatchBlockNum = startBlock
|
||||
pipeline.state.lastScheduledL1BatchBlockNum = startBlock
|
||||
stats.Sync.LastL1BatchBlock = startBlock - 10
|
||||
|
||||
// We are are one block before the timeout range * 0.5
|
||||
@@ -128,6 +136,11 @@ func preloadSync(t *testing.T, ethClient *test.Client, sync *synchronizer.Synchr
|
||||
blocks, err := tc.GenerateBlocksFromInstructions(set)
|
||||
require.NoError(t, err)
|
||||
require.NotNil(t, blocks)
|
||||
// Set StateRoots for batches manually (til doesn't set it)
|
||||
blocks[0].Rollup.Batches[0].Batch.StateRoot =
|
||||
newBigInt("0")
|
||||
blocks[0].Rollup.Batches[1].Batch.StateRoot =
|
||||
newBigInt("6860514559199319426609623120853503165917774887908204288119245630904770452486")
|
||||
|
||||
ethAddTokens(blocks, ethClient)
|
||||
err = ethClient.CtlAddBlocks(blocks)
|
||||
@@ -135,7 +148,7 @@ func preloadSync(t *testing.T, ethClient *test.Client, sync *synchronizer.Synchr
|
||||
|
||||
ctx := context.Background()
|
||||
for {
|
||||
syncBlock, discards, err := sync.Sync2(ctx, nil)
|
||||
syncBlock, discards, err := sync.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
if syncBlock == nil {
|
||||
@@ -172,7 +185,7 @@ func TestPipelineForgeBatchWithTxs(t *testing.T) {
|
||||
// users with positive balances
|
||||
tilCtx := preloadSync(t, ethClient, sync, modules.historyDB, modules.stateDB)
|
||||
syncStats := sync.Stats()
|
||||
batchNum := common.BatchNum(syncStats.Sync.LastBatch)
|
||||
batchNum := syncStats.Sync.LastBatch.BatchNum
|
||||
syncSCVars := sync.SCVars()
|
||||
|
||||
pipeline, err := coord.newPipeline(ctx)
|
||||
|
||||
@@ -13,13 +13,23 @@ import (
|
||||
|
||||
// PurgerCfg is the purger configuration
|
||||
type PurgerCfg struct {
|
||||
// PurgeBatchDelay is the delay between batches to purge outdated transactions
|
||||
// PurgeBatchDelay is the delay between batches to purge outdated
|
||||
// transactions. Outdated L2Txs are those that have been forged or
|
||||
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||
// that have been in the pool for longer than TTL once there are
|
||||
// MaxTxs.
|
||||
PurgeBatchDelay int64
|
||||
// InvalidateBatchDelay is the delay between batches to mark invalid transactions
|
||||
// InvalidateBatchDelay is the delay between batches to mark invalid
|
||||
// transactions due to nonce lower than the account nonce.
|
||||
InvalidateBatchDelay int64
|
||||
// PurgeBlockDelay is the delay between blocks to purge outdated transactions
|
||||
// PurgeBlockDelay is the delay between blocks to purge outdated
|
||||
// transactions. Outdated L2Txs are those that have been forged or
|
||||
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||
// that have been in the pool for longer than TTL once there are
|
||||
// MaxTxs.
|
||||
PurgeBlockDelay int64
|
||||
// InvalidateBlockDelay is the delay between blocks to mark invalid transactions
|
||||
// InvalidateBlockDelay is the delay between blocks to mark invalid
|
||||
// transactions due to nonce lower than the account nonce.
|
||||
InvalidateBlockDelay int64
|
||||
}
|
||||
|
||||
|
||||
@@ -21,19 +21,21 @@ func newL2DB(t *testing.T) *l2db.L2DB {
|
||||
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
test.WipeDB(db)
|
||||
return l2db.NewL2DB(db, 10, 100, 24*time.Hour)
|
||||
return l2db.NewL2DB(db, db, 10, 100, 0.0, 24*time.Hour, nil)
|
||||
}
|
||||
|
||||
func newStateDB(t *testing.T) *statedb.LocalStateDB {
|
||||
syncDBPath, err := ioutil.TempDir("", "tmpSyncDB")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, syncDBPath)
|
||||
syncStateDB, err := statedb.NewStateDB(syncDBPath, 128, statedb.TypeSynchronizer, 48)
|
||||
syncStateDB, err := statedb.NewStateDB(statedb.Config{Path: syncDBPath, Keep: 128,
|
||||
Type: statedb.TypeSynchronizer, NLevels: 48})
|
||||
assert.NoError(t, err)
|
||||
stateDBPath, err := ioutil.TempDir("", "tmpStateDB")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, stateDBPath)
|
||||
stateDB, err := statedb.NewLocalStateDB(stateDBPath, 128, syncStateDB, statedb.TypeTxSelector, 0)
|
||||
stateDB, err := statedb.NewLocalStateDB(statedb.Config{Path: stateDBPath, Keep: 128,
|
||||
Type: statedb.TypeTxSelector, NLevels: 0}, syncStateDB)
|
||||
require.NoError(t, err)
|
||||
return stateDB
|
||||
}
|
||||
|
||||
@@ -4,11 +4,13 @@ import (
|
||||
"context"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/ethereum/go-ethereum"
|
||||
"github.com/ethereum/go-ethereum/accounts"
|
||||
"github.com/ethereum/go-ethereum/accounts/abi/bind"
|
||||
"github.com/ethereum/go-ethereum/core"
|
||||
"github.com/ethereum/go-ethereum/core/types"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
@@ -35,17 +37,26 @@ type TxManager struct {
|
||||
vars synchronizer.SCVariables
|
||||
statsVarsCh chan statsVars
|
||||
|
||||
queue []*BatchInfo
|
||||
discardPipelineCh chan int // int refers to the pipelineNum
|
||||
|
||||
minPipelineNum int
|
||||
queue Queue
|
||||
// lastSuccessBatch stores the last BatchNum that who's forge call was confirmed
|
||||
lastSuccessBatch common.BatchNum
|
||||
lastPendingBatch common.BatchNum
|
||||
lastSuccessNonce uint64
|
||||
lastPendingNonce uint64
|
||||
// lastPendingBatch common.BatchNum
|
||||
// accNonce is the account nonce in the last mined block (due to mined txs)
|
||||
accNonce uint64
|
||||
// accNextNonce is the nonce that we should use to send the next tx.
|
||||
// In some cases this will be a reused nonce of an already pending tx.
|
||||
accNextNonce uint64
|
||||
|
||||
lastSentL1BatchBlockNum int64
|
||||
}
|
||||
|
||||
// NewTxManager creates a new TxManager
|
||||
func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterface, l2DB *l2db.L2DB,
|
||||
coord *Coordinator, scConsts *synchronizer.SCConsts, initSCVars *synchronizer.SCVariables) (*TxManager, error) {
|
||||
coord *Coordinator, scConsts *synchronizer.SCConsts, initSCVars *synchronizer.SCVariables) (
|
||||
*TxManager, error) {
|
||||
chainID, err := ethClient.EthChainID()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -54,26 +65,19 @@ func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterfac
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
lastSuccessNonce, err := ethClient.EthNonceAt(ctx, *address, nil)
|
||||
accNonce, err := ethClient.EthNonceAt(ctx, *address, nil)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
lastPendingNonce, err := ethClient.EthPendingNonceAt(ctx, *address)
|
||||
if err != nil {
|
||||
return nil, err
|
||||
}
|
||||
if lastSuccessNonce != lastPendingNonce {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("lastSuccessNonce (%v) != lastPendingNonce (%v)",
|
||||
lastSuccessNonce, lastPendingNonce))
|
||||
}
|
||||
log.Infow("TxManager started", "nonce", lastSuccessNonce)
|
||||
log.Infow("TxManager started", "nonce", accNonce)
|
||||
return &TxManager{
|
||||
cfg: *cfg,
|
||||
ethClient: ethClient,
|
||||
l2DB: l2DB,
|
||||
coord: coord,
|
||||
batchCh: make(chan *BatchInfo, queueLen),
|
||||
statsVarsCh: make(chan statsVars, queueLen),
|
||||
cfg: *cfg,
|
||||
ethClient: ethClient,
|
||||
l2DB: l2DB,
|
||||
coord: coord,
|
||||
batchCh: make(chan *BatchInfo, queueLen),
|
||||
statsVarsCh: make(chan statsVars, queueLen),
|
||||
discardPipelineCh: make(chan int, queueLen),
|
||||
account: accounts.Account{
|
||||
Address: *address,
|
||||
},
|
||||
@@ -82,8 +86,10 @@ func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterfac
|
||||
|
||||
vars: *initSCVars,
|
||||
|
||||
lastSuccessNonce: lastSuccessNonce,
|
||||
lastPendingNonce: lastPendingNonce,
|
||||
minPipelineNum: 0,
|
||||
queue: NewQueue(),
|
||||
accNonce: accNonce,
|
||||
accNextNonce: accNonce,
|
||||
}, nil
|
||||
}
|
||||
|
||||
@@ -97,35 +103,41 @@ func (t *TxManager) AddBatch(ctx context.Context, batchInfo *BatchInfo) {
|
||||
}
|
||||
|
||||
// SetSyncStatsVars is a thread safe method to sets the synchronizer Stats
|
||||
func (t *TxManager) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats, vars *synchronizer.SCVariablesPtr) {
|
||||
func (t *TxManager) SetSyncStatsVars(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars *synchronizer.SCVariablesPtr) {
|
||||
select {
|
||||
case t.statsVarsCh <- statsVars{Stats: *stats, Vars: *vars}:
|
||||
case <-ctx.Done():
|
||||
}
|
||||
}
|
||||
|
||||
func (t *TxManager) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
if vars.Rollup != nil {
|
||||
t.vars.Rollup = *vars.Rollup
|
||||
}
|
||||
if vars.Auction != nil {
|
||||
t.vars.Auction = *vars.Auction
|
||||
}
|
||||
if vars.WDelayer != nil {
|
||||
t.vars.WDelayer = *vars.WDelayer
|
||||
// DiscardPipeline is a thread safe method to notify about a discarded pipeline
|
||||
// due to a reorg
|
||||
func (t *TxManager) DiscardPipeline(ctx context.Context, pipelineNum int) {
|
||||
select {
|
||||
case t.discardPipelineCh <- pipelineNum:
|
||||
case <-ctx.Done():
|
||||
}
|
||||
}
|
||||
|
||||
func (t *TxManager) syncSCVars(vars synchronizer.SCVariablesPtr) {
|
||||
updateSCVars(&t.vars, vars)
|
||||
}
|
||||
|
||||
// NewAuth generates a new auth object for an ethereum transaction
|
||||
func (t *TxManager) NewAuth(ctx context.Context) (*bind.TransactOpts, error) {
|
||||
func (t *TxManager) NewAuth(ctx context.Context, batchInfo *BatchInfo) (*bind.TransactOpts, error) {
|
||||
gasPrice, err := t.ethClient.EthSuggestGasPrice(ctx)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
inc := new(big.Int).Set(gasPrice)
|
||||
const gasPriceDiv = 100
|
||||
inc.Div(inc, new(big.Int).SetUint64(gasPriceDiv))
|
||||
gasPrice.Add(gasPrice, inc)
|
||||
if t.cfg.GasPriceIncPerc != 0 {
|
||||
inc := new(big.Int).Set(gasPrice)
|
||||
inc.Mul(inc, new(big.Int).SetInt64(t.cfg.GasPriceIncPerc))
|
||||
// nolint reason: to calculate percentages we use 100
|
||||
inc.Div(inc, new(big.Int).SetUint64(100)) //nolint:gomnd
|
||||
gasPrice.Add(gasPrice, inc)
|
||||
}
|
||||
|
||||
// log.Debugw("TxManager: transaction metadata", "gasPrice", gasPrice)
|
||||
|
||||
auth, err := bind.NewKeyStoreTransactorWithChainID(t.ethClient.EthKeyStore(), t.account, t.chainID)
|
||||
@@ -133,42 +145,106 @@ func (t *TxManager) NewAuth(ctx context.Context) (*bind.TransactOpts, error) {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auth.Value = big.NewInt(0) // in wei
|
||||
// TODO: Calculate GasLimit based on the contents of the ForgeBatchArgs
|
||||
auth.GasLimit = 1000000
|
||||
|
||||
gasLimit := t.cfg.ForgeBatchGasCost.Fixed +
|
||||
uint64(len(batchInfo.L1UserTxsExtra))*t.cfg.ForgeBatchGasCost.L1UserTx +
|
||||
uint64(len(batchInfo.L1CoordTxs))*t.cfg.ForgeBatchGasCost.L1CoordTx +
|
||||
uint64(len(batchInfo.L2Txs))*t.cfg.ForgeBatchGasCost.L2Tx
|
||||
auth.GasLimit = gasLimit
|
||||
auth.GasPrice = gasPrice
|
||||
auth.Nonce = nil
|
||||
|
||||
return auth, nil
|
||||
}
|
||||
|
||||
func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchInfo) error {
|
||||
// TODO: Check if we can forge in the next blockNum, abort if we can't
|
||||
batchInfo.Debug.Status = StatusSent
|
||||
batchInfo.Debug.SendBlockNum = t.stats.Eth.LastBlock.Num + 1
|
||||
batchInfo.Debug.SendTimestamp = time.Now()
|
||||
batchInfo.Debug.StartToSendDelay = batchInfo.Debug.SendTimestamp.Sub(
|
||||
batchInfo.Debug.StartTimestamp).Seconds()
|
||||
func (t *TxManager) shouldSendRollupForgeBatch(batchInfo *BatchInfo) error {
|
||||
nextBlock := t.stats.Eth.LastBlock.Num + 1
|
||||
if !t.canForgeAt(nextBlock) {
|
||||
return tracerr.Wrap(fmt.Errorf("can't forge in the next block: %v", nextBlock))
|
||||
}
|
||||
if t.mustL1L2Batch(nextBlock) && !batchInfo.L1Batch {
|
||||
return tracerr.Wrap(fmt.Errorf("can't forge non-L1Batch in the next block: %v", nextBlock))
|
||||
}
|
||||
margin := t.cfg.SendBatchBlocksMarginCheck
|
||||
if margin != 0 {
|
||||
if !t.canForgeAt(nextBlock + margin) {
|
||||
return tracerr.Wrap(fmt.Errorf("can't forge after %v blocks: %v",
|
||||
margin, nextBlock))
|
||||
}
|
||||
if t.mustL1L2Batch(nextBlock+margin) && !batchInfo.L1Batch {
|
||||
return tracerr.Wrap(fmt.Errorf("can't forge non-L1Batch after %v blocks: %v",
|
||||
margin, nextBlock))
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func addPerc(v *big.Int, p int64) *big.Int {
|
||||
r := new(big.Int).Set(v)
|
||||
r.Mul(r, big.NewInt(p))
|
||||
// nolint reason: to calculate percentages we divide by 100
|
||||
r.Div(r, big.NewInt(100)) //nolit:gomnd
|
||||
// If the increase is 0, force it to be 1 so that a gas increase
|
||||
// doesn't result in the same value, making the transaction to be equal
|
||||
// than before.
|
||||
if r.Cmp(big.NewInt(0)) == 0 {
|
||||
r = big.NewInt(1)
|
||||
}
|
||||
return r.Add(v, r)
|
||||
}
|
||||
|
||||
func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchInfo,
|
||||
resend bool) error {
|
||||
var ethTx *types.Transaction
|
||||
var err error
|
||||
auth, err := t.NewAuth(ctx)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
auth.Nonce = big.NewInt(int64(t.lastPendingNonce))
|
||||
t.lastPendingNonce++
|
||||
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ {
|
||||
ethTx, err = t.ethClient.RollupForgeBatch(batchInfo.ForgeBatchArgs, auth)
|
||||
var auth *bind.TransactOpts
|
||||
if resend {
|
||||
auth = batchInfo.Auth
|
||||
auth.GasPrice = addPerc(auth.GasPrice, 10)
|
||||
} else {
|
||||
auth, err = t.NewAuth(ctx, batchInfo)
|
||||
if err != nil {
|
||||
// if strings.Contains(err.Error(), common.AuctionErrMsgCannotForge) {
|
||||
// log.Errorw("TxManager ethClient.RollupForgeBatch", "err", err,
|
||||
// "block", t.stats.Eth.LastBlock.Num+1)
|
||||
// return tracerr.Wrap(err)
|
||||
// }
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
batchInfo.Auth = auth
|
||||
auth.Nonce = big.NewInt(int64(t.accNextNonce))
|
||||
}
|
||||
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ {
|
||||
if auth.GasPrice.Cmp(t.cfg.MaxGasPrice) > 0 {
|
||||
return tracerr.Wrap(fmt.Errorf("calculated gasPrice (%v) > maxGasPrice (%v)",
|
||||
auth.GasPrice, t.cfg.MaxGasPrice))
|
||||
}
|
||||
// RollupForgeBatch() calls ethclient.SendTransaction()
|
||||
ethTx, err = t.ethClient.RollupForgeBatch(batchInfo.ForgeBatchArgs, auth)
|
||||
// We check the errors via strings because we match the
|
||||
// definition of the error from geth, with the string returned
|
||||
// via RPC obtained by the client.
|
||||
if err == nil {
|
||||
break
|
||||
} else if strings.Contains(err.Error(), core.ErrNonceTooLow.Error()) {
|
||||
log.Warnw("TxManager ethClient.RollupForgeBatch incrementing nonce",
|
||||
"err", err, "nonce", auth.Nonce, "batchNum", batchInfo.BatchNum)
|
||||
auth.Nonce.Add(auth.Nonce, big.NewInt(1))
|
||||
attempt--
|
||||
} else if strings.Contains(err.Error(), core.ErrNonceTooHigh.Error()) {
|
||||
log.Warnw("TxManager ethClient.RollupForgeBatch decrementing nonce",
|
||||
"err", err, "nonce", auth.Nonce, "batchNum", batchInfo.BatchNum)
|
||||
auth.Nonce.Sub(auth.Nonce, big.NewInt(1))
|
||||
attempt--
|
||||
} else if strings.Contains(err.Error(), core.ErrReplaceUnderpriced.Error()) {
|
||||
log.Warnw("TxManager ethClient.RollupForgeBatch incrementing gasPrice",
|
||||
"err", err, "gasPrice", auth.GasPrice, "batchNum", batchInfo.BatchNum)
|
||||
auth.GasPrice = addPerc(auth.GasPrice, 10)
|
||||
attempt--
|
||||
} else if strings.Contains(err.Error(), core.ErrUnderpriced.Error()) {
|
||||
log.Warnw("TxManager ethClient.RollupForgeBatch incrementing gasPrice",
|
||||
"err", err, "gasPrice", auth.GasPrice, "batchNum", batchInfo.BatchNum)
|
||||
auth.GasPrice = addPerc(auth.GasPrice, 10)
|
||||
attempt--
|
||||
} else {
|
||||
log.Errorw("TxManager ethClient.RollupForgeBatch",
|
||||
"attempt", attempt, "err", err, "block", t.stats.Eth.LastBlock.Num+1,
|
||||
"batchNum", batchInfo.BatchNum)
|
||||
} else {
|
||||
break
|
||||
}
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
@@ -179,11 +255,31 @@ func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchIn
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("reached max attempts for ethClient.RollupForgeBatch: %w", err))
|
||||
}
|
||||
if !resend {
|
||||
t.accNextNonce = auth.Nonce.Uint64() + 1
|
||||
}
|
||||
batchInfo.EthTx = ethTx
|
||||
log.Infow("TxManager ethClient.RollupForgeBatch", "batch", batchInfo.BatchNum, "tx", ethTx.Hash().Hex())
|
||||
log.Infow("TxManager ethClient.RollupForgeBatch", "batch", batchInfo.BatchNum, "tx", ethTx.Hash())
|
||||
now := time.Now()
|
||||
batchInfo.SendTimestamp = now
|
||||
|
||||
if resend {
|
||||
batchInfo.Debug.ResendNum++
|
||||
}
|
||||
batchInfo.Debug.Status = StatusSent
|
||||
batchInfo.Debug.SendBlockNum = t.stats.Eth.LastBlock.Num + 1
|
||||
batchInfo.Debug.SendTimestamp = batchInfo.SendTimestamp
|
||||
batchInfo.Debug.StartToSendDelay = batchInfo.Debug.SendTimestamp.Sub(
|
||||
batchInfo.Debug.StartTimestamp).Seconds()
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
t.lastPendingBatch = batchInfo.BatchNum
|
||||
if err := t.l2DB.DoneForging(common.TxIDsFromL2Txs(batchInfo.L2Txs), batchInfo.BatchNum); err != nil {
|
||||
|
||||
if !resend {
|
||||
if batchInfo.L1Batch {
|
||||
t.lastSentL1BatchBlockNum = t.stats.Eth.LastBlock.Num + 1
|
||||
}
|
||||
}
|
||||
if err := t.l2DB.DoneForging(common.TxIDsFromL2Txs(batchInfo.L2Txs),
|
||||
batchInfo.BatchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
@@ -215,7 +311,9 @@ func (t *TxManager) checkEthTransactionReceipt(ctx context.Context, batchInfo *B
|
||||
}
|
||||
}
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w", err))
|
||||
return tracerr.Wrap(
|
||||
fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w",
|
||||
err))
|
||||
}
|
||||
batchInfo.Receipt = receipt
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
@@ -225,13 +323,20 @@ func (t *TxManager) checkEthTransactionReceipt(ctx context.Context, batchInfo *B
|
||||
func (t *TxManager) handleReceipt(ctx context.Context, batchInfo *BatchInfo) (*int64, error) {
|
||||
receipt := batchInfo.Receipt
|
||||
if receipt != nil {
|
||||
if batchInfo.EthTx.Nonce()+1 > t.accNonce {
|
||||
t.accNonce = batchInfo.EthTx.Nonce() + 1
|
||||
}
|
||||
if receipt.Status == types.ReceiptStatusFailed {
|
||||
batchInfo.Debug.Status = StatusFailed
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
_, err := t.ethClient.EthCall(ctx, batchInfo.EthTx, receipt.BlockNumber)
|
||||
log.Warnw("TxManager receipt status is failed", "tx", receipt.TxHash.Hex(),
|
||||
log.Warnw("TxManager receipt status is failed", "tx", receipt.TxHash,
|
||||
"batch", batchInfo.BatchNum, "block", receipt.BlockNumber.Int64(),
|
||||
"err", err)
|
||||
batchInfo.EthTxErr = err
|
||||
if batchInfo.BatchNum <= t.lastSuccessBatch {
|
||||
t.lastSuccessBatch = batchInfo.BatchNum - 1
|
||||
}
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
return nil, tracerr.Wrap(fmt.Errorf(
|
||||
"ethereum transaction receipt status is failed: %w", err))
|
||||
} else if receipt.Status == types.ReceiptStatusSuccessful {
|
||||
@@ -239,6 +344,17 @@ func (t *TxManager) handleReceipt(ctx context.Context, batchInfo *BatchInfo) (*i
|
||||
batchInfo.Debug.MineBlockNum = receipt.BlockNumber.Int64()
|
||||
batchInfo.Debug.StartToMineBlocksDelay = batchInfo.Debug.MineBlockNum -
|
||||
batchInfo.Debug.StartBlockNum
|
||||
if batchInfo.Debug.StartToMineDelay == 0 {
|
||||
if block, err := t.ethClient.EthBlockByNumber(ctx,
|
||||
receipt.BlockNumber.Int64()); err != nil {
|
||||
log.Warnw("TxManager: ethClient.EthBlockByNumber", "err", err)
|
||||
} else {
|
||||
batchInfo.Debug.SendToMineDelay = block.Timestamp.Sub(
|
||||
batchInfo.Debug.SendTimestamp).Seconds()
|
||||
batchInfo.Debug.StartToMineDelay = block.Timestamp.Sub(
|
||||
batchInfo.Debug.StartTimestamp).Seconds()
|
||||
}
|
||||
}
|
||||
t.cfg.debugBatchStore(batchInfo)
|
||||
if batchInfo.BatchNum > t.lastSuccessBatch {
|
||||
t.lastSuccessBatch = batchInfo.BatchNum
|
||||
@@ -250,11 +366,72 @@ func (t *TxManager) handleReceipt(ctx context.Context, batchInfo *BatchInfo) (*i
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
// TODO:
|
||||
// - After sending a message: CancelPipeline, stop all consecutive pending Batches (transactions)
|
||||
|
||||
// Queue of BatchInfos
|
||||
type Queue struct {
|
||||
list []*BatchInfo
|
||||
// nonceByBatchNum map[common.BatchNum]uint64
|
||||
next int
|
||||
}
|
||||
|
||||
// NewQueue returns a new queue
|
||||
func NewQueue() Queue {
|
||||
return Queue{
|
||||
list: make([]*BatchInfo, 0),
|
||||
// nonceByBatchNum: make(map[common.BatchNum]uint64),
|
||||
next: 0,
|
||||
}
|
||||
}
|
||||
|
||||
// Len is the length of the queue
|
||||
func (q *Queue) Len() int {
|
||||
return len(q.list)
|
||||
}
|
||||
|
||||
// At returns the BatchInfo at position (or nil if position is out of bounds)
|
||||
func (q *Queue) At(position int) *BatchInfo {
|
||||
if position >= len(q.list) {
|
||||
return nil
|
||||
}
|
||||
return q.list[position]
|
||||
}
|
||||
|
||||
// Next returns the next BatchInfo (or nil if queue is empty)
|
||||
func (q *Queue) Next() (int, *BatchInfo) {
|
||||
if len(q.list) == 0 {
|
||||
return 0, nil
|
||||
}
|
||||
defer func() { q.next = (q.next + 1) % len(q.list) }()
|
||||
return q.next, q.list[q.next]
|
||||
}
|
||||
|
||||
// Remove removes the BatchInfo at position
|
||||
func (q *Queue) Remove(position int) {
|
||||
// batchInfo := q.list[position]
|
||||
// delete(q.nonceByBatchNum, batchInfo.BatchNum)
|
||||
q.list = append(q.list[:position], q.list[position+1:]...)
|
||||
if len(q.list) == 0 {
|
||||
q.next = 0
|
||||
} else {
|
||||
q.next = position % len(q.list)
|
||||
}
|
||||
}
|
||||
|
||||
// Push adds a new BatchInfo
|
||||
func (q *Queue) Push(batchInfo *BatchInfo) {
|
||||
q.list = append(q.list, batchInfo)
|
||||
// q.nonceByBatchNum[batchInfo.BatchNum] = batchInfo.EthTx.Nonce()
|
||||
}
|
||||
|
||||
// func (q *Queue) NonceByBatchNum(batchNum common.BatchNum) (uint64, bool) {
|
||||
// nonce, ok := q.nonceByBatchNum[batchNum]
|
||||
// return nonce, ok
|
||||
// }
|
||||
|
||||
// Run the TxManager
|
||||
func (t *TxManager) Run(ctx context.Context) {
|
||||
next := 0
|
||||
waitDuration := longWaitDuration
|
||||
|
||||
var statsVars statsVars
|
||||
select {
|
||||
case statsVars = <-t.statsVarsCh:
|
||||
@@ -263,8 +440,9 @@ func (t *TxManager) Run(ctx context.Context) {
|
||||
t.stats = statsVars.Stats
|
||||
t.syncSCVars(statsVars.Vars)
|
||||
log.Infow("TxManager: received initial statsVars",
|
||||
"block", t.stats.Eth.LastBlock.Num, "batch", t.stats.Eth.LastBatch)
|
||||
"block", t.stats.Eth.LastBlock.Num, "batch", t.stats.Eth.LastBatchNum)
|
||||
|
||||
timer := time.NewTimer(longWaitDuration)
|
||||
for {
|
||||
select {
|
||||
case <-ctx.Done():
|
||||
@@ -273,8 +451,27 @@ func (t *TxManager) Run(ctx context.Context) {
|
||||
case statsVars := <-t.statsVarsCh:
|
||||
t.stats = statsVars.Stats
|
||||
t.syncSCVars(statsVars.Vars)
|
||||
case pipelineNum := <-t.discardPipelineCh:
|
||||
t.minPipelineNum = pipelineNum + 1
|
||||
if err := t.removeBadBatchInfos(ctx); ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
log.Errorw("TxManager: removeBadBatchInfos", "err", err)
|
||||
continue
|
||||
}
|
||||
case batchInfo := <-t.batchCh:
|
||||
if err := t.sendRollupForgeBatch(ctx, batchInfo); ctx.Err() != nil {
|
||||
if batchInfo.PipelineNum < t.minPipelineNum {
|
||||
log.Warnw("TxManager: batchInfo received pipelineNum < minPipelineNum",
|
||||
"num", batchInfo.PipelineNum, "minNum", t.minPipelineNum)
|
||||
}
|
||||
if err := t.shouldSendRollupForgeBatch(batchInfo); err != nil {
|
||||
log.Warnw("TxManager: shouldSend", "err", err,
|
||||
"batch", batchInfo.BatchNum)
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf("forgeBatch shouldSend: %v", err)})
|
||||
continue
|
||||
}
|
||||
if err := t.sendRollupForgeBatch(ctx, batchInfo, false); ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
// If we reach here it's because our ethNode has
|
||||
@@ -282,29 +479,35 @@ func (t *TxManager) Run(ctx context.Context) {
|
||||
// ethereum. This could be due to the ethNode
|
||||
// failure, or an invalid transaction (that
|
||||
// can't be mined)
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch send: %v", err)})
|
||||
log.Warnw("TxManager: forgeBatch send failed", "err", err,
|
||||
"batch", batchInfo.BatchNum)
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf("forgeBatch send: %v", err)})
|
||||
continue
|
||||
}
|
||||
t.queue = append(t.queue, batchInfo)
|
||||
waitDuration = t.cfg.TxManagerCheckInterval
|
||||
case <-time.After(waitDuration):
|
||||
if len(t.queue) == 0 {
|
||||
waitDuration = longWaitDuration
|
||||
t.queue.Push(batchInfo)
|
||||
if !timer.Stop() {
|
||||
<-timer.C
|
||||
}
|
||||
timer.Reset(t.cfg.TxManagerCheckInterval)
|
||||
case <-timer.C:
|
||||
queuePosition, batchInfo := t.queue.Next()
|
||||
if batchInfo == nil {
|
||||
timer.Reset(longWaitDuration)
|
||||
continue
|
||||
}
|
||||
current := next
|
||||
next = (current + 1) % len(t.queue)
|
||||
batchInfo := t.queue[current]
|
||||
timer.Reset(t.cfg.TxManagerCheckInterval)
|
||||
if err := t.checkEthTransactionReceipt(ctx, batchInfo); ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil { //nolint:staticcheck
|
||||
// Our ethNode is giving an error different
|
||||
// than "not found" when getting the receipt
|
||||
// for the transaction, so we can't figure out
|
||||
// if it was not mined, mined and succesfull or
|
||||
// if it was not mined, mined and successful or
|
||||
// mined and failed. This could be due to the
|
||||
// ethNode failure.
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch receipt: %v", err)})
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf("forgeBatch receipt: %v", err)})
|
||||
}
|
||||
|
||||
confirm, err := t.handleReceipt(ctx, batchInfo)
|
||||
@@ -312,32 +515,106 @@ func (t *TxManager) Run(ctx context.Context) {
|
||||
continue
|
||||
} else if err != nil { //nolint:staticcheck
|
||||
// Transaction was rejected
|
||||
t.queue = append(t.queue[:current], t.queue[current+1:]...)
|
||||
if len(t.queue) == 0 {
|
||||
next = 0
|
||||
} else {
|
||||
next = current % len(t.queue)
|
||||
if err := t.removeBadBatchInfos(ctx); ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
log.Errorw("TxManager: removeBadBatchInfos", "err", err)
|
||||
continue
|
||||
}
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch reject: %v", err)})
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf("forgeBatch reject: %v", err)})
|
||||
continue
|
||||
}
|
||||
if confirm != nil && *confirm >= t.cfg.ConfirmBlocks {
|
||||
log.Debugw("TxManager tx for RollupForgeBatch confirmed",
|
||||
"batch", batchInfo.BatchNum)
|
||||
t.queue = append(t.queue[:current], t.queue[current+1:]...)
|
||||
if len(t.queue) == 0 {
|
||||
next = 0
|
||||
} else {
|
||||
next = current % len(t.queue)
|
||||
now := time.Now()
|
||||
if !t.cfg.EthNoReuseNonce && confirm == nil &&
|
||||
now.Sub(batchInfo.SendTimestamp) > t.cfg.EthTxResendTimeout {
|
||||
log.Infow("TxManager: forgeBatch tx not been mined timeout, resending",
|
||||
"tx", batchInfo.EthTx.Hash(), "batch", batchInfo.BatchNum)
|
||||
if err := t.sendRollupForgeBatch(ctx, batchInfo, true); ctx.Err() != nil {
|
||||
continue
|
||||
} else if err != nil {
|
||||
// If we reach here it's because our ethNode has
|
||||
// been unable to send the transaction to
|
||||
// ethereum. This could be due to the ethNode
|
||||
// failure, or an invalid transaction (that
|
||||
// can't be mined)
|
||||
log.Warnw("TxManager: forgeBatch resend failed", "err", err,
|
||||
"batch", batchInfo.BatchNum)
|
||||
t.coord.SendMsg(ctx, MsgStopPipeline{
|
||||
Reason: fmt.Sprintf("forgeBatch resend: %v", err)})
|
||||
continue
|
||||
}
|
||||
}
|
||||
|
||||
if confirm != nil && *confirm >= t.cfg.ConfirmBlocks {
|
||||
log.Debugw("TxManager: forgeBatch tx confirmed",
|
||||
"tx", batchInfo.EthTx.Hash(), "batch", batchInfo.BatchNum)
|
||||
t.queue.Remove(queuePosition)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// nolint reason: this function will be used in the future
|
||||
//nolint:unused
|
||||
func (t *TxManager) canForge(stats *synchronizer.Stats, blockNum int64) bool {
|
||||
func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
||||
next := 0
|
||||
for {
|
||||
batchInfo := t.queue.At(next)
|
||||
if batchInfo == nil {
|
||||
break
|
||||
}
|
||||
if err := t.checkEthTransactionReceipt(ctx, batchInfo); ctx.Err() != nil {
|
||||
return nil
|
||||
} else if err != nil {
|
||||
// Our ethNode is giving an error different
|
||||
// than "not found" when getting the receipt
|
||||
// for the transaction, so we can't figure out
|
||||
// if it was not mined, mined and successful or
|
||||
// mined and failed. This could be due to the
|
||||
// ethNode failure.
|
||||
next++
|
||||
continue
|
||||
}
|
||||
confirm, err := t.handleReceipt(ctx, batchInfo)
|
||||
if ctx.Err() != nil {
|
||||
return nil
|
||||
} else if err != nil {
|
||||
// Transaction was rejected
|
||||
if t.minPipelineNum <= batchInfo.PipelineNum {
|
||||
t.minPipelineNum = batchInfo.PipelineNum + 1
|
||||
}
|
||||
t.queue.Remove(next)
|
||||
continue
|
||||
}
|
||||
// If tx is pending but is from a cancelled pipeline, remove it
|
||||
// from the queue
|
||||
if confirm == nil {
|
||||
if batchInfo.PipelineNum < t.minPipelineNum {
|
||||
t.queue.Remove(next)
|
||||
continue
|
||||
}
|
||||
}
|
||||
next++
|
||||
}
|
||||
accNonce, err := t.ethClient.EthNonceAt(ctx, t.account.Address, nil)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
if !t.cfg.EthNoReuseNonce {
|
||||
t.accNextNonce = accNonce
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func (t *TxManager) canForgeAt(blockNum int64) bool {
|
||||
return canForge(&t.consts.Auction, &t.vars.Auction,
|
||||
&stats.Sync.Auction.CurrentSlot, &stats.Sync.Auction.NextSlot,
|
||||
&t.stats.Sync.Auction.CurrentSlot, &t.stats.Sync.Auction.NextSlot,
|
||||
t.cfg.ForgerAddress, blockNum)
|
||||
}
|
||||
|
||||
func (t *TxManager) mustL1L2Batch(blockNum int64) bool {
|
||||
lastL1BatchBlockNum := t.lastSentL1BatchBlockNum
|
||||
if t.stats.Sync.LastL1BatchBlock > lastL1BatchBlockNum {
|
||||
lastL1BatchBlockNum = t.stats.Sync.LastL1BatchBlock
|
||||
}
|
||||
return blockNum-lastL1BatchBlockNum >= t.vars.Rollup.ForgeL1L2BatchTimeout-1
|
||||
}
|
||||
|
||||
15
coordinator/txmanager_test.go
Normal file
15
coordinator/txmanager_test.go
Normal file
@@ -0,0 +1,15 @@
|
||||
package coordinator
|
||||
|
||||
import (
|
||||
"math/big"
|
||||
"testing"
|
||||
|
||||
"github.com/stretchr/testify/assert"
|
||||
)
|
||||
|
||||
func TestAddPerc(t *testing.T) {
|
||||
assert.Equal(t, "110", addPerc(big.NewInt(100), 10).String())
|
||||
assert.Equal(t, "101", addPerc(big.NewInt(100), 1).String())
|
||||
assert.Equal(t, "12", addPerc(big.NewInt(10), 20).String())
|
||||
assert.Equal(t, "1500", addPerc(big.NewInt(1000), 50).String())
|
||||
}
|
||||
1077
db/historydb/apiqueries.go
Normal file
1077
db/historydb/apiqueries.go
Normal file
File diff suppressed because it is too large
Load Diff
File diff suppressed because it is too large
Load Diff
@@ -22,6 +22,7 @@ import (
|
||||
)
|
||||
|
||||
var historyDB *HistoryDB
|
||||
var historyDBWithACC *HistoryDB
|
||||
|
||||
// In order to run the test you need to run a Posgres DB with
|
||||
// a database named "history" that is accessible by
|
||||
@@ -38,10 +39,12 @@ func TestMain(m *testing.M) {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
historyDB = NewHistoryDB(db)
|
||||
historyDB = NewHistoryDB(db, db, nil)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
apiConnCon := dbUtils.NewAPIConnectionController(1, time.Second)
|
||||
historyDBWithACC = NewHistoryDB(db, db, apiConnCon)
|
||||
// Run tests
|
||||
result := m.Run()
|
||||
// Close DB
|
||||
@@ -85,7 +88,7 @@ func TestBlocks(t *testing.T) {
|
||||
blocks...,
|
||||
)
|
||||
// Get all blocks from DB
|
||||
fetchedBlocks, err := historyDB.GetBlocks(fromBlock, toBlock)
|
||||
fetchedBlocks, err := historyDB.getBlocks(fromBlock, toBlock)
|
||||
assert.Equal(t, len(blocks), len(fetchedBlocks))
|
||||
// Compare generated vs getted blocks
|
||||
assert.NoError(t, err)
|
||||
@@ -200,6 +203,10 @@ func TestBatches(t *testing.T) {
|
||||
fetchedLastBatchNum, err := historyDB.GetLastBatchNum()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, batches[len(batches)-1].BatchNum, fetchedLastBatchNum)
|
||||
// Test GetLastBatch
|
||||
fetchedLastBatch, err := historyDB.GetLastBatch()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, &batches[len(batches)-1], fetchedLastBatch)
|
||||
// Test GetLastL1TxsNum
|
||||
fetchedLastL1TxsNum, err := historyDB.GetLastL1TxsNum()
|
||||
assert.NoError(t, err)
|
||||
@@ -208,6 +215,12 @@ func TestBatches(t *testing.T) {
|
||||
fetchedLastL1BatchBlockNum, err := historyDB.GetLastL1BatchBlockNum()
|
||||
assert.NoError(t, err)
|
||||
assert.Equal(t, lastL1BatchBlockNum, fetchedLastL1BatchBlockNum)
|
||||
// Test GetBatch
|
||||
fetchedBatch, err := historyDB.GetBatch(1)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, &batches[0], fetchedBatch)
|
||||
_, err = historyDB.GetBatch(common.BatchNum(len(batches) + 1))
|
||||
assert.Equal(t, sql.ErrNoRows, tracerr.Unwrap(err))
|
||||
}
|
||||
|
||||
func TestBids(t *testing.T) {
|
||||
@@ -245,9 +258,8 @@ func TestTokens(t *testing.T) {
|
||||
err := historyDB.AddTokens(tokens)
|
||||
assert.NoError(t, err)
|
||||
tokens = append([]common.Token{ethToken}, tokens...)
|
||||
limit := uint(10)
|
||||
// Fetch tokens
|
||||
fetchedTokens, _, err := historyDB.GetTokens(nil, nil, "", nil, &limit, OrderAsc)
|
||||
fetchedTokens, err := historyDB.GetTokensTest()
|
||||
assert.NoError(t, err)
|
||||
// Compare fetched tokens vs generated tokens
|
||||
// All the tokens should have USDUpdate setted by the DB trigger
|
||||
@@ -267,7 +279,7 @@ func TestTokens(t *testing.T) {
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.Symbol, value))
|
||||
}
|
||||
// Fetch tokens
|
||||
fetchedTokens, _, err = historyDB.GetTokens(nil, nil, "", nil, &limit, OrderAsc)
|
||||
fetchedTokens, err = historyDB.GetTokensTest()
|
||||
assert.NoError(t, err)
|
||||
// Compare fetched tokens vs generated tokens
|
||||
// All the tokens should have USDUpdate setted by the DB trigger
|
||||
@@ -302,9 +314,8 @@ func TestTokensUTF8(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
// Work with nonUTFTokens as tokens one gets updated and non UTF-8 characters are lost
|
||||
nonUTFTokens = append([]common.Token{ethToken}, nonUTFTokens...)
|
||||
limit := uint(10)
|
||||
// Fetch tokens
|
||||
fetchedTokens, _, err := historyDB.GetTokens(nil, nil, "", nil, &limit, OrderAsc)
|
||||
fetchedTokens, err := historyDB.GetTokensTest()
|
||||
assert.NoError(t, err)
|
||||
// Compare fetched tokens vs generated tokens
|
||||
// All the tokens should have USDUpdate setted by the DB trigger
|
||||
@@ -324,7 +335,7 @@ func TestTokensUTF8(t *testing.T) {
|
||||
assert.NoError(t, historyDB.UpdateTokenValue(token.Symbol, value))
|
||||
}
|
||||
// Fetch tokens
|
||||
fetchedTokens, _, err = historyDB.GetTokens(nil, nil, "", nil, &limit, OrderAsc)
|
||||
fetchedTokens, err = historyDB.GetTokensTest()
|
||||
assert.NoError(t, err)
|
||||
// Compare fetched tokens vs generated tokens
|
||||
// All the tokens should have USDUpdate setted by the DB trigger
|
||||
@@ -366,6 +377,22 @@ func TestAccounts(t *testing.T) {
|
||||
accs[i].Balance = nil
|
||||
assert.Equal(t, accs[i], acc)
|
||||
}
|
||||
// Test AccountBalances
|
||||
accUpdates := make([]common.AccountUpdate, len(accs))
|
||||
for i, acc := range accs {
|
||||
accUpdates[i] = common.AccountUpdate{
|
||||
EthBlockNum: batches[acc.BatchNum-1].EthBlockNum,
|
||||
BatchNum: acc.BatchNum,
|
||||
Idx: acc.Idx,
|
||||
Nonce: common.Nonce(i),
|
||||
Balance: big.NewInt(int64(i)),
|
||||
}
|
||||
}
|
||||
err = historyDB.AddAccountUpdates(accUpdates)
|
||||
require.NoError(t, err)
|
||||
fetchedAccBalances, err := historyDB.GetAllAccountUpdates()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, accUpdates, fetchedAccBalances)
|
||||
}
|
||||
|
||||
func TestTxs(t *testing.T) {
|
||||
@@ -610,10 +637,10 @@ func TestTxs(t *testing.T) {
|
||||
assert.Equal(t, common.TxTypeExit, dbL2Txs[3].Type)
|
||||
|
||||
// Tx ID
|
||||
assert.Equal(t, "0x02d709307533c4e3c03f20751fc4d72bc18b225d14f9616525540a64342c7c350d", dbL2Txs[0].TxID.String())
|
||||
assert.Equal(t, "0x02e88bc5503f282cca045847668511290e642410a459bb67b1fafcd1b6097c149c", dbL2Txs[1].TxID.String())
|
||||
assert.Equal(t, "0x027911262b43315c0b24942a02fe228274b6e4d57a476bfcdd7a324b3091362c7d", dbL2Txs[2].TxID.String())
|
||||
assert.Equal(t, "0x02f572b63f2a5c302e1b9337ea6944bfbac3d199e4ddd262b5a53759c72ec10ee6", dbL2Txs[3].TxID.String())
|
||||
assert.Equal(t, "0x024e555248100b69a8aabf6d31719b9fe8a60dcc6c3407904a93c8d2d9ade18ee5", dbL2Txs[0].TxID.String())
|
||||
assert.Equal(t, "0x021ae87ca34d50ff35d98dfc0d7c95f2bf2e4ffeebb82ea71f43a8b0dfa5d36d89", dbL2Txs[1].TxID.String())
|
||||
assert.Equal(t, "0x024abce7f3f2382dc520ed557593f11dea1ee197e55b60402e664facc27aa19774", dbL2Txs[2].TxID.String())
|
||||
assert.Equal(t, "0x02f921ad9e7a6e59606570fe12a7dde0e36014197de0363b9b45e5097d6f2b1dd0", dbL2Txs[3].TxID.String())
|
||||
|
||||
// Tx From and To IDx
|
||||
assert.Equal(t, dbL2Txs[0].ToIdx, dbL2Txs[2].FromIdx)
|
||||
@@ -693,6 +720,10 @@ func TestGetUnforgedL1UserTxs(t *testing.T) {
|
||||
assert.Equal(t, 5, len(l1UserTxs))
|
||||
assert.Equal(t, blocks[0].Rollup.L1UserTxs, l1UserTxs)
|
||||
|
||||
count, err := historyDB.GetUnforgedL1UserTxsCount()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, 5, count)
|
||||
|
||||
// No l1UserTxs for this toForgeL1TxsNum
|
||||
l1UserTxs, err = historyDB.GetUnforgedL1UserTxs(2)
|
||||
require.NoError(t, err)
|
||||
@@ -790,11 +821,11 @@ func TestSetExtraInfoForgedL1UserTxs(t *testing.T) {
|
||||
}
|
||||
// Add second batch to trigger the update of the batch_num,
|
||||
// while avoiding the implicit call of setExtraInfoForgedL1UserTxs
|
||||
err = historyDB.addBlock(historyDB.db, &blocks[1].Block)
|
||||
err = historyDB.addBlock(historyDB.dbWrite, &blocks[1].Block)
|
||||
require.NoError(t, err)
|
||||
err = historyDB.addBatch(historyDB.db, &blocks[1].Rollup.Batches[0].Batch)
|
||||
err = historyDB.addBatch(historyDB.dbWrite, &blocks[1].Rollup.Batches[0].Batch)
|
||||
require.NoError(t, err)
|
||||
err = historyDB.addAccounts(historyDB.db, blocks[1].Rollup.Batches[0].CreatedAccounts)
|
||||
err = historyDB.addAccounts(historyDB.dbWrite, blocks[1].Rollup.Batches[0].CreatedAccounts)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Set the Effective{Amount,DepositAmount} of the L1UserTxs that are forged in the second block
|
||||
@@ -804,7 +835,7 @@ func TestSetExtraInfoForgedL1UserTxs(t *testing.T) {
|
||||
l1Txs[1].EffectiveAmount = big.NewInt(0)
|
||||
l1Txs[2].EffectiveDepositAmount = big.NewInt(0)
|
||||
l1Txs[2].EffectiveAmount = big.NewInt(0)
|
||||
err = historyDB.setExtraInfoForgedL1UserTxs(historyDB.db, l1Txs)
|
||||
err = historyDB.setExtraInfoForgedL1UserTxs(historyDB.dbWrite, l1Txs)
|
||||
require.NoError(t, err)
|
||||
|
||||
dbL1Txs, err := historyDB.GetAllL1UserTxs()
|
||||
@@ -891,10 +922,10 @@ func TestUpdateExitTree(t *testing.T) {
|
||||
common.WithdrawInfo{Idx: 259, NumExitRoot: 3, InstantWithdraw: false,
|
||||
Owner: tc.UsersByIdx[259].Addr, Token: tokenAddr},
|
||||
)
|
||||
err = historyDB.addBlock(historyDB.db, &block.Block)
|
||||
err = historyDB.addBlock(historyDB.dbWrite, &block.Block)
|
||||
require.NoError(t, err)
|
||||
|
||||
err = historyDB.updateExitTree(historyDB.db, block.Block.Num,
|
||||
err = historyDB.updateExitTree(historyDB.dbWrite, block.Block.Num,
|
||||
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
||||
require.NoError(t, err)
|
||||
|
||||
@@ -924,10 +955,10 @@ func TestUpdateExitTree(t *testing.T) {
|
||||
Token: tokenAddr,
|
||||
Amount: big.NewInt(80),
|
||||
})
|
||||
err = historyDB.addBlock(historyDB.db, &block.Block)
|
||||
err = historyDB.addBlock(historyDB.dbWrite, &block.Block)
|
||||
require.NoError(t, err)
|
||||
|
||||
err = historyDB.updateExitTree(historyDB.db, block.Block.Num,
|
||||
err = historyDB.updateExitTree(historyDB.dbWrite, block.Block.Num,
|
||||
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
||||
require.NoError(t, err)
|
||||
|
||||
@@ -970,7 +1001,7 @@ func TestGetBestBidCoordinator(t *testing.T) {
|
||||
URL: "bar",
|
||||
},
|
||||
}
|
||||
err = historyDB.addCoordinators(historyDB.db, coords)
|
||||
err = historyDB.addCoordinators(historyDB.dbWrite, coords)
|
||||
require.NoError(t, err)
|
||||
|
||||
bids := []common.Bid{
|
||||
@@ -988,7 +1019,7 @@ func TestGetBestBidCoordinator(t *testing.T) {
|
||||
},
|
||||
}
|
||||
|
||||
err = historyDB.addBids(historyDB.db, bids)
|
||||
err = historyDB.addBids(historyDB.dbWrite, bids)
|
||||
require.NoError(t, err)
|
||||
|
||||
forger10, err := historyDB.GetBestBidCoordinator(10)
|
||||
@@ -1026,7 +1057,7 @@ func TestAddBucketUpdates(t *testing.T) {
|
||||
Withdrawals: big.NewInt(42),
|
||||
},
|
||||
}
|
||||
err := historyDB.addBucketUpdates(historyDB.db, bucketUpdates)
|
||||
err := historyDB.addBucketUpdates(historyDB.dbWrite, bucketUpdates)
|
||||
require.NoError(t, err)
|
||||
dbBucketUpdates, err := historyDB.GetAllBucketUpdates()
|
||||
require.NoError(t, err)
|
||||
@@ -1051,7 +1082,7 @@ func TestAddTokenExchanges(t *testing.T) {
|
||||
ValueUSD: 67890,
|
||||
},
|
||||
}
|
||||
err := historyDB.addTokenExchanges(historyDB.db, tokenExchanges)
|
||||
err := historyDB.addTokenExchanges(historyDB.dbWrite, tokenExchanges)
|
||||
require.NoError(t, err)
|
||||
dbTokenExchanges, err := historyDB.GetAllTokenExchanges()
|
||||
require.NoError(t, err)
|
||||
@@ -1080,16 +1111,15 @@ func TestAddEscapeHatchWithdrawals(t *testing.T) {
|
||||
Amount: big.NewInt(20003),
|
||||
},
|
||||
}
|
||||
err := historyDB.addEscapeHatchWithdrawals(historyDB.db, escapeHatchWithdrawals)
|
||||
err := historyDB.addEscapeHatchWithdrawals(historyDB.dbWrite, escapeHatchWithdrawals)
|
||||
require.NoError(t, err)
|
||||
dbEscapeHatchWithdrawals, err := historyDB.GetAllEscapeHatchWithdrawals()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, escapeHatchWithdrawals, dbEscapeHatchWithdrawals)
|
||||
}
|
||||
|
||||
func TestGetMetrics(t *testing.T) {
|
||||
func TestGetMetricsAPI(t *testing.T) {
|
||||
test.WipeDB(historyDB.DB())
|
||||
|
||||
set := `
|
||||
Type: Blockchain
|
||||
|
||||
@@ -1146,26 +1176,22 @@ func TestGetMetrics(t *testing.T) {
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
res, err := historyDB.GetMetrics(common.BatchNum(numBatches))
|
||||
res, err := historyDBWithACC.GetMetricsAPI(common.BatchNum(numBatches))
|
||||
assert.NoError(t, err)
|
||||
|
||||
assert.Equal(t, float64(numTx)/float64(numBatches-1), res.TransactionsPerBatch)
|
||||
assert.Equal(t, float64(numTx)/float64(numBatches), res.TransactionsPerBatch)
|
||||
|
||||
// Frequency is not exactly the desired one, some decimals may appear
|
||||
assert.GreaterOrEqual(t, res.BatchFrequency, float64(frequency))
|
||||
assert.Less(t, res.BatchFrequency, float64(frequency+1))
|
||||
// Truncate frecuency into an int to do an exact check
|
||||
assert.Equal(t, frequency, int(res.BatchFrequency))
|
||||
// This may also be different in some decimals
|
||||
// Truncate it to the third decimal to compare
|
||||
assert.Equal(t, math.Trunc((float64(numTx)/float64(frequency*blockNum-frequency))/0.001)*0.001, math.Trunc(res.TransactionsPerSecond/0.001)*0.001)
|
||||
// There is a -2 as time for first and last batch is not taken into account
|
||||
assert.InEpsilon(t, float64(frequency)*float64(numBatches-2)/float64(numBatches), res.BatchFrequency, 0.01)
|
||||
assert.InEpsilon(t, float64(numTx)/float64(frequency*blockNum-frequency), res.TransactionsPerSecond, 0.01)
|
||||
assert.Equal(t, int64(3), res.TotalAccounts)
|
||||
assert.Equal(t, int64(3), res.TotalBJJs)
|
||||
// Til does not set fees
|
||||
assert.Equal(t, float64(0), res.AvgTransactionFee)
|
||||
}
|
||||
|
||||
func TestGetMetricsMoreThan24Hours(t *testing.T) {
|
||||
func TestGetMetricsAPIMoreThan24Hours(t *testing.T) {
|
||||
test.WipeDB(historyDB.DB())
|
||||
|
||||
testUsersLen := 3
|
||||
@@ -1185,7 +1211,8 @@ func TestGetMetricsMoreThan24Hours(t *testing.T) {
|
||||
set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
||||
|
||||
// Transfers
|
||||
for x := 0; x < 6000; x++ {
|
||||
const numBlocks int = 30
|
||||
for x := 0; x < numBlocks; x++ {
|
||||
set = append(set, til.Instruction{
|
||||
Typ: common.TxTypeTransfer,
|
||||
TokenID: common.TokenID(0),
|
||||
@@ -1209,51 +1236,46 @@ func TestGetMetricsMoreThan24Hours(t *testing.T) {
|
||||
err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
||||
require.NoError(t, err)
|
||||
|
||||
const numBatches int = 6002
|
||||
const numTx int = 6003
|
||||
const blockNum = 6005 - 1
|
||||
const numBatches int = 2 + numBlocks
|
||||
const blockNum = 4 + numBlocks
|
||||
|
||||
// Sanity check
|
||||
require.Equal(t, blockNum, len(blocks))
|
||||
|
||||
// Adding one batch per block
|
||||
// batch frequency can be chosen
|
||||
const frequency int = 15
|
||||
const blockTime time.Duration = 3600 * time.Second
|
||||
now := time.Now()
|
||||
require.NoError(t, err)
|
||||
|
||||
for i := range blocks {
|
||||
blocks[i].Block.Timestamp = time.Now().Add(-time.Second * time.Duration(frequency*(len(blocks)-i)))
|
||||
blocks[i].Block.Timestamp = now.Add(-time.Duration(len(blocks)-1-i) * blockTime)
|
||||
err = historyDB.AddBlockSCData(&blocks[i])
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
res, err := historyDB.GetMetrics(common.BatchNum(numBatches))
|
||||
res, err := historyDBWithACC.GetMetricsAPI(common.BatchNum(numBatches))
|
||||
assert.NoError(t, err)
|
||||
|
||||
assert.Equal(t, math.Trunc((float64(numTx)/float64(numBatches-1))/0.001)*0.001, math.Trunc(res.TransactionsPerBatch/0.001)*0.001)
|
||||
assert.InEpsilon(t, 1.0, res.TransactionsPerBatch, 0.1)
|
||||
|
||||
// Frequency is not exactly the desired one, some decimals may appear
|
||||
assert.GreaterOrEqual(t, res.BatchFrequency, float64(frequency))
|
||||
assert.Less(t, res.BatchFrequency, float64(frequency+1))
|
||||
// Truncate frecuency into an int to do an exact check
|
||||
assert.Equal(t, frequency, int(res.BatchFrequency))
|
||||
// This may also be different in some decimals
|
||||
// Truncate it to the third decimal to compare
|
||||
assert.Equal(t, math.Trunc((float64(numTx)/float64(frequency*blockNum-frequency))/0.001)*0.001, math.Trunc(res.TransactionsPerSecond/0.001)*0.001)
|
||||
assert.InEpsilon(t, res.BatchFrequency, float64(blockTime/time.Second), 0.1)
|
||||
assert.InEpsilon(t, 1.0/float64(blockTime/time.Second), res.TransactionsPerSecond, 0.1)
|
||||
assert.Equal(t, int64(3), res.TotalAccounts)
|
||||
assert.Equal(t, int64(3), res.TotalBJJs)
|
||||
// Til does not set fees
|
||||
assert.Equal(t, float64(0), res.AvgTransactionFee)
|
||||
}
|
||||
|
||||
func TestGetMetricsEmpty(t *testing.T) {
|
||||
func TestGetMetricsAPIEmpty(t *testing.T) {
|
||||
test.WipeDB(historyDB.DB())
|
||||
_, err := historyDB.GetMetrics(0)
|
||||
_, err := historyDBWithACC.GetMetricsAPI(0)
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
func TestGetAvgTxFeeEmpty(t *testing.T) {
|
||||
test.WipeDB(historyDB.DB())
|
||||
_, err := historyDB.GetAvgTxFee()
|
||||
_, err := historyDBWithACC.GetAvgTxFeeAPI()
|
||||
assert.NoError(t, err)
|
||||
}
|
||||
|
||||
|
||||
@@ -147,6 +147,12 @@ type txWrite struct {
|
||||
Nonce *common.Nonce `meddler:"nonce"`
|
||||
}
|
||||
|
||||
// TokenSymbolAndAddr token representation with only Eth addr and symbol
|
||||
type TokenSymbolAndAddr struct {
|
||||
Symbol string `meddler:"symbol"`
|
||||
Addr ethCommon.Address `meddler:"eth_addr"`
|
||||
}
|
||||
|
||||
// TokenWithUSD add USD info to common.Token
|
||||
type TokenWithUSD struct {
|
||||
ItemID uint64 `json:"itemId" meddler:"item_id"`
|
||||
@@ -239,8 +245,8 @@ type AccountAPI struct {
|
||||
BatchNum common.BatchNum `meddler:"batch_num"`
|
||||
PublicKey apitypes.HezBJJ `meddler:"bjj"`
|
||||
EthAddr apitypes.HezEthAddr `meddler:"eth_addr"`
|
||||
Nonce common.Nonce `meddler:"-"` // max of 40 bits used
|
||||
Balance *apitypes.BigIntStr `meddler:"-"` // max of 192 bits used
|
||||
Nonce common.Nonce `meddler:"nonce"` // max of 40 bits used
|
||||
Balance *apitypes.BigIntStr `meddler:"balance"` // max of 192 bits used
|
||||
TotalItems uint64 `meddler:"total_items"`
|
||||
FirstItem uint64 `meddler:"first_item"`
|
||||
LastItem uint64 `meddler:"last_item"`
|
||||
@@ -304,12 +310,13 @@ type BatchAPI struct {
|
||||
|
||||
// Metrics define metrics of the network
|
||||
type Metrics struct {
|
||||
TransactionsPerBatch float64 `json:"transactionsPerBatch"`
|
||||
BatchFrequency float64 `json:"batchFrequency"`
|
||||
TransactionsPerSecond float64 `json:"transactionsPerSecond"`
|
||||
TotalAccounts int64 `json:"totalAccounts" meddler:"total_accounts"`
|
||||
TotalBJJs int64 `json:"totalBJJs" meddler:"total_bjjs"`
|
||||
AvgTransactionFee float64 `json:"avgTransactionFee"`
|
||||
TransactionsPerBatch float64 `json:"transactionsPerBatch"`
|
||||
BatchFrequency float64 `json:"batchFrequency"`
|
||||
TransactionsPerSecond float64 `json:"transactionsPerSecond"`
|
||||
TotalAccounts int64 `json:"totalAccounts" meddler:"total_accounts"`
|
||||
TotalBJJs int64 `json:"totalBJJs" meddler:"total_bjjs"`
|
||||
AvgTransactionFee float64 `json:"avgTransactionFee"`
|
||||
EstimatedTimeToForgeL1 float64 `json:"estimatedTimeToForgeL1" meddler:"estimated_time_to_forge_l1"`
|
||||
}
|
||||
|
||||
// MetricsTotals is used to get temporal information from HistoryDB
|
||||
|
||||
287
db/kvdb/kvdb.go
287
db/kvdb/kvdb.go
@@ -27,6 +27,8 @@ const (
|
||||
// PathLast defines the subpath of the last Batch in the subpath
|
||||
// of the StateDB
|
||||
PathLast = "last"
|
||||
// DefaultKeep is the default value for the Keep parameter
|
||||
DefaultKeep = 128
|
||||
)
|
||||
|
||||
var (
|
||||
@@ -34,16 +36,18 @@ var (
|
||||
KeyCurrentBatch = []byte("k:currentbatch")
|
||||
// keyCurrentIdx is used as key in the db to store the CurrentIdx
|
||||
keyCurrentIdx = []byte("k:idx")
|
||||
// ErrNoLast is returned when the KVDB has been configured to not have
|
||||
// a Last checkpoint but a Last method is used
|
||||
ErrNoLast = fmt.Errorf("no last checkpoint")
|
||||
)
|
||||
|
||||
// KVDB represents the Key-Value DB object
|
||||
type KVDB struct {
|
||||
path string
|
||||
db *pebble.Storage
|
||||
cfg Config
|
||||
db *pebble.Storage
|
||||
// CurrentIdx holds the current Idx that the BatchBuilder is using
|
||||
CurrentIdx common.Idx
|
||||
CurrentBatch common.BatchNum
|
||||
keep int
|
||||
m sync.Mutex
|
||||
last *Last
|
||||
}
|
||||
@@ -61,13 +65,13 @@ func (k *Last) setNew() error {
|
||||
defer k.rw.Unlock()
|
||||
if k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
lastPath := path.Join(k.path, PathLast)
|
||||
err := os.RemoveAll(lastPath)
|
||||
if err != nil {
|
||||
if err := os.RemoveAll(lastPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
db, err := pebble.NewPebbleStorage(path.Join(k.path, lastPath), false)
|
||||
db, err := pebble.NewPebbleStorage(lastPath, false)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -80,6 +84,7 @@ func (k *Last) set(kvdb *KVDB, batchNum common.BatchNum) error {
|
||||
defer k.rw.Unlock()
|
||||
if k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
lastPath := path.Join(k.path, PathLast)
|
||||
if err := kvdb.MakeCheckpointFromTo(batchNum, lastPath); err != nil {
|
||||
@@ -96,26 +101,48 @@ func (k *Last) set(kvdb *KVDB, batchNum common.BatchNum) error {
|
||||
func (k *Last) close() {
|
||||
k.rw.Lock()
|
||||
defer k.rw.Unlock()
|
||||
k.db.Close()
|
||||
if k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
}
|
||||
|
||||
// Config of the KVDB
|
||||
type Config struct {
|
||||
// Path where the checkpoints will be stored
|
||||
Path string
|
||||
// Keep is the number of old checkpoints to keep. If 0, all
|
||||
// checkpoints are kept.
|
||||
Keep int
|
||||
// At every checkpoint, check that there are no gaps between the
|
||||
// checkpoints
|
||||
NoGapsCheck bool
|
||||
// NoLast skips having an opened DB with a checkpoint to the last
|
||||
// batchNum for thread-safe reads.
|
||||
NoLast bool
|
||||
}
|
||||
|
||||
// NewKVDB creates a new KVDB, allowing to use an in-memory or in-disk storage.
|
||||
// Checkpoints older than the value defined by `keep` will be deleted.
|
||||
func NewKVDB(pathDB string, keep int) (*KVDB, error) {
|
||||
// func NewKVDB(pathDB string, keep int) (*KVDB, error) {
|
||||
func NewKVDB(cfg Config) (*KVDB, error) {
|
||||
var sto *pebble.Storage
|
||||
var err error
|
||||
sto, err = pebble.NewPebbleStorage(path.Join(pathDB, PathCurrent), false)
|
||||
sto, err = pebble.NewPebbleStorage(path.Join(cfg.Path, PathCurrent), false)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var last *Last
|
||||
if !cfg.NoLast {
|
||||
last = &Last{
|
||||
path: cfg.Path,
|
||||
}
|
||||
}
|
||||
kvdb := &KVDB{
|
||||
path: pathDB,
|
||||
cfg: cfg,
|
||||
db: sto,
|
||||
keep: keep,
|
||||
last: &Last{
|
||||
path: pathDB,
|
||||
},
|
||||
last: last,
|
||||
}
|
||||
// load currentBatch
|
||||
kvdb.CurrentBatch, err = kvdb.GetCurrentBatch()
|
||||
@@ -133,29 +160,32 @@ func NewKVDB(pathDB string, keep int) (*KVDB, error) {
|
||||
}
|
||||
|
||||
// LastRead is a thread-safe method to query the last KVDB
|
||||
func (kvdb *KVDB) LastRead(fn func(db *pebble.Storage) error) error {
|
||||
kvdb.last.rw.RLock()
|
||||
defer kvdb.last.rw.RUnlock()
|
||||
return fn(kvdb.last.db)
|
||||
func (k *KVDB) LastRead(fn func(db *pebble.Storage) error) error {
|
||||
if k.last == nil {
|
||||
return tracerr.Wrap(ErrNoLast)
|
||||
}
|
||||
k.last.rw.RLock()
|
||||
defer k.last.rw.RUnlock()
|
||||
return fn(k.last.db)
|
||||
}
|
||||
|
||||
// DB returns the *pebble.Storage from the KVDB
|
||||
func (kvdb *KVDB) DB() *pebble.Storage {
|
||||
return kvdb.db
|
||||
func (k *KVDB) DB() *pebble.Storage {
|
||||
return k.db
|
||||
}
|
||||
|
||||
// StorageWithPrefix returns the db.Storage with the given prefix from the
|
||||
// current KVDB
|
||||
func (kvdb *KVDB) StorageWithPrefix(prefix []byte) db.Storage {
|
||||
return kvdb.db.WithPrefix(prefix)
|
||||
func (k *KVDB) StorageWithPrefix(prefix []byte) db.Storage {
|
||||
return k.db.WithPrefix(prefix)
|
||||
}
|
||||
|
||||
// Reset resets the KVDB to the checkpoint at the given batchNum. Reset does
|
||||
// not delete the checkpoints between old current and the new current, those
|
||||
// checkpoints will remain in the storage, and eventually will be deleted when
|
||||
// MakeCheckpoint overwrites them.
|
||||
func (kvdb *KVDB) Reset(batchNum common.BatchNum) error {
|
||||
return kvdb.reset(batchNum, true)
|
||||
func (k *KVDB) Reset(batchNum common.BatchNum) error {
|
||||
return k.reset(batchNum, true)
|
||||
}
|
||||
|
||||
// reset resets the KVDB to the checkpoint at the given batchNum. Reset does
|
||||
@@ -163,21 +193,19 @@ func (kvdb *KVDB) Reset(batchNum common.BatchNum) error {
|
||||
// checkpoints will remain in the storage, and eventually will be deleted when
|
||||
// MakeCheckpoint overwrites them. `closeCurrent` will close the currently
|
||||
// opened db before doing the reset.
|
||||
func (kvdb *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
currentPath := path.Join(kvdb.path, PathCurrent)
|
||||
func (k *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
currentPath := path.Join(k.cfg.Path, PathCurrent)
|
||||
|
||||
if closeCurrent {
|
||||
if err := kvdb.db.Pebble().Close(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if closeCurrent && k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
// remove 'current'
|
||||
err := os.RemoveAll(currentPath)
|
||||
if err != nil {
|
||||
if err := os.RemoveAll(currentPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// remove all checkpoints > batchNum
|
||||
list, err := kvdb.ListCheckpoints()
|
||||
list, err := k.ListCheckpoints()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -190,7 +218,7 @@ func (kvdb *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
}
|
||||
}
|
||||
for _, bn := range list[start:] {
|
||||
if err := kvdb.DeleteCheckpoint(common.BatchNum(bn)); err != nil {
|
||||
if err := k.DeleteCheckpoint(common.BatchNum(bn)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
@@ -201,23 +229,27 @@ func (kvdb *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
kvdb.db = sto
|
||||
kvdb.CurrentIdx = common.RollupConstReservedIDx // 255
|
||||
kvdb.CurrentBatch = 0
|
||||
if err := kvdb.last.setNew(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
k.db = sto
|
||||
k.CurrentIdx = common.RollupConstReservedIDx // 255
|
||||
k.CurrentBatch = 0
|
||||
if k.last != nil {
|
||||
if err := k.last.setNew(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// copy 'batchNum' to 'current'
|
||||
if err := kvdb.MakeCheckpointFromTo(batchNum, currentPath); err != nil {
|
||||
if err := k.MakeCheckpointFromTo(batchNum, currentPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// copy 'batchNum' to 'last'
|
||||
if err := kvdb.last.set(kvdb, batchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
if k.last != nil {
|
||||
if err := k.last.set(k, batchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
|
||||
// open the new 'current'
|
||||
@@ -225,15 +257,15 @@ func (kvdb *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
kvdb.db = sto
|
||||
k.db = sto
|
||||
|
||||
// get currentBatch num
|
||||
kvdb.CurrentBatch, err = kvdb.GetCurrentBatch()
|
||||
k.CurrentBatch, err = k.GetCurrentBatch()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// idx is obtained from the statedb reset
|
||||
kvdb.CurrentIdx, err = kvdb.GetCurrentIdx()
|
||||
k.CurrentIdx, err = k.GetCurrentIdx()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -243,28 +275,28 @@ func (kvdb *KVDB) reset(batchNum common.BatchNum, closeCurrent bool) error {
|
||||
|
||||
// ResetFromSynchronizer performs a reset in the KVDB getting the state from
|
||||
// synchronizerKVDB for the given batchNum.
|
||||
func (kvdb *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKVDB *KVDB) error {
|
||||
func (k *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKVDB *KVDB) error {
|
||||
if synchronizerKVDB == nil {
|
||||
return tracerr.Wrap(fmt.Errorf("synchronizerKVDB can not be nil"))
|
||||
}
|
||||
|
||||
currentPath := path.Join(kvdb.path, PathCurrent)
|
||||
if err := kvdb.db.Pebble().Close(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
currentPath := path.Join(k.cfg.Path, PathCurrent)
|
||||
if k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
|
||||
// remove 'current'
|
||||
err := os.RemoveAll(currentPath)
|
||||
if err != nil {
|
||||
if err := os.RemoveAll(currentPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// remove all checkpoints
|
||||
list, err := kvdb.ListCheckpoints()
|
||||
list, err := k.ListCheckpoints()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
for _, bn := range list {
|
||||
if err := kvdb.DeleteCheckpoint(common.BatchNum(bn)); err != nil {
|
||||
if err := k.DeleteCheckpoint(common.BatchNum(bn)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
@@ -275,22 +307,22 @@ func (kvdb *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKV
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
kvdb.db = sto
|
||||
kvdb.CurrentIdx = common.RollupConstReservedIDx // 255
|
||||
kvdb.CurrentBatch = 0
|
||||
k.db = sto
|
||||
k.CurrentIdx = common.RollupConstReservedIDx // 255
|
||||
k.CurrentBatch = 0
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
checkpointPath := path.Join(kvdb.path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
checkpointPath := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
|
||||
// copy synchronizer'BatchNumX' to 'BatchNumX'
|
||||
// copy synchronizer 'BatchNumX' to 'BatchNumX'
|
||||
if err := synchronizerKVDB.MakeCheckpointFromTo(batchNum, checkpointPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// copy 'BatchNumX' to 'current'
|
||||
err = kvdb.MakeCheckpointFromTo(batchNum, currentPath)
|
||||
err = k.MakeCheckpointFromTo(batchNum, currentPath)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -300,15 +332,15 @@ func (kvdb *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKV
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
kvdb.db = sto
|
||||
k.db = sto
|
||||
|
||||
// get currentBatch num
|
||||
kvdb.CurrentBatch, err = kvdb.GetCurrentBatch()
|
||||
k.CurrentBatch, err = k.GetCurrentBatch()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// get currentIdx
|
||||
kvdb.CurrentIdx, err = kvdb.GetCurrentIdx()
|
||||
k.CurrentIdx, err = k.GetCurrentIdx()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -317,8 +349,8 @@ func (kvdb *KVDB) ResetFromSynchronizer(batchNum common.BatchNum, synchronizerKV
|
||||
}
|
||||
|
||||
// GetCurrentBatch returns the current BatchNum stored in the KVDB
|
||||
func (kvdb *KVDB) GetCurrentBatch() (common.BatchNum, error) {
|
||||
cbBytes, err := kvdb.db.Get(KeyCurrentBatch)
|
||||
func (k *KVDB) GetCurrentBatch() (common.BatchNum, error) {
|
||||
cbBytes, err := k.db.Get(KeyCurrentBatch)
|
||||
if tracerr.Unwrap(err) == db.ErrNotFound {
|
||||
return 0, nil
|
||||
}
|
||||
@@ -329,12 +361,12 @@ func (kvdb *KVDB) GetCurrentBatch() (common.BatchNum, error) {
|
||||
}
|
||||
|
||||
// setCurrentBatch stores the current BatchNum in the KVDB
|
||||
func (kvdb *KVDB) setCurrentBatch() error {
|
||||
tx, err := kvdb.db.NewTx()
|
||||
func (k *KVDB) setCurrentBatch() error {
|
||||
tx, err := k.db.NewTx()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
err = tx.Put(KeyCurrentBatch, kvdb.CurrentBatch.Bytes())
|
||||
err = tx.Put(KeyCurrentBatch, k.CurrentBatch.Bytes())
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -345,9 +377,9 @@ func (kvdb *KVDB) setCurrentBatch() error {
|
||||
}
|
||||
|
||||
// GetCurrentIdx returns the stored Idx from the KVDB, which is the last Idx
|
||||
// used for an Account in the KVDB.
|
||||
func (kvdb *KVDB) GetCurrentIdx() (common.Idx, error) {
|
||||
idxBytes, err := kvdb.db.Get(keyCurrentIdx)
|
||||
// used for an Account in the k.
|
||||
func (k *KVDB) GetCurrentIdx() (common.Idx, error) {
|
||||
idxBytes, err := k.db.Get(keyCurrentIdx)
|
||||
if tracerr.Unwrap(err) == db.ErrNotFound {
|
||||
return common.RollupConstReservedIDx, nil // 255, nil
|
||||
}
|
||||
@@ -358,10 +390,10 @@ func (kvdb *KVDB) GetCurrentIdx() (common.Idx, error) {
|
||||
}
|
||||
|
||||
// SetCurrentIdx stores Idx in the KVDB
|
||||
func (kvdb *KVDB) SetCurrentIdx(idx common.Idx) error {
|
||||
kvdb.CurrentIdx = idx
|
||||
func (k *KVDB) SetCurrentIdx(idx common.Idx) error {
|
||||
k.CurrentIdx = idx
|
||||
|
||||
tx, err := kvdb.db.NewTx()
|
||||
tx, err := k.db.NewTx()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -381,49 +413,64 @@ func (kvdb *KVDB) SetCurrentIdx(idx common.Idx) error {
|
||||
|
||||
// MakeCheckpoint does a checkpoint at the given batchNum in the defined path.
|
||||
// Internally this advances & stores the current BatchNum, and then stores a
|
||||
// Checkpoint of the current state of the KVDB.
|
||||
func (kvdb *KVDB) MakeCheckpoint() error {
|
||||
// Checkpoint of the current state of the k.
|
||||
func (k *KVDB) MakeCheckpoint() error {
|
||||
// advance currentBatch
|
||||
kvdb.CurrentBatch++
|
||||
k.CurrentBatch++
|
||||
|
||||
checkpointPath := path.Join(kvdb.path, fmt.Sprintf("%s%d", PathBatchNum, kvdb.CurrentBatch))
|
||||
checkpointPath := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, k.CurrentBatch))
|
||||
|
||||
if err := kvdb.setCurrentBatch(); err != nil {
|
||||
if err := k.setCurrentBatch(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// if checkpoint BatchNum already exist in disk, delete it
|
||||
if _, err := os.Stat(checkpointPath); !os.IsNotExist(err) {
|
||||
err := os.RemoveAll(checkpointPath)
|
||||
if err != nil {
|
||||
if _, err := os.Stat(checkpointPath); os.IsNotExist(err) {
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
} else {
|
||||
if err := os.RemoveAll(checkpointPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
} else if err != nil && !os.IsNotExist(err) {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// execute Checkpoint
|
||||
if err := kvdb.db.Pebble().Checkpoint(checkpointPath); err != nil {
|
||||
if err := k.db.Pebble().Checkpoint(checkpointPath); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// copy 'CurrentBatch' to 'last'
|
||||
if err := kvdb.last.set(kvdb, kvdb.CurrentBatch); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
if k.last != nil {
|
||||
if err := k.last.set(k, k.CurrentBatch); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
// delete old checkpoints
|
||||
if err := kvdb.deleteOldCheckpoints(); err != nil {
|
||||
if err := k.deleteOldCheckpoints(); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// CheckpointExists returns true if the checkpoint exists
|
||||
func (k *KVDB) CheckpointExists(batchNum common.BatchNum) (bool, error) {
|
||||
source := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
if _, err := os.Stat(source); os.IsNotExist(err) {
|
||||
return false, nil
|
||||
} else if err != nil {
|
||||
return false, tracerr.Wrap(err)
|
||||
}
|
||||
return true, nil
|
||||
}
|
||||
|
||||
// DeleteCheckpoint removes if exist the checkpoint of the given batchNum
|
||||
func (kvdb *KVDB) DeleteCheckpoint(batchNum common.BatchNum) error {
|
||||
checkpointPath := path.Join(kvdb.path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
func (k *KVDB) DeleteCheckpoint(batchNum common.BatchNum) error {
|
||||
checkpointPath := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, batchNum))
|
||||
|
||||
if _, err := os.Stat(checkpointPath); os.IsNotExist(err) {
|
||||
return tracerr.Wrap(fmt.Errorf("Checkpoint with batchNum %d does not exist in DB", batchNum))
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
return os.RemoveAll(checkpointPath)
|
||||
@@ -431,8 +478,8 @@ func (kvdb *KVDB) DeleteCheckpoint(batchNum common.BatchNum) error {
|
||||
|
||||
// ListCheckpoints returns the list of batchNums of the checkpoints, sorted.
|
||||
// If there's a gap between the list of checkpoints, an error is returned.
|
||||
func (kvdb *KVDB) ListCheckpoints() ([]int, error) {
|
||||
files, err := ioutil.ReadDir(kvdb.path)
|
||||
func (k *KVDB) ListCheckpoints() ([]int, error) {
|
||||
files, err := ioutil.ReadDir(k.cfg.Path)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -449,12 +496,12 @@ func (kvdb *KVDB) ListCheckpoints() ([]int, error) {
|
||||
}
|
||||
}
|
||||
sort.Ints(checkpoints)
|
||||
if len(checkpoints) > 0 {
|
||||
if !k.cfg.NoGapsCheck && len(checkpoints) > 0 {
|
||||
first := checkpoints[0]
|
||||
for _, checkpoint := range checkpoints[1:] {
|
||||
first++
|
||||
if checkpoint != first {
|
||||
log.Errorw("GAP", "checkpoints", checkpoints)
|
||||
log.Errorw("gap between checkpoints", "checkpoints", checkpoints)
|
||||
return nil, tracerr.Wrap(fmt.Errorf("checkpoint gap at %v", checkpoint))
|
||||
}
|
||||
}
|
||||
@@ -464,14 +511,14 @@ func (kvdb *KVDB) ListCheckpoints() ([]int, error) {
|
||||
|
||||
// deleteOldCheckpoints deletes old checkpoints when there are more than
|
||||
// `s.keep` checkpoints
|
||||
func (kvdb *KVDB) deleteOldCheckpoints() error {
|
||||
list, err := kvdb.ListCheckpoints()
|
||||
func (k *KVDB) deleteOldCheckpoints() error {
|
||||
list, err := k.ListCheckpoints()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if len(list) > kvdb.keep {
|
||||
for _, checkpoint := range list[:len(list)-kvdb.keep] {
|
||||
if err := kvdb.DeleteCheckpoint(common.BatchNum(checkpoint)); err != nil {
|
||||
if k.cfg.Keep > 0 && len(list) > k.cfg.Keep {
|
||||
for _, checkpoint := range list[:len(list)-k.cfg.Keep] {
|
||||
if err := k.DeleteCheckpoint(common.BatchNum(checkpoint)); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
@@ -482,43 +529,42 @@ func (kvdb *KVDB) deleteOldCheckpoints() error {
|
||||
// MakeCheckpointFromTo makes a checkpoint from the current db at fromBatchNum
|
||||
// to the dest folder. This method is locking, so it can be called from
|
||||
// multiple places at the same time.
|
||||
func (kvdb *KVDB) MakeCheckpointFromTo(fromBatchNum common.BatchNum, dest string) error {
|
||||
source := path.Join(kvdb.path, fmt.Sprintf("%s%d", PathBatchNum, fromBatchNum))
|
||||
func (k *KVDB) MakeCheckpointFromTo(fromBatchNum common.BatchNum, dest string) error {
|
||||
source := path.Join(k.cfg.Path, fmt.Sprintf("%s%d", PathBatchNum, fromBatchNum))
|
||||
if _, err := os.Stat(source); os.IsNotExist(err) {
|
||||
// if kvdb does not have checkpoint at batchNum, return err
|
||||
return tracerr.Wrap(fmt.Errorf("Checkpoint \"%v\" does not exist", source))
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// By locking we allow calling MakeCheckpointFromTo from multiple
|
||||
// places at the same time for the same stateDB. This allows the
|
||||
// synchronizer to do a reset to a batchNum at the same time as the
|
||||
// pipeline is doing a txSelector.Reset and batchBuilder.Reset from
|
||||
// synchronizer to the same batchNum
|
||||
kvdb.m.Lock()
|
||||
defer kvdb.m.Unlock()
|
||||
return pebbleMakeCheckpoint(source, dest)
|
||||
k.m.Lock()
|
||||
defer k.m.Unlock()
|
||||
return PebbleMakeCheckpoint(source, dest)
|
||||
}
|
||||
|
||||
func pebbleMakeCheckpoint(source, dest string) error {
|
||||
// PebbleMakeCheckpoint is a hepler function to make a pebble checkpoint from
|
||||
// source to dest.
|
||||
func PebbleMakeCheckpoint(source, dest string) error {
|
||||
// Remove dest folder (if it exists) before doing the checkpoint
|
||||
if _, err := os.Stat(dest); !os.IsNotExist(err) {
|
||||
err := os.RemoveAll(dest)
|
||||
if err != nil {
|
||||
if _, err := os.Stat(dest); os.IsNotExist(err) {
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
} else {
|
||||
if err := os.RemoveAll(dest); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
} else if err != nil && !os.IsNotExist(err) {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
sto, err := pebble.NewPebbleStorage(source, false)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
defer func() {
|
||||
errClose := sto.Pebble().Close()
|
||||
if errClose != nil {
|
||||
log.Errorw("Pebble.Close", "err", errClose)
|
||||
}
|
||||
}()
|
||||
defer sto.Close()
|
||||
|
||||
// execute Checkpoint
|
||||
err = sto.Pebble().Checkpoint(dest)
|
||||
@@ -530,7 +576,12 @@ func pebbleMakeCheckpoint(source, dest string) error {
|
||||
}
|
||||
|
||||
// Close the DB
|
||||
func (kvdb *KVDB) Close() {
|
||||
kvdb.db.Close()
|
||||
kvdb.last.close()
|
||||
func (k *KVDB) Close() {
|
||||
if k.db != nil {
|
||||
k.db.Close()
|
||||
k.db = nil
|
||||
}
|
||||
if k.last != nil {
|
||||
k.last.close()
|
||||
}
|
||||
}
|
||||
|
||||
@@ -37,7 +37,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
db, err := NewKVDB(dir, 128)
|
||||
db, err := NewKVDB(Config{Path: dir, Keep: 128})
|
||||
require.NoError(t, err)
|
||||
|
||||
// add test key-values
|
||||
@@ -72,7 +72,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
err = db.Reset(3)
|
||||
require.NoError(t, err)
|
||||
|
||||
printCheckpoints(t, db.path)
|
||||
printCheckpoints(t, db.cfg.Path)
|
||||
|
||||
// check that currentBatch is as expected after Reset
|
||||
cb, err = db.GetCurrentBatch()
|
||||
@@ -99,7 +99,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal, err := ioutil.TempDir("", "ldb")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal))
|
||||
ldb, err := NewKVDB(dirLocal, 128)
|
||||
ldb, err := NewKVDB(Config{Path: dirLocal, Keep: 128})
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
@@ -120,7 +120,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal2, err := ioutil.TempDir("", "ldb2")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal2))
|
||||
ldb2, err := NewKVDB(dirLocal2, 128)
|
||||
ldb2, err := NewKVDB(Config{Path: dirLocal2, Keep: 128})
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
@@ -139,9 +139,9 @@ func TestCheckpoints(t *testing.T) {
|
||||
|
||||
debug := false
|
||||
if debug {
|
||||
printCheckpoints(t, db.path)
|
||||
printCheckpoints(t, ldb.path)
|
||||
printCheckpoints(t, ldb2.path)
|
||||
printCheckpoints(t, db.cfg.Path)
|
||||
printCheckpoints(t, ldb.cfg.Path)
|
||||
printCheckpoints(t, ldb2.cfg.Path)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -150,7 +150,7 @@ func TestListCheckpoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
db, err := NewKVDB(dir, 128)
|
||||
db, err := NewKVDB(Config{Path: dir, Keep: 128})
|
||||
require.NoError(t, err)
|
||||
|
||||
numCheckpoints := 16
|
||||
@@ -181,7 +181,7 @@ func TestDeleteOldCheckpoints(t *testing.T) {
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
keep := 16
|
||||
db, err := NewKVDB(dir, keep)
|
||||
db, err := NewKVDB(Config{Path: dir, Keep: keep})
|
||||
require.NoError(t, err)
|
||||
|
||||
numCheckpoints := 32
|
||||
@@ -202,7 +202,7 @@ func TestGetCurrentIdx(t *testing.T) {
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
keep := 16
|
||||
db, err := NewKVDB(dir, keep)
|
||||
db, err := NewKVDB(Config{Path: dir, Keep: keep})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx, err := db.GetCurrentIdx()
|
||||
@@ -211,7 +211,7 @@ func TestGetCurrentIdx(t *testing.T) {
|
||||
|
||||
db.Close()
|
||||
|
||||
db, err = NewKVDB(dir, keep)
|
||||
db, err = NewKVDB(Config{Path: dir, Keep: keep})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx, err = db.GetCurrentIdx()
|
||||
@@ -227,7 +227,7 @@ func TestGetCurrentIdx(t *testing.T) {
|
||||
|
||||
db.Close()
|
||||
|
||||
db, err = NewKVDB(dir, keep)
|
||||
db, err = NewKVDB(Config{Path: dir, Keep: keep})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx, err = db.GetCurrentIdx()
|
||||
|
||||
125
db/l2db/apiqueries.go
Normal file
125
db/l2db/apiqueries.go
Normal file
@@ -0,0 +1,125 @@
|
||||
package l2db
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/russross/meddler"
|
||||
)
|
||||
|
||||
var (
|
||||
errPoolFull = fmt.Errorf("the pool is at full capacity. More transactions are not accepted currently")
|
||||
)
|
||||
|
||||
// AddAccountCreationAuthAPI inserts an account creation authorization into the DB
|
||||
func (l2db *L2DB) AddAccountCreationAuthAPI(auth *common.AccountCreationAuth) error {
|
||||
cancel, err := l2db.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
defer l2db.apiConnCon.Release()
|
||||
return l2db.AddAccountCreationAuth(auth)
|
||||
}
|
||||
|
||||
// GetAccountCreationAuthAPI returns an account creation authorization from the DB
|
||||
func (l2db *L2DB) GetAccountCreationAuthAPI(addr ethCommon.Address) (*AccountCreationAuthAPI, error) {
|
||||
cancel, err := l2db.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
defer l2db.apiConnCon.Release()
|
||||
auth := new(AccountCreationAuthAPI)
|
||||
return auth, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.dbRead, auth,
|
||||
"SELECT * FROM account_creation_auth WHERE eth_addr = $1;",
|
||||
addr,
|
||||
))
|
||||
}
|
||||
|
||||
// AddTxAPI inserts a tx to the pool
|
||||
func (l2db *L2DB) AddTxAPI(tx *PoolL2TxWrite) error {
|
||||
cancel, err := l2db.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
defer l2db.apiConnCon.Release()
|
||||
|
||||
row := l2db.dbRead.QueryRow(`SELECT
|
||||
($1::NUMERIC * COALESCE(token.usd, 0) * fee_percentage($2::NUMERIC)) /
|
||||
(10.0 ^ token.decimals::NUMERIC)
|
||||
FROM token WHERE token.token_id = $3;`,
|
||||
tx.AmountFloat, tx.Fee, tx.TokenID)
|
||||
var feeUSD float64
|
||||
if err := row.Scan(&feeUSD); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if feeUSD < l2db.minFeeUSD {
|
||||
return tracerr.Wrap(fmt.Errorf("tx.feeUSD (%v) < minFeeUSD (%v)",
|
||||
feeUSD, l2db.minFeeUSD))
|
||||
}
|
||||
|
||||
// Prepare insert SQL query argument parameters
|
||||
namesPart, err := meddler.Default.ColumnsQuoted(tx, false)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
valuesPart, err := meddler.Default.PlaceholdersString(tx, false)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
values, err := meddler.Default.Values(tx, false)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
q := fmt.Sprintf(
|
||||
`INSERT INTO tx_pool (%s)
|
||||
SELECT %s
|
||||
WHERE (SELECT COUNT(*) FROM tx_pool WHERE state = $%v) < $%v;`,
|
||||
namesPart, valuesPart,
|
||||
len(values)+1, len(values)+2) //nolint:gomnd
|
||||
values = append(values, common.PoolL2TxStatePending, l2db.maxTxs)
|
||||
res, err := l2db.dbWrite.Exec(q, values...)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
rowsAffected, err := res.RowsAffected()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if rowsAffected == 0 {
|
||||
return tracerr.Wrap(errPoolFull)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// selectPoolTxAPI select part of queries to get PoolL2TxRead
|
||||
const selectPoolTxAPI = `SELECT tx_pool.tx_id, hez_idx(tx_pool.from_idx, token.symbol) AS from_idx, tx_pool.effective_from_eth_addr,
|
||||
tx_pool.effective_from_bjj, hez_idx(tx_pool.to_idx, token.symbol) AS to_idx, tx_pool.effective_to_eth_addr,
|
||||
tx_pool.effective_to_bjj, tx_pool.token_id, tx_pool.amount, tx_pool.fee, tx_pool.nonce,
|
||||
tx_pool.state, tx_pool.info, tx_pool.signature, tx_pool.timestamp, tx_pool.batch_num, hez_idx(tx_pool.rq_from_idx, token.symbol) AS rq_from_idx,
|
||||
hez_idx(tx_pool.rq_to_idx, token.symbol) AS rq_to_idx, tx_pool.rq_to_eth_addr, tx_pool.rq_to_bjj, tx_pool.rq_token_id, tx_pool.rq_amount,
|
||||
tx_pool.rq_fee, tx_pool.rq_nonce, tx_pool.tx_type,
|
||||
token.item_id AS token_item_id, token.eth_block_num, token.eth_addr, token.name, token.symbol, token.decimals, token.usd, token.usd_update
|
||||
FROM tx_pool INNER JOIN token ON tx_pool.token_id = token.token_id `
|
||||
|
||||
// GetTxAPI return the specified Tx in PoolTxAPI format
|
||||
func (l2db *L2DB) GetTxAPI(txID common.TxID) (*PoolTxAPI, error) {
|
||||
cancel, err := l2db.apiConnCon.Acquire()
|
||||
defer cancel()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
defer l2db.apiConnCon.Release()
|
||||
tx := new(PoolTxAPI)
|
||||
return tx, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.dbRead, tx,
|
||||
selectPoolTxAPI+"WHERE tx_id = $1;",
|
||||
txID,
|
||||
))
|
||||
}
|
||||
151
db/l2db/l2db.go
151
db/l2db/l2db.go
@@ -21,34 +21,52 @@ import (
|
||||
// L2DB stores L2 txs and authorization registers received by the coordinator and keeps them until they are no longer relevant
|
||||
// due to them being forged or invalid after a safety period
|
||||
type L2DB struct {
|
||||
db *sqlx.DB
|
||||
dbRead *sqlx.DB
|
||||
dbWrite *sqlx.DB
|
||||
safetyPeriod common.BatchNum
|
||||
ttl time.Duration
|
||||
maxTxs uint32 // limit of txs that are accepted in the pool
|
||||
minFeeUSD float64
|
||||
apiConnCon *db.APIConnectionController
|
||||
}
|
||||
|
||||
// NewL2DB creates a L2DB.
|
||||
// To create it, it's needed db connection, safety period expressed in batches,
|
||||
// maxTxs that the DB should have and TTL (time to live) for pending txs.
|
||||
func NewL2DB(db *sqlx.DB, safetyPeriod common.BatchNum, maxTxs uint32, TTL time.Duration) *L2DB {
|
||||
func NewL2DB(
|
||||
dbRead, dbWrite *sqlx.DB,
|
||||
safetyPeriod common.BatchNum,
|
||||
maxTxs uint32,
|
||||
minFeeUSD float64,
|
||||
TTL time.Duration,
|
||||
apiConnCon *db.APIConnectionController,
|
||||
) *L2DB {
|
||||
return &L2DB{
|
||||
db: db,
|
||||
dbRead: dbRead,
|
||||
dbWrite: dbWrite,
|
||||
safetyPeriod: safetyPeriod,
|
||||
ttl: TTL,
|
||||
maxTxs: maxTxs,
|
||||
minFeeUSD: minFeeUSD,
|
||||
apiConnCon: apiConnCon,
|
||||
}
|
||||
}
|
||||
|
||||
// DB returns a pointer to the L2DB.db. This method should be used only for
|
||||
// internal testing purposes.
|
||||
func (l2db *L2DB) DB() *sqlx.DB {
|
||||
return l2db.db
|
||||
return l2db.dbWrite
|
||||
}
|
||||
|
||||
// MinFeeUSD returns the minimum fee in USD that is required to accept txs into
|
||||
// the pool
|
||||
func (l2db *L2DB) MinFeeUSD() float64 {
|
||||
return l2db.minFeeUSD
|
||||
}
|
||||
|
||||
// AddAccountCreationAuth inserts an account creation authorization into the DB
|
||||
func (l2db *L2DB) AddAccountCreationAuth(auth *common.AccountCreationAuth) error {
|
||||
// return meddler.Insert(l2db.db, "account_creation_auth", auth)
|
||||
_, err := l2db.db.Exec(
|
||||
_, err := l2db.dbWrite.Exec(
|
||||
`INSERT INTO account_creation_auth (eth_addr, bjj, signature)
|
||||
VALUES ($1, $2, $3);`,
|
||||
auth.EthAddr, auth.BJJ, auth.Signature,
|
||||
@@ -56,44 +74,26 @@ func (l2db *L2DB) AddAccountCreationAuth(auth *common.AccountCreationAuth) error
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// AddManyAccountCreationAuth inserts a batch of accounts creation authorization
|
||||
// if not exist into the DB
|
||||
func (l2db *L2DB) AddManyAccountCreationAuth(auths []common.AccountCreationAuth) error {
|
||||
_, err := sqlx.NamedExec(l2db.dbWrite,
|
||||
`INSERT INTO account_creation_auth (eth_addr, bjj, signature)
|
||||
VALUES (:ethaddr, :bjj, :signature)
|
||||
ON CONFLICT (eth_addr) DO NOTHING`, auths)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// GetAccountCreationAuth returns an account creation authorization from the DB
|
||||
func (l2db *L2DB) GetAccountCreationAuth(addr ethCommon.Address) (*common.AccountCreationAuth, error) {
|
||||
auth := new(common.AccountCreationAuth)
|
||||
return auth, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.db, auth,
|
||||
l2db.dbRead, auth,
|
||||
"SELECT * FROM account_creation_auth WHERE eth_addr = $1;",
|
||||
addr,
|
||||
))
|
||||
}
|
||||
|
||||
// GetAccountCreationAuthAPI returns an account creation authorization from the DB
|
||||
func (l2db *L2DB) GetAccountCreationAuthAPI(addr ethCommon.Address) (*AccountCreationAuthAPI, error) {
|
||||
auth := new(AccountCreationAuthAPI)
|
||||
return auth, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.db, auth,
|
||||
"SELECT * FROM account_creation_auth WHERE eth_addr = $1;",
|
||||
addr,
|
||||
))
|
||||
}
|
||||
|
||||
// AddTx inserts a tx to the pool
|
||||
func (l2db *L2DB) AddTx(tx *PoolL2TxWrite) error {
|
||||
row := l2db.db.QueryRow(
|
||||
"SELECT COUNT(*) FROM tx_pool WHERE state = $1;",
|
||||
common.PoolL2TxStatePending,
|
||||
)
|
||||
var totalTxs uint32
|
||||
if err := row.Scan(&totalTxs); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if totalTxs >= l2db.maxTxs {
|
||||
return tracerr.New(
|
||||
"The pool is at full capacity. More transactions are not accepted currently",
|
||||
)
|
||||
}
|
||||
return tracerr.Wrap(meddler.Insert(l2db.db, "tx_pool", tx))
|
||||
}
|
||||
|
||||
// UpdateTxsInfo updates the parameter Info of the pool transactions
|
||||
func (l2db *L2DB) UpdateTxsInfo(txs []common.PoolL2Tx) error {
|
||||
if len(txs) == 0 {
|
||||
@@ -117,7 +117,7 @@ func (l2db *L2DB) UpdateTxsInfo(txs []common.PoolL2Tx) error {
|
||||
WHERE tx_pool.tx_id = tx_update.id;
|
||||
`
|
||||
if len(txUpdates) > 0 {
|
||||
if _, err := sqlx.NamedExec(l2db.db, query, txUpdates); err != nil {
|
||||
if _, err := sqlx.NamedExec(l2db.dbWrite, query, txUpdates); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
@@ -125,9 +125,8 @@ func (l2db *L2DB) UpdateTxsInfo(txs []common.PoolL2Tx) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// AddTxTest inserts a tx into the L2DB. This is useful for test purposes,
|
||||
// but in production txs will only be inserted through the API
|
||||
func (l2db *L2DB) AddTxTest(tx *common.PoolL2Tx) error {
|
||||
// NewPoolL2TxWriteFromPoolL2Tx creates a new PoolL2TxWrite from a PoolL2Tx
|
||||
func NewPoolL2TxWriteFromPoolL2Tx(tx *common.PoolL2Tx) *PoolL2TxWrite {
|
||||
// transform tx from *common.PoolL2Tx to PoolL2TxWrite
|
||||
insertTx := &PoolL2TxWrite{
|
||||
TxID: tx.TxID,
|
||||
@@ -169,19 +168,16 @@ func (l2db *L2DB) AddTxTest(tx *common.PoolL2Tx) error {
|
||||
f := new(big.Float).SetInt(tx.Amount)
|
||||
amountF, _ := f.Float64()
|
||||
insertTx.AmountFloat = amountF
|
||||
// insert tx
|
||||
return tracerr.Wrap(meddler.Insert(l2db.db, "tx_pool", insertTx))
|
||||
return insertTx
|
||||
}
|
||||
|
||||
// selectPoolTxAPI select part of queries to get PoolL2TxRead
|
||||
const selectPoolTxAPI = `SELECT tx_pool.tx_id, hez_idx(tx_pool.from_idx, token.symbol) AS from_idx, tx_pool.effective_from_eth_addr,
|
||||
tx_pool.effective_from_bjj, hez_idx(tx_pool.to_idx, token.symbol) AS to_idx, tx_pool.effective_to_eth_addr,
|
||||
tx_pool.effective_to_bjj, tx_pool.token_id, tx_pool.amount, tx_pool.fee, tx_pool.nonce,
|
||||
tx_pool.state, tx_pool.info, tx_pool.signature, tx_pool.timestamp, tx_pool.batch_num, hez_idx(tx_pool.rq_from_idx, token.symbol) AS rq_from_idx,
|
||||
hez_idx(tx_pool.rq_to_idx, token.symbol) AS rq_to_idx, tx_pool.rq_to_eth_addr, tx_pool.rq_to_bjj, tx_pool.rq_token_id, tx_pool.rq_amount,
|
||||
tx_pool.rq_fee, tx_pool.rq_nonce, tx_pool.tx_type,
|
||||
token.item_id AS token_item_id, token.eth_block_num, token.eth_addr, token.name, token.symbol, token.decimals, token.usd, token.usd_update
|
||||
FROM tx_pool INNER JOIN token ON tx_pool.token_id = token.token_id `
|
||||
// AddTxTest inserts a tx into the L2DB. This is useful for test purposes,
|
||||
// but in production txs will only be inserted through the API
|
||||
func (l2db *L2DB) AddTxTest(tx *common.PoolL2Tx) error {
|
||||
insertTx := NewPoolL2TxWriteFromPoolL2Tx(tx)
|
||||
// insert tx
|
||||
return tracerr.Wrap(meddler.Insert(l2db.dbWrite, "tx_pool", insertTx))
|
||||
}
|
||||
|
||||
// selectPoolTxCommon select part of queries to get common.PoolL2Tx
|
||||
const selectPoolTxCommon = `SELECT tx_pool.tx_id, from_idx, to_idx, tx_pool.to_eth_addr,
|
||||
@@ -189,34 +185,25 @@ tx_pool.to_bjj, tx_pool.token_id, tx_pool.amount, tx_pool.fee, tx_pool.nonce,
|
||||
tx_pool.state, tx_pool.info, tx_pool.signature, tx_pool.timestamp, rq_from_idx,
|
||||
rq_to_idx, tx_pool.rq_to_eth_addr, tx_pool.rq_to_bjj, tx_pool.rq_token_id, tx_pool.rq_amount,
|
||||
tx_pool.rq_fee, tx_pool.rq_nonce, tx_pool.tx_type,
|
||||
fee_percentage(tx_pool.fee::NUMERIC) * token.usd * tx_pool.amount_f AS fee_usd, token.usd_update
|
||||
(fee_percentage(tx_pool.fee::NUMERIC) * token.usd * tx_pool.amount_f) /
|
||||
(10.0 ^ token.decimals::NUMERIC) AS fee_usd, token.usd_update
|
||||
FROM tx_pool INNER JOIN token ON tx_pool.token_id = token.token_id `
|
||||
|
||||
// GetTx return the specified Tx in common.PoolL2Tx format
|
||||
func (l2db *L2DB) GetTx(txID common.TxID) (*common.PoolL2Tx, error) {
|
||||
tx := new(common.PoolL2Tx)
|
||||
return tx, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.db, tx,
|
||||
l2db.dbRead, tx,
|
||||
selectPoolTxCommon+"WHERE tx_id = $1;",
|
||||
txID,
|
||||
))
|
||||
}
|
||||
|
||||
// GetTxAPI return the specified Tx in PoolTxAPI format
|
||||
func (l2db *L2DB) GetTxAPI(txID common.TxID) (*PoolTxAPI, error) {
|
||||
tx := new(PoolTxAPI)
|
||||
return tx, tracerr.Wrap(meddler.QueryRow(
|
||||
l2db.db, tx,
|
||||
selectPoolTxAPI+"WHERE tx_id = $1;",
|
||||
txID,
|
||||
))
|
||||
}
|
||||
|
||||
// GetPendingTxs return all the pending txs of the L2DB, that have a non NULL AbsoluteFee
|
||||
func (l2db *L2DB) GetPendingTxs() ([]common.PoolL2Tx, error) {
|
||||
var txs []*common.PoolL2Tx
|
||||
err := meddler.QueryAll(
|
||||
l2db.db, &txs,
|
||||
l2db.dbRead, &txs,
|
||||
selectPoolTxCommon+"WHERE state = $1",
|
||||
common.PoolL2TxStatePending,
|
||||
)
|
||||
@@ -241,8 +228,8 @@ func (l2db *L2DB) StartForging(txIDs []common.TxID, batchNum common.BatchNum) er
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
query = l2db.db.Rebind(query)
|
||||
_, err = l2db.db.Exec(query, args...)
|
||||
query = l2db.dbWrite.Rebind(query)
|
||||
_, err = l2db.dbWrite.Exec(query, args...)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -264,8 +251,8 @@ func (l2db *L2DB) DoneForging(txIDs []common.TxID, batchNum common.BatchNum) err
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
query = l2db.db.Rebind(query)
|
||||
_, err = l2db.db.Exec(query, args...)
|
||||
query = l2db.dbWrite.Rebind(query)
|
||||
_, err = l2db.dbWrite.Exec(query, args...)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -286,8 +273,8 @@ func (l2db *L2DB) InvalidateTxs(txIDs []common.TxID, batchNum common.BatchNum) e
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
query = l2db.db.Rebind(query)
|
||||
_, err = l2db.db.Exec(query, args...)
|
||||
query = l2db.dbWrite.Rebind(query)
|
||||
_, err = l2db.dbWrite.Exec(query, args...)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -295,7 +282,7 @@ func (l2db *L2DB) InvalidateTxs(txIDs []common.TxID, batchNum common.BatchNum) e
|
||||
// of unique FromIdx
|
||||
func (l2db *L2DB) GetPendingUniqueFromIdxs() ([]common.Idx, error) {
|
||||
var idxs []common.Idx
|
||||
rows, err := l2db.db.Query(`SELECT DISTINCT from_idx FROM tx_pool
|
||||
rows, err := l2db.dbRead.Query(`SELECT DISTINCT from_idx FROM tx_pool
|
||||
WHERE state = $1;`, common.PoolL2TxStatePending)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -333,10 +320,10 @@ func (l2db *L2DB) InvalidateOldNonces(updatedAccounts []common.IdxNonce, batchNu
|
||||
return nil
|
||||
}
|
||||
// Fill the batch_num in the query with Sprintf because we are using a
|
||||
// named query which works with slices, and doens't handle an extra
|
||||
// named query which works with slices, and doesn't handle an extra
|
||||
// individual argument.
|
||||
query := fmt.Sprintf(invalidateOldNoncesQuery, batchNum)
|
||||
if _, err := sqlx.NamedExec(l2db.db, query, updatedAccounts); err != nil {
|
||||
if _, err := sqlx.NamedExec(l2db.dbWrite, query, updatedAccounts); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
@@ -345,10 +332,11 @@ func (l2db *L2DB) InvalidateOldNonces(updatedAccounts []common.IdxNonce, batchNu
|
||||
// Reorg updates the state of txs that were updated in a batch that has been discarted due to a blockchain reorg.
|
||||
// The state of the affected txs can change form Forged -> Pending or from Invalid -> Pending
|
||||
func (l2db *L2DB) Reorg(lastValidBatch common.BatchNum) error {
|
||||
_, err := l2db.db.Exec(
|
||||
`UPDATE tx_pool SET batch_num = NULL, state = $1
|
||||
WHERE (state = $2 OR state = $3) AND batch_num > $4`,
|
||||
_, err := l2db.dbWrite.Exec(
|
||||
`UPDATE tx_pool SET batch_num = NULL, state = $1
|
||||
WHERE (state = $2 OR state = $3 OR state = $4) AND batch_num > $5`,
|
||||
common.PoolL2TxStatePending,
|
||||
common.PoolL2TxStateForging,
|
||||
common.PoolL2TxStateForged,
|
||||
common.PoolL2TxStateInvalid,
|
||||
lastValidBatch,
|
||||
@@ -360,7 +348,7 @@ func (l2db *L2DB) Reorg(lastValidBatch common.BatchNum) error {
|
||||
// it also deletes pending txs that have been in the L2DB for longer than the ttl if maxTxs has been exceeded
|
||||
func (l2db *L2DB) Purge(currentBatchNum common.BatchNum) (err error) {
|
||||
now := time.Now().UTC().Unix()
|
||||
_, err = l2db.db.Exec(
|
||||
_, err = l2db.dbWrite.Exec(
|
||||
`DELETE FROM tx_pool WHERE (
|
||||
batch_num < $1 AND (state = $2 OR state = $3)
|
||||
) OR (
|
||||
@@ -376,3 +364,14 @@ func (l2db *L2DB) Purge(currentBatchNum common.BatchNum) (err error) {
|
||||
)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// PurgeByExternalDelete deletes all pending transactions marked with true in
|
||||
// the `external_delete` column. An external process can set this column to
|
||||
// true to instruct the coordinator to delete the tx when possible.
|
||||
func (l2db *L2DB) PurgeByExternalDelete() error {
|
||||
_, err := l2db.dbWrite.Exec(
|
||||
`DELETE from tx_pool WHERE (external_delete = true AND state = $1);`,
|
||||
common.PoolL2TxStatePending,
|
||||
)
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -1,8 +1,8 @@
|
||||
package l2db
|
||||
|
||||
import (
|
||||
"math"
|
||||
"math/big"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"os"
|
||||
"testing"
|
||||
"time"
|
||||
@@ -20,11 +20,14 @@ import (
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
var decimals = uint64(3)
|
||||
var tokenValue = 1.0 // The price update gives a value of 1.0 USD to the token
|
||||
var l2DB *L2DB
|
||||
var l2DBWithACC *L2DB
|
||||
var historyDB *historydb.HistoryDB
|
||||
var tc *til.Context
|
||||
var tokens map[common.TokenID]historydb.TokenWithUSD
|
||||
var tokensValue map[common.TokenID]float64
|
||||
|
||||
var accs map[common.Idx]common.Account
|
||||
|
||||
func TestMain(m *testing.M) {
|
||||
@@ -34,9 +37,11 @@ func TestMain(m *testing.M) {
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
l2DB = NewL2DB(db, 10, 1000, 24*time.Hour)
|
||||
l2DB = NewL2DB(db, db, 10, 1000, 0.0, 24*time.Hour, nil)
|
||||
apiConnCon := dbUtils.NewAPIConnectionController(1, time.Second)
|
||||
l2DBWithACC = NewL2DB(db, db, 10, 1000, 0.0, 24*time.Hour, apiConnCon)
|
||||
test.WipeDB(l2DB.DB())
|
||||
historyDB = historydb.NewHistoryDB(db)
|
||||
historyDB = historydb.NewHistoryDB(db, db, nil)
|
||||
// Run tests
|
||||
result := m.Run()
|
||||
// Close DB
|
||||
@@ -55,10 +60,10 @@ func prepareHistoryDB(historyDB *historydb.HistoryDB) error {
|
||||
|
||||
AddToken(1)
|
||||
AddToken(2)
|
||||
CreateAccountDeposit(1) A: 2000
|
||||
CreateAccountDeposit(2) A: 2000
|
||||
CreateAccountDeposit(1) B: 1000
|
||||
CreateAccountDeposit(2) B: 1000
|
||||
CreateAccountDeposit(1) A: 20000
|
||||
CreateAccountDeposit(2) A: 20000
|
||||
CreateAccountDeposit(1) B: 10000
|
||||
CreateAccountDeposit(2) B: 10000
|
||||
> batchL1
|
||||
> batchL1
|
||||
> block
|
||||
@@ -79,15 +84,23 @@ func prepareHistoryDB(historyDB *historydb.HistoryDB) error {
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
for i := range blocks {
|
||||
block := &blocks[i]
|
||||
for j := range block.Rollup.AddedTokens {
|
||||
token := &block.Rollup.AddedTokens[j]
|
||||
token.Name = fmt.Sprintf("Token %d", token.TokenID)
|
||||
token.Symbol = fmt.Sprintf("TK%d", token.TokenID)
|
||||
token.Decimals = decimals
|
||||
}
|
||||
}
|
||||
|
||||
tokens = make(map[common.TokenID]historydb.TokenWithUSD)
|
||||
tokensValue = make(map[common.TokenID]float64)
|
||||
// tokensValue = make(map[common.TokenID]float64)
|
||||
accs = make(map[common.Idx]common.Account)
|
||||
value := 5 * 5.389329
|
||||
now := time.Now().UTC()
|
||||
// Add all blocks except for the last one
|
||||
for i := range blocks[:len(blocks)-1] {
|
||||
err = historyDB.AddBlockSCData(&blocks[i])
|
||||
if err != nil {
|
||||
if err := historyDB.AddBlockSCData(&blocks[i]); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
for _, batch := range blocks[i].Rollup.Batches {
|
||||
@@ -103,39 +116,38 @@ func prepareHistoryDB(historyDB *historydb.HistoryDB) error {
|
||||
Name: token.Name,
|
||||
Symbol: token.Symbol,
|
||||
Decimals: token.Decimals,
|
||||
USD: &tokenValue,
|
||||
USDUpdate: &now,
|
||||
}
|
||||
tokensValue[token.TokenID] = value / math.Pow(10, float64(token.Decimals))
|
||||
readToken.USDUpdate = &now
|
||||
readToken.USD = &value
|
||||
tokens[token.TokenID] = readToken
|
||||
}
|
||||
// Set value to the tokens (tokens have no symbol)
|
||||
tokenSymbol := ""
|
||||
err := historyDB.UpdateTokenValue(tokenSymbol, value)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
// Set value to the tokens
|
||||
err := historyDB.UpdateTokenValue(readToken.Symbol, *readToken.USD)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
func generatePoolL2Txs() ([]common.PoolL2Tx, error) {
|
||||
// Fee = 126 corresponds to ~10%
|
||||
setPool := `
|
||||
Type: PoolL2
|
||||
PoolTransfer(1) A-B: 6 (4)
|
||||
PoolTransfer(2) A-B: 3 (1)
|
||||
PoolTransfer(1) B-A: 5 (2)
|
||||
PoolTransfer(2) B-A: 10 (3)
|
||||
PoolTransfer(1) A-B: 7 (2)
|
||||
PoolTransfer(2) A-B: 2 (1)
|
||||
PoolTransfer(1) B-A: 8 (2)
|
||||
PoolTransfer(2) B-A: 1 (1)
|
||||
PoolTransfer(1) A-B: 3 (1)
|
||||
PoolTransferToEthAddr(2) B-A: 5 (2)
|
||||
PoolTransferToBJJ(2) B-A: 5 (2)
|
||||
PoolTransfer(1) A-B: 6000 (126)
|
||||
PoolTransfer(2) A-B: 3000 (126)
|
||||
PoolTransfer(1) B-A: 5000 (126)
|
||||
PoolTransfer(2) B-A: 10000 (126)
|
||||
PoolTransfer(1) A-B: 7000 (126)
|
||||
PoolTransfer(2) A-B: 2000 (126)
|
||||
PoolTransfer(1) B-A: 8000 (126)
|
||||
PoolTransfer(2) B-A: 1000 (126)
|
||||
PoolTransfer(1) A-B: 3000 (126)
|
||||
PoolTransferToEthAddr(2) B-A: 5000 (126)
|
||||
PoolTransferToBJJ(2) B-A: 5000 (126)
|
||||
|
||||
PoolExit(1) A: 5 (2)
|
||||
PoolExit(2) B: 3 (1)
|
||||
PoolExit(1) A: 5000 (126)
|
||||
PoolExit(2) B: 3000 (126)
|
||||
`
|
||||
poolL2Txs, err := tc.GeneratePoolL2Txs(setPool)
|
||||
if err != nil {
|
||||
@@ -150,25 +162,74 @@ func TestAddTxTest(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
fetchedTx, err := l2DB.GetTx(poolL2Txs[i].TxID)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
assertTx(t, &poolL2Txs[i], fetchedTx)
|
||||
nameZone, offset := fetchedTx.Timestamp.Zone()
|
||||
assert.Equal(t, "UTC", nameZone)
|
||||
assert.Equal(t, 0, offset)
|
||||
}
|
||||
}
|
||||
|
||||
func TestAddTxAPI(t *testing.T) {
|
||||
err := prepareHistoryDB(historyDB)
|
||||
if err != nil {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
|
||||
oldMaxTxs := l2DBWithACC.maxTxs
|
||||
// set max number of pending txs that can be kept in the pool to 5
|
||||
l2DBWithACC.maxTxs = 5
|
||||
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
txs := make([]*PoolL2TxWrite, len(poolL2Txs))
|
||||
for i := range poolL2Txs {
|
||||
txs[i] = NewPoolL2TxWriteFromPoolL2Tx(&poolL2Txs[i])
|
||||
}
|
||||
require.NoError(t, err)
|
||||
require.GreaterOrEqual(t, len(poolL2Txs), 8)
|
||||
for i := range txs[:5] {
|
||||
err := l2DBWithACC.AddTxAPI(txs[i])
|
||||
require.NoError(t, err)
|
||||
fetchedTx, err := l2DB.GetTx(poolL2Txs[i].TxID)
|
||||
require.NoError(t, err)
|
||||
assertTx(t, &poolL2Txs[i], fetchedTx)
|
||||
nameZone, offset := fetchedTx.Timestamp.Zone()
|
||||
assert.Equal(t, "UTC", nameZone)
|
||||
assert.Equal(t, 0, offset)
|
||||
}
|
||||
err = l2DBWithACC.AddTxAPI(txs[5])
|
||||
assert.Equal(t, errPoolFull, tracerr.Unwrap(err))
|
||||
// reset maxTxs to original value
|
||||
l2DBWithACC.maxTxs = oldMaxTxs
|
||||
|
||||
// set minFeeUSD to a high value than the tx feeUSD to test the error
|
||||
// of inserting a tx with lower than min fee
|
||||
oldMinFeeUSD := l2DBWithACC.minFeeUSD
|
||||
tx := txs[5]
|
||||
feeAmount, err := common.CalcFeeAmount(tx.Amount, tx.Fee)
|
||||
require.NoError(t, err)
|
||||
feeAmountUSD := common.TokensToUSD(feeAmount, decimals, tokenValue)
|
||||
// set minFeeUSD higher than the tx fee to trigger the error
|
||||
l2DBWithACC.minFeeUSD = feeAmountUSD + 1
|
||||
err = l2DBWithACC.AddTxAPI(tx)
|
||||
require.Error(t, err)
|
||||
assert.Regexp(t, "tx.feeUSD (.*) < minFeeUSD (.*)", err.Error())
|
||||
// reset minFeeUSD to original value
|
||||
l2DBWithACC.minFeeUSD = oldMinFeeUSD
|
||||
}
|
||||
|
||||
func TestUpdateTxsInfo(t *testing.T) {
|
||||
err := prepareHistoryDB(historyDB)
|
||||
if err != nil {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
require.NoError(t, err)
|
||||
@@ -182,7 +243,7 @@ func TestUpdateTxsInfo(t *testing.T) {
|
||||
|
||||
for i := range poolL2Txs {
|
||||
fetchedTx, err := l2DB.GetTx(poolL2Txs[i].TxID)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, "test", fetchedTx.Info)
|
||||
}
|
||||
}
|
||||
@@ -200,9 +261,8 @@ func assertTx(t *testing.T, expected, actual *common.PoolL2Tx) {
|
||||
assert.Less(t, token.USDUpdate.Unix()-3, actual.AbsoluteFeeUpdate.Unix())
|
||||
expected.AbsoluteFeeUpdate = actual.AbsoluteFeeUpdate
|
||||
// Set expected fee
|
||||
f := new(big.Float).SetInt(expected.Amount)
|
||||
amountF, _ := f.Float64()
|
||||
expected.AbsoluteFee = *token.USD * amountF * expected.Fee.Percentage()
|
||||
amountUSD := common.TokensToUSD(expected.Amount, token.Decimals, *token.USD)
|
||||
expected.AbsoluteFee = amountUSD * expected.Fee.Percentage()
|
||||
test.AssertUSD(t, &expected.AbsoluteFee, &actual.AbsoluteFee)
|
||||
}
|
||||
assert.Equal(t, expected, actual)
|
||||
@@ -227,19 +287,28 @@ func TestGetPending(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
var pendingTxs []*common.PoolL2Tx
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
pendingTxs = append(pendingTxs, &poolL2Txs[i])
|
||||
}
|
||||
fetchedTxs, err := l2DB.GetPendingTxs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, len(pendingTxs), len(fetchedTxs))
|
||||
for i := range fetchedTxs {
|
||||
assertTx(t, pendingTxs[i], &fetchedTxs[i])
|
||||
}
|
||||
// Check AbsoluteFee amount
|
||||
for i := range fetchedTxs {
|
||||
tx := &fetchedTxs[i]
|
||||
feeAmount, err := common.CalcFeeAmount(tx.Amount, tx.Fee)
|
||||
require.NoError(t, err)
|
||||
feeAmountUSD := common.TokensToUSD(feeAmount,
|
||||
tokens[tx.TokenID].Decimals, *tokens[tx.TokenID].USD)
|
||||
assert.InEpsilon(t, feeAmountUSD, tx.AbsoluteFee, 0.01)
|
||||
}
|
||||
}
|
||||
|
||||
func TestStartForging(t *testing.T) {
|
||||
@@ -250,13 +319,13 @@ func TestStartForging(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
var startForgingTxIDs []common.TxID
|
||||
randomizer := 0
|
||||
// Add txs to DB
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
if poolL2Txs[i].State == common.PoolL2TxStatePending && randomizer%2 == 0 {
|
||||
startForgingTxIDs = append(startForgingTxIDs, poolL2Txs[i].TxID)
|
||||
}
|
||||
@@ -264,11 +333,11 @@ func TestStartForging(t *testing.T) {
|
||||
}
|
||||
// Start forging txs
|
||||
err = l2DB.StartForging(startForgingTxIDs, fakeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Fetch txs and check that they've been updated correctly
|
||||
for _, id := range startForgingTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, common.PoolL2TxStateForging, fetchedTx.State)
|
||||
assert.Equal(t, &fakeBatchNum, fetchedTx.BatchNum)
|
||||
}
|
||||
@@ -282,13 +351,13 @@ func TestDoneForging(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
var startForgingTxIDs []common.TxID
|
||||
randomizer := 0
|
||||
// Add txs to DB
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
if poolL2Txs[i].State == common.PoolL2TxStatePending && randomizer%2 == 0 {
|
||||
startForgingTxIDs = append(startForgingTxIDs, poolL2Txs[i].TxID)
|
||||
}
|
||||
@@ -296,7 +365,7 @@ func TestDoneForging(t *testing.T) {
|
||||
}
|
||||
// Start forging txs
|
||||
err = l2DB.StartForging(startForgingTxIDs, fakeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
var doneForgingTxIDs []common.TxID
|
||||
randomizer = 0
|
||||
@@ -308,12 +377,12 @@ func TestDoneForging(t *testing.T) {
|
||||
}
|
||||
// Done forging txs
|
||||
err = l2DB.DoneForging(doneForgingTxIDs, fakeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Fetch txs and check that they've been updated correctly
|
||||
for _, id := range doneForgingTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, common.PoolL2TxStateForged, fetchedTx.State)
|
||||
assert.Equal(t, &fakeBatchNum, fetchedTx.BatchNum)
|
||||
}
|
||||
@@ -327,13 +396,13 @@ func TestInvalidate(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
var invalidTxIDs []common.TxID
|
||||
randomizer := 0
|
||||
// Add txs to DB
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
if poolL2Txs[i].State != common.PoolL2TxStateInvalid && randomizer%2 == 0 {
|
||||
randomizer++
|
||||
invalidTxIDs = append(invalidTxIDs, poolL2Txs[i].TxID)
|
||||
@@ -341,11 +410,11 @@ func TestInvalidate(t *testing.T) {
|
||||
}
|
||||
// Invalidate txs
|
||||
err = l2DB.InvalidateTxs(invalidTxIDs, fakeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Fetch txs and check that they've been updated correctly
|
||||
for _, id := range invalidTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, common.PoolL2TxStateInvalid, fetchedTx.State)
|
||||
assert.Equal(t, &fakeBatchNum, fetchedTx.BatchNum)
|
||||
}
|
||||
@@ -359,7 +428,7 @@ func TestInvalidateOldNonces(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Update Accounts currentNonce
|
||||
var updateAccounts []common.IdxNonce
|
||||
var currentNonce = common.Nonce(1)
|
||||
@@ -376,16 +445,16 @@ func TestInvalidateOldNonces(t *testing.T) {
|
||||
invalidTxIDs = append(invalidTxIDs, poolL2Txs[i].TxID)
|
||||
}
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
// sanity check
|
||||
require.Greater(t, len(invalidTxIDs), 0)
|
||||
|
||||
err = l2DB.InvalidateOldNonces(updateAccounts, fakeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Fetch txs and check that they've been updated correctly
|
||||
for _, id := range invalidTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, common.PoolL2TxStateInvalid, fetchedTx.State)
|
||||
assert.Equal(t, &fakeBatchNum, fetchedTx.BatchNum)
|
||||
@@ -404,7 +473,7 @@ func TestReorg(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
reorgedTxIDs := []common.TxID{}
|
||||
nonReorgedTxIDs := []common.TxID{}
|
||||
@@ -415,7 +484,7 @@ func TestReorg(t *testing.T) {
|
||||
// Add txs to DB
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
if poolL2Txs[i].State == common.PoolL2TxStatePending && randomizer%2 == 0 {
|
||||
startForgingTxIDs = append(startForgingTxIDs, poolL2Txs[i].TxID)
|
||||
allTxRandomize = append(allTxRandomize, poolL2Txs[i].TxID)
|
||||
@@ -427,7 +496,7 @@ func TestReorg(t *testing.T) {
|
||||
}
|
||||
// Start forging txs
|
||||
err = l2DB.StartForging(startForgingTxIDs, lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
var doneForgingTxIDs []common.TxID
|
||||
randomizer = 0
|
||||
@@ -452,22 +521,22 @@ func TestReorg(t *testing.T) {
|
||||
|
||||
// Invalidate txs BEFORE reorgBatch --> nonReorg
|
||||
err = l2DB.InvalidateTxs(invalidTxIDs, lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Done forging txs in reorgBatch --> Reorg
|
||||
err = l2DB.DoneForging(doneForgingTxIDs, reorgBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
err = l2DB.Reorg(lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
for _, id := range reorgedTxIDs {
|
||||
tx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
tx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Nil(t, tx.BatchNum)
|
||||
assert.Equal(t, common.PoolL2TxStatePending, tx.State)
|
||||
}
|
||||
for _, id := range nonReorgedTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, lastValidBatch, *fetchedTx.BatchNum)
|
||||
}
|
||||
}
|
||||
@@ -484,7 +553,7 @@ func TestReorg2(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
reorgedTxIDs := []common.TxID{}
|
||||
nonReorgedTxIDs := []common.TxID{}
|
||||
@@ -495,7 +564,7 @@ func TestReorg2(t *testing.T) {
|
||||
// Add txs to DB
|
||||
for i := range poolL2Txs {
|
||||
err := l2DB.AddTxTest(&poolL2Txs[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
if poolL2Txs[i].State == common.PoolL2TxStatePending && randomizer%2 == 0 {
|
||||
startForgingTxIDs = append(startForgingTxIDs, poolL2Txs[i].TxID)
|
||||
allTxRandomize = append(allTxRandomize, poolL2Txs[i].TxID)
|
||||
@@ -507,7 +576,7 @@ func TestReorg2(t *testing.T) {
|
||||
}
|
||||
// Start forging txs
|
||||
err = l2DB.StartForging(startForgingTxIDs, lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
var doneForgingTxIDs []common.TxID
|
||||
randomizer = 0
|
||||
@@ -529,22 +598,22 @@ func TestReorg2(t *testing.T) {
|
||||
}
|
||||
// Done forging txs BEFORE reorgBatch --> nonReorg
|
||||
err = l2DB.DoneForging(doneForgingTxIDs, lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Invalidate txs in reorgBatch --> Reorg
|
||||
err = l2DB.InvalidateTxs(invalidTxIDs, reorgBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
err = l2DB.Reorg(lastValidBatch)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
for _, id := range reorgedTxIDs {
|
||||
tx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
tx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Nil(t, tx.BatchNum)
|
||||
assert.Equal(t, common.PoolL2TxStatePending, tx.State)
|
||||
}
|
||||
for _, id := range nonReorgedTxIDs {
|
||||
fetchedTx, err := l2DB.GetTxAPI(id)
|
||||
assert.NoError(t, err)
|
||||
fetchedTx, err := l2DBWithACC.GetTxAPI(id)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, lastValidBatch, *fetchedTx.BatchNum)
|
||||
}
|
||||
}
|
||||
@@ -560,7 +629,7 @@ func TestPurge(t *testing.T) {
|
||||
var poolL2Tx []common.PoolL2Tx
|
||||
for i := 0; i < generateTx; i++ {
|
||||
poolL2TxAux, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
poolL2Tx = append(poolL2Tx, poolL2TxAux...)
|
||||
}
|
||||
|
||||
@@ -587,39 +656,39 @@ func TestPurge(t *testing.T) {
|
||||
deletedIDs = append(deletedIDs, poolL2Tx[i].TxID)
|
||||
}
|
||||
err := l2DB.AddTxTest(&tx)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
// Set batchNum keeped txs
|
||||
for i := range keepedIDs {
|
||||
_, err = l2DB.db.Exec(
|
||||
_, err = l2DB.dbWrite.Exec(
|
||||
"UPDATE tx_pool SET batch_num = $1 WHERE tx_id = $2;",
|
||||
safeBatchNum, keepedIDs[i],
|
||||
)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
// Start forging txs and set batchNum
|
||||
err = l2DB.StartForging(doneForgingTxIDs, toDeleteBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Done forging txs and set batchNum
|
||||
err = l2DB.DoneForging(doneForgingTxIDs, toDeleteBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Invalidate txs and set batchNum
|
||||
err = l2DB.InvalidateTxs(invalidTxIDs, toDeleteBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Update timestamp of afterTTL txs
|
||||
deleteTimestamp := time.Unix(time.Now().UTC().Unix()-int64(l2DB.ttl.Seconds()+float64(4*time.Second)), 0)
|
||||
for _, id := range afterTTLIDs {
|
||||
// Set timestamp
|
||||
_, err = l2DB.db.Exec(
|
||||
_, err = l2DB.dbWrite.Exec(
|
||||
"UPDATE tx_pool SET timestamp = $1, state = $2 WHERE tx_id = $3;",
|
||||
deleteTimestamp, common.PoolL2TxStatePending, id,
|
||||
)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
// Purge txs
|
||||
err = l2DB.Purge(safeBatchNum)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Check results
|
||||
for _, id := range deletedIDs {
|
||||
_, err := l2DB.GetTx(id)
|
||||
@@ -627,7 +696,7 @@ func TestPurge(t *testing.T) {
|
||||
}
|
||||
for _, id := range keepedIDs {
|
||||
_, err := l2DB.GetTx(id)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -641,10 +710,47 @@ func TestAuth(t *testing.T) {
|
||||
for i := 0; i < len(auths); i++ {
|
||||
// Add to the DB
|
||||
err := l2DB.AddAccountCreationAuth(auths[i])
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Fetch from DB
|
||||
auth, err := l2DB.GetAccountCreationAuth(auths[i].EthAddr)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
// Check fetched vs generated
|
||||
assert.Equal(t, auths[i].EthAddr, auth.EthAddr)
|
||||
assert.Equal(t, auths[i].BJJ, auth.BJJ)
|
||||
assert.Equal(t, auths[i].Signature, auth.Signature)
|
||||
assert.Equal(t, auths[i].Timestamp.Unix(), auths[i].Timestamp.Unix())
|
||||
nameZone, offset := auths[i].Timestamp.Zone()
|
||||
assert.Equal(t, "UTC", nameZone)
|
||||
assert.Equal(t, 0, offset)
|
||||
}
|
||||
}
|
||||
|
||||
func TestManyAuth(t *testing.T) {
|
||||
test.WipeDB(l2DB.DB())
|
||||
const nAuths = 5
|
||||
chainID := uint16(0)
|
||||
hermezContractAddr := ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
||||
// Generate authorizations
|
||||
genAuths := test.GenAuths(nAuths, chainID, hermezContractAddr)
|
||||
auths := make([]common.AccountCreationAuth, len(genAuths))
|
||||
// Convert to a non-pointer slice
|
||||
for i := 0; i < len(genAuths); i++ {
|
||||
auths[i] = *genAuths[i]
|
||||
}
|
||||
|
||||
// Add a duplicate one to check the not exist condition
|
||||
err := l2DB.AddAccountCreationAuth(genAuths[0])
|
||||
require.NoError(t, err)
|
||||
|
||||
// Add to the DB
|
||||
err = l2DB.AddManyAccountCreationAuth(auths)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Assert the result
|
||||
for i := 0; i < len(auths); i++ {
|
||||
// Fetch from DB
|
||||
auth, err := l2DB.GetAccountCreationAuth(auths[i].EthAddr)
|
||||
require.NoError(t, err)
|
||||
// Check fetched vs generated
|
||||
assert.Equal(t, auths[i].EthAddr, auth.EthAddr)
|
||||
assert.Equal(t, auths[i].BJJ, auth.BJJ)
|
||||
@@ -662,7 +768,7 @@ func TestAddGet(t *testing.T) {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
poolL2Txs, err := generatePoolL2Txs()
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
|
||||
// We will work with only 3 txs
|
||||
require.GreaterOrEqual(t, len(poolL2Txs), 3)
|
||||
@@ -698,3 +804,56 @@ func TestAddGet(t *testing.T) {
|
||||
assert.Equal(t, txs[i], *dbTx)
|
||||
}
|
||||
}
|
||||
|
||||
func TestPurgeByExternalDelete(t *testing.T) {
|
||||
err := prepareHistoryDB(historyDB)
|
||||
if err != nil {
|
||||
log.Error("Error prepare historyDB", err)
|
||||
}
|
||||
txs, err := generatePoolL2Txs()
|
||||
require.NoError(t, err)
|
||||
|
||||
// We will work with 8 txs
|
||||
require.GreaterOrEqual(t, len(txs), 8)
|
||||
txs = txs[:8]
|
||||
for i := range txs {
|
||||
require.NoError(t, l2DB.AddTxTest(&txs[i]))
|
||||
}
|
||||
|
||||
// We will recreate this scenario:
|
||||
// tx index, status , external_delete
|
||||
// 0 , pending, false
|
||||
// 1 , pending, false
|
||||
// 2 , pending, true // will be deleted
|
||||
// 3 , pending, true // will be deleted
|
||||
// 4 , fging , false
|
||||
// 5 , fging , false
|
||||
// 6 , fging , true
|
||||
// 7 , fging , true
|
||||
|
||||
require.NoError(t, l2DB.StartForging(
|
||||
[]common.TxID{txs[4].TxID, txs[5].TxID, txs[6].TxID, txs[7].TxID},
|
||||
1))
|
||||
_, err = l2DB.dbWrite.Exec(
|
||||
`UPDATE tx_pool SET external_delete = true WHERE
|
||||
tx_id IN ($1, $2, $3, $4)
|
||||
;`,
|
||||
txs[2].TxID, txs[3].TxID, txs[6].TxID, txs[7].TxID,
|
||||
)
|
||||
require.NoError(t, err)
|
||||
require.NoError(t, l2DB.PurgeByExternalDelete())
|
||||
|
||||
// Query txs that are have been not deleted
|
||||
for _, i := range []int{0, 1, 4, 5, 6, 7} {
|
||||
txID := txs[i].TxID
|
||||
_, err := l2DB.GetTx(txID)
|
||||
require.NoError(t, err)
|
||||
}
|
||||
|
||||
// Query txs that have been deleted
|
||||
for _, i := range []int{2, 3} {
|
||||
txID := txs[i].TxID
|
||||
_, err := l2DB.GetTx(txID)
|
||||
require.Equal(t, sql.ErrNoRows, tracerr.Unwrap(err))
|
||||
}
|
||||
}
|
||||
|
||||
@@ -34,6 +34,7 @@ type PoolL2TxWrite struct {
|
||||
RqFee *common.FeeSelector `meddler:"rq_fee"`
|
||||
RqNonce *common.Nonce `meddler:"rq_nonce"`
|
||||
Type common.TxType `meddler:"tx_type"`
|
||||
ClientIP string `meddler:"client_ip"`
|
||||
}
|
||||
|
||||
// PoolTxAPI represents a L2 Tx pool with extra metadata used by the API
|
||||
@@ -94,7 +95,6 @@ func (tx PoolTxAPI) MarshalJSON() ([]byte, error) {
|
||||
"info": tx.Info,
|
||||
"signature": tx.Signature,
|
||||
"timestamp": tx.Timestamp,
|
||||
"batchNum": tx.BatchNum,
|
||||
"requestFromAccountIndex": tx.RqFromIdx,
|
||||
"requestToAccountIndex": tx.RqToIdx,
|
||||
"requestToHezEthereumAddress": tx.RqToEthAddr,
|
||||
|
||||
@@ -47,7 +47,7 @@ CREATE TABLE token (
|
||||
name VARCHAR(20) NOT NULL,
|
||||
symbol VARCHAR(10) NOT NULL,
|
||||
decimals INT NOT NULL,
|
||||
usd NUMERIC,
|
||||
usd NUMERIC, -- value of a normalized token (1 token = 10^decimals units)
|
||||
usd_update TIMESTAMP WITHOUT TIME ZONE
|
||||
);
|
||||
|
||||
@@ -100,6 +100,15 @@ CREATE TABLE account (
|
||||
eth_addr BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE account_update (
|
||||
item_id SERIAL,
|
||||
eth_block_num BIGINT NOT NULL REFERENCES block (eth_block_num) ON DELETE CASCADE,
|
||||
batch_num BIGINT NOT NULL REFERENCES batch (batch_num) ON DELETE CASCADE,
|
||||
idx BIGINT NOT NULL REFERENCES account (idx) ON DELETE CASCADE,
|
||||
nonce BIGINT NOT NULL,
|
||||
balance BYTEA NOT NULL
|
||||
);
|
||||
|
||||
CREATE TABLE exit_tree (
|
||||
item_id SERIAL PRIMARY KEY,
|
||||
batch_num BIGINT REFERENCES batch (batch_num) ON DELETE CASCADE,
|
||||
@@ -618,7 +627,9 @@ CREATE TABLE tx_pool (
|
||||
rq_amount BYTEA,
|
||||
rq_fee SMALLINT,
|
||||
rq_nonce BIGINT,
|
||||
tx_type VARCHAR(40) NOT NULL
|
||||
tx_type VARCHAR(40) NOT NULL,
|
||||
client_ip VARCHAR,
|
||||
external_delete BOOLEAN NOT NULL DEFAULT false
|
||||
);
|
||||
|
||||
-- +migrate StatementBegin
|
||||
@@ -651,34 +662,35 @@ CREATE TABLE account_creation_auth (
|
||||
);
|
||||
|
||||
-- +migrate Down
|
||||
-- drop triggers
|
||||
DROP TRIGGER trigger_token_usd_update ON token;
|
||||
DROP TRIGGER trigger_set_tx ON tx;
|
||||
DROP TRIGGER trigger_forge_l1_txs ON batch;
|
||||
DROP TRIGGER trigger_set_pool_tx ON tx_pool;
|
||||
-- drop functions
|
||||
DROP FUNCTION hez_idx;
|
||||
DROP FUNCTION set_token_usd_update;
|
||||
DROP FUNCTION fee_percentage;
|
||||
DROP FUNCTION set_tx;
|
||||
DROP FUNCTION forge_l1_user_txs;
|
||||
DROP FUNCTION set_pool_tx;
|
||||
-- drop tables
|
||||
DROP TABLE account_creation_auth;
|
||||
DROP TABLE tx_pool;
|
||||
DROP TABLE auction_vars;
|
||||
DROP TABLE rollup_vars;
|
||||
DROP TABLE escape_hatch_withdrawal;
|
||||
DROP TABLE bucket_update;
|
||||
DROP TABLE token_exchange;
|
||||
DROP TABLE wdelayer_vars;
|
||||
DROP TABLE tx;
|
||||
DROP TABLE exit_tree;
|
||||
DROP TABLE account;
|
||||
DROP TABLE token;
|
||||
DROP TABLE bid;
|
||||
DROP TABLE batch;
|
||||
DROP TABLE coordinator;
|
||||
DROP TABLE block;
|
||||
-- drop sequences
|
||||
DROP SEQUENCE tx_item_id;
|
||||
-- triggers
|
||||
DROP TRIGGER IF EXISTS trigger_token_usd_update ON token;
|
||||
DROP TRIGGER IF EXISTS trigger_set_tx ON tx;
|
||||
DROP TRIGGER IF EXISTS trigger_forge_l1_txs ON batch;
|
||||
DROP TRIGGER IF EXISTS trigger_set_pool_tx ON tx_pool;
|
||||
-- functions
|
||||
DROP FUNCTION IF EXISTS hez_idx;
|
||||
DROP FUNCTION IF EXISTS set_token_usd_update;
|
||||
DROP FUNCTION IF EXISTS fee_percentage;
|
||||
DROP FUNCTION IF EXISTS set_tx;
|
||||
DROP FUNCTION IF EXISTS forge_l1_user_txs;
|
||||
DROP FUNCTION IF EXISTS set_pool_tx;
|
||||
-- drop tables IF EXISTS
|
||||
DROP TABLE IF EXISTS account_creation_auth;
|
||||
DROP TABLE IF EXISTS tx_pool;
|
||||
DROP TABLE IF EXISTS auction_vars;
|
||||
DROP TABLE IF EXISTS rollup_vars;
|
||||
DROP TABLE IF EXISTS escape_hatch_withdrawal;
|
||||
DROP TABLE IF EXISTS bucket_update;
|
||||
DROP TABLE IF EXISTS token_exchange;
|
||||
DROP TABLE IF EXISTS wdelayer_vars;
|
||||
DROP TABLE IF EXISTS tx;
|
||||
DROP TABLE IF EXISTS exit_tree;
|
||||
DROP TABLE IF EXISTS account_update;
|
||||
DROP TABLE IF EXISTS account;
|
||||
DROP TABLE IF EXISTS token;
|
||||
DROP TABLE IF EXISTS bid;
|
||||
DROP TABLE IF EXISTS batch;
|
||||
DROP TABLE IF EXISTS coordinator;
|
||||
DROP TABLE IF EXISTS block;
|
||||
-- sequences
|
||||
DROP SEQUENCE IF EXISTS tx_item_id;
|
||||
|
||||
@@ -17,7 +17,8 @@ import (
|
||||
var (
|
||||
// ErrStateDBWithoutMT is used when a method that requires a MerkleTree
|
||||
// is called in a StateDB that does not have a MerkleTree defined
|
||||
ErrStateDBWithoutMT = errors.New("Can not call method to use MerkleTree in a StateDB without MerkleTree")
|
||||
ErrStateDBWithoutMT = errors.New(
|
||||
"Can not call method to use MerkleTree in a StateDB without MerkleTree")
|
||||
|
||||
// ErrAccountAlreadyExists is used when CreateAccount is called and the
|
||||
// Account already exists
|
||||
@@ -28,7 +29,8 @@ var (
|
||||
ErrIdxNotFound = errors.New("Idx can not be found")
|
||||
// ErrGetIdxNoCase is used when trying to get the Idx from EthAddr &
|
||||
// BJJ with not compatible combination
|
||||
ErrGetIdxNoCase = errors.New("Can not get Idx due unexpected combination of ethereum Address & BabyJubJub PublicKey")
|
||||
ErrGetIdxNoCase = errors.New(
|
||||
"Can not get Idx due unexpected combination of ethereum Address & BabyJubJub PublicKey")
|
||||
|
||||
// PrefixKeyIdx is the key prefix for idx in the db
|
||||
PrefixKeyIdx = []byte("i:")
|
||||
@@ -52,19 +54,40 @@ const (
|
||||
// TypeBatchBuilder defines a StateDB used by the BatchBuilder, that
|
||||
// generates the ExitTree and the ZKInput when processing the txs
|
||||
TypeBatchBuilder = "batchbuilder"
|
||||
// MaxNLevels is the maximum value of NLevels for the merkle tree,
|
||||
// which comes from the fact that AccountIdx has 48 bits.
|
||||
MaxNLevels = 48
|
||||
)
|
||||
|
||||
// TypeStateDB determines the type of StateDB
|
||||
type TypeStateDB string
|
||||
|
||||
// Config of the StateDB
|
||||
type Config struct {
|
||||
// Path where the checkpoints will be stored
|
||||
Path string
|
||||
// Keep is the number of old checkpoints to keep. If 0, all
|
||||
// checkpoints are kept.
|
||||
Keep int
|
||||
// NoLast skips having an opened DB with a checkpoint to the last
|
||||
// batchNum for thread-safe reads.
|
||||
NoLast bool
|
||||
// Type of StateDB (
|
||||
Type TypeStateDB
|
||||
// NLevels is the number of merkle tree levels in case the Type uses a
|
||||
// merkle tree. If the Type doesn't use a merkle tree, NLevels should
|
||||
// be 0.
|
||||
NLevels int
|
||||
// At every checkpoint, check that there are no gaps between the
|
||||
// checkpoints
|
||||
noGapsCheck bool
|
||||
}
|
||||
|
||||
// StateDB represents the StateDB object
|
||||
type StateDB struct {
|
||||
path string
|
||||
Typ TypeStateDB
|
||||
db *kvdb.KVDB
|
||||
nLevels int
|
||||
MT *merkletree.MerkleTree
|
||||
keep int
|
||||
cfg Config
|
||||
db *kvdb.KVDB
|
||||
MT *merkletree.MerkleTree
|
||||
}
|
||||
|
||||
// Last offers a subset of view methods of the StateDB that can be
|
||||
@@ -104,36 +127,41 @@ func (s *Last) GetAccounts() ([]common.Account, error) {
|
||||
// NewStateDB creates a new StateDB, allowing to use an in-memory or in-disk
|
||||
// storage. Checkpoints older than the value defined by `keep` will be
|
||||
// deleted.
|
||||
func NewStateDB(pathDB string, keep int, typ TypeStateDB, nLevels int) (*StateDB, error) {
|
||||
// func NewStateDB(pathDB string, keep int, typ TypeStateDB, nLevels int) (*StateDB, error) {
|
||||
func NewStateDB(cfg Config) (*StateDB, error) {
|
||||
var kv *kvdb.KVDB
|
||||
var err error
|
||||
|
||||
kv, err = kvdb.NewKVDB(pathDB, keep)
|
||||
kv, err = kvdb.NewKVDB(kvdb.Config{Path: cfg.Path, Keep: cfg.Keep,
|
||||
NoGapsCheck: cfg.noGapsCheck, NoLast: cfg.NoLast})
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
var mt *merkletree.MerkleTree = nil
|
||||
if typ == TypeSynchronizer || typ == TypeBatchBuilder {
|
||||
mt, err = merkletree.NewMerkleTree(kv.StorageWithPrefix(PrefixKeyMT), nLevels)
|
||||
if cfg.Type == TypeSynchronizer || cfg.Type == TypeBatchBuilder {
|
||||
mt, err = merkletree.NewMerkleTree(kv.StorageWithPrefix(PrefixKeyMT), cfg.NLevels)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
if typ == TypeTxSelector && nLevels != 0 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid StateDB parameters: StateDB type==TypeStateDB can not have nLevels!=0"))
|
||||
if cfg.Type == TypeTxSelector && cfg.NLevels != 0 {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("invalid StateDB parameters: StateDB type==TypeStateDB can not have nLevels!=0"))
|
||||
}
|
||||
|
||||
return &StateDB{
|
||||
path: pathDB,
|
||||
db: kv,
|
||||
nLevels: nLevels,
|
||||
MT: mt,
|
||||
Typ: typ,
|
||||
keep: keep,
|
||||
cfg: cfg,
|
||||
db: kv,
|
||||
MT: mt,
|
||||
}, nil
|
||||
}
|
||||
|
||||
// Type returns the StateDB configured Type
|
||||
func (s *StateDB) Type() TypeStateDB {
|
||||
return s.cfg.Type
|
||||
}
|
||||
|
||||
// LastRead is a thread-safe method to query the last checkpoint of the StateDB
|
||||
// via the Last type methods
|
||||
func (s *StateDB) LastRead(fn func(sdbLast *Last) error) error {
|
||||
@@ -179,7 +207,7 @@ func (s *StateDB) LastGetCurrentBatch() (common.BatchNum, error) {
|
||||
func (s *StateDB) LastMTGetRoot() (*big.Int, error) {
|
||||
var root *big.Int
|
||||
if err := s.LastRead(func(sdb *Last) error {
|
||||
mt, err := merkletree.NewMerkleTree(sdb.DB().WithPrefix(PrefixKeyMT), s.nLevels)
|
||||
mt, err := merkletree.NewMerkleTree(sdb.DB().WithPrefix(PrefixKeyMT), s.cfg.NLevels)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -195,7 +223,7 @@ func (s *StateDB) LastMTGetRoot() (*big.Int, error) {
|
||||
// Internally this advances & stores the current BatchNum, and then stores a
|
||||
// Checkpoint of the current state of the StateDB.
|
||||
func (s *StateDB) MakeCheckpoint() error {
|
||||
log.Debugw("Making StateDB checkpoint", "batch", s.CurrentBatch()+1, "type", s.Typ)
|
||||
log.Debugw("Making StateDB checkpoint", "batch", s.CurrentBatch()+1, "type", s.cfg.Type)
|
||||
return s.db.MakeCheckpoint()
|
||||
}
|
||||
|
||||
@@ -230,8 +258,8 @@ func (s *StateDB) SetCurrentIdx(idx common.Idx) error {
|
||||
// those checkpoints will remain in the storage, and eventually will be
|
||||
// deleted when MakeCheckpoint overwrites them.
|
||||
func (s *StateDB) Reset(batchNum common.BatchNum) error {
|
||||
err := s.db.Reset(batchNum)
|
||||
if err != nil {
|
||||
log.Debugw("Making StateDB Reset", "batch", batchNum, "type", s.cfg.Type)
|
||||
if err := s.db.Reset(batchNum); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if s.MT != nil {
|
||||
@@ -242,7 +270,6 @@ func (s *StateDB) Reset(batchNum common.BatchNum) error {
|
||||
}
|
||||
s.MT = mt
|
||||
}
|
||||
log.Debugw("Making StateDB Reset", "batch", batchNum)
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -323,7 +350,8 @@ func GetAccountInTreeDB(sto db.Storage, idx common.Idx) (*common.Account, error)
|
||||
// CreateAccount creates a new Account in the StateDB for the given Idx. If
|
||||
// StateDB.MT==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (
|
||||
*merkletree.CircomProcessorProof, error) {
|
||||
cpp, err := CreateAccountInTreeDB(s.db.DB(), s.MT, idx, account)
|
||||
if err != nil {
|
||||
return cpp, tracerr.Wrap(err)
|
||||
@@ -337,7 +365,8 @@ func (s *StateDB) CreateAccount(idx common.Idx, account *common.Account) (*merkl
|
||||
// from ExitTree. Creates a new Account in the StateDB for the given Idx. If
|
||||
// StateDB.MT==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx,
|
||||
account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
// store at the DB the key: v, and value: leaf.Bytes()
|
||||
v, err := account.HashValue()
|
||||
if err != nil {
|
||||
@@ -386,7 +415,8 @@ func CreateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common
|
||||
// UpdateAccount updates the Account in the StateDB for the given Idx. If
|
||||
// StateDB.mt==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (
|
||||
*merkletree.CircomProcessorProof, error) {
|
||||
return UpdateAccountInTreeDB(s.db.DB(), s.MT, idx, account)
|
||||
}
|
||||
|
||||
@@ -394,7 +424,8 @@ func (s *StateDB) UpdateAccount(idx common.Idx, account *common.Account) (*merkl
|
||||
// from ExitTree. Updates the Account in the StateDB for the given Idx. If
|
||||
// StateDB.mt==nil, MerkleTree is not affected, otherwise updates the
|
||||
// MerkleTree, returning a CircomProcessorProof.
|
||||
func UpdateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx, account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
func UpdateAccountInTreeDB(sto db.Storage, mt *merkletree.MerkleTree, idx common.Idx,
|
||||
account *common.Account) (*merkletree.CircomProcessorProof, error) {
|
||||
// store at the DB the key: v, and value: account.Bytes()
|
||||
v, err := account.HashValue()
|
||||
if err != nil {
|
||||
@@ -461,9 +492,10 @@ type LocalStateDB struct {
|
||||
// NewLocalStateDB returns a new LocalStateDB connected to the given
|
||||
// synchronizerDB. Checkpoints older than the value defined by `keep` will be
|
||||
// deleted.
|
||||
func NewLocalStateDB(path string, keep int, synchronizerDB *StateDB, typ TypeStateDB,
|
||||
nLevels int) (*LocalStateDB, error) {
|
||||
s, err := NewStateDB(path, keep, typ, nLevels)
|
||||
func NewLocalStateDB(cfg Config, synchronizerDB *StateDB) (*LocalStateDB, error) {
|
||||
cfg.noGapsCheck = true
|
||||
cfg.NoLast = true
|
||||
s, err := NewStateDB(cfg)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -473,18 +505,24 @@ func NewLocalStateDB(path string, keep int, synchronizerDB *StateDB, typ TypeSta
|
||||
}, nil
|
||||
}
|
||||
|
||||
// Reset performs a reset in the LocaStateDB. If fromSynchronizer is true, it
|
||||
// CheckpointExists returns true if the checkpoint exists
|
||||
func (l *LocalStateDB) CheckpointExists(batchNum common.BatchNum) (bool, error) {
|
||||
return l.db.CheckpointExists(batchNum)
|
||||
}
|
||||
|
||||
// Reset performs a reset in the LocalStateDB. If fromSynchronizer is true, it
|
||||
// gets the state from LocalStateDB.synchronizerStateDB for the given batchNum.
|
||||
// If fromSynchronizer is false, get the state from LocalStateDB checkpoints.
|
||||
func (l *LocalStateDB) Reset(batchNum common.BatchNum, fromSynchronizer bool) error {
|
||||
if fromSynchronizer {
|
||||
err := l.db.ResetFromSynchronizer(batchNum, l.synchronizerStateDB.db)
|
||||
if err != nil {
|
||||
log.Debugw("Making StateDB ResetFromSynchronizer", "batch", batchNum, "type", l.cfg.Type)
|
||||
if err := l.db.ResetFromSynchronizer(batchNum, l.synchronizerStateDB.db); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
// open the MT for the current s.db
|
||||
if l.MT != nil {
|
||||
mt, err := merkletree.NewMerkleTree(l.db.StorageWithPrefix(PrefixKeyMT), l.MT.MaxLevels())
|
||||
mt, err := merkletree.NewMerkleTree(l.db.StorageWithPrefix(PrefixKeyMT),
|
||||
l.MT.MaxLevels())
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -22,7 +22,8 @@ import (
|
||||
|
||||
func newAccount(t *testing.T, i int) *common.Account {
|
||||
var sk babyjub.PrivateKey
|
||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
_, err := hex.Decode(sk[:],
|
||||
[]byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||
require.NoError(t, err)
|
||||
pk := sk.Public()
|
||||
|
||||
@@ -45,7 +46,7 @@ func TestNewStateDBIntermediateState(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
require.NoError(t, err)
|
||||
|
||||
// test values
|
||||
@@ -78,7 +79,7 @@ func TestNewStateDBIntermediateState(t *testing.T) {
|
||||
|
||||
// call NewStateDB which should get the db at the last checkpoint state
|
||||
// executing a Reset (discarding the last 'testkey0'&'testvalue0' data)
|
||||
sdb, err = NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err = NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
require.NoError(t, err)
|
||||
v, err = sdb.db.DB().Get(k0)
|
||||
assert.NotNil(t, err)
|
||||
@@ -116,7 +117,7 @@ func TestNewStateDBIntermediateState(t *testing.T) {
|
||||
bn, err := sdb.getCurrentBatch()
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, common.BatchNum(0), bn)
|
||||
err = sdb.db.MakeCheckpoint()
|
||||
err = sdb.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
bn, err = sdb.getCurrentBatch()
|
||||
require.NoError(t, err)
|
||||
@@ -158,7 +159,7 @@ func TestNewStateDBIntermediateState(t *testing.T) {
|
||||
|
||||
// call NewStateDB which should get the db at the last checkpoint state
|
||||
// executing a Reset (discarding the last 'testkey1'&'testvalue1' data)
|
||||
sdb, err = NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err = NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
require.NoError(t, err)
|
||||
|
||||
bn, err = sdb.getCurrentBatch()
|
||||
@@ -182,7 +183,7 @@ func TestStateDBWithoutMT(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
require.NoError(t, err)
|
||||
|
||||
// create test accounts
|
||||
@@ -236,7 +237,7 @@ func TestStateDBWithMT(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
// create test accounts
|
||||
@@ -290,7 +291,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
err = sdb.Reset(0)
|
||||
@@ -335,7 +336,7 @@ func TestCheckpoints(t *testing.T) {
|
||||
assert.Equal(t, common.BatchNum(i+1), cb)
|
||||
}
|
||||
|
||||
// printCheckpoints(t, sdb.path)
|
||||
// printCheckpoints(t, sdb.cfg.Path)
|
||||
|
||||
// reset checkpoint
|
||||
err = sdb.Reset(3)
|
||||
@@ -371,7 +372,8 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal, err := ioutil.TempDir("", "ldb")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal))
|
||||
ldb, err := NewLocalStateDB(dirLocal, 128, sdb, TypeBatchBuilder, 32)
|
||||
ldb, err := NewLocalStateDB(Config{Path: dirLocal, Keep: 128, Type: TypeBatchBuilder,
|
||||
NLevels: 32}, sdb)
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
@@ -392,28 +394,27 @@ func TestCheckpoints(t *testing.T) {
|
||||
dirLocal2, err := ioutil.TempDir("", "ldb2")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dirLocal2))
|
||||
ldb2, err := NewLocalStateDB(dirLocal2, 128, sdb, TypeBatchBuilder, 32)
|
||||
ldb2, err := NewLocalStateDB(Config{Path: dirLocal2, Keep: 128, Type: TypeBatchBuilder,
|
||||
NLevels: 32}, sdb)
|
||||
require.NoError(t, err)
|
||||
|
||||
// get checkpoint 4 from sdb (StateDB) to ldb (LocalStateDB)
|
||||
err = ldb2.Reset(4, true)
|
||||
require.NoError(t, err)
|
||||
// check that currentBatch is 4 after the Reset
|
||||
cb, err = ldb2.db.GetCurrentBatch()
|
||||
require.NoError(t, err)
|
||||
cb = ldb2.CurrentBatch()
|
||||
assert.Equal(t, common.BatchNum(4), cb)
|
||||
// advance one checkpoint in ldb2
|
||||
err = ldb2.db.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
cb, err = ldb2.db.GetCurrentBatch()
|
||||
err = ldb2.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
cb = ldb2.CurrentBatch()
|
||||
assert.Equal(t, common.BatchNum(5), cb)
|
||||
|
||||
debug := false
|
||||
if debug {
|
||||
printCheckpoints(t, sdb.path)
|
||||
printCheckpoints(t, ldb.path)
|
||||
printCheckpoints(t, ldb2.path)
|
||||
printCheckpoints(t, sdb.cfg.Path)
|
||||
printCheckpoints(t, ldb.cfg.Path)
|
||||
printCheckpoints(t, ldb2.cfg.Path)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -421,7 +422,7 @@ func TestStateDBGetAccounts(t *testing.T) {
|
||||
dir, err := ioutil.TempDir("", "tmpdb")
|
||||
require.NoError(t, err)
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
require.NoError(t, err)
|
||||
|
||||
// create test accounts
|
||||
@@ -468,12 +469,13 @@ func TestCheckAccountsTreeTestVectors(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
ay0 := new(big.Int).Sub(new(big.Int).Exp(big.NewInt(2), big.NewInt(253), nil), big.NewInt(1))
|
||||
// test value from js version (compatibility-canary)
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff", (hex.EncodeToString(ay0.Bytes())))
|
||||
assert.Equal(t, "1fffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffff",
|
||||
(hex.EncodeToString(ay0.Bytes())))
|
||||
bjjPoint0Comp := babyjub.PackSignY(true, ay0)
|
||||
bjj0 := babyjub.PublicKeyComp(bjjPoint0Comp)
|
||||
|
||||
@@ -532,7 +534,9 @@ func TestCheckAccountsTreeTestVectors(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
}
|
||||
// root value generated by js version:
|
||||
assert.Equal(t, "17298264051379321456969039521810887093935433569451713402227686942080129181291", sdb.MT.Root().BigInt().String())
|
||||
assert.Equal(t,
|
||||
"13174362770971232417413036794215823584762073355951212910715422236001731746065",
|
||||
sdb.MT.Root().BigInt().String())
|
||||
}
|
||||
|
||||
// TestListCheckpoints performs almost the same test than kvdb/kvdb_test.go
|
||||
@@ -542,7 +546,7 @@ func TestListCheckpoints(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
numCheckpoints := 16
|
||||
@@ -575,7 +579,7 @@ func TestDeleteOldCheckpoints(t *testing.T) {
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
keep := 16
|
||||
sdb, err := NewStateDB(dir, keep, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: keep, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
numCheckpoints := 32
|
||||
@@ -596,7 +600,7 @@ func TestCurrentIdx(t *testing.T) {
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
keep := 16
|
||||
sdb, err := NewStateDB(dir, keep, TypeSynchronizer, 32)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: keep, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx := sdb.CurrentIdx()
|
||||
@@ -604,7 +608,7 @@ func TestCurrentIdx(t *testing.T) {
|
||||
|
||||
sdb.Close()
|
||||
|
||||
sdb, err = NewStateDB(dir, keep, TypeSynchronizer, 32)
|
||||
sdb, err = NewStateDB(Config{Path: dir, Keep: keep, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx = sdb.CurrentIdx()
|
||||
@@ -618,9 +622,30 @@ func TestCurrentIdx(t *testing.T) {
|
||||
|
||||
sdb.Close()
|
||||
|
||||
sdb, err = NewStateDB(dir, keep, TypeSynchronizer, 32)
|
||||
sdb, err = NewStateDB(Config{Path: dir, Keep: keep, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
idx = sdb.CurrentIdx()
|
||||
assert.Equal(t, common.Idx(255), idx)
|
||||
}
|
||||
|
||||
func TestResetFromBadCheckpoint(t *testing.T) {
|
||||
dir, err := ioutil.TempDir("", "tmpdb")
|
||||
require.NoError(t, err)
|
||||
defer require.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
keep := 16
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: keep, Type: TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
err = sdb.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
err = sdb.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
err = sdb.MakeCheckpoint()
|
||||
require.NoError(t, err)
|
||||
|
||||
// reset from a checkpoint that doesn't exist
|
||||
err = sdb.Reset(10)
|
||||
require.Error(t, err)
|
||||
}
|
||||
|
||||
@@ -18,7 +18,8 @@ func concatEthAddrTokenID(addr ethCommon.Address, tokenID common.TokenID) []byte
|
||||
b = append(b[:], tokenID.Bytes()[:]...)
|
||||
return b
|
||||
}
|
||||
func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp, tokenID common.TokenID) []byte {
|
||||
func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp,
|
||||
tokenID common.TokenID) []byte {
|
||||
pkComp := pk
|
||||
var b []byte
|
||||
b = append(b, addr.Bytes()...)
|
||||
@@ -32,7 +33,8 @@ func concatEthAddrBJJTokenID(addr ethCommon.Address, pk babyjub.PublicKeyComp, t
|
||||
// - key: EthAddr & BabyJubJub PublicKey Compressed, value: idx
|
||||
// If Idx already exist for the given EthAddr & BJJ, the remaining Idx will be
|
||||
// always the smallest one.
|
||||
func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address, pk babyjub.PublicKeyComp, tokenID common.TokenID) error {
|
||||
func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address,
|
||||
pk babyjub.PublicKeyComp, tokenID common.TokenID) error {
|
||||
oldIdx, err := s.GetIdxByEthAddrBJJ(addr, pk, tokenID)
|
||||
if err == nil {
|
||||
// EthAddr & BJJ already have an Idx
|
||||
@@ -40,7 +42,8 @@ func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address, pk
|
||||
// if new idx is smaller, store the new one
|
||||
// if new idx is bigger, don't store and return, as the used one will be the old
|
||||
if idx >= oldIdx {
|
||||
log.Debug("StateDB.setIdxByEthAddrBJJ: Idx not stored because there already exist a smaller Idx for the given EthAddr & BJJ")
|
||||
log.Debug("StateDB.setIdxByEthAddrBJJ: Idx not stored because there " +
|
||||
"already exist a smaller Idx for the given EthAddr & BJJ")
|
||||
return nil
|
||||
}
|
||||
}
|
||||
@@ -80,7 +83,8 @@ func (s *StateDB) setIdxByEthAddrBJJ(idx common.Idx, addr ethCommon.Address, pk
|
||||
// GetIdxByEthAddr returns the smallest Idx in the StateDB for the given
|
||||
// Ethereum Address. Will return common.Idx(0) and error in case that Idx is
|
||||
// not found in the StateDB.
|
||||
func (s *StateDB) GetIdxByEthAddr(addr ethCommon.Address, tokenID common.TokenID) (common.Idx, error) {
|
||||
func (s *StateDB) GetIdxByEthAddr(addr ethCommon.Address, tokenID common.TokenID) (common.Idx,
|
||||
error) {
|
||||
k := concatEthAddrTokenID(addr, tokenID)
|
||||
b, err := s.db.DB().Get(append(PrefixKeyAddr, k...))
|
||||
if err != nil {
|
||||
@@ -116,18 +120,22 @@ func (s *StateDB) GetIdxByEthAddrBJJ(addr ethCommon.Address, pk babyjub.PublicKe
|
||||
return common.Idx(0), tracerr.Wrap(ErrIdxNotFound)
|
||||
} else if err != nil {
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrIdxNotFound, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
ErrIdxNotFound, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
idx, err := common.IdxFromBytes(b)
|
||||
if err != nil {
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", err, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
err, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
return idx, nil
|
||||
}
|
||||
// rest of cases (included case ToEthAddr==0) are not possible
|
||||
return common.Idx(0),
|
||||
tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
||||
tracerr.Wrap(
|
||||
fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d",
|
||||
ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
||||
}
|
||||
|
||||
// GetTokenIDsFromIdxs returns a map containing the common.TokenID with its
|
||||
@@ -137,7 +145,9 @@ func (s *StateDB) GetTokenIDsFromIdxs(idxs []common.Idx) (map[common.TokenID]com
|
||||
for i := 0; i < len(idxs); i++ {
|
||||
a, err := s.GetAccount(idxs[i])
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("GetTokenIDsFromIdxs error on GetAccount with Idx==%d: %s", idxs[i], err.Error()))
|
||||
return nil,
|
||||
tracerr.Wrap(fmt.Errorf("GetTokenIDsFromIdxs error on GetAccount with Idx==%d: %s",
|
||||
idxs[i], err.Error()))
|
||||
}
|
||||
m[a.TokenID] = idxs[i]
|
||||
}
|
||||
|
||||
@@ -18,7 +18,7 @@ func TestGetIdx(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
defer assert.NoError(t, os.RemoveAll(dir))
|
||||
|
||||
sdb, err := NewStateDB(dir, 128, TypeTxSelector, 0)
|
||||
sdb, err := NewStateDB(Config{Path: dir, Keep: 128, Type: TypeTxSelector, NLevels: 0})
|
||||
assert.NoError(t, err)
|
||||
|
||||
var sk babyjub.PrivateKey
|
||||
|
||||
29
db/utils.go
29
db/utils.go
@@ -1,11 +1,13 @@
|
||||
package db
|
||||
|
||||
import (
|
||||
"context"
|
||||
"database/sql"
|
||||
"fmt"
|
||||
"math/big"
|
||||
"reflect"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/gobuffalo/packr/v2"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
@@ -13,6 +15,7 @@ import (
|
||||
"github.com/jmoiron/sqlx"
|
||||
migrate "github.com/rubenv/sql-migrate"
|
||||
"github.com/russross/meddler"
|
||||
"golang.org/x/sync/semaphore"
|
||||
)
|
||||
|
||||
var migrations *migrate.PackrMigrationSource
|
||||
@@ -84,6 +87,32 @@ func InitSQLDB(port int, host, user, password, name string) (*sqlx.DB, error) {
|
||||
return db, nil
|
||||
}
|
||||
|
||||
// APIConnectionController is used to limit the SQL open connections used by the API
|
||||
type APIConnectionController struct {
|
||||
smphr *semaphore.Weighted
|
||||
timeout time.Duration
|
||||
}
|
||||
|
||||
// NewAPIConnectionController initialize APIConnectionController
|
||||
func NewAPIConnectionController(maxConnections int, timeout time.Duration) *APIConnectionController {
|
||||
return &APIConnectionController{
|
||||
smphr: semaphore.NewWeighted(int64(maxConnections)),
|
||||
timeout: timeout,
|
||||
}
|
||||
}
|
||||
|
||||
// Acquire reserves a SQL connection. If the connection is not acquired
|
||||
// within the timeout, the function will return an error
|
||||
func (acc *APIConnectionController) Acquire() (context.CancelFunc, error) {
|
||||
ctx, cancel := context.WithTimeout(context.Background(), acc.timeout) //nolint:govet
|
||||
return cancel, acc.smphr.Acquire(ctx, 1)
|
||||
}
|
||||
|
||||
// Release frees a SQL connection
|
||||
func (acc *APIConnectionController) Release() {
|
||||
acc.smphr.Release(1)
|
||||
}
|
||||
|
||||
// initMeddler registers tags to be used to read/write from SQL DBs using meddler
|
||||
func initMeddler() {
|
||||
meddler.Register("bigint", BigIntMeddler{})
|
||||
|
||||
242
eth/auction.go
242
eth/auction.go
@@ -70,7 +70,8 @@ type AuctionEventInitialize struct {
|
||||
}
|
||||
|
||||
// AuctionVariables returns the AuctionVariables from the initialize event
|
||||
func (ei *AuctionEventInitialize) AuctionVariables(InitialMinimalBidding *big.Int) *common.AuctionVariables {
|
||||
func (ei *AuctionEventInitialize) AuctionVariables(
|
||||
InitialMinimalBidding *big.Int) *common.AuctionVariables {
|
||||
return &common.AuctionVariables{
|
||||
EthBlockNum: 0,
|
||||
DonationAddress: ei.DonationAddress,
|
||||
@@ -222,12 +223,15 @@ type AuctionInterface interface {
|
||||
AuctionGetAllocationRatio() ([3]uint16, error)
|
||||
AuctionSetDonationAddress(newDonationAddress ethCommon.Address) (*types.Transaction, error)
|
||||
AuctionGetDonationAddress() (*ethCommon.Address, error)
|
||||
AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address, newBootCoordinatorURL string) (*types.Transaction, error)
|
||||
AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address,
|
||||
newBootCoordinatorURL string) (*types.Transaction, error)
|
||||
AuctionGetBootCoordinator() (*ethCommon.Address, error)
|
||||
AuctionChangeDefaultSlotSetBid(slotSet int64, newInitialMinBid *big.Int) (*types.Transaction, error)
|
||||
AuctionChangeDefaultSlotSetBid(slotSet int64,
|
||||
newInitialMinBid *big.Int) (*types.Transaction, error)
|
||||
|
||||
// Coordinator Management
|
||||
AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (*types.Transaction, error)
|
||||
AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (*types.Transaction,
|
||||
error)
|
||||
|
||||
// Slot Info
|
||||
AuctionGetSlotNumber(blockNum int64) (int64, error)
|
||||
@@ -237,7 +241,8 @@ type AuctionInterface interface {
|
||||
AuctionGetSlotSet(slot int64) (*big.Int, error)
|
||||
|
||||
// Bidding
|
||||
AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (
|
||||
tx *types.Transaction, err error)
|
||||
AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64, slotSets [6]bool,
|
||||
maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
|
||||
@@ -254,7 +259,7 @@ type AuctionInterface interface {
|
||||
//
|
||||
|
||||
AuctionConstants() (*common.AuctionConstants, error)
|
||||
AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *ethCommon.Hash, error)
|
||||
AuctionEventsByBlock(blockNum int64, blockHash *ethCommon.Hash) (*AuctionEvents, error)
|
||||
AuctionEventInit() (*AuctionEventInitialize, int64, error)
|
||||
}
|
||||
|
||||
@@ -275,8 +280,10 @@ type AuctionClient struct {
|
||||
}
|
||||
|
||||
// NewAuctionClient creates a new AuctionClient. `tokenAddress` is the address of the HEZ tokens.
|
||||
func NewAuctionClient(client *EthereumClient, address ethCommon.Address, tokenHEZCfg TokenConfig) (*AuctionClient, error) {
|
||||
contractAbi, err := abi.JSON(strings.NewReader(string(HermezAuctionProtocol.HermezAuctionProtocolABI)))
|
||||
func NewAuctionClient(client *EthereumClient, address ethCommon.Address,
|
||||
tokenHEZCfg TokenConfig) (*AuctionClient, error) {
|
||||
contractAbi, err :=
|
||||
abi.JSON(strings.NewReader(string(HermezAuctionProtocol.HermezAuctionProtocolABI)))
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -331,7 +338,8 @@ func (c *AuctionClient) AuctionGetSlotDeadline() (slotDeadline uint8, err error)
|
||||
}
|
||||
|
||||
// AuctionSetOpenAuctionSlots is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetOpenAuctionSlots(newOpenAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetOpenAuctionSlots(
|
||||
newOpenAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -355,7 +363,8 @@ func (c *AuctionClient) AuctionGetOpenAuctionSlots() (openAuctionSlots uint16, e
|
||||
}
|
||||
|
||||
// AuctionSetClosedAuctionSlots is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetClosedAuctionSlots(newClosedAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetClosedAuctionSlots(
|
||||
newClosedAuctionSlots uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -379,7 +388,8 @@ func (c *AuctionClient) AuctionGetClosedAuctionSlots() (closedAuctionSlots uint1
|
||||
}
|
||||
|
||||
// AuctionSetOutbidding is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetOutbidding(newOutbidding uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetOutbidding(newOutbidding uint16) (tx *types.Transaction,
|
||||
err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
12500000, //nolint:gomnd
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -403,7 +413,8 @@ func (c *AuctionClient) AuctionGetOutbidding() (outbidding uint16, err error) {
|
||||
}
|
||||
|
||||
// AuctionSetAllocationRatio is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetAllocationRatio(newAllocationRatio [3]uint16) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetAllocationRatio(
|
||||
newAllocationRatio [3]uint16) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -427,7 +438,8 @@ func (c *AuctionClient) AuctionGetAllocationRatio() (allocationRation [3]uint16,
|
||||
}
|
||||
|
||||
// AuctionSetDonationAddress is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetDonationAddress(newDonationAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetDonationAddress(
|
||||
newDonationAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -440,7 +452,8 @@ func (c *AuctionClient) AuctionSetDonationAddress(newDonationAddress ethCommon.A
|
||||
}
|
||||
|
||||
// AuctionGetDonationAddress is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.Address, err error) {
|
||||
func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.Address,
|
||||
err error) {
|
||||
var _donationAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_donationAddress, err = c.auction.GetDonationAddress(c.opts)
|
||||
@@ -452,11 +465,13 @@ func (c *AuctionClient) AuctionGetDonationAddress() (donationAddress *ethCommon.
|
||||
}
|
||||
|
||||
// AuctionSetBootCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address, newBootCoordinatorURL string) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.Address,
|
||||
newBootCoordinatorURL string) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
return c.auction.SetBootCoordinator(auth, newBootCoordinator, newBootCoordinatorURL)
|
||||
return c.auction.SetBootCoordinator(auth, newBootCoordinator,
|
||||
newBootCoordinatorURL)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed setting bootCoordinator: %w", err))
|
||||
@@ -465,7 +480,8 @@ func (c *AuctionClient) AuctionSetBootCoordinator(newBootCoordinator ethCommon.A
|
||||
}
|
||||
|
||||
// AuctionGetBootCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.Address, err error) {
|
||||
func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.Address,
|
||||
err error) {
|
||||
var _bootCoordinator ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_bootCoordinator, err = c.auction.GetBootCoordinator(c.opts)
|
||||
@@ -477,7 +493,8 @@ func (c *AuctionClient) AuctionGetBootCoordinator() (bootCoordinator *ethCommon.
|
||||
}
|
||||
|
||||
// AuctionChangeDefaultSlotSetBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64, newInitialMinBid *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64,
|
||||
newInitialMinBid *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -491,7 +508,8 @@ func (c *AuctionClient) AuctionChangeDefaultSlotSetBid(slotSet int64, newInitial
|
||||
}
|
||||
|
||||
// AuctionGetClaimableHEZ is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetClaimableHEZ(claimAddress ethCommon.Address) (claimableHEZ *big.Int, err error) {
|
||||
func (c *AuctionClient) AuctionGetClaimableHEZ(
|
||||
claimAddress ethCommon.Address) (claimableHEZ *big.Int, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
claimableHEZ, err = c.auction.GetClaimableHEZ(c.opts, claimAddress)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -502,7 +520,8 @@ func (c *AuctionClient) AuctionGetClaimableHEZ(claimAddress ethCommon.Address) (
|
||||
}
|
||||
|
||||
// AuctionSetCoordinator is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionSetCoordinator(forger ethCommon.Address, coordinatorURL string) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionSetCoordinator(forger ethCommon.Address,
|
||||
coordinatorURL string) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -551,7 +570,8 @@ func (c *AuctionClient) AuctionGetSlotSet(slot int64) (slotSet *big.Int, err err
|
||||
}
|
||||
|
||||
// AuctionGetDefaultSlotSetBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionGetDefaultSlotSetBid(slotSet uint8) (minBidSlotSet *big.Int, err error) {
|
||||
func (c *AuctionClient) AuctionGetDefaultSlotSetBid(slotSet uint8) (minBidSlotSet *big.Int,
|
||||
err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
minBidSlotSet, err = c.auction.GetDefaultSlotSetBid(c.opts, slotSet)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -574,7 +594,8 @@ func (c *AuctionClient) AuctionGetSlotNumber(blockNum int64) (slot int64, err er
|
||||
}
|
||||
|
||||
// AuctionBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.Int,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -586,7 +607,8 @@ func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.I
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
_slot := big.NewInt(slot)
|
||||
@@ -599,8 +621,8 @@ func (c *AuctionClient) AuctionBid(amount *big.Int, slot int64, bidAmount *big.I
|
||||
}
|
||||
|
||||
// AuctionMultiBid is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64, slotSets [6]bool,
|
||||
maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlot int64,
|
||||
slotSets [6]bool, maxBid, minBid, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
1000000, //nolint:gomnd
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -613,12 +635,14 @@ func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlo
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
_startingSlot := big.NewInt(startingSlot)
|
||||
_endingSlot := big.NewInt(endingSlot)
|
||||
return c.auction.ProcessMultiBid(auth, amount, _startingSlot, _endingSlot, slotSets, maxBid, minBid, permit)
|
||||
return c.auction.ProcessMultiBid(auth, amount, _startingSlot, _endingSlot,
|
||||
slotSets, maxBid, minBid, permit)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed multibid: %w", err))
|
||||
@@ -627,7 +651,8 @@ func (c *AuctionClient) AuctionMultiBid(amount *big.Int, startingSlot, endingSlo
|
||||
}
|
||||
|
||||
// AuctionCanForge is the interface to call the smart contract function
|
||||
func (c *AuctionClient) AuctionCanForge(forger ethCommon.Address, blockNum int64) (canForge bool, err error) {
|
||||
func (c *AuctionClient) AuctionCanForge(forger ethCommon.Address, blockNum int64) (canForge bool,
|
||||
err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
canForge, err = c.auction.CanForge(c.opts, forger, big.NewInt(blockNum))
|
||||
return tracerr.Wrap(err)
|
||||
@@ -680,7 +705,8 @@ func (c *AuctionClient) AuctionConstants() (auctionConstants *common.AuctionCons
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
auctionConstants.InitialMinimalBidding, err = c.auction.INITIALMINIMALBIDDING(c.opts)
|
||||
auctionConstants.InitialMinimalBidding, err =
|
||||
c.auction.INITIALMINIMALBIDDING(c.opts)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
@@ -751,21 +777,35 @@ func (c *AuctionClient) AuctionVariables() (auctionVariables *common.AuctionVari
|
||||
}
|
||||
|
||||
var (
|
||||
logAuctionNewBid = crypto.Keccak256Hash([]byte("NewBid(uint128,uint128,address)"))
|
||||
logAuctionNewSlotDeadline = crypto.Keccak256Hash([]byte("NewSlotDeadline(uint8)"))
|
||||
logAuctionNewClosedAuctionSlots = crypto.Keccak256Hash([]byte("NewClosedAuctionSlots(uint16)"))
|
||||
logAuctionNewOutbidding = crypto.Keccak256Hash([]byte("NewOutbidding(uint16)"))
|
||||
logAuctionNewDonationAddress = crypto.Keccak256Hash([]byte("NewDonationAddress(address)"))
|
||||
logAuctionNewBootCoordinator = crypto.Keccak256Hash([]byte("NewBootCoordinator(address,string)"))
|
||||
logAuctionNewOpenAuctionSlots = crypto.Keccak256Hash([]byte("NewOpenAuctionSlots(uint16)"))
|
||||
logAuctionNewAllocationRatio = crypto.Keccak256Hash([]byte("NewAllocationRatio(uint16[3])"))
|
||||
logAuctionSetCoordinator = crypto.Keccak256Hash([]byte("SetCoordinator(address,address,string)"))
|
||||
logAuctionNewForgeAllocated = crypto.Keccak256Hash([]byte("NewForgeAllocated(address,address,uint128,uint128,uint128,uint128)"))
|
||||
logAuctionNewDefaultSlotSetBid = crypto.Keccak256Hash([]byte("NewDefaultSlotSetBid(uint128,uint128)"))
|
||||
logAuctionNewForge = crypto.Keccak256Hash([]byte("NewForge(address,uint128)"))
|
||||
logAuctionHEZClaimed = crypto.Keccak256Hash([]byte("HEZClaimed(address,uint128)"))
|
||||
logAuctionInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezAuctionProtocolEvent(address,address,string,uint16,uint8,uint16,uint16,uint16[3])"))
|
||||
logAuctionNewBid = crypto.Keccak256Hash([]byte(
|
||||
"NewBid(uint128,uint128,address)"))
|
||||
logAuctionNewSlotDeadline = crypto.Keccak256Hash([]byte(
|
||||
"NewSlotDeadline(uint8)"))
|
||||
logAuctionNewClosedAuctionSlots = crypto.Keccak256Hash([]byte(
|
||||
"NewClosedAuctionSlots(uint16)"))
|
||||
logAuctionNewOutbidding = crypto.Keccak256Hash([]byte(
|
||||
"NewOutbidding(uint16)"))
|
||||
logAuctionNewDonationAddress = crypto.Keccak256Hash([]byte(
|
||||
"NewDonationAddress(address)"))
|
||||
logAuctionNewBootCoordinator = crypto.Keccak256Hash([]byte(
|
||||
"NewBootCoordinator(address,string)"))
|
||||
logAuctionNewOpenAuctionSlots = crypto.Keccak256Hash([]byte(
|
||||
"NewOpenAuctionSlots(uint16)"))
|
||||
logAuctionNewAllocationRatio = crypto.Keccak256Hash([]byte(
|
||||
"NewAllocationRatio(uint16[3])"))
|
||||
logAuctionSetCoordinator = crypto.Keccak256Hash([]byte(
|
||||
"SetCoordinator(address,address,string)"))
|
||||
logAuctionNewForgeAllocated = crypto.Keccak256Hash([]byte(
|
||||
"NewForgeAllocated(address,address,uint128,uint128,uint128,uint128)"))
|
||||
logAuctionNewDefaultSlotSetBid = crypto.Keccak256Hash([]byte(
|
||||
"NewDefaultSlotSetBid(uint128,uint128)"))
|
||||
logAuctionNewForge = crypto.Keccak256Hash([]byte(
|
||||
"NewForge(address,uint128)"))
|
||||
logAuctionHEZClaimed = crypto.Keccak256Hash([]byte(
|
||||
"HEZClaimed(address,uint128)"))
|
||||
logAuctionInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezAuctionProtocolEvent(address,address,string," +
|
||||
"uint16,uint8,uint16,uint16,uint16[3])"))
|
||||
)
|
||||
|
||||
// AuctionEventInit returns the initialize event with its corresponding block number
|
||||
@@ -781,7 +821,8 @@ func (c *AuctionClient) AuctionEventInit() (*AuctionEventInitialize, int64, erro
|
||||
return nil, 0, tracerr.Wrap(err)
|
||||
}
|
||||
if len(logs) != 1 {
|
||||
return nil, 0, tracerr.Wrap(fmt.Errorf("no event of type InitializeHermezAuctionProtocolEvent found"))
|
||||
return nil, 0,
|
||||
tracerr.Wrap(fmt.Errorf("no event of type InitializeHermezAuctionProtocolEvent found"))
|
||||
}
|
||||
vLog := logs[0]
|
||||
if vLog.Topics[0] != logAuctionInitialize {
|
||||
@@ -797,15 +838,22 @@ func (c *AuctionClient) AuctionEventInit() (*AuctionEventInitialize, int64, erro
|
||||
}
|
||||
|
||||
// AuctionEventsByBlock returns the events in a block that happened in the
|
||||
// Auction Smart Contract and the blockHash where the eents happened. If there
|
||||
// are no events in that block, blockHash is nil.
|
||||
func (c *AuctionClient) AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *ethCommon.Hash, error) {
|
||||
// Auction Smart Contract.
|
||||
// To query by blockNum, set blockNum >= 0 and blockHash == nil.
|
||||
// To query by blockHash set blockHash != nil, and blockNum will be ignored.
|
||||
// If there are no events in that block the result is nil.
|
||||
func (c *AuctionClient) AuctionEventsByBlock(blockNum int64,
|
||||
blockHash *ethCommon.Hash) (*AuctionEvents, error) {
|
||||
var auctionEvents AuctionEvents
|
||||
var blockHash *ethCommon.Hash
|
||||
|
||||
var blockNumBigInt *big.Int
|
||||
if blockHash == nil {
|
||||
blockNumBigInt = big.NewInt(blockNum)
|
||||
}
|
||||
query := ethereum.FilterQuery{
|
||||
FromBlock: big.NewInt(blockNum),
|
||||
ToBlock: big.NewInt(blockNum),
|
||||
BlockHash: blockHash,
|
||||
FromBlock: blockNumBigInt,
|
||||
ToBlock: blockNumBigInt,
|
||||
Addresses: []ethCommon.Address{
|
||||
c.address,
|
||||
},
|
||||
@@ -814,15 +862,17 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *e
|
||||
|
||||
logs, err := c.client.client.FilterLogs(context.TODO(), query)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if len(logs) > 0 {
|
||||
blockHash = &logs[0].BlockHash
|
||||
if len(logs) == 0 {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
for _, vLog := range logs {
|
||||
if vLog.BlockHash != *blockHash {
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got", vLog.BlockHash.String())
|
||||
return nil, nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
if blockHash != nil && vLog.BlockHash != *blockHash {
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got",
|
||||
vLog.BlockHash.String())
|
||||
return nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
}
|
||||
switch vLog.Topics[0] {
|
||||
case logAuctionNewBid:
|
||||
@@ -832,8 +882,9 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *e
|
||||
Address ethCommon.Address
|
||||
}
|
||||
var newBid AuctionEventNewBid
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewBid, "NewBid", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewBid, "NewBid",
|
||||
vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newBid.BidAmount = auxNewBid.BidAmount
|
||||
newBid.Slot = new(big.Int).SetBytes(vLog.Topics[1][:]).Int64()
|
||||
@@ -841,74 +892,90 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *e
|
||||
auctionEvents.NewBid = append(auctionEvents.NewBid, newBid)
|
||||
case logAuctionNewSlotDeadline:
|
||||
var newSlotDeadline AuctionEventNewSlotDeadline
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newSlotDeadline, "NewSlotDeadline", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newSlotDeadline,
|
||||
"NewSlotDeadline", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewSlotDeadline = append(auctionEvents.NewSlotDeadline, newSlotDeadline)
|
||||
case logAuctionNewClosedAuctionSlots:
|
||||
var newClosedAuctionSlots AuctionEventNewClosedAuctionSlots
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newClosedAuctionSlots, "NewClosedAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newClosedAuctionSlots,
|
||||
"NewClosedAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewClosedAuctionSlots = append(auctionEvents.NewClosedAuctionSlots, newClosedAuctionSlots)
|
||||
auctionEvents.NewClosedAuctionSlots =
|
||||
append(auctionEvents.NewClosedAuctionSlots, newClosedAuctionSlots)
|
||||
case logAuctionNewOutbidding:
|
||||
var newOutbidding AuctionEventNewOutbidding
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOutbidding, "NewOutbidding", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOutbidding, "NewOutbidding",
|
||||
vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewOutbidding = append(auctionEvents.NewOutbidding, newOutbidding)
|
||||
case logAuctionNewDonationAddress:
|
||||
var newDonationAddress AuctionEventNewDonationAddress
|
||||
newDonationAddress.NewDonationAddress = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
auctionEvents.NewDonationAddress = append(auctionEvents.NewDonationAddress, newDonationAddress)
|
||||
auctionEvents.NewDonationAddress = append(auctionEvents.NewDonationAddress,
|
||||
newDonationAddress)
|
||||
case logAuctionNewBootCoordinator:
|
||||
var newBootCoordinator AuctionEventNewBootCoordinator
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newBootCoordinator, "NewBootCoordinator", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newBootCoordinator,
|
||||
"NewBootCoordinator", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newBootCoordinator.NewBootCoordinator = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
auctionEvents.NewBootCoordinator = append(auctionEvents.NewBootCoordinator, newBootCoordinator)
|
||||
auctionEvents.NewBootCoordinator = append(auctionEvents.NewBootCoordinator,
|
||||
newBootCoordinator)
|
||||
case logAuctionNewOpenAuctionSlots:
|
||||
var newOpenAuctionSlots AuctionEventNewOpenAuctionSlots
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOpenAuctionSlots, "NewOpenAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newOpenAuctionSlots,
|
||||
"NewOpenAuctionSlots", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewOpenAuctionSlots = append(auctionEvents.NewOpenAuctionSlots, newOpenAuctionSlots)
|
||||
auctionEvents.NewOpenAuctionSlots =
|
||||
append(auctionEvents.NewOpenAuctionSlots, newOpenAuctionSlots)
|
||||
case logAuctionNewAllocationRatio:
|
||||
var newAllocationRatio AuctionEventNewAllocationRatio
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newAllocationRatio, "NewAllocationRatio", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newAllocationRatio,
|
||||
"NewAllocationRatio", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionEvents.NewAllocationRatio = append(auctionEvents.NewAllocationRatio, newAllocationRatio)
|
||||
auctionEvents.NewAllocationRatio = append(auctionEvents.NewAllocationRatio,
|
||||
newAllocationRatio)
|
||||
case logAuctionSetCoordinator:
|
||||
var setCoordinator AuctionEventSetCoordinator
|
||||
if err := c.contractAbi.UnpackIntoInterface(&setCoordinator, "SetCoordinator", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&setCoordinator,
|
||||
"SetCoordinator", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
setCoordinator.BidderAddress = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
setCoordinator.ForgerAddress = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
auctionEvents.SetCoordinator = append(auctionEvents.SetCoordinator, setCoordinator)
|
||||
case logAuctionNewForgeAllocated:
|
||||
var newForgeAllocated AuctionEventNewForgeAllocated
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newForgeAllocated, "NewForgeAllocated", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&newForgeAllocated,
|
||||
"NewForgeAllocated", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newForgeAllocated.Bidder = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
newForgeAllocated.Forger = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
newForgeAllocated.SlotToForge = new(big.Int).SetBytes(vLog.Topics[3][:]).Int64()
|
||||
auctionEvents.NewForgeAllocated = append(auctionEvents.NewForgeAllocated, newForgeAllocated)
|
||||
auctionEvents.NewForgeAllocated = append(auctionEvents.NewForgeAllocated,
|
||||
newForgeAllocated)
|
||||
case logAuctionNewDefaultSlotSetBid:
|
||||
var auxNewDefaultSlotSetBid struct {
|
||||
SlotSet *big.Int
|
||||
NewInitialMinBid *big.Int
|
||||
}
|
||||
var newDefaultSlotSetBid AuctionEventNewDefaultSlotSetBid
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewDefaultSlotSetBid, "NewDefaultSlotSetBid", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&auxNewDefaultSlotSetBid,
|
||||
"NewDefaultSlotSetBid", vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
newDefaultSlotSetBid.NewInitialMinBid = auxNewDefaultSlotSetBid.NewInitialMinBid
|
||||
newDefaultSlotSetBid.SlotSet = auxNewDefaultSlotSetBid.SlotSet.Int64()
|
||||
auctionEvents.NewDefaultSlotSetBid = append(auctionEvents.NewDefaultSlotSetBid, newDefaultSlotSetBid)
|
||||
auctionEvents.NewDefaultSlotSetBid =
|
||||
append(auctionEvents.NewDefaultSlotSetBid, newDefaultSlotSetBid)
|
||||
case logAuctionNewForge:
|
||||
var newForge AuctionEventNewForge
|
||||
newForge.Forger = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
@@ -916,12 +983,13 @@ func (c *AuctionClient) AuctionEventsByBlock(blockNum int64) (*AuctionEvents, *e
|
||||
auctionEvents.NewForge = append(auctionEvents.NewForge, newForge)
|
||||
case logAuctionHEZClaimed:
|
||||
var HEZClaimed AuctionEventHEZClaimed
|
||||
if err := c.contractAbi.UnpackIntoInterface(&HEZClaimed, "HEZClaimed", vLog.Data); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
if err := c.contractAbi.UnpackIntoInterface(&HEZClaimed, "HEZClaimed",
|
||||
vLog.Data); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
HEZClaimed.Owner = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
auctionEvents.HEZClaimed = append(auctionEvents.HEZClaimed, HEZClaimed)
|
||||
}
|
||||
}
|
||||
return &auctionEvents, blockHash, nil
|
||||
return &auctionEvents, nil
|
||||
}
|
||||
|
||||
@@ -58,7 +58,8 @@ func TestAuctionConstants(t *testing.T) {
|
||||
func TestAuctionVariables(t *testing.T) {
|
||||
INITMINBID := new(big.Int)
|
||||
INITMINBID.SetString(minBidStr, 10)
|
||||
defaultSlotSetBid := [6]*big.Int{INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID}
|
||||
defaultSlotSetBid := [6]*big.Int{INITMINBID, INITMINBID, INITMINBID, INITMINBID, INITMINBID,
|
||||
INITMINBID}
|
||||
|
||||
auctionVariables, err := auctionClientTest.AuctionVariables()
|
||||
require.Nil(t, err)
|
||||
@@ -88,7 +89,7 @@ func TestAuctionSetSlotDeadline(t *testing.T) {
|
||||
assert.Equal(t, newSlotDeadline, slotDeadline)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newSlotDeadline, auctionEvents.NewSlotDeadline[0].NewSlotDeadline)
|
||||
}
|
||||
@@ -109,7 +110,7 @@ func TestAuctionSetOpenAuctionSlots(t *testing.T) {
|
||||
assert.Equal(t, newOpenAuctionSlots, openAuctionSlots)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newOpenAuctionSlots, auctionEvents.NewOpenAuctionSlots[0].NewOpenAuctionSlots)
|
||||
}
|
||||
@@ -130,9 +131,10 @@ func TestAuctionSetClosedAuctionSlots(t *testing.T) {
|
||||
assert.Equal(t, newClosedAuctionSlots, closedAuctionSlots)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newClosedAuctionSlots, auctionEvents.NewClosedAuctionSlots[0].NewClosedAuctionSlots)
|
||||
assert.Equal(t, newClosedAuctionSlots,
|
||||
auctionEvents.NewClosedAuctionSlots[0].NewClosedAuctionSlots)
|
||||
_, err = auctionClientTest.AuctionSetClosedAuctionSlots(closedAuctionSlots)
|
||||
require.Nil(t, err)
|
||||
}
|
||||
@@ -153,7 +155,7 @@ func TestAuctionSetOutbidding(t *testing.T) {
|
||||
assert.Equal(t, newOutbidding, outbidding)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newOutbidding, auctionEvents.NewOutbidding[0].NewOutbidding)
|
||||
_, err = auctionClientTest.AuctionSetOutbidding(outbiddingConst)
|
||||
@@ -176,7 +178,7 @@ func TestAuctionSetAllocationRatio(t *testing.T) {
|
||||
assert.Equal(t, newAllocationRatio, allocationRatio)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newAllocationRatio, auctionEvents.NewAllocationRatio[0].NewAllocationRatio)
|
||||
_, err = auctionClientTest.AuctionSetAllocationRatio(allocationRatioConst)
|
||||
@@ -205,7 +207,7 @@ func TestAuctionSetDonationAddress(t *testing.T) {
|
||||
assert.Equal(t, &newDonationAddress, donationAddress)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newDonationAddress, auctionEvents.NewDonationAddress[0].NewDonationAddress)
|
||||
_, err = auctionClientTest.AuctionSetDonationAddress(donationAddressConst)
|
||||
@@ -224,11 +226,12 @@ func TestAuctionSetBootCoordinator(t *testing.T) {
|
||||
assert.Equal(t, &newBootCoordinator, bootCoordinator)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, newBootCoordinator, auctionEvents.NewBootCoordinator[0].NewBootCoordinator)
|
||||
assert.Equal(t, newBootCoordinatorURL, auctionEvents.NewBootCoordinator[0].NewBootCoordinatorURL)
|
||||
_, err = auctionClientTest.AuctionSetBootCoordinator(bootCoordinatorAddressConst, bootCoordinatorURL)
|
||||
_, err = auctionClientTest.AuctionSetBootCoordinator(bootCoordinatorAddressConst,
|
||||
bootCoordinatorURL)
|
||||
require.Nil(t, err)
|
||||
}
|
||||
|
||||
@@ -261,7 +264,7 @@ func TestAuctionChangeDefaultSlotSetBid(t *testing.T) {
|
||||
assert.Equal(t, minBid, newInitialMinBid)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, slotSet, auctionEvents.NewDefaultSlotSetBid[0].SlotSet)
|
||||
assert.Equal(t, newInitialMinBid, auctionEvents.NewDefaultSlotSetBid[0].NewInitialMinBid)
|
||||
@@ -287,7 +290,7 @@ func TestAuctionRegisterCoordinator(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, forgerAddress, auctionEvents.SetCoordinator[0].ForgerAddress)
|
||||
assert.Equal(t, bidderAddress, auctionEvents.SetCoordinator[0].BidderAddress)
|
||||
@@ -306,7 +309,7 @@ func TestAuctionBid(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, bidAmount, auctionEvents.NewBid[0].BidAmount)
|
||||
assert.Equal(t, bidderAddress, auctionEvents.NewBid[0].Bidder)
|
||||
@@ -342,11 +345,12 @@ func TestAuctionMultiBid(t *testing.T) {
|
||||
budget := new(big.Int)
|
||||
budget.SetString("45200000000000000000", 10)
|
||||
bidderAddress := governanceAddressConst
|
||||
_, err = auctionClientTest.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet, maxBid, minBid, deadline)
|
||||
_, err = auctionClientTest.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet,
|
||||
maxBid, minBid, deadline)
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, bidderAddress, auctionEvents.NewBid[0].Bidder)
|
||||
assert.Equal(t, currentSlot+4, auctionEvents.NewBid[0].Slot)
|
||||
@@ -376,14 +380,15 @@ func TestAuctionClaimHEZ(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := auctionClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
auctionEvents, _, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum)
|
||||
auctionEvents, err := auctionClientTest.AuctionEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, amount, auctionEvents.HEZClaimed[0].Amount)
|
||||
assert.Equal(t, governanceAddressConst, auctionEvents.HEZClaimed[0].Owner)
|
||||
}
|
||||
|
||||
func TestAuctionForge(t *testing.T) {
|
||||
auctionClientTestHermez, err := NewAuctionClient(ethereumClientHermez, auctionTestAddressConst, tokenHEZ)
|
||||
auctionClientTestHermez, err := NewAuctionClient(ethereumClientHermez,
|
||||
auctionTestAddressConst, tokenHEZ)
|
||||
require.Nil(t, err)
|
||||
slotConst := 4
|
||||
blockNum := int64(int(blocksPerSlot)*slotConst + int(genesisBlock))
|
||||
|
||||
@@ -64,16 +64,19 @@ type ClientConfig struct {
|
||||
}
|
||||
|
||||
// NewClient creates a new Client to interact with Ethereum and the Hermez smart contracts.
|
||||
func NewClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore, cfg *ClientConfig) (*Client, error) {
|
||||
func NewClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore,
|
||||
cfg *ClientConfig) (*Client, error) {
|
||||
ethereumClient, err := NewEthereumClient(client, account, ks, &cfg.Ethereum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auctionClient, err := NewAuctionClient(ethereumClient, cfg.Auction.Address, cfg.Auction.TokenHEZ)
|
||||
auctionClient, err := NewAuctionClient(ethereumClient, cfg.Auction.Address,
|
||||
cfg.Auction.TokenHEZ)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupClient, err := NewRollupClient(ethereumClient, cfg.Rollup.Address, cfg.Auction.TokenHEZ)
|
||||
rollupClient, err := NewRollupClient(ethereumClient, cfg.Rollup.Address,
|
||||
cfg.Auction.TokenHEZ)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -64,7 +64,8 @@ type EthereumConfig struct {
|
||||
GasPriceDiv uint64
|
||||
}
|
||||
|
||||
// EthereumClient is an ethereum client to call Smart Contract methods and check blockchain information.
|
||||
// EthereumClient is an ethereum client to call Smart Contract methods and check blockchain
|
||||
// information.
|
||||
type EthereumClient struct {
|
||||
client *ethclient.Client
|
||||
chainID *big.Int
|
||||
@@ -76,7 +77,8 @@ type EthereumClient struct {
|
||||
|
||||
// NewEthereumClient creates a EthereumClient instance. The account is not mandatory (it can
|
||||
// be nil). If the account is nil, CallAuth will fail with ErrAccountNil.
|
||||
func NewEthereumClient(client *ethclient.Client, account *accounts.Account, ks *ethKeystore.KeyStore, config *EthereumConfig) (*EthereumClient, error) {
|
||||
func NewEthereumClient(client *ethclient.Client, account *accounts.Account,
|
||||
ks *ethKeystore.KeyStore, config *EthereumConfig) (*EthereumClient, error) {
|
||||
if config == nil {
|
||||
config = &EthereumConfig{
|
||||
CallGasLimit: defaultCallGasLimit,
|
||||
@@ -166,7 +168,8 @@ func (c *EthereumClient) NewAuth() (*bind.TransactOpts, error) {
|
||||
// This call requires a valid account with Ether that can be spend during the
|
||||
// call.
|
||||
func (c *EthereumClient) CallAuth(gasLimit uint64,
|
||||
fn func(*ethclient.Client, *bind.TransactOpts) (*types.Transaction, error)) (*types.Transaction, error) {
|
||||
fn func(*ethclient.Client, *bind.TransactOpts) (*types.Transaction, error)) (*types.Transaction,
|
||||
error) {
|
||||
if c.account == nil {
|
||||
return nil, tracerr.Wrap(ErrAccountNil)
|
||||
}
|
||||
@@ -212,7 +215,8 @@ func (c *EthereumClient) Call(fn func(*ethclient.Client) error) error {
|
||||
}
|
||||
|
||||
// EthTransactionReceipt returns the transaction receipt of the given txHash
|
||||
func (c *EthereumClient) EthTransactionReceipt(ctx context.Context, txHash ethCommon.Hash) (*types.Receipt, error) {
|
||||
func (c *EthereumClient) EthTransactionReceipt(ctx context.Context,
|
||||
txHash ethCommon.Hash) (*types.Receipt, error) {
|
||||
return c.client.TransactionReceipt(ctx, txHash)
|
||||
}
|
||||
|
||||
@@ -228,13 +232,15 @@ func (c *EthereumClient) EthLastBlock() (int64, error) {
|
||||
}
|
||||
|
||||
// EthHeaderByNumber internally calls ethclient.Client HeaderByNumber
|
||||
// func (c *EthereumClient) EthHeaderByNumber(ctx context.Context, number *big.Int) (*types.Header, error) {
|
||||
// func (c *EthereumClient) EthHeaderByNumber(ctx context.Context, number *big.Int) (*types.Header,
|
||||
// error) {
|
||||
// return c.client.HeaderByNumber(ctx, number)
|
||||
// }
|
||||
|
||||
// EthBlockByNumber internally calls ethclient.Client BlockByNumber and returns
|
||||
// *common.Block. If number == -1, the latests known block is returned.
|
||||
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block, error) {
|
||||
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block,
|
||||
error) {
|
||||
blockNum := big.NewInt(number)
|
||||
if number == -1 {
|
||||
blockNum = nil
|
||||
@@ -324,5 +330,6 @@ func (c *EthereumClient) EthCall(ctx context.Context, tx *types.Transaction,
|
||||
Value: tx.Value(),
|
||||
Data: tx.Data(),
|
||||
}
|
||||
return c.client.CallContract(ctx, msg, blockNum)
|
||||
result, err := c.client.CallContract(ctx, msg, blockNum)
|
||||
return result, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
@@ -14,7 +14,8 @@ import (
|
||||
func addBlock(url string) {
|
||||
method := "POST"
|
||||
|
||||
payload := strings.NewReader("{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_mine\",\n \"params\":[],\n \"id\":1\n}")
|
||||
payload := strings.NewReader(
|
||||
"{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_mine\",\n \"params\":[],\n \"id\":1\n}")
|
||||
|
||||
client := &http.Client{}
|
||||
req, err := http.NewRequest(method, url, payload)
|
||||
@@ -45,7 +46,9 @@ func addTime(seconds float64, url string) {
|
||||
secondsStr := strconv.FormatFloat(seconds, 'E', -1, 32)
|
||||
|
||||
method := "POST"
|
||||
payload := strings.NewReader("{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_increaseTime\",\n \"params\":[" + secondsStr + "],\n \"id\":1\n}")
|
||||
payload := strings.NewReader(
|
||||
"{\n \"jsonrpc\":\"2.0\",\n \"method\":\"evm_increaseTime\",\n \"params\":[" +
|
||||
secondsStr + "],\n \"id\":1\n}")
|
||||
|
||||
client := &http.Client{}
|
||||
req, err := http.NewRequest(method, url, payload)
|
||||
@@ -66,13 +69,16 @@ func addTime(seconds float64, url string) {
|
||||
}()
|
||||
}
|
||||
|
||||
func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, value, nonce, deadline *big.Int, tokenName string) ([]byte, error) {
|
||||
func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, value, nonce,
|
||||
deadline *big.Int, tokenName string) ([]byte, error) {
|
||||
// NOTE: We ignore hash.Write errors because we are writing to a memory
|
||||
// buffer and don't expect any errors to occur.
|
||||
abiPermit := []byte("Permit(address owner,address spender,uint256 value,uint256 nonce,uint256 deadline)")
|
||||
abiPermit :=
|
||||
[]byte("Permit(address owner,address spender,uint256 value,uint256 nonce,uint256 deadline)")
|
||||
hashPermit := sha3.NewLegacyKeccak256()
|
||||
hashPermit.Write(abiPermit) //nolint:errcheck,gosec
|
||||
abiEIP712Domain := []byte("EIP712Domain(string name,string version,uint256 chainId,address verifyingContract)")
|
||||
abiEIP712Domain :=
|
||||
[]byte("EIP712Domain(string name,string version,uint256 chainId,address verifyingContract)")
|
||||
hashEIP712Domain := sha3.NewLegacyKeccak256()
|
||||
hashEIP712Domain.Write(abiEIP712Domain) //nolint:errcheck,gosec
|
||||
var encodeBytes []byte
|
||||
@@ -124,7 +130,8 @@ func createPermitDigest(tokenAddr, owner, spender ethCommon.Address, chainID, va
|
||||
return hashBytes2.Sum(nil), nil
|
||||
}
|
||||
|
||||
func createPermit(owner, spender ethCommon.Address, amount, deadline *big.Int, digest, signature []byte) []byte {
|
||||
func createPermit(owner, spender ethCommon.Address, amount, deadline *big.Int, digest,
|
||||
signature []byte) []byte {
|
||||
r := signature[0:32]
|
||||
s := signature[32:64]
|
||||
v := signature[64] + byte(27) //nolint:gomnd
|
||||
|
||||
@@ -26,7 +26,8 @@ var (
|
||||
mnemonic = "explain tackle mirror kit van hammer degree position ginger unfair soup bonus"
|
||||
)
|
||||
|
||||
func genAcc(w *hdwallet.Wallet, ks *keystore.KeyStore, i int) (*accounts.Account, ethCommon.Address) {
|
||||
func genAcc(w *hdwallet.Wallet, ks *keystore.KeyStore, i int) (*accounts.Account,
|
||||
ethCommon.Address) {
|
||||
path := hdwallet.MustParseDerivationPath(fmt.Sprintf("m/44'/60'/0'/0/%d", i))
|
||||
account, err := w.Derive(path, false)
|
||||
if err != nil {
|
||||
@@ -111,7 +112,9 @@ func getEnvVariables() {
|
||||
if err != nil {
|
||||
log.Fatal(errEnvVar)
|
||||
}
|
||||
if auctionAddressStr == "" || auctionTestAddressStr == "" || tokenHEZAddressStr == "" || hermezRollupAddressStr == "" || wdelayerAddressStr == "" || wdelayerTestAddressStr == "" || genesisBlockEnv == "" {
|
||||
if auctionAddressStr == "" || auctionTestAddressStr == "" || tokenHEZAddressStr == "" ||
|
||||
hermezRollupAddressStr == "" || wdelayerAddressStr == "" || wdelayerTestAddressStr == "" ||
|
||||
genesisBlockEnv == "" {
|
||||
log.Fatal(errEnvVar)
|
||||
}
|
||||
|
||||
@@ -189,7 +192,8 @@ func TestMain(m *testing.M) {
|
||||
log.Fatal(err)
|
||||
}
|
||||
|
||||
ethereumClientEmergencyCouncil, err = NewEthereumClient(ethClient, emergencyCouncilAccount, ks, nil)
|
||||
ethereumClientEmergencyCouncil, err = NewEthereumClient(ethClient,
|
||||
emergencyCouncilAccount, ks, nil)
|
||||
if err != nil {
|
||||
log.Fatal(err)
|
||||
}
|
||||
|
||||
243
eth/rollup.go
243
eth/rollup.go
@@ -243,13 +243,20 @@ type RollupInterface interface {
|
||||
// Public Functions
|
||||
|
||||
RollupForgeBatch(*RollupForgeBatchArgs, *bind.TransactOpts) (*types.Transaction, error)
|
||||
RollupAddToken(tokenAddress ethCommon.Address, feeAddToken, deadline *big.Int) (*types.Transaction, error)
|
||||
RollupAddToken(tokenAddress ethCommon.Address, feeAddToken,
|
||||
deadline *big.Int) (*types.Transaction, error)
|
||||
|
||||
RollupWithdrawMerkleProof(babyPubKey babyjub.PublicKeyComp, tokenID uint32, numExitRoot, idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
RollupWithdrawMerkleProof(babyPubKey babyjub.PublicKeyComp, tokenID uint32, numExitRoot,
|
||||
idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (*types.Transaction,
|
||||
error)
|
||||
RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32,
|
||||
numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error)
|
||||
|
||||
RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (*types.Transaction, error)
|
||||
RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64, deadline *big.Int) (tx *types.Transaction, err error)
|
||||
RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int,
|
||||
amount *big.Int, tokenID uint32, toIdx int64) (*types.Transaction, error)
|
||||
RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64,
|
||||
deadline *big.Int) (tx *types.Transaction, err error)
|
||||
|
||||
// Governance Public Functions
|
||||
RollupUpdateForgeL1L2BatchTimeout(newForgeL1L2BatchTimeout int64) (*types.Transaction, error)
|
||||
@@ -264,7 +271,7 @@ type RollupInterface interface {
|
||||
//
|
||||
|
||||
RollupConstants() (*common.RollupConstants, error)
|
||||
RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethCommon.Hash, error)
|
||||
RollupEventsByBlock(blockNum int64, blockHash *ethCommon.Hash) (*RollupEvents, error)
|
||||
RollupForgeBatchArgs(ethCommon.Hash, uint16) (*RollupForgeBatchArgs, *ethCommon.Address, error)
|
||||
RollupEventInit() (*RollupEventInitialize, int64, error)
|
||||
}
|
||||
@@ -287,7 +294,8 @@ type RollupClient struct {
|
||||
}
|
||||
|
||||
// NewRollupClient creates a new RollupClient
|
||||
func NewRollupClient(client *EthereumClient, address ethCommon.Address, tokenHEZCfg TokenConfig) (*RollupClient, error) {
|
||||
func NewRollupClient(client *EthereumClient, address ethCommon.Address,
|
||||
tokenHEZCfg TokenConfig) (*RollupClient, error) {
|
||||
contractAbi, err := abi.JSON(strings.NewReader(string(Hermez.HermezABI)))
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -316,18 +324,19 @@ func NewRollupClient(client *EthereumClient, address ethCommon.Address, tokenHEZ
|
||||
}
|
||||
consts, err := c.RollupConstants()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(fmt.Errorf("RollupConstants at %v: %w", address, err))
|
||||
}
|
||||
c.consts = consts
|
||||
return c, nil
|
||||
}
|
||||
|
||||
// RollupForgeBatch is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs, auth *bind.TransactOpts) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs,
|
||||
auth *bind.TransactOpts) (tx *types.Transaction, err error) {
|
||||
if auth == nil {
|
||||
auth, err = c.client.NewAuth()
|
||||
if err != nil {
|
||||
return nil, err
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
auth.GasLimit = 1000000
|
||||
}
|
||||
@@ -393,7 +402,7 @@ func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs, auth *bind.T
|
||||
l1CoordinatorBytes, l1l2TxData, feeIdxCoordinator, args.VerifierIdx, args.L1Batch,
|
||||
args.ProofA, args.ProofB, args.ProofC)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed Hermez.ForgeBatch: %w", err))
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Hermez.ForgeBatch: %w", err))
|
||||
}
|
||||
return tx, nil
|
||||
}
|
||||
@@ -401,7 +410,8 @@ func (c *RollupClient) RollupForgeBatch(args *RollupForgeBatchArgs, auth *bind.T
|
||||
// RollupAddToken is the interface to call the smart contract function.
|
||||
// `feeAddToken` is the amount of HEZ tokens that will be paid to add the
|
||||
// token. `feeAddToken` must match the public value of the smart contract.
|
||||
func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToken, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToken,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -413,9 +423,11 @@ func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToke
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, feeAddToken, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
feeAddToken, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, feeAddToken, deadline, digest, signature)
|
||||
permit := createPermit(owner, spender, feeAddToken, deadline, digest,
|
||||
signature)
|
||||
|
||||
return c.hermez.AddToken(auth, tokenAddress, permit)
|
||||
},
|
||||
@@ -426,7 +438,9 @@ func (c *RollupClient) RollupAddToken(tokenAddress ethCommon.Address, feeAddToke
|
||||
}
|
||||
|
||||
// RollupWithdrawMerkleProof is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp, tokenID uint32, numExitRoot, idx int64, amount *big.Int, siblings []*big.Int, instantWithdraw bool) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp, tokenID uint32,
|
||||
numExitRoot, idx int64, amount *big.Int, siblings []*big.Int,
|
||||
instantWithdraw bool) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -434,7 +448,8 @@ func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp,
|
||||
babyPubKey := new(big.Int).SetBytes(pkCompB)
|
||||
numExitRootB := uint32(numExitRoot)
|
||||
idxBig := big.NewInt(idx)
|
||||
return c.hermez.WithdrawMerkleProof(auth, tokenID, amount, babyPubKey, numExitRootB, siblings, idxBig, instantWithdraw)
|
||||
return c.hermez.WithdrawMerkleProof(auth, tokenID, amount, babyPubKey,
|
||||
numExitRootB, siblings, idxBig, instantWithdraw)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed update WithdrawMerkleProof: %w", err))
|
||||
@@ -443,13 +458,17 @@ func (c *RollupClient) RollupWithdrawMerkleProof(fromBJJ babyjub.PublicKeyComp,
|
||||
}
|
||||
|
||||
// RollupWithdrawCircuit is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int, tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction, error) {
|
||||
func (c *RollupClient) RollupWithdrawCircuit(proofA, proofC [2]*big.Int, proofB [2][2]*big.Int,
|
||||
tokenID uint32, numExitRoot, idx int64, amount *big.Int, instantWithdraw bool) (*types.Transaction,
|
||||
error) {
|
||||
log.Error("TODO")
|
||||
return nil, tracerr.Wrap(errTODO)
|
||||
}
|
||||
|
||||
// RollupL1UserTxERC20ETH is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64) (tx *types.Transaction,
|
||||
err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -462,11 +481,11 @@ func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fro
|
||||
}
|
||||
fromIdxBig := big.NewInt(fromIdx)
|
||||
toIdxBig := big.NewInt(toIdx)
|
||||
depositAmountF, err := common.NewFloat16(depositAmount)
|
||||
depositAmountF, err := common.NewFloat40(depositAmount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
amountF, err := common.NewFloat16(amount)
|
||||
amountF, err := common.NewFloat40(amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -484,7 +503,9 @@ func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fro
|
||||
}
|
||||
|
||||
// RollupL1UserTxERC20Permit is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64, depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64, deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp, fromIdx int64,
|
||||
depositAmount *big.Int, amount *big.Int, tokenID uint32, toIdx int64,
|
||||
deadline *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -497,11 +518,11 @@ func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp,
|
||||
}
|
||||
fromIdxBig := big.NewInt(fromIdx)
|
||||
toIdxBig := big.NewInt(toIdx)
|
||||
depositAmountF, err := common.NewFloat16(depositAmount)
|
||||
depositAmountF, err := common.NewFloat40(depositAmount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
amountF, err := common.NewFloat16(amount)
|
||||
amountF, err := common.NewFloat40(amount)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -516,11 +537,12 @@ func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp,
|
||||
}
|
||||
tokenName := c.tokenHEZCfg.Name
|
||||
tokenAddr := c.tokenHEZCfg.Address
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID, amount, nonce, deadline, tokenName)
|
||||
digest, _ := createPermitDigest(tokenAddr, owner, spender, c.chainID,
|
||||
amount, nonce, deadline, tokenName)
|
||||
signature, _ := c.client.ks.SignHash(*c.client.account, digest)
|
||||
permit := createPermit(owner, spender, amount, deadline, digest, signature)
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig, uint16(depositAmountF),
|
||||
uint16(amountF), tokenID, toIdxBig, permit)
|
||||
return c.hermez.AddL1Transaction(auth, babyPubKey, fromIdxBig,
|
||||
uint16(depositAmountF), uint16(amountF), tokenID, toIdxBig, permit)
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed add L1 Tx ERC20Permit: %w", err))
|
||||
@@ -552,11 +574,13 @@ func (c *RollupClient) RollupLastForgedBatch() (lastForgedBatch int64, err error
|
||||
}
|
||||
|
||||
// RollupUpdateForgeL1L2BatchTimeout is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(newForgeL1L2BatchTimeout int64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(
|
||||
newForgeL1L2BatchTimeout int64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
return c.hermez.UpdateForgeL1L2BatchTimeout(auth, uint8(newForgeL1L2BatchTimeout))
|
||||
return c.hermez.UpdateForgeL1L2BatchTimeout(auth,
|
||||
uint8(newForgeL1L2BatchTimeout))
|
||||
},
|
||||
); err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Failed update ForgeL1L2BatchTimeout: %w", err))
|
||||
@@ -565,7 +589,8 @@ func (c *RollupClient) RollupUpdateForgeL1L2BatchTimeout(newForgeL1L2BatchTimeou
|
||||
}
|
||||
|
||||
// RollupUpdateFeeAddToken is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateFeeAddToken(newFeeAddToken *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateFeeAddToken(newFeeAddToken *big.Int) (tx *types.Transaction,
|
||||
err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -600,7 +625,8 @@ func (c *RollupClient) RollupUpdateBucketsParameters(
|
||||
}
|
||||
|
||||
// RollupUpdateTokenExchange is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Address, valueArray []uint64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Address,
|
||||
valueArray []uint64) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -613,7 +639,8 @@ func (c *RollupClient) RollupUpdateTokenExchange(addressArray []ethCommon.Addres
|
||||
}
|
||||
|
||||
// RollupUpdateWithdrawalDelay is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupUpdateWithdrawalDelay(newWithdrawalDelay int64) (tx *types.Transaction, err error) {
|
||||
func (c *RollupClient) RollupUpdateWithdrawalDelay(newWithdrawalDelay int64) (tx *types.Transaction,
|
||||
err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -639,7 +666,8 @@ func (c *RollupClient) RollupSafeMode() (tx *types.Transaction, err error) {
|
||||
}
|
||||
|
||||
// RollupInstantWithdrawalViewer is the interface to call the smart contract function
|
||||
func (c *RollupClient) RollupInstantWithdrawalViewer(tokenAddress ethCommon.Address, amount *big.Int) (instantAllowed bool, err error) {
|
||||
func (c *RollupClient) RollupInstantWithdrawalViewer(tokenAddress ethCommon.Address,
|
||||
amount *big.Int) (instantAllowed bool, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
instantAllowed, err = c.hermez.InstantWithdrawalViewer(c.opts, tokenAddress, amount)
|
||||
return tracerr.Wrap(err)
|
||||
@@ -674,7 +702,8 @@ func (c *RollupClient) RollupConstants() (rollupConstants *common.RollupConstant
|
||||
}
|
||||
newRollupVerifier.MaxTx = rollupVerifier.MaxTx.Int64()
|
||||
newRollupVerifier.NLevels = rollupVerifier.NLevels.Int64()
|
||||
rollupConstants.Verifiers = append(rollupConstants.Verifiers, newRollupVerifier)
|
||||
rollupConstants.Verifiers = append(rollupConstants.Verifiers,
|
||||
newRollupVerifier)
|
||||
}
|
||||
rollupConstants.HermezAuctionContract, err = c.hermez.HermezAuctionContract(c.opts)
|
||||
if err != nil {
|
||||
@@ -693,19 +722,30 @@ func (c *RollupClient) RollupConstants() (rollupConstants *common.RollupConstant
|
||||
}
|
||||
|
||||
var (
|
||||
logHermezL1UserTxEvent = crypto.Keccak256Hash([]byte("L1UserTxEvent(uint32,uint8,bytes)"))
|
||||
logHermezAddToken = crypto.Keccak256Hash([]byte("AddToken(address,uint32)"))
|
||||
logHermezForgeBatch = crypto.Keccak256Hash([]byte("ForgeBatch(uint32,uint16)"))
|
||||
logHermezUpdateForgeL1L2BatchTimeout = crypto.Keccak256Hash([]byte("UpdateForgeL1L2BatchTimeout(uint8)"))
|
||||
logHermezUpdateFeeAddToken = crypto.Keccak256Hash([]byte("UpdateFeeAddToken(uint256)"))
|
||||
logHermezWithdrawEvent = crypto.Keccak256Hash([]byte("WithdrawEvent(uint48,uint32,bool)"))
|
||||
logHermezUpdateBucketWithdraw = crypto.Keccak256Hash([]byte("UpdateBucketWithdraw(uint8,uint256,uint256)"))
|
||||
logHermezUpdateWithdrawalDelay = crypto.Keccak256Hash([]byte("UpdateWithdrawalDelay(uint64)"))
|
||||
logHermezUpdateBucketsParameters = crypto.Keccak256Hash([]byte("UpdateBucketsParameters(uint256[4][" +
|
||||
strconv.Itoa(common.RollupConstNumBuckets) + "])"))
|
||||
logHermezUpdateTokenExchange = crypto.Keccak256Hash([]byte("UpdateTokenExchange(address[],uint64[])"))
|
||||
logHermezSafeMode = crypto.Keccak256Hash([]byte("SafeMode()"))
|
||||
logHermezInitialize = crypto.Keccak256Hash([]byte("InitializeHermezEvent(uint8,uint256,uint64)"))
|
||||
logHermezL1UserTxEvent = crypto.Keccak256Hash([]byte(
|
||||
"L1UserTxEvent(uint32,uint8,bytes)"))
|
||||
logHermezAddToken = crypto.Keccak256Hash([]byte(
|
||||
"AddToken(address,uint32)"))
|
||||
logHermezForgeBatch = crypto.Keccak256Hash([]byte(
|
||||
"ForgeBatch(uint32,uint16)"))
|
||||
logHermezUpdateForgeL1L2BatchTimeout = crypto.Keccak256Hash([]byte(
|
||||
"UpdateForgeL1L2BatchTimeout(uint8)"))
|
||||
logHermezUpdateFeeAddToken = crypto.Keccak256Hash([]byte(
|
||||
"UpdateFeeAddToken(uint256)"))
|
||||
logHermezWithdrawEvent = crypto.Keccak256Hash([]byte(
|
||||
"WithdrawEvent(uint48,uint32,bool)"))
|
||||
logHermezUpdateBucketWithdraw = crypto.Keccak256Hash([]byte(
|
||||
"UpdateBucketWithdraw(uint8,uint256,uint256)"))
|
||||
logHermezUpdateWithdrawalDelay = crypto.Keccak256Hash([]byte(
|
||||
"UpdateWithdrawalDelay(uint64)"))
|
||||
logHermezUpdateBucketsParameters = crypto.Keccak256Hash([]byte(
|
||||
"UpdateBucketsParameters(uint256[4][" + strconv.Itoa(common.RollupConstNumBuckets) + "])"))
|
||||
logHermezUpdateTokenExchange = crypto.Keccak256Hash([]byte(
|
||||
"UpdateTokenExchange(address[],uint64[])"))
|
||||
logHermezSafeMode = crypto.Keccak256Hash([]byte(
|
||||
"SafeMode()"))
|
||||
logHermezInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeHermezEvent(uint8,uint256,uint64)"))
|
||||
)
|
||||
|
||||
// RollupEventInit returns the initialize event with its corresponding block number
|
||||
@@ -729,37 +769,47 @@ func (c *RollupClient) RollupEventInit() (*RollupEventInitialize, int64, error)
|
||||
}
|
||||
|
||||
var rollupInit RollupEventInitialize
|
||||
if err := c.contractAbi.UnpackIntoInterface(&rollupInit, "InitializeHermezEvent", vLog.Data); err != nil {
|
||||
if err := c.contractAbi.UnpackIntoInterface(&rollupInit, "InitializeHermezEvent",
|
||||
vLog.Data); err != nil {
|
||||
return nil, 0, tracerr.Wrap(err)
|
||||
}
|
||||
return &rollupInit, int64(vLog.BlockNumber), tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// RollupEventsByBlock returns the events in a block that happened in the Rollup Smart Contract
|
||||
func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethCommon.Hash, error) {
|
||||
// RollupEventsByBlock returns the events in a block that happened in the
|
||||
// Rollup Smart Contract.
|
||||
// To query by blockNum, set blockNum >= 0 and blockHash == nil.
|
||||
// To query by blockHash set blockHash != nil, and blockNum will be ignored.
|
||||
// If there are no events in that block the result is nil.
|
||||
func (c *RollupClient) RollupEventsByBlock(blockNum int64,
|
||||
blockHash *ethCommon.Hash) (*RollupEvents, error) {
|
||||
var rollupEvents RollupEvents
|
||||
var blockHash *ethCommon.Hash
|
||||
|
||||
var blockNumBigInt *big.Int
|
||||
if blockHash == nil {
|
||||
blockNumBigInt = big.NewInt(blockNum)
|
||||
}
|
||||
query := ethereum.FilterQuery{
|
||||
FromBlock: big.NewInt(blockNum),
|
||||
ToBlock: big.NewInt(blockNum),
|
||||
BlockHash: blockHash,
|
||||
FromBlock: blockNumBigInt,
|
||||
ToBlock: blockNumBigInt,
|
||||
Addresses: []ethCommon.Address{
|
||||
c.address,
|
||||
},
|
||||
BlockHash: nil,
|
||||
Topics: [][]ethCommon.Hash{},
|
||||
Topics: [][]ethCommon.Hash{},
|
||||
}
|
||||
logs, err := c.client.client.FilterLogs(context.Background(), query)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if len(logs) > 0 {
|
||||
blockHash = &logs[0].BlockHash
|
||||
if len(logs) == 0 {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
for _, vLog := range logs {
|
||||
if vLog.BlockHash != *blockHash {
|
||||
if blockHash != nil && vLog.BlockHash != *blockHash {
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got", vLog.BlockHash.String())
|
||||
return nil, nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
return nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
}
|
||||
switch vLog.Topics[0] {
|
||||
case logHermezL1UserTxEvent:
|
||||
@@ -767,11 +817,11 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
var L1UserTx RollupEventL1UserTx
|
||||
err := c.contractAbi.UnpackIntoInterface(&L1UserTxAux, "L1UserTxEvent", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
L1Tx, err := common.L1UserTxFromBytes(L1UserTxAux.L1UserTx)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
toForgeL1TxsNum := new(big.Int).SetBytes(vLog.Topics[1][:]).Int64()
|
||||
L1Tx.ToForgeL1TxsNum = &toForgeL1TxsNum
|
||||
@@ -783,7 +833,7 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
var addToken RollupEventAddToken
|
||||
err := c.contractAbi.UnpackIntoInterface(&addToken, "AddToken", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
addToken.TokenAddress = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
rollupEvents.AddToken = append(rollupEvents.AddToken, addToken)
|
||||
@@ -791,7 +841,7 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
var forgeBatch RollupEventForgeBatch
|
||||
err := c.contractAbi.UnpackIntoInterface(&forgeBatch, "ForgeBatch", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
forgeBatch.BatchNum = new(big.Int).SetBytes(vLog.Topics[1][:]).Int64()
|
||||
forgeBatch.EthTxHash = vLog.TxHash
|
||||
@@ -801,9 +851,10 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
var updateForgeL1L2BatchTimeout struct {
|
||||
NewForgeL1L2BatchTimeout uint8
|
||||
}
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateForgeL1L2BatchTimeout, "UpdateForgeL1L2BatchTimeout", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateForgeL1L2BatchTimeout,
|
||||
"UpdateForgeL1L2BatchTimeout", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupEvents.UpdateForgeL1L2BatchTimeout = append(rollupEvents.UpdateForgeL1L2BatchTimeout,
|
||||
RollupEventUpdateForgeL1L2BatchTimeout{
|
||||
@@ -813,7 +864,7 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
var updateFeeAddToken RollupEventUpdateFeeAddToken
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateFeeAddToken, "UpdateFeeAddToken", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupEvents.UpdateFeeAddToken = append(rollupEvents.UpdateFeeAddToken, updateFeeAddToken)
|
||||
case logHermezWithdrawEvent:
|
||||
@@ -829,28 +880,31 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
case logHermezUpdateBucketWithdraw:
|
||||
var updateBucketWithdrawAux rollupEventUpdateBucketWithdrawAux
|
||||
var updateBucketWithdraw RollupEventUpdateBucketWithdraw
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateBucketWithdrawAux, "UpdateBucketWithdraw", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&updateBucketWithdrawAux,
|
||||
"UpdateBucketWithdraw", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
updateBucketWithdraw.Withdrawals = updateBucketWithdrawAux.Withdrawals
|
||||
updateBucketWithdraw.NumBucket = int(new(big.Int).SetBytes(vLog.Topics[1][:]).Int64())
|
||||
updateBucketWithdraw.BlockStamp = new(big.Int).SetBytes(vLog.Topics[2][:]).Int64()
|
||||
rollupEvents.UpdateBucketWithdraw = append(rollupEvents.UpdateBucketWithdraw, updateBucketWithdraw)
|
||||
rollupEvents.UpdateBucketWithdraw =
|
||||
append(rollupEvents.UpdateBucketWithdraw, updateBucketWithdraw)
|
||||
|
||||
case logHermezUpdateWithdrawalDelay:
|
||||
var withdrawalDelay RollupEventUpdateWithdrawalDelay
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdrawalDelay, "UpdateWithdrawalDelay", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupEvents.UpdateWithdrawalDelay = append(rollupEvents.UpdateWithdrawalDelay, withdrawalDelay)
|
||||
case logHermezUpdateBucketsParameters:
|
||||
var bucketsParametersAux rollupEventUpdateBucketsParametersAux
|
||||
var bucketsParameters RollupEventUpdateBucketsParameters
|
||||
err := c.contractAbi.UnpackIntoInterface(&bucketsParametersAux, "UpdateBucketsParameters", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&bucketsParametersAux,
|
||||
"UpdateBucketsParameters", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
for i, bucket := range bucketsParametersAux.ArrayBuckets {
|
||||
bucketsParameters.ArrayBuckets[i].CeilUSD = bucket[0]
|
||||
@@ -858,12 +912,13 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
bucketsParameters.ArrayBuckets[i].BlockWithdrawalRate = bucket[2]
|
||||
bucketsParameters.ArrayBuckets[i].MaxWithdrawals = bucket[3]
|
||||
}
|
||||
rollupEvents.UpdateBucketsParameters = append(rollupEvents.UpdateBucketsParameters, bucketsParameters)
|
||||
rollupEvents.UpdateBucketsParameters =
|
||||
append(rollupEvents.UpdateBucketsParameters, bucketsParameters)
|
||||
case logHermezUpdateTokenExchange:
|
||||
var tokensExchange RollupEventUpdateTokenExchange
|
||||
err := c.contractAbi.UnpackIntoInterface(&tokensExchange, "UpdateTokenExchange", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupEvents.UpdateTokenExchange = append(rollupEvents.UpdateTokenExchange, tokensExchange)
|
||||
case logHermezSafeMode:
|
||||
@@ -885,15 +940,16 @@ func (c *RollupClient) RollupEventsByBlock(blockNum int64) (*RollupEvents, *ethC
|
||||
bucketsParameters)
|
||||
}
|
||||
}
|
||||
return &rollupEvents, blockHash, nil
|
||||
return &rollupEvents, nil
|
||||
}
|
||||
|
||||
// RollupForgeBatchArgs returns the arguments used in a ForgeBatch call in the
|
||||
// Rollup Smart Contract in the given transaction, and the sender address.
|
||||
func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsLen uint16) (*RollupForgeBatchArgs, *ethCommon.Address, error) {
|
||||
func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash,
|
||||
l1UserTxsLen uint16) (*RollupForgeBatchArgs, *ethCommon.Address, error) {
|
||||
tx, _, err := c.client.client.TransactionByHash(context.Background(), ethTxHash)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("TransactionByHash: %w", err))
|
||||
}
|
||||
txData := tx.Data()
|
||||
|
||||
@@ -905,7 +961,8 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
sender, err := c.client.client.TransactionSender(context.Background(), tx, receipt.Logs[0].BlockHash, receipt.Logs[0].Index)
|
||||
sender, err := c.client.client.TransactionSender(context.Background(), tx,
|
||||
receipt.Logs[0].BlockHash, receipt.Logs[0].Index)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -930,9 +987,9 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
||||
FeeIdxCoordinator: []common.Idx{},
|
||||
}
|
||||
nLevels := c.consts.Verifiers[rollupForgeBatchArgs.VerifierIdx].NLevels
|
||||
lenL1L2TxsBytes := int((nLevels/8)*2 + 2 + 1)
|
||||
lenL1L2TxsBytes := int((nLevels/8)*2 + common.Float40BytesLength + 1) //nolint:gomnd
|
||||
numBytesL1TxUser := int(l1UserTxsLen) * lenL1L2TxsBytes
|
||||
numTxsL1Coord := len(aux.EncodedL1CoordinatorTx) / common.L1CoordinatorTxBytesLen
|
||||
numTxsL1Coord := len(aux.EncodedL1CoordinatorTx) / common.RollupConstL1CoordinatorTotalBytes
|
||||
numBytesL1TxCoord := numTxsL1Coord * lenL1L2TxsBytes
|
||||
numBeginL2Tx := numBytesL1TxCoord + numBytesL1TxUser
|
||||
l1UserTxsData := []byte{}
|
||||
@@ -940,7 +997,9 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
||||
l1UserTxsData = aux.L1L2TxsData[:numBytesL1TxUser]
|
||||
}
|
||||
for i := 0; i < int(l1UserTxsLen); i++ {
|
||||
l1Tx, err := common.L1TxFromDataAvailability(l1UserTxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes], uint32(nLevels))
|
||||
l1Tx, err :=
|
||||
common.L1TxFromDataAvailability(l1UserTxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes],
|
||||
uint32(nLevels))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -952,14 +1011,17 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
||||
}
|
||||
numTxsL2 := len(l2TxsData) / lenL1L2TxsBytes
|
||||
for i := 0; i < numTxsL2; i++ {
|
||||
l2Tx, err := common.L2TxFromBytesDataAvailability(l2TxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes], int(nLevels))
|
||||
l2Tx, err :=
|
||||
common.L2TxFromBytesDataAvailability(l2TxsData[i*lenL1L2TxsBytes:(i+1)*lenL1L2TxsBytes],
|
||||
int(nLevels))
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupForgeBatchArgs.L2TxsData = append(rollupForgeBatchArgs.L2TxsData, *l2Tx)
|
||||
}
|
||||
for i := 0; i < numTxsL1Coord; i++ {
|
||||
bytesL1Coordinator := aux.EncodedL1CoordinatorTx[i*common.L1CoordinatorTxBytesLen : (i+1)*common.L1CoordinatorTxBytesLen]
|
||||
bytesL1Coordinator :=
|
||||
aux.EncodedL1CoordinatorTx[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes] //nolint:lll
|
||||
var signature []byte
|
||||
v := bytesL1Coordinator[0]
|
||||
s := bytesL1Coordinator[1:33]
|
||||
@@ -972,24 +1034,29 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
rollupForgeBatchArgs.L1CoordinatorTxs = append(rollupForgeBatchArgs.L1CoordinatorTxs, *l1Tx)
|
||||
rollupForgeBatchArgs.L1CoordinatorTxsAuths = append(rollupForgeBatchArgs.L1CoordinatorTxsAuths, signature)
|
||||
rollupForgeBatchArgs.L1CoordinatorTxsAuths =
|
||||
append(rollupForgeBatchArgs.L1CoordinatorTxsAuths, signature)
|
||||
}
|
||||
lenFeeIdxCoordinatorBytes := int(nLevels / 8) //nolint:gomnd
|
||||
numFeeIdxCoordinator := len(aux.FeeIdxCoordinator) / lenFeeIdxCoordinatorBytes
|
||||
for i := 0; i < numFeeIdxCoordinator; i++ {
|
||||
var paddedFeeIdx [6]byte
|
||||
// TODO: This check is not necessary: the first case will always work. Test it before removing the if.
|
||||
// TODO: This check is not necessary: the first case will always work. Test it
|
||||
// before removing the if.
|
||||
if lenFeeIdxCoordinatorBytes < common.IdxBytesLen {
|
||||
copy(paddedFeeIdx[6-lenFeeIdxCoordinatorBytes:], aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
copy(paddedFeeIdx[6-lenFeeIdxCoordinatorBytes:],
|
||||
aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
} else {
|
||||
copy(paddedFeeIdx[:], aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
copy(paddedFeeIdx[:],
|
||||
aux.FeeIdxCoordinator[i*lenFeeIdxCoordinatorBytes:(i+1)*lenFeeIdxCoordinatorBytes])
|
||||
}
|
||||
feeIdxCoordinator, err := common.IdxFromBytes(paddedFeeIdx[:])
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
if feeIdxCoordinator != common.Idx(0) {
|
||||
rollupForgeBatchArgs.FeeIdxCoordinator = append(rollupForgeBatchArgs.FeeIdxCoordinator, feeIdxCoordinator)
|
||||
rollupForgeBatchArgs.FeeIdxCoordinator =
|
||||
append(rollupForgeBatchArgs.FeeIdxCoordinator, feeIdxCoordinator)
|
||||
}
|
||||
}
|
||||
return &rollupForgeBatchArgs, &sender, nil
|
||||
|
||||
@@ -91,7 +91,7 @@ func TestRollupAddToken(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, tokenHEZAddressConst, rollupEvents.AddToken[0].TokenAddress)
|
||||
@@ -116,7 +116,8 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
minBid.SetString("11000000000000000000", 10)
|
||||
budget := new(big.Int)
|
||||
budget.SetString("45200000000000000000", 10)
|
||||
_, err = auctionClient.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet, maxBid, minBid, deadline)
|
||||
_, err = auctionClient.AuctionMultiBid(budget, currentSlot+4, currentSlot+10, slotSet,
|
||||
maxBid, minBid, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Add Blocks
|
||||
@@ -128,12 +129,18 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
|
||||
// Forge Batch 1
|
||||
args := new(RollupForgeBatchArgs)
|
||||
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
l1CoordinatorBytes, err := hex.DecodeString("1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
||||
// When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
args.FeeIdxCoordinator = []common.Idx{}
|
||||
l1CoordinatorBytes, err := hex.DecodeString(
|
||||
"1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf" +
|
||||
"42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230" +
|
||||
"de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
||||
require.NoError(t, err)
|
||||
numTxsL1 := len(l1CoordinatorBytes) / common.L1CoordinatorTxBytesLen
|
||||
numTxsL1 := len(l1CoordinatorBytes) / common.RollupConstL1CoordinatorTotalBytes
|
||||
for i := 0; i < numTxsL1; i++ {
|
||||
bytesL1Coordinator := l1CoordinatorBytes[i*common.L1CoordinatorTxBytesLen : (i+1)*common.L1CoordinatorTxBytesLen]
|
||||
bytesL1Coordinator :=
|
||||
l1CoordinatorBytes[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*
|
||||
common.RollupConstL1CoordinatorTotalBytes]
|
||||
var signature []byte
|
||||
v := bytesL1Coordinator[0]
|
||||
s := bytesL1Coordinator[1:33]
|
||||
@@ -149,9 +156,12 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
args.L1UserTxs = []common.L1Tx{}
|
||||
args.L2TxsData = []common.L2Tx{}
|
||||
newStateRoot := new(big.Int)
|
||||
newStateRoot.SetString("18317824016047294649053625209337295956588174734569560016974612130063629505228", 10)
|
||||
newStateRoot.SetString(
|
||||
"18317824016047294649053625209337295956588174734569560016974612130063629505228",
|
||||
10)
|
||||
newExitRoot := new(big.Int)
|
||||
bytesNumExitRoot, err := hex.DecodeString("10a89d5fe8d488eda1ba371d633515739933c706c210c604f5bd209180daa43b")
|
||||
bytesNumExitRoot, err := hex.DecodeString(
|
||||
"10a89d5fe8d488eda1ba371d633515739933c706c210c604f5bd209180daa43b")
|
||||
require.NoError(t, err)
|
||||
newExitRoot.SetBytes(bytesNumExitRoot)
|
||||
args.NewLastIdx = int64(300)
|
||||
@@ -174,7 +184,7 @@ func TestRollupForgeBatch(t *testing.T) {
|
||||
|
||||
currentBlockNum, err = rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, int64(1), rollupEvents.ForgeBatch[0].BatchNum)
|
||||
@@ -203,10 +213,11 @@ func TestRollupUpdateForgeL1L2BatchTimeout(t *testing.T) {
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, newForgeL1L2BatchTimeout, rollupEvents.UpdateForgeL1L2BatchTimeout[0].NewForgeL1L2BatchTimeout)
|
||||
assert.Equal(t, newForgeL1L2BatchTimeout,
|
||||
rollupEvents.UpdateForgeL1L2BatchTimeout[0].NewForgeL1L2BatchTimeout)
|
||||
}
|
||||
|
||||
func TestRollupUpdateFeeAddToken(t *testing.T) {
|
||||
@@ -216,7 +227,7 @@ func TestRollupUpdateFeeAddToken(t *testing.T) {
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, newFeeAddToken, rollupEvents.UpdateFeeAddToken[0].NewFeeAddToken)
|
||||
@@ -235,7 +246,7 @@ func TestRollupUpdateBucketsParameters(t *testing.T) {
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
blockStampBucket = currentBlockNum
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, bucketsParameters, rollupEvents.UpdateBucketsParameters[0].ArrayBuckets)
|
||||
}
|
||||
@@ -246,9 +257,10 @@ func TestRollupUpdateWithdrawalDelay(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, newWithdrawalDelay, int64(rollupEvents.UpdateWithdrawalDelay[0].NewWithdrawalDelay))
|
||||
assert.Equal(t, newWithdrawalDelay,
|
||||
int64(rollupEvents.UpdateWithdrawalDelay[0].NewWithdrawalDelay))
|
||||
}
|
||||
|
||||
func TestRollupUpdateTokenExchange(t *testing.T) {
|
||||
@@ -263,7 +275,7 @@ func TestRollupUpdateTokenExchange(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, addressArray, rollupEvents.UpdateTokenExchange[0].AddressArray)
|
||||
assert.Equal(t, valueArray, rollupEvents.UpdateTokenExchange[0].ValueArray)
|
||||
@@ -287,23 +299,26 @@ func TestRollupL1UserTxETHCreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.FromBJJ, rollupEvents.L1UserTx[0].L1UserTx.FromBJJ)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20CreateAccountDeposit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
key := genKeysBjj(1)
|
||||
fromIdxInt64 := int64(0)
|
||||
@@ -319,23 +334,26 @@ func TestRollupL1UserTxERC20CreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.FromBJJ, rollupEvents.L1UserTx[0].L1UserTx.FromBJJ)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitCreateAccountDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
key := genKeysBjj(3)
|
||||
fromIdxInt64 := int64(0)
|
||||
@@ -351,23 +369,26 @@ func TestRollupL1UserTxERC20PermitCreateAccountDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.FromBJJ, rollupEvents.L1UserTx[0].L1UserTx.FromBJJ)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -383,22 +404,25 @@ func TestRollupL1UserTxETHDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20Deposit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -413,22 +437,25 @@ func TestRollupL1UserTxERC20Deposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitDeposit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(0)
|
||||
@@ -442,22 +469,25 @@ func TestRollupL1UserTxERC20PermitDeposit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -473,22 +503,25 @@ func TestRollupL1UserTxETHDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20DepositTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -503,22 +536,25 @@ func TestRollupL1UserTxERC20DepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(259)
|
||||
@@ -533,22 +569,25 @@ func TestRollupL1UserTxERC20PermitDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHCreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -564,22 +603,25 @@ func TestRollupL1UserTxETHCreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20CreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -594,22 +636,25 @@ func TestRollupL1UserTxERC20CreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitCreateAccountDepositTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(259)
|
||||
@@ -624,22 +669,25 @@ func TestRollupL1UserTxERC20PermitCreateAccountDepositTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHForceTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(257)
|
||||
@@ -654,22 +702,25 @@ func TestRollupL1UserTxETHForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20ForceTransfer(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(258)
|
||||
@@ -683,22 +734,25 @@ func TestRollupL1UserTxERC20ForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitForceTransfer(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(259)
|
||||
toIdxInt64 := int64(260)
|
||||
@@ -712,22 +766,25 @@ func TestRollupL1UserTxERC20PermitForceTransfer(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxETHForceExit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(256)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -742,22 +799,25 @@ func TestRollupL1UserTxETHForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDUint32, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20ForceExit(t *testing.T) {
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux2, err := NewRollupClient(ethereumClientAux2, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(257)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -771,22 +831,25 @@ func TestRollupL1UserTxERC20ForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
_, err = rollupClientAux2.RollupL1UserTxERC20ETH(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenHEZID, toIdxInt64)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux2.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupL1UserTxERC20PermitForceExit(t *testing.T) {
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst, tokenHEZ)
|
||||
rollupClientAux, err := NewRollupClient(ethereumClientAux, hermezRollupAddressConst,
|
||||
tokenHEZ)
|
||||
require.NoError(t, err)
|
||||
fromIdxInt64 := int64(258)
|
||||
toIdxInt64 := int64(1)
|
||||
@@ -802,18 +865,20 @@ func TestRollupL1UserTxERC20PermitForceExit(t *testing.T) {
|
||||
}
|
||||
L1UserTxs = append(L1UserTxs, l1Tx)
|
||||
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64, l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
_, err = rollupClientAux.RollupL1UserTxERC20Permit(l1Tx.FromBJJ, fromIdxInt64,
|
||||
l1Tx.DepositAmount, l1Tx.Amount, tokenIDERC777, toIdxInt64, deadline)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
assert.Equal(t, l1Tx.ToIdx, rollupEvents.L1UserTx[0].L1UserTx.ToIdx)
|
||||
assert.Equal(t, l1Tx.DepositAmount, rollupEvents.L1UserTx[0].L1UserTx.DepositAmount)
|
||||
assert.Equal(t, l1Tx.TokenID, rollupEvents.L1UserTx[0].L1UserTx.TokenID)
|
||||
assert.Equal(t, l1Tx.Amount, rollupEvents.L1UserTx[0].L1UserTx.Amount)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address, rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
assert.Equal(t, rollupClientAux.client.account.Address,
|
||||
rollupEvents.L1UserTx[0].L1UserTx.FromEthAddr)
|
||||
}
|
||||
|
||||
func TestRollupForgeBatch2(t *testing.T) {
|
||||
@@ -822,14 +887,15 @@ func TestRollupForgeBatch2(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, int64(2), rollupEvents.ForgeBatch[0].BatchNum)
|
||||
|
||||
// Forge Batch 3
|
||||
args := new(RollupForgeBatchArgs)
|
||||
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
// When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||
args.FeeIdxCoordinator = []common.Idx{}
|
||||
args.L1CoordinatorTxs = argsForge.L1CoordinatorTxs
|
||||
args.L1CoordinatorTxsAuths = argsForge.L1CoordinatorTxsAuths
|
||||
for i := 0; i < len(L1UserTxs); i++ {
|
||||
@@ -837,14 +903,19 @@ func TestRollupForgeBatch2(t *testing.T) {
|
||||
l1UserTx.EffectiveAmount = l1UserTx.Amount
|
||||
l1Bytes, err := l1UserTx.BytesDataAvailability(uint32(nLevels))
|
||||
require.NoError(t, err)
|
||||
l1UserTxDataAvailability, err := common.L1TxFromDataAvailability(l1Bytes, uint32(nLevels))
|
||||
l1UserTxDataAvailability, err := common.L1TxFromDataAvailability(l1Bytes,
|
||||
uint32(nLevels))
|
||||
require.NoError(t, err)
|
||||
args.L1UserTxs = append(args.L1UserTxs, *l1UserTxDataAvailability)
|
||||
}
|
||||
newStateRoot := new(big.Int)
|
||||
newStateRoot.SetString("18317824016047294649053625209337295956588174734569560016974612130063629505228", 10)
|
||||
newStateRoot.SetString(
|
||||
"18317824016047294649053625209337295956588174734569560016974612130063629505228",
|
||||
10)
|
||||
newExitRoot := new(big.Int)
|
||||
newExitRoot.SetString("1114281409737474688393837964161044726766678436313681099613347372031079422302", 10)
|
||||
newExitRoot.SetString(
|
||||
"1114281409737474688393837964161044726766678436313681099613347372031079422302",
|
||||
10)
|
||||
amount := new(big.Int)
|
||||
amount.SetString("79000000", 10)
|
||||
l2Tx := common.L2Tx{
|
||||
@@ -876,7 +947,7 @@ func TestRollupForgeBatch2(t *testing.T) {
|
||||
|
||||
currentBlockNum, err = rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err = rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err = rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, int64(3), rollupEvents.ForgeBatch[0].BatchNum)
|
||||
@@ -904,7 +975,8 @@ func TestRollupWithdrawMerkleProof(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
var pkComp babyjub.PublicKeyComp
|
||||
pkCompBE, err := hex.DecodeString("adc3b754f8da621967b073a787bef8eec7052f2ba712b23af57d98f65beea8b2")
|
||||
pkCompBE, err :=
|
||||
hex.DecodeString("adc3b754f8da621967b073a787bef8eec7052f2ba712b23af57d98f65beea8b2")
|
||||
require.NoError(t, err)
|
||||
pkCompLE := common.SwapEndianness(pkCompBE)
|
||||
copy(pkComp[:], pkCompLE)
|
||||
@@ -914,21 +986,25 @@ func TestRollupWithdrawMerkleProof(t *testing.T) {
|
||||
numExitRoot := int64(3)
|
||||
fromIdx := int64(256)
|
||||
amount, _ := new(big.Int).SetString("20000000000000000000", 10)
|
||||
// siblingBytes0, err := new(big.Int).SetString("19508838618377323910556678335932426220272947530531646682154552299216398748115", 10)
|
||||
// siblingBytes0, err := new(big.Int).SetString(
|
||||
// "19508838618377323910556678335932426220272947530531646682154552299216398748115",
|
||||
// 10)
|
||||
// require.NoError(t, err)
|
||||
// siblingBytes1, err := new(big.Int).SetString("15198806719713909654457742294233381653226080862567104272457668857208564789571", 10)
|
||||
// siblingBytes1, err := new(big.Int).SetString(
|
||||
// "15198806719713909654457742294233381653226080862567104272457668857208564789571", 10)
|
||||
// require.NoError(t, err)
|
||||
var siblings []*big.Int
|
||||
// siblings = append(siblings, siblingBytes0)
|
||||
// siblings = append(siblings, siblingBytes1)
|
||||
instantWithdraw := true
|
||||
|
||||
_, err = rollupClientAux.RollupWithdrawMerkleProof(pkComp, tokenID, numExitRoot, fromIdx, amount, siblings, instantWithdraw)
|
||||
_, err = rollupClientAux.RollupWithdrawMerkleProof(pkComp, tokenID, numExitRoot, fromIdx,
|
||||
amount, siblings, instantWithdraw)
|
||||
require.NoError(t, err)
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
|
||||
assert.Equal(t, uint64(fromIdx), rollupEvents.Withdraw[0].Idx)
|
||||
@@ -951,7 +1027,7 @@ func TestRollupSafeMode(t *testing.T) {
|
||||
|
||||
currentBlockNum, err := rollupClient.client.EthLastBlock()
|
||||
require.NoError(t, err)
|
||||
rollupEvents, _, err := rollupClient.RollupEventsByBlock(currentBlockNum)
|
||||
rollupEvents, err := rollupClient.RollupEventsByBlock(currentBlockNum, nil)
|
||||
require.NoError(t, err)
|
||||
auxEvent := new(RollupEventSafeMode)
|
||||
assert.Equal(t, auxEvent, &rollupEvents.SafeMode[0])
|
||||
|
||||
135
eth/wdelayer.go
135
eth/wdelayer.go
@@ -132,9 +132,10 @@ type WDelayerInterface interface {
|
||||
WDelayerDepositInfo(owner, token ethCommon.Address) (depositInfo DepositState, err error)
|
||||
WDelayerDeposit(onwer, token ethCommon.Address, amount *big.Int) (*types.Transaction, error)
|
||||
WDelayerWithdrawal(owner, token ethCommon.Address) (*types.Transaction, error)
|
||||
WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address, amount *big.Int) (*types.Transaction, error)
|
||||
WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address,
|
||||
amount *big.Int) (*types.Transaction, error)
|
||||
|
||||
WDelayerEventsByBlock(blockNum int64) (*WDelayerEvents, *ethCommon.Hash, error)
|
||||
WDelayerEventsByBlock(blockNum int64, blockHash *ethCommon.Hash) (*WDelayerEvents, error)
|
||||
WDelayerConstants() (*common.WDelayerConstants, error)
|
||||
WDelayerEventInit() (*WDelayerEventInitialize, int64, error)
|
||||
}
|
||||
@@ -143,7 +144,8 @@ type WDelayerInterface interface {
|
||||
// Implementation
|
||||
//
|
||||
|
||||
// WDelayerClient is the implementation of the interface to the WithdrawDelayer Smart Contract in ethereum.
|
||||
// WDelayerClient is the implementation of the interface to the WithdrawDelayer
|
||||
// Smart Contract in ethereum.
|
||||
type WDelayerClient struct {
|
||||
client *EthereumClient
|
||||
address ethCommon.Address
|
||||
@@ -172,7 +174,8 @@ func NewWDelayerClient(client *EthereumClient, address ethCommon.Address) (*WDel
|
||||
}
|
||||
|
||||
// WDelayerGetHermezGovernanceAddress is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (hermezGovernanceAddress *ethCommon.Address, err error) {
|
||||
func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (
|
||||
hermezGovernanceAddress *ethCommon.Address, err error) {
|
||||
var _hermezGovernanceAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_hermezGovernanceAddress, err = c.wdelayer.GetHermezGovernanceAddress(c.opts)
|
||||
@@ -184,7 +187,8 @@ func (c *WDelayerClient) WDelayerGetHermezGovernanceAddress() (hermezGovernanceA
|
||||
}
|
||||
|
||||
// WDelayerTransferGovernance is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerTransferGovernance(newAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerTransferGovernance(newAddress ethCommon.Address) (
|
||||
tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -210,7 +214,8 @@ func (c *WDelayerClient) WDelayerClaimGovernance() (tx *types.Transaction, err e
|
||||
}
|
||||
|
||||
// WDelayerGetEmergencyCouncil is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress *ethCommon.Address, err error) {
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress *ethCommon.Address,
|
||||
err error) {
|
||||
var _emergencyCouncilAddress ethCommon.Address
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_emergencyCouncilAddress, err = c.wdelayer.GetEmergencyCouncil(c.opts)
|
||||
@@ -222,7 +227,8 @@ func (c *WDelayerClient) WDelayerGetEmergencyCouncil() (emergencyCouncilAddress
|
||||
}
|
||||
|
||||
// WDelayerTransferEmergencyCouncil is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerTransferEmergencyCouncil(newAddress ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerTransferEmergencyCouncil(newAddress ethCommon.Address) (
|
||||
tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -271,7 +277,8 @@ func (c *WDelayerClient) WDelayerGetWithdrawalDelay() (withdrawalDelay int64, er
|
||||
}
|
||||
|
||||
// WDelayerGetEmergencyModeStartingTime is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyModeStartingTime() (emergencyModeStartingTime int64, err error) {
|
||||
func (c *WDelayerClient) WDelayerGetEmergencyModeStartingTime() (emergencyModeStartingTime int64,
|
||||
err error) {
|
||||
var _emergencyModeStartingTime uint64
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
_emergencyModeStartingTime, err = c.wdelayer.GetEmergencyModeStartingTime(c.opts)
|
||||
@@ -296,7 +303,8 @@ func (c *WDelayerClient) WDelayerEnableEmergencyMode() (tx *types.Transaction, e
|
||||
}
|
||||
|
||||
// WDelayerChangeWithdrawalDelay is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64) (
|
||||
tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -309,7 +317,8 @@ func (c *WDelayerClient) WDelayerChangeWithdrawalDelay(newWithdrawalDelay uint64
|
||||
}
|
||||
|
||||
// WDelayerDepositInfo is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (depositInfo DepositState, err error) {
|
||||
func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (
|
||||
depositInfo DepositState, err error) {
|
||||
if err := c.client.Call(func(ec *ethclient.Client) error {
|
||||
amount, depositTimestamp, err := c.wdelayer.DepositInfo(c.opts, owner, token)
|
||||
depositInfo.Amount = amount
|
||||
@@ -322,7 +331,8 @@ func (c *WDelayerClient) WDelayerDepositInfo(owner, token ethCommon.Address) (de
|
||||
}
|
||||
|
||||
// WDelayerDeposit is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount *big.Int) (
|
||||
tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -335,7 +345,8 @@ func (c *WDelayerClient) WDelayerDeposit(owner, token ethCommon.Address, amount
|
||||
}
|
||||
|
||||
// WDelayerWithdrawal is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx *types.Transaction,
|
||||
err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -348,7 +359,8 @@ func (c *WDelayerClient) WDelayerWithdrawal(owner, token ethCommon.Address) (tx
|
||||
}
|
||||
|
||||
// WDelayerEscapeHatchWithdrawal is the interface to call the smart contract function
|
||||
func (c *WDelayerClient) WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address, amount *big.Int) (tx *types.Transaction, err error) {
|
||||
func (c *WDelayerClient) WDelayerEscapeHatchWithdrawal(to, token ethCommon.Address,
|
||||
amount *big.Int) (tx *types.Transaction, err error) {
|
||||
if tx, err = c.client.CallAuth(
|
||||
0,
|
||||
func(ec *ethclient.Client, auth *bind.TransactOpts) (*types.Transaction, error) {
|
||||
@@ -384,14 +396,21 @@ func (c *WDelayerClient) WDelayerConstants() (constants *common.WDelayerConstant
|
||||
}
|
||||
|
||||
var (
|
||||
logWDelayerDeposit = crypto.Keccak256Hash([]byte("Deposit(address,address,uint192,uint64)"))
|
||||
logWDelayerWithdraw = crypto.Keccak256Hash([]byte("Withdraw(address,address,uint192)"))
|
||||
logWDelayerEmergencyModeEnabled = crypto.Keccak256Hash([]byte("EmergencyModeEnabled()"))
|
||||
logWDelayerNewWithdrawalDelay = crypto.Keccak256Hash([]byte("NewWithdrawalDelay(uint64)"))
|
||||
logWDelayerEscapeHatchWithdrawal = crypto.Keccak256Hash([]byte("EscapeHatchWithdrawal(address,address,address,uint256)"))
|
||||
logWDelayerNewEmergencyCouncil = crypto.Keccak256Hash([]byte("NewEmergencyCouncil(address)"))
|
||||
logWDelayerNewHermezGovernanceAddress = crypto.Keccak256Hash([]byte("NewHermezGovernanceAddress(address)"))
|
||||
logWDelayerInitialize = crypto.Keccak256Hash([]byte(
|
||||
logWDelayerDeposit = crypto.Keccak256Hash([]byte(
|
||||
"Deposit(address,address,uint192,uint64)"))
|
||||
logWDelayerWithdraw = crypto.Keccak256Hash([]byte(
|
||||
"Withdraw(address,address,uint192)"))
|
||||
logWDelayerEmergencyModeEnabled = crypto.Keccak256Hash([]byte(
|
||||
"EmergencyModeEnabled()"))
|
||||
logWDelayerNewWithdrawalDelay = crypto.Keccak256Hash([]byte(
|
||||
"NewWithdrawalDelay(uint64)"))
|
||||
logWDelayerEscapeHatchWithdrawal = crypto.Keccak256Hash([]byte(
|
||||
"EscapeHatchWithdrawal(address,address,address,uint256)"))
|
||||
logWDelayerNewEmergencyCouncil = crypto.Keccak256Hash([]byte(
|
||||
"NewEmergencyCouncil(address)"))
|
||||
logWDelayerNewHermezGovernanceAddress = crypto.Keccak256Hash([]byte(
|
||||
"NewHermezGovernanceAddress(address)"))
|
||||
logWDelayerInitialize = crypto.Keccak256Hash([]byte(
|
||||
"InitializeWithdrawalDelayerEvent(uint64,address,address)"))
|
||||
)
|
||||
|
||||
@@ -424,40 +443,47 @@ func (c *WDelayerClient) WDelayerEventInit() (*WDelayerEventInitialize, int64, e
|
||||
}
|
||||
|
||||
// WDelayerEventsByBlock returns the events in a block that happened in the
|
||||
// WDelayer Smart Contract and the blockHash where the eents happened. If
|
||||
// there are no events in that block, blockHash is nil.
|
||||
func (c *WDelayerClient) WDelayerEventsByBlock(blockNum int64) (*WDelayerEvents, *ethCommon.Hash, error) {
|
||||
// WDelayer Smart Contract.
|
||||
// To query by blockNum, set blockNum >= 0 and blockHash == nil.
|
||||
// To query by blockHash set blockHash != nil, and blockNum will be ignored.
|
||||
// If there are no events in that block the result is nil.
|
||||
func (c *WDelayerClient) WDelayerEventsByBlock(blockNum int64,
|
||||
blockHash *ethCommon.Hash) (*WDelayerEvents, error) {
|
||||
var wdelayerEvents WDelayerEvents
|
||||
var blockHash *ethCommon.Hash
|
||||
|
||||
var blockNumBigInt *big.Int
|
||||
if blockHash == nil {
|
||||
blockNumBigInt = big.NewInt(blockNum)
|
||||
}
|
||||
query := ethereum.FilterQuery{
|
||||
FromBlock: big.NewInt(blockNum),
|
||||
ToBlock: big.NewInt(blockNum),
|
||||
BlockHash: blockHash,
|
||||
FromBlock: blockNumBigInt,
|
||||
ToBlock: blockNumBigInt,
|
||||
Addresses: []ethCommon.Address{
|
||||
c.address,
|
||||
},
|
||||
BlockHash: nil,
|
||||
Topics: [][]ethCommon.Hash{},
|
||||
Topics: [][]ethCommon.Hash{},
|
||||
}
|
||||
|
||||
logs, err := c.client.client.FilterLogs(context.Background(), query)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if len(logs) > 0 {
|
||||
blockHash = &logs[0].BlockHash
|
||||
if len(logs) == 0 {
|
||||
return nil, nil
|
||||
}
|
||||
|
||||
for _, vLog := range logs {
|
||||
if vLog.BlockHash != *blockHash {
|
||||
if blockHash != nil && vLog.BlockHash != *blockHash {
|
||||
log.Errorw("Block hash mismatch", "expected", blockHash.String(), "got", vLog.BlockHash.String())
|
||||
return nil, nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
return nil, tracerr.Wrap(ErrBlockHashMismatchEvent)
|
||||
}
|
||||
switch vLog.Topics[0] {
|
||||
case logWDelayerDeposit:
|
||||
var deposit WDelayerEventDeposit
|
||||
err := c.contractAbi.UnpackIntoInterface(&deposit, "Deposit", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
deposit.Owner = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
deposit.Token = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
@@ -468,7 +494,7 @@ func (c *WDelayerClient) WDelayerEventsByBlock(blockNum int64) (*WDelayerEvents,
|
||||
var withdraw WDelayerEventWithdraw
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdraw, "Withdraw", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
withdraw.Token = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
withdraw.Owner = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
@@ -476,43 +502,52 @@ func (c *WDelayerClient) WDelayerEventsByBlock(blockNum int64) (*WDelayerEvents,
|
||||
|
||||
case logWDelayerEmergencyModeEnabled:
|
||||
var emergencyModeEnabled WDelayerEventEmergencyModeEnabled
|
||||
wdelayerEvents.EmergencyModeEnabled = append(wdelayerEvents.EmergencyModeEnabled, emergencyModeEnabled)
|
||||
wdelayerEvents.EmergencyModeEnabled =
|
||||
append(wdelayerEvents.EmergencyModeEnabled, emergencyModeEnabled)
|
||||
|
||||
case logWDelayerNewWithdrawalDelay:
|
||||
var withdrawalDelay WDelayerEventNewWithdrawalDelay
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdrawalDelay, "NewWithdrawalDelay", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&withdrawalDelay,
|
||||
"NewWithdrawalDelay", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewWithdrawalDelay = append(wdelayerEvents.NewWithdrawalDelay, withdrawalDelay)
|
||||
wdelayerEvents.NewWithdrawalDelay =
|
||||
append(wdelayerEvents.NewWithdrawalDelay, withdrawalDelay)
|
||||
|
||||
case logWDelayerEscapeHatchWithdrawal:
|
||||
var escapeHatchWithdrawal WDelayerEventEscapeHatchWithdrawal
|
||||
err := c.contractAbi.UnpackIntoInterface(&escapeHatchWithdrawal, "EscapeHatchWithdrawal", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&escapeHatchWithdrawal,
|
||||
"EscapeHatchWithdrawal", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
escapeHatchWithdrawal.Who = ethCommon.BytesToAddress(vLog.Topics[1].Bytes())
|
||||
escapeHatchWithdrawal.To = ethCommon.BytesToAddress(vLog.Topics[2].Bytes())
|
||||
escapeHatchWithdrawal.Token = ethCommon.BytesToAddress(vLog.Topics[3].Bytes())
|
||||
wdelayerEvents.EscapeHatchWithdrawal = append(wdelayerEvents.EscapeHatchWithdrawal, escapeHatchWithdrawal)
|
||||
wdelayerEvents.EscapeHatchWithdrawal =
|
||||
append(wdelayerEvents.EscapeHatchWithdrawal, escapeHatchWithdrawal)
|
||||
|
||||
case logWDelayerNewEmergencyCouncil:
|
||||
var emergencyCouncil WDelayerEventNewEmergencyCouncil
|
||||
err := c.contractAbi.UnpackIntoInterface(&emergencyCouncil, "NewEmergencyCouncil", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&emergencyCouncil,
|
||||
"NewEmergencyCouncil", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewEmergencyCouncil = append(wdelayerEvents.NewEmergencyCouncil, emergencyCouncil)
|
||||
wdelayerEvents.NewEmergencyCouncil =
|
||||
append(wdelayerEvents.NewEmergencyCouncil, emergencyCouncil)
|
||||
|
||||
case logWDelayerNewHermezGovernanceAddress:
|
||||
var governanceAddress WDelayerEventNewHermezGovernanceAddress
|
||||
err := c.contractAbi.UnpackIntoInterface(&governanceAddress, "NewHermezGovernanceAddress", vLog.Data)
|
||||
err := c.contractAbi.UnpackIntoInterface(&governanceAddress,
|
||||
"NewHermezGovernanceAddress", vLog.Data)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
wdelayerEvents.NewHermezGovernanceAddress = append(wdelayerEvents.NewHermezGovernanceAddress, governanceAddress)
|
||||
wdelayerEvents.NewHermezGovernanceAddress =
|
||||
append(wdelayerEvents.NewHermezGovernanceAddress, governanceAddress)
|
||||
}
|
||||
}
|
||||
return &wdelayerEvents, blockHash, nil
|
||||
return &wdelayerEvents, nil
|
||||
}
|
||||
|
||||
@@ -52,9 +52,10 @@ func TestWDelayerSetHermezGovernanceAddress(t *testing.T) {
|
||||
assert.Equal(t, &auxAddressConst, auxAddress)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.NewHermezGovernanceAddress[0].NewHermezGovernanceAddress)
|
||||
assert.Equal(t, auxAddressConst,
|
||||
wdelayerEvents.NewHermezGovernanceAddress[0].NewHermezGovernanceAddress)
|
||||
_, err = wdelayerClientAux.WDelayerTransferGovernance(governanceAddressConst)
|
||||
require.Nil(t, err)
|
||||
_, err = wdelayerClientTest.WDelayerClaimGovernance()
|
||||
@@ -68,7 +69,8 @@ func TestWDelayerGetEmergencyCouncil(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestWDelayerSetEmergencyCouncil(t *testing.T) {
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil, wdelayerTestAddressConst)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil,
|
||||
wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
wdelayerClientAux, err := NewWDelayerClient(ethereumClientAux, wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
@@ -81,7 +83,7 @@ func TestWDelayerSetEmergencyCouncil(t *testing.T) {
|
||||
assert.Equal(t, &auxAddressConst, auxAddress)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.NewEmergencyCouncil[0].NewEmergencyCouncil)
|
||||
_, err = wdelayerClientAux.WDelayerTransferEmergencyCouncil(emergencyCouncilAddressConst)
|
||||
@@ -110,7 +112,7 @@ func TestWDelayerChangeWithdrawalDelay(t *testing.T) {
|
||||
assert.Equal(t, newWithdrawalDelay, withdrawalDelay)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, uint64(newWithdrawalDelay), wdelayerEvents.NewWithdrawalDelay[0].WithdrawalDelay)
|
||||
}
|
||||
@@ -124,7 +126,7 @@ func TestWDelayerDeposit(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, amount, wdelayerEvents.Deposit[0].Amount)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.Deposit[0].Owner)
|
||||
@@ -150,7 +152,7 @@ func TestWDelayerWithdrawal(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, amount, wdelayerEvents.Withdraw[0].Amount)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.Withdraw[0].Owner)
|
||||
@@ -166,7 +168,7 @@ func TestWDelayerSecondDeposit(t *testing.T) {
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, amount, wdelayerEvents.Deposit[0].Amount)
|
||||
assert.Equal(t, auxAddressConst, wdelayerEvents.Deposit[0].Owner)
|
||||
@@ -181,7 +183,7 @@ func TestWDelayerEnableEmergencyMode(t *testing.T) {
|
||||
assert.Equal(t, true, emergencyMode)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
auxEvent := new(WDelayerEventEmergencyModeEnabled)
|
||||
assert.Equal(t, auxEvent, &wdelayerEvents.EmergencyModeEnabled[0])
|
||||
@@ -200,17 +202,22 @@ func TestWDelayerGetEmergencyModeStartingTime(t *testing.T) {
|
||||
func TestWDelayerEscapeHatchWithdrawal(t *testing.T) {
|
||||
amount := new(big.Int)
|
||||
amount.SetString("10000000000000000", 10)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil, wdelayerTestAddressConst)
|
||||
wdelayerClientEmergencyCouncil, err := NewWDelayerClient(ethereumClientEmergencyCouncil,
|
||||
wdelayerTestAddressConst)
|
||||
require.Nil(t, err)
|
||||
_, err = wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst, tokenHEZAddressConst, amount)
|
||||
_, err =
|
||||
wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst,
|
||||
tokenHEZAddressConst, amount)
|
||||
require.Contains(t, err.Error(), "NO_MAX_EMERGENCY_MODE_TIME")
|
||||
seconds := maxEmergencyModeTime.Seconds()
|
||||
addTime(seconds, ethClientDialURL)
|
||||
_, err = wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst, tokenHEZAddressConst, amount)
|
||||
_, err =
|
||||
wdelayerClientEmergencyCouncil.WDelayerEscapeHatchWithdrawal(governanceAddressConst,
|
||||
tokenHEZAddressConst, amount)
|
||||
require.Nil(t, err)
|
||||
currentBlockNum, err := wdelayerClientTest.client.EthLastBlock()
|
||||
require.Nil(t, err)
|
||||
wdelayerEvents, _, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum)
|
||||
wdelayerEvents, err := wdelayerClientTest.WDelayerEventsByBlock(currentBlockNum, nil)
|
||||
require.Nil(t, err)
|
||||
assert.Equal(t, tokenHEZAddressConst, wdelayerEvents.EscapeHatchWithdrawal[0].Token)
|
||||
assert.Equal(t, governanceAddressConst, wdelayerEvents.EscapeHatchWithdrawal[0].To)
|
||||
|
||||
6
go.mod
6
go.mod
@@ -11,8 +11,8 @@ require (
|
||||
github.com/gin-gonic/gin v1.5.0
|
||||
github.com/gobuffalo/packr/v2 v2.8.1
|
||||
github.com/hermeznetwork/tracerr v0.3.1-0.20210120162744-5da60b576169
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189
|
||||
github.com/jinzhu/copier v0.0.0-20190924061706-b57f9002281a
|
||||
github.com/jmoiron/sqlx v1.2.1-0.20200615141059-0794cb1f47ee
|
||||
github.com/joho/godotenv v1.3.0
|
||||
@@ -21,6 +21,7 @@ require (
|
||||
github.com/miguelmota/go-ethereum-hdwallet v0.0.0-20200123000308-a60dcd172b4c
|
||||
github.com/mitchellh/copystructure v1.0.0
|
||||
github.com/mitchellh/mapstructure v1.3.0
|
||||
github.com/prometheus/client_golang v1.3.0
|
||||
github.com/rubenv/sql-migrate v0.0.0-20200616145509-8d140a17f351
|
||||
github.com/russross/meddler v1.0.0
|
||||
github.com/stretchr/testify v1.6.1
|
||||
@@ -28,5 +29,6 @@ require (
|
||||
go.uber.org/zap v1.16.0
|
||||
golang.org/x/crypto v0.0.0-20201217014255-9d1352758620
|
||||
golang.org/x/net v0.0.0-20200822124328-c89045814202
|
||||
golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9
|
||||
gopkg.in/go-playground/validator.v9 v9.29.1
|
||||
)
|
||||
|
||||
15
go.sum
15
go.sum
@@ -66,6 +66,7 @@ github.com/aws/aws-sdk-go-v2 v0.18.0/go.mod h1:JWVYvqSMppoMJC0x5wdwiImzgXTI9FuZw
|
||||
github.com/aymerick/raymond v2.0.3-0.20180322193309-b565731e1464+incompatible/go.mod h1:osfaiScAUVup+UC9Nfq76eWqDhXlp+4UYaA8uhTBO6g=
|
||||
github.com/beorn7/perks v0.0.0-20180321164747-3a771d992973/go.mod h1:Dwedo/Wpr24TaqPxmxbtue+5NUziq4I4S80YR8gNf3Q=
|
||||
github.com/beorn7/perks v1.0.0/go.mod h1:KWe93zE9D1o94FZ5RNwFwVgaQK1VOXiVxmqh+CedLV8=
|
||||
github.com/beorn7/perks v1.0.1 h1:VlbKKnNfV8bJzeqoa4cOKqO6bYr3WgKZxO8Z16+hsOM=
|
||||
github.com/beorn7/perks v1.0.1/go.mod h1:G2ZrVWU2WbWT9wwq4/hrbKbnv/1ERSJQ0ibhJ6rlkpw=
|
||||
github.com/bgentry/speakeasy v0.1.0/go.mod h1:+zsyZBPWlz7T6j88CTgSN5bM796AkVf0kBD4zp0CCIs=
|
||||
github.com/btcsuite/btcd v0.0.0-20171128150713-2e60448ffcc6/go.mod h1:Dmm/EzmjnCiweXmzRIAiUWCInVmPgjkzgv5k4tVyXiQ=
|
||||
@@ -329,11 +330,10 @@ github.com/huin/goupnp v1.0.0 h1:wg75sLpL6DZqwHQN6E1Cfk6mtfzS45z8OV+ic+DtHRo=
|
||||
github.com/huin/goupnp v1.0.0/go.mod h1:n9v9KO1tAxYH82qOn+UTIFQDmx5n1Zxd/ClZDMX7Bnc=
|
||||
github.com/huin/goutil v0.0.0-20170803182201-1ca381bf3150/go.mod h1:PpLOETDnJ0o3iZrZfqZzyLl6l7F3c6L1oWn7OICBi6o=
|
||||
github.com/hydrogen18/memlistener v0.0.0-20141126152155-54553eb933fb/go.mod h1:qEIFzExnS6016fRpRfxrExeVn2gbClQA99gQhnIcdhE=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201218111145-a2015adb2f1b/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c h1:D2u8FFYey6iFXLsqqJZ8R7ch8gZum+/b98whvoSDbyg=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20201221160344-58e589b6eb4c/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12 h1:DXWT0BLCSm7cJmTMQy7+iOlxkA1/5ADglufhLK52e10=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210119155851-bb53e6ad1a12/go.mod h1:FdUFTW2qJiwHyy5R70uErwq7Kaq1uskyFdTfodcUJqA=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef h1:72PG9b2eDlLqKszJVLrsoJbpt4CtgJLhKOjH1MJqCVY=
|
||||
github.com/iden3/go-iden3-crypto v0.0.6-0.20210308142348-8f85683b2cef/go.mod h1:oBgthFLboAWi9feaBUFy7OxEcyn9vA1khHSL/WwWFyg=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189 h1:hoarWk/SwNwnMXE0kiskcZULW0XBLQIUGYK4C39ozfs=
|
||||
github.com/iden3/go-merkletree v0.0.0-20210308143313-8b63ca866189/go.mod h1:56abMeBKD4BIFe346rk+yuJ4MQgfMHe28sRx4o2gOpk=
|
||||
github.com/imkira/go-interpol v1.1.0/go.mod h1:z0h2/2T3XF8kyEPpRgJ3kmNv+C43p+I/CoI+jC3w2iA=
|
||||
github.com/inconshreveable/mousetrap v1.0.0/go.mod h1:PxqpIevigyE2G7u3NXJIT2ANytuPF1OarO4DADm73n8=
|
||||
github.com/influxdata/influxdb v1.2.3-0.20180221223340-01288bdb0883/go.mod h1:qZna6X/4elxqT3yI9iZYdZrWWdeFOOprn86kgg4+IzY=
|
||||
@@ -438,6 +438,7 @@ github.com/mattn/go-sqlite3 v1.12.0/go.mod h1:FPy6KqzDD04eiIsT53CuJW3U88zkxoIYsO
|
||||
github.com/mattn/go-sqlite3 v2.0.3+incompatible h1:gXHsfypPkaMZrKbD5209QV9jbUTJKjyR5WD3HYQSd+U=
|
||||
github.com/mattn/go-sqlite3 v2.0.3+incompatible/go.mod h1:FPy6KqzDD04eiIsT53CuJW3U88zkxoIYsOqkbpncsNc=
|
||||
github.com/mattn/goveralls v0.0.2/go.mod h1:8d1ZMHsd7fW6IRPKQh46F2WRpyib5/X4FOpevwGNQEw=
|
||||
github.com/matttproud/golang_protobuf_extensions v1.0.1 h1:4hp9jkHxhMHkqkrB3Ix0jegS5sx/RkqARlsWZ6pIwiU=
|
||||
github.com/matttproud/golang_protobuf_extensions v1.0.1/go.mod h1:D8He9yQNgCq6Z5Ld7szi9bcBfOoFv/3dc6xSMkL2PC0=
|
||||
github.com/mediocregopher/mediocre-go-lib v0.0.0-20181029021733-cb65787f37ed/go.mod h1:dSsfyI2zABAdhcbvkXqgxOxrCsbYeHCPgrZkku60dSg=
|
||||
github.com/mediocregopher/radix/v3 v3.3.0/go.mod h1:EmfVyvspXz1uZEyPBMyGK+kjWiKQGvsUt6O3Pj+LDCQ=
|
||||
@@ -538,23 +539,27 @@ github.com/prometheus/client_golang v0.9.3-0.20190127221311-3c4408c8b829/go.mod
|
||||
github.com/prometheus/client_golang v0.9.3/go.mod h1:/TN21ttK/J9q6uSwhBd54HahCDft0ttaMvbicHlPoso=
|
||||
github.com/prometheus/client_golang v1.0.0/go.mod h1:db9x61etRT2tGnBNRi70OPL5FsnadC4Ky3P0J6CfImo=
|
||||
github.com/prometheus/client_golang v1.1.0/go.mod h1:I1FGZT9+L76gKKOs5djB6ezCbFQP1xR9D75/vuwEF3g=
|
||||
github.com/prometheus/client_golang v1.3.0 h1:miYCvYqFXtl/J9FIy8eNpBfYthAEFg+Ys0XyUVEcDsc=
|
||||
github.com/prometheus/client_golang v1.3.0/go.mod h1:hJaj2vgQTGQmVCsAACORcieXFeDPbaTKGT+JTgUa3og=
|
||||
github.com/prometheus/client_model v0.0.0-20180712105110-5c3871d89910/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
||||
github.com/prometheus/client_model v0.0.0-20190115171406-56726106282f/go.mod h1:MbSGuTsp3dbXC40dX6PRTWyKYBIrTGTE9sqQNg2J8bo=
|
||||
github.com/prometheus/client_model v0.0.0-20190129233127-fd36f4220a90/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||
github.com/prometheus/client_model v0.0.0-20190812154241-14fe0d1b01d4/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||
github.com/prometheus/client_model v0.1.0 h1:ElTg5tNp4DqfV7UQjDqv2+RJlNzsDtvNAWccbItceIE=
|
||||
github.com/prometheus/client_model v0.1.0/go.mod h1:xMI15A0UPsDsEKsMN9yxemIoYk6Tm2C1GtYGdfGttqA=
|
||||
github.com/prometheus/common v0.0.0-20181113130724-41aa239b4cce/go.mod h1:daVV7qP5qjZbuso7PdcryaAu0sAZbrN9i7WWcTMWvro=
|
||||
github.com/prometheus/common v0.2.0/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4=
|
||||
github.com/prometheus/common v0.4.0/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4=
|
||||
github.com/prometheus/common v0.4.1/go.mod h1:TNfzLD0ON7rHzMJeJkieUDPYmFC7Snx/y86RQel1bk4=
|
||||
github.com/prometheus/common v0.6.0/go.mod h1:eBmuwkDJBwy6iBfxCBob6t6dR6ENT/y+J+Zk0j9GMYc=
|
||||
github.com/prometheus/common v0.7.0 h1:L+1lyG48J1zAQXA3RBX/nG/B3gjlHq0zTt2tlbJLyCY=
|
||||
github.com/prometheus/common v0.7.0/go.mod h1:DjGbpBbp5NYNiECxcL/VnbXCCaQpKd3tt26CguLLsqA=
|
||||
github.com/prometheus/procfs v0.0.0-20181005140218-185b4288413d/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
||||
github.com/prometheus/procfs v0.0.0-20190117184657-bf6a532e95b1/go.mod h1:c3At6R/oaqEKCNdg8wHV1ftS6bRYblBhIjjI8uT2IGk=
|
||||
github.com/prometheus/procfs v0.0.0-20190507164030-5867b95ac084/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsTZCD3I8kEA=
|
||||
github.com/prometheus/procfs v0.0.2/go.mod h1:TjEm7ze935MbeOT/UhFTIMYKhuLP4wbCsTZCD3I8kEA=
|
||||
github.com/prometheus/procfs v0.0.3/go.mod h1:4A/X28fw3Fc593LaREMrKMqOKvUAntwMDaekg4FpcdQ=
|
||||
github.com/prometheus/procfs v0.0.8 h1:+fpWZdT24pJBiqJdAwYBjPSk+5YmQzYNPYzQsdzLkt8=
|
||||
github.com/prometheus/procfs v0.0.8/go.mod h1:7Qr8sr6344vo1JqZ6HhLceV9o3AJ1Ff+GxbHq6oeK9A=
|
||||
github.com/prometheus/tsdb v0.6.2-0.20190402121629-4f204dcbc150/go.mod h1:qhTCs0VvXwvX/y3TZrWD7rabWM+ijKTux40TwIPHuXU=
|
||||
github.com/prometheus/tsdb v0.7.1 h1:YZcsG11NqnK4czYLrWd9mpEuAJIHVQLwdrleYfszMAA=
|
||||
|
||||
@@ -67,6 +67,11 @@ func Init(levelStr string, outputs []string) {
|
||||
|
||||
func sprintStackTrace(st []tracerr.Frame) string {
|
||||
builder := strings.Builder{}
|
||||
// Skip deepest frame because it belongs to the go runtime and we don't
|
||||
// care about it.
|
||||
if len(st) > 0 {
|
||||
st = st[:len(st)-1]
|
||||
}
|
||||
for _, f := range st {
|
||||
builder.WriteString(fmt.Sprintf("\n%s:%d %s()", f.Path, f.Line, f.Func))
|
||||
}
|
||||
|
||||
223
node/node.go
223
node/node.go
@@ -2,7 +2,9 @@ package node
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"fmt"
|
||||
"net"
|
||||
"net/http"
|
||||
"sync"
|
||||
"time"
|
||||
@@ -61,30 +63,57 @@ type Node struct {
|
||||
sync *synchronizer.Synchronizer
|
||||
|
||||
// General
|
||||
cfg *config.Node
|
||||
mode Mode
|
||||
sqlConn *sqlx.DB
|
||||
ctx context.Context
|
||||
wg sync.WaitGroup
|
||||
cancel context.CancelFunc
|
||||
cfg *config.Node
|
||||
mode Mode
|
||||
sqlConnRead *sqlx.DB
|
||||
sqlConnWrite *sqlx.DB
|
||||
ctx context.Context
|
||||
wg sync.WaitGroup
|
||||
cancel context.CancelFunc
|
||||
}
|
||||
|
||||
// NewNode creates a Node
|
||||
func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
meddler.Debug = cfg.Debug.MeddlerLogs
|
||||
// Stablish DB connection
|
||||
db, err := dbUtils.InitSQLDB(
|
||||
cfg.PostgreSQL.Port,
|
||||
cfg.PostgreSQL.Host,
|
||||
cfg.PostgreSQL.User,
|
||||
cfg.PostgreSQL.Password,
|
||||
cfg.PostgreSQL.Name,
|
||||
dbWrite, err := dbUtils.InitSQLDB(
|
||||
cfg.PostgreSQL.PortWrite,
|
||||
cfg.PostgreSQL.HostWrite,
|
||||
cfg.PostgreSQL.UserWrite,
|
||||
cfg.PostgreSQL.PasswordWrite,
|
||||
cfg.PostgreSQL.NameWrite,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||
}
|
||||
var dbRead *sqlx.DB
|
||||
if cfg.PostgreSQL.HostRead == "" {
|
||||
dbRead = dbWrite
|
||||
} else if cfg.PostgreSQL.HostRead == cfg.PostgreSQL.HostWrite {
|
||||
return nil, tracerr.Wrap(fmt.Errorf(
|
||||
"PostgreSQL.HostRead and PostgreSQL.HostWrite must be different",
|
||||
))
|
||||
} else {
|
||||
dbRead, err = dbUtils.InitSQLDB(
|
||||
cfg.PostgreSQL.PortRead,
|
||||
cfg.PostgreSQL.HostRead,
|
||||
cfg.PostgreSQL.UserRead,
|
||||
cfg.PostgreSQL.PasswordRead,
|
||||
cfg.PostgreSQL.NameRead,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||
}
|
||||
}
|
||||
var apiConnCon *dbUtils.APIConnectionController
|
||||
if cfg.API.Explorer || mode == ModeCoordinator {
|
||||
apiConnCon = dbUtils.NewAPIConnectionController(
|
||||
cfg.API.MaxSQLConnections,
|
||||
cfg.API.SQLConnectionTimeout.Duration,
|
||||
)
|
||||
}
|
||||
|
||||
historyDB := historydb.NewHistoryDB(db)
|
||||
historyDB := historydb.NewHistoryDB(dbRead, dbWrite, apiConnCon)
|
||||
|
||||
ethClient, err := ethclient.Dial(cfg.Web3.URL)
|
||||
if err != nil {
|
||||
@@ -95,8 +124,8 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
var keyStore *ethKeystore.KeyStore
|
||||
if mode == ModeCoordinator {
|
||||
ethCfg = eth.EthereumConfig{
|
||||
CallGasLimit: cfg.Coordinator.EthClient.CallGasLimit,
|
||||
GasPriceDiv: cfg.Coordinator.EthClient.GasPriceDiv,
|
||||
CallGasLimit: 0, // cfg.Coordinator.EthClient.CallGasLimit,
|
||||
GasPriceDiv: 0, // cfg.Coordinator.EthClient.GasPriceDiv,
|
||||
}
|
||||
|
||||
scryptN := ethKeystore.StandardScryptN
|
||||
@@ -108,6 +137,23 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
keyStore = ethKeystore.NewKeyStore(cfg.Coordinator.EthClient.Keystore.Path,
|
||||
scryptN, scryptP)
|
||||
|
||||
balance, err := ethClient.BalanceAt(context.TODO(), cfg.Coordinator.ForgerAddress, nil)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
minForgeBalance := cfg.Coordinator.MinimumForgeAddressBalance
|
||||
if minForgeBalance != nil && balance.Cmp(minForgeBalance) == -1 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf(
|
||||
"forger account balance is less than cfg.Coordinator.MinimumForgeAddressBalance: %v < %v",
|
||||
balance, minForgeBalance))
|
||||
}
|
||||
log.Infow("forger ethereum account balance",
|
||||
"addr", cfg.Coordinator.ForgerAddress,
|
||||
"balance", balance,
|
||||
"minForgeBalance", minForgeBalance,
|
||||
)
|
||||
|
||||
// Unlock Coordinator ForgerAddr in the keystore to make calls
|
||||
// to ForgeBatch in the smart contract
|
||||
if !keyStore.HasAddress(cfg.Coordinator.ForgerAddress) {
|
||||
@@ -164,13 +210,29 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("cfg.StateDB.Keep = %v < %v, which is unsafe",
|
||||
cfg.StateDB.Keep, safeStateDBKeep))
|
||||
}
|
||||
stateDB, err := statedb.NewStateDB(cfg.StateDB.Path, cfg.StateDB.Keep,
|
||||
statedb.TypeSynchronizer, 32)
|
||||
stateDB, err := statedb.NewStateDB(statedb.Config{
|
||||
Path: cfg.StateDB.Path,
|
||||
Keep: cfg.StateDB.Keep,
|
||||
Type: statedb.TypeSynchronizer,
|
||||
NLevels: statedb.MaxNLevels,
|
||||
})
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
sync, err := synchronizer.NewSynchronizer(client, historyDB, stateDB, synchronizer.Config{
|
||||
var l2DB *l2db.L2DB
|
||||
if mode == ModeCoordinator {
|
||||
l2DB = l2db.NewL2DB(
|
||||
dbRead, dbWrite,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
apiConnCon,
|
||||
)
|
||||
}
|
||||
|
||||
sync, err := synchronizer.NewSynchronizer(client, historyDB, l2DB, stateDB, synchronizer.Config{
|
||||
StatsRefreshPeriod: cfg.Synchronizer.StatsRefreshPeriod.Duration,
|
||||
ChainID: chainIDU16,
|
||||
})
|
||||
@@ -186,15 +248,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
}
|
||||
|
||||
var coord *coordinator.Coordinator
|
||||
var l2DB *l2db.L2DB
|
||||
if mode == ModeCoordinator {
|
||||
l2DB = l2db.NewL2DB(
|
||||
db,
|
||||
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||
cfg.Coordinator.L2DB.MaxTxs,
|
||||
cfg.Coordinator.L2DB.TTL.Duration,
|
||||
)
|
||||
|
||||
// Unlock FeeAccount EthAddr in the keystore to generate the
|
||||
// account creation authorization
|
||||
if !keyStore.HasAddress(cfg.Coordinator.FeeAccount.Address) {
|
||||
@@ -216,14 +270,15 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
if err := auth.Sign(func(msg []byte) ([]byte, error) {
|
||||
return keyStore.SignHash(feeAccount, msg)
|
||||
}, chainIDU16, cfg.SmartContracts.Rollup); err != nil {
|
||||
return nil, err
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
coordAccount := &txselector.CoordAccount{
|
||||
Addr: cfg.Coordinator.FeeAccount.Address,
|
||||
BJJ: cfg.Coordinator.FeeAccount.BJJ,
|
||||
AccountCreationAuth: auth.Signature,
|
||||
}
|
||||
txSelector, err := txselector.NewTxSelector(coordAccount, cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
||||
txSelector, err := txselector.NewTxSelector(coordAccount,
|
||||
cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -232,9 +287,6 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
serverProofs := make([]prover.Client, len(cfg.Coordinator.ServerProofs))
|
||||
for i, serverProofCfg := range cfg.Coordinator.ServerProofs {
|
||||
serverProofs[i] = prover.NewProofServerClient(serverProofCfg.URL,
|
||||
@@ -248,14 +300,37 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
MaxFeeTx: common.RollupConstMaxFeeIdxCoordinator,
|
||||
MaxL1Tx: common.RollupConstMaxL1Tx,
|
||||
}
|
||||
verifierIdx, err := scConsts.Rollup.FindVerifierIdx(
|
||||
cfg.Coordinator.Circuit.MaxTx,
|
||||
cfg.Coordinator.Circuit.NLevels,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
var verifierIdx int
|
||||
if cfg.Coordinator.Debug.RollupVerifierIndex == nil {
|
||||
verifierIdx, err = scConsts.Rollup.FindVerifierIdx(
|
||||
cfg.Coordinator.Circuit.MaxTx,
|
||||
cfg.Coordinator.Circuit.NLevels,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
log.Infow("Found verifier that matches circuit config", "verifierIdx", verifierIdx)
|
||||
} else {
|
||||
verifierIdx = *cfg.Coordinator.Debug.RollupVerifierIndex
|
||||
log.Infow("Using debug verifier index from config", "verifierIdx", verifierIdx)
|
||||
if verifierIdx >= len(scConsts.Rollup.Verifiers) {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("verifierIdx (%v) >= "+
|
||||
"len(scConsts.Rollup.Verifiers) (%v)",
|
||||
verifierIdx, len(scConsts.Rollup.Verifiers)))
|
||||
}
|
||||
verifier := scConsts.Rollup.Verifiers[verifierIdx]
|
||||
if verifier.MaxTx != cfg.Coordinator.Circuit.MaxTx ||
|
||||
verifier.NLevels != cfg.Coordinator.Circuit.NLevels {
|
||||
return nil, tracerr.Wrap(
|
||||
fmt.Errorf("Circuit config and verifier params don't match. "+
|
||||
"circuit.MaxTx = %v, circuit.NLevels = %v, "+
|
||||
"verifier.MaxTx = %v, verifier.NLevels = %v",
|
||||
cfg.Coordinator.Circuit.MaxTx, cfg.Coordinator.Circuit.NLevels,
|
||||
verifier.MaxTx, verifier.NLevels,
|
||||
))
|
||||
}
|
||||
}
|
||||
log.Infow("Found verifier that matches circuit config", "verifierIdx", verifierIdx)
|
||||
|
||||
coord, err = coordinator.NewCoordinator(
|
||||
coordinator.Config{
|
||||
@@ -263,9 +338,16 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
ConfirmBlocks: cfg.Coordinator.ConfirmBlocks,
|
||||
L1BatchTimeoutPerc: cfg.Coordinator.L1BatchTimeoutPerc,
|
||||
ForgeRetryInterval: cfg.Coordinator.ForgeRetryInterval.Duration,
|
||||
ForgeDelay: cfg.Coordinator.ForgeDelay.Duration,
|
||||
ForgeNoTxsDelay: cfg.Coordinator.ForgeNoTxsDelay.Duration,
|
||||
SyncRetryInterval: cfg.Coordinator.SyncRetryInterval.Duration,
|
||||
PurgeByExtDelInterval: cfg.Coordinator.PurgeByExtDelInterval.Duration,
|
||||
EthClientAttempts: cfg.Coordinator.EthClient.Attempts,
|
||||
EthClientAttemptsDelay: cfg.Coordinator.EthClient.AttemptsDelay.Duration,
|
||||
EthNoReuseNonce: cfg.Coordinator.EthClient.NoReuseNonce,
|
||||
EthTxResendTimeout: cfg.Coordinator.EthClient.TxResendTimeout.Duration,
|
||||
MaxGasPrice: cfg.Coordinator.EthClient.MaxGasPrice,
|
||||
GasPriceIncPerc: cfg.Coordinator.EthClient.GasPriceIncPerc,
|
||||
TxManagerCheckInterval: cfg.Coordinator.EthClient.CheckLoopInterval.Duration,
|
||||
DebugBatchPath: cfg.Coordinator.Debug.BatchPath,
|
||||
Purger: coordinator.PurgerCfg{
|
||||
@@ -274,6 +356,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
PurgeBlockDelay: cfg.Coordinator.L2DB.PurgeBlockDelay,
|
||||
InvalidateBlockDelay: cfg.Coordinator.L2DB.InvalidateBlockDelay,
|
||||
},
|
||||
ForgeBatchGasCost: cfg.Coordinator.EthClient.ForgeBatchGasCost,
|
||||
VerifierIdx: uint8(verifierIdx),
|
||||
TxProcessorConfig: txProcessorCfg,
|
||||
},
|
||||
@@ -296,6 +379,11 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
}
|
||||
var nodeAPI *NodeAPI
|
||||
if cfg.API.Address != "" {
|
||||
if cfg.Debug.GinDebugMode {
|
||||
gin.SetMode(gin.DebugMode)
|
||||
} else {
|
||||
gin.SetMode(gin.ReleaseMode)
|
||||
}
|
||||
if cfg.API.UpdateMetricsInterval.Duration == 0 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("invalid cfg.API.UpdateMetricsInterval: %v",
|
||||
cfg.API.UpdateMetricsInterval.Duration))
|
||||
@@ -324,6 +412,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
ChainID: chainIDU16,
|
||||
HermezAddress: cfg.SmartContracts.Rollup,
|
||||
},
|
||||
cfg.Coordinator.ForgeDelay.Duration,
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -350,7 +439,8 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
||||
sync: sync,
|
||||
cfg: cfg,
|
||||
mode: mode,
|
||||
sqlConn: db,
|
||||
sqlConnRead: dbRead,
|
||||
sqlConnWrite: dbWrite,
|
||||
ctx: ctx,
|
||||
cancel: cancel,
|
||||
}, nil
|
||||
@@ -378,6 +468,7 @@ func NewNodeAPI(
|
||||
sdb *statedb.StateDB,
|
||||
l2db *l2db.L2DB,
|
||||
config *api.Config,
|
||||
forgeDelay time.Duration,
|
||||
) (*NodeAPI, error) {
|
||||
engine := gin.Default()
|
||||
engine.NoRoute(handleNoRoute)
|
||||
@@ -386,9 +477,11 @@ func NewNodeAPI(
|
||||
coordinatorEndpoints, explorerEndpoints,
|
||||
engine,
|
||||
hdb,
|
||||
sdb,
|
||||
l2db,
|
||||
config,
|
||||
&api.NodeConfig{
|
||||
ForgeDelay: forgeDelay.Seconds(),
|
||||
},
|
||||
)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
@@ -400,20 +493,24 @@ func NewNodeAPI(
|
||||
}, nil
|
||||
}
|
||||
|
||||
// Run starts the http server of the NodeAPI. To stop it, pass a context with
|
||||
// cancelation.
|
||||
// Run starts the http server of the NodeAPI. To stop it, pass a context
|
||||
// with cancellation.
|
||||
func (a *NodeAPI) Run(ctx context.Context) error {
|
||||
server := &http.Server{
|
||||
Addr: a.addr,
|
||||
Handler: a.engine,
|
||||
// TODO: Figure out best parameters for production
|
||||
ReadTimeout: 30 * time.Second, //nolint:gomnd
|
||||
WriteTimeout: 30 * time.Second, //nolint:gomnd
|
||||
MaxHeaderBytes: 1 << 20, //nolint:gomnd
|
||||
}
|
||||
listener, err := net.Listen("tcp", a.addr)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
log.Infof("NodeAPI is ready at %v", a.addr)
|
||||
go func() {
|
||||
log.Infof("NodeAPI is ready at %v", a.addr)
|
||||
if err := server.ListenAndServe(); err != nil && tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
if err := server.Serve(listener); err != nil &&
|
||||
tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
log.Fatalf("Listen: %s\n", err)
|
||||
}
|
||||
}()
|
||||
@@ -429,8 +526,8 @@ func (a *NodeAPI) Run(ctx context.Context) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, vars synchronizer.SCVariablesPtr,
|
||||
batches []common.BatchData) {
|
||||
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars synchronizer.SCVariablesPtr, batches []common.BatchData) {
|
||||
if n.mode == ModeCoordinator {
|
||||
n.coord.SendMsg(ctx, coordinator.MsgSyncBlock{
|
||||
Stats: *stats,
|
||||
@@ -452,16 +549,21 @@ func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, va
|
||||
if stats.Synced() {
|
||||
if err := n.nodeAPI.api.UpdateNetworkInfo(
|
||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||
common.BatchNum(stats.Eth.LastBatch),
|
||||
common.BatchNum(stats.Eth.LastBatchNum),
|
||||
stats.Sync.Auction.CurrentSlot.SlotNum,
|
||||
); err != nil {
|
||||
log.Errorw("API.UpdateNetworkInfo", "err", err)
|
||||
}
|
||||
} else {
|
||||
n.nodeAPI.api.UpdateNetworkInfoBlock(
|
||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||
)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats, vars synchronizer.SCVariablesPtr) {
|
||||
func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats,
|
||||
vars synchronizer.SCVariablesPtr) {
|
||||
if n.mode == ModeCoordinator {
|
||||
n.coord.SendMsg(ctx, coordinator.MsgSyncReorg{
|
||||
Stats: *stats,
|
||||
@@ -481,8 +583,9 @@ func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats, vars
|
||||
|
||||
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
||||
// don't have to pass it around.
|
||||
func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common.Block, time.Duration, error) {
|
||||
blockData, discarded, err := n.sync.Sync2(ctx, lastBlock)
|
||||
func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common.Block,
|
||||
time.Duration, error) {
|
||||
blockData, discarded, err := n.sync.Sync(ctx, lastBlock)
|
||||
stats := n.sync.Stats()
|
||||
if err != nil {
|
||||
// case: error
|
||||
@@ -538,7 +641,13 @@ func (n *Node) StartSynchronizer() {
|
||||
if n.ctx.Err() != nil {
|
||||
continue
|
||||
}
|
||||
log.Errorw("Synchronizer.Sync", "err", err)
|
||||
if errors.Is(err, eth.ErrBlockHashMismatchEvent) {
|
||||
log.Warnw("Synchronizer.Sync", "err", err)
|
||||
} else if errors.Is(err, synchronizer.ErrUnknownBlock) {
|
||||
log.Warnw("Synchronizer.Sync", "err", err)
|
||||
} else {
|
||||
log.Errorw("Synchronizer.Sync", "err", err)
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -599,6 +708,10 @@ func (n *Node) StartNodeAPI() {
|
||||
|
||||
n.wg.Add(1)
|
||||
go func() {
|
||||
// Do an initial update on startup
|
||||
if err := n.nodeAPI.api.UpdateMetrics(); err != nil {
|
||||
log.Errorw("API.UpdateMetrics", "err", err)
|
||||
}
|
||||
for {
|
||||
select {
|
||||
case <-n.ctx.Done():
|
||||
@@ -615,6 +728,10 @@ func (n *Node) StartNodeAPI() {
|
||||
|
||||
n.wg.Add(1)
|
||||
go func() {
|
||||
// Do an initial update on startup
|
||||
if err := n.nodeAPI.api.UpdateRecommendedFee(); err != nil {
|
||||
log.Errorw("API.UpdateRecommendedFee", "err", err)
|
||||
}
|
||||
for {
|
||||
select {
|
||||
case <-n.ctx.Done():
|
||||
|
||||
@@ -4,9 +4,12 @@ import (
|
||||
"context"
|
||||
"fmt"
|
||||
"net/http"
|
||||
"strings"
|
||||
"time"
|
||||
|
||||
"github.com/dghubble/sling"
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
@@ -23,12 +26,16 @@ type APIType string
|
||||
const (
|
||||
// APITypeBitFinexV2 is the http API used by bitfinex V2
|
||||
APITypeBitFinexV2 APIType = "bitfinexV2"
|
||||
// APITypeCoingeckoV3 is the http API used by copingecko V3
|
||||
APITypeCoingeckoV3 APIType = "coingeckoV3"
|
||||
)
|
||||
|
||||
func (t *APIType) valid() bool {
|
||||
switch *t {
|
||||
case APITypeBitFinexV2:
|
||||
return true
|
||||
case APITypeCoingeckoV3:
|
||||
return true
|
||||
default:
|
||||
return false
|
||||
}
|
||||
@@ -36,23 +43,23 @@ func (t *APIType) valid() bool {
|
||||
|
||||
// PriceUpdater definition
|
||||
type PriceUpdater struct {
|
||||
db *historydb.HistoryDB
|
||||
apiURL string
|
||||
apiType APIType
|
||||
tokenSymbols []string
|
||||
db *historydb.HistoryDB
|
||||
apiURL string
|
||||
apiType APIType
|
||||
tokens []historydb.TokenSymbolAndAddr
|
||||
}
|
||||
|
||||
// NewPriceUpdater is the constructor for the updater
|
||||
func NewPriceUpdater(apiURL string, apiType APIType, db *historydb.HistoryDB) (*PriceUpdater, error) {
|
||||
tokenSymbols := []string{}
|
||||
func NewPriceUpdater(apiURL string, apiType APIType, db *historydb.HistoryDB) (*PriceUpdater,
|
||||
error) {
|
||||
if !apiType.valid() {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Invalid apiType: %v", apiType))
|
||||
}
|
||||
return &PriceUpdater{
|
||||
db: db,
|
||||
apiURL: apiURL,
|
||||
apiType: apiType,
|
||||
tokenSymbols: tokenSymbols,
|
||||
db: db,
|
||||
apiURL: apiURL,
|
||||
apiType: apiType,
|
||||
tokens: []historydb.TokenSymbolAndAddr{},
|
||||
}, nil
|
||||
}
|
||||
|
||||
@@ -73,7 +80,39 @@ func getTokenPriceBitfinex(ctx context.Context, client *sling.Sling,
|
||||
return state[6], nil
|
||||
}
|
||||
|
||||
// UpdatePrices is triggered by the Coordinator, and internally will update the token prices in the db
|
||||
func getTokenPriceCoingecko(ctx context.Context, client *sling.Sling,
|
||||
tokenAddr ethCommon.Address) (float64, error) {
|
||||
responseObject := make(map[string]map[string]float64)
|
||||
var url string
|
||||
var id string
|
||||
if tokenAddr == common.EmptyAddr { // Special case for Ether
|
||||
url = "simple/price?ids=ethereum&vs_currencies=usd"
|
||||
id = "ethereum"
|
||||
} else { // Common case (ERC20)
|
||||
id = strings.ToLower(tokenAddr.String())
|
||||
url = "simple/token_price/ethereum?contract_addresses=" +
|
||||
id + "&vs_currencies=usd"
|
||||
}
|
||||
req, err := client.New().Get(url).Request()
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
res, err := client.Do(req.WithContext(ctx), &responseObject, nil)
|
||||
if err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
if res.StatusCode != http.StatusOK {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("http response is not is %v", res.StatusCode))
|
||||
}
|
||||
price := responseObject[id]["usd"]
|
||||
if price <= 0 {
|
||||
return 0, tracerr.Wrap(fmt.Errorf("price not found for %v", id))
|
||||
}
|
||||
return price, nil
|
||||
}
|
||||
|
||||
// UpdatePrices is triggered by the Coordinator, and internally will update the
|
||||
// token prices in the db
|
||||
func (p *PriceUpdater) UpdatePrices(ctx context.Context) {
|
||||
tr := &http.Transport{
|
||||
MaxIdleConns: defaultMaxIdleConns,
|
||||
@@ -83,33 +122,35 @@ func (p *PriceUpdater) UpdatePrices(ctx context.Context) {
|
||||
httpClient := &http.Client{Transport: tr}
|
||||
client := sling.New().Base(p.apiURL).Client(httpClient)
|
||||
|
||||
for _, tokenSymbol := range p.tokenSymbols {
|
||||
for _, token := range p.tokens {
|
||||
var tokenPrice float64
|
||||
var err error
|
||||
switch p.apiType {
|
||||
case APITypeBitFinexV2:
|
||||
tokenPrice, err = getTokenPriceBitfinex(ctx, client, tokenSymbol)
|
||||
tokenPrice, err = getTokenPriceBitfinex(ctx, client, token.Symbol)
|
||||
case APITypeCoingeckoV3:
|
||||
tokenPrice, err = getTokenPriceCoingecko(ctx, client, token.Addr)
|
||||
}
|
||||
if ctx.Err() != nil {
|
||||
return
|
||||
}
|
||||
if err != nil {
|
||||
log.Warnw("token price not updated (get error)",
|
||||
"err", err, "token", tokenSymbol, "apiType", p.apiType)
|
||||
"err", err, "token", token.Symbol, "apiType", p.apiType)
|
||||
}
|
||||
if err = p.db.UpdateTokenValue(tokenSymbol, tokenPrice); err != nil {
|
||||
if err = p.db.UpdateTokenValue(token.Symbol, tokenPrice); err != nil {
|
||||
log.Errorw("token price not updated (db error)",
|
||||
"err", err, "token", tokenSymbol, "apiType", p.apiType)
|
||||
"err", err, "token", token.Symbol, "apiType", p.apiType)
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// UpdateTokenList get the registered token symbols from HistoryDB
|
||||
func (p *PriceUpdater) UpdateTokenList() error {
|
||||
tokenSymbols, err := p.db.GetTokenSymbols()
|
||||
tokens, err := p.db.GetTokenSymbolsAndAddrs()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
p.tokenSymbols = tokenSymbols
|
||||
p.tokens = tokens
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -2,7 +2,6 @@ package priceupdater
|
||||
|
||||
import (
|
||||
"context"
|
||||
"math/big"
|
||||
"os"
|
||||
"testing"
|
||||
|
||||
@@ -15,29 +14,45 @@ import (
|
||||
"github.com/stretchr/testify/require"
|
||||
)
|
||||
|
||||
func TestPriceUpdater(t *testing.T) {
|
||||
var historyDB *historydb.HistoryDB
|
||||
|
||||
func TestMain(m *testing.M) {
|
||||
// Init DB
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
assert.NoError(t, err)
|
||||
historyDB := historydb.NewHistoryDB(db)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
historyDB = historydb.NewHistoryDB(db, db, nil)
|
||||
// Clean DB
|
||||
test.WipeDB(historyDB.DB())
|
||||
// Populate DB
|
||||
// Gen blocks and add them to DB
|
||||
blocks := test.GenBlocks(1, 2)
|
||||
assert.NoError(t, historyDB.AddBlocks(blocks))
|
||||
err = historyDB.AddBlocks(blocks)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
// Gen tokens and add them to DB
|
||||
tokens := []common.Token{}
|
||||
tokens = append(tokens, common.Token{
|
||||
TokenID: 1,
|
||||
EthBlockNum: blocks[0].Num,
|
||||
EthAddr: ethCommon.BigToAddress(big.NewInt(2)),
|
||||
EthAddr: ethCommon.HexToAddress("0x6b175474e89094c44da98b954eedeac495271d0f"),
|
||||
Name: "DAI",
|
||||
Symbol: "DAI",
|
||||
Decimals: 18,
|
||||
})
|
||||
assert.NoError(t, historyDB.AddTokens(tokens))
|
||||
err = historyDB.AddTokens(tokens)
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
|
||||
result := m.Run()
|
||||
os.Exit(result)
|
||||
}
|
||||
|
||||
func TestPriceUpdaterBitfinex(t *testing.T) {
|
||||
// Init price updater
|
||||
pu, err := NewPriceUpdater("https://api-pub.bitfinex.com/v2/", APITypeBitFinexV2, historyDB)
|
||||
require.NoError(t, err)
|
||||
@@ -45,14 +60,29 @@ func TestPriceUpdater(t *testing.T) {
|
||||
assert.NoError(t, pu.UpdateTokenList())
|
||||
// Update prices
|
||||
pu.UpdatePrices(context.Background())
|
||||
assertTokenHasPriceAndClean(t)
|
||||
}
|
||||
|
||||
func TestPriceUpdaterCoingecko(t *testing.T) {
|
||||
// Init price updater
|
||||
pu, err := NewPriceUpdater("https://api.coingecko.com/api/v3/", APITypeCoingeckoV3, historyDB)
|
||||
require.NoError(t, err)
|
||||
// Update token list
|
||||
assert.NoError(t, pu.UpdateTokenList())
|
||||
// Update prices
|
||||
pu.UpdatePrices(context.Background())
|
||||
assertTokenHasPriceAndClean(t)
|
||||
}
|
||||
|
||||
func assertTokenHasPriceAndClean(t *testing.T) {
|
||||
// Check that prices have been updated
|
||||
limit := uint(10)
|
||||
fetchedTokens, _, err := historyDB.GetTokens(nil, nil, "", nil, &limit, historydb.OrderAsc)
|
||||
fetchedTokens, err := historyDB.GetTokensTest()
|
||||
require.NoError(t, err)
|
||||
// TokenID 0 (ETH) is always on the DB
|
||||
assert.Equal(t, 2, len(fetchedTokens))
|
||||
for _, token := range fetchedTokens {
|
||||
assert.NotNil(t, token.USD)
|
||||
assert.NotNil(t, token.USDUpdate)
|
||||
require.NotNil(t, token.USD)
|
||||
require.NotNil(t, token.USDUpdate)
|
||||
assert.Greater(t, *token.USD, 0.0)
|
||||
}
|
||||
}
|
||||
|
||||
@@ -235,7 +235,7 @@ func (p *ProofServerClient) CalculateProof(ctx context.Context, zkInputs *common
|
||||
return tracerr.Wrap(p.apiInput(ctx, zkInputs))
|
||||
}
|
||||
|
||||
// GetProof retreives the Proof and Public Data (public inputs) from the
|
||||
// GetProof retrieves the Proof and Public Data (public inputs) from the
|
||||
// ServerProof, blocking until the proof is ready.
|
||||
func (p *ProofServerClient) GetProof(ctx context.Context) (*Proof, []*big.Int, error) {
|
||||
if err := p.WaitReady(ctx); err != nil {
|
||||
@@ -256,7 +256,8 @@ func (p *ProofServerClient) GetProof(ctx context.Context) (*Proof, []*big.Int, e
|
||||
}
|
||||
return &proof, pubInputs, nil
|
||||
}
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("status != %v, status = %v", StatusCodeSuccess, status.Status))
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("status != %v, status = %v", StatusCodeSuccess,
|
||||
status.Status))
|
||||
}
|
||||
|
||||
// Cancel cancels any current proof computation
|
||||
@@ -297,7 +298,7 @@ func (p *MockClient) CalculateProof(ctx context.Context, zkInputs *common.ZKInpu
|
||||
return nil
|
||||
}
|
||||
|
||||
// GetProof retreives the Proof from the ServerProof
|
||||
// GetProof retrieves the Proof from the ServerProof
|
||||
func (p *MockClient) GetProof(ctx context.Context) (*Proof, []*big.Int, error) {
|
||||
// Simulate a delay
|
||||
select {
|
||||
|
||||
44
synchronizer/metrics.go
Normal file
44
synchronizer/metrics.go
Normal file
@@ -0,0 +1,44 @@
|
||||
package synchronizer
|
||||
|
||||
import "github.com/prometheus/client_golang/prometheus"
|
||||
|
||||
var (
|
||||
metricReorgsCount = prometheus.NewCounter(
|
||||
prometheus.CounterOpts{
|
||||
Name: "sync_reorgs",
|
||||
Help: "",
|
||||
},
|
||||
)
|
||||
metricSyncedLastBlockNum = prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Name: "sync_synced_last_block_num",
|
||||
Help: "",
|
||||
},
|
||||
)
|
||||
metricEthLastBlockNum = prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Name: "sync_eth_last_block_num",
|
||||
Help: "",
|
||||
},
|
||||
)
|
||||
metricSyncedLastBatchNum = prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Name: "sync_synced_last_batch_num",
|
||||
Help: "",
|
||||
},
|
||||
)
|
||||
metricEthLastBatchNum = prometheus.NewGauge(
|
||||
prometheus.GaugeOpts{
|
||||
Name: "sync_eth_last_batch_num",
|
||||
Help: "",
|
||||
},
|
||||
)
|
||||
)
|
||||
|
||||
func init() {
|
||||
prometheus.MustRegister(metricReorgsCount)
|
||||
prometheus.MustRegister(metricSyncedLastBlockNum)
|
||||
prometheus.MustRegister(metricEthLastBlockNum)
|
||||
prometheus.MustRegister(metricSyncedLastBatchNum)
|
||||
prometheus.MustRegister(metricEthLastBatchNum)
|
||||
}
|
||||
@@ -11,6 +11,7 @@ import (
|
||||
"github.com/ethereum/go-ethereum"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
@@ -18,19 +19,32 @@ import (
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
)
|
||||
|
||||
// Stats of the syncrhonizer
|
||||
const (
|
||||
// errStrUnknownBlock is the string returned by geth when querying an
|
||||
// unknown block
|
||||
errStrUnknownBlock = "unknown block"
|
||||
)
|
||||
|
||||
var (
|
||||
// ErrUnknownBlock is the error returned by the Synchronizer when a
|
||||
// block is queried by hash but the ethereum node doesn't find it due
|
||||
// to it being discarded from a reorg.
|
||||
ErrUnknownBlock = fmt.Errorf("unknown block")
|
||||
)
|
||||
|
||||
// Stats of the synchronizer
|
||||
type Stats struct {
|
||||
Eth struct {
|
||||
RefreshPeriod time.Duration
|
||||
Updated time.Time
|
||||
FirstBlockNum int64
|
||||
LastBlock common.Block
|
||||
LastBatch int64
|
||||
LastBatchNum int64
|
||||
}
|
||||
Sync struct {
|
||||
Updated time.Time
|
||||
LastBlock common.Block
|
||||
LastBatch int64
|
||||
LastBatch common.Batch
|
||||
// LastL1BatchBlock is the last ethereum block in which an
|
||||
// l1Batch was forged
|
||||
LastL1BatchBlock int64
|
||||
@@ -77,13 +91,13 @@ func (s *StatsHolder) UpdateCurrentNextSlot(current *common.Slot, next *common.S
|
||||
}
|
||||
|
||||
// UpdateSync updates the synchronizer stats
|
||||
func (s *StatsHolder) UpdateSync(lastBlock *common.Block, lastBatch *common.BatchNum,
|
||||
func (s *StatsHolder) UpdateSync(lastBlock *common.Block, lastBatch *common.Batch,
|
||||
lastL1BatchBlock *int64, lastForgeL1TxsNum *int64) {
|
||||
now := time.Now()
|
||||
s.rw.Lock()
|
||||
s.Sync.LastBlock = *lastBlock
|
||||
if lastBatch != nil {
|
||||
s.Sync.LastBatch = int64(*lastBatch)
|
||||
s.Sync.LastBatch = *lastBatch
|
||||
}
|
||||
if lastL1BatchBlock != nil {
|
||||
s.Sync.LastL1BatchBlock = *lastL1BatchBlock
|
||||
@@ -105,16 +119,16 @@ func (s *StatsHolder) UpdateEth(ethClient eth.ClientInterface) error {
|
||||
|
||||
lastBlock, err := ethClient.EthBlockByNumber(context.TODO(), -1)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
return tracerr.Wrap(fmt.Errorf("EthBlockByNumber: %w", err))
|
||||
}
|
||||
lastBatch, err := ethClient.RollupLastForgedBatch()
|
||||
lastBatchNum, err := ethClient.RollupLastForgedBatch()
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
return tracerr.Wrap(fmt.Errorf("RollupLastForgedBatch: %w", err))
|
||||
}
|
||||
s.rw.Lock()
|
||||
s.Eth.Updated = now
|
||||
s.Eth.LastBlock = *lastBlock
|
||||
s.Eth.LastBatch = lastBatch
|
||||
s.Eth.LastBatchNum = lastBatchNum
|
||||
s.rw.Unlock()
|
||||
return nil
|
||||
}
|
||||
@@ -139,6 +153,10 @@ func (s *StatsHolder) CopyStats() *Stats {
|
||||
sCopy.Sync.Auction.NextSlot.DefaultSlotBid =
|
||||
common.CopyBigInt(s.Sync.Auction.NextSlot.DefaultSlotBid)
|
||||
}
|
||||
if s.Sync.LastBatch.StateRoot != nil {
|
||||
sCopy.Sync.LastBatch.StateRoot =
|
||||
common.CopyBigInt(s.Sync.LastBatch.StateRoot)
|
||||
}
|
||||
s.rw.RUnlock()
|
||||
return &sCopy
|
||||
}
|
||||
@@ -152,9 +170,9 @@ func (s *StatsHolder) blocksPerc() float64 {
|
||||
float64(s.Eth.LastBlock.Num-(s.Eth.FirstBlockNum-1))
|
||||
}
|
||||
|
||||
func (s *StatsHolder) batchesPerc(batchNum int64) float64 {
|
||||
func (s *StatsHolder) batchesPerc(batchNum common.BatchNum) float64 {
|
||||
return float64(batchNum) * 100.0 /
|
||||
float64(s.Eth.LastBatch)
|
||||
float64(s.Eth.LastBatchNum)
|
||||
}
|
||||
|
||||
// StartBlockNums sets the first block used to start tracking the smart
|
||||
@@ -195,20 +213,22 @@ type Config struct {
|
||||
|
||||
// Synchronizer implements the Synchronizer type
|
||||
type Synchronizer struct {
|
||||
ethClient eth.ClientInterface
|
||||
consts SCConsts
|
||||
historyDB *historydb.HistoryDB
|
||||
stateDB *statedb.StateDB
|
||||
cfg Config
|
||||
initVars SCVariables
|
||||
startBlockNum int64
|
||||
vars SCVariables
|
||||
stats *StatsHolder
|
||||
ethClient eth.ClientInterface
|
||||
consts SCConsts
|
||||
historyDB *historydb.HistoryDB
|
||||
l2DB *l2db.L2DB
|
||||
stateDB *statedb.StateDB
|
||||
cfg Config
|
||||
initVars SCVariables
|
||||
startBlockNum int64
|
||||
vars SCVariables
|
||||
stats *StatsHolder
|
||||
resetStateFailed bool
|
||||
}
|
||||
|
||||
// NewSynchronizer creates a new Synchronizer
|
||||
func NewSynchronizer(ethClient eth.ClientInterface, historyDB *historydb.HistoryDB,
|
||||
stateDB *statedb.StateDB, cfg Config) (*Synchronizer, error) {
|
||||
l2DB *l2db.L2DB, stateDB *statedb.StateDB, cfg Config) (*Synchronizer, error) {
|
||||
auctionConstants, err := ethClient.AuctionConstants()
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("NewSynchronizer ethClient.AuctionConstants(): %w",
|
||||
@@ -253,6 +273,7 @@ func NewSynchronizer(ethClient eth.ClientInterface, historyDB *historydb.History
|
||||
ethClient: ethClient,
|
||||
consts: consts,
|
||||
historyDB: historyDB,
|
||||
l2DB: l2DB,
|
||||
stateDB: stateDB,
|
||||
cfg: cfg,
|
||||
initVars: *initVars,
|
||||
@@ -328,23 +349,25 @@ func (s *Synchronizer) setSlotCoordinator(slot *common.Slot) error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// firstBatchBlockNum is the blockNum of first batch in that block, if any
|
||||
func (s *Synchronizer) getCurrentSlot(reset bool, firstBatchBlockNum *int64) (*common.Slot, error) {
|
||||
slot := common.Slot{
|
||||
SlotNum: s.stats.Sync.Auction.CurrentSlot.SlotNum,
|
||||
ForgerCommitment: s.stats.Sync.Auction.CurrentSlot.ForgerCommitment,
|
||||
}
|
||||
// updateCurrentSlot updates the slot with information of the current slot.
|
||||
// The information about which coordinator is allowed to forge is only updated
|
||||
// when we are Synced.
|
||||
// hasBatch is true when the last synced block contained at least one batch.
|
||||
func (s *Synchronizer) updateCurrentSlot(slot *common.Slot, reset bool, hasBatch bool) error {
|
||||
// We want the next block because the current one is already mined
|
||||
blockNum := s.stats.Sync.LastBlock.Num + 1
|
||||
slotNum := s.consts.Auction.SlotNum(blockNum)
|
||||
firstBatchBlockNum := s.stats.Sync.LastBlock.Num
|
||||
if reset {
|
||||
// Using this query only to know if there
|
||||
dbFirstBatchBlockNum, err := s.historyDB.GetFirstBatchBlockNumBySlot(slotNum)
|
||||
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("historyDB.GetFirstBatchBySlot: %w", err))
|
||||
return tracerr.Wrap(fmt.Errorf("historyDB.GetFirstBatchBySlot: %w", err))
|
||||
} else if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
firstBatchBlockNum = nil
|
||||
hasBatch = false
|
||||
} else {
|
||||
firstBatchBlockNum = &dbFirstBatchBlockNum
|
||||
hasBatch = true
|
||||
firstBatchBlockNum = dbFirstBatchBlockNum
|
||||
}
|
||||
slot.ForgerCommitment = false
|
||||
} else if slotNum > slot.SlotNum {
|
||||
@@ -355,11 +378,11 @@ func (s *Synchronizer) getCurrentSlot(reset bool, firstBatchBlockNum *int64) (*c
|
||||
slot.StartBlock, slot.EndBlock = s.consts.Auction.SlotBlocks(slot.SlotNum)
|
||||
// If Synced, update the current coordinator
|
||||
if s.stats.Synced() && blockNum >= s.consts.Auction.GenesisBlockNum {
|
||||
if err := s.setSlotCoordinator(&slot); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
if err := s.setSlotCoordinator(slot); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
if firstBatchBlockNum != nil &&
|
||||
s.consts.Auction.RelativeBlock(*firstBatchBlockNum) <
|
||||
if hasBatch &&
|
||||
s.consts.Auction.RelativeBlock(firstBatchBlockNum) <
|
||||
int64(s.vars.Auction.SlotDeadline) {
|
||||
slot.ForgerCommitment = true
|
||||
}
|
||||
@@ -368,57 +391,61 @@ func (s *Synchronizer) getCurrentSlot(reset bool, firstBatchBlockNum *int64) (*c
|
||||
// BEGIN SANITY CHECK
|
||||
canForge, err := s.ethClient.AuctionCanForge(slot.Forger, blockNum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return tracerr.Wrap(fmt.Errorf("AuctionCanForge: %w", err))
|
||||
}
|
||||
if !canForge {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Synchronized value of forger address for closed slot "+
|
||||
return tracerr.Wrap(fmt.Errorf("Synchronized value of forger address for closed slot "+
|
||||
"differs from smart contract: %+v", slot))
|
||||
}
|
||||
// END SANITY CHECK
|
||||
}
|
||||
return &slot, nil
|
||||
return nil
|
||||
}
|
||||
|
||||
func (s *Synchronizer) getNextSlot() (*common.Slot, error) {
|
||||
// updateNextSlot updates the slot with information of the next slot.
|
||||
// The information about which coordinator is allowed to forge is only updated
|
||||
// when we are Synced.
|
||||
func (s *Synchronizer) updateNextSlot(slot *common.Slot) error {
|
||||
// We want the next block because the current one is already mined
|
||||
blockNum := s.stats.Sync.LastBlock.Num + 1
|
||||
slotNum := s.consts.Auction.SlotNum(blockNum) + 1
|
||||
slot := common.Slot{
|
||||
SlotNum: slotNum,
|
||||
ForgerCommitment: false,
|
||||
}
|
||||
slot.SlotNum = slotNum
|
||||
slot.ForgerCommitment = false
|
||||
slot.StartBlock, slot.EndBlock = s.consts.Auction.SlotBlocks(slot.SlotNum)
|
||||
// If Synced, update the current coordinator
|
||||
if s.stats.Synced() && blockNum >= s.consts.Auction.GenesisBlockNum {
|
||||
if err := s.setSlotCoordinator(&slot); err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
if err := s.setSlotCoordinator(slot); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
// TODO: Remove this SANITY CHECK once this code is tested enough
|
||||
// BEGIN SANITY CHECK
|
||||
canForge, err := s.ethClient.AuctionCanForge(slot.Forger, slot.StartBlock)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return tracerr.Wrap(fmt.Errorf("AuctionCanForge: %w", err))
|
||||
}
|
||||
if !canForge {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("Synchronized value of forger address for closed slot "+
|
||||
return tracerr.Wrap(fmt.Errorf("Synchronized value of forger address for closed slot "+
|
||||
"differs from smart contract: %+v", slot))
|
||||
}
|
||||
// END SANITY CHECK
|
||||
}
|
||||
return &slot, nil
|
||||
return nil
|
||||
}
|
||||
|
||||
func (s *Synchronizer) updateCurrentNextSlotIfSync(reset bool, firstBatchBlockNum *int64) error {
|
||||
current, err := s.getCurrentSlot(reset, firstBatchBlockNum)
|
||||
if err != nil {
|
||||
// updateCurrentNextSlotIfSync updates the current and next slot. Information
|
||||
// about forger address that is allowed to forge is only updated if we are
|
||||
// Synced.
|
||||
func (s *Synchronizer) updateCurrentNextSlotIfSync(reset bool, hasBatch bool) error {
|
||||
current := s.stats.Sync.Auction.CurrentSlot
|
||||
next := s.stats.Sync.Auction.NextSlot
|
||||
if err := s.updateCurrentSlot(¤t, reset, hasBatch); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
next, err := s.getNextSlot()
|
||||
if err != nil {
|
||||
if err := s.updateNextSlot(&next); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
s.stats.UpdateCurrentNextSlot(current, next)
|
||||
s.stats.UpdateCurrentNextSlot(¤t, &next)
|
||||
return nil
|
||||
}
|
||||
|
||||
@@ -445,8 +472,10 @@ func (s *Synchronizer) init() error {
|
||||
lastBlock = lastSavedBlock
|
||||
}
|
||||
if err := s.resetState(lastBlock); err != nil {
|
||||
s.resetStateFailed = true
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
s.resetStateFailed = false
|
||||
|
||||
log.Infow("Sync init block",
|
||||
"syncLastBlock", s.stats.Sync.LastBlock,
|
||||
@@ -455,23 +484,44 @@ func (s *Synchronizer) init() error {
|
||||
"ethLastBlock", s.stats.Eth.LastBlock,
|
||||
)
|
||||
log.Infow("Sync init batch",
|
||||
"syncLastBatch", s.stats.Sync.LastBatch,
|
||||
"syncBatchesPerc", s.stats.batchesPerc(s.stats.Sync.LastBatch),
|
||||
"ethLastBatch", s.stats.Eth.LastBatch,
|
||||
"syncLastBatch", s.stats.Sync.LastBatch.BatchNum,
|
||||
"syncBatchesPerc", s.stats.batchesPerc(s.stats.Sync.LastBatch.BatchNum),
|
||||
"ethLastBatch", s.stats.Eth.LastBatchNum,
|
||||
)
|
||||
return nil
|
||||
}
|
||||
|
||||
// Sync2 attems to synchronize an ethereum block starting from lastSavedBlock.
|
||||
func (s *Synchronizer) resetIntermediateState() error {
|
||||
lastBlock, err := s.historyDB.GetLastBlock()
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
lastBlock = &common.Block{}
|
||||
} else if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("historyDB.GetLastBlock: %w", err))
|
||||
}
|
||||
if err := s.resetState(lastBlock); err != nil {
|
||||
s.resetStateFailed = true
|
||||
return tracerr.Wrap(fmt.Errorf("resetState at block %v: %w", lastBlock.Num, err))
|
||||
}
|
||||
s.resetStateFailed = false
|
||||
return nil
|
||||
}
|
||||
|
||||
// Sync attempts to synchronize an ethereum block starting from lastSavedBlock.
|
||||
// If lastSavedBlock is nil, the lastSavedBlock value is obtained from de DB.
|
||||
// If a block is synched, it will be returned and also stored in the DB. If a
|
||||
// If a block is synced, it will be returned and also stored in the DB. If a
|
||||
// reorg is detected, the number of discarded blocks will be returned and no
|
||||
// synchronization will be made.
|
||||
// TODO: Be smart about locking: only lock during the read/write operations
|
||||
func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block) (*common.BlockData, *int64, error) {
|
||||
func (s *Synchronizer) Sync(ctx context.Context,
|
||||
lastSavedBlock *common.Block) (blockData *common.BlockData, discarded *int64, err error) {
|
||||
if s.resetStateFailed {
|
||||
if err := s.resetIntermediateState(); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
|
||||
var nextBlockNum int64 // next block number to sync
|
||||
if lastSavedBlock == nil {
|
||||
var err error
|
||||
// Get lastSavedBlock from History DB
|
||||
lastSavedBlock, err = s.historyDB.GetLastBlock()
|
||||
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
||||
@@ -497,7 +547,7 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
||||
if tracerr.Unwrap(err) == ethereum.NotFound {
|
||||
return nil, nil, nil
|
||||
} else if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("EthBlockByNumber: %w", err))
|
||||
}
|
||||
log.Debugf("ethBlock: num: %v, parent: %v, hash: %v",
|
||||
ethBlock.Num, ethBlock.ParentHash.String(), ethBlock.Hash.String())
|
||||
@@ -511,7 +561,7 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
||||
"ethLastBlock", s.stats.Eth.LastBlock,
|
||||
)
|
||||
|
||||
// Check that the obtianed ethBlock.ParentHash == prevEthBlock.Hash; if not, reorg!
|
||||
// Check that the obtained ethBlock.ParentHash == prevEthBlock.Hash; if not, reorg!
|
||||
if lastSavedBlock != nil {
|
||||
if lastSavedBlock.Hash != ethBlock.ParentHash {
|
||||
// Reorg detected
|
||||
@@ -527,6 +577,20 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
||||
}
|
||||
}
|
||||
|
||||
defer func() {
|
||||
// If there was an error during sync, reset to the last block
|
||||
// in the historyDB because the historyDB is written last in
|
||||
// the Sync method and is the source of consistency. This
|
||||
// allows resetting the stateDB in the case a batch was
|
||||
// processed but the historyDB block was not committed due to an
|
||||
// error.
|
||||
if err != nil {
|
||||
if err2 := s.resetIntermediateState(); err2 != nil {
|
||||
log.Errorw("sync revert", "err", err2)
|
||||
}
|
||||
}
|
||||
}()
|
||||
|
||||
// Get data from the rollup contract
|
||||
rollupData, err := s.rollupSync(ethBlock)
|
||||
if err != nil {
|
||||
@@ -564,14 +628,14 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
||||
}
|
||||
|
||||
// Group all the block data into the structs to save into HistoryDB
|
||||
blockData := common.BlockData{
|
||||
blockData = &common.BlockData{
|
||||
Block: *ethBlock,
|
||||
Rollup: *rollupData,
|
||||
Auction: *auctionData,
|
||||
WDelayer: *wDelayerData,
|
||||
}
|
||||
|
||||
err = s.historyDB.AddBlockSCData(&blockData)
|
||||
err = s.historyDB.AddBlockSCData(blockData)
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
@@ -589,31 +653,35 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
||||
}
|
||||
}
|
||||
s.stats.UpdateSync(ethBlock,
|
||||
&rollupData.Batches[batchesLen-1].Batch.BatchNum,
|
||||
&rollupData.Batches[batchesLen-1].Batch,
|
||||
lastL1BatchBlock, lastForgeL1TxsNum)
|
||||
}
|
||||
var firstBatchBlockNum *int64
|
||||
hasBatch := false
|
||||
if len(rollupData.Batches) > 0 {
|
||||
firstBatchBlockNum = &rollupData.Batches[0].Batch.EthBlockNum
|
||||
hasBatch = true
|
||||
}
|
||||
if err := s.updateCurrentNextSlotIfSync(false, firstBatchBlockNum); err != nil {
|
||||
if err = s.updateCurrentNextSlotIfSync(false, hasBatch); err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
for _, batchData := range rollupData.Batches {
|
||||
metricSyncedLastBatchNum.Set(float64(batchData.Batch.BatchNum))
|
||||
metricEthLastBatchNum.Set(float64(s.stats.Eth.LastBatchNum))
|
||||
log.Debugw("Synced batch",
|
||||
"syncLastBatch", batchData.Batch.BatchNum,
|
||||
"syncBatchesPerc", s.stats.batchesPerc(batchData.Batch.BatchNum),
|
||||
"ethLastBatch", s.stats.Eth.LastBatchNum,
|
||||
)
|
||||
}
|
||||
metricSyncedLastBlockNum.Set(float64(s.stats.Sync.LastBlock.Num))
|
||||
metricEthLastBlockNum.Set(float64(s.stats.Eth.LastBlock.Num))
|
||||
log.Debugw("Synced block",
|
||||
"syncLastBlockNum", s.stats.Sync.LastBlock.Num,
|
||||
"syncBlocksPerc", s.stats.blocksPerc(),
|
||||
"ethLastBlockNum", s.stats.Eth.LastBlock.Num,
|
||||
)
|
||||
for _, batchData := range rollupData.Batches {
|
||||
log.Debugw("Synced batch",
|
||||
"syncLastBatch", batchData.Batch.BatchNum,
|
||||
"syncBatchesPerc", s.stats.batchesPerc(int64(batchData.Batch.BatchNum)),
|
||||
"ethLastBatch", s.stats.Eth.LastBatch,
|
||||
)
|
||||
}
|
||||
|
||||
return &blockData, nil, nil
|
||||
return blockData, nil, nil
|
||||
}
|
||||
|
||||
// reorg manages a reorg, updating History and State DB as needed. Keeps
|
||||
@@ -645,14 +713,15 @@ func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
||||
log.Debugw("Discarding blocks", "total", total, "from", uncleBlock.Num, "to", block.Num+1)
|
||||
|
||||
// Set History DB and State DB to the correct state
|
||||
err := s.historyDB.Reorg(block.Num)
|
||||
if err != nil {
|
||||
if err := s.historyDB.Reorg(block.Num); err != nil {
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
|
||||
if err := s.resetState(block); err != nil {
|
||||
s.resetStateFailed = true
|
||||
return 0, tracerr.Wrap(err)
|
||||
}
|
||||
s.resetStateFailed = false
|
||||
|
||||
return block.Num, nil
|
||||
}
|
||||
@@ -661,15 +730,15 @@ func getInitialVariables(ethClient eth.ClientInterface,
|
||||
consts *SCConsts) (*SCVariables, *StartBlockNums, error) {
|
||||
rollupInit, rollupInitBlock, err := ethClient.RollupEventInit()
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("RollupEventInit: %w", err))
|
||||
}
|
||||
auctionInit, auctionInitBlock, err := ethClient.AuctionEventInit()
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("AuctionEventInit: %w", err))
|
||||
}
|
||||
wDelayerInit, wDelayerInitBlock, err := ethClient.WDelayerEventInit()
|
||||
if err != nil {
|
||||
return nil, nil, tracerr.Wrap(err)
|
||||
return nil, nil, tracerr.Wrap(fmt.Errorf("WDelayerEventInit: %w", err))
|
||||
}
|
||||
rollupVars := rollupInit.RollupVariables()
|
||||
auctionVars := auctionInit.AuctionVariables(consts.Auction.InitialMinimalBidding)
|
||||
@@ -714,12 +783,17 @@ func (s *Synchronizer) resetState(block *common.Block) error {
|
||||
s.vars.WDelayer = *wDelayer
|
||||
}
|
||||
|
||||
batchNum, err := s.historyDB.GetLastBatchNum()
|
||||
batch, err := s.historyDB.GetLastBatch()
|
||||
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
||||
return tracerr.Wrap(fmt.Errorf("historyDB.GetLastBatchNum: %w", err))
|
||||
}
|
||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||
batchNum = 0
|
||||
batch = &common.Batch{}
|
||||
}
|
||||
|
||||
err = s.stateDB.Reset(batch.BatchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("stateDB.Reset: %w", err))
|
||||
}
|
||||
|
||||
lastL1BatchBlockNum, err := s.historyDB.GetLastL1BatchBlockNum()
|
||||
@@ -739,20 +813,15 @@ func (s *Synchronizer) resetState(block *common.Block) error {
|
||||
lastForgeL1TxsNum = &n
|
||||
}
|
||||
|
||||
err = s.stateDB.Reset(batchNum)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(fmt.Errorf("stateDB.Reset: %w", err))
|
||||
}
|
||||
s.stats.UpdateSync(block, batch, &lastL1BatchBlockNum, lastForgeL1TxsNum)
|
||||
|
||||
s.stats.UpdateSync(block, &batchNum, &lastL1BatchBlockNum, lastForgeL1TxsNum)
|
||||
|
||||
if err := s.updateCurrentNextSlotIfSync(true, nil); err != nil {
|
||||
if err := s.updateCurrentNextSlotIfSync(true, false); err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
return nil
|
||||
}
|
||||
|
||||
// rollupSync retreives all the Rollup Smart Contract Data that happened at
|
||||
// rollupSync retrieves all the Rollup Smart Contract Data that happened at
|
||||
// ethBlock.blockNum with ethBlock.Hash.
|
||||
func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, error) {
|
||||
blockNum := ethBlock.Num
|
||||
@@ -761,19 +830,16 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
|
||||
// Get rollup events in the block, and make sure the block hash matches
|
||||
// the expected one.
|
||||
rollupEvents, blockHash, err := s.ethClient.RollupEventsByBlock(blockNum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
rollupEvents, err := s.ethClient.RollupEventsByBlock(blockNum, ðBlock.Hash)
|
||||
if err != nil && err.Error() == errStrUnknownBlock {
|
||||
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||
} else if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("RollupEventsByBlock: %w", err))
|
||||
}
|
||||
// No events in this block
|
||||
if blockHash == nil {
|
||||
if rollupEvents == nil {
|
||||
return &rollupData, nil
|
||||
}
|
||||
if *blockHash != ethBlock.Hash {
|
||||
log.Errorw("Block hash mismatch in Rollup events", "expected", ethBlock.Hash.String(),
|
||||
"got", blockHash.String())
|
||||
return nil, tracerr.Wrap(eth.ErrBlockHashMismatchEvent)
|
||||
}
|
||||
|
||||
var nextForgeL1TxsNum int64 // forgeL1TxsNum for the next L1Batch
|
||||
nextForgeL1TxsNumPtr, err := s.historyDB.GetLastL1TxsNum()
|
||||
@@ -801,7 +867,7 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
forgeBatchArgs, sender, err := s.ethClient.RollupForgeBatchArgs(evtForgeBatch.EthTxHash,
|
||||
evtForgeBatch.L1UserTxsLen)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
return nil, tracerr.Wrap(fmt.Errorf("RollupForgeBatchArgs: %w", err))
|
||||
}
|
||||
|
||||
batchNum := common.BatchNum(evtForgeBatch.BatchNum)
|
||||
@@ -832,6 +898,9 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
|
||||
position = len(l1UserTxs)
|
||||
}
|
||||
|
||||
l1TxsAuth := make([]common.AccountCreationAuth,
|
||||
0, len(forgeBatchArgs.L1CoordinatorTxsAuths))
|
||||
// Get L1 Coordinator Txs
|
||||
for i := range forgeBatchArgs.L1CoordinatorTxs {
|
||||
l1CoordinatorTx := forgeBatchArgs.L1CoordinatorTxs[i]
|
||||
@@ -847,9 +916,30 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
|
||||
batchData.L1CoordinatorTxs = append(batchData.L1CoordinatorTxs, *l1Tx)
|
||||
position++
|
||||
|
||||
// Create a slice of account creation auth to be
|
||||
// inserted later if not exists
|
||||
if l1CoordinatorTx.FromEthAddr != common.RollupConstEthAddressInternalOnly {
|
||||
l1CoordinatorTxAuth := forgeBatchArgs.L1CoordinatorTxsAuths[i]
|
||||
l1TxsAuth = append(l1TxsAuth, common.AccountCreationAuth{
|
||||
EthAddr: l1CoordinatorTx.FromEthAddr,
|
||||
BJJ: l1CoordinatorTx.FromBJJ,
|
||||
Signature: l1CoordinatorTxAuth,
|
||||
})
|
||||
}
|
||||
|
||||
// fmt.Println("DGB l1coordtx")
|
||||
}
|
||||
|
||||
// Insert the slice of account creation auth
|
||||
// only if the node run as a coordinator
|
||||
if s.l2DB != nil && len(l1TxsAuth) > 0 {
|
||||
err = s.l2DB.AddManyAccountCreationAuth(l1TxsAuth)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
}
|
||||
|
||||
// Insert all the txs forged in this batch (l1UserTxs,
|
||||
// L1CoordinatorTxs, PoolL2Txs) into stateDB so that they are
|
||||
// processed.
|
||||
@@ -862,7 +952,8 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
}
|
||||
|
||||
// Transform L2 txs to PoolL2Txs
|
||||
poolL2Txs := common.L2TxsToPoolL2Txs(forgeBatchArgs.L2TxsData) // NOTE: This is a big ugly, find a better way
|
||||
// NOTE: This is a big ugly, find a better way
|
||||
poolL2Txs := common.L2TxsToPoolL2Txs(forgeBatchArgs.L2TxsData)
|
||||
|
||||
if int(forgeBatchArgs.VerifierIdx) >= len(s.consts.Rollup.Verifiers) {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("forgeBatchArgs.VerifierIdx (%v) >= "+
|
||||
@@ -884,6 +975,16 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
}
|
||||
if s.stateDB.CurrentBatch() != batchNum {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("stateDB.BatchNum (%v) != "+
|
||||
"evtForgeBatch.BatchNum = (%v)",
|
||||
s.stateDB.CurrentBatch(), batchNum))
|
||||
}
|
||||
if s.stateDB.MT.Root().BigInt().Cmp(forgeBatchArgs.NewStRoot) != 0 {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("stateDB.MTRoot (%v) != "+
|
||||
"forgeBatchArgs.NewStRoot (%v)",
|
||||
s.stateDB.MT.Root().BigInt(), forgeBatchArgs.NewStRoot))
|
||||
}
|
||||
|
||||
// Transform processed PoolL2 txs to L2 and store in BatchData
|
||||
l2Txs, err := common.PoolL2TxsToL2Txs(poolL2Txs) // NOTE: This is a big uggly, find a better way
|
||||
@@ -924,6 +1025,19 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
||||
}
|
||||
batchData.CreatedAccounts = processTxsOut.CreatedAccounts
|
||||
|
||||
batchData.UpdatedAccounts = make([]common.AccountUpdate, 0,
|
||||
len(processTxsOut.UpdatedAccounts))
|
||||
for _, acc := range processTxsOut.UpdatedAccounts {
|
||||
batchData.UpdatedAccounts = append(batchData.UpdatedAccounts,
|
||||
common.AccountUpdate{
|
||||
EthBlockNum: blockNum,
|
||||
BatchNum: batchNum,
|
||||
Idx: acc.Idx,
|
||||
Nonce: acc.Nonce,
|
||||
Balance: acc.Balance,
|
||||
})
|
||||
}
|
||||
|
||||
slotNum := int64(0)
|
||||
if ethBlock.Num >= s.consts.Auction.GenesisBlockNum {
|
||||
slotNum = (ethBlock.Num - s.consts.Auction.GenesisBlockNum) /
|
||||
@@ -1066,19 +1180,16 @@ func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData,
|
||||
var auctionData = common.NewAuctionData()
|
||||
|
||||
// Get auction events in the block
|
||||
auctionEvents, blockHash, err := s.ethClient.AuctionEventsByBlock(blockNum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
auctionEvents, err := s.ethClient.AuctionEventsByBlock(blockNum, ðBlock.Hash)
|
||||
if err != nil && err.Error() == errStrUnknownBlock {
|
||||
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||
} else if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("AuctionEventsByBlock: %w", err))
|
||||
}
|
||||
// No events in this block
|
||||
if blockHash == nil {
|
||||
if auctionEvents == nil {
|
||||
return &auctionData, nil
|
||||
}
|
||||
if *blockHash != ethBlock.Hash {
|
||||
log.Errorw("Block hash mismatch in Auction events", "expected", ethBlock.Hash.String(),
|
||||
"got", blockHash.String())
|
||||
return nil, tracerr.Wrap(eth.ErrBlockHashMismatchEvent)
|
||||
}
|
||||
|
||||
// Get bids
|
||||
for _, evt := range auctionEvents.NewBid {
|
||||
@@ -1168,19 +1279,16 @@ func (s *Synchronizer) wdelayerSync(ethBlock *common.Block) (*common.WDelayerDat
|
||||
wDelayerData := common.NewWDelayerData()
|
||||
|
||||
// Get wDelayer events in the block
|
||||
wDelayerEvents, blockHash, err := s.ethClient.WDelayerEventsByBlock(blockNum)
|
||||
if err != nil {
|
||||
return nil, tracerr.Wrap(err)
|
||||
wDelayerEvents, err := s.ethClient.WDelayerEventsByBlock(blockNum, ðBlock.Hash)
|
||||
if err != nil && err.Error() == errStrUnknownBlock {
|
||||
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||
} else if err != nil {
|
||||
return nil, tracerr.Wrap(fmt.Errorf("WDelayerEventsByBlock: %w", err))
|
||||
}
|
||||
// No events in this block
|
||||
if blockHash == nil {
|
||||
if wDelayerEvents == nil {
|
||||
return &wDelayerData, nil
|
||||
}
|
||||
if *blockHash != ethBlock.Hash {
|
||||
log.Errorw("Block hash mismatch in WDelayer events", "expected", ethBlock.Hash.String(),
|
||||
"got", blockHash.String())
|
||||
return nil, tracerr.Wrap(eth.ErrBlockHashMismatchEvent)
|
||||
}
|
||||
|
||||
for _, evt := range wDelayerEvents.Deposit {
|
||||
wDelayerData.Deposits = append(wDelayerData.Deposits, common.WDelayerTransfer{
|
||||
|
||||
@@ -15,9 +15,9 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/test"
|
||||
"github.com/hermeznetwork/hermez-node/test/til"
|
||||
"github.com/jinzhu/copier"
|
||||
@@ -43,7 +43,8 @@ func accountsCmp(accounts []common.Account) func(i, j int) bool {
|
||||
|
||||
// Check Sync output and HistoryDB state against expected values generated by
|
||||
// til
|
||||
func checkSyncBlock(t *testing.T, s *Synchronizer, blockNum int, block, syncBlock *common.BlockData) {
|
||||
func checkSyncBlock(t *testing.T, s *Synchronizer, blockNum int, block,
|
||||
syncBlock *common.BlockData) {
|
||||
// Check Blocks
|
||||
dbBlocks, err := s.historyDB.GetAllBlocks()
|
||||
require.NoError(t, err)
|
||||
@@ -172,6 +173,8 @@ func checkSyncBlock(t *testing.T, s *Synchronizer, blockNum int, block, syncBloc
|
||||
*exit = syncBatch.ExitTree[j]
|
||||
}
|
||||
assert.Equal(t, batch.Batch, syncBatch.Batch)
|
||||
// Ignore updated accounts
|
||||
syncBatch.UpdatedAccounts = nil
|
||||
assert.Equal(t, batch, syncBatch)
|
||||
assert.Equal(t, &batch.Batch, dbBatch) //nolint:gosec
|
||||
|
||||
@@ -301,24 +304,36 @@ func TestMain(m *testing.M) {
|
||||
os.Exit(exitVal)
|
||||
}
|
||||
|
||||
func newTestModules(t *testing.T) (*statedb.StateDB, *historydb.HistoryDB) {
|
||||
func newTestModules(t *testing.T) (*statedb.StateDB, *historydb.HistoryDB, *l2db.L2DB) {
|
||||
// Int State DB
|
||||
dir, err := ioutil.TempDir("", "tmpdb")
|
||||
require.NoError(t, err)
|
||||
deleteme = append(deleteme, dir)
|
||||
|
||||
stateDB, err := statedb.NewStateDB(dir, 128, statedb.TypeSynchronizer, 32)
|
||||
stateDB, err := statedb.NewStateDB(statedb.Config{Path: dir, Keep: 128,
|
||||
Type: statedb.TypeSynchronizer, NLevels: 32})
|
||||
require.NoError(t, err)
|
||||
|
||||
// Init History DB
|
||||
pass := os.Getenv("POSTGRES_PASS")
|
||||
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||
require.NoError(t, err)
|
||||
historyDB := historydb.NewHistoryDB(db)
|
||||
historyDB := historydb.NewHistoryDB(db, db, nil)
|
||||
// Clear DB
|
||||
test.WipeDB(historyDB.DB())
|
||||
|
||||
return stateDB, historyDB
|
||||
// Init L2 DB
|
||||
l2DB := l2db.NewL2DB(db, db, 10, 100, 0.0, 24*time.Hour, nil)
|
||||
|
||||
return stateDB, historyDB, l2DB
|
||||
}
|
||||
|
||||
func newBigInt(s string) *big.Int {
|
||||
v, ok := new(big.Int).SetString(s, 10)
|
||||
if !ok {
|
||||
panic(fmt.Errorf("Can't set big.Int from %s", s))
|
||||
}
|
||||
return v
|
||||
}
|
||||
|
||||
func TestSyncGeneral(t *testing.T) {
|
||||
@@ -326,7 +341,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
// Setup
|
||||
//
|
||||
|
||||
stateDB, historyDB := newTestModules(t)
|
||||
stateDB, historyDB, l2DB := newTestModules(t)
|
||||
|
||||
// Init eth client
|
||||
var timer timer
|
||||
@@ -336,10 +351,9 @@ func TestSyncGeneral(t *testing.T) {
|
||||
client := test.NewClient(true, &timer, ðCommon.Address{}, clientSetup)
|
||||
|
||||
// Create Synchronizer
|
||||
s, err := NewSynchronizer(client, historyDB, stateDB, Config{
|
||||
s, err := NewSynchronizer(client, historyDB, l2DB, stateDB, Config{
|
||||
StatsRefreshPeriod: 0 * time.Second,
|
||||
})
|
||||
log.Error(err)
|
||||
require.NoError(t, err)
|
||||
|
||||
ctx := context.Background()
|
||||
@@ -351,7 +365,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
assert.Equal(t, false, stats.Synced())
|
||||
|
||||
// Test Sync for rollup genesis block
|
||||
syncBlock, discards, err := s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err := s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.NotNil(t, syncBlock)
|
||||
@@ -374,7 +388,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
assert.Equal(t, int64(1), dbBlocks[1].Num)
|
||||
|
||||
// Sync again and expect no new blocks
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.Nil(t, syncBlock)
|
||||
@@ -434,12 +448,22 @@ func TestSyncGeneral(t *testing.T) {
|
||||
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches[0].L1CoordinatorTxs))
|
||||
// Set StateRoots for batches manually (til doesn't set it)
|
||||
blocks[i].Rollup.Batches[0].Batch.StateRoot =
|
||||
newBigInt("11432094872416618651837327395264042968926668786266585816625577088890451620254")
|
||||
blocks[i].Rollup.Batches[1].Batch.StateRoot =
|
||||
newBigInt("16914212635847451457076355431350059348585556180740555407203882688922702410093")
|
||||
// blocks 1 (blockNum=3)
|
||||
i = 1
|
||||
require.Equal(t, 3, int(blocks[i].Block.Num))
|
||||
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||
require.Equal(t, 3, len(blocks[i].Rollup.Batches[0].L2Txs))
|
||||
// Set StateRoots for batches manually (til doesn't set it)
|
||||
blocks[i].Rollup.Batches[0].Batch.StateRoot =
|
||||
newBigInt("13535760140937349829640752733057594576151546047374619177689224612061148090678")
|
||||
blocks[i].Rollup.Batches[1].Batch.StateRoot =
|
||||
newBigInt("19413739476363469870744893742469056615496274423228302914851564791727474664804")
|
||||
|
||||
// Generate extra required data
|
||||
ethAddTokens(blocks, client)
|
||||
@@ -461,7 +485,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
|
||||
// Block 2
|
||||
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.NotNil(t, syncBlock)
|
||||
@@ -478,7 +502,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
|
||||
// Block 3
|
||||
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
assert.NoError(t, err)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
@@ -496,13 +520,15 @@ func TestSyncGeneral(t *testing.T) {
|
||||
|
||||
// Block 4
|
||||
// Generate 2 withdraws manually
|
||||
_, err = client.RollupWithdrawMerkleProof(tc.Users["A"].BJJ.Public().Compress(), 1, 4, 256, big.NewInt(100), []*big.Int{}, true)
|
||||
_, err = client.RollupWithdrawMerkleProof(tc.Users["A"].BJJ.Public().Compress(), 1, 4, 256,
|
||||
big.NewInt(100), []*big.Int{}, true)
|
||||
require.NoError(t, err)
|
||||
_, err = client.RollupWithdrawMerkleProof(tc.Users["C"].BJJ.Public().Compress(), 1, 3, 258, big.NewInt(50), []*big.Int{}, false)
|
||||
_, err = client.RollupWithdrawMerkleProof(tc.Users["C"].BJJ.Public().Compress(), 1, 3, 258,
|
||||
big.NewInt(50), []*big.Int{}, false)
|
||||
require.NoError(t, err)
|
||||
client.CtlMineBlock()
|
||||
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.NotNil(t, syncBlock)
|
||||
@@ -553,7 +579,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
|
||||
client.CtlMineBlock()
|
||||
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.NotNil(t, syncBlock)
|
||||
@@ -614,6 +640,12 @@ func TestSyncGeneral(t *testing.T) {
|
||||
blocks, err = tc.GenerateBlocks(set2)
|
||||
require.NoError(t, err)
|
||||
|
||||
// Set StateRoots for batches manually (til doesn't set it)
|
||||
blocks[0].Rollup.Batches[0].Batch.StateRoot =
|
||||
newBigInt("14095767774967159269372103336737817266053275274769794195030162905513860477094")
|
||||
blocks[0].Rollup.Batches[1].Batch.StateRoot =
|
||||
newBigInt("2095674348545184674850951945506660952512376416769035169971006930847780339914")
|
||||
|
||||
for i := 0; i < 4; i++ {
|
||||
client.CtlRollback()
|
||||
}
|
||||
@@ -632,7 +664,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
require.NoError(t, err)
|
||||
|
||||
// First sync detects the reorg and discards 4 blocks
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
expetedDiscards := int64(4)
|
||||
require.Equal(t, &expetedDiscards, discards)
|
||||
@@ -660,7 +692,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
|
||||
// Sync blocks 2-6
|
||||
for i := 0; i < 5; i++ {
|
||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err = s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
require.NotNil(t, syncBlock)
|
||||
@@ -699,7 +731,7 @@ func TestSyncGeneral(t *testing.T) {
|
||||
}
|
||||
|
||||
func TestSyncForgerCommitment(t *testing.T) {
|
||||
stateDB, historyDB := newTestModules(t)
|
||||
stateDB, historyDB, l2DB := newTestModules(t)
|
||||
|
||||
// Init eth client
|
||||
var timer timer
|
||||
@@ -712,7 +744,7 @@ func TestSyncForgerCommitment(t *testing.T) {
|
||||
client := test.NewClient(true, &timer, ðCommon.Address{}, clientSetup)
|
||||
|
||||
// Create Synchronizer
|
||||
s, err := NewSynchronizer(client, historyDB, stateDB, Config{
|
||||
s, err := NewSynchronizer(client, historyDB, l2DB, stateDB, Config{
|
||||
StatsRefreshPeriod: 0 * time.Second,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
@@ -783,7 +815,7 @@ func TestSyncForgerCommitment(t *testing.T) {
|
||||
|
||||
// be in sync
|
||||
for {
|
||||
syncBlock, discards, err := s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err := s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
if syncBlock == nil {
|
||||
@@ -795,14 +827,14 @@ func TestSyncForgerCommitment(t *testing.T) {
|
||||
|
||||
// Store ForgerComitmnent observed at every block by the live synchronizer
|
||||
syncCommitment := map[int64]bool{}
|
||||
// Store ForgerComitmnent observed at every block by a syncrhonizer that is restarted
|
||||
// Store ForgerComitmnent observed at every block by a synchronizer that is restarted
|
||||
syncRestartedCommitment := map[int64]bool{}
|
||||
for _, block := range blocks {
|
||||
// Add block data to the smart contracts
|
||||
err = client.CtlAddBlocks([]common.BlockData{block})
|
||||
require.NoError(t, err)
|
||||
|
||||
syncBlock, discards, err := s.Sync2(ctx, nil)
|
||||
syncBlock, discards, err := s.Sync(ctx, nil)
|
||||
require.NoError(t, err)
|
||||
require.Nil(t, discards)
|
||||
if syncBlock == nil {
|
||||
@@ -812,7 +844,7 @@ func TestSyncForgerCommitment(t *testing.T) {
|
||||
require.True(t, stats.Synced())
|
||||
syncCommitment[syncBlock.Block.Num] = stats.Sync.Auction.CurrentSlot.ForgerCommitment
|
||||
|
||||
s2, err := NewSynchronizer(client, historyDB, stateDB, Config{
|
||||
s2, err := NewSynchronizer(client, historyDB, l2DB, stateDB, Config{
|
||||
StatsRefreshPeriod: 0 * time.Second,
|
||||
})
|
||||
require.NoError(t, err)
|
||||
|
||||
@@ -2,6 +2,7 @@ package debugapi
|
||||
|
||||
import (
|
||||
"context"
|
||||
"net"
|
||||
"net/http"
|
||||
"time"
|
||||
|
||||
@@ -12,6 +13,7 @@ import (
|
||||
"github.com/hermeznetwork/hermez-node/log"
|
||||
"github.com/hermeznetwork/hermez-node/synchronizer"
|
||||
"github.com/hermeznetwork/tracerr"
|
||||
"github.com/prometheus/client_golang/prometheus/promhttp"
|
||||
)
|
||||
|
||||
func handleNoRoute(c *gin.Context) {
|
||||
@@ -99,14 +101,16 @@ func (a *DebugAPI) handleSyncStats(c *gin.Context) {
|
||||
c.JSON(http.StatusOK, stats)
|
||||
}
|
||||
|
||||
// Run starts the http server of the DebugAPI. To stop it, pass a context with
|
||||
// cancelation (see `debugapi_test.go` for an example).
|
||||
// Run starts the http server of the DebugAPI. To stop it, pass a context
|
||||
// with cancellation (see `debugapi_test.go` for an example).
|
||||
func (a *DebugAPI) Run(ctx context.Context) error {
|
||||
api := gin.Default()
|
||||
api.NoRoute(handleNoRoute)
|
||||
api.Use(cors.Default())
|
||||
debugAPI := api.Group("/debug")
|
||||
|
||||
debugAPI.GET("/metrics", gin.WrapH(promhttp.Handler()))
|
||||
|
||||
debugAPI.GET("sdb/batchnum", a.handleCurrentBatch)
|
||||
debugAPI.GET("sdb/mtroot", a.handleMTRoot)
|
||||
// Accounts returned by these endpoints will always have BatchNum = 0,
|
||||
@@ -118,16 +122,20 @@ func (a *DebugAPI) Run(ctx context.Context) error {
|
||||
debugAPI.GET("sync/stats", a.handleSyncStats)
|
||||
|
||||
debugAPIServer := &http.Server{
|
||||
Addr: a.addr,
|
||||
Handler: api,
|
||||
// Use some hardcoded numberes that are suitable for testing
|
||||
// Use some hardcoded numbers that are suitable for testing
|
||||
ReadTimeout: 30 * time.Second, //nolint:gomnd
|
||||
WriteTimeout: 30 * time.Second, //nolint:gomnd
|
||||
MaxHeaderBytes: 1 << 20, //nolint:gomnd
|
||||
}
|
||||
listener, err := net.Listen("tcp", a.addr)
|
||||
if err != nil {
|
||||
return tracerr.Wrap(err)
|
||||
}
|
||||
log.Infof("DebugAPI is ready at %v", a.addr)
|
||||
go func() {
|
||||
log.Infof("DebugAPI is ready at %v", a.addr)
|
||||
if err := debugAPIServer.ListenAndServe(); err != nil && tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
if err := debugAPIServer.Serve(listener); err != nil &&
|
||||
tracerr.Unwrap(err) != http.ErrServerClosed {
|
||||
log.Fatalf("Listen: %s\n", err)
|
||||
}
|
||||
}()
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user