mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 03:16:45 +01:00
Update coordinator, call all api update functions
- Common: - Rename Block.EthBlockNum to Block.Num to avoid unneeded repetition - API: - Add UpdateNetworkInfoBlock to update just block information, to be used when the node is not yet synchronized - Node: - Call API.UpdateMetrics and UpdateRecommendedFee in a loop, with configurable time intervals - Synchronizer: - When mapping events by TxHash, use an array to support the possibility of multiple calls of the same function happening in the same transaction (for example, a smart contract in a single transaction could call withdraw with delay twice, which would generate 2 withdraw events, and 2 deposit events). - In Stats, keep entire LastBlock instead of just the blockNum - In Stats, add lastL1BatchBlock - Test Stats and SCVars - Coordinator: - Enable writing the BatchInfo in every step of the pipeline to disk (with JSON text files) for debugging purposes. - Move the Pipeline functionality from the Coordinator to its own struct (Pipeline) - Implement shouldL1lL2Batch - In TxManager, implement logic to perform several attempts when doing ethereum node RPC calls before considering the error. (Both for calls to forgeBatch and transaction receipt) - In TxManager, reorganize the flow and note the specific points in which actions are made when err != nil - HistoryDB: - Implement GetLastL1BatchBlockNum: returns the blockNum of the latest forged l1Batch, to help the coordinator decide when to forge an L1Batch. - EthereumClient and test.Client: - Update EthBlockByNumber to return the last block when the passed number is -1.
This commit is contained in:
@@ -412,7 +412,7 @@ func TestMain(m *testing.M) {
|
|||||||
bids: testBids,
|
bids: testBids,
|
||||||
slots: api.genTestSlots(
|
slots: api.genTestSlots(
|
||||||
20,
|
20,
|
||||||
commonBlocks[len(commonBlocks)-1].EthBlockNum,
|
commonBlocks[len(commonBlocks)-1].Num,
|
||||||
testBids,
|
testBids,
|
||||||
auctionVars,
|
auctionVars,
|
||||||
),
|
),
|
||||||
@@ -611,7 +611,7 @@ func doBadReq(method, path string, reqBody io.Reader, expectedResponseCode int)
|
|||||||
|
|
||||||
func getTimestamp(blockNum int64, blocks []common.Block) time.Time {
|
func getTimestamp(blockNum int64, blocks []common.Block) time.Time {
|
||||||
for i := 0; i < len(blocks); i++ {
|
for i := 0; i < len(blocks); i++ {
|
||||||
if blocks[i].EthBlockNum == blockNum {
|
if blocks[i].Num == blockNum {
|
||||||
return blocks[i].Timestamp
|
return blocks[i].Timestamp
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -647,7 +647,7 @@ func getAccountByIdx(idx common.Idx, accs []common.Account) *common.Account {
|
|||||||
|
|
||||||
func getBlockByNum(ethBlockNum int64, blocks []common.Block) common.Block {
|
func getBlockByNum(ethBlockNum int64, blocks []common.Block) common.Block {
|
||||||
for _, b := range blocks {
|
for _, b := range blocks {
|
||||||
if b.EthBlockNum == ethBlockNum {
|
if b.Num == ethBlockNum {
|
||||||
return b
|
return b
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ import (
|
|||||||
func AddAditionalInformation(blocks []common.BlockData) {
|
func AddAditionalInformation(blocks []common.BlockData) {
|
||||||
for i := range blocks {
|
for i := range blocks {
|
||||||
blocks[i].Block.Timestamp = time.Now().Add(time.Second * 13).UTC()
|
blocks[i].Block.Timestamp = time.Now().Add(time.Second * 13).UTC()
|
||||||
blocks[i].Block.Hash = ethCommon.BigToHash(big.NewInt(blocks[i].Block.EthBlockNum))
|
blocks[i].Block.Hash = ethCommon.BigToHash(big.NewInt(blocks[i].Block.Num))
|
||||||
for j := range blocks[i].Rollup.AddedTokens {
|
for j := range blocks[i].Rollup.AddedTokens {
|
||||||
blocks[i].Rollup.AddedTokens[j].Name = "NAME" + strconv.Itoa(int(blocks[i].Rollup.AddedTokens[j].TokenID))
|
blocks[i].Rollup.AddedTokens[j].Name = "NAME" + strconv.Itoa(int(blocks[i].Rollup.AddedTokens[j].TokenID))
|
||||||
blocks[i].Rollup.AddedTokens[j].Symbol = strconv.Itoa(int(blocks[i].Rollup.AddedTokens[j].TokenID))
|
blocks[i].Rollup.AddedTokens[j].Symbol = strconv.Itoa(int(blocks[i].Rollup.AddedTokens[j].TokenID))
|
||||||
|
|||||||
@@ -60,7 +60,7 @@ func genTestBatches(
|
|||||||
block := common.Block{}
|
block := common.Block{}
|
||||||
found := false
|
found := false
|
||||||
for _, b := range blocks {
|
for _, b := range blocks {
|
||||||
if b.EthBlockNum == cBatches[i].EthBlockNum {
|
if b.Num == cBatches[i].EthBlockNum {
|
||||||
block = b
|
block = b
|
||||||
found = true
|
found = true
|
||||||
break
|
break
|
||||||
|
|||||||
10
api/slots.go
10
api/slots.go
@@ -116,9 +116,9 @@ func (a *API) getSlot(c *gin.Context) {
|
|||||||
|
|
||||||
var slot SlotAPI
|
var slot SlotAPI
|
||||||
if err == sql.ErrNoRows {
|
if err == sql.ErrNoRows {
|
||||||
slot = a.newSlotAPI(slotNum, currentBlock.EthBlockNum, nil, auctionVars)
|
slot = a.newSlotAPI(slotNum, currentBlock.Num, nil, auctionVars)
|
||||||
} else {
|
} else {
|
||||||
slot = a.newSlotAPI(bid.SlotNum, currentBlock.EthBlockNum, &bid, auctionVars)
|
slot = a.newSlotAPI(bid.SlotNum, currentBlock.Num, &bid, auctionVars)
|
||||||
}
|
}
|
||||||
|
|
||||||
// JSON response
|
// JSON response
|
||||||
@@ -221,7 +221,7 @@ func (a *API) getSlots(c *gin.Context) {
|
|||||||
retBadReq(err, c)
|
retBadReq(err, c)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
currentSlot := a.getCurrentSlot(currentBlock.EthBlockNum)
|
currentSlot := a.getCurrentSlot(currentBlock.Num)
|
||||||
auctionVars, err := a.h.GetAuctionVars()
|
auctionVars, err := a.h.GetAuctionVars()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
retBadReq(err, c)
|
retBadReq(err, c)
|
||||||
@@ -268,7 +268,7 @@ func (a *API) getSlots(c *gin.Context) {
|
|||||||
// Build the slot information with previous bids
|
// Build the slot information with previous bids
|
||||||
var slotsBids []SlotAPI
|
var slotsBids []SlotAPI
|
||||||
if len(bids) > 0 {
|
if len(bids) > 0 {
|
||||||
slotsBids = a.newSlotsAPIFromWinnerBids(fromItem, order, bids, currentBlock.EthBlockNum, auctionVars)
|
slotsBids = a.newSlotsAPIFromWinnerBids(fromItem, order, bids, currentBlock.Num, auctionVars)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
retBadReq(err, c)
|
retBadReq(err, c)
|
||||||
return
|
return
|
||||||
@@ -296,7 +296,7 @@ func (a *API) getSlots(c *gin.Context) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if !found {
|
if !found {
|
||||||
slots, err = a.addEmptySlot(slots, i, currentBlock.EthBlockNum, auctionVars, fromItem, order)
|
slots, err = a.addEmptySlot(slots, i, currentBlock.Num, auctionVars, fromItem, order)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
retBadReq(err, c)
|
retBadReq(err, c)
|
||||||
return
|
return
|
||||||
|
|||||||
@@ -181,7 +181,7 @@ func TestGetSlots(t *testing.T) {
|
|||||||
err = doGoodReqPaginated(path, historydb.OrderAsc, &testSlotsResponse{}, appendIter)
|
err = doGoodReqPaginated(path, historydb.OrderAsc, &testSlotsResponse{}, appendIter)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
currentSlot := api.getCurrentSlot(tc.blocks[len(tc.blocks)-1].EthBlockNum)
|
currentSlot := api.getCurrentSlot(tc.blocks[len(tc.blocks)-1].Num)
|
||||||
finishedAuctionSlots := []testSlot{}
|
finishedAuctionSlots := []testSlot{}
|
||||||
for i := 0; i < len(tc.slots); i++ {
|
for i := 0; i < len(tc.slots); i++ {
|
||||||
finishAuction := currentSlot + int64(tc.auctionVars.ClosedAuctionSlots)
|
finishAuction := currentSlot + int64(tc.auctionVars.ClosedAuctionSlots)
|
||||||
|
|||||||
16
api/state.go
16
api/state.go
@@ -65,13 +65,21 @@ func (a *API) SetAuctionVariables(auctionVariables common.AuctionVariables) {
|
|||||||
|
|
||||||
// Network
|
// Network
|
||||||
|
|
||||||
|
// UpdateNetworkInfoBlock update Status.Network block related information
|
||||||
|
func (a *API) UpdateNetworkInfoBlock(
|
||||||
|
lastEthBlock, lastSyncBlock common.Block,
|
||||||
|
) {
|
||||||
|
a.status.Network.LastSyncBlock = lastSyncBlock.Num
|
||||||
|
a.status.Network.LastEthBlock = lastEthBlock.Num
|
||||||
|
}
|
||||||
|
|
||||||
// UpdateNetworkInfo update Status.Network information
|
// UpdateNetworkInfo update Status.Network information
|
||||||
func (a *API) UpdateNetworkInfo(
|
func (a *API) UpdateNetworkInfo(
|
||||||
lastEthBlock, lastSyncBlock common.Block,
|
lastEthBlock, lastSyncBlock common.Block,
|
||||||
lastBatchNum common.BatchNum, currentSlot int64,
|
lastBatchNum common.BatchNum, currentSlot int64,
|
||||||
) error {
|
) error {
|
||||||
a.status.Network.LastSyncBlock = lastSyncBlock.EthBlockNum
|
a.status.Network.LastSyncBlock = lastSyncBlock.Num
|
||||||
a.status.Network.LastEthBlock = lastEthBlock.EthBlockNum
|
a.status.Network.LastEthBlock = lastEthBlock.Num
|
||||||
lastBatch, err := a.h.GetBatchAPI(lastBatchNum)
|
lastBatch, err := a.h.GetBatchAPI(lastBatchNum)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -106,8 +114,8 @@ func (a *API) GetNextForgers(lastBlock common.Block, currentSlot, lastClosedSlot
|
|||||||
SlotNum: i,
|
SlotNum: i,
|
||||||
FromBlock: fromBlock,
|
FromBlock: fromBlock,
|
||||||
ToBlock: toBlock,
|
ToBlock: toBlock,
|
||||||
FromTimestamp: lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(fromBlock-lastBlock.EthBlockNum))),
|
FromTimestamp: lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(fromBlock-lastBlock.Num))),
|
||||||
ToTimestamp: lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(toBlock-lastBlock.EthBlockNum))),
|
ToTimestamp: lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(toBlock-lastBlock.Num))),
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
foundBid := false
|
foundBid := false
|
||||||
|
|||||||
@@ -69,8 +69,8 @@ func TestNextForgers(t *testing.T) {
|
|||||||
assert.Equal(t, bootCoordinator.Bidder, nextForgers[q].Coordinator.Bidder)
|
assert.Equal(t, bootCoordinator.Bidder, nextForgers[q].Coordinator.Bidder)
|
||||||
}
|
}
|
||||||
firstBlockSlot, lastBlockSlot := api.getFirstLastBlock(j)
|
firstBlockSlot, lastBlockSlot := api.getFirstLastBlock(j)
|
||||||
fromTimestamp := lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(firstBlockSlot-lastBlock.EthBlockNum)))
|
fromTimestamp := lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(firstBlockSlot-lastBlock.Num)))
|
||||||
toTimestamp := lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(lastBlockSlot-lastBlock.EthBlockNum)))
|
toTimestamp := lastBlock.Timestamp.Add(time.Second * time.Duration(secondsPerBlock*(lastBlockSlot-lastBlock.Num)))
|
||||||
assert.Equal(t, fromTimestamp.Unix(), nextForgers[q].Period.FromTimestamp.Unix())
|
assert.Equal(t, fromTimestamp.Unix(), nextForgers[q].Period.FromTimestamp.Unix())
|
||||||
assert.Equal(t, toTimestamp.Unix(), nextForgers[q].Period.ToTimestamp.Unix())
|
assert.Equal(t, toTimestamp.Unix(), nextForgers[q].Period.ToTimestamp.Unix())
|
||||||
}
|
}
|
||||||
@@ -90,7 +90,7 @@ func TestUpdateNetworkInfo(t *testing.T) {
|
|||||||
currentSlotNum := int64(1)
|
currentSlotNum := int64(1)
|
||||||
err := api.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
err := api.UpdateNetworkInfo(lastBlock, lastBlock, lastBatchNum, currentSlotNum)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, lastBlock.EthBlockNum, api.status.Network.LastSyncBlock)
|
assert.Equal(t, lastBlock.Num, api.status.Network.LastSyncBlock)
|
||||||
assert.Equal(t, lastBatchNum, api.status.Network.LastBatch.BatchNum)
|
assert.Equal(t, lastBatchNum, api.status.Network.LastBatch.BatchNum)
|
||||||
assert.Equal(t, currentSlotNum, api.status.Network.CurrentSlot)
|
assert.Equal(t, currentSlotNum, api.status.Network.CurrentSlot)
|
||||||
assert.Equal(t, int(api.status.Auction.ClosedAuctionSlots)+1, len(api.status.Network.NextForgers))
|
assert.Equal(t, int(api.status.Auction.ClosedAuctionSlots)+1, len(api.status.Network.NextForgers))
|
||||||
@@ -146,8 +146,8 @@ func TestGetState(t *testing.T) {
|
|||||||
assert.Equal(t, tc.rollupVars, status.Rollup)
|
assert.Equal(t, tc.rollupVars, status.Rollup)
|
||||||
assert.Equal(t, tc.auctionVars, status.Auction)
|
assert.Equal(t, tc.auctionVars, status.Auction)
|
||||||
assert.Equal(t, tc.wdelayerVars, status.WithdrawalDelayer)
|
assert.Equal(t, tc.wdelayerVars, status.WithdrawalDelayer)
|
||||||
assert.Equal(t, lastBlock.EthBlockNum, status.Network.LastEthBlock)
|
assert.Equal(t, lastBlock.Num, status.Network.LastEthBlock)
|
||||||
assert.Equal(t, lastBlock.EthBlockNum, status.Network.LastSyncBlock)
|
assert.Equal(t, lastBlock.Num, status.Network.LastSyncBlock)
|
||||||
assert.Equal(t, lastBatchNum, status.Network.LastBatch.BatchNum)
|
assert.Equal(t, lastBatchNum, status.Network.LastBatch.BatchNum)
|
||||||
assert.Equal(t, currentSlotNum, status.Network.CurrentSlot)
|
assert.Equal(t, currentSlotNum, status.Network.CurrentSlot)
|
||||||
assert.Equal(t, int(api.status.Auction.ClosedAuctionSlots)+1, len(status.Network.NextForgers))
|
assert.Equal(t, int(api.status.Auction.ClosedAuctionSlots)+1, len(status.Network.NextForgers))
|
||||||
|
|||||||
@@ -1,6 +1,8 @@
|
|||||||
[API]
|
[API]
|
||||||
Address = "localhost:8086"
|
Address = "localhost:8086"
|
||||||
Explorer = true
|
Explorer = true
|
||||||
|
UpdateMetricsInterval = "10s"
|
||||||
|
UpdateRecommendedFeeInterval = "10s"
|
||||||
|
|
||||||
[Debug]
|
[Debug]
|
||||||
APIAddress = "localhost:12345"
|
APIAddress = "localhost:12345"
|
||||||
|
|||||||
@@ -33,6 +33,9 @@ type Slot struct {
|
|||||||
BatchesLen int
|
BatchesLen int
|
||||||
BidValue *big.Int
|
BidValue *big.Int
|
||||||
BootCoord bool
|
BootCoord bool
|
||||||
|
// Bidder, Forer and URL correspond to the winner of the slot (which is
|
||||||
|
// not always the highest bidder). These are the values of the
|
||||||
|
// coordinator that is able to forge exclusively before the deadline.
|
||||||
Bidder ethCommon.Address
|
Bidder ethCommon.Address
|
||||||
Forger ethCommon.Address
|
Forger ethCommon.Address
|
||||||
URL string
|
URL string
|
||||||
|
|||||||
@@ -9,10 +9,10 @@ import (
|
|||||||
|
|
||||||
// Block represents of an Ethereum block
|
// Block represents of an Ethereum block
|
||||||
type Block struct {
|
type Block struct {
|
||||||
EthBlockNum int64 `meddler:"eth_block_num"`
|
Num int64 `meddler:"eth_block_num"`
|
||||||
Timestamp time.Time `meddler:"timestamp,utctime"`
|
Timestamp time.Time `meddler:"timestamp,utctime"`
|
||||||
Hash ethCommon.Hash `meddler:"hash"`
|
Hash ethCommon.Hash `meddler:"hash"`
|
||||||
ParentHash ethCommon.Hash `meddler:"-"`
|
ParentHash ethCommon.Hash `meddler:"-" json:"-"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// RollupData contains information returned by the Rollup smart contract
|
// RollupData contains information returned by the Rollup smart contract
|
||||||
@@ -65,7 +65,8 @@ type WDelayerTransfer struct {
|
|||||||
type WDelayerData struct {
|
type WDelayerData struct {
|
||||||
Vars *WDelayerVariables
|
Vars *WDelayerVariables
|
||||||
Deposits []WDelayerTransfer
|
Deposits []WDelayerTransfer
|
||||||
DepositsByTxHash map[ethCommon.Hash]*WDelayerTransfer
|
// We use an array because there can be multiple deposits in a single eth transaction
|
||||||
|
DepositsByTxHash map[ethCommon.Hash][]*WDelayerTransfer
|
||||||
Withdrawals []WDelayerTransfer
|
Withdrawals []WDelayerTransfer
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -74,7 +75,7 @@ func NewWDelayerData() WDelayerData {
|
|||||||
return WDelayerData{
|
return WDelayerData{
|
||||||
Vars: nil,
|
Vars: nil,
|
||||||
Deposits: make([]WDelayerTransfer, 0),
|
Deposits: make([]WDelayerTransfer, 0),
|
||||||
DepositsByTxHash: make(map[ethCommon.Hash]*WDelayerTransfer),
|
DepositsByTxHash: make(map[ethCommon.Hash][]*WDelayerTransfer),
|
||||||
Withdrawals: make([]WDelayerTransfer, 0),
|
Withdrawals: make([]WDelayerTransfer, 0),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -92,6 +92,8 @@ type Node struct {
|
|||||||
API struct {
|
API struct {
|
||||||
Address string
|
Address string
|
||||||
Explorer bool
|
Explorer bool
|
||||||
|
UpdateMetricsInterval Duration
|
||||||
|
UpdateRecommendedFeeInterval Duration
|
||||||
} `validate:"required"`
|
} `validate:"required"`
|
||||||
Debug struct {
|
Debug struct {
|
||||||
APIAddress string
|
APIAddress string
|
||||||
|
|||||||
@@ -1,6 +1,11 @@
|
|||||||
package coordinator
|
package coordinator
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"encoding/json"
|
||||||
|
"fmt"
|
||||||
|
"io/ioutil"
|
||||||
|
"path"
|
||||||
|
|
||||||
"github.com/ethereum/go-ethereum/core/types"
|
"github.com/ethereum/go-ethereum/core/types"
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
"github.com/hermeznetwork/hermez-node/eth"
|
"github.com/hermeznetwork/hermez-node/eth"
|
||||||
@@ -33,4 +38,22 @@ type BatchInfo struct {
|
|||||||
// FeesInfo
|
// FeesInfo
|
||||||
TxStatus TxStatus
|
TxStatus TxStatus
|
||||||
EthTx *types.Transaction
|
EthTx *types.Transaction
|
||||||
|
Receipt *types.Receipt
|
||||||
|
}
|
||||||
|
|
||||||
|
// DebugStore is a debug function to store the BatchInfo as a json text file in
|
||||||
|
// storePath
|
||||||
|
func (b *BatchInfo) DebugStore(storePath string) error {
|
||||||
|
batchJSON, err := json.Marshal(b)
|
||||||
|
if err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
oldStateRoot := "null"
|
||||||
|
if b.ZKInputs != nil && b.ZKInputs.OldStateRoot != nil {
|
||||||
|
oldStateRoot = b.ZKInputs.OldStateRoot.String()
|
||||||
|
}
|
||||||
|
filename := fmt.Sprintf("%010d-%s.json", b.BatchNum, oldStateRoot)
|
||||||
|
// nolint reason: 0640 allows rw to owner and r to group
|
||||||
|
//nolint:gosec
|
||||||
|
return ioutil.WriteFile(path.Join(storePath, filename), batchJSON, 0640)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -25,71 +25,98 @@ var ErrDone = fmt.Errorf("done")
|
|||||||
|
|
||||||
// Config contains the Coordinator configuration
|
// Config contains the Coordinator configuration
|
||||||
type Config struct {
|
type Config struct {
|
||||||
|
// ForgerAddress is the address under which this coordinator is forging
|
||||||
ForgerAddress ethCommon.Address
|
ForgerAddress ethCommon.Address
|
||||||
|
// ConfirmBlocks is the number of confirmation blocks to wait for sent
|
||||||
|
// ethereum transactions before forgetting about them
|
||||||
ConfirmBlocks int64
|
ConfirmBlocks int64
|
||||||
|
// L1BatchTimeoutPerc is the portion of the range before the L1Batch
|
||||||
|
// timeout that will trigger a schedule to forge an L1Batch
|
||||||
|
L1BatchTimeoutPerc float64
|
||||||
|
// EthClientAttempts is the number of attempts to do an eth client RPC
|
||||||
|
// call before giving up
|
||||||
|
EthClientAttempts int
|
||||||
|
// EthClientAttemptsDelay is delay between attempts do do an eth client
|
||||||
|
// RPC call
|
||||||
|
EthClientAttemptsDelay time.Duration
|
||||||
|
// TxManagerCheckInterval is the waiting interval between receipt
|
||||||
|
// checks of ethereum transactions in the TxManager
|
||||||
|
TxManagerCheckInterval time.Duration
|
||||||
|
// DebugBatchPath if set, specifies the path where batchInfo is stored
|
||||||
|
// in JSON in every step/update of the pipeline
|
||||||
|
DebugBatchPath string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Config) debugBatchStore(batchInfo *BatchInfo) {
|
||||||
|
if c.DebugBatchPath != "" {
|
||||||
|
if err := batchInfo.DebugStore(c.DebugBatchPath); err != nil {
|
||||||
|
log.Warnw("Error storing debug BatchInfo",
|
||||||
|
"path", c.DebugBatchPath, "err", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Coordinator implements the Coordinator type
|
// Coordinator implements the Coordinator type
|
||||||
type Coordinator struct {
|
type Coordinator struct {
|
||||||
// State
|
// State
|
||||||
forging bool
|
|
||||||
batchNum common.BatchNum
|
batchNum common.BatchNum
|
||||||
serverProofPool *ServerProofPool
|
serverProofs []ServerProofInterface
|
||||||
consts synchronizer.SCConsts
|
consts synchronizer.SCConsts
|
||||||
vars synchronizer.SCVariables
|
vars synchronizer.SCVariables
|
||||||
|
started bool
|
||||||
|
|
||||||
cfg Config
|
cfg Config
|
||||||
|
|
||||||
hdb *historydb.HistoryDB
|
historyDB *historydb.HistoryDB
|
||||||
txsel *txselector.TxSelector
|
txSelector *txselector.TxSelector
|
||||||
batchBuilder *batchbuilder.BatchBuilder
|
batchBuilder *batchbuilder.BatchBuilder
|
||||||
|
|
||||||
ethClient eth.ClientInterface
|
|
||||||
|
|
||||||
msgCh chan interface{}
|
msgCh chan interface{}
|
||||||
ctx context.Context
|
ctx context.Context
|
||||||
wg sync.WaitGroup
|
wg sync.WaitGroup
|
||||||
cancel context.CancelFunc
|
cancel context.CancelFunc
|
||||||
|
|
||||||
pipelineCtx context.Context
|
pipeline *Pipeline
|
||||||
pipelineWg sync.WaitGroup
|
|
||||||
pipelineCancel context.CancelFunc
|
|
||||||
|
|
||||||
txManager *TxManager
|
txManager *TxManager
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewCoordinator creates a new Coordinator
|
// NewCoordinator creates a new Coordinator
|
||||||
func NewCoordinator(cfg Config,
|
func NewCoordinator(cfg Config,
|
||||||
hdb *historydb.HistoryDB,
|
historyDB *historydb.HistoryDB,
|
||||||
txsel *txselector.TxSelector,
|
txSelector *txselector.TxSelector,
|
||||||
bb *batchbuilder.BatchBuilder,
|
batchBuilder *batchbuilder.BatchBuilder,
|
||||||
serverProofs []ServerProofInterface,
|
serverProofs []ServerProofInterface,
|
||||||
ethClient eth.ClientInterface,
|
ethClient eth.ClientInterface,
|
||||||
scConsts *synchronizer.SCConsts,
|
scConsts *synchronizer.SCConsts,
|
||||||
initSCVars *synchronizer.SCVariables,
|
initSCVars *synchronizer.SCVariables,
|
||||||
) *Coordinator { // once synchronizer is ready, synchronizer.Synchronizer will be passed as parameter here
|
) (*Coordinator, error) {
|
||||||
serverProofPool := NewServerProofPool(len(serverProofs))
|
// nolint reason: hardcoded `1.0`, by design the percentage can't be over 100%
|
||||||
for _, serverProof := range serverProofs {
|
if cfg.L1BatchTimeoutPerc >= 1.0 { //nolint:gomnd
|
||||||
serverProofPool.Add(serverProof)
|
return nil, fmt.Errorf("invalid value for Config.L1BatchTimeoutPerc (%v >= 1.0)",
|
||||||
|
cfg.L1BatchTimeoutPerc)
|
||||||
|
}
|
||||||
|
if cfg.EthClientAttempts < 1 {
|
||||||
|
return nil, fmt.Errorf("invalid value for Config.EthClientAttempts (%v < 1)",
|
||||||
|
cfg.EthClientAttempts)
|
||||||
}
|
}
|
||||||
|
|
||||||
txManager := NewTxManager(ethClient, cfg.ConfirmBlocks)
|
txManager := NewTxManager(&cfg, ethClient)
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
c := Coordinator{
|
c := Coordinator{
|
||||||
forging: false,
|
|
||||||
batchNum: -1,
|
batchNum: -1,
|
||||||
serverProofPool: serverProofPool,
|
serverProofs: serverProofs,
|
||||||
consts: *scConsts,
|
consts: *scConsts,
|
||||||
vars: *initSCVars,
|
vars: *initSCVars,
|
||||||
|
|
||||||
cfg: cfg,
|
cfg: cfg,
|
||||||
|
|
||||||
hdb: hdb,
|
historyDB: historyDB,
|
||||||
txsel: txsel,
|
txSelector: txSelector,
|
||||||
batchBuilder: bb,
|
batchBuilder: batchBuilder,
|
||||||
|
|
||||||
ethClient: ethClient,
|
// ethClient: ethClient,
|
||||||
|
|
||||||
msgCh: make(chan interface{}),
|
msgCh: make(chan interface{}),
|
||||||
ctx: ctx,
|
ctx: ctx,
|
||||||
@@ -98,7 +125,12 @@ func NewCoordinator(cfg Config,
|
|||||||
|
|
||||||
txManager: txManager,
|
txManager: txManager,
|
||||||
}
|
}
|
||||||
return &c
|
return &c, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (c *Coordinator) newPipeline() *Pipeline {
|
||||||
|
return NewPipeline(c.cfg, c.historyDB, c.txSelector, c.batchBuilder,
|
||||||
|
c.txManager, c.serverProofs, &c.consts)
|
||||||
}
|
}
|
||||||
|
|
||||||
// MsgSyncStats indicates an update to the Synchronizer stats
|
// MsgSyncStats indicates an update to the Synchronizer stats
|
||||||
@@ -134,50 +166,53 @@ func (c *Coordinator) handleMsgSyncSCVars(msg *MsgSyncSCVars) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (c *Coordinator) canForge(stats *synchronizer.Stats) bool {
|
||||||
|
anyoneForge := false
|
||||||
|
if stats.Sync.Auction.CurrentSlot.BatchesLen == 0 &&
|
||||||
|
c.consts.Auction.RelativeBlock(stats.Eth.LastBlock.Num+1) > int64(c.vars.Auction.SlotDeadline) {
|
||||||
|
log.Debug("Coordinator: anyone can forge in the current slot (slotDeadline passed)")
|
||||||
|
anyoneForge = true
|
||||||
|
}
|
||||||
|
if stats.Sync.Auction.CurrentSlot.Forger == c.cfg.ForgerAddress || anyoneForge {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
|
}
|
||||||
|
|
||||||
func (c *Coordinator) handleMsgSyncStats(stats *synchronizer.Stats) error {
|
func (c *Coordinator) handleMsgSyncStats(stats *synchronizer.Stats) error {
|
||||||
if !stats.Synced() {
|
if !stats.Synced() {
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
c.txManager.SetLastBlock(stats.Eth.LastBlock)
|
c.txManager.SetLastBlock(stats.Eth.LastBlock.Num)
|
||||||
|
|
||||||
anyoneForge := false
|
canForge := c.canForge(stats)
|
||||||
if stats.Sync.Auction.CurrentSlot.BatchesLen == 0 &&
|
if c.pipeline == nil {
|
||||||
c.consts.Auction.RelativeBlock(stats.Eth.LastBlock) > int64(c.vars.Auction.SlotDeadline) {
|
if canForge {
|
||||||
log.Debug("Coordinator: anyone can forge in the current slot (slotDeadline passed)")
|
|
||||||
anyoneForge = true
|
|
||||||
}
|
|
||||||
if stats.Sync.Auction.CurrentSlot.Forger != c.cfg.ForgerAddress && !anyoneForge {
|
|
||||||
if c.forging {
|
|
||||||
log.Info("Coordinator: forging state end")
|
|
||||||
c.forging = false
|
|
||||||
c.PipelineStop()
|
|
||||||
}
|
|
||||||
// log.Debug("Coordinator: not in forge time") // DBG
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
// log.Debug("Coordinator: forge time") // DBG
|
|
||||||
if !c.forging {
|
|
||||||
// Start pipeline from a batchNum state taken from synchronizer
|
|
||||||
log.Info("Coordinator: forging state begin")
|
log.Info("Coordinator: forging state begin")
|
||||||
c.batchNum = common.BatchNum(stats.Sync.LastBatch)
|
batchNum := common.BatchNum(stats.Sync.LastBatch)
|
||||||
err := c.txsel.Reset(c.batchNum)
|
c.pipeline = c.newPipeline()
|
||||||
if err != nil {
|
if err := c.pipeline.Start(batchNum, stats, &c.vars); err != nil {
|
||||||
log.Errorw("Coordinator: TxSelector.Reset", "error", err)
|
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
err = c.batchBuilder.Reset(c.batchNum, true)
|
|
||||||
if err != nil {
|
|
||||||
log.Errorw("Coordinator: BatchBuilder.Reset", "error", err)
|
|
||||||
return err
|
|
||||||
}
|
}
|
||||||
c.forging = true
|
} else {
|
||||||
c.PipelineStart()
|
if canForge {
|
||||||
|
c.pipeline.SetSyncStats(stats)
|
||||||
|
} else {
|
||||||
|
log.Info("Coordinator: forging state end")
|
||||||
|
c.pipeline.Stop()
|
||||||
|
c.pipeline = nil
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// Start the coordinator
|
// Start the coordinator
|
||||||
func (c *Coordinator) Start() {
|
func (c *Coordinator) Start() {
|
||||||
|
if c.started {
|
||||||
|
log.Fatal("Coordinator already started")
|
||||||
|
}
|
||||||
|
c.started = true
|
||||||
c.wg.Add(1)
|
c.wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
c.txManager.Run(c.ctx)
|
c.txManager.Run(c.ctx)
|
||||||
@@ -198,10 +233,12 @@ func (c *Coordinator) Start() {
|
|||||||
stats := msg.Stats
|
stats := msg.Stats
|
||||||
if err := c.handleMsgSyncStats(&stats); err != nil {
|
if err := c.handleMsgSyncStats(&stats); err != nil {
|
||||||
log.Errorw("Coordinator.handleMsgSyncStats error", "err", err)
|
log.Errorw("Coordinator.handleMsgSyncStats error", "err", err)
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
case MsgSyncReorg:
|
case MsgSyncReorg:
|
||||||
if err := c.handleReorg(); err != nil {
|
if err := c.handleReorg(); err != nil {
|
||||||
log.Errorw("Coordinator.handleReorg error", "err", err)
|
log.Errorw("Coordinator.handleReorg error", "err", err)
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
case MsgSyncSCVars:
|
case MsgSyncSCVars:
|
||||||
c.handleMsgSyncSCVars(&msg)
|
c.handleMsgSyncSCVars(&msg)
|
||||||
@@ -215,96 +252,44 @@ func (c *Coordinator) Start() {
|
|||||||
|
|
||||||
// Stop the coordinator
|
// Stop the coordinator
|
||||||
func (c *Coordinator) Stop() {
|
func (c *Coordinator) Stop() {
|
||||||
log.Infow("Stopping coordinator...")
|
if !c.started {
|
||||||
|
log.Fatal("Coordinator already stopped")
|
||||||
|
}
|
||||||
|
c.started = false
|
||||||
|
log.Infow("Stopping Coordinator...")
|
||||||
c.cancel()
|
c.cancel()
|
||||||
c.wg.Wait()
|
c.wg.Wait()
|
||||||
if c.forging {
|
if c.pipeline != nil {
|
||||||
c.forging = false
|
c.pipeline.Stop()
|
||||||
c.PipelineStop()
|
c.pipeline = nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// PipelineStart starts the forging pipeline
|
func (c *Coordinator) handleReorg() error {
|
||||||
func (c *Coordinator) PipelineStart() {
|
return nil // TODO
|
||||||
c.pipelineCtx, c.pipelineCancel = context.WithCancel(context.Background())
|
|
||||||
|
|
||||||
queueSize := 1
|
|
||||||
batchChSentServerProof := make(chan *BatchInfo, queueSize)
|
|
||||||
|
|
||||||
c.pipelineWg.Add(1)
|
|
||||||
go func() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-c.pipelineCtx.Done():
|
|
||||||
log.Debug("Pipeline forgeSendServerProof loop done")
|
|
||||||
c.pipelineWg.Done()
|
|
||||||
return
|
|
||||||
default:
|
|
||||||
c.batchNum = c.batchNum + 1
|
|
||||||
batchInfo, err := c.forgeSendServerProof(c.pipelineCtx, c.batchNum)
|
|
||||||
if err == ErrDone {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
log.Errorw("forgeSendServerProof", "err", err)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
batchChSentServerProof <- batchInfo
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
|
|
||||||
c.pipelineWg.Add(1)
|
|
||||||
go func() {
|
|
||||||
for {
|
|
||||||
select {
|
|
||||||
case <-c.pipelineCtx.Done():
|
|
||||||
log.Debug("Pipeline waitServerProofSendEth loop done")
|
|
||||||
c.pipelineWg.Done()
|
|
||||||
return
|
|
||||||
case batchInfo := <-batchChSentServerProof:
|
|
||||||
err := c.waitServerProof(c.pipelineCtx, batchInfo)
|
|
||||||
if err == ErrDone {
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
if err != nil {
|
|
||||||
log.Errorw("waitServerProof", "err", err)
|
|
||||||
continue
|
|
||||||
}
|
|
||||||
c.txManager.AddBatch(batchInfo)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
}
|
|
||||||
|
|
||||||
// PipelineStop stops the forging pipeline
|
|
||||||
func (c *Coordinator) PipelineStop() {
|
|
||||||
log.Debug("Stopping pipeline...")
|
|
||||||
c.pipelineCancel()
|
|
||||||
c.pipelineWg.Wait()
|
|
||||||
}
|
}
|
||||||
|
|
||||||
// TxManager handles everything related to ethereum transactions: It makes the
|
// TxManager handles everything related to ethereum transactions: It makes the
|
||||||
// call to forge, waits for transaction confirmation, and keeps checking them
|
// call to forge, waits for transaction confirmation, and keeps checking them
|
||||||
// until a number of confirmed blocks have passed.
|
// until a number of confirmed blocks have passed.
|
||||||
type TxManager struct {
|
type TxManager struct {
|
||||||
|
cfg Config
|
||||||
ethClient eth.ClientInterface
|
ethClient eth.ClientInterface
|
||||||
batchCh chan *BatchInfo
|
batchCh chan *BatchInfo
|
||||||
lastBlockCh chan int64
|
lastBlockCh chan int64
|
||||||
queue []*BatchInfo
|
queue []*BatchInfo
|
||||||
confirmation int64
|
|
||||||
lastBlock int64
|
lastBlock int64
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewTxManager creates a new TxManager
|
// NewTxManager creates a new TxManager
|
||||||
func NewTxManager(ethClient eth.ClientInterface, confirmation int64) *TxManager {
|
func NewTxManager(cfg *Config, ethClient eth.ClientInterface) *TxManager {
|
||||||
return &TxManager{
|
return &TxManager{
|
||||||
|
cfg: *cfg,
|
||||||
ethClient: ethClient,
|
ethClient: ethClient,
|
||||||
// TODO: Find best queue size
|
// TODO: Find best queue size
|
||||||
batchCh: make(chan *BatchInfo, 16), //nolint:gomnd
|
batchCh: make(chan *BatchInfo, 16), //nolint:gomnd
|
||||||
// TODO: Find best queue size
|
// TODO: Find best queue size
|
||||||
lastBlockCh: make(chan int64, 16), //nolint:gomnd
|
lastBlockCh: make(chan int64, 16), //nolint:gomnd
|
||||||
confirmation: confirmation,
|
|
||||||
lastBlock: -1,
|
lastBlock: -1,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -320,13 +305,77 @@ func (t *TxManager) SetLastBlock(lastBlock int64) {
|
|||||||
t.lastBlockCh <- lastBlock
|
t.lastBlockCh <- lastBlock
|
||||||
}
|
}
|
||||||
|
|
||||||
const waitTime = 200 * time.Millisecond
|
func (t *TxManager) rollupForgeBatch(ctx context.Context, batchInfo *BatchInfo) error {
|
||||||
|
var ethTx *types.Transaction
|
||||||
|
var err error
|
||||||
|
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ {
|
||||||
|
ethTx, err = t.ethClient.RollupForgeBatch(batchInfo.ForgeBatchArgs)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("TxManager ethClient.RollupForgeBatch",
|
||||||
|
"attempt", attempt, "err", err)
|
||||||
|
} else {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return ErrDone
|
||||||
|
case <-time.After(t.cfg.EthClientAttemptsDelay):
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("reached max attempts for ethClient.RollupForgeBatch: %w", err)
|
||||||
|
}
|
||||||
|
batchInfo.EthTx = ethTx
|
||||||
|
t.cfg.debugBatchStore(batchInfo)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *TxManager) ethTransactionReceipt(ctx context.Context, batchInfo *BatchInfo) error {
|
||||||
|
txHash := batchInfo.EthTx.Hash()
|
||||||
|
var receipt *types.Receipt
|
||||||
|
var err error
|
||||||
|
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ {
|
||||||
|
receipt, err = t.ethClient.EthTransactionReceipt(ctx, txHash)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("TxManager ethClient.EthTransactionReceipt",
|
||||||
|
"attempt", attempt, "err", err)
|
||||||
|
} else {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
select {
|
||||||
|
case <-ctx.Done():
|
||||||
|
return ErrDone
|
||||||
|
case <-time.After(t.cfg.EthClientAttemptsDelay):
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w", err)
|
||||||
|
}
|
||||||
|
batchInfo.Receipt = receipt
|
||||||
|
t.cfg.debugBatchStore(batchInfo)
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *TxManager) handleReceipt(batchInfo *BatchInfo) (*int64, error) {
|
||||||
|
receipt := batchInfo.Receipt
|
||||||
|
if receipt != nil {
|
||||||
|
if receipt.Status == types.ReceiptStatusFailed {
|
||||||
|
log.Errorw("TxManager receipt status is failed", "receipt", receipt)
|
||||||
|
return nil, fmt.Errorf("ethereum transaction receipt statis is failed")
|
||||||
|
} else if receipt.Status == types.ReceiptStatusSuccessful {
|
||||||
|
confirm := t.lastBlock - receipt.BlockNumber.Int64()
|
||||||
|
return &confirm, nil
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nil, nil
|
||||||
|
}
|
||||||
|
|
||||||
const longWaitTime = 999 * time.Hour
|
const longWaitTime = 999 * time.Hour
|
||||||
|
|
||||||
// Run the TxManager
|
// Run the TxManager
|
||||||
func (t *TxManager) Run(ctx context.Context) {
|
func (t *TxManager) Run(ctx context.Context) {
|
||||||
next := 0
|
next := 0
|
||||||
d := time.Duration(longWaitTime)
|
waitTime := time.Duration(longWaitTime)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
@@ -335,41 +384,41 @@ func (t *TxManager) Run(ctx context.Context) {
|
|||||||
case lastBlock := <-t.lastBlockCh:
|
case lastBlock := <-t.lastBlockCh:
|
||||||
t.lastBlock = lastBlock
|
t.lastBlock = lastBlock
|
||||||
case batchInfo := <-t.batchCh:
|
case batchInfo := <-t.batchCh:
|
||||||
ethTx, err := t.ethClient.RollupForgeBatch(batchInfo.ForgeBatchArgs)
|
if err := t.rollupForgeBatch(ctx, batchInfo); err == ErrDone {
|
||||||
if err != nil {
|
continue
|
||||||
// TODO: Figure out different error cases and handle them properly
|
} else if err != nil {
|
||||||
log.Errorw("TxManager ethClient.RollupForgeBatch", "err", err)
|
// TODO: Reset pipeline
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
log.Debugf("ethClient ForgeCall sent, batchNum: %d", batchInfo.BatchNum)
|
log.Debugf("ethClient ForgeCall sent, batchNum: %d", batchInfo.BatchNum)
|
||||||
batchInfo.EthTx = ethTx
|
|
||||||
t.queue = append(t.queue, batchInfo)
|
t.queue = append(t.queue, batchInfo)
|
||||||
d = waitTime
|
waitTime = t.cfg.TxManagerCheckInterval
|
||||||
case <-time.After(d):
|
case <-time.After(waitTime):
|
||||||
if len(t.queue) == 0 {
|
if len(t.queue) == 0 {
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
batchInfo := t.queue[next]
|
batchInfo := t.queue[next]
|
||||||
txID := batchInfo.EthTx.Hash()
|
err := t.ethTransactionReceipt(ctx, batchInfo)
|
||||||
receipt, err := t.ethClient.EthTransactionReceipt(ctx, txID)
|
if err == ErrDone {
|
||||||
if err != nil {
|
|
||||||
log.Errorw("TxManager ethClient.EthTransactionReceipt", "err", err)
|
|
||||||
// TODO: Figure out different error cases and handle them properly
|
|
||||||
// TODO: Notify the Coordinator to maybe reset the pipeline
|
|
||||||
continue
|
continue
|
||||||
|
} else if err != nil { //nolint:staticcheck
|
||||||
|
// We can't get the receipt for the
|
||||||
|
// transaction, so we can't confirm if it was
|
||||||
|
// mined
|
||||||
|
// TODO: Reset pipeline
|
||||||
}
|
}
|
||||||
|
|
||||||
if receipt != nil {
|
confirm, err := t.handleReceipt(batchInfo)
|
||||||
if receipt.Status == types.ReceiptStatusFailed {
|
if err != nil { //nolint:staticcheck
|
||||||
log.Errorw("TxManager receipt status is failed", "receipt", receipt)
|
// Transaction was rejected
|
||||||
} else if receipt.Status == types.ReceiptStatusSuccessful {
|
// TODO: Reset pipeline
|
||||||
if t.lastBlock-receipt.BlockNumber.Int64() >= t.confirmation {
|
}
|
||||||
log.Debugw("TxManager tx for RollupForgeBatch confirmed", "batchNum", batchInfo.BatchNum)
|
if confirm != nil && *confirm >= t.cfg.ConfirmBlocks {
|
||||||
|
log.Debugw("TxManager tx for RollupForgeBatch confirmed",
|
||||||
|
"batchNum", batchInfo.BatchNum)
|
||||||
t.queue = t.queue[1:]
|
t.queue = t.queue[1:]
|
||||||
if len(t.queue) == 0 {
|
if len(t.queue) == 0 {
|
||||||
d = longWaitTime
|
waitTime = longWaitTime
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if len(t.queue) == 0 {
|
if len(t.queue) == 0 {
|
||||||
@@ -381,11 +430,157 @@ func (t *TxManager) Run(ctx context.Context) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Pipeline manages the forging of batches with parallel server proofs
|
||||||
|
type Pipeline struct {
|
||||||
|
cfg Config
|
||||||
|
consts synchronizer.SCConsts
|
||||||
|
|
||||||
|
// state
|
||||||
|
batchNum common.BatchNum
|
||||||
|
vars synchronizer.SCVariables
|
||||||
|
lastScheduledL1BatchBlockNum int64
|
||||||
|
started bool
|
||||||
|
|
||||||
|
serverProofPool *ServerProofPool
|
||||||
|
txManager *TxManager
|
||||||
|
historyDB *historydb.HistoryDB
|
||||||
|
txSelector *txselector.TxSelector
|
||||||
|
batchBuilder *batchbuilder.BatchBuilder
|
||||||
|
|
||||||
|
stats synchronizer.Stats
|
||||||
|
statsCh chan synchronizer.Stats
|
||||||
|
|
||||||
|
ctx context.Context
|
||||||
|
wg sync.WaitGroup
|
||||||
|
cancel context.CancelFunc
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewPipeline creates a new Pipeline
|
||||||
|
func NewPipeline(cfg Config,
|
||||||
|
historyDB *historydb.HistoryDB,
|
||||||
|
txSelector *txselector.TxSelector,
|
||||||
|
batchBuilder *batchbuilder.BatchBuilder,
|
||||||
|
txManager *TxManager,
|
||||||
|
serverProofs []ServerProofInterface,
|
||||||
|
scConsts *synchronizer.SCConsts,
|
||||||
|
) *Pipeline {
|
||||||
|
serverProofPool := NewServerProofPool(len(serverProofs))
|
||||||
|
for _, serverProof := range serverProofs {
|
||||||
|
serverProofPool.Add(serverProof)
|
||||||
|
}
|
||||||
|
return &Pipeline{
|
||||||
|
cfg: cfg,
|
||||||
|
historyDB: historyDB,
|
||||||
|
txSelector: txSelector,
|
||||||
|
batchBuilder: batchBuilder,
|
||||||
|
serverProofPool: serverProofPool,
|
||||||
|
txManager: txManager,
|
||||||
|
consts: *scConsts,
|
||||||
|
// TODO: Find best queue size
|
||||||
|
statsCh: make(chan synchronizer.Stats, 16), //nolint:gomnd
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SetSyncStats is a thread safe method to sets the synchronizer Stats
|
||||||
|
func (p *Pipeline) SetSyncStats(stats *synchronizer.Stats) {
|
||||||
|
p.statsCh <- *stats
|
||||||
|
}
|
||||||
|
|
||||||
|
// Start the forging pipeline
|
||||||
|
func (p *Pipeline) Start(batchNum common.BatchNum,
|
||||||
|
syncStats *synchronizer.Stats, initSCVars *synchronizer.SCVariables) error {
|
||||||
|
if p.started {
|
||||||
|
log.Fatal("Pipeline already started")
|
||||||
|
}
|
||||||
|
p.started = true
|
||||||
|
|
||||||
|
// Reset pipeline state
|
||||||
|
p.batchNum = batchNum
|
||||||
|
p.vars = *initSCVars
|
||||||
|
p.lastScheduledL1BatchBlockNum = 0
|
||||||
|
|
||||||
|
p.ctx, p.cancel = context.WithCancel(context.Background())
|
||||||
|
|
||||||
|
err := p.txSelector.Reset(p.batchNum)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("Pipeline: TxSelector.Reset", "error", err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
err = p.batchBuilder.Reset(p.batchNum, true)
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("Pipeline: BatchBuilder.Reset", "error", err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
queueSize := 1
|
||||||
|
batchChSentServerProof := make(chan *BatchInfo, queueSize)
|
||||||
|
|
||||||
|
p.wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-p.ctx.Done():
|
||||||
|
log.Debug("Pipeline forgeSendServerProof loop done")
|
||||||
|
p.wg.Done()
|
||||||
|
return
|
||||||
|
case syncStats := <-p.statsCh:
|
||||||
|
p.stats = syncStats
|
||||||
|
default:
|
||||||
|
p.batchNum = p.batchNum + 1
|
||||||
|
batchInfo, err := p.forgeSendServerProof(p.ctx, p.batchNum)
|
||||||
|
if err == ErrDone {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("forgeSendServerProof", "err", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
batchChSentServerProof <- batchInfo
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
p.wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-p.ctx.Done():
|
||||||
|
log.Debug("Pipeline waitServerProofSendEth loop done")
|
||||||
|
p.wg.Done()
|
||||||
|
return
|
||||||
|
case batchInfo := <-batchChSentServerProof:
|
||||||
|
err := p.waitServerProof(p.ctx, batchInfo)
|
||||||
|
if err == ErrDone {
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
log.Errorw("waitServerProof", "err", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
p.txManager.AddBatch(batchInfo)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Stop the forging pipeline
|
||||||
|
func (p *Pipeline) Stop() {
|
||||||
|
if !p.started {
|
||||||
|
log.Fatal("Pipeline already stopped")
|
||||||
|
}
|
||||||
|
p.started = false
|
||||||
|
log.Debug("Stopping Pipeline...")
|
||||||
|
p.cancel()
|
||||||
|
p.wg.Wait()
|
||||||
|
// TODO: Cancel all proofServers with pending proofs
|
||||||
|
}
|
||||||
|
|
||||||
// forgeSendServerProof the next batch, wait for a proof server to be available and send the
|
// forgeSendServerProof the next batch, wait for a proof server to be available and send the
|
||||||
// circuit inputs to the proof server.
|
// circuit inputs to the proof server.
|
||||||
func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.BatchNum) (*BatchInfo, error) {
|
func (p *Pipeline) forgeSendServerProof(ctx context.Context, batchNum common.BatchNum) (*BatchInfo, error) {
|
||||||
// remove transactions from the pool that have been there for too long
|
// remove transactions from the pool that have been there for too long
|
||||||
err := c.purgeRemoveByTimeout()
|
err := p.purgeRemoveByTimeout()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@@ -396,17 +591,18 @@ func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.
|
|||||||
// var feesInfo
|
// var feesInfo
|
||||||
var l1UserTxsExtra, l1OperatorTxs []common.L1Tx
|
var l1UserTxsExtra, l1OperatorTxs []common.L1Tx
|
||||||
// 1. Decide if we forge L2Tx or L1+L2Tx
|
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||||
if c.shouldL1L2Batch() {
|
if p.shouldL1L2Batch() {
|
||||||
|
p.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBatch
|
||||||
// 2a: L1+L2 txs
|
// 2a: L1+L2 txs
|
||||||
// l1UserTxs, toForgeL1TxsNumber := c.hdb.GetNextL1UserTxs() // TODO once HistoryDB is ready, uncomment
|
// l1UserTxs, toForgeL1TxsNumber := c.historyDB.GetNextL1UserTxs() // TODO once HistoryDB is ready, uncomment
|
||||||
var l1UserTxs []common.L1Tx = nil // tmp, depends on HistoryDB
|
var l1UserTxs []common.L1Tx = nil // tmp, depends on HistoryDB
|
||||||
l1UserTxsExtra, l1OperatorTxs, poolL2Txs, err = c.txsel.GetL1L2TxSelection([]common.Idx{}, batchNum, l1UserTxs) // TODO once feesInfo is added to method return, add the var
|
l1UserTxsExtra, l1OperatorTxs, poolL2Txs, err = p.txSelector.GetL1L2TxSelection([]common.Idx{}, batchNum, l1UserTxs) // TODO once feesInfo is added to method return, add the var
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
// 2b: only L2 txs
|
// 2b: only L2 txs
|
||||||
_, poolL2Txs, err = c.txsel.GetL2TxSelection([]common.Idx{}, batchNum) // TODO once feesInfo is added to method return, add the var
|
_, poolL2Txs, err = p.txSelector.GetL2TxSelection([]common.Idx{}, batchNum) // TODO once feesInfo is added to method return, add the var
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@@ -418,7 +614,7 @@ func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.
|
|||||||
// the poolL2Txs selected. Will mark as invalid the txs that have a
|
// the poolL2Txs selected. Will mark as invalid the txs that have a
|
||||||
// (fromIdx, nonce) which already appears in the selected txs (includes
|
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||||
// all the nonces smaller than the current one)
|
// all the nonces smaller than the current one)
|
||||||
err = c.purgeInvalidDueToL2TxsSelection(poolL2Txs)
|
err = p.purgeInvalidDueToL2TxsSelection(poolL2Txs)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@@ -431,18 +627,19 @@ func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.
|
|||||||
|
|
||||||
// 4. Call BatchBuilder with TxSelector output
|
// 4. Call BatchBuilder with TxSelector output
|
||||||
configBatch := &batchbuilder.ConfigBatch{
|
configBatch := &batchbuilder.ConfigBatch{
|
||||||
ForgerAddress: c.cfg.ForgerAddress,
|
ForgerAddress: p.cfg.ForgerAddress,
|
||||||
}
|
}
|
||||||
zkInputs, err := c.batchBuilder.BuildBatch([]common.Idx{}, configBatch, l1UserTxsExtra, l1OperatorTxs, poolL2Txs, nil) // TODO []common.TokenID --> feesInfo
|
zkInputs, err := p.batchBuilder.BuildBatch([]common.Idx{}, configBatch, l1UserTxsExtra, l1OperatorTxs, poolL2Txs, nil) // TODO []common.TokenID --> feesInfo
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// 5. Save metadata from BatchBuilder output for BatchNum
|
// 5. Save metadata from BatchBuilder output for BatchNum
|
||||||
batchInfo.ZKInputs = zkInputs
|
batchInfo.ZKInputs = zkInputs
|
||||||
|
p.cfg.debugBatchStore(&batchInfo)
|
||||||
|
|
||||||
// 6. Wait for an available server proof blocking call
|
// 6. Wait for an available server proof blocking call
|
||||||
serverProof, err := c.serverProofPool.Get(ctx)
|
serverProof, err := p.serverProofPool.Get(ctx)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
@@ -451,9 +648,10 @@ func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.
|
|||||||
// If there's an error further on, add the serverProof back to
|
// If there's an error further on, add the serverProof back to
|
||||||
// the pool
|
// the pool
|
||||||
if err != nil {
|
if err != nil {
|
||||||
c.serverProofPool.Add(serverProof)
|
p.serverProofPool.Add(serverProof)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
p.cfg.debugBatchStore(&batchInfo)
|
||||||
|
|
||||||
// 7. Call the selected idle server proof with BatchBuilder output,
|
// 7. Call the selected idle server proof with BatchBuilder output,
|
||||||
// save server proof info for batchNum
|
// save server proof info for batchNum
|
||||||
@@ -466,35 +664,20 @@ func (c *Coordinator) forgeSendServerProof(ctx context.Context, batchNum common.
|
|||||||
}
|
}
|
||||||
|
|
||||||
// waitServerProof gets the generated zkProof & sends it to the SmartContract
|
// waitServerProof gets the generated zkProof & sends it to the SmartContract
|
||||||
func (c *Coordinator) waitServerProof(ctx context.Context, batchInfo *BatchInfo) error {
|
func (p *Pipeline) waitServerProof(ctx context.Context, batchInfo *BatchInfo) error {
|
||||||
proof, err := batchInfo.ServerProof.GetProof(ctx) // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
proof, err := batchInfo.ServerProof.GetProof(ctx) // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
c.serverProofPool.Add(batchInfo.ServerProof)
|
p.serverProofPool.Add(batchInfo.ServerProof)
|
||||||
batchInfo.ServerProof = nil
|
batchInfo.ServerProof = nil
|
||||||
batchInfo.Proof = proof
|
batchInfo.Proof = proof
|
||||||
batchInfo.ForgeBatchArgs = c.prepareForgeBatchArgs(batchInfo)
|
batchInfo.ForgeBatchArgs = p.prepareForgeBatchArgs(batchInfo)
|
||||||
batchInfo.TxStatus = TxStatusPending
|
batchInfo.TxStatus = TxStatusPending
|
||||||
|
p.cfg.debugBatchStore(batchInfo)
|
||||||
// TODO(FUTURE) once tx data type is defined, store ethTx (returned by ForgeCall)
|
|
||||||
// TBD if use ethTxStore as a disk k-v database, or use a Queue
|
|
||||||
// tx, err := c.ethTxStore.NewTx()
|
|
||||||
// if err != nil {
|
|
||||||
// return err
|
|
||||||
// }
|
|
||||||
// tx.Put(ethTx.Hash(), ethTx.Bytes())
|
|
||||||
// if err := tx.Commit(); err!=nil {
|
|
||||||
// return nil
|
|
||||||
// }
|
|
||||||
|
|
||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Coordinator) handleReorg() error {
|
|
||||||
return nil // TODO
|
|
||||||
}
|
|
||||||
|
|
||||||
// isForgeSequence returns true if the node is the Forger in the current ethereum block
|
// isForgeSequence returns true if the node is the Forger in the current ethereum block
|
||||||
// func (c *Coordinator) isForgeSequence() (bool, error) {
|
// func (c *Coordinator) isForgeSequence() (bool, error) {
|
||||||
// // TODO: Consider checking if we can forge by quering the Synchronizer instead of using ethClient
|
// // TODO: Consider checking if we can forge by quering the Synchronizer instead of using ethClient
|
||||||
@@ -509,19 +692,31 @@ func (c *Coordinator) handleReorg() error {
|
|||||||
// return c.ethClient.AuctionCanForge(*addr, blockNum+1)
|
// return c.ethClient.AuctionCanForge(*addr, blockNum+1)
|
||||||
// }
|
// }
|
||||||
|
|
||||||
func (c *Coordinator) purgeRemoveByTimeout() error {
|
func (p *Pipeline) purgeRemoveByTimeout() error {
|
||||||
return nil // TODO
|
return nil // TODO
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Coordinator) purgeInvalidDueToL2TxsSelection(l2Txs []common.PoolL2Tx) error {
|
func (p *Pipeline) purgeInvalidDueToL2TxsSelection(l2Txs []common.PoolL2Tx) error {
|
||||||
return nil // TODO
|
return nil // TODO
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Coordinator) shouldL1L2Batch() bool {
|
func (p *Pipeline) shouldL1L2Batch() bool {
|
||||||
return false // TODO
|
// Take the lastL1BatchBlockNum as the biggest between the last
|
||||||
|
// scheduled one, and the synchronized one.
|
||||||
|
lastL1BatchBlockNum := p.lastScheduledL1BatchBlockNum
|
||||||
|
if p.stats.Sync.LastL1BatchBlock > lastL1BatchBlockNum {
|
||||||
|
lastL1BatchBlockNum = p.stats.Sync.LastL1BatchBlock
|
||||||
|
}
|
||||||
|
// Return true if we have passed the l1BatchTimeoutPerc portion of the
|
||||||
|
// range before the l1batch timeout.
|
||||||
|
if p.stats.Eth.LastBlock.Num-lastL1BatchBlockNum >=
|
||||||
|
int64(float64(p.vars.Rollup.ForgeL1L2BatchTimeout)*p.cfg.L1BatchTimeoutPerc) {
|
||||||
|
return true
|
||||||
|
}
|
||||||
|
return false
|
||||||
}
|
}
|
||||||
|
|
||||||
func (c *Coordinator) prepareForgeBatchArgs(batchInfo *BatchInfo) *eth.RollupForgeBatchArgs {
|
func (p *Pipeline) prepareForgeBatchArgs(batchInfo *BatchInfo) *eth.RollupForgeBatchArgs {
|
||||||
// TODO
|
// TODO
|
||||||
return ð.RollupForgeBatchArgs{}
|
return ð.RollupForgeBatchArgs{}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -10,7 +10,6 @@ import (
|
|||||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
||||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
|
||||||
"github.com/hermeznetwork/hermez-node/db/l2db"
|
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||||
"github.com/hermeznetwork/hermez-node/log"
|
"github.com/hermeznetwork/hermez-node/log"
|
||||||
@@ -21,13 +20,25 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var deleteme = []string{}
|
||||||
|
|
||||||
|
func TestMain(m *testing.M) {
|
||||||
|
exitVal := m.Run()
|
||||||
|
for _, dir := range deleteme {
|
||||||
|
if err := os.RemoveAll(dir); err != nil {
|
||||||
|
panic(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
os.Exit(exitVal)
|
||||||
|
}
|
||||||
|
|
||||||
func newTestModules(t *testing.T) (*txselector.TxSelector, *batchbuilder.BatchBuilder) { // FUTURE once Synchronizer is ready, should return it also
|
func newTestModules(t *testing.T) (*txselector.TxSelector, *batchbuilder.BatchBuilder) { // FUTURE once Synchronizer is ready, should return it also
|
||||||
nLevels := 32
|
nLevels := 32
|
||||||
|
|
||||||
synchDBPath, err := ioutil.TempDir("", "tmpSynchDB")
|
syncDBPath, err := ioutil.TempDir("", "tmpSyncDB")
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
defer assert.Nil(t, os.RemoveAll(synchDBPath))
|
deleteme = append(deleteme, syncDBPath)
|
||||||
synchSdb, err := statedb.NewStateDB(synchDBPath, statedb.TypeSynchronizer, nLevels)
|
syncSdb, err := statedb.NewStateDB(syncDBPath, statedb.TypeSynchronizer, nLevels)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
|
|
||||||
pass := os.Getenv("POSTGRES_PASS")
|
pass := os.Getenv("POSTGRES_PASS")
|
||||||
@@ -37,14 +48,14 @@ func newTestModules(t *testing.T) (*txselector.TxSelector, *batchbuilder.BatchBu
|
|||||||
|
|
||||||
txselDir, err := ioutil.TempDir("", "tmpTxSelDB")
|
txselDir, err := ioutil.TempDir("", "tmpTxSelDB")
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
defer assert.Nil(t, os.RemoveAll(txselDir))
|
deleteme = append(deleteme, txselDir)
|
||||||
txsel, err := txselector.NewTxSelector(txselDir, synchSdb, l2DB, 10, 10, 10)
|
txsel, err := txselector.NewTxSelector(txselDir, syncSdb, l2DB, 10, 10, 10)
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
|
|
||||||
bbDir, err := ioutil.TempDir("", "tmpBatchBuilderDB")
|
bbDir, err := ioutil.TempDir("", "tmpBatchBuilderDB")
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
defer assert.Nil(t, os.RemoveAll(bbDir))
|
deleteme = append(deleteme, bbDir)
|
||||||
bb, err := batchbuilder.NewBatchBuilder(bbDir, synchSdb, nil, 0, uint64(nLevels))
|
bb, err := batchbuilder.NewBatchBuilder(bbDir, syncSdb, nil, 0, uint64(nLevels))
|
||||||
assert.Nil(t, err)
|
assert.Nil(t, err)
|
||||||
|
|
||||||
// l1Txs, coordinatorL1Txs, poolL2Txs := test.GenerateTestTxsFromSet(t, test.SetTest0)
|
// l1Txs, coordinatorL1Txs, poolL2Txs := test.GenerateTestTxsFromSet(t, test.SetTest0)
|
||||||
@@ -62,58 +73,27 @@ func (t *timer) Time() int64 {
|
|||||||
return currentTime
|
return currentTime
|
||||||
}
|
}
|
||||||
|
|
||||||
var forger ethCommon.Address
|
var bidder = ethCommon.HexToAddress("0x6b175474e89094c44da98b954eedeac495271d0f")
|
||||||
var bidder ethCommon.Address
|
var forger = ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
||||||
|
|
||||||
func waitForSlot(t *testing.T, coord *Coordinator, c *test.Client, slot int64) {
|
func newTestCoordinator(t *testing.T, forgerAddr ethCommon.Address, ethClient *test.Client, ethClientSetup *test.ClientSetup) *Coordinator {
|
||||||
for {
|
|
||||||
blockNum, err := c.EthLastBlock()
|
|
||||||
require.Nil(t, err)
|
|
||||||
nextBlockSlot, err := c.AuctionGetSlotNumber(blockNum + 1)
|
|
||||||
require.Nil(t, err)
|
|
||||||
if nextBlockSlot == slot {
|
|
||||||
break
|
|
||||||
}
|
|
||||||
c.CtlMineBlock()
|
|
||||||
time.Sleep(100 * time.Millisecond)
|
|
||||||
var stats synchronizer.Stats
|
|
||||||
stats.Eth.LastBlock = c.CtlLastBlock()
|
|
||||||
stats.Sync.LastBlock = c.CtlLastBlock()
|
|
||||||
canForge, err := c.AuctionCanForge(forger, blockNum+1)
|
|
||||||
require.Nil(t, err)
|
|
||||||
if canForge {
|
|
||||||
// fmt.Println("DBG canForge")
|
|
||||||
stats.Sync.Auction.CurrentSlot.Forger = forger
|
|
||||||
}
|
|
||||||
coord.SendMsg(MsgSyncStats{
|
|
||||||
Stats: stats,
|
|
||||||
})
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestCoordinator(t *testing.T) {
|
|
||||||
txsel, bb := newTestModules(t)
|
txsel, bb := newTestModules(t)
|
||||||
bidder = ethCommon.HexToAddress("0x6b175474e89094c44da98b954eedeac495271d0f")
|
|
||||||
forger = ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
debugBatchPath, err := ioutil.TempDir("", "tmpDebugBatch")
|
||||||
|
require.Nil(t, err)
|
||||||
|
deleteme = append(deleteme, debugBatchPath)
|
||||||
|
|
||||||
conf := Config{
|
conf := Config{
|
||||||
ForgerAddress: forger,
|
ForgerAddress: forgerAddr,
|
||||||
|
ConfirmBlocks: 5,
|
||||||
|
L1BatchTimeoutPerc: 0.5,
|
||||||
|
EthClientAttempts: 5,
|
||||||
|
EthClientAttemptsDelay: 100 * time.Millisecond,
|
||||||
|
TxManagerCheckInterval: 500 * time.Millisecond,
|
||||||
|
DebugBatchPath: debugBatchPath,
|
||||||
}
|
}
|
||||||
hdb := &historydb.HistoryDB{}
|
|
||||||
serverProofs := []ServerProofInterface{&ServerProofMock{}, &ServerProofMock{}}
|
serverProofs := []ServerProofInterface{&ServerProofMock{}, &ServerProofMock{}}
|
||||||
|
|
||||||
var timer timer
|
|
||||||
ethClientSetup := test.NewClientSetupExample()
|
|
||||||
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
|
||||||
|
|
||||||
// Bid for slot 2 and 4
|
|
||||||
_, err := ethClient.AuctionSetCoordinator(forger, "https://foo.bar")
|
|
||||||
require.Nil(t, err)
|
|
||||||
_, err = ethClient.AuctionBidSimple(2, big.NewInt(9999))
|
|
||||||
require.Nil(t, err)
|
|
||||||
_, err = ethClient.AuctionBidSimple(4, big.NewInt(9999))
|
|
||||||
require.Nil(t, err)
|
|
||||||
|
|
||||||
scConsts := &synchronizer.SCConsts{
|
scConsts := &synchronizer.SCConsts{
|
||||||
Rollup: *ethClientSetup.RollupConstants,
|
Rollup: *ethClientSetup.RollupConstants,
|
||||||
Auction: *ethClientSetup.AuctionConstants,
|
Auction: *ethClientSetup.AuctionConstants,
|
||||||
@@ -124,30 +104,253 @@ func TestCoordinator(t *testing.T) {
|
|||||||
Auction: *ethClientSetup.AuctionVariables,
|
Auction: *ethClientSetup.AuctionVariables,
|
||||||
WDelayer: *ethClientSetup.WDelayerVariables,
|
WDelayer: *ethClientSetup.WDelayerVariables,
|
||||||
}
|
}
|
||||||
c := NewCoordinator(conf, hdb, txsel, bb, serverProofs, ethClient, scConsts, initSCVars)
|
coord, err := NewCoordinator(conf, nil, txsel, bb, serverProofs, ethClient, scConsts, initSCVars)
|
||||||
c.Start()
|
require.Nil(t, err)
|
||||||
|
return coord
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCoordinatorFlow(t *testing.T) {
|
||||||
|
if os.Getenv("TEST_COORD_FLOW") == "" {
|
||||||
|
return
|
||||||
|
}
|
||||||
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
|
var timer timer
|
||||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
||||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup)
|
||||||
|
|
||||||
|
// Bid for slot 2 and 4
|
||||||
|
_, err := ethClient.AuctionSetCoordinator(forger, "https://foo.bar")
|
||||||
|
require.Nil(t, err)
|
||||||
|
_, err = ethClient.AuctionBidSimple(2, big.NewInt(9999))
|
||||||
|
require.Nil(t, err)
|
||||||
|
_, err = ethClient.AuctionBidSimple(4, big.NewInt(9999))
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
coord.Start()
|
||||||
time.Sleep(1 * time.Second)
|
time.Sleep(1 * time.Second)
|
||||||
|
|
||||||
|
waitForSlot := func(slot int64) {
|
||||||
|
for {
|
||||||
|
blockNum, err := ethClient.EthLastBlock()
|
||||||
|
require.Nil(t, err)
|
||||||
|
nextBlockSlot, err := ethClient.AuctionGetSlotNumber(blockNum + 1)
|
||||||
|
require.Nil(t, err)
|
||||||
|
if nextBlockSlot == slot {
|
||||||
|
break
|
||||||
|
}
|
||||||
|
ethClient.CtlMineBlock()
|
||||||
|
time.Sleep(100 * time.Millisecond)
|
||||||
|
var stats synchronizer.Stats
|
||||||
|
stats.Eth.LastBlock = *ethClient.CtlLastBlock()
|
||||||
|
stats.Sync.LastBlock = *ethClient.CtlLastBlock()
|
||||||
|
canForge, err := ethClient.AuctionCanForge(forger, blockNum+1)
|
||||||
|
require.Nil(t, err)
|
||||||
|
if canForge {
|
||||||
|
// fmt.Println("DBG canForge")
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = forger
|
||||||
|
}
|
||||||
|
coord.SendMsg(MsgSyncStats{
|
||||||
|
Stats: stats,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// NOTE: With the current test, the coordinator will enter in forge
|
// NOTE: With the current test, the coordinator will enter in forge
|
||||||
// time before the bidded slot because no one else is forging in the
|
// time before the bidded slot because no one else is forging in the
|
||||||
// other slots before the slot deadline.
|
// other slots before the slot deadline.
|
||||||
// simulate forgeSequence time
|
// simulate forgeSequence time
|
||||||
waitForSlot(t, c, ethClient, 2)
|
waitForSlot(2)
|
||||||
log.Info("~~~ simulate entering in forge time")
|
log.Info("~~~ simulate entering in forge time")
|
||||||
time.Sleep(1 * time.Second)
|
time.Sleep(1 * time.Second)
|
||||||
|
|
||||||
// simulate going out from forgeSequence
|
// simulate going out from forgeSequence
|
||||||
waitForSlot(t, c, ethClient, 3)
|
waitForSlot(3)
|
||||||
log.Info("~~~ simulate going out from forge time")
|
log.Info("~~~ simulate going out from forge time")
|
||||||
time.Sleep(1 * time.Second)
|
time.Sleep(1 * time.Second)
|
||||||
|
|
||||||
// simulate entering forgeSequence time again
|
// simulate entering forgeSequence time again
|
||||||
waitForSlot(t, c, ethClient, 4)
|
waitForSlot(4)
|
||||||
log.Info("~~~ simulate entering in forge time again")
|
log.Info("~~~ simulate entering in forge time again")
|
||||||
time.Sleep(2 * time.Second)
|
time.Sleep(2 * time.Second)
|
||||||
|
|
||||||
// simulate stopping forgerLoop by channel
|
// simulate stopping forgerLoop by channel
|
||||||
log.Info("~~~ simulate stopping forgerLoop by closing coordinator stopch")
|
log.Info("~~~ simulate stopping forgerLoop by closing coordinator stopch")
|
||||||
c.Stop()
|
coord.Stop()
|
||||||
time.Sleep(1 * time.Second)
|
time.Sleep(1 * time.Second)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestCoordinatorStartStop(t *testing.T) {
|
||||||
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
|
var timer timer
|
||||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
||||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup)
|
||||||
|
coord.Start()
|
||||||
|
coord.Stop()
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCoordCanForge(t *testing.T) {
|
||||||
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
|
bootForger := ethClientSetup.AuctionVariables.BootCoordinator
|
||||||
|
|
||||||
|
var timer timer
|
||||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
||||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup)
|
||||||
|
_, err := ethClient.AuctionSetCoordinator(forger, "https://foo.bar")
|
||||||
|
require.Nil(t, err)
|
||||||
|
_, err = ethClient.AuctionBidSimple(2, big.NewInt(9999))
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
bootCoord := newTestCoordinator(t, bootForger, ethClient, ethClientSetup)
|
||||||
|
|
||||||
|
assert.Equal(t, forger, coord.cfg.ForgerAddress)
|
||||||
|
assert.Equal(t, bootForger, bootCoord.cfg.ForgerAddress)
|
||||||
|
ethBootCoord, err := ethClient.AuctionGetBootCoordinator()
|
||||||
|
require.Nil(t, err)
|
||||||
|
assert.Equal(t, &bootForger, ethBootCoord)
|
||||||
|
|
||||||
|
var stats synchronizer.Stats
|
||||||
|
|
||||||
|
// Slot 0. No bid, so the winner is the boot coordinator
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, false, coord.canForge(&stats))
|
||||||
|
assert.Equal(t, true, bootCoord.canForge(&stats))
|
||||||
|
|
||||||
|
// Slot 0. No bid, and we reach the deadline, so anyone can forge
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||||
|
int64(ethClientSetup.AuctionVariables.SlotDeadline)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, true, coord.canForge(&stats))
|
||||||
|
assert.Equal(t, true, bootCoord.canForge(&stats))
|
||||||
|
|
||||||
|
// Slot 1. coordinator bid, so the winner is the coordinator
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||||
|
1*int64(ethClientSetup.AuctionConstants.BlocksPerSlot)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = forger
|
||||||
|
assert.Equal(t, true, coord.canForge(&stats))
|
||||||
|
assert.Equal(t, false, bootCoord.canForge(&stats))
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestCoordHandleMsgSyncStats(t *testing.T) {
|
||||||
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
|
bootForger := ethClientSetup.AuctionVariables.BootCoordinator
|
||||||
|
|
||||||
|
var timer timer
|
||||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
||||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup)
|
||||||
|
_, err := ethClient.AuctionSetCoordinator(forger, "https://foo.bar")
|
||||||
|
require.Nil(t, err)
|
||||||
|
_, err = ethClient.AuctionBidSimple(2, big.NewInt(9999))
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
var stats synchronizer.Stats
|
||||||
|
|
||||||
|
// Slot 0. No bid, so the winner is the boot coordinator
|
||||||
|
// pipelineStarted: false -> false
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, false, coord.canForge(&stats))
|
||||||
|
require.Nil(t, coord.handleMsgSyncStats(&stats))
|
||||||
|
assert.Nil(t, coord.pipeline)
|
||||||
|
|
||||||
|
// Slot 0. No bid, and we reach the deadline, so anyone can forge
|
||||||
|
// pipelineStarted: false -> true
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||||
|
int64(ethClientSetup.AuctionVariables.SlotDeadline)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, true, coord.canForge(&stats))
|
||||||
|
require.Nil(t, coord.handleMsgSyncStats(&stats))
|
||||||
|
assert.NotNil(t, coord.pipeline)
|
||||||
|
|
||||||
|
// Slot 0. No bid, and we reach the deadline, so anyone can forge
|
||||||
|
// pipelineStarted: true -> true
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||||
|
int64(ethClientSetup.AuctionVariables.SlotDeadline) + 1
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, true, coord.canForge(&stats))
|
||||||
|
require.Nil(t, coord.handleMsgSyncStats(&stats))
|
||||||
|
assert.NotNil(t, coord.pipeline)
|
||||||
|
|
||||||
|
// Slot 0. No bid, so the winner is the boot coordinator
|
||||||
|
// pipelineStarted: true -> false
|
||||||
|
stats.Eth.LastBlock.Num = ethClientSetup.AuctionConstants.GenesisBlockNum +
|
||||||
|
1*int64(ethClientSetup.AuctionConstants.BlocksPerSlot)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.Auction.CurrentSlot.Forger = bootForger
|
||||||
|
assert.Equal(t, false, coord.canForge(&stats))
|
||||||
|
require.Nil(t, coord.handleMsgSyncStats(&stats))
|
||||||
|
assert.Nil(t, coord.pipeline)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestPipelineShouldL1L2Batch(t *testing.T) {
|
||||||
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
|
|
||||||
|
var timer timer
|
||||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup)
|
||||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup)
|
||||||
|
pipeline := coord.newPipeline()
|
||||||
|
pipeline.vars = coord.vars
|
||||||
|
|
||||||
|
// Check that the parameters are the ones we expect and use in this test
|
||||||
|
require.Equal(t, 0.5, pipeline.cfg.L1BatchTimeoutPerc)
|
||||||
|
require.Equal(t, int64(9), ethClientSetup.RollupVariables.ForgeL1L2BatchTimeout)
|
||||||
|
l1BatchTimeoutPerc := pipeline.cfg.L1BatchTimeoutPerc
|
||||||
|
l1BatchTimeout := ethClientSetup.RollupVariables.ForgeL1L2BatchTimeout
|
||||||
|
|
||||||
|
var stats synchronizer.Stats
|
||||||
|
|
||||||
|
startBlock := int64(100)
|
||||||
|
|
||||||
|
//
|
||||||
|
// No scheduled L1Batch
|
||||||
|
//
|
||||||
|
|
||||||
|
// Last L1Batch was a long time ago
|
||||||
|
stats.Eth.LastBlock.Num = startBlock
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
stats.Sync.LastL1BatchBlock = 0
|
||||||
|
pipeline.stats = stats
|
||||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch())
|
||||||
|
|
||||||
|
stats.Sync.LastL1BatchBlock = startBlock
|
||||||
|
|
||||||
|
// We are are one block before the timeout range * 0.5
|
||||||
|
stats.Eth.LastBlock.Num = startBlock + int64(float64(l1BatchTimeout)*l1BatchTimeoutPerc) - 1
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
pipeline.stats = stats
|
||||||
|
assert.Equal(t, false, pipeline.shouldL1L2Batch())
|
||||||
|
|
||||||
|
// We are are at timeout range * 0.5
|
||||||
|
stats.Eth.LastBlock.Num = startBlock + int64(float64(l1BatchTimeout)*l1BatchTimeoutPerc)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
pipeline.stats = stats
|
||||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch())
|
||||||
|
|
||||||
|
//
|
||||||
|
// Scheduled L1Batch
|
||||||
|
//
|
||||||
|
pipeline.lastScheduledL1BatchBlockNum = startBlock
|
||||||
|
stats.Sync.LastL1BatchBlock = startBlock - 10
|
||||||
|
|
||||||
|
// We are are one block before the timeout range * 0.5
|
||||||
|
stats.Eth.LastBlock.Num = startBlock + int64(float64(l1BatchTimeout)*l1BatchTimeoutPerc) - 1
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
pipeline.stats = stats
|
||||||
|
assert.Equal(t, false, pipeline.shouldL1L2Batch())
|
||||||
|
|
||||||
|
// We are are at timeout range * 0.5
|
||||||
|
stats.Eth.LastBlock.Num = startBlock + int64(float64(l1BatchTimeout)*l1BatchTimeoutPerc)
|
||||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock
|
||||||
|
pipeline.stats = stats
|
||||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch())
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: Test Reorg
|
||||||
|
// TODO: Test Pipeline
|
||||||
|
// TODO: Test TxMonitor
|
||||||
|
|||||||
@@ -304,6 +304,15 @@ func (hdb *HistoryDB) GetLastBatchNum() (common.BatchNum, error) {
|
|||||||
return batchNum, row.Scan(&batchNum)
|
return batchNum, row.Scan(&batchNum)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetLastL1BatchBlockNum returns the blockNum of the latest forged l1Batch
|
||||||
|
func (hdb *HistoryDB) GetLastL1BatchBlockNum() (int64, error) {
|
||||||
|
row := hdb.db.QueryRow(`SELECT eth_block_num FROM batch
|
||||||
|
WHERE forge_l1_txs_num IS NOT NULL
|
||||||
|
ORDER BY batch_num DESC LIMIT 1;`)
|
||||||
|
var blockNum int64
|
||||||
|
return blockNum, row.Scan(&blockNum)
|
||||||
|
}
|
||||||
|
|
||||||
// GetLastL1TxsNum returns the greatest ForgeL1TxsNum in the DB from forged
|
// GetLastL1TxsNum returns the greatest ForgeL1TxsNum in the DB from forged
|
||||||
// batches. If there's no batch in the DB (nil, nil) is returned.
|
// batches. If there's no batch in the DB (nil, nil) is returned.
|
||||||
func (hdb *HistoryDB) GetLastL1TxsNum() (*int64, error) {
|
func (hdb *HistoryDB) GetLastL1TxsNum() (*int64, error) {
|
||||||
@@ -1410,7 +1419,7 @@ func (hdb *HistoryDB) AddBlockSCData(blockData *common.BlockData) (err error) {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := hdb.updateExitTree(txn, blockData.Block.EthBlockNum,
|
if err := hdb.updateExitTree(txn, blockData.Block.Num,
|
||||||
blockData.Rollup.Withdrawals, blockData.WDelayer.Withdrawals); err != nil {
|
blockData.Rollup.Withdrawals, blockData.WDelayer.Withdrawals); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -107,7 +107,7 @@ func TestBlocks(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func assertEqualBlock(t *testing.T, expected *common.Block, actual *common.Block) {
|
func assertEqualBlock(t *testing.T, expected *common.Block, actual *common.Block) {
|
||||||
assert.Equal(t, expected.EthBlockNum, actual.EthBlockNum)
|
assert.Equal(t, expected.Num, actual.Num)
|
||||||
assert.Equal(t, expected.Hash, actual.Hash)
|
assert.Equal(t, expected.Hash, actual.Hash)
|
||||||
assert.Equal(t, expected.Timestamp.Unix(), actual.Timestamp.Unix())
|
assert.Equal(t, expected.Timestamp.Unix(), actual.Timestamp.Unix())
|
||||||
}
|
}
|
||||||
@@ -150,6 +150,7 @@ func TestBatches(t *testing.T) {
|
|||||||
batches := []common.Batch{}
|
batches := []common.Batch{}
|
||||||
tokensValue := make(map[common.TokenID]float64)
|
tokensValue := make(map[common.TokenID]float64)
|
||||||
lastL1TxsNum := new(int64)
|
lastL1TxsNum := new(int64)
|
||||||
|
lastL1BatchBlockNum := int64(0)
|
||||||
for _, block := range blocks {
|
for _, block := range blocks {
|
||||||
// Insert block
|
// Insert block
|
||||||
assert.NoError(t, historyDB.AddBlock(&block.Block))
|
assert.NoError(t, historyDB.AddBlock(&block.Block))
|
||||||
@@ -169,6 +170,7 @@ func TestBatches(t *testing.T) {
|
|||||||
forgeTxsNum := batch.Batch.ForgeL1TxsNum
|
forgeTxsNum := batch.Batch.ForgeL1TxsNum
|
||||||
if forgeTxsNum != nil && (lastL1TxsNum == nil || *lastL1TxsNum < *forgeTxsNum) {
|
if forgeTxsNum != nil && (lastL1TxsNum == nil || *lastL1TxsNum < *forgeTxsNum) {
|
||||||
*lastL1TxsNum = *forgeTxsNum
|
*lastL1TxsNum = *forgeTxsNum
|
||||||
|
lastL1BatchBlockNum = batch.Batch.EthBlockNum
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -199,6 +201,10 @@ func TestBatches(t *testing.T) {
|
|||||||
fetchedLastL1TxsNum, err := historyDB.GetLastL1TxsNum()
|
fetchedLastL1TxsNum, err := historyDB.GetLastL1TxsNum()
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, lastL1TxsNum, fetchedLastL1TxsNum)
|
assert.Equal(t, lastL1TxsNum, fetchedLastL1TxsNum)
|
||||||
|
// Test GetLastL1BatchBlockNum
|
||||||
|
fetchedLastL1BatchBlockNum, err := historyDB.GetLastL1BatchBlockNum()
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, lastL1BatchBlockNum, fetchedLastL1BatchBlockNum)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestBids(t *testing.T) {
|
func TestBids(t *testing.T) {
|
||||||
@@ -735,7 +741,7 @@ func TestUpdateExitTree(t *testing.T) {
|
|||||||
|
|
||||||
// Add withdraws to the second-to-last block, and insert block into the DB
|
// Add withdraws to the second-to-last block, and insert block into the DB
|
||||||
block := &blocks[len(blocks)-2]
|
block := &blocks[len(blocks)-2]
|
||||||
require.Equal(t, int64(4), block.Block.EthBlockNum)
|
require.Equal(t, int64(4), block.Block.Num)
|
||||||
tokenAddr := blocks[0].Rollup.AddedTokens[0].EthAddr
|
tokenAddr := blocks[0].Rollup.AddedTokens[0].EthAddr
|
||||||
// block.WDelayer.Deposits = append(block.WDelayer.Deposits,
|
// block.WDelayer.Deposits = append(block.WDelayer.Deposits,
|
||||||
// common.WDelayerTransfer{Owner: tc.UsersByIdx[257].Addr, Token: tokenAddr, Amount: big.NewInt(80)}, // 257
|
// common.WDelayerTransfer{Owner: tc.UsersByIdx[257].Addr, Token: tokenAddr, Amount: big.NewInt(80)}, // 257
|
||||||
@@ -752,7 +758,7 @@ func TestUpdateExitTree(t *testing.T) {
|
|||||||
err = historyDB.addBlock(historyDB.db, &block.Block)
|
err = historyDB.addBlock(historyDB.db, &block.Block)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
|
|
||||||
err = historyDB.updateExitTree(historyDB.db, block.Block.EthBlockNum,
|
err = historyDB.updateExitTree(historyDB.db, block.Block.Num,
|
||||||
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
|
|
||||||
@@ -767,15 +773,15 @@ func TestUpdateExitTree(t *testing.T) {
|
|||||||
for _, withdraw := range block.Rollup.Withdrawals {
|
for _, withdraw := range block.Rollup.Withdrawals {
|
||||||
assert.Equal(t, withdraw.NumExitRoot, dbExitsByIdx[withdraw.Idx].BatchNum)
|
assert.Equal(t, withdraw.NumExitRoot, dbExitsByIdx[withdraw.Idx].BatchNum)
|
||||||
if withdraw.InstantWithdraw {
|
if withdraw.InstantWithdraw {
|
||||||
assert.Equal(t, &block.Block.EthBlockNum, dbExitsByIdx[withdraw.Idx].InstantWithdrawn)
|
assert.Equal(t, &block.Block.Num, dbExitsByIdx[withdraw.Idx].InstantWithdrawn)
|
||||||
} else {
|
} else {
|
||||||
assert.Equal(t, &block.Block.EthBlockNum, dbExitsByIdx[withdraw.Idx].DelayedWithdrawRequest)
|
assert.Equal(t, &block.Block.Num, dbExitsByIdx[withdraw.Idx].DelayedWithdrawRequest)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Add delayed withdraw to the last block, and insert block into the DB
|
// Add delayed withdraw to the last block, and insert block into the DB
|
||||||
block = &blocks[len(blocks)-1]
|
block = &blocks[len(blocks)-1]
|
||||||
require.Equal(t, int64(5), block.Block.EthBlockNum)
|
require.Equal(t, int64(5), block.Block.Num)
|
||||||
block.WDelayer.Withdrawals = append(block.WDelayer.Withdrawals,
|
block.WDelayer.Withdrawals = append(block.WDelayer.Withdrawals,
|
||||||
common.WDelayerTransfer{
|
common.WDelayerTransfer{
|
||||||
Owner: tc.UsersByIdx[257].Addr,
|
Owner: tc.UsersByIdx[257].Addr,
|
||||||
@@ -785,7 +791,7 @@ func TestUpdateExitTree(t *testing.T) {
|
|||||||
err = historyDB.addBlock(historyDB.db, &block.Block)
|
err = historyDB.addBlock(historyDB.db, &block.Block)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
|
|
||||||
err = historyDB.updateExitTree(historyDB.db, block.Block.EthBlockNum,
|
err = historyDB.updateExitTree(historyDB.db, block.Block.Num,
|
||||||
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
block.Rollup.Withdrawals, block.WDelayer.Withdrawals)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
|
|
||||||
@@ -795,7 +801,7 @@ func TestUpdateExitTree(t *testing.T) {
|
|||||||
for _, dbExit := range dbExits {
|
for _, dbExit := range dbExits {
|
||||||
dbExitsByIdx[dbExit.AccountIdx] = dbExit
|
dbExitsByIdx[dbExit.AccountIdx] = dbExit
|
||||||
}
|
}
|
||||||
require.Equal(t, &block.Block.EthBlockNum, dbExitsByIdx[257].DelayedWithdrawn)
|
require.Equal(t, &block.Block.Num, dbExitsByIdx[257].DelayedWithdrawn)
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestGetBestBidCoordinator(t *testing.T) {
|
func TestGetBestBidCoordinator(t *testing.T) {
|
||||||
|
|||||||
@@ -257,10 +257,11 @@ func (c *EthereumClient) EthLastBlock() (int64, error) {
|
|||||||
// return c.client.HeaderByNumber(ctx, number)
|
// return c.client.HeaderByNumber(ctx, number)
|
||||||
// }
|
// }
|
||||||
|
|
||||||
// EthBlockByNumber internally calls ethclient.Client BlockByNumber and returns *common.Block
|
// EthBlockByNumber internally calls ethclient.Client BlockByNumber and returns
|
||||||
|
// *common.Block. If number == -1, the latests known block is returned.
|
||||||
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block, error) {
|
func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*common.Block, error) {
|
||||||
blockNum := big.NewInt(number)
|
blockNum := big.NewInt(number)
|
||||||
if number == 0 {
|
if number == -1 {
|
||||||
blockNum = nil
|
blockNum = nil
|
||||||
}
|
}
|
||||||
block, err := c.client.BlockByNumber(ctx, blockNum)
|
block, err := c.client.BlockByNumber(ctx, blockNum)
|
||||||
@@ -268,7 +269,7 @@ func (c *EthereumClient) EthBlockByNumber(ctx context.Context, number int64) (*c
|
|||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
b := &common.Block{
|
b := &common.Block{
|
||||||
EthBlockNum: block.Number().Int64(),
|
Num: block.Number().Int64(),
|
||||||
Timestamp: time.Unix(int64(block.Time()), 0),
|
Timestamp: time.Unix(int64(block.Time()), 0),
|
||||||
ParentHash: block.ParentHash(),
|
ParentHash: block.ParentHash(),
|
||||||
Hash: block.Hash(),
|
Hash: block.Hash(),
|
||||||
|
|||||||
68
node/node.go
68
node/node.go
@@ -2,6 +2,7 @@ package node
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"fmt"
|
||||||
"net/http"
|
"net/http"
|
||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
@@ -165,7 +166,7 @@ func NewNode(mode Mode, cfg *config.Node, coordCfg *config.Coordinator) (*Node,
|
|||||||
serverProofs[i] = coordinator.NewServerProof(serverProofCfg.URL)
|
serverProofs[i] = coordinator.NewServerProof(serverProofCfg.URL)
|
||||||
}
|
}
|
||||||
|
|
||||||
coord = coordinator.NewCoordinator(
|
coord, err = coordinator.NewCoordinator(
|
||||||
coordinator.Config{
|
coordinator.Config{
|
||||||
ForgerAddress: coordCfg.ForgerAddress,
|
ForgerAddress: coordCfg.ForgerAddress,
|
||||||
ConfirmBlocks: coordCfg.ConfirmBlocks,
|
ConfirmBlocks: coordCfg.ConfirmBlocks,
|
||||||
@@ -178,9 +179,20 @@ func NewNode(mode Mode, cfg *config.Node, coordCfg *config.Coordinator) (*Node,
|
|||||||
&scConsts,
|
&scConsts,
|
||||||
&initSCVars,
|
&initSCVars,
|
||||||
)
|
)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
}
|
}
|
||||||
var nodeAPI *NodeAPI
|
var nodeAPI *NodeAPI
|
||||||
if cfg.API.Address != "" {
|
if cfg.API.Address != "" {
|
||||||
|
if cfg.API.UpdateMetricsInterval.Duration == 0 {
|
||||||
|
return nil, fmt.Errorf("invalid cfg.API.UpdateMetricsInterval: %v",
|
||||||
|
cfg.API.UpdateMetricsInterval.Duration)
|
||||||
|
}
|
||||||
|
if cfg.API.UpdateRecommendedFeeInterval.Duration == 0 {
|
||||||
|
return nil, fmt.Errorf("invalid cfg.API.UpdateRecommendedFeeInterval: %v",
|
||||||
|
cfg.API.UpdateRecommendedFeeInterval.Duration)
|
||||||
|
}
|
||||||
server := gin.Default()
|
server := gin.Default()
|
||||||
coord := false
|
coord := false
|
||||||
if mode == ModeCoordinator {
|
if mode == ModeCoordinator {
|
||||||
@@ -303,9 +315,11 @@ func (a *NodeAPI) Run(ctx context.Context) error {
|
|||||||
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
||||||
// don't have to pass it around.
|
// don't have to pass it around.
|
||||||
func (n *Node) syncLoopFn(lastBlock *common.Block) (*common.Block, time.Duration) {
|
func (n *Node) syncLoopFn(lastBlock *common.Block) (*common.Block, time.Duration) {
|
||||||
if blockData, discarded, err := n.sync.Sync2(n.ctx, lastBlock); err != nil {
|
blockData, discarded, err := n.sync.Sync2(n.ctx, lastBlock)
|
||||||
|
stats := n.sync.Stats()
|
||||||
|
if err != nil {
|
||||||
// case: error
|
// case: error
|
||||||
log.Errorw("Synchronizer.Sync", "error", err)
|
log.Errorw("Synchronizer.Sync", "err", err)
|
||||||
return nil, n.cfg.Synchronizer.SyncLoopInterval.Duration
|
return nil, n.cfg.Synchronizer.SyncLoopInterval.Duration
|
||||||
} else if discarded != nil {
|
} else if discarded != nil {
|
||||||
// case: reorg
|
// case: reorg
|
||||||
@@ -318,13 +332,13 @@ func (n *Node) syncLoopFn(lastBlock *common.Block) (*common.Block, time.Duration
|
|||||||
n.nodeAPI.api.SetRollupVariables(*rollup)
|
n.nodeAPI.api.SetRollupVariables(*rollup)
|
||||||
n.nodeAPI.api.SetAuctionVariables(*auction)
|
n.nodeAPI.api.SetAuctionVariables(*auction)
|
||||||
n.nodeAPI.api.SetWDelayerVariables(*wDelayer)
|
n.nodeAPI.api.SetWDelayerVariables(*wDelayer)
|
||||||
|
n.nodeAPI.api.UpdateNetworkInfoBlock(
|
||||||
// TODO: n.nodeAPI.api.UpdateNetworkInfo()
|
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||||
|
)
|
||||||
}
|
}
|
||||||
return nil, time.Duration(0)
|
return nil, time.Duration(0)
|
||||||
} else if blockData != nil {
|
} else if blockData != nil {
|
||||||
// case: new block
|
// case: new block
|
||||||
stats := n.sync.Stats()
|
|
||||||
if n.mode == ModeCoordinator {
|
if n.mode == ModeCoordinator {
|
||||||
if stats.Synced() && (blockData.Rollup.Vars != nil ||
|
if stats.Synced() && (blockData.Rollup.Vars != nil ||
|
||||||
blockData.Auction.Vars != nil ||
|
blockData.Auction.Vars != nil ||
|
||||||
@@ -350,7 +364,15 @@ func (n *Node) syncLoopFn(lastBlock *common.Block) (*common.Block, time.Duration
|
|||||||
n.nodeAPI.api.SetWDelayerVariables(*blockData.WDelayer.Vars)
|
n.nodeAPI.api.SetWDelayerVariables(*blockData.WDelayer.Vars)
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO: n.nodeAPI.api.UpdateNetworkInfo()
|
if stats.Synced() {
|
||||||
|
if err := n.nodeAPI.api.UpdateNetworkInfo(
|
||||||
|
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||||
|
common.BatchNum(stats.Eth.LastBatch),
|
||||||
|
stats.Sync.Auction.CurrentSlot.SlotNum,
|
||||||
|
); err != nil {
|
||||||
|
log.Errorw("API.UpdateNetworkInfo", "err", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
return &blockData.Block, time.Duration(0)
|
return &blockData.Block, time.Duration(0)
|
||||||
} else {
|
} else {
|
||||||
@@ -408,6 +430,38 @@ func (n *Node) StartNodeAPI() {
|
|||||||
log.Fatalw("NodeAPI.Run", "err", err)
|
log.Fatalw("NodeAPI.Run", "err", err)
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|
||||||
|
n.wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-n.ctx.Done():
|
||||||
|
log.Info("API.UpdateMetrics loop done")
|
||||||
|
n.wg.Done()
|
||||||
|
return
|
||||||
|
case <-time.After(n.cfg.API.UpdateMetricsInterval.Duration):
|
||||||
|
if err := n.nodeAPI.api.UpdateMetrics(); err != nil {
|
||||||
|
log.Errorw("API.UpdateMetrics", "err", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
|
|
||||||
|
n.wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
for {
|
||||||
|
select {
|
||||||
|
case <-n.ctx.Done():
|
||||||
|
log.Info("API.UpdateRecommendedFee loop done")
|
||||||
|
n.wg.Done()
|
||||||
|
return
|
||||||
|
case <-time.After(n.cfg.API.UpdateRecommendedFeeInterval.Duration):
|
||||||
|
if err := n.nodeAPI.api.UpdateRecommendedFee(); err != nil {
|
||||||
|
log.Errorw("API.UpdateRecommendedFee", "err", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}()
|
||||||
}
|
}
|
||||||
|
|
||||||
// Start the node
|
// Start the node
|
||||||
|
|||||||
@@ -29,7 +29,7 @@ func TestPriceUpdater(t *testing.T) {
|
|||||||
tokens := []common.Token{}
|
tokens := []common.Token{}
|
||||||
tokens = append(tokens, common.Token{
|
tokens = append(tokens, common.Token{
|
||||||
TokenID: 1,
|
TokenID: 1,
|
||||||
EthBlockNum: blocks[0].EthBlockNum,
|
EthBlockNum: blocks[0].Num,
|
||||||
EthAddr: ethCommon.BigToAddress(big.NewInt(2)),
|
EthAddr: ethCommon.BigToAddress(big.NewInt(2)),
|
||||||
Name: "DAI",
|
Name: "DAI",
|
||||||
Symbol: "DAI",
|
Symbol: "DAI",
|
||||||
|
|||||||
@@ -46,14 +46,17 @@ type Stats struct {
|
|||||||
Eth struct {
|
Eth struct {
|
||||||
RefreshPeriod time.Duration
|
RefreshPeriod time.Duration
|
||||||
Updated time.Time
|
Updated time.Time
|
||||||
FirstBlock int64
|
FirstBlockNum int64
|
||||||
LastBlock int64
|
LastBlock common.Block
|
||||||
LastBatch int64
|
LastBatch int64
|
||||||
}
|
}
|
||||||
Sync struct {
|
Sync struct {
|
||||||
Updated time.Time
|
Updated time.Time
|
||||||
LastBlock int64
|
LastBlock common.Block
|
||||||
LastBatch int64
|
LastBatch int64
|
||||||
|
// LastL1BatchBlock is the last ethereum block in which an
|
||||||
|
// l1Batch was forged
|
||||||
|
LastL1BatchBlock int64
|
||||||
Auction struct {
|
Auction struct {
|
||||||
CurrentSlot common.Slot
|
CurrentSlot common.Slot
|
||||||
}
|
}
|
||||||
@@ -78,10 +81,10 @@ type StatsHolder struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// NewStatsHolder creates a new StatsHolder
|
// NewStatsHolder creates a new StatsHolder
|
||||||
func NewStatsHolder(firstBlock int64, refreshPeriod time.Duration) *StatsHolder {
|
func NewStatsHolder(firstBlockNum int64, refreshPeriod time.Duration) *StatsHolder {
|
||||||
stats := Stats{}
|
stats := Stats{}
|
||||||
stats.Eth.RefreshPeriod = refreshPeriod
|
stats.Eth.RefreshPeriod = refreshPeriod
|
||||||
stats.Eth.FirstBlock = firstBlock
|
stats.Eth.FirstBlockNum = firstBlockNum
|
||||||
return &StatsHolder{Stats: stats}
|
return &StatsHolder{Stats: stats}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -93,13 +96,16 @@ func (s *StatsHolder) UpdateCurrentSlot(slot common.Slot) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// UpdateSync updates the synchronizer stats
|
// UpdateSync updates the synchronizer stats
|
||||||
func (s *StatsHolder) UpdateSync(lastBlock int64, lastBatch *common.BatchNum) {
|
func (s *StatsHolder) UpdateSync(lastBlock *common.Block, lastBatch *common.BatchNum, lastL1BatchBlock *int64) {
|
||||||
now := time.Now()
|
now := time.Now()
|
||||||
s.rw.Lock()
|
s.rw.Lock()
|
||||||
s.Sync.LastBlock = lastBlock
|
s.Sync.LastBlock = *lastBlock
|
||||||
if lastBatch != nil {
|
if lastBatch != nil {
|
||||||
s.Sync.LastBatch = int64(*lastBatch)
|
s.Sync.LastBatch = int64(*lastBatch)
|
||||||
}
|
}
|
||||||
|
if lastL1BatchBlock != nil {
|
||||||
|
s.Sync.LastL1BatchBlock = *lastL1BatchBlock
|
||||||
|
}
|
||||||
s.Sync.Updated = now
|
s.Sync.Updated = now
|
||||||
s.rw.Unlock()
|
s.rw.Unlock()
|
||||||
}
|
}
|
||||||
@@ -114,7 +120,7 @@ func (s *StatsHolder) UpdateEth(ethClient eth.ClientInterface) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
lastBlock, err := ethClient.EthLastBlock()
|
lastBlock, err := ethClient.EthBlockByNumber(context.TODO(), -1)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@@ -124,7 +130,7 @@ func (s *StatsHolder) UpdateEth(ethClient eth.ClientInterface) error {
|
|||||||
}
|
}
|
||||||
s.rw.Lock()
|
s.rw.Lock()
|
||||||
s.Eth.Updated = now
|
s.Eth.Updated = now
|
||||||
s.Eth.LastBlock = lastBlock
|
s.Eth.LastBlock = *lastBlock
|
||||||
s.Eth.LastBatch = lastBatch
|
s.Eth.LastBatch = lastBatch
|
||||||
s.rw.Unlock()
|
s.rw.Unlock()
|
||||||
return nil
|
return nil
|
||||||
@@ -138,17 +144,21 @@ func (s *StatsHolder) CopyStats() *Stats {
|
|||||||
sCopy.Sync.Auction.CurrentSlot.BidValue =
|
sCopy.Sync.Auction.CurrentSlot.BidValue =
|
||||||
common.CopyBigInt(s.Sync.Auction.CurrentSlot.BidValue)
|
common.CopyBigInt(s.Sync.Auction.CurrentSlot.BidValue)
|
||||||
}
|
}
|
||||||
|
if s.Sync.Auction.CurrentSlot.DefaultSlotBid != nil {
|
||||||
|
sCopy.Sync.Auction.CurrentSlot.DefaultSlotBid =
|
||||||
|
common.CopyBigInt(s.Sync.Auction.CurrentSlot.DefaultSlotBid)
|
||||||
|
}
|
||||||
s.rw.RUnlock()
|
s.rw.RUnlock()
|
||||||
return &sCopy
|
return &sCopy
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *StatsHolder) blocksPerc() float64 {
|
func (s *StatsHolder) blocksPerc() float64 {
|
||||||
syncLastBlock := s.Sync.LastBlock
|
syncLastBlockNum := s.Sync.LastBlock.Num
|
||||||
if s.Sync.LastBlock == 0 {
|
if s.Sync.LastBlock.Num == 0 {
|
||||||
syncLastBlock = s.Eth.FirstBlock - 1
|
syncLastBlockNum = s.Eth.FirstBlockNum - 1
|
||||||
}
|
}
|
||||||
return float64(syncLastBlock-(s.Eth.FirstBlock-1)) * 100.0 /
|
return float64(syncLastBlockNum-(s.Eth.FirstBlockNum-1)) * 100.0 /
|
||||||
float64(s.Eth.LastBlock-(s.Eth.FirstBlock-1))
|
float64(s.Eth.LastBlock.Num-(s.Eth.FirstBlockNum-1))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *StatsHolder) batchesPerc(batchNum int64) float64 {
|
func (s *StatsHolder) batchesPerc(batchNum int64) float64 {
|
||||||
@@ -279,7 +289,7 @@ func (s *Synchronizer) updateCurrentSlotIfSync(batchesLen int) error {
|
|||||||
BatchesLen: int(s.stats.Sync.Auction.CurrentSlot.BatchesLen),
|
BatchesLen: int(s.stats.Sync.Auction.CurrentSlot.BatchesLen),
|
||||||
}
|
}
|
||||||
// We want the next block because the current one is already mined
|
// We want the next block because the current one is already mined
|
||||||
blockNum := s.stats.Sync.LastBlock + 1
|
blockNum := s.stats.Sync.LastBlock.Num + 1
|
||||||
slotNum := s.consts.Auction.SlotNum(blockNum)
|
slotNum := s.consts.Auction.SlotNum(blockNum)
|
||||||
if batchesLen == -1 {
|
if batchesLen == -1 {
|
||||||
dbBatchesLen, err := s.historyDB.GetBatchesLen(slotNum)
|
dbBatchesLen, err := s.historyDB.GetBatchesLen(slotNum)
|
||||||
@@ -310,8 +320,11 @@ func (s *Synchronizer) updateCurrentSlotIfSync(batchesLen int) error {
|
|||||||
slot.URL = "???"
|
slot.URL = "???"
|
||||||
} else if err == nil {
|
} else if err == nil {
|
||||||
slot.BidValue = bidCoord.BidValue
|
slot.BidValue = bidCoord.BidValue
|
||||||
defaultSlotBid := bidCoord.DefaultSlotSetBid[slot.SlotNum%6]
|
slot.DefaultSlotBid = bidCoord.DefaultSlotSetBid[slot.SlotNum%6]
|
||||||
if slot.BidValue.Cmp(defaultSlotBid) >= 0 {
|
// Only if the highest bid value is higher than the
|
||||||
|
// default slot bid, the bidder is the winner of the
|
||||||
|
// slot. Otherwise the boot coordinator is the winner.
|
||||||
|
if slot.BidValue.Cmp(slot.DefaultSlotBid) >= 0 {
|
||||||
slot.Bidder = bidCoord.Bidder
|
slot.Bidder = bidCoord.Bidder
|
||||||
slot.Forger = bidCoord.Forger
|
slot.Forger = bidCoord.Forger
|
||||||
slot.URL = bidCoord.URL
|
slot.URL = bidCoord.URL
|
||||||
@@ -344,28 +357,28 @@ func (s *Synchronizer) init() error {
|
|||||||
if err := s.stats.UpdateEth(s.ethClient); err != nil {
|
if err := s.stats.UpdateEth(s.ethClient); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
var lastBlockNum int64
|
lastBlock := &common.Block{}
|
||||||
lastSavedBlock, err := s.historyDB.GetLastBlock()
|
lastSavedBlock, err := s.historyDB.GetLastBlock()
|
||||||
if err != nil && err != sql.ErrNoRows {
|
if err != nil && err != sql.ErrNoRows {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
// If there's no block in the DB (or we only have the default block 0),
|
// If there's no block in the DB (or we only have the default block 0),
|
||||||
// make sure that the stateDB is clean
|
// make sure that the stateDB is clean
|
||||||
if err == sql.ErrNoRows || lastSavedBlock.EthBlockNum == 0 {
|
if err == sql.ErrNoRows || lastSavedBlock.Num == 0 {
|
||||||
if err := s.stateDB.Reset(0); err != nil {
|
if err := s.stateDB.Reset(0); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
lastBlockNum = lastSavedBlock.EthBlockNum
|
lastBlock = lastSavedBlock
|
||||||
}
|
}
|
||||||
if err := s.resetState(lastBlockNum); err != nil {
|
if err := s.resetState(lastBlock); err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Infow("Sync init block",
|
log.Infow("Sync init block",
|
||||||
"syncLastBlock", s.stats.Sync.LastBlock,
|
"syncLastBlock", s.stats.Sync.LastBlock,
|
||||||
"syncBlocksPerc", s.stats.blocksPerc(),
|
"syncBlocksPerc", s.stats.blocksPerc(),
|
||||||
"ethFirstBlock", s.stats.Eth.FirstBlock,
|
"ethFirstBlockNum", s.stats.Eth.FirstBlockNum,
|
||||||
"ethLastBlock", s.stats.Eth.LastBlock,
|
"ethLastBlock", s.stats.Eth.LastBlock,
|
||||||
)
|
)
|
||||||
log.Infow("Sync init batch",
|
log.Infow("Sync init batch",
|
||||||
@@ -393,13 +406,13 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
}
|
}
|
||||||
// If we don't have any stored block, we must do a full sync
|
// If we don't have any stored block, we must do a full sync
|
||||||
// starting from the startBlockNum
|
// starting from the startBlockNum
|
||||||
if err == sql.ErrNoRows || lastSavedBlock.EthBlockNum == 0 {
|
if err == sql.ErrNoRows || lastSavedBlock.Num == 0 {
|
||||||
nextBlockNum = s.startBlockNum
|
nextBlockNum = s.startBlockNum
|
||||||
lastSavedBlock = nil
|
lastSavedBlock = nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if lastSavedBlock != nil {
|
if lastSavedBlock != nil {
|
||||||
nextBlockNum = lastSavedBlock.EthBlockNum + 1
|
nextBlockNum = lastSavedBlock.Num + 1
|
||||||
}
|
}
|
||||||
|
|
||||||
ethBlock, err := s.ethClient.EthBlockByNumber(ctx, nextBlockNum)
|
ethBlock, err := s.ethClient.EthBlockByNumber(ctx, nextBlockNum)
|
||||||
@@ -408,7 +421,8 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
} else if err != nil {
|
} else if err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
log.Debugf("ethBlock: num: %v, parent: %v, hash: %v", ethBlock.EthBlockNum, ethBlock.ParentHash.String(), ethBlock.Hash.String())
|
log.Debugf("ethBlock: num: %v, parent: %v, hash: %v",
|
||||||
|
ethBlock.Num, ethBlock.ParentHash.String(), ethBlock.Hash.String())
|
||||||
|
|
||||||
if err := s.stats.UpdateEth(s.ethClient); err != nil {
|
if err := s.stats.UpdateEth(s.ethClient); err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
@@ -424,13 +438,13 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
if lastSavedBlock.Hash != ethBlock.ParentHash {
|
if lastSavedBlock.Hash != ethBlock.ParentHash {
|
||||||
// Reorg detected
|
// Reorg detected
|
||||||
log.Debugw("Reorg Detected",
|
log.Debugw("Reorg Detected",
|
||||||
"blockNum", ethBlock.EthBlockNum,
|
"blockNum", ethBlock.Num,
|
||||||
"block.parent(got)", ethBlock.ParentHash, "parent.hash(exp)", lastSavedBlock.Hash)
|
"block.parent(got)", ethBlock.ParentHash, "parent.hash(exp)", lastSavedBlock.Hash)
|
||||||
lastDBBlockNum, err := s.reorg(lastSavedBlock)
|
lastDBBlockNum, err := s.reorg(lastSavedBlock)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
discarded := lastSavedBlock.EthBlockNum - lastDBBlockNum
|
discarded := lastSavedBlock.Num - lastDBBlockNum
|
||||||
return nil, &discarded, nil
|
return nil, &discarded, nil
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -456,11 +470,16 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
for i := range rollupData.Withdrawals {
|
for i := range rollupData.Withdrawals {
|
||||||
withdrawal := &rollupData.Withdrawals[i]
|
withdrawal := &rollupData.Withdrawals[i]
|
||||||
if !withdrawal.InstantWithdraw {
|
if !withdrawal.InstantWithdraw {
|
||||||
wDelayerTransfer, ok := wDelayerData.DepositsByTxHash[withdrawal.TxHash]
|
wDelayerTransfers := wDelayerData.DepositsByTxHash[withdrawal.TxHash]
|
||||||
if !ok {
|
if len(wDelayerTransfers) == 0 {
|
||||||
return nil, nil, fmt.Errorf("WDelayer deposit corresponding to " +
|
return nil, nil, fmt.Errorf("WDelayer deposit corresponding to " +
|
||||||
"non-instant rollup withdrawal not found")
|
"non-instant rollup withdrawal not found")
|
||||||
}
|
}
|
||||||
|
// Pop the first wDelayerTransfer to consume them in chronological order
|
||||||
|
wDelayerTransfer := wDelayerTransfers[0]
|
||||||
|
wDelayerData.DepositsByTxHash[withdrawal.TxHash] =
|
||||||
|
wDelayerData.DepositsByTxHash[withdrawal.TxHash][1:]
|
||||||
|
|
||||||
withdrawal.Owner = wDelayerTransfer.Owner
|
withdrawal.Owner = wDelayerTransfer.Owner
|
||||||
withdrawal.Token = wDelayerTransfer.Token
|
withdrawal.Token = wDelayerTransfer.Token
|
||||||
}
|
}
|
||||||
@@ -486,18 +505,25 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
|
|
||||||
batchesLen := len(rollupData.Batches)
|
batchesLen := len(rollupData.Batches)
|
||||||
if batchesLen == 0 {
|
if batchesLen == 0 {
|
||||||
s.stats.UpdateSync(ethBlock.EthBlockNum, nil)
|
s.stats.UpdateSync(ethBlock, nil, nil)
|
||||||
} else {
|
} else {
|
||||||
s.stats.UpdateSync(ethBlock.EthBlockNum,
|
var lastL1BatchBlock *int64
|
||||||
&rollupData.Batches[batchesLen-1].Batch.BatchNum)
|
for _, batchData := range rollupData.Batches {
|
||||||
|
if batchData.L1Batch {
|
||||||
|
lastL1BatchBlock = &batchData.Batch.EthBlockNum
|
||||||
|
}
|
||||||
|
}
|
||||||
|
s.stats.UpdateSync(ethBlock,
|
||||||
|
&rollupData.Batches[batchesLen-1].Batch.BatchNum, lastL1BatchBlock)
|
||||||
}
|
}
|
||||||
if err := s.updateCurrentSlotIfSync(len(rollupData.Batches)); err != nil {
|
if err := s.updateCurrentSlotIfSync(len(rollupData.Batches)); err != nil {
|
||||||
return nil, nil, err
|
return nil, nil, err
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Debugw("Synced block",
|
log.Debugw("Synced block",
|
||||||
"syncLastBlock", s.stats.Sync.LastBlock,
|
"syncLastBlockNum", s.stats.Sync.LastBlock.Num,
|
||||||
"syncBlocksPerc", s.stats.blocksPerc(),
|
"syncBlocksPerc", s.stats.blocksPerc(),
|
||||||
"ethLastBlock", s.stats.Eth.LastBlock,
|
"ethLastBlockNum", s.stats.Eth.LastBlock.Num,
|
||||||
)
|
)
|
||||||
for _, batchData := range rollupData.Batches {
|
for _, batchData := range rollupData.Batches {
|
||||||
log.Debugw("Synced batch",
|
log.Debugw("Synced batch",
|
||||||
@@ -516,8 +542,9 @@ func (s *Synchronizer) Sync2(ctx context.Context, lastSavedBlock *common.Block)
|
|||||||
// corresponding batches in StateBD are discarded. Returns the last valid
|
// corresponding batches in StateBD are discarded. Returns the last valid
|
||||||
// blockNum from the HistoryDB.
|
// blockNum from the HistoryDB.
|
||||||
func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
||||||
blockNum := uncleBlock.EthBlockNum
|
blockNum := uncleBlock.Num
|
||||||
|
|
||||||
|
var block *common.Block
|
||||||
for blockNum >= s.startBlockNum {
|
for blockNum >= s.startBlockNum {
|
||||||
ethBlock, err := s.ethClient.EthBlockByNumber(context.Background(), blockNum)
|
ethBlock, err := s.ethClient.EthBlockByNumber(context.Background(), blockNum)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -525,7 +552,7 @@ func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
|||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
|
||||||
block, err := s.historyDB.GetBlock(blockNum)
|
block, err = s.historyDB.GetBlock(blockNum)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorw("historyDB.GetBlock", "err", err)
|
log.Errorw("historyDB.GetBlock", "err", err)
|
||||||
return 0, err
|
return 0, err
|
||||||
@@ -536,23 +563,23 @@ func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
|||||||
}
|
}
|
||||||
blockNum--
|
blockNum--
|
||||||
}
|
}
|
||||||
total := uncleBlock.EthBlockNum - blockNum
|
total := uncleBlock.Num - block.Num
|
||||||
log.Debugw("Discarding blocks", "total", total, "from", uncleBlock.EthBlockNum, "to", blockNum+1)
|
log.Debugw("Discarding blocks", "total", total, "from", uncleBlock.Num, "to", block.Num+1)
|
||||||
|
|
||||||
// Set History DB and State DB to the correct state
|
// Set History DB and State DB to the correct state
|
||||||
err := s.historyDB.Reorg(blockNum)
|
err := s.historyDB.Reorg(block.Num)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := s.resetState(blockNum); err != nil {
|
if err := s.resetState(block); err != nil {
|
||||||
return 0, err
|
return 0, err
|
||||||
}
|
}
|
||||||
|
|
||||||
return blockNum, nil
|
return block.Num, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *Synchronizer) resetState(blockNum int64) error {
|
func (s *Synchronizer) resetState(block *common.Block) error {
|
||||||
rollup, auction, wDelayer, err := s.historyDB.GetSCVars()
|
rollup, auction, wDelayer, err := s.historyDB.GetSCVars()
|
||||||
// If SCVars are not in the HistoryDB, this is probably the first run
|
// If SCVars are not in the HistoryDB, this is probably the first run
|
||||||
// of the Synchronizer: store the initial vars taken from config
|
// of the Synchronizer: store the initial vars taken from config
|
||||||
@@ -578,13 +605,23 @@ func (s *Synchronizer) resetState(blockNum int64) error {
|
|||||||
if err == sql.ErrNoRows {
|
if err == sql.ErrNoRows {
|
||||||
batchNum = 0
|
batchNum = 0
|
||||||
}
|
}
|
||||||
|
|
||||||
|
lastL1BatchBlockNum, err := s.historyDB.GetLastL1BatchBlockNum()
|
||||||
|
if err != nil && err != sql.ErrNoRows {
|
||||||
|
log.Errorw("historyDB.GetLastL1BatchBlockNum", "err", err)
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
if err == sql.ErrNoRows {
|
||||||
|
lastL1BatchBlockNum = 0
|
||||||
|
}
|
||||||
|
|
||||||
err = s.stateDB.Reset(batchNum)
|
err = s.stateDB.Reset(batchNum)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Errorw("stateDB.Reset", "err", err)
|
log.Errorw("stateDB.Reset", "err", err)
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
|
|
||||||
s.stats.UpdateSync(blockNum, &batchNum)
|
s.stats.UpdateSync(block, &batchNum, &lastL1BatchBlockNum) // TODO
|
||||||
|
|
||||||
if err := s.updateCurrentSlotIfSync(-1); err != nil {
|
if err := s.updateCurrentSlotIfSync(-1); err != nil {
|
||||||
return err
|
return err
|
||||||
@@ -640,7 +677,7 @@ func (s *Synchronizer) Status() (*common.SyncStatus, error) {
|
|||||||
// rollupSync retreives all the Rollup Smart Contract Data that happened at
|
// rollupSync retreives all the Rollup Smart Contract Data that happened at
|
||||||
// ethBlock.blockNum with ethBlock.Hash.
|
// ethBlock.blockNum with ethBlock.Hash.
|
||||||
func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, error) {
|
func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, error) {
|
||||||
blockNum := ethBlock.EthBlockNum
|
blockNum := ethBlock.Num
|
||||||
var rollupData = common.NewRollupData()
|
var rollupData = common.NewRollupData()
|
||||||
// var forgeL1TxsNum int64
|
// var forgeL1TxsNum int64
|
||||||
|
|
||||||
@@ -788,8 +825,8 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
|||||||
batchData.CreatedAccounts = processTxsOut.CreatedAccounts
|
batchData.CreatedAccounts = processTxsOut.CreatedAccounts
|
||||||
|
|
||||||
slotNum := int64(0)
|
slotNum := int64(0)
|
||||||
if ethBlock.EthBlockNum >= s.consts.Auction.GenesisBlockNum {
|
if ethBlock.Num >= s.consts.Auction.GenesisBlockNum {
|
||||||
slotNum = (ethBlock.EthBlockNum - s.consts.Auction.GenesisBlockNum) /
|
slotNum = (ethBlock.Num - s.consts.Auction.GenesisBlockNum) /
|
||||||
int64(s.consts.Auction.BlocksPerSlot)
|
int64(s.consts.Auction.BlocksPerSlot)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -880,7 +917,7 @@ func cutStringMax(s string, max int) string {
|
|||||||
|
|
||||||
// auctionSync gets information from the Auction Contract
|
// auctionSync gets information from the Auction Contract
|
||||||
func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData, error) {
|
func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData, error) {
|
||||||
blockNum := ethBlock.EthBlockNum
|
blockNum := ethBlock.Num
|
||||||
var auctionData = common.NewAuctionData()
|
var auctionData = common.NewAuctionData()
|
||||||
|
|
||||||
// Get auction events in the block
|
// Get auction events in the block
|
||||||
@@ -955,7 +992,8 @@ func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData,
|
|||||||
"auctionEvents.NewDefaultSlotSetBid: %v", evt.SlotSet)
|
"auctionEvents.NewDefaultSlotSetBid: %v", evt.SlotSet)
|
||||||
}
|
}
|
||||||
s.vars.Auction.DefaultSlotSetBid[evt.SlotSet] = evt.NewInitialMinBid
|
s.vars.Auction.DefaultSlotSetBid[evt.SlotSet] = evt.NewInitialMinBid
|
||||||
s.vars.Auction.DefaultSlotSetBidSlotNum = s.consts.Auction.SlotNum(blockNum) + int64(s.vars.Auction.ClosedAuctionSlots) + 1
|
s.vars.Auction.DefaultSlotSetBidSlotNum = s.consts.Auction.SlotNum(blockNum) +
|
||||||
|
int64(s.vars.Auction.ClosedAuctionSlots) + 1
|
||||||
varsUpdate = true
|
varsUpdate = true
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -973,7 +1011,7 @@ func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData,
|
|||||||
|
|
||||||
// wdelayerSync gets information from the Withdrawal Delayer Contract
|
// wdelayerSync gets information from the Withdrawal Delayer Contract
|
||||||
func (s *Synchronizer) wdelayerSync(ethBlock *common.Block) (*common.WDelayerData, error) {
|
func (s *Synchronizer) wdelayerSync(ethBlock *common.Block) (*common.WDelayerData, error) {
|
||||||
blockNum := ethBlock.EthBlockNum
|
blockNum := ethBlock.Num
|
||||||
wDelayerData := common.NewWDelayerData()
|
wDelayerData := common.NewWDelayerData()
|
||||||
|
|
||||||
// Get wDelayer events in the block
|
// Get wDelayer events in the block
|
||||||
@@ -998,7 +1036,8 @@ func (s *Synchronizer) wdelayerSync(ethBlock *common.Block) (*common.WDelayerDat
|
|||||||
Amount: evt.Amount,
|
Amount: evt.Amount,
|
||||||
})
|
})
|
||||||
wDelayerData.DepositsByTxHash[evt.TxHash] =
|
wDelayerData.DepositsByTxHash[evt.TxHash] =
|
||||||
&wDelayerData.Deposits[len(wDelayerData.Deposits)-1]
|
append(wDelayerData.DepositsByTxHash[evt.TxHash],
|
||||||
|
&wDelayerData.Deposits[len(wDelayerData.Deposits)-1])
|
||||||
}
|
}
|
||||||
for _, evt := range wDelayerEvents.Withdraw {
|
for _, evt := range wDelayerEvents.Withdraw {
|
||||||
wDelayerData.Withdrawals = append(wDelayerData.Withdrawals, common.WDelayerTransfer{
|
wDelayerData.Withdrawals = append(wDelayerData.Withdrawals, common.WDelayerTransfer{
|
||||||
|
|||||||
@@ -47,7 +47,7 @@ func checkSyncBlock(t *testing.T, s *Synchronizer, blockNum int, block, syncBloc
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
dbBlocks = dbBlocks[1:] // ignore block 0, added by default in the DB
|
dbBlocks = dbBlocks[1:] // ignore block 0, added by default in the DB
|
||||||
assert.Equal(t, blockNum, len(dbBlocks))
|
assert.Equal(t, blockNum, len(dbBlocks))
|
||||||
assert.Equal(t, int64(blockNum), dbBlocks[blockNum-1].EthBlockNum)
|
assert.Equal(t, int64(blockNum), dbBlocks[blockNum-1].Num)
|
||||||
assert.NotEqual(t, dbBlocks[blockNum-1].Hash, dbBlocks[blockNum-2].Hash)
|
assert.NotEqual(t, dbBlocks[blockNum-1].Hash, dbBlocks[blockNum-2].Hash)
|
||||||
assert.Greater(t, dbBlocks[blockNum-1].Timestamp.Unix(), dbBlocks[blockNum-2].Timestamp.Unix())
|
assert.Greater(t, dbBlocks[blockNum-1].Timestamp.Unix(), dbBlocks[blockNum-2].Timestamp.Unix())
|
||||||
|
|
||||||
@@ -60,7 +60,7 @@ func checkSyncBlock(t *testing.T, s *Synchronizer, blockNum int, block, syncBloc
|
|||||||
dbToken := dbTokens[i]
|
dbToken := dbTokens[i]
|
||||||
syncToken := syncBlock.Rollup.AddedTokens[i]
|
syncToken := syncBlock.Rollup.AddedTokens[i]
|
||||||
|
|
||||||
assert.Equal(t, block.Block.EthBlockNum, syncToken.EthBlockNum)
|
assert.Equal(t, block.Block.Num, syncToken.EthBlockNum)
|
||||||
assert.Equal(t, token.TokenID, syncToken.TokenID)
|
assert.Equal(t, token.TokenID, syncToken.TokenID)
|
||||||
assert.Equal(t, token.EthAddr, syncToken.EthAddr)
|
assert.Equal(t, token.EthAddr, syncToken.EthAddr)
|
||||||
tokenConst := tokenConsts[token.TokenID]
|
tokenConst := tokenConsts[token.TokenID]
|
||||||
@@ -321,17 +321,36 @@ func TestSync(t *testing.T) {
|
|||||||
//
|
//
|
||||||
// First Sync from an initial state
|
// First Sync from an initial state
|
||||||
//
|
//
|
||||||
|
var vars struct {
|
||||||
|
Rollup *common.RollupVariables
|
||||||
|
Auction *common.AuctionVariables
|
||||||
|
WDelayer *common.WDelayerVariables
|
||||||
|
}
|
||||||
|
stats := s.Stats()
|
||||||
|
assert.Equal(t, false, stats.Synced())
|
||||||
|
|
||||||
// Test Sync for rollup genesis block
|
// Test Sync for rollup genesis block
|
||||||
syncBlock, discards, err := s.Sync2(ctx, nil)
|
syncBlock, discards, err := s.Sync2(ctx, nil)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(1), syncBlock.Block.EthBlockNum)
|
require.Nil(t, syncBlock.Rollup.Vars)
|
||||||
|
require.Nil(t, syncBlock.Auction.Vars)
|
||||||
|
require.Nil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(1), syncBlock.Block.Num)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(1), stats.Sync.LastBlock.Num)
|
||||||
|
vars.Rollup, vars.Auction, vars.WDelayer = s.SCVars()
|
||||||
|
assert.Equal(t, clientSetup.RollupVariables, vars.Rollup)
|
||||||
|
assert.Equal(t, clientSetup.AuctionVariables, vars.Auction)
|
||||||
|
assert.Equal(t, clientSetup.WDelayerVariables, vars.WDelayer)
|
||||||
|
|
||||||
dbBlocks, err := s.historyDB.GetAllBlocks()
|
dbBlocks, err := s.historyDB.GetAllBlocks()
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
assert.Equal(t, 2, len(dbBlocks))
|
assert.Equal(t, 2, len(dbBlocks))
|
||||||
assert.Equal(t, int64(1), dbBlocks[1].EthBlockNum)
|
assert.Equal(t, int64(1), dbBlocks[1].Num)
|
||||||
|
|
||||||
// Sync again and expect no new blocks
|
// Sync again and expect no new blocks
|
||||||
syncBlock, discards, err = s.Sync2(ctx, nil)
|
syncBlock, discards, err = s.Sync2(ctx, nil)
|
||||||
@@ -388,14 +407,14 @@ func TestSync(t *testing.T) {
|
|||||||
require.Equal(t, 2, len(blocks))
|
require.Equal(t, 2, len(blocks))
|
||||||
// blocks 0 (blockNum=2)
|
// blocks 0 (blockNum=2)
|
||||||
i := 0
|
i := 0
|
||||||
require.Equal(t, 2, int(blocks[i].Block.EthBlockNum))
|
require.Equal(t, 2, int(blocks[i].Block.Num))
|
||||||
require.Equal(t, 3, len(blocks[i].Rollup.AddedTokens))
|
require.Equal(t, 3, len(blocks[i].Rollup.AddedTokens))
|
||||||
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches[0].L1CoordinatorTxs))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches[0].L1CoordinatorTxs))
|
||||||
// blocks 1 (blockNum=3)
|
// blocks 1 (blockNum=3)
|
||||||
i = 1
|
i = 1
|
||||||
require.Equal(t, 3, int(blocks[i].Block.EthBlockNum))
|
require.Equal(t, 3, int(blocks[i].Block.Num))
|
||||||
require.Equal(t, 4, len(blocks[i].Rollup.L1UserTxs))
|
require.Equal(t, 4, len(blocks[i].Rollup.L1UserTxs))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||||
require.Equal(t, 3, len(blocks[i].Rollup.Batches[0].L2Txs))
|
require.Equal(t, 3, len(blocks[i].Rollup.Batches[0].L2Txs))
|
||||||
@@ -421,7 +440,14 @@ func TestSync(t *testing.T) {
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(2), syncBlock.Block.EthBlockNum)
|
assert.Nil(t, syncBlock.Rollup.Vars)
|
||||||
|
assert.Nil(t, syncBlock.Auction.Vars)
|
||||||
|
assert.Nil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(2), syncBlock.Block.Num)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(3), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(2), stats.Sync.LastBlock.Num)
|
||||||
|
|
||||||
checkSyncBlock(t, s, 2, &blocks[0], syncBlock)
|
checkSyncBlock(t, s, 2, &blocks[0], syncBlock)
|
||||||
|
|
||||||
@@ -431,7 +457,14 @@ func TestSync(t *testing.T) {
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(3), syncBlock.Block.EthBlockNum)
|
assert.Nil(t, syncBlock.Rollup.Vars)
|
||||||
|
assert.Nil(t, syncBlock.Auction.Vars)
|
||||||
|
assert.Nil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(3), syncBlock.Block.Num)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(3), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(3), stats.Sync.LastBlock.Num)
|
||||||
|
|
||||||
checkSyncBlock(t, s, 3, &blocks[1], syncBlock)
|
checkSyncBlock(t, s, 3, &blocks[1], syncBlock)
|
||||||
|
|
||||||
@@ -447,7 +480,19 @@ func TestSync(t *testing.T) {
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(4), syncBlock.Block.EthBlockNum)
|
assert.Nil(t, syncBlock.Rollup.Vars)
|
||||||
|
assert.Nil(t, syncBlock.Auction.Vars)
|
||||||
|
assert.Nil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(4), syncBlock.Block.Num)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(4), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(4), stats.Sync.LastBlock.Num)
|
||||||
|
vars.Rollup, vars.Auction, vars.WDelayer = s.SCVars()
|
||||||
|
assert.Equal(t, clientSetup.RollupVariables, vars.Rollup)
|
||||||
|
assert.Equal(t, clientSetup.AuctionVariables, vars.Auction)
|
||||||
|
assert.Equal(t, clientSetup.WDelayerVariables, vars.WDelayer)
|
||||||
|
|
||||||
dbExits, err := s.historyDB.GetAllExits()
|
dbExits, err := s.historyDB.GetAllExits()
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
foundA1, foundC1 := false, false
|
foundA1, foundC1 := false, false
|
||||||
@@ -486,14 +531,25 @@ func TestSync(t *testing.T) {
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(5), syncBlock.Block.EthBlockNum)
|
assert.NotNil(t, syncBlock.Rollup.Vars)
|
||||||
|
assert.NotNil(t, syncBlock.Auction.Vars)
|
||||||
|
assert.NotNil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(5), syncBlock.Block.Num)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(5), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(5), stats.Sync.LastBlock.Num)
|
||||||
|
vars.Rollup, vars.Auction, vars.WDelayer = s.SCVars()
|
||||||
|
assert.NotEqual(t, clientSetup.RollupVariables, vars.Rollup)
|
||||||
|
assert.NotEqual(t, clientSetup.AuctionVariables, vars.Auction)
|
||||||
|
assert.NotEqual(t, clientSetup.WDelayerVariables, vars.WDelayer)
|
||||||
|
|
||||||
dbRollupVars, dbAuctionVars, dbWDelayerVars, err := s.historyDB.GetSCVars()
|
dbRollupVars, dbAuctionVars, dbWDelayerVars, err := s.historyDB.GetSCVars()
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
// Set EthBlockNum for Vars to the blockNum in which they were updated (should be 5)
|
// Set EthBlockNum for Vars to the blockNum in which they were updated (should be 5)
|
||||||
rollupVars.EthBlockNum = syncBlock.Block.EthBlockNum
|
rollupVars.EthBlockNum = syncBlock.Block.Num
|
||||||
auctionVars.EthBlockNum = syncBlock.Block.EthBlockNum
|
auctionVars.EthBlockNum = syncBlock.Block.Num
|
||||||
wDelayerVars.EthBlockNum = syncBlock.Block.EthBlockNum
|
wDelayerVars.EthBlockNum = syncBlock.Block.Num
|
||||||
assert.Equal(t, rollupVars, dbRollupVars)
|
assert.Equal(t, rollupVars, dbRollupVars)
|
||||||
assert.Equal(t, auctionVars, dbAuctionVars)
|
assert.Equal(t, auctionVars, dbAuctionVars)
|
||||||
assert.Equal(t, wDelayerVars, dbWDelayerVars)
|
assert.Equal(t, wDelayerVars, dbWDelayerVars)
|
||||||
@@ -535,8 +591,8 @@ func TestSync(t *testing.T) {
|
|||||||
for i := 0; i < 4; i++ {
|
for i := 0; i < 4; i++ {
|
||||||
client.CtlRollback()
|
client.CtlRollback()
|
||||||
}
|
}
|
||||||
blockNum := client.CtlLastBlock()
|
block := client.CtlLastBlock()
|
||||||
require.Equal(t, int64(1), blockNum)
|
require.Equal(t, int64(1), block.Num)
|
||||||
|
|
||||||
// Generate extra required data
|
// Generate extra required data
|
||||||
ethAddTokens(blocks, client)
|
ethAddTokens(blocks, client)
|
||||||
@@ -554,11 +610,18 @@ func TestSync(t *testing.T) {
|
|||||||
expetedDiscards := int64(4)
|
expetedDiscards := int64(4)
|
||||||
require.Equal(t, &expetedDiscards, discards)
|
require.Equal(t, &expetedDiscards, discards)
|
||||||
require.Nil(t, syncBlock)
|
require.Nil(t, syncBlock)
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, false, stats.Synced())
|
||||||
|
assert.Equal(t, int64(6), stats.Eth.LastBlock.Num)
|
||||||
|
vars.Rollup, vars.Auction, vars.WDelayer = s.SCVars()
|
||||||
|
assert.Equal(t, clientSetup.RollupVariables, vars.Rollup)
|
||||||
|
assert.Equal(t, clientSetup.AuctionVariables, vars.Auction)
|
||||||
|
assert.Equal(t, clientSetup.WDelayerVariables, vars.WDelayer)
|
||||||
|
|
||||||
// At this point, the DB only has data up to block 1
|
// At this point, the DB only has data up to block 1
|
||||||
dbBlock, err := s.historyDB.GetLastBlock()
|
dbBlock, err := s.historyDB.GetLastBlock()
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
assert.Equal(t, int64(1), dbBlock.EthBlockNum)
|
assert.Equal(t, int64(1), dbBlock.Num)
|
||||||
|
|
||||||
// Accounts in HistoryDB and StateDB must be empty
|
// Accounts in HistoryDB and StateDB must be empty
|
||||||
dbAccounts, err := s.historyDB.GetAllAccounts()
|
dbAccounts, err := s.historyDB.GetAllAccounts()
|
||||||
@@ -574,12 +637,30 @@ func TestSync(t *testing.T) {
|
|||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
require.Nil(t, discards)
|
require.Nil(t, discards)
|
||||||
require.NotNil(t, syncBlock)
|
require.NotNil(t, syncBlock)
|
||||||
assert.Equal(t, int64(2+i), syncBlock.Block.EthBlockNum)
|
assert.Nil(t, syncBlock.Rollup.Vars)
|
||||||
|
assert.Nil(t, syncBlock.Auction.Vars)
|
||||||
|
assert.Nil(t, syncBlock.WDelayer.Vars)
|
||||||
|
assert.Equal(t, int64(2+i), syncBlock.Block.Num)
|
||||||
|
|
||||||
|
stats = s.Stats()
|
||||||
|
assert.Equal(t, int64(1), stats.Eth.FirstBlockNum)
|
||||||
|
assert.Equal(t, int64(6), stats.Eth.LastBlock.Num)
|
||||||
|
assert.Equal(t, int64(2+i), stats.Sync.LastBlock.Num)
|
||||||
|
if i == 4 {
|
||||||
|
assert.Equal(t, true, stats.Synced())
|
||||||
|
} else {
|
||||||
|
assert.Equal(t, false, stats.Synced())
|
||||||
|
}
|
||||||
|
|
||||||
|
vars.Rollup, vars.Auction, vars.WDelayer = s.SCVars()
|
||||||
|
assert.Equal(t, clientSetup.RollupVariables, vars.Rollup)
|
||||||
|
assert.Equal(t, clientSetup.AuctionVariables, vars.Auction)
|
||||||
|
assert.Equal(t, clientSetup.WDelayerVariables, vars.WDelayer)
|
||||||
}
|
}
|
||||||
|
|
||||||
dbBlock, err = s.historyDB.GetLastBlock()
|
dbBlock, err = s.historyDB.GetLastBlock()
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
assert.Equal(t, int64(6), dbBlock.EthBlockNum)
|
assert.Equal(t, int64(6), dbBlock.Num)
|
||||||
|
|
||||||
// Accounts in HistoryDB and StateDB is only 2 entries
|
// Accounts in HistoryDB and StateDB is only 2 entries
|
||||||
dbAccounts, err = s.historyDB.GetAllAccounts()
|
dbAccounts, err = s.historyDB.GetAllAccounts()
|
||||||
|
|||||||
@@ -553,10 +553,17 @@ func (c *Client) CtlRollback() {
|
|||||||
//
|
//
|
||||||
|
|
||||||
// CtlLastBlock returns the last blockNum without checks
|
// CtlLastBlock returns the last blockNum without checks
|
||||||
func (c *Client) CtlLastBlock() int64 {
|
func (c *Client) CtlLastBlock() *common.Block {
|
||||||
c.rw.RLock()
|
c.rw.RLock()
|
||||||
defer c.rw.RUnlock()
|
defer c.rw.RUnlock()
|
||||||
return c.blockNum
|
|
||||||
|
block := c.blocks[c.blockNum]
|
||||||
|
return &common.Block{
|
||||||
|
Num: c.blockNum,
|
||||||
|
Timestamp: time.Unix(block.Eth.Time, 0),
|
||||||
|
Hash: block.Eth.Hash,
|
||||||
|
ParentHash: block.Eth.ParentHash,
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// EthLastBlock returns the last blockNum
|
// EthLastBlock returns the last blockNum
|
||||||
@@ -626,7 +633,7 @@ func (c *Client) EthERC20Consts(tokenAddr ethCommon.Address) (*eth.ERC20Consts,
|
|||||||
// }
|
// }
|
||||||
|
|
||||||
// EthBlockByNumber returns the *common.Block for the given block number in a
|
// EthBlockByNumber returns the *common.Block for the given block number in a
|
||||||
// deterministic way.
|
// deterministic way. If number == -1, the latests known block is returned.
|
||||||
func (c *Client) EthBlockByNumber(ctx context.Context, blockNum int64) (*common.Block, error) {
|
func (c *Client) EthBlockByNumber(ctx context.Context, blockNum int64) (*common.Block, error) {
|
||||||
c.rw.RLock()
|
c.rw.RLock()
|
||||||
defer c.rw.RUnlock()
|
defer c.rw.RUnlock()
|
||||||
@@ -634,9 +641,12 @@ func (c *Client) EthBlockByNumber(ctx context.Context, blockNum int64) (*common.
|
|||||||
if blockNum > c.blockNum {
|
if blockNum > c.blockNum {
|
||||||
return nil, ethereum.NotFound
|
return nil, ethereum.NotFound
|
||||||
}
|
}
|
||||||
|
if blockNum == -1 {
|
||||||
|
blockNum = c.blockNum
|
||||||
|
}
|
||||||
block := c.blocks[blockNum]
|
block := c.blocks[blockNum]
|
||||||
return &common.Block{
|
return &common.Block{
|
||||||
EthBlockNum: blockNum,
|
Num: blockNum,
|
||||||
Timestamp: time.Unix(block.Eth.Time, 0),
|
Timestamp: time.Unix(block.Eth.Time, 0),
|
||||||
Hash: block.Eth.Hash,
|
Hash: block.Eth.Hash,
|
||||||
ParentHash: block.Eth.ParentHash,
|
ParentHash: block.Eth.ParentHash,
|
||||||
|
|||||||
@@ -46,7 +46,7 @@ func TestClientEth(t *testing.T) {
|
|||||||
|
|
||||||
block, err := c.EthBlockByNumber(context.TODO(), 0)
|
block, err := c.EthBlockByNumber(context.TODO(), 0)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
assert.Equal(t, int64(0), block.EthBlockNum)
|
assert.Equal(t, int64(0), block.Num)
|
||||||
assert.Equal(t, time.Unix(0, 0), block.Timestamp)
|
assert.Equal(t, time.Unix(0, 0), block.Timestamp)
|
||||||
assert.Equal(t, "0x0000000000000000000000000000000000000000000000000000000000000000", block.Hash.Hex())
|
assert.Equal(t, "0x0000000000000000000000000000000000000000000000000000000000000000", block.Hash.Hex())
|
||||||
|
|
||||||
@@ -60,7 +60,7 @@ func TestClientEth(t *testing.T) {
|
|||||||
|
|
||||||
block, err = c.EthBlockByNumber(context.TODO(), 2)
|
block, err = c.EthBlockByNumber(context.TODO(), 2)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
assert.Equal(t, int64(2), block.EthBlockNum)
|
assert.Equal(t, int64(2), block.Num)
|
||||||
assert.Equal(t, time.Unix(2, 0), block.Timestamp)
|
assert.Equal(t, time.Unix(2, 0), block.Timestamp)
|
||||||
|
|
||||||
// Add a token
|
// Add a token
|
||||||
|
|||||||
@@ -15,7 +15,7 @@ import (
|
|||||||
// Block0 represents Ethereum's genesis block,
|
// Block0 represents Ethereum's genesis block,
|
||||||
// which is stored by default at HistoryDB
|
// which is stored by default at HistoryDB
|
||||||
var Block0 common.Block = common.Block{
|
var Block0 common.Block = common.Block{
|
||||||
EthBlockNum: 0,
|
Num: 0,
|
||||||
Hash: ethCommon.Hash([32]byte{
|
Hash: ethCommon.Hash([32]byte{
|
||||||
212, 229, 103, 64, 248, 118, 174, 248,
|
212, 229, 103, 64, 248, 118, 174, 248,
|
||||||
192, 16, 184, 106, 64, 213, 245, 103,
|
192, 16, 184, 106, 64, 213, 245, 103,
|
||||||
@@ -44,7 +44,7 @@ func GenBlocks(from, to int64) []common.Block {
|
|||||||
var blocks []common.Block
|
var blocks []common.Block
|
||||||
for i := from; i < to; i++ {
|
for i := from; i < to; i++ {
|
||||||
blocks = append(blocks, common.Block{
|
blocks = append(blocks, common.Block{
|
||||||
EthBlockNum: i,
|
Num: i,
|
||||||
//nolint:gomnd
|
//nolint:gomnd
|
||||||
Timestamp: time.Now().Add(time.Second * 13).UTC(),
|
Timestamp: time.Now().Add(time.Second * 13).UTC(),
|
||||||
Hash: ethCommon.BigToHash(big.NewInt(int64(i))),
|
Hash: ethCommon.BigToHash(big.NewInt(int64(i))),
|
||||||
@@ -62,7 +62,7 @@ func GenTokens(nTokens int, blocks []common.Block) (tokensToAddInDB []common.Tok
|
|||||||
Name: "NAME" + fmt.Sprint(i),
|
Name: "NAME" + fmt.Sprint(i),
|
||||||
Symbol: fmt.Sprint(i),
|
Symbol: fmt.Sprint(i),
|
||||||
Decimals: uint64(i + 1),
|
Decimals: uint64(i + 1),
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
EthAddr: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
EthAddr: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
||||||
}
|
}
|
||||||
tokensToAddInDB = append(tokensToAddInDB, token)
|
tokensToAddInDB = append(tokensToAddInDB, token)
|
||||||
@@ -87,7 +87,7 @@ func GenBatches(nBatches int, blocks []common.Block) []common.Batch {
|
|||||||
for i := 0; i < nBatches; i++ {
|
for i := 0; i < nBatches; i++ {
|
||||||
batch := common.Batch{
|
batch := common.Batch{
|
||||||
BatchNum: common.BatchNum(i + 1),
|
BatchNum: common.BatchNum(i + 1),
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
//nolint:gomnd
|
//nolint:gomnd
|
||||||
ForgerAddr: ethCommon.BigToAddress(big.NewInt(6886723)),
|
ForgerAddr: ethCommon.BigToAddress(big.NewInt(6886723)),
|
||||||
CollectedFees: collectedFees,
|
CollectedFees: collectedFees,
|
||||||
@@ -161,7 +161,7 @@ func GenL1Txs(
|
|||||||
TokenID: token.TokenID,
|
TokenID: token.TokenID,
|
||||||
Amount: amount,
|
Amount: amount,
|
||||||
LoadAmount: amount,
|
LoadAmount: amount,
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
}
|
}
|
||||||
if tx.UserOrigin {
|
if tx.UserOrigin {
|
||||||
n := nextTxsNum
|
n := nextTxsNum
|
||||||
@@ -287,7 +287,7 @@ func GenL2Txs(
|
|||||||
Amount: amount,
|
Amount: amount,
|
||||||
Fee: fee,
|
Fee: fee,
|
||||||
Nonce: common.Nonce(i + 1),
|
Nonce: common.Nonce(i + 1),
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
Type: randomTxType(i),
|
Type: randomTxType(i),
|
||||||
}
|
}
|
||||||
if i < nUserTxs {
|
if i < nUserTxs {
|
||||||
@@ -341,7 +341,7 @@ func GenCoordinators(nCoords int, blocks []common.Block) []common.Coordinator {
|
|||||||
coords := []common.Coordinator{}
|
coords := []common.Coordinator{}
|
||||||
for i := 0; i < nCoords; i++ {
|
for i := 0; i < nCoords; i++ {
|
||||||
coords = append(coords, common.Coordinator{
|
coords = append(coords, common.Coordinator{
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
Forger: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
Forger: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
||||||
Bidder: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
Bidder: ethCommon.BigToAddress(big.NewInt(int64(i))),
|
||||||
URL: "https://foo.bar",
|
URL: "https://foo.bar",
|
||||||
@@ -363,7 +363,7 @@ func GenBids(nBids int, blocks []common.Block, coords []common.Coordinator) []co
|
|||||||
bids = append(bids, common.Bid{
|
bids = append(bids, common.Bid{
|
||||||
SlotNum: slotNum,
|
SlotNum: slotNum,
|
||||||
BidValue: big.NewInt(int64(i)),
|
BidValue: big.NewInt(int64(i)),
|
||||||
EthBlockNum: blocks[i%len(blocks)].EthBlockNum,
|
EthBlockNum: blocks[i%len(blocks)].Num,
|
||||||
Bidder: coords[i%len(blocks)].Bidder,
|
Bidder: coords[i%len(blocks)].Bidder,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
@@ -397,13 +397,13 @@ func GenExitTree(n int, batches []common.Batch, accounts []common.Account, block
|
|||||||
Balance: big.NewInt(int64(i) * 1000),
|
Balance: big.NewInt(int64(i) * 1000),
|
||||||
}
|
}
|
||||||
if i%2 == 0 {
|
if i%2 == 0 {
|
||||||
instant := int64(blocks[i%len(blocks)].EthBlockNum)
|
instant := int64(blocks[i%len(blocks)].Num)
|
||||||
exitTree[i].InstantWithdrawn = &instant
|
exitTree[i].InstantWithdrawn = &instant
|
||||||
} else if i%3 == 0 {
|
} else if i%3 == 0 {
|
||||||
delayedReq := int64(blocks[i%len(blocks)].EthBlockNum)
|
delayedReq := int64(blocks[i%len(blocks)].Num)
|
||||||
exitTree[i].DelayedWithdrawRequest = &delayedReq
|
exitTree[i].DelayedWithdrawRequest = &delayedReq
|
||||||
if i%9 == 0 {
|
if i%9 == 0 {
|
||||||
delayed := int64(blocks[i%len(blocks)].EthBlockNum)
|
delayed := int64(blocks[i%len(blocks)].Num)
|
||||||
exitTree[i].DelayedWithdrawn = &delayed
|
exitTree[i].DelayedWithdrawn = &delayed
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ func newBatchData(batchNum int) common.BatchData {
|
|||||||
func newBlock(blockNum int64) common.BlockData {
|
func newBlock(blockNum int64) common.BlockData {
|
||||||
return common.BlockData{
|
return common.BlockData{
|
||||||
Block: common.Block{
|
Block: common.Block{
|
||||||
EthBlockNum: blockNum,
|
Num: blockNum,
|
||||||
},
|
},
|
||||||
Rollup: common.RollupData{
|
Rollup: common.RollupData{
|
||||||
L1UserTxs: []common.L1Tx{},
|
L1UserTxs: []common.L1Tx{},
|
||||||
@@ -711,7 +711,7 @@ func (tc *Context) FillBlocksExtra(blocks []common.BlockData, cfg *ConfigExtra)
|
|||||||
block := &blocks[i]
|
block := &blocks[i]
|
||||||
for j := range block.Rollup.Batches {
|
for j := range block.Rollup.Batches {
|
||||||
batch := &block.Rollup.Batches[j]
|
batch := &block.Rollup.Batches[j]
|
||||||
batch.Batch.EthBlockNum = block.Block.EthBlockNum
|
batch.Batch.EthBlockNum = block.Block.Num
|
||||||
// til doesn't fill the batch forger addr
|
// til doesn't fill the batch forger addr
|
||||||
batch.Batch.ForgerAddr = cfg.BootCoordAddr
|
batch.Batch.ForgerAddr = cfg.BootCoordAddr
|
||||||
if batch.L1Batch {
|
if batch.L1Batch {
|
||||||
|
|||||||
@@ -51,7 +51,7 @@ func addTokens(t *testing.T, tokens []common.Token, db *sqlx.DB) {
|
|||||||
hdb := historydb.NewHistoryDB(db)
|
hdb := historydb.NewHistoryDB(db)
|
||||||
test.WipeDB(hdb.DB())
|
test.WipeDB(hdb.DB())
|
||||||
assert.Nil(t, hdb.AddBlock(&common.Block{
|
assert.Nil(t, hdb.AddBlock(&common.Block{
|
||||||
EthBlockNum: 1,
|
Num: 1,
|
||||||
}))
|
}))
|
||||||
assert.Nil(t, hdb.AddTokens(tokens))
|
assert.Nil(t, hdb.AddTokens(tokens))
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user