mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 03:16:45 +01:00
WIP3
This commit is contained in:
@@ -180,7 +180,7 @@ type testCommon struct {
|
|||||||
auctionVars common.AuctionVariables
|
auctionVars common.AuctionVariables
|
||||||
rollupVars common.RollupVariables
|
rollupVars common.RollupVariables
|
||||||
wdelayerVars common.WDelayerVariables
|
wdelayerVars common.WDelayerVariables
|
||||||
nextForgers []historydb.NextForger
|
nextForgers []historydb.NextForgerAPI
|
||||||
}
|
}
|
||||||
|
|
||||||
var tc testCommon
|
var tc testCommon
|
||||||
@@ -455,14 +455,14 @@ func TestMain(m *testing.M) {
|
|||||||
if err = api.h.AddBids(bids); err != nil {
|
if err = api.h.AddBids(bids); err != nil {
|
||||||
panic(err)
|
panic(err)
|
||||||
}
|
}
|
||||||
bootForger := historydb.NextForger{
|
bootForger := historydb.NextForgerAPI{
|
||||||
Coordinator: historydb.CoordinatorAPI{
|
Coordinator: historydb.CoordinatorAPI{
|
||||||
Forger: auctionVars.BootCoordinator,
|
Forger: auctionVars.BootCoordinator,
|
||||||
URL: auctionVars.BootCoordinatorURL,
|
URL: auctionVars.BootCoordinatorURL,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
// Set next forgers: set all as boot coordinator then replace the non boot coordinators
|
// Set next forgers: set all as boot coordinator then replace the non boot coordinators
|
||||||
nextForgers := []historydb.NextForger{}
|
nextForgers := []historydb.NextForgerAPI{}
|
||||||
var initBlock int64 = 140
|
var initBlock int64 = 140
|
||||||
var deltaBlocks int64 = 40
|
var deltaBlocks int64 = 40
|
||||||
for i := 1; i < int(auctionVars.ClosedAuctionSlots)+2; i++ {
|
for i := 1; i < int(auctionVars.ClosedAuctionSlots)+2; i++ {
|
||||||
|
|||||||
111
api/state.go
111
api/state.go
@@ -1,16 +1,123 @@
|
|||||||
package api
|
package api
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"database/sql"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
|
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||||
|
"github.com/hermeznetwork/tracerr"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (a *API) getState(c *gin.Context) {
|
func (a *API) getState(c *gin.Context) {
|
||||||
ni, err := a.h.GetNodeInfoAPI()
|
stateAPI, err := a.h.GetStateAPI()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
retBadReq(err, c)
|
retBadReq(err, c)
|
||||||
return
|
return
|
||||||
}
|
}
|
||||||
c.JSON(http.StatusOK, ni.APIState)
|
c.JSON(http.StatusOK, stateAPI)
|
||||||
|
}
|
||||||
|
|
||||||
|
type APIStateUpdater struct {
|
||||||
|
hdb *historydb.HistoryDB
|
||||||
|
state historydb.StateAPI
|
||||||
|
config historydb.NodeConfig
|
||||||
|
vars common.SCVariablesPtr
|
||||||
|
consts historydb.Constants
|
||||||
|
}
|
||||||
|
|
||||||
|
func NewAPIStateUpdater(hdb *historydb.HistoryDB, config *historydb.NodeConfig, vars *common.SCVariables,
|
||||||
|
consts *historydb.Constants) *APIStateUpdater {
|
||||||
|
u := APIStateUpdater{
|
||||||
|
hdb: hdb,
|
||||||
|
config: *config,
|
||||||
|
consts: *consts,
|
||||||
|
}
|
||||||
|
u.SetSCVars(&common.SCVariablesPtr{&vars.Rollup, &vars.Auction, &vars.WDelayer})
|
||||||
|
return &u
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u *APIStateUpdater) Store() error {
|
||||||
|
return tracerr.Wrap(u.hdb.SetAPIState(&u.state))
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u *APIStateUpdater) SetSCVars(vars *common.SCVariablesPtr) {
|
||||||
|
if vars.Rollup != nil {
|
||||||
|
u.vars.Rollup = vars.Rollup
|
||||||
|
rollupVars := historydb.NewRollupVariablesAPI(u.vars.Rollup)
|
||||||
|
u.state.Rollup = *rollupVars
|
||||||
|
}
|
||||||
|
if vars.Auction != nil {
|
||||||
|
u.vars.Auction = vars.Auction
|
||||||
|
auctionVars := historydb.NewAuctionVariablesAPI(u.vars.Auction)
|
||||||
|
u.state.Auction = *auctionVars
|
||||||
|
}
|
||||||
|
if vars.WDelayer != nil {
|
||||||
|
u.vars.WDelayer = vars.WDelayer
|
||||||
|
u.state.WithdrawalDelayer = *u.vars.WDelayer
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u *APIStateUpdater) UpdateMetrics() error {
|
||||||
|
if u.state.Network.LastBatch == nil {
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
lastBatchNum := u.state.Network.LastBatch.BatchNum
|
||||||
|
metrics, err := u.hdb.GetMetricsInternalAPI(lastBatchNum)
|
||||||
|
if err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
u.state.Metrics = *metrics
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u *APIStateUpdater) UpdateNetworkInfoBlock(lastEthBlock, lastSyncBlock common.Block) {
|
||||||
|
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
||||||
|
u.state.Network.LastEthBlock = lastEthBlock.Num
|
||||||
|
}
|
||||||
|
|
||||||
|
func (u *APIStateUpdater) UpdateNetworkInfo(
|
||||||
|
lastEthBlock, lastSyncBlock common.Block,
|
||||||
|
lastBatchNum common.BatchNum, currentSlot int64,
|
||||||
|
) error {
|
||||||
|
// Get last batch in API format
|
||||||
|
lastBatch, err := u.hdb.GetBatchInternalAPI(lastBatchNum)
|
||||||
|
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||||
|
lastBatch = nil
|
||||||
|
} else if err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Get next forgers
|
||||||
|
lastClosedSlot := currentSlot + int64(u.state.Auction.ClosedAuctionSlots)
|
||||||
|
nextForgers, err := u.hdb.GetNextForgersInternalAPI(u.vars.Auction, &u.consts.Auction,
|
||||||
|
lastSyncBlock, currentSlot, lastClosedSlot)
|
||||||
|
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
||||||
|
nextForgers = nil
|
||||||
|
} else if err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
bucketUpdates, err := u.hdb.GetBucketUpdatesInternalAPI()
|
||||||
|
if err == sql.ErrNoRows {
|
||||||
|
bucketUpdates = nil
|
||||||
|
} else if err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Update NodeInfo struct
|
||||||
|
for i, bucketParams := range u.state.Rollup.Buckets {
|
||||||
|
for _, bucketUpdate := range bucketUpdates {
|
||||||
|
if bucketUpdate.NumBucket == i {
|
||||||
|
bucketParams.Withdrawals = bucketUpdate.Withdrawals
|
||||||
|
u.state.Rollup.Buckets[i] = bucketParams
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
||||||
|
u.state.Network.LastEthBlock = lastEthBlock.Num
|
||||||
|
u.state.Network.LastBatch = lastBatch
|
||||||
|
u.state.Network.CurrentSlot = currentSlot
|
||||||
|
u.state.Network.NextForgers = nextForgers
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -13,7 +13,7 @@ import (
|
|||||||
|
|
||||||
type testStatus struct {
|
type testStatus struct {
|
||||||
Network testNetwork `json:"network"`
|
Network testNetwork `json:"network"`
|
||||||
Metrics historydb.Metrics `json:"metrics"`
|
Metrics historydb.MetricsAPI `json:"metrics"`
|
||||||
Rollup historydb.RollupVariablesAPI `json:"rollup"`
|
Rollup historydb.RollupVariablesAPI `json:"rollup"`
|
||||||
Auction historydb.AuctionVariablesAPI `json:"auction"`
|
Auction historydb.AuctionVariablesAPI `json:"auction"`
|
||||||
WithdrawalDelayer common.WDelayerVariables `json:"withdrawalDelayer"`
|
WithdrawalDelayer common.WDelayerVariables `json:"withdrawalDelayer"`
|
||||||
@@ -21,11 +21,11 @@ type testStatus struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type testNetwork struct {
|
type testNetwork struct {
|
||||||
LastEthBlock int64 `json:"lastEthereumBlock"`
|
LastEthBlock int64 `json:"lastEthereumBlock"`
|
||||||
LastSyncBlock int64 `json:"lastSynchedBlock"`
|
LastSyncBlock int64 `json:"lastSynchedBlock"`
|
||||||
LastBatch testBatch `json:"lastBatch"`
|
LastBatch testBatch `json:"lastBatch"`
|
||||||
CurrentSlot int64 `json:"currentSlot"`
|
CurrentSlot int64 `json:"currentSlot"`
|
||||||
NextForgers []historydb.NextForger `json:"nextForgers"`
|
NextForgers []historydb.NextForgerAPI `json:"nextForgers"`
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestSetRollupVariables(t *testing.T) {
|
func TestSetRollupVariables(t *testing.T) {
|
||||||
@@ -211,7 +211,7 @@ func TestGetState(t *testing.T) {
|
|||||||
// status.RecommendedFee.ExistingAccount*createAccountInternalExtraFeePercentage)
|
// status.RecommendedFee.ExistingAccount*createAccountInternalExtraFeePercentage)
|
||||||
}
|
}
|
||||||
|
|
||||||
func assertNextForgers(t *testing.T, expected, actual []historydb.NextForger) {
|
func assertNextForgers(t *testing.T, expected, actual []historydb.NextForgerAPI) {
|
||||||
assert.Equal(t, len(expected), len(actual))
|
assert.Equal(t, len(expected), len(actual))
|
||||||
for i := range expected {
|
for i := range expected {
|
||||||
// ignore timestamps and other metadata
|
// ignore timestamps and other metadata
|
||||||
|
|||||||
31
common/eth.go
Normal file
31
common/eth.go
Normal file
@@ -0,0 +1,31 @@
|
|||||||
|
package common
|
||||||
|
|
||||||
|
// SCVariables joins all the smart contract variables in a single struct
|
||||||
|
type SCVariables struct {
|
||||||
|
Rollup RollupVariables `validate:"required"`
|
||||||
|
Auction AuctionVariables `validate:"required"`
|
||||||
|
WDelayer WDelayerVariables `validate:"required"`
|
||||||
|
}
|
||||||
|
|
||||||
|
func (v *SCVariables) AsPtr() *SCVariablesPtr {
|
||||||
|
return &SCVariablesPtr{
|
||||||
|
Rollup: &v.Rollup,
|
||||||
|
Auction: &v.Auction,
|
||||||
|
WDelayer: &v.WDelayer,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// SCVariablesPtr joins all the smart contract variables as pointers in a single
|
||||||
|
// struct
|
||||||
|
type SCVariablesPtr struct {
|
||||||
|
Rollup *RollupVariables `validate:"required"`
|
||||||
|
Auction *AuctionVariables `validate:"required"`
|
||||||
|
WDelayer *WDelayerVariables `validate:"required"`
|
||||||
|
}
|
||||||
|
|
||||||
|
// SCConsts joins all the smart contract constants in a single struct
|
||||||
|
type SCConsts struct {
|
||||||
|
Rollup RollupConstants
|
||||||
|
Auction AuctionConstants
|
||||||
|
WDelayer WDelayerConstants
|
||||||
|
}
|
||||||
@@ -185,8 +185,8 @@ func NewCoordinator(cfg Config,
|
|||||||
batchBuilder *batchbuilder.BatchBuilder,
|
batchBuilder *batchbuilder.BatchBuilder,
|
||||||
serverProofs []prover.Client,
|
serverProofs []prover.Client,
|
||||||
ethClient eth.ClientInterface,
|
ethClient eth.ClientInterface,
|
||||||
scConsts *synchronizer.SCConsts,
|
scConsts *common.SCConsts,
|
||||||
initSCVars *synchronizer.SCVariables,
|
initSCVars *common.SCVariables,
|
||||||
) (*Coordinator, error) {
|
) (*Coordinator, error) {
|
||||||
// nolint reason: hardcoded `1.0`, by design the percentage can't be over 100%
|
// nolint reason: hardcoded `1.0`, by design the percentage can't be over 100%
|
||||||
if cfg.L1BatchTimeoutPerc >= 1.0 { //nolint:gomnd
|
if cfg.L1BatchTimeoutPerc >= 1.0 { //nolint:gomnd
|
||||||
|
|||||||
@@ -1,8 +1,11 @@
|
|||||||
package historydb
|
package historydb
|
||||||
|
|
||||||
import (
|
import (
|
||||||
|
"database/sql"
|
||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
|
"math/big"
|
||||||
|
"time"
|
||||||
|
|
||||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
@@ -34,6 +37,12 @@ func (hdb *HistoryDB) GetBatchAPI(batchNum common.BatchNum) (*BatchAPI, error) {
|
|||||||
defer hdb.apiConnCon.Release()
|
defer hdb.apiConnCon.Release()
|
||||||
return hdb.getBatchAPI(hdb.dbRead, batchNum)
|
return hdb.getBatchAPI(hdb.dbRead, batchNum)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// GetBatchAPI return the batch with the given batchNum
|
||||||
|
func (hdb *HistoryDB) GetBatchInternalAPI(batchNum common.BatchNum) (*BatchAPI, error) {
|
||||||
|
return hdb.getBatchAPI(hdb.dbRead, batchNum)
|
||||||
|
}
|
||||||
|
|
||||||
func (hdb *HistoryDB) getBatchAPI(d meddler.DB, batchNum common.BatchNum) (*BatchAPI, error) {
|
func (hdb *HistoryDB) getBatchAPI(d meddler.DB, batchNum common.BatchNum) (*BatchAPI, error) {
|
||||||
batch := &BatchAPI{}
|
batch := &BatchAPI{}
|
||||||
return batch, tracerr.Wrap(meddler.QueryRow(
|
return batch, tracerr.Wrap(meddler.QueryRow(
|
||||||
@@ -954,3 +963,220 @@ func (hdb *HistoryDB) GetNodeInfoAPI() (*NodeInfo, error) {
|
|||||||
defer hdb.apiConnCon.Release()
|
defer hdb.apiConnCon.Release()
|
||||||
return hdb.GetNodeInfo()
|
return hdb.GetNodeInfo()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func (hdb *HistoryDB) GetBucketUpdatesInternalAPI() ([]BucketUpdateAPI, error) {
|
||||||
|
var bucketUpdates []*BucketUpdateAPI
|
||||||
|
// var bucketUpdates []*common.BucketUpdate
|
||||||
|
err := meddler.QueryAll(
|
||||||
|
hdb.dbRead, &bucketUpdates,
|
||||||
|
`SELECT num_bucket, withdrawals FROM bucket_update
|
||||||
|
WHERE item_id in(SELECT max(item_id) FROM bucket_update
|
||||||
|
group by num_bucket)
|
||||||
|
ORDER BY num_bucket ASC;`,
|
||||||
|
)
|
||||||
|
return db.SlicePtrsToSlice(bucketUpdates).([]BucketUpdateAPI), tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// getNextForgers returns next forgers
|
||||||
|
func (hdb *HistoryDB) GetNextForgersInternalAPI(auctionVars *common.AuctionVariables,
|
||||||
|
auctionConsts *common.AuctionConstants,
|
||||||
|
lastBlock common.Block, currentSlot, lastClosedSlot int64) ([]NextForgerAPI, error) {
|
||||||
|
secondsPerBlock := int64(15) //nolint:gomnd
|
||||||
|
// currentSlot and lastClosedSlot included
|
||||||
|
limit := uint(lastClosedSlot - currentSlot + 1)
|
||||||
|
bids, _, err := hdb.getBestBidsAPI(hdb.dbRead, ¤tSlot, &lastClosedSlot, nil, &limit, "ASC")
|
||||||
|
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
nextForgers := []NextForgerAPI{}
|
||||||
|
// Get min bid info
|
||||||
|
var minBidInfo []MinBidInfo
|
||||||
|
if currentSlot >= auctionVars.DefaultSlotSetBidSlotNum {
|
||||||
|
// All min bids can be calculated with the last update of AuctionVariables
|
||||||
|
|
||||||
|
minBidInfo = []MinBidInfo{{
|
||||||
|
DefaultSlotSetBid: auctionVars.DefaultSlotSetBid,
|
||||||
|
DefaultSlotSetBidSlotNum: auctionVars.DefaultSlotSetBidSlotNum,
|
||||||
|
}}
|
||||||
|
} else {
|
||||||
|
// Get all the relevant updates from the DB
|
||||||
|
minBidInfo, err = hdb.getMinBidInfo(hdb.dbRead, currentSlot, lastClosedSlot)
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// Create nextForger for each slot
|
||||||
|
for i := currentSlot; i <= lastClosedSlot; i++ {
|
||||||
|
fromBlock := i*int64(auctionConsts.BlocksPerSlot) +
|
||||||
|
auctionConsts.GenesisBlockNum
|
||||||
|
toBlock := (i+1)*int64(auctionConsts.BlocksPerSlot) +
|
||||||
|
auctionConsts.GenesisBlockNum - 1
|
||||||
|
nextForger := NextForgerAPI{
|
||||||
|
Period: Period{
|
||||||
|
SlotNum: i,
|
||||||
|
FromBlock: fromBlock,
|
||||||
|
ToBlock: toBlock,
|
||||||
|
FromTimestamp: lastBlock.Timestamp.Add(time.Second *
|
||||||
|
time.Duration(secondsPerBlock*(fromBlock-lastBlock.Num))),
|
||||||
|
ToTimestamp: lastBlock.Timestamp.Add(time.Second *
|
||||||
|
time.Duration(secondsPerBlock*(toBlock-lastBlock.Num))),
|
||||||
|
},
|
||||||
|
}
|
||||||
|
foundForger := false
|
||||||
|
// If there is a bid for a slot, get forger (coordinator)
|
||||||
|
for j := range bids {
|
||||||
|
slotNum := bids[j].SlotNum
|
||||||
|
if slotNum == i {
|
||||||
|
// There's a bid for the slot
|
||||||
|
// Check if the bid is greater than the minimum required
|
||||||
|
for i := 0; i < len(minBidInfo); i++ {
|
||||||
|
// Find the most recent update
|
||||||
|
if slotNum >= minBidInfo[i].DefaultSlotSetBidSlotNum {
|
||||||
|
// Get min bid
|
||||||
|
minBidSelector := slotNum % int64(len(auctionVars.DefaultSlotSetBid))
|
||||||
|
minBid := minBidInfo[i].DefaultSlotSetBid[minBidSelector]
|
||||||
|
// Check if the bid has beaten the minimum
|
||||||
|
bid, ok := new(big.Int).SetString(string(bids[j].BidValue), 10)
|
||||||
|
if !ok {
|
||||||
|
return nil, tracerr.New("Wrong bid value, error parsing it as big.Int")
|
||||||
|
}
|
||||||
|
if minBid.Cmp(bid) == 1 {
|
||||||
|
// Min bid is greater than bid, the slot will be forged by boot coordinator
|
||||||
|
break
|
||||||
|
}
|
||||||
|
foundForger = true
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if !foundForger { // There is no bid or it's smaller than the minimum
|
||||||
|
break
|
||||||
|
}
|
||||||
|
coordinator, err := hdb.getCoordinatorAPI(hdb.dbRead, bids[j].Bidder)
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
nextForger.Coordinator = *coordinator
|
||||||
|
break
|
||||||
|
}
|
||||||
|
}
|
||||||
|
// If there is no bid, the coordinator that will forge is boot coordinator
|
||||||
|
if !foundForger {
|
||||||
|
nextForger.Coordinator = CoordinatorAPI{
|
||||||
|
Forger: auctionVars.BootCoordinator,
|
||||||
|
URL: auctionVars.BootCoordinatorURL,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
nextForgers = append(nextForgers, nextForger)
|
||||||
|
}
|
||||||
|
return nextForgers, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// UpdateMetrics update Status.Metrics information
|
||||||
|
func (hdb *HistoryDB) GetMetricsInternalAPI(lastBatchNum common.BatchNum) (*MetricsAPI, error) {
|
||||||
|
var metrics MetricsAPI
|
||||||
|
// Get the first and last batch of the last 24h and their timestamps
|
||||||
|
// if u.state.Network.LastBatch == nil {
|
||||||
|
// return &metrics, nil
|
||||||
|
// }
|
||||||
|
type period struct {
|
||||||
|
FromBatchNum common.BatchNum `meddler:"from_batch_num"`
|
||||||
|
FromTimestamp time.Time `meddler:"from_timestamp"`
|
||||||
|
ToBatchNum common.BatchNum `meddler:"-"`
|
||||||
|
ToTimestamp time.Time `meddler:"to_timestamp"`
|
||||||
|
}
|
||||||
|
p := &period{
|
||||||
|
ToBatchNum: lastBatchNum,
|
||||||
|
}
|
||||||
|
if err := meddler.QueryRow(
|
||||||
|
hdb.dbRead, p, `SELECT
|
||||||
|
COALESCE (MIN(batch.batch_num), 0) as from_batch_num,
|
||||||
|
COALESCE (MIN(block.timestamp), NOW()) AS from_timestamp,
|
||||||
|
COALESCE (MAX(block.timestamp), NOW()) AS to_timestamp
|
||||||
|
FROM batch INNER JOIN block ON batch.eth_block_num = block.eth_block_num
|
||||||
|
WHERE block.timestamp >= NOW() - INTERVAL '24 HOURS';`,
|
||||||
|
); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Get the amount of txs of that period
|
||||||
|
row := hdb.dbRead.QueryRow(
|
||||||
|
`SELECT COUNT(*) as total_txs FROM tx WHERE tx.batch_num between $1 AND $2;`,
|
||||||
|
p.FromBatchNum, p.ToBatchNum,
|
||||||
|
)
|
||||||
|
var nTxs int
|
||||||
|
if err := row.Scan(&nTxs); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Set txs/s
|
||||||
|
seconds := p.ToTimestamp.Sub(p.FromTimestamp).Seconds()
|
||||||
|
if seconds == 0 { // Avoid dividing by 0
|
||||||
|
seconds++
|
||||||
|
}
|
||||||
|
metrics.TransactionsPerSecond = float64(nTxs) / seconds
|
||||||
|
// Set txs/batch
|
||||||
|
nBatches := p.ToBatchNum - p.FromBatchNum + 1
|
||||||
|
if nBatches == 0 { // Avoid dividing by 0
|
||||||
|
nBatches++
|
||||||
|
}
|
||||||
|
if (p.ToBatchNum - p.FromBatchNum) > 0 {
|
||||||
|
fmt.Printf("DBG ntxs: %v, nBatches: %v\n", nTxs, nBatches)
|
||||||
|
metrics.TransactionsPerBatch = float64(nTxs) /
|
||||||
|
float64(nBatches)
|
||||||
|
} else {
|
||||||
|
metrics.TransactionsPerBatch = 0
|
||||||
|
}
|
||||||
|
// Get total fee of that period
|
||||||
|
row = hdb.dbRead.QueryRow(
|
||||||
|
`SELECT COALESCE (SUM(total_fees_usd), 0) FROM batch WHERE batch_num between $1 AND $2;`,
|
||||||
|
p.FromBatchNum, p.ToBatchNum,
|
||||||
|
)
|
||||||
|
var totalFee float64
|
||||||
|
if err := row.Scan(&totalFee); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Set batch frequency
|
||||||
|
metrics.BatchFrequency = seconds / float64(nBatches)
|
||||||
|
if nTxs > 0 {
|
||||||
|
metrics.AvgTransactionFee = totalFee / float64(nTxs)
|
||||||
|
} else {
|
||||||
|
metrics.AvgTransactionFee = 0
|
||||||
|
}
|
||||||
|
// Get and set amount of registered accounts
|
||||||
|
type registeredAccounts struct {
|
||||||
|
TotalIdx int64 `meddler:"total_idx"`
|
||||||
|
TotalBJJ int64 `meddler:"total_bjj"`
|
||||||
|
}
|
||||||
|
ra := ®isteredAccounts{}
|
||||||
|
if err := meddler.QueryRow(
|
||||||
|
hdb.dbRead, ra,
|
||||||
|
`SELECT COUNT(*) AS total_bjj, COUNT(DISTINCT(bjj)) AS total_idx FROM account;`,
|
||||||
|
); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
metrics.TotalAccounts = ra.TotalIdx
|
||||||
|
metrics.TotalBJJs = ra.TotalBJJ
|
||||||
|
// Get and set estimated time to forge L1 tx
|
||||||
|
row = hdb.dbRead.QueryRow(
|
||||||
|
`SELECT COALESCE (AVG(EXTRACT(EPOCH FROM (forged.timestamp - added.timestamp))), 0) FROM tx
|
||||||
|
INNER JOIN block AS added ON tx.eth_block_num = added.eth_block_num
|
||||||
|
INNER JOIN batch AS forged_batch ON tx.batch_num = forged_batch.batch_num
|
||||||
|
INNER JOIN block AS forged ON forged_batch.eth_block_num = forged.eth_block_num
|
||||||
|
WHERE tx.batch_num between $1 and $2 AND tx.is_l1 AND tx.user_origin;`,
|
||||||
|
p.FromBatchNum, p.ToBatchNum,
|
||||||
|
)
|
||||||
|
var timeToForgeL1 float64
|
||||||
|
if err := row.Scan(&timeToForgeL1); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
metrics.EstimatedTimeToForgeL1 = timeToForgeL1
|
||||||
|
return &metrics, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (hdb *HistoryDB) GetStateAPI() (*StateAPI, error) {
|
||||||
|
cancel, err := hdb.apiConnCon.Acquire()
|
||||||
|
defer cancel()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
defer hdb.apiConnCon.Release()
|
||||||
|
return hdb.getStateAPI(hdb.dbRead)
|
||||||
|
}
|
||||||
|
|||||||
@@ -841,20 +841,7 @@ func (hdb *HistoryDB) GetAllBucketUpdates() ([]common.BucketUpdate, error) {
|
|||||||
return db.SlicePtrsToSlice(bucketUpdates).([]common.BucketUpdate), tracerr.Wrap(err)
|
return db.SlicePtrsToSlice(bucketUpdates).([]common.BucketUpdate), tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (hdb *HistoryDB) getBucketUpdatesAPI(txn *sqlx.Tx) ([]BucketUpdateAPI, error) {
|
func (hdb *HistoryDB) getMinBidInfo(d meddler.DB,
|
||||||
var bucketUpdates []*BucketUpdateAPI
|
|
||||||
// var bucketUpdates []*common.BucketUpdate
|
|
||||||
err := meddler.QueryAll(
|
|
||||||
txn, &bucketUpdates,
|
|
||||||
`SELECT num_bucket, withdrawals FROM bucket_update
|
|
||||||
WHERE item_id in(SELECT max(item_id) FROM bucket_update
|
|
||||||
group by num_bucket)
|
|
||||||
ORDER BY num_bucket ASC;`,
|
|
||||||
)
|
|
||||||
return db.SlicePtrsToSlice(bucketUpdates).([]BucketUpdateAPI), tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (hdb *HistoryDB) getMinBidInfo(txn *sqlx.Tx,
|
|
||||||
currentSlot, lastClosedSlot int64) ([]MinBidInfo, error) {
|
currentSlot, lastClosedSlot int64) ([]MinBidInfo, error) {
|
||||||
minBidInfo := []*MinBidInfo{}
|
minBidInfo := []*MinBidInfo{}
|
||||||
query := `
|
query := `
|
||||||
@@ -862,7 +849,7 @@ func (hdb *HistoryDB) getMinBidInfo(txn *sqlx.Tx,
|
|||||||
WHERE default_slot_set_bid_slot_num < $1
|
WHERE default_slot_set_bid_slot_num < $1
|
||||||
ORDER BY default_slot_set_bid_slot_num DESC
|
ORDER BY default_slot_set_bid_slot_num DESC
|
||||||
LIMIT $2;`
|
LIMIT $2;`
|
||||||
err := meddler.QueryAll(txn, &minBidInfo, query, lastClosedSlot, int(lastClosedSlot-currentSlot)+1)
|
err := meddler.QueryAll(d, &minBidInfo, query, lastClosedSlot, int(lastClosedSlot-currentSlot)+1)
|
||||||
return db.SlicePtrsToSlice(minBidInfo).([]MinBidInfo), tracerr.Wrap(err)
|
return db.SlicePtrsToSlice(minBidInfo).([]MinBidInfo), tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1183,3 +1170,49 @@ func (hdb *HistoryDB) GetTokensTest() ([]TokenWithUSD, error) {
|
|||||||
}
|
}
|
||||||
return db.SlicePtrsToSlice(tokens).([]TokenWithUSD), nil
|
return db.SlicePtrsToSlice(tokens).([]TokenWithUSD), nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// UpdateRecommendedFee update Status.RecommendedFee information
|
||||||
|
func (hdb *HistoryDB) GetRecommendedFee(minFeeUSD float64) (*common.RecommendedFee, error) {
|
||||||
|
var recommendedFee common.RecommendedFee
|
||||||
|
// Get total txs and the batch of the first selected tx of the last hour
|
||||||
|
type totalTxsSinceBatchNum struct {
|
||||||
|
TotalTxs int `meddler:"total_txs"`
|
||||||
|
FirstBatchNum common.BatchNum `meddler:"batch_num"`
|
||||||
|
}
|
||||||
|
ttsbn := &totalTxsSinceBatchNum{}
|
||||||
|
if err := meddler.QueryRow(
|
||||||
|
hdb.dbRead, ttsbn, `SELECT COUNT(tx.*) as total_txs,
|
||||||
|
COALESCE (MIN(tx.batch_num), 0) as batch_num
|
||||||
|
FROM tx INNER JOIN block ON tx.eth_block_num = block.eth_block_num
|
||||||
|
WHERE block.timestamp >= NOW() - INTERVAL '1 HOURS';`,
|
||||||
|
); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Get the amount of batches and acumulated fees for the last hour
|
||||||
|
type totalBatchesAndFee struct {
|
||||||
|
TotalBatches int `meddler:"total_batches"`
|
||||||
|
TotalFees float64 `meddler:"total_fees"`
|
||||||
|
}
|
||||||
|
tbf := &totalBatchesAndFee{}
|
||||||
|
if err := meddler.QueryRow(
|
||||||
|
hdb.dbRead, tbf, `SELECT COUNT(*) AS total_batches,
|
||||||
|
COALESCE (SUM(total_fees_usd), 0) AS total_fees FROM batch
|
||||||
|
WHERE batch_num > $1;`, ttsbn.FirstBatchNum,
|
||||||
|
); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// Update NodeInfo struct
|
||||||
|
var avgTransactionFee float64
|
||||||
|
if ttsbn.TotalTxs > 0 {
|
||||||
|
avgTransactionFee = tbf.TotalFees / float64(ttsbn.TotalTxs)
|
||||||
|
} else {
|
||||||
|
avgTransactionFee = 0
|
||||||
|
}
|
||||||
|
recommendedFee.ExistingAccount =
|
||||||
|
math.Max(avgTransactionFee, minFeeUSD)
|
||||||
|
recommendedFee.CreatesAccount =
|
||||||
|
math.Max(createAccountExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||||
|
recommendedFee.CreatesAccountAndRegister =
|
||||||
|
math.Max(createAccountInternalExtraFeePercentage*avgTransactionFee, minFeeUSD)
|
||||||
|
return &recommendedFee, nil
|
||||||
|
}
|
||||||
|
|||||||
@@ -1114,166 +1114,169 @@ func TestAddEscapeHatchWithdrawals(t *testing.T) {
|
|||||||
assert.Equal(t, escapeHatchWithdrawals, dbEscapeHatchWithdrawals)
|
assert.Equal(t, escapeHatchWithdrawals, dbEscapeHatchWithdrawals)
|
||||||
}
|
}
|
||||||
|
|
||||||
// func TestGetMetricsAPI(t *testing.T) {
|
func TestGetMetricsAPI(t *testing.T) {
|
||||||
// test.WipeDB(historyDB.DB())
|
test.WipeDB(historyDB.DB())
|
||||||
// set := `
|
set := `
|
||||||
// Type: Blockchain
|
Type: Blockchain
|
||||||
//
|
|
||||||
// AddToken(1)
|
AddToken(1)
|
||||||
//
|
|
||||||
// CreateAccountDeposit(1) A: 1000 // numTx=1
|
CreateAccountDeposit(1) A: 1000 // numTx=1
|
||||||
// CreateAccountDeposit(1) B: 2000 // numTx=2
|
CreateAccountDeposit(1) B: 2000 // numTx=2
|
||||||
// CreateAccountDeposit(1) C: 3000 //numTx=3
|
CreateAccountDeposit(1) C: 3000 //numTx=3
|
||||||
//
|
|
||||||
// // block 0 is stored as default in the DB
|
// block 0 is stored as default in the DB
|
||||||
// // block 1 does not exist
|
// block 1 does not exist
|
||||||
// > batchL1 // numBatches=1
|
> batchL1 // numBatches=1
|
||||||
// > batchL1 // numBatches=2
|
> batchL1 // numBatches=2
|
||||||
// > block // blockNum=2
|
> block // blockNum=2
|
||||||
//
|
|
||||||
// Transfer(1) C-A : 10 (1) // numTx=4
|
Transfer(1) C-A : 10 (1) // numTx=4
|
||||||
// > batch // numBatches=3
|
> batch // numBatches=3
|
||||||
// > block // blockNum=3
|
> block // blockNum=3
|
||||||
// Transfer(1) B-C : 10 (1) // numTx=5
|
Transfer(1) B-C : 10 (1) // numTx=5
|
||||||
// > batch // numBatches=5
|
> batch // numBatches=5
|
||||||
// > block // blockNum=4
|
> block // blockNum=4
|
||||||
// Transfer(1) A-B : 10 (1) // numTx=6
|
Transfer(1) A-B : 10 (1) // numTx=6
|
||||||
// > batch // numBatches=5
|
> batch // numBatches=5
|
||||||
// > block // blockNum=5
|
> block // blockNum=5
|
||||||
// Transfer(1) A-B : 10 (1) // numTx=7
|
Transfer(1) A-B : 10 (1) // numTx=7
|
||||||
// > batch // numBatches=6
|
> batch // numBatches=6
|
||||||
// > block // blockNum=6
|
> block // blockNum=6
|
||||||
// `
|
`
|
||||||
//
|
|
||||||
// const numBatches int = 6
|
const numBatches int = 6
|
||||||
// const numTx int = 7
|
const numTx int = 7
|
||||||
// const blockNum = 6 - 1
|
const blockNum = 6 - 1
|
||||||
//
|
|
||||||
// tc := til.NewContext(uint16(0), common.RollupConstMaxL1UserTx)
|
tc := til.NewContext(uint16(0), common.RollupConstMaxL1UserTx)
|
||||||
// tilCfgExtra := til.ConfigExtra{
|
tilCfgExtra := til.ConfigExtra{
|
||||||
// BootCoordAddr: ethCommon.HexToAddress("0xE39fEc6224708f0772D2A74fd3f9055A90E0A9f2"),
|
BootCoordAddr: ethCommon.HexToAddress("0xE39fEc6224708f0772D2A74fd3f9055A90E0A9f2"),
|
||||||
// CoordUser: "A",
|
CoordUser: "A",
|
||||||
// }
|
}
|
||||||
// blocks, err := tc.GenerateBlocks(set)
|
blocks, err := tc.GenerateBlocks(set)
|
||||||
// require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
// err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
||||||
// require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
//
|
|
||||||
// // Sanity check
|
// Sanity check
|
||||||
// require.Equal(t, blockNum, len(blocks))
|
require.Equal(t, blockNum, len(blocks))
|
||||||
//
|
|
||||||
// // Adding one batch per block
|
// Adding one batch per block
|
||||||
// // batch frequency can be chosen
|
// batch frequency can be chosen
|
||||||
// const frequency int = 15
|
const frequency int = 15
|
||||||
//
|
|
||||||
// for i := range blocks {
|
for i := range blocks {
|
||||||
// blocks[i].Block.Timestamp = time.Now().Add(-time.Second * time.Duration(frequency*(len(blocks)-i)))
|
blocks[i].Block.Timestamp = time.Now().Add(-time.Second * time.Duration(frequency*(len(blocks)-i)))
|
||||||
// err = historyDB.AddBlockSCData(&blocks[i])
|
err = historyDB.AddBlockSCData(&blocks[i])
|
||||||
// assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
// }
|
}
|
||||||
//
|
|
||||||
// res, err := historyDBWithACC.GetMetricsAPI(common.BatchNum(numBatches))
|
// clientSetupExample := test.NewClientSetupExample()
|
||||||
// assert.NoError(t, err)
|
// apiStateUpdater := NewAPIStateUpdater(historyDB, &NodeConfig{1000, 0.5},
|
||||||
//
|
// &Constants{
|
||||||
// assert.Equal(t, float64(numTx)/float64(numBatches), res.TransactionsPerBatch)
|
// RollupConstants: *clientSetupExample.RollupConstants,
|
||||||
//
|
// AuctionConstants: *clientSetupExample.AuctionConstants,
|
||||||
// // Frequency is not exactly the desired one, some decimals may appear
|
// WDelayerConstants: *clientSetupExample.WDelayerConstants,
|
||||||
// // There is a -2 as time for first and last batch is not taken into account
|
// ChainID: uint16(clientSetupExample.ChainID.Int64()),
|
||||||
// assert.InEpsilon(t, float64(frequency)*float64(numBatches-2)/float64(numBatches), res.BatchFrequency, 0.01)
|
// HermezAddress: clientSetupExample.AuctionConstants.HermezRollup,
|
||||||
// assert.InEpsilon(t, float64(numTx)/float64(frequency*blockNum-frequency), res.TransactionsPerSecond, 0.01)
|
// })
|
||||||
// assert.Equal(t, int64(3), res.TotalAccounts)
|
res, err := historyDB.GetMetricsInternalAPI(common.BatchNum(numBatches))
|
||||||
// assert.Equal(t, int64(3), res.TotalBJJs)
|
assert.NoError(t, err)
|
||||||
// // Til does not set fees
|
|
||||||
// assert.Equal(t, float64(0), res.AvgTransactionFee)
|
assert.Equal(t, float64(numTx)/float64(numBatches), res.TransactionsPerBatch)
|
||||||
// }
|
|
||||||
//
|
// Frequency is not exactly the desired one, some decimals may appear
|
||||||
// func TestGetMetricsAPIMoreThan24Hours(t *testing.T) {
|
// There is a -2 as time for first and last batch is not taken into account
|
||||||
// test.WipeDB(historyDB.DB())
|
assert.InEpsilon(t, float64(frequency)*float64(numBatches-2)/float64(numBatches), res.BatchFrequency, 0.01)
|
||||||
//
|
assert.InEpsilon(t, float64(numTx)/float64(frequency*blockNum-frequency), res.TransactionsPerSecond, 0.01)
|
||||||
// testUsersLen := 3
|
assert.Equal(t, int64(3), res.TotalAccounts)
|
||||||
// var set []til.Instruction
|
assert.Equal(t, int64(3), res.TotalBJJs)
|
||||||
// for user := 0; user < testUsersLen; user++ {
|
// Til does not set fees
|
||||||
// set = append(set, til.Instruction{
|
assert.Equal(t, float64(0), res.AvgTransactionFee)
|
||||||
// Typ: common.TxTypeCreateAccountDeposit,
|
}
|
||||||
// TokenID: common.TokenID(0),
|
|
||||||
// DepositAmount: big.NewInt(1000000),
|
func TestGetMetricsAPIMoreThan24Hours(t *testing.T) {
|
||||||
// Amount: big.NewInt(0),
|
test.WipeDB(historyDB.DB())
|
||||||
// From: fmt.Sprintf("User%02d", user),
|
|
||||||
// })
|
testUsersLen := 3
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
var set []til.Instruction
|
||||||
// }
|
for user := 0; user < testUsersLen; user++ {
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBatchL1})
|
set = append(set, til.Instruction{
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBatchL1})
|
Typ: common.TxTypeCreateAccountDeposit,
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
TokenID: common.TokenID(0),
|
||||||
//
|
DepositAmount: big.NewInt(1000000),
|
||||||
// // Transfers
|
Amount: big.NewInt(0),
|
||||||
// const numBlocks int = 30
|
From: fmt.Sprintf("User%02d", user),
|
||||||
// for x := 0; x < numBlocks; x++ {
|
})
|
||||||
// set = append(set, til.Instruction{
|
set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
||||||
// Typ: common.TxTypeTransfer,
|
}
|
||||||
// TokenID: common.TokenID(0),
|
set = append(set, til.Instruction{Typ: til.TypeNewBatchL1})
|
||||||
// DepositAmount: big.NewInt(1),
|
set = append(set, til.Instruction{Typ: til.TypeNewBatchL1})
|
||||||
// Amount: big.NewInt(0),
|
set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
||||||
// From: "User00",
|
|
||||||
// To: "User01",
|
// Transfers
|
||||||
// })
|
const numBlocks int = 30
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBatch})
|
for x := 0; x < numBlocks; x++ {
|
||||||
// set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
set = append(set, til.Instruction{
|
||||||
// }
|
Typ: common.TxTypeTransfer,
|
||||||
//
|
TokenID: common.TokenID(0),
|
||||||
// var chainID uint16 = 0
|
DepositAmount: big.NewInt(1),
|
||||||
// tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
Amount: big.NewInt(0),
|
||||||
// blocks, err := tc.GenerateBlocksFromInstructions(set)
|
From: "User00",
|
||||||
// assert.NoError(t, err)
|
To: "User01",
|
||||||
//
|
})
|
||||||
// tilCfgExtra := til.ConfigExtra{
|
set = append(set, til.Instruction{Typ: til.TypeNewBatch})
|
||||||
// CoordUser: "A",
|
set = append(set, til.Instruction{Typ: til.TypeNewBlock})
|
||||||
// }
|
}
|
||||||
// err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
|
||||||
// require.NoError(t, err)
|
var chainID uint16 = 0
|
||||||
//
|
tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||||
// const numBatches int = 2 + numBlocks
|
blocks, err := tc.GenerateBlocksFromInstructions(set)
|
||||||
// const blockNum = 4 + numBlocks
|
assert.NoError(t, err)
|
||||||
//
|
|
||||||
// // Sanity check
|
tilCfgExtra := til.ConfigExtra{
|
||||||
// require.Equal(t, blockNum, len(blocks))
|
CoordUser: "A",
|
||||||
//
|
}
|
||||||
// // Adding one batch per block
|
err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
||||||
// // batch frequency can be chosen
|
require.NoError(t, err)
|
||||||
// const blockTime time.Duration = 3600 * time.Second
|
|
||||||
// now := time.Now()
|
const numBatches int = 2 + numBlocks
|
||||||
// require.NoError(t, err)
|
const blockNum = 4 + numBlocks
|
||||||
//
|
|
||||||
// for i := range blocks {
|
// Sanity check
|
||||||
// blocks[i].Block.Timestamp = now.Add(-time.Duration(len(blocks)-1-i) * blockTime)
|
require.Equal(t, blockNum, len(blocks))
|
||||||
// err = historyDB.AddBlockSCData(&blocks[i])
|
|
||||||
// assert.NoError(t, err)
|
// Adding one batch per block
|
||||||
// }
|
// batch frequency can be chosen
|
||||||
//
|
const blockTime time.Duration = 3600 * time.Second
|
||||||
// res, err := historyDBWithACC.GetMetricsAPI(common.BatchNum(numBatches))
|
now := time.Now()
|
||||||
// assert.NoError(t, err)
|
require.NoError(t, err)
|
||||||
//
|
|
||||||
// assert.InEpsilon(t, 1.0, res.TransactionsPerBatch, 0.1)
|
for i := range blocks {
|
||||||
//
|
blocks[i].Block.Timestamp = now.Add(-time.Duration(len(blocks)-1-i) * blockTime)
|
||||||
// assert.InEpsilon(t, res.BatchFrequency, float64(blockTime/time.Second), 0.1)
|
err = historyDB.AddBlockSCData(&blocks[i])
|
||||||
// assert.InEpsilon(t, 1.0/float64(blockTime/time.Second), res.TransactionsPerSecond, 0.1)
|
assert.NoError(t, err)
|
||||||
// assert.Equal(t, int64(3), res.TotalAccounts)
|
}
|
||||||
// assert.Equal(t, int64(3), res.TotalBJJs)
|
|
||||||
// // Til does not set fees
|
res, err := historyDBWithACC.GetMetricsInternalAPI(common.BatchNum(numBatches))
|
||||||
// assert.Equal(t, float64(0), res.AvgTransactionFee)
|
assert.NoError(t, err)
|
||||||
// }
|
|
||||||
//
|
assert.InEpsilon(t, 1.0, res.TransactionsPerBatch, 0.1)
|
||||||
// func TestGetMetricsAPIEmpty(t *testing.T) {
|
|
||||||
// test.WipeDB(historyDB.DB())
|
assert.InEpsilon(t, res.BatchFrequency, float64(blockTime/time.Second), 0.1)
|
||||||
// _, err := historyDBWithACC.GetMetricsAPI(0)
|
assert.InEpsilon(t, 1.0/float64(blockTime/time.Second), res.TransactionsPerSecond, 0.1)
|
||||||
// assert.NoError(t, err)
|
assert.Equal(t, int64(3), res.TotalAccounts)
|
||||||
// }
|
assert.Equal(t, int64(3), res.TotalBJJs)
|
||||||
//
|
// Til does not set fees
|
||||||
// func TestGetAvgTxFeeEmpty(t *testing.T) {
|
assert.Equal(t, float64(0), res.AvgTransactionFee)
|
||||||
// test.WipeDB(historyDB.DB())
|
}
|
||||||
// _, err := historyDBWithACC.GetAvgTxFeeAPI()
|
|
||||||
// assert.NoError(t, err)
|
func TestGetMetricsAPIEmpty(t *testing.T) {
|
||||||
// }
|
test.WipeDB(historyDB.DB())
|
||||||
|
_, err := historyDBWithACC.GetMetricsInternalAPI(0)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
func TestGetLastL1TxsNum(t *testing.T) {
|
func TestGetLastL1TxsNum(t *testing.T) {
|
||||||
test.WipeDB(historyDB.DB())
|
test.WipeDB(historyDB.DB())
|
||||||
@@ -1464,30 +1467,32 @@ func setTestBlocks(from, to int64) []common.Block {
|
|||||||
func TestNodeInfo(t *testing.T) {
|
func TestNodeInfo(t *testing.T) {
|
||||||
test.WipeDB(historyDB.DB())
|
test.WipeDB(historyDB.DB())
|
||||||
|
|
||||||
err := historyDB.SetAPIState(&APIState{})
|
err := historyDB.SetAPIState(&StateAPI{})
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
clientSetup := test.NewClientSetupExample()
|
clientSetup := test.NewClientSetupExample()
|
||||||
constants := &Constants{
|
constants := &Constants{
|
||||||
RollupConstants: *clientSetup.RollupConstants,
|
SCConsts: common.SCConsts{
|
||||||
AuctionConstants: *clientSetup.AuctionConstants,
|
Rollup: *clientSetup.RollupConstants,
|
||||||
WDelayerConstants: *clientSetup.WDelayerConstants,
|
Auction: *clientSetup.AuctionConstants,
|
||||||
ChainID: 42,
|
WDelayer: *clientSetup.WDelayerConstants,
|
||||||
HermezAddress: clientSetup.AuctionConstants.HermezRollup,
|
},
|
||||||
|
ChainID: 42,
|
||||||
|
HermezAddress: clientSetup.AuctionConstants.HermezRollup,
|
||||||
}
|
}
|
||||||
err = historyDB.SetConstants(constants)
|
err = historyDB.SetConstants(constants)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
// Test parameters
|
// Test parameters
|
||||||
apiState := &APIState{
|
stateAPI := &StateAPI{
|
||||||
NodePublicConfig: NodePublicConfig{
|
NodePublicConfig: NodePublicConfig{
|
||||||
ForgeDelay: 3.1,
|
ForgeDelay: 3.1,
|
||||||
},
|
},
|
||||||
Network: Network{
|
Network: NetworkAPI{
|
||||||
LastEthBlock: 12,
|
LastEthBlock: 12,
|
||||||
LastSyncBlock: 34,
|
LastSyncBlock: 34,
|
||||||
},
|
},
|
||||||
Metrics: Metrics{
|
Metrics: MetricsAPI{
|
||||||
TransactionsPerBatch: 1.1,
|
TransactionsPerBatch: 1.1,
|
||||||
TotalAccounts: 42,
|
TotalAccounts: 42,
|
||||||
},
|
},
|
||||||
@@ -1498,7 +1503,7 @@ func TestNodeInfo(t *testing.T) {
|
|||||||
ExistingAccount: 0.15,
|
ExistingAccount: 0.15,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
err = historyDB.SetAPIState(apiState)
|
err = historyDB.SetAPIState(stateAPI)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
nodeConfig := &NodeConfig{
|
nodeConfig := &NodeConfig{
|
||||||
@@ -1516,7 +1521,7 @@ func TestNodeInfo(t *testing.T) {
|
|||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, nodeConfig, dbNodeConfig)
|
assert.Equal(t, nodeConfig, dbNodeConfig)
|
||||||
|
|
||||||
dbAPIState, err := historyDB.GetAPIState()
|
dbStateAPI, err := historyDB.GetStateAPI()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, apiState, dbAPIState)
|
assert.Equal(t, stateAPI, dbStateAPI)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,18 +1,11 @@
|
|||||||
package historydb
|
package historydb
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"database/sql"
|
|
||||||
"fmt"
|
|
||||||
"math"
|
|
||||||
"math/big"
|
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/hermeznetwork/hermez-node/apitypes"
|
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
"github.com/hermeznetwork/hermez-node/db"
|
|
||||||
"github.com/hermeznetwork/tracerr"
|
"github.com/hermeznetwork/tracerr"
|
||||||
"github.com/jmoiron/sqlx"
|
|
||||||
"github.com/russross/meddler"
|
"github.com/russross/meddler"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -29,17 +22,17 @@ type Period struct {
|
|||||||
ToTimestamp time.Time `json:"toTimestamp"`
|
ToTimestamp time.Time `json:"toTimestamp"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type NextForger struct {
|
type NextForgerAPI struct {
|
||||||
Coordinator CoordinatorAPI `json:"coordinator"`
|
Coordinator CoordinatorAPI `json:"coordinator"`
|
||||||
Period Period `json:"period"`
|
Period Period `json:"period"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type Network struct {
|
type NetworkAPI struct {
|
||||||
LastEthBlock int64 `json:"lastEthereumBlock"`
|
LastEthBlock int64 `json:"lastEthereumBlock"`
|
||||||
LastSyncBlock int64 `json:"lastSynchedBlock"`
|
LastSyncBlock int64 `json:"lastSynchedBlock"`
|
||||||
LastBatch *BatchAPI `json:"lastBatch"`
|
LastBatch *BatchAPI `json:"lastBatch"`
|
||||||
CurrentSlot int64 `json:"currentSlot"`
|
CurrentSlot int64 `json:"currentSlot"`
|
||||||
NextForgers []NextForger `json:"nextForgers"`
|
NextForgers []NextForgerAPI `json:"nextForgers"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// NodePublicConfig is the configuration of the node that is exposed via API
|
// NodePublicConfig is the configuration of the node that is exposed via API
|
||||||
@@ -48,11 +41,11 @@ type NodePublicConfig struct {
|
|||||||
ForgeDelay float64 `json:"forgeDelay"`
|
ForgeDelay float64 `json:"forgeDelay"`
|
||||||
}
|
}
|
||||||
|
|
||||||
type APIState struct {
|
type StateAPI struct {
|
||||||
// NodePublicConfig is the configuration of the node that is exposed via API
|
// NodePublicConfig is the configuration of the node that is exposed via API
|
||||||
NodePublicConfig NodePublicConfig `json:"nodeConfig"`
|
NodePublicConfig NodePublicConfig `json:"nodeConfig"`
|
||||||
Network Network `json:"network"`
|
Network NetworkAPI `json:"network"`
|
||||||
Metrics Metrics `json:"metrics"`
|
Metrics MetricsAPI `json:"metrics"`
|
||||||
Rollup RollupVariablesAPI `json:"rollup"`
|
Rollup RollupVariablesAPI `json:"rollup"`
|
||||||
Auction AuctionVariablesAPI `json:"auction"`
|
Auction AuctionVariablesAPI `json:"auction"`
|
||||||
WithdrawalDelayer common.WDelayerVariables `json:"withdrawalDelayer"`
|
WithdrawalDelayer common.WDelayerVariables `json:"withdrawalDelayer"`
|
||||||
@@ -60,11 +53,12 @@ type APIState struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
type Constants struct {
|
type Constants struct {
|
||||||
RollupConstants common.RollupConstants
|
// RollupConstants common.RollupConstants
|
||||||
AuctionConstants common.AuctionConstants
|
// AuctionConstants common.AuctionConstants
|
||||||
WDelayerConstants common.WDelayerConstants
|
// WDelayerConstants common.WDelayerConstants
|
||||||
ChainID uint16
|
common.SCConsts
|
||||||
HermezAddress ethCommon.Address
|
ChainID uint16
|
||||||
|
HermezAddress ethCommon.Address
|
||||||
}
|
}
|
||||||
|
|
||||||
type NodeConfig struct {
|
type NodeConfig struct {
|
||||||
@@ -74,7 +68,7 @@ type NodeConfig struct {
|
|||||||
|
|
||||||
type NodeInfo struct {
|
type NodeInfo struct {
|
||||||
ItemID int `meddler:"item_id,pk"`
|
ItemID int `meddler:"item_id,pk"`
|
||||||
APIState *APIState `meddler:"state,json"`
|
APIState *StateAPI `meddler:"state,json"`
|
||||||
NodeConfig *NodeConfig `meddler:"config,json"`
|
NodeConfig *NodeConfig `meddler:"config,json"`
|
||||||
Constants *Constants `meddler:"constants,json"`
|
Constants *Constants `meddler:"constants,json"`
|
||||||
}
|
}
|
||||||
@@ -111,18 +105,22 @@ func (hdb *HistoryDB) SetConstants(constants *Constants) error {
|
|||||||
return tracerr.Wrap(err)
|
return tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (hdb *HistoryDB) GetAPIState() (*APIState, error) {
|
func (hdb *HistoryDB) GetStateInternalAPI() (*StateAPI, error) {
|
||||||
|
return hdb.getStateAPI(hdb.dbRead)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (hdb *HistoryDB) getStateAPI(d meddler.DB) (*StateAPI, error) {
|
||||||
var nodeInfo NodeInfo
|
var nodeInfo NodeInfo
|
||||||
err := meddler.QueryRow(
|
err := meddler.QueryRow(
|
||||||
hdb.dbRead, &nodeInfo,
|
d, &nodeInfo,
|
||||||
"SELECT state FROM node_info WHERE item_id = 1;",
|
"SELECT state FROM node_info WHERE item_id = 1;",
|
||||||
)
|
)
|
||||||
return nodeInfo.APIState, tracerr.Wrap(err)
|
return nodeInfo.APIState, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
func (hdb *HistoryDB) SetAPIState(apiState *APIState) error {
|
func (hdb *HistoryDB) SetAPIState(apiState *StateAPI) error {
|
||||||
_apiState := struct {
|
_apiState := struct {
|
||||||
APIState *APIState `meddler:"state,json"`
|
APIState *StateAPI `meddler:"state,json"`
|
||||||
}{apiState}
|
}{apiState}
|
||||||
values, err := meddler.Default.Values(&_apiState, false)
|
values, err := meddler.Default.Values(&_apiState, false)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -168,474 +166,50 @@ func (hdb *HistoryDB) SetNodeConfig(nodeConfig *NodeConfig) error {
|
|||||||
// return tracerr.Wrap(meddler.Insert(hdb.dbWrite, "node_info", ni))
|
// return tracerr.Wrap(meddler.Insert(hdb.dbWrite, "node_info", ni))
|
||||||
// }
|
// }
|
||||||
|
|
||||||
type APIStateUpdater struct {
|
|
||||||
hdb *HistoryDB
|
|
||||||
state APIState
|
|
||||||
config NodeConfig
|
|
||||||
constants Constants
|
|
||||||
}
|
|
||||||
|
|
||||||
func (u *APIStateUpdater) SetSCVars(rollupVariables *common.RollupVariables,
|
|
||||||
auctionVariables *common.AuctionVariables,
|
|
||||||
wDelayerVariables *common.WDelayerVariables) {
|
|
||||||
if rollupVariables != nil {
|
|
||||||
rollupVars := NewRollupVariablesAPI(rollupVariables)
|
|
||||||
u.state.Rollup = *rollupVars
|
|
||||||
}
|
|
||||||
if auctionVariables != nil {
|
|
||||||
auctionVars := NewAuctionVariablesAPI(auctionVariables)
|
|
||||||
u.state.Auction = *auctionVars
|
|
||||||
}
|
|
||||||
if wDelayerVariables != nil {
|
|
||||||
u.state.WithdrawalDelayer = *wDelayerVariables
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func (u *APIStateUpdater) UpdateNetworkInfoBlock(lastEthBlock, lastSyncBlock common.Block) {
|
|
||||||
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
|
||||||
u.state.Network.LastEthBlock = lastEthBlock.Num
|
|
||||||
}
|
|
||||||
|
|
||||||
func (u *APIStateUpdater) UpdateNetworkInfo(
|
|
||||||
txn *sqlx.Tx,
|
|
||||||
lastEthBlock, lastSyncBlock common.Block,
|
|
||||||
lastBatchNum common.BatchNum, currentSlot int64,
|
|
||||||
) error {
|
|
||||||
// Get last batch in API format
|
|
||||||
lastBatch, err := u.hdb.getBatchAPI(txn, lastBatchNum)
|
|
||||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
|
||||||
lastBatch = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Get next forrgers
|
|
||||||
lastClosedSlot := currentSlot + int64(u.state.Auction.ClosedAuctionSlots)
|
|
||||||
nextForgers, err := u.getNextForgers(txn, lastSyncBlock, currentSlot, lastClosedSlot)
|
|
||||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
|
||||||
nextForgers = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
bucketUpdates, err := u.hdb.getBucketUpdatesAPI(txn)
|
|
||||||
if err == sql.ErrNoRows {
|
|
||||||
bucketUpdates = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Update NodeInfo struct
|
|
||||||
for i, bucketParams := range u.state.Rollup.Buckets {
|
|
||||||
for _, bucketUpdate := range bucketUpdates {
|
|
||||||
if bucketUpdate.NumBucket == i {
|
|
||||||
bucketParams.Withdrawals = bucketUpdate.Withdrawals
|
|
||||||
u.state.Rollup.Buckets[i] = bucketParams
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
u.state.Network.LastSyncBlock = lastSyncBlock.Num
|
|
||||||
u.state.Network.LastEthBlock = lastEthBlock.Num
|
|
||||||
u.state.Network.LastBatch = lastBatch
|
|
||||||
u.state.Network.CurrentSlot = currentSlot
|
|
||||||
u.state.Network.NextForgers = nextForgers
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Remove
|
|
||||||
// SetRollupVariables set Status.Rollup variables
|
|
||||||
func (hdb *HistoryDB) SetRollupVariables(rollupVariables *common.RollupVariables) error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
rollupVars := NewRollupVariablesAPI(rollupVariables)
|
|
||||||
ni.APIState.Rollup = *rollupVars
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Remove
|
|
||||||
// SetWDelayerVariables set Status.WithdrawalDelayer variables
|
|
||||||
func (hdb *HistoryDB) SetWDelayerVariables(wDelayerVariables *common.WDelayerVariables) error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
ni.APIState.WithdrawalDelayer = *wDelayerVariables
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Remove
|
|
||||||
// SetAuctionVariables set Status.Auction variables
|
|
||||||
func (hdb *HistoryDB) SetAuctionVariables(auctionVariables *common.AuctionVariables) error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
auctionVars := NewAuctionVariablesAPI(auctionVariables)
|
|
||||||
ni.APIState.Auction = *auctionVars
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Remove
|
|
||||||
// UpdateNetworkInfoBlock update Status.Network block related information
|
|
||||||
func (hdb *HistoryDB) UpdateNetworkInfoBlock(
|
|
||||||
lastEthBlock, lastSyncBlock common.Block,
|
|
||||||
) error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
ni.APIState.Network.LastSyncBlock = lastSyncBlock.Num
|
|
||||||
ni.APIState.Network.LastEthBlock = lastEthBlock.Num
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// UpdateNetworkInfo update Status.Network information
|
|
||||||
func (hdb *HistoryDB) UpdateNetworkInfo(
|
|
||||||
lastEthBlock, lastSyncBlock common.Block,
|
|
||||||
lastBatchNum common.BatchNum, currentSlot int64,
|
|
||||||
) error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
// Get last batch in API format
|
|
||||||
lastBatch, err := hdb.getBatchAPI(txn, lastBatchNum)
|
|
||||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
|
||||||
lastBatch = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Get next forrgers
|
|
||||||
lastClosedSlot := currentSlot + int64(ni.APIState.Auction.ClosedAuctionSlots)
|
|
||||||
nextForgers, err := hdb.getNextForgers(txn, ni, lastSyncBlock, currentSlot, lastClosedSlot)
|
|
||||||
if tracerr.Unwrap(err) == sql.ErrNoRows {
|
|
||||||
nextForgers = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get buckets withdrawals
|
|
||||||
var bucketUpdatesPtrs []*BucketUpdateAPI
|
|
||||||
var bucketUpdates []BucketUpdateAPI
|
|
||||||
err = meddler.QueryAll(
|
|
||||||
txn, &bucketUpdatesPtrs,
|
|
||||||
`SELECT num_bucket, withdrawals FROM bucket_update
|
|
||||||
WHERE item_id in(SELECT max(item_id) FROM bucket_update
|
|
||||||
group by num_bucket)
|
|
||||||
ORDER BY num_bucket ASC;`,
|
|
||||||
)
|
|
||||||
if err == sql.ErrNoRows {
|
|
||||||
bucketUpdates = nil
|
|
||||||
} else if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
} else {
|
|
||||||
bucketUpdates = db.SlicePtrsToSlice(bucketUpdatesPtrs).([]BucketUpdateAPI)
|
|
||||||
}
|
|
||||||
// Update NodeInfo struct
|
|
||||||
for i, bucketParams := range ni.APIState.Rollup.Buckets {
|
|
||||||
for _, bucketUpdate := range bucketUpdates {
|
|
||||||
if bucketUpdate.NumBucket == i {
|
|
||||||
bucketParams.Withdrawals = bucketUpdate.Withdrawals
|
|
||||||
ni.APIState.Rollup.Buckets[i] = bucketParams
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
ni.APIState.Network.LastSyncBlock = lastSyncBlock.Num
|
|
||||||
ni.APIState.Network.LastEthBlock = lastEthBlock.Num
|
|
||||||
ni.APIState.Network.LastBatch = lastBatch
|
|
||||||
ni.APIState.Network.CurrentSlot = currentSlot
|
|
||||||
ni.APIState.Network.NextForgers = nextForgers
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// apiSlotToBigInts converts from [6]*apitypes.BigIntStr to [6]*big.Int
|
// apiSlotToBigInts converts from [6]*apitypes.BigIntStr to [6]*big.Int
|
||||||
func apiSlotToBigInts(defaultSlotSetBid [6]*apitypes.BigIntStr) ([6]*big.Int, error) {
|
// func apiSlotToBigInts(defaultSlotSetBid [6]*apitypes.BigIntStr) ([6]*big.Int, error) {
|
||||||
var slots [6]*big.Int
|
// var slots [6]*big.Int
|
||||||
|
//
|
||||||
|
// for i, slot := range defaultSlotSetBid {
|
||||||
|
// bigInt, ok := new(big.Int).SetString(string(*slot), 10)
|
||||||
|
// if !ok {
|
||||||
|
// return slots, tracerr.Wrap(fmt.Errorf("can't convert %T into big.Int", slot))
|
||||||
|
// }
|
||||||
|
// slots[i] = bigInt
|
||||||
|
// }
|
||||||
|
//
|
||||||
|
// return slots, nil
|
||||||
|
// }
|
||||||
|
|
||||||
for i, slot := range defaultSlotSetBid {
|
// func (hdb *HistoryDB) updateNodeInfo(setUpdatedNodeInfo func(*sqlx.Tx, *NodeInfo) error) error {
|
||||||
bigInt, ok := new(big.Int).SetString(string(*slot), 10)
|
// // Create a SQL transaction or read and update atomicaly
|
||||||
if !ok {
|
// txn, err := hdb.dbWrite.Beginx()
|
||||||
return slots, tracerr.Wrap(fmt.Errorf("can't convert %T into big.Int", slot))
|
// if err != nil {
|
||||||
}
|
// return tracerr.Wrap(err)
|
||||||
slots[i] = bigInt
|
// }
|
||||||
}
|
// defer func() {
|
||||||
|
// if err != nil {
|
||||||
return slots, nil
|
// db.Rollback(txn)
|
||||||
}
|
// }
|
||||||
|
// }()
|
||||||
// getNextForgers returns next forgers
|
// // Read current node info
|
||||||
func (u *APIStateUpdater) getNextForgers(txn *sqlx.Tx,
|
// ni := &NodeInfo{}
|
||||||
lastBlock common.Block, currentSlot, lastClosedSlot int64) ([]NextForger, error) {
|
// if err := meddler.QueryRow(
|
||||||
secondsPerBlock := int64(15) //nolint:gomnd
|
// txn, ni, "SELECT * FROM node_info;",
|
||||||
// currentSlot and lastClosedSlot included
|
// ); err != nil {
|
||||||
limit := uint(lastClosedSlot - currentSlot + 1)
|
// return tracerr.Wrap(err)
|
||||||
bids, _, err := u.hdb.getBestBidsAPI(txn, ¤tSlot, &lastClosedSlot, nil, &limit, "ASC")
|
// }
|
||||||
if err != nil && tracerr.Unwrap(err) != sql.ErrNoRows {
|
// // Update NodeInfo struct
|
||||||
return nil, tracerr.Wrap(err)
|
// if err := setUpdatedNodeInfo(txn, ni); err != nil {
|
||||||
}
|
// return tracerr.Wrap(err)
|
||||||
nextForgers := []NextForger{}
|
// }
|
||||||
// Get min bid info
|
// // Update NodeInfo at DB
|
||||||
var minBidInfo []MinBidInfo
|
// if _, err := txn.Exec("DELETE FROM node_info;"); err != nil {
|
||||||
if currentSlot >= u.state.Auction.DefaultSlotSetBidSlotNum {
|
// return tracerr.Wrap(err)
|
||||||
// All min bids can be calculated with the last update of AuctionVariables
|
// }
|
||||||
bigIntSlots, err := apiSlotToBigInts(u.state.Auction.DefaultSlotSetBid)
|
// if err := meddler.Insert(txn, "node_info", ni); err != nil {
|
||||||
if err != nil {
|
// return tracerr.Wrap(err)
|
||||||
return nil, tracerr.Wrap(err)
|
// }
|
||||||
}
|
// // Commit NodeInfo update
|
||||||
|
// return tracerr.Wrap(txn.Commit())
|
||||||
minBidInfo = []MinBidInfo{{
|
// }
|
||||||
DefaultSlotSetBid: bigIntSlots,
|
|
||||||
DefaultSlotSetBidSlotNum: u.state.Auction.DefaultSlotSetBidSlotNum,
|
|
||||||
}}
|
|
||||||
} else {
|
|
||||||
// Get all the relevant updates from the DB
|
|
||||||
minBidInfo, err = u.hdb.getMinBidInfo(txn, currentSlot, lastClosedSlot)
|
|
||||||
if err != nil {
|
|
||||||
return nil, tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// Create nextForger for each slot
|
|
||||||
for i := currentSlot; i <= lastClosedSlot; i++ {
|
|
||||||
fromBlock := i*int64(u.constants.AuctionConstants.BlocksPerSlot) +
|
|
||||||
u.constants.AuctionConstants.GenesisBlockNum
|
|
||||||
toBlock := (i+1)*int64(u.constants.AuctionConstants.BlocksPerSlot) +
|
|
||||||
u.constants.AuctionConstants.GenesisBlockNum - 1
|
|
||||||
nextForger := NextForger{
|
|
||||||
Period: Period{
|
|
||||||
SlotNum: i,
|
|
||||||
FromBlock: fromBlock,
|
|
||||||
ToBlock: toBlock,
|
|
||||||
FromTimestamp: lastBlock.Timestamp.Add(time.Second *
|
|
||||||
time.Duration(secondsPerBlock*(fromBlock-lastBlock.Num))),
|
|
||||||
ToTimestamp: lastBlock.Timestamp.Add(time.Second *
|
|
||||||
time.Duration(secondsPerBlock*(toBlock-lastBlock.Num))),
|
|
||||||
},
|
|
||||||
}
|
|
||||||
foundForger := false
|
|
||||||
// If there is a bid for a slot, get forger (coordinator)
|
|
||||||
for j := range bids {
|
|
||||||
slotNum := bids[j].SlotNum
|
|
||||||
if slotNum == i {
|
|
||||||
// There's a bid for the slot
|
|
||||||
// Check if the bid is greater than the minimum required
|
|
||||||
for i := 0; i < len(minBidInfo); i++ {
|
|
||||||
// Find the most recent update
|
|
||||||
if slotNum >= minBidInfo[i].DefaultSlotSetBidSlotNum {
|
|
||||||
// Get min bid
|
|
||||||
minBidSelector := slotNum % int64(len(u.state.Auction.DefaultSlotSetBid))
|
|
||||||
minBid := minBidInfo[i].DefaultSlotSetBid[minBidSelector]
|
|
||||||
// Check if the bid has beaten the minimum
|
|
||||||
bid, ok := new(big.Int).SetString(string(bids[j].BidValue), 10)
|
|
||||||
if !ok {
|
|
||||||
return nil, tracerr.New("Wrong bid value, error parsing it as big.Int")
|
|
||||||
}
|
|
||||||
if minBid.Cmp(bid) == 1 {
|
|
||||||
// Min bid is greater than bid, the slot will be forged by boot coordinator
|
|
||||||
break
|
|
||||||
}
|
|
||||||
foundForger = true
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
if !foundForger { // There is no bid or it's smaller than the minimum
|
|
||||||
break
|
|
||||||
}
|
|
||||||
coordinator, err := u.hdb.GetCoordinatorAPI(bids[j].Bidder)
|
|
||||||
if err != nil {
|
|
||||||
return nil, tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
nextForger.Coordinator = *coordinator
|
|
||||||
break
|
|
||||||
}
|
|
||||||
}
|
|
||||||
// If there is no bid, the coordinator that will forge is boot coordinator
|
|
||||||
if !foundForger {
|
|
||||||
nextForger.Coordinator = CoordinatorAPI{
|
|
||||||
Forger: u.state.Auction.BootCoordinator,
|
|
||||||
URL: u.state.Auction.BootCoordinatorURL,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
nextForgers = append(nextForgers, nextForger)
|
|
||||||
}
|
|
||||||
return nextForgers, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
// TODO: Rename to getMetrics and don't write anything
|
|
||||||
// UpdateMetrics update Status.Metrics information
|
|
||||||
func (hdb *HistoryDB) UpdateMetrics() error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
// Get the first and last batch of the last 24h and their timestamps
|
|
||||||
if ni.APIState.Network.LastBatch == nil {
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
type period struct {
|
|
||||||
FromBatchNum common.BatchNum `meddler:"from_batch_num"`
|
|
||||||
FromTimestamp time.Time `meddler:"from_timestamp"`
|
|
||||||
ToBatchNum common.BatchNum `meddler:"-"`
|
|
||||||
ToTimestamp time.Time `meddler:"to_timestamp"`
|
|
||||||
}
|
|
||||||
p := &period{
|
|
||||||
ToBatchNum: ni.APIState.Network.LastBatch.BatchNum,
|
|
||||||
}
|
|
||||||
if err := meddler.QueryRow(
|
|
||||||
txn, p, `SELECT
|
|
||||||
COALESCE (MIN(batch.batch_num), 0) as from_batch_num,
|
|
||||||
COALESCE (MIN(block.timestamp), NOW()) AS from_timestamp,
|
|
||||||
COALESCE (MAX(block.timestamp), NOW()) AS to_timestamp
|
|
||||||
FROM batch INNER JOIN block ON batch.eth_block_num = block.eth_block_num
|
|
||||||
WHERE block.timestamp >= NOW() - INTERVAL '24 HOURS';`,
|
|
||||||
); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Get the amount of txs of that period
|
|
||||||
row := txn.QueryRow(
|
|
||||||
`SELECT COUNT(*) as total_txs FROM tx WHERE tx.batch_num between $1 AND $2;`,
|
|
||||||
p.FromBatchNum, p.ToBatchNum,
|
|
||||||
)
|
|
||||||
var nTxs int
|
|
||||||
if err := row.Scan(&nTxs); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Set txs/s
|
|
||||||
seconds := p.ToTimestamp.Sub(p.FromTimestamp).Seconds()
|
|
||||||
if seconds == 0 { // Avoid dividing by 0
|
|
||||||
seconds++
|
|
||||||
}
|
|
||||||
ni.APIState.Metrics.TransactionsPerSecond = float64(nTxs) / seconds
|
|
||||||
// Set txs/batch
|
|
||||||
nBatches := p.ToBatchNum - p.FromBatchNum
|
|
||||||
if nBatches == 0 { // Avoid dividing by 0
|
|
||||||
nBatches++
|
|
||||||
}
|
|
||||||
if (p.ToBatchNum - p.FromBatchNum) > 0 {
|
|
||||||
ni.APIState.Metrics.TransactionsPerBatch = float64(nTxs) /
|
|
||||||
float64(nBatches)
|
|
||||||
} else {
|
|
||||||
ni.APIState.Metrics.TransactionsPerBatch = 0
|
|
||||||
}
|
|
||||||
// Get total fee of that period
|
|
||||||
row = txn.QueryRow(
|
|
||||||
`SELECT COALESCE (SUM(total_fees_usd), 0) FROM batch WHERE batch_num between $1 AND $2;`,
|
|
||||||
p.FromBatchNum, p.ToBatchNum,
|
|
||||||
)
|
|
||||||
var totalFee float64
|
|
||||||
if err := row.Scan(&totalFee); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Set batch frequency
|
|
||||||
ni.APIState.Metrics.BatchFrequency = seconds / float64(nBatches)
|
|
||||||
if nTxs > 0 {
|
|
||||||
ni.APIState.Metrics.AvgTransactionFee = totalFee / float64(nTxs)
|
|
||||||
} else {
|
|
||||||
ni.APIState.Metrics.AvgTransactionFee = 0
|
|
||||||
}
|
|
||||||
// Get and set amount of registered accounts
|
|
||||||
type registeredAccounts struct {
|
|
||||||
TotalIdx int64 `meddler:"total_idx"`
|
|
||||||
TotalBJJ int64 `meddler:"total_bjj"`
|
|
||||||
}
|
|
||||||
ra := ®isteredAccounts{}
|
|
||||||
if err := meddler.QueryRow(
|
|
||||||
txn, ra,
|
|
||||||
`SELECT COUNT(*) AS total_bjj, COUNT(DISTINCT(bjj)) AS total_idx FROM account;`,
|
|
||||||
); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
ni.APIState.Metrics.TotalAccounts = ra.TotalIdx
|
|
||||||
ni.APIState.Metrics.TotalBJJs = ra.TotalBJJ
|
|
||||||
// Get and set estimated time to forge L1 tx
|
|
||||||
row = txn.QueryRow(
|
|
||||||
`SELECT COALESCE (AVG(EXTRACT(EPOCH FROM (forged.timestamp - added.timestamp))), 0) FROM tx
|
|
||||||
INNER JOIN block AS added ON tx.eth_block_num = added.eth_block_num
|
|
||||||
INNER JOIN batch AS forged_batch ON tx.batch_num = forged_batch.batch_num
|
|
||||||
INNER JOIN block AS forged ON forged_batch.eth_block_num = forged.eth_block_num
|
|
||||||
WHERE tx.batch_num between $1 and $2 AND tx.is_l1 AND tx.user_origin;`,
|
|
||||||
p.FromBatchNum, p.ToBatchNum,
|
|
||||||
)
|
|
||||||
var timeToForgeL1 float64
|
|
||||||
if err := row.Scan(&timeToForgeL1); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
ni.APIState.Metrics.EstimatedTimeToForgeL1 = timeToForgeL1
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
// UpdateRecommendedFee update Status.RecommendedFee information
|
|
||||||
func (hdb *HistoryDB) UpdateRecommendedFee() error {
|
|
||||||
setUpdatedNodeInfo := func(txn *sqlx.Tx, ni *NodeInfo) error {
|
|
||||||
// Get total txs and the batch of the first selected tx of the last hour
|
|
||||||
type totalTxsSinceBatchNum struct {
|
|
||||||
TotalTxs int `meddler:"total_txs"`
|
|
||||||
FirstBatchNum common.BatchNum `meddler:"batch_num"`
|
|
||||||
}
|
|
||||||
ttsbn := &totalTxsSinceBatchNum{}
|
|
||||||
if err := meddler.QueryRow(
|
|
||||||
txn, ttsbn, `SELECT COUNT(tx.*) as total_txs,
|
|
||||||
COALESCE (MIN(tx.batch_num), 0) as batch_num
|
|
||||||
FROM tx INNER JOIN block ON tx.eth_block_num = block.eth_block_num
|
|
||||||
WHERE block.timestamp >= NOW() - INTERVAL '1 HOURS';`,
|
|
||||||
); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Get the amount of batches and acumulated fees for the last hour
|
|
||||||
type totalBatchesAndFee struct {
|
|
||||||
TotalBatches int `meddler:"total_batches"`
|
|
||||||
TotalFees float64 `meddler:"total_fees"`
|
|
||||||
}
|
|
||||||
tbf := &totalBatchesAndFee{}
|
|
||||||
if err := meddler.QueryRow(
|
|
||||||
txn, tbf, `SELECT COUNT(*) AS total_batches,
|
|
||||||
COALESCE (SUM(total_fees_usd), 0) AS total_fees FROM batch
|
|
||||||
WHERE batch_num > $1;`, ttsbn.FirstBatchNum,
|
|
||||||
); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Update NodeInfo struct
|
|
||||||
var avgTransactionFee float64
|
|
||||||
if ttsbn.TotalTxs > 0 {
|
|
||||||
avgTransactionFee = tbf.TotalFees / float64(ttsbn.TotalTxs)
|
|
||||||
} else {
|
|
||||||
avgTransactionFee = 0
|
|
||||||
}
|
|
||||||
ni.APIState.RecommendedFee.ExistingAccount =
|
|
||||||
math.Max(avgTransactionFee, *ni.MinFeeUSD)
|
|
||||||
ni.APIState.RecommendedFee.CreatesAccount =
|
|
||||||
math.Max(createAccountExtraFeePercentage*avgTransactionFee, *ni.MinFeeUSD)
|
|
||||||
ni.APIState.RecommendedFee.CreatesAccountAndRegister =
|
|
||||||
math.Max(createAccountInternalExtraFeePercentage*avgTransactionFee, *ni.MinFeeUSD)
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
return hdb.updateNodeInfo(setUpdatedNodeInfo)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (hdb *HistoryDB) updateNodeInfo(setUpdatedNodeInfo func(*sqlx.Tx, *NodeInfo) error) error {
|
|
||||||
// Create a SQL transaction or read and update atomicaly
|
|
||||||
txn, err := hdb.dbWrite.Beginx()
|
|
||||||
if err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
defer func() {
|
|
||||||
if err != nil {
|
|
||||||
db.Rollback(txn)
|
|
||||||
}
|
|
||||||
}()
|
|
||||||
// Read current node info
|
|
||||||
ni := &NodeInfo{}
|
|
||||||
if err := meddler.QueryRow(
|
|
||||||
txn, ni, "SELECT * FROM node_info;",
|
|
||||||
); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Update NodeInfo struct
|
|
||||||
if err := setUpdatedNodeInfo(txn, ni); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Update NodeInfo at DB
|
|
||||||
if _, err := txn.Exec("DELETE FROM node_info;"); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
if err := meddler.Insert(txn, "node_info", ni); err != nil {
|
|
||||||
return tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
// Commit NodeInfo update
|
|
||||||
return tracerr.Wrap(txn.Commit())
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -302,8 +302,8 @@ type BatchAPI struct {
|
|||||||
LastItem uint64 `json:"-" meddler:"last_item"`
|
LastItem uint64 `json:"-" meddler:"last_item"`
|
||||||
}
|
}
|
||||||
|
|
||||||
// Metrics define metrics of the network
|
// MetricsAPI define metrics of the network
|
||||||
type Metrics struct {
|
type MetricsAPI struct {
|
||||||
TransactionsPerBatch float64 `json:"transactionsPerBatch"`
|
TransactionsPerBatch float64 `json:"transactionsPerBatch"`
|
||||||
BatchFrequency float64 `json:"batchFrequency"`
|
BatchFrequency float64 `json:"batchFrequency"`
|
||||||
TransactionsPerSecond float64 `json:"transactionsPerSecond"`
|
TransactionsPerSecond float64 `json:"transactionsPerSecond"`
|
||||||
|
|||||||
154
node/node.go
154
node/node.go
@@ -53,9 +53,10 @@ const (
|
|||||||
|
|
||||||
// Node is the Hermez Node
|
// Node is the Hermez Node
|
||||||
type Node struct {
|
type Node struct {
|
||||||
nodeAPI *NodeAPI
|
nodeAPI *NodeAPI
|
||||||
debugAPI *debugapi.DebugAPI
|
apiStateUpdater *api.APIStateUpdater
|
||||||
priceUpdater *priceupdater.PriceUpdater
|
debugAPI *debugapi.DebugAPI
|
||||||
|
priceUpdater *priceupdater.PriceUpdater
|
||||||
// Coordinator
|
// Coordinator
|
||||||
coord *coordinator.Coordinator
|
coord *coordinator.Coordinator
|
||||||
|
|
||||||
@@ -230,25 +231,33 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
}
|
}
|
||||||
initSCVars := sync.SCVars()
|
initSCVars := sync.SCVars()
|
||||||
|
|
||||||
scConsts := synchronizer.SCConsts{
|
scConsts := common.SCConsts{
|
||||||
Rollup: *sync.RollupConstants(),
|
Rollup: *sync.RollupConstants(),
|
||||||
Auction: *sync.AuctionConstants(),
|
Auction: *sync.AuctionConstants(),
|
||||||
WDelayer: *sync.WDelayerConstants(),
|
WDelayer: *sync.WDelayerConstants(),
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := historyDB.SetInitialNodeInfo(
|
hdbNodeCfg := historydb.NodeConfig{
|
||||||
cfg.Coordinator.L2DB.MaxTxs,
|
MaxPoolTxs: cfg.Coordinator.L2DB.MaxTxs,
|
||||||
cfg.Coordinator.L2DB.MinFeeUSD,
|
MinFeeUSD: cfg.Coordinator.L2DB.MinFeeUSD,
|
||||||
&historydb.Constants{
|
}
|
||||||
RollupConstants: scConsts.Rollup,
|
if err := historyDB.SetNodeConfig(&hdbNodeCfg); err != nil {
|
||||||
AuctionConstants: scConsts.Auction,
|
|
||||||
WDelayerConstants: scConsts.WDelayer,
|
|
||||||
ChainID: chainIDU16,
|
|
||||||
HermezAddress: cfg.SmartContracts.Rollup,
|
|
||||||
},
|
|
||||||
); err != nil {
|
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
hdbConsts := historydb.Constants{
|
||||||
|
SCConsts: common.SCConsts{
|
||||||
|
Rollup: scConsts.Rollup,
|
||||||
|
Auction: scConsts.Auction,
|
||||||
|
WDelayer: scConsts.WDelayer,
|
||||||
|
},
|
||||||
|
ChainID: chainIDU16,
|
||||||
|
HermezAddress: cfg.SmartContracts.Rollup,
|
||||||
|
}
|
||||||
|
if err := historyDB.SetConstants(&hdbConsts); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
apiStateUpdater := api.NewAPIStateUpdater(historyDB, &hdbNodeCfg, initSCVars, &hdbConsts)
|
||||||
|
|
||||||
var coord *coordinator.Coordinator
|
var coord *coordinator.Coordinator
|
||||||
var l2DB *l2db.L2DB
|
var l2DB *l2db.L2DB
|
||||||
@@ -379,11 +388,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
serverProofs,
|
serverProofs,
|
||||||
client,
|
client,
|
||||||
&scConsts,
|
&scConsts,
|
||||||
&synchronizer.SCVariables{
|
initSCVars,
|
||||||
Rollup: *initSCVars.Rollup,
|
|
||||||
Auction: *initSCVars.Auction,
|
|
||||||
WDelayer: *initSCVars.WDelayer,
|
|
||||||
},
|
|
||||||
)
|
)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
@@ -432,18 +437,19 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
}
|
}
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
return &Node{
|
return &Node{
|
||||||
nodeAPI: nodeAPI,
|
apiStateUpdater: apiStateUpdater,
|
||||||
debugAPI: debugAPI,
|
nodeAPI: nodeAPI,
|
||||||
priceUpdater: priceUpdater,
|
debugAPI: debugAPI,
|
||||||
coord: coord,
|
priceUpdater: priceUpdater,
|
||||||
sync: sync,
|
coord: coord,
|
||||||
cfg: cfg,
|
sync: sync,
|
||||||
mode: mode,
|
cfg: cfg,
|
||||||
sqlConnRead: dbRead,
|
mode: mode,
|
||||||
sqlConnWrite: dbWrite,
|
sqlConnRead: dbRead,
|
||||||
historyDB: historyDB,
|
sqlConnWrite: dbWrite,
|
||||||
ctx: ctx,
|
historyDB: historyDB,
|
||||||
cancel: cancel,
|
ctx: ctx,
|
||||||
|
cancel: cancel,
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -616,8 +622,8 @@ func (a *NodeAPI) Run(ctx context.Context) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, vars synchronizer.SCVariablesPtr,
|
func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, vars *common.SCVariablesPtr,
|
||||||
batches []common.BatchData) {
|
batches []common.BatchData) error {
|
||||||
if n.mode == ModeCoordinator {
|
if n.mode == ModeCoordinator {
|
||||||
n.coord.SendMsg(ctx, coordinator.MsgSyncBlock{
|
n.coord.SendMsg(ctx, coordinator.MsgSyncBlock{
|
||||||
Stats: *stats,
|
Stats: *stats,
|
||||||
@@ -625,47 +631,42 @@ func (n *Node) handleNewBlock(ctx context.Context, stats *synchronizer.Stats, va
|
|||||||
Batches: batches,
|
Batches: batches,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
if n.nodeAPI != nil {
|
n.apiStateUpdater.SetSCVars(vars)
|
||||||
if vars.Rollup != nil {
|
if stats.Synced() {
|
||||||
n.historyDB.SetRollupVariables(vars.Rollup)
|
if err := n.apiStateUpdater.UpdateNetworkInfo(
|
||||||
}
|
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||||
if vars.Auction != nil {
|
common.BatchNum(stats.Eth.LastBatchNum),
|
||||||
n.historyDB.SetAuctionVariables(vars.Auction)
|
stats.Sync.Auction.CurrentSlot.SlotNum,
|
||||||
}
|
); err != nil {
|
||||||
if vars.WDelayer != nil {
|
log.Errorw("ApiStateUpdater.UpdateNetworkInfo", "err", err)
|
||||||
n.historyDB.SetWDelayerVariables(vars.WDelayer)
|
|
||||||
}
|
|
||||||
|
|
||||||
if stats.Synced() {
|
|
||||||
if err := n.historyDB.UpdateNetworkInfo(
|
|
||||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
|
||||||
common.BatchNum(stats.Eth.LastBatchNum),
|
|
||||||
stats.Sync.Auction.CurrentSlot.SlotNum,
|
|
||||||
); err != nil {
|
|
||||||
log.Errorw("API.UpdateNetworkInfo", "err", err)
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
n.historyDB.UpdateNetworkInfoBlock(
|
|
||||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
|
||||||
)
|
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
n.apiStateUpdater.UpdateNetworkInfoBlock(
|
||||||
|
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||||
|
)
|
||||||
}
|
}
|
||||||
|
if err := n.apiStateUpdater.Store(); err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats, vars synchronizer.SCVariablesPtr) {
|
func (n *Node) handleReorg(ctx context.Context, stats *synchronizer.Stats,
|
||||||
|
vars *common.SCVariables) error {
|
||||||
if n.mode == ModeCoordinator {
|
if n.mode == ModeCoordinator {
|
||||||
n.coord.SendMsg(ctx, coordinator.MsgSyncReorg{
|
n.coord.SendMsg(ctx, coordinator.MsgSyncReorg{
|
||||||
Stats: *stats,
|
Stats: *stats,
|
||||||
Vars: vars,
|
Vars: vars,
|
||||||
})
|
})
|
||||||
}
|
}
|
||||||
vars = n.sync.SCVars()
|
n.apiStateUpdater.SetSCVars(vars.AsPtr())
|
||||||
n.historyDB.SetRollupVariables(vars.Rollup)
|
n.apiStateUpdater.UpdateNetworkInfoBlock(
|
||||||
n.historyDB.SetAuctionVariables(vars.Auction)
|
|
||||||
n.historyDB.SetWDelayerVariables(vars.WDelayer)
|
|
||||||
n.historyDB.UpdateNetworkInfoBlock(
|
|
||||||
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
stats.Eth.LastBlock, stats.Sync.LastBlock,
|
||||||
)
|
)
|
||||||
|
if err := n.apiStateUpdater.Store(); err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
// TODO(Edu): Consider keeping the `lastBlock` inside synchronizer so that we
|
||||||
@@ -680,16 +681,20 @@ func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common
|
|||||||
// case: reorg
|
// case: reorg
|
||||||
log.Infow("Synchronizer.Sync reorg", "discarded", *discarded)
|
log.Infow("Synchronizer.Sync reorg", "discarded", *discarded)
|
||||||
vars := n.sync.SCVars()
|
vars := n.sync.SCVars()
|
||||||
n.handleReorg(ctx, stats, vars)
|
if err := n.handleReorg(ctx, stats, vars); err != nil {
|
||||||
|
return nil, time.Duration(0), tracerr.Wrap(err)
|
||||||
|
}
|
||||||
return nil, time.Duration(0), nil
|
return nil, time.Duration(0), nil
|
||||||
} else if blockData != nil {
|
} else if blockData != nil {
|
||||||
// case: new block
|
// case: new block
|
||||||
vars := synchronizer.SCVariablesPtr{
|
vars := common.SCVariablesPtr{
|
||||||
Rollup: blockData.Rollup.Vars,
|
Rollup: blockData.Rollup.Vars,
|
||||||
Auction: blockData.Auction.Vars,
|
Auction: blockData.Auction.Vars,
|
||||||
WDelayer: blockData.WDelayer.Vars,
|
WDelayer: blockData.WDelayer.Vars,
|
||||||
}
|
}
|
||||||
n.handleNewBlock(ctx, stats, vars, blockData.Rollup.Batches)
|
if err := n.handleNewBlock(ctx, stats, &vars, blockData.Rollup.Batches); err != nil {
|
||||||
|
return nil, time.Duration(0), tracerr.Wrap(err)
|
||||||
|
}
|
||||||
return &blockData.Block, time.Duration(0), nil
|
return &blockData.Block, time.Duration(0), nil
|
||||||
} else {
|
} else {
|
||||||
// case: no block
|
// case: no block
|
||||||
@@ -708,7 +713,7 @@ func (n *Node) StartSynchronizer() {
|
|||||||
// the last synced one) is synchronized
|
// the last synced one) is synchronized
|
||||||
stats := n.sync.Stats()
|
stats := n.sync.Stats()
|
||||||
vars := n.sync.SCVars()
|
vars := n.sync.SCVars()
|
||||||
n.handleNewBlock(n.ctx, stats, vars, []common.BatchData{})
|
n.handleNewBlock(n.ctx, stats, vars.AsPtr(), []common.BatchData{})
|
||||||
|
|
||||||
n.wg.Add(1)
|
n.wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
@@ -795,8 +800,11 @@ func (n *Node) StartNodeAPI() {
|
|||||||
n.wg.Add(1)
|
n.wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
// Do an initial update on startup
|
// Do an initial update on startup
|
||||||
if err := n.historyDB.UpdateMetrics(); err != nil {
|
if err := n.apiStateUpdater.UpdateMetrics(); err != nil {
|
||||||
log.Errorw("API.UpdateMetrics", "err", err)
|
log.Errorw("ApiStateUpdater.UpdateMetrics", "err", err)
|
||||||
|
}
|
||||||
|
if err := n.apiStateUpdater.Store(); err != nil {
|
||||||
|
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||||
}
|
}
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
@@ -805,8 +813,12 @@ func (n *Node) StartNodeAPI() {
|
|||||||
n.wg.Done()
|
n.wg.Done()
|
||||||
return
|
return
|
||||||
case <-time.After(n.cfg.API.UpdateMetricsInterval.Duration):
|
case <-time.After(n.cfg.API.UpdateMetricsInterval.Duration):
|
||||||
if err := n.historyDB.UpdateMetrics(); err != nil {
|
if err := n.apiStateUpdater.UpdateMetrics(); err != nil {
|
||||||
log.Errorw("API.UpdateMetrics", "err", err)
|
log.Errorw("ApiStateUpdater.UpdateMetrics", "err", err)
|
||||||
|
continue
|
||||||
|
}
|
||||||
|
if err := n.apiStateUpdater.Store(); err != nil {
|
||||||
|
log.Errorw("ApiStateUpdater.Store", "err", err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -183,26 +183,26 @@ type StartBlockNums struct {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// SCVariables joins all the smart contract variables in a single struct
|
// SCVariables joins all the smart contract variables in a single struct
|
||||||
type SCVariables struct {
|
// type SCVariables struct {
|
||||||
Rollup common.RollupVariables `validate:"required"`
|
// Rollup common.RollupVariables `validate:"required"`
|
||||||
Auction common.AuctionVariables `validate:"required"`
|
// Auction common.AuctionVariables `validate:"required"`
|
||||||
WDelayer common.WDelayerVariables `validate:"required"`
|
// WDelayer common.WDelayerVariables `validate:"required"`
|
||||||
}
|
// }
|
||||||
|
//
|
||||||
// SCVariablesPtr joins all the smart contract variables as pointers in a single
|
// // SCVariablesPtr joins all the smart contract variables as pointers in a single
|
||||||
// struct
|
// // struct
|
||||||
type SCVariablesPtr struct {
|
// type SCVariablesPtr struct {
|
||||||
Rollup *common.RollupVariables `validate:"required"`
|
// Rollup *common.RollupVariables `validate:"required"`
|
||||||
Auction *common.AuctionVariables `validate:"required"`
|
// Auction *common.AuctionVariables `validate:"required"`
|
||||||
WDelayer *common.WDelayerVariables `validate:"required"`
|
// WDelayer *common.WDelayerVariables `validate:"required"`
|
||||||
}
|
// }
|
||||||
|
//
|
||||||
// SCConsts joins all the smart contract constants in a single struct
|
// // SCConsts joins all the smart contract constants in a single struct
|
||||||
type SCConsts struct {
|
// type SCConsts struct {
|
||||||
Rollup common.RollupConstants
|
// Rollup common.RollupConstants
|
||||||
Auction common.AuctionConstants
|
// Auction common.AuctionConstants
|
||||||
WDelayer common.WDelayerConstants
|
// WDelayer common.WDelayerConstants
|
||||||
}
|
// }
|
||||||
|
|
||||||
// Config is the Synchronizer configuration
|
// Config is the Synchronizer configuration
|
||||||
type Config struct {
|
type Config struct {
|
||||||
@@ -213,13 +213,13 @@ type Config struct {
|
|||||||
// Synchronizer implements the Synchronizer type
|
// Synchronizer implements the Synchronizer type
|
||||||
type Synchronizer struct {
|
type Synchronizer struct {
|
||||||
ethClient eth.ClientInterface
|
ethClient eth.ClientInterface
|
||||||
consts SCConsts
|
consts common.SCConsts
|
||||||
historyDB *historydb.HistoryDB
|
historyDB *historydb.HistoryDB
|
||||||
stateDB *statedb.StateDB
|
stateDB *statedb.StateDB
|
||||||
cfg Config
|
cfg Config
|
||||||
initVars SCVariables
|
initVars common.SCVariables
|
||||||
startBlockNum int64
|
startBlockNum int64
|
||||||
vars SCVariables
|
vars common.SCVariables
|
||||||
stats *StatsHolder
|
stats *StatsHolder
|
||||||
resetStateFailed bool
|
resetStateFailed bool
|
||||||
}
|
}
|
||||||
@@ -242,7 +242,7 @@ func NewSynchronizer(ethClient eth.ClientInterface, historyDB *historydb.History
|
|||||||
return nil, tracerr.Wrap(fmt.Errorf("NewSynchronizer ethClient.WDelayerConstants(): %w",
|
return nil, tracerr.Wrap(fmt.Errorf("NewSynchronizer ethClient.WDelayerConstants(): %w",
|
||||||
err))
|
err))
|
||||||
}
|
}
|
||||||
consts := SCConsts{
|
consts := common.SCConsts{
|
||||||
Rollup: *rollupConstants,
|
Rollup: *rollupConstants,
|
||||||
Auction: *auctionConstants,
|
Auction: *auctionConstants,
|
||||||
WDelayer: *wDelayerConstants,
|
WDelayer: *wDelayerConstants,
|
||||||
@@ -307,11 +307,11 @@ func (s *Synchronizer) WDelayerConstants() *common.WDelayerConstants {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// SCVars returns a copy of the Smart Contract Variables
|
// SCVars returns a copy of the Smart Contract Variables
|
||||||
func (s *Synchronizer) SCVars() SCVariablesPtr {
|
func (s *Synchronizer) SCVars() *common.SCVariables {
|
||||||
return SCVariablesPtr{
|
return &common.SCVariables{
|
||||||
Rollup: s.vars.Rollup.Copy(),
|
Rollup: *s.vars.Rollup.Copy(),
|
||||||
Auction: s.vars.Auction.Copy(),
|
Auction: *s.vars.Auction.Copy(),
|
||||||
WDelayer: s.vars.WDelayer.Copy(),
|
WDelayer: *s.vars.WDelayer.Copy(),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -724,7 +724,7 @@ func (s *Synchronizer) reorg(uncleBlock *common.Block) (int64, error) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func getInitialVariables(ethClient eth.ClientInterface,
|
func getInitialVariables(ethClient eth.ClientInterface,
|
||||||
consts *SCConsts) (*SCVariables, *StartBlockNums, error) {
|
consts *common.SCConsts) (*common.SCVariables, *StartBlockNums, error) {
|
||||||
rollupInit, rollupInitBlock, err := ethClient.RollupEventInit()
|
rollupInit, rollupInitBlock, err := ethClient.RollupEventInit()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, tracerr.Wrap(fmt.Errorf("RollupEventInit: %w", err))
|
return nil, nil, tracerr.Wrap(fmt.Errorf("RollupEventInit: %w", err))
|
||||||
@@ -740,7 +740,7 @@ func getInitialVariables(ethClient eth.ClientInterface,
|
|||||||
rollupVars := rollupInit.RollupVariables()
|
rollupVars := rollupInit.RollupVariables()
|
||||||
auctionVars := auctionInit.AuctionVariables(consts.Auction.InitialMinimalBidding)
|
auctionVars := auctionInit.AuctionVariables(consts.Auction.InitialMinimalBidding)
|
||||||
wDelayerVars := wDelayerInit.WDelayerVariables()
|
wDelayerVars := wDelayerInit.WDelayerVariables()
|
||||||
return &SCVariables{
|
return &common.SCVariables{
|
||||||
Rollup: *rollupVars,
|
Rollup: *rollupVars,
|
||||||
Auction: *auctionVars,
|
Auction: *auctionVars,
|
||||||
WDelayer: *wDelayerVars,
|
WDelayer: *wDelayerVars,
|
||||||
|
|||||||
Reference in New Issue
Block a user