Fix forging L1Batch too earlyfeature/sql-semaphore1
@ -0,0 +1,428 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"context" |
||||
|
"fmt" |
||||
|
"math/big" |
||||
|
"sync" |
||||
|
"time" |
||||
|
|
||||
|
"github.com/hermeznetwork/hermez-node/batchbuilder" |
||||
|
"github.com/hermeznetwork/hermez-node/common" |
||||
|
"github.com/hermeznetwork/hermez-node/db/historydb" |
||||
|
"github.com/hermeznetwork/hermez-node/db/l2db" |
||||
|
"github.com/hermeznetwork/hermez-node/eth" |
||||
|
"github.com/hermeznetwork/hermez-node/log" |
||||
|
"github.com/hermeznetwork/hermez-node/prover" |
||||
|
"github.com/hermeznetwork/hermez-node/synchronizer" |
||||
|
"github.com/hermeznetwork/hermez-node/txselector" |
||||
|
"github.com/hermeznetwork/tracerr" |
||||
|
) |
||||
|
|
||||
|
type statsVars struct { |
||||
|
Stats synchronizer.Stats |
||||
|
Vars synchronizer.SCVariablesPtr |
||||
|
} |
||||
|
|
||||
|
// Pipeline manages the forging of batches with parallel server proofs
|
||||
|
type Pipeline struct { |
||||
|
cfg Config |
||||
|
consts synchronizer.SCConsts |
||||
|
|
||||
|
// state
|
||||
|
batchNum common.BatchNum |
||||
|
lastScheduledL1BatchBlockNum int64 |
||||
|
lastForgeL1TxsNum int64 |
||||
|
started bool |
||||
|
|
||||
|
proversPool *ProversPool |
||||
|
provers []prover.Client |
||||
|
txManager *TxManager |
||||
|
historyDB *historydb.HistoryDB |
||||
|
l2DB *l2db.L2DB |
||||
|
txSelector *txselector.TxSelector |
||||
|
batchBuilder *batchbuilder.BatchBuilder |
||||
|
purger *Purger |
||||
|
|
||||
|
stats synchronizer.Stats |
||||
|
vars synchronizer.SCVariables |
||||
|
statsVarsCh chan statsVars |
||||
|
|
||||
|
ctx context.Context |
||||
|
wg sync.WaitGroup |
||||
|
cancel context.CancelFunc |
||||
|
} |
||||
|
|
||||
|
// NewPipeline creates a new Pipeline
|
||||
|
func NewPipeline(ctx context.Context, |
||||
|
cfg Config, |
||||
|
historyDB *historydb.HistoryDB, |
||||
|
l2DB *l2db.L2DB, |
||||
|
txSelector *txselector.TxSelector, |
||||
|
batchBuilder *batchbuilder.BatchBuilder, |
||||
|
purger *Purger, |
||||
|
txManager *TxManager, |
||||
|
provers []prover.Client, |
||||
|
scConsts *synchronizer.SCConsts, |
||||
|
) (*Pipeline, error) { |
||||
|
proversPool := NewProversPool(len(provers)) |
||||
|
proversPoolSize := 0 |
||||
|
for _, prover := range provers { |
||||
|
if err := prover.WaitReady(ctx); err != nil { |
||||
|
log.Errorw("prover.WaitReady", "err", err) |
||||
|
} else { |
||||
|
proversPool.Add(prover) |
||||
|
proversPoolSize++ |
||||
|
} |
||||
|
} |
||||
|
if proversPoolSize == 0 { |
||||
|
return nil, tracerr.Wrap(fmt.Errorf("no provers in the pool")) |
||||
|
} |
||||
|
return &Pipeline{ |
||||
|
cfg: cfg, |
||||
|
historyDB: historyDB, |
||||
|
l2DB: l2DB, |
||||
|
txSelector: txSelector, |
||||
|
batchBuilder: batchBuilder, |
||||
|
provers: provers, |
||||
|
proversPool: proversPool, |
||||
|
purger: purger, |
||||
|
txManager: txManager, |
||||
|
consts: *scConsts, |
||||
|
statsVarsCh: make(chan statsVars, queueLen), |
||||
|
}, nil |
||||
|
} |
||||
|
|
||||
|
// SetSyncStatsVars is a thread safe method to sets the synchronizer Stats
|
||||
|
func (p *Pipeline) SetSyncStatsVars(stats *synchronizer.Stats, vars *synchronizer.SCVariablesPtr) { |
||||
|
p.statsVarsCh <- statsVars{Stats: *stats, Vars: *vars} |
||||
|
} |
||||
|
|
||||
|
// reset pipeline state
|
||||
|
func (p *Pipeline) reset(batchNum common.BatchNum, |
||||
|
stats *synchronizer.Stats, vars *synchronizer.SCVariables) error { |
||||
|
p.batchNum = batchNum |
||||
|
p.lastForgeL1TxsNum = stats.Sync.LastForgeL1TxsNum |
||||
|
p.stats = *stats |
||||
|
p.vars = *vars |
||||
|
p.lastScheduledL1BatchBlockNum = 0 |
||||
|
|
||||
|
err := p.txSelector.Reset(p.batchNum) |
||||
|
if err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
err = p.batchBuilder.Reset(p.batchNum, true) |
||||
|
if err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (p *Pipeline) syncSCVars(vars synchronizer.SCVariablesPtr) { |
||||
|
if vars.Rollup != nil { |
||||
|
p.vars.Rollup = *vars.Rollup |
||||
|
} |
||||
|
if vars.Auction != nil { |
||||
|
p.vars.Auction = *vars.Auction |
||||
|
} |
||||
|
if vars.WDelayer != nil { |
||||
|
p.vars.WDelayer = *vars.WDelayer |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
// handleForgeBatch calls p.forgeBatch to forge the batch and get the zkInputs,
|
||||
|
// and then waits for an available proof server and sends the zkInputs to it so
|
||||
|
// that the proof computation begins.
|
||||
|
func (p *Pipeline) handleForgeBatch(ctx context.Context, batchNum common.BatchNum) (*BatchInfo, error) { |
||||
|
batchInfo, err := p.forgeBatch(batchNum) |
||||
|
if ctx.Err() != nil { |
||||
|
return nil, ctx.Err() |
||||
|
} else if err != nil { |
||||
|
if tracerr.Unwrap(err) == errLastL1BatchNotSynced { |
||||
|
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err, |
||||
|
"lastForgeL1TxsNum", p.lastForgeL1TxsNum, |
||||
|
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum) |
||||
|
} else { |
||||
|
log.Errorw("forgeBatch", "err", err) |
||||
|
} |
||||
|
return nil, err |
||||
|
} |
||||
|
// 6. Wait for an available server proof (blocking call)
|
||||
|
serverProof, err := p.proversPool.Get(ctx) |
||||
|
if ctx.Err() != nil { |
||||
|
return nil, ctx.Err() |
||||
|
} else if err != nil { |
||||
|
log.Errorw("proversPool.Get", "err", err) |
||||
|
return nil, err |
||||
|
} |
||||
|
batchInfo.ServerProof = serverProof |
||||
|
if err := p.sendServerProof(ctx, batchInfo); ctx.Err() != nil { |
||||
|
return nil, ctx.Err() |
||||
|
} else if err != nil { |
||||
|
log.Errorw("sendServerProof", "err", err) |
||||
|
batchInfo.ServerProof = nil |
||||
|
p.proversPool.Add(serverProof) |
||||
|
return nil, err |
||||
|
} |
||||
|
return batchInfo, nil |
||||
|
} |
||||
|
|
||||
|
// Start the forging pipeline
|
||||
|
func (p *Pipeline) Start(batchNum common.BatchNum, |
||||
|
stats *synchronizer.Stats, vars *synchronizer.SCVariables) error { |
||||
|
if p.started { |
||||
|
log.Fatal("Pipeline already started") |
||||
|
} |
||||
|
p.started = true |
||||
|
|
||||
|
if err := p.reset(batchNum, stats, vars); err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
p.ctx, p.cancel = context.WithCancel(context.Background()) |
||||
|
|
||||
|
queueSize := 1 |
||||
|
batchChSentServerProof := make(chan *BatchInfo, queueSize) |
||||
|
|
||||
|
p.wg.Add(1) |
||||
|
go func() { |
||||
|
waitDuration := zeroDuration |
||||
|
for { |
||||
|
select { |
||||
|
case <-p.ctx.Done(): |
||||
|
log.Info("Pipeline forgeBatch loop done") |
||||
|
p.wg.Done() |
||||
|
return |
||||
|
case statsVars := <-p.statsVarsCh: |
||||
|
p.stats = statsVars.Stats |
||||
|
p.syncSCVars(statsVars.Vars) |
||||
|
case <-time.After(waitDuration): |
||||
|
batchNum = p.batchNum + 1 |
||||
|
if batchInfo, err := p.handleForgeBatch(p.ctx, batchNum); err != nil { |
||||
|
waitDuration = p.cfg.SyncRetryInterval |
||||
|
continue |
||||
|
} else { |
||||
|
p.batchNum = batchNum |
||||
|
batchChSentServerProof <- batchInfo |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
}() |
||||
|
|
||||
|
p.wg.Add(1) |
||||
|
go func() { |
||||
|
for { |
||||
|
select { |
||||
|
case <-p.ctx.Done(): |
||||
|
log.Info("Pipeline waitServerProofSendEth loop done") |
||||
|
p.wg.Done() |
||||
|
return |
||||
|
case batchInfo := <-batchChSentServerProof: |
||||
|
err := p.waitServerProof(p.ctx, batchInfo) |
||||
|
// We are done with this serverProof, add it back to the pool
|
||||
|
p.proversPool.Add(batchInfo.ServerProof) |
||||
|
batchInfo.ServerProof = nil |
||||
|
if p.ctx.Err() != nil { |
||||
|
continue |
||||
|
} |
||||
|
if err != nil { |
||||
|
log.Errorw("waitServerProof", "err", err) |
||||
|
continue |
||||
|
} |
||||
|
p.txManager.AddBatch(batchInfo) |
||||
|
} |
||||
|
} |
||||
|
}() |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
// Stop the forging pipeline
|
||||
|
func (p *Pipeline) Stop(ctx context.Context) { |
||||
|
if !p.started { |
||||
|
log.Fatal("Pipeline already stopped") |
||||
|
} |
||||
|
p.started = false |
||||
|
log.Info("Stopping Pipeline...") |
||||
|
p.cancel() |
||||
|
p.wg.Wait() |
||||
|
for _, prover := range p.provers { |
||||
|
if err := prover.Cancel(ctx); ctx.Err() != nil { |
||||
|
continue |
||||
|
} else if err != nil { |
||||
|
log.Errorw("prover.Cancel", "err", err) |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
// sendServerProof sends the circuit inputs to the proof server
|
||||
|
func (p *Pipeline) sendServerProof(ctx context.Context, batchInfo *BatchInfo) error { |
||||
|
p.cfg.debugBatchStore(batchInfo) |
||||
|
|
||||
|
// 7. Call the selected idle server proof with BatchBuilder output,
|
||||
|
// save server proof info for batchNum
|
||||
|
if err := batchInfo.ServerProof.CalculateProof(ctx, batchInfo.ZKInputs); err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
// forgeBatch forges the batchNum batch.
|
||||
|
func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, err error) { |
||||
|
// remove transactions from the pool that have been there for too long
|
||||
|
_, err = p.purger.InvalidateMaybe(p.l2DB, p.txSelector.LocalAccountsDB(), |
||||
|
p.stats.Sync.LastBlock.Num, int64(batchNum)) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
_, err = p.purger.PurgeMaybe(p.l2DB, p.stats.Sync.LastBlock.Num, int64(batchNum)) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
|
||||
|
batchInfo = &BatchInfo{BatchNum: batchNum} // to accumulate metadata of the batch
|
||||
|
batchInfo.Debug.StartTimestamp = time.Now() |
||||
|
batchInfo.Debug.StartBlockNum = p.stats.Eth.LastBlock.Num + 1 |
||||
|
|
||||
|
selectionCfg := &txselector.SelectionConfig{ |
||||
|
MaxL1UserTxs: common.RollupConstMaxL1UserTx, |
||||
|
TxProcessorConfig: p.cfg.TxProcessorConfig, |
||||
|
} |
||||
|
|
||||
|
var poolL2Txs []common.PoolL2Tx |
||||
|
var l1UserTxsExtra, l1CoordTxs []common.L1Tx |
||||
|
var auths [][]byte |
||||
|
var coordIdxs []common.Idx |
||||
|
|
||||
|
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||
|
if p.shouldL1L2Batch(batchInfo) { |
||||
|
batchInfo.L1Batch = true |
||||
|
defer func() { |
||||
|
// If there's no error, update the parameters related
|
||||
|
// to the last L1Batch forged
|
||||
|
if err == nil { |
||||
|
p.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBlock.Num + 1 |
||||
|
p.lastForgeL1TxsNum++ |
||||
|
} |
||||
|
}() |
||||
|
if p.lastForgeL1TxsNum != p.stats.Sync.LastForgeL1TxsNum { |
||||
|
return nil, tracerr.Wrap(errLastL1BatchNotSynced) |
||||
|
} |
||||
|
// 2a: L1+L2 txs
|
||||
|
l1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(p.lastForgeL1TxsNum + 1) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
coordIdxs, auths, l1UserTxsExtra, l1CoordTxs, poolL2Txs, err = |
||||
|
p.txSelector.GetL1L2TxSelection(selectionCfg, l1UserTxs) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
} else { |
||||
|
// 2b: only L2 txs
|
||||
|
coordIdxs, auths, l1CoordTxs, poolL2Txs, err = |
||||
|
p.txSelector.GetL2TxSelection(selectionCfg) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
l1UserTxsExtra = nil |
||||
|
} |
||||
|
|
||||
|
// 3. Save metadata from TxSelector output for BatchNum
|
||||
|
batchInfo.L1UserTxsExtra = l1UserTxsExtra |
||||
|
batchInfo.L1CoordTxs = l1CoordTxs |
||||
|
batchInfo.L1CoordinatorTxsAuths = auths |
||||
|
batchInfo.CoordIdxs = coordIdxs |
||||
|
batchInfo.VerifierIdx = p.cfg.VerifierIdx |
||||
|
|
||||
|
if err := p.l2DB.StartForging(common.TxIDsFromPoolL2Txs(poolL2Txs), batchInfo.BatchNum); err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
|
||||
|
// Invalidate transactions that become invalid beause of
|
||||
|
// the poolL2Txs selected. Will mark as invalid the txs that have a
|
||||
|
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||
|
// all the nonces smaller than the current one)
|
||||
|
err = p.l2DB.InvalidateOldNonces(idxsNonceFromPoolL2Txs(poolL2Txs), batchInfo.BatchNum) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
|
||||
|
// 4. Call BatchBuilder with TxSelector output
|
||||
|
configBatch := &batchbuilder.ConfigBatch{ |
||||
|
ForgerAddress: p.cfg.ForgerAddress, |
||||
|
TxProcessorConfig: p.cfg.TxProcessorConfig, |
||||
|
} |
||||
|
zkInputs, err := p.batchBuilder.BuildBatch(coordIdxs, configBatch, l1UserTxsExtra, |
||||
|
l1CoordTxs, poolL2Txs, nil) |
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
l2Txs, err := common.PoolL2TxsToL2Txs(poolL2Txs) // NOTE: This is a big uggly, find a better way
|
||||
|
if err != nil { |
||||
|
return nil, tracerr.Wrap(err) |
||||
|
} |
||||
|
batchInfo.L2Txs = l2Txs |
||||
|
|
||||
|
// 5. Save metadata from BatchBuilder output for BatchNum
|
||||
|
batchInfo.ZKInputs = zkInputs |
||||
|
batchInfo.Debug.Status = StatusForged |
||||
|
p.cfg.debugBatchStore(batchInfo) |
||||
|
|
||||
|
return batchInfo, nil |
||||
|
} |
||||
|
|
||||
|
// waitServerProof gets the generated zkProof & sends it to the SmartContract
|
||||
|
func (p *Pipeline) waitServerProof(ctx context.Context, batchInfo *BatchInfo) error { |
||||
|
proof, pubInputs, err := batchInfo.ServerProof.GetProof(ctx) // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
|
if err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
batchInfo.Proof = proof |
||||
|
batchInfo.PublicInputs = pubInputs |
||||
|
batchInfo.ForgeBatchArgs = prepareForgeBatchArgs(batchInfo) |
||||
|
batchInfo.Debug.Status = StatusProof |
||||
|
p.cfg.debugBatchStore(batchInfo) |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (p *Pipeline) shouldL1L2Batch(batchInfo *BatchInfo) bool { |
||||
|
// Take the lastL1BatchBlockNum as the biggest between the last
|
||||
|
// scheduled one, and the synchronized one.
|
||||
|
lastL1BatchBlockNum := p.lastScheduledL1BatchBlockNum |
||||
|
if p.stats.Sync.LastL1BatchBlock > lastL1BatchBlockNum { |
||||
|
lastL1BatchBlockNum = p.stats.Sync.LastL1BatchBlock |
||||
|
} |
||||
|
// Set Debug information
|
||||
|
batchInfo.Debug.LastScheduledL1BatchBlockNum = p.lastScheduledL1BatchBlockNum |
||||
|
batchInfo.Debug.LastL1BatchBlock = p.stats.Sync.LastL1BatchBlock |
||||
|
batchInfo.Debug.LastL1BatchBlockDelta = p.stats.Eth.LastBlock.Num + 1 - lastL1BatchBlockNum |
||||
|
batchInfo.Debug.L1BatchBlockScheduleDeadline = |
||||
|
int64(float64(p.vars.Rollup.ForgeL1L2BatchTimeout-1) * p.cfg.L1BatchTimeoutPerc) |
||||
|
// Return true if we have passed the l1BatchTimeoutPerc portion of the
|
||||
|
// range before the l1batch timeout.
|
||||
|
return p.stats.Eth.LastBlock.Num+1-lastL1BatchBlockNum >= |
||||
|
int64(float64(p.vars.Rollup.ForgeL1L2BatchTimeout-1)*p.cfg.L1BatchTimeoutPerc) |
||||
|
} |
||||
|
|
||||
|
func prepareForgeBatchArgs(batchInfo *BatchInfo) *eth.RollupForgeBatchArgs { |
||||
|
proof := batchInfo.Proof |
||||
|
zki := batchInfo.ZKInputs |
||||
|
return ð.RollupForgeBatchArgs{ |
||||
|
NewLastIdx: int64(zki.Metadata.NewLastIdxRaw), |
||||
|
NewStRoot: zki.Metadata.NewStateRootRaw.BigInt(), |
||||
|
NewExitRoot: zki.Metadata.NewExitRootRaw.BigInt(), |
||||
|
L1UserTxs: batchInfo.L1UserTxsExtra, |
||||
|
L1CoordinatorTxs: batchInfo.L1CoordTxs, |
||||
|
L1CoordinatorTxsAuths: batchInfo.L1CoordinatorTxsAuths, |
||||
|
L2TxsData: batchInfo.L2Txs, |
||||
|
FeeIdxCoordinator: batchInfo.CoordIdxs, |
||||
|
// Circuit selector
|
||||
|
VerifierIdx: batchInfo.VerifierIdx, |
||||
|
L1Batch: batchInfo.L1Batch, |
||||
|
ProofA: [2]*big.Int{proof.PiA[0], proof.PiA[1]}, |
||||
|
ProofB: [2][2]*big.Int{ |
||||
|
{proof.PiB[0][0], proof.PiB[0][1]}, |
||||
|
{proof.PiB[1][0], proof.PiB[1][1]}, |
||||
|
}, |
||||
|
ProofC: [2]*big.Int{proof.PiC[0], proof.PiC[1]}, |
||||
|
} |
||||
|
} |
@ -0,0 +1,297 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"context" |
||||
|
"fmt" |
||||
|
"io/ioutil" |
||||
|
"math/big" |
||||
|
"os" |
||||
|
"testing" |
||||
|
|
||||
|
ethKeystore "github.com/ethereum/go-ethereum/accounts/keystore" |
||||
|
ethCommon "github.com/ethereum/go-ethereum/common" |
||||
|
"github.com/ethereum/go-ethereum/crypto" |
||||
|
"github.com/ethereum/go-ethereum/ethclient" |
||||
|
"github.com/hermeznetwork/hermez-node/common" |
||||
|
"github.com/hermeznetwork/hermez-node/db/historydb" |
||||
|
"github.com/hermeznetwork/hermez-node/db/statedb" |
||||
|
"github.com/hermeznetwork/hermez-node/eth" |
||||
|
"github.com/hermeznetwork/hermez-node/prover" |
||||
|
"github.com/hermeznetwork/hermez-node/synchronizer" |
||||
|
"github.com/hermeznetwork/hermez-node/test" |
||||
|
"github.com/hermeznetwork/hermez-node/test/til" |
||||
|
"github.com/iden3/go-merkletree" |
||||
|
"github.com/stretchr/testify/assert" |
||||
|
"github.com/stretchr/testify/require" |
||||
|
) |
||||
|
|
||||
|
func TestPipelineShouldL1L2Batch(t *testing.T) { |
||||
|
ethClientSetup := test.NewClientSetupExample() |
||||
|
ethClientSetup.ChainID = big.NewInt(int64(chainID)) |
||||
|
|
||||
|
var timer timer |
||||
|
ctx := context.Background() |
||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup) |
||||
|
modules := newTestModules(t) |
||||
|
var stats synchronizer.Stats |
||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup, modules) |
||||
|
pipeline, err := coord.newPipeline(ctx) |
||||
|
require.NoError(t, err) |
||||
|
pipeline.vars = coord.vars |
||||
|
|
||||
|
// Check that the parameters are the ones we expect and use in this test
|
||||
|
require.Equal(t, 0.5, pipeline.cfg.L1BatchTimeoutPerc) |
||||
|
require.Equal(t, int64(10), ethClientSetup.RollupVariables.ForgeL1L2BatchTimeout) |
||||
|
l1BatchTimeoutPerc := pipeline.cfg.L1BatchTimeoutPerc |
||||
|
l1BatchTimeout := ethClientSetup.RollupVariables.ForgeL1L2BatchTimeout |
||||
|
|
||||
|
startBlock := int64(100) |
||||
|
// Empty batchInfo to pass to shouldL1L2Batch() which sets debug information
|
||||
|
batchInfo := BatchInfo{} |
||||
|
|
||||
|
//
|
||||
|
// No scheduled L1Batch
|
||||
|
//
|
||||
|
|
||||
|
// Last L1Batch was a long time ago
|
||||
|
stats.Eth.LastBlock.Num = startBlock |
||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock |
||||
|
stats.Sync.LastL1BatchBlock = 0 |
||||
|
pipeline.stats = stats |
||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch(&batchInfo)) |
||||
|
|
||||
|
stats.Sync.LastL1BatchBlock = startBlock |
||||
|
|
||||
|
// We are are one block before the timeout range * 0.5
|
||||
|
stats.Eth.LastBlock.Num = startBlock - 1 + int64(float64(l1BatchTimeout-1)*l1BatchTimeoutPerc) - 1 |
||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock |
||||
|
pipeline.stats = stats |
||||
|
assert.Equal(t, false, pipeline.shouldL1L2Batch(&batchInfo)) |
||||
|
|
||||
|
// We are are at timeout range * 0.5
|
||||
|
stats.Eth.LastBlock.Num = startBlock - 1 + int64(float64(l1BatchTimeout-1)*l1BatchTimeoutPerc) |
||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock |
||||
|
pipeline.stats = stats |
||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch(&batchInfo)) |
||||
|
|
||||
|
//
|
||||
|
// Scheduled L1Batch
|
||||
|
//
|
||||
|
pipeline.lastScheduledL1BatchBlockNum = startBlock |
||||
|
stats.Sync.LastL1BatchBlock = startBlock - 10 |
||||
|
|
||||
|
// We are are one block before the timeout range * 0.5
|
||||
|
stats.Eth.LastBlock.Num = startBlock - 1 + int64(float64(l1BatchTimeout-1)*l1BatchTimeoutPerc) - 1 |
||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock |
||||
|
pipeline.stats = stats |
||||
|
assert.Equal(t, false, pipeline.shouldL1L2Batch(&batchInfo)) |
||||
|
|
||||
|
// We are are at timeout range * 0.5
|
||||
|
stats.Eth.LastBlock.Num = startBlock - 1 + int64(float64(l1BatchTimeout-1)*l1BatchTimeoutPerc) |
||||
|
stats.Sync.LastBlock = stats.Eth.LastBlock |
||||
|
pipeline.stats = stats |
||||
|
assert.Equal(t, true, pipeline.shouldL1L2Batch(&batchInfo)) |
||||
|
} |
||||
|
|
||||
|
const testTokensLen = 3 |
||||
|
const testUsersLen = 4 |
||||
|
|
||||
|
func preloadSync(t *testing.T, ethClient *test.Client, sync *synchronizer.Synchronizer, |
||||
|
historyDB *historydb.HistoryDB, stateDB *statedb.StateDB) *til.Context { |
||||
|
// Create a set with `testTokensLen` tokens and for each token
|
||||
|
// `testUsersLen` accounts.
|
||||
|
var set []til.Instruction |
||||
|
// set = append(set, til.Instruction{Typ: "Blockchain"})
|
||||
|
for tokenID := 1; tokenID < testTokensLen; tokenID++ { |
||||
|
set = append(set, til.Instruction{ |
||||
|
Typ: til.TypeAddToken, |
||||
|
TokenID: common.TokenID(tokenID), |
||||
|
}) |
||||
|
} |
||||
|
depositAmount, ok := new(big.Int).SetString("10225000000000000000000000000000000", 10) |
||||
|
require.True(t, ok) |
||||
|
for tokenID := 0; tokenID < testTokensLen; tokenID++ { |
||||
|
for user := 0; user < testUsersLen; user++ { |
||||
|
set = append(set, til.Instruction{ |
||||
|
Typ: common.TxTypeCreateAccountDeposit, |
||||
|
TokenID: common.TokenID(tokenID), |
||||
|
DepositAmount: depositAmount, |
||||
|
From: fmt.Sprintf("User%d", user), |
||||
|
}) |
||||
|
} |
||||
|
} |
||||
|
set = append(set, til.Instruction{Typ: til.TypeNewBatchL1}) |
||||
|
set = append(set, til.Instruction{Typ: til.TypeNewBatchL1}) |
||||
|
set = append(set, til.Instruction{Typ: til.TypeNewBlock}) |
||||
|
|
||||
|
tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx) |
||||
|
blocks, err := tc.GenerateBlocksFromInstructions(set) |
||||
|
require.NoError(t, err) |
||||
|
require.NotNil(t, blocks) |
||||
|
|
||||
|
ethAddTokens(blocks, ethClient) |
||||
|
err = ethClient.CtlAddBlocks(blocks) |
||||
|
require.NoError(t, err) |
||||
|
|
||||
|
ctx := context.Background() |
||||
|
for { |
||||
|
syncBlock, discards, err := sync.Sync2(ctx, nil) |
||||
|
require.NoError(t, err) |
||||
|
require.Nil(t, discards) |
||||
|
if syncBlock == nil { |
||||
|
break |
||||
|
} |
||||
|
} |
||||
|
dbTokens, err := historyDB.GetAllTokens() |
||||
|
require.Nil(t, err) |
||||
|
require.Equal(t, testTokensLen, len(dbTokens)) |
||||
|
|
||||
|
dbAccounts, err := historyDB.GetAllAccounts() |
||||
|
require.Nil(t, err) |
||||
|
require.Equal(t, testTokensLen*testUsersLen, len(dbAccounts)) |
||||
|
|
||||
|
sdbAccounts, err := stateDB.GetAccounts() |
||||
|
require.Nil(t, err) |
||||
|
require.Equal(t, testTokensLen*testUsersLen, len(sdbAccounts)) |
||||
|
|
||||
|
return tc |
||||
|
} |
||||
|
|
||||
|
func TestPipelineForgeBatchWithTxs(t *testing.T) { |
||||
|
ethClientSetup := test.NewClientSetupExample() |
||||
|
ethClientSetup.ChainID = big.NewInt(int64(chainID)) |
||||
|
|
||||
|
var timer timer |
||||
|
ctx := context.Background() |
||||
|
ethClient := test.NewClient(true, &timer, &bidder, ethClientSetup) |
||||
|
modules := newTestModules(t) |
||||
|
coord := newTestCoordinator(t, forger, ethClient, ethClientSetup, modules) |
||||
|
sync := newTestSynchronizer(t, ethClient, ethClientSetup, modules) |
||||
|
|
||||
|
// preload the synchronier (via the test ethClient) some tokens and
|
||||
|
// users with positive balances
|
||||
|
tilCtx := preloadSync(t, ethClient, sync, modules.historyDB, modules.stateDB) |
||||
|
syncStats := sync.Stats() |
||||
|
batchNum := common.BatchNum(syncStats.Sync.LastBatch) |
||||
|
syncSCVars := sync.SCVars() |
||||
|
|
||||
|
pipeline, err := coord.newPipeline(ctx) |
||||
|
require.NoError(t, err) |
||||
|
|
||||
|
// Insert some l2txs in the Pool
|
||||
|
setPool := ` |
||||
|
Type: PoolL2 |
||||
|
|
||||
|
PoolTransfer(0) User0-User1: 100 (126) |
||||
|
PoolTransfer(0) User1-User2: 200 (126) |
||||
|
PoolTransfer(0) User2-User3: 300 (126) |
||||
|
` |
||||
|
l2txs, err := tilCtx.GeneratePoolL2Txs(setPool) |
||||
|
require.NoError(t, err) |
||||
|
for _, tx := range l2txs { |
||||
|
err := modules.l2DB.AddTxTest(&tx) //nolint:gosec
|
||||
|
require.NoError(t, err) |
||||
|
} |
||||
|
|
||||
|
err = pipeline.reset(batchNum, syncStats, &synchronizer.SCVariables{ |
||||
|
Rollup: *syncSCVars.Rollup, |
||||
|
Auction: *syncSCVars.Auction, |
||||
|
WDelayer: *syncSCVars.WDelayer, |
||||
|
}) |
||||
|
require.NoError(t, err) |
||||
|
// Sanity check
|
||||
|
sdbAccounts, err := pipeline.txSelector.LocalAccountsDB().GetAccounts() |
||||
|
require.Nil(t, err) |
||||
|
require.Equal(t, testTokensLen*testUsersLen, len(sdbAccounts)) |
||||
|
|
||||
|
// Sanity check
|
||||
|
sdbAccounts, err = pipeline.batchBuilder.LocalStateDB().GetAccounts() |
||||
|
require.Nil(t, err) |
||||
|
require.Equal(t, testTokensLen*testUsersLen, len(sdbAccounts)) |
||||
|
|
||||
|
// Sanity check
|
||||
|
require.Equal(t, modules.stateDB.MT.Root(), |
||||
|
pipeline.batchBuilder.LocalStateDB().MT.Root()) |
||||
|
|
||||
|
batchNum++ |
||||
|
|
||||
|
batchInfo, err := pipeline.forgeBatch(batchNum) |
||||
|
require.NoError(t, err) |
||||
|
assert.Equal(t, 3, len(batchInfo.L2Txs)) |
||||
|
|
||||
|
batchNum++ |
||||
|
batchInfo, err = pipeline.forgeBatch(batchNum) |
||||
|
require.NoError(t, err) |
||||
|
assert.Equal(t, 0, len(batchInfo.L2Txs)) |
||||
|
} |
||||
|
|
||||
|
func TestEthRollupForgeBatch(t *testing.T) { |
||||
|
if os.Getenv("TEST_ROLLUP_FORGE_BATCH") == "" { |
||||
|
return |
||||
|
} |
||||
|
const web3URL = "http://localhost:8545" |
||||
|
const password = "test" |
||||
|
addr := ethCommon.HexToAddress("0xb4124ceb3451635dacedd11767f004d8a28c6ee7") |
||||
|
sk, err := crypto.HexToECDSA( |
||||
|
"a8a54b2d8197bc0b19bb8a084031be71835580a01e70a45a13babd16c9bc1563") |
||||
|
require.NoError(t, err) |
||||
|
rollupAddr := ethCommon.HexToAddress("0x8EEaea23686c319133a7cC110b840d1591d9AeE0") |
||||
|
pathKeystore, err := ioutil.TempDir("", "tmpKeystore") |
||||
|
require.NoError(t, err) |
||||
|
deleteme = append(deleteme, pathKeystore) |
||||
|
ctx := context.Background() |
||||
|
batchInfo := &BatchInfo{} |
||||
|
proofClient := &prover.MockClient{} |
||||
|
chainID := uint16(0) |
||||
|
|
||||
|
ethClient, err := ethclient.Dial(web3URL) |
||||
|
require.NoError(t, err) |
||||
|
ethCfg := eth.EthereumConfig{ |
||||
|
CallGasLimit: 300000, |
||||
|
GasPriceDiv: 100, |
||||
|
} |
||||
|
scryptN := ethKeystore.LightScryptN |
||||
|
scryptP := ethKeystore.LightScryptP |
||||
|
keyStore := ethKeystore.NewKeyStore(pathKeystore, |
||||
|
scryptN, scryptP) |
||||
|
account, err := keyStore.ImportECDSA(sk, password) |
||||
|
require.NoError(t, err) |
||||
|
require.Equal(t, account.Address, addr) |
||||
|
err = keyStore.Unlock(account, password) |
||||
|
require.NoError(t, err) |
||||
|
|
||||
|
client, err := eth.NewClient(ethClient, &account, keyStore, ð.ClientConfig{ |
||||
|
Ethereum: ethCfg, |
||||
|
Rollup: eth.RollupConfig{ |
||||
|
Address: rollupAddr, |
||||
|
}, |
||||
|
Auction: eth.AuctionConfig{ |
||||
|
Address: ethCommon.Address{}, |
||||
|
TokenHEZ: eth.TokenConfig{ |
||||
|
Address: ethCommon.Address{}, |
||||
|
Name: "HEZ", |
||||
|
}, |
||||
|
}, |
||||
|
WDelayer: eth.WDelayerConfig{ |
||||
|
Address: ethCommon.Address{}, |
||||
|
}, |
||||
|
}) |
||||
|
require.NoError(t, err) |
||||
|
|
||||
|
zkInputs := common.NewZKInputs(chainID, 100, 24, 512, 32, big.NewInt(1)) |
||||
|
zkInputs.Metadata.NewStateRootRaw = &merkletree.Hash{1} |
||||
|
zkInputs.Metadata.NewExitRootRaw = &merkletree.Hash{2} |
||||
|
batchInfo.ZKInputs = zkInputs |
||||
|
err = proofClient.CalculateProof(ctx, batchInfo.ZKInputs) |
||||
|
require.NoError(t, err) |
||||
|
|
||||
|
proof, pubInputs, err := proofClient.GetProof(ctx) |
||||
|
require.NoError(t, err) |
||||
|
batchInfo.Proof = proof |
||||
|
batchInfo.PublicInputs = pubInputs |
||||
|
|
||||
|
batchInfo.ForgeBatchArgs = prepareForgeBatchArgs(batchInfo) |
||||
|
_, err = client.RollupForgeBatch(batchInfo.ForgeBatchArgs) |
||||
|
require.NoError(t, err) |
||||
|
batchInfo.Proof = proof |
||||
|
} |
@ -0,0 +1,207 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"context" |
||||
|
"fmt" |
||||
|
"strings" |
||||
|
"time" |
||||
|
|
||||
|
"github.com/ethereum/go-ethereum/core/types" |
||||
|
"github.com/hermeznetwork/hermez-node/common" |
||||
|
"github.com/hermeznetwork/hermez-node/db/l2db" |
||||
|
"github.com/hermeznetwork/hermez-node/eth" |
||||
|
"github.com/hermeznetwork/hermez-node/log" |
||||
|
"github.com/hermeznetwork/tracerr" |
||||
|
) |
||||
|
|
||||
|
// TxManager handles everything related to ethereum transactions: It makes the
|
||||
|
// call to forge, waits for transaction confirmation, and keeps checking them
|
||||
|
// until a number of confirmed blocks have passed.
|
||||
|
type TxManager struct { |
||||
|
cfg Config |
||||
|
ethClient eth.ClientInterface |
||||
|
l2DB *l2db.L2DB // Used only to mark forged txs as forged in the L2DB
|
||||
|
coord *Coordinator // Used only to send messages to stop the pipeline
|
||||
|
batchCh chan *BatchInfo |
||||
|
lastBlockCh chan int64 |
||||
|
queue []*BatchInfo |
||||
|
lastBlock int64 |
||||
|
// lastConfirmedBatch stores the last BatchNum that who's forge call was confirmed
|
||||
|
lastConfirmedBatch common.BatchNum |
||||
|
} |
||||
|
|
||||
|
// NewTxManager creates a new TxManager
|
||||
|
func NewTxManager(cfg *Config, ethClient eth.ClientInterface, l2DB *l2db.L2DB, |
||||
|
coord *Coordinator) *TxManager { |
||||
|
return &TxManager{ |
||||
|
cfg: *cfg, |
||||
|
ethClient: ethClient, |
||||
|
l2DB: l2DB, |
||||
|
coord: coord, |
||||
|
batchCh: make(chan *BatchInfo, queueLen), |
||||
|
lastBlockCh: make(chan int64, queueLen), |
||||
|
lastBlock: -1, |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
// AddBatch is a thread safe method to pass a new batch TxManager to be sent to
|
||||
|
// the smart contract via the forge call
|
||||
|
func (t *TxManager) AddBatch(batchInfo *BatchInfo) { |
||||
|
t.batchCh <- batchInfo |
||||
|
} |
||||
|
|
||||
|
// SetLastBlock is a thread safe method to pass the lastBlock to the TxManager
|
||||
|
func (t *TxManager) SetLastBlock(lastBlock int64) { |
||||
|
t.lastBlockCh <- lastBlock |
||||
|
} |
||||
|
|
||||
|
func (t *TxManager) callRollupForgeBatch(ctx context.Context, batchInfo *BatchInfo) error { |
||||
|
batchInfo.Debug.Status = StatusSent |
||||
|
batchInfo.Debug.SendBlockNum = t.lastBlock + 1 |
||||
|
batchInfo.Debug.SendTimestamp = time.Now() |
||||
|
batchInfo.Debug.StartToSendDelay = batchInfo.Debug.SendTimestamp.Sub( |
||||
|
batchInfo.Debug.StartTimestamp).Seconds() |
||||
|
var ethTx *types.Transaction |
||||
|
var err error |
||||
|
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ { |
||||
|
ethTx, err = t.ethClient.RollupForgeBatch(batchInfo.ForgeBatchArgs) |
||||
|
if err != nil { |
||||
|
if strings.Contains(err.Error(), common.AuctionErrMsgCannotForge) { |
||||
|
log.Debugw("TxManager ethClient.RollupForgeBatch", "err", err, |
||||
|
"block", t.lastBlock+1) |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
log.Errorw("TxManager ethClient.RollupForgeBatch", |
||||
|
"attempt", attempt, "err", err, "block", t.lastBlock+1, |
||||
|
"batchNum", batchInfo.BatchNum) |
||||
|
} else { |
||||
|
break |
||||
|
} |
||||
|
select { |
||||
|
case <-ctx.Done(): |
||||
|
return tracerr.Wrap(common.ErrDone) |
||||
|
case <-time.After(t.cfg.EthClientAttemptsDelay): |
||||
|
} |
||||
|
} |
||||
|
if err != nil { |
||||
|
return tracerr.Wrap(fmt.Errorf("reached max attempts for ethClient.RollupForgeBatch: %w", err)) |
||||
|
} |
||||
|
batchInfo.EthTx = ethTx |
||||
|
log.Infow("TxManager ethClient.RollupForgeBatch", "batch", batchInfo.BatchNum, "tx", ethTx.Hash().Hex()) |
||||
|
t.cfg.debugBatchStore(batchInfo) |
||||
|
if err := t.l2DB.DoneForging(common.TxIDsFromL2Txs(batchInfo.L2Txs), batchInfo.BatchNum); err != nil { |
||||
|
return tracerr.Wrap(err) |
||||
|
} |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (t *TxManager) checkEthTransactionReceipt(ctx context.Context, batchInfo *BatchInfo) error { |
||||
|
txHash := batchInfo.EthTx.Hash() |
||||
|
var receipt *types.Receipt |
||||
|
var err error |
||||
|
for attempt := 0; attempt < t.cfg.EthClientAttempts; attempt++ { |
||||
|
receipt, err = t.ethClient.EthTransactionReceipt(ctx, txHash) |
||||
|
if ctx.Err() != nil { |
||||
|
continue |
||||
|
} |
||||
|
if err != nil { |
||||
|
log.Errorw("TxManager ethClient.EthTransactionReceipt", |
||||
|
"attempt", attempt, "err", err) |
||||
|
} else { |
||||
|
break |
||||
|
} |
||||
|
select { |
||||
|
case <-ctx.Done(): |
||||
|
return tracerr.Wrap(common.ErrDone) |
||||
|
case <-time.After(t.cfg.EthClientAttemptsDelay): |
||||
|
} |
||||
|
} |
||||
|
if err != nil { |
||||
|
return tracerr.Wrap(fmt.Errorf("reached max attempts for ethClient.EthTransactionReceipt: %w", err)) |
||||
|
} |
||||
|
batchInfo.Receipt = receipt |
||||
|
t.cfg.debugBatchStore(batchInfo) |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (t *TxManager) handleReceipt(batchInfo *BatchInfo) (*int64, error) { |
||||
|
receipt := batchInfo.Receipt |
||||
|
if receipt != nil { |
||||
|
if receipt.Status == types.ReceiptStatusFailed { |
||||
|
batchInfo.Debug.Status = StatusFailed |
||||
|
t.cfg.debugBatchStore(batchInfo) |
||||
|
log.Errorw("TxManager receipt status is failed", "receipt", receipt) |
||||
|
return nil, tracerr.Wrap(fmt.Errorf("ethereum transaction receipt statis is failed")) |
||||
|
} else if receipt.Status == types.ReceiptStatusSuccessful { |
||||
|
batchInfo.Debug.Status = StatusMined |
||||
|
batchInfo.Debug.MineBlockNum = receipt.BlockNumber.Int64() |
||||
|
batchInfo.Debug.StartToMineBlocksDelay = batchInfo.Debug.MineBlockNum - |
||||
|
batchInfo.Debug.StartBlockNum |
||||
|
t.cfg.debugBatchStore(batchInfo) |
||||
|
if batchInfo.BatchNum > t.lastConfirmedBatch { |
||||
|
t.lastConfirmedBatch = batchInfo.BatchNum |
||||
|
} |
||||
|
confirm := t.lastBlock - receipt.BlockNumber.Int64() |
||||
|
return &confirm, nil |
||||
|
} |
||||
|
} |
||||
|
return nil, nil |
||||
|
} |
||||
|
|
||||
|
// Run the TxManager
|
||||
|
func (t *TxManager) Run(ctx context.Context) { |
||||
|
next := 0 |
||||
|
waitDuration := longWaitDuration |
||||
|
|
||||
|
for { |
||||
|
select { |
||||
|
case <-ctx.Done(): |
||||
|
log.Info("TxManager done") |
||||
|
return |
||||
|
case lastBlock := <-t.lastBlockCh: |
||||
|
t.lastBlock = lastBlock |
||||
|
case batchInfo := <-t.batchCh: |
||||
|
if err := t.callRollupForgeBatch(ctx, batchInfo); ctx.Err() != nil { |
||||
|
continue |
||||
|
} else if err != nil { |
||||
|
t.coord.SendMsg(MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch call: %v", err)}) |
||||
|
continue |
||||
|
} |
||||
|
log.Debugf("ethClient ForgeCall sent, batchNum: %d", batchInfo.BatchNum) |
||||
|
t.queue = append(t.queue, batchInfo) |
||||
|
waitDuration = t.cfg.TxManagerCheckInterval |
||||
|
case <-time.After(waitDuration): |
||||
|
if len(t.queue) == 0 { |
||||
|
continue |
||||
|
} |
||||
|
current := next |
||||
|
next = (current + 1) % len(t.queue) |
||||
|
batchInfo := t.queue[current] |
||||
|
if err := t.checkEthTransactionReceipt(ctx, batchInfo); ctx.Err() != nil { |
||||
|
continue |
||||
|
} else if err != nil { //nolint:staticcheck
|
||||
|
// We can't get the receipt for the
|
||||
|
// transaction, so we can't confirm if it was
|
||||
|
// mined
|
||||
|
t.coord.SendMsg(MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch receipt: %v", err)}) |
||||
|
} |
||||
|
|
||||
|
confirm, err := t.handleReceipt(batchInfo) |
||||
|
if err != nil { //nolint:staticcheck
|
||||
|
// Transaction was rejected
|
||||
|
t.coord.SendMsg(MsgStopPipeline{Reason: fmt.Sprintf("forgeBatch reject: %v", err)}) |
||||
|
} |
||||
|
if confirm != nil && *confirm >= t.cfg.ConfirmBlocks { |
||||
|
log.Debugw("TxManager tx for RollupForgeBatch confirmed", |
||||
|
"batch", batchInfo.BatchNum) |
||||
|
t.queue = append(t.queue[:current], t.queue[current+1:]...) |
||||
|
if len(t.queue) == 0 { |
||||
|
waitDuration = longWaitDuration |
||||
|
next = 0 |
||||
|
} else { |
||||
|
next = current % len(t.queue) |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
} |
||||
|
} |