|
|
package coordinator
import ( "time"
ethCommon "github.com/ethereum/go-ethereum/common" "github.com/hermeznetwork/hermez-node/batchbuilder" "github.com/hermeznetwork/hermez-node/common" "github.com/hermeznetwork/hermez-node/eth" "github.com/hermeznetwork/hermez-node/txselector" kvdb "github.com/iden3/go-merkletree/db" "github.com/iden3/go-merkletree/db/memory" )
// CoordinatorConfig contains the Coordinator configuration
type CoordinatorConfig struct { ForgerAddress ethCommon.Address }
// Coordinator implements the Coordinator type
type Coordinator struct { config CoordinatorConfig
batchNum uint64 batchQueue *BatchQueue serverProofPool ServerProofPool
// synchronizer *synchronizer.Synchronizer
txsel *txselector.TxSelector batchBuilder *batchbuilder.BatchBuilder
ethClient *eth.EthClient ethTxStore kvdb.Storage }
// NewCoordinator creates a new Coordinator
func NewCoordinator() *Coordinator { // once synchronizer is ready, synchronizer.Synchronizer will be passed as parameter here
var c *Coordinator c.ethClient = eth.NewEthClient() // TBD
c.ethTxStore = memory.NewMemoryStorage() return c }
// Start starts the Coordinator service
func (c *Coordinator) Start() { // TODO TBD note: the sequences & loops & errors & logging & goroutines
// & channels approach still needs to be defined, the current code is a
// wip draft
// TBD: goroutines strategy
// if in Forge Sequence:
if c.isForgeSequence() { // c.batchNum = c.synchronizer.LastBatchNum()
_ = c.txsel.Reset(c.batchNum) _ = c.batchBuilder.Reset(c.batchNum, true) c.batchQueue = NewBatchQueue() go func() { for { _ = c.forgeSequence() time.Sleep(1 * time.Second) } }() go func() { for { _ = c.proveSequence() time.Sleep(1 * time.Second) } }() go func() { for { _ = c.forgeConfirmationSequence() time.Sleep(1 * time.Second) } }() } }
func (c *Coordinator) forgeSequence() error { // TODO once synchronizer has this method ready:
// If there's been a reorg, handle it
// handleReorg() function decides if the reorg must restart the pipeline or not
// if c.synchronizer.Reorg():
_ = c.handleReorg()
// 0. If there's an available server proof: Start pipeline for batchNum = batchNum + 1
serverProofInfo, err := c.serverProofPool.GetNextAvailable() // blocking call, returns when a server proof is available
if err != nil { return err }
// remove transactions from the pool that have been there for too long
err = c.purgeRemoveByTimeout() if err != nil { return err }
c.batchNum = c.batchNum + 1 batchInfo := NewBatchInfo(c.batchNum, serverProofInfo) // to accumulate metadata of the batch
var l2Txs []*common.PoolL2Tx // var feesInfo
var l1UserTxsExtra, l1OperatorTxs []*common.L1Tx // 1. Decide if we forge L2Tx or L1+L2Tx
if c.shouldL1L2Batch() { // 2a: L1+L2 txs
// l1UserTxs, toForgeL1TxsNumber := c.synchronizer.GetNextL1UserTxs() // TODO once synchronizer is ready, uncomment
var l1UserTxs []*common.L1Tx = nil // tmp, depends on synchronizer
l1UserTxsExtra, l1OperatorTxs, l2Txs, err = c.txsel.GetL1L2TxSelection(c.batchNum, l1UserTxs) // TODO once feesInfo is added to method return, add the var
if err != nil { return err } } else { // 2b: only L2 txs
l2Txs, err = c.txsel.GetL2TxSelection(c.batchNum) // TODO once feesInfo is added to method return, add the var
if err != nil { return err } l1UserTxsExtra = nil l1OperatorTxs = nil }
// Run purger to invalidate transactions that become invalid beause of
// the l2Txs selected. Will mark as invalid the txs that have a
// (fromIdx, nonce) which already appears in the selected txs (includes
// all the nonces smaller than the current one)
err = c.purgeInvalidDueToL2TxsSelection(l2Txs) if err != nil { return err }
// 3. Save metadata from TxSelector output for BatchNum
batchInfo.SetTxsInfo(l1UserTxsExtra, l1OperatorTxs, l2Txs) // TODO feesInfo
// 4. Call BatchBuilder with TxSelector output
configBatch := batchbuilder.ConfigBatch{ ForgerAddress: c.config.ForgerAddress, } zkInputs, err := c.batchBuilder.BuildBatch(configBatch, l1UserTxsExtra, l1OperatorTxs, l2Txs, nil) // TODO []common.TokenID --> feesInfo
if err != nil { return err }
// 5. Save metadata from BatchBuilder output for BatchNum
batchInfo.SetZKInputs(zkInputs)
// 6. Call an idle server proof with BatchBuilder output, save server proof info for batchNum
err = batchInfo.serverProof.CalculateProof(zkInputs) if err != nil { return err } c.batchQueue.Push(&batchInfo)
return nil }
// proveSequence gets the generated zkProof & sends it to the SmartContract
func (c *Coordinator) proveSequence() error { batchInfo := c.batchQueue.Pop() if batchInfo == nil { // no batches in queue, return
return common.ErrBatchQueueEmpty } serverProofInfo := batchInfo.serverProof proof, err := serverProofInfo.GetProof() // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
if err != nil { return err } batchInfo.SetProof(proof) callData := c.prepareCallDataForge(batchInfo) _, err = c.ethClient.ForgeCall(callData) if err != nil { return err } // TODO once tx data type is defined, store ethTx (returned by ForgeCall)
// TBD if use ethTxStore as a disk k-v database, or use a Queue
// tx, err := c.ethTxStore.NewTx()
// if err != nil {
// return err
// }
// tx.Put(ethTx.Hash(), ethTx.Bytes())
// if err := tx.Commit(); err!=nil {
// return nil
// }
return nil }
func (c *Coordinator) forgeConfirmationSequence() error { // TODO strategy of this sequence TBD
// confirm eth txs and mark them as accepted sequence
// ethTx := ethTxStore.GetFirstPending()
// waitForAccepted(ethTx) // blocking call, returns once the ethTx is mined
// ethTxStore.MarkAccepted(ethTx)
return nil }
func (c *Coordinator) handleReorg() error { return nil }
// isForgeSequence returns true if the node is the Forger in the current ethereum block
func (c *Coordinator) isForgeSequence() bool {
return false }
func (c *Coordinator) purgeRemoveByTimeout() error {
return nil }
func (c *Coordinator) purgeInvalidDueToL2TxsSelection(l2Txs []*common.PoolL2Tx) error {
return nil }
func (c *Coordinator) shouldL1L2Batch() bool {
return false }
func (c *Coordinator) prepareCallDataForge(batchInfo *BatchInfo) *common.CallDataForge { return nil }
|