mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 11:26:44 +01:00
Add initial structure Coordinator Forge Sequence
This commit is contained in:
225
coordinator/coordinator.go
Normal file
225
coordinator/coordinator.go
Normal file
@@ -0,0 +1,225 @@
|
||||
package coordinator
|
||||
|
||||
import (
|
||||
"time"
|
||||
|
||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
||||
"github.com/hermeznetwork/hermez-node/common"
|
||||
"github.com/hermeznetwork/hermez-node/eth"
|
||||
"github.com/hermeznetwork/hermez-node/txselector"
|
||||
kvdb "github.com/iden3/go-merkletree/db"
|
||||
"github.com/iden3/go-merkletree/db/memory"
|
||||
)
|
||||
|
||||
// CoordinatorConfig contains the Coordinator configuration
|
||||
type CoordinatorConfig struct {
|
||||
ForgerAddress ethCommon.Address
|
||||
}
|
||||
|
||||
// Coordinator implements the Coordinator type
|
||||
type Coordinator struct {
|
||||
config CoordinatorConfig
|
||||
|
||||
batchNum uint64
|
||||
batchQueue *BatchQueue
|
||||
serverProofPool ServerProofPool
|
||||
|
||||
// synchronizer *synchronizer.Synchronizer
|
||||
txsel *txselector.TxSelector
|
||||
batchBuilder *batchbuilder.BatchBuilder
|
||||
|
||||
ethClient *eth.EthClient
|
||||
ethTxStore kvdb.Storage
|
||||
}
|
||||
|
||||
// NewCoordinator creates a new Coordinator
|
||||
func NewCoordinator() *Coordinator { // once synchronizer is ready, synchronizer.Synchronizer will be passed as parameter here
|
||||
var c *Coordinator
|
||||
c.ethClient = eth.NewEthClient() // TBD
|
||||
c.ethTxStore = memory.NewMemoryStorage()
|
||||
return c
|
||||
}
|
||||
|
||||
// Start starts the Coordinator service
|
||||
func (c *Coordinator) Start() {
|
||||
// TODO TBD note: the sequences & loops & errors & logging & goroutines
|
||||
// & channels approach still needs to be defined, the current code is a
|
||||
// wip draft
|
||||
|
||||
// TBD: goroutines strategy
|
||||
|
||||
// if in Forge Sequence:
|
||||
if c.isForgeSequence() {
|
||||
// c.batchNum = c.synchronizer.LastBatchNum()
|
||||
_ = c.txsel.Reset(c.batchNum)
|
||||
_ = c.batchBuilder.Reset(c.batchNum, true)
|
||||
c.batchQueue = NewBatchQueue()
|
||||
go func() {
|
||||
for {
|
||||
_ = c.forgeSequence()
|
||||
time.Sleep(1 * time.Second)
|
||||
}
|
||||
}()
|
||||
go func() {
|
||||
for {
|
||||
_ = c.proveSequence()
|
||||
time.Sleep(1 * time.Second)
|
||||
}
|
||||
}()
|
||||
go func() {
|
||||
for {
|
||||
_ = c.forgeConfirmationSequence()
|
||||
time.Sleep(1 * time.Second)
|
||||
}
|
||||
}()
|
||||
}
|
||||
}
|
||||
|
||||
func (c *Coordinator) forgeSequence() error {
|
||||
// TODO once synchronizer has this method ready:
|
||||
// If there's been a reorg, handle it
|
||||
// handleReorg() function decides if the reorg must restart the pipeline or not
|
||||
// if c.synchronizer.Reorg():
|
||||
_ = c.handleReorg()
|
||||
|
||||
// 0. If there's an available server proof: Start pipeline for batchNum = batchNum + 1
|
||||
serverProofInfo, err := c.serverProofPool.GetNextAvailable() // blocking call, returns when a server proof is available
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// remove transactions from the pool that have been there for too long
|
||||
err = c.purgeRemoveByTimeout()
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
c.batchNum = c.batchNum + 1
|
||||
batchInfo := NewBatchInfo(c.batchNum, serverProofInfo) // to accumulate metadata of the batch
|
||||
|
||||
var l2Txs []common.PoolL2Tx
|
||||
// var feesInfo
|
||||
var l1UserTxsExtra, l1OperatorTxs []common.L1Tx
|
||||
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||
if c.shouldL1L2Batch() {
|
||||
// 2a: L1+L2 txs
|
||||
// l1UserTxs, toForgeL1TxsNumber := c.synchronizer.GetNextL1UserTxs() // TODO once synchronizer is ready, uncomment
|
||||
var l1UserTxs []common.L1Tx = nil // tmp, depends on synchronizer
|
||||
l1UserTxsExtra, l1OperatorTxs, l2Txs, err = c.txsel.GetL1L2TxSelection(c.batchNum, l1UserTxs) // TODO once feesInfo is added to method return, add the var
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
} else {
|
||||
// 2b: only L2 txs
|
||||
l2Txs, err = c.txsel.GetL2TxSelection(c.batchNum) // TODO once feesInfo is added to method return, add the var
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
l1UserTxsExtra = nil
|
||||
l1OperatorTxs = nil
|
||||
}
|
||||
|
||||
// Run purger to invalidate transactions that become invalid beause of
|
||||
// the l2Txs selected. Will mark as invalid the txs that have a
|
||||
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||
// all the nonces smaller than the current one)
|
||||
err = c.purgeInvalidDueToL2TxsSelection(l2Txs)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// 3. Save metadata from TxSelector output for BatchNum
|
||||
batchInfo.SetTxsInfo(l1UserTxsExtra, l1OperatorTxs, l2Txs) // TODO feesInfo
|
||||
|
||||
// 4. Call BatchBuilder with TxSelector output
|
||||
configBatch := batchbuilder.ConfigBatch{
|
||||
ForgerAddress: c.config.ForgerAddress,
|
||||
}
|
||||
zkInputs, err := c.batchBuilder.BuildBatch(configBatch, l1UserTxsExtra, l1OperatorTxs, l2Txs, nil) // TODO []common.TokenID --> feesInfo
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
|
||||
// 5. Save metadata from BatchBuilder output for BatchNum
|
||||
batchInfo.SetZKInputs(zkInputs)
|
||||
|
||||
// 6. Call an idle server proof with BatchBuilder output, save server proof info for batchNum
|
||||
err = batchInfo.serverProof.CalculateProof(zkInputs)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
c.batchQueue.Push(&batchInfo)
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
// proveSequence gets the generated zkProof & sends it to the SmartContract
|
||||
func (c *Coordinator) proveSequence() error {
|
||||
batchInfo := c.batchQueue.Pop()
|
||||
if batchInfo == nil {
|
||||
// no batches in queue, return
|
||||
return common.ErrBatchQueueEmpty
|
||||
}
|
||||
serverProofInfo := batchInfo.serverProof
|
||||
proof, err := serverProofInfo.GetProof() // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
batchInfo.SetProof(proof)
|
||||
callData := c.prepareCallDataForge(batchInfo)
|
||||
_, err = c.ethClient.ForgeCall(callData)
|
||||
if err != nil {
|
||||
return err
|
||||
}
|
||||
// TODO once tx data type is defined, store ethTx (returned by ForgeCall)
|
||||
// TBD if use ethTxStore as a disk k-v database, or use a Queue
|
||||
// tx, err := c.ethTxStore.NewTx()
|
||||
// if err != nil {
|
||||
// return err
|
||||
// }
|
||||
// tx.Put(ethTx.Hash(), ethTx.Bytes())
|
||||
// if err := tx.Commit(); err!=nil {
|
||||
// return nil
|
||||
// }
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *Coordinator) forgeConfirmationSequence() error {
|
||||
// TODO strategy of this sequence TBD
|
||||
// confirm eth txs and mark them as accepted sequence
|
||||
// ethTx := ethTxStore.GetFirstPending()
|
||||
// waitForAccepted(ethTx) // blocking call, returns once the ethTx is mined
|
||||
// ethTxStore.MarkAccepted(ethTx)
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *Coordinator) handleReorg() error {
|
||||
return nil
|
||||
}
|
||||
|
||||
// isForgeSequence returns true if the node is the Forger in the current ethereum block
|
||||
func (c *Coordinator) isForgeSequence() bool {
|
||||
|
||||
return false
|
||||
}
|
||||
|
||||
func (c *Coordinator) purgeRemoveByTimeout() error {
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *Coordinator) purgeInvalidDueToL2TxsSelection(l2Txs []common.PoolL2Tx) error {
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func (c *Coordinator) shouldL1L2Batch() bool {
|
||||
|
||||
return false
|
||||
}
|
||||
|
||||
func (c *Coordinator) prepareCallDataForge(batchInfo *BatchInfo) *common.CallDataForge {
|
||||
return nil
|
||||
}
|
||||
Reference in New Issue
Block a user