Initial Coordinator (OpManager) structurefeature/sql-semaphore1
@ -0,0 +1,80 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"github.com/hermeznetwork/hermez-node/common" |
||||
|
) |
||||
|
|
||||
|
type Proof struct { |
||||
|
// TBD this type will be got from the proof server
|
||||
|
} |
||||
|
|
||||
|
// BatchInfo contans the Batch information
|
||||
|
type BatchInfo struct { |
||||
|
batchNum uint64 |
||||
|
serverProof *ServerProofInfo |
||||
|
zkInputs *common.ZKInputs |
||||
|
proof *Proof |
||||
|
L1UserTxsExtra []common.L1Tx |
||||
|
L1OperatorTxs []common.L1Tx |
||||
|
L2Txs []common.PoolL2Tx |
||||
|
// FeesInfo
|
||||
|
} |
||||
|
|
||||
|
// NewBatchInfo creates a new BatchInfo with the given batchNum &
|
||||
|
// ServerProofInfo
|
||||
|
func NewBatchInfo(batchNum uint64, serverProof *ServerProofInfo) BatchInfo { |
||||
|
return BatchInfo{ |
||||
|
batchNum: batchNum, |
||||
|
serverProof: serverProof, |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
// SetTxsInfo sets the l1UserTxs, l1OperatorTxs and l2Txs to the BatchInfo data
|
||||
|
// structure
|
||||
|
func (bi *BatchInfo) SetTxsInfo(l1UserTxsExtra, l1OperatorTxs []common.L1Tx, l2Txs []common.PoolL2Tx) { |
||||
|
// TBD parameter: feesInfo
|
||||
|
bi.L1UserTxsExtra = l1UserTxsExtra |
||||
|
bi.L1OperatorTxs = l1OperatorTxs |
||||
|
bi.L2Txs = l2Txs |
||||
|
} |
||||
|
|
||||
|
// SetZKInputs sets the ZKInputs to the BatchInfo data structure
|
||||
|
func (bi *BatchInfo) SetZKInputs(zkInputs *common.ZKInputs) { |
||||
|
bi.zkInputs = zkInputs |
||||
|
} |
||||
|
|
||||
|
// SetServerProof sets the ServerProofInfo to the BatchInfo data structure
|
||||
|
func (bi *BatchInfo) SetServerProof(serverProof *ServerProofInfo) { |
||||
|
bi.serverProof = serverProof |
||||
|
} |
||||
|
|
||||
|
// SetProof sets the Proof to the BatchInfo data structure
|
||||
|
func (bi *BatchInfo) SetProof(proof *Proof) { |
||||
|
bi.proof = proof |
||||
|
} |
||||
|
|
||||
|
// BatchQueue implements a FIFO queue of BatchInfo
|
||||
|
type BatchQueue struct { |
||||
|
queue []*BatchInfo |
||||
|
} |
||||
|
|
||||
|
func NewBatchQueue() *BatchQueue { |
||||
|
return &BatchQueue{ |
||||
|
queue: []*BatchInfo{}, |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
// Push adds the given BatchInfo to the BatchQueue
|
||||
|
func (bq *BatchQueue) Push(b *BatchInfo) { |
||||
|
bq.queue = append(bq.queue, b) |
||||
|
} |
||||
|
|
||||
|
// Pop pops the first BatchInfo from the BatchQueue
|
||||
|
func (bq *BatchQueue) Pop() *BatchInfo { |
||||
|
if len(bq.queue) == 0 { |
||||
|
return nil |
||||
|
} |
||||
|
b := bq.queue[0] |
||||
|
bq.queue = bq.queue[1:] |
||||
|
return b |
||||
|
} |
@ -0,0 +1,26 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"testing" |
||||
|
|
||||
|
"github.com/stretchr/testify/assert" |
||||
|
) |
||||
|
|
||||
|
func TestBatchQueue(t *testing.T) { |
||||
|
bq := BatchQueue{} |
||||
|
|
||||
|
bq.Push(&BatchInfo{ |
||||
|
batchNum: 0, |
||||
|
}) |
||||
|
bq.Push(&BatchInfo{ |
||||
|
batchNum: 2, |
||||
|
}) |
||||
|
bq.Push(&BatchInfo{ |
||||
|
batchNum: 1, |
||||
|
}) |
||||
|
|
||||
|
assert.Equal(t, uint64(0), bq.Pop().batchNum) |
||||
|
assert.Equal(t, uint64(2), bq.Pop().batchNum) |
||||
|
assert.Equal(t, uint64(1), bq.Pop().batchNum) |
||||
|
assert.Nil(t, bq.Pop()) |
||||
|
} |
@ -0,0 +1,225 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import ( |
||||
|
"time" |
||||
|
|
||||
|
ethCommon "github.com/ethereum/go-ethereum/common" |
||||
|
"github.com/hermeznetwork/hermez-node/batchbuilder" |
||||
|
"github.com/hermeznetwork/hermez-node/common" |
||||
|
"github.com/hermeznetwork/hermez-node/eth" |
||||
|
"github.com/hermeznetwork/hermez-node/txselector" |
||||
|
kvdb "github.com/iden3/go-merkletree/db" |
||||
|
"github.com/iden3/go-merkletree/db/memory" |
||||
|
) |
||||
|
|
||||
|
// CoordinatorConfig contains the Coordinator configuration
|
||||
|
type CoordinatorConfig struct { |
||||
|
ForgerAddress ethCommon.Address |
||||
|
} |
||||
|
|
||||
|
// Coordinator implements the Coordinator type
|
||||
|
type Coordinator struct { |
||||
|
config CoordinatorConfig |
||||
|
|
||||
|
batchNum uint64 |
||||
|
batchQueue *BatchQueue |
||||
|
serverProofPool ServerProofPool |
||||
|
|
||||
|
// synchronizer *synchronizer.Synchronizer
|
||||
|
txsel *txselector.TxSelector |
||||
|
batchBuilder *batchbuilder.BatchBuilder |
||||
|
|
||||
|
ethClient *eth.EthClient |
||||
|
ethTxStore kvdb.Storage |
||||
|
} |
||||
|
|
||||
|
// NewCoordinator creates a new Coordinator
|
||||
|
func NewCoordinator() *Coordinator { // once synchronizer is ready, synchronizer.Synchronizer will be passed as parameter here
|
||||
|
var c *Coordinator |
||||
|
c.ethClient = eth.NewEthClient() // TBD
|
||||
|
c.ethTxStore = memory.NewMemoryStorage() |
||||
|
return c |
||||
|
} |
||||
|
|
||||
|
// Start starts the Coordinator service
|
||||
|
func (c *Coordinator) Start() { |
||||
|
// TODO TBD note: the sequences & loops & errors & logging & goroutines
|
||||
|
// & channels approach still needs to be defined, the current code is a
|
||||
|
// wip draft
|
||||
|
|
||||
|
// TBD: goroutines strategy
|
||||
|
|
||||
|
// if in Forge Sequence:
|
||||
|
if c.isForgeSequence() { |
||||
|
// c.batchNum = c.synchronizer.LastBatchNum()
|
||||
|
_ = c.txsel.Reset(c.batchNum) |
||||
|
_ = c.batchBuilder.Reset(c.batchNum, true) |
||||
|
c.batchQueue = NewBatchQueue() |
||||
|
go func() { |
||||
|
for { |
||||
|
_ = c.forgeSequence() |
||||
|
time.Sleep(1 * time.Second) |
||||
|
} |
||||
|
}() |
||||
|
go func() { |
||||
|
for { |
||||
|
_ = c.proveSequence() |
||||
|
time.Sleep(1 * time.Second) |
||||
|
} |
||||
|
}() |
||||
|
go func() { |
||||
|
for { |
||||
|
_ = c.forgeConfirmationSequence() |
||||
|
time.Sleep(1 * time.Second) |
||||
|
} |
||||
|
}() |
||||
|
} |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) forgeSequence() error { |
||||
|
// TODO once synchronizer has this method ready:
|
||||
|
// If there's been a reorg, handle it
|
||||
|
// handleReorg() function decides if the reorg must restart the pipeline or not
|
||||
|
// if c.synchronizer.Reorg():
|
||||
|
_ = c.handleReorg() |
||||
|
|
||||
|
// 0. If there's an available server proof: Start pipeline for batchNum = batchNum + 1
|
||||
|
serverProofInfo, err := c.serverProofPool.GetNextAvailable() // blocking call, returns when a server proof is available
|
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
// remove transactions from the pool that have been there for too long
|
||||
|
err = c.purgeRemoveByTimeout() |
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
c.batchNum = c.batchNum + 1 |
||||
|
batchInfo := NewBatchInfo(c.batchNum, serverProofInfo) // to accumulate metadata of the batch
|
||||
|
|
||||
|
var l2Txs []common.PoolL2Tx |
||||
|
// var feesInfo
|
||||
|
var l1UserTxsExtra, l1OperatorTxs []common.L1Tx |
||||
|
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||
|
if c.shouldL1L2Batch() { |
||||
|
// 2a: L1+L2 txs
|
||||
|
// l1UserTxs, toForgeL1TxsNumber := c.synchronizer.GetNextL1UserTxs() // TODO once synchronizer is ready, uncomment
|
||||
|
var l1UserTxs []common.L1Tx = nil // tmp, depends on synchronizer
|
||||
|
l1UserTxsExtra, l1OperatorTxs, l2Txs, err = c.txsel.GetL1L2TxSelection(c.batchNum, l1UserTxs) // TODO once feesInfo is added to method return, add the var
|
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
} else { |
||||
|
// 2b: only L2 txs
|
||||
|
l2Txs, err = c.txsel.GetL2TxSelection(c.batchNum) // TODO once feesInfo is added to method return, add the var
|
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
l1UserTxsExtra = nil |
||||
|
l1OperatorTxs = nil |
||||
|
} |
||||
|
|
||||
|
// Run purger to invalidate transactions that become invalid beause of
|
||||
|
// the l2Txs selected. Will mark as invalid the txs that have a
|
||||
|
// (fromIdx, nonce) which already appears in the selected txs (includes
|
||||
|
// all the nonces smaller than the current one)
|
||||
|
err = c.purgeInvalidDueToL2TxsSelection(l2Txs) |
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
// 3. Save metadata from TxSelector output for BatchNum
|
||||
|
batchInfo.SetTxsInfo(l1UserTxsExtra, l1OperatorTxs, l2Txs) // TODO feesInfo
|
||||
|
|
||||
|
// 4. Call BatchBuilder with TxSelector output
|
||||
|
configBatch := batchbuilder.ConfigBatch{ |
||||
|
ForgerAddress: c.config.ForgerAddress, |
||||
|
} |
||||
|
zkInputs, err := c.batchBuilder.BuildBatch(configBatch, l1UserTxsExtra, l1OperatorTxs, l2Txs, nil) // TODO []common.TokenID --> feesInfo
|
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
|
||||
|
// 5. Save metadata from BatchBuilder output for BatchNum
|
||||
|
batchInfo.SetZKInputs(zkInputs) |
||||
|
|
||||
|
// 6. Call an idle server proof with BatchBuilder output, save server proof info for batchNum
|
||||
|
err = batchInfo.serverProof.CalculateProof(zkInputs) |
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
c.batchQueue.Push(&batchInfo) |
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
// proveSequence gets the generated zkProof & sends it to the SmartContract
|
||||
|
func (c *Coordinator) proveSequence() error { |
||||
|
batchInfo := c.batchQueue.Pop() |
||||
|
if batchInfo == nil { |
||||
|
// no batches in queue, return
|
||||
|
return common.ErrBatchQueueEmpty |
||||
|
} |
||||
|
serverProofInfo := batchInfo.serverProof |
||||
|
proof, err := serverProofInfo.GetProof() // blocking call, until not resolved don't continue. Returns when the proof server has calculated the proof
|
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
batchInfo.SetProof(proof) |
||||
|
callData := c.prepareCallDataForge(batchInfo) |
||||
|
_, err = c.ethClient.ForgeCall(callData) |
||||
|
if err != nil { |
||||
|
return err |
||||
|
} |
||||
|
// TODO once tx data type is defined, store ethTx (returned by ForgeCall)
|
||||
|
// TBD if use ethTxStore as a disk k-v database, or use a Queue
|
||||
|
// tx, err := c.ethTxStore.NewTx()
|
||||
|
// if err != nil {
|
||||
|
// return err
|
||||
|
// }
|
||||
|
// tx.Put(ethTx.Hash(), ethTx.Bytes())
|
||||
|
// if err := tx.Commit(); err!=nil {
|
||||
|
// return nil
|
||||
|
// }
|
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) forgeConfirmationSequence() error { |
||||
|
// TODO strategy of this sequence TBD
|
||||
|
// confirm eth txs and mark them as accepted sequence
|
||||
|
// ethTx := ethTxStore.GetFirstPending()
|
||||
|
// waitForAccepted(ethTx) // blocking call, returns once the ethTx is mined
|
||||
|
// ethTxStore.MarkAccepted(ethTx)
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) handleReorg() error { |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
// isForgeSequence returns true if the node is the Forger in the current ethereum block
|
||||
|
func (c *Coordinator) isForgeSequence() bool { |
||||
|
|
||||
|
return false |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) purgeRemoveByTimeout() error { |
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) purgeInvalidDueToL2TxsSelection(l2Txs []common.PoolL2Tx) error { |
||||
|
|
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) shouldL1L2Batch() bool { |
||||
|
|
||||
|
return false |
||||
|
} |
||||
|
|
||||
|
func (c *Coordinator) prepareCallDataForge(batchInfo *BatchInfo) *common.CallDataForge { |
||||
|
return nil |
||||
|
} |
@ -0,0 +1,25 @@ |
|||||
|
package coordinator |
||||
|
|
||||
|
import "github.com/hermeznetwork/hermez-node/common" |
||||
|
|
||||
|
type ServerProofInfo struct { |
||||
|
// TODO
|
||||
|
Available bool |
||||
|
} |
||||
|
|
||||
|
func (p *ServerProofInfo) CalculateProof(zkInputs *common.ZKInputs) error { |
||||
|
return nil |
||||
|
} |
||||
|
|
||||
|
func (p *ServerProofInfo) GetProof() (*Proof, error) { |
||||
|
return nil, nil |
||||
|
} |
||||
|
|
||||
|
type ServerProofPool struct { |
||||
|
// pool []ServerProofInfo
|
||||
|
} |
||||
|
|
||||
|
func (p *ServerProofPool) GetNextAvailable() (*ServerProofInfo, error) { |
||||
|
|
||||
|
return nil, nil |
||||
|
} |
@ -0,0 +1,15 @@ |
|||||
|
package eth |
||||
|
|
||||
|
import "github.com/hermeznetwork/hermez-node/common" |
||||
|
|
||||
|
type EthClient struct { |
||||
|
} |
||||
|
|
||||
|
func NewEthClient() *EthClient { |
||||
|
// TODO
|
||||
|
return &EthClient{} |
||||
|
} |
||||
|
func (ec *EthClient) ForgeCall(callData *common.CallDataForge) ([]byte, error) { |
||||
|
// TODO this depends on the smart contracts, once are ready this will be updated
|
||||
|
return nil, nil |
||||
|
} |
@ -0,0 +1,13 @@ |
|||||
|
package node |
||||
|
|
||||
|
type mode string |
||||
|
|
||||
|
// ModeCoordinator defines the mode of the HermezNode as Coordinator, which
|
||||
|
// means that the node is set to forge (which also will be synchronizing with
|
||||
|
// the L1 blockchain state)
|
||||
|
const ModeCoordinator mode = "coordinator" |
||||
|
|
||||
|
// ModeSynchronizer defines the mode of the HermezNode as Synchronizer, which
|
||||
|
// means that the node is set to only synchronize with the L1 blockchain state
|
||||
|
// and will not forge
|
||||
|
const ModeSynchronizer mode = "synchronizer" |