Browse Source

Merge pull request #231 from hermeznetwork/feature/statedb-createdaccounts

ProcessTxs return an array of created accounts
feature/sql-semaphore1
Eduard S 4 years ago
committed by GitHub
parent
commit
d6d1a9c898
No known key found for this signature in database GPG Key ID: 4AEE18F83AFDEB23
5 changed files with 95 additions and 54 deletions
  1. +2
    -2
      batchbuilder/batchbuilder.go
  2. +75
    -39
      db/statedb/txprocessors.go
  3. +14
    -9
      db/statedb/txprocessors_test.go
  4. +2
    -2
      synchronizer/synchronizer.go
  5. +2
    -2
      txselector/txselector.go

+ 2
- 2
batchbuilder/batchbuilder.go

@ -52,10 +52,10 @@ func (bb *BatchBuilder) Reset(batchNum common.BatchNum, fromSynchronizer bool) e
// BuildBatch takes the transactions and returns the common.ZKInputs of the next batch // BuildBatch takes the transactions and returns the common.ZKInputs of the next batch
func (bb *BatchBuilder) BuildBatch(coordIdxs []common.Idx, configBatch *ConfigBatch, l1usertxs, l1coordinatortxs []common.L1Tx, pooll2txs []common.PoolL2Tx, tokenIDs []common.TokenID) (*common.ZKInputs, error) { func (bb *BatchBuilder) BuildBatch(coordIdxs []common.Idx, configBatch *ConfigBatch, l1usertxs, l1coordinatortxs []common.L1Tx, pooll2txs []common.PoolL2Tx, tokenIDs []common.TokenID) (*common.ZKInputs, error) {
zkInputs, _, err := bb.localStateDB.ProcessTxs(coordIdxs, l1usertxs, l1coordinatortxs, pooll2txs)
ptOut, err := bb.localStateDB.ProcessTxs(coordIdxs, l1usertxs, l1coordinatortxs, pooll2txs)
if err != nil { if err != nil {
return nil, err return nil, err
} }
err = bb.localStateDB.MakeCheckpoint() err = bb.localStateDB.MakeCheckpoint()
return zkInputs, err
return ptOut.ZKInputs, err
} }

+ 75
- 39
db/statedb/txprocessors.go

@ -21,7 +21,7 @@ var (
func (s *StateDB) resetZKInputs() { func (s *StateDB) resetZKInputs() {
s.zki = nil s.zki = nil
s.i = 0
s.i = 0 // initialize current transaction index in the ZKInputs generation
} }
type processedExit struct { type processedExit struct {
@ -31,6 +31,14 @@ type processedExit struct {
acc common.Account acc common.Account
} }
// ProcessTxOutput contains the output of the ProcessTxs method
type ProcessTxOutput struct {
ZKInputs *common.ZKInputs
ExitInfos []common.ExitInfo
CreatedAccounts []common.Account
CoordinatorIdxsMap map[common.TokenID]common.Idx
}
// ProcessTxs process the given L1Txs & L2Txs applying the needed updates to // ProcessTxs process the given L1Txs & L2Txs applying the needed updates to
// the StateDB depending on the transaction Type. If StateDB // the StateDB depending on the transaction Type. If StateDB
// type==TypeBatchBuilder, returns the common.ZKInputs to generate the // type==TypeBatchBuilder, returns the common.ZKInputs to generate the
@ -38,26 +46,34 @@ type processedExit struct {
// type==TypeSynchronizer, assumes that the call is done from the Synchronizer, // type==TypeSynchronizer, assumes that the call is done from the Synchronizer,
// returns common.ExitTreeLeaf that is later used by the Synchronizer to update // returns common.ExitTreeLeaf that is later used by the Synchronizer to update
// the HistoryDB, and adds Nonce & TokenID to the L2Txs. // the HistoryDB, and adds Nonce & TokenID to the L2Txs.
func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs []common.L1Tx, l2txs []common.PoolL2Tx) (*common.ZKInputs, []common.ExitInfo, error) {
// And if TypeSynchronizer returns an array of common.Account with all the
// created accounts.
func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs []common.L1Tx, l2txs []common.PoolL2Tx) (*ProcessTxOutput, error) {
var err error var err error
var exitTree *merkletree.MerkleTree var exitTree *merkletree.MerkleTree
var createdAccounts []common.Account
if s.zki != nil { if s.zki != nil {
return nil, nil, errors.New("Expected StateDB.zki==nil, something went wrong and it's not empty")
return nil, errors.New("Expected StateDB.zki==nil, something went wrong and it's not empty")
} }
defer s.resetZKInputs() defer s.resetZKInputs()
nTx := len(l1usertxs) + len(l1coordinatortxs) + len(l2txs) nTx := len(l1usertxs) + len(l1coordinatortxs) + len(l2txs)
if nTx == 0 { if nTx == 0 {
// TODO return ZKInputs of batch without txs // TODO return ZKInputs of batch without txs
return nil, nil, nil
return &ProcessTxOutput{
ZKInputs: nil,
ExitInfos: nil,
CreatedAccounts: nil,
CoordinatorIdxsMap: nil,
}, nil
} }
exits := make([]processedExit, nTx) exits := make([]processedExit, nTx)
// get TokenIDs of coordIdxs // get TokenIDs of coordIdxs
coordIdxsMap, err := s.getTokenIDsFromIdxs(coordIdxs) coordIdxsMap, err := s.getTokenIDsFromIdxs(coordIdxs)
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
if s.typ == TypeBatchBuilder { if s.typ == TypeBatchBuilder {
@ -71,7 +87,7 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder { if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
tmpDir, err := ioutil.TempDir("", "hermez-statedb-exittree") tmpDir, err := ioutil.TempDir("", "hermez-statedb-exittree")
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
defer func() { defer func() {
if err := os.RemoveAll(tmpDir); err != nil { if err := os.RemoveAll(tmpDir); err != nil {
@ -80,19 +96,19 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
}() }()
sto, err := pebble.NewPebbleStorage(tmpDir, false) sto, err := pebble.NewPebbleStorage(tmpDir, false)
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
exitTree, err = merkletree.NewMerkleTree(sto, s.mt.MaxLevels()) exitTree, err = merkletree.NewMerkleTree(sto, s.mt.MaxLevels())
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
} }
// assumption: l1usertx are sorted by L1Tx.Position // assumption: l1usertx are sorted by L1Tx.Position
for i := 0; i < len(l1usertxs); i++ { for i := 0; i < len(l1usertxs); i++ {
exitIdx, exitAccount, newExit, err := s.processL1Tx(exitTree, &l1usertxs[i])
exitIdx, exitAccount, newExit, createdAccount, err := s.processL1Tx(exitTree, &l1usertxs[i])
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder { if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
if exitIdx != nil && exitTree != nil { if exitIdx != nil && exitTree != nil {
@ -105,31 +121,26 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
} }
s.i++ s.i++
} }
if s.typ == TypeSynchronizer && createdAccount != nil {
createdAccounts = append(createdAccounts, *createdAccount)
}
} }
for i := 0; i < len(l1coordinatortxs); i++ { for i := 0; i < len(l1coordinatortxs); i++ {
exitIdx, exitAccount, newExit, err := s.processL1Tx(exitTree, &l1coordinatortxs[i])
exitIdx, _, _, createdAccount, err := s.processL1Tx(exitTree, &l1coordinatortxs[i])
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
if exitIdx != nil { if exitIdx != nil {
log.Error("Unexpected Exit in L1CoordinatorTx") log.Error("Unexpected Exit in L1CoordinatorTx")
} }
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
if exitIdx != nil && exitTree != nil {
exits[s.i] = processedExit{
exit: true,
newExit: newExit,
idx: *exitIdx,
acc: *exitAccount,
}
}
s.i++
if s.typ == TypeSynchronizer && createdAccount != nil {
createdAccounts = append(createdAccounts, *createdAccount)
} }
} }
for i := 0; i < len(l2txs); i++ { for i := 0; i < len(l2txs); i++ {
exitIdx, exitAccount, newExit, err := s.processL2Tx(coordIdxsMap, exitTree, &l2txs[i]) exitIdx, exitAccount, newExit, err := s.processL2Tx(coordIdxsMap, exitTree, &l2txs[i])
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder { if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
if exitIdx != nil && exitTree != nil { if exitIdx != nil && exitTree != nil {
@ -145,12 +156,13 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
} }
if s.typ == TypeTxSelector { if s.typ == TypeTxSelector {
return nil, nil, nil
return nil, nil
} }
// once all txs processed (exitTree root frozen), for each Exit, // once all txs processed (exitTree root frozen), for each Exit,
// generate common.ExitInfo data // generate common.ExitInfo data
var exitInfos []common.ExitInfo var exitInfos []common.ExitInfo
// exitInfos := []common.ExitInfo{}
for i := 0; i < nTx; i++ { for i := 0; i < nTx; i++ {
if !exits[i].exit { if !exits[i].exit {
continue continue
@ -161,7 +173,7 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
// 0. generate MerkleProof // 0. generate MerkleProof
p, err := exitTree.GenerateCircomVerifierProof(exitIdx.BigInt(), nil) p, err := exitTree.GenerateCircomVerifierProof(exitIdx.BigInt(), nil)
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
// 1. generate common.ExitInfo // 1. generate common.ExitInfo
ei := common.ExitInfo{ ei := common.ExitInfo{
@ -192,7 +204,14 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
} }
} }
if s.typ == TypeSynchronizer { if s.typ == TypeSynchronizer {
return nil, exitInfos, nil
// return exitInfos and createdAccounts, so Synchronizer will
// be able to store it into HistoryDB for the concrete BatchNum
return &ProcessTxOutput{
ZKInputs: nil,
ExitInfos: exitInfos,
CreatedAccounts: createdAccounts,
CoordinatorIdxsMap: coordIdxsMap,
}, nil
} }
// compute last ZKInputs parameters // compute last ZKInputs parameters
@ -200,7 +219,7 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
// zki.FeeIdxs = ? // TODO, this will be get from the config file // zki.FeeIdxs = ? // TODO, this will be get from the config file
tokenIDs, err := s.getTokenIDsBigInt(l1usertxs, l1coordinatortxs, l2txs) tokenIDs, err := s.getTokenIDsBigInt(l1usertxs, l1coordinatortxs, l2txs)
if err != nil { if err != nil {
return nil, nil, err
return nil, err
} }
s.zki.FeePlanTokens = tokenIDs s.zki.FeePlanTokens = tokenIDs
@ -209,9 +228,13 @@ func (s *StateDB) ProcessTxs(coordIdxs []common.Idx, l1usertxs, l1coordinatortxs
// TODO once the Node Config sets the Accounts where to send the Fees // TODO once the Node Config sets the Accounts where to send the Fees
// compute fees & update ZKInputs // compute fees & update ZKInputs
// return exitInfos, so Synchronizer will be able to store it into
// HistoryDB for the concrete BatchNum
return s.zki, exitInfos, nil
// return ZKInputs as the BatchBuilder will return it to forge the Batch
return &ProcessTxOutput{
ZKInputs: s.zki,
ExitInfos: nil,
CreatedAccounts: nil,
CoordinatorIdxsMap: coordIdxsMap,
}, nil
} }
// getTokenIDsBigInt returns the list of TokenIDs in *big.Int format // getTokenIDsBigInt returns the list of TokenIDs in *big.Int format
@ -243,7 +266,10 @@ func (s *StateDB) getTokenIDsBigInt(l1usertxs, l1coordinatortxs []common.L1Tx, l
// StateDB depending on the transaction Type. It returns the 3 parameters // StateDB depending on the transaction Type. It returns the 3 parameters
// related to the Exit (in case of): Idx, ExitAccount, boolean determining if // related to the Exit (in case of): Idx, ExitAccount, boolean determining if
// the Exit created a new Leaf in the ExitTree. // the Exit created a new Leaf in the ExitTree.
func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx) (*common.Idx, *common.Account, bool, error) {
// And another *common.Account parameter which contains the created account in
// case that has been a new created account and that the StateDB is of type
// TypeSynchronizer.
func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx) (*common.Idx, *common.Account, bool, *common.Account, error) {
// ZKInputs // ZKInputs
if s.zki != nil { if s.zki != nil {
// Txs // Txs
@ -273,14 +299,14 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
err := s.applyTransfer(nil, tx.Tx(), 0) err := s.applyTransfer(nil, tx.Tx(), 0)
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
case common.TxTypeCreateAccountDeposit: case common.TxTypeCreateAccountDeposit:
// add new account to the MT, update balance of the MT account // add new account to the MT, update balance of the MT account
err := s.applyCreateAccount(tx) err := s.applyCreateAccount(tx)
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
// TODO applyCreateAccount will return the created account, // TODO applyCreateAccount will return the created account,
// which in the case type==TypeSynchronizer will be added to an // which in the case type==TypeSynchronizer will be added to an
@ -295,7 +321,7 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
err := s.applyDeposit(tx, false) err := s.applyDeposit(tx, false)
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
case common.TxTypeDepositTransfer: case common.TxTypeDepositTransfer:
// update balance in MT account, update balance & nonce of sender // update balance in MT account, update balance & nonce of sender
@ -303,7 +329,7 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
err := s.applyDeposit(tx, true) err := s.applyDeposit(tx, true)
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
case common.TxTypeCreateAccountDepositTransfer: case common.TxTypeCreateAccountDepositTransfer:
// add new account to the merkletree, update balance in MT account, // add new account to the merkletree, update balance in MT account,
@ -311,7 +337,7 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
err := s.applyCreateAccountDepositTransfer(tx) err := s.applyCreateAccountDepositTransfer(tx)
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
if s.zki != nil { if s.zki != nil {
@ -324,13 +350,23 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
exitAccount, newExit, err := s.applyExit(nil, exitTree, tx.Tx()) exitAccount, newExit, err := s.applyExit(nil, exitTree, tx.Tx())
if err != nil { if err != nil {
log.Error(err) log.Error(err)
return nil, nil, false, err
return nil, nil, false, nil, err
} }
return &tx.FromIdx, exitAccount, newExit, nil
return &tx.FromIdx, exitAccount, newExit, nil, nil
default: default:
} }
return nil, nil, false, nil
var createdAccount *common.Account
if s.typ == TypeSynchronizer && (tx.Type == common.TxTypeCreateAccountDeposit || tx.Type == common.TxTypeCreateAccountDepositTransfer) {
var err error
createdAccount, err = s.GetAccount(s.idx)
if err != nil {
log.Error(err)
return nil, nil, false, nil, err
}
}
return nil, nil, false, createdAccount, nil
} }
// processL2Tx process the given L2Tx applying the needed updates to the // processL2Tx process the given L2Tx applying the needed updates to the

+ 14
- 9
db/statedb/txprocessors_test.go

@ -45,45 +45,50 @@ func TestProcessTxsSynchronizer(t *testing.T) {
idxA1 := tc.Users["A"].Accounts[common.TokenID(1)].Idx idxA1 := tc.Users["A"].Accounts[common.TokenID(1)].Idx
log.Debug("1st batch, 1st block, only L1CoordinatorTxs") log.Debug("1st batch, 1st block, only L1CoordinatorTxs")
_, _, err = sdb.ProcessTxs(nil, nil, blocks[0].Batches[0].L1CoordinatorTxs, nil)
ptOut, err := sdb.ProcessTxs(nil, nil, blocks[0].Batches[0].L1CoordinatorTxs, nil)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, 4, len(ptOut.CreatedAccounts))
log.Debug("2nd batch, 1st block") log.Debug("2nd batch, 1st block")
l2Txs := common.L2TxsToPoolL2Txs(blocks[0].Batches[1].L2Txs) l2Txs := common.L2TxsToPoolL2Txs(blocks[0].Batches[1].L2Txs)
_, exitInfos, err := sdb.ProcessTxs(coordIdxs, blocks[0].L1UserTxs, blocks[0].Batches[1].L1CoordinatorTxs, l2Txs)
ptOut, err = sdb.ProcessTxs(coordIdxs, blocks[0].L1UserTxs, blocks[0].Batches[1].L1CoordinatorTxs, l2Txs)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, 0, len(exitInfos))
assert.Equal(t, 0, len(ptOut.ExitInfos))
assert.Equal(t, 31, len(ptOut.CreatedAccounts))
acc, err := sdb.GetAccount(idxA1) acc, err := sdb.GetAccount(idxA1)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, "50", acc.Balance.String()) assert.Equal(t, "50", acc.Balance.String())
log.Debug("3rd batch, 1st block") log.Debug("3rd batch, 1st block")
l2Txs = common.L2TxsToPoolL2Txs(blocks[0].Batches[2].L2Txs) l2Txs = common.L2TxsToPoolL2Txs(blocks[0].Batches[2].L2Txs)
_, exitInfos, err = sdb.ProcessTxs(coordIdxs, nil, blocks[0].Batches[2].L1CoordinatorTxs, l2Txs)
ptOut, err = sdb.ProcessTxs(coordIdxs, nil, blocks[0].Batches[2].L1CoordinatorTxs, l2Txs)
require.Nil(t, err) require.Nil(t, err)
// TODO once TTGL is updated, add a check that a input poolL2Tx with // TODO once TTGL is updated, add a check that a input poolL2Tx with
// Nonce & TokenID =0, after ProcessTxs call has the expected value // Nonce & TokenID =0, after ProcessTxs call has the expected value
assert.Equal(t, 0, len(exitInfos))
assert.Equal(t, 0, len(ptOut.ExitInfos))
assert.Equal(t, 0, len(ptOut.CreatedAccounts))
acc, err = sdb.GetAccount(idxA1) acc, err = sdb.GetAccount(idxA1)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, "28", acc.Balance.String()) assert.Equal(t, "28", acc.Balance.String())
log.Debug("1st batch, 2nd block") log.Debug("1st batch, 2nd block")
l2Txs = common.L2TxsToPoolL2Txs(blocks[1].Batches[0].L2Txs) l2Txs = common.L2TxsToPoolL2Txs(blocks[1].Batches[0].L2Txs)
_, exitInfos, err = sdb.ProcessTxs(coordIdxs, nil, blocks[1].Batches[0].L1CoordinatorTxs, l2Txs)
ptOut, err = sdb.ProcessTxs(coordIdxs, nil, blocks[1].Batches[0].L1CoordinatorTxs, l2Txs)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, 4, len(exitInfos)) // the 'ForceExit(1)' is not computed yet, as the batch is without L1UserTxs
assert.Equal(t, 4, len(ptOut.ExitInfos)) // the 'ForceExit(1)' is not computed yet, as the batch is without L1UserTxs
assert.Equal(t, 1, len(ptOut.CreatedAccounts))
acc, err = sdb.GetAccount(idxA1) acc, err = sdb.GetAccount(idxA1)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, "53", acc.Balance.String()) assert.Equal(t, "53", acc.Balance.String())
log.Debug("2nd batch, 2nd block") log.Debug("2nd batch, 2nd block")
l2Txs = common.L2TxsToPoolL2Txs(blocks[1].Batches[1].L2Txs) l2Txs = common.L2TxsToPoolL2Txs(blocks[1].Batches[1].L2Txs)
_, exitInfos, err = sdb.ProcessTxs(coordIdxs, blocks[1].L1UserTxs, blocks[1].Batches[1].L1CoordinatorTxs, l2Txs)
ptOut, err = sdb.ProcessTxs(coordIdxs, blocks[1].L1UserTxs, blocks[1].Batches[1].L1CoordinatorTxs, l2Txs)
require.Nil(t, err) require.Nil(t, err)
assert.Equal(t, 2, len(exitInfos)) // 2, as previous batch was without L1UserTxs, and has pending the 'ForceExit(1) A: 5'
assert.Equal(t, 2, len(ptOut.ExitInfos)) // 2, as previous batch was without L1UserTxs, and has pending the 'ForceExit(1) A: 5'
assert.Equal(t, 1, len(ptOut.CreatedAccounts))
acc, err = sdb.GetAccount(idxA1) acc, err = sdb.GetAccount(idxA1)
assert.Nil(t, err) assert.Nil(t, err)
assert.Equal(t, "78", acc.Balance.String()) assert.Equal(t, "78", acc.Balance.String())

+ 2
- 2
synchronizer/synchronizer.go

@ -405,7 +405,7 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*rollupData, error) {
// TODO: Get CollectedFees from ProcessTxs() // TODO: Get CollectedFees from ProcessTxs()
// TODO: Pass forgeBatchArgs.FeeIdxCoordinator to ProcessTxs() // TODO: Pass forgeBatchArgs.FeeIdxCoordinator to ProcessTxs()
// ProcessTxs updates poolL2Txs adding: Nonce, TokenID // ProcessTxs updates poolL2Txs adding: Nonce, TokenID
_, exitInfo, err := s.stateDB.ProcessTxs([]common.Idx{}, l1UserTxs, batchData.L1CoordinatorTxs, poolL2Txs)
ptOut, err := s.stateDB.ProcessTxs([]common.Idx{}, l1UserTxs, batchData.L1CoordinatorTxs, poolL2Txs)
if err != nil { if err != nil {
return nil, err return nil, err
} }
@ -429,7 +429,7 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*rollupData, error) {
position++ position++
} }
batchData.ExitTree = exitInfo
batchData.ExitTree = ptOut.ExitInfos
slotNum := int64(0) slotNum := int64(0)
if ethBlock.EthBlockNum >= s.auctionConstants.GenesisBlockNum { if ethBlock.EthBlockNum >= s.auctionConstants.GenesisBlockNum {

+ 2
- 2
txselector/txselector.go

@ -89,7 +89,7 @@ func (txsel *TxSelector) GetL2TxSelection(coordIdxs []common.Idx, batchNum commo
txs := txsel.getL2Profitable(validTxs, txsel.MaxTxs) txs := txsel.getL2Profitable(validTxs, txsel.MaxTxs)
// process the txs in the local AccountsDB // process the txs in the local AccountsDB
_, _, err = txsel.localAccountsDB.ProcessTxs(coordIdxs, nil, nil, txs)
_, err = txsel.localAccountsDB.ProcessTxs(coordIdxs, nil, nil, txs)
if err != nil { if err != nil {
return nil, err return nil, err
} }
@ -238,7 +238,7 @@ func (txsel *TxSelector) GetL1L2TxSelection(coordIdxs []common.Idx, batchNum com
l2Txs := txsel.getL2Profitable(validTxs, maxL2Txs) l2Txs := txsel.getL2Profitable(validTxs, maxL2Txs)
// process the txs in the local AccountsDB // process the txs in the local AccountsDB
_, _, err = txsel.localAccountsDB.ProcessTxs(coordIdxs, l1Txs, l1CoordinatorTxs, l2Txs)
_, err = txsel.localAccountsDB.ProcessTxs(coordIdxs, l1Txs, l1CoordinatorTxs, l2Txs)
if err != nil { if err != nil {
return nil, nil, nil, err return nil, nil, nil, err
} }

Loading…
Cancel
Save