mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 11:26:44 +01:00
TxSelector update txs processing logic
This commit is contained in:
@@ -113,7 +113,7 @@ func newTestModules(t *testing.T) modules {
|
|||||||
|
|
||||||
coordAccount := &txselector.CoordAccount{ // TODO TMP
|
coordAccount := &txselector.CoordAccount{ // TODO TMP
|
||||||
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
||||||
BJJ: nil,
|
BJJ: common.EmptyBJJComp,
|
||||||
AccountCreationAuth: nil,
|
AccountCreationAuth: nil,
|
||||||
}
|
}
|
||||||
txSelector, err := txselector.NewTxSelector(coordAccount, txSelDBPath, syncStateDB, l2DB)
|
txSelector, err := txselector.NewTxSelector(coordAccount, txSelDBPath, syncStateDB, l2DB)
|
||||||
|
|||||||
@@ -81,9 +81,9 @@ type StateDB struct {
|
|||||||
zki *common.ZKInputs
|
zki *common.ZKInputs
|
||||||
// i is the current transaction index in the ZKInputs generation (zki)
|
// i is the current transaction index in the ZKInputs generation (zki)
|
||||||
i int
|
i int
|
||||||
// accumulatedFees contains the accumulated fees for each token (Coord
|
// AccumulatedFees contains the accumulated fees for each token (Coord
|
||||||
// Idx) in the processed batch
|
// Idx) in the processed batch
|
||||||
accumulatedFees map[common.Idx]*big.Int
|
AccumulatedFees map[common.Idx]*big.Int
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewStateDB creates a new StateDB, allowing to use an in-memory or in-disk
|
// NewStateDB creates a new StateDB, allowing to use an in-memory or in-disk
|
||||||
|
|||||||
@@ -79,7 +79,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
return nil, tracerr.Wrap(fmt.Errorf("CoordIdxs (%d) length must be smaller than MaxFeeTx (%d)", len(coordIdxs), ptc.MaxFeeTx))
|
return nil, tracerr.Wrap(fmt.Errorf("CoordIdxs (%d) length must be smaller than MaxFeeTx (%d)", len(coordIdxs), ptc.MaxFeeTx))
|
||||||
}
|
}
|
||||||
|
|
||||||
s.accumulatedFees = make(map[common.Idx]*big.Int)
|
s.AccumulatedFees = make(map[common.Idx]*big.Int)
|
||||||
|
|
||||||
nTx := len(l1usertxs) + len(l1coordinatortxs) + len(l2txs)
|
nTx := len(l1usertxs) + len(l1coordinatortxs) + len(l2txs)
|
||||||
|
|
||||||
@@ -124,7 +124,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
// Process L1UserTxs
|
// Process L1UserTxs
|
||||||
for i := 0; i < len(l1usertxs); i++ {
|
for i := 0; i < len(l1usertxs); i++ {
|
||||||
// assumption: l1usertx are sorted by L1Tx.Position
|
// assumption: l1usertx are sorted by L1Tx.Position
|
||||||
exitIdx, exitAccount, newExit, createdAccount, err := s.processL1Tx(exitTree, &l1usertxs[i])
|
exitIdx, exitAccount, newExit, createdAccount, err := s.ProcessL1Tx(exitTree, &l1usertxs[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -166,7 +166,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
|
|
||||||
// Process L1CoordinatorTxs
|
// Process L1CoordinatorTxs
|
||||||
for i := 0; i < len(l1coordinatortxs); i++ {
|
for i := 0; i < len(l1coordinatortxs); i++ {
|
||||||
exitIdx, _, _, createdAccount, err := s.processL1Tx(exitTree, &l1coordinatortxs[i])
|
exitIdx, _, _, createdAccount, err := s.ProcessL1Tx(exitTree, &l1coordinatortxs[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -194,15 +194,15 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
s.accumulatedFees = make(map[common.Idx]*big.Int)
|
s.AccumulatedFees = make(map[common.Idx]*big.Int)
|
||||||
for _, idx := range coordIdxs {
|
for _, idx := range coordIdxs {
|
||||||
s.accumulatedFees[idx] = big.NewInt(0)
|
s.AccumulatedFees[idx] = big.NewInt(0)
|
||||||
}
|
}
|
||||||
|
|
||||||
// once L1UserTxs & L1CoordinatorTxs are processed, get TokenIDs of
|
// once L1UserTxs & L1CoordinatorTxs are processed, get TokenIDs of
|
||||||
// coordIdxs. In this way, if a coordIdx uses an Idx that is being
|
// coordIdxs. In this way, if a coordIdx uses an Idx that is being
|
||||||
// created in the current batch, at this point the Idx will be created
|
// created in the current batch, at this point the Idx will be created
|
||||||
coordIdxsMap, err := s.getTokenIDsFromIdxs(coordIdxs)
|
coordIdxsMap, err := s.GetTokenIDsFromIdxs(coordIdxs)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -228,7 +228,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
|
|
||||||
// Process L2Txs
|
// Process L2Txs
|
||||||
for i := 0; i < len(l2txs); i++ {
|
for i := 0; i < len(l2txs); i++ {
|
||||||
exitIdx, exitAccount, newExit, err := s.processL2Tx(coordIdxsMap, collectedFees, exitTree, &l2txs[i])
|
exitIdx, exitAccount, newExit, err := s.ProcessL2Tx(coordIdxsMap, collectedFees, exitTree, &l2txs[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -287,7 +287,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
// distribute the AccumulatedFees from the processed L2Txs into the
|
// distribute the AccumulatedFees from the processed L2Txs into the
|
||||||
// Coordinator Idxs
|
// Coordinator Idxs
|
||||||
iFee := 0
|
iFee := 0
|
||||||
for idx, accumulatedFee := range s.accumulatedFees {
|
for idx, accumulatedFee := range s.AccumulatedFees {
|
||||||
cmp := accumulatedFee.Cmp(big.NewInt(0))
|
cmp := accumulatedFee.Cmp(big.NewInt(0))
|
||||||
if cmp == 1 { // accumulatedFee>0
|
if cmp == 1 { // accumulatedFee>0
|
||||||
// send the fee to the Idx of the Coordinator for the TokenID
|
// send the fee to the Idx of the Coordinator for the TokenID
|
||||||
@@ -321,7 +321,7 @@ func (s *StateDB) ProcessTxs(ptc ProcessTxsConfig, coordIdxs []common.Idx, l1use
|
|||||||
iFee++
|
iFee++
|
||||||
}
|
}
|
||||||
if s.zki != nil {
|
if s.zki != nil {
|
||||||
for i := len(s.accumulatedFees); i < int(ptc.MaxFeeTx)-1; i++ {
|
for i := len(s.AccumulatedFees); i < int(ptc.MaxFeeTx)-1; i++ {
|
||||||
s.zki.ISStateRootFee[i] = s.mt.Root().BigInt()
|
s.zki.ISStateRootFee[i] = s.mt.Root().BigInt()
|
||||||
}
|
}
|
||||||
// add Coord Idx to ZKInputs.FeeTxsData
|
// add Coord Idx to ZKInputs.FeeTxsData
|
||||||
@@ -400,14 +400,14 @@ func (s *StateDB) getFeePlanTokens(coordIdxs []common.Idx) ([]*big.Int, error) {
|
|||||||
return tBI, nil
|
return tBI, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// processL1Tx process the given L1Tx applying the needed updates to the
|
// ProcessL1Tx process the given L1Tx applying the needed updates to the
|
||||||
// StateDB depending on the transaction Type. It returns the 3 parameters
|
// StateDB depending on the transaction Type. It returns the 3 parameters
|
||||||
// related to the Exit (in case of): Idx, ExitAccount, boolean determining if
|
// related to the Exit (in case of): Idx, ExitAccount, boolean determining if
|
||||||
// the Exit created a new Leaf in the ExitTree.
|
// the Exit created a new Leaf in the ExitTree.
|
||||||
// And another *common.Account parameter which contains the created account in
|
// And another *common.Account parameter which contains the created account in
|
||||||
// case that has been a new created account and that the StateDB is of type
|
// case that has been a new created account and that the StateDB is of type
|
||||||
// TypeSynchronizer.
|
// TypeSynchronizer.
|
||||||
func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx) (*common.Idx, *common.Account, bool, *common.Account, error) {
|
func (s *StateDB) ProcessL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx) (*common.Idx, *common.Account, bool, *common.Account, error) {
|
||||||
// ZKInputs
|
// ZKInputs
|
||||||
if s.zki != nil {
|
if s.zki != nil {
|
||||||
// Txs
|
// Txs
|
||||||
@@ -520,11 +520,11 @@ func (s *StateDB) processL1Tx(exitTree *merkletree.MerkleTree, tx *common.L1Tx)
|
|||||||
return nil, nil, false, createdAccount, nil
|
return nil, nil, false, createdAccount, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// processL2Tx process the given L2Tx applying the needed updates to the
|
// ProcessL2Tx process the given L2Tx applying the needed updates to the
|
||||||
// StateDB depending on the transaction Type. It returns the 3 parameters
|
// StateDB depending on the transaction Type. It returns the 3 parameters
|
||||||
// related to the Exit (in case of): Idx, ExitAccount, boolean determining if
|
// related to the Exit (in case of): Idx, ExitAccount, boolean determining if
|
||||||
// the Exit created a new Leaf in the ExitTree.
|
// the Exit created a new Leaf in the ExitTree.
|
||||||
func (s *StateDB) processL2Tx(coordIdxsMap map[common.TokenID]common.Idx, collectedFees map[common.TokenID]*big.Int,
|
func (s *StateDB) ProcessL2Tx(coordIdxsMap map[common.TokenID]common.Idx, collectedFees map[common.TokenID]*big.Int,
|
||||||
exitTree *merkletree.MerkleTree, tx *common.PoolL2Tx) (*common.Idx, *common.Account, bool, error) {
|
exitTree *merkletree.MerkleTree, tx *common.PoolL2Tx) (*common.Idx, *common.Account, bool, error) {
|
||||||
var err error
|
var err error
|
||||||
// if tx.ToIdx==0, get toIdx by ToEthAddr or ToBJJ
|
// if tx.ToIdx==0, get toIdx by ToEthAddr or ToBJJ
|
||||||
@@ -794,7 +794,7 @@ func (s *StateDB) applyTransfer(coordIdxsMap map[common.TokenID]common.Idx,
|
|||||||
return tracerr.Wrap(fmt.Errorf("Can not use CoordIdx that does not exist in the tree. TokenID: %d, CoordIdx: %d", accSender.TokenID, coordIdxsMap[accSender.TokenID]))
|
return tracerr.Wrap(fmt.Errorf("Can not use CoordIdx that does not exist in the tree. TokenID: %d, CoordIdx: %d", accSender.TokenID, coordIdxsMap[accSender.TokenID]))
|
||||||
}
|
}
|
||||||
// accumulate the fee for the Coord account
|
// accumulate the fee for the Coord account
|
||||||
accumulated := s.accumulatedFees[accCoord.Idx]
|
accumulated := s.AccumulatedFees[accCoord.Idx]
|
||||||
accumulated.Add(accumulated, fee)
|
accumulated.Add(accumulated, fee)
|
||||||
|
|
||||||
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
|
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
|
||||||
@@ -991,7 +991,7 @@ func (s *StateDB) applyExit(coordIdxsMap map[common.TokenID]common.Idx,
|
|||||||
return nil, false, tracerr.Wrap(fmt.Errorf("Can not use CoordIdx that does not exist in the tree. TokenID: %d, CoordIdx: %d", acc.TokenID, coordIdxsMap[acc.TokenID]))
|
return nil, false, tracerr.Wrap(fmt.Errorf("Can not use CoordIdx that does not exist in the tree. TokenID: %d, CoordIdx: %d", acc.TokenID, coordIdxsMap[acc.TokenID]))
|
||||||
}
|
}
|
||||||
// accumulate the fee for the Coord account
|
// accumulate the fee for the Coord account
|
||||||
accumulated := s.accumulatedFees[accCoord.Idx]
|
accumulated := s.AccumulatedFees[accCoord.Idx]
|
||||||
accumulated.Add(accumulated, fee)
|
accumulated.Add(accumulated, fee)
|
||||||
|
|
||||||
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
|
if s.typ == TypeSynchronizer || s.typ == TypeBatchBuilder {
|
||||||
|
|||||||
@@ -123,12 +123,14 @@ func (s *StateDB) GetIdxByEthAddrBJJ(addr ethCommon.Address, pk babyjub.PublicKe
|
|||||||
return common.Idx(0), tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
return common.Idx(0), tracerr.Wrap(fmt.Errorf("GetIdxByEthAddrBJJ: Not found, %s: ToEthAddr: %s, ToBJJ: %s, TokenID: %d", ErrGetIdxNoCase, addr.Hex(), pk, tokenID))
|
||||||
}
|
}
|
||||||
|
|
||||||
func (s *StateDB) getTokenIDsFromIdxs(idxs []common.Idx) (map[common.TokenID]common.Idx, error) {
|
// GetTokenIDsFromIdxs returns a map containing the common.TokenID with its
|
||||||
|
// respective common.Idx for a given slice of common.Idx
|
||||||
|
func (s *StateDB) GetTokenIDsFromIdxs(idxs []common.Idx) (map[common.TokenID]common.Idx, error) {
|
||||||
m := make(map[common.TokenID]common.Idx)
|
m := make(map[common.TokenID]common.Idx)
|
||||||
for i := 0; i < len(idxs); i++ {
|
for i := 0; i < len(idxs); i++ {
|
||||||
a, err := s.GetAccount(idxs[i])
|
a, err := s.GetAccount(idxs[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("getTokenIDsFromIdxs error on GetAccount with Idx==%d: %s", idxs[i], err.Error()))
|
return nil, tracerr.Wrap(fmt.Errorf("GetTokenIDsFromIdxs error on GetAccount with Idx==%d: %s", idxs[i], err.Error()))
|
||||||
}
|
}
|
||||||
m[a.TokenID] = idxs[i]
|
m[a.TokenID] = idxs[i]
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -148,7 +148,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
// TODO: Get (maxL1UserTxs, maxL1OperatorTxs, maxTxs) from the smart contract
|
// TODO: Get (maxL1UserTxs, maxL1OperatorTxs, maxTxs) from the smart contract
|
||||||
coordAccount := &txselector.CoordAccount{ // TODO TMP
|
coordAccount := &txselector.CoordAccount{ // TODO TMP
|
||||||
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
||||||
BJJ: nil,
|
BJJ: common.EmptyBJJComp,
|
||||||
AccountCreationAuth: nil,
|
AccountCreationAuth: nil,
|
||||||
}
|
}
|
||||||
txSelector, err := txselector.NewTxSelector(coordAccount, cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
txSelector, err := txselector.NewTxSelector(coordAccount, cfg.Coordinator.TxSelector.Path, stateDB, l2DB)
|
||||||
|
|||||||
@@ -15,6 +15,7 @@ import (
|
|||||||
"github.com/hermeznetwork/hermez-node/log"
|
"github.com/hermeznetwork/hermez-node/log"
|
||||||
"github.com/hermeznetwork/tracerr"
|
"github.com/hermeznetwork/tracerr"
|
||||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||||
|
"github.com/iden3/go-merkletree/db"
|
||||||
"github.com/iden3/go-merkletree/db/pebble"
|
"github.com/iden3/go-merkletree/db/pebble"
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -42,7 +43,7 @@ func (t txs) Less(i, j int) bool {
|
|||||||
// accounts for the Coordinator to receive the fees.
|
// accounts for the Coordinator to receive the fees.
|
||||||
type CoordAccount struct {
|
type CoordAccount struct {
|
||||||
Addr ethCommon.Address
|
Addr ethCommon.Address
|
||||||
BJJ *babyjub.PublicKey
|
BJJ babyjub.PublicKeyComp
|
||||||
AccountCreationAuth []byte
|
AccountCreationAuth []byte
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -148,19 +149,54 @@ func (txsel *TxSelector) GetCoordIdxs() (map[common.TokenID]common.Idx, error) {
|
|||||||
return r, tracerr.Wrap(err)
|
return r, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
//nolint:unused
|
||||||
|
func (txsel *TxSelector) coordAccountForTokenID(l1CoordinatorTxs []common.L1Tx, tokenID common.TokenID, positionL1 int) (*common.L1Tx, int, error) {
|
||||||
|
// check if CoordinatorAccount for TokenID is already pending to create
|
||||||
|
if checkAlreadyPendingToCreate(l1CoordinatorTxs, tokenID, txsel.coordAccount.Addr, txsel.coordAccount.BJJ) {
|
||||||
|
return nil, positionL1, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
_, err := txsel.coordIdxsDB.Get(tokenID.Bytes())
|
||||||
|
if tracerr.Unwrap(err) == db.ErrNotFound {
|
||||||
|
// create L1CoordinatorTx to create new CoordAccount for TokenID
|
||||||
|
l1CoordinatorTx := common.L1Tx{
|
||||||
|
Position: positionL1,
|
||||||
|
UserOrigin: false,
|
||||||
|
FromEthAddr: txsel.coordAccount.Addr,
|
||||||
|
FromBJJ: txsel.coordAccount.BJJ,
|
||||||
|
TokenID: tokenID,
|
||||||
|
DepositAmount: big.NewInt(0),
|
||||||
|
Type: common.TxTypeCreateAccountDeposit,
|
||||||
|
}
|
||||||
|
positionL1++
|
||||||
|
|
||||||
|
return &l1CoordinatorTx, positionL1, nil
|
||||||
|
}
|
||||||
|
if err != nil {
|
||||||
|
return nil, positionL1, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// CoordAccount for TokenID already exists
|
||||||
|
return nil, positionL1, nil
|
||||||
|
}
|
||||||
|
|
||||||
// GetL2TxSelection returns the L1CoordinatorTxs and a selection of the L2Txs
|
// GetL2TxSelection returns the L1CoordinatorTxs and a selection of the L2Txs
|
||||||
// for the next batch, from the L2DB pool
|
// for the next batch, from the L2DB pool
|
||||||
func (txsel *TxSelector) GetL2TxSelection(selectionConfig *SelectionConfig,
|
func (txsel *TxSelector) GetL2TxSelection(selectionConfig *SelectionConfig,
|
||||||
batchNum common.BatchNum) ([]common.Idx, [][]byte, []common.L1Tx, []common.PoolL2Tx, error) {
|
batchNum common.BatchNum) ([]common.Idx, [][]byte, []common.L1Tx, []common.PoolL2Tx, error) {
|
||||||
coordIdxs, auths, _, l1CoordinatorTxs, l2Txs, err := txsel.GetL1L2TxSelection(selectionConfig, batchNum,
|
coordIdxs, accCreationAuths, _, l1CoordinatorTxs, l2Txs, err := txsel.GetL1L2TxSelection(selectionConfig, batchNum,
|
||||||
[]common.L1Tx{})
|
[]common.L1Tx{})
|
||||||
return coordIdxs, auths, l1CoordinatorTxs, l2Txs, tracerr.Wrap(err)
|
return coordIdxs, accCreationAuths, l1CoordinatorTxs, l2Txs, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// GetL1L2TxSelection returns the selection of L1 + L2 txs
|
// GetL1L2TxSelection returns the selection of L1 + L2 txs
|
||||||
func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
||||||
batchNum common.BatchNum, l1Txs []common.L1Tx) ([]common.Idx, [][]byte, []common.L1Tx, []common.L1Tx,
|
batchNum common.BatchNum, l1Txs []common.L1Tx) ([]common.Idx, [][]byte, []common.L1Tx, []common.L1Tx,
|
||||||
[]common.PoolL2Tx, error) {
|
[]common.PoolL2Tx, error) {
|
||||||
|
// TODO WIP this method uses a 'cherry-pick' of internal calls of the
|
||||||
|
// StateDB, a refactor of the StateDB to reorganize it internally is
|
||||||
|
// planned once the main functionallities are covered, with that
|
||||||
|
// refactor the TxSelector will be updated also
|
||||||
|
|
||||||
// apply l1-user-tx to localAccountDB
|
// apply l1-user-tx to localAccountDB
|
||||||
// create new leaves
|
// create new leaves
|
||||||
// update balances
|
// update balances
|
||||||
@@ -186,6 +222,19 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
var l1CoordinatorTxs []common.L1Tx
|
var l1CoordinatorTxs []common.L1Tx
|
||||||
positionL1 := len(l1Txs)
|
positionL1 := len(l1Txs)
|
||||||
|
|
||||||
|
// Process L1UserTxs
|
||||||
|
for i := 0; i < len(l1Txs); i++ {
|
||||||
|
// assumption: l1usertx are sorted by L1Tx.Position
|
||||||
|
_, _, _, _, err := txsel.localAccountsDB.ProcessL1Tx(nil, &l1Txs[i])
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// get last idx from LocalStateDB
|
||||||
|
// lastIdx := txsel.localStateDB.idx
|
||||||
|
// update lastIdx with the L1UserTxs (of account creation)
|
||||||
|
|
||||||
for i := 0; i < len(l2TxsRaw); i++ {
|
for i := 0; i < len(l2TxsRaw); i++ {
|
||||||
// If tx.ToIdx>=256, tx.ToIdx should exist to localAccountsDB,
|
// If tx.ToIdx>=256, tx.ToIdx should exist to localAccountsDB,
|
||||||
// if so, tx is used. If tx.ToIdx==0, for an L2Tx will be the
|
// if so, tx is used. If tx.ToIdx==0, for an L2Tx will be the
|
||||||
@@ -201,6 +250,7 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
positionL1, l2TxsRaw[i])
|
positionL1, l2TxsRaw[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Debug(err)
|
log.Debug(err)
|
||||||
|
continue
|
||||||
}
|
}
|
||||||
} else if l2TxsRaw[i].ToIdx >= common.IdxUserThreshold {
|
} else if l2TxsRaw[i].ToIdx >= common.IdxUserThreshold {
|
||||||
_, err = txsel.localAccountsDB.GetAccount(l2TxsRaw[i].ToIdx)
|
_, err = txsel.localAccountsDB.GetAccount(l2TxsRaw[i].ToIdx)
|
||||||
@@ -219,24 +269,55 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
// valid txs (of Exit type)
|
// valid txs (of Exit type)
|
||||||
validTxs = append(validTxs, l2TxsRaw[i])
|
validTxs = append(validTxs, l2TxsRaw[i])
|
||||||
}
|
}
|
||||||
|
// TODO if needed add L1CoordinatorTx to create a Coordinator
|
||||||
|
// account for the new TokenID
|
||||||
|
// var newL1CoordTx *common.L1Tx
|
||||||
|
// newL1CoordTx, positionL1, err = txsel.coordAccountForTokenID(l1CoordinatorTxs, l2TxsRaw[i].TokenID, positionL1)
|
||||||
|
// if err != nil {
|
||||||
|
// return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
|
// }
|
||||||
|
// if newL1CoordTx != nil {
|
||||||
|
// l1CoordinatorTxs = append(l1CoordinatorTxs, *newL1CoordTx)
|
||||||
|
// }
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process L1CoordinatorTxs
|
||||||
|
for i := 0; i < len(l1CoordinatorTxs); i++ {
|
||||||
|
fmt.Println("PRINT", i, &l1CoordinatorTxs[i])
|
||||||
|
_, _, _, _, err := txsel.localAccountsDB.ProcessL1Tx(nil, &l1CoordinatorTxs[i])
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
txsel.localAccountsDB.AccumulatedFees = make(map[common.Idx]*big.Int)
|
||||||
|
for _, idx := range coordIdxs {
|
||||||
|
txsel.localAccountsDB.AccumulatedFees[idx] = big.NewInt(0)
|
||||||
|
}
|
||||||
|
|
||||||
|
// once L1UserTxs & L1CoordinatorTxs are processed, get TokenIDs of
|
||||||
|
// coordIdxs. In this way, if a coordIdx uses an Idx that is being
|
||||||
|
// created in the current batch, at this point the Idx will be created
|
||||||
|
coordIdxsMap, err = txsel.localAccountsDB.GetTokenIDsFromIdxs(coordIdxs)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
// get most profitable L2-tx
|
// get most profitable L2-tx
|
||||||
maxL2Txs := selectionConfig.ProcessTxsConfig.MaxTx - uint32(len(l1CoordinatorTxs)) // - len(l1UserTxs) // TODO if there are L1UserTxs take them in to account
|
maxL2Txs := selectionConfig.ProcessTxsConfig.MaxTx - uint32(len(l1CoordinatorTxs)) // - len(l1UserTxs) // TODO if there are L1UserTxs take them in to account
|
||||||
l2Txs := txsel.getL2Profitable(validTxs, maxL2Txs)
|
l2Txs := txsel.getL2Profitable(validTxs, maxL2Txs)
|
||||||
|
|
||||||
//nolint:gomnd
|
// Process L2Txs
|
||||||
ptc := statedb.ProcessTxsConfig{ // TODO TMP
|
for i := 0; i < len(l2Txs); i++ {
|
||||||
NLevels: 32,
|
_, _, _, err = txsel.localAccountsDB.ProcessL2Tx(coordIdxsMap, nil, nil, &l2Txs[i])
|
||||||
MaxFeeTx: 64,
|
|
||||||
MaxTx: 512,
|
|
||||||
MaxL1Tx: 64,
|
|
||||||
}
|
|
||||||
// process the txs in the local AccountsDB
|
|
||||||
_, err = txsel.localAccountsDB.ProcessTxs(ptc, coordIdxs, l1Txs, l1CoordinatorTxs, l2Txs)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
err = txsel.AddCoordIdxs(coordIdxsMap)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
err = txsel.localAccountsDB.MakeCheckpoint()
|
err = txsel.localAccountsDB.MakeCheckpoint()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
return nil, nil, nil, nil, nil, tracerr.Wrap(err)
|
||||||
@@ -263,7 +344,7 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs txs, l1CoordinatorTxs []
|
|||||||
// L1CoordinatorTx of this type, in the DB there still seem that needs
|
// L1CoordinatorTx of this type, in the DB there still seem that needs
|
||||||
// to create a new L1CoordinatorTx, but as is already created, the tx
|
// to create a new L1CoordinatorTx, but as is already created, the tx
|
||||||
// is valid
|
// is valid
|
||||||
if checkAlreadyPendingToCreate(l1CoordinatorTxs, l2Tx.ToEthAddr, l2Tx.ToBJJ) {
|
if checkAlreadyPendingToCreate(l1CoordinatorTxs, l2Tx.TokenID, l2Tx.ToEthAddr, l2Tx.ToBJJ) {
|
||||||
validTxs = append(validTxs, l2Tx)
|
validTxs = append(validTxs, l2Tx)
|
||||||
return validTxs, l1CoordinatorTxs, positionL1, nil
|
return validTxs, l1CoordinatorTxs, positionL1, nil
|
||||||
}
|
}
|
||||||
@@ -362,17 +443,14 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs txs, l1CoordinatorTxs []
|
|||||||
return validTxs, l1CoordinatorTxs, positionL1, nil
|
return validTxs, l1CoordinatorTxs, positionL1, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
func checkAlreadyPendingToCreate(l1CoordinatorTxs []common.L1Tx,
|
func checkAlreadyPendingToCreate(l1CoordinatorTxs []common.L1Tx, tokenID common.TokenID,
|
||||||
addr ethCommon.Address, bjj babyjub.PublicKeyComp) bool {
|
addr ethCommon.Address, bjj babyjub.PublicKeyComp) bool {
|
||||||
for i := 0; i < len(l1CoordinatorTxs); i++ {
|
for i := 0; i < len(l1CoordinatorTxs); i++ {
|
||||||
if bytes.Equal(l1CoordinatorTxs[i].FromEthAddr.Bytes(), addr.Bytes()) {
|
if bytes.Equal(l1CoordinatorTxs[i].FromEthAddr.Bytes(), addr.Bytes()) &&
|
||||||
if bjj == common.EmptyBJJComp {
|
l1CoordinatorTxs[i].TokenID == tokenID &&
|
||||||
|
l1CoordinatorTxs[i].FromBJJ == bjj {
|
||||||
return true
|
return true
|
||||||
}
|
}
|
||||||
if l1CoordinatorTxs[i].FromBJJ == bjj {
|
|
||||||
return true
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
return false
|
return false
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -38,7 +38,7 @@ func initTest(t *testing.T, testSet string) *TxSelector {
|
|||||||
|
|
||||||
coordAccount := &CoordAccount{ // TODO TMP
|
coordAccount := &CoordAccount{ // TODO TMP
|
||||||
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
Addr: ethCommon.HexToAddress("0xc58d29fA6e86E4FAe04DDcEd660d45BCf3Cb2370"),
|
||||||
BJJ: nil,
|
BJJ: common.EmptyBJJComp,
|
||||||
AccountCreationAuth: nil,
|
AccountCreationAuth: nil,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
Reference in New Issue
Block a user