mirror of
https://github.com/arnaucube/hermez-node.git
synced 2026-02-07 03:16:45 +01:00
Compare commits
20 Commits
feature/up
...
feature/to
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
5ccea68905 | ||
|
|
bb8d81c3aa | ||
|
|
af6f114667 | ||
|
|
e2376980f8 | ||
|
|
264f01b572 | ||
|
|
a21793b2b0 | ||
|
|
7b01f6a288 | ||
|
|
f2e5800ebd | ||
|
|
f0e79f3d55 | ||
|
|
26fbeb5c68 | ||
|
|
05104b0565 | ||
|
|
53507edabb | ||
|
|
729966f854 | ||
|
|
1c10a01cf7 | ||
|
|
1d0abe438f | ||
|
|
6a01c0ac14 | ||
|
|
63151a285c | ||
|
|
52d4197330 | ||
|
|
ea63cba62a | ||
|
|
c7e6267189 |
@@ -4,6 +4,10 @@ Go implementation of the Hermez node.
|
|||||||
|
|
||||||
## Developing
|
## Developing
|
||||||
|
|
||||||
|
### Go version
|
||||||
|
|
||||||
|
The `hermez-node` has been tested with go version 1.14
|
||||||
|
|
||||||
### Unit testing
|
### Unit testing
|
||||||
|
|
||||||
Running the unit tests requires a connection to a PostgreSQL database. You can
|
Running the unit tests requires a connection to a PostgreSQL database. You can
|
||||||
|
|||||||
@@ -2916,7 +2916,7 @@ components:
|
|||||||
example: 101
|
example: 101
|
||||||
l1UserTotalBytes:
|
l1UserTotalBytes:
|
||||||
type: integer
|
type: integer
|
||||||
description: Number of bytes that a L1 user transaction has ([20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx + [2 bytes] depositAmountFloat16 + [2 bytes] amountFloat16 + [4 bytes] tokenId + [6 bytes] toIdx).
|
description: Number of bytes that a L1 user transaction has ([20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx + [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes] tokenId + [6 bytes] toIdx).
|
||||||
example: 72
|
example: 72
|
||||||
maxL1UserTx:
|
maxL1UserTx:
|
||||||
type: integer
|
type: integer
|
||||||
|
|||||||
@@ -2,6 +2,10 @@
|
|||||||
|
|
||||||
This is the main cli for the node
|
This is the main cli for the node
|
||||||
|
|
||||||
|
## Go version
|
||||||
|
|
||||||
|
The `hermez-node` has been tested with go version 1.14
|
||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
|
|
||||||
```
|
```
|
||||||
@@ -65,29 +69,64 @@ when running the coordinator in sync mode
|
|||||||
- The node requires a PostgreSQL database. The parameters of the server and
|
- The node requires a PostgreSQL database. The parameters of the server and
|
||||||
database must be set in the `PostgreSQL` section.
|
database must be set in the `PostgreSQL` section.
|
||||||
|
|
||||||
|
## Building
|
||||||
|
|
||||||
|
*All commands assume you are at the `cli/node` directory.*
|
||||||
|
|
||||||
|
Building the node requires using the packr utility to bundle the database
|
||||||
|
migrations inside the resulting binary. Install the packr utility with:
|
||||||
|
```
|
||||||
|
cd /tmp && go get -u github.com/gobuffalo/packr/v2/packr2 && cd -
|
||||||
|
```
|
||||||
|
|
||||||
|
Make sure your `$PATH` contains `$GOPATH/bin`, otherwise the packr utility will
|
||||||
|
not be found.
|
||||||
|
|
||||||
|
Now build the node executable:
|
||||||
|
```
|
||||||
|
cd ../../db && packr2 && cd -
|
||||||
|
go build .
|
||||||
|
cd ../../db && packr2 clean && cd -
|
||||||
|
```
|
||||||
|
|
||||||
|
The executable is `node`.
|
||||||
|
|
||||||
## Usage Examples
|
## Usage Examples
|
||||||
|
|
||||||
|
The following commands assume you have built the node previously. You can also
|
||||||
|
run the following examples by replacing `./node` with `go run .` and executing
|
||||||
|
them in the `cli/node` directory to build from source and run at the same time.
|
||||||
|
|
||||||
Run the node in mode synchronizer:
|
Run the node in mode synchronizer:
|
||||||
```
|
```
|
||||||
go run . --mode sync --cfg cfg.buidler.toml run
|
./node --mode sync --cfg cfg.buidler.toml run
|
||||||
```
|
```
|
||||||
|
|
||||||
Run the node in mode coordinator:
|
Run the node in mode coordinator:
|
||||||
```
|
```
|
||||||
go run . --mode coord --cfg cfg.buidler.toml run
|
./node --mode coord --cfg cfg.buidler.toml run
|
||||||
```
|
```
|
||||||
|
|
||||||
Import an ethereum private key into the keystore:
|
Import an ethereum private key into the keystore:
|
||||||
```
|
```
|
||||||
go run . --mode coord --cfg cfg.buidler.toml importkey --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
./node --mode coord --cfg cfg.buidler.toml importkey --privatekey 0x618b35096c477aab18b11a752be619f0023a539bb02dd6c813477a6211916cde
|
||||||
```
|
```
|
||||||
|
|
||||||
Generate a new BabyJubJub key pair:
|
Generate a new BabyJubJub key pair:
|
||||||
```
|
```
|
||||||
go run . --mode coord --cfg cfg.buidler.toml genbjj
|
./node --mode coord --cfg cfg.buidler.toml genbjj
|
||||||
```
|
```
|
||||||
|
|
||||||
Wipe the entier SQL database (this will destroy all synchronized and pool data):
|
Wipe the entier SQL database (this will destroy all synchronized and pool
|
||||||
|
data):
|
||||||
```
|
```
|
||||||
go run . --mode coord --cfg cfg.buidler.toml wipesql
|
./node --mode coord --cfg cfg.buidler.toml wipesql
|
||||||
|
```
|
||||||
|
|
||||||
|
Discard all synchronized blocks and associated state up to a given block
|
||||||
|
number. This command is useful in case the synchronizer reaches an invalid
|
||||||
|
state and you want to roll back a few blocks and try again (maybe with some
|
||||||
|
fixes in the code).
|
||||||
|
```
|
||||||
|
./node --mode coord --cfg cfg.buidler.toml discard --block 8061330
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -53,6 +53,8 @@ SendBatchBlocksMarginCheck = 1
|
|||||||
ProofServerPollInterval = "1s"
|
ProofServerPollInterval = "1s"
|
||||||
ForgeRetryInterval = "500ms"
|
ForgeRetryInterval = "500ms"
|
||||||
SyncRetryInterval = "1s"
|
SyncRetryInterval = "1s"
|
||||||
|
ForgeDelay = "10s"
|
||||||
|
ForgeNoTxsDelay = "0s"
|
||||||
|
|
||||||
[Coordinator.FeeAccount]
|
[Coordinator.FeeAccount]
|
||||||
Address = "0x56232B1c5B10038125Bc7345664B4AFD745bcF8E"
|
Address = "0x56232B1c5B10038125Bc7345664B4AFD745bcF8E"
|
||||||
@@ -83,16 +85,13 @@ MaxTx = 512
|
|||||||
NLevels = 32
|
NLevels = 32
|
||||||
|
|
||||||
[Coordinator.EthClient]
|
[Coordinator.EthClient]
|
||||||
ReceiptTimeout = "60s"
|
|
||||||
ReceiptLoopInterval = "500ms"
|
|
||||||
CheckLoopInterval = "500ms"
|
CheckLoopInterval = "500ms"
|
||||||
Attempts = 4
|
Attempts = 4
|
||||||
AttemptsDelay = "500ms"
|
AttemptsDelay = "500ms"
|
||||||
TxResendTimeout = "2m"
|
TxResendTimeout = "2m"
|
||||||
NoReuseNonce = false
|
NoReuseNonce = false
|
||||||
CallGasLimit = 300000
|
|
||||||
GasPriceDiv = 100
|
|
||||||
MaxGasPrice = "5000000000"
|
MaxGasPrice = "5000000000"
|
||||||
|
GasPriceIncPerc = 10
|
||||||
|
|
||||||
[Coordinator.EthClient.Keystore]
|
[Coordinator.EthClient.Keystore]
|
||||||
Path = "/tmp/iden3-test/hermez/ethkeystore"
|
Path = "/tmp/iden3-test/hermez/ethkeystore"
|
||||||
|
|||||||
@@ -11,6 +11,8 @@ import (
|
|||||||
"github.com/ethereum/go-ethereum/crypto"
|
"github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/hermeznetwork/hermez-node/config"
|
"github.com/hermeznetwork/hermez-node/config"
|
||||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||||
|
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||||
|
"github.com/hermeznetwork/hermez-node/db/l2db"
|
||||||
"github.com/hermeznetwork/hermez-node/log"
|
"github.com/hermeznetwork/hermez-node/log"
|
||||||
"github.com/hermeznetwork/hermez-node/node"
|
"github.com/hermeznetwork/hermez-node/node"
|
||||||
"github.com/hermeznetwork/tracerr"
|
"github.com/hermeznetwork/tracerr"
|
||||||
@@ -23,6 +25,7 @@ const (
|
|||||||
flagMode = "mode"
|
flagMode = "mode"
|
||||||
flagSK = "privatekey"
|
flagSK = "privatekey"
|
||||||
flagYes = "yes"
|
flagYes = "yes"
|
||||||
|
flagBlock = "block"
|
||||||
modeSync = "sync"
|
modeSync = "sync"
|
||||||
modeCoord = "coord"
|
modeCoord = "coord"
|
||||||
)
|
)
|
||||||
@@ -139,6 +142,47 @@ func cmdRun(c *cli.Context) error {
|
|||||||
return nil
|
return nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func cmdDiscard(c *cli.Context) error {
|
||||||
|
_cfg, err := parseCli(c)
|
||||||
|
if err != nil {
|
||||||
|
return tracerr.Wrap(fmt.Errorf("error parsing flags and config: %w", err))
|
||||||
|
}
|
||||||
|
cfg := _cfg.node
|
||||||
|
blockNum := c.Int64(flagBlock)
|
||||||
|
log.Infof("Discarding all blocks up to block %v...", blockNum)
|
||||||
|
|
||||||
|
db, err := dbUtils.InitSQLDB(
|
||||||
|
cfg.PostgreSQL.Port,
|
||||||
|
cfg.PostgreSQL.Host,
|
||||||
|
cfg.PostgreSQL.User,
|
||||||
|
cfg.PostgreSQL.Password,
|
||||||
|
cfg.PostgreSQL.Name,
|
||||||
|
)
|
||||||
|
if err != nil {
|
||||||
|
return tracerr.Wrap(fmt.Errorf("dbUtils.InitSQLDB: %w", err))
|
||||||
|
}
|
||||||
|
historyDB := historydb.NewHistoryDB(db, nil)
|
||||||
|
if err := historyDB.Reorg(blockNum); err != nil {
|
||||||
|
return tracerr.Wrap(fmt.Errorf("historyDB.Reorg: %w", err))
|
||||||
|
}
|
||||||
|
batchNum, err := historyDB.GetLastBatchNum()
|
||||||
|
if err != nil {
|
||||||
|
return tracerr.Wrap(fmt.Errorf("historyDB.GetLastBatchNum: %w", err))
|
||||||
|
}
|
||||||
|
l2DB := l2db.NewL2DB(
|
||||||
|
db,
|
||||||
|
cfg.Coordinator.L2DB.SafetyPeriod,
|
||||||
|
cfg.Coordinator.L2DB.MaxTxs,
|
||||||
|
cfg.Coordinator.L2DB.TTL.Duration,
|
||||||
|
nil,
|
||||||
|
)
|
||||||
|
if err := l2DB.Reorg(batchNum); err != nil {
|
||||||
|
return tracerr.Wrap(fmt.Errorf("l2DB.Reorg: %w", err))
|
||||||
|
}
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// Config is the configuration of the hermez node execution
|
// Config is the configuration of the hermez node execution
|
||||||
type Config struct {
|
type Config struct {
|
||||||
mode node.Mode
|
mode node.Mode
|
||||||
@@ -239,6 +283,18 @@ func main() {
|
|||||||
Usage: "Run the hermez-node in the indicated mode",
|
Usage: "Run the hermez-node in the indicated mode",
|
||||||
Action: cmdRun,
|
Action: cmdRun,
|
||||||
},
|
},
|
||||||
|
{
|
||||||
|
Name: "discard",
|
||||||
|
Aliases: []string{},
|
||||||
|
Usage: "Discard blocks up to a specified block number",
|
||||||
|
Action: cmdDiscard,
|
||||||
|
Flags: []cli.Flag{
|
||||||
|
&cli.Int64Flag{
|
||||||
|
Name: flagBlock,
|
||||||
|
Usage: "last block number to keep",
|
||||||
|
Required: false,
|
||||||
|
}},
|
||||||
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
err := app.Run(os.Args)
|
err := app.Run(os.Args)
|
||||||
|
|||||||
@@ -17,6 +17,11 @@ const AccountCreationAuthMsg = "I authorize this babyjubjub key for hermez rollu
|
|||||||
// EthMsgPrefix is the prefix for message signing at the Ethereum ecosystem
|
// EthMsgPrefix is the prefix for message signing at the Ethereum ecosystem
|
||||||
const EthMsgPrefix = "\x19Ethereum Signed Message:\n"
|
const EthMsgPrefix = "\x19Ethereum Signed Message:\n"
|
||||||
|
|
||||||
|
var (
|
||||||
|
// EmptyEthSignature is an ethereum signature of all zeroes
|
||||||
|
EmptyEthSignature = make([]byte, 65)
|
||||||
|
)
|
||||||
|
|
||||||
// AccountCreationAuth authorizations sent by users to the L2DB, to be used for
|
// AccountCreationAuth authorizations sent by users to the L2DB, to be used for
|
||||||
// account creations when necessary
|
// account creations when necessary
|
||||||
type AccountCreationAuth struct {
|
type AccountCreationAuth struct {
|
||||||
|
|||||||
@@ -93,23 +93,3 @@ func NewBatchData() *BatchData {
|
|||||||
Batch: Batch{},
|
Batch: Batch{},
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// BatchSync is a subset of Batch that contains fileds needed for the
|
|
||||||
// synchronizer and coordinator
|
|
||||||
// type BatchSync struct {
|
|
||||||
// BatchNum BatchNum `meddler:"batch_num"`
|
|
||||||
// EthBlockNum int64 `meddler:"eth_block_num"` // Ethereum block in which the batch is forged
|
|
||||||
// ForgerAddr ethCommon.Address `meddler:"forger_addr"`
|
|
||||||
// StateRoot *big.Int `meddler:"state_root,bigint"`
|
|
||||||
// SlotNum int64 `meddler:"slot_num"` // Slot in which the batch is forged
|
|
||||||
// }
|
|
||||||
//
|
|
||||||
// func NewBatchSync() *BatchSync {
|
|
||||||
// return &BatchSync{
|
|
||||||
// BatchNum: 0,
|
|
||||||
// EthBlockNum: 0,
|
|
||||||
// ForgerAddr: ethCommon.Address,
|
|
||||||
// StateRoot: big.NewInt(0),
|
|
||||||
// SlotNum: 0,
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|||||||
@@ -24,8 +24,8 @@ const (
|
|||||||
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes] compressedSignature
|
// RollupConstL1CoordinatorTotalBytes [4 bytes] token + [32 bytes] babyjub + [65 bytes] compressedSignature
|
||||||
RollupConstL1CoordinatorTotalBytes = 101
|
RollupConstL1CoordinatorTotalBytes = 101
|
||||||
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx +
|
// RollupConstL1UserTotalBytes [20 bytes] fromEthAddr + [32 bytes] fromBjj-compressed + [6 bytes] fromIdx +
|
||||||
// [2 bytes] depositAmountFloat16 + [2 bytes] amountFloat16 + [4 bytes] tokenId + [6 bytes] toIdx
|
// [5 bytes] depositAmountFloat40 + [5 bytes] amountFloat40 + [4 bytes] tokenId + [6 bytes] toIdx
|
||||||
RollupConstL1UserTotalBytes = 72
|
RollupConstL1UserTotalBytes = 78
|
||||||
// RollupConstMaxL1UserTx Maximum L1-user transactions allowed to be queued in a batch
|
// RollupConstMaxL1UserTx Maximum L1-user transactions allowed to be queued in a batch
|
||||||
RollupConstMaxL1UserTx = 128
|
RollupConstMaxL1UserTx = 128
|
||||||
// RollupConstMaxL1Tx Maximum L1 transactions allowed to be queued in a batch
|
// RollupConstMaxL1Tx Maximum L1 transactions allowed to be queued in a batch
|
||||||
|
|||||||
@@ -1,131 +0,0 @@
|
|||||||
// Package common Float16 provides methods to work with Hermez custom half float
|
|
||||||
// precision, 16 bits, codification internally called Float16 has been adopted
|
|
||||||
// to encode large integers. This is done in order to save bits when L2
|
|
||||||
// transactions are published.
|
|
||||||
//nolint:gomnd
|
|
||||||
package common
|
|
||||||
|
|
||||||
import (
|
|
||||||
"encoding/binary"
|
|
||||||
"errors"
|
|
||||||
"math/big"
|
|
||||||
|
|
||||||
"github.com/hermeznetwork/tracerr"
|
|
||||||
)
|
|
||||||
|
|
||||||
var (
|
|
||||||
// ErrRoundingLoss is used when converted big.Int to Float16 causes rounding loss
|
|
||||||
ErrRoundingLoss = errors.New("input value causes rounding loss")
|
|
||||||
)
|
|
||||||
|
|
||||||
// Float16 represents a float in a 16 bit format
|
|
||||||
type Float16 uint16
|
|
||||||
|
|
||||||
// Bytes return a byte array of length 2 with the Float16 value encoded in BigEndian
|
|
||||||
func (f16 Float16) Bytes() []byte {
|
|
||||||
var b [2]byte
|
|
||||||
binary.BigEndian.PutUint16(b[:], uint16(f16))
|
|
||||||
return b[:]
|
|
||||||
}
|
|
||||||
|
|
||||||
// Float16FromBytes returns a Float16 from a byte array of 2 bytes.
|
|
||||||
func Float16FromBytes(b []byte) *Float16 {
|
|
||||||
f16 := Float16(binary.BigEndian.Uint16(b[:2]))
|
|
||||||
return &f16
|
|
||||||
}
|
|
||||||
|
|
||||||
// BigInt converts the Float16 to a *big.Int integer
|
|
||||||
func (f16 *Float16) BigInt() *big.Int {
|
|
||||||
fl := int64(*f16)
|
|
||||||
|
|
||||||
m := big.NewInt(fl & 0x3FF)
|
|
||||||
e := big.NewInt(fl >> 11)
|
|
||||||
e5 := (fl >> 10) & 0x01
|
|
||||||
|
|
||||||
exp := big.NewInt(0).Exp(big.NewInt(10), e, nil)
|
|
||||||
res := m.Mul(m, exp)
|
|
||||||
|
|
||||||
if e5 != 0 && e.Cmp(big.NewInt(0)) != 0 {
|
|
||||||
res.Add(res, exp.Div(exp, big.NewInt(2)))
|
|
||||||
}
|
|
||||||
return res
|
|
||||||
}
|
|
||||||
|
|
||||||
// floorFix2Float converts a fix to a float, always rounding down
|
|
||||||
func floorFix2Float(_f *big.Int) Float16 {
|
|
||||||
zero := big.NewInt(0)
|
|
||||||
ten := big.NewInt(10)
|
|
||||||
e := int64(0)
|
|
||||||
|
|
||||||
m := big.NewInt(0)
|
|
||||||
m.Set(_f)
|
|
||||||
|
|
||||||
if m.Cmp(zero) == 0 {
|
|
||||||
return 0
|
|
||||||
}
|
|
||||||
|
|
||||||
s := big.NewInt(0).Rsh(m, 10)
|
|
||||||
|
|
||||||
for s.Cmp(zero) != 0 {
|
|
||||||
m.Div(m, ten)
|
|
||||||
s.Rsh(m, 10)
|
|
||||||
e++
|
|
||||||
}
|
|
||||||
|
|
||||||
return Float16(m.Int64() | e<<11)
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewFloat16 encodes a *big.Int integer as a Float16, returning error in
|
|
||||||
// case of loss during the encoding.
|
|
||||||
func NewFloat16(f *big.Int) (Float16, error) {
|
|
||||||
fl1 := floorFix2Float(f)
|
|
||||||
fi1 := fl1.BigInt()
|
|
||||||
fl2 := fl1 | 0x400
|
|
||||||
fi2 := fl2.BigInt()
|
|
||||||
|
|
||||||
m3 := (fl1 & 0x3FF) + 1
|
|
||||||
e3 := fl1 >> 11
|
|
||||||
|
|
||||||
if m3&0x400 == 0 {
|
|
||||||
m3 = 0x66
|
|
||||||
e3++
|
|
||||||
}
|
|
||||||
|
|
||||||
fl3 := m3 + e3<<11
|
|
||||||
fi3 := fl3.BigInt()
|
|
||||||
|
|
||||||
res := fl1
|
|
||||||
|
|
||||||
d := big.NewInt(0).Abs(fi1.Sub(fi1, f))
|
|
||||||
d2 := big.NewInt(0).Abs(fi2.Sub(fi2, f))
|
|
||||||
|
|
||||||
if d.Cmp(d2) == 1 {
|
|
||||||
res = fl2
|
|
||||||
d = d2
|
|
||||||
}
|
|
||||||
|
|
||||||
d3 := big.NewInt(0).Abs(fi3.Sub(fi3, f))
|
|
||||||
|
|
||||||
if d.Cmp(d3) == 1 {
|
|
||||||
res = fl3
|
|
||||||
}
|
|
||||||
|
|
||||||
// Do rounding check
|
|
||||||
if res.BigInt().Cmp(f) == 0 {
|
|
||||||
return res, nil
|
|
||||||
}
|
|
||||||
return res, tracerr.Wrap(ErrRoundingLoss)
|
|
||||||
}
|
|
||||||
|
|
||||||
// NewFloat16Floor encodes a big.Int integer as a Float16, rounding down in
|
|
||||||
// case of loss during the encoding.
|
|
||||||
func NewFloat16Floor(f *big.Int) Float16 {
|
|
||||||
fl1 := floorFix2Float(f)
|
|
||||||
fl2 := fl1 | 0x400
|
|
||||||
fi2 := fl2.BigInt()
|
|
||||||
|
|
||||||
if fi2.Cmp(f) < 1 {
|
|
||||||
return fl2
|
|
||||||
}
|
|
||||||
return fl1
|
|
||||||
}
|
|
||||||
@@ -1,132 +0,0 @@
|
|||||||
package common
|
|
||||||
|
|
||||||
import (
|
|
||||||
"math/big"
|
|
||||||
"testing"
|
|
||||||
|
|
||||||
"github.com/hermeznetwork/tracerr"
|
|
||||||
"github.com/stretchr/testify/assert"
|
|
||||||
)
|
|
||||||
|
|
||||||
func TestConversions(t *testing.T) {
|
|
||||||
testVector := map[Float16]string{
|
|
||||||
0x307B: "123000000",
|
|
||||||
0x1DC6: "454500",
|
|
||||||
0xFFFF: "10235000000000000000000000000000000",
|
|
||||||
0x0000: "0",
|
|
||||||
0x0400: "0",
|
|
||||||
0x0001: "1",
|
|
||||||
0x0401: "1",
|
|
||||||
0x0800: "0",
|
|
||||||
0x0c00: "5",
|
|
||||||
0x0801: "10",
|
|
||||||
0x0c01: "15",
|
|
||||||
}
|
|
||||||
|
|
||||||
for test := range testVector {
|
|
||||||
fix := test.BigInt()
|
|
||||||
|
|
||||||
assert.Equal(t, fix.String(), testVector[test])
|
|
||||||
|
|
||||||
bi := big.NewInt(0)
|
|
||||||
bi.SetString(testVector[test], 10)
|
|
||||||
|
|
||||||
fl, err := NewFloat16(bi)
|
|
||||||
assert.Equal(t, nil, err)
|
|
||||||
|
|
||||||
fx2 := fl.BigInt()
|
|
||||||
assert.Equal(t, fx2.String(), testVector[test])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestFloorFix2Float(t *testing.T) {
|
|
||||||
testVector := map[string]Float16{
|
|
||||||
"87999990000000000": 0x776f,
|
|
||||||
"87950000000000001": 0x776f,
|
|
||||||
"87950000000000000": 0x776f,
|
|
||||||
"87949999999999999": 0x736f,
|
|
||||||
}
|
|
||||||
|
|
||||||
for test := range testVector {
|
|
||||||
bi := big.NewInt(0)
|
|
||||||
bi.SetString(test, 10)
|
|
||||||
|
|
||||||
testFloat := NewFloat16Floor(bi)
|
|
||||||
|
|
||||||
assert.Equal(t, testFloat, testVector[test])
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestConversionLosses(t *testing.T) {
|
|
||||||
a := big.NewInt(1000)
|
|
||||||
b, err := NewFloat16(a)
|
|
||||||
assert.Equal(t, nil, err)
|
|
||||||
c := b.BigInt()
|
|
||||||
assert.Equal(t, c, a)
|
|
||||||
|
|
||||||
a = big.NewInt(1024)
|
|
||||||
b, err = NewFloat16(a)
|
|
||||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
|
||||||
c = b.BigInt()
|
|
||||||
assert.NotEqual(t, c, a)
|
|
||||||
|
|
||||||
a = big.NewInt(32767)
|
|
||||||
b, err = NewFloat16(a)
|
|
||||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
|
||||||
c = b.BigInt()
|
|
||||||
assert.NotEqual(t, c, a)
|
|
||||||
|
|
||||||
a = big.NewInt(32768)
|
|
||||||
b, err = NewFloat16(a)
|
|
||||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
|
||||||
c = b.BigInt()
|
|
||||||
assert.NotEqual(t, c, a)
|
|
||||||
|
|
||||||
a = big.NewInt(65536000)
|
|
||||||
b, err = NewFloat16(a)
|
|
||||||
assert.Equal(t, ErrRoundingLoss, tracerr.Unwrap(err))
|
|
||||||
c = b.BigInt()
|
|
||||||
assert.NotEqual(t, c, a)
|
|
||||||
}
|
|
||||||
|
|
||||||
func BenchmarkFloat16(b *testing.B) {
|
|
||||||
newBigInt := func(s string) *big.Int {
|
|
||||||
bigInt, ok := new(big.Int).SetString(s, 10)
|
|
||||||
if !ok {
|
|
||||||
panic("Bad big int")
|
|
||||||
}
|
|
||||||
return bigInt
|
|
||||||
}
|
|
||||||
type pair struct {
|
|
||||||
Float16 Float16
|
|
||||||
BigInt *big.Int
|
|
||||||
}
|
|
||||||
testVector := []pair{
|
|
||||||
{0x307B, newBigInt("123000000")},
|
|
||||||
{0x1DC6, newBigInt("454500")},
|
|
||||||
{0xFFFF, newBigInt("10235000000000000000000000000000000")},
|
|
||||||
{0x0000, newBigInt("0")},
|
|
||||||
{0x0400, newBigInt("0")},
|
|
||||||
{0x0001, newBigInt("1")},
|
|
||||||
{0x0401, newBigInt("1")},
|
|
||||||
{0x0800, newBigInt("0")},
|
|
||||||
{0x0c00, newBigInt("5")},
|
|
||||||
{0x0801, newBigInt("10")},
|
|
||||||
{0x0c01, newBigInt("15")},
|
|
||||||
}
|
|
||||||
b.Run("floorFix2Float()", func(b *testing.B) {
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
NewFloat16Floor(testVector[i%len(testVector)].BigInt)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
b.Run("NewFloat16()", func(b *testing.B) {
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
_, _ = NewFloat16(testVector[i%len(testVector)].BigInt)
|
|
||||||
}
|
|
||||||
})
|
|
||||||
b.Run("Float16.BigInt()", func(b *testing.B) {
|
|
||||||
for i := 0; i < b.N; i++ {
|
|
||||||
testVector[i%len(testVector)].Float16.BigInt()
|
|
||||||
}
|
|
||||||
})
|
|
||||||
}
|
|
||||||
103
common/float40.go
Normal file
103
common/float40.go
Normal file
@@ -0,0 +1,103 @@
|
|||||||
|
// Package common Float40 provides methods to work with Hermez custom half
|
||||||
|
// float precision, 40 bits, codification internally called Float40 has been
|
||||||
|
// adopted to encode large integers. This is done in order to save bits when L2
|
||||||
|
// transactions are published.
|
||||||
|
//nolint:gomnd
|
||||||
|
package common
|
||||||
|
|
||||||
|
import (
|
||||||
|
"encoding/binary"
|
||||||
|
"errors"
|
||||||
|
"math/big"
|
||||||
|
|
||||||
|
"github.com/hermeznetwork/tracerr"
|
||||||
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// maxFloat40Value is the maximum value that the Float40 can have
|
||||||
|
// (40 bits: maxFloat40Value=2**40-1)
|
||||||
|
maxFloat40Value = 0xffffffffff
|
||||||
|
// Float40BytesLength defines the length of the Float40 values
|
||||||
|
// represented as byte arrays
|
||||||
|
Float40BytesLength = 5
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrFloat40Overflow is used when a given Float40 overflows the
|
||||||
|
// maximum capacity of the Float40 (2**40-1)
|
||||||
|
ErrFloat40Overflow = errors.New("Float40 overflow, max value: 2**40 -1")
|
||||||
|
// ErrFloat40E31 is used when the e > 31 when trying to convert a
|
||||||
|
// *big.Int to Float40
|
||||||
|
ErrFloat40E31 = errors.New("Float40 error, e > 31")
|
||||||
|
// ErrFloat40NotEnoughPrecission is used when the given *big.Int can
|
||||||
|
// not be represented as Float40 due not enough precission
|
||||||
|
ErrFloat40NotEnoughPrecission = errors.New("Float40 error, not enough precission")
|
||||||
|
)
|
||||||
|
|
||||||
|
// Float40 represents a float in a 64 bit format
|
||||||
|
type Float40 uint64
|
||||||
|
|
||||||
|
// Bytes return a byte array of length 5 with the Float40 value encoded in
|
||||||
|
// BigEndian
|
||||||
|
func (f40 Float40) Bytes() ([]byte, error) {
|
||||||
|
if f40 > maxFloat40Value {
|
||||||
|
return []byte{}, tracerr.Wrap(ErrFloat40Overflow)
|
||||||
|
}
|
||||||
|
|
||||||
|
var f40Bytes [8]byte
|
||||||
|
binary.BigEndian.PutUint64(f40Bytes[:], uint64(f40))
|
||||||
|
var b [5]byte
|
||||||
|
copy(b[:], f40Bytes[3:])
|
||||||
|
return b[:], nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// Float40FromBytes returns a Float40 from a byte array of 5 bytes in Bigendian
|
||||||
|
// representation.
|
||||||
|
func Float40FromBytes(b []byte) Float40 {
|
||||||
|
var f40Bytes [8]byte
|
||||||
|
copy(f40Bytes[3:], b[:])
|
||||||
|
f40 := binary.BigEndian.Uint64(f40Bytes[:])
|
||||||
|
return Float40(f40)
|
||||||
|
}
|
||||||
|
|
||||||
|
// BigInt converts the Float40 to a *big.Int v, where v = m * 10^e, being:
|
||||||
|
// [ e | m ]
|
||||||
|
// [ 5 bits | 35 bits ]
|
||||||
|
func (f40 Float40) BigInt() (*big.Int, error) {
|
||||||
|
// take the 5 used bytes (FF * 5)
|
||||||
|
var f40Uint64 uint64 = uint64(f40) & 0x00_00_00_FF_FF_FF_FF_FF
|
||||||
|
f40Bytes, err := f40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
e := f40Bytes[0] & 0xF8 >> 3 // take first 5 bits
|
||||||
|
m := f40Uint64 & 0x07_FF_FF_FF_FF // take the others 35 bits
|
||||||
|
|
||||||
|
exp := new(big.Int).Exp(big.NewInt(10), big.NewInt(int64(e)), nil)
|
||||||
|
r := new(big.Int).Mul(big.NewInt(int64(m)), exp)
|
||||||
|
return r, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
// NewFloat40 encodes a *big.Int integer as a Float40, returning error in case
|
||||||
|
// of loss during the encoding.
|
||||||
|
func NewFloat40(f *big.Int) (Float40, error) {
|
||||||
|
m := f
|
||||||
|
e := big.NewInt(0)
|
||||||
|
zero := big.NewInt(0)
|
||||||
|
ten := big.NewInt(10)
|
||||||
|
thres := big.NewInt(0x08_00_00_00_00)
|
||||||
|
for new(big.Int).Mod(m, ten).Cmp(zero) == 0 && m.Cmp(thres) >= 0 {
|
||||||
|
m = new(big.Int).Div(m, ten)
|
||||||
|
e = new(big.Int).Add(e, big.NewInt(1))
|
||||||
|
}
|
||||||
|
if e.Int64() > 31 {
|
||||||
|
return 0, ErrFloat40E31
|
||||||
|
}
|
||||||
|
if m.Cmp(thres) >= 0 {
|
||||||
|
return 0, ErrFloat40NotEnoughPrecission
|
||||||
|
}
|
||||||
|
r := new(big.Int).Add(m,
|
||||||
|
new(big.Int).Mul(e, thres))
|
||||||
|
return Float40(r.Uint64()), nil
|
||||||
|
}
|
||||||
95
common/float40_test.go
Normal file
95
common/float40_test.go
Normal file
@@ -0,0 +1,95 @@
|
|||||||
|
package common
|
||||||
|
|
||||||
|
import (
|
||||||
|
"math/big"
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
"github.com/stretchr/testify/assert"
|
||||||
|
"github.com/stretchr/testify/require"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestConversionsFloat40(t *testing.T) {
|
||||||
|
testVector := map[Float40]string{
|
||||||
|
6*0x800000000 + 123: "123000000",
|
||||||
|
2*0x800000000 + 4545: "454500",
|
||||||
|
30*0x800000000 + 10235: "10235000000000000000000000000000000",
|
||||||
|
0x000000000: "0",
|
||||||
|
0x800000000: "0",
|
||||||
|
0x0001: "1",
|
||||||
|
0x0401: "1025",
|
||||||
|
0x800000000 + 1: "10",
|
||||||
|
0xFFFFFFFFFF: "343597383670000000000000000000000000000000",
|
||||||
|
}
|
||||||
|
|
||||||
|
for test := range testVector {
|
||||||
|
fix, err := test.BigInt()
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, fix.String(), testVector[test])
|
||||||
|
|
||||||
|
bi, ok := new(big.Int).SetString(testVector[test], 10)
|
||||||
|
require.True(t, ok)
|
||||||
|
|
||||||
|
fl, err := NewFloat40(bi)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
fx2, err := fl.BigInt()
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, fx2.String(), testVector[test])
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestExpectError(t *testing.T) {
|
||||||
|
testVector := map[string]error{
|
||||||
|
"9922334455000000000000000000000000000000": nil,
|
||||||
|
"9922334455000000000000000000000000000001": ErrFloat40NotEnoughPrecission,
|
||||||
|
"9922334454999999999999999999999999999999": ErrFloat40NotEnoughPrecission,
|
||||||
|
"42949672950000000000000000000000000000000": nil,
|
||||||
|
"99223344556573838487575": ErrFloat40NotEnoughPrecission,
|
||||||
|
"992233445500000000000000000000000000000000": ErrFloat40E31,
|
||||||
|
"343597383670000000000000000000000000000000": nil,
|
||||||
|
"343597383680000000000000000000000000000000": ErrFloat40NotEnoughPrecission,
|
||||||
|
"343597383690000000000000000000000000000000": ErrFloat40NotEnoughPrecission,
|
||||||
|
"343597383700000000000000000000000000000000": ErrFloat40E31,
|
||||||
|
}
|
||||||
|
for test := range testVector {
|
||||||
|
bi, ok := new(big.Int).SetString(test, 10)
|
||||||
|
require.True(t, ok)
|
||||||
|
_, err := NewFloat40(bi)
|
||||||
|
assert.Equal(t, testVector[test], err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func BenchmarkFloat40(b *testing.B) {
|
||||||
|
newBigInt := func(s string) *big.Int {
|
||||||
|
bigInt, ok := new(big.Int).SetString(s, 10)
|
||||||
|
if !ok {
|
||||||
|
panic("Can not convert string to *big.Int")
|
||||||
|
}
|
||||||
|
return bigInt
|
||||||
|
}
|
||||||
|
type pair struct {
|
||||||
|
Float40 Float40
|
||||||
|
BigInt *big.Int
|
||||||
|
}
|
||||||
|
testVector := []pair{
|
||||||
|
{6*0x800000000 + 123, newBigInt("123000000")},
|
||||||
|
{2*0x800000000 + 4545, newBigInt("454500")},
|
||||||
|
{30*0x800000000 + 10235, newBigInt("10235000000000000000000000000000000")},
|
||||||
|
{0x000000000, newBigInt("0")},
|
||||||
|
{0x800000000, newBigInt("0")},
|
||||||
|
{0x0001, newBigInt("1")},
|
||||||
|
{0x0401, newBigInt("1025")},
|
||||||
|
{0x800000000 + 1, newBigInt("10")},
|
||||||
|
{0xFFFFFFFFFF, newBigInt("343597383670000000000000000000000000000000")},
|
||||||
|
}
|
||||||
|
b.Run("NewFloat40()", func(b *testing.B) {
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
_, _ = NewFloat40(testVector[i%len(testVector)].BigInt)
|
||||||
|
}
|
||||||
|
})
|
||||||
|
b.Run("Float40.BigInt()", func(b *testing.B) {
|
||||||
|
for i := 0; i < b.N; i++ {
|
||||||
|
_, _ = testVector[i%len(testVector)].Float40.BigInt()
|
||||||
|
}
|
||||||
|
})
|
||||||
|
}
|
||||||
@@ -11,18 +11,11 @@ import (
|
|||||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
|
||||||
// L1UserTxBytesLen is the length of the byte array that represents the L1Tx
|
|
||||||
L1UserTxBytesLen = 72
|
|
||||||
// L1CoordinatorTxBytesLen is the length of the byte array that represents the L1CoordinatorTx
|
|
||||||
L1CoordinatorTxBytesLen = 101
|
|
||||||
)
|
|
||||||
|
|
||||||
// L1Tx is a struct that represents a L1 tx
|
// L1Tx is a struct that represents a L1 tx
|
||||||
type L1Tx struct {
|
type L1Tx struct {
|
||||||
// Stored in DB: mandatory fileds
|
// Stored in DB: mandatory fileds
|
||||||
|
|
||||||
// TxID (12 bytes) for L1Tx is:
|
// TxID (32 bytes) for L1Tx is the Keccak256 (ethereum) hash of:
|
||||||
// bytes: | 1 | 8 | 2 | 1 |
|
// bytes: | 1 | 8 | 2 | 1 |
|
||||||
// values: | type | ToForgeL1TxsNum | Position | 0 (padding) |
|
// values: | type | ToForgeL1TxsNum | Position | 0 (padding) |
|
||||||
// where type:
|
// where type:
|
||||||
@@ -179,45 +172,38 @@ func (tx L1Tx) Tx() Tx {
|
|||||||
// [ 8 bits ] empty (userFee) // 1 byte
|
// [ 8 bits ] empty (userFee) // 1 byte
|
||||||
// [ 40 bits ] empty (nonce) // 5 bytes
|
// [ 40 bits ] empty (nonce) // 5 bytes
|
||||||
// [ 32 bits ] tokenID // 4 bytes
|
// [ 32 bits ] tokenID // 4 bytes
|
||||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
|
||||||
// [ 48 bits ] toIdx // 6 bytes
|
// [ 48 bits ] toIdx // 6 bytes
|
||||||
// [ 48 bits ] fromIdx // 6 bytes
|
// [ 48 bits ] fromIdx // 6 bytes
|
||||||
// [ 16 bits ] chainId // 2 bytes
|
// [ 16 bits ] chainId // 2 bytes
|
||||||
// [ 32 bits ] empty (signatureConstant) // 4 bytes
|
// [ 32 bits ] empty (signatureConstant) // 4 bytes
|
||||||
// Total bits compressed data: 241 bits // 31 bytes in *big.Int representation
|
// Total bits compressed data: 225 bits // 29 bytes in *big.Int representation
|
||||||
func (tx L1Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
func (tx L1Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
var b [29]byte
|
||||||
if err != nil {
|
|
||||||
return nil, tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
var b [31]byte
|
|
||||||
// b[0:7] empty: no ToBJJSign, no fee, no nonce
|
// b[0:7] empty: no ToBJJSign, no fee, no nonce
|
||||||
copy(b[7:11], tx.TokenID.Bytes())
|
copy(b[7:11], tx.TokenID.Bytes())
|
||||||
copy(b[11:13], amountFloat16.Bytes())
|
|
||||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[13:19], toIdxBytes[:])
|
copy(b[11:17], toIdxBytes[:])
|
||||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[19:25], fromIdxBytes[:])
|
copy(b[17:23], fromIdxBytes[:])
|
||||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||||
copy(b[27:31], SignatureConstantBytes[:])
|
copy(b[25:29], SignatureConstantBytes[:])
|
||||||
|
|
||||||
bi := new(big.Int).SetBytes(b[:])
|
bi := new(big.Int).SetBytes(b[:])
|
||||||
return bi, nil
|
return bi, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
// BytesDataAvailability encodes a L1Tx into []byte for the Data Availability
|
// BytesDataAvailability encodes a L1Tx into []byte for the Data Availability
|
||||||
// [ fromIdx | toIdx | amountFloat16 | Fee ]
|
// [ fromIdx | toIdx | amountFloat40 | Fee ]
|
||||||
func (tx *L1Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
func (tx *L1Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||||
idxLen := nLevels / 8 //nolint:gomnd
|
idxLen := nLevels / 8 //nolint:gomnd
|
||||||
|
|
||||||
b := make([]byte, ((nLevels*2)+16+8)/8) //nolint:gomnd
|
b := make([]byte, ((nLevels*2)+40+8)/8) //nolint:gomnd
|
||||||
|
|
||||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -231,13 +217,17 @@ func (tx *L1Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
|||||||
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
||||||
|
|
||||||
if tx.EffectiveAmount != nil {
|
if tx.EffectiveAmount != nil {
|
||||||
amountFloat16, err := NewFloat16(tx.EffectiveAmount)
|
amountFloat40, err := NewFloat40(tx.EffectiveAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[idxLen*2:idxLen*2+2], amountFloat16.Bytes())
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
// fee = 0 (as is L1Tx) b[10:11]
|
copy(b[idxLen*2:idxLen*2+Float40BytesLength], amountFloat40Bytes)
|
||||||
|
}
|
||||||
|
// fee = 0 (as is L1Tx)
|
||||||
return b[:], nil
|
return b[:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -247,7 +237,7 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
|||||||
|
|
||||||
fromIdxBytes := b[0:idxLen]
|
fromIdxBytes := b[0:idxLen]
|
||||||
toIdxBytes := b[idxLen : idxLen*2]
|
toIdxBytes := b[idxLen : idxLen*2]
|
||||||
amountBytes := b[idxLen*2 : idxLen*2+2]
|
amountBytes := b[idxLen*2 : idxLen*2+Float40BytesLength]
|
||||||
|
|
||||||
l1tx := L1Tx{}
|
l1tx := L1Tx{}
|
||||||
fromIdx, err := IdxFromBytes(ethCommon.LeftPadBytes(fromIdxBytes, 6))
|
fromIdx, err := IdxFromBytes(ethCommon.LeftPadBytes(fromIdxBytes, 6))
|
||||||
@@ -260,8 +250,8 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
|||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
l1tx.ToIdx = toIdx
|
l1tx.ToIdx = toIdx
|
||||||
l1tx.EffectiveAmount = Float16FromBytes(amountBytes).BigInt()
|
l1tx.EffectiveAmount, err = Float40FromBytes(amountBytes).BigInt()
|
||||||
return &l1tx, nil
|
return &l1tx, err
|
||||||
}
|
}
|
||||||
|
|
||||||
// BytesGeneric returns the generic representation of a L1Tx. This method is
|
// BytesGeneric returns the generic representation of a L1Tx. This method is
|
||||||
@@ -269,7 +259,7 @@ func L1TxFromDataAvailability(b []byte, nLevels uint32) (*L1Tx, error) {
|
|||||||
// the L1TxData for the ZKInputs (at the HashGlobalInputs), using this method
|
// the L1TxData for the ZKInputs (at the HashGlobalInputs), using this method
|
||||||
// for L1CoordinatorTxs & L1UserTxs (for the ZKInputs case).
|
// for L1CoordinatorTxs & L1UserTxs (for the ZKInputs case).
|
||||||
func (tx *L1Tx) BytesGeneric() ([]byte, error) {
|
func (tx *L1Tx) BytesGeneric() ([]byte, error) {
|
||||||
var b [L1UserTxBytesLen]byte
|
var b [RollupConstL1UserTotalBytes]byte
|
||||||
copy(b[0:20], tx.FromEthAddr.Bytes())
|
copy(b[0:20], tx.FromEthAddr.Bytes())
|
||||||
if tx.FromBJJ != EmptyBJJComp {
|
if tx.FromBJJ != EmptyBJJComp {
|
||||||
pkCompL := tx.FromBJJ
|
pkCompL := tx.FromBJJ
|
||||||
@@ -281,22 +271,33 @@ func (tx *L1Tx) BytesGeneric() ([]byte, error) {
|
|||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[52:58], fromIdxBytes[:])
|
copy(b[52:58], fromIdxBytes[:])
|
||||||
depositAmountFloat16, err := NewFloat16(tx.DepositAmount)
|
|
||||||
|
depositAmountFloat40, err := NewFloat40(tx.DepositAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[58:60], depositAmountFloat16.Bytes())
|
depositAmountFloat40Bytes, err := depositAmountFloat40.Bytes()
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[60:62], amountFloat16.Bytes())
|
copy(b[58:63], depositAmountFloat40Bytes)
|
||||||
copy(b[62:66], tx.TokenID.Bytes())
|
|
||||||
|
amountFloat40, err := NewFloat40(tx.Amount)
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
copy(b[63:68], amountFloat40Bytes)
|
||||||
|
|
||||||
|
copy(b[68:72], tx.TokenID.Bytes())
|
||||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[66:72], toIdxBytes[:])
|
copy(b[72:78], toIdxBytes[:])
|
||||||
return b[:], nil
|
return b[:], nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -313,7 +314,7 @@ func (tx *L1Tx) BytesCoordinatorTx(compressedSignatureBytes []byte) ([]byte, err
|
|||||||
if tx.UserOrigin {
|
if tx.UserOrigin {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("Can not calculate BytesCoordinatorTx() for a L1UserTx"))
|
return nil, tracerr.Wrap(fmt.Errorf("Can not calculate BytesCoordinatorTx() for a L1UserTx"))
|
||||||
}
|
}
|
||||||
var b [L1CoordinatorTxBytesLen]byte
|
var b [RollupConstL1CoordinatorTotalBytes]byte
|
||||||
v := compressedSignatureBytes[64]
|
v := compressedSignatureBytes[64]
|
||||||
s := compressedSignatureBytes[32:64]
|
s := compressedSignatureBytes[32:64]
|
||||||
r := compressedSignatureBytes[0:32]
|
r := compressedSignatureBytes[0:32]
|
||||||
@@ -329,7 +330,7 @@ func (tx *L1Tx) BytesCoordinatorTx(compressedSignatureBytes []byte) ([]byte, err
|
|||||||
|
|
||||||
// L1UserTxFromBytes decodes a L1Tx from []byte
|
// L1UserTxFromBytes decodes a L1Tx from []byte
|
||||||
func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
||||||
if len(b) != L1UserTxBytesLen {
|
if len(b) != RollupConstL1UserTotalBytes {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d", 68, len(b)))
|
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1Tx bytes, expected length %d, current: %d", 68, len(b)))
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -347,13 +348,19 @@ func L1UserTxFromBytes(b []byte) (*L1Tx, error) {
|
|||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
tx.FromIdx = fromIdx
|
tx.FromIdx = fromIdx
|
||||||
tx.DepositAmount = Float16FromBytes(b[58:60]).BigInt()
|
tx.DepositAmount, err = Float40FromBytes(b[58:63]).BigInt()
|
||||||
tx.Amount = Float16FromBytes(b[60:62]).BigInt()
|
|
||||||
tx.TokenID, err = TokenIDFromBytes(b[62:66])
|
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
tx.ToIdx, err = IdxFromBytes(b[66:72])
|
tx.Amount, err = Float40FromBytes(b[63:68]).BigInt()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
tx.TokenID, err = TokenIDFromBytes(b[68:72])
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
tx.ToIdx, err = IdxFromBytes(b[72:78])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -368,7 +375,7 @@ func signHash(data []byte) []byte {
|
|||||||
|
|
||||||
// L1CoordinatorTxFromBytes decodes a L1Tx from []byte
|
// L1CoordinatorTxFromBytes decodes a L1Tx from []byte
|
||||||
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx, error) {
|
func L1CoordinatorTxFromBytes(b []byte, chainID *big.Int, hermezAddress ethCommon.Address) (*L1Tx, error) {
|
||||||
if len(b) != L1CoordinatorTxBytesLen {
|
if len(b) != RollupConstL1CoordinatorTotalBytes {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d", 101, len(b)))
|
return nil, tracerr.Wrap(fmt.Errorf("Can not parse L1CoordinatorTx bytes, expected length %d, current: %d", 101, len(b)))
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -50,64 +50,110 @@ func TestNewL1CoordinatorTx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func TestL1TxCompressedData(t *testing.T) {
|
func TestL1TxCompressedData(t *testing.T) {
|
||||||
|
// test vectors values generated from javascript implementation (using
|
||||||
|
// PoolL2Tx values)
|
||||||
|
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||||
|
require.True(t, ok)
|
||||||
tx := L1Tx{
|
tx := L1Tx{
|
||||||
FromIdx: 2,
|
FromIdx: (1 << 48) - 1,
|
||||||
ToIdx: 3,
|
ToIdx: (1 << 48) - 1,
|
||||||
Amount: big.NewInt(4),
|
Amount: amount,
|
||||||
TokenID: 5,
|
TokenID: (1 << 32) - 1,
|
||||||
}
|
}
|
||||||
chainID := uint16(0)
|
txCompressedData, err := tx.TxCompressedData(uint16((1 << 16) - 1))
|
||||||
txCompressedData, err := tx.TxCompressedData(chainID)
|
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
expectedStr := "ffffffffffffffffffffffffffffffffffffc60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
// test vector value generated from javascript implementation
|
tx = L1Tx{
|
||||||
expectedStr := "7307597389635308713748674793997299267459594577423"
|
FromIdx: 0,
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
ToIdx: 0,
|
||||||
assert.Equal(t, "0500040000000000030000000000020000c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
Amount: big.NewInt(0),
|
||||||
|
TokenID: 0,
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||||
|
assert.NoError(t, err)
|
||||||
|
expectedStr = "c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
|
amount, ok = new(big.Int).SetString("63000000000000000", 10)
|
||||||
|
require.True(t, ok)
|
||||||
|
tx = L1Tx{
|
||||||
|
FromIdx: 324,
|
||||||
|
ToIdx: 256,
|
||||||
|
Amount: amount,
|
||||||
|
TokenID: 123,
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(1))
|
||||||
|
assert.NoError(t, err)
|
||||||
|
expectedStr = "7b0000000001000000000001440001c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
|
tx = L1Tx{
|
||||||
|
FromIdx: 1,
|
||||||
|
ToIdx: 2,
|
||||||
|
TokenID: 3,
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||||
|
assert.NoError(t, err)
|
||||||
|
expectedStr = "030000000000020000000000010000c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestBytesDataAvailability(t *testing.T) {
|
func TestBytesDataAvailability(t *testing.T) {
|
||||||
|
// test vectors values generated from javascript implementation
|
||||||
|
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||||
|
require.True(t, ok)
|
||||||
tx := L1Tx{
|
tx := L1Tx{
|
||||||
FromIdx: 2,
|
ToIdx: (1 << 16) - 1,
|
||||||
ToIdx: 3,
|
FromIdx: (1 << 16) - 1,
|
||||||
Amount: big.NewInt(4),
|
EffectiveAmount: amount,
|
||||||
TokenID: 5,
|
|
||||||
}
|
}
|
||||||
txCompressedData, err := tx.BytesDataAvailability(32)
|
txCompressedData, err := tx.BytesDataAvailability(16)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0000000200000003000000", hex.EncodeToString(txCompressedData))
|
assert.Equal(t, "ffffffffffffffffff00", hex.EncodeToString(txCompressedData))
|
||||||
|
l1tx, err := L1TxFromDataAvailability(txCompressedData, 16)
|
||||||
tx = L1Tx{
|
|
||||||
FromIdx: 2,
|
|
||||||
ToIdx: 3,
|
|
||||||
EffectiveAmount: big.NewInt(4),
|
|
||||||
TokenID: 5,
|
|
||||||
}
|
|
||||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
|
||||||
assert.NoError(t, err)
|
|
||||||
assert.Equal(t, "0000000200000003000400", hex.EncodeToString(txCompressedData))
|
|
||||||
}
|
|
||||||
|
|
||||||
func TestL1TxFromDataAvailability(t *testing.T) {
|
|
||||||
tx := L1Tx{
|
|
||||||
FromIdx: 2,
|
|
||||||
ToIdx: 3,
|
|
||||||
Amount: big.NewInt(4),
|
|
||||||
}
|
|
||||||
txCompressedData, err := tx.BytesDataAvailability(32)
|
|
||||||
assert.NoError(t, err)
|
|
||||||
l1tx, err := L1TxFromDataAvailability(txCompressedData, 32)
|
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||||
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||||
|
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||||
|
|
||||||
tx = L1Tx{
|
tx = L1Tx{
|
||||||
FromIdx: 2,
|
ToIdx: (1 << 32) - 1,
|
||||||
ToIdx: 3,
|
FromIdx: (1 << 32) - 1,
|
||||||
EffectiveAmount: big.NewInt(4),
|
EffectiveAmount: amount,
|
||||||
}
|
}
|
||||||
txCompressedData, err = tx.BytesDataAvailability(32)
|
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, "ffffffffffffffffffffffffff00", hex.EncodeToString(txCompressedData))
|
||||||
|
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||||
|
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||||
|
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||||
|
|
||||||
|
tx = L1Tx{
|
||||||
|
ToIdx: 0,
|
||||||
|
FromIdx: 0,
|
||||||
|
EffectiveAmount: big.NewInt(0),
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, "0000000000000000000000000000", hex.EncodeToString(txCompressedData))
|
||||||
|
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||||
|
assert.Equal(t, tx.ToIdx, l1tx.ToIdx)
|
||||||
|
assert.Equal(t, tx.EffectiveAmount, l1tx.EffectiveAmount)
|
||||||
|
|
||||||
|
tx = L1Tx{
|
||||||
|
ToIdx: 635,
|
||||||
|
FromIdx: 296,
|
||||||
|
EffectiveAmount: big.NewInt(1000000000000000000),
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.BytesDataAvailability(32)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
assert.Equal(t, "000001280000027b42540be40000", hex.EncodeToString(txCompressedData))
|
||||||
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
l1tx, err = L1TxFromDataAvailability(txCompressedData, 32)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
assert.Equal(t, tx.FromIdx, l1tx.FromIdx)
|
||||||
@@ -172,12 +218,10 @@ func TestL1TxByteParsersCompatibility(t *testing.T) {
|
|||||||
UserOrigin: true,
|
UserOrigin: true,
|
||||||
}
|
}
|
||||||
|
|
||||||
expected, err := utils.HexDecode("85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a000001c638db8be880f00020039c0000053cb88d")
|
|
||||||
require.NoError(t, err)
|
|
||||||
|
|
||||||
encodedData, err := l1Tx.BytesUser()
|
encodedData, err := l1Tx.BytesUser()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, expected, encodedData)
|
expected := "85dab5b9e2e361d0c208d77be90efcc0439b0a530dd02deb2c81068e7a0f7e327df80b4ab79ee1f41a7def613e73a20c32eece5a000001c638db52540be400459682f0000020039c0000053cb88d"
|
||||||
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestL1CoordinatorTxByteParsers(t *testing.T) {
|
func TestL1CoordinatorTxByteParsers(t *testing.T) {
|
||||||
|
|||||||
@@ -89,11 +89,15 @@ func (tx L2Tx) CalculateTxID() ([TxIDLen]byte, error) {
|
|||||||
// TokenID
|
// TokenID
|
||||||
b = append(b, tx.TokenID.Bytes()[:]...)
|
b = append(b, tx.TokenID.Bytes()[:]...)
|
||||||
// Amount
|
// Amount
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
amountFloat40, err := NewFloat40(tx.Amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return txID, tracerr.Wrap(fmt.Errorf("%s: %d", err, tx.Amount))
|
return txID, tracerr.Wrap(fmt.Errorf("%s: %d", err, tx.Amount))
|
||||||
}
|
}
|
||||||
b = append(b, amountFloat16.Bytes()...)
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return txID, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
b = append(b, amountFloat40Bytes...)
|
||||||
// Nonce
|
// Nonce
|
||||||
nonceBytes, err := tx.Nonce.Bytes()
|
nonceBytes, err := tx.Nonce.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -170,11 +174,11 @@ func TxIDsFromL2Txs(txs []L2Tx) []TxID {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// BytesDataAvailability encodes a L2Tx into []byte for the Data Availability
|
// BytesDataAvailability encodes a L2Tx into []byte for the Data Availability
|
||||||
// [ fromIdx | toIdx | amountFloat16 | Fee ]
|
// [ fromIdx | toIdx | amountFloat40 | Fee ]
|
||||||
func (tx L2Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
func (tx L2Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
||||||
idxLen := nLevels / 8 //nolint:gomnd
|
idxLen := nLevels / 8 //nolint:gomnd
|
||||||
|
|
||||||
b := make([]byte, ((nLevels*2)+16+8)/8) //nolint:gomnd
|
b := make([]byte, ((nLevels*2)+40+8)/8) //nolint:gomnd
|
||||||
|
|
||||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
@@ -188,13 +192,16 @@ func (tx L2Tx) BytesDataAvailability(nLevels uint32) ([]byte, error) {
|
|||||||
}
|
}
|
||||||
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
copy(b[idxLen:idxLen*2], toIdxBytes[6-idxLen:])
|
||||||
|
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
amountFloat40, err := NewFloat40(tx.Amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
copy(b[idxLen*2:idxLen*2+2], amountFloat16.Bytes())
|
if err != nil {
|
||||||
b[idxLen*2+2] = byte(tx.Fee)
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
copy(b[idxLen*2:idxLen*2+Float40BytesLength], amountFloat40Bytes)
|
||||||
|
b[idxLen*2+Float40BytesLength] = byte(tx.Fee)
|
||||||
|
|
||||||
return b[:], nil
|
return b[:], nil
|
||||||
}
|
}
|
||||||
@@ -219,7 +226,10 @@ func L2TxFromBytesDataAvailability(b []byte, nLevels int) (*L2Tx, error) {
|
|||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
tx.Amount = Float16FromBytes(b[idxLen*2 : idxLen*2+2]).BigInt()
|
tx.Amount, err = Float40FromBytes(b[idxLen*2 : idxLen*2+Float40BytesLength]).BigInt()
|
||||||
tx.Fee = FeeSelector(b[idxLen*2+2])
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
tx.Fee = FeeSelector(b[idxLen*2+Float40BytesLength])
|
||||||
return tx, nil
|
return tx, nil
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -19,7 +19,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err := NewL2Tx(l2Tx)
|
l2Tx, err := NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x02fb52b5d0b9ef2626c11701bb751b2720c76d59946b9a48146ac153bb6e63bf6a", l2Tx.TxID.String())
|
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 87654,
|
FromIdx: 87654,
|
||||||
@@ -30,7 +30,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x0276114a8f666fa1ff7dbf34b4a9da577808dc501e3b2760d01fe3ef5473f5737f", l2Tx.TxID.String())
|
assert.Equal(t, "0x029e7499a830f8f5eb17c07da48cf91415710f1bcbe0169d363ff91e81faf92fc2", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 87654,
|
FromIdx: 87654,
|
||||||
@@ -42,7 +42,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x025afb63126d3067f61f633d13e5a51da0551af3a4567a9af2db5321ed04214ff4", l2Tx.TxID.String())
|
assert.Equal(t, "0x0255c70ed20e1b8935232e1b9c5884dbcc88a6e1a3454d24f2d77252eb2bb0b64e", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 87654,
|
FromIdx: 87654,
|
||||||
@@ -54,7 +54,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x02cf390157041c3b1b59f0aaed4da464f0d0d48f1d026e46fd89c7fe1e5aed7fcf", l2Tx.TxID.String())
|
assert.Equal(t, "0x0206b372f967061d1148bbcff679de38120e075141a80a07326d0f514c2efc6ca9", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 1,
|
FromIdx: 1,
|
||||||
@@ -66,7 +66,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x020ec18eaae67fcd545998841a9c4be09ee3083e12db6ae5e5213a2ecaaa52d5cf", l2Tx.TxID.String())
|
assert.Equal(t, "0x0236f7ea5bccf78ba60baf56c058d235a844f9b09259fd0efa4f5f72a7d4a26618", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 999,
|
FromIdx: 999,
|
||||||
@@ -78,7 +78,7 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x02f036223e79fac776de107f50822552cc964ee9fc4caa304613285f6976bcc940", l2Tx.TxID.String())
|
assert.Equal(t, "0x02ac122f5b709ce190129fecbbe35bfd30c70e6433dbd85a8eb743d110906a1dc1", l2Tx.TxID.String())
|
||||||
|
|
||||||
l2Tx = &L2Tx{
|
l2Tx = &L2Tx{
|
||||||
FromIdx: 4444,
|
FromIdx: 4444,
|
||||||
@@ -90,25 +90,85 @@ func TestNewL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
l2Tx, err = NewL2Tx(l2Tx)
|
l2Tx, err = NewL2Tx(l2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x029c8aef9ef24531e4cf84e78cbab1018ba1626a5a10afb6b7c356be1b5c28e92c", l2Tx.TxID.String())
|
assert.Equal(t, "0x02c674951a81881b7bc50db3b9e5efd97ac88550c7426ac548720e5057cfba515a", l2Tx.TxID.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestL2TxByteParsers(t *testing.T) {
|
func TestL2TxByteParsers(t *testing.T) {
|
||||||
amount := new(big.Int)
|
// test vectors values generated from javascript implementation
|
||||||
amount.SetString("79000000", 10)
|
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||||
|
require.True(t, ok)
|
||||||
l2Tx := &L2Tx{
|
l2Tx := &L2Tx{
|
||||||
ToIdx: 256,
|
ToIdx: (1 << 16) - 1,
|
||||||
|
FromIdx: (1 << 16) - 1,
|
||||||
Amount: amount,
|
Amount: amount,
|
||||||
FromIdx: 257,
|
Fee: (1 << 8) - 1,
|
||||||
Fee: 201,
|
|
||||||
}
|
}
|
||||||
// Data from the compatibility test
|
expected := "ffffffffffffffffffff"
|
||||||
expected := "00000101000001002b16c9"
|
encodedData, err := l2Tx.BytesDataAvailability(16)
|
||||||
encodedData, err := l2Tx.BytesDataAvailability(32)
|
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
|
|
||||||
decodedData, err := L2TxFromBytesDataAvailability(encodedData, 32)
|
decodedData, err := L2TxFromBytesDataAvailability(encodedData, 16)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, l2Tx, decodedData)
|
||||||
|
|
||||||
|
l2Tx = &L2Tx{
|
||||||
|
ToIdx: (1 << 32) - 1,
|
||||||
|
FromIdx: (1 << 32) - 1,
|
||||||
|
Amount: amount,
|
||||||
|
Fee: (1 << 8) - 1,
|
||||||
|
}
|
||||||
|
expected = "ffffffffffffffffffffffffffff"
|
||||||
|
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
|
|
||||||
|
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, l2Tx, decodedData)
|
||||||
|
|
||||||
|
l2Tx = &L2Tx{
|
||||||
|
ToIdx: 0,
|
||||||
|
FromIdx: 0,
|
||||||
|
Amount: big.NewInt(0),
|
||||||
|
Fee: 0,
|
||||||
|
}
|
||||||
|
expected = "0000000000000000000000000000"
|
||||||
|
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
|
|
||||||
|
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, l2Tx, decodedData)
|
||||||
|
|
||||||
|
l2Tx = &L2Tx{
|
||||||
|
ToIdx: 0,
|
||||||
|
FromIdx: 1061,
|
||||||
|
Amount: big.NewInt(420000000000),
|
||||||
|
Fee: 127,
|
||||||
|
}
|
||||||
|
expected = "000004250000000010fa56ea007f"
|
||||||
|
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
|
|
||||||
|
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, l2Tx, decodedData)
|
||||||
|
|
||||||
|
l2Tx = &L2Tx{
|
||||||
|
ToIdx: 256,
|
||||||
|
FromIdx: 257,
|
||||||
|
Amount: big.NewInt(79000000),
|
||||||
|
Fee: 201,
|
||||||
|
}
|
||||||
|
expected = "00000101000001000004b571c0c9"
|
||||||
|
encodedData, err = l2Tx.BytesDataAvailability(32)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, expected, hex.EncodeToString(encodedData))
|
||||||
|
|
||||||
|
decodedData, err = L2TxFromBytesDataAvailability(encodedData, 32)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, l2Tx, decodedData)
|
assert.Equal(t, l2Tx, decodedData)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -36,7 +36,7 @@ type PoolL2Tx struct {
|
|||||||
ToEthAddr ethCommon.Address `meddler:"to_eth_addr,zeroisnull"`
|
ToEthAddr ethCommon.Address `meddler:"to_eth_addr,zeroisnull"`
|
||||||
ToBJJ babyjub.PublicKeyComp `meddler:"to_bjj,zeroisnull"`
|
ToBJJ babyjub.PublicKeyComp `meddler:"to_bjj,zeroisnull"`
|
||||||
TokenID TokenID `meddler:"token_id"`
|
TokenID TokenID `meddler:"token_id"`
|
||||||
Amount *big.Int `meddler:"amount,bigint"` // TODO: change to float16
|
Amount *big.Int `meddler:"amount,bigint"`
|
||||||
Fee FeeSelector `meddler:"fee"`
|
Fee FeeSelector `meddler:"fee"`
|
||||||
Nonce Nonce `meddler:"nonce"` // effective 40 bits used
|
Nonce Nonce `meddler:"nonce"` // effective 40 bits used
|
||||||
State PoolL2TxState `meddler:"state"`
|
State PoolL2TxState `meddler:"state"`
|
||||||
@@ -53,7 +53,7 @@ type PoolL2Tx struct {
|
|||||||
RqToEthAddr ethCommon.Address `meddler:"rq_to_eth_addr,zeroisnull"`
|
RqToEthAddr ethCommon.Address `meddler:"rq_to_eth_addr,zeroisnull"`
|
||||||
RqToBJJ babyjub.PublicKeyComp `meddler:"rq_to_bjj,zeroisnull"`
|
RqToBJJ babyjub.PublicKeyComp `meddler:"rq_to_bjj,zeroisnull"`
|
||||||
RqTokenID TokenID `meddler:"rq_token_id,zeroisnull"`
|
RqTokenID TokenID `meddler:"rq_token_id,zeroisnull"`
|
||||||
RqAmount *big.Int `meddler:"rq_amount,bigintnull"` // TODO: change to float16
|
RqAmount *big.Int `meddler:"rq_amount,bigintnull"`
|
||||||
RqFee FeeSelector `meddler:"rq_fee,zeroisnull"`
|
RqFee FeeSelector `meddler:"rq_fee,zeroisnull"`
|
||||||
RqNonce Nonce `meddler:"rq_nonce,zeroisnull"` // effective 48 bits used
|
RqNonce Nonce `meddler:"rq_nonce,zeroisnull"` // effective 48 bits used
|
||||||
AbsoluteFee float64 `meddler:"fee_usd,zeroisnull"`
|
AbsoluteFee float64 `meddler:"fee_usd,zeroisnull"`
|
||||||
@@ -122,18 +122,13 @@ func (tx *PoolL2Tx) SetID() error {
|
|||||||
// [ 8 bits ] userFee // 1 byte
|
// [ 8 bits ] userFee // 1 byte
|
||||||
// [ 40 bits ] nonce // 5 bytes
|
// [ 40 bits ] nonce // 5 bytes
|
||||||
// [ 32 bits ] tokenID // 4 bytes
|
// [ 32 bits ] tokenID // 4 bytes
|
||||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
|
||||||
// [ 48 bits ] toIdx // 6 bytes
|
// [ 48 bits ] toIdx // 6 bytes
|
||||||
// [ 48 bits ] fromIdx // 6 bytes
|
// [ 48 bits ] fromIdx // 6 bytes
|
||||||
// [ 16 bits ] chainId // 2 bytes
|
// [ 16 bits ] chainId // 2 bytes
|
||||||
// [ 32 bits ] signatureConstant // 4 bytes
|
// [ 32 bits ] signatureConstant // 4 bytes
|
||||||
// Total bits compressed data: 241 bits // 31 bytes in *big.Int representation
|
// Total bits compressed data: 225 bits // 29 bytes in *big.Int representation
|
||||||
func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
var b [29]byte
|
||||||
if err != nil {
|
|
||||||
return nil, tracerr.Wrap(err)
|
|
||||||
}
|
|
||||||
var b [31]byte
|
|
||||||
|
|
||||||
toBJJSign := byte(0)
|
toBJJSign := byte(0)
|
||||||
pkSign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
pkSign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
||||||
@@ -149,19 +144,18 @@ func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
|||||||
}
|
}
|
||||||
copy(b[2:7], nonceBytes[:])
|
copy(b[2:7], nonceBytes[:])
|
||||||
copy(b[7:11], tx.TokenID.Bytes())
|
copy(b[7:11], tx.TokenID.Bytes())
|
||||||
copy(b[11:13], amountFloat16.Bytes())
|
|
||||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[13:19], toIdxBytes[:])
|
copy(b[11:17], toIdxBytes[:])
|
||||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[19:25], fromIdxBytes[:])
|
copy(b[17:23], fromIdxBytes[:])
|
||||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||||
copy(b[27:31], SignatureConstantBytes[:])
|
copy(b[25:29], SignatureConstantBytes[:])
|
||||||
|
|
||||||
bi := new(big.Int).SetBytes(b[:])
|
bi := new(big.Int).SetBytes(b[:])
|
||||||
return bi, nil
|
return bi, nil
|
||||||
@@ -170,9 +164,9 @@ func (tx *PoolL2Tx) TxCompressedData(chainID uint16) (*big.Int, error) {
|
|||||||
// TxCompressedDataEmpty calculates the TxCompressedData of an empty
|
// TxCompressedDataEmpty calculates the TxCompressedData of an empty
|
||||||
// transaction
|
// transaction
|
||||||
func TxCompressedDataEmpty(chainID uint16) *big.Int {
|
func TxCompressedDataEmpty(chainID uint16) *big.Int {
|
||||||
var b [31]byte
|
var b [29]byte
|
||||||
binary.BigEndian.PutUint16(b[25:27], chainID)
|
binary.BigEndian.PutUint16(b[23:25], chainID)
|
||||||
copy(b[27:31], SignatureConstantBytes[:])
|
copy(b[25:29], SignatureConstantBytes[:])
|
||||||
bi := new(big.Int).SetBytes(b[:])
|
bi := new(big.Int).SetBytes(b[:])
|
||||||
return bi
|
return bi
|
||||||
}
|
}
|
||||||
@@ -182,19 +176,24 @@ func TxCompressedDataEmpty(chainID uint16) *big.Int {
|
|||||||
// [ 8 bits ] userFee // 1 byte
|
// [ 8 bits ] userFee // 1 byte
|
||||||
// [ 40 bits ] nonce // 5 bytes
|
// [ 40 bits ] nonce // 5 bytes
|
||||||
// [ 32 bits ] tokenID // 4 bytes
|
// [ 32 bits ] tokenID // 4 bytes
|
||||||
// [ 16 bits ] amountFloat16 // 2 bytes
|
// [ 40 bits ] amountFloat40 // 5 bytes
|
||||||
// [ 48 bits ] toIdx // 6 bytes
|
// [ 48 bits ] toIdx // 6 bytes
|
||||||
// [ 48 bits ] fromIdx // 6 bytes
|
// [ 48 bits ] fromIdx // 6 bytes
|
||||||
// Total bits compressed data: 193 bits // 25 bytes in *big.Int representation
|
// Total bits compressed data: 217 bits // 28 bytes in *big.Int representation
|
||||||
func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
||||||
if tx.Amount == nil {
|
if tx.Amount == nil {
|
||||||
tx.Amount = big.NewInt(0)
|
tx.Amount = big.NewInt(0)
|
||||||
}
|
}
|
||||||
amountFloat16, err := NewFloat16(tx.Amount)
|
amountFloat40, err := NewFloat40(tx.Amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
var b [25]byte
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
var b [28]byte
|
||||||
toBJJSign := byte(0)
|
toBJJSign := byte(0)
|
||||||
if tx.ToBJJ != EmptyBJJComp {
|
if tx.ToBJJ != EmptyBJJComp {
|
||||||
sign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
sign, _ := babyjub.UnpackSignY(tx.ToBJJ)
|
||||||
@@ -210,17 +209,17 @@ func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
|||||||
}
|
}
|
||||||
copy(b[2:7], nonceBytes[:])
|
copy(b[2:7], nonceBytes[:])
|
||||||
copy(b[7:11], tx.TokenID.Bytes())
|
copy(b[7:11], tx.TokenID.Bytes())
|
||||||
copy(b[11:13], amountFloat16.Bytes())
|
copy(b[11:16], amountFloat40Bytes)
|
||||||
toIdxBytes, err := tx.ToIdx.Bytes()
|
toIdxBytes, err := tx.ToIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[13:19], toIdxBytes[:])
|
copy(b[16:22], toIdxBytes[:])
|
||||||
fromIdxBytes, err := tx.FromIdx.Bytes()
|
fromIdxBytes, err := tx.FromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[19:25], fromIdxBytes[:])
|
copy(b[22:28], fromIdxBytes[:])
|
||||||
|
|
||||||
bi := new(big.Int).SetBytes(b[:])
|
bi := new(big.Int).SetBytes(b[:])
|
||||||
return bi, nil
|
return bi, nil
|
||||||
@@ -236,19 +235,24 @@ func (tx *PoolL2Tx) TxCompressedDataV2() (*big.Int, error) {
|
|||||||
// [ 8 bits ] rqUserFee // 1 byte
|
// [ 8 bits ] rqUserFee // 1 byte
|
||||||
// [ 40 bits ] rqNonce // 5 bytes
|
// [ 40 bits ] rqNonce // 5 bytes
|
||||||
// [ 32 bits ] rqTokenID // 4 bytes
|
// [ 32 bits ] rqTokenID // 4 bytes
|
||||||
// [ 16 bits ] rqAmountFloat16 // 2 bytes
|
// [ 40 bits ] rqAmountFloat40 // 5 bytes
|
||||||
// [ 48 bits ] rqToIdx // 6 bytes
|
// [ 48 bits ] rqToIdx // 6 bytes
|
||||||
// [ 48 bits ] rqFromIdx // 6 bytes
|
// [ 48 bits ] rqFromIdx // 6 bytes
|
||||||
// Total bits compressed data: 193 bits // 25 bytes in *big.Int representation
|
// Total bits compressed data: 217 bits // 28 bytes in *big.Int representation
|
||||||
func (tx *PoolL2Tx) RqTxCompressedDataV2() (*big.Int, error) {
|
func (tx *PoolL2Tx) RqTxCompressedDataV2() (*big.Int, error) {
|
||||||
if tx.RqAmount == nil {
|
if tx.RqAmount == nil {
|
||||||
tx.RqAmount = big.NewInt(0)
|
tx.RqAmount = big.NewInt(0)
|
||||||
}
|
}
|
||||||
amountFloat16, err := NewFloat16(tx.RqAmount)
|
amountFloat40, err := NewFloat40(tx.RqAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
var b [25]byte
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
var b [28]byte
|
||||||
rqToBJJSign := byte(0)
|
rqToBJJSign := byte(0)
|
||||||
if tx.RqToBJJ != EmptyBJJComp {
|
if tx.RqToBJJ != EmptyBJJComp {
|
||||||
sign, _ := babyjub.UnpackSignY(tx.RqToBJJ)
|
sign, _ := babyjub.UnpackSignY(tx.RqToBJJ)
|
||||||
@@ -264,17 +268,17 @@ func (tx *PoolL2Tx) RqTxCompressedDataV2() (*big.Int, error) {
|
|||||||
}
|
}
|
||||||
copy(b[2:7], nonceBytes[:])
|
copy(b[2:7], nonceBytes[:])
|
||||||
copy(b[7:11], tx.RqTokenID.Bytes())
|
copy(b[7:11], tx.RqTokenID.Bytes())
|
||||||
copy(b[11:13], amountFloat16.Bytes())
|
copy(b[11:16], amountFloat40Bytes)
|
||||||
toIdxBytes, err := tx.RqToIdx.Bytes()
|
toIdxBytes, err := tx.RqToIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[13:19], toIdxBytes[:])
|
copy(b[16:22], toIdxBytes[:])
|
||||||
fromIdxBytes, err := tx.RqFromIdx.Bytes()
|
fromIdxBytes, err := tx.RqFromIdx.Bytes()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
copy(b[19:25], fromIdxBytes[:])
|
copy(b[22:28], fromIdxBytes[:])
|
||||||
|
|
||||||
bi := new(big.Int).SetBytes(b[:])
|
bi := new(big.Int).SetBytes(b[:])
|
||||||
return bi, nil
|
return bi, nil
|
||||||
@@ -287,7 +291,22 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// e1: [5 bytes AmountFloat40 | 20 bytes ToEthAddr]
|
||||||
|
var e1B [25]byte
|
||||||
|
amountFloat40, err := NewFloat40(tx.Amount)
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
amountFloat40Bytes, err := amountFloat40.Bytes()
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
copy(e1B[0:5], amountFloat40Bytes)
|
||||||
toEthAddr := EthAddrToBigInt(tx.ToEthAddr)
|
toEthAddr := EthAddrToBigInt(tx.ToEthAddr)
|
||||||
|
copy(e1B[5:25], toEthAddr.Bytes())
|
||||||
|
e1 := new(big.Int).SetBytes(e1B[:])
|
||||||
|
|
||||||
rqToEthAddr := EthAddrToBigInt(tx.RqToEthAddr)
|
rqToEthAddr := EthAddrToBigInt(tx.RqToEthAddr)
|
||||||
|
|
||||||
_, toBJJY := babyjub.UnpackSignY(tx.ToBJJ)
|
_, toBJJY := babyjub.UnpackSignY(tx.ToBJJ)
|
||||||
@@ -299,7 +318,7 @@ func (tx *PoolL2Tx) HashToSign(chainID uint16) (*big.Int, error) {
|
|||||||
|
|
||||||
_, rqToBJJY := babyjub.UnpackSignY(tx.RqToBJJ)
|
_, rqToBJJY := babyjub.UnpackSignY(tx.RqToBJJ)
|
||||||
|
|
||||||
return poseidon.Hash([]*big.Int{toCompressedData, toEthAddr, toBJJY, rqTxCompressedDataV2, rqToEthAddr, rqToBJJY})
|
return poseidon.Hash([]*big.Int{toCompressedData, e1, toBJJY, rqTxCompressedDataV2, rqToEthAddr, rqToBJJY})
|
||||||
}
|
}
|
||||||
|
|
||||||
// VerifySignature returns true if the signature verification is correct for the given PublicKeyComp
|
// VerifySignature returns true if the signature verification is correct for the given PublicKeyComp
|
||||||
|
|||||||
@@ -21,80 +21,104 @@ func TestNewPoolL2Tx(t *testing.T) {
|
|||||||
}
|
}
|
||||||
poolL2Tx, err := NewPoolL2Tx(poolL2Tx)
|
poolL2Tx, err := NewPoolL2Tx(poolL2Tx)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "0x02fb52b5d0b9ef2626c11701bb751b2720c76d59946b9a48146ac153bb6e63bf6a", poolL2Tx.TxID.String())
|
assert.Equal(t, "0x022669acda59b827d20ef5354a3eebd1dffb3972b0a6bf89d18bfd2efa0ab9f41e", poolL2Tx.TxID.String())
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestTxCompressedData(t *testing.T) {
|
func TestTxCompressedDataAndTxCompressedDataV2JSVectors(t *testing.T) {
|
||||||
chainID := uint16(0)
|
// test vectors values generated from javascript implementation
|
||||||
var sk babyjub.PrivateKey
|
var skPositive babyjub.PrivateKey // 'Positive' refers to the sign
|
||||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
_, err := hex.Decode(skPositive[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
var skNegative babyjub.PrivateKey // 'Negative' refers to the sign
|
||||||
|
_, err = hex.Decode(skNegative[:], []byte("0001020304050607080900010203040506070809000102030405060708090002"))
|
||||||
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
amount, ok := new(big.Int).SetString("343597383670000000000000000000000000000000", 10)
|
||||||
|
require.True(t, ok)
|
||||||
tx := PoolL2Tx{
|
tx := PoolL2Tx{
|
||||||
FromIdx: 2,
|
FromIdx: (1 << 48) - 1,
|
||||||
ToIdx: 3,
|
ToIdx: (1 << 48) - 1,
|
||||||
Amount: big.NewInt(4),
|
Amount: amount,
|
||||||
TokenID: 5,
|
TokenID: (1 << 32) - 1,
|
||||||
Nonce: 6,
|
Nonce: (1 << 40) - 1,
|
||||||
ToBJJ: sk.Public().Compress(),
|
Fee: (1 << 3) - 1,
|
||||||
|
ToBJJ: skPositive.Public().Compress(),
|
||||||
}
|
}
|
||||||
txCompressedData, err := tx.TxCompressedData(chainID)
|
txCompressedData, err := tx.TxCompressedData(uint16((1 << 16) - 1))
|
||||||
assert.NoError(t, err)
|
require.NoError(t, err)
|
||||||
// test vector value generated from javascript implementation
|
expectedStr := "0107ffffffffffffffffffffffffffffffffffffffffffffffc60be60f"
|
||||||
expectedStr := "1766847064778421992193717128424891165872736891548909569553540445094274575"
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
|
||||||
assert.Equal(t, "010000000000060000000500040000000000030000000000020000c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
txCompressedDataV2, err := tx.TxCompressedDataV2()
|
||||||
// using a different chainID
|
require.NoError(t, err)
|
||||||
txCompressedData, err = tx.TxCompressedData(uint16(100))
|
expectedStr = "0107ffffffffffffffffffffffffffffffffffffffffffffffffffff"
|
||||||
assert.NoError(t, err)
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedDataV2.Bytes()))
|
||||||
expectedStr = "1766847064778421992193717128424891165872736891548909569553540874591004175"
|
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
|
||||||
assert.Equal(t, "010000000000060000000500040000000000030000000000020064c60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
|
||||||
txCompressedData, err = tx.TxCompressedData(uint16(65535))
|
|
||||||
assert.NoError(t, err)
|
|
||||||
expectedStr = "1766847064778421992193717128424891165872736891548909569553821915776017935"
|
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
|
||||||
assert.Equal(t, "01000000000006000000050004000000000003000000000002ffffc60be60f", hex.EncodeToString(txCompressedData.Bytes()))
|
|
||||||
|
|
||||||
tx = PoolL2Tx{
|
tx = PoolL2Tx{
|
||||||
RqFromIdx: 7,
|
FromIdx: 0,
|
||||||
RqToIdx: 8,
|
ToIdx: 0,
|
||||||
RqAmount: big.NewInt(9),
|
Amount: big.NewInt(0),
|
||||||
RqTokenID: 10,
|
TokenID: 0,
|
||||||
RqNonce: 11,
|
Nonce: 0,
|
||||||
RqFee: 12,
|
Fee: 0,
|
||||||
RqToBJJ: sk.Public().Compress(),
|
ToBJJ: skNegative.Public().Compress(),
|
||||||
}
|
|
||||||
rqTxCompressedData, err := tx.RqTxCompressedDataV2()
|
|
||||||
assert.NoError(t, err)
|
|
||||||
// test vector value generated from javascript implementation
|
|
||||||
expectedStr = "6571340879233176732837827812956721483162819083004853354503"
|
|
||||||
assert.Equal(t, expectedStr, rqTxCompressedData.String())
|
|
||||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(rqTxCompressedData.Bytes()))
|
|
||||||
}
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||||
|
require.NoError(t, err)
|
||||||
|
expectedStr = "c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
func TestTxCompressedDataV2(t *testing.T) {
|
txCompressedDataV2, err = tx.TxCompressedDataV2()
|
||||||
var sk babyjub.PrivateKey
|
require.NoError(t, err)
|
||||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
assert.Equal(t, "0", txCompressedDataV2.String())
|
||||||
assert.NoError(t, err)
|
|
||||||
tx := PoolL2Tx{
|
|
||||||
FromIdx: 7,
|
|
||||||
ToIdx: 8,
|
|
||||||
Amount: big.NewInt(9),
|
|
||||||
TokenID: 10,
|
|
||||||
Nonce: 11,
|
|
||||||
Fee: 12,
|
|
||||||
ToBJJ: sk.Public().Compress(),
|
|
||||||
}
|
|
||||||
txCompressedData, err := tx.TxCompressedDataV2()
|
|
||||||
assert.NoError(t, err)
|
|
||||||
// test vector value generated from javascript implementation
|
|
||||||
expectedStr := "6571340879233176732837827812956721483162819083004853354503"
|
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
|
||||||
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
|
||||||
assert.True(t, ok)
|
|
||||||
|
|
||||||
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
amount, ok = new(big.Int).SetString("63000000000000000", 10)
|
||||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
require.True(t, ok)
|
||||||
|
tx = PoolL2Tx{
|
||||||
|
FromIdx: 324,
|
||||||
|
ToIdx: 256,
|
||||||
|
Amount: amount,
|
||||||
|
TokenID: 123,
|
||||||
|
Nonce: 76,
|
||||||
|
Fee: 214,
|
||||||
|
ToBJJ: skNegative.Public().Compress(),
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(1))
|
||||||
|
require.NoError(t, err)
|
||||||
|
expectedStr = "d6000000004c0000007b0000000001000000000001440001c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
|
txCompressedDataV2, err = tx.TxCompressedDataV2()
|
||||||
|
require.NoError(t, err)
|
||||||
|
expectedStr = "d6000000004c0000007b3977825f00000000000100000000000144"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedDataV2.Bytes()))
|
||||||
|
|
||||||
|
tx = PoolL2Tx{
|
||||||
|
FromIdx: 1,
|
||||||
|
ToIdx: 2,
|
||||||
|
TokenID: 3,
|
||||||
|
Nonce: 4,
|
||||||
|
Fee: 5,
|
||||||
|
ToBJJ: skNegative.Public().Compress(),
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||||
|
require.NoError(t, err)
|
||||||
|
expectedStr = "050000000004000000030000000000020000000000010000c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
|
|
||||||
|
tx = PoolL2Tx{
|
||||||
|
FromIdx: 2,
|
||||||
|
ToIdx: 3,
|
||||||
|
TokenID: 4,
|
||||||
|
Nonce: 5,
|
||||||
|
Fee: 6,
|
||||||
|
ToBJJ: skPositive.Public().Compress(),
|
||||||
|
}
|
||||||
|
txCompressedData, err = tx.TxCompressedData(uint16(0))
|
||||||
|
require.NoError(t, err)
|
||||||
|
expectedStr = "01060000000005000000040000000000030000000000020000c60be60f"
|
||||||
|
assert.Equal(t, expectedStr, hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestRqTxCompressedDataV2(t *testing.T) {
|
func TestRqTxCompressedDataV2(t *testing.T) {
|
||||||
@@ -113,19 +137,16 @@ func TestRqTxCompressedDataV2(t *testing.T) {
|
|||||||
txCompressedData, err := tx.RqTxCompressedDataV2()
|
txCompressedData, err := tx.RqTxCompressedDataV2()
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
// test vector value generated from javascript implementation
|
// test vector value generated from javascript implementation
|
||||||
expectedStr := "6571340879233176732837827812956721483162819083004853354503"
|
expectedStr := "110248805340524920412994530176819463725852160917809517418728390663"
|
||||||
assert.Equal(t, expectedStr, txCompressedData.String())
|
assert.Equal(t, expectedStr, txCompressedData.String())
|
||||||
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
expected, ok := new(big.Int).SetString(expectedStr, 10)
|
||||||
assert.True(t, ok)
|
assert.True(t, ok)
|
||||||
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
assert.Equal(t, expected.Bytes(), txCompressedData.Bytes())
|
||||||
assert.Equal(t, "010c000000000b0000000a0009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
assert.Equal(t, "010c000000000b0000000a0000000009000000000008000000000007", hex.EncodeToString(txCompressedData.Bytes()))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestHashToSign(t *testing.T) {
|
func TestHashToSign(t *testing.T) {
|
||||||
chainID := uint16(0)
|
chainID := uint16(0)
|
||||||
var sk babyjub.PrivateKey
|
|
||||||
_, err := hex.Decode(sk[:], []byte("0001020304050607080900010203040506070809000102030405060708090001"))
|
|
||||||
assert.NoError(t, err)
|
|
||||||
tx := PoolL2Tx{
|
tx := PoolL2Tx{
|
||||||
FromIdx: 2,
|
FromIdx: 2,
|
||||||
ToIdx: 3,
|
ToIdx: 3,
|
||||||
@@ -136,7 +157,7 @@ func TestHashToSign(t *testing.T) {
|
|||||||
}
|
}
|
||||||
toSign, err := tx.HashToSign(chainID)
|
toSign, err := tx.HashToSign(chainID)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "1469900657138253851938022936440971384682713995864967090251961124784132925291", toSign.String())
|
assert.Equal(t, "2d49ce1d4136e06f64e3eb1f79a346e6ee3e93ceeac909a57806a8d87005c263", hex.EncodeToString(toSign.Bytes()))
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestVerifyTxSignature(t *testing.T) {
|
func TestVerifyTxSignature(t *testing.T) {
|
||||||
@@ -156,7 +177,7 @@ func TestVerifyTxSignature(t *testing.T) {
|
|||||||
}
|
}
|
||||||
toSign, err := tx.HashToSign(chainID)
|
toSign, err := tx.HashToSign(chainID)
|
||||||
assert.NoError(t, err)
|
assert.NoError(t, err)
|
||||||
assert.Equal(t, "18645218094210271622244722988708640202588315450486586312909439859037906375295", toSign.String())
|
assert.Equal(t, "1571327027383224465388301747239444557034990637650927918405777653988509342917", toSign.String())
|
||||||
|
|
||||||
sig := sk.SignPoseidon(toSign)
|
sig := sk.SignPoseidon(toSign)
|
||||||
tx.Signature = sig.Compress()
|
tx.Signature = sig.Compress()
|
||||||
|
|||||||
15
common/zk.go
15
common/zk.go
@@ -102,6 +102,8 @@ type ZKInputs struct {
|
|||||||
ToBJJAy []*big.Int `json:"toBjjAy"` // big.Int, len: [maxTx]
|
ToBJJAy []*big.Int `json:"toBjjAy"` // big.Int, len: [maxTx]
|
||||||
// ToEthAddr
|
// ToEthAddr
|
||||||
ToEthAddr []*big.Int `json:"toEthAddr"` // ethCommon.Address, len: [maxTx]
|
ToEthAddr []*big.Int `json:"toEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||||
|
// AmountF encoded as float40
|
||||||
|
AmountF []*big.Int `json:"amountF"` // uint40 len: [maxTx]
|
||||||
|
|
||||||
// OnChain determines if is L1 (1/true) or L2 (0/false)
|
// OnChain determines if is L1 (1/true) or L2 (0/false)
|
||||||
OnChain []*big.Int `json:"onChain"` // bool, len: [maxTx]
|
OnChain []*big.Int `json:"onChain"` // bool, len: [maxTx]
|
||||||
@@ -112,8 +114,8 @@ type ZKInputs struct {
|
|||||||
// NewAccount boolean (0/1) flag set 'true' when L1 tx creates a new
|
// NewAccount boolean (0/1) flag set 'true' when L1 tx creates a new
|
||||||
// account (fromIdx==0)
|
// account (fromIdx==0)
|
||||||
NewAccount []*big.Int `json:"newAccount"` // bool, len: [maxTx]
|
NewAccount []*big.Int `json:"newAccount"` // bool, len: [maxTx]
|
||||||
// DepositAmountF encoded as float16
|
// DepositAmountF encoded as float40
|
||||||
DepositAmountF []*big.Int `json:"loadAmountF"` // uint16, len: [maxTx]
|
DepositAmountF []*big.Int `json:"loadAmountF"` // uint40, len: [maxTx]
|
||||||
// FromEthAddr
|
// FromEthAddr
|
||||||
FromEthAddr []*big.Int `json:"fromEthAddr"` // ethCommon.Address, len: [maxTx]
|
FromEthAddr []*big.Int `json:"fromEthAddr"` // ethCommon.Address, len: [maxTx]
|
||||||
// FromBJJCompressed boolean encoded where each value is a *big.Int
|
// FromBJJCompressed boolean encoded where each value is a *big.Int
|
||||||
@@ -326,6 +328,7 @@ func NewZKInputs(chainID uint16, maxTx, maxL1Tx, maxFeeIdxs, nLevels uint32, cur
|
|||||||
zki.AuxToIdx = newSlice(maxTx)
|
zki.AuxToIdx = newSlice(maxTx)
|
||||||
zki.ToBJJAy = newSlice(maxTx)
|
zki.ToBJJAy = newSlice(maxTx)
|
||||||
zki.ToEthAddr = newSlice(maxTx)
|
zki.ToEthAddr = newSlice(maxTx)
|
||||||
|
zki.AmountF = newSlice(maxTx)
|
||||||
zki.OnChain = newSlice(maxTx)
|
zki.OnChain = newSlice(maxTx)
|
||||||
zki.NewAccount = newSlice(maxTx)
|
zki.NewAccount = newSlice(maxTx)
|
||||||
|
|
||||||
@@ -476,8 +479,8 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
|||||||
copy(newExitRoot, z.Metadata.NewExitRootRaw.Bytes())
|
copy(newExitRoot, z.Metadata.NewExitRootRaw.Bytes())
|
||||||
b = append(b, newExitRoot...)
|
b = append(b, newExitRoot...)
|
||||||
|
|
||||||
// [MAX_L1_TX * (2 * MAX_NLEVELS + 480) bits] L1TxsData
|
// [MAX_L1_TX * (2 * MAX_NLEVELS + 528) bits] L1TxsData
|
||||||
l1TxDataLen := (2*z.Metadata.MaxLevels + 480)
|
l1TxDataLen := (2*z.Metadata.MaxLevels + 528)
|
||||||
l1TxsDataLen := (z.Metadata.MaxL1Tx * l1TxDataLen)
|
l1TxsDataLen := (z.Metadata.MaxL1Tx * l1TxDataLen)
|
||||||
l1TxsData := make([]byte, l1TxsDataLen/8) //nolint:gomnd
|
l1TxsData := make([]byte, l1TxsDataLen/8) //nolint:gomnd
|
||||||
for i := 0; i < len(z.Metadata.L1TxsData); i++ {
|
for i := 0; i < len(z.Metadata.L1TxsData); i++ {
|
||||||
@@ -494,9 +497,9 @@ func (z ZKInputs) ToHashGlobalData() ([]byte, error) {
|
|||||||
}
|
}
|
||||||
b = append(b, l1TxsDataAvailability...)
|
b = append(b, l1TxsDataAvailability...)
|
||||||
|
|
||||||
// [MAX_TX*(2*NLevels + 24) bits] L2TxsData
|
// [MAX_TX*(2*NLevels + 48) bits] L2TxsData
|
||||||
var l2TxsData []byte
|
var l2TxsData []byte
|
||||||
l2TxDataLen := 2*z.Metadata.NLevels + 24 //nolint:gomnd
|
l2TxDataLen := 2*z.Metadata.NLevels + 48 //nolint:gomnd
|
||||||
l2TxsDataLen := (z.Metadata.MaxTx * l2TxDataLen)
|
l2TxsDataLen := (z.Metadata.MaxTx * l2TxDataLen)
|
||||||
expectedL2TxsDataLen := l2TxsDataLen / 8 //nolint:gomnd
|
expectedL2TxsDataLen := l2TxsDataLen / 8 //nolint:gomnd
|
||||||
for i := 0; i < len(z.Metadata.L2TxsData); i++ {
|
for i := 0; i < len(z.Metadata.L2TxsData); i++ {
|
||||||
|
|||||||
@@ -79,6 +79,15 @@ type Coordinator struct {
|
|||||||
// ForgeRetryInterval is the waiting interval between calls forge a
|
// ForgeRetryInterval is the waiting interval between calls forge a
|
||||||
// batch after an error
|
// batch after an error
|
||||||
ForgeRetryInterval Duration `validate:"required"`
|
ForgeRetryInterval Duration `validate:"required"`
|
||||||
|
// ForgeDelay is the delay after which a batch is forged if the slot is
|
||||||
|
// already committed. If set to 0s, the coordinator will continuously
|
||||||
|
// forge at the maximum rate.
|
||||||
|
ForgeDelay Duration `validate:"-"`
|
||||||
|
// ForgeNoTxsDelay is the delay after which a batch is forged even if
|
||||||
|
// there are no txs to forge if the slot is already committed. If set
|
||||||
|
// to 0s, the coordinator will continuously forge even if the batches
|
||||||
|
// are empty.
|
||||||
|
ForgeNoTxsDelay Duration `validate:"-"`
|
||||||
// SyncRetryInterval is the waiting interval between calls to the main
|
// SyncRetryInterval is the waiting interval between calls to the main
|
||||||
// handler of a synced block after an error
|
// handler of a synced block after an error
|
||||||
SyncRetryInterval Duration `validate:"required"`
|
SyncRetryInterval Duration `validate:"required"`
|
||||||
@@ -87,19 +96,33 @@ type Coordinator struct {
|
|||||||
// SafetyPeriod is the number of batches after which
|
// SafetyPeriod is the number of batches after which
|
||||||
// non-pending L2Txs are deleted from the pool
|
// non-pending L2Txs are deleted from the pool
|
||||||
SafetyPeriod common.BatchNum `validate:"required"`
|
SafetyPeriod common.BatchNum `validate:"required"`
|
||||||
// MaxTxs is the number of L2Txs that once reached triggers
|
// MaxTxs is the maximum number of pending L2Txs that can be
|
||||||
// deletion of old L2Txs
|
// stored in the pool. Once this number of pending L2Txs is
|
||||||
|
// reached, inserts to the pool will be denied until some of
|
||||||
|
// the pending txs are forged.
|
||||||
MaxTxs uint32 `validate:"required"`
|
MaxTxs uint32 `validate:"required"`
|
||||||
// TTL is the Time To Live for L2Txs in the pool. Once MaxTxs
|
// TTL is the Time To Live for L2Txs in the pool. Once MaxTxs
|
||||||
// L2Txs is reached, L2Txs older than TTL will be deleted.
|
// L2Txs is reached, L2Txs older than TTL will be deleted.
|
||||||
TTL Duration `validate:"required"`
|
TTL Duration `validate:"required"`
|
||||||
// PurgeBatchDelay is the delay between batches to purge outdated transactions
|
// PurgeBatchDelay is the delay between batches to purge
|
||||||
|
// outdated transactions. Oudated L2Txs are those that have
|
||||||
|
// been forged or marked as invalid for longer than the
|
||||||
|
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||||
|
// for longer than TTL once there are MaxTxs.
|
||||||
PurgeBatchDelay int64 `validate:"required"`
|
PurgeBatchDelay int64 `validate:"required"`
|
||||||
// InvalidateBatchDelay is the delay between batches to mark invalid transactions
|
// InvalidateBatchDelay is the delay between batches to mark
|
||||||
|
// invalid transactions due to nonce lower than the account
|
||||||
|
// nonce.
|
||||||
InvalidateBatchDelay int64 `validate:"required"`
|
InvalidateBatchDelay int64 `validate:"required"`
|
||||||
// PurgeBlockDelay is the delay between blocks to purge outdated transactions
|
// PurgeBlockDelay is the delay between blocks to purge
|
||||||
|
// outdated transactions. Oudated L2Txs are those that have
|
||||||
|
// been forged or marked as invalid for longer than the
|
||||||
|
// SafetyPeriod and pending L2Txs that have been in the pool
|
||||||
|
// for longer than TTL once there are MaxTxs.
|
||||||
PurgeBlockDelay int64 `validate:"required"`
|
PurgeBlockDelay int64 `validate:"required"`
|
||||||
// InvalidateBlockDelay is the delay between blocks to mark invalid transactions
|
// InvalidateBlockDelay is the delay between blocks to mark
|
||||||
|
// invalid transactions due to nonce lower than the account
|
||||||
|
// nonce.
|
||||||
InvalidateBlockDelay int64 `validate:"required"`
|
InvalidateBlockDelay int64 `validate:"required"`
|
||||||
} `validate:"required"`
|
} `validate:"required"`
|
||||||
TxSelector struct {
|
TxSelector struct {
|
||||||
@@ -119,15 +142,13 @@ type Coordinator struct {
|
|||||||
NLevels int64 `validate:"required"`
|
NLevels int64 `validate:"required"`
|
||||||
} `validate:"required"`
|
} `validate:"required"`
|
||||||
EthClient struct {
|
EthClient struct {
|
||||||
// CallGasLimit is the default gas limit set for ethereum
|
|
||||||
// calls, except for methods where a particular gas limit is
|
|
||||||
// harcoded because it's known to be a big value
|
|
||||||
CallGasLimit uint64 `validate:"required"`
|
|
||||||
// MaxGasPrice is the maximum gas price allowed for ethereum
|
// MaxGasPrice is the maximum gas price allowed for ethereum
|
||||||
// transactions
|
// transactions
|
||||||
MaxGasPrice *big.Int `validate:"required"`
|
MaxGasPrice *big.Int `validate:"required"`
|
||||||
// GasPriceDiv is the gas price division
|
// GasPriceIncPerc is the percentage increase of gas price set
|
||||||
GasPriceDiv uint64 `validate:"required"`
|
// in an ethereum transaction from the suggested gas price by
|
||||||
|
// the ehtereum node
|
||||||
|
GasPriceIncPerc int64
|
||||||
// CheckLoopInterval is the waiting interval between receipt
|
// CheckLoopInterval is the waiting interval between receipt
|
||||||
// checks of ethereum transactions in the TxManager
|
// checks of ethereum transactions in the TxManager
|
||||||
CheckLoopInterval Duration `validate:"required"`
|
CheckLoopInterval Duration `validate:"required"`
|
||||||
|
|||||||
@@ -92,6 +92,7 @@ type BatchInfo struct {
|
|||||||
ForgeBatchArgs *eth.RollupForgeBatchArgs
|
ForgeBatchArgs *eth.RollupForgeBatchArgs
|
||||||
// FeesInfo
|
// FeesInfo
|
||||||
EthTx *types.Transaction
|
EthTx *types.Transaction
|
||||||
|
EthTxErr error
|
||||||
// SendTimestamp the time of batch sent to ethereum
|
// SendTimestamp the time of batch sent to ethereum
|
||||||
SendTimestamp time.Time
|
SendTimestamp time.Time
|
||||||
Receipt *types.Receipt
|
Receipt *types.Receipt
|
||||||
|
|||||||
@@ -24,6 +24,8 @@ import (
|
|||||||
|
|
||||||
var (
|
var (
|
||||||
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
errLastL1BatchNotSynced = fmt.Errorf("last L1Batch not synced yet")
|
||||||
|
errForgeNoTxsBeforeDelay = fmt.Errorf("no txs to forge and we haven't reached the forge no txs delay")
|
||||||
|
errForgeBeforeDelay = fmt.Errorf("we haven't reached the forge delay")
|
||||||
)
|
)
|
||||||
|
|
||||||
const (
|
const (
|
||||||
@@ -71,6 +73,15 @@ type Config struct {
|
|||||||
// ForgeRetryInterval is the waiting interval between calls forge a
|
// ForgeRetryInterval is the waiting interval between calls forge a
|
||||||
// batch after an error
|
// batch after an error
|
||||||
ForgeRetryInterval time.Duration
|
ForgeRetryInterval time.Duration
|
||||||
|
// ForgeDelay is the delay after which a batch is forged if the slot is
|
||||||
|
// already committed. If set to 0s, the coordinator will continuously
|
||||||
|
// forge at the maximum rate.
|
||||||
|
ForgeDelay time.Duration
|
||||||
|
// ForgeNoTxsDelay is the delay after which a batch is forged even if
|
||||||
|
// there are no txs to forge if the slot is already committed. If set
|
||||||
|
// to 0s, the coordinator will continuously forge even if the batches
|
||||||
|
// are empty.
|
||||||
|
ForgeNoTxsDelay time.Duration
|
||||||
// SyncRetryInterval is the waiting interval between calls to the main
|
// SyncRetryInterval is the waiting interval between calls to the main
|
||||||
// handler of a synced block after an error
|
// handler of a synced block after an error
|
||||||
SyncRetryInterval time.Duration
|
SyncRetryInterval time.Duration
|
||||||
@@ -87,6 +98,10 @@ type Config struct {
|
|||||||
// MaxGasPrice is the maximum gas price allowed for ethereum
|
// MaxGasPrice is the maximum gas price allowed for ethereum
|
||||||
// transactions
|
// transactions
|
||||||
MaxGasPrice *big.Int
|
MaxGasPrice *big.Int
|
||||||
|
// GasPriceIncPerc is the percentage increase of gas price set in an
|
||||||
|
// ethereum transaction from the suggested gas price by the ehtereum
|
||||||
|
// node
|
||||||
|
GasPriceIncPerc int64
|
||||||
// TxManagerCheckInterval is the waiting interval between receipt
|
// TxManagerCheckInterval is the waiting interval between receipt
|
||||||
// checks of ethereum transactions in the TxManager
|
// checks of ethereum transactions in the TxManager
|
||||||
TxManagerCheckInterval time.Duration
|
TxManagerCheckInterval time.Duration
|
||||||
@@ -254,7 +269,7 @@ type MsgSyncReorg struct {
|
|||||||
// MsgStopPipeline indicates a signal to reset the pipeline
|
// MsgStopPipeline indicates a signal to reset the pipeline
|
||||||
type MsgStopPipeline struct {
|
type MsgStopPipeline struct {
|
||||||
Reason string
|
Reason string
|
||||||
// FailedBatchNum indicates the first batchNum that faile in the
|
// FailedBatchNum indicates the first batchNum that failed in the
|
||||||
// pipeline. If FailedBatchNum is 0, it should be ignored.
|
// pipeline. If FailedBatchNum is 0, it should be ignored.
|
||||||
FailedBatchNum common.BatchNum
|
FailedBatchNum common.BatchNum
|
||||||
}
|
}
|
||||||
@@ -338,25 +353,31 @@ func (c *Coordinator) syncStats(ctx context.Context, stats *synchronizer.Stats)
|
|||||||
if c.pipeline == nil {
|
if c.pipeline == nil {
|
||||||
relativeBlock := c.consts.Auction.RelativeBlock(nextBlock)
|
relativeBlock := c.consts.Auction.RelativeBlock(nextBlock)
|
||||||
if canForge && relativeBlock < c.cfg.StartSlotBlocksDelay {
|
if canForge && relativeBlock < c.cfg.StartSlotBlocksDelay {
|
||||||
log.Debugw("Coordinator: delaying pipeline start due to "+
|
log.Debugf("Coordinator: delaying pipeline start due to "+
|
||||||
"relativeBlock (%v) < cfg.StartSlotBlocksDelay (%v)",
|
"relativeBlock (%v) < cfg.StartSlotBlocksDelay (%v)",
|
||||||
relativeBlock, c.cfg.StartSlotBlocksDelay)
|
relativeBlock, c.cfg.StartSlotBlocksDelay)
|
||||||
} else if canForge {
|
} else if canForge {
|
||||||
log.Infow("Coordinator: forging state begin", "block",
|
log.Infow("Coordinator: forging state begin", "block",
|
||||||
stats.Eth.LastBlock.Num+1, "batch", stats.Sync.LastBatch.BatchNum)
|
stats.Eth.LastBlock.Num+1, "batch", stats.Sync.LastBatch.BatchNum)
|
||||||
batchNum := stats.Sync.LastBatch.BatchNum
|
fromBatch := fromBatch{
|
||||||
if c.lastNonFailedBatchNum > batchNum {
|
BatchNum: stats.Sync.LastBatch.BatchNum,
|
||||||
batchNum = c.lastNonFailedBatchNum
|
ForgerAddr: stats.Sync.LastBatch.ForgerAddr,
|
||||||
|
StateRoot: stats.Sync.LastBatch.StateRoot,
|
||||||
|
}
|
||||||
|
if c.lastNonFailedBatchNum > fromBatch.BatchNum {
|
||||||
|
fromBatch.BatchNum = c.lastNonFailedBatchNum
|
||||||
|
fromBatch.ForgerAddr = c.cfg.ForgerAddress
|
||||||
|
fromBatch.StateRoot = big.NewInt(0)
|
||||||
}
|
}
|
||||||
var err error
|
var err error
|
||||||
if c.pipeline, err = c.newPipeline(ctx); err != nil {
|
if c.pipeline, err = c.newPipeline(ctx); err != nil {
|
||||||
return tracerr.Wrap(err)
|
return tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
if err := c.pipeline.Start(batchNum, stats, &c.vars); err != nil {
|
c.pipelineFromBatch = fromBatch
|
||||||
|
if err := c.pipeline.Start(fromBatch.BatchNum, stats, &c.vars); err != nil {
|
||||||
c.pipeline = nil
|
c.pipeline = nil
|
||||||
return tracerr.Wrap(err)
|
return tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
// c.pipelineBatchNum = batchNum
|
|
||||||
}
|
}
|
||||||
} else {
|
} else {
|
||||||
if !canForge {
|
if !canForge {
|
||||||
@@ -366,18 +387,6 @@ func (c *Coordinator) syncStats(ctx context.Context, stats *synchronizer.Stats)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
if c.pipeline == nil {
|
if c.pipeline == nil {
|
||||||
// Mark invalid in Pool due to forged L2Txs
|
|
||||||
// for _, batch := range batches {
|
|
||||||
// if err := c.l2DB.InvalidateOldNonces(
|
|
||||||
// idxsNonceFromL2Txs(batch.L2Txs), batch.Batch.BatchNum); err != nil {
|
|
||||||
// return err
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// if c.purger.CanInvalidate(stats.Sync.LastBlock.Num, int64(stats.Sync.LastBatch.BatchNum)) {
|
|
||||||
// if err := c.txSelector.Reset(stats.Sync.LastBatch.BatchNum); err != nil {
|
|
||||||
// return tracerr.Wrap(err)
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
if _, err := c.purger.InvalidateMaybe(c.l2DB, c.txSelector.LocalAccountsDB(),
|
if _, err := c.purger.InvalidateMaybe(c.l2DB, c.txSelector.LocalAccountsDB(),
|
||||||
stats.Sync.LastBlock.Num, int64(stats.Sync.LastBatch.BatchNum)); err != nil {
|
stats.Sync.LastBlock.Num, int64(stats.Sync.LastBatch.BatchNum)); err != nil {
|
||||||
return tracerr.Wrap(err)
|
return tracerr.Wrap(err)
|
||||||
@@ -411,7 +420,8 @@ func (c *Coordinator) handleReorg(ctx context.Context, msg *MsgSyncReorg) error
|
|||||||
c.pipeline.SetSyncStatsVars(ctx, &msg.Stats, &msg.Vars)
|
c.pipeline.SetSyncStatsVars(ctx, &msg.Stats, &msg.Vars)
|
||||||
}
|
}
|
||||||
if c.stats.Sync.LastBatch.ForgerAddr != c.cfg.ForgerAddress &&
|
if c.stats.Sync.LastBatch.ForgerAddr != c.cfg.ForgerAddress &&
|
||||||
c.stats.Sync.LastBatch.StateRoot.Cmp(c.pipelineFromBatch.StateRoot) != 0 {
|
(c.stats.Sync.LastBatch.StateRoot == nil || c.pipelineFromBatch.StateRoot == nil ||
|
||||||
|
c.stats.Sync.LastBatch.StateRoot.Cmp(c.pipelineFromBatch.StateRoot) != 0) {
|
||||||
// There's been a reorg and the batch state root from which the
|
// There's been a reorg and the batch state root from which the
|
||||||
// pipeline was started has changed (probably because it was in
|
// pipeline was started has changed (probably because it was in
|
||||||
// a block that was discarded), and it was sent by a different
|
// a block that was discarded), and it was sent by a different
|
||||||
@@ -484,7 +494,7 @@ func (c *Coordinator) Start() {
|
|||||||
|
|
||||||
c.wg.Add(1)
|
c.wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
waitDuration := longWaitDuration
|
waitCh := time.After(longWaitDuration)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-c.ctx.Done():
|
case <-c.ctx.Done():
|
||||||
@@ -496,23 +506,24 @@ func (c *Coordinator) Start() {
|
|||||||
continue
|
continue
|
||||||
} else if err != nil {
|
} else if err != nil {
|
||||||
log.Errorw("Coordinator.handleMsg", "err", err)
|
log.Errorw("Coordinator.handleMsg", "err", err)
|
||||||
waitDuration = c.cfg.SyncRetryInterval
|
waitCh = time.After(c.cfg.SyncRetryInterval)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
case <-time.After(waitDuration):
|
case <-waitCh:
|
||||||
if !c.stats.Synced() {
|
if !c.stats.Synced() {
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := c.syncStats(c.ctx, &c.stats); c.ctx.Err() != nil {
|
if err := c.syncStats(c.ctx, &c.stats); c.ctx.Err() != nil {
|
||||||
|
waitCh = time.After(longWaitDuration)
|
||||||
continue
|
continue
|
||||||
} else if err != nil {
|
} else if err != nil {
|
||||||
log.Errorw("Coordinator.syncStats", "err", err)
|
log.Errorw("Coordinator.syncStats", "err", err)
|
||||||
waitDuration = c.cfg.SyncRetryInterval
|
waitCh = time.After(c.cfg.SyncRetryInterval)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
|
|||||||
@@ -2,7 +2,6 @@ package coordinator
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
"errors"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"math/big"
|
"math/big"
|
||||||
@@ -12,7 +11,6 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
ethCommon "github.com/ethereum/go-ethereum/common"
|
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||||
"github.com/ethereum/go-ethereum/core"
|
|
||||||
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
"github.com/hermeznetwork/hermez-node/batchbuilder"
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||||
@@ -568,8 +566,3 @@ func TestCoordinatorStress(t *testing.T) {
|
|||||||
// TODO: Test forgeBatch
|
// TODO: Test forgeBatch
|
||||||
// TODO: Test waitServerProof
|
// TODO: Test waitServerProof
|
||||||
// TODO: Test handleReorg
|
// TODO: Test handleReorg
|
||||||
|
|
||||||
func TestFoo(t *testing.T) {
|
|
||||||
a := tracerr.Wrap(fmt.Errorf("AAA: %w", core.ErrNonceTooLow))
|
|
||||||
fmt.Println(errors.Is(a, core.ErrNonceTooLow))
|
|
||||||
}
|
|
||||||
|
|||||||
@@ -29,6 +29,7 @@ type state struct {
|
|||||||
batchNum common.BatchNum
|
batchNum common.BatchNum
|
||||||
lastScheduledL1BatchBlockNum int64
|
lastScheduledL1BatchBlockNum int64
|
||||||
lastForgeL1TxsNum int64
|
lastForgeL1TxsNum int64
|
||||||
|
lastSlotForged int64
|
||||||
}
|
}
|
||||||
|
|
||||||
// Pipeline manages the forging of batches with parallel server proofs
|
// Pipeline manages the forging of batches with parallel server proofs
|
||||||
@@ -39,12 +40,10 @@ type Pipeline struct {
|
|||||||
|
|
||||||
// state
|
// state
|
||||||
state state
|
state state
|
||||||
// batchNum common.BatchNum
|
|
||||||
// lastScheduledL1BatchBlockNum int64
|
|
||||||
// lastForgeL1TxsNum int64
|
|
||||||
started bool
|
started bool
|
||||||
rw sync.RWMutex
|
rw sync.RWMutex
|
||||||
errAtBatchNum common.BatchNum
|
errAtBatchNum common.BatchNum
|
||||||
|
lastForgeTime time.Time
|
||||||
|
|
||||||
proversPool *ProversPool
|
proversPool *ProversPool
|
||||||
provers []prover.Client
|
provers []prover.Client
|
||||||
@@ -136,6 +135,7 @@ func (p *Pipeline) reset(batchNum common.BatchNum,
|
|||||||
batchNum: batchNum,
|
batchNum: batchNum,
|
||||||
lastForgeL1TxsNum: stats.Sync.LastForgeL1TxsNum,
|
lastForgeL1TxsNum: stats.Sync.LastForgeL1TxsNum,
|
||||||
lastScheduledL1BatchBlockNum: 0,
|
lastScheduledL1BatchBlockNum: 0,
|
||||||
|
lastSlotForged: -1,
|
||||||
}
|
}
|
||||||
p.stats = *stats
|
p.stats = *stats
|
||||||
p.vars = *vars
|
p.vars = *vars
|
||||||
@@ -207,6 +207,9 @@ func (p *Pipeline) handleForgeBatch(ctx context.Context, batchNum common.BatchNu
|
|||||||
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err,
|
log.Warnw("forgeBatch: scheduled L1Batch too early", "err", err,
|
||||||
"lastForgeL1TxsNum", p.state.lastForgeL1TxsNum,
|
"lastForgeL1TxsNum", p.state.lastForgeL1TxsNum,
|
||||||
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum)
|
"syncLastForgeL1TxsNum", p.stats.Sync.LastForgeL1TxsNum)
|
||||||
|
} else if tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||||
|
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||||
|
// no log
|
||||||
} else {
|
} else {
|
||||||
log.Errorw("forgeBatch", "err", err)
|
log.Errorw("forgeBatch", "err", err)
|
||||||
}
|
}
|
||||||
@@ -250,7 +253,7 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
|||||||
|
|
||||||
p.wg.Add(1)
|
p.wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
waitDuration := zeroDuration
|
waitCh := time.After(zeroDuration)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-p.ctx.Done():
|
case <-p.ctx.Done():
|
||||||
@@ -260,37 +263,42 @@ func (p *Pipeline) Start(batchNum common.BatchNum,
|
|||||||
case statsVars := <-p.statsVarsCh:
|
case statsVars := <-p.statsVarsCh:
|
||||||
p.stats = statsVars.Stats
|
p.stats = statsVars.Stats
|
||||||
p.syncSCVars(statsVars.Vars)
|
p.syncSCVars(statsVars.Vars)
|
||||||
case <-time.After(waitDuration):
|
case <-waitCh:
|
||||||
// Once errAtBatchNum != 0, we stop forging
|
// Once errAtBatchNum != 0, we stop forging
|
||||||
// batches because there's been an error and we
|
// batches because there's been an error and we
|
||||||
// wait for the pipeline to be stopped.
|
// wait for the pipeline to be stopped.
|
||||||
if p.getErrAtBatchNum() != 0 {
|
if p.getErrAtBatchNum() != 0 {
|
||||||
waitDuration = p.cfg.ForgeRetryInterval
|
waitCh = time.After(p.cfg.ForgeRetryInterval)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
batchNum = p.state.batchNum + 1
|
batchNum = p.state.batchNum + 1
|
||||||
batchInfo, err := p.handleForgeBatch(p.ctx, batchNum)
|
batchInfo, err := p.handleForgeBatch(p.ctx, batchNum)
|
||||||
if p.ctx.Err() != nil {
|
if p.ctx.Err() != nil {
|
||||||
|
waitCh = time.After(p.cfg.ForgeRetryInterval)
|
||||||
continue
|
continue
|
||||||
} else if tracerr.Unwrap(err) == errLastL1BatchNotSynced {
|
} else if tracerr.Unwrap(err) == errLastL1BatchNotSynced ||
|
||||||
waitDuration = p.cfg.ForgeRetryInterval
|
tracerr.Unwrap(err) == errForgeNoTxsBeforeDelay ||
|
||||||
|
tracerr.Unwrap(err) == errForgeBeforeDelay {
|
||||||
|
waitCh = time.After(p.cfg.ForgeRetryInterval)
|
||||||
continue
|
continue
|
||||||
} else if err != nil {
|
} else if err != nil {
|
||||||
p.setErrAtBatchNum(batchNum)
|
p.setErrAtBatchNum(batchNum)
|
||||||
waitDuration = p.cfg.ForgeRetryInterval
|
|
||||||
p.coord.SendMsg(p.ctx, MsgStopPipeline{
|
p.coord.SendMsg(p.ctx, MsgStopPipeline{
|
||||||
Reason: fmt.Sprintf(
|
Reason: fmt.Sprintf(
|
||||||
"Pipeline.handleForgBatch: %v", err),
|
"Pipeline.handleForgBatch: %v", err),
|
||||||
FailedBatchNum: batchNum,
|
FailedBatchNum: batchNum,
|
||||||
})
|
})
|
||||||
|
waitCh = time.After(p.cfg.ForgeRetryInterval)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
p.lastForgeTime = time.Now()
|
||||||
|
|
||||||
p.state.batchNum = batchNum
|
p.state.batchNum = batchNum
|
||||||
select {
|
select {
|
||||||
case batchChSentServerProof <- batchInfo:
|
case batchChSentServerProof <- batchInfo:
|
||||||
case <-p.ctx.Done():
|
case <-p.ctx.Done():
|
||||||
}
|
}
|
||||||
|
waitCh = time.After(zeroDuration)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}()
|
}()
|
||||||
@@ -376,8 +384,9 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
|||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
// Structure to accumulate data and metadata of the batch
|
// Structure to accumulate data and metadata of the batch
|
||||||
|
now := time.Now()
|
||||||
batchInfo = &BatchInfo{PipelineNum: p.num, BatchNum: batchNum}
|
batchInfo = &BatchInfo{PipelineNum: p.num, BatchNum: batchNum}
|
||||||
batchInfo.Debug.StartTimestamp = time.Now()
|
batchInfo.Debug.StartTimestamp = now
|
||||||
batchInfo.Debug.StartBlockNum = p.stats.Eth.LastBlock.Num + 1
|
batchInfo.Debug.StartBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||||
|
|
||||||
selectionCfg := &txselector.SelectionConfig{
|
selectionCfg := &txselector.SelectionConfig{
|
||||||
@@ -391,10 +400,17 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
|||||||
var auths [][]byte
|
var auths [][]byte
|
||||||
var coordIdxs []common.Idx
|
var coordIdxs []common.Idx
|
||||||
|
|
||||||
// TODO: If there are no txs and we are behind the timeout, skip
|
// Check if the slot is not yet fulfilled
|
||||||
// forging a batch and return a particular error that can be handleded
|
slotCommitted := false
|
||||||
// in the loop where handleForgeBatch is called to retry after an
|
if p.stats.Sync.Auction.CurrentSlot.ForgerCommitment ||
|
||||||
// interval
|
p.stats.Sync.Auction.CurrentSlot.SlotNum == p.state.lastSlotForged {
|
||||||
|
slotCommitted = true
|
||||||
|
}
|
||||||
|
|
||||||
|
// If we haven't reached the ForgeDelay, skip forging the batch
|
||||||
|
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeDelay {
|
||||||
|
return nil, errForgeBeforeDelay
|
||||||
|
}
|
||||||
|
|
||||||
// 1. Decide if we forge L2Tx or L1+L2Tx
|
// 1. Decide if we forge L2Tx or L1+L2Tx
|
||||||
if p.shouldL1L2Batch(batchInfo) {
|
if p.shouldL1L2Batch(batchInfo) {
|
||||||
@@ -412,9 +428,6 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
p.state.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBlock.Num + 1
|
|
||||||
p.state.lastForgeL1TxsNum++
|
|
||||||
} else {
|
} else {
|
||||||
// 2b: only L2 txs
|
// 2b: only L2 txs
|
||||||
coordIdxs, auths, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
coordIdxs, auths, l1CoordTxs, poolL2Txs, discardedL2Txs, err =
|
||||||
@@ -425,6 +438,43 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
|||||||
l1UserTxsExtra = nil
|
l1UserTxsExtra = nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// If there are no txs to forge, no l1UserTxs in the open queue to
|
||||||
|
// freeze, and we haven't reached the ForgeNoTxsDelay, skip forging the
|
||||||
|
// batch.
|
||||||
|
if slotCommitted && now.Sub(p.lastForgeTime) < p.cfg.ForgeNoTxsDelay {
|
||||||
|
noTxs := false
|
||||||
|
if len(l1UserTxsExtra) == 0 && len(l1CoordTxs) == 0 && len(poolL2Txs) == 0 {
|
||||||
|
if batchInfo.L1Batch {
|
||||||
|
// Query the L1UserTxs in the queue following
|
||||||
|
// the one we are trying to forge.
|
||||||
|
nextL1UserTxs, err := p.historyDB.GetUnforgedL1UserTxs(
|
||||||
|
p.state.lastForgeL1TxsNum + 1)
|
||||||
|
if err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
// If there are future L1UserTxs, we forge a
|
||||||
|
// batch to advance the queues and forge the
|
||||||
|
// L1UserTxs in the future. Otherwise, skip.
|
||||||
|
if len(nextL1UserTxs) == 0 {
|
||||||
|
noTxs = true
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
noTxs = true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if noTxs {
|
||||||
|
if err := p.txSelector.Reset(batchInfo.BatchNum-1, false); err != nil {
|
||||||
|
return nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
return nil, errForgeNoTxsBeforeDelay
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if batchInfo.L1Batch {
|
||||||
|
p.state.lastScheduledL1BatchBlockNum = p.stats.Eth.LastBlock.Num + 1
|
||||||
|
p.state.lastForgeL1TxsNum++
|
||||||
|
}
|
||||||
|
|
||||||
// 3. Save metadata from TxSelector output for BatchNum
|
// 3. Save metadata from TxSelector output for BatchNum
|
||||||
batchInfo.L1UserTxsExtra = l1UserTxsExtra
|
batchInfo.L1UserTxsExtra = l1UserTxsExtra
|
||||||
batchInfo.L1CoordTxs = l1CoordTxs
|
batchInfo.L1CoordTxs = l1CoordTxs
|
||||||
@@ -469,6 +519,8 @@ func (p *Pipeline) forgeBatch(batchNum common.BatchNum) (batchInfo *BatchInfo, e
|
|||||||
p.cfg.debugBatchStore(batchInfo)
|
p.cfg.debugBatchStore(batchInfo)
|
||||||
log.Infow("Pipeline: batch forged internally", "batch", batchInfo.BatchNum)
|
log.Infow("Pipeline: batch forged internally", "batch", batchInfo.BatchNum)
|
||||||
|
|
||||||
|
p.state.lastSlotForged = p.stats.Sync.Auction.CurrentSlot.SlotNum
|
||||||
|
|
||||||
return batchInfo, nil
|
return batchInfo, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -25,6 +25,14 @@ import (
|
|||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
func newBigInt(s string) *big.Int {
|
||||||
|
v, ok := new(big.Int).SetString(s, 10)
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Errorf("Can't set big.Int from %s", s))
|
||||||
|
}
|
||||||
|
return v
|
||||||
|
}
|
||||||
|
|
||||||
func TestPipelineShouldL1L2Batch(t *testing.T) {
|
func TestPipelineShouldL1L2Batch(t *testing.T) {
|
||||||
ethClientSetup := test.NewClientSetupExample()
|
ethClientSetup := test.NewClientSetupExample()
|
||||||
ethClientSetup.ChainID = big.NewInt(int64(chainID))
|
ethClientSetup.ChainID = big.NewInt(int64(chainID))
|
||||||
@@ -128,6 +136,11 @@ func preloadSync(t *testing.T, ethClient *test.Client, sync *synchronizer.Synchr
|
|||||||
blocks, err := tc.GenerateBlocksFromInstructions(set)
|
blocks, err := tc.GenerateBlocksFromInstructions(set)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
require.NotNil(t, blocks)
|
require.NotNil(t, blocks)
|
||||||
|
// Set StateRoots for batches manually (til doesn't set it)
|
||||||
|
blocks[0].Rollup.Batches[0].Batch.StateRoot =
|
||||||
|
newBigInt("0")
|
||||||
|
blocks[0].Rollup.Batches[1].Batch.StateRoot =
|
||||||
|
newBigInt("10941365282189107056349764238909072001483688090878331371699519307087372995595")
|
||||||
|
|
||||||
ethAddTokens(blocks, ethClient)
|
ethAddTokens(blocks, ethClient)
|
||||||
err = ethClient.CtlAddBlocks(blocks)
|
err = ethClient.CtlAddBlocks(blocks)
|
||||||
|
|||||||
@@ -13,13 +13,23 @@ import (
|
|||||||
|
|
||||||
// PurgerCfg is the purger configuration
|
// PurgerCfg is the purger configuration
|
||||||
type PurgerCfg struct {
|
type PurgerCfg struct {
|
||||||
// PurgeBatchDelay is the delay between batches to purge outdated transactions
|
// PurgeBatchDelay is the delay between batches to purge outdated
|
||||||
|
// transactions. Oudated L2Txs are those that have been forged or
|
||||||
|
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||||
|
// that have been in the pool for longer than TTL once there are
|
||||||
|
// MaxTxs.
|
||||||
PurgeBatchDelay int64
|
PurgeBatchDelay int64
|
||||||
// InvalidateBatchDelay is the delay between batches to mark invalid transactions
|
// InvalidateBatchDelay is the delay between batches to mark invalid
|
||||||
|
// transactions due to nonce lower than the account nonce.
|
||||||
InvalidateBatchDelay int64
|
InvalidateBatchDelay int64
|
||||||
// PurgeBlockDelay is the delay between blocks to purge outdated transactions
|
// PurgeBlockDelay is the delay between blocks to purge outdated
|
||||||
|
// transactions. Oudated L2Txs are those that have been forged or
|
||||||
|
// marked as invalid for longer than the SafetyPeriod and pending L2Txs
|
||||||
|
// that have been in the pool for longer than TTL once there are
|
||||||
|
// MaxTxs.
|
||||||
PurgeBlockDelay int64
|
PurgeBlockDelay int64
|
||||||
// InvalidateBlockDelay is the delay between blocks to mark invalid transactions
|
// InvalidateBlockDelay is the delay between blocks to mark invalid
|
||||||
|
// transactions due to nonce lower than the account nonce.
|
||||||
InvalidateBlockDelay int64
|
InvalidateBlockDelay int64
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -49,8 +49,6 @@ type TxManager struct {
|
|||||||
// accNextNonce is the nonce that we should use to send the next tx.
|
// accNextNonce is the nonce that we should use to send the next tx.
|
||||||
// In some cases this will be a reused nonce of an already pending tx.
|
// In some cases this will be a reused nonce of an already pending tx.
|
||||||
accNextNonce uint64
|
accNextNonce uint64
|
||||||
// accPendingNonce is the pending nonce of the account due to pending txs
|
|
||||||
// accPendingNonce uint64
|
|
||||||
|
|
||||||
lastSentL1BatchBlockNum int64
|
lastSentL1BatchBlockNum int64
|
||||||
}
|
}
|
||||||
@@ -70,14 +68,6 @@ func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterfac
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, err
|
return nil, err
|
||||||
}
|
}
|
||||||
// accPendingNonce, err := ethClient.EthPendingNonceAt(ctx, *address)
|
|
||||||
// if err != nil {
|
|
||||||
// return nil, err
|
|
||||||
// }
|
|
||||||
// if accNonce != accPendingNonce {
|
|
||||||
// return nil, tracerr.Wrap(fmt.Errorf("currentNonce (%v) != accPendingNonce (%v)",
|
|
||||||
// accNonce, accPendingNonce))
|
|
||||||
// }
|
|
||||||
log.Infow("TxManager started", "nonce", accNonce)
|
log.Infow("TxManager started", "nonce", accNonce)
|
||||||
return &TxManager{
|
return &TxManager{
|
||||||
cfg: *cfg,
|
cfg: *cfg,
|
||||||
@@ -99,7 +89,6 @@ func NewTxManager(ctx context.Context, cfg *Config, ethClient eth.ClientInterfac
|
|||||||
queue: NewQueue(),
|
queue: NewQueue(),
|
||||||
accNonce: accNonce,
|
accNonce: accNonce,
|
||||||
accNextNonce: accNonce,
|
accNextNonce: accNonce,
|
||||||
// accPendingNonce: accPendingNonce,
|
|
||||||
}, nil
|
}, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -139,11 +128,14 @@ func (t *TxManager) NewAuth(ctx context.Context) (*bind.TransactOpts, error) {
|
|||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
if t.cfg.GasPriceIncPerc != 0 {
|
||||||
inc := new(big.Int).Set(gasPrice)
|
inc := new(big.Int).Set(gasPrice)
|
||||||
// TODO: Replace this by a value of percentage
|
inc.Mul(inc, new(big.Int).SetInt64(t.cfg.GasPriceIncPerc))
|
||||||
const gasPriceDiv = 100
|
// nolint reason: to calculate percentages we use 100
|
||||||
inc.Div(inc, new(big.Int).SetUint64(gasPriceDiv))
|
inc.Div(inc, new(big.Int).SetUint64(100)) //nolint:gomnd
|
||||||
gasPrice.Add(gasPrice, inc)
|
gasPrice.Add(gasPrice, inc)
|
||||||
|
}
|
||||||
|
|
||||||
// log.Debugw("TxManager: transaction metadata", "gasPrice", gasPrice)
|
// log.Debugw("TxManager: transaction metadata", "gasPrice", gasPrice)
|
||||||
|
|
||||||
auth, err := bind.NewKeyStoreTransactorWithChainID(t.ethClient.EthKeyStore(), t.account, t.chainID)
|
auth, err := bind.NewKeyStoreTransactorWithChainID(t.ethClient.EthKeyStore(), t.account, t.chainID)
|
||||||
@@ -152,6 +144,13 @@ func (t *TxManager) NewAuth(ctx context.Context) (*bind.TransactOpts, error) {
|
|||||||
}
|
}
|
||||||
auth.Value = big.NewInt(0) // in wei
|
auth.Value = big.NewInt(0) // in wei
|
||||||
// TODO: Calculate GasLimit based on the contents of the ForgeBatchArgs
|
// TODO: Calculate GasLimit based on the contents of the ForgeBatchArgs
|
||||||
|
// This requires a function that estimates the gas usage of the
|
||||||
|
// forgeBatch call based on the contents of the ForgeBatch args:
|
||||||
|
// - length of l2txs
|
||||||
|
// - length of l1Usertxs
|
||||||
|
// - length of l1CoordTxs with authorization signature
|
||||||
|
// - length of l1CoordTxs without authoriation signature
|
||||||
|
// - etc.
|
||||||
auth.GasLimit = 1000000
|
auth.GasLimit = 1000000
|
||||||
auth.GasPrice = gasPrice
|
auth.GasPrice = gasPrice
|
||||||
auth.Nonce = nil
|
auth.Nonce = nil
|
||||||
@@ -184,7 +183,7 @@ func (t *TxManager) shouldSendRollupForgeBatch(batchInfo *BatchInfo) error {
|
|||||||
func addPerc(v *big.Int, p int64) *big.Int {
|
func addPerc(v *big.Int, p int64) *big.Int {
|
||||||
r := new(big.Int).Set(v)
|
r := new(big.Int).Set(v)
|
||||||
r.Mul(r, big.NewInt(p))
|
r.Mul(r, big.NewInt(p))
|
||||||
// nolint reason: to calculate percetnages we divide by 100
|
// nolint reason: to calculate percentages we divide by 100
|
||||||
r.Div(r, big.NewInt(100)) //nolit:gomnd
|
r.Div(r, big.NewInt(100)) //nolit:gomnd
|
||||||
return r.Add(v, r)
|
return r.Add(v, r)
|
||||||
}
|
}
|
||||||
@@ -261,7 +260,6 @@ func (t *TxManager) sendRollupForgeBatch(ctx context.Context, batchInfo *BatchIn
|
|||||||
batchInfo.Debug.StartTimestamp).Seconds()
|
batchInfo.Debug.StartTimestamp).Seconds()
|
||||||
t.cfg.debugBatchStore(batchInfo)
|
t.cfg.debugBatchStore(batchInfo)
|
||||||
|
|
||||||
// t.lastPendingBatch = batchInfo.BatchNum
|
|
||||||
if !resend {
|
if !resend {
|
||||||
if batchInfo.L1Batch {
|
if batchInfo.L1Batch {
|
||||||
t.lastSentL1BatchBlockNum = t.stats.Eth.LastBlock.Num + 1
|
t.lastSentL1BatchBlockNum = t.stats.Eth.LastBlock.Num + 1
|
||||||
@@ -314,14 +312,15 @@ func (t *TxManager) handleReceipt(ctx context.Context, batchInfo *BatchInfo) (*i
|
|||||||
}
|
}
|
||||||
if receipt.Status == types.ReceiptStatusFailed {
|
if receipt.Status == types.ReceiptStatusFailed {
|
||||||
batchInfo.Debug.Status = StatusFailed
|
batchInfo.Debug.Status = StatusFailed
|
||||||
t.cfg.debugBatchStore(batchInfo)
|
|
||||||
_, err := t.ethClient.EthCall(ctx, batchInfo.EthTx, receipt.BlockNumber)
|
_, err := t.ethClient.EthCall(ctx, batchInfo.EthTx, receipt.BlockNumber)
|
||||||
log.Warnw("TxManager receipt status is failed", "tx", receipt.TxHash,
|
log.Warnw("TxManager receipt status is failed", "tx", receipt.TxHash,
|
||||||
"batch", batchInfo.BatchNum, "block", receipt.BlockNumber.Int64(),
|
"batch", batchInfo.BatchNum, "block", receipt.BlockNumber.Int64(),
|
||||||
"err", err)
|
"err", err)
|
||||||
|
batchInfo.EthTxErr = err
|
||||||
if batchInfo.BatchNum <= t.lastSuccessBatch {
|
if batchInfo.BatchNum <= t.lastSuccessBatch {
|
||||||
t.lastSuccessBatch = batchInfo.BatchNum - 1
|
t.lastSuccessBatch = batchInfo.BatchNum - 1
|
||||||
}
|
}
|
||||||
|
t.cfg.debugBatchStore(batchInfo)
|
||||||
return nil, tracerr.Wrap(fmt.Errorf(
|
return nil, tracerr.Wrap(fmt.Errorf(
|
||||||
"ethereum transaction receipt status is failed: %w", err))
|
"ethereum transaction receipt status is failed: %w", err))
|
||||||
} else if receipt.Status == types.ReceiptStatusSuccessful {
|
} else if receipt.Status == types.ReceiptStatusSuccessful {
|
||||||
@@ -417,7 +416,7 @@ func (q *Queue) Push(batchInfo *BatchInfo) {
|
|||||||
|
|
||||||
// Run the TxManager
|
// Run the TxManager
|
||||||
func (t *TxManager) Run(ctx context.Context) {
|
func (t *TxManager) Run(ctx context.Context) {
|
||||||
waitDuration := longWaitDuration
|
waitCh := time.After(longWaitDuration)
|
||||||
|
|
||||||
var statsVars statsVars
|
var statsVars statsVars
|
||||||
select {
|
select {
|
||||||
@@ -472,11 +471,11 @@ func (t *TxManager) Run(ctx context.Context) {
|
|||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
t.queue.Push(batchInfo)
|
t.queue.Push(batchInfo)
|
||||||
waitDuration = t.cfg.TxManagerCheckInterval
|
waitCh = time.After(t.cfg.TxManagerCheckInterval)
|
||||||
case <-time.After(waitDuration):
|
case <-waitCh:
|
||||||
queuePosition, batchInfo := t.queue.Next()
|
queuePosition, batchInfo := t.queue.Next()
|
||||||
if batchInfo == nil {
|
if batchInfo == nil {
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if err := t.checkEthTransactionReceipt(ctx, batchInfo); ctx.Err() != nil {
|
if err := t.checkEthTransactionReceipt(ctx, batchInfo); ctx.Err() != nil {
|
||||||
@@ -539,7 +538,6 @@ func (t *TxManager) Run(ctx context.Context) {
|
|||||||
|
|
||||||
func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
||||||
next := 0
|
next := 0
|
||||||
// batchNum := 0
|
|
||||||
for {
|
for {
|
||||||
batchInfo := t.queue.At(next)
|
batchInfo := t.queue.At(next)
|
||||||
if batchInfo == nil {
|
if batchInfo == nil {
|
||||||
@@ -572,7 +570,6 @@ func (t *TxManager) removeBadBatchInfos(ctx context.Context) error {
|
|||||||
// from the queue
|
// from the queue
|
||||||
if confirm == nil {
|
if confirm == nil {
|
||||||
if batchInfo.PipelineNum < t.minPipelineNum {
|
if batchInfo.PipelineNum < t.minPipelineNum {
|
||||||
// batchNum++
|
|
||||||
t.queue.Remove(next)
|
t.queue.Remove(next)
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -621,10 +621,10 @@ func TestTxs(t *testing.T) {
|
|||||||
assert.Equal(t, common.TxTypeExit, dbL2Txs[3].Type)
|
assert.Equal(t, common.TxTypeExit, dbL2Txs[3].Type)
|
||||||
|
|
||||||
// Tx ID
|
// Tx ID
|
||||||
assert.Equal(t, "0x02d709307533c4e3c03f20751fc4d72bc18b225d14f9616525540a64342c7c350d", dbL2Txs[0].TxID.String())
|
assert.Equal(t, "0x024e555248100b69a8aabf6d31719b9fe8a60dcc6c3407904a93c8d2d9ade18ee5", dbL2Txs[0].TxID.String())
|
||||||
assert.Equal(t, "0x02e88bc5503f282cca045847668511290e642410a459bb67b1fafcd1b6097c149c", dbL2Txs[1].TxID.String())
|
assert.Equal(t, "0x021ae87ca34d50ff35d98dfc0d7c95f2bf2e4ffeebb82ea71f43a8b0dfa5d36d89", dbL2Txs[1].TxID.String())
|
||||||
assert.Equal(t, "0x027911262b43315c0b24942a02fe228274b6e4d57a476bfcdd7a324b3091362c7d", dbL2Txs[2].TxID.String())
|
assert.Equal(t, "0x024abce7f3f2382dc520ed557593f11dea1ee197e55b60402e664facc27aa19774", dbL2Txs[2].TxID.String())
|
||||||
assert.Equal(t, "0x02f572b63f2a5c302e1b9337ea6944bfbac3d199e4ddd262b5a53759c72ec10ee6", dbL2Txs[3].TxID.String())
|
assert.Equal(t, "0x02f921ad9e7a6e59606570fe12a7dde0e36014197de0363b9b45e5097d6f2b1dd0", dbL2Txs[3].TxID.String())
|
||||||
|
|
||||||
// Tx From and To IDx
|
// Tx From and To IDx
|
||||||
assert.Equal(t, dbL2Txs[0].ToIdx, dbL2Txs[2].FromIdx)
|
assert.Equal(t, dbL2Txs[0].ToIdx, dbL2Txs[2].FromIdx)
|
||||||
|
|||||||
@@ -275,7 +275,8 @@ func (s *StateDB) GetAccount(idx common.Idx) (*common.Account, error) {
|
|||||||
return GetAccountInTreeDB(s.db.DB(), idx)
|
return GetAccountInTreeDB(s.db.DB(), idx)
|
||||||
}
|
}
|
||||||
|
|
||||||
func accountsIter(db db.Storage, fn func(a *common.Account) (bool, error)) error {
|
// AccountsIter iterates over all the accounts in db, calling fn for each one
|
||||||
|
func AccountsIter(db db.Storage, fn func(a *common.Account) (bool, error)) error {
|
||||||
idxDB := db.WithPrefix(PrefixKeyIdx)
|
idxDB := db.WithPrefix(PrefixKeyIdx)
|
||||||
if err := idxDB.Iterate(func(k []byte, v []byte) (bool, error) {
|
if err := idxDB.Iterate(func(k []byte, v []byte) (bool, error) {
|
||||||
idx, err := common.IdxFromBytes(k)
|
idx, err := common.IdxFromBytes(k)
|
||||||
|
|||||||
@@ -13,9 +13,9 @@ import (
|
|||||||
"github.com/hermeznetwork/hermez-node/log"
|
"github.com/hermeznetwork/hermez-node/log"
|
||||||
"github.com/hermeznetwork/tracerr"
|
"github.com/hermeznetwork/tracerr"
|
||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
"github.com/marusama/semaphore/v2"
|
|
||||||
migrate "github.com/rubenv/sql-migrate"
|
migrate "github.com/rubenv/sql-migrate"
|
||||||
"github.com/russross/meddler"
|
"github.com/russross/meddler"
|
||||||
|
"golang.org/x/sync/semaphore"
|
||||||
)
|
)
|
||||||
|
|
||||||
var migrations *migrate.PackrMigrationSource
|
var migrations *migrate.PackrMigrationSource
|
||||||
@@ -89,14 +89,14 @@ func InitSQLDB(port int, host, user, password, name string) (*sqlx.DB, error) {
|
|||||||
|
|
||||||
// APIConnectionController is used to limit the SQL open connections used by the API
|
// APIConnectionController is used to limit the SQL open connections used by the API
|
||||||
type APIConnectionController struct {
|
type APIConnectionController struct {
|
||||||
smphr semaphore.Semaphore
|
smphr *semaphore.Weighted
|
||||||
timeout time.Duration
|
timeout time.Duration
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewAPICnnectionController initialize APIConnectionController
|
// NewAPICnnectionController initialize APIConnectionController
|
||||||
func NewAPICnnectionController(maxConnections int, timeout time.Duration) *APIConnectionController {
|
func NewAPICnnectionController(maxConnections int, timeout time.Duration) *APIConnectionController {
|
||||||
return &APIConnectionController{
|
return &APIConnectionController{
|
||||||
smphr: semaphore.New(maxConnections),
|
smphr: semaphore.NewWeighted(int64(maxConnections)),
|
||||||
timeout: timeout,
|
timeout: timeout,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -316,7 +316,7 @@ func NewRollupClient(client *EthereumClient, address ethCommon.Address, tokenHEZ
|
|||||||
}
|
}
|
||||||
consts, err := c.RollupConstants()
|
consts, err := c.RollupConstants()
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(fmt.Errorf("RollupConstants at %v: %w", address, err))
|
||||||
}
|
}
|
||||||
c.consts = consts
|
c.consts = consts
|
||||||
return c, nil
|
return c, nil
|
||||||
@@ -462,11 +462,11 @@ func (c *RollupClient) RollupL1UserTxERC20ETH(fromBJJ babyjub.PublicKeyComp, fro
|
|||||||
}
|
}
|
||||||
fromIdxBig := big.NewInt(fromIdx)
|
fromIdxBig := big.NewInt(fromIdx)
|
||||||
toIdxBig := big.NewInt(toIdx)
|
toIdxBig := big.NewInt(toIdx)
|
||||||
depositAmountF, err := common.NewFloat16(depositAmount)
|
depositAmountF, err := common.NewFloat40(depositAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
amountF, err := common.NewFloat16(amount)
|
amountF, err := common.NewFloat40(amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -497,11 +497,11 @@ func (c *RollupClient) RollupL1UserTxERC20Permit(fromBJJ babyjub.PublicKeyComp,
|
|||||||
}
|
}
|
||||||
fromIdxBig := big.NewInt(fromIdx)
|
fromIdxBig := big.NewInt(fromIdx)
|
||||||
toIdxBig := big.NewInt(toIdx)
|
toIdxBig := big.NewInt(toIdx)
|
||||||
depositAmountF, err := common.NewFloat16(depositAmount)
|
depositAmountF, err := common.NewFloat40(depositAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
amountF, err := common.NewFloat16(amount)
|
amountF, err := common.NewFloat40(amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
@@ -939,9 +939,9 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
|||||||
FeeIdxCoordinator: []common.Idx{},
|
FeeIdxCoordinator: []common.Idx{},
|
||||||
}
|
}
|
||||||
nLevels := c.consts.Verifiers[rollupForgeBatchArgs.VerifierIdx].NLevels
|
nLevels := c.consts.Verifiers[rollupForgeBatchArgs.VerifierIdx].NLevels
|
||||||
lenL1L2TxsBytes := int((nLevels/8)*2 + 2 + 1)
|
lenL1L2TxsBytes := int((nLevels/8)*2 + common.Float40BytesLength + 1)
|
||||||
numBytesL1TxUser := int(l1UserTxsLen) * lenL1L2TxsBytes
|
numBytesL1TxUser := int(l1UserTxsLen) * lenL1L2TxsBytes
|
||||||
numTxsL1Coord := len(aux.EncodedL1CoordinatorTx) / common.L1CoordinatorTxBytesLen
|
numTxsL1Coord := len(aux.EncodedL1CoordinatorTx) / common.RollupConstL1CoordinatorTotalBytes
|
||||||
numBytesL1TxCoord := numTxsL1Coord * lenL1L2TxsBytes
|
numBytesL1TxCoord := numTxsL1Coord * lenL1L2TxsBytes
|
||||||
numBeginL2Tx := numBytesL1TxCoord + numBytesL1TxUser
|
numBeginL2Tx := numBytesL1TxCoord + numBytesL1TxUser
|
||||||
l1UserTxsData := []byte{}
|
l1UserTxsData := []byte{}
|
||||||
@@ -968,7 +968,7 @@ func (c *RollupClient) RollupForgeBatchArgs(ethTxHash ethCommon.Hash, l1UserTxsL
|
|||||||
rollupForgeBatchArgs.L2TxsData = append(rollupForgeBatchArgs.L2TxsData, *l2Tx)
|
rollupForgeBatchArgs.L2TxsData = append(rollupForgeBatchArgs.L2TxsData, *l2Tx)
|
||||||
}
|
}
|
||||||
for i := 0; i < numTxsL1Coord; i++ {
|
for i := 0; i < numTxsL1Coord; i++ {
|
||||||
bytesL1Coordinator := aux.EncodedL1CoordinatorTx[i*common.L1CoordinatorTxBytesLen : (i+1)*common.L1CoordinatorTxBytesLen]
|
bytesL1Coordinator := aux.EncodedL1CoordinatorTx[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes]
|
||||||
var signature []byte
|
var signature []byte
|
||||||
v := bytesL1Coordinator[0]
|
v := bytesL1Coordinator[0]
|
||||||
s := bytesL1Coordinator[1:33]
|
s := bytesL1Coordinator[1:33]
|
||||||
|
|||||||
@@ -131,9 +131,9 @@ func TestRollupForgeBatch(t *testing.T) {
|
|||||||
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
args.FeeIdxCoordinator = []common.Idx{} // When encoded, 64 times the 0 idx means that no idx to collect fees is specified.
|
||||||
l1CoordinatorBytes, err := hex.DecodeString("1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
l1CoordinatorBytes, err := hex.DecodeString("1c660323607bb113e586183609964a333d07ebe4bef3be82ec13af453bae9590bd7711cdb6abf42f176eadfbe5506fbef5e092e5543733f91b0061d9a7747fa10694a915a6470fa230de387b51e6f4db0b09787867778687b55197ad6d6a86eac000000001")
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
numTxsL1 := len(l1CoordinatorBytes) / common.L1CoordinatorTxBytesLen
|
numTxsL1 := len(l1CoordinatorBytes) / common.RollupConstL1CoordinatorTotalBytes
|
||||||
for i := 0; i < numTxsL1; i++ {
|
for i := 0; i < numTxsL1; i++ {
|
||||||
bytesL1Coordinator := l1CoordinatorBytes[i*common.L1CoordinatorTxBytesLen : (i+1)*common.L1CoordinatorTxBytesLen]
|
bytesL1Coordinator := l1CoordinatorBytes[i*common.RollupConstL1CoordinatorTotalBytes : (i+1)*common.RollupConstL1CoordinatorTotalBytes]
|
||||||
var signature []byte
|
var signature []byte
|
||||||
v := bytesL1Coordinator[0]
|
v := bytesL1Coordinator[0]
|
||||||
s := bytesL1Coordinator[1:33]
|
s := bytesL1Coordinator[1:33]
|
||||||
|
|||||||
5
go.mod
5
go.mod
@@ -6,7 +6,6 @@ require (
|
|||||||
github.com/BurntSushi/toml v0.3.1
|
github.com/BurntSushi/toml v0.3.1
|
||||||
github.com/dghubble/sling v1.3.0
|
github.com/dghubble/sling v1.3.0
|
||||||
github.com/ethereum/go-ethereum v1.9.25
|
github.com/ethereum/go-ethereum v1.9.25
|
||||||
github.com/fjl/memsize v0.0.0-20190710130421-bcb5799ab5e5 // indirect
|
|
||||||
github.com/getkin/kin-openapi v0.22.0
|
github.com/getkin/kin-openapi v0.22.0
|
||||||
github.com/gin-contrib/cors v1.3.1
|
github.com/gin-contrib/cors v1.3.1
|
||||||
github.com/gin-gonic/gin v1.5.0
|
github.com/gin-gonic/gin v1.5.0
|
||||||
@@ -18,19 +17,17 @@ require (
|
|||||||
github.com/jmoiron/sqlx v1.2.1-0.20200615141059-0794cb1f47ee
|
github.com/jmoiron/sqlx v1.2.1-0.20200615141059-0794cb1f47ee
|
||||||
github.com/joho/godotenv v1.3.0
|
github.com/joho/godotenv v1.3.0
|
||||||
github.com/lib/pq v1.8.0
|
github.com/lib/pq v1.8.0
|
||||||
github.com/marusama/semaphore/v2 v2.4.1
|
|
||||||
github.com/mattn/go-sqlite3 v2.0.3+incompatible
|
github.com/mattn/go-sqlite3 v2.0.3+incompatible
|
||||||
github.com/miguelmota/go-ethereum-hdwallet v0.0.0-20200123000308-a60dcd172b4c
|
github.com/miguelmota/go-ethereum-hdwallet v0.0.0-20200123000308-a60dcd172b4c
|
||||||
github.com/mitchellh/copystructure v1.0.0
|
github.com/mitchellh/copystructure v1.0.0
|
||||||
github.com/mitchellh/mapstructure v1.3.0
|
github.com/mitchellh/mapstructure v1.3.0
|
||||||
github.com/rubenv/sql-migrate v0.0.0-20200616145509-8d140a17f351
|
github.com/rubenv/sql-migrate v0.0.0-20200616145509-8d140a17f351
|
||||||
github.com/russross/meddler v1.0.0
|
github.com/russross/meddler v1.0.0
|
||||||
github.com/sirupsen/logrus v1.5.0 // indirect
|
|
||||||
github.com/status-im/keycard-go v0.0.0-20190424133014-d95853db0f48 // indirect
|
|
||||||
github.com/stretchr/testify v1.6.1
|
github.com/stretchr/testify v1.6.1
|
||||||
github.com/urfave/cli/v2 v2.2.0
|
github.com/urfave/cli/v2 v2.2.0
|
||||||
go.uber.org/zap v1.16.0
|
go.uber.org/zap v1.16.0
|
||||||
golang.org/x/crypto v0.0.0-20201217014255-9d1352758620
|
golang.org/x/crypto v0.0.0-20201217014255-9d1352758620
|
||||||
golang.org/x/net v0.0.0-20200822124328-c89045814202
|
golang.org/x/net v0.0.0-20200822124328-c89045814202
|
||||||
|
golang.org/x/sync v0.0.0-20201020160332-67f06af15bc9
|
||||||
gopkg.in/go-playground/validator.v9 v9.29.1
|
gopkg.in/go-playground/validator.v9 v9.29.1
|
||||||
)
|
)
|
||||||
|
|||||||
3
go.sum
3
go.sum
@@ -421,9 +421,6 @@ github.com/markbates/oncer v1.0.0 h1:E83IaVAHygyndzPimgUYJjbshhDTALZyXxvk9FOlQRY
|
|||||||
github.com/markbates/oncer v1.0.0/go.mod h1:Z59JA581E9GP6w96jai+TGqafHPW+cPfRxz2aSZ0mcI=
|
github.com/markbates/oncer v1.0.0/go.mod h1:Z59JA581E9GP6w96jai+TGqafHPW+cPfRxz2aSZ0mcI=
|
||||||
github.com/markbates/safe v1.0.1 h1:yjZkbvRM6IzKj9tlu/zMJLS0n/V351OZWRnF3QfaUxI=
|
github.com/markbates/safe v1.0.1 h1:yjZkbvRM6IzKj9tlu/zMJLS0n/V351OZWRnF3QfaUxI=
|
||||||
github.com/markbates/safe v1.0.1/go.mod h1:nAqgmRi7cY2nqMc92/bSEeQA+R4OheNU2T1kNSCBdG0=
|
github.com/markbates/safe v1.0.1/go.mod h1:nAqgmRi7cY2nqMc92/bSEeQA+R4OheNU2T1kNSCBdG0=
|
||||||
github.com/marusama/semaphore v0.0.0-20190110074507-6952cef993b2 h1:sq+a5mb8zHbmHhrIH06oqIMGsanjpbxNgxEgZVfgpvQ=
|
|
||||||
github.com/marusama/semaphore/v2 v2.4.1 h1:Y29DhhFMvreVgoqF9EtaSJAF9t2E7Sk7i5VW81sqB8I=
|
|
||||||
github.com/marusama/semaphore/v2 v2.4.1/go.mod h1:z9nMiNUekt/LTpTUQdpp+4sJeYqUGpwMHfW0Z8V8fnQ=
|
|
||||||
github.com/mattn/go-colorable v0.0.9/go.mod h1:9vuHe8Xs5qXnSaW/c/ABM9alt+Vo+STaOChaDxuIBZU=
|
github.com/mattn/go-colorable v0.0.9/go.mod h1:9vuHe8Xs5qXnSaW/c/ABM9alt+Vo+STaOChaDxuIBZU=
|
||||||
github.com/mattn/go-colorable v0.1.0/go.mod h1:9vuHe8Xs5qXnSaW/c/ABM9alt+Vo+STaOChaDxuIBZU=
|
github.com/mattn/go-colorable v0.1.0/go.mod h1:9vuHe8Xs5qXnSaW/c/ABM9alt+Vo+STaOChaDxuIBZU=
|
||||||
github.com/mattn/go-colorable v0.1.2 h1:/bC9yWikZXAL9uJdulbSfyVNIR3n3trXl+v8+1sx8mU=
|
github.com/mattn/go-colorable v0.1.2 h1:/bC9yWikZXAL9uJdulbSfyVNIR3n3trXl+v8+1sx8mU=
|
||||||
|
|||||||
14
node/node.go
14
node/node.go
@@ -103,8 +103,8 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
var keyStore *ethKeystore.KeyStore
|
var keyStore *ethKeystore.KeyStore
|
||||||
if mode == ModeCoordinator {
|
if mode == ModeCoordinator {
|
||||||
ethCfg = eth.EthereumConfig{
|
ethCfg = eth.EthereumConfig{
|
||||||
CallGasLimit: cfg.Coordinator.EthClient.CallGasLimit,
|
CallGasLimit: 0, // cfg.Coordinator.EthClient.CallGasLimit,
|
||||||
GasPriceDiv: cfg.Coordinator.EthClient.GasPriceDiv,
|
GasPriceDiv: 0, // cfg.Coordinator.EthClient.GasPriceDiv,
|
||||||
}
|
}
|
||||||
|
|
||||||
scryptN := ethKeystore.StandardScryptN
|
scryptN := ethKeystore.StandardScryptN
|
||||||
@@ -299,12 +299,15 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
ConfirmBlocks: cfg.Coordinator.ConfirmBlocks,
|
ConfirmBlocks: cfg.Coordinator.ConfirmBlocks,
|
||||||
L1BatchTimeoutPerc: cfg.Coordinator.L1BatchTimeoutPerc,
|
L1BatchTimeoutPerc: cfg.Coordinator.L1BatchTimeoutPerc,
|
||||||
ForgeRetryInterval: cfg.Coordinator.ForgeRetryInterval.Duration,
|
ForgeRetryInterval: cfg.Coordinator.ForgeRetryInterval.Duration,
|
||||||
|
ForgeDelay: cfg.Coordinator.ForgeDelay.Duration,
|
||||||
|
ForgeNoTxsDelay: cfg.Coordinator.ForgeNoTxsDelay.Duration,
|
||||||
SyncRetryInterval: cfg.Coordinator.SyncRetryInterval.Duration,
|
SyncRetryInterval: cfg.Coordinator.SyncRetryInterval.Duration,
|
||||||
EthClientAttempts: cfg.Coordinator.EthClient.Attempts,
|
EthClientAttempts: cfg.Coordinator.EthClient.Attempts,
|
||||||
EthClientAttemptsDelay: cfg.Coordinator.EthClient.AttemptsDelay.Duration,
|
EthClientAttemptsDelay: cfg.Coordinator.EthClient.AttemptsDelay.Duration,
|
||||||
EthNoReuseNonce: cfg.Coordinator.EthClient.NoReuseNonce,
|
EthNoReuseNonce: cfg.Coordinator.EthClient.NoReuseNonce,
|
||||||
EthTxResendTimeout: cfg.Coordinator.EthClient.TxResendTimeout.Duration,
|
EthTxResendTimeout: cfg.Coordinator.EthClient.TxResendTimeout.Duration,
|
||||||
MaxGasPrice: cfg.Coordinator.EthClient.MaxGasPrice,
|
MaxGasPrice: cfg.Coordinator.EthClient.MaxGasPrice,
|
||||||
|
GasPriceIncPerc: cfg.Coordinator.EthClient.GasPriceIncPerc,
|
||||||
TxManagerCheckInterval: cfg.Coordinator.EthClient.CheckLoopInterval.Duration,
|
TxManagerCheckInterval: cfg.Coordinator.EthClient.CheckLoopInterval.Duration,
|
||||||
DebugBatchPath: cfg.Coordinator.Debug.BatchPath,
|
DebugBatchPath: cfg.Coordinator.Debug.BatchPath,
|
||||||
Purger: coordinator.PurgerCfg{
|
Purger: coordinator.PurgerCfg{
|
||||||
@@ -373,7 +376,7 @@ func NewNode(mode Mode, cfg *config.Node) (*Node, error) {
|
|||||||
}
|
}
|
||||||
var debugAPI *debugapi.DebugAPI
|
var debugAPI *debugapi.DebugAPI
|
||||||
if cfg.Debug.APIAddress != "" {
|
if cfg.Debug.APIAddress != "" {
|
||||||
debugAPI = debugapi.NewDebugAPI(cfg.Debug.APIAddress, stateDB, sync)
|
debugAPI = debugapi.NewDebugAPI(cfg.Debug.APIAddress, historyDB, stateDB, sync)
|
||||||
}
|
}
|
||||||
priceUpdater, err := priceupdater.NewPriceUpdater(cfg.PriceUpdater.URL,
|
priceUpdater, err := priceupdater.NewPriceUpdater(cfg.PriceUpdater.URL,
|
||||||
priceupdater.APIType(cfg.PriceUpdater.Type), historyDB)
|
priceupdater.APIType(cfg.PriceUpdater.Type), historyDB)
|
||||||
@@ -544,6 +547,9 @@ func (n *Node) syncLoopFn(ctx context.Context, lastBlock *common.Block) (*common
|
|||||||
WDelayer: blockData.WDelayer.Vars,
|
WDelayer: blockData.WDelayer.Vars,
|
||||||
}
|
}
|
||||||
n.handleNewBlock(ctx, stats, vars, blockData.Rollup.Batches)
|
n.handleNewBlock(ctx, stats, vars, blockData.Rollup.Batches)
|
||||||
|
if n.debugAPI != nil {
|
||||||
|
n.debugAPI.SyncBlockHook()
|
||||||
|
}
|
||||||
return &blockData.Block, time.Duration(0), nil
|
return &blockData.Block, time.Duration(0), nil
|
||||||
} else {
|
} else {
|
||||||
// case: no block
|
// case: no block
|
||||||
@@ -583,6 +589,8 @@ func (n *Node) StartSynchronizer() {
|
|||||||
}
|
}
|
||||||
if errors.Is(err, eth.ErrBlockHashMismatchEvent) {
|
if errors.Is(err, eth.ErrBlockHashMismatchEvent) {
|
||||||
log.Warnw("Synchronizer.Sync", "err", err)
|
log.Warnw("Synchronizer.Sync", "err", err)
|
||||||
|
} else if errors.Is(err, synchronizer.ErrUnknownBlock) {
|
||||||
|
log.Warnw("Synchronizer.Sync", "err", err)
|
||||||
} else {
|
} else {
|
||||||
log.Errorw("Synchronizer.Sync", "err", err)
|
log.Errorw("Synchronizer.Sync", "err", err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -18,6 +18,19 @@ import (
|
|||||||
"github.com/hermeznetwork/tracerr"
|
"github.com/hermeznetwork/tracerr"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
const (
|
||||||
|
// errStrUnknownBlock is the string returned by geth when querying an
|
||||||
|
// unknown block
|
||||||
|
errStrUnknownBlock = "unknown block"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
// ErrUnknownBlock is the error returned by the Synchronizer when a
|
||||||
|
// block is queried by hash but the ethereum node doesn't find it due
|
||||||
|
// to it being discarded from a reorg.
|
||||||
|
ErrUnknownBlock = fmt.Errorf("unknown block")
|
||||||
|
)
|
||||||
|
|
||||||
// Stats of the syncrhonizer
|
// Stats of the syncrhonizer
|
||||||
type Stats struct {
|
type Stats struct {
|
||||||
Eth struct {
|
Eth struct {
|
||||||
@@ -648,11 +661,6 @@ func (s *Synchronizer) Sync2(ctx context.Context,
|
|||||||
return nil, nil, tracerr.Wrap(err)
|
return nil, nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
log.Debugw("Synced block",
|
|
||||||
"syncLastBlockNum", s.stats.Sync.LastBlock.Num,
|
|
||||||
"syncBlocksPerc", s.stats.blocksPerc(),
|
|
||||||
"ethLastBlockNum", s.stats.Eth.LastBlock.Num,
|
|
||||||
)
|
|
||||||
for _, batchData := range rollupData.Batches {
|
for _, batchData := range rollupData.Batches {
|
||||||
log.Debugw("Synced batch",
|
log.Debugw("Synced batch",
|
||||||
"syncLastBatch", batchData.Batch.BatchNum,
|
"syncLastBatch", batchData.Batch.BatchNum,
|
||||||
@@ -660,6 +668,11 @@ func (s *Synchronizer) Sync2(ctx context.Context,
|
|||||||
"ethLastBatch", s.stats.Eth.LastBatchNum,
|
"ethLastBatch", s.stats.Eth.LastBatchNum,
|
||||||
)
|
)
|
||||||
}
|
}
|
||||||
|
log.Debugw("Synced block",
|
||||||
|
"syncLastBlockNum", s.stats.Sync.LastBlock.Num,
|
||||||
|
"syncBlocksPerc", s.stats.blocksPerc(),
|
||||||
|
"ethLastBlockNum", s.stats.Eth.LastBlock.Num,
|
||||||
|
)
|
||||||
|
|
||||||
return blockData, nil, nil
|
return blockData, nil, nil
|
||||||
}
|
}
|
||||||
@@ -811,7 +824,9 @@ func (s *Synchronizer) rollupSync(ethBlock *common.Block) (*common.RollupData, e
|
|||||||
// Get rollup events in the block, and make sure the block hash matches
|
// Get rollup events in the block, and make sure the block hash matches
|
||||||
// the expected one.
|
// the expected one.
|
||||||
rollupEvents, err := s.ethClient.RollupEventsByBlock(blockNum, ðBlock.Hash)
|
rollupEvents, err := s.ethClient.RollupEventsByBlock(blockNum, ðBlock.Hash)
|
||||||
if err != nil {
|
if err != nil && err.Error() == errStrUnknownBlock {
|
||||||
|
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||||
|
} else if err != nil {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("RollupEventsByBlock: %w", err))
|
return nil, tracerr.Wrap(fmt.Errorf("RollupEventsByBlock: %w", err))
|
||||||
}
|
}
|
||||||
// No events in this block
|
// No events in this block
|
||||||
@@ -1121,7 +1136,9 @@ func (s *Synchronizer) auctionSync(ethBlock *common.Block) (*common.AuctionData,
|
|||||||
|
|
||||||
// Get auction events in the block
|
// Get auction events in the block
|
||||||
auctionEvents, err := s.ethClient.AuctionEventsByBlock(blockNum, ðBlock.Hash)
|
auctionEvents, err := s.ethClient.AuctionEventsByBlock(blockNum, ðBlock.Hash)
|
||||||
if err != nil {
|
if err != nil && err.Error() == errStrUnknownBlock {
|
||||||
|
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||||
|
} else if err != nil {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("AuctionEventsByBlock: %w", err))
|
return nil, tracerr.Wrap(fmt.Errorf("AuctionEventsByBlock: %w", err))
|
||||||
}
|
}
|
||||||
// No events in this block
|
// No events in this block
|
||||||
@@ -1218,7 +1235,9 @@ func (s *Synchronizer) wdelayerSync(ethBlock *common.Block) (*common.WDelayerDat
|
|||||||
|
|
||||||
// Get wDelayer events in the block
|
// Get wDelayer events in the block
|
||||||
wDelayerEvents, err := s.ethClient.WDelayerEventsByBlock(blockNum, ðBlock.Hash)
|
wDelayerEvents, err := s.ethClient.WDelayerEventsByBlock(blockNum, ðBlock.Hash)
|
||||||
if err != nil {
|
if err != nil && err.Error() == errStrUnknownBlock {
|
||||||
|
return nil, tracerr.Wrap(ErrUnknownBlock)
|
||||||
|
} else if err != nil {
|
||||||
return nil, tracerr.Wrap(fmt.Errorf("WDelayerEventsByBlock: %w", err))
|
return nil, tracerr.Wrap(fmt.Errorf("WDelayerEventsByBlock: %w", err))
|
||||||
}
|
}
|
||||||
// No events in this block
|
// No events in this block
|
||||||
|
|||||||
@@ -17,7 +17,6 @@ import (
|
|||||||
"github.com/hermeznetwork/hermez-node/db/historydb"
|
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||||
"github.com/hermeznetwork/hermez-node/eth"
|
"github.com/hermeznetwork/hermez-node/eth"
|
||||||
"github.com/hermeznetwork/hermez-node/log"
|
|
||||||
"github.com/hermeznetwork/hermez-node/test"
|
"github.com/hermeznetwork/hermez-node/test"
|
||||||
"github.com/hermeznetwork/hermez-node/test/til"
|
"github.com/hermeznetwork/hermez-node/test/til"
|
||||||
"github.com/jinzhu/copier"
|
"github.com/jinzhu/copier"
|
||||||
@@ -321,6 +320,14 @@ func newTestModules(t *testing.T) (*statedb.StateDB, *historydb.HistoryDB) {
|
|||||||
return stateDB, historyDB
|
return stateDB, historyDB
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func newBigInt(s string) *big.Int {
|
||||||
|
v, ok := new(big.Int).SetString(s, 10)
|
||||||
|
if !ok {
|
||||||
|
panic(fmt.Errorf("Can't set big.Int from %s", s))
|
||||||
|
}
|
||||||
|
return v
|
||||||
|
}
|
||||||
|
|
||||||
func TestSyncGeneral(t *testing.T) {
|
func TestSyncGeneral(t *testing.T) {
|
||||||
//
|
//
|
||||||
// Setup
|
// Setup
|
||||||
@@ -339,7 +346,6 @@ func TestSyncGeneral(t *testing.T) {
|
|||||||
s, err := NewSynchronizer(client, historyDB, stateDB, Config{
|
s, err := NewSynchronizer(client, historyDB, stateDB, Config{
|
||||||
StatsRefreshPeriod: 0 * time.Second,
|
StatsRefreshPeriod: 0 * time.Second,
|
||||||
})
|
})
|
||||||
log.Error(err)
|
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
ctx := context.Background()
|
ctx := context.Background()
|
||||||
@@ -434,12 +440,22 @@ func TestSyncGeneral(t *testing.T) {
|
|||||||
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches[0].L1CoordinatorTxs))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches[0].L1CoordinatorTxs))
|
||||||
|
// Set StateRoots for batches manually (til doesn't set it)
|
||||||
|
blocks[i].Rollup.Batches[0].Batch.StateRoot =
|
||||||
|
newBigInt("18906357591508007884273218035694076596537737437965299189312069102730480717391")
|
||||||
|
blocks[i].Rollup.Batches[1].Batch.StateRoot =
|
||||||
|
newBigInt("9513185123401321669660637227182204000277156839501731093239187625486561933297")
|
||||||
// blocks 1 (blockNum=3)
|
// blocks 1 (blockNum=3)
|
||||||
i = 1
|
i = 1
|
||||||
require.Equal(t, 3, int(blocks[i].Block.Num))
|
require.Equal(t, 3, int(blocks[i].Block.Num))
|
||||||
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
require.Equal(t, 5, len(blocks[i].Rollup.L1UserTxs))
|
||||||
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
require.Equal(t, 2, len(blocks[i].Rollup.Batches))
|
||||||
require.Equal(t, 3, len(blocks[i].Rollup.Batches[0].L2Txs))
|
require.Equal(t, 3, len(blocks[i].Rollup.Batches[0].L2Txs))
|
||||||
|
// Set StateRoots for batches manually (til doesn't set it)
|
||||||
|
blocks[i].Rollup.Batches[0].Batch.StateRoot =
|
||||||
|
newBigInt("13060270878200012606074130020925677466793317216609491464427188889005039616594")
|
||||||
|
blocks[i].Rollup.Batches[1].Batch.StateRoot =
|
||||||
|
newBigInt("21427104994652624302859637783375978708867165042357535792408500519060088086054")
|
||||||
|
|
||||||
// Generate extra required data
|
// Generate extra required data
|
||||||
ethAddTokens(blocks, client)
|
ethAddTokens(blocks, client)
|
||||||
@@ -614,6 +630,12 @@ func TestSyncGeneral(t *testing.T) {
|
|||||||
blocks, err = tc.GenerateBlocks(set2)
|
blocks, err = tc.GenerateBlocks(set2)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// Set StateRoots for batches manually (til doesn't set it)
|
||||||
|
blocks[0].Rollup.Batches[0].Batch.StateRoot =
|
||||||
|
newBigInt("11218510534825843475100588932060366395781087435899915642332104464234485046683")
|
||||||
|
blocks[0].Rollup.Batches[1].Batch.StateRoot =
|
||||||
|
newBigInt("20283020730369146334077598087403837297563965802277806438205710455191646998983")
|
||||||
|
|
||||||
for i := 0; i < 4; i++ {
|
for i := 0; i < 4; i++ {
|
||||||
client.CtlRollback()
|
client.CtlRollback()
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -2,12 +2,16 @@ package debugapi
|
|||||||
|
|
||||||
import (
|
import (
|
||||||
"context"
|
"context"
|
||||||
|
"fmt"
|
||||||
|
"math/big"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/gin-contrib/cors"
|
"github.com/gin-contrib/cors"
|
||||||
"github.com/gin-gonic/gin"
|
"github.com/gin-gonic/gin"
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
|
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||||
"github.com/hermeznetwork/hermez-node/log"
|
"github.com/hermeznetwork/hermez-node/log"
|
||||||
"github.com/hermeznetwork/hermez-node/synchronizer"
|
"github.com/hermeznetwork/hermez-node/synchronizer"
|
||||||
@@ -31,22 +35,109 @@ func badReq(err error, c *gin.Context) {
|
|||||||
})
|
})
|
||||||
}
|
}
|
||||||
|
|
||||||
|
const (
|
||||||
|
statusUpdating = "updating"
|
||||||
|
statusOK = "ok"
|
||||||
|
)
|
||||||
|
|
||||||
|
type tokenBalances struct {
|
||||||
|
sync.RWMutex
|
||||||
|
Value struct {
|
||||||
|
Status string
|
||||||
|
Block *common.Block
|
||||||
|
Batch *common.Batch
|
||||||
|
Balances map[common.TokenID]*big.Int
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (t *tokenBalances) Update(historyDB *historydb.HistoryDB, sdb *statedb.StateDB) (err error) {
|
||||||
|
var block *common.Block
|
||||||
|
var batch *common.Batch
|
||||||
|
var balances map[common.TokenID]*big.Int
|
||||||
|
defer func() {
|
||||||
|
t.Lock()
|
||||||
|
if err == nil {
|
||||||
|
t.Value.Status = statusOK
|
||||||
|
t.Value.Block = block
|
||||||
|
t.Value.Batch = batch
|
||||||
|
t.Value.Balances = balances
|
||||||
|
} else {
|
||||||
|
t.Value.Status = fmt.Sprintf("tokenBalances.Update: %v", err)
|
||||||
|
t.Value.Block = nil
|
||||||
|
t.Value.Batch = nil
|
||||||
|
t.Value.Balances = nil
|
||||||
|
}
|
||||||
|
t.Unlock()
|
||||||
|
}()
|
||||||
|
|
||||||
|
if block, err = historyDB.GetLastBlock(); err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
if batch, err = historyDB.GetLastBatch(); err != nil {
|
||||||
|
return tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
balances = make(map[common.TokenID]*big.Int)
|
||||||
|
sdb.LastRead(func(sdbLast *statedb.Last) error {
|
||||||
|
return tracerr.Wrap(
|
||||||
|
statedb.AccountsIter(sdbLast.DB(), func(a *common.Account) (bool, error) {
|
||||||
|
if balance, ok := balances[a.TokenID]; !ok {
|
||||||
|
balances[a.TokenID] = a.Balance
|
||||||
|
} else {
|
||||||
|
balance.Add(balance, a.Balance)
|
||||||
|
}
|
||||||
|
return true, nil
|
||||||
|
}),
|
||||||
|
)
|
||||||
|
})
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
// DebugAPI is an http API with debugging endpoints
|
// DebugAPI is an http API with debugging endpoints
|
||||||
type DebugAPI struct {
|
type DebugAPI struct {
|
||||||
addr string
|
addr string
|
||||||
|
historyDB *historydb.HistoryDB
|
||||||
stateDB *statedb.StateDB // synchronizer statedb
|
stateDB *statedb.StateDB // synchronizer statedb
|
||||||
sync *synchronizer.Synchronizer
|
sync *synchronizer.Synchronizer
|
||||||
|
tokenBalances tokenBalances
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewDebugAPI creates a new DebugAPI
|
// NewDebugAPI creates a new DebugAPI
|
||||||
func NewDebugAPI(addr string, stateDB *statedb.StateDB, sync *synchronizer.Synchronizer) *DebugAPI {
|
func NewDebugAPI(addr string, historyDB *historydb.HistoryDB, stateDB *statedb.StateDB,
|
||||||
|
sync *synchronizer.Synchronizer) *DebugAPI {
|
||||||
return &DebugAPI{
|
return &DebugAPI{
|
||||||
addr: addr,
|
addr: addr,
|
||||||
|
historyDB: historyDB,
|
||||||
stateDB: stateDB,
|
stateDB: stateDB,
|
||||||
sync: sync,
|
sync: sync,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// SyncBlockHook is a hook function that the node will call after every new synchronized block
|
||||||
|
func (a *DebugAPI) SyncBlockHook() {
|
||||||
|
a.tokenBalances.RLock()
|
||||||
|
updateTokenBalances := a.tokenBalances.Value.Status == statusUpdating
|
||||||
|
a.tokenBalances.RUnlock()
|
||||||
|
if updateTokenBalances {
|
||||||
|
if err := a.tokenBalances.Update(a.historyDB, a.stateDB); err != nil {
|
||||||
|
log.Errorw("DebugAPI.tokenBalances.Upate", "err", err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *DebugAPI) handleTokenBalances(c *gin.Context) {
|
||||||
|
a.tokenBalances.RLock()
|
||||||
|
tokenBalances := a.tokenBalances.Value
|
||||||
|
a.tokenBalances.RUnlock()
|
||||||
|
c.JSON(http.StatusOK, tokenBalances)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (a *DebugAPI) handlePostTokenBalances(c *gin.Context) {
|
||||||
|
a.tokenBalances.Lock()
|
||||||
|
a.tokenBalances.Value.Status = statusUpdating
|
||||||
|
a.tokenBalances.Unlock()
|
||||||
|
c.JSON(http.StatusOK, nil)
|
||||||
|
}
|
||||||
|
|
||||||
func (a *DebugAPI) handleAccount(c *gin.Context) {
|
func (a *DebugAPI) handleAccount(c *gin.Context) {
|
||||||
uri := struct {
|
uri := struct {
|
||||||
Idx uint32
|
Idx uint32
|
||||||
@@ -114,6 +205,8 @@ func (a *DebugAPI) Run(ctx context.Context) error {
|
|||||||
// is created.
|
// is created.
|
||||||
debugAPI.GET("sdb/accounts", a.handleAccounts)
|
debugAPI.GET("sdb/accounts", a.handleAccounts)
|
||||||
debugAPI.GET("sdb/accounts/:Idx", a.handleAccount)
|
debugAPI.GET("sdb/accounts/:Idx", a.handleAccount)
|
||||||
|
debugAPI.POST("sdb/tokenbalances", a.handlePostTokenBalances)
|
||||||
|
debugAPI.GET("sdb/tokenbalances", a.handleTokenBalances)
|
||||||
|
|
||||||
debugAPI.GET("sync/stats", a.handleSyncStats)
|
debugAPI.GET("sync/stats", a.handleSyncStats)
|
||||||
|
|
||||||
|
|||||||
@@ -6,13 +6,21 @@ import (
|
|||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"math/big"
|
"math/big"
|
||||||
"net/http"
|
"net/http"
|
||||||
|
"os"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
"sync"
|
||||||
"testing"
|
"testing"
|
||||||
|
|
||||||
"github.com/dghubble/sling"
|
"github.com/dghubble/sling"
|
||||||
|
ethCommon "github.com/ethereum/go-ethereum/common"
|
||||||
ethCrypto "github.com/ethereum/go-ethereum/crypto"
|
ethCrypto "github.com/ethereum/go-ethereum/crypto"
|
||||||
"github.com/hermeznetwork/hermez-node/common"
|
"github.com/hermeznetwork/hermez-node/common"
|
||||||
|
dbUtils "github.com/hermeznetwork/hermez-node/db"
|
||||||
|
"github.com/hermeznetwork/hermez-node/db/historydb"
|
||||||
"github.com/hermeznetwork/hermez-node/db/statedb"
|
"github.com/hermeznetwork/hermez-node/db/statedb"
|
||||||
|
"github.com/hermeznetwork/hermez-node/test"
|
||||||
|
"github.com/hermeznetwork/hermez-node/test/til"
|
||||||
|
"github.com/hermeznetwork/hermez-node/txprocessor"
|
||||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||||
"github.com/stretchr/testify/assert"
|
"github.com/stretchr/testify/assert"
|
||||||
"github.com/stretchr/testify/require"
|
"github.com/stretchr/testify/require"
|
||||||
@@ -51,12 +59,15 @@ func TestDebugAPI(t *testing.T) {
|
|||||||
|
|
||||||
addr := "localhost:12345"
|
addr := "localhost:12345"
|
||||||
// We won't test the sync/stats endpoint, so we can se the Syncrhonizer to nil
|
// We won't test the sync/stats endpoint, so we can se the Syncrhonizer to nil
|
||||||
debugAPI := NewDebugAPI(addr, sdb, nil)
|
debugAPI := NewDebugAPI(addr, nil, sdb, nil)
|
||||||
|
|
||||||
ctx, cancel := context.WithCancel(context.Background())
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
wg := sync.WaitGroup{}
|
||||||
|
wg.Add(1)
|
||||||
go func() {
|
go func() {
|
||||||
err := debugAPI.Run(ctx)
|
err := debugAPI.Run(ctx)
|
||||||
require.Nil(t, err)
|
require.Nil(t, err)
|
||||||
|
wg.Done()
|
||||||
}()
|
}()
|
||||||
|
|
||||||
var accounts []common.Account
|
var accounts []common.Account
|
||||||
@@ -102,4 +113,130 @@ func TestDebugAPI(t *testing.T) {
|
|||||||
assert.Equal(t, accounts, accountsAPI)
|
assert.Equal(t, accounts, accountsAPI)
|
||||||
|
|
||||||
cancel()
|
cancel()
|
||||||
|
wg.Wait()
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestDebugAPITokenBalances(t *testing.T) {
|
||||||
|
dir, err := ioutil.TempDir("", "tmpdb")
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
sdb, err := statedb.NewStateDB(statedb.Config{Path: dir, Keep: 128, Type: statedb.TypeSynchronizer, NLevels: 32})
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
// Init History DB
|
||||||
|
pass := os.Getenv("POSTGRES_PASS")
|
||||||
|
db, err := dbUtils.InitSQLDB(5432, "localhost", "hermez", pass, "hermez")
|
||||||
|
require.NoError(t, err)
|
||||||
|
historyDB := historydb.NewHistoryDB(db, nil)
|
||||||
|
// Clear DB
|
||||||
|
test.WipeDB(historyDB.DB())
|
||||||
|
|
||||||
|
set := `
|
||||||
|
Type: Blockchain
|
||||||
|
|
||||||
|
AddToken(1)
|
||||||
|
AddToken(2)
|
||||||
|
AddToken(3)
|
||||||
|
|
||||||
|
CreateAccountDeposit(1) A: 1000
|
||||||
|
CreateAccountDeposit(2) A: 2000
|
||||||
|
CreateAccountDeposit(1) B: 100
|
||||||
|
CreateAccountDeposit(2) B: 200
|
||||||
|
CreateAccountDeposit(2) C: 400
|
||||||
|
|
||||||
|
> batchL1 // forge L1UserTxs{nil}, freeze defined L1UserTxs{5}
|
||||||
|
> batchL1 // forge defined L1UserTxs{5}, freeze L1UserTxs{nil}
|
||||||
|
> block // blockNum=2
|
||||||
|
`
|
||||||
|
|
||||||
|
chainID := uint16(0)
|
||||||
|
tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||||
|
tilCfgExtra := til.ConfigExtra{
|
||||||
|
BootCoordAddr: ethCommon.HexToAddress("0xE39fEc6224708f0772D2A74fd3f9055A90E0A9f2"),
|
||||||
|
CoordUser: "A",
|
||||||
|
}
|
||||||
|
blocks, err := tc.GenerateBlocks(set)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
err = tc.FillBlocksExtra(blocks, &tilCfgExtra)
|
||||||
|
require.NoError(t, err)
|
||||||
|
tc.FillBlocksL1UserTxsBatchNum(blocks)
|
||||||
|
err = tc.FillBlocksForgedL1UserTxs(blocks)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
tpc := txprocessor.Config{
|
||||||
|
NLevels: 32,
|
||||||
|
MaxTx: 100,
|
||||||
|
ChainID: chainID,
|
||||||
|
MaxFeeTx: common.RollupConstMaxFeeIdxCoordinator,
|
||||||
|
MaxL1Tx: common.RollupConstMaxL1Tx,
|
||||||
|
}
|
||||||
|
tp := txprocessor.NewTxProcessor(sdb, tpc)
|
||||||
|
|
||||||
|
for _, block := range blocks {
|
||||||
|
require.NoError(t, historyDB.AddBlockSCData(&block))
|
||||||
|
for _, batch := range block.Rollup.Batches {
|
||||||
|
_, err := tp.ProcessTxs(batch.Batch.FeeIdxsCoordinator,
|
||||||
|
batch.L1UserTxs, batch.L1CoordinatorTxs, []common.PoolL2Tx{})
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
addr := "localhost:12345"
|
||||||
|
// We won't test the sync/stats endpoint, so we can se the Syncrhonizer to nil
|
||||||
|
debugAPI := NewDebugAPI(addr, historyDB, sdb, nil)
|
||||||
|
|
||||||
|
ctx, cancel := context.WithCancel(context.Background())
|
||||||
|
wg := sync.WaitGroup{}
|
||||||
|
wg.Add(1)
|
||||||
|
go func() {
|
||||||
|
err := debugAPI.Run(ctx)
|
||||||
|
require.Nil(t, err)
|
||||||
|
wg.Done()
|
||||||
|
}()
|
||||||
|
|
||||||
|
var accounts []common.Account
|
||||||
|
for i := 0; i < 16; i++ {
|
||||||
|
account := newAccount(t, i)
|
||||||
|
accounts = append(accounts, *account)
|
||||||
|
_, err = sdb.CreateAccount(account.Idx, account)
|
||||||
|
require.Nil(t, err)
|
||||||
|
}
|
||||||
|
// Make a checkpoint (batchNum 2) to make the accounts available in Last
|
||||||
|
err = sdb.MakeCheckpoint()
|
||||||
|
require.Nil(t, err)
|
||||||
|
|
||||||
|
url := fmt.Sprintf("http://%v/debug/", addr)
|
||||||
|
|
||||||
|
var batchNum common.BatchNum
|
||||||
|
req, err := sling.New().Get(url).Path("sdb/batchnum").ReceiveSuccess(&batchNum)
|
||||||
|
require.Equal(t, http.StatusOK, req.StatusCode)
|
||||||
|
require.Nil(t, err)
|
||||||
|
assert.Equal(t, common.BatchNum(2), batchNum)
|
||||||
|
|
||||||
|
var mtroot *big.Int
|
||||||
|
req, err = sling.New().Get(url).Path("sdb/mtroot").ReceiveSuccess(&mtroot)
|
||||||
|
require.Equal(t, http.StatusOK, req.StatusCode)
|
||||||
|
require.Nil(t, err)
|
||||||
|
// Testing against a hardcoded value obtained by running the test and
|
||||||
|
// printing the value previously.
|
||||||
|
assert.Equal(t, "21765339739823365993496282904432398015268846626944509989242908567129545640185",
|
||||||
|
mtroot.String())
|
||||||
|
|
||||||
|
var accountAPI common.Account
|
||||||
|
req, err = sling.New().Get(url).
|
||||||
|
Path(fmt.Sprintf("sdb/accounts/%v", accounts[0].Idx)).
|
||||||
|
ReceiveSuccess(&accountAPI)
|
||||||
|
require.Equal(t, http.StatusOK, req.StatusCode)
|
||||||
|
require.Nil(t, err)
|
||||||
|
assert.Equal(t, accounts[0], accountAPI)
|
||||||
|
|
||||||
|
var accountsAPI []common.Account
|
||||||
|
req, err = sling.New().Get(url).Path("sdb/accounts").ReceiveSuccess(&accountsAPI)
|
||||||
|
require.Equal(t, http.StatusOK, req.StatusCode)
|
||||||
|
require.Nil(t, err)
|
||||||
|
assert.Equal(t, accounts, accountsAPI)
|
||||||
|
|
||||||
|
cancel()
|
||||||
|
wg.Wait()
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -797,11 +797,11 @@ func (c *Client) RollupL1UserTxERC20ETH(
|
|||||||
cpy := c.nextBlock().copy()
|
cpy := c.nextBlock().copy()
|
||||||
defer func() { c.revertIfErr(err, cpy) }()
|
defer func() { c.revertIfErr(err, cpy) }()
|
||||||
|
|
||||||
_, err = common.NewFloat16(amount)
|
_, err = common.NewFloat40(amount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
_, err = common.NewFloat16(depositAmount)
|
_, err = common.NewFloat40(depositAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, tracerr.Wrap(err)
|
return nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -145,7 +145,7 @@ const longWaitDuration = 999 * time.Hour
|
|||||||
// const provingDuration = 2 * time.Second
|
// const provingDuration = 2 * time.Second
|
||||||
|
|
||||||
func (s *Mock) runProver(ctx context.Context) {
|
func (s *Mock) runProver(ctx context.Context) {
|
||||||
waitDuration := longWaitDuration
|
waitCh := time.After(longWaitDuration)
|
||||||
for {
|
for {
|
||||||
select {
|
select {
|
||||||
case <-ctx.Done():
|
case <-ctx.Done():
|
||||||
@@ -153,21 +153,21 @@ func (s *Mock) runProver(ctx context.Context) {
|
|||||||
case msg := <-s.msgCh:
|
case msg := <-s.msgCh:
|
||||||
switch msg.value {
|
switch msg.value {
|
||||||
case "cancel":
|
case "cancel":
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
s.Lock()
|
s.Lock()
|
||||||
if !s.status.IsReady() {
|
if !s.status.IsReady() {
|
||||||
s.status = prover.StatusCodeAborted
|
s.status = prover.StatusCodeAborted
|
||||||
}
|
}
|
||||||
s.Unlock()
|
s.Unlock()
|
||||||
case "prove":
|
case "prove":
|
||||||
waitDuration = s.provingDuration
|
waitCh = time.After(s.provingDuration)
|
||||||
s.Lock()
|
s.Lock()
|
||||||
s.status = prover.StatusCodeBusy
|
s.status = prover.StatusCodeBusy
|
||||||
s.Unlock()
|
s.Unlock()
|
||||||
}
|
}
|
||||||
msg.ackCh <- true
|
msg.ackCh <- true
|
||||||
case <-time.After(waitDuration):
|
case <-waitCh:
|
||||||
waitDuration = longWaitDuration
|
waitCh = time.After(longWaitDuration)
|
||||||
s.Lock()
|
s.Lock()
|
||||||
if s.status != prover.StatusCodeBusy {
|
if s.status != prover.StatusCodeBusy {
|
||||||
s.Unlock()
|
s.Unlock()
|
||||||
|
|||||||
@@ -142,10 +142,12 @@ func GenerateTxsZKInputs0(t *testing.T, chainID uint16) (users []til.User, coord
|
|||||||
// same values than in the js test
|
// same values than in the js test
|
||||||
users = GenerateJsUsers(t)
|
users = GenerateJsUsers(t)
|
||||||
|
|
||||||
|
depositAmount, err := common.Float40(10400).BigInt()
|
||||||
|
require.Nil(t, err)
|
||||||
l1UserTxs = []common.L1Tx{
|
l1UserTxs = []common.L1Tx{
|
||||||
{
|
{
|
||||||
FromIdx: 0,
|
FromIdx: 0,
|
||||||
DepositAmount: big.NewInt(16000000),
|
DepositAmount: depositAmount,
|
||||||
Amount: big.NewInt(0),
|
Amount: big.NewInt(0),
|
||||||
TokenID: 1,
|
TokenID: 1,
|
||||||
FromBJJ: users[0].BJJ.Public().Compress(),
|
FromBJJ: users[0].BJJ.Public().Compress(),
|
||||||
|
|||||||
@@ -311,7 +311,7 @@ func TestZKInputsExitWithFee0(t *testing.T) {
|
|||||||
assert.Equal(t, "8737171572459172806192626402462788826264011087579491137542380589998149683116", bb.LocalStateDB().MT.Root().BigInt().String())
|
assert.Equal(t, "8737171572459172806192626402462788826264011087579491137542380589998149683116", bb.LocalStateDB().MT.Root().BigInt().String())
|
||||||
h, err := zki.HashGlobalData()
|
h, err := zki.HashGlobalData()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, "9971598169768987067017223790214537222850903267980994716992761290793474746117", h.String())
|
assert.Equal(t, "18608843755023673022528019960628191162333429206359207449879743919826610006009", h.String())
|
||||||
sendProofAndCheckResp(t, zki)
|
sendProofAndCheckResp(t, zki)
|
||||||
|
|
||||||
// batch3
|
// batch3
|
||||||
@@ -334,7 +334,7 @@ func TestZKInputsExitWithFee0(t *testing.T) {
|
|||||||
assert.Equal(t, "18306761925365215381387147754881756804475668085493847010988306480531520370130", bb.LocalStateDB().MT.Root().BigInt().String())
|
assert.Equal(t, "18306761925365215381387147754881756804475668085493847010988306480531520370130", bb.LocalStateDB().MT.Root().BigInt().String())
|
||||||
h, err = zki.HashGlobalData()
|
h, err = zki.HashGlobalData()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
assert.Equal(t, "7992262236065691439683036344554725221924027193771770363772735722054938818364", h.String())
|
assert.Equal(t, "6651837443119278772088559395433504719862425648816904171510845286897104469889", h.String())
|
||||||
assert.Equal(t, common.EthAddrToBigInt(tc.Users["Coord"].Addr), zki.EthAddr3[0])
|
assert.Equal(t, common.EthAddrToBigInt(tc.Users["Coord"].Addr), zki.EthAddr3[0])
|
||||||
assert.Equal(t, "0", zki.EthAddr3[1].String())
|
assert.Equal(t, "0", zki.EthAddr3[1].String())
|
||||||
sendProofAndCheckResp(t, zki)
|
sendProofAndCheckResp(t, zki)
|
||||||
|
|||||||
@@ -31,7 +31,7 @@ func TestMain(m *testing.M) {
|
|||||||
os.Exit(exitVal)
|
os.Exit(exitVal)
|
||||||
}
|
}
|
||||||
|
|
||||||
const MaxTx = 376
|
const MaxTx = 352
|
||||||
const NLevels = 32
|
const NLevels = 32
|
||||||
const MaxL1Tx = 256
|
const MaxL1Tx = 256
|
||||||
const MaxFeeTx = 64
|
const MaxFeeTx = 64
|
||||||
@@ -61,6 +61,7 @@ func sendProofAndCheckResp(t *testing.T, zki *common.ZKInputs) {
|
|||||||
return
|
return
|
||||||
}
|
}
|
||||||
|
|
||||||
|
log.Infof("sending proof to %s", proofServerURL)
|
||||||
// Store zkinputs json for debugging purposes
|
// Store zkinputs json for debugging purposes
|
||||||
zkInputsJSON, err := json.Marshal(zki)
|
zkInputsJSON, err := json.Marshal(zki)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|||||||
@@ -33,9 +33,11 @@ type TxProcessor struct {
|
|||||||
// Config contains the TxProcessor configuration parameters
|
// Config contains the TxProcessor configuration parameters
|
||||||
type Config struct {
|
type Config struct {
|
||||||
NLevels uint32
|
NLevels uint32
|
||||||
|
// MaxFeeTx is the maximum number of coordinator accounts that can receive fees
|
||||||
MaxFeeTx uint32
|
MaxFeeTx uint32
|
||||||
MaxTx uint32
|
MaxTx uint32
|
||||||
MaxL1Tx uint32
|
MaxL1Tx uint32
|
||||||
|
// ChainID of the blockchain
|
||||||
ChainID uint16
|
ChainID uint16
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -501,11 +503,11 @@ func (tp *TxProcessor) ProcessL1Tx(exitTree *merkletree.MerkleTree, tx *common.L
|
|||||||
tp.zki.OnChain[tp.i] = big.NewInt(1)
|
tp.zki.OnChain[tp.i] = big.NewInt(1)
|
||||||
|
|
||||||
// L1Txs
|
// L1Txs
|
||||||
depositAmountF16, err := common.NewFloat16(tx.DepositAmount)
|
depositAmountF40, err := common.NewFloat40(tx.DepositAmount)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return nil, nil, false, nil, tracerr.Wrap(err)
|
return nil, nil, false, nil, tracerr.Wrap(err)
|
||||||
}
|
}
|
||||||
tp.zki.DepositAmountF[tp.i] = big.NewInt(int64(depositAmountF16))
|
tp.zki.DepositAmountF[tp.i] = big.NewInt(int64(depositAmountF40))
|
||||||
tp.zki.FromEthAddr[tp.i] = common.EthAddrToBigInt(tx.FromEthAddr)
|
tp.zki.FromEthAddr[tp.i] = common.EthAddrToBigInt(tx.FromEthAddr)
|
||||||
if tx.FromBJJ != common.EmptyBJJComp {
|
if tx.FromBJJ != common.EmptyBJJComp {
|
||||||
tp.zki.FromBJJCompressed[tp.i] = BJJCompressedTo256BigInts(tx.FromBJJ)
|
tp.zki.FromBJJCompressed[tp.i] = BJJCompressedTo256BigInts(tx.FromBJJ)
|
||||||
@@ -515,6 +517,20 @@ func (tp *TxProcessor) ProcessL1Tx(exitTree *merkletree.MerkleTree, tx *common.L
|
|||||||
if tp.i < len(tp.zki.ISOnChain) { // len(tp.zki.ISOnChain) == nTx
|
if tp.i < len(tp.zki.ISOnChain) { // len(tp.zki.ISOnChain) == nTx
|
||||||
tp.zki.ISOnChain[tp.i] = big.NewInt(1)
|
tp.zki.ISOnChain[tp.i] = big.NewInt(1)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if tx.Type == common.TxTypeForceTransfer ||
|
||||||
|
tx.Type == common.TxTypeDepositTransfer ||
|
||||||
|
tx.Type == common.TxTypeCreateAccountDepositTransfer ||
|
||||||
|
tx.Type == common.TxTypeForceExit {
|
||||||
|
// in the cases where at L1Tx there is usage of the
|
||||||
|
// Amount parameter, add it at the ZKInputs.AmountF
|
||||||
|
// slot
|
||||||
|
amountF40, err := common.NewFloat40(tx.Amount)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, false, nil, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
tp.zki.AmountF[tp.i] = big.NewInt(int64(amountF40))
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
switch tx.Type {
|
switch tx.Type {
|
||||||
@@ -657,6 +673,11 @@ func (tp *TxProcessor) ProcessL2Tx(coordIdxsMap map[common.TokenID]common.Idx,
|
|||||||
tp.zki.ToEthAddr[tp.i] = common.EthAddrToBigInt(tx.ToEthAddr)
|
tp.zki.ToEthAddr[tp.i] = common.EthAddrToBigInt(tx.ToEthAddr)
|
||||||
|
|
||||||
tp.zki.OnChain[tp.i] = big.NewInt(0)
|
tp.zki.OnChain[tp.i] = big.NewInt(0)
|
||||||
|
amountF40, err := common.NewFloat40(tx.Amount)
|
||||||
|
if err != nil {
|
||||||
|
return nil, nil, false, tracerr.Wrap(err)
|
||||||
|
}
|
||||||
|
tp.zki.AmountF[tp.i] = big.NewInt(int64(amountF40))
|
||||||
tp.zki.NewAccount[tp.i] = big.NewInt(0)
|
tp.zki.NewAccount[tp.i] = big.NewInt(0)
|
||||||
|
|
||||||
// L2Txs
|
// L2Txs
|
||||||
|
|||||||
@@ -1,8 +1,6 @@
|
|||||||
package txprocessor
|
package txprocessor
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"encoding/binary"
|
|
||||||
"encoding/hex"
|
|
||||||
"io/ioutil"
|
"io/ioutil"
|
||||||
"math/big"
|
"math/big"
|
||||||
"os"
|
"os"
|
||||||
@@ -642,17 +640,16 @@ func TestCreateAccountDepositMaxValue(t *testing.T) {
|
|||||||
|
|
||||||
users := txsets.GenerateJsUsers(t)
|
users := txsets.GenerateJsUsers(t)
|
||||||
|
|
||||||
daMaxHex, err := hex.DecodeString("FFFF")
|
daMaxF40 := common.Float40(0xFFFFFFFFFF)
|
||||||
|
daMaxBI, err := daMaxF40.BigInt()
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
daMaxF16 := common.Float16(binary.BigEndian.Uint16(daMaxHex))
|
assert.Equal(t, "343597383670000000000000000000000000000000", daMaxBI.String())
|
||||||
daMaxBI := daMaxF16.BigInt()
|
|
||||||
assert.Equal(t, "10235000000000000000000000000000000", daMaxBI.String())
|
|
||||||
|
|
||||||
daMax1Hex, err := hex.DecodeString("FFFE")
|
daMax1F40 := common.Float40(0xFFFFFFFFFE)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
daMax1F16 := common.Float16(binary.BigEndian.Uint16(daMax1Hex))
|
daMax1BI, err := daMax1F40.BigInt()
|
||||||
daMax1BI := daMax1F16.BigInt()
|
require.NoError(t, err)
|
||||||
assert.Equal(t, "10225000000000000000000000000000000", daMax1BI.String())
|
assert.Equal(t, "343597383660000000000000000000000000000000", daMax1BI.String())
|
||||||
|
|
||||||
l1Txs := []common.L1Tx{
|
l1Txs := []common.L1Tx{
|
||||||
{
|
{
|
||||||
|
|||||||
File diff suppressed because one or more lines are too long
@@ -3,7 +3,6 @@ package txselector
|
|||||||
// current: very simple version of TxSelector
|
// current: very simple version of TxSelector
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bytes"
|
|
||||||
"fmt"
|
"fmt"
|
||||||
"math/big"
|
"math/big"
|
||||||
"sort"
|
"sort"
|
||||||
@@ -19,7 +18,8 @@ import (
|
|||||||
"github.com/iden3/go-iden3-crypto/babyjub"
|
"github.com/iden3/go-iden3-crypto/babyjub"
|
||||||
)
|
)
|
||||||
|
|
||||||
// txs implements the interface Sort for an array of Tx
|
// txs implements the interface Sort for an array of Tx, and sorts the txs by
|
||||||
|
// absolute fee
|
||||||
type txs []common.PoolL2Tx
|
type txs []common.PoolL2Tx
|
||||||
|
|
||||||
func (t txs) Len() int {
|
func (t txs) Len() int {
|
||||||
@@ -232,7 +232,8 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
if len(l1CoordinatorTxs) >= int(selectionConfig.MaxL1UserTxs)-len(l1UserTxs) {
|
if len(l1CoordinatorTxs) >= int(selectionConfig.MaxL1UserTxs)-len(l1UserTxs) {
|
||||||
// discard L2Tx, and update Info parameter of
|
// discard L2Tx, and update Info parameter of
|
||||||
// the tx, and add it to the discardedTxs array
|
// the tx, and add it to the discardedTxs array
|
||||||
l2Txs0[i].Info = "Tx not selected due the L2Tx depends on a L1CoordinatorTx and there is not enough space for L1Coordinator"
|
l2Txs0[i].Info = "Tx not selected because the L2Tx depends on a " +
|
||||||
|
"L1CoordinatorTx and there is not enough space for L1Coordinator"
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs0[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs0[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -257,7 +258,9 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
// not valid Amount with current Balance. Discard L2Tx,
|
// not valid Amount with current Balance. Discard L2Tx,
|
||||||
// and update Info parameter of the tx, and add it to
|
// and update Info parameter of the tx, and add it to
|
||||||
// the discardedTxs array
|
// the discardedTxs array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected due not enough Balance at the sender. Current sender account Balance: %s, Amount+Fee: %s", balance.String(), feeAndAmount.String())
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected due to not enough Balance at the sender. "+
|
||||||
|
"Current sender account Balance: %s, Amount+Fee: %s",
|
||||||
|
balance.String(), feeAndAmount.String())
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -269,7 +272,8 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
// not valid Nonce at tx. Discard L2Tx, and update Info
|
// not valid Nonce at tx. Discard L2Tx, and update Info
|
||||||
// parameter of the tx, and add it to the discardedTxs
|
// parameter of the tx, and add it to the discardedTxs
|
||||||
// array
|
// array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected due not current Nonce. Tx.Nonce: %d, Account.Nonce: %d", l2Txs[i].Nonce, nonce)
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected due to not current Nonce. "+
|
||||||
|
"Tx.Nonce: %d, Account.Nonce: %d", l2Txs[i].Nonce, nonce)
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -287,18 +291,31 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
txsel.processTxToEthAddrBJJ(validTxs, selectionConfig,
|
txsel.processTxToEthAddrBJJ(validTxs, selectionConfig,
|
||||||
len(l1UserTxs), l1CoordinatorTxs, positionL1, l2Txs[i])
|
len(l1UserTxs), l1CoordinatorTxs, positionL1, l2Txs[i])
|
||||||
if err != nil {
|
if err != nil {
|
||||||
log.Debug(err)
|
log.Debugw("txsel.processTxToEthAddrBJJ", "err", err)
|
||||||
// Discard L2Tx, and update Info parameter of
|
// Discard L2Tx, and update Info parameter of
|
||||||
// the tx, and add it to the discardedTxs array
|
// the tx, and add it to the discardedTxs array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected (in processTxToEthAddrBJJ) due %s", err.Error())
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected (in processTxToEthAddrBJJ) due to %s",
|
||||||
|
err.Error())
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
if accAuth != nil && l1CoordinatorTx != nil {
|
if l1CoordinatorTx != nil {
|
||||||
|
// If ToEthAddr == 0xff.. this means that we
|
||||||
|
// are handling a TransferToBJJ, which doesn't
|
||||||
|
// require an authorization because it doesn't
|
||||||
|
// contain a valid ethereum address.
|
||||||
|
// Otherwise only create the account if we have
|
||||||
|
// the corresponding authorization
|
||||||
|
if validL2Tx.ToEthAddr == common.FFAddr {
|
||||||
|
accAuths = append(accAuths, common.EmptyEthSignature)
|
||||||
|
l1CoordinatorTxs = append(l1CoordinatorTxs, *l1CoordinatorTx)
|
||||||
|
positionL1++
|
||||||
|
} else if accAuth != nil {
|
||||||
accAuths = append(accAuths, accAuth.Signature)
|
accAuths = append(accAuths, accAuth.Signature)
|
||||||
l1CoordinatorTxs = append(l1CoordinatorTxs, *l1CoordinatorTx)
|
l1CoordinatorTxs = append(l1CoordinatorTxs, *l1CoordinatorTx)
|
||||||
positionL1++
|
positionL1++
|
||||||
}
|
}
|
||||||
|
}
|
||||||
if validL2Tx != nil {
|
if validL2Tx != nil {
|
||||||
validTxs = append(validTxs, *validL2Tx)
|
validTxs = append(validTxs, *validL2Tx)
|
||||||
}
|
}
|
||||||
@@ -310,8 +327,8 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
"ToIdx", l2Txs[i].ToIdx)
|
"ToIdx", l2Txs[i].ToIdx)
|
||||||
// Discard L2Tx, and update Info parameter of
|
// Discard L2Tx, and update Info parameter of
|
||||||
// the tx, and add it to the discardedTxs array
|
// the tx, and add it to the discardedTxs array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected due tx.ToIdx not found in StateDB. ToIdx: %d",
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected due to tx.ToIdx not found in StateDB. "+
|
||||||
l2Txs[i].ToIdx)
|
"ToIdx: %d", l2Txs[i].ToIdx)
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -323,7 +340,9 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
// Discard L2Tx, and update Info
|
// Discard L2Tx, and update Info
|
||||||
// parameter of the tx, and add it to
|
// parameter of the tx, and add it to
|
||||||
// the discardedTxs array
|
// the discardedTxs array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected due ToEthAddr does not correspond to the Account.EthAddr. tx.ToIdx: %d, tx.ToEthAddr: %s, account.EthAddr: %s",
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected because ToEthAddr "+
|
||||||
|
"does not correspond to the Account.EthAddr. "+
|
||||||
|
"tx.ToIdx: %d, tx.ToEthAddr: %s, account.EthAddr: %s",
|
||||||
l2Txs[i].ToIdx, l2Txs[i].ToEthAddr, receiverAcc.EthAddr)
|
l2Txs[i].ToIdx, l2Txs[i].ToEthAddr, receiverAcc.EthAddr)
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
@@ -337,7 +356,9 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
// Discard L2Tx, and update Info
|
// Discard L2Tx, and update Info
|
||||||
// parameter of the tx, and add it to
|
// parameter of the tx, and add it to
|
||||||
// the discardedTxs array
|
// the discardedTxs array
|
||||||
l2Txs[i].Info = fmt.Sprintf("Tx not selected due tx.ToBJJ does not correspond to the Account.BJJ. tx.ToIdx: %d, tx.ToEthAddr: %s, tx.ToBJJ: %s, account.BJJ: %s",
|
l2Txs[i].Info = fmt.Sprintf("Tx not selected because tx.ToBJJ "+
|
||||||
|
"does not correspond to the Account.BJJ. "+
|
||||||
|
"tx.ToIdx: %d, tx.ToEthAddr: %s, tx.ToBJJ: %s, account.BJJ: %s",
|
||||||
l2Txs[i].ToIdx, l2Txs[i].ToEthAddr, l2Txs[i].ToBJJ, receiverAcc.BJJ)
|
l2Txs[i].ToIdx, l2Txs[i].ToEthAddr, l2Txs[i].ToBJJ, receiverAcc.BJJ)
|
||||||
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, l2Txs[i])
|
||||||
continue
|
continue
|
||||||
@@ -411,7 +432,7 @@ func (txsel *TxSelector) GetL1L2TxSelection(selectionConfig *SelectionConfig,
|
|||||||
log.Error(err)
|
log.Error(err)
|
||||||
// Discard L2Tx, and update Info parameter of the tx,
|
// Discard L2Tx, and update Info parameter of the tx,
|
||||||
// and add it to the discardedTxs array
|
// and add it to the discardedTxs array
|
||||||
selectedL2Txs[i].Info = fmt.Sprintf("Tx not selected (in ProcessL2Tx) due %s", err.Error())
|
selectedL2Txs[i].Info = fmt.Sprintf("Tx not selected (in ProcessL2Tx) due to %s", err.Error())
|
||||||
discardedL2Txs = append(discardedL2Txs, selectedL2Txs[i])
|
discardedL2Txs = append(discardedL2Txs, selectedL2Txs[i])
|
||||||
continue
|
continue
|
||||||
}
|
}
|
||||||
@@ -467,8 +488,7 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs []common.PoolL2Tx,
|
|||||||
|
|
||||||
var l1CoordinatorTx *common.L1Tx
|
var l1CoordinatorTx *common.L1Tx
|
||||||
var accAuth *common.AccountCreationAuth
|
var accAuth *common.AccountCreationAuth
|
||||||
if !bytes.Equal(l2Tx.ToEthAddr.Bytes(), common.EmptyAddr.Bytes()) &&
|
if l2Tx.ToEthAddr != common.EmptyAddr && l2Tx.ToEthAddr != common.FFAddr {
|
||||||
!bytes.Equal(l2Tx.ToEthAddr.Bytes(), common.FFAddr.Bytes()) {
|
|
||||||
// case: ToEthAddr != 0x00 neither 0xff
|
// case: ToEthAddr != 0x00 neither 0xff
|
||||||
if l2Tx.ToBJJ != common.EmptyBJJComp {
|
if l2Tx.ToBJJ != common.EmptyBJJComp {
|
||||||
// case: ToBJJ!=0:
|
// case: ToBJJ!=0:
|
||||||
@@ -524,8 +544,7 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs []common.PoolL2Tx,
|
|||||||
DepositAmount: big.NewInt(0),
|
DepositAmount: big.NewInt(0),
|
||||||
Type: common.TxTypeCreateAccountDeposit,
|
Type: common.TxTypeCreateAccountDeposit,
|
||||||
}
|
}
|
||||||
} else if bytes.Equal(l2Tx.ToEthAddr.Bytes(), common.FFAddr.Bytes()) &&
|
} else if l2Tx.ToEthAddr == common.FFAddr && l2Tx.ToBJJ != common.EmptyBJJComp {
|
||||||
l2Tx.ToBJJ != common.EmptyBJJComp {
|
|
||||||
// if idx exist for EthAddr&BJJ use it
|
// if idx exist for EthAddr&BJJ use it
|
||||||
_, err := txsel.localAccountsDB.GetIdxByEthAddrBJJ(l2Tx.ToEthAddr, l2Tx.ToBJJ,
|
_, err := txsel.localAccountsDB.GetIdxByEthAddrBJJ(l2Tx.ToEthAddr, l2Tx.ToBJJ,
|
||||||
l2Tx.TokenID)
|
l2Tx.TokenID)
|
||||||
@@ -551,7 +570,8 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs []common.PoolL2Tx,
|
|||||||
}
|
}
|
||||||
if len(l1CoordinatorTxs) >= int(selectionConfig.MaxL1UserTxs)-nL1UserTxs {
|
if len(l1CoordinatorTxs) >= int(selectionConfig.MaxL1UserTxs)-nL1UserTxs {
|
||||||
// L2Tx discarded
|
// L2Tx discarded
|
||||||
return nil, nil, nil, tracerr.Wrap(fmt.Errorf("L2Tx discarded due not slots for L1CoordinatorTx to create a new account for receiver of L2Tx"))
|
return nil, nil, nil, tracerr.Wrap(fmt.Errorf("L2Tx discarded due to no available slots " +
|
||||||
|
"for L1CoordinatorTx to create a new account for receiver of L2Tx"))
|
||||||
}
|
}
|
||||||
|
|
||||||
return &l2Tx, l1CoordinatorTx, accAuth, nil
|
return &l2Tx, l1CoordinatorTx, accAuth, nil
|
||||||
@@ -560,7 +580,7 @@ func (txsel *TxSelector) processTxToEthAddrBJJ(validTxs []common.PoolL2Tx,
|
|||||||
func checkAlreadyPendingToCreate(l1CoordinatorTxs []common.L1Tx, tokenID common.TokenID,
|
func checkAlreadyPendingToCreate(l1CoordinatorTxs []common.L1Tx, tokenID common.TokenID,
|
||||||
addr ethCommon.Address, bjj babyjub.PublicKeyComp) bool {
|
addr ethCommon.Address, bjj babyjub.PublicKeyComp) bool {
|
||||||
for i := 0; i < len(l1CoordinatorTxs); i++ {
|
for i := 0; i < len(l1CoordinatorTxs); i++ {
|
||||||
if bytes.Equal(l1CoordinatorTxs[i].FromEthAddr.Bytes(), addr.Bytes()) &&
|
if l1CoordinatorTxs[i].FromEthAddr == addr &&
|
||||||
l1CoordinatorTxs[i].TokenID == tokenID &&
|
l1CoordinatorTxs[i].TokenID == tokenID &&
|
||||||
l1CoordinatorTxs[i].FromBJJ == bjj {
|
l1CoordinatorTxs[i].FromBJJ == bjj {
|
||||||
return true
|
return true
|
||||||
@@ -571,11 +591,12 @@ func checkAlreadyPendingToCreate(l1CoordinatorTxs []common.L1Tx, tokenID common.
|
|||||||
|
|
||||||
// getL2Profitable returns the profitable selection of L2Txssorted by Nonce
|
// getL2Profitable returns the profitable selection of L2Txssorted by Nonce
|
||||||
func (txsel *TxSelector) getL2Profitable(l2Txs []common.PoolL2Tx, max uint32) []common.PoolL2Tx {
|
func (txsel *TxSelector) getL2Profitable(l2Txs []common.PoolL2Tx, max uint32) []common.PoolL2Tx {
|
||||||
|
// Sort by absolute fee
|
||||||
sort.Sort(txs(l2Txs))
|
sort.Sort(txs(l2Txs))
|
||||||
if len(l2Txs) < int(max) {
|
|
||||||
return l2Txs
|
if len(l2Txs) > int(max) {
|
||||||
}
|
|
||||||
l2Txs = l2Txs[:max]
|
l2Txs = l2Txs[:max]
|
||||||
|
}
|
||||||
|
|
||||||
// sort l2Txs by Nonce. This can be done in many different ways, what
|
// sort l2Txs by Nonce. This can be done in many different ways, what
|
||||||
// is needed is to output the l2Txs where the Nonce of l2Txs for each
|
// is needed is to output the l2Txs where the Nonce of l2Txs for each
|
||||||
|
|||||||
@@ -48,7 +48,7 @@ func initTest(t *testing.T, chainID uint16, hermezContractAddr ethCommon.Address
|
|||||||
BJJ: coordUser.BJJ.Public().Compress(),
|
BJJ: coordUser.BJJ.Public().Compress(),
|
||||||
AccountCreationAuth: nil,
|
AccountCreationAuth: nil,
|
||||||
}
|
}
|
||||||
fmt.Printf("%v", coordAccount)
|
// fmt.Printf("%v\n", coordAccount)
|
||||||
auth := common.AccountCreationAuth{
|
auth := common.AccountCreationAuth{
|
||||||
EthAddr: coordUser.Addr,
|
EthAddr: coordUser.Addr,
|
||||||
BJJ: coordUser.BJJ.Public().Compress(),
|
BJJ: coordUser.BJJ.Public().Compress(),
|
||||||
@@ -424,9 +424,9 @@ func TestPoolL2TxsWithoutEnoughBalance(t *testing.T) {
|
|||||||
_, _, _, _, _, _, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
_, _, _, _, _, _, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
|
|
||||||
expectedTxID0 := "0x0248bae02b5c8c3847d312bfac3a33ae790616e888f2f711f22aeaff007cde92c2" // 1st TransferToEthAddr
|
expectedTxID0 := "0x028847b86613c0b70be18c8622119ed045b42e4e47d7938fa90bb3d1dc14928965" // 1st TransferToEthAddr
|
||||||
expectedTxID1 := "0x0249af018311a393c337ab9174ca2466cba489e49942b4ca4e5c530903671c4aef" // 1st Exit
|
expectedTxID1 := "0x0200b18773dcf56f770d65870fb02041cb59a088fd35b7c3f3df69f8a250b99a42" // 1st Exit
|
||||||
expectedTxID2 := "0x0228b93a261a0cdc62f35588c03bd179d31a0807c28afffdb6a7aaf0c4f017e4cf" // 2nd TransferToEthAddr
|
expectedTxID2 := "0x029720ff506153f970f120ac638cd7ee759eeff2c2012e7634a78e4fdc05c04a90" // 2nd TransferToEthAddr
|
||||||
|
|
||||||
// batch2
|
// batch2
|
||||||
// prepare the PoolL2Txs
|
// prepare the PoolL2Txs
|
||||||
@@ -497,3 +497,210 @@ func TestPoolL2TxsWithoutEnoughBalance(t *testing.T) {
|
|||||||
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
||||||
require.NoError(t, err)
|
require.NoError(t, err)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
func TestTransferToBjj(t *testing.T) {
|
||||||
|
set := `
|
||||||
|
Type: Blockchain
|
||||||
|
AddToken(1)
|
||||||
|
|
||||||
|
CreateAccountDeposit(0) Coord: 0
|
||||||
|
CreateAccountDeposit(0) A: 1000
|
||||||
|
CreateAccountDeposit(0) B: 1000
|
||||||
|
CreateAccountDeposit(1) B: 1000
|
||||||
|
|
||||||
|
> batchL1 // freeze L1User{1}
|
||||||
|
> batchL1 // forge L1User{1}
|
||||||
|
> block
|
||||||
|
`
|
||||||
|
|
||||||
|
chainID := uint16(0)
|
||||||
|
tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||||
|
blocks, err := tc.GenerateBlocks(set)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
hermezContractAddr := ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
||||||
|
txsel := initTest(t, chainID, hermezContractAddr, tc.Users["Coord"])
|
||||||
|
|
||||||
|
// restart nonces of TilContext, as will be set by generating directly
|
||||||
|
// the PoolL2Txs for each specific batch with tc.GeneratePoolL2Txs
|
||||||
|
tc.RestartNonces()
|
||||||
|
|
||||||
|
addTokens(t, tc, txsel.l2db.DB())
|
||||||
|
|
||||||
|
tpc := txprocessor.Config{
|
||||||
|
NLevels: 16,
|
||||||
|
MaxFeeTx: 10,
|
||||||
|
MaxTx: 20,
|
||||||
|
MaxL1Tx: 10,
|
||||||
|
ChainID: chainID,
|
||||||
|
}
|
||||||
|
selectionConfig := &SelectionConfig{
|
||||||
|
MaxL1UserTxs: 5,
|
||||||
|
TxProcessorConfig: tpc,
|
||||||
|
}
|
||||||
|
// batch1 to freeze L1UserTxs that will create some accounts with
|
||||||
|
// positive balance
|
||||||
|
l1UserTxs := []common.L1Tx{}
|
||||||
|
_, _, _, _, _, _, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// Transfer is ToBJJ to a BJJ-only account that doesn't exist
|
||||||
|
// and the coordinator will create it via L1CoordTx.
|
||||||
|
|
||||||
|
batchPoolL2 := `
|
||||||
|
Type: PoolL2
|
||||||
|
PoolTransferToBJJ(0) A-B: 50 (126)
|
||||||
|
`
|
||||||
|
poolL2Txs, err := tc.GeneratePoolL2Txs(batchPoolL2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// add the PoolL2Txs to the l2DB
|
||||||
|
addL2Txs(t, txsel, poolL2Txs)
|
||||||
|
|
||||||
|
l1UserTxs = til.L1TxsToCommonL1Txs(tc.Queues[*blocks[0].Rollup.Batches[1].Batch.ForgeL1TxsNum])
|
||||||
|
_, _, oL1UserTxs, oL1CoordTxs, oL2Txs, discardedL2Txs, err := txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, 4, len(oL1UserTxs))
|
||||||
|
// We expect the coordinator to add an L1CoordTx to create an account for the recipient of the l2tx
|
||||||
|
require.Equal(t, 1, len(oL1CoordTxs))
|
||||||
|
assert.Equal(t, poolL2Txs[0].ToEthAddr, oL1CoordTxs[0].FromEthAddr)
|
||||||
|
assert.Equal(t, poolL2Txs[0].ToBJJ, oL1CoordTxs[0].FromBJJ)
|
||||||
|
// fmt.Printf("DBG l1CoordTx[0]: %+v\n", oL1CoordTxs[0])
|
||||||
|
assert.Equal(t, 1, len(oL2Txs))
|
||||||
|
assert.Equal(t, 0, len(discardedL2Txs))
|
||||||
|
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// Now the BJJ-only account for B is already created, so the transfer
|
||||||
|
// happens without an L1CoordTx that creates the user account.
|
||||||
|
|
||||||
|
batchPoolL2 = `
|
||||||
|
Type: PoolL2
|
||||||
|
PoolTransferToBJJ(0) A-B: 50 (126)
|
||||||
|
`
|
||||||
|
|
||||||
|
poolL2Txs, err = tc.GeneratePoolL2Txs(batchPoolL2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
addL2Txs(t, txsel, poolL2Txs)
|
||||||
|
|
||||||
|
l1UserTxs = []common.L1Tx{}
|
||||||
|
_, _, oL1UserTxs, oL1CoordTxs, oL2Txs, discardedL2Txs, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, 0, len(oL1UserTxs))
|
||||||
|
// Since the BJJ-only account B already exists, the coordinator doesn't add any L1CoordTxs
|
||||||
|
assert.Equal(t, 0, len(oL1CoordTxs))
|
||||||
|
assert.Equal(t, 1, len(oL2Txs))
|
||||||
|
assert.Equal(t, 0, len(discardedL2Txs))
|
||||||
|
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// The transfer now is ToBJJ to a BJJ-only account that doesn't exist
|
||||||
|
// and the coordinator will create it via L1CoordTx. Since it's a
|
||||||
|
// transfer of a token for which the coordinator doesn't have a fee
|
||||||
|
// account, another L1CoordTx will be created for the coordinator to
|
||||||
|
// receive the fees.
|
||||||
|
|
||||||
|
batchPoolL2 = `
|
||||||
|
Type: PoolL2
|
||||||
|
PoolTransferToBJJ(1) B-A: 50 (126)
|
||||||
|
`
|
||||||
|
|
||||||
|
poolL2Txs, err = tc.GeneratePoolL2Txs(batchPoolL2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
addL2Txs(t, txsel, poolL2Txs)
|
||||||
|
|
||||||
|
l1UserTxs = []common.L1Tx{}
|
||||||
|
_, _, oL1UserTxs, oL1CoordTxs, oL2Txs, discardedL2Txs, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, 0, len(oL1UserTxs))
|
||||||
|
// We expect the coordinator to add an L1CoordTx to create an account
|
||||||
|
// to receive the fees by the coordinator and another one for the
|
||||||
|
// recipient of the l2tx
|
||||||
|
assert.Equal(t, 2, len(oL1CoordTxs))
|
||||||
|
// [0] Coordinator account cration for token 1
|
||||||
|
assert.Equal(t, tc.Users["Coord"].Addr, oL1CoordTxs[0].FromEthAddr)
|
||||||
|
// [1] User A BJJ-only account creation for token 1
|
||||||
|
assert.Equal(t, poolL2Txs[0].ToEthAddr, oL1CoordTxs[1].FromEthAddr)
|
||||||
|
assert.Equal(t, poolL2Txs[0].ToBJJ, oL1CoordTxs[1].FromBJJ)
|
||||||
|
assert.Equal(t, common.TokenID(1), oL1CoordTxs[1].TokenID)
|
||||||
|
|
||||||
|
assert.Equal(t, 1, len(oL2Txs))
|
||||||
|
assert.Equal(t, 0, len(discardedL2Txs))
|
||||||
|
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|
||||||
|
func TestTransferManyFromSameAccount(t *testing.T) {
|
||||||
|
set := `
|
||||||
|
Type: Blockchain
|
||||||
|
|
||||||
|
CreateAccountDeposit(0) Coord: 0
|
||||||
|
CreateAccountDeposit(0) A: 1000
|
||||||
|
CreateAccountDeposit(0) B: 1000
|
||||||
|
|
||||||
|
> batchL1 // freeze L1User{1}
|
||||||
|
> batchL1 // forge L1User{1}
|
||||||
|
> block
|
||||||
|
`
|
||||||
|
|
||||||
|
chainID := uint16(0)
|
||||||
|
tc := til.NewContext(chainID, common.RollupConstMaxL1UserTx)
|
||||||
|
blocks, err := tc.GenerateBlocks(set)
|
||||||
|
assert.NoError(t, err)
|
||||||
|
|
||||||
|
hermezContractAddr := ethCommon.HexToAddress("0xc344E203a046Da13b0B4467EB7B3629D0C99F6E6")
|
||||||
|
txsel := initTest(t, chainID, hermezContractAddr, tc.Users["Coord"])
|
||||||
|
|
||||||
|
// restart nonces of TilContext, as will be set by generating directly
|
||||||
|
// the PoolL2Txs for each specific batch with tc.GeneratePoolL2Txs
|
||||||
|
tc.RestartNonces()
|
||||||
|
|
||||||
|
tpc := txprocessor.Config{
|
||||||
|
NLevels: 16,
|
||||||
|
MaxFeeTx: 10,
|
||||||
|
MaxTx: 20,
|
||||||
|
MaxL1Tx: 10,
|
||||||
|
ChainID: chainID,
|
||||||
|
}
|
||||||
|
selectionConfig := &SelectionConfig{
|
||||||
|
MaxL1UserTxs: 5,
|
||||||
|
TxProcessorConfig: tpc,
|
||||||
|
}
|
||||||
|
// batch1 to freeze L1UserTxs
|
||||||
|
l1UserTxs := []common.L1Tx{}
|
||||||
|
_, _, _, _, _, _, err = txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// 8 transfers from the same account
|
||||||
|
|
||||||
|
batchPoolL2 := `
|
||||||
|
Type: PoolL2
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 1
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 2
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 3
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 4
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 5
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 6
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 7
|
||||||
|
PoolTransfer(0) A-B: 10 (126) // 8
|
||||||
|
`
|
||||||
|
poolL2Txs, err := tc.GeneratePoolL2Txs(batchPoolL2)
|
||||||
|
require.NoError(t, err)
|
||||||
|
|
||||||
|
// reorder poolL2Txs so that nonces are not sorted
|
||||||
|
poolL2Txs[0], poolL2Txs[7] = poolL2Txs[7], poolL2Txs[0]
|
||||||
|
poolL2Txs[1], poolL2Txs[6] = poolL2Txs[6], poolL2Txs[1]
|
||||||
|
|
||||||
|
// add the PoolL2Txs to the l2DB
|
||||||
|
addL2Txs(t, txsel, poolL2Txs)
|
||||||
|
// batch 2 to crate some accounts with positive balance, and do 8 L2Tx transfers from account A
|
||||||
|
l1UserTxs = til.L1TxsToCommonL1Txs(tc.Queues[*blocks[0].Rollup.Batches[1].Batch.ForgeL1TxsNum])
|
||||||
|
_, _, oL1UserTxs, oL1CoordTxs, oL2Txs, discardedL2Txs, err := txsel.GetL1L2TxSelection(selectionConfig, l1UserTxs)
|
||||||
|
require.NoError(t, err)
|
||||||
|
assert.Equal(t, 3, len(oL1UserTxs))
|
||||||
|
require.Equal(t, 0, len(oL1CoordTxs))
|
||||||
|
assert.Equal(t, 8, len(oL2Txs))
|
||||||
|
assert.Equal(t, 0, len(discardedL2Txs))
|
||||||
|
err = txsel.l2db.StartForging(common.TxIDsFromPoolL2Txs(oL2Txs), txsel.localAccountsDB.CurrentBatch())
|
||||||
|
require.NoError(t, err)
|
||||||
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user