You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

267 lines
9.9 KiB

4 years ago
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Update coordinator, call all api update functions - Common: - Rename Block.EthBlockNum to Block.Num to avoid unneeded repetition - API: - Add UpdateNetworkInfoBlock to update just block information, to be used when the node is not yet synchronized - Node: - Call API.UpdateMetrics and UpdateRecommendedFee in a loop, with configurable time intervals - Synchronizer: - When mapping events by TxHash, use an array to support the possibility of multiple calls of the same function happening in the same transaction (for example, a smart contract in a single transaction could call withdraw with delay twice, which would generate 2 withdraw events, and 2 deposit events). - In Stats, keep entire LastBlock instead of just the blockNum - In Stats, add lastL1BatchBlock - Test Stats and SCVars - Coordinator: - Enable writing the BatchInfo in every step of the pipeline to disk (with JSON text files) for debugging purposes. - Move the Pipeline functionality from the Coordinator to its own struct (Pipeline) - Implement shouldL1lL2Batch - In TxManager, implement logic to perform several attempts when doing ethereum node RPC calls before considering the error. (Both for calls to forgeBatch and transaction receipt) - In TxManager, reorganize the flow and note the specific points in which actions are made when err != nil - HistoryDB: - Implement GetLastL1BatchBlockNum: returns the blockNum of the latest forged l1Batch, to help the coordinator decide when to forge an L1Batch. - EthereumClient and test.Client: - Update EthBlockByNumber to return the last block when the passed number is -1.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
  1. package config
  2. import (
  3. "fmt"
  4. "io/ioutil"
  5. "time"
  6. "github.com/BurntSushi/toml"
  7. ethCommon "github.com/ethereum/go-ethereum/common"
  8. "github.com/hermeznetwork/hermez-node/common"
  9. "github.com/hermeznetwork/tracerr"
  10. "github.com/iden3/go-iden3-crypto/babyjub"
  11. "gopkg.in/go-playground/validator.v9"
  12. )
  13. // Duration is a wrapper type that parses time duration from text.
  14. type Duration struct {
  15. time.Duration `validate:"required"`
  16. }
  17. // UnmarshalText unmarshalls time duration from text.
  18. func (d *Duration) UnmarshalText(data []byte) error {
  19. duration, err := time.ParseDuration(string(data))
  20. if err != nil {
  21. return tracerr.Wrap(err)
  22. }
  23. d.Duration = duration
  24. return nil
  25. }
  26. // ServerProof is the server proof configuration data.
  27. type ServerProof struct {
  28. // URL is the server proof API URL
  29. URL string `validate:"required"`
  30. }
  31. // Coordinator is the coordinator specific configuration.
  32. type Coordinator struct {
  33. // ForgerAddress is the address under which this coordinator is forging
  34. ForgerAddress ethCommon.Address `validate:"required"`
  35. // FeeAccount is the Hermez account that the coordinator uses to receive fees
  36. FeeAccount struct {
  37. // Address is the ethereum address of the account to receive fees
  38. Address ethCommon.Address `validate:"required"`
  39. // BJJ is the baby jub jub public key of the account to receive fees
  40. BJJ babyjub.PublicKeyComp `validate:"required"`
  41. } `validate:"required"`
  42. // ConfirmBlocks is the number of confirmation blocks to wait for sent
  43. // ethereum transactions before forgetting about them
  44. ConfirmBlocks int64 `validate:"required"`
  45. // L1BatchTimeoutPerc is the portion of the range before the L1Batch
  46. // timeout that will trigger a schedule to forge an L1Batch
  47. L1BatchTimeoutPerc float64 `validate:"required"`
  48. // ProofServerPollInterval is the waiting interval between polling the
  49. // ProofServer while waiting for a particular status
  50. ProofServerPollInterval Duration `validate:"required"`
  51. // ForgeRetryInterval is the waiting interval between calls forge a
  52. // batch after an error
  53. ForgeRetryInterval Duration `validate:"required"`
  54. // SyncRetryInterval is the waiting interval between calls to the main
  55. // handler of a synced block after an error
  56. SyncRetryInterval Duration `validate:"required"`
  57. // L2DB is the DB that holds the pool of L2Txs
  58. L2DB struct {
  59. // SafetyPeriod is the number of batches after which
  60. // non-pending L2Txs are deleted from the pool
  61. SafetyPeriod common.BatchNum `validate:"required"`
  62. // MaxTxs is the number of L2Txs that once reached triggers
  63. // deletion of old L2Txs
  64. MaxTxs uint32 `validate:"required"`
  65. // TTL is the Time To Live for L2Txs in the pool. Once MaxTxs
  66. // L2Txs is reached, L2Txs older than TTL will be deleted.
  67. TTL Duration `validate:"required"`
  68. // PurgeBatchDelay is the delay between batches to purge outdated transactions
  69. PurgeBatchDelay int64 `validate:"required"`
  70. // InvalidateBatchDelay is the delay between batches to mark invalid transactions
  71. InvalidateBatchDelay int64 `validate:"required"`
  72. // PurgeBlockDelay is the delay between blocks to purge outdated transactions
  73. PurgeBlockDelay int64 `validate:"required"`
  74. // InvalidateBlockDelay is the delay between blocks to mark invalid transactions
  75. InvalidateBlockDelay int64 `validate:"required"`
  76. } `validate:"required"`
  77. TxSelector struct {
  78. // Path where the TxSelector StateDB is stored
  79. Path string `validate:"required"`
  80. } `validate:"required"`
  81. BatchBuilder struct {
  82. // Path where the BatchBuilder StateDB is stored
  83. Path string `validate:"required"`
  84. } `validate:"required"`
  85. ServerProofs []ServerProof `validate:"required"`
  86. Circuit struct {
  87. // MaxTx is the maximum number of txs supported by the circuit
  88. MaxTx int64 `validate:"required"`
  89. // NLevels is the maximum number of merkle tree levels
  90. // supported by the circuit
  91. NLevels int64 `validate:"required"`
  92. } `validate:"required"`
  93. EthClient struct {
  94. // CallGasLimit is the default gas limit set for ethereum
  95. // calls, except for methods where a particular gas limit is
  96. // harcoded because it's known to be a big value
  97. CallGasLimit uint64 `validate:"required"`
  98. // GasPriceDiv is the gas price division
  99. GasPriceDiv uint64 `validate:"required"`
  100. // CheckLoopInterval is the waiting interval between receipt
  101. // checks of ethereum transactions in the TxManager
  102. CheckLoopInterval Duration `validate:"required"`
  103. // Attempts is the number of attempts to do an eth client RPC
  104. // call before giving up
  105. Attempts int `validate:"required"`
  106. // AttemptsDelay is delay between attempts do do an eth client
  107. // RPC call
  108. AttemptsDelay Duration `validate:"required"`
  109. // Keystore is the ethereum keystore where private keys are kept
  110. Keystore struct {
  111. // Path to the keystore
  112. Path string `validate:"required"`
  113. // Password used to decrypt the keys in the keystore
  114. Password string `validate:"required"`
  115. } `validate:"required"`
  116. } `validate:"required"`
  117. API struct {
  118. // Coordinator enables the coordinator API endpoints
  119. Coordinator bool
  120. } `validate:"required"`
  121. Debug struct {
  122. // BatchPath if set, specifies the path where batchInfo is stored
  123. // in JSON in every step/update of the pipeline
  124. BatchPath string
  125. // LightScrypt if set, uses light parameters for the ethereum
  126. // keystore encryption algorithm.
  127. LightScrypt bool
  128. // RollupVerifierIndex is the index of the verifier to use in
  129. // the Rollup smart contract. The verifier chosen by index
  130. // must match with the Circuit parameters.
  131. RollupVerifierIndex *int
  132. }
  133. }
  134. // Node is the hermez node configuration.
  135. type Node struct {
  136. PriceUpdater struct {
  137. // Interval between price updater calls
  138. Interval Duration `valudate:"required"`
  139. // URL of the token prices provider
  140. URL string `valudate:"required"`
  141. // Type of the API of the token prices provider
  142. Type string `valudate:"required"`
  143. } `validate:"required"`
  144. StateDB struct {
  145. // Path where the synchronizer StateDB is stored
  146. Path string `validate:"required"`
  147. // Keep is the number of checkpoints to keep
  148. Keep int `validate:"required"`
  149. } `validate:"required"`
  150. PostgreSQL struct {
  151. // Port of the PostgreSQL server
  152. Port int `validate:"required"`
  153. // Host of the PostgreSQL server
  154. Host string `validate:"required"`
  155. // User of the PostgreSQL server
  156. User string `validate:"required"`
  157. // Password of the PostgreSQL server
  158. Password string `validate:"required"`
  159. // Name of the PostgreSQL server database
  160. Name string `validate:"required"`
  161. } `validate:"required"`
  162. Web3 struct {
  163. // URL is the URL of the web3 ethereum-node RPC server
  164. URL string `validate:"required"`
  165. } `validate:"required"`
  166. Synchronizer struct {
  167. // SyncLoopInterval is the interval between attempts to
  168. // synchronize a new block from an ethereum node
  169. SyncLoopInterval Duration `validate:"required"`
  170. // StatsRefreshPeriod is the interval between updates of the
  171. // synchronizer state Eth parameters (`Eth.LastBlock` and
  172. // `Eth.LastBatch`). This value only affects the reported % of
  173. // synchronization of blocks and batches, nothing else.
  174. StatsRefreshPeriod Duration `validate:"required"`
  175. } `validate:"required"`
  176. SmartContracts struct {
  177. // Rollup is the address of the Hermez.sol smart contract
  178. Rollup ethCommon.Address `validate:"required"`
  179. // Rollup is the address of the HermezAuctionProtocol.sol smart
  180. // contract
  181. Auction ethCommon.Address `validate:"required"`
  182. // WDelayer is the address of the WithdrawalDelayer.sol smart
  183. // contract
  184. WDelayer ethCommon.Address `validate:"required"`
  185. // TokenHEZ is the address of the HEZTokenFull.sol smart
  186. // contract
  187. TokenHEZ ethCommon.Address `validate:"required"`
  188. // TokenHEZName is the name of the HEZ token deployed at
  189. // TokenHEZ address
  190. TokenHEZName string `validate:"required"`
  191. } `validate:"required"`
  192. API struct {
  193. // Address where the API will listen if set
  194. Address string
  195. // Explorer enables the Explorer API endpoints
  196. Explorer bool
  197. // UpdateMetricsInterval is the interval between updates of the
  198. // API metrics
  199. UpdateMetricsInterval Duration
  200. // UpdateRecommendedFeeInterval is the interval between updates of the
  201. // recommended fees
  202. UpdateRecommendedFeeInterval Duration
  203. // Maximum concurrent connections allowed between API and SQL
  204. MaxSQLConnections int `validate:"required"`
  205. // SQLConnectionTimeout is the maximum amount of time that an API request
  206. // can wait to stablish a SQL connection
  207. SQLConnectionTimeout Duration
  208. } `validate:"required"`
  209. Debug struct {
  210. // APIAddress is the address where the debugAPI will listen if
  211. // set
  212. APIAddress string
  213. // MeddlerLogs enables meddler debug mode, where unused columns and struct
  214. // fields will be logged
  215. MeddlerLogs bool
  216. }
  217. Coordinator Coordinator `validate:"-"`
  218. }
  219. // Load loads a generic config.
  220. func Load(path string, cfg interface{}) error {
  221. bs, err := ioutil.ReadFile(path) //nolint:gosec
  222. if err != nil {
  223. return tracerr.Wrap(err)
  224. }
  225. cfgToml := string(bs)
  226. if _, err := toml.Decode(cfgToml, cfg); err != nil {
  227. return tracerr.Wrap(err)
  228. }
  229. return nil
  230. }
  231. // LoadCoordinator loads the Coordinator configuration from path.
  232. func LoadCoordinator(path string) (*Node, error) {
  233. var cfg Node
  234. if err := Load(path, &cfg); err != nil {
  235. return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
  236. }
  237. validate := validator.New()
  238. if err := validate.Struct(cfg); err != nil {
  239. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  240. }
  241. if err := validate.Struct(cfg.Coordinator); err != nil {
  242. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  243. }
  244. return &cfg, nil
  245. }
  246. // LoadNode loads the Node configuration from path.
  247. func LoadNode(path string) (*Node, error) {
  248. var cfg Node
  249. if err := Load(path, &cfg); err != nil {
  250. return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
  251. }
  252. validate := validator.New()
  253. if err := validate.Struct(cfg); err != nil {
  254. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  255. }
  256. return &cfg, nil
  257. }