You can not select more than 25 topics Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.

247 lines
8.8 KiB

4 years ago
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
Redo coordinator structure, connect API to node - API: - Modify the constructor so that hardcoded rollup constants don't need to be passed (introduce a `Config` and use `configAPI` internally) - Common: - Update rollup constants with proper *big.Int when required - Add BidCoordinator and Slot structs used by the HistoryDB and Synchronizer. - Add helper methods to AuctionConstants - AuctionVariables: Add column `DefaultSlotSetBidSlotNum` (in the SQL table: `default_slot_set_bid_slot_num`), which indicates at which slotNum does the `DefaultSlotSetBid` specified starts applying. - Config: - Move coordinator exclusive configuration from the node config to the coordinator config - Coordinator: - Reorganize the code towards having the goroutines started and stopped from the coordinator itself instead of the node. - Remove all stop and stopped channels, and use context.Context and sync.WaitGroup instead. - Remove BatchInfo setters and assing variables directly - In ServerProof and ServerProofPool use context instead stop channel. - Use message passing to notify the coordinator about sync updates and reorgs - Introduce the Pipeline, which can be started and stopped by the Coordinator - Introduce the TxManager, which manages ethereum transactions (the TxManager is also in charge of making the forge call to the rollup smart contract). The TxManager keeps ethereum transactions and: 1. Waits for the transaction to be accepted 2. Waits for the transaction to be confirmed for N blocks - In forge logic, first prepare a batch and then wait for an available server proof to have all work ready once the proof server is ready. - Remove the `isForgeSequence` method which was querying the smart contract, and instead use notifications sent by the Synchronizer to figure out if it's forging time. - Update test (which is a minimal test to manually see if the coordinator starts) - HistoryDB: - Add method to get the number of batches in a slot (used to detect when a slot has passed the bid winner forging deadline) - Add method to get the best bid and associated coordinator of a slot (used to detect the forgerAddress that can forge the slot) - General: - Rename some instances of `currentBlock` to `lastBlock` to be more clear. - Node: - Connect the API to the node and call the methods to update cached state when the sync advances blocks. - Call methods to update Coordinator state when the sync advances blocks and finds reorgs. - Synchronizer: - Add Auction field in the Stats, which contain the current slot with info about highest bidder and other related info required to know who can forge in the current block. - Better organization of cached state: - On Sync, update the internal cached state - On Init or Reorg, load the state from HistoryDB into the internal cached state.
4 years ago
Update coordinator, call all api update functions - Common: - Rename Block.EthBlockNum to Block.Num to avoid unneeded repetition - API: - Add UpdateNetworkInfoBlock to update just block information, to be used when the node is not yet synchronized - Node: - Call API.UpdateMetrics and UpdateRecommendedFee in a loop, with configurable time intervals - Synchronizer: - When mapping events by TxHash, use an array to support the possibility of multiple calls of the same function happening in the same transaction (for example, a smart contract in a single transaction could call withdraw with delay twice, which would generate 2 withdraw events, and 2 deposit events). - In Stats, keep entire LastBlock instead of just the blockNum - In Stats, add lastL1BatchBlock - Test Stats and SCVars - Coordinator: - Enable writing the BatchInfo in every step of the pipeline to disk (with JSON text files) for debugging purposes. - Move the Pipeline functionality from the Coordinator to its own struct (Pipeline) - Implement shouldL1lL2Batch - In TxManager, implement logic to perform several attempts when doing ethereum node RPC calls before considering the error. (Both for calls to forgeBatch and transaction receipt) - In TxManager, reorganize the flow and note the specific points in which actions are made when err != nil - HistoryDB: - Implement GetLastL1BatchBlockNum: returns the blockNum of the latest forged l1Batch, to help the coordinator decide when to forge an L1Batch. - EthereumClient and test.Client: - Update EthBlockByNumber to return the last block when the passed number is -1.
4 years ago
Update missing parts, improve til, and more - Node - Updated configuration to initialize the interface to all the smart contracts - Common - Moved BlockData and BatchData types to common so that they can be shared among: historydb, til and synchronizer - Remove hash.go (it was never used) - Remove slot.go (it was never used) - Remove smartcontractparams.go (it was never used, and appropriate structs are defined in `eth/`) - Comment state / status method until requirements of this method are properly defined, and move it to Synchronizer - Synchronizer - Simplify `Sync` routine to only sync one block per call, and return useful information. - Use BlockData and BatchData from common - Check that events belong to the expected block hash - In L1Batch, query L1UserTxs from HistoryDB - Fill ERC20 token information - Test AddTokens with test.Client - HistryDB - Use BlockData and BatchData from common - Add `GetAllTokens` method - Uncomment and update GetL1UserTxs (with corresponding tests) - Til - Rename all instances of RegisterToken to AddToken (to follow the smart contract implementation naming) - Use BlockData and BatchData from common - Move testL1CoordinatorTxs and testL2Txs to a separate struct from BatchData in Context - Start Context with BatchNum = 1 (which the protocol defines to be the first batchNum) - In every Batch, set StateRoot and ExitRoot to a non-nil big.Int (zero). - In all L1Txs, if LoadAmount is not used, set it to 0; if Amount is not used, set it to 0; so that no *big.Int is nil. - In L1UserTx, don't set BatchNum, because when L1UserTxs are created and obtained by the synchronizer, the BatchNum is not known yet (it's a synchronizer job to set it) - In L1UserTxs, set `UserOrigin` and set `ToForgeL1TxsNum`.
4 years ago
  1. package config
  2. import (
  3. "fmt"
  4. "io/ioutil"
  5. "time"
  6. "github.com/BurntSushi/toml"
  7. ethCommon "github.com/ethereum/go-ethereum/common"
  8. "github.com/hermeznetwork/hermez-node/common"
  9. "github.com/hermeznetwork/tracerr"
  10. "gopkg.in/go-playground/validator.v9"
  11. )
  12. // Duration is a wrapper type that parses time duration from text.
  13. type Duration struct {
  14. time.Duration `validate:"required"`
  15. }
  16. // UnmarshalText unmarshalls time duration from text.
  17. func (d *Duration) UnmarshalText(data []byte) error {
  18. duration, err := time.ParseDuration(string(data))
  19. if err != nil {
  20. return tracerr.Wrap(err)
  21. }
  22. d.Duration = duration
  23. return nil
  24. }
  25. // ServerProof is the server proof configuration data.
  26. type ServerProof struct {
  27. // URL is the server proof API URL
  28. URL string `validate:"required"`
  29. }
  30. // Coordinator is the coordinator specific configuration.
  31. type Coordinator struct {
  32. // ForgerAddress is the address under which this coordinator is forging
  33. ForgerAddress ethCommon.Address `validate:"required"`
  34. // ConfirmBlocks is the number of confirmation blocks to wait for sent
  35. // ethereum transactions before forgetting about them
  36. ConfirmBlocks int64 `validate:"required"`
  37. // L1BatchTimeoutPerc is the portion of the range before the L1Batch
  38. // timeout that will trigger a schedule to forge an L1Batch
  39. L1BatchTimeoutPerc float64 `validate:"required"`
  40. // ProofServerPollInterval is the waiting interval between polling the
  41. // ProofServer while waiting for a particular status
  42. ProofServerPollInterval Duration `validate:"required"`
  43. // SyncRetryInterval is the waiting interval between calls to the main
  44. // handler of a synced block after an error
  45. SyncRetryInterval Duration `validate:"required"`
  46. // L2DB is the DB that holds the pool of L2Txs
  47. L2DB struct {
  48. // SafetyPeriod is the number of batches after which
  49. // non-pending L2Txs are deleted from the pool
  50. SafetyPeriod common.BatchNum `validate:"required"`
  51. // MaxTxs is the number of L2Txs that once reached triggers
  52. // deletion of old L2Txs
  53. MaxTxs uint32 `validate:"required"`
  54. // TTL is the Time To Live for L2Txs in the pool. Once MaxTxs
  55. // L2Txs is reached, L2Txs older than TTL will be deleted.
  56. TTL Duration `validate:"required"`
  57. // PurgeBatchDelay is the delay between batches to purge outdated transactions
  58. PurgeBatchDelay int64 `validate:"required"`
  59. // InvalidateBatchDelay is the delay between batches to mark invalid transactions
  60. InvalidateBatchDelay int64 `validate:"required"`
  61. // PurgeBlockDelay is the delay between blocks to purge outdated transactions
  62. PurgeBlockDelay int64 `validate:"required"`
  63. // InvalidateBlockDelay is the delay between blocks to mark invalid transactions
  64. InvalidateBlockDelay int64 `validate:"required"`
  65. } `validate:"required"`
  66. TxSelector struct {
  67. // Path where the TxSelector StateDB is stored
  68. Path string `validate:"required"`
  69. } `validate:"required"`
  70. BatchBuilder struct {
  71. // Path where the BatchBuilder StateDB is stored
  72. Path string `validate:"required"`
  73. } `validate:"required"`
  74. ServerProofs []ServerProof `validate:"required"`
  75. Circuit struct {
  76. // VerifierIdx uint8 `validate:"required"`
  77. // MaxTx is the maximum number of txs supported by the circuit
  78. MaxTx int64 `validate:"required"`
  79. // NLevels is the maximum number of merkle tree levels
  80. // supported by the circuit
  81. NLevels int64 `validate:"required"`
  82. } `validate:"required"`
  83. EthClient struct {
  84. // CallGasLimit is the default gas limit set for ethereum
  85. // calls, except for methods where a particular gas limit is
  86. // harcoded because it's known to be a big value
  87. CallGasLimit uint64 `validate:"required"`
  88. // GasPriceDiv is the gas price division
  89. GasPriceDiv uint64 `validate:"required"`
  90. // CheckLoopInterval is the waiting interval between receipt
  91. // checks of ethereum transactions in the TxManager
  92. CheckLoopInterval Duration `validate:"required"`
  93. // Attempts is the number of attempts to do an eth client RPC
  94. // call before giving up
  95. Attempts int `validate:"required"`
  96. // AttemptsDelay is delay between attempts do do an eth client
  97. // RPC call
  98. AttemptsDelay Duration `validate:"required"`
  99. // Keystore is the ethereum keystore where private keys are kept
  100. Keystore struct {
  101. // Path to the keystore
  102. Path string `validate:"required"`
  103. // Password used to decrypt the keys in the keystore
  104. Password string `validate:"required"`
  105. } `validate:"required"`
  106. } `validate:"required"`
  107. API struct {
  108. // Coordinator enables the coordinator API endpoints
  109. Coordinator bool
  110. } `validate:"required"`
  111. Debug struct {
  112. // BatchPath if set, specifies the path where batchInfo is stored
  113. // in JSON in every step/update of the pipeline
  114. BatchPath string
  115. // LightScrypt if set, uses light parameters for the ethereum
  116. // keystore encryption algorithm.
  117. LightScrypt bool
  118. }
  119. }
  120. // Node is the hermez node configuration.
  121. type Node struct {
  122. PriceUpdater struct {
  123. // Interval between price updater calls
  124. Interval Duration `valudate:"required"`
  125. // URL of the token prices provider
  126. URL string `valudate:"required"`
  127. // Type of the API of the token prices provider
  128. Type string `valudate:"required"`
  129. } `validate:"required"`
  130. StateDB struct {
  131. // Path where the synchronizer StateDB is stored
  132. Path string `validate:"required"`
  133. // Keep is the number of checkpoints to keep
  134. Keep int `validate:"required"`
  135. } `validate:"required"`
  136. PostgreSQL struct {
  137. // Port of the PostgreSQL server
  138. Port int `validate:"required"`
  139. // Host of the PostgreSQL server
  140. Host string `validate:"required"`
  141. // User of the PostgreSQL server
  142. User string `validate:"required"`
  143. // Password of the PostgreSQL server
  144. Password string `validate:"required"`
  145. // Name of the PostgreSQL server database
  146. Name string `validate:"required"`
  147. } `validate:"required"`
  148. Web3 struct {
  149. // URL is the URL of the web3 ethereum-node RPC server
  150. URL string `validate:"required"`
  151. } `validate:"required"`
  152. Synchronizer struct {
  153. // SyncLoopInterval is the interval between attempts to
  154. // synchronize a new block from an ethereum node
  155. SyncLoopInterval Duration `validate:"required"`
  156. // StatsRefreshPeriod is the interval between updates of the
  157. // synchronizer state Eth parameters (`Eth.LastBlock` and
  158. // `Eth.LastBatch`)
  159. StatsRefreshPeriod Duration `validate:"required"`
  160. } `validate:"required"`
  161. SmartContracts struct {
  162. // Rollup is the address of the Hermez.sol smart contract
  163. Rollup ethCommon.Address `validate:"required"`
  164. // Rollup is the address of the HermezAuctionProtocol.sol smart
  165. // contract
  166. Auction ethCommon.Address `validate:"required"`
  167. // WDelayer is the address of the WithdrawalDelayer.sol smart
  168. // contract
  169. WDelayer ethCommon.Address `validate:"required"`
  170. // TokenHEZ is the address of the HEZTokenFull.sol smart
  171. // contract
  172. TokenHEZ ethCommon.Address `validate:"required"`
  173. // TokenHEZName is the name of the HEZ token deployed at
  174. // TokenHEZ address
  175. TokenHEZName string `validate:"required"`
  176. } `validate:"required"`
  177. API struct {
  178. // Address where the API will listen if set
  179. Address string
  180. // Explorer enables the Explorer API endpoints
  181. Explorer bool
  182. // UpdateMetricsInterval is the interval between updates of the
  183. // API metrics
  184. UpdateMetricsInterval Duration
  185. // UpdateMetricsInterval is the interval between updates of the
  186. // recommended fees
  187. UpdateRecommendedFeeInterval Duration
  188. } `validate:"required"`
  189. Debug struct {
  190. // APIAddress is the address where the debugAPI will listen if
  191. // set
  192. APIAddress string
  193. // MeddlerLogs enables meddler debug mode, where unused columns and struct
  194. // fields will be logged
  195. MeddlerLogs bool
  196. }
  197. Coordinator Coordinator `validate:"-"`
  198. }
  199. // Load loads a generic config.
  200. func Load(path string, cfg interface{}) error {
  201. bs, err := ioutil.ReadFile(path) //nolint:gosec
  202. if err != nil {
  203. return tracerr.Wrap(err)
  204. }
  205. cfgToml := string(bs)
  206. if _, err := toml.Decode(cfgToml, cfg); err != nil {
  207. return tracerr.Wrap(err)
  208. }
  209. return nil
  210. }
  211. // LoadCoordinator loads the Coordinator configuration from path.
  212. func LoadCoordinator(path string) (*Node, error) {
  213. var cfg Node
  214. if err := Load(path, &cfg); err != nil {
  215. return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
  216. }
  217. validate := validator.New()
  218. if err := validate.Struct(cfg); err != nil {
  219. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  220. }
  221. if err := validate.Struct(cfg.Coordinator); err != nil {
  222. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  223. }
  224. return &cfg, nil
  225. }
  226. // LoadNode loads the Node configuration from path.
  227. func LoadNode(path string) (*Node, error) {
  228. var cfg Node
  229. if err := Load(path, &cfg); err != nil {
  230. return nil, tracerr.Wrap(fmt.Errorf("error loading node configuration file: %w", err))
  231. }
  232. validate := validator.New()
  233. if err := validate.Struct(cfg); err != nil {
  234. return nil, tracerr.Wrap(fmt.Errorf("error validating configuration file: %w", err))
  235. }
  236. return &cfg, nil
  237. }