@aztec/p2p 0.0.0-test.1 → 0.0.1-commit.1142ef1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dest/bootstrap/bootstrap.d.ts +1 -1
- package/dest/bootstrap/bootstrap.d.ts.map +1 -1
- package/dest/bootstrap/bootstrap.js +22 -9
- package/dest/client/factory.d.ts +15 -5
- package/dest/client/factory.d.ts.map +1 -1
- package/dest/client/factory.js +60 -25
- package/dest/client/index.d.ts +2 -1
- package/dest/client/index.d.ts.map +1 -1
- package/dest/client/index.js +1 -0
- package/dest/client/interface.d.ts +170 -0
- package/dest/client/interface.d.ts.map +1 -0
- package/dest/client/interface.js +9 -0
- package/dest/client/p2p_client.d.ts +75 -193
- package/dest/client/p2p_client.d.ts.map +1 -1
- package/dest/client/p2p_client.js +757 -228
- package/dest/config.d.ts +148 -125
- package/dest/config.d.ts.map +1 -1
- package/dest/config.js +180 -34
- package/dest/enr/generate-enr.d.ts +11 -3
- package/dest/enr/generate-enr.d.ts.map +1 -1
- package/dest/enr/generate-enr.js +27 -5
- package/dest/enr/index.d.ts +1 -1
- package/dest/errors/attestation-pool.error.d.ts +7 -0
- package/dest/errors/attestation-pool.error.d.ts.map +1 -0
- package/dest/errors/attestation-pool.error.js +12 -0
- package/dest/errors/reqresp.error.d.ts +1 -1
- package/dest/errors/reqresp.error.d.ts.map +1 -1
- package/dest/index.d.ts +4 -1
- package/dest/index.d.ts.map +1 -1
- package/dest/index.js +2 -0
- package/dest/mem_pools/attestation_pool/attestation_pool.d.ts +104 -25
- package/dest/mem_pools/attestation_pool/attestation_pool.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.d.ts +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/attestation_pool_test_suite.js +288 -174
- package/dest/mem_pools/attestation_pool/index.d.ts +1 -1
- package/dest/mem_pools/attestation_pool/kv_attestation_pool.d.ts +29 -11
- package/dest/mem_pools/attestation_pool/kv_attestation_pool.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/kv_attestation_pool.js +168 -62
- package/dest/mem_pools/attestation_pool/memory_attestation_pool.d.ts +24 -10
- package/dest/mem_pools/attestation_pool/memory_attestation_pool.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/memory_attestation_pool.js +133 -82
- package/dest/mem_pools/attestation_pool/mocks.d.ts +232 -11
- package/dest/mem_pools/attestation_pool/mocks.d.ts.map +1 -1
- package/dest/mem_pools/attestation_pool/mocks.js +15 -20
- package/dest/mem_pools/index.d.ts +1 -1
- package/dest/mem_pools/instrumentation.d.ts +16 -12
- package/dest/mem_pools/instrumentation.d.ts.map +1 -1
- package/dest/mem_pools/instrumentation.js +55 -40
- package/dest/mem_pools/interface.d.ts +3 -4
- package/dest/mem_pools/interface.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/aztec_kv_tx_pool.d.ts +70 -16
- package/dest/mem_pools/tx_pool/aztec_kv_tx_pool.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/aztec_kv_tx_pool.js +452 -142
- package/dest/mem_pools/tx_pool/eviction/eviction_manager.d.ts +18 -0
- package/dest/mem_pools/tx_pool/eviction/eviction_manager.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/eviction_manager.js +56 -0
- package/dest/mem_pools/tx_pool/eviction/eviction_strategy.d.ts +83 -0
- package/dest/mem_pools/tx_pool/eviction/eviction_strategy.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/eviction_strategy.js +5 -0
- package/dest/mem_pools/tx_pool/eviction/insufficient_fee_payer_balance_rule.d.ts +15 -0
- package/dest/mem_pools/tx_pool/eviction/insufficient_fee_payer_balance_rule.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/insufficient_fee_payer_balance_rule.js +88 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_mining_rule.d.ts +17 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_mining_rule.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_mining_rule.js +84 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_reorg_rule.d.ts +19 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_reorg_rule.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/invalid_txs_after_reorg_rule.js +76 -0
- package/dest/mem_pools/tx_pool/eviction/low_priority_eviction_rule.d.ts +26 -0
- package/dest/mem_pools/tx_pool/eviction/low_priority_eviction_rule.d.ts.map +1 -0
- package/dest/mem_pools/tx_pool/eviction/low_priority_eviction_rule.js +84 -0
- package/dest/mem_pools/tx_pool/index.d.ts +1 -2
- package/dest/mem_pools/tx_pool/index.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/index.js +0 -1
- package/dest/mem_pools/tx_pool/priority.d.ts +5 -1
- package/dest/mem_pools/tx_pool/priority.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/priority.js +7 -2
- package/dest/mem_pools/tx_pool/tx_pool.d.ts +72 -11
- package/dest/mem_pools/tx_pool/tx_pool.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/tx_pool_test_suite.d.ts +1 -1
- package/dest/mem_pools/tx_pool/tx_pool_test_suite.d.ts.map +1 -1
- package/dest/mem_pools/tx_pool/tx_pool_test_suite.js +276 -45
- package/dest/msg_validators/attestation_validator/attestation_validator.d.ts +7 -5
- package/dest/msg_validators/attestation_validator/attestation_validator.d.ts.map +1 -1
- package/dest/msg_validators/attestation_validator/attestation_validator.js +48 -10
- package/dest/msg_validators/attestation_validator/fisherman_attestation_validator.d.ts +20 -0
- package/dest/msg_validators/attestation_validator/fisherman_attestation_validator.d.ts.map +1 -0
- package/dest/msg_validators/attestation_validator/fisherman_attestation_validator.js +64 -0
- package/dest/msg_validators/attestation_validator/index.d.ts +2 -1
- package/dest/msg_validators/attestation_validator/index.d.ts.map +1 -1
- package/dest/msg_validators/attestation_validator/index.js +1 -0
- package/dest/msg_validators/index.d.ts +2 -2
- package/dest/msg_validators/index.d.ts.map +1 -1
- package/dest/msg_validators/index.js +1 -1
- package/dest/msg_validators/msg_seen_validator/msg_seen_validator.d.ts +10 -0
- package/dest/msg_validators/msg_seen_validator/msg_seen_validator.d.ts.map +1 -0
- package/dest/msg_validators/msg_seen_validator/msg_seen_validator.js +36 -0
- package/dest/msg_validators/proposal_validator/block_proposal_validator.d.ts +9 -0
- package/dest/msg_validators/proposal_validator/block_proposal_validator.d.ts.map +1 -0
- package/dest/msg_validators/proposal_validator/block_proposal_validator.js +6 -0
- package/dest/msg_validators/proposal_validator/checkpoint_proposal_validator.d.ts +9 -0
- package/dest/msg_validators/proposal_validator/checkpoint_proposal_validator.d.ts.map +1 -0
- package/dest/msg_validators/proposal_validator/checkpoint_proposal_validator.js +6 -0
- package/dest/msg_validators/proposal_validator/index.d.ts +4 -0
- package/dest/msg_validators/proposal_validator/index.d.ts.map +1 -0
- package/dest/msg_validators/proposal_validator/index.js +3 -0
- package/dest/msg_validators/proposal_validator/proposal_validator.d.ts +13 -0
- package/dest/msg_validators/proposal_validator/proposal_validator.d.ts.map +1 -0
- package/dest/msg_validators/proposal_validator/proposal_validator.js +80 -0
- package/dest/msg_validators/proposal_validator/proposal_validator_test_suite.d.ts +23 -0
- package/dest/msg_validators/proposal_validator/proposal_validator_test_suite.d.ts.map +1 -0
- package/dest/msg_validators/proposal_validator/proposal_validator_test_suite.js +183 -0
- package/dest/msg_validators/tx_validator/aggregate_tx_validator.d.ts +1 -1
- package/dest/msg_validators/tx_validator/aggregate_tx_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/allowed_public_setup.d.ts +3 -0
- package/dest/msg_validators/tx_validator/allowed_public_setup.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/allowed_public_setup.js +27 -0
- package/dest/msg_validators/tx_validator/archive_cache.d.ts +14 -0
- package/dest/msg_validators/tx_validator/archive_cache.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/archive_cache.js +22 -0
- package/dest/msg_validators/tx_validator/block_header_validator.d.ts +2 -2
- package/dest/msg_validators/tx_validator/block_header_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/block_header_validator.js +4 -4
- package/dest/msg_validators/tx_validator/data_validator.d.ts +1 -1
- package/dest/msg_validators/tx_validator/data_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/data_validator.js +56 -86
- package/dest/msg_validators/tx_validator/double_spend_validator.d.ts +1 -3
- package/dest/msg_validators/tx_validator/double_spend_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/double_spend_validator.js +21 -27
- package/dest/msg_validators/tx_validator/factory.d.ts +16 -0
- package/dest/msg_validators/tx_validator/factory.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/factory.js +74 -0
- package/dest/msg_validators/tx_validator/gas_validator.d.ts +11 -0
- package/dest/msg_validators/tx_validator/gas_validator.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/gas_validator.js +115 -0
- package/dest/msg_validators/tx_validator/index.d.ts +8 -1
- package/dest/msg_validators/tx_validator/index.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/index.js +7 -0
- package/dest/msg_validators/tx_validator/metadata_validator.d.ts +9 -5
- package/dest/msg_validators/tx_validator/metadata_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/metadata_validator.js +39 -20
- package/dest/msg_validators/tx_validator/phases_validator.d.ts +14 -0
- package/dest/msg_validators/tx_validator/phases_validator.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/phases_validator.js +93 -0
- package/dest/msg_validators/tx_validator/test_utils.d.ts +17 -0
- package/dest/msg_validators/tx_validator/test_utils.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/test_utils.js +22 -0
- package/dest/msg_validators/tx_validator/timestamp_validator.d.ts +13 -0
- package/dest/msg_validators/tx_validator/timestamp_validator.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/timestamp_validator.js +32 -0
- package/dest/msg_validators/tx_validator/tx_permitted_validator.d.ts +8 -0
- package/dest/msg_validators/tx_validator/tx_permitted_validator.d.ts.map +1 -0
- package/dest/msg_validators/tx_validator/tx_permitted_validator.js +24 -0
- package/dest/msg_validators/tx_validator/tx_proof_validator.d.ts +1 -1
- package/dest/msg_validators/tx_validator/tx_proof_validator.d.ts.map +1 -1
- package/dest/msg_validators/tx_validator/tx_proof_validator.js +6 -5
- package/dest/services/data_store.d.ts +1 -1
- package/dest/services/data_store.d.ts.map +1 -1
- package/dest/services/discv5/discV5_service.d.ts +10 -9
- package/dest/services/discv5/discV5_service.d.ts.map +1 -1
- package/dest/services/discv5/discV5_service.js +63 -36
- package/dest/services/dummy_service.d.ts +54 -11
- package/dest/services/dummy_service.d.ts.map +1 -1
- package/dest/services/dummy_service.js +91 -5
- package/dest/services/encoding.d.ts +26 -7
- package/dest/services/encoding.d.ts.map +1 -1
- package/dest/services/encoding.js +76 -6
- package/dest/services/gossipsub/scoring.d.ts +1 -1
- package/dest/services/index.d.ts +5 -1
- package/dest/services/index.d.ts.map +1 -1
- package/dest/services/index.js +4 -0
- package/dest/services/libp2p/instrumentation.d.ts +20 -0
- package/dest/services/libp2p/instrumentation.d.ts.map +1 -0
- package/dest/services/libp2p/instrumentation.js +111 -0
- package/dest/services/libp2p/libp2p_service.d.ts +102 -96
- package/dest/services/libp2p/libp2p_service.d.ts.map +1 -1
- package/dest/services/libp2p/libp2p_service.js +1307 -301
- package/dest/services/peer-manager/interface.d.ts +23 -0
- package/dest/services/peer-manager/interface.d.ts.map +1 -0
- package/dest/services/peer-manager/interface.js +1 -0
- package/dest/services/peer-manager/metrics.d.ts +11 -2
- package/dest/services/peer-manager/metrics.d.ts.map +1 -1
- package/dest/services/peer-manager/metrics.js +29 -12
- package/dest/services/peer-manager/peer_manager.d.ts +103 -23
- package/dest/services/peer-manager/peer_manager.d.ts.map +1 -1
- package/dest/services/peer-manager/peer_manager.js +551 -82
- package/dest/services/peer-manager/peer_scoring.d.ts +7 -2
- package/dest/services/peer-manager/peer_scoring.d.ts.map +1 -1
- package/dest/services/peer-manager/peer_scoring.js +37 -2
- package/dest/services/reqresp/config.d.ts +11 -9
- package/dest/services/reqresp/config.d.ts.map +1 -1
- package/dest/services/reqresp/config.js +18 -4
- package/dest/services/reqresp/connection-sampler/batch_connection_sampler.d.ts +2 -2
- package/dest/services/reqresp/connection-sampler/batch_connection_sampler.d.ts.map +1 -1
- package/dest/services/reqresp/connection-sampler/batch_connection_sampler.js +10 -6
- package/dest/services/reqresp/connection-sampler/connection_sampler.d.ts +31 -17
- package/dest/services/reqresp/connection-sampler/connection_sampler.d.ts.map +1 -1
- package/dest/services/reqresp/connection-sampler/connection_sampler.js +142 -84
- package/dest/services/reqresp/index.d.ts +3 -2
- package/dest/services/reqresp/index.d.ts.map +1 -1
- package/dest/services/reqresp/index.js +2 -1
- package/dest/services/reqresp/interface.d.ts +73 -24
- package/dest/services/reqresp/interface.d.ts.map +1 -1
- package/dest/services/reqresp/interface.js +46 -27
- package/dest/services/reqresp/metrics.d.ts +1 -1
- package/dest/services/reqresp/metrics.d.ts.map +1 -1
- package/dest/services/reqresp/metrics.js +5 -21
- package/dest/services/reqresp/protocols/auth.d.ts +43 -0
- package/dest/services/reqresp/protocols/auth.d.ts.map +1 -0
- package/dest/services/reqresp/protocols/auth.js +71 -0
- package/dest/services/reqresp/protocols/block.d.ts +6 -1
- package/dest/services/reqresp/protocols/block.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/block.js +30 -6
- package/dest/services/reqresp/protocols/block_txs/bitvector.d.ts +30 -0
- package/dest/services/reqresp/protocols/block_txs/bitvector.d.ts.map +1 -0
- package/dest/services/reqresp/protocols/block_txs/bitvector.js +75 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_handler.d.ts +11 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_handler.d.ts.map +1 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_handler.js +39 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.d.ts +47 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.d.ts.map +1 -0
- package/dest/services/reqresp/protocols/block_txs/block_txs_reqresp.js +75 -0
- package/dest/services/reqresp/protocols/block_txs/index.d.ts +4 -0
- package/dest/services/reqresp/protocols/block_txs/index.d.ts.map +1 -0
- package/dest/services/reqresp/protocols/block_txs/index.js +3 -0
- package/dest/services/reqresp/protocols/goodbye.d.ts +3 -5
- package/dest/services/reqresp/protocols/goodbye.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/goodbye.js +7 -7
- package/dest/services/reqresp/protocols/index.d.ts +3 -1
- package/dest/services/reqresp/protocols/index.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/index.js +2 -0
- package/dest/services/reqresp/protocols/ping.d.ts +1 -3
- package/dest/services/reqresp/protocols/ping.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/status.d.ts +40 -7
- package/dest/services/reqresp/protocols/status.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/status.js +75 -5
- package/dest/services/reqresp/protocols/tx.d.ts +14 -4
- package/dest/services/reqresp/protocols/tx.d.ts.map +1 -1
- package/dest/services/reqresp/protocols/tx.js +34 -6
- package/dest/services/reqresp/rate-limiter/index.d.ts +1 -1
- package/dest/services/reqresp/rate-limiter/rate_limiter.d.ts +6 -4
- package/dest/services/reqresp/rate-limiter/rate_limiter.d.ts.map +1 -1
- package/dest/services/reqresp/rate-limiter/rate_limiter.js +10 -2
- package/dest/services/reqresp/rate-limiter/rate_limits.d.ts +1 -1
- package/dest/services/reqresp/rate-limiter/rate_limits.d.ts.map +1 -1
- package/dest/services/reqresp/rate-limiter/rate_limits.js +21 -1
- package/dest/services/reqresp/reqresp.d.ts +24 -66
- package/dest/services/reqresp/reqresp.d.ts.map +1 -1
- package/dest/services/reqresp/reqresp.js +699 -230
- package/dest/services/reqresp/status.d.ts +10 -4
- package/dest/services/reqresp/status.d.ts.map +1 -1
- package/dest/services/reqresp/status.js +9 -2
- package/dest/services/service.d.ts +37 -20
- package/dest/services/service.d.ts.map +1 -1
- package/dest/services/tx_collection/config.d.ts +25 -0
- package/dest/services/tx_collection/config.d.ts.map +1 -0
- package/dest/services/tx_collection/config.js +58 -0
- package/dest/services/tx_collection/fast_tx_collection.d.ts +51 -0
- package/dest/services/tx_collection/fast_tx_collection.d.ts.map +1 -0
- package/dest/services/tx_collection/fast_tx_collection.js +300 -0
- package/dest/services/tx_collection/index.d.ts +3 -0
- package/dest/services/tx_collection/index.d.ts.map +1 -0
- package/dest/services/tx_collection/index.js +2 -0
- package/dest/services/tx_collection/instrumentation.d.ts +10 -0
- package/dest/services/tx_collection/instrumentation.d.ts.map +1 -0
- package/dest/services/tx_collection/instrumentation.js +24 -0
- package/dest/services/tx_collection/slow_tx_collection.d.ts +53 -0
- package/dest/services/tx_collection/slow_tx_collection.d.ts.map +1 -0
- package/dest/services/tx_collection/slow_tx_collection.js +177 -0
- package/dest/services/tx_collection/tx_collection.d.ts +110 -0
- package/dest/services/tx_collection/tx_collection.d.ts.map +1 -0
- package/dest/services/tx_collection/tx_collection.js +128 -0
- package/dest/services/tx_collection/tx_collection_sink.d.ts +30 -0
- package/dest/services/tx_collection/tx_collection_sink.d.ts.map +1 -0
- package/dest/services/tx_collection/tx_collection_sink.js +111 -0
- package/dest/services/tx_collection/tx_source.d.ts +18 -0
- package/dest/services/tx_collection/tx_source.d.ts.map +1 -0
- package/dest/services/tx_collection/tx_source.js +31 -0
- package/dest/services/tx_provider.d.ts +51 -0
- package/dest/services/tx_provider.d.ts.map +1 -0
- package/dest/services/tx_provider.js +219 -0
- package/dest/services/tx_provider_instrumentation.d.ts +16 -0
- package/dest/services/tx_provider_instrumentation.d.ts.map +1 -0
- package/dest/services/tx_provider_instrumentation.js +34 -0
- package/dest/test-helpers/generate-peer-id-private-keys.d.ts +1 -1
- package/dest/test-helpers/get-ports.d.ts +1 -1
- package/dest/test-helpers/get-ports.d.ts.map +1 -1
- package/dest/test-helpers/index.d.ts +2 -1
- package/dest/test-helpers/index.d.ts.map +1 -1
- package/dest/test-helpers/index.js +1 -0
- package/dest/test-helpers/make-enrs.d.ts +1 -1
- package/dest/test-helpers/make-enrs.d.ts.map +1 -1
- package/dest/test-helpers/make-enrs.js +4 -5
- package/dest/test-helpers/make-test-p2p-clients.d.ts +33 -5
- package/dest/test-helpers/make-test-p2p-clients.d.ts.map +1 -1
- package/dest/test-helpers/make-test-p2p-clients.js +86 -16
- package/dest/test-helpers/mock-pubsub.d.ts +59 -0
- package/dest/test-helpers/mock-pubsub.d.ts.map +1 -0
- package/dest/test-helpers/mock-pubsub.js +130 -0
- package/dest/test-helpers/mock-tx-helpers.d.ts +12 -0
- package/dest/test-helpers/mock-tx-helpers.d.ts.map +1 -0
- package/dest/test-helpers/mock-tx-helpers.js +19 -0
- package/dest/test-helpers/reqresp-nodes.d.ts +15 -11
- package/dest/test-helpers/reqresp-nodes.d.ts.map +1 -1
- package/dest/test-helpers/reqresp-nodes.js +62 -28
- package/dest/testbench/p2p_client_testbench_worker.d.ts +1 -1
- package/dest/testbench/p2p_client_testbench_worker.js +124 -35
- package/dest/testbench/parse_log_file.d.ts +1 -1
- package/dest/testbench/parse_log_file.js +4 -4
- package/dest/testbench/testbench.d.ts +1 -1
- package/dest/testbench/testbench.js +4 -4
- package/dest/testbench/worker_client_manager.d.ts +1 -6
- package/dest/testbench/worker_client_manager.d.ts.map +1 -1
- package/dest/testbench/worker_client_manager.js +17 -20
- package/dest/types/index.d.ts +4 -2
- package/dest/types/index.d.ts.map +1 -1
- package/dest/types/index.js +2 -0
- package/dest/util.d.ts +24 -16
- package/dest/util.d.ts.map +1 -1
- package/dest/util.js +75 -69
- package/dest/versioning.d.ts +4 -4
- package/dest/versioning.d.ts.map +1 -1
- package/dest/versioning.js +8 -3
- package/package.json +32 -27
- package/src/bootstrap/bootstrap.ts +27 -11
- package/src/client/factory.ts +139 -53
- package/src/client/index.ts +1 -0
- package/src/client/interface.ts +213 -0
- package/src/client/p2p_client.ts +471 -385
- package/src/config.ts +299 -134
- package/src/enr/generate-enr.ts +39 -6
- package/src/errors/attestation-pool.error.ts +13 -0
- package/src/index.ts +4 -0
- package/src/mem_pools/attestation_pool/attestation_pool.ts +119 -24
- package/src/mem_pools/attestation_pool/attestation_pool_test_suite.ts +344 -201
- package/src/mem_pools/attestation_pool/kv_attestation_pool.ts +233 -72
- package/src/mem_pools/attestation_pool/memory_attestation_pool.ts +199 -96
- package/src/mem_pools/attestation_pool/mocks.ts +21 -16
- package/src/mem_pools/instrumentation.ts +71 -48
- package/src/mem_pools/interface.ts +2 -4
- package/src/mem_pools/tx_pool/README.md +255 -0
- package/src/mem_pools/tx_pool/aztec_kv_tx_pool.ts +537 -155
- package/src/mem_pools/tx_pool/eviction/eviction_manager.ts +71 -0
- package/src/mem_pools/tx_pool/eviction/eviction_strategy.ts +93 -0
- package/src/mem_pools/tx_pool/eviction/insufficient_fee_payer_balance_rule.ts +108 -0
- package/src/mem_pools/tx_pool/eviction/invalid_txs_after_mining_rule.ts +104 -0
- package/src/mem_pools/tx_pool/eviction/invalid_txs_after_reorg_rule.ts +91 -0
- package/src/mem_pools/tx_pool/eviction/low_priority_eviction_rule.ts +106 -0
- package/src/mem_pools/tx_pool/index.ts +0 -1
- package/src/mem_pools/tx_pool/priority.ts +9 -2
- package/src/mem_pools/tx_pool/tx_pool.ts +75 -10
- package/src/mem_pools/tx_pool/tx_pool_test_suite.ts +225 -36
- package/src/msg_validators/attestation_validator/attestation_validator.ts +60 -14
- package/src/msg_validators/attestation_validator/fisherman_attestation_validator.ts +88 -0
- package/src/msg_validators/attestation_validator/index.ts +1 -0
- package/src/msg_validators/index.ts +1 -1
- package/src/msg_validators/msg_seen_validator/msg_seen_validator.ts +36 -0
- package/src/msg_validators/proposal_validator/block_proposal_validator.ts +10 -0
- package/src/msg_validators/proposal_validator/checkpoint_proposal_validator.ts +13 -0
- package/src/msg_validators/proposal_validator/index.ts +3 -0
- package/src/msg_validators/proposal_validator/proposal_validator.ts +92 -0
- package/src/msg_validators/proposal_validator/proposal_validator_test_suite.ts +206 -0
- package/src/msg_validators/tx_validator/allowed_public_setup.ts +35 -0
- package/src/msg_validators/tx_validator/archive_cache.ts +28 -0
- package/src/msg_validators/tx_validator/block_header_validator.ts +5 -5
- package/src/msg_validators/tx_validator/data_validator.ts +89 -69
- package/src/msg_validators/tx_validator/double_spend_validator.ts +19 -17
- package/src/msg_validators/tx_validator/factory.ts +110 -0
- package/src/msg_validators/tx_validator/gas_validator.ts +134 -0
- package/src/msg_validators/tx_validator/index.ts +7 -0
- package/src/msg_validators/tx_validator/metadata_validator.ts +67 -22
- package/src/msg_validators/tx_validator/phases_validator.ts +116 -0
- package/src/msg_validators/tx_validator/test_utils.ts +43 -0
- package/src/msg_validators/tx_validator/timestamp_validator.ts +49 -0
- package/src/msg_validators/tx_validator/tx_permitted_validator.ts +17 -0
- package/src/msg_validators/tx_validator/tx_proof_validator.ts +6 -5
- package/src/services/discv5/discV5_service.ts +84 -38
- package/src/services/dummy_service.ts +153 -9
- package/src/services/encoding.ts +83 -6
- package/src/services/index.ts +4 -0
- package/src/services/libp2p/instrumentation.ts +113 -0
- package/src/services/libp2p/libp2p_service.ts +1120 -329
- package/src/services/peer-manager/interface.ts +29 -0
- package/src/services/peer-manager/metrics.ts +38 -12
- package/src/services/peer-manager/peer_manager.ts +657 -80
- package/src/services/peer-manager/peer_scoring.ts +42 -3
- package/src/services/reqresp/config.ts +26 -9
- package/src/services/reqresp/connection-sampler/batch_connection_sampler.ts +12 -6
- package/src/services/reqresp/connection-sampler/connection_sampler.ts +150 -95
- package/src/services/reqresp/index.ts +2 -0
- package/src/services/reqresp/interface.ts +92 -37
- package/src/services/reqresp/metrics.ts +11 -24
- package/src/services/reqresp/protocols/auth.ts +83 -0
- package/src/services/reqresp/protocols/block.ts +26 -4
- package/src/services/reqresp/protocols/block_txs/bitvector.ts +90 -0
- package/src/services/reqresp/protocols/block_txs/block_txs_handler.ts +53 -0
- package/src/services/reqresp/protocols/block_txs/block_txs_reqresp.ts +79 -0
- package/src/services/reqresp/protocols/block_txs/index.ts +3 -0
- package/src/services/reqresp/protocols/goodbye.ts +9 -7
- package/src/services/reqresp/protocols/index.ts +2 -0
- package/src/services/reqresp/protocols/status.ts +119 -5
- package/src/services/reqresp/protocols/tx.ts +36 -8
- package/src/services/reqresp/rate-limiter/rate_limiter.ts +12 -3
- package/src/services/reqresp/rate-limiter/rate_limits.ts +21 -1
- package/src/services/reqresp/reqresp.ts +387 -256
- package/src/services/reqresp/status.ts +12 -3
- package/src/services/service.ts +61 -22
- package/src/services/tx_collection/config.ts +84 -0
- package/src/services/tx_collection/fast_tx_collection.ts +341 -0
- package/src/services/tx_collection/index.ts +2 -0
- package/src/services/tx_collection/instrumentation.ts +26 -0
- package/src/services/tx_collection/slow_tx_collection.ts +233 -0
- package/src/services/tx_collection/tx_collection.ts +216 -0
- package/src/services/tx_collection/tx_collection_sink.ts +129 -0
- package/src/services/tx_collection/tx_source.ts +37 -0
- package/src/services/tx_provider.ts +232 -0
- package/src/services/tx_provider_instrumentation.ts +48 -0
- package/src/test-helpers/index.ts +1 -0
- package/src/test-helpers/make-enrs.ts +4 -5
- package/src/test-helpers/make-test-p2p-clients.ts +111 -21
- package/src/test-helpers/mock-pubsub.ts +188 -0
- package/src/test-helpers/mock-tx-helpers.ts +24 -0
- package/src/test-helpers/reqresp-nodes.ts +87 -36
- package/src/testbench/p2p_client_testbench_worker.ts +182 -32
- package/src/testbench/parse_log_file.ts +4 -4
- package/src/testbench/testbench.ts +4 -4
- package/src/testbench/worker_client_manager.ts +23 -24
- package/src/types/index.ts +2 -0
- package/src/util.ts +105 -91
- package/src/versioning.ts +11 -4
- package/dest/mem_pools/tx_pool/memory_tx_pool.d.ts +0 -56
- package/dest/mem_pools/tx_pool/memory_tx_pool.d.ts.map +0 -1
- package/dest/mem_pools/tx_pool/memory_tx_pool.js +0 -141
- package/dest/msg_validators/block_proposal_validator/block_proposal_validator.d.ts +0 -8
- package/dest/msg_validators/block_proposal_validator/block_proposal_validator.d.ts.map +0 -1
- package/dest/msg_validators/block_proposal_validator/block_proposal_validator.js +0 -21
- package/dest/msg_validators/block_proposal_validator/index.d.ts +0 -2
- package/dest/msg_validators/block_proposal_validator/index.d.ts.map +0 -1
- package/dest/msg_validators/block_proposal_validator/index.js +0 -1
- package/src/mem_pools/tx_pool/memory_tx_pool.ts +0 -174
- package/src/msg_validators/block_proposal_validator/block_proposal_validator.ts +0 -29
- package/src/msg_validators/block_proposal_validator/index.ts +0 -1
|
@@ -1,23 +1,403 @@
|
|
|
1
1
|
// @attribution: lodestar impl for inspiration
|
|
2
|
-
function
|
|
3
|
-
|
|
4
|
-
|
|
5
|
-
|
|
6
|
-
|
|
2
|
+
function applyDecs2203RFactory() {
|
|
3
|
+
function createAddInitializerMethod(initializers, decoratorFinishedRef) {
|
|
4
|
+
return function addInitializer(initializer) {
|
|
5
|
+
assertNotFinished(decoratorFinishedRef, "addInitializer");
|
|
6
|
+
assertCallable(initializer, "An initializer");
|
|
7
|
+
initializers.push(initializer);
|
|
8
|
+
};
|
|
9
|
+
}
|
|
10
|
+
function memberDec(dec, name, desc, initializers, kind, isStatic, isPrivate, metadata, value) {
|
|
11
|
+
var kindStr;
|
|
12
|
+
switch(kind){
|
|
13
|
+
case 1:
|
|
14
|
+
kindStr = "accessor";
|
|
15
|
+
break;
|
|
16
|
+
case 2:
|
|
17
|
+
kindStr = "method";
|
|
18
|
+
break;
|
|
19
|
+
case 3:
|
|
20
|
+
kindStr = "getter";
|
|
21
|
+
break;
|
|
22
|
+
case 4:
|
|
23
|
+
kindStr = "setter";
|
|
24
|
+
break;
|
|
25
|
+
default:
|
|
26
|
+
kindStr = "field";
|
|
27
|
+
}
|
|
28
|
+
var ctx = {
|
|
29
|
+
kind: kindStr,
|
|
30
|
+
name: isPrivate ? "#" + name : name,
|
|
31
|
+
static: isStatic,
|
|
32
|
+
private: isPrivate,
|
|
33
|
+
metadata: metadata
|
|
34
|
+
};
|
|
35
|
+
var decoratorFinishedRef = {
|
|
36
|
+
v: false
|
|
37
|
+
};
|
|
38
|
+
ctx.addInitializer = createAddInitializerMethod(initializers, decoratorFinishedRef);
|
|
39
|
+
var get, set;
|
|
40
|
+
if (kind === 0) {
|
|
41
|
+
if (isPrivate) {
|
|
42
|
+
get = desc.get;
|
|
43
|
+
set = desc.set;
|
|
44
|
+
} else {
|
|
45
|
+
get = function() {
|
|
46
|
+
return this[name];
|
|
47
|
+
};
|
|
48
|
+
set = function(v) {
|
|
49
|
+
this[name] = v;
|
|
50
|
+
};
|
|
51
|
+
}
|
|
52
|
+
} else if (kind === 2) {
|
|
53
|
+
get = function() {
|
|
54
|
+
return desc.value;
|
|
55
|
+
};
|
|
56
|
+
} else {
|
|
57
|
+
if (kind === 1 || kind === 3) {
|
|
58
|
+
get = function() {
|
|
59
|
+
return desc.get.call(this);
|
|
60
|
+
};
|
|
61
|
+
}
|
|
62
|
+
if (kind === 1 || kind === 4) {
|
|
63
|
+
set = function(v) {
|
|
64
|
+
desc.set.call(this, v);
|
|
65
|
+
};
|
|
66
|
+
}
|
|
67
|
+
}
|
|
68
|
+
ctx.access = get && set ? {
|
|
69
|
+
get: get,
|
|
70
|
+
set: set
|
|
71
|
+
} : get ? {
|
|
72
|
+
get: get
|
|
73
|
+
} : {
|
|
74
|
+
set: set
|
|
75
|
+
};
|
|
76
|
+
try {
|
|
77
|
+
return dec(value, ctx);
|
|
78
|
+
} finally{
|
|
79
|
+
decoratorFinishedRef.v = true;
|
|
80
|
+
}
|
|
81
|
+
}
|
|
82
|
+
function assertNotFinished(decoratorFinishedRef, fnName) {
|
|
83
|
+
if (decoratorFinishedRef.v) {
|
|
84
|
+
throw new Error("attempted to call " + fnName + " after decoration was finished");
|
|
85
|
+
}
|
|
86
|
+
}
|
|
87
|
+
function assertCallable(fn, hint) {
|
|
88
|
+
if (typeof fn !== "function") {
|
|
89
|
+
throw new TypeError(hint + " must be a function");
|
|
90
|
+
}
|
|
91
|
+
}
|
|
92
|
+
function assertValidReturnValue(kind, value) {
|
|
93
|
+
var type = typeof value;
|
|
94
|
+
if (kind === 1) {
|
|
95
|
+
if (type !== "object" || value === null) {
|
|
96
|
+
throw new TypeError("accessor decorators must return an object with get, set, or init properties or void 0");
|
|
97
|
+
}
|
|
98
|
+
if (value.get !== undefined) {
|
|
99
|
+
assertCallable(value.get, "accessor.get");
|
|
100
|
+
}
|
|
101
|
+
if (value.set !== undefined) {
|
|
102
|
+
assertCallable(value.set, "accessor.set");
|
|
103
|
+
}
|
|
104
|
+
if (value.init !== undefined) {
|
|
105
|
+
assertCallable(value.init, "accessor.init");
|
|
106
|
+
}
|
|
107
|
+
} else if (type !== "function") {
|
|
108
|
+
var hint;
|
|
109
|
+
if (kind === 0) {
|
|
110
|
+
hint = "field";
|
|
111
|
+
} else if (kind === 10) {
|
|
112
|
+
hint = "class";
|
|
113
|
+
} else {
|
|
114
|
+
hint = "method";
|
|
115
|
+
}
|
|
116
|
+
throw new TypeError(hint + " decorators must return a function or void 0");
|
|
117
|
+
}
|
|
118
|
+
}
|
|
119
|
+
function applyMemberDec(ret, base, decInfo, name, kind, isStatic, isPrivate, initializers, metadata) {
|
|
120
|
+
var decs = decInfo[0];
|
|
121
|
+
var desc, init, value;
|
|
122
|
+
if (isPrivate) {
|
|
123
|
+
if (kind === 0 || kind === 1) {
|
|
124
|
+
desc = {
|
|
125
|
+
get: decInfo[3],
|
|
126
|
+
set: decInfo[4]
|
|
127
|
+
};
|
|
128
|
+
} else if (kind === 3) {
|
|
129
|
+
desc = {
|
|
130
|
+
get: decInfo[3]
|
|
131
|
+
};
|
|
132
|
+
} else if (kind === 4) {
|
|
133
|
+
desc = {
|
|
134
|
+
set: decInfo[3]
|
|
135
|
+
};
|
|
136
|
+
} else {
|
|
137
|
+
desc = {
|
|
138
|
+
value: decInfo[3]
|
|
139
|
+
};
|
|
140
|
+
}
|
|
141
|
+
} else if (kind !== 0) {
|
|
142
|
+
desc = Object.getOwnPropertyDescriptor(base, name);
|
|
143
|
+
}
|
|
144
|
+
if (kind === 1) {
|
|
145
|
+
value = {
|
|
146
|
+
get: desc.get,
|
|
147
|
+
set: desc.set
|
|
148
|
+
};
|
|
149
|
+
} else if (kind === 2) {
|
|
150
|
+
value = desc.value;
|
|
151
|
+
} else if (kind === 3) {
|
|
152
|
+
value = desc.get;
|
|
153
|
+
} else if (kind === 4) {
|
|
154
|
+
value = desc.set;
|
|
155
|
+
}
|
|
156
|
+
var newValue, get, set;
|
|
157
|
+
if (typeof decs === "function") {
|
|
158
|
+
newValue = memberDec(decs, name, desc, initializers, kind, isStatic, isPrivate, metadata, value);
|
|
159
|
+
if (newValue !== void 0) {
|
|
160
|
+
assertValidReturnValue(kind, newValue);
|
|
161
|
+
if (kind === 0) {
|
|
162
|
+
init = newValue;
|
|
163
|
+
} else if (kind === 1) {
|
|
164
|
+
init = newValue.init;
|
|
165
|
+
get = newValue.get || value.get;
|
|
166
|
+
set = newValue.set || value.set;
|
|
167
|
+
value = {
|
|
168
|
+
get: get,
|
|
169
|
+
set: set
|
|
170
|
+
};
|
|
171
|
+
} else {
|
|
172
|
+
value = newValue;
|
|
173
|
+
}
|
|
174
|
+
}
|
|
175
|
+
} else {
|
|
176
|
+
for(var i = decs.length - 1; i >= 0; i--){
|
|
177
|
+
var dec = decs[i];
|
|
178
|
+
newValue = memberDec(dec, name, desc, initializers, kind, isStatic, isPrivate, metadata, value);
|
|
179
|
+
if (newValue !== void 0) {
|
|
180
|
+
assertValidReturnValue(kind, newValue);
|
|
181
|
+
var newInit;
|
|
182
|
+
if (kind === 0) {
|
|
183
|
+
newInit = newValue;
|
|
184
|
+
} else if (kind === 1) {
|
|
185
|
+
newInit = newValue.init;
|
|
186
|
+
get = newValue.get || value.get;
|
|
187
|
+
set = newValue.set || value.set;
|
|
188
|
+
value = {
|
|
189
|
+
get: get,
|
|
190
|
+
set: set
|
|
191
|
+
};
|
|
192
|
+
} else {
|
|
193
|
+
value = newValue;
|
|
194
|
+
}
|
|
195
|
+
if (newInit !== void 0) {
|
|
196
|
+
if (init === void 0) {
|
|
197
|
+
init = newInit;
|
|
198
|
+
} else if (typeof init === "function") {
|
|
199
|
+
init = [
|
|
200
|
+
init,
|
|
201
|
+
newInit
|
|
202
|
+
];
|
|
203
|
+
} else {
|
|
204
|
+
init.push(newInit);
|
|
205
|
+
}
|
|
206
|
+
}
|
|
207
|
+
}
|
|
208
|
+
}
|
|
209
|
+
}
|
|
210
|
+
if (kind === 0 || kind === 1) {
|
|
211
|
+
if (init === void 0) {
|
|
212
|
+
init = function(instance, init) {
|
|
213
|
+
return init;
|
|
214
|
+
};
|
|
215
|
+
} else if (typeof init !== "function") {
|
|
216
|
+
var ownInitializers = init;
|
|
217
|
+
init = function(instance, init) {
|
|
218
|
+
var value = init;
|
|
219
|
+
for(var i = 0; i < ownInitializers.length; i++){
|
|
220
|
+
value = ownInitializers[i].call(instance, value);
|
|
221
|
+
}
|
|
222
|
+
return value;
|
|
223
|
+
};
|
|
224
|
+
} else {
|
|
225
|
+
var originalInitializer = init;
|
|
226
|
+
init = function(instance, init) {
|
|
227
|
+
return originalInitializer.call(instance, init);
|
|
228
|
+
};
|
|
229
|
+
}
|
|
230
|
+
ret.push(init);
|
|
231
|
+
}
|
|
232
|
+
if (kind !== 0) {
|
|
233
|
+
if (kind === 1) {
|
|
234
|
+
desc.get = value.get;
|
|
235
|
+
desc.set = value.set;
|
|
236
|
+
} else if (kind === 2) {
|
|
237
|
+
desc.value = value;
|
|
238
|
+
} else if (kind === 3) {
|
|
239
|
+
desc.get = value;
|
|
240
|
+
} else if (kind === 4) {
|
|
241
|
+
desc.set = value;
|
|
242
|
+
}
|
|
243
|
+
if (isPrivate) {
|
|
244
|
+
if (kind === 1) {
|
|
245
|
+
ret.push(function(instance, args) {
|
|
246
|
+
return value.get.call(instance, args);
|
|
247
|
+
});
|
|
248
|
+
ret.push(function(instance, args) {
|
|
249
|
+
return value.set.call(instance, args);
|
|
250
|
+
});
|
|
251
|
+
} else if (kind === 2) {
|
|
252
|
+
ret.push(value);
|
|
253
|
+
} else {
|
|
254
|
+
ret.push(function(instance, args) {
|
|
255
|
+
return value.call(instance, args);
|
|
256
|
+
});
|
|
257
|
+
}
|
|
258
|
+
} else {
|
|
259
|
+
Object.defineProperty(base, name, desc);
|
|
260
|
+
}
|
|
261
|
+
}
|
|
262
|
+
}
|
|
263
|
+
function applyMemberDecs(Class, decInfos, metadata) {
|
|
264
|
+
var ret = [];
|
|
265
|
+
var protoInitializers;
|
|
266
|
+
var staticInitializers;
|
|
267
|
+
var existingProtoNonFields = new Map();
|
|
268
|
+
var existingStaticNonFields = new Map();
|
|
269
|
+
for(var i = 0; i < decInfos.length; i++){
|
|
270
|
+
var decInfo = decInfos[i];
|
|
271
|
+
if (!Array.isArray(decInfo)) continue;
|
|
272
|
+
var kind = decInfo[1];
|
|
273
|
+
var name = decInfo[2];
|
|
274
|
+
var isPrivate = decInfo.length > 3;
|
|
275
|
+
var isStatic = kind >= 5;
|
|
276
|
+
var base;
|
|
277
|
+
var initializers;
|
|
278
|
+
if (isStatic) {
|
|
279
|
+
base = Class;
|
|
280
|
+
kind = kind - 5;
|
|
281
|
+
staticInitializers = staticInitializers || [];
|
|
282
|
+
initializers = staticInitializers;
|
|
283
|
+
} else {
|
|
284
|
+
base = Class.prototype;
|
|
285
|
+
protoInitializers = protoInitializers || [];
|
|
286
|
+
initializers = protoInitializers;
|
|
287
|
+
}
|
|
288
|
+
if (kind !== 0 && !isPrivate) {
|
|
289
|
+
var existingNonFields = isStatic ? existingStaticNonFields : existingProtoNonFields;
|
|
290
|
+
var existingKind = existingNonFields.get(name) || 0;
|
|
291
|
+
if (existingKind === true || existingKind === 3 && kind !== 4 || existingKind === 4 && kind !== 3) {
|
|
292
|
+
throw new Error("Attempted to decorate a public method/accessor that has the same name as a previously decorated public method/accessor. This is not currently supported by the decorators plugin. Property name was: " + name);
|
|
293
|
+
} else if (!existingKind && kind > 2) {
|
|
294
|
+
existingNonFields.set(name, kind);
|
|
295
|
+
} else {
|
|
296
|
+
existingNonFields.set(name, true);
|
|
297
|
+
}
|
|
298
|
+
}
|
|
299
|
+
applyMemberDec(ret, base, decInfo, name, kind, isStatic, isPrivate, initializers, metadata);
|
|
300
|
+
}
|
|
301
|
+
pushInitializers(ret, protoInitializers);
|
|
302
|
+
pushInitializers(ret, staticInitializers);
|
|
303
|
+
return ret;
|
|
304
|
+
}
|
|
305
|
+
function pushInitializers(ret, initializers) {
|
|
306
|
+
if (initializers) {
|
|
307
|
+
ret.push(function(instance) {
|
|
308
|
+
for(var i = 0; i < initializers.length; i++){
|
|
309
|
+
initializers[i].call(instance);
|
|
310
|
+
}
|
|
311
|
+
return instance;
|
|
312
|
+
});
|
|
313
|
+
}
|
|
314
|
+
}
|
|
315
|
+
function applyClassDecs(targetClass, classDecs, metadata) {
|
|
316
|
+
if (classDecs.length > 0) {
|
|
317
|
+
var initializers = [];
|
|
318
|
+
var newClass = targetClass;
|
|
319
|
+
var name = targetClass.name;
|
|
320
|
+
for(var i = classDecs.length - 1; i >= 0; i--){
|
|
321
|
+
var decoratorFinishedRef = {
|
|
322
|
+
v: false
|
|
323
|
+
};
|
|
324
|
+
try {
|
|
325
|
+
var nextNewClass = classDecs[i](newClass, {
|
|
326
|
+
kind: "class",
|
|
327
|
+
name: name,
|
|
328
|
+
addInitializer: createAddInitializerMethod(initializers, decoratorFinishedRef),
|
|
329
|
+
metadata
|
|
330
|
+
});
|
|
331
|
+
} finally{
|
|
332
|
+
decoratorFinishedRef.v = true;
|
|
333
|
+
}
|
|
334
|
+
if (nextNewClass !== undefined) {
|
|
335
|
+
assertValidReturnValue(10, nextNewClass);
|
|
336
|
+
newClass = nextNewClass;
|
|
337
|
+
}
|
|
338
|
+
}
|
|
339
|
+
return [
|
|
340
|
+
defineMetadata(newClass, metadata),
|
|
341
|
+
function() {
|
|
342
|
+
for(var i = 0; i < initializers.length; i++){
|
|
343
|
+
initializers[i].call(newClass);
|
|
344
|
+
}
|
|
345
|
+
}
|
|
346
|
+
];
|
|
347
|
+
}
|
|
348
|
+
}
|
|
349
|
+
function defineMetadata(Class, metadata) {
|
|
350
|
+
return Object.defineProperty(Class, Symbol.metadata || Symbol.for("Symbol.metadata"), {
|
|
351
|
+
configurable: true,
|
|
352
|
+
enumerable: true,
|
|
353
|
+
value: metadata
|
|
354
|
+
});
|
|
355
|
+
}
|
|
356
|
+
return function applyDecs2203R(targetClass, memberDecs, classDecs, parentClass) {
|
|
357
|
+
if (parentClass !== void 0) {
|
|
358
|
+
var parentMetadata = parentClass[Symbol.metadata || Symbol.for("Symbol.metadata")];
|
|
359
|
+
}
|
|
360
|
+
var metadata = Object.create(parentMetadata === void 0 ? null : parentMetadata);
|
|
361
|
+
var e = applyMemberDecs(targetClass, memberDecs, metadata);
|
|
362
|
+
if (!classDecs.length) defineMetadata(targetClass, metadata);
|
|
363
|
+
return {
|
|
364
|
+
e: e,
|
|
365
|
+
get c () {
|
|
366
|
+
return applyClassDecs(targetClass, classDecs, metadata);
|
|
367
|
+
}
|
|
368
|
+
};
|
|
369
|
+
};
|
|
370
|
+
}
|
|
371
|
+
function _apply_decs_2203_r(targetClass, memberDecs, classDecs, parentClass) {
|
|
372
|
+
return (_apply_decs_2203_r = applyDecs2203RFactory())(targetClass, memberDecs, classDecs, parentClass);
|
|
7
373
|
}
|
|
374
|
+
var _dec, _dec1, _dec2, _initProto;
|
|
375
|
+
import { compactArray } from '@aztec/foundation/collection';
|
|
376
|
+
import { AbortError, TimeoutError } from '@aztec/foundation/error';
|
|
8
377
|
import { createLogger } from '@aztec/foundation/log';
|
|
9
378
|
import { executeTimeout } from '@aztec/foundation/timer';
|
|
10
379
|
import { PeerErrorSeverity } from '@aztec/stdlib/p2p';
|
|
11
380
|
import { Attributes, getTelemetryClient, trackSpan } from '@aztec/telemetry-client';
|
|
12
|
-
import {
|
|
381
|
+
import { pipeline } from 'node:stream/promises';
|
|
13
382
|
import { CollectiveReqRespTimeoutError, IndividualReqRespTimeoutError, InvalidResponseError } from '../../errors/reqresp.error.js';
|
|
14
383
|
import { SnappyTransform } from '../encoding.js';
|
|
384
|
+
import { DEFAULT_INDIVIDUAL_REQUEST_TIMEOUT_MS, DEFAULT_REQRESP_DIAL_TIMEOUT_MS } from './config.js';
|
|
15
385
|
import { BatchConnectionSampler } from './connection-sampler/batch_connection_sampler.js';
|
|
16
|
-
import { ConnectionSampler } from './connection-sampler/connection_sampler.js';
|
|
17
|
-
import {
|
|
386
|
+
import { ConnectionSampler, RandomSampler } from './connection-sampler/connection_sampler.js';
|
|
387
|
+
import { DEFAULT_SUB_PROTOCOL_VALIDATORS, ReqRespSubProtocol, responseFromBuffer } from './interface.js';
|
|
18
388
|
import { ReqRespMetrics } from './metrics.js';
|
|
19
389
|
import { RateLimitStatus, RequestResponseRateLimiter, prettyPrintRateLimitStatus } from './rate-limiter/rate_limiter.js';
|
|
20
390
|
import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqRespStatus } from './status.js';
|
|
391
|
+
_dec = trackSpan('ReqResp.sendBatchRequest', (subProtocol, requests)=>({
|
|
392
|
+
[Attributes.P2P_REQ_RESP_PROTOCOL]: subProtocol,
|
|
393
|
+
[Attributes.P2P_REQ_RESP_BATCH_REQUESTS_COUNT]: requests.length
|
|
394
|
+
})), _dec1 = trackSpan('ReqResp.sendRequestToPeer', (peerId, subProtocol, _)=>({
|
|
395
|
+
[Attributes.P2P_ID]: peerId.toString(),
|
|
396
|
+
[Attributes.P2P_REQ_RESP_PROTOCOL]: subProtocol
|
|
397
|
+
})), _dec2 = trackSpan('ReqResp.streamHandler', (protocol, { connection })=>({
|
|
398
|
+
[Attributes.P2P_REQ_RESP_PROTOCOL]: protocol,
|
|
399
|
+
[Attributes.P2P_ID]: connection.remotePeer.toString()
|
|
400
|
+
}));
|
|
21
401
|
/**
|
|
22
402
|
* The Request Response Service
|
|
23
403
|
*
|
|
@@ -35,43 +415,77 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
35
415
|
libp2p;
|
|
36
416
|
peerScoring;
|
|
37
417
|
logger;
|
|
38
|
-
|
|
418
|
+
static{
|
|
419
|
+
({ e: [_initProto] } = _apply_decs_2203_r(this, [
|
|
420
|
+
[
|
|
421
|
+
_dec,
|
|
422
|
+
2,
|
|
423
|
+
"sendBatchRequest"
|
|
424
|
+
],
|
|
425
|
+
[
|
|
426
|
+
_dec1,
|
|
427
|
+
2,
|
|
428
|
+
"sendRequestToPeer"
|
|
429
|
+
],
|
|
430
|
+
[
|
|
431
|
+
_dec2,
|
|
432
|
+
2,
|
|
433
|
+
"streamHandler"
|
|
434
|
+
]
|
|
435
|
+
], []));
|
|
436
|
+
}
|
|
39
437
|
individualRequestTimeoutMs;
|
|
40
|
-
|
|
438
|
+
dialTimeoutMs;
|
|
41
439
|
subProtocolHandlers;
|
|
42
440
|
subProtocolValidators;
|
|
43
441
|
connectionSampler;
|
|
44
442
|
rateLimiter;
|
|
45
443
|
snappyTransform;
|
|
46
444
|
metrics;
|
|
47
|
-
constructor(config, libp2p, peerScoring, telemetryClient = getTelemetryClient()){
|
|
445
|
+
constructor(config, libp2p, peerScoring, logger = createLogger('p2p:reqresp'), rateLimits = {}, telemetryClient = getTelemetryClient()){
|
|
48
446
|
this.libp2p = libp2p;
|
|
49
447
|
this.peerScoring = peerScoring;
|
|
50
|
-
this.
|
|
51
|
-
this.
|
|
52
|
-
this.
|
|
53
|
-
this.
|
|
54
|
-
this.
|
|
55
|
-
this.
|
|
448
|
+
this.logger = logger;
|
|
449
|
+
this.individualRequestTimeoutMs = (_initProto(this), DEFAULT_INDIVIDUAL_REQUEST_TIMEOUT_MS);
|
|
450
|
+
this.dialTimeoutMs = DEFAULT_REQRESP_DIAL_TIMEOUT_MS;
|
|
451
|
+
this.subProtocolHandlers = {};
|
|
452
|
+
this.subProtocolValidators = {};
|
|
453
|
+
this.updateConfig(config);
|
|
454
|
+
this.rateLimiter = new RequestResponseRateLimiter(peerScoring, rateLimits);
|
|
56
455
|
// Connection sampler is used to sample our connected peers
|
|
57
|
-
this.connectionSampler = new ConnectionSampler(libp2p);
|
|
456
|
+
this.connectionSampler = new ConnectionSampler(libp2p, new RandomSampler(), createLogger(`${logger.module}:connection-sampler`), config);
|
|
58
457
|
this.snappyTransform = new SnappyTransform();
|
|
59
458
|
this.metrics = new ReqRespMetrics(telemetryClient);
|
|
60
459
|
}
|
|
460
|
+
updateConfig(config) {
|
|
461
|
+
if (typeof config.individualRequestTimeoutMs === 'number') {
|
|
462
|
+
this.individualRequestTimeoutMs = config.individualRequestTimeoutMs;
|
|
463
|
+
}
|
|
464
|
+
if (typeof config.dialTimeoutMs === 'number') {
|
|
465
|
+
this.dialTimeoutMs = config.dialTimeoutMs;
|
|
466
|
+
}
|
|
467
|
+
}
|
|
61
468
|
get tracer() {
|
|
62
469
|
return this.metrics.tracer;
|
|
63
470
|
}
|
|
64
471
|
/**
|
|
65
472
|
* Start the reqresp service
|
|
66
473
|
*/ async start(subProtocolHandlers, subProtocolValidators) {
|
|
67
|
-
this.subProtocolHandlers
|
|
68
|
-
this.subProtocolValidators
|
|
474
|
+
Object.assign(this.subProtocolHandlers, subProtocolHandlers);
|
|
475
|
+
Object.assign(this.subProtocolValidators, subProtocolValidators);
|
|
69
476
|
// Register all protocol handlers
|
|
70
|
-
for (const subProtocol of Object.keys(
|
|
477
|
+
for (const subProtocol of Object.keys(subProtocolHandlers)){
|
|
478
|
+
this.logger.debug(`Registering handler for sub protocol ${subProtocol}`);
|
|
71
479
|
await this.libp2p.handle(subProtocol, (data)=>void this.streamHandler(subProtocol, data).catch((err)=>this.logger.error(`Error on libp2p subprotocol ${subProtocol} handler`, err)));
|
|
72
480
|
}
|
|
73
481
|
this.rateLimiter.start();
|
|
74
482
|
}
|
|
483
|
+
async addSubProtocol(subProtocol, handler, validator = DEFAULT_SUB_PROTOCOL_VALIDATORS[subProtocol]) {
|
|
484
|
+
this.subProtocolHandlers[subProtocol] = handler;
|
|
485
|
+
this.subProtocolValidators[subProtocol] = validator;
|
|
486
|
+
this.logger.debug(`Registering handler for sub protocol ${subProtocol}`);
|
|
487
|
+
await this.libp2p.handle(subProtocol, (data)=>void this.streamHandler(subProtocol, data).catch((err)=>this.logger.error(`Error on libp2p subprotocol ${subProtocol} handler`, err)));
|
|
488
|
+
}
|
|
75
489
|
/**
|
|
76
490
|
* Stop the reqresp service
|
|
77
491
|
*/ async stop() {
|
|
@@ -90,81 +504,6 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
90
504
|
// NOTE: We assume libp2p instance is managed by the caller
|
|
91
505
|
}
|
|
92
506
|
/**
|
|
93
|
-
* Send a request to peers, returns the first response
|
|
94
|
-
*
|
|
95
|
-
* @param subProtocol - The protocol being requested
|
|
96
|
-
* @param request - The request to send
|
|
97
|
-
* @returns - The response from the peer, otherwise undefined
|
|
98
|
-
*
|
|
99
|
-
* @description
|
|
100
|
-
* This method attempts to send a request to all active peers using the specified sub-protocol.
|
|
101
|
-
* It opens a stream with each peer, sends the request, and awaits a response.
|
|
102
|
-
* If a valid response is received, it returns the response; otherwise, it continues to the next peer.
|
|
103
|
-
* If no response is received from any peer, it returns undefined.
|
|
104
|
-
*
|
|
105
|
-
* The method performs the following steps:
|
|
106
|
-
* - Sample a peer to send the request to.
|
|
107
|
-
* - Opens a stream with the peer using the specified sub-protocol.
|
|
108
|
-
*
|
|
109
|
-
* When a response is received, it is validated using the given sub protocols response validator.
|
|
110
|
-
* To see the interface for the response validator - see `interface.ts`
|
|
111
|
-
*
|
|
112
|
-
* Failing a response validation requests in a severe peer penalty, and will
|
|
113
|
-
* prompt the node to continue to search to the next peer.
|
|
114
|
-
* For example, a transaction request validator will check that the payload returned does in fact
|
|
115
|
-
* match the txHash that was requested. A peer that fails this check an only be an extremely naughty peer.
|
|
116
|
-
*
|
|
117
|
-
* This entire operation is wrapped in an overall timeout, that is independent of the
|
|
118
|
-
* peer it is requesting data from.
|
|
119
|
-
*
|
|
120
|
-
*/ async sendRequest(subProtocol, request) {
|
|
121
|
-
const responseValidator = this.subProtocolValidators[subProtocol];
|
|
122
|
-
const requestBuffer = request.toBuffer();
|
|
123
|
-
const requestFunction = async ()=>{
|
|
124
|
-
// Attempt to ask all of our peers, but sampled in a random order
|
|
125
|
-
// This function is wrapped in a timeout, so we will exit the loop if we have not received a response
|
|
126
|
-
const numberOfPeers = this.libp2p.getPeers().length;
|
|
127
|
-
if (numberOfPeers === 0) {
|
|
128
|
-
this.logger.debug('No active peers to send requests to');
|
|
129
|
-
return undefined;
|
|
130
|
-
}
|
|
131
|
-
const attemptedPeers = new Map();
|
|
132
|
-
for(let i = 0; i < numberOfPeers; i++){
|
|
133
|
-
// Sample a peer to make a request to
|
|
134
|
-
const peer = this.connectionSampler.getPeer(attemptedPeers);
|
|
135
|
-
this.logger.trace(`Attempting to send request to peer: ${peer?.toString()}`);
|
|
136
|
-
if (!peer) {
|
|
137
|
-
this.logger.debug('No peers available to send requests to');
|
|
138
|
-
return undefined;
|
|
139
|
-
}
|
|
140
|
-
attemptedPeers.set(peer.toString(), true);
|
|
141
|
-
this.logger.trace(`Sending request to peer: ${peer.toString()}`);
|
|
142
|
-
const response = await this.sendRequestToPeer(peer, subProtocol, requestBuffer);
|
|
143
|
-
if (response && response.status !== ReqRespStatus.SUCCESS) {
|
|
144
|
-
this.logger.debug(`Request to peer ${peer.toString()} failed with status ${prettyPrintReqRespStatus(response.status)}`);
|
|
145
|
-
continue;
|
|
146
|
-
}
|
|
147
|
-
// If we get a response, return it, otherwise we iterate onto the next peer
|
|
148
|
-
// We do not consider it a success if we have an empty buffer
|
|
149
|
-
if (response && response.data.length > 0) {
|
|
150
|
-
const object = subProtocolMap[subProtocol].response.fromBuffer(response.data);
|
|
151
|
-
// The response validator handles peer punishment within
|
|
152
|
-
const isValid = await responseValidator(request, object, peer);
|
|
153
|
-
if (!isValid) {
|
|
154
|
-
throw new InvalidResponseError();
|
|
155
|
-
}
|
|
156
|
-
return object;
|
|
157
|
-
}
|
|
158
|
-
}
|
|
159
|
-
};
|
|
160
|
-
try {
|
|
161
|
-
return await executeTimeout(requestFunction, this.overallRequestTimeoutMs, ()=>new CollectiveReqRespTimeoutError());
|
|
162
|
-
} catch (e) {
|
|
163
|
-
this.logger.debug(`${e.message} | subProtocol: ${subProtocol}`);
|
|
164
|
-
return undefined;
|
|
165
|
-
}
|
|
166
|
-
}
|
|
167
|
-
/**
|
|
168
507
|
* Request multiple messages over the same sub protocol, balancing the requests across peers.
|
|
169
508
|
*
|
|
170
509
|
* @devnote
|
|
@@ -185,17 +524,19 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
185
524
|
* @returns
|
|
186
525
|
*
|
|
187
526
|
* @throws {CollectiveReqRespTimeoutError} - If the request batch exceeds the specified timeout (`timeoutMs`).
|
|
188
|
-
*/ async sendBatchRequest(subProtocol, requests, timeoutMs = 10000, maxPeers = Math.
|
|
189
|
-
const responseValidator = this.subProtocolValidators[subProtocol];
|
|
527
|
+
*/ async sendBatchRequest(subProtocol, requests, pinnedPeer, timeoutMs = 10000, maxPeers = Math.max(10, Math.ceil(requests.length / 3)), maxRetryAttempts = 3) {
|
|
528
|
+
const responseValidator = this.subProtocolValidators[subProtocol] ?? DEFAULT_SUB_PROTOCOL_VALIDATORS[subProtocol];
|
|
190
529
|
const responses = new Array(requests.length);
|
|
191
530
|
const requestBuffers = requests.map((req)=>req.toBuffer());
|
|
192
|
-
const requestFunction = async ()=>{
|
|
531
|
+
const requestFunction = async (signal)=>{
|
|
193
532
|
// Track which requests still need to be processed
|
|
194
533
|
const pendingRequestIndices = new Set(requestBuffers.map((_, i)=>i));
|
|
195
534
|
// Create batch sampler with the total number of requests and max peers
|
|
196
|
-
const batchSampler = new BatchConnectionSampler(this.connectionSampler, requests.length, maxPeers
|
|
197
|
-
|
|
198
|
-
|
|
535
|
+
const batchSampler = new BatchConnectionSampler(this.connectionSampler, requests.length, maxPeers, compactArray([
|
|
536
|
+
pinnedPeer
|
|
537
|
+
]), createLogger(`${this.logger.module}:batch-connection-sampler`));
|
|
538
|
+
if (batchSampler.activePeerCount === 0 && !pinnedPeer) {
|
|
539
|
+
this.logger.warn('No active peers to send requests to');
|
|
199
540
|
return [];
|
|
200
541
|
}
|
|
201
542
|
// This is where it gets fun
|
|
@@ -206,7 +547,9 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
206
547
|
// We will continue to retry until we have processed all indices, or we have reached the max retry attempts
|
|
207
548
|
let retryAttempts = 0;
|
|
208
549
|
while(pendingRequestIndices.size > 0 && batchSampler.activePeerCount > 0 && retryAttempts < maxRetryAttempts){
|
|
209
|
-
|
|
550
|
+
if (signal.aborted) {
|
|
551
|
+
throw new AbortError('Batch request aborted');
|
|
552
|
+
}
|
|
210
553
|
const requestBatches = new Map();
|
|
211
554
|
// Group requests by peer
|
|
212
555
|
for (const requestIndex of pendingRequestIndices){
|
|
@@ -214,10 +557,23 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
214
557
|
if (!peer) {
|
|
215
558
|
break;
|
|
216
559
|
}
|
|
217
|
-
|
|
218
|
-
|
|
560
|
+
const peerAsString = peer.toString();
|
|
561
|
+
if (!requestBatches.has(peerAsString)) {
|
|
562
|
+
requestBatches.set(peerAsString, {
|
|
563
|
+
peerId: peer,
|
|
564
|
+
indices: []
|
|
565
|
+
});
|
|
219
566
|
}
|
|
220
|
-
requestBatches.get(
|
|
567
|
+
requestBatches.get(peerAsString).indices.push(requestIndex);
|
|
568
|
+
}
|
|
569
|
+
// If there is a pinned peer, we will always send every request to that peer
|
|
570
|
+
// We use the default limits for the subprotocol to avoid hitting the rate limiter
|
|
571
|
+
if (pinnedPeer) {
|
|
572
|
+
const limit = this.rateLimiter.getRateLimits(subProtocol).peerLimit.quotaCount;
|
|
573
|
+
requestBatches.set(pinnedPeer.toString(), {
|
|
574
|
+
peerId: pinnedPeer,
|
|
575
|
+
indices: Array.from(pendingRequestIndices.values()).slice(0, limit)
|
|
576
|
+
});
|
|
221
577
|
}
|
|
222
578
|
// Make parallel requests for each peer's batch
|
|
223
579
|
// A batch entry will look something like this:
|
|
@@ -225,15 +581,16 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
225
581
|
// PeerId1: [4, 5, 6, 7]
|
|
226
582
|
// Peer Id 0 will send requests 0, 1, 2, 3 in serial
|
|
227
583
|
// while simultaneously Peer Id 1 will send requests 4, 5, 6, 7 in serial
|
|
228
|
-
const batchResults = await Promise.all(Array.from(requestBatches.entries()).map(async ([peer, indices])=>{
|
|
584
|
+
const batchResults = await Promise.all(Array.from(requestBatches.entries()).map(async ([peerAsString, { peerId: peer, indices }])=>{
|
|
229
585
|
try {
|
|
230
586
|
// Requests all going to the same peer are sent synchronously
|
|
231
587
|
const peerResults = [];
|
|
232
588
|
for (const index of indices){
|
|
589
|
+
this.logger.trace(`Sending request ${index} to peer ${peerAsString}`);
|
|
233
590
|
const response = await this.sendRequestToPeer(peer, subProtocol, requestBuffers[index]);
|
|
234
591
|
// Check the status of the response buffer
|
|
235
|
-
if (response
|
|
236
|
-
this.logger.debug(`Request to peer ${
|
|
592
|
+
if (response.status !== ReqRespStatus.SUCCESS) {
|
|
593
|
+
this.logger.debug(`Request to peer ${peerAsString} failed with status ${prettyPrintReqRespStatus(response.status)}`);
|
|
237
594
|
// If we hit a rate limit or some failure, we remove the peer and return the results,
|
|
238
595
|
// they will be split among remaining peers and the new sampled peer
|
|
239
596
|
batchSampler.removePeerAndReplace(peer);
|
|
@@ -243,7 +600,7 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
243
600
|
};
|
|
244
601
|
}
|
|
245
602
|
if (response && response.data.length > 0) {
|
|
246
|
-
const object =
|
|
603
|
+
const object = responseFromBuffer(subProtocol, response.data);
|
|
247
604
|
const isValid = await responseValidator(requests[index], object, peer);
|
|
248
605
|
if (isValid) {
|
|
249
606
|
peerResults.push({
|
|
@@ -258,7 +615,7 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
258
615
|
results: peerResults
|
|
259
616
|
};
|
|
260
617
|
} catch (error) {
|
|
261
|
-
this.logger.debug(`Failed batch request to peer ${
|
|
618
|
+
this.logger.debug(`Failed batch request to peer ${peerAsString}:`, error);
|
|
262
619
|
batchSampler.removePeerAndReplace(peer);
|
|
263
620
|
return {
|
|
264
621
|
peer,
|
|
@@ -299,6 +656,7 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
299
656
|
* @param peerId - The peer to send the request to
|
|
300
657
|
* @param subProtocol - The protocol to use to request
|
|
301
658
|
* @param payload - The payload to send
|
|
659
|
+
* @param dialTimeout - If establishing a stream takes longer than this an error will be thrown
|
|
302
660
|
* @returns If the request is successful, the response is returned, otherwise undefined
|
|
303
661
|
*
|
|
304
662
|
* @description
|
|
@@ -312,24 +670,46 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
312
670
|
*
|
|
313
671
|
* If the stream is not closed by the dialled peer, and a timeout occurs, then
|
|
314
672
|
* the stream is closed on the requester's end and sender (us) updates its peer score
|
|
315
|
-
*/ async sendRequestToPeer(peerId, subProtocol, payload) {
|
|
673
|
+
*/ async sendRequestToPeer(peerId, subProtocol, payload, dialTimeout = this.dialTimeoutMs) {
|
|
316
674
|
let stream;
|
|
317
675
|
try {
|
|
318
676
|
this.metrics.recordRequestSent(subProtocol);
|
|
319
|
-
|
|
320
|
-
|
|
321
|
-
|
|
322
|
-
|
|
323
|
-
|
|
324
|
-
|
|
677
|
+
this.logger.trace(`Sending request to peer ${peerId.toString()} on sub protocol ${subProtocol}`);
|
|
678
|
+
stream = await this.connectionSampler.dialProtocol(peerId, subProtocol, dialTimeout);
|
|
679
|
+
this.logger.trace(`Opened stream ${stream.id} for sending request to peer ${peerId.toString()} on sub protocol ${subProtocol}`);
|
|
680
|
+
const timeoutErr = new IndividualReqRespTimeoutError();
|
|
681
|
+
const [_, resp] = await executeTimeout((signal)=>Promise.all([
|
|
682
|
+
pipeline([
|
|
683
|
+
payload
|
|
684
|
+
], stream.sink, {
|
|
685
|
+
signal
|
|
686
|
+
}),
|
|
687
|
+
pipeline(stream.source, this.readMessage.bind(this), {
|
|
688
|
+
signal
|
|
689
|
+
})
|
|
690
|
+
]), this.individualRequestTimeoutMs, ()=>timeoutErr);
|
|
691
|
+
return resp;
|
|
325
692
|
} catch (e) {
|
|
693
|
+
// On error we immediately abort the stream, this is preferred way,
|
|
694
|
+
// because it signals to the sender that error happened, whereas
|
|
695
|
+
// closing the stream only closes our side and is much slower
|
|
696
|
+
if (stream) {
|
|
697
|
+
stream.abort(e);
|
|
698
|
+
}
|
|
326
699
|
this.metrics.recordRequestError(subProtocol);
|
|
327
700
|
this.handleResponseError(e, peerId, subProtocol);
|
|
701
|
+
// If there is an exception, we return an unknown response
|
|
702
|
+
this.logger.debug(`Error sending request to peer ${peerId.toString()} on sub protocol ${subProtocol}: ${e}`);
|
|
703
|
+
return {
|
|
704
|
+
status: ReqRespStatus.FAILURE
|
|
705
|
+
};
|
|
328
706
|
} finally{
|
|
329
707
|
// Only close the stream if we created it
|
|
708
|
+
// Note even if we aborted the stream, calling close on it is ok, it's just a no-op
|
|
330
709
|
if (stream) {
|
|
331
710
|
try {
|
|
332
|
-
|
|
711
|
+
this.logger.trace(`Closing stream ${stream.id} for request to peer ${peerId.toString()} on sub protocol ${subProtocol}`);
|
|
712
|
+
await this.connectionSampler.close(stream);
|
|
333
713
|
} catch (closeError) {
|
|
334
714
|
this.logger.error(`Error closing stream: ${closeError instanceof Error ? closeError.message : 'Unknown error'}`);
|
|
335
715
|
}
|
|
@@ -346,75 +726,41 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
346
726
|
* @param subProtocol - The sub protocol
|
|
347
727
|
* @returns If the error is non pubishable, then undefined is returned, otherwise the peer is penalized
|
|
348
728
|
*/ handleResponseError(e, peerId, subProtocol) {
|
|
349
|
-
const severity = this.
|
|
729
|
+
const severity = this.categorizeResponseError(e, peerId, subProtocol);
|
|
350
730
|
if (severity) {
|
|
351
731
|
this.peerScoring.penalizePeer(peerId, severity);
|
|
352
732
|
}
|
|
353
733
|
}
|
|
354
734
|
/**
|
|
355
|
-
* Categorize the error and log it.
|
|
356
|
-
*/ categorizeError(e, peerId, subProtocol) {
|
|
357
|
-
// Non punishable errors - we do not expect a response for goodbye messages
|
|
358
|
-
if (subProtocol === ReqRespSubProtocol.GOODBYE) {
|
|
359
|
-
this.logger.debug('Error encountered on goodbye sub protocol, no penalty', {
|
|
360
|
-
peerId: peerId.toString(),
|
|
361
|
-
subProtocol
|
|
362
|
-
});
|
|
363
|
-
return undefined;
|
|
364
|
-
}
|
|
365
|
-
// We do not punish a collective timeout, as the node triggers this interupt, independent of the peer's behaviour
|
|
366
|
-
const logTags = {
|
|
367
|
-
peerId: peerId.toString(),
|
|
368
|
-
subProtocol
|
|
369
|
-
};
|
|
370
|
-
if (e instanceof CollectiveReqRespTimeoutError || e instanceof InvalidResponseError) {
|
|
371
|
-
this.logger.debug(`Non-punishable error: ${e.message} | peerId: ${peerId.toString()} | subProtocol: ${subProtocol}`, logTags);
|
|
372
|
-
return undefined;
|
|
373
|
-
}
|
|
374
|
-
// Pubishable errors
|
|
375
|
-
// Connection reset errors in the networking stack are punished with high severity
|
|
376
|
-
// it just signals an unreliable peer
|
|
377
|
-
// We assume that the requesting node has a functioning networking stack.
|
|
378
|
-
if (e?.code === 'ECONNRESET' || e?.code === 'EPIPE') {
|
|
379
|
-
this.logger.debug(`Connection reset: ${peerId.toString()}`, logTags);
|
|
380
|
-
return PeerErrorSeverity.HighToleranceError;
|
|
381
|
-
}
|
|
382
|
-
if (e?.code === 'ECONNREFUSED') {
|
|
383
|
-
this.logger.debug(`Connection refused: ${peerId.toString()}`, logTags);
|
|
384
|
-
return PeerErrorSeverity.HighToleranceError;
|
|
385
|
-
}
|
|
386
|
-
// Timeout errors are punished with high tolerance, they can be due to a geogrpahically far away peer or an
|
|
387
|
-
// overloaded peer
|
|
388
|
-
if (e instanceof IndividualReqRespTimeoutError) {
|
|
389
|
-
this.logger.debug(`Timeout error: ${e.message} | peerId: ${peerId.toString()} | subProtocol: ${subProtocol}`, logTags);
|
|
390
|
-
return PeerErrorSeverity.HighToleranceError;
|
|
391
|
-
}
|
|
392
|
-
// Catch all error
|
|
393
|
-
this.logger.error(`Unexpected error sending request to peer`, e, logTags);
|
|
394
|
-
return PeerErrorSeverity.HighToleranceError;
|
|
395
|
-
}
|
|
396
|
-
/**
|
|
397
735
|
* Read a message returned from a stream into a single buffer
|
|
398
736
|
*
|
|
399
737
|
* The message is split into two components
|
|
400
738
|
* - The first chunk should contain a control byte, indicating the status of the response see `ReqRespStatus`
|
|
401
739
|
* - The second chunk should contain the response data
|
|
402
740
|
*/ async readMessage(source) {
|
|
403
|
-
let
|
|
741
|
+
let status;
|
|
404
742
|
const chunks = [];
|
|
405
743
|
try {
|
|
406
744
|
for await (const chunk of source){
|
|
407
|
-
|
|
408
|
-
|
|
409
|
-
statusBuffer = parseStatusChunk(firstChunkBuffer);
|
|
410
|
-
} else {
|
|
745
|
+
const statusParsed = status !== undefined;
|
|
746
|
+
if (statusParsed) {
|
|
411
747
|
chunks.push(chunk.subarray());
|
|
748
|
+
continue;
|
|
749
|
+
}
|
|
750
|
+
const firstChunkBuffer = chunk.subarray();
|
|
751
|
+
status = parseStatusChunk(firstChunkBuffer);
|
|
752
|
+
// In case status is not SUCCESS, we do not expect any data in the response
|
|
753
|
+
// we can return early
|
|
754
|
+
if (status !== ReqRespStatus.SUCCESS) {
|
|
755
|
+
return {
|
|
756
|
+
status
|
|
757
|
+
};
|
|
412
758
|
}
|
|
413
759
|
}
|
|
414
760
|
const messageData = Buffer.concat(chunks);
|
|
415
|
-
const message = this.snappyTransform.
|
|
761
|
+
const message = this.snappyTransform.inboundTransformData(messageData);
|
|
416
762
|
return {
|
|
417
|
-
status:
|
|
763
|
+
status: status ?? ReqRespStatus.UNKNOWN,
|
|
418
764
|
data: message
|
|
419
765
|
};
|
|
420
766
|
} catch (e) {
|
|
@@ -424,8 +770,7 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
424
770
|
status = e.status;
|
|
425
771
|
}
|
|
426
772
|
return {
|
|
427
|
-
status
|
|
428
|
-
data: Buffer.from([])
|
|
773
|
+
status
|
|
429
774
|
};
|
|
430
775
|
}
|
|
431
776
|
}
|
|
@@ -433,7 +778,8 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
433
778
|
* Stream Handler
|
|
434
779
|
* Reads the incoming stream, determines the protocol, then triggers the appropriate handler
|
|
435
780
|
*
|
|
436
|
-
* @param
|
|
781
|
+
* @param protocol - The sub protocol to handle
|
|
782
|
+
* @param incomingStream - The incoming stream data containing the stream and connection
|
|
437
783
|
*
|
|
438
784
|
* @description
|
|
439
785
|
* An individual stream handler will be bound to each sub protocol, and handles returning data back
|
|
@@ -444,73 +790,196 @@ import { ReqRespStatus, ReqRespStatusError, parseStatusChunk, prettyPrintReqResp
|
|
|
444
790
|
*
|
|
445
791
|
* We check rate limits for each peer, note the peer will be penalised within the rate limiter implementation
|
|
446
792
|
* if they exceed their peer specific limits.
|
|
447
|
-
*/ async streamHandler(protocol,
|
|
448
|
-
|
|
793
|
+
*/ async streamHandler(protocol, incomingStream) {
|
|
794
|
+
const { stream, connection } = incomingStream;
|
|
449
795
|
try {
|
|
450
|
-
|
|
796
|
+
this.metrics.recordRequestReceived(protocol);
|
|
451
797
|
const rateLimitStatus = this.rateLimiter.allow(protocol, connection.remotePeer);
|
|
452
|
-
if (rateLimitStatus
|
|
453
|
-
this.logger.
|
|
798
|
+
if (rateLimitStatus !== RateLimitStatus.Allowed) {
|
|
799
|
+
this.logger.verbose(`Rate limit exceeded ${prettyPrintRateLimitStatus(rateLimitStatus)} for ${protocol} from ${connection.remotePeer}`);
|
|
454
800
|
throw new ReqRespStatusError(ReqRespStatus.RATE_LIMIT_EXCEEDED);
|
|
455
801
|
}
|
|
456
|
-
|
|
457
|
-
|
|
458
|
-
await pipe(stream, async function*(source) {
|
|
459
|
-
for await (const chunkList of source){
|
|
460
|
-
const msg = Buffer.from(chunkList.subarray());
|
|
461
|
-
const response = await handler(connection.remotePeer, msg);
|
|
462
|
-
if (protocol === ReqRespSubProtocol.GOODBYE) {
|
|
463
|
-
// Don't respond
|
|
464
|
-
await stream.close();
|
|
465
|
-
return;
|
|
466
|
-
}
|
|
467
|
-
// Send success code first, then the response
|
|
468
|
-
const successChunk = Buffer.from([
|
|
469
|
-
ReqRespStatus.SUCCESS
|
|
470
|
-
]);
|
|
471
|
-
yield new Uint8Array(successChunk);
|
|
472
|
-
yield new Uint8Array(transform.outboundTransformNoTopic(response));
|
|
473
|
-
}
|
|
474
|
-
}, stream);
|
|
475
|
-
} catch (e) {
|
|
476
|
-
this.logger.warn('Reqresp Response error: ', e);
|
|
802
|
+
await this.processStream(protocol, incomingStream);
|
|
803
|
+
} catch (err) {
|
|
477
804
|
this.metrics.recordResponseError(protocol);
|
|
478
|
-
|
|
479
|
-
|
|
480
|
-
|
|
481
|
-
|
|
805
|
+
this.handleRequestError(err, connection.remotePeer, protocol);
|
|
806
|
+
if (err instanceof ReqRespStatusError) {
|
|
807
|
+
const errorSent = await this.trySendError(stream, connection.remotePeer, protocol, err.status);
|
|
808
|
+
const logMessage = errorSent ? 'Protocol error sent successfully.' : 'Stream already closed or poisoned, not sending error response.';
|
|
809
|
+
const isRateLimit = err.status === ReqRespStatus.RATE_LIMIT_EXCEEDED;
|
|
810
|
+
const level = isRateLimit ? 'debug' : 'warn';
|
|
811
|
+
this.logger[level](logMessage + ` Status: ${ReqRespStatus[err.status]}`, {
|
|
812
|
+
protocol,
|
|
813
|
+
err,
|
|
814
|
+
errorStatus: err.status,
|
|
815
|
+
cause: err.cause ?? 'Cause unknown'
|
|
816
|
+
});
|
|
817
|
+
} else {
|
|
818
|
+
// In erroneous case we abort the stream, this will signal the peer that something went wrong
|
|
819
|
+
// and that this stream should be dropped
|
|
820
|
+
const isMessageToNotWarn = err instanceof Error && [
|
|
821
|
+
'stream reset',
|
|
822
|
+
'Cannot push value onto an ended pushable'
|
|
823
|
+
].some((msg)=>err.message.includes(msg));
|
|
824
|
+
const level = isMessageToNotWarn ? 'debug' : 'warn';
|
|
825
|
+
this.logger[level]('Unknown stream error while handling the stream, aborting', {
|
|
826
|
+
protocol,
|
|
827
|
+
err
|
|
828
|
+
});
|
|
829
|
+
stream.abort(err);
|
|
482
830
|
}
|
|
483
|
-
const sendErrorChunk = this.sendErrorChunk(errorStatus);
|
|
484
|
-
// Return and yield the response chunk
|
|
485
|
-
await pipe(stream, async function*(_source) {
|
|
486
|
-
yield* sendErrorChunk;
|
|
487
|
-
}, stream);
|
|
488
831
|
} finally{
|
|
832
|
+
//NOTE: This is idempotent action, so it's ok to call it even if stream was aborted
|
|
489
833
|
await stream.close();
|
|
490
834
|
}
|
|
491
835
|
}
|
|
492
|
-
|
|
493
|
-
|
|
494
|
-
|
|
495
|
-
|
|
496
|
-
|
|
836
|
+
/**
|
|
837
|
+
* Reads incoming data from the stream, processes it according to the sub protocol,
|
|
838
|
+
* and puts response back into the stream.
|
|
839
|
+
*
|
|
840
|
+
* @param protocol - The sub protocol to use for processing the stream
|
|
841
|
+
* @param incomingStream - The incoming stream data containing the stream and connection
|
|
842
|
+
*
|
|
843
|
+
* */ async processStream(protocol, { stream, connection }) {
|
|
844
|
+
const handler = this.subProtocolHandlers[protocol];
|
|
845
|
+
if (!handler) {
|
|
846
|
+
throw new Error(`No handler defined for reqresp subprotocol ${protocol}`);
|
|
847
|
+
}
|
|
848
|
+
const snappy = this.snappyTransform;
|
|
849
|
+
const SUCCESS = Uint8Array.of(ReqRespStatus.SUCCESS);
|
|
850
|
+
await pipeline(stream.source, async function*(source) {
|
|
851
|
+
for await (const chunk of source){
|
|
852
|
+
const response = await handler(connection.remotePeer, chunk.subarray());
|
|
853
|
+
if (protocol === ReqRespSubProtocol.GOODBYE) {
|
|
854
|
+
// NOTE: The stream was already closed by Goodbye handler
|
|
855
|
+
// peerManager.goodbyeReceived(peerId, reason); will call libp2p.hangUp closing all active streams and connections
|
|
856
|
+
// Don't try to respond
|
|
857
|
+
return;
|
|
858
|
+
}
|
|
859
|
+
stream.metadata.written = true; // Mark the stream as written to;
|
|
860
|
+
yield SUCCESS;
|
|
861
|
+
yield snappy.outboundTransformData(response);
|
|
862
|
+
}
|
|
863
|
+
}, stream.sink);
|
|
864
|
+
}
|
|
865
|
+
/**
|
|
866
|
+
* Try to send error status to the peer. We say try, because the stream,
|
|
867
|
+
* might already be closed
|
|
868
|
+
* @param stream - The stream opened between us and the peer
|
|
869
|
+
* @param status - The error status to send back to the peer
|
|
870
|
+
* @returns true if error was sent successfully, otherwise false
|
|
871
|
+
*
|
|
872
|
+
*/ async trySendError(stream, peerId, protocol, status) {
|
|
873
|
+
const canWriteToStream = // 'writing' is a bit weird naming, but it actually means that the stream is ready to write
|
|
874
|
+
// 'ready' means that stream ready to be opened for writing
|
|
875
|
+
stream.status === 'open' && (stream.writeStatus === 'writing' || stream.writeStatus === 'ready');
|
|
876
|
+
// Stream was already written to, we consider it poisoned, in a sense,
|
|
877
|
+
// that even if we write an error response, it will not be interpreted correctly by the peer
|
|
878
|
+
const streamPoisoned = stream.metadata.written === true;
|
|
879
|
+
const shouldWriteToStream = canWriteToStream && !streamPoisoned;
|
|
880
|
+
if (!shouldWriteToStream) {
|
|
881
|
+
return false;
|
|
882
|
+
}
|
|
883
|
+
try {
|
|
884
|
+
await pipeline(function*() {
|
|
885
|
+
yield Uint8Array.of(status);
|
|
886
|
+
}, stream.sink);
|
|
887
|
+
return true;
|
|
888
|
+
} catch (e) {
|
|
889
|
+
this.logger.warn('Error while sending error response', e);
|
|
890
|
+
stream.abort(e);
|
|
891
|
+
this.handleRequestError(e, peerId, protocol);
|
|
892
|
+
return false;
|
|
893
|
+
}
|
|
894
|
+
}
|
|
895
|
+
handleRequestError(e, peerId, subProtocol) {
|
|
896
|
+
const severity = this.categorizeRequestError(e, peerId, subProtocol);
|
|
897
|
+
if (severity) {
|
|
898
|
+
this.peerScoring.penalizePeer(peerId, severity);
|
|
899
|
+
}
|
|
900
|
+
}
|
|
901
|
+
/**
|
|
902
|
+
* Categorize the request error and log it.
|
|
903
|
+
*
|
|
904
|
+
* @returns Severity of the error, or undefined if the error is not punishable.
|
|
905
|
+
*/ categorizeRequestError(e, peerId, subProtocol) {
|
|
906
|
+
const logTags = {
|
|
907
|
+
peerId: peerId.toString(),
|
|
908
|
+
subProtocol
|
|
909
|
+
};
|
|
910
|
+
//Punishable error - peer should never send badly formed request
|
|
911
|
+
if (e instanceof ReqRespStatusError && e.status === ReqRespStatus.BADLY_FORMED_REQUEST) {
|
|
912
|
+
this.logger.debug(`Punishable error in ${subProtocol}: ${e.cause}`, logTags);
|
|
913
|
+
return PeerErrorSeverity.LowToleranceError;
|
|
914
|
+
}
|
|
915
|
+
//TODO: (mralj): think if we should penalize peer here based on connection errors
|
|
916
|
+
return undefined;
|
|
917
|
+
}
|
|
918
|
+
/**
|
|
919
|
+
* Categorize the response error and log it.
|
|
920
|
+
*
|
|
921
|
+
* @returns Severity of the error, or undefined if the error is not punishable.
|
|
922
|
+
*/ categorizeResponseError(e, peerId, subProtocol) {
|
|
923
|
+
const logTags = {
|
|
924
|
+
peerId: peerId.toString(),
|
|
925
|
+
subProtocol
|
|
926
|
+
};
|
|
927
|
+
// Non punishable errors - we do not expect a response for goodbye messages
|
|
928
|
+
if (subProtocol === ReqRespSubProtocol.GOODBYE) {
|
|
929
|
+
this.logger.debug('Error encountered on goodbye sub protocol, no penalty', logTags);
|
|
930
|
+
return undefined;
|
|
931
|
+
}
|
|
932
|
+
// We do not punish a collective timeout, as the node triggers this interupt, independent of the peer's behaviour
|
|
933
|
+
if (e instanceof CollectiveReqRespTimeoutError || e instanceof InvalidResponseError) {
|
|
934
|
+
this.logger.debug(`Non-punishable error in ${subProtocol}: ${e.message}`, logTags);
|
|
935
|
+
return undefined;
|
|
936
|
+
}
|
|
937
|
+
return this.categorizeConnectionErrors(e, peerId, subProtocol);
|
|
938
|
+
}
|
|
939
|
+
/*
|
|
940
|
+
* Errors specific to connection handling
|
|
941
|
+
* These can happen both when sending request and response*/ categorizeConnectionErrors(e, peerId, subProtocol) {
|
|
942
|
+
const logTags = {
|
|
943
|
+
peerId: peerId.toString(),
|
|
944
|
+
subProtocol
|
|
945
|
+
};
|
|
946
|
+
// Do not punish if we are stopping the service
|
|
947
|
+
if (e instanceof AbortError || e?.code == 'ABORT_ERR') {
|
|
948
|
+
this.logger.debug(`Request aborted: ${e.message}`, logTags);
|
|
949
|
+
return undefined;
|
|
950
|
+
}
|
|
951
|
+
// Do not punish if we are the ones closing the connection
|
|
952
|
+
if (e?.code === 'ERR_CONNECTION_BEING_CLOSED' || e?.code === 'ERR_CONNECTION_CLOSED' || e?.code === 'ERR_TRANSIENT_CONNECTION' || e?.message?.includes('Muxer already closed') || e?.message?.includes('muxer closed') || e?.message?.includes('ended pushable')) {
|
|
953
|
+
this.logger.debug(`Connection closed to peer from our side: ${peerId.toString()} (${e?.message ?? 'missing error message'})`, logTags);
|
|
954
|
+
return undefined;
|
|
955
|
+
}
|
|
956
|
+
// Pubishable errors
|
|
957
|
+
// Connection reset errors in the networking stack are punished with high severity
|
|
958
|
+
// it just signals an unreliable peer
|
|
959
|
+
// We assume that the requesting node has a functioning networking stack.
|
|
960
|
+
if (e?.code === 'ECONNRESET' || e?.code === 'EPIPE') {
|
|
961
|
+
this.logger.debug(`Connection reset: ${peerId.toString()}`, logTags);
|
|
962
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
963
|
+
}
|
|
964
|
+
if (e?.code === 'ECONNREFUSED') {
|
|
965
|
+
this.logger.debug(`Connection refused: ${peerId.toString()}`, logTags);
|
|
966
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
967
|
+
}
|
|
968
|
+
if (e?.code === 'ERR_UNEXPECTED_EOF') {
|
|
969
|
+
this.logger.debug(`Connection unexpected EOF: ${peerId.toString()}`, logTags);
|
|
970
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
971
|
+
}
|
|
972
|
+
if (e?.code === 'ERR_UNSUPPORTED_PROTOCOL') {
|
|
973
|
+
this.logger.debug(`Sub protocol not supported by peer: ${peerId.toString()}`, logTags);
|
|
974
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
975
|
+
}
|
|
976
|
+
// Timeout errors are punished with high tolerance, they can be due to a geographically far away or overloaded peer
|
|
977
|
+
if (e instanceof IndividualReqRespTimeoutError || e instanceof TimeoutError) {
|
|
978
|
+
this.logger.debug(`Timeout error in ${subProtocol}: ${e.message}`, logTags);
|
|
979
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
980
|
+
}
|
|
981
|
+
// Catch all error
|
|
982
|
+
this.logger.error(`Unexpected error in ReqResp protocol`, e, logTags);
|
|
983
|
+
return PeerErrorSeverity.HighToleranceError;
|
|
497
984
|
}
|
|
498
985
|
}
|
|
499
|
-
_ts_decorate([
|
|
500
|
-
trackSpan('ReqResp.sendBatchRequest', (subProtocol, requests)=>({
|
|
501
|
-
[Attributes.P2P_REQ_RESP_PROTOCOL]: subProtocol,
|
|
502
|
-
[Attributes.P2P_REQ_RESP_BATCH_REQUESTS_COUNT]: requests.length
|
|
503
|
-
}))
|
|
504
|
-
], ReqResp.prototype, "sendBatchRequest", null);
|
|
505
|
-
_ts_decorate([
|
|
506
|
-
trackSpan('ReqResp.sendRequestToPeer', (peerId, subProtocol, _)=>({
|
|
507
|
-
[Attributes.P2P_ID]: peerId.toString(),
|
|
508
|
-
[Attributes.P2P_REQ_RESP_PROTOCOL]: subProtocol
|
|
509
|
-
}))
|
|
510
|
-
], ReqResp.prototype, "sendRequestToPeer", null);
|
|
511
|
-
_ts_decorate([
|
|
512
|
-
trackSpan('ReqResp.streamHandler', (protocol, { connection })=>({
|
|
513
|
-
[Attributes.P2P_REQ_RESP_PROTOCOL]: protocol,
|
|
514
|
-
[Attributes.P2P_ID]: connection.remotePeer.toString()
|
|
515
|
-
}))
|
|
516
|
-
], ReqResp.prototype, "streamHandler", null);
|