mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-21 22:04:44 +00:00
* some blob reprocessing work
* remove ForceBlockLookup
* reorder enum match arms in sync manager
* a lot more reprocessing work
* impl logic for triggerng blob lookups along with block lookups
* deal with rpc blobs in groups per block in the da checker. don't cache missing blob ids in the da checker.
* make single block lookup generic
* more work
* add delayed processing logic and combine some requests
* start fixing some compile errors
* fix compilation in main block lookup mod
* much work
* get things compiling
* parent blob lookups
* fix compile
* revert red/stevie changes
* fix up sync manager delay message logic
* add peer usefulness enum
* should remove lookup refactor
* consolidate retry error handling
* improve peer scoring during certain failures in parent lookups
* improve retry code
* drop parent lookup if either req has a peer disconnect during download
* refactor single block processed method
* processing peer refactor
* smol bugfix
* fix some todos
* fix lints
* fix lints
* fix compile in lookup tests
* fix lints
* fix lints
* fix existing block lookup tests
* renamings
* fix after merge
* cargo fmt
* compilation fix in beacon chain tests
* fix
* refactor lookup tests to work with multiple forks and response types
* make tests into macros
* wrap availability check error
* fix compile after merge
* add random blobs
* start fixing up lookup verify error handling
* some bug fixes and the start of deneb only tests
* make tests work for all forks
* track information about peer source
* error refactoring
* improve peer scoring
* fix test compilation
* make sure blobs are sent for processing after stream termination, delete copied tests
* add some tests and fix a bug
* smol bugfixes and moar tests
* add tests and fix some things
* compile after merge
* lots of refactoring
* retry on invalid block/blob
* merge unknown parent messages before current slot lookup
* get tests compiling
* penalize blob peer on invalid blobs
* Check disk on in-memory cache miss
* Update beacon_node/beacon_chain/src/data_availability_checker/overflow_lru_cache.rs
* Update beacon_node/network/src/sync/network_context.rs
Co-authored-by: Divma <26765164+divagant-martian@users.noreply.github.com>
* fix bug in matching blocks and blobs in range sync
* pr feedback
* fix conflicts
* upgrade logs from warn to crit when we receive incorrect response in range
* synced_and_connected_within_tolerance -> should_search_for_block
* remove todo
* add data gas used and update excess data gas to u64
* Fix Broken Overflow Tests
* payload verification with commitments
* fix merge conflicts
* restore payload file
* Restore payload file
* remove todo
* add max blob commitments per block
* c-kzg lib update
* Fix ef tests
* Abstract over minimal/mainnet spec in kzg crate
* Start integrating new KZG
* checkpoint sync without alignment
* checkpoint sync without alignment
* add import
* add import
* query for checkpoint state by slot rather than state root (teku doesn't serve by state root)
* query for checkpoint state by slot rather than state root (teku doesn't serve by state root)
* loosen check
* get state first and query by most recent block root
* Revert "loosen check"
This reverts commit 069d13dd63.
* get state first and query by most recent block root
* merge max blobs change
* simplify delay logic
* rename unknown parent sync message variants
* rename parameter, block_slot -> slot
* add some docs to the lookup module
* use interval instead of sleep
* drop request if blocks and blobs requests both return `None` for `Id`
* clean up `find_single_lookup` logic
* add lookup source enum
* clean up `find_single_lookup` logic
* add docs to find_single_lookup_request
* move LookupSource our of param where unnecessary
* remove unnecessary todo
* query for block by `state.latest_block_header.slot`
* fix lint
* fix merge transition ef tests
* fix test
* fix test
* fix observed blob sidecars test
* Add some metrics (#33)
* fix protocol limits for blobs by root
* Update Engine API for 1:1 Structure Method
* make beacon chain tests to fix devnet 6 changes
* get ckzg working and fix some tests
* fix remaining tests
* fix lints
* Fix KZG linking issues
* remove unused dep
* lockfile
* test fixes
* remove dbgs
* remove unwrap
* cleanup tx generator
* small fixes
* fixing fixes
* more self reivew
* more self review
* refactor genesis header initialization
* refactor mock el instantiations
* fix compile
* fix network test, make sure they run for each fork
* pr feedback
* fix last test (hopefully)
---------
Co-authored-by: Pawan Dhananjay <pawandhananjay@gmail.com>
Co-authored-by: Mark Mackey <mark@sigmaprime.io>
Co-authored-by: Divma <26765164+divagant-martian@users.noreply.github.com>
Co-authored-by: Michael Sproul <michael@sigmaprime.io>
247 lines
8.1 KiB
Rust
247 lines
8.1 KiB
Rust
#![cfg(not(debug_assertions))]
|
|
|
|
use beacon_chain::blob_verification::BlockWrapper;
|
|
use beacon_chain::test_utils::{AttestationStrategy, BeaconChainHarness, BlockStrategy};
|
|
use beacon_chain::{StateSkipConfig, WhenSlotSkipped};
|
|
use lazy_static::lazy_static;
|
|
use std::sync::Arc;
|
|
use tree_hash::TreeHash;
|
|
use types::{AggregateSignature, EthSpec, Keypair, MainnetEthSpec, RelativeEpoch, Slot};
|
|
|
|
pub const VALIDATOR_COUNT: usize = 16;
|
|
|
|
lazy_static! {
|
|
/// A cached set of keys.
|
|
static ref KEYPAIRS: Vec<Keypair> = types::test_utils::generate_deterministic_keypairs(VALIDATOR_COUNT);
|
|
}
|
|
|
|
/// This test builds a chain that is just long enough to finalize an epoch then it produces an
|
|
/// attestation at each slot from genesis through to three epochs past the head.
|
|
///
|
|
/// It checks the produced attestation against some locally computed values.
|
|
#[tokio::test]
|
|
async fn produces_attestations() {
|
|
let num_blocks_produced = MainnetEthSpec::slots_per_epoch() * 4;
|
|
let additional_slots_tested = MainnetEthSpec::slots_per_epoch() * 3;
|
|
|
|
let harness = BeaconChainHarness::builder(MainnetEthSpec)
|
|
.default_spec()
|
|
.keypairs(KEYPAIRS[..].to_vec())
|
|
.fresh_ephemeral_store()
|
|
.mock_execution_layer()
|
|
.build();
|
|
|
|
let chain = &harness.chain;
|
|
|
|
// Test all valid committee indices for all slots in the chain.
|
|
// for slot in 0..=current_slot.as_u64() + MainnetEthSpec::slots_per_epoch() * 3 {
|
|
for slot in 0..=num_blocks_produced + additional_slots_tested {
|
|
if slot > 0 && slot <= num_blocks_produced {
|
|
harness.advance_slot();
|
|
|
|
harness
|
|
.extend_chain(
|
|
1,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
)
|
|
.await;
|
|
}
|
|
|
|
let slot = Slot::from(slot);
|
|
let mut state = chain
|
|
.state_at_slot(slot, StateSkipConfig::WithStateRoots)
|
|
.expect("should get state");
|
|
|
|
let block_slot = if slot <= num_blocks_produced {
|
|
slot
|
|
} else {
|
|
Slot::from(num_blocks_produced)
|
|
};
|
|
|
|
let blinded_block = chain
|
|
.block_at_slot(block_slot, WhenSlotSkipped::Prev)
|
|
.expect("should get block")
|
|
.expect("block should not be skipped");
|
|
let block_root = blinded_block.message().tree_hash_root();
|
|
let block = chain
|
|
.store
|
|
.make_full_block(&block_root, blinded_block)
|
|
.unwrap();
|
|
let blobs = chain.get_blobs(&block_root).unwrap();
|
|
|
|
let epoch_boundary_slot = state
|
|
.current_epoch()
|
|
.start_slot(MainnetEthSpec::slots_per_epoch());
|
|
let target_root = if state.slot() == epoch_boundary_slot {
|
|
block_root
|
|
} else {
|
|
*state
|
|
.get_block_root(epoch_boundary_slot)
|
|
.expect("should get target block root")
|
|
};
|
|
|
|
state
|
|
.build_committee_cache(RelativeEpoch::Current, &harness.chain.spec)
|
|
.unwrap();
|
|
let committee_cache = state
|
|
.committee_cache(RelativeEpoch::Current)
|
|
.expect("should get committee_cache");
|
|
|
|
let committee_count = committee_cache.committees_per_slot();
|
|
|
|
for index in 0..committee_count {
|
|
let committee_len = committee_cache
|
|
.get_beacon_committee(slot, index)
|
|
.expect("should get committee for slot")
|
|
.committee
|
|
.len();
|
|
|
|
let attestation = chain
|
|
.produce_unaggregated_attestation(slot, index)
|
|
.expect("should produce attestation");
|
|
|
|
let data = &attestation.data;
|
|
|
|
assert_eq!(
|
|
attestation.aggregation_bits.len(),
|
|
committee_len,
|
|
"bad committee len"
|
|
);
|
|
assert!(
|
|
attestation.aggregation_bits.is_zero(),
|
|
"some committee bits are set"
|
|
);
|
|
assert_eq!(
|
|
attestation.signature,
|
|
AggregateSignature::empty(),
|
|
"bad signature"
|
|
);
|
|
assert_eq!(data.index, index, "bad index");
|
|
assert_eq!(data.slot, slot, "bad slot");
|
|
assert_eq!(data.beacon_block_root, block_root, "bad block root");
|
|
assert_eq!(
|
|
data.source,
|
|
state.current_justified_checkpoint(),
|
|
"bad source"
|
|
);
|
|
assert_eq!(
|
|
data.source,
|
|
state.current_justified_checkpoint(),
|
|
"bad source"
|
|
);
|
|
assert_eq!(data.target.epoch, state.current_epoch(), "bad target epoch");
|
|
assert_eq!(data.target.root, target_root, "bad target root");
|
|
|
|
let block_wrapper =
|
|
BlockWrapper::<MainnetEthSpec>::new(Arc::new(block.clone()), blobs.clone());
|
|
let beacon_chain::blob_verification::MaybeAvailableBlock::Available(available_block) = chain
|
|
.data_availability_checker
|
|
.check_availability(block_wrapper)
|
|
.unwrap()
|
|
else {
|
|
panic!("block should be available")
|
|
};
|
|
|
|
let early_attestation = {
|
|
let proto_block = chain
|
|
.canonical_head
|
|
.fork_choice_read_lock()
|
|
.get_block(&block_root)
|
|
.unwrap();
|
|
chain
|
|
.early_attester_cache
|
|
.add_head_block(
|
|
block_root,
|
|
available_block,
|
|
proto_block,
|
|
&state,
|
|
&chain.spec,
|
|
)
|
|
.unwrap();
|
|
chain
|
|
.early_attester_cache
|
|
.try_attest(slot, index, &chain.spec)
|
|
.unwrap()
|
|
.unwrap()
|
|
};
|
|
|
|
assert_eq!(
|
|
attestation, early_attestation,
|
|
"early attester cache inconsistent"
|
|
);
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Ensures that the early attester cache wont create an attestation to a block in a later slot than
|
|
/// the one requested.
|
|
#[tokio::test]
|
|
async fn early_attester_cache_old_request() {
|
|
let harness = BeaconChainHarness::builder(MainnetEthSpec)
|
|
.default_spec()
|
|
.keypairs(KEYPAIRS[..].to_vec())
|
|
.fresh_ephemeral_store()
|
|
.mock_execution_layer()
|
|
.build();
|
|
|
|
harness.advance_slot();
|
|
|
|
harness
|
|
.extend_chain(
|
|
2,
|
|
BlockStrategy::OnCanonicalHead,
|
|
AttestationStrategy::AllValidators,
|
|
)
|
|
.await;
|
|
|
|
let head = harness.chain.head_snapshot();
|
|
assert_eq!(head.beacon_block.slot(), 2);
|
|
let head_proto_block = harness
|
|
.chain
|
|
.canonical_head
|
|
.fork_choice_read_lock()
|
|
.get_block(&head.beacon_block_root)
|
|
.unwrap();
|
|
|
|
let head_blobs = harness
|
|
.chain
|
|
.get_blobs(&head.beacon_block_root)
|
|
.expect("should get blobs");
|
|
|
|
let block_wrapper = BlockWrapper::<MainnetEthSpec>::new(head.beacon_block.clone(), head_blobs);
|
|
let beacon_chain::blob_verification::MaybeAvailableBlock::Available(available_block) = harness.chain
|
|
.data_availability_checker
|
|
.check_availability(block_wrapper)
|
|
.unwrap()
|
|
else {
|
|
panic!("block should be available")
|
|
};
|
|
|
|
harness
|
|
.chain
|
|
.early_attester_cache
|
|
.add_head_block(
|
|
head.beacon_block_root,
|
|
available_block,
|
|
head_proto_block,
|
|
&head.beacon_state,
|
|
&harness.chain.spec,
|
|
)
|
|
.unwrap();
|
|
|
|
let attest_slot = head.beacon_block.slot() - 1;
|
|
let attestation = harness
|
|
.chain
|
|
.produce_unaggregated_attestation(attest_slot, 0)
|
|
.unwrap();
|
|
|
|
assert_eq!(attestation.data.slot, attest_slot);
|
|
let attested_block = harness
|
|
.chain
|
|
.get_blinded_block(&attestation.data.beacon_block_root)
|
|
.unwrap()
|
|
.unwrap();
|
|
assert_eq!(attested_block.slot(), attest_slot);
|
|
}
|