mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-20 21:34:46 +00:00
Merge remote-tracking branch 'origin/unstable' into tree-states
This commit is contained in:
@@ -26,6 +26,7 @@ fn produces_attestations() {
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[..].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
let chain = &harness.chain;
|
||||
|
||||
@@ -42,6 +42,7 @@ fn get_harness(validator_count: usize) -> BeaconChainHarness<EphemeralHarnessTyp
|
||||
.spec(spec)
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
harness.advance_slot();
|
||||
|
||||
@@ -54,6 +54,7 @@ fn get_harness(validator_count: usize) -> BeaconChainHarness<EphemeralHarnessTyp
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
harness.advance_slot();
|
||||
@@ -839,6 +840,7 @@ fn verify_block_for_gossip_slashing_detection() {
|
||||
.keypairs(KEYPAIRS.to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.initial_mutator(Box::new(move |builder| builder.slasher(inner_slasher)))
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
harness.advance_slot();
|
||||
|
||||
@@ -918,6 +920,7 @@ fn add_base_block_to_altair_chain() {
|
||||
.spec(spec)
|
||||
.keypairs(KEYPAIRS[..].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
// Move out of the genesis slot.
|
||||
@@ -1037,6 +1040,7 @@ fn add_altair_block_to_base_chain() {
|
||||
.spec(spec)
|
||||
.keypairs(KEYPAIRS[..].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
// Move out of the genesis slot.
|
||||
|
||||
@@ -3,6 +3,7 @@ mod attestation_verification;
|
||||
mod block_verification;
|
||||
mod merge;
|
||||
mod op_verification;
|
||||
mod payload_invalidation;
|
||||
mod store_tests;
|
||||
mod sync_committee_verification;
|
||||
mod tests;
|
||||
|
||||
@@ -13,7 +13,7 @@ fn verify_execution_payload_chain<T: EthSpec>(chain: &[ExecutionPayload<T>]) {
|
||||
|
||||
for ep in chain {
|
||||
assert!(*ep != ExecutionPayload::default());
|
||||
assert!(ep.block_hash != Hash256::zero());
|
||||
assert!(ep.block_hash != ExecutionBlockHash::zero());
|
||||
|
||||
// Check against previous `ExecutionPayload`.
|
||||
if let Some(prev_ep) = prev_ep {
|
||||
@@ -40,7 +40,7 @@ fn merge_with_terminal_block_hash_override() {
|
||||
spec.terminal_total_difficulty,
|
||||
DEFAULT_TERMINAL_BLOCK,
|
||||
0,
|
||||
Hash256::zero(),
|
||||
ExecutionBlockHash::zero(),
|
||||
)
|
||||
.unwrap()
|
||||
.block_hash;
|
||||
|
||||
@@ -40,6 +40,7 @@ fn get_harness(store: Arc<HotColdDB>, validator_count: usize) -> TestHarness {
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_disk_store(store)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
harness.advance_slot();
|
||||
harness
|
||||
|
||||
601
beacon_node/beacon_chain/tests/payload_invalidation.rs
Normal file
601
beacon_node/beacon_chain/tests/payload_invalidation.rs
Normal file
@@ -0,0 +1,601 @@
|
||||
#![cfg(not(debug_assertions))]
|
||||
|
||||
use beacon_chain::{
|
||||
test_utils::{BeaconChainHarness, EphemeralHarnessType},
|
||||
BeaconChainError, BlockError, ExecutionPayloadError, HeadInfo, StateSkipConfig,
|
||||
WhenSlotSkipped, INVALID_JUSTIFIED_PAYLOAD_SHUTDOWN_REASON,
|
||||
};
|
||||
use proto_array::ExecutionStatus;
|
||||
use task_executor::ShutdownReason;
|
||||
use types::*;
|
||||
|
||||
const VALIDATOR_COUNT: usize = 32;
|
||||
|
||||
type E = MainnetEthSpec;
|
||||
|
||||
#[derive(PartialEq, Clone)]
|
||||
enum Payload {
|
||||
Valid,
|
||||
Invalid {
|
||||
latest_valid_hash: Option<ExecutionBlockHash>,
|
||||
},
|
||||
Syncing,
|
||||
}
|
||||
|
||||
struct InvalidPayloadRig {
|
||||
harness: BeaconChainHarness<EphemeralHarnessType<E>>,
|
||||
enable_attestations: bool,
|
||||
}
|
||||
|
||||
impl InvalidPayloadRig {
|
||||
fn new() -> Self {
|
||||
let mut spec = E::default_spec();
|
||||
spec.altair_fork_epoch = Some(Epoch::new(0));
|
||||
spec.bellatrix_fork_epoch = Some(Epoch::new(0));
|
||||
|
||||
let harness = BeaconChainHarness::builder(MainnetEthSpec)
|
||||
.spec(spec)
|
||||
.deterministic_keypairs(VALIDATOR_COUNT)
|
||||
.mock_execution_layer()
|
||||
.fresh_ephemeral_store()
|
||||
.build();
|
||||
|
||||
// Move to slot 1.
|
||||
harness.advance_slot();
|
||||
|
||||
Self {
|
||||
harness,
|
||||
enable_attestations: false,
|
||||
}
|
||||
}
|
||||
|
||||
fn enable_attestations(mut self) -> Self {
|
||||
self.enable_attestations = true;
|
||||
self
|
||||
}
|
||||
|
||||
fn block_hash(&self, block_root: Hash256) -> ExecutionBlockHash {
|
||||
self.harness
|
||||
.chain
|
||||
.get_block(&block_root)
|
||||
.unwrap()
|
||||
.unwrap()
|
||||
.message()
|
||||
.body()
|
||||
.execution_payload()
|
||||
.unwrap()
|
||||
.block_hash
|
||||
}
|
||||
|
||||
fn execution_status(&self, block_root: Hash256) -> ExecutionStatus {
|
||||
self.harness
|
||||
.chain
|
||||
.fork_choice
|
||||
.read()
|
||||
.get_block(&block_root)
|
||||
.unwrap()
|
||||
.execution_status
|
||||
}
|
||||
|
||||
fn fork_choice(&self) {
|
||||
self.harness.chain.fork_choice().unwrap();
|
||||
}
|
||||
|
||||
fn head_info(&self) -> HeadInfo {
|
||||
self.harness.chain.head_info().unwrap()
|
||||
}
|
||||
|
||||
fn move_to_terminal_block(&self) {
|
||||
let mock_execution_layer = self.harness.mock_execution_layer.as_ref().unwrap();
|
||||
mock_execution_layer
|
||||
.server
|
||||
.execution_block_generator()
|
||||
.move_to_terminal_block()
|
||||
.unwrap();
|
||||
}
|
||||
|
||||
fn build_blocks(&mut self, num_blocks: u64, is_valid: Payload) -> Vec<Hash256> {
|
||||
(0..num_blocks)
|
||||
.map(|_| self.import_block(is_valid.clone()))
|
||||
.collect()
|
||||
}
|
||||
|
||||
fn move_to_first_justification(&mut self, is_valid: Payload) {
|
||||
let slots_till_justification = E::slots_per_epoch() * 3;
|
||||
self.build_blocks(slots_till_justification, is_valid);
|
||||
|
||||
let justified_checkpoint = self.head_info().current_justified_checkpoint;
|
||||
assert_eq!(justified_checkpoint.epoch, 2);
|
||||
}
|
||||
|
||||
fn import_block(&mut self, is_valid: Payload) -> Hash256 {
|
||||
self.import_block_parametric(is_valid, |error| {
|
||||
matches!(
|
||||
error,
|
||||
BlockError::ExecutionPayloadError(
|
||||
ExecutionPayloadError::RejectedByExecutionEngine { .. }
|
||||
)
|
||||
)
|
||||
})
|
||||
}
|
||||
|
||||
fn block_root_at_slot(&self, slot: Slot) -> Option<Hash256> {
|
||||
self.harness
|
||||
.chain
|
||||
.block_root_at_slot(slot, WhenSlotSkipped::None)
|
||||
.unwrap()
|
||||
}
|
||||
|
||||
fn import_block_parametric<F: Fn(&BlockError<E>) -> bool>(
|
||||
&mut self,
|
||||
is_valid: Payload,
|
||||
evaluate_error: F,
|
||||
) -> Hash256 {
|
||||
let mock_execution_layer = self.harness.mock_execution_layer.as_ref().unwrap();
|
||||
|
||||
let head = self.harness.chain.head().unwrap();
|
||||
let state = head.beacon_state;
|
||||
let slot = state.slot() + 1;
|
||||
let (block, post_state) = self.harness.make_block(state, slot);
|
||||
let block_root = block.canonical_root();
|
||||
|
||||
match is_valid {
|
||||
Payload::Valid | Payload::Syncing => {
|
||||
if is_valid == Payload::Syncing {
|
||||
// Importing a payload whilst returning `SYNCING` simulates an EE that obtains
|
||||
// the block via it's own means (e.g., devp2p).
|
||||
let should_import_payload = true;
|
||||
mock_execution_layer
|
||||
.server
|
||||
.all_payloads_syncing(should_import_payload);
|
||||
} else {
|
||||
mock_execution_layer.server.full_payload_verification();
|
||||
}
|
||||
let root = self.harness.process_block(slot, block.clone()).unwrap();
|
||||
|
||||
if self.enable_attestations {
|
||||
let all_validators: Vec<usize> = (0..VALIDATOR_COUNT).collect();
|
||||
self.harness.attest_block(
|
||||
&post_state,
|
||||
block.state_root(),
|
||||
block_root.into(),
|
||||
&block,
|
||||
&all_validators,
|
||||
);
|
||||
}
|
||||
|
||||
let execution_status = self.execution_status(root.into());
|
||||
|
||||
match is_valid {
|
||||
Payload::Syncing => assert!(execution_status.is_not_verified()),
|
||||
Payload::Valid => assert!(execution_status.is_valid()),
|
||||
Payload::Invalid { .. } => unreachable!(),
|
||||
}
|
||||
|
||||
assert_eq!(
|
||||
self.harness.chain.get_block(&block_root).unwrap().unwrap(),
|
||||
block,
|
||||
"block from db must match block imported"
|
||||
);
|
||||
}
|
||||
Payload::Invalid { latest_valid_hash } => {
|
||||
let latest_valid_hash = latest_valid_hash
|
||||
.unwrap_or_else(|| self.block_hash(block.message().parent_root()));
|
||||
|
||||
mock_execution_layer
|
||||
.server
|
||||
.all_payloads_invalid(latest_valid_hash);
|
||||
|
||||
match self.harness.process_block(slot, block) {
|
||||
Err(error) if evaluate_error(&error) => (),
|
||||
Err(other) => {
|
||||
panic!("evaluate_error returned false with {:?}", other)
|
||||
}
|
||||
Ok(_) => panic!("block with invalid payload was imported"),
|
||||
};
|
||||
|
||||
assert!(
|
||||
self.harness
|
||||
.chain
|
||||
.fork_choice
|
||||
.read()
|
||||
.get_block(&block_root)
|
||||
.is_none(),
|
||||
"invalid block must not exist in fork choice"
|
||||
);
|
||||
assert!(
|
||||
self.harness.chain.get_block(&block_root).unwrap().is_none(),
|
||||
"invalid block cannot be accessed via get_block"
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
block_root
|
||||
}
|
||||
}
|
||||
|
||||
/// Simple test of the different import types.
|
||||
#[test]
|
||||
fn valid_invalid_syncing() {
|
||||
let mut rig = InvalidPayloadRig::new();
|
||||
rig.move_to_terminal_block();
|
||||
|
||||
rig.import_block(Payload::Valid);
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: None,
|
||||
});
|
||||
rig.import_block(Payload::Syncing);
|
||||
}
|
||||
|
||||
/// Ensure that an invalid payload can invalidate its parent too (given the right
|
||||
/// `latest_valid_hash`.
|
||||
#[test]
|
||||
fn invalid_payload_invalidates_parent() {
|
||||
let mut rig = InvalidPayloadRig::new();
|
||||
rig.move_to_terminal_block();
|
||||
|
||||
let roots = vec![
|
||||
rig.import_block(Payload::Syncing),
|
||||
rig.import_block(Payload::Syncing),
|
||||
rig.import_block(Payload::Syncing),
|
||||
];
|
||||
|
||||
let latest_valid_hash = rig.block_hash(roots[0]);
|
||||
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(latest_valid_hash),
|
||||
});
|
||||
|
||||
assert!(rig.execution_status(roots[0]).is_valid());
|
||||
assert!(rig.execution_status(roots[1]).is_invalid());
|
||||
assert!(rig.execution_status(roots[2]).is_invalid());
|
||||
|
||||
assert_eq!(rig.head_info().block_root, roots[0]);
|
||||
}
|
||||
|
||||
/// Ensure the client tries to exit when the justified checkpoint is invalidated.
|
||||
#[test]
|
||||
fn justified_checkpoint_becomes_invalid() {
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
rig.move_to_first_justification(Payload::Syncing);
|
||||
|
||||
let justified_checkpoint = rig.head_info().current_justified_checkpoint;
|
||||
let parent_root_of_justified = rig
|
||||
.harness
|
||||
.chain
|
||||
.get_block(&justified_checkpoint.root)
|
||||
.unwrap()
|
||||
.unwrap()
|
||||
.parent_root();
|
||||
let parent_hash_of_justified = rig.block_hash(parent_root_of_justified);
|
||||
|
||||
// No service should have triggered a shutdown, yet.
|
||||
assert!(rig.harness.shutdown_reasons().is_empty());
|
||||
|
||||
// Import a block that will invalidate the justified checkpoint.
|
||||
rig.import_block_parametric(
|
||||
Payload::Invalid {
|
||||
latest_valid_hash: Some(parent_hash_of_justified),
|
||||
},
|
||||
|error| {
|
||||
matches!(
|
||||
error,
|
||||
// The block import should fail since the beacon chain knows the justified payload
|
||||
// is invalid.
|
||||
BlockError::BeaconChainError(BeaconChainError::JustifiedPayloadInvalid { .. })
|
||||
)
|
||||
},
|
||||
);
|
||||
|
||||
// The beacon chain should have triggered a shutdown.
|
||||
assert_eq!(
|
||||
rig.harness.shutdown_reasons(),
|
||||
vec![ShutdownReason::Failure(
|
||||
INVALID_JUSTIFIED_PAYLOAD_SHUTDOWN_REASON
|
||||
)]
|
||||
);
|
||||
}
|
||||
|
||||
/// Ensure that a `latest_valid_hash` for a pre-finality block only reverts a single block.
|
||||
#[test]
|
||||
fn pre_finalized_latest_valid_hash() {
|
||||
let num_blocks = E::slots_per_epoch() * 4;
|
||||
let finalized_epoch = 2;
|
||||
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
let blocks = rig.build_blocks(num_blocks, Payload::Syncing);
|
||||
|
||||
assert_eq!(rig.head_info().finalized_checkpoint.epoch, finalized_epoch);
|
||||
|
||||
let pre_finalized_block_root = rig.block_root_at_slot(Slot::new(1)).unwrap();
|
||||
let pre_finalized_block_hash = rig.block_hash(pre_finalized_block_root);
|
||||
|
||||
// No service should have triggered a shutdown, yet.
|
||||
assert!(rig.harness.shutdown_reasons().is_empty());
|
||||
|
||||
// Import a pre-finalized block.
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(pre_finalized_block_hash),
|
||||
});
|
||||
|
||||
// The latest imported block should be the head.
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
// The beacon chain should *not* have triggered a shutdown.
|
||||
assert_eq!(rig.harness.shutdown_reasons(), vec![]);
|
||||
|
||||
// All blocks should still be unverified.
|
||||
for i in E::slots_per_epoch() * finalized_epoch..num_blocks {
|
||||
let slot = Slot::new(i);
|
||||
let root = rig.block_root_at_slot(slot).unwrap();
|
||||
assert!(rig.execution_status(root).is_not_verified());
|
||||
}
|
||||
}
|
||||
|
||||
/// Ensure that a `latest_valid_hash` will:
|
||||
///
|
||||
/// - Invalidate descendants of `latest_valid_root`.
|
||||
/// - Validate `latest_valid_root` and its ancestors.
|
||||
#[test]
|
||||
fn latest_valid_hash_will_validate() {
|
||||
const LATEST_VALID_SLOT: u64 = 3;
|
||||
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
let blocks = rig.build_blocks(4, Payload::Syncing);
|
||||
|
||||
let latest_valid_root = rig
|
||||
.block_root_at_slot(Slot::new(LATEST_VALID_SLOT))
|
||||
.unwrap();
|
||||
let latest_valid_hash = rig.block_hash(latest_valid_root);
|
||||
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(latest_valid_hash),
|
||||
});
|
||||
|
||||
assert_eq!(rig.head_info().slot, LATEST_VALID_SLOT);
|
||||
|
||||
for slot in 0..=4 {
|
||||
let slot = Slot::new(slot);
|
||||
let root = if slot > 0 {
|
||||
// If not the genesis slot, check the blocks we just produced.
|
||||
blocks[slot.as_usize() - 1]
|
||||
} else {
|
||||
// Genesis slot
|
||||
rig.block_root_at_slot(slot).unwrap()
|
||||
};
|
||||
let execution_status = rig.execution_status(root);
|
||||
|
||||
if slot > LATEST_VALID_SLOT {
|
||||
assert!(execution_status.is_invalid())
|
||||
} else if slot == 0 {
|
||||
assert!(execution_status.is_irrelevant())
|
||||
} else {
|
||||
assert!(execution_status.is_valid())
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Check behaviour when the `latest_valid_hash` is a junk value.
|
||||
#[test]
|
||||
fn latest_valid_hash_is_junk() {
|
||||
let num_blocks = E::slots_per_epoch() * 5;
|
||||
let finalized_epoch = 3;
|
||||
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
let blocks = rig.build_blocks(num_blocks, Payload::Syncing);
|
||||
|
||||
assert_eq!(rig.head_info().finalized_checkpoint.epoch, finalized_epoch);
|
||||
|
||||
// No service should have triggered a shutdown, yet.
|
||||
assert!(rig.harness.shutdown_reasons().is_empty());
|
||||
|
||||
let junk_hash = ExecutionBlockHash::repeat_byte(42);
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(junk_hash),
|
||||
});
|
||||
|
||||
// The latest imported block should be the head.
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
// The beacon chain should *not* have triggered a shutdown.
|
||||
assert_eq!(rig.harness.shutdown_reasons(), vec![]);
|
||||
|
||||
// All blocks should still be unverified.
|
||||
for i in E::slots_per_epoch() * finalized_epoch..num_blocks {
|
||||
let slot = Slot::new(i);
|
||||
let root = rig.block_root_at_slot(slot).unwrap();
|
||||
assert!(rig.execution_status(root).is_not_verified());
|
||||
}
|
||||
}
|
||||
|
||||
/// Check that descendants of invalid blocks are also invalidated.
|
||||
#[test]
|
||||
fn invalidates_all_descendants() {
|
||||
let num_blocks = E::slots_per_epoch() * 4 + E::slots_per_epoch() / 2;
|
||||
let finalized_epoch = 2;
|
||||
let finalized_slot = E::slots_per_epoch() * 2;
|
||||
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
let blocks = rig.build_blocks(num_blocks, Payload::Syncing);
|
||||
|
||||
assert_eq!(rig.head_info().finalized_checkpoint.epoch, finalized_epoch);
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
// Apply a block which conflicts with the canonical chain.
|
||||
let fork_slot = Slot::new(4 * E::slots_per_epoch() + 3);
|
||||
let fork_parent_slot = fork_slot - 1;
|
||||
let fork_parent_state = rig
|
||||
.harness
|
||||
.chain
|
||||
.state_at_slot(fork_parent_slot, StateSkipConfig::WithStateRoots)
|
||||
.unwrap();
|
||||
assert_eq!(fork_parent_state.slot(), fork_parent_slot);
|
||||
let (fork_block, _fork_post_state) = rig.harness.make_block(fork_parent_state, fork_slot);
|
||||
let fork_block_root = rig.harness.chain.process_block(fork_block).unwrap();
|
||||
rig.fork_choice();
|
||||
|
||||
// The latest valid hash will be set to the grandparent of the fork block. This means that the
|
||||
// parent of the fork block will become invalid.
|
||||
let latest_valid_slot = fork_parent_slot - 1;
|
||||
let latest_valid_root = rig
|
||||
.harness
|
||||
.chain
|
||||
.block_root_at_slot(latest_valid_slot, WhenSlotSkipped::None)
|
||||
.unwrap()
|
||||
.unwrap();
|
||||
assert!(blocks.contains(&latest_valid_root));
|
||||
let latest_valid_hash = rig.block_hash(latest_valid_root);
|
||||
|
||||
// The new block should not become the head, the old head should remain.
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(latest_valid_hash),
|
||||
});
|
||||
|
||||
// The block before the fork should become the head.
|
||||
assert_eq!(rig.head_info().block_root, latest_valid_root);
|
||||
|
||||
// The fork block should be invalidated, even though it's not an ancestor of the block that
|
||||
// triggered the INVALID response from the EL.
|
||||
assert!(rig.execution_status(fork_block_root).is_invalid());
|
||||
|
||||
for root in blocks {
|
||||
let slot = rig.harness.chain.get_block(&root).unwrap().unwrap().slot();
|
||||
|
||||
// Fork choice doesn't have info about pre-finalization, nothing to check here.
|
||||
if slot < finalized_slot {
|
||||
continue;
|
||||
}
|
||||
|
||||
let execution_status = rig.execution_status(root);
|
||||
if slot <= latest_valid_slot {
|
||||
// Blocks prior to the latest valid hash are valid.
|
||||
assert!(execution_status.is_valid());
|
||||
} else {
|
||||
// Blocks after the latest valid hash are invalid.
|
||||
assert!(execution_status.is_invalid());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Check that the head will switch after the canonical branch is invalidated.
|
||||
#[test]
|
||||
fn switches_heads() {
|
||||
let num_blocks = E::slots_per_epoch() * 4 + E::slots_per_epoch() / 2;
|
||||
let finalized_epoch = 2;
|
||||
let finalized_slot = E::slots_per_epoch() * 2;
|
||||
|
||||
let mut rig = InvalidPayloadRig::new().enable_attestations();
|
||||
rig.move_to_terminal_block();
|
||||
let blocks = rig.build_blocks(num_blocks, Payload::Syncing);
|
||||
|
||||
assert_eq!(rig.head_info().finalized_checkpoint.epoch, finalized_epoch);
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
// Apply a block which conflicts with the canonical chain.
|
||||
let fork_slot = Slot::new(4 * E::slots_per_epoch() + 3);
|
||||
let fork_parent_slot = fork_slot - 1;
|
||||
let fork_parent_state = rig
|
||||
.harness
|
||||
.chain
|
||||
.state_at_slot(fork_parent_slot, StateSkipConfig::WithStateRoots)
|
||||
.unwrap();
|
||||
assert_eq!(fork_parent_state.slot(), fork_parent_slot);
|
||||
let (fork_block, _fork_post_state) = rig.harness.make_block(fork_parent_state, fork_slot);
|
||||
let fork_parent_root = fork_block.parent_root();
|
||||
let fork_block_root = rig.harness.chain.process_block(fork_block).unwrap();
|
||||
rig.fork_choice();
|
||||
|
||||
let latest_valid_slot = fork_parent_slot;
|
||||
let latest_valid_hash = rig.block_hash(fork_parent_root);
|
||||
|
||||
// The new block should not become the head, the old head should remain.
|
||||
assert_eq!(rig.head_info().block_root, *blocks.last().unwrap());
|
||||
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(latest_valid_hash),
|
||||
});
|
||||
|
||||
// The fork block should become the head.
|
||||
assert_eq!(rig.head_info().block_root, fork_block_root);
|
||||
|
||||
// The fork block has not yet been validated.
|
||||
assert!(rig.execution_status(fork_block_root).is_not_verified());
|
||||
|
||||
for root in blocks {
|
||||
let slot = rig.harness.chain.get_block(&root).unwrap().unwrap().slot();
|
||||
|
||||
// Fork choice doesn't have info about pre-finalization, nothing to check here.
|
||||
if slot < finalized_slot {
|
||||
continue;
|
||||
}
|
||||
|
||||
let execution_status = rig.execution_status(root);
|
||||
if slot <= latest_valid_slot {
|
||||
// Blocks prior to the latest valid hash are valid.
|
||||
assert!(execution_status.is_valid());
|
||||
} else {
|
||||
// Blocks after the latest valid hash are invalid.
|
||||
assert!(execution_status.is_invalid());
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn invalid_during_processing() {
|
||||
let mut rig = InvalidPayloadRig::new();
|
||||
rig.move_to_terminal_block();
|
||||
|
||||
let roots = &[
|
||||
rig.import_block(Payload::Valid),
|
||||
rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: None,
|
||||
}),
|
||||
rig.import_block(Payload::Valid),
|
||||
];
|
||||
|
||||
// 0 should be present in the chain.
|
||||
assert!(rig.harness.chain.get_block(&roots[0]).unwrap().is_some());
|
||||
// 1 should *not* be present in the chain.
|
||||
assert_eq!(rig.harness.chain.get_block(&roots[1]).unwrap(), None);
|
||||
// 2 should be the head.
|
||||
let head = rig.harness.chain.head_info().unwrap();
|
||||
assert_eq!(head.block_root, roots[2]);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn invalid_after_optimistic_sync() {
|
||||
let mut rig = InvalidPayloadRig::new();
|
||||
rig.move_to_terminal_block();
|
||||
|
||||
let mut roots = vec![
|
||||
rig.import_block(Payload::Syncing),
|
||||
rig.import_block(Payload::Syncing),
|
||||
rig.import_block(Payload::Syncing),
|
||||
];
|
||||
|
||||
for root in &roots {
|
||||
assert!(rig.harness.chain.get_block(root).unwrap().is_some());
|
||||
}
|
||||
|
||||
// 2 should be the head.
|
||||
let head = rig.harness.chain.head_info().unwrap();
|
||||
assert_eq!(head.block_root, roots[2]);
|
||||
|
||||
roots.push(rig.import_block(Payload::Invalid {
|
||||
latest_valid_hash: Some(rig.block_hash(roots[1])),
|
||||
}));
|
||||
|
||||
// Running fork choice is necessary since a block has been invalidated.
|
||||
rig.fork_choice();
|
||||
|
||||
// 1 should be the head, since 2 was invalidated.
|
||||
let head = rig.harness.chain.head_info().unwrap();
|
||||
assert_eq!(head.block_root, roots[1]);
|
||||
}
|
||||
@@ -65,6 +65,7 @@ fn get_harness(
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_disk_store(store)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
harness.advance_slot();
|
||||
harness
|
||||
@@ -542,6 +543,7 @@ fn delete_blocks_and_states() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_disk_store(store.clone())
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
let unforked_blocks: u64 = 4 * E::slots_per_epoch();
|
||||
@@ -668,6 +670,7 @@ fn multi_epoch_fork_valid_blocks_test(
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_disk_store(store)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
let num_fork1_blocks: u64 = num_fork1_blocks_.try_into().unwrap();
|
||||
@@ -962,6 +965,7 @@ fn prunes_abandoned_fork_between_two_finalized_checkpoints() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let slots_per_epoch = rig.slots_per_epoch();
|
||||
let (mut state, state_root) = rig.get_current_state_and_root();
|
||||
@@ -1071,6 +1075,7 @@ fn pruning_does_not_touch_abandoned_block_shared_with_canonical_chain() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let slots_per_epoch = rig.slots_per_epoch();
|
||||
let (state, state_root) = rig.get_current_state_and_root();
|
||||
@@ -1200,6 +1205,7 @@ fn pruning_does_not_touch_blocks_prior_to_finalization() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let slots_per_epoch = rig.slots_per_epoch();
|
||||
let (mut state, state_root) = rig.get_current_state_and_root();
|
||||
@@ -1294,6 +1300,7 @@ fn prunes_fork_growing_past_youngest_finalized_checkpoint() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let (state, state_root) = rig.get_current_state_and_root();
|
||||
|
||||
@@ -1436,6 +1443,7 @@ fn prunes_skipped_slots_states() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let (state, state_root) = rig.get_current_state_and_root();
|
||||
|
||||
@@ -1559,6 +1567,7 @@ fn finalizes_non_epoch_start_slot() {
|
||||
.default_spec()
|
||||
.keypairs(validators_keypairs)
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
let (state, state_root) = rig.get_current_state_and_root();
|
||||
|
||||
@@ -2128,6 +2137,7 @@ fn finalizes_after_resuming_from_db() {
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_disk_store(store.clone())
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
harness.advance_slot();
|
||||
@@ -2171,6 +2181,7 @@ fn finalizes_after_resuming_from_db() {
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.resumed_disk_store(store)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
assert_chains_pretty_much_the_same(&original_chain, &resumed_harness.chain);
|
||||
@@ -2242,6 +2253,7 @@ fn revert_minority_fork_on_resume() {
|
||||
.spec(spec1)
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_disk_store(store1)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
// Chain with fork epoch configured.
|
||||
@@ -2251,6 +2263,7 @@ fn revert_minority_fork_on_resume() {
|
||||
.spec(spec2.clone())
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_disk_store(store2)
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
// Apply the same blocks to both chains initially.
|
||||
@@ -2346,6 +2359,7 @@ fn revert_minority_fork_on_resume() {
|
||||
.set_slot(end_slot.as_u64());
|
||||
builder
|
||||
}))
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
// Head should now be just before the fork.
|
||||
|
||||
@@ -30,6 +30,7 @@ fn get_harness(validator_count: usize) -> BeaconChainHarness<EphemeralHarnessTyp
|
||||
.spec(spec)
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
harness.advance_slot();
|
||||
@@ -179,7 +180,7 @@ fn aggregated_gossip_verification() {
|
||||
get_valid_sync_contribution(&harness, RelativeSyncCommittee::Current);
|
||||
|
||||
macro_rules! assert_invalid {
|
||||
($desc: tt, $attn_getter: expr, $($error: pat) |+ $( if $guard: expr )?) => {
|
||||
($desc: tt, $attn_getter: expr, $($error: pat_param) |+ $( if $guard: expr )?) => {
|
||||
assert!(
|
||||
matches!(
|
||||
harness
|
||||
@@ -505,7 +506,7 @@ fn unaggregated_gossip_verification() {
|
||||
get_valid_sync_committee_message(&harness, current_slot, RelativeSyncCommittee::Current);
|
||||
|
||||
macro_rules! assert_invalid {
|
||||
($desc: tt, $attn_getter: expr, $subnet_getter: expr, $($error: pat) |+ $( if $guard: expr )?) => {
|
||||
($desc: tt, $attn_getter: expr, $subnet_getter: expr, $($error: pat_param) |+ $( if $guard: expr )?) => {
|
||||
assert!(
|
||||
matches!(
|
||||
harness
|
||||
|
||||
@@ -28,6 +28,7 @@ fn get_harness(validator_count: usize) -> BeaconChainHarness<EphemeralHarnessTyp
|
||||
.default_spec()
|
||||
.keypairs(KEYPAIRS[0..validator_count].to_vec())
|
||||
.fresh_ephemeral_store()
|
||||
.mock_execution_layer()
|
||||
.build();
|
||||
|
||||
harness.advance_slot();
|
||||
|
||||
Reference in New Issue
Block a user