mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-16 03:12:41 +00:00
* rebase and add comment * conditional test * test * optimistic chould be working now * finality should be working now * try again * try again * clippy fix * add lc bootstrap beacon api * add lc optimistic/finality update to events * fmt * That error isn't occuring on my computer but I think this should fix it * Merge branch 'unstable' into light_client_beacon_api_1 # Conflicts: # beacon_node/beacon_chain/src/events.rs # beacon_node/http_api/src/lib.rs # beacon_node/http_api/src/test_utils.rs # beacon_node/http_api/tests/main.rs # beacon_node/lighthouse_network/src/rpc/codec/ssz_snappy.rs # beacon_node/lighthouse_network/src/rpc/methods.rs # beacon_node/lighthouse_network/src/service/api_types.rs # beacon_node/network/src/beacon_processor/worker/rpc_methods.rs # beacon_node/tests/test.rs # common/eth2/src/types.rs # lighthouse/src/main.rs * Add missing test file * Update light client types to comply with Altair light client spec. * Fix test compilation * Merge branch 'unstable' into light_client_beacon_api_1 * Support deserializing light client structures for the Bellatrix fork * Move `get_light_client_bootstrap` logic to `BeaconChain`. `LightClientBootstrap` API to return `ForkVersionedResponse`. * Misc fixes. - log cleanup - move http_api config mutation to `config::get_config` for consistency - fix light client API responses * Add light client bootstrap API test and fix existing ones. * Merge branch 'unstable' into light_client_beacon_api_1 * Fix test for `light-client-server` http api config. * Appease clippy * Add Altair light client SSZ tests * Merge branch 'unstable' of https://github.com/sigp/lighthouse into light_client_beacon_api_1 * updates to light client header * light client header from signed beacon block * using options * implement helper functions * placeholder conversion from vec hash256 to exec branch * add deneb * using fixed vector * remove unwraps * by epoch * compute merkle proof * merkle proof * update comments * resolve merge conflicts * linting * Merge branch 'unstable' into light-client-ssz-tests # Conflicts: # beacon_node/beacon_chain/src/beacon_chain.rs # consensus/types/src/light_client_bootstrap.rs # consensus/types/src/light_client_header.rs * superstruct attempt * superstruct changes * lint * altair * update * update * changes to light_client_optimistic_ and finality * merge unstable * refactor * resolved merge conflicts * Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types * block_to_light_client_header fork aware * fmt * comment fix * comment fix * include merge fork, update deserialize_by_fork, refactor * fmt * pass by ref to prevent clone * rename merkle proof fn * add FIXME * LightClientHeader TestRandom * fix comments * fork version deserialize * merge unstable * move fn arguments, fork name calc * use task executor * remove unneeded fns * remove dead code * add manual ssz decoding/encoding and add ssz_tests_by_fork macro * merge deneb types with tests * merge ssz tests, revert code deletion, cleanup * move chainspec * update ssz tests * fmt * light client ssz tests * change to superstruct * changes from feedback * linting * Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types * test fix * cleanup * Remove unused `derive`. * Merge branch 'unstable' of https://github.com/sigp/lighthouse into capella_deneb_light_client_types * beta compiler fix * merge
336 lines
12 KiB
Rust
336 lines
12 KiB
Rust
use crate::local_network::LocalNetwork;
|
|
use node_test_rig::eth2::types::{BlockId, FinalityCheckpointsData, StateId};
|
|
use std::time::Duration;
|
|
use types::{Epoch, EthSpec, ExecPayload, ExecutionBlockHash, Hash256, Slot, Unsigned};
|
|
|
|
/// Checks that all of the validators have on-boarded by the start of the second eth1 voting
|
|
/// period.
|
|
pub async fn verify_initial_validator_count<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
slot_duration: Duration,
|
|
initial_validator_count: usize,
|
|
) -> Result<(), String> {
|
|
slot_delay(Slot::new(1), slot_duration).await;
|
|
verify_validator_count(network, initial_validator_count).await?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Checks that all of the validators have on-boarded by the start of the second eth1 voting
|
|
/// period.
|
|
pub async fn verify_validator_onboarding<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
slot_duration: Duration,
|
|
expected_validator_count: usize,
|
|
) -> Result<(), String> {
|
|
slot_delay(
|
|
Slot::new(E::SlotsPerEth1VotingPeriod::to_u64()),
|
|
slot_duration,
|
|
)
|
|
.await;
|
|
verify_validator_count(network, expected_validator_count).await?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Checks that the chain has made the first possible finalization.
|
|
///
|
|
/// Intended to be run as soon as chain starts.
|
|
pub async fn verify_first_finalization<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
slot_duration: Duration,
|
|
) -> Result<(), String> {
|
|
epoch_delay(Epoch::new(4), slot_duration, E::slots_per_epoch()).await;
|
|
verify_all_finalized_at(network, Epoch::new(2)).await?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Delays for `epochs`, plus half a slot extra.
|
|
pub async fn epoch_delay(epochs: Epoch, slot_duration: Duration, slots_per_epoch: u64) {
|
|
let duration = slot_duration * (epochs.as_u64() * slots_per_epoch) as u32 + slot_duration / 2;
|
|
tokio::time::sleep(duration).await
|
|
}
|
|
|
|
/// Delays for `slots`, plus half a slot extra.
|
|
async fn slot_delay(slots: Slot, slot_duration: Duration) {
|
|
let duration = slot_duration * slots.as_u64() as u32 + slot_duration / 2;
|
|
tokio::time::sleep(duration).await;
|
|
}
|
|
|
|
/// Verifies that all beacon nodes in the given network have a head state that has a finalized
|
|
/// epoch of `epoch`.
|
|
pub async fn verify_all_finalized_at<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
epoch: Epoch,
|
|
) -> Result<(), String> {
|
|
let epochs = {
|
|
let mut epochs = Vec::new();
|
|
for remote_node in network.remote_nodes()? {
|
|
epochs.push(
|
|
remote_node
|
|
.get_beacon_states_finality_checkpoints(StateId::Head)
|
|
.await
|
|
.map(|body| body.unwrap().data.finalized.epoch)
|
|
.map_err(|e| format!("Get head via http failed: {:?}", e))?,
|
|
);
|
|
}
|
|
epochs
|
|
};
|
|
|
|
if epochs.iter().any(|node_epoch| *node_epoch != epoch) {
|
|
Err(format!(
|
|
"Nodes are not finalized at epoch {}. Finalized epochs: {:?}",
|
|
epoch, epochs
|
|
))
|
|
} else {
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
/// Verifies that all beacon nodes in the given `network` have a head state that contains
|
|
/// `expected_count` validators.
|
|
async fn verify_validator_count<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
expected_count: usize,
|
|
) -> Result<(), String> {
|
|
let validator_counts = {
|
|
let mut validator_counts = Vec::new();
|
|
for remote_node in network.remote_nodes()? {
|
|
let vc = remote_node
|
|
.get_debug_beacon_states::<E>(StateId::Head)
|
|
.await
|
|
.map(|body| body.unwrap().data)
|
|
.map_err(|e| format!("Get state root via http failed: {:?}", e))?
|
|
.validators()
|
|
.len();
|
|
validator_counts.push(vc);
|
|
}
|
|
validator_counts
|
|
};
|
|
|
|
if validator_counts
|
|
.iter()
|
|
.any(|count| *count != expected_count)
|
|
{
|
|
Err(format!(
|
|
"Nodes do not all have {} validators in their state. Validator counts: {:?}",
|
|
expected_count, validator_counts
|
|
))
|
|
} else {
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
/// Verifies that there's been a block produced at every slot up to and including `slot`.
|
|
pub async fn verify_full_block_production_up_to<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
slot: Slot,
|
|
slot_duration: Duration,
|
|
) -> Result<(), String> {
|
|
slot_delay(slot, slot_duration).await;
|
|
let beacon_nodes = network.beacon_nodes.read();
|
|
let beacon_chain = beacon_nodes[0].client.beacon_chain().unwrap();
|
|
let num_blocks = beacon_chain
|
|
.chain_dump()
|
|
.unwrap()
|
|
.iter()
|
|
.take_while(|s| s.beacon_block.slot() <= slot)
|
|
.count();
|
|
if num_blocks != slot.as_usize() + 1 {
|
|
return Err(format!(
|
|
"There wasn't a block produced at every slot, got: {}, expected: {}",
|
|
num_blocks,
|
|
slot.as_usize() + 1
|
|
));
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
/// Verify that all nodes have the correct fork version after the `fork_epoch`.
|
|
pub async fn verify_fork_version<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
fork_epoch: Epoch,
|
|
slot_duration: Duration,
|
|
fork_version: [u8; 4],
|
|
) -> Result<(), String> {
|
|
epoch_delay(fork_epoch, slot_duration, E::slots_per_epoch()).await;
|
|
for remote_node in network.remote_nodes()? {
|
|
let remote_fork_version = remote_node
|
|
.get_beacon_states_fork(StateId::Head)
|
|
.await
|
|
.map(|resp| resp.unwrap().data.current_version)
|
|
.map_err(|e| format!("Failed to get fork from beacon node: {:?}", e))?;
|
|
if fork_version != remote_fork_version {
|
|
return Err(format!(
|
|
"Fork version after FORK_EPOCH is incorrect, got: {:?}, expected: {:?}",
|
|
remote_fork_version, fork_version,
|
|
));
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
/// Verify that all sync aggregates from `sync_committee_start_slot` until `upto_slot`
|
|
/// have full aggregates.
|
|
pub async fn verify_full_sync_aggregates_up_to<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
sync_committee_start_slot: Slot,
|
|
upto_slot: Slot,
|
|
slot_duration: Duration,
|
|
) -> Result<(), String> {
|
|
slot_delay(upto_slot, slot_duration).await;
|
|
let remote_nodes = network.remote_nodes()?;
|
|
let remote_node = remote_nodes.first().unwrap();
|
|
|
|
for slot in sync_committee_start_slot.as_u64()..=upto_slot.as_u64() {
|
|
let sync_aggregate_count = remote_node
|
|
.get_beacon_blocks::<E>(BlockId::Slot(Slot::new(slot)))
|
|
.await
|
|
.map(|resp| {
|
|
resp.unwrap()
|
|
.data
|
|
.message()
|
|
.body()
|
|
.sync_aggregate()
|
|
.map(|agg| agg.num_set_bits())
|
|
})
|
|
.map_err(|e| format!("Error while getting beacon block: {:?}", e))?
|
|
.map_err(|_| format!("Altair block {} should have sync aggregate", slot))?;
|
|
|
|
if sync_aggregate_count != E::sync_committee_size() {
|
|
return Err(format!(
|
|
"Sync aggregate at slot {} was not full, got: {}, expected: {}",
|
|
slot,
|
|
sync_aggregate_count,
|
|
E::sync_committee_size()
|
|
));
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Verify that the first merged PoS block got finalized.
|
|
pub async fn verify_transition_block_finalized<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
transition_epoch: Epoch,
|
|
slot_duration: Duration,
|
|
should_verify: bool,
|
|
) -> Result<(), String> {
|
|
if !should_verify {
|
|
return Ok(());
|
|
}
|
|
epoch_delay(transition_epoch + 2, slot_duration, E::slots_per_epoch()).await;
|
|
let mut block_hashes = Vec::new();
|
|
for remote_node in network.remote_nodes()?.iter() {
|
|
let execution_block_hash: ExecutionBlockHash = remote_node
|
|
.get_beacon_blocks::<E>(BlockId::Finalized)
|
|
.await
|
|
.map(|body| body.unwrap().data)
|
|
.map_err(|e| format!("Get state root via http failed: {:?}", e))?
|
|
.message()
|
|
.execution_payload()
|
|
.map(|payload| payload.block_hash())
|
|
.map_err(|e| format!("Execution payload does not exist: {:?}", e))?;
|
|
block_hashes.push(execution_block_hash);
|
|
}
|
|
|
|
let first = block_hashes[0];
|
|
if first.into_root() != Hash256::zero() && block_hashes.iter().all(|&item| item == first) {
|
|
Ok(())
|
|
} else {
|
|
Err(format!(
|
|
"Terminal block not finalized on all nodes Finalized block hashes:{:?}",
|
|
block_hashes
|
|
))
|
|
}
|
|
}
|
|
|
|
pub(crate) async fn verify_light_client_updates<E: EthSpec>(
|
|
network: LocalNetwork<E>,
|
|
start_slot: Slot,
|
|
end_slot: Slot,
|
|
slot_duration: Duration,
|
|
) -> Result<(), String> {
|
|
slot_delay(start_slot, slot_duration).await;
|
|
|
|
// Tolerance of 2 slot allows for 1 single missed slot.
|
|
let light_client_update_slot_tolerance = Slot::new(2);
|
|
let remote_nodes = network.remote_nodes()?;
|
|
let client = remote_nodes.first().unwrap();
|
|
let mut have_seen_block = false;
|
|
let mut have_achieved_finality = false;
|
|
|
|
for slot in start_slot.as_u64()..=end_slot.as_u64() {
|
|
slot_delay(Slot::new(1), slot_duration).await;
|
|
let slot = Slot::new(slot);
|
|
let previous_slot = slot - 1;
|
|
|
|
let previous_slot_block = client
|
|
.get_beacon_blocks::<E>(BlockId::Slot(previous_slot))
|
|
.await
|
|
.map_err(|e| {
|
|
format!("Unable to get beacon block for previous slot {previous_slot:?}: {e:?}")
|
|
})?;
|
|
let previous_slot_has_block = previous_slot_block.is_some();
|
|
|
|
if !have_seen_block {
|
|
// Make sure we have seen the first block in Altair, to make sure we have sync aggregates available.
|
|
if previous_slot_has_block {
|
|
have_seen_block = true;
|
|
}
|
|
// Wait for another slot before we check the first update to avoid race condition.
|
|
continue;
|
|
}
|
|
|
|
// Make sure previous slot has a block, otherwise skip checking for the signature slot distance
|
|
if !previous_slot_has_block {
|
|
continue;
|
|
}
|
|
|
|
// Verify light client optimistic update. `signature_slot_distance` should be 1 in the ideal scenario.
|
|
let signature_slot = *client
|
|
.get_beacon_light_client_optimistic_update::<E>()
|
|
.await
|
|
.map_err(|e| format!("Error while getting light client updates: {:?}", e))?
|
|
.ok_or(format!("Light client optimistic update not found {slot:?}"))?
|
|
.data
|
|
.signature_slot();
|
|
let signature_slot_distance = slot - signature_slot;
|
|
if signature_slot_distance > light_client_update_slot_tolerance {
|
|
return Err(format!("Existing optimistic update too old: signature slot {signature_slot}, current slot {slot:?}"));
|
|
}
|
|
|
|
// Verify light client finality update. `signature_slot_distance` should be 1 in the ideal scenario.
|
|
// NOTE: Currently finality updates are produced as long as the finalized block is known, even if the finalized header
|
|
// sync committee period does not match the signature slot committee period.
|
|
// TODO: This complies with the current spec, but we should check if this is a bug.
|
|
if !have_achieved_finality {
|
|
let FinalityCheckpointsData { finalized, .. } = client
|
|
.get_beacon_states_finality_checkpoints(StateId::Head)
|
|
.await
|
|
.map_err(|e| format!("Unable to get beacon state finality checkpoint: {e:?}"))?
|
|
.ok_or("Unable to get head state".to_string())?
|
|
.data;
|
|
if !finalized.root.is_zero() {
|
|
// Wait for another slot before we check the first finality update to avoid race condition.
|
|
have_achieved_finality = true;
|
|
}
|
|
continue;
|
|
}
|
|
let signature_slot = *client
|
|
.get_beacon_light_client_finality_update::<E>()
|
|
.await
|
|
.map_err(|e| format!("Error while getting light client updates: {:?}", e))?
|
|
.ok_or(format!("Light client finality update not found {slot:?}"))?
|
|
.data
|
|
.signature_slot();
|
|
let signature_slot_distance = slot - signature_slot;
|
|
if signature_slot_distance > light_client_update_slot_tolerance {
|
|
return Err(format!(
|
|
"Existing finality update too old: signature slot {signature_slot}, current slot {slot:?}"
|
|
));
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|