mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-15 10:52:43 +00:00
524 lines
18 KiB
Rust
524 lines
18 KiB
Rust
use crate::blob_verification::{
|
|
verify_kzg_for_blob, verify_kzg_for_blob_list, GossipVerifiedBlob, KzgVerifiedBlob,
|
|
};
|
|
use crate::block_verification_types::{
|
|
AvailabilityPendingExecutedBlock, AvailableExecutedBlock, RpcBlock,
|
|
};
|
|
use crate::data_availability_checker::overflow_lru_cache::OverflowLRUCache;
|
|
use crate::{BeaconChain, BeaconChainTypes, BeaconStore};
|
|
use kzg::Error as KzgError;
|
|
use kzg::Kzg;
|
|
use slog::{debug, error};
|
|
use slot_clock::SlotClock;
|
|
use ssz_types::{Error, VariableList};
|
|
use std::collections::HashSet;
|
|
use std::fmt;
|
|
use std::fmt::Debug;
|
|
use std::sync::Arc;
|
|
use strum::IntoStaticStr;
|
|
use task_executor::TaskExecutor;
|
|
use types::blob_sidecar::{BlobIdentifier, BlobSidecar, FixedBlobSidecarList};
|
|
use types::consts::deneb::MIN_EPOCHS_FOR_BLOB_SIDECARS_REQUESTS;
|
|
use types::{BlobSidecarList, ChainSpec, Epoch, EthSpec, Hash256, SignedBeaconBlock, Slot};
|
|
|
|
mod overflow_lru_cache;
|
|
|
|
/// The LRU Cache stores `PendingComponents` which can store up to
|
|
/// `MAX_BLOBS_PER_BLOCK = 4` blobs each. A `BlobSidecar` is 0.131256 MB. So
|
|
/// the maximum size of a `PendingComponents` is ~ 0.525024 MB. Setting this
|
|
/// to 1024 means the maximum size of the cache is ~ 0.5 GB. But the cache
|
|
/// will target a size of less than 75% of capacity.
|
|
pub const OVERFLOW_LRU_CAPACITY: usize = 1024;
|
|
|
|
#[derive(Debug, IntoStaticStr)]
|
|
pub enum AvailabilityCheckError {
|
|
Kzg(KzgError),
|
|
KzgNotInitialized,
|
|
KzgVerificationFailed,
|
|
SszTypes(ssz_types::Error),
|
|
NumBlobsMismatch {
|
|
num_kzg_commitments: usize,
|
|
num_blobs: usize,
|
|
},
|
|
MissingBlobs,
|
|
TxKzgCommitmentMismatch(String),
|
|
KzgCommitmentMismatch {
|
|
blob_index: u64,
|
|
},
|
|
IncorrectFork,
|
|
BlobIndexInvalid(u64),
|
|
UnorderedBlobs {
|
|
blob_index: u64,
|
|
expected_index: u64,
|
|
},
|
|
StoreError(store::Error),
|
|
DecodeError(ssz::DecodeError),
|
|
BlockBlobRootMismatch {
|
|
block_root: Hash256,
|
|
blob_block_root: Hash256,
|
|
},
|
|
BlockBlobSlotMismatch {
|
|
block_slot: Slot,
|
|
blob_slot: Slot,
|
|
},
|
|
}
|
|
|
|
impl From<ssz_types::Error> for AvailabilityCheckError {
|
|
fn from(value: Error) -> Self {
|
|
Self::SszTypes(value)
|
|
}
|
|
}
|
|
|
|
impl From<store::Error> for AvailabilityCheckError {
|
|
fn from(value: store::Error) -> Self {
|
|
Self::StoreError(value)
|
|
}
|
|
}
|
|
|
|
impl From<ssz::DecodeError> for AvailabilityCheckError {
|
|
fn from(value: ssz::DecodeError) -> Self {
|
|
Self::DecodeError(value)
|
|
}
|
|
}
|
|
|
|
/// This cache contains
|
|
/// - blobs that have been gossip verified
|
|
/// - commitments for blocks that have been gossip verified, but the commitments themselves
|
|
/// have not been verified against blobs
|
|
/// - blocks that have been fully verified and only require a data availability check
|
|
pub struct DataAvailabilityChecker<T: BeaconChainTypes> {
|
|
availability_cache: Arc<OverflowLRUCache<T>>,
|
|
slot_clock: T::SlotClock,
|
|
kzg: Option<Arc<Kzg<<T::EthSpec as EthSpec>::Kzg>>>,
|
|
spec: ChainSpec,
|
|
}
|
|
|
|
/// This type is returned after adding a block / blob to the `DataAvailabilityChecker`.
|
|
///
|
|
/// Indicates if the block is fully `Available` or if we need blobs or blocks
|
|
/// to "complete" the requirements for an `AvailableBlock`.
|
|
#[derive(PartialEq)]
|
|
pub enum Availability<T: EthSpec> {
|
|
MissingComponents(Hash256),
|
|
Available(Box<AvailableExecutedBlock<T>>),
|
|
}
|
|
|
|
impl<T: EthSpec> Debug for Availability<T> {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
match self {
|
|
Self::MissingComponents(block_root) => {
|
|
write!(f, "MissingComponents({})", block_root)
|
|
}
|
|
Self::Available(block) => write!(f, "Available({:?})", block.import_data.block_root),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<T: EthSpec> Availability<T> {
|
|
/// Returns all the blob identifiers associated with an `AvailableBlock`.
|
|
/// Returns `None` if avaiability hasn't been fully satisfied yet.
|
|
pub fn get_available_blob_ids(&self) -> Option<Vec<BlobIdentifier>> {
|
|
if let Self::Available(block) = self {
|
|
Some(block.get_all_blob_ids())
|
|
} else {
|
|
None
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
|
|
pub fn new(
|
|
slot_clock: T::SlotClock,
|
|
kzg: Option<Arc<Kzg<<T::EthSpec as EthSpec>::Kzg>>>,
|
|
store: BeaconStore<T>,
|
|
spec: ChainSpec,
|
|
) -> Result<Self, AvailabilityCheckError> {
|
|
let overflow_cache = OverflowLRUCache::new(OVERFLOW_LRU_CAPACITY, store)?;
|
|
Ok(Self {
|
|
availability_cache: Arc::new(overflow_cache),
|
|
slot_clock,
|
|
kzg,
|
|
spec,
|
|
})
|
|
}
|
|
|
|
pub fn has_block(&self, block_root: &Hash256) -> bool {
|
|
self.availability_cache.has_block(block_root)
|
|
}
|
|
|
|
pub fn get_missing_blob_ids_checking_cache(
|
|
&self,
|
|
block_root: Hash256,
|
|
) -> Option<Vec<BlobIdentifier>> {
|
|
let (block, blob_indices) = self.availability_cache.get_missing_blob_info(block_root);
|
|
self.get_missing_blob_ids(block_root, block.as_ref(), Some(blob_indices))
|
|
}
|
|
|
|
/// A `None` indicates blobs are not required.
|
|
///
|
|
/// If there's no block, all possible ids will be returned that don't exist in the given blobs.
|
|
/// If there no blobs, all possible ids will be returned.
|
|
pub fn get_missing_blob_ids(
|
|
&self,
|
|
block_root: Hash256,
|
|
block_opt: Option<&Arc<SignedBeaconBlock<T::EthSpec>>>,
|
|
blobs_opt: Option<HashSet<usize>>,
|
|
) -> Option<Vec<BlobIdentifier>> {
|
|
let epoch = self.slot_clock.now()?.epoch(T::EthSpec::slots_per_epoch());
|
|
|
|
self.da_check_required(epoch).then(|| {
|
|
block_opt
|
|
.map(|block| {
|
|
block.get_filtered_blob_ids(Some(block_root), |i, _| {
|
|
blobs_opt.as_ref().map_or(true, |blobs| !blobs.contains(&i))
|
|
})
|
|
})
|
|
.unwrap_or_else(|| {
|
|
let mut blob_ids = Vec::with_capacity(T::EthSpec::max_blobs_per_block());
|
|
for i in 0..T::EthSpec::max_blobs_per_block() {
|
|
if blobs_opt.as_ref().map_or(true, |blobs| !blobs.contains(&i)) {
|
|
blob_ids.push(BlobIdentifier {
|
|
block_root,
|
|
index: i as u64,
|
|
});
|
|
}
|
|
}
|
|
blob_ids
|
|
})
|
|
})
|
|
}
|
|
|
|
/// Get a blob from the availability cache.
|
|
pub fn get_blob(
|
|
&self,
|
|
blob_id: &BlobIdentifier,
|
|
) -> Result<Option<Arc<BlobSidecar<T::EthSpec>>>, AvailabilityCheckError> {
|
|
self.availability_cache.peek_blob(blob_id)
|
|
}
|
|
|
|
pub fn put_rpc_blobs(
|
|
&self,
|
|
block_root: Hash256,
|
|
blobs: FixedBlobSidecarList<T::EthSpec>,
|
|
) -> Result<Availability<T::EthSpec>, AvailabilityCheckError> {
|
|
let mut verified_blobs = vec![];
|
|
if let Some(kzg) = self.kzg.as_ref() {
|
|
for blob in blobs.iter().flatten() {
|
|
verified_blobs.push(verify_kzg_for_blob(blob.clone(), kzg)?)
|
|
}
|
|
} else {
|
|
return Err(AvailabilityCheckError::KzgNotInitialized);
|
|
};
|
|
self.availability_cache
|
|
.put_kzg_verified_blobs(block_root, &verified_blobs)
|
|
}
|
|
|
|
/// This first validates the KZG commitments included in the blob sidecar.
|
|
/// Check if we've cached other blobs for this block. If it completes a set and we also
|
|
/// have a block cached, return the `Availability` variant triggering block import.
|
|
/// Otherwise cache the blob sidecar.
|
|
///
|
|
/// This should only accept gossip verified blobs, so we should not have to worry about dupes.
|
|
pub fn put_gossip_blob(
|
|
&self,
|
|
gossip_blob: GossipVerifiedBlob<T>,
|
|
) -> Result<Availability<T::EthSpec>, AvailabilityCheckError> {
|
|
// Verify the KZG commitments.
|
|
let kzg_verified_blob = if let Some(kzg) = self.kzg.as_ref() {
|
|
verify_kzg_for_blob(gossip_blob.to_blob(), kzg)?
|
|
} else {
|
|
return Err(AvailabilityCheckError::KzgNotInitialized);
|
|
};
|
|
|
|
self.availability_cache
|
|
.put_kzg_verified_blobs(kzg_verified_blob.block_root(), &[kzg_verified_blob])
|
|
}
|
|
|
|
/// Check if we have all the blobs for a block. If we do, return the Availability variant that
|
|
/// triggers import of the block.
|
|
pub fn put_pending_executed_block(
|
|
&self,
|
|
executed_block: AvailabilityPendingExecutedBlock<T::EthSpec>,
|
|
) -> Result<Availability<T::EthSpec>, AvailabilityCheckError> {
|
|
self.availability_cache
|
|
.put_pending_executed_block(executed_block)
|
|
}
|
|
|
|
/// Checks if a block is available, returns a `MaybeAvailableBlock` that may include the fully
|
|
/// available block.
|
|
pub fn check_rpc_block_availability(
|
|
&self,
|
|
block: RpcBlock<T::EthSpec>,
|
|
) -> Result<MaybeAvailableBlock<T::EthSpec>, AvailabilityCheckError> {
|
|
let (block, blobs) = block.deconstruct();
|
|
match blobs {
|
|
None => {
|
|
if self.blobs_required_for_block(&block) {
|
|
Ok(MaybeAvailableBlock::AvailabilityPending(block))
|
|
} else {
|
|
Ok(MaybeAvailableBlock::Available(AvailableBlock {
|
|
block,
|
|
blobs: None,
|
|
}))
|
|
}
|
|
}
|
|
Some(blob_list) => {
|
|
let verified_blobs = if self.blobs_required_for_block(&block) {
|
|
let kzg = self
|
|
.kzg
|
|
.as_ref()
|
|
.ok_or(AvailabilityCheckError::KzgNotInitialized)?;
|
|
verify_kzg_for_blob_list(&blob_list, kzg)?;
|
|
Some(blob_list)
|
|
} else {
|
|
None
|
|
};
|
|
Ok(MaybeAvailableBlock::Available(AvailableBlock {
|
|
block,
|
|
blobs: verified_blobs,
|
|
}))
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Determines the blob requirements for a block. Answers the question: "Does this block require
|
|
/// blobs?".
|
|
fn blobs_required_for_block(&self, block: &SignedBeaconBlock<T::EthSpec>) -> bool {
|
|
let block_within_da_period = self.da_check_required(block.epoch());
|
|
let block_has_kzg_commitments = block
|
|
.message()
|
|
.body()
|
|
.blob_kzg_commitments()
|
|
.map_or(false, |commitments| !commitments.is_empty());
|
|
block_within_da_period && block_has_kzg_commitments
|
|
}
|
|
|
|
/// The epoch at which we require a data availability check in block processing.
|
|
/// `None` if the `Deneb` fork is disabled.
|
|
pub fn data_availability_boundary(&self) -> Option<Epoch> {
|
|
self.spec.deneb_fork_epoch.and_then(|fork_epoch| {
|
|
self.slot_clock
|
|
.now()
|
|
.map(|slot| slot.epoch(T::EthSpec::slots_per_epoch()))
|
|
.map(|current_epoch| {
|
|
std::cmp::max(
|
|
fork_epoch,
|
|
current_epoch.saturating_sub(*MIN_EPOCHS_FOR_BLOB_SIDECARS_REQUESTS),
|
|
)
|
|
})
|
|
})
|
|
}
|
|
|
|
/// Returns true if the given epoch lies within the da boundary and false otherwise.
|
|
pub fn da_check_required(&self, block_epoch: Epoch) -> bool {
|
|
self.data_availability_boundary()
|
|
.map_or(false, |da_epoch| block_epoch >= da_epoch)
|
|
}
|
|
|
|
/// Persist all in memory components to disk
|
|
pub fn persist_all(&self) -> Result<(), AvailabilityCheckError> {
|
|
self.availability_cache.write_all_to_disk()
|
|
}
|
|
}
|
|
|
|
/// Verifies an `SignedBeaconBlock` against a set of KZG verified blobs.
|
|
/// This does not check whether a block *should* have blobs, these checks should have been
|
|
/// completed when producing the `AvailabilityPendingBlock`.
|
|
pub fn make_available<T: EthSpec>(
|
|
block: Arc<SignedBeaconBlock<T>>,
|
|
blobs: Vec<KzgVerifiedBlob<T>>,
|
|
) -> Result<AvailableBlock<T>, AvailabilityCheckError> {
|
|
let blobs = VariableList::new(blobs.into_iter().map(|blob| blob.to_blob()).collect())?;
|
|
|
|
consistency_checks(&block, &blobs)?;
|
|
|
|
Ok(AvailableBlock {
|
|
block,
|
|
blobs: Some(blobs),
|
|
})
|
|
}
|
|
|
|
/// Makes the following checks to ensure that the list of blobs correspond block:
|
|
///
|
|
/// * Check that a block is post-deneb
|
|
/// * Checks that the number of blobs is equal to the length of kzg commitments in the list
|
|
/// * Checks that the index, slot, root and kzg_commitment in the block match the blobs in the correct order
|
|
///
|
|
/// Returns `Ok(())` if all consistency checks pass and an error otherwise.
|
|
pub fn consistency_checks<T: EthSpec>(
|
|
block: &SignedBeaconBlock<T>,
|
|
blobs: &[Arc<BlobSidecar<T>>],
|
|
) -> Result<(), AvailabilityCheckError> {
|
|
let Ok(block_kzg_commitments) = block
|
|
.message()
|
|
.body()
|
|
.blob_kzg_commitments() else {
|
|
return Ok(())
|
|
};
|
|
|
|
if blobs.len() != block_kzg_commitments.len() {
|
|
return Err(AvailabilityCheckError::NumBlobsMismatch {
|
|
num_kzg_commitments: block_kzg_commitments.len(),
|
|
num_blobs: blobs.len(),
|
|
});
|
|
}
|
|
|
|
if block_kzg_commitments.is_empty() {
|
|
return Ok(());
|
|
}
|
|
|
|
let block_root = blobs
|
|
.first()
|
|
.map(|blob| blob.block_root)
|
|
.unwrap_or(block.canonical_root());
|
|
for (index, (block_commitment, blob)) in
|
|
block_kzg_commitments.iter().zip(blobs.iter()).enumerate()
|
|
{
|
|
let index = index as u64;
|
|
if index != blob.index {
|
|
return Err(AvailabilityCheckError::UnorderedBlobs {
|
|
blob_index: blob.index,
|
|
expected_index: index,
|
|
});
|
|
}
|
|
if block_root != blob.block_root {
|
|
return Err(AvailabilityCheckError::BlockBlobRootMismatch {
|
|
block_root,
|
|
blob_block_root: blob.block_root,
|
|
});
|
|
}
|
|
if block.slot() != blob.slot {
|
|
return Err(AvailabilityCheckError::BlockBlobSlotMismatch {
|
|
block_slot: block.slot(),
|
|
blob_slot: blob.slot,
|
|
});
|
|
}
|
|
if *block_commitment != blob.kzg_commitment {
|
|
return Err(AvailabilityCheckError::KzgCommitmentMismatch {
|
|
blob_index: blob.index,
|
|
});
|
|
}
|
|
}
|
|
Ok(())
|
|
}
|
|
|
|
pub fn start_availability_cache_maintenance_service<T: BeaconChainTypes>(
|
|
executor: TaskExecutor,
|
|
chain: Arc<BeaconChain<T>>,
|
|
) {
|
|
// this cache only needs to be maintained if deneb is configured
|
|
if chain.spec.deneb_fork_epoch.is_some() {
|
|
let overflow_cache = chain.data_availability_checker.availability_cache.clone();
|
|
executor.spawn(
|
|
async move { availability_cache_maintenance_service(chain, overflow_cache).await },
|
|
"availability_cache_service",
|
|
);
|
|
} else {
|
|
debug!(
|
|
chain.log,
|
|
"Deneb fork not configured, not starting availability cache maintenance service"
|
|
);
|
|
}
|
|
}
|
|
|
|
async fn availability_cache_maintenance_service<T: BeaconChainTypes>(
|
|
chain: Arc<BeaconChain<T>>,
|
|
overflow_cache: Arc<OverflowLRUCache<T>>,
|
|
) {
|
|
let epoch_duration = chain.slot_clock.slot_duration() * T::EthSpec::slots_per_epoch() as u32;
|
|
loop {
|
|
match chain
|
|
.slot_clock
|
|
.duration_to_next_epoch(T::EthSpec::slots_per_epoch())
|
|
{
|
|
Some(duration) => {
|
|
// this service should run 3/4 of the way through the epoch
|
|
let additional_delay = (epoch_duration * 3) / 4;
|
|
tokio::time::sleep(duration + additional_delay).await;
|
|
|
|
let deneb_fork_epoch = match chain.spec.deneb_fork_epoch {
|
|
Some(epoch) => epoch,
|
|
None => break, // shutdown service if deneb fork epoch not set
|
|
};
|
|
|
|
debug!(
|
|
chain.log,
|
|
"Availability cache maintenance service firing";
|
|
);
|
|
|
|
let current_epoch = match chain
|
|
.slot_clock
|
|
.now()
|
|
.map(|slot| slot.epoch(T::EthSpec::slots_per_epoch()))
|
|
{
|
|
Some(epoch) => epoch,
|
|
None => continue, // we'll have to try again next time I suppose..
|
|
};
|
|
|
|
if current_epoch < deneb_fork_epoch {
|
|
// we are not in deneb yet
|
|
continue;
|
|
}
|
|
|
|
let finalized_epoch = chain
|
|
.canonical_head
|
|
.fork_choice_read_lock()
|
|
.finalized_checkpoint()
|
|
.epoch;
|
|
// any data belonging to an epoch before this should be pruned
|
|
let cutoff_epoch = std::cmp::max(
|
|
finalized_epoch + 1,
|
|
std::cmp::max(
|
|
current_epoch.saturating_sub(*MIN_EPOCHS_FOR_BLOB_SIDECARS_REQUESTS),
|
|
deneb_fork_epoch,
|
|
),
|
|
);
|
|
|
|
if let Err(e) = overflow_cache.do_maintenance(cutoff_epoch) {
|
|
error!(chain.log, "Failed to maintain availability cache"; "error" => ?e);
|
|
}
|
|
}
|
|
None => {
|
|
error!(chain.log, "Failed to read slot clock");
|
|
// If we can't read the slot clock, just wait another slot.
|
|
tokio::time::sleep(chain.slot_clock.slot_duration()).await;
|
|
}
|
|
};
|
|
}
|
|
}
|
|
|
|
/// A fully available block that is ready to be imported into fork choice.
|
|
#[derive(Clone, Debug, PartialEq)]
|
|
pub struct AvailableBlock<E: EthSpec> {
|
|
block: Arc<SignedBeaconBlock<E>>,
|
|
blobs: Option<BlobSidecarList<E>>,
|
|
}
|
|
|
|
impl<E: EthSpec> AvailableBlock<E> {
|
|
pub fn block(&self) -> &SignedBeaconBlock<E> {
|
|
&self.block
|
|
}
|
|
pub fn block_cloned(&self) -> Arc<SignedBeaconBlock<E>> {
|
|
self.block.clone()
|
|
}
|
|
|
|
pub fn blobs(&self) -> Option<&BlobSidecarList<E>> {
|
|
self.blobs.as_ref()
|
|
}
|
|
|
|
pub fn deconstruct(self) -> (Arc<SignedBeaconBlock<E>>, Option<BlobSidecarList<E>>) {
|
|
let AvailableBlock { block, blobs } = self;
|
|
(block, blobs)
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, Clone)]
|
|
pub enum MaybeAvailableBlock<E: EthSpec> {
|
|
/// This variant is fully available.
|
|
/// i.e. for pre-deneb blocks, it contains a (`SignedBeaconBlock`, `Blobs::None`) and for
|
|
/// post-4844 blocks, it contains a `SignedBeaconBlock` and a Blobs variant other than `Blobs::None`.
|
|
Available(AvailableBlock<E>),
|
|
/// This variant is not fully available and requires blobs to become fully available.
|
|
AvailabilityPending(Arc<SignedBeaconBlock<E>>),
|
|
}
|