mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-03 08:41:43 +00:00
Compare commits
3 Commits
progressiv
...
copilot/fi
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
ebb8bea461 | ||
|
|
8debf97d3d | ||
|
|
dfeda940fe |
@@ -3777,17 +3777,26 @@ impl<T: BeaconChainTypes> BeaconChain<T> {
|
||||
consensus_context,
|
||||
} = import_data;
|
||||
|
||||
// Record the time at which this block's blobs became available.
|
||||
// Record the time at which this block's blobs or data columns became available.
|
||||
if let Some(blobs_available) = block.blobs_available_timestamp() {
|
||||
self.block_times_cache.write().set_time_blob_observed(
|
||||
block_root,
|
||||
block.slot(),
|
||||
blobs_available,
|
||||
);
|
||||
// Check if we're using data columns (post-Fulu) by looking at the block data
|
||||
match block.data() {
|
||||
AvailableBlockData::DataColumns(_) => {
|
||||
self.block_times_cache
|
||||
.write()
|
||||
.set_time_data_column_observed(block_root, block.slot(), blobs_available);
|
||||
}
|
||||
AvailableBlockData::Blobs(_) => {
|
||||
self.block_times_cache
|
||||
.write()
|
||||
.set_time_blob_observed(block_root, block.slot(), blobs_available);
|
||||
}
|
||||
AvailableBlockData::NoData => {
|
||||
// No data to record
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// TODO(das) record custody column available timestamp
|
||||
|
||||
let block_root = {
|
||||
// Capture the current span before moving into the blocking task
|
||||
let current_span = tracing::Span::current();
|
||||
|
||||
@@ -19,6 +19,7 @@ type BlockRoot = Hash256;
|
||||
pub struct Timestamps {
|
||||
pub observed: Option<Duration>,
|
||||
pub all_blobs_observed: Option<Duration>,
|
||||
pub all_data_columns_observed: Option<Duration>,
|
||||
pub consensus_verified: Option<Duration>,
|
||||
pub started_execution: Option<Duration>,
|
||||
pub executed: Option<Duration>,
|
||||
@@ -34,13 +35,15 @@ pub struct BlockDelays {
|
||||
pub observed: Option<Duration>,
|
||||
/// The time after the start of the slot we saw all blobs.
|
||||
pub all_blobs_observed: Option<Duration>,
|
||||
/// The time after the start of the slot we saw all data columns.
|
||||
pub all_data_columns_observed: Option<Duration>,
|
||||
/// The time it took to complete consensus verification of the block.
|
||||
pub consensus_verification_time: Option<Duration>,
|
||||
/// The time it took to complete execution verification of the block.
|
||||
pub execution_time: Option<Duration>,
|
||||
/// The delay from the start of the slot before the block became available
|
||||
///
|
||||
/// Equal to max(`observed + execution_time`, `all_blobs_observed`).
|
||||
/// Equal to max(`executed`, `all_blobs_observed` or `all_data_columns_observed`) minus slot start time.
|
||||
pub available: Option<Duration>,
|
||||
/// Time after `available`.
|
||||
pub attestable: Option<Duration>,
|
||||
@@ -62,6 +65,11 @@ impl BlockDelays {
|
||||
let all_blobs_observed = times
|
||||
.all_blobs_observed
|
||||
.and_then(|all_blobs_observed| all_blobs_observed.checked_sub(slot_start_time));
|
||||
let all_data_columns_observed = times
|
||||
.all_data_columns_observed
|
||||
.and_then(|all_data_columns_observed| {
|
||||
all_data_columns_observed.checked_sub(slot_start_time)
|
||||
});
|
||||
let consensus_verification_time = times
|
||||
.consensus_verified
|
||||
.and_then(|consensus_verified| consensus_verified.checked_sub(times.observed?));
|
||||
@@ -69,9 +77,14 @@ impl BlockDelays {
|
||||
.executed
|
||||
.and_then(|executed| executed.checked_sub(times.started_execution?));
|
||||
// Duration since UNIX epoch at which block became available.
|
||||
// Use data columns if present, otherwise fall back to blobs
|
||||
let availability_timestamp = times
|
||||
.all_data_columns_observed
|
||||
.or(times.all_blobs_observed)
|
||||
.unwrap_or_default();
|
||||
let available_time = times
|
||||
.executed
|
||||
.map(|executed| std::cmp::max(executed, times.all_blobs_observed.unwrap_or_default()));
|
||||
.map(|executed| std::cmp::max(executed, availability_timestamp));
|
||||
// Duration from the start of the slot until the block became available.
|
||||
let available_delay =
|
||||
available_time.and_then(|available_time| available_time.checked_sub(slot_start_time));
|
||||
@@ -87,6 +100,7 @@ impl BlockDelays {
|
||||
BlockDelays {
|
||||
observed,
|
||||
all_blobs_observed,
|
||||
all_data_columns_observed,
|
||||
consensus_verification_time,
|
||||
execution_time,
|
||||
available: available_delay,
|
||||
@@ -179,6 +193,28 @@ impl BlockTimesCache {
|
||||
}
|
||||
}
|
||||
|
||||
pub fn set_time_data_column_observed(
|
||||
&mut self,
|
||||
block_root: BlockRoot,
|
||||
slot: Slot,
|
||||
timestamp: Duration,
|
||||
) {
|
||||
// Unlike other functions in this file, we update the data column observed time only if it
|
||||
// is *greater* than existing data column observation times. This allows us to know the
|
||||
// observation time of the last data column to arrive.
|
||||
let block_times = self
|
||||
.cache
|
||||
.entry(block_root)
|
||||
.or_insert_with(|| BlockTimesCacheValue::new(slot));
|
||||
if block_times
|
||||
.timestamps
|
||||
.all_data_columns_observed
|
||||
.is_none_or(|prev| timestamp > prev)
|
||||
{
|
||||
block_times.timestamps.all_data_columns_observed = Some(timestamp);
|
||||
}
|
||||
}
|
||||
|
||||
/// Set the timestamp for `field` if that timestamp is less than any previously known value.
|
||||
///
|
||||
/// If no previous value is known for the field, then the supplied timestamp will always be
|
||||
|
||||
@@ -1359,12 +1359,12 @@ fn observe_head_block_delays<E: EthSpec, S: SlotClock>(
|
||||
.as_millis() as i64,
|
||||
);
|
||||
|
||||
// The time from the start of the slot when all blobs have been observed. Technically this
|
||||
// is the time we last saw a blob related to this block/slot.
|
||||
// The time from the start of the slot when all data columns have been observed. Technically
|
||||
// this is the time we last saw a data column related to this block/slot.
|
||||
metrics::set_gauge(
|
||||
&metrics::BEACON_BLOB_DELAY_ALL_OBSERVED_SLOT_START,
|
||||
block_delays
|
||||
.all_blobs_observed
|
||||
.all_data_columns_observed
|
||||
.unwrap_or_else(|| Duration::from_secs(0))
|
||||
.as_millis() as i64,
|
||||
);
|
||||
@@ -1445,7 +1445,7 @@ fn observe_head_block_delays<E: EthSpec, S: SlotClock>(
|
||||
slot = %head_block_slot,
|
||||
total_delay_ms = block_delay_total.as_millis(),
|
||||
observed_delay_ms = format_delay(&block_delays.observed),
|
||||
blob_delay_ms = format_delay(&block_delays.all_blobs_observed),
|
||||
data_column_delay_ms = format_delay(&block_delays.all_data_columns_observed),
|
||||
consensus_time_ms = format_delay(&block_delays.consensus_verification_time),
|
||||
execution_time_ms = format_delay(&block_delays.execution_time),
|
||||
available_delay_ms = format_delay(&block_delays.available),
|
||||
@@ -1479,7 +1479,7 @@ fn observe_head_block_delays<E: EthSpec, S: SlotClock>(
|
||||
slot = %head_block_slot,
|
||||
total_delay_ms = block_delay_total.as_millis(),
|
||||
observed_delay_ms = format_delay(&block_delays.observed),
|
||||
blob_delay_ms = format_delay(&block_delays.all_blobs_observed),
|
||||
data_column_delay_ms = format_delay(&block_delays.all_data_columns_observed),
|
||||
consensus_time_ms = format_delay(&block_delays.consensus_verification_time),
|
||||
execution_time_ms = format_delay(&block_delays.execution_time),
|
||||
available_delay_ms = format_delay(&block_delays.available),
|
||||
|
||||
@@ -240,10 +240,18 @@ impl<T: BeaconChainTypes> DataAvailabilityChecker<T> {
|
||||
slot: Slot,
|
||||
custody_columns: DataColumnSidecarList<T::EthSpec>,
|
||||
) -> Result<Availability<T::EthSpec>, AvailabilityCheckError> {
|
||||
let seen_timestamp = self
|
||||
.slot_clock
|
||||
.now_duration()
|
||||
.unwrap_or(Duration::from_secs(0));
|
||||
|
||||
// Attributes fault to the specific peer that sent an invalid column
|
||||
let kzg_verified_columns =
|
||||
KzgVerifiedDataColumn::from_batch_with_scoring(custody_columns, &self.kzg)
|
||||
.map_err(AvailabilityCheckError::InvalidColumn)?;
|
||||
let kzg_verified_columns = KzgVerifiedDataColumn::from_batch_with_scoring(
|
||||
custody_columns,
|
||||
&self.kzg,
|
||||
seen_timestamp,
|
||||
)
|
||||
.map_err(AvailabilityCheckError::InvalidColumn)?;
|
||||
|
||||
// Filter out columns that aren't required for custody for this slot
|
||||
// This is required because `data_columns_by_root` requests the **latest** CGC that _may_
|
||||
|
||||
@@ -282,8 +282,11 @@ impl<E: EthSpec> PendingComponents<E> {
|
||||
.flatten()
|
||||
.map(|blob| blob.seen_timestamp())
|
||||
.max(),
|
||||
// TODO(das): To be fixed with https://github.com/sigp/lighthouse/pull/6850
|
||||
AvailableBlockData::DataColumns(_) => None,
|
||||
AvailableBlockData::DataColumns(_) => self
|
||||
.verified_data_columns
|
||||
.iter()
|
||||
.map(|col| col.seen_timestamp())
|
||||
.max(),
|
||||
};
|
||||
|
||||
let AvailabilityPendingExecutedBlock {
|
||||
|
||||
@@ -16,6 +16,7 @@ use ssz_types::VariableList;
|
||||
use std::iter;
|
||||
use std::marker::PhantomData;
|
||||
use std::sync::Arc;
|
||||
use std::time::{Duration, SystemTime, UNIX_EPOCH};
|
||||
use tracing::{debug, instrument};
|
||||
use types::data::ColumnIndex;
|
||||
use types::{
|
||||
@@ -254,6 +255,8 @@ impl<T: BeaconChainTypes, O: ObservationStrategy> GossipVerifiedDataColumn<T, O>
|
||||
) -> Result<Self, GossipDataColumnError> {
|
||||
verify_data_column_sidecar(&column_sidecar, &chain.spec)?;
|
||||
|
||||
let seen_timestamp = chain.slot_clock.now_duration().unwrap_or_default();
|
||||
|
||||
// Check if the data column is already in the DA checker cache. This happens when data columns
|
||||
// are made available through the `engine_getBlobs` method. If it exists in the cache, we know
|
||||
// it has already passed the gossip checks, even though this particular instance hasn't been
|
||||
@@ -274,7 +277,10 @@ impl<T: BeaconChainTypes, O: ObservationStrategy> GossipVerifiedDataColumn<T, O>
|
||||
|
||||
Ok(Self {
|
||||
block_root: column_sidecar.block_root(),
|
||||
data_column: KzgVerifiedDataColumn::from_execution_verified(column_sidecar),
|
||||
data_column: KzgVerifiedDataColumn::from_execution_verified(
|
||||
column_sidecar,
|
||||
seen_timestamp,
|
||||
),
|
||||
_phantom: Default::default(),
|
||||
})
|
||||
}
|
||||
@@ -320,35 +326,51 @@ impl<T: BeaconChainTypes, O: ObservationStrategy> GossipVerifiedDataColumn<T, O>
|
||||
#[ssz(struct_behaviour = "transparent")]
|
||||
pub struct KzgVerifiedDataColumn<E: EthSpec> {
|
||||
data: Arc<DataColumnSidecar<E>>,
|
||||
#[ssz(skip_serializing, skip_deserializing)]
|
||||
seen_timestamp: Duration,
|
||||
}
|
||||
|
||||
impl<E: EthSpec> KzgVerifiedDataColumn<E> {
|
||||
pub fn new(
|
||||
data_column: Arc<DataColumnSidecar<E>>,
|
||||
kzg: &Kzg,
|
||||
seen_timestamp: Duration,
|
||||
) -> Result<Self, (Option<ColumnIndex>, KzgError)> {
|
||||
verify_kzg_for_data_column(data_column, kzg)
|
||||
verify_kzg_for_data_column(data_column, kzg, seen_timestamp)
|
||||
}
|
||||
|
||||
/// Mark a data column as KZG verified. Caller must ONLY use this on columns constructed
|
||||
/// from EL blobs.
|
||||
pub fn from_execution_verified(data_column: Arc<DataColumnSidecar<E>>) -> Self {
|
||||
Self { data: data_column }
|
||||
pub fn from_execution_verified(
|
||||
data_column: Arc<DataColumnSidecar<E>>,
|
||||
seen_timestamp: Duration,
|
||||
) -> Self {
|
||||
Self {
|
||||
data: data_column,
|
||||
seen_timestamp,
|
||||
}
|
||||
}
|
||||
|
||||
/// Create a `KzgVerifiedDataColumn` from `DataColumnSidecar` for testing ONLY.
|
||||
pub(crate) fn __new_for_testing(data_column: Arc<DataColumnSidecar<E>>) -> Self {
|
||||
Self { data: data_column }
|
||||
Self {
|
||||
data: data_column,
|
||||
seen_timestamp: Duration::from_secs(0),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn from_batch_with_scoring(
|
||||
data_columns: Vec<Arc<DataColumnSidecar<E>>>,
|
||||
kzg: &Kzg,
|
||||
seen_timestamp: Duration,
|
||||
) -> Result<Vec<Self>, (Option<ColumnIndex>, KzgError)> {
|
||||
verify_kzg_for_data_column_list(data_columns.iter(), kzg)?;
|
||||
Ok(data_columns
|
||||
.into_iter()
|
||||
.map(|column| Self { data: column })
|
||||
.map(|column| Self {
|
||||
data: column,
|
||||
seen_timestamp,
|
||||
})
|
||||
.collect())
|
||||
}
|
||||
|
||||
@@ -366,6 +388,10 @@ impl<E: EthSpec> KzgVerifiedDataColumn<E> {
|
||||
pub fn index(&self) -> ColumnIndex {
|
||||
*self.data.index()
|
||||
}
|
||||
|
||||
pub fn seen_timestamp(&self) -> Duration {
|
||||
self.seen_timestamp
|
||||
}
|
||||
}
|
||||
|
||||
pub type CustodyDataColumnList<E> =
|
||||
@@ -407,14 +433,18 @@ impl<E: EthSpec> CustodyDataColumn<E> {
|
||||
#[ssz(struct_behaviour = "transparent")]
|
||||
pub struct KzgVerifiedCustodyDataColumn<E: EthSpec> {
|
||||
data: Arc<DataColumnSidecar<E>>,
|
||||
#[ssz(skip_serializing, skip_deserializing)]
|
||||
seen_timestamp: Duration,
|
||||
}
|
||||
|
||||
impl<E: EthSpec> KzgVerifiedCustodyDataColumn<E> {
|
||||
/// Mark a column as custody column. Caller must ensure that our current custody requirements
|
||||
/// include this column
|
||||
pub fn from_asserted_custody(kzg_verified: KzgVerifiedDataColumn<E>) -> Self {
|
||||
let seen_timestamp = kzg_verified.seen_timestamp();
|
||||
Self {
|
||||
data: kzg_verified.to_data_column(),
|
||||
seen_timestamp,
|
||||
}
|
||||
}
|
||||
|
||||
@@ -422,10 +452,12 @@ impl<E: EthSpec> KzgVerifiedCustodyDataColumn<E> {
|
||||
pub fn new(
|
||||
data_column: CustodyDataColumn<E>,
|
||||
kzg: &Kzg,
|
||||
seen_timestamp: Duration,
|
||||
) -> Result<Self, (Option<ColumnIndex>, KzgError)> {
|
||||
verify_kzg_for_data_column(data_column.clone_arc(), kzg)?;
|
||||
verify_kzg_for_data_column(data_column.clone_arc(), kzg, seen_timestamp)?;
|
||||
Ok(Self {
|
||||
data: data_column.data,
|
||||
seen_timestamp,
|
||||
})
|
||||
}
|
||||
|
||||
@@ -443,10 +475,16 @@ impl<E: EthSpec> KzgVerifiedCustodyDataColumn<E> {
|
||||
spec,
|
||||
)?;
|
||||
|
||||
// Use the current time as the seen_timestamp for reconstructed columns
|
||||
let seen_timestamp = SystemTime::now()
|
||||
.duration_since(UNIX_EPOCH)
|
||||
.unwrap_or_else(|_| Duration::from_secs(0));
|
||||
|
||||
Ok(all_data_columns
|
||||
.into_iter()
|
||||
.map(|data| {
|
||||
KzgVerifiedCustodyDataColumn::from_asserted_custody(KzgVerifiedDataColumn { data })
|
||||
.map(|data| KzgVerifiedCustodyDataColumn {
|
||||
data,
|
||||
seen_timestamp,
|
||||
})
|
||||
.collect::<Vec<_>>())
|
||||
}
|
||||
@@ -464,6 +502,10 @@ impl<E: EthSpec> KzgVerifiedCustodyDataColumn<E> {
|
||||
pub fn index(&self) -> ColumnIndex {
|
||||
*self.data.index()
|
||||
}
|
||||
|
||||
pub fn seen_timestamp(&self) -> Duration {
|
||||
self.seen_timestamp
|
||||
}
|
||||
}
|
||||
|
||||
/// Complete kzg verification for a `DataColumnSidecar`.
|
||||
@@ -473,10 +515,14 @@ impl<E: EthSpec> KzgVerifiedCustodyDataColumn<E> {
|
||||
pub fn verify_kzg_for_data_column<E: EthSpec>(
|
||||
data_column: Arc<DataColumnSidecar<E>>,
|
||||
kzg: &Kzg,
|
||||
seen_timestamp: Duration,
|
||||
) -> Result<KzgVerifiedDataColumn<E>, (Option<ColumnIndex>, KzgError)> {
|
||||
let _timer = metrics::start_timer(&metrics::KZG_VERIFICATION_DATA_COLUMN_SINGLE_TIMES);
|
||||
validate_data_columns(kzg, iter::once(&data_column))?;
|
||||
Ok(KzgVerifiedDataColumn { data: data_column })
|
||||
Ok(KzgVerifiedDataColumn {
|
||||
data: data_column,
|
||||
seen_timestamp,
|
||||
})
|
||||
}
|
||||
|
||||
/// Complete kzg verification for a list of `DataColumnSidecar`s.
|
||||
@@ -518,6 +564,8 @@ pub fn validate_data_column_sidecar_for_gossip_fulu<T: BeaconChainTypes, O: Obse
|
||||
verify_slot_greater_than_latest_finalized_slot(chain, column_slot)?;
|
||||
verify_is_unknown_sidecar(chain, &data_column)?;
|
||||
|
||||
let seen_timestamp = chain.slot_clock.now_duration().unwrap_or_default();
|
||||
|
||||
// Check if the data column is already in the DA checker cache. This happens when data columns
|
||||
// are made available through the `engine_getBlobs` method. If it exists in the cache, we know
|
||||
// it has already passed the gossip checks, even though this particular instance hasn't been
|
||||
@@ -539,8 +587,9 @@ pub fn validate_data_column_sidecar_for_gossip_fulu<T: BeaconChainTypes, O: Obse
|
||||
verify_slot_higher_than_parent(&parent_block, column_slot)?;
|
||||
verify_proposer_and_signature(data_column_fulu, &parent_block, chain)?;
|
||||
let kzg = &chain.kzg;
|
||||
let kzg_verified_data_column = verify_kzg_for_data_column(data_column.clone(), kzg)
|
||||
.map_err(|(_, e)| GossipDataColumnError::InvalidKzgProof(e))?;
|
||||
let kzg_verified_data_column =
|
||||
verify_kzg_for_data_column(data_column.clone(), kzg, seen_timestamp)
|
||||
.map_err(|(_, e)| GossipDataColumnError::InvalidKzgProof(e))?;
|
||||
|
||||
chain
|
||||
.observed_slashable
|
||||
|
||||
@@ -377,6 +377,7 @@ async fn compute_custody_columns_to_import<T: BeaconChainTypes>(
|
||||
// This filtering ensures we only import and publish the custody columns.
|
||||
// `DataAvailabilityChecker` requires a strict match on custody columns count to
|
||||
// consider a block available.
|
||||
let seen_timestamp = timestamp_now();
|
||||
let mut custody_columns = data_columns_result
|
||||
.map(|data_columns| {
|
||||
data_columns
|
||||
@@ -384,7 +385,10 @@ async fn compute_custody_columns_to_import<T: BeaconChainTypes>(
|
||||
.filter(|col| custody_columns_indices.contains(col.index()))
|
||||
.map(|col| {
|
||||
KzgVerifiedCustodyDataColumn::from_asserted_custody(
|
||||
KzgVerifiedDataColumn::from_execution_verified(col),
|
||||
KzgVerifiedDataColumn::from_execution_verified(
|
||||
col,
|
||||
seen_timestamp,
|
||||
),
|
||||
)
|
||||
})
|
||||
.collect::<Vec<_>>()
|
||||
|
||||
Reference in New Issue
Block a user