mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-12 02:14:10 +00:00
* Attestation superstruct changes for EIP 7549 (#5644)
* update
* experiment
* superstruct changes
* revert
* superstruct changes
* fix tests
* indexed attestation
* indexed attestation superstruct
* updated TODOs
* `superstruct` the `AttesterSlashing` (#5636)
* `superstruct` Attester Fork Variants
* Push a little further
* Deal with Encode / Decode of AttesterSlashing
* not so sure about this..
* Stop Encode/Decode Bounds from Propagating Out
* Tons of Changes..
* More Conversions to AttestationRef
* Add AsReference trait (#15)
* Add AsReference trait
* Fix some snafus
* Got it Compiling! :D
* Got Tests Building
* Get beacon chain tests compiling
---------
Co-authored-by: Michael Sproul <micsproul@gmail.com>
* Merge remote-tracking branch 'upstream/unstable' into electra_attestation_changes
* Make EF Tests Fork-Agnostic (#5713)
* Finish EF Test Fork Agnostic (#5714)
* Superstruct `AggregateAndProof` (#5715)
* Upgrade `superstruct` to `0.8.0`
* superstruct `AggregateAndProof`
* Merge remote-tracking branch 'sigp/unstable' into electra_attestation_changes
* cargo fmt
* Merge pull request #5726 from realbigsean/electra_attestation_changes
Merge unstable into Electra attestation changes
* EIP7549 `get_attestation_indices` (#5657)
* get attesting indices electra impl
* fmt
* get tests to pass
* fmt
* fix some beacon chain tests
* fmt
* fix slasher test
* fmt got me again
* fix more tests
* fix tests
* Some small changes (#5739)
* cargo fmt (#5740)
* Sketch op pool changes
* fix get attesting indices (#5742)
* fix get attesting indices
* better errors
* fix compile
* only get committee index once
* Ef test fixes (#5753)
* attestation related ef test fixes
* delete commented out stuff
* Fix Aggregation Pool for Electra (#5754)
* Fix Aggregation Pool for Electra
* Remove Outdated Interface
* fix ssz (#5755)
* Get `electra_op_pool` up to date (#5756)
* fix get attesting indices (#5742)
* fix get attesting indices
* better errors
* fix compile
* only get committee index once
* Ef test fixes (#5753)
* attestation related ef test fixes
* delete commented out stuff
* Fix Aggregation Pool for Electra (#5754)
* Fix Aggregation Pool for Electra
* Remove Outdated Interface
* fix ssz (#5755)
---------
Co-authored-by: realbigsean <sean@sigmaprime.io>
* Revert "Get `electra_op_pool` up to date (#5756)" (#5757)
This reverts commit ab9e58aa3d.
* Merge branch 'electra_attestation_changes' of https://github.com/sigp/lighthouse into electra_op_pool
* Compute on chain aggregate impl (#5752)
* add compute_on_chain_agg impl to op pool changes
* fmt
* get op pool tests to pass
* update the naive agg pool interface (#5760)
* Fix bugs in cross-committee aggregation
* Add comment to max cover optimisation
* Fix assert
* Merge pull request #5749 from sigp/electra_op_pool
Optimise Electra op pool aggregation
* update committee offset
* Fix Electra Fork Choice Tests (#5764)
* Subscribe to the correct subnets for electra attestations (#5782)
* subscribe to the correct att subnets for electra
* subscribe to the correct att subnets for electra
* cargo fmt
* fix slashing handling
* Merge remote-tracking branch 'upstream/unstable'
* Send unagg attestation based on fork
* Publish all aggregates
* just one more check bro plz..
* Merge pull request #5832 from ethDreamer/electra_attestation_changes_merge_unstable
Merge `unstable` into `electra_attestation_changes`
* Merge pull request #5835 from realbigsean/fix-validator-logic
Fix validator logic
* Merge pull request #5816 from realbigsean/electra-attestation-slashing-handling
Electra slashing handling
* Electra attestation changes rm decode impl (#5856)
* Remove Crappy Decode impl for Attestation
* Remove Inefficient Attestation Decode impl
* Implement Schema Upgrade / Downgrade
* Update beacon_node/beacon_chain/src/schema_change/migration_schema_v20.rs
Co-authored-by: Michael Sproul <micsproul@gmail.com>
---------
Co-authored-by: Michael Sproul <micsproul@gmail.com>
* Fix failing attestation tests and misc electra attestation cleanup (#5810)
* - get attestation related beacon chain tests to pass
- observed attestations are now keyed off of data + committee index
- rename op pool attestationref to compactattestationref
- remove unwraps in agg pool and use options instead
- cherry pick some changes from ef-tests-electra
* cargo fmt
* fix failing test
* Revert dockerfile changes
* make committee_index return option
* function args shouldnt be a ref to attestation ref
* fmt
* fix dup imports
---------
Co-authored-by: realbigsean <seananderson33@GMAIL.com>
* fix some todos (#5817)
* Merge branch 'unstable' of https://github.com/sigp/lighthouse into electra_attestation_changes
* add consolidations to merkle calc for inclusion proof
* Remove Duplicate KZG Commitment Merkle Proof Code (#5874)
* Remove Duplicate KZG Commitment Merkle Proof Code
* s/tree_lists/fields/
* Merge branch 'unstable' of https://github.com/sigp/lighthouse into electra_attestation_changes
* fix compile
* Fix slasher tests (#5906)
* Fix electra tests
* Add electra attestations to double vote tests
* Update superstruct to 0.8
* Merge remote-tracking branch 'origin/unstable' into electra_attestation_changes
* Small cleanup in slasher tests
* Clean up Electra observed aggregates (#5929)
* Use consistent key in observed_attestations
* Remove unwraps from observed aggregates
* Merge branch 'unstable' of https://github.com/sigp/lighthouse into electra_attestation_changes
* De-dup attestation constructor logic
* Remove unwraps in Attestation construction
* Dedup match_attestation_data
* Remove outdated TODO
* Use ForkName Ord in fork-choice tests
* Use ForkName Ord in BeaconBlockBody
* Make to_electra not fallible
* Remove TestRandom impl for IndexedAttestation
* Remove IndexedAttestation faulty Decode impl
* Drop TestRandom impl
* Add PendingAttestationInElectra
* Indexed att on disk (#35)
* indexed att on disk
* fix lints
* Update slasher/src/migrate.rs
Co-authored-by: ethDreamer <37123614+ethDreamer@users.noreply.github.com>
---------
Co-authored-by: Lion - dapplion <35266934+dapplion@users.noreply.github.com>
Co-authored-by: ethDreamer <37123614+ethDreamer@users.noreply.github.com>
* add electra fork enabled fn to ForkName impl (#36)
* add electra fork enabled fn to ForkName impl
* remove inadvertent file
* Update common/eth2/src/types.rs
Co-authored-by: ethDreamer <37123614+ethDreamer@users.noreply.github.com>
* Dedup attestation constructor logic in attester cache
* Use if let Ok for committee_bits
* Dedup Attestation constructor code
* Diff reduction in tests
* Fix beacon_chain tests
* Diff reduction
* Use Ord for ForkName in pubsub
* Resolve into_attestation_and_indices todo
* Remove stale TODO
* Fix beacon_chain tests
* Test spec invariant
* Use electra_enabled in pubsub
* Remove get_indexed_attestation_from_signed_aggregate
* Use ok_or instead of if let else
* committees are sorted
* remove dup method `get_indexed_attestation_from_committees`
* Merge pull request #5940 from dapplion/electra_attestation_changes_lionreview
Electra attestations #5712 review
* update default persisted op pool deserialization
* ensure aggregate and proof uses serde untagged on ref
* Fork aware ssz static attestation tests
* Electra attestation changes from Lions review (#5971)
* dedup/cleanup and remove unneeded hashset use
* remove irrelevant TODOs
* Merge branch 'unstable' of https://github.com/sigp/lighthouse into electra_attestation_changes
* Electra attestation changes sean review (#5972)
* instantiate empty bitlist in unreachable code
* clean up error conversion
* fork enabled bool cleanup
* remove a couple todos
* return bools instead of options in `aggregate` and use the result
* delete commented out code
* use map macros in simple transformations
* remove signers_disjoint_from
* get ef tests compiling
* get ef tests compiling
* update intentionally excluded files
* Avoid changing slasher schema for Electra
* Delete slasher schema v4
* Fix clippy
* Fix compilation of beacon_chain tests
* Update database.rs
* Add electra lightclient types
* Update slasher/src/database.rs
* fix imports
* Merge pull request #5980 from dapplion/electra-lightclient
Add electra lightclient types
* Merge pull request #5975 from michaelsproul/electra-slasher-no-migration
Avoid changing slasher schema for Electra
* Update beacon_node/beacon_chain/src/attestation_verification.rs
* Update beacon_node/beacon_chain/src/attestation_verification.rs
953 lines
33 KiB
Rust
953 lines
33 KiB
Rust
use crate::beacon_node_fallback::{Error as FallbackError, Errors};
|
|
use crate::{
|
|
beacon_node_fallback::{ApiTopic, BeaconNodeFallback, RequireSynced},
|
|
determine_graffiti,
|
|
graffiti_file::GraffitiFile,
|
|
OfflineOnFailure,
|
|
};
|
|
use crate::{
|
|
http_metrics::metrics,
|
|
validator_store::{Error as ValidatorStoreError, ValidatorStore},
|
|
};
|
|
use bls::SignatureBytes;
|
|
use environment::RuntimeContext;
|
|
use eth2::types::{FullBlockContents, PublishBlockRequest};
|
|
use eth2::{BeaconNodeHttpClient, StatusCode};
|
|
use slog::{crit, debug, error, info, trace, warn, Logger};
|
|
use slot_clock::SlotClock;
|
|
use std::fmt::Debug;
|
|
use std::future::Future;
|
|
use std::ops::Deref;
|
|
use std::sync::Arc;
|
|
use std::time::Duration;
|
|
use tokio::sync::mpsc;
|
|
use types::{
|
|
BlindedBeaconBlock, BlockType, EthSpec, Graffiti, PublicKeyBytes, SignedBlindedBeaconBlock,
|
|
Slot,
|
|
};
|
|
|
|
#[derive(Debug)]
|
|
pub enum BlockError {
|
|
/// A recoverable error that can be retried, as the validator has not signed anything.
|
|
Recoverable(String),
|
|
/// An irrecoverable error has occurred during block proposal and should not be retried, as a
|
|
/// block may have already been signed.
|
|
Irrecoverable(String),
|
|
}
|
|
|
|
impl From<Errors<BlockError>> for BlockError {
|
|
fn from(e: Errors<BlockError>) -> Self {
|
|
if e.0.iter().any(|(_, error)| {
|
|
matches!(
|
|
error,
|
|
FallbackError::RequestFailed(BlockError::Irrecoverable(_))
|
|
)
|
|
}) {
|
|
BlockError::Irrecoverable(e.to_string())
|
|
} else {
|
|
BlockError::Recoverable(e.to_string())
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Builds a `BlockService`.
|
|
pub struct BlockServiceBuilder<T, E: EthSpec> {
|
|
validator_store: Option<Arc<ValidatorStore<T, E>>>,
|
|
slot_clock: Option<Arc<T>>,
|
|
beacon_nodes: Option<Arc<BeaconNodeFallback<T, E>>>,
|
|
proposer_nodes: Option<Arc<BeaconNodeFallback<T, E>>>,
|
|
context: Option<RuntimeContext<E>>,
|
|
graffiti: Option<Graffiti>,
|
|
graffiti_file: Option<GraffitiFile>,
|
|
}
|
|
|
|
impl<T: SlotClock + 'static, E: EthSpec> BlockServiceBuilder<T, E> {
|
|
pub fn new() -> Self {
|
|
Self {
|
|
validator_store: None,
|
|
slot_clock: None,
|
|
beacon_nodes: None,
|
|
proposer_nodes: None,
|
|
context: None,
|
|
graffiti: None,
|
|
graffiti_file: None,
|
|
}
|
|
}
|
|
|
|
pub fn validator_store(mut self, store: Arc<ValidatorStore<T, E>>) -> Self {
|
|
self.validator_store = Some(store);
|
|
self
|
|
}
|
|
|
|
pub fn slot_clock(mut self, slot_clock: T) -> Self {
|
|
self.slot_clock = Some(Arc::new(slot_clock));
|
|
self
|
|
}
|
|
|
|
pub fn beacon_nodes(mut self, beacon_nodes: Arc<BeaconNodeFallback<T, E>>) -> Self {
|
|
self.beacon_nodes = Some(beacon_nodes);
|
|
self
|
|
}
|
|
|
|
pub fn proposer_nodes(mut self, proposer_nodes: Arc<BeaconNodeFallback<T, E>>) -> Self {
|
|
self.proposer_nodes = Some(proposer_nodes);
|
|
self
|
|
}
|
|
|
|
pub fn runtime_context(mut self, context: RuntimeContext<E>) -> Self {
|
|
self.context = Some(context);
|
|
self
|
|
}
|
|
|
|
pub fn graffiti(mut self, graffiti: Option<Graffiti>) -> Self {
|
|
self.graffiti = graffiti;
|
|
self
|
|
}
|
|
|
|
pub fn graffiti_file(mut self, graffiti_file: Option<GraffitiFile>) -> Self {
|
|
self.graffiti_file = graffiti_file;
|
|
self
|
|
}
|
|
|
|
pub fn build(self) -> Result<BlockService<T, E>, String> {
|
|
Ok(BlockService {
|
|
inner: Arc::new(Inner {
|
|
validator_store: self
|
|
.validator_store
|
|
.ok_or("Cannot build BlockService without validator_store")?,
|
|
slot_clock: self
|
|
.slot_clock
|
|
.ok_or("Cannot build BlockService without slot_clock")?,
|
|
beacon_nodes: self
|
|
.beacon_nodes
|
|
.ok_or("Cannot build BlockService without beacon_node")?,
|
|
context: self
|
|
.context
|
|
.ok_or("Cannot build BlockService without runtime_context")?,
|
|
proposer_nodes: self.proposer_nodes,
|
|
graffiti: self.graffiti,
|
|
graffiti_file: self.graffiti_file,
|
|
}),
|
|
})
|
|
}
|
|
}
|
|
|
|
// Combines a set of non-block-proposing `beacon_nodes` and only-block-proposing
|
|
// `proposer_nodes`.
|
|
pub struct ProposerFallback<T, E: EthSpec> {
|
|
beacon_nodes: Arc<BeaconNodeFallback<T, E>>,
|
|
proposer_nodes: Option<Arc<BeaconNodeFallback<T, E>>>,
|
|
}
|
|
|
|
impl<T: SlotClock, E: EthSpec> ProposerFallback<T, E> {
|
|
// Try `func` on `self.proposer_nodes` first. If that doesn't work, try `self.beacon_nodes`.
|
|
pub async fn request_proposers_first<'a, F, Err, R>(
|
|
&'a self,
|
|
require_synced: RequireSynced,
|
|
offline_on_failure: OfflineOnFailure,
|
|
func: F,
|
|
) -> Result<(), Errors<Err>>
|
|
where
|
|
F: Fn(&'a BeaconNodeHttpClient) -> R + Clone,
|
|
R: Future<Output = Result<(), Err>>,
|
|
Err: Debug,
|
|
{
|
|
// If there are proposer nodes, try calling `func` on them and return early if they are successful.
|
|
if let Some(proposer_nodes) = &self.proposer_nodes {
|
|
if proposer_nodes
|
|
.request(
|
|
require_synced,
|
|
offline_on_failure,
|
|
ApiTopic::Blocks,
|
|
func.clone(),
|
|
)
|
|
.await
|
|
.is_ok()
|
|
{
|
|
return Ok(());
|
|
}
|
|
}
|
|
|
|
// If the proposer nodes failed, try on the non-proposer nodes.
|
|
self.beacon_nodes
|
|
.request(require_synced, offline_on_failure, ApiTopic::Blocks, func)
|
|
.await
|
|
}
|
|
|
|
// Try `func` on `self.beacon_nodes` first. If that doesn't work, try `self.proposer_nodes`.
|
|
pub async fn request_proposers_last<'a, F, O, Err, R>(
|
|
&'a self,
|
|
require_synced: RequireSynced,
|
|
offline_on_failure: OfflineOnFailure,
|
|
func: F,
|
|
) -> Result<O, Errors<Err>>
|
|
where
|
|
F: Fn(&'a BeaconNodeHttpClient) -> R + Clone,
|
|
R: Future<Output = Result<O, Err>>,
|
|
Err: Debug,
|
|
{
|
|
// Try running `func` on the non-proposer beacon nodes.
|
|
let beacon_nodes_result = self
|
|
.beacon_nodes
|
|
.first_success(require_synced, offline_on_failure, func.clone())
|
|
.await;
|
|
|
|
match (beacon_nodes_result, &self.proposer_nodes) {
|
|
// The non-proposer node call succeed, return the result.
|
|
(Ok(success), _) => Ok(success),
|
|
// The non-proposer node call failed, but we don't have any proposer nodes. Return an error.
|
|
(Err(e), None) => Err(e),
|
|
// The non-proposer node call failed, try the same call on the proposer nodes.
|
|
(Err(_), Some(proposer_nodes)) => {
|
|
proposer_nodes
|
|
.first_success(require_synced, offline_on_failure, func)
|
|
.await
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
/// Helper to minimise `Arc` usage.
|
|
pub struct Inner<T, E: EthSpec> {
|
|
validator_store: Arc<ValidatorStore<T, E>>,
|
|
slot_clock: Arc<T>,
|
|
beacon_nodes: Arc<BeaconNodeFallback<T, E>>,
|
|
proposer_nodes: Option<Arc<BeaconNodeFallback<T, E>>>,
|
|
context: RuntimeContext<E>,
|
|
graffiti: Option<Graffiti>,
|
|
graffiti_file: Option<GraffitiFile>,
|
|
}
|
|
|
|
/// Attempts to produce attestations for any block producer(s) at the start of the epoch.
|
|
pub struct BlockService<T, E: EthSpec> {
|
|
inner: Arc<Inner<T, E>>,
|
|
}
|
|
|
|
impl<T, E: EthSpec> Clone for BlockService<T, E> {
|
|
fn clone(&self) -> Self {
|
|
Self {
|
|
inner: self.inner.clone(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl<T, E: EthSpec> Deref for BlockService<T, E> {
|
|
type Target = Inner<T, E>;
|
|
|
|
fn deref(&self) -> &Self::Target {
|
|
self.inner.deref()
|
|
}
|
|
}
|
|
|
|
/// Notification from the duties service that we should try to produce a block.
|
|
pub struct BlockServiceNotification {
|
|
pub slot: Slot,
|
|
pub block_proposers: Vec<PublicKeyBytes>,
|
|
}
|
|
|
|
impl<T: SlotClock + 'static, E: EthSpec> BlockService<T, E> {
|
|
pub fn start_update_service(
|
|
self,
|
|
mut notification_rx: mpsc::Receiver<BlockServiceNotification>,
|
|
) -> Result<(), String> {
|
|
let log = self.context.log().clone();
|
|
|
|
info!(log, "Block production service started");
|
|
|
|
let executor = self.inner.context.executor.clone();
|
|
|
|
executor.spawn(
|
|
async move {
|
|
while let Some(notif) = notification_rx.recv().await {
|
|
self.do_update(notif).await.ok();
|
|
}
|
|
debug!(log, "Block service shutting down");
|
|
},
|
|
"block_service",
|
|
);
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Attempt to produce a block for any block producers in the `ValidatorStore`.
|
|
async fn do_update(&self, notification: BlockServiceNotification) -> Result<(), ()> {
|
|
let log = self.context.log();
|
|
let _timer =
|
|
metrics::start_timer_vec(&metrics::BLOCK_SERVICE_TIMES, &[metrics::FULL_UPDATE]);
|
|
|
|
let slot = self.slot_clock.now().ok_or_else(move || {
|
|
crit!(log, "Duties manager failed to read slot clock");
|
|
})?;
|
|
|
|
if notification.slot != slot {
|
|
warn!(
|
|
log,
|
|
"Skipping block production for expired slot";
|
|
"current_slot" => slot.as_u64(),
|
|
"notification_slot" => notification.slot.as_u64(),
|
|
"info" => "Your machine could be overloaded"
|
|
);
|
|
return Ok(());
|
|
}
|
|
|
|
if slot == self.context.eth2_config.spec.genesis_slot {
|
|
debug!(
|
|
log,
|
|
"Not producing block at genesis slot";
|
|
"proposers" => format!("{:?}", notification.block_proposers),
|
|
);
|
|
return Ok(());
|
|
}
|
|
|
|
trace!(
|
|
log,
|
|
"Block service update started";
|
|
"slot" => slot.as_u64()
|
|
);
|
|
|
|
let proposers = notification.block_proposers;
|
|
|
|
if proposers.is_empty() {
|
|
trace!(
|
|
log,
|
|
"No local block proposers for this slot";
|
|
"slot" => slot.as_u64()
|
|
)
|
|
} else if proposers.len() > 1 {
|
|
error!(
|
|
log,
|
|
"Multiple block proposers for this slot";
|
|
"action" => "producing blocks for all proposers",
|
|
"num_proposers" => proposers.len(),
|
|
"slot" => slot.as_u64(),
|
|
)
|
|
}
|
|
|
|
if self.validator_store.produce_block_v3() {
|
|
for validator_pubkey in proposers {
|
|
let builder_boost_factor = self.get_builder_boost_factor(&validator_pubkey);
|
|
let service = self.clone();
|
|
let log = log.clone();
|
|
self.inner.context.executor.spawn(
|
|
async move {
|
|
let result = service
|
|
.publish_block_v3(slot, validator_pubkey, builder_boost_factor)
|
|
.await;
|
|
|
|
match result {
|
|
Ok(_) => {}
|
|
Err(BlockError::Recoverable(e)) | Err(BlockError::Irrecoverable(e)) => {
|
|
error!(
|
|
log,
|
|
"Error whilst producing block";
|
|
"error" => ?e,
|
|
"block_slot" => ?slot,
|
|
"info" => "block v3 proposal failed, this error may or may not result in a missed block"
|
|
);
|
|
}
|
|
}
|
|
},
|
|
"block service",
|
|
)
|
|
}
|
|
} else {
|
|
for validator_pubkey in proposers {
|
|
let builder_proposals = self
|
|
.validator_store
|
|
.get_builder_proposals(&validator_pubkey);
|
|
let service = self.clone();
|
|
let log = log.clone();
|
|
self.inner.context.executor.spawn(
|
|
async move {
|
|
if builder_proposals {
|
|
let result = service
|
|
.publish_block(slot, validator_pubkey, true)
|
|
.await;
|
|
|
|
match result {
|
|
Err(BlockError::Recoverable(e)) => {
|
|
error!(
|
|
log,
|
|
"Error whilst producing block";
|
|
"error" => ?e,
|
|
"block_slot" => ?slot,
|
|
"info" => "blinded proposal failed, attempting full block"
|
|
);
|
|
if let Err(e) = service
|
|
.publish_block(slot, validator_pubkey, false)
|
|
.await
|
|
{
|
|
// Log a `crit` since a full block
|
|
// (non-builder) proposal failed.
|
|
crit!(
|
|
log,
|
|
"Error whilst producing block";
|
|
"error" => ?e,
|
|
"block_slot" => ?slot,
|
|
"info" => "full block attempted after a blinded failure",
|
|
);
|
|
}
|
|
}
|
|
Err(BlockError::Irrecoverable(e)) => {
|
|
// Only log an `error` since it's common for
|
|
// builders to timeout on their response, only
|
|
// to publish the block successfully themselves.
|
|
error!(
|
|
log,
|
|
"Error whilst producing block";
|
|
"error" => ?e,
|
|
"block_slot" => ?slot,
|
|
"info" => "this error may or may not result in a missed block",
|
|
)
|
|
}
|
|
Ok(_) => {}
|
|
};
|
|
} else if let Err(e) = service
|
|
.publish_block(slot, validator_pubkey, false)
|
|
.await
|
|
{
|
|
// Log a `crit` since a full block (non-builder)
|
|
// proposal failed.
|
|
crit!(
|
|
log,
|
|
"Error whilst producing block";
|
|
"message" => ?e,
|
|
"block_slot" => ?slot,
|
|
"info" => "proposal did not use a builder",
|
|
);
|
|
}
|
|
},
|
|
"block service",
|
|
)
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[allow(clippy::too_many_arguments)]
|
|
async fn sign_and_publish_block(
|
|
&self,
|
|
proposer_fallback: ProposerFallback<T, E>,
|
|
slot: Slot,
|
|
graffiti: Option<Graffiti>,
|
|
validator_pubkey: &PublicKeyBytes,
|
|
unsigned_block: UnsignedBlock<E>,
|
|
) -> Result<(), BlockError> {
|
|
let log = self.context.log();
|
|
let signing_timer = metrics::start_timer(&metrics::BLOCK_SIGNING_TIMES);
|
|
|
|
let res = match unsigned_block {
|
|
UnsignedBlock::Full(block_contents) => {
|
|
let (block, maybe_blobs) = block_contents.deconstruct();
|
|
self.validator_store
|
|
.sign_block(*validator_pubkey, block, slot)
|
|
.await
|
|
.map(|b| SignedBlock::Full(PublishBlockRequest::new(Arc::new(b), maybe_blobs)))
|
|
}
|
|
UnsignedBlock::Blinded(block) => self
|
|
.validator_store
|
|
.sign_block(*validator_pubkey, block, slot)
|
|
.await
|
|
.map(Arc::new)
|
|
.map(SignedBlock::Blinded),
|
|
};
|
|
|
|
let signed_block = match res {
|
|
Ok(block) => block,
|
|
Err(ValidatorStoreError::UnknownPubkey(pubkey)) => {
|
|
// A pubkey can be missing when a validator was recently removed
|
|
// via the API.
|
|
warn!(
|
|
log,
|
|
"Missing pubkey for block";
|
|
"info" => "a validator may have recently been removed from this VC",
|
|
"pubkey" => ?pubkey,
|
|
"slot" => ?slot
|
|
);
|
|
return Ok(());
|
|
}
|
|
Err(e) => {
|
|
return Err(BlockError::Recoverable(format!(
|
|
"Unable to sign block: {:?}",
|
|
e
|
|
)))
|
|
}
|
|
};
|
|
|
|
let signing_time_ms =
|
|
Duration::from_secs_f64(signing_timer.map_or(0.0, |t| t.stop_and_record())).as_millis();
|
|
|
|
info!(
|
|
log,
|
|
"Publishing signed block";
|
|
"slot" => slot.as_u64(),
|
|
"signing_time_ms" => signing_time_ms,
|
|
);
|
|
|
|
// Publish block with first available beacon node.
|
|
//
|
|
// Try the proposer nodes first, since we've likely gone to efforts to
|
|
// protect them from DoS attacks and they're most likely to successfully
|
|
// publish a block.
|
|
proposer_fallback
|
|
.request_proposers_first(
|
|
RequireSynced::No,
|
|
OfflineOnFailure::Yes,
|
|
|beacon_node| async {
|
|
self.publish_signed_block_contents(&signed_block, beacon_node)
|
|
.await
|
|
},
|
|
)
|
|
.await?;
|
|
|
|
info!(
|
|
log,
|
|
"Successfully published block";
|
|
"block_type" => ?signed_block.block_type(),
|
|
"deposits" => signed_block.num_deposits(),
|
|
"attestations" => signed_block.num_attestations(),
|
|
"graffiti" => ?graffiti.map(|g| g.as_utf8_lossy()),
|
|
"slot" => signed_block.slot().as_u64(),
|
|
);
|
|
Ok(())
|
|
}
|
|
|
|
async fn publish_block_v3(
|
|
self,
|
|
slot: Slot,
|
|
validator_pubkey: PublicKeyBytes,
|
|
builder_boost_factor: Option<u64>,
|
|
) -> Result<(), BlockError> {
|
|
let log = self.context.log();
|
|
let _timer =
|
|
metrics::start_timer_vec(&metrics::BLOCK_SERVICE_TIMES, &[metrics::BEACON_BLOCK]);
|
|
|
|
let randao_reveal = match self
|
|
.validator_store
|
|
.randao_reveal(validator_pubkey, slot.epoch(E::slots_per_epoch()))
|
|
.await
|
|
{
|
|
Ok(signature) => signature.into(),
|
|
Err(ValidatorStoreError::UnknownPubkey(pubkey)) => {
|
|
// A pubkey can be missing when a validator was recently removed
|
|
// via the API.
|
|
warn!(
|
|
log,
|
|
"Missing pubkey for block randao";
|
|
"info" => "a validator may have recently been removed from this VC",
|
|
"pubkey" => ?pubkey,
|
|
"slot" => ?slot
|
|
);
|
|
return Ok(());
|
|
}
|
|
Err(e) => {
|
|
return Err(BlockError::Recoverable(format!(
|
|
"Unable to produce randao reveal signature: {:?}",
|
|
e
|
|
)))
|
|
}
|
|
};
|
|
|
|
let graffiti = determine_graffiti(
|
|
&validator_pubkey,
|
|
log,
|
|
self.graffiti_file.clone(),
|
|
self.validator_store.graffiti(&validator_pubkey),
|
|
self.graffiti,
|
|
);
|
|
|
|
let randao_reveal_ref = &randao_reveal;
|
|
let self_ref = &self;
|
|
let proposer_index = self.validator_store.validator_index(&validator_pubkey);
|
|
let proposer_fallback = ProposerFallback {
|
|
beacon_nodes: self.beacon_nodes.clone(),
|
|
proposer_nodes: self.proposer_nodes.clone(),
|
|
};
|
|
|
|
info!(
|
|
log,
|
|
"Requesting unsigned block";
|
|
"slot" => slot.as_u64(),
|
|
);
|
|
|
|
// Request block from first responsive beacon node.
|
|
//
|
|
// Try the proposer nodes last, since it's likely that they don't have a
|
|
// great view of attestations on the network.
|
|
let unsigned_block = proposer_fallback
|
|
.request_proposers_last(
|
|
RequireSynced::No,
|
|
OfflineOnFailure::Yes,
|
|
|beacon_node| async move {
|
|
let _get_timer = metrics::start_timer_vec(
|
|
&metrics::BLOCK_SERVICE_TIMES,
|
|
&[metrics::BEACON_BLOCK_HTTP_GET],
|
|
);
|
|
let block_response = Self::get_validator_block_v3(
|
|
beacon_node,
|
|
slot,
|
|
randao_reveal_ref,
|
|
graffiti,
|
|
proposer_index,
|
|
builder_boost_factor,
|
|
log,
|
|
)
|
|
.await
|
|
.map_err(|e| {
|
|
BlockError::Recoverable(format!(
|
|
"Error from beacon node when producing block: {:?}",
|
|
e
|
|
))
|
|
});
|
|
|
|
Ok::<_, BlockError>(block_response)
|
|
},
|
|
)
|
|
.await??;
|
|
|
|
self_ref
|
|
.sign_and_publish_block(
|
|
proposer_fallback,
|
|
slot,
|
|
graffiti,
|
|
&validator_pubkey,
|
|
unsigned_block,
|
|
)
|
|
.await?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Produce a block at the given slot for validator_pubkey
|
|
async fn publish_block(
|
|
&self,
|
|
slot: Slot,
|
|
validator_pubkey: PublicKeyBytes,
|
|
builder_proposal: bool,
|
|
) -> Result<(), BlockError> {
|
|
let log = self.context.log();
|
|
let _timer =
|
|
metrics::start_timer_vec(&metrics::BLOCK_SERVICE_TIMES, &[metrics::BEACON_BLOCK]);
|
|
|
|
let randao_reveal = match self
|
|
.validator_store
|
|
.randao_reveal(validator_pubkey, slot.epoch(E::slots_per_epoch()))
|
|
.await
|
|
{
|
|
Ok(signature) => signature.into(),
|
|
Err(ValidatorStoreError::UnknownPubkey(pubkey)) => {
|
|
// A pubkey can be missing when a validator was recently removed
|
|
// via the API.
|
|
warn!(
|
|
log,
|
|
"Missing pubkey for block";
|
|
"info" => "a validator may have recently been removed from this VC",
|
|
"pubkey" => ?pubkey,
|
|
"slot" => ?slot
|
|
);
|
|
return Ok(());
|
|
}
|
|
Err(e) => {
|
|
return Err(BlockError::Recoverable(format!(
|
|
"Unable to sign block: {:?}",
|
|
e
|
|
)))
|
|
}
|
|
};
|
|
|
|
let graffiti = determine_graffiti(
|
|
&validator_pubkey,
|
|
log,
|
|
self.graffiti_file.clone(),
|
|
self.validator_store.graffiti(&validator_pubkey),
|
|
self.graffiti,
|
|
);
|
|
|
|
let randao_reveal_ref = &randao_reveal;
|
|
let self_ref = &self;
|
|
let proposer_index = self.validator_store.validator_index(&validator_pubkey);
|
|
let proposer_fallback = ProposerFallback {
|
|
beacon_nodes: self.beacon_nodes.clone(),
|
|
proposer_nodes: self.proposer_nodes.clone(),
|
|
};
|
|
|
|
info!(
|
|
log,
|
|
"Requesting unsigned block";
|
|
"slot" => slot.as_u64(),
|
|
);
|
|
|
|
// Request block from first responsive beacon node.
|
|
//
|
|
// Try the proposer nodes last, since it's likely that they don't have a
|
|
// great view of attestations on the network.
|
|
let unsigned_block = proposer_fallback
|
|
.request_proposers_last(
|
|
RequireSynced::No,
|
|
OfflineOnFailure::Yes,
|
|
move |beacon_node| {
|
|
Self::get_validator_block(
|
|
beacon_node,
|
|
slot,
|
|
randao_reveal_ref,
|
|
graffiti,
|
|
proposer_index,
|
|
builder_proposal,
|
|
log,
|
|
)
|
|
},
|
|
)
|
|
.await?;
|
|
|
|
self_ref
|
|
.sign_and_publish_block(
|
|
proposer_fallback,
|
|
slot,
|
|
graffiti,
|
|
&validator_pubkey,
|
|
unsigned_block,
|
|
)
|
|
.await?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
async fn publish_signed_block_contents(
|
|
&self,
|
|
signed_block: &SignedBlock<E>,
|
|
beacon_node: &BeaconNodeHttpClient,
|
|
) -> Result<(), BlockError> {
|
|
let log = self.context.log();
|
|
let slot = signed_block.slot();
|
|
match signed_block {
|
|
SignedBlock::Full(signed_block) => {
|
|
let _post_timer = metrics::start_timer_vec(
|
|
&metrics::BLOCK_SERVICE_TIMES,
|
|
&[metrics::BEACON_BLOCK_HTTP_POST],
|
|
);
|
|
beacon_node
|
|
.post_beacon_blocks(signed_block)
|
|
.await
|
|
.or_else(|e| handle_block_post_error(e, slot, log))?
|
|
}
|
|
SignedBlock::Blinded(signed_block) => {
|
|
let _post_timer = metrics::start_timer_vec(
|
|
&metrics::BLOCK_SERVICE_TIMES,
|
|
&[metrics::BLINDED_BEACON_BLOCK_HTTP_POST],
|
|
);
|
|
beacon_node
|
|
.post_beacon_blinded_blocks(signed_block)
|
|
.await
|
|
.or_else(|e| handle_block_post_error(e, slot, log))?
|
|
}
|
|
}
|
|
Ok::<_, BlockError>(())
|
|
}
|
|
|
|
async fn get_validator_block_v3(
|
|
beacon_node: &BeaconNodeHttpClient,
|
|
slot: Slot,
|
|
randao_reveal_ref: &SignatureBytes,
|
|
graffiti: Option<Graffiti>,
|
|
proposer_index: Option<u64>,
|
|
builder_boost_factor: Option<u64>,
|
|
log: &Logger,
|
|
) -> Result<UnsignedBlock<E>, BlockError> {
|
|
let (block_response, _) = beacon_node
|
|
.get_validator_blocks_v3::<E>(
|
|
slot,
|
|
randao_reveal_ref,
|
|
graffiti.as_ref(),
|
|
builder_boost_factor,
|
|
)
|
|
.await
|
|
.map_err(|e| {
|
|
BlockError::Recoverable(format!(
|
|
"Error from beacon node when producing block: {:?}",
|
|
e
|
|
))
|
|
})?;
|
|
|
|
let unsigned_block = match block_response.data {
|
|
eth2::types::ProduceBlockV3Response::Full(block) => UnsignedBlock::Full(block),
|
|
eth2::types::ProduceBlockV3Response::Blinded(block) => UnsignedBlock::Blinded(block),
|
|
};
|
|
|
|
info!(
|
|
log,
|
|
"Received unsigned block";
|
|
"slot" => slot.as_u64(),
|
|
);
|
|
if proposer_index != Some(unsigned_block.proposer_index()) {
|
|
return Err(BlockError::Recoverable(
|
|
"Proposer index does not match block proposer. Beacon chain re-orged".to_string(),
|
|
));
|
|
}
|
|
|
|
Ok::<_, BlockError>(unsigned_block)
|
|
}
|
|
|
|
async fn get_validator_block(
|
|
beacon_node: &BeaconNodeHttpClient,
|
|
slot: Slot,
|
|
randao_reveal_ref: &SignatureBytes,
|
|
graffiti: Option<Graffiti>,
|
|
proposer_index: Option<u64>,
|
|
builder_proposal: bool,
|
|
log: &Logger,
|
|
) -> Result<UnsignedBlock<E>, BlockError> {
|
|
let unsigned_block = if !builder_proposal {
|
|
let _get_timer = metrics::start_timer_vec(
|
|
&metrics::BLOCK_SERVICE_TIMES,
|
|
&[metrics::BEACON_BLOCK_HTTP_GET],
|
|
);
|
|
UnsignedBlock::Full(
|
|
beacon_node
|
|
.get_validator_blocks::<E>(slot, randao_reveal_ref, graffiti.as_ref())
|
|
.await
|
|
.map_err(|e| {
|
|
BlockError::Recoverable(format!(
|
|
"Error from beacon node when producing block: {:?}",
|
|
e
|
|
))
|
|
})?
|
|
.data,
|
|
)
|
|
} else {
|
|
let _get_timer = metrics::start_timer_vec(
|
|
&metrics::BLOCK_SERVICE_TIMES,
|
|
&[metrics::BLINDED_BEACON_BLOCK_HTTP_GET],
|
|
);
|
|
UnsignedBlock::Blinded(
|
|
beacon_node
|
|
.get_validator_blinded_blocks::<E>(slot, randao_reveal_ref, graffiti.as_ref())
|
|
.await
|
|
.map_err(|e| {
|
|
BlockError::Recoverable(format!(
|
|
"Error from beacon node when producing block: {:?}",
|
|
e
|
|
))
|
|
})?
|
|
.data,
|
|
)
|
|
};
|
|
|
|
info!(
|
|
log,
|
|
"Received unsigned block";
|
|
"slot" => slot.as_u64(),
|
|
);
|
|
if proposer_index != Some(unsigned_block.proposer_index()) {
|
|
return Err(BlockError::Recoverable(
|
|
"Proposer index does not match block proposer. Beacon chain re-orged".to_string(),
|
|
));
|
|
}
|
|
|
|
Ok::<_, BlockError>(unsigned_block)
|
|
}
|
|
|
|
/// Returns the builder boost factor of the given public key.
|
|
/// The priority order for fetching this value is:
|
|
///
|
|
/// 1. validator_definitions.yml
|
|
/// 2. process level flag
|
|
fn get_builder_boost_factor(&self, validator_pubkey: &PublicKeyBytes) -> Option<u64> {
|
|
// Apply per validator configuration first.
|
|
let validator_builder_boost_factor = self
|
|
.validator_store
|
|
.determine_validator_builder_boost_factor(validator_pubkey);
|
|
|
|
// Fallback to process-wide configuration if needed.
|
|
let maybe_builder_boost_factor = validator_builder_boost_factor.or_else(|| {
|
|
self.validator_store
|
|
.determine_default_builder_boost_factor()
|
|
});
|
|
|
|
if let Some(builder_boost_factor) = maybe_builder_boost_factor {
|
|
// if builder boost factor is set to 100 it should be treated
|
|
// as None to prevent unnecessary calculations that could
|
|
// lead to loss of information.
|
|
if builder_boost_factor == 100 {
|
|
return None;
|
|
}
|
|
return Some(builder_boost_factor);
|
|
}
|
|
|
|
None
|
|
}
|
|
}
|
|
|
|
pub enum UnsignedBlock<E: EthSpec> {
|
|
Full(FullBlockContents<E>),
|
|
Blinded(BlindedBeaconBlock<E>),
|
|
}
|
|
|
|
impl<E: EthSpec> UnsignedBlock<E> {
|
|
pub fn proposer_index(&self) -> u64 {
|
|
match self {
|
|
UnsignedBlock::Full(block) => block.block().proposer_index(),
|
|
UnsignedBlock::Blinded(block) => block.proposer_index(),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
pub enum SignedBlock<E: EthSpec> {
|
|
Full(PublishBlockRequest<E>),
|
|
Blinded(Arc<SignedBlindedBeaconBlock<E>>),
|
|
}
|
|
|
|
impl<E: EthSpec> SignedBlock<E> {
|
|
pub fn block_type(&self) -> BlockType {
|
|
match self {
|
|
SignedBlock::Full(_) => BlockType::Full,
|
|
SignedBlock::Blinded(_) => BlockType::Blinded,
|
|
}
|
|
}
|
|
pub fn slot(&self) -> Slot {
|
|
match self {
|
|
SignedBlock::Full(block) => block.signed_block().message().slot(),
|
|
SignedBlock::Blinded(block) => block.message().slot(),
|
|
}
|
|
}
|
|
pub fn num_deposits(&self) -> usize {
|
|
match self {
|
|
SignedBlock::Full(block) => block.signed_block().message().body().deposits().len(),
|
|
SignedBlock::Blinded(block) => block.message().body().deposits().len(),
|
|
}
|
|
}
|
|
pub fn num_attestations(&self) -> usize {
|
|
match self {
|
|
SignedBlock::Full(block) => block.signed_block().message().body().attestations_len(),
|
|
SignedBlock::Blinded(block) => block.message().body().attestations_len(),
|
|
}
|
|
}
|
|
}
|
|
|
|
fn handle_block_post_error(err: eth2::Error, slot: Slot, log: &Logger) -> Result<(), BlockError> {
|
|
// Handle non-200 success codes.
|
|
if let Some(status) = err.status() {
|
|
if status == StatusCode::ACCEPTED {
|
|
info!(
|
|
log,
|
|
"Block is already known to BN or might be invalid";
|
|
"slot" => slot,
|
|
"status_code" => status.as_u16(),
|
|
);
|
|
return Ok(());
|
|
} else if status.is_success() {
|
|
debug!(
|
|
log,
|
|
"Block published with non-standard success code";
|
|
"slot" => slot,
|
|
"status_code" => status.as_u16(),
|
|
);
|
|
return Ok(());
|
|
}
|
|
}
|
|
Err(BlockError::Irrecoverable(format!(
|
|
"Error from beacon node when publishing block: {err:?}",
|
|
)))
|
|
}
|