mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-03 00:31:50 +00:00
* Upgrade discv5 to v0.8 * Rename some logs * Improve the NAT reporting with new discv5 metrics * Merge branch 'unstable' into discv5-v8 * Limited Cargo.lock update * Update yanked futures-* crates
1339 lines
54 KiB
Rust
1339 lines
54 KiB
Rust
//! The discovery sub-behaviour of Lighthouse.
|
||
//!
|
||
//! This module creates a libp2p dummy-behaviour built around the discv5 protocol. It handles
|
||
//! queries and manages access to the discovery routing table.
|
||
|
||
pub(crate) mod enr;
|
||
pub mod enr_ext;
|
||
|
||
// Allow external use of the lighthouse ENR builder
|
||
use crate::service::TARGET_SUBNET_PEERS;
|
||
use crate::{error, Enr, NetworkConfig, NetworkGlobals, Subnet, SubnetDiscovery};
|
||
use crate::{metrics, ClearDialError};
|
||
use discv5::{enr::NodeId, Discv5};
|
||
pub use enr::{build_enr, load_enr_from_disk, use_or_load_enr, CombinedKey, Eth2Enr};
|
||
pub use enr_ext::{peer_id_to_node_id, CombinedKeyExt, EnrExt};
|
||
pub use libp2p::identity::{Keypair, PublicKey};
|
||
|
||
use alloy_rlp::bytes::Bytes;
|
||
use enr::{ATTESTATION_BITFIELD_ENR_KEY, ETH2_ENR_KEY, SYNC_COMMITTEE_BITFIELD_ENR_KEY};
|
||
use futures::prelude::*;
|
||
use futures::stream::FuturesUnordered;
|
||
use libp2p::core::transport::PortUse;
|
||
use libp2p::multiaddr::Protocol;
|
||
use libp2p::swarm::behaviour::{DialFailure, FromSwarm};
|
||
use libp2p::swarm::THandlerInEvent;
|
||
pub use libp2p::{
|
||
core::{transport::ListenerId, ConnectedPoint, Multiaddr},
|
||
identity::PeerId,
|
||
swarm::{
|
||
dummy::ConnectionHandler, ConnectionId, DialError, NetworkBehaviour, NotifyHandler,
|
||
SubstreamProtocol, ToSwarm,
|
||
},
|
||
};
|
||
use lru::LruCache;
|
||
use slog::{crit, debug, error, info, trace, warn};
|
||
use ssz::Encode;
|
||
use std::num::NonZeroUsize;
|
||
use std::{
|
||
collections::{HashMap, VecDeque},
|
||
net::{IpAddr, SocketAddr},
|
||
path::Path,
|
||
pin::Pin,
|
||
sync::Arc,
|
||
task::{Context, Poll},
|
||
time::{Duration, Instant},
|
||
};
|
||
use tokio::sync::mpsc;
|
||
use types::{ChainSpec, EnrForkId, EthSpec};
|
||
|
||
mod subnet_predicate;
|
||
pub use subnet_predicate::subnet_predicate;
|
||
use types::non_zero_usize::new_non_zero_usize;
|
||
|
||
/// Local ENR storage filename.
|
||
pub const ENR_FILENAME: &str = "enr.dat";
|
||
/// Target number of peers to search for given a grouped subnet query.
|
||
const TARGET_PEERS_FOR_GROUPED_QUERY: usize = 6;
|
||
/// Number of times to attempt a discovery request.
|
||
const MAX_DISCOVERY_RETRY: usize = 3;
|
||
/// The maximum number of concurrent subnet discovery queries.
|
||
/// Note: we always allow a single FindPeers query, so we would be
|
||
/// running a maximum of `MAX_CONCURRENT_SUBNET_QUERIES + 1`
|
||
/// discovery queries at a time.
|
||
const MAX_CONCURRENT_SUBNET_QUERIES: usize = 4;
|
||
/// The max number of subnets to search for in a single subnet discovery query.
|
||
const MAX_SUBNETS_IN_QUERY: usize = 3;
|
||
/// The number of closest peers to search for when doing a regular peer search.
|
||
///
|
||
/// We could reduce this constant to speed up queries however at the cost of security. It will
|
||
/// make it easier to peers to eclipse this node. Kademlia suggests a value of 16.
|
||
pub const FIND_NODE_QUERY_CLOSEST_PEERS: usize = 16;
|
||
/// The threshold for updating `min_ttl` on a connected peer.
|
||
const DURATION_DIFFERENCE: Duration = Duration::from_millis(1);
|
||
/// The capacity of the Discovery ENR cache.
|
||
const ENR_CACHE_CAPACITY: NonZeroUsize = new_non_zero_usize(50);
|
||
|
||
/// A query has completed. This result contains a mapping of discovered peer IDs to the `min_ttl`
|
||
/// of the peer if it is specified.
|
||
#[derive(Debug)]
|
||
pub struct DiscoveredPeers {
|
||
pub peers: HashMap<Enr, Option<Instant>>,
|
||
}
|
||
|
||
/// Specifies which port numbers should be modified after start of the discovery service
|
||
#[derive(Debug)]
|
||
pub struct UpdatePorts {
|
||
/// TCP port associated wih IPv4 address (if present)
|
||
pub tcp4: bool,
|
||
/// TCP port associated wih IPv6 address (if present)
|
||
pub tcp6: bool,
|
||
/// QUIC port associated wih IPv4 address (if present)
|
||
pub quic4: bool,
|
||
/// QUIC port associated wih IPv6 address (if present)
|
||
pub quic6: bool,
|
||
}
|
||
|
||
#[derive(Clone, PartialEq)]
|
||
struct SubnetQuery {
|
||
subnet: Subnet,
|
||
min_ttl: Option<Instant>,
|
||
retries: usize,
|
||
}
|
||
|
||
impl SubnetQuery {
|
||
/// Returns true if this query has expired.
|
||
pub fn expired(&self) -> bool {
|
||
if let Some(ttl) = self.min_ttl {
|
||
ttl < Instant::now()
|
||
}
|
||
// `None` corresponds to long lived subnet discovery requests.
|
||
else {
|
||
false
|
||
}
|
||
}
|
||
}
|
||
|
||
impl std::fmt::Debug for SubnetQuery {
|
||
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
|
||
let min_ttl_secs = self
|
||
.min_ttl
|
||
.map(|ttl| ttl.saturating_duration_since(Instant::now()).as_secs());
|
||
f.debug_struct("SubnetQuery")
|
||
.field("subnet", &self.subnet)
|
||
.field("min_ttl_secs", &min_ttl_secs)
|
||
.field("retries", &self.retries)
|
||
.finish()
|
||
}
|
||
}
|
||
|
||
#[derive(Debug, Clone, PartialEq)]
|
||
enum QueryType {
|
||
/// We are searching for subnet peers.
|
||
Subnet(Vec<SubnetQuery>),
|
||
/// We are searching for more peers without ENR or time constraints.
|
||
FindPeers,
|
||
}
|
||
|
||
/// The result of a query.
|
||
struct QueryResult {
|
||
query_type: QueryType,
|
||
result: Result<Vec<Enr>, discv5::QueryError>,
|
||
}
|
||
|
||
// Awaiting the event stream future
|
||
enum EventStream {
|
||
/// Awaiting an event stream to be generated. This is required due to the poll nature of
|
||
/// `Discovery`
|
||
Awaiting(
|
||
Pin<Box<dyn Future<Output = Result<mpsc::Receiver<discv5::Event>, discv5::Error>> + Send>>,
|
||
),
|
||
/// The future has completed.
|
||
Present(mpsc::Receiver<discv5::Event>),
|
||
// The future has failed or discv5 has been disabled. There are no events from discv5.
|
||
InActive,
|
||
}
|
||
|
||
/// The main discovery service. This can be disabled via CLI arguements. When disabled the
|
||
/// underlying processes are not started, but this struct still maintains our current ENR.
|
||
pub struct Discovery<E: EthSpec> {
|
||
/// A collection of seen live ENRs for quick lookup and to map peer-id's to ENRs.
|
||
cached_enrs: LruCache<PeerId, Enr>,
|
||
|
||
/// The directory where the ENR is stored.
|
||
enr_dir: String,
|
||
|
||
/// The handle for the underlying discv5 Server.
|
||
///
|
||
/// This is behind a Reference counter to allow for futures to be spawned and polled with a
|
||
/// static lifetime.
|
||
discv5: Discv5,
|
||
|
||
/// A collection of network constants that can be read from other threads.
|
||
network_globals: Arc<NetworkGlobals<E>>,
|
||
|
||
/// Indicates if we are actively searching for peers. We only allow a single FindPeers query at
|
||
/// a time, regardless of the query concurrency.
|
||
find_peer_active: bool,
|
||
|
||
/// A queue of subnet queries to be processed.
|
||
queued_queries: VecDeque<SubnetQuery>,
|
||
|
||
/// Active discovery queries.
|
||
active_queries: FuturesUnordered<std::pin::Pin<Box<dyn Future<Output = QueryResult> + Send>>>,
|
||
|
||
/// The discv5 event stream.
|
||
event_stream: EventStream,
|
||
|
||
/// Indicates if the discovery service has been started. When the service is disabled, this is
|
||
/// always false.
|
||
pub started: bool,
|
||
|
||
/// Specifies whether various port numbers should be updated after the discovery service has been started
|
||
update_ports: UpdatePorts,
|
||
|
||
/// Logger for the discovery behaviour.
|
||
log: slog::Logger,
|
||
spec: Arc<ChainSpec>,
|
||
}
|
||
|
||
impl<E: EthSpec> Discovery<E> {
|
||
/// NOTE: Creating discovery requires running within a tokio execution environment.
|
||
pub async fn new(
|
||
local_key: Keypair,
|
||
config: &NetworkConfig,
|
||
network_globals: Arc<NetworkGlobals<E>>,
|
||
log: &slog::Logger,
|
||
spec: &ChainSpec,
|
||
) -> error::Result<Self> {
|
||
let log = log.clone();
|
||
|
||
let enr_dir = match config.network_dir.to_str() {
|
||
Some(path) => String::from(path),
|
||
None => String::from(""),
|
||
};
|
||
|
||
let local_enr = network_globals.local_enr.read().clone();
|
||
let local_node_id = local_enr.node_id();
|
||
|
||
info!(log, "ENR Initialised"; "enr" => local_enr.to_base64(), "seq" => local_enr.seq(), "id"=> %local_enr.node_id(),
|
||
"ip4" => ?local_enr.ip4(), "udp4"=> ?local_enr.udp4(), "tcp4" => ?local_enr.tcp4(), "tcp6" => ?local_enr.tcp6(), "udp6" => ?local_enr.udp6(),
|
||
"quic4" => ?local_enr.quic4(), "quic6" => ?local_enr.quic6()
|
||
);
|
||
|
||
// convert the keypair into an ENR key
|
||
let enr_key: CombinedKey = CombinedKey::from_libp2p(local_key)?;
|
||
|
||
let mut discv5 = Discv5::new(local_enr, enr_key, config.discv5_config.clone())
|
||
.map_err(|e| format!("Discv5 service failed. Error: {:?}", e))?;
|
||
|
||
// Add bootnodes to routing table
|
||
for bootnode_enr in config.boot_nodes_enr.clone() {
|
||
if bootnode_enr.node_id() == local_node_id {
|
||
// If we are a boot node, ignore adding it to the routing table
|
||
continue;
|
||
}
|
||
debug!(
|
||
log,
|
||
"Adding node to routing table";
|
||
"node_id" => %bootnode_enr.node_id(),
|
||
"peer_id" => %bootnode_enr.peer_id(),
|
||
"ip" => ?bootnode_enr.ip4(),
|
||
"udp" => ?bootnode_enr.udp4(),
|
||
"tcp" => ?bootnode_enr.tcp4(),
|
||
"quic" => ?bootnode_enr.quic4()
|
||
);
|
||
let repr = bootnode_enr.to_string();
|
||
let _ = discv5.add_enr(bootnode_enr).map_err(|e| {
|
||
error!(
|
||
log,
|
||
"Could not add peer to the local routing table";
|
||
"addr" => repr,
|
||
"error" => e.to_string(),
|
||
)
|
||
});
|
||
}
|
||
|
||
// Start the discv5 service and obtain an event stream
|
||
let event_stream = if !config.disable_discovery {
|
||
discv5.start().map_err(|e| e.to_string()).await?;
|
||
debug!(log, "Discovery service started");
|
||
EventStream::Awaiting(Box::pin(discv5.event_stream()))
|
||
} else {
|
||
EventStream::InActive
|
||
};
|
||
|
||
if !config.boot_nodes_multiaddr.is_empty() {
|
||
info!(log, "Contacting Multiaddr boot-nodes for their ENR");
|
||
}
|
||
|
||
// get futures for requesting the Enrs associated to these multiaddr and wait for their
|
||
// completion
|
||
let mut fut_coll = config
|
||
.boot_nodes_multiaddr
|
||
.iter()
|
||
.map(|addr| addr.to_string())
|
||
// request the ENR for this multiaddr and keep the original for logging
|
||
.map(|addr| {
|
||
futures::future::join(
|
||
discv5.request_enr(addr.clone()),
|
||
futures::future::ready(addr),
|
||
)
|
||
})
|
||
.collect::<FuturesUnordered<_>>();
|
||
|
||
while let Some((result, original_addr)) = fut_coll.next().await {
|
||
match result {
|
||
Ok(enr) => {
|
||
debug!(
|
||
log,
|
||
"Adding node to routing table";
|
||
"node_id" => %enr.node_id(),
|
||
"peer_id" => %enr.peer_id(),
|
||
"ip" => ?enr.ip4(),
|
||
"udp" => ?enr.udp4(),
|
||
"tcp" => ?enr.tcp4(),
|
||
"quic" => ?enr.quic4()
|
||
);
|
||
let _ = discv5.add_enr(enr).map_err(|e| {
|
||
error!(
|
||
log,
|
||
"Could not add peer to the local routing table";
|
||
"addr" => original_addr.to_string(),
|
||
"error" => e.to_string(),
|
||
)
|
||
});
|
||
}
|
||
Err(e) => {
|
||
error!(log, "Error getting mapping to ENR"; "multiaddr" => original_addr.to_string(), "error" => e.to_string())
|
||
}
|
||
}
|
||
}
|
||
|
||
let update_ports = UpdatePorts {
|
||
tcp4: config.enr_tcp4_port.is_none(),
|
||
tcp6: config.enr_tcp6_port.is_none(),
|
||
quic4: config.enr_quic4_port.is_none(),
|
||
quic6: config.enr_quic6_port.is_none(),
|
||
};
|
||
|
||
Ok(Self {
|
||
cached_enrs: LruCache::new(ENR_CACHE_CAPACITY),
|
||
network_globals,
|
||
find_peer_active: false,
|
||
queued_queries: VecDeque::with_capacity(10),
|
||
active_queries: FuturesUnordered::new(),
|
||
discv5,
|
||
event_stream,
|
||
started: !config.disable_discovery,
|
||
update_ports,
|
||
log,
|
||
enr_dir,
|
||
spec: Arc::new(spec.clone()),
|
||
})
|
||
}
|
||
|
||
/// Return the nodes local ENR.
|
||
pub fn local_enr(&self) -> Enr {
|
||
self.discv5.local_enr()
|
||
}
|
||
|
||
/// Return the cached enrs.
|
||
pub fn cached_enrs(&self) -> impl Iterator<Item = (&PeerId, &Enr)> {
|
||
self.cached_enrs.iter()
|
||
}
|
||
|
||
/// Removes a cached ENR from the list.
|
||
pub fn remove_cached_enr(&mut self, peer_id: &PeerId) -> Option<Enr> {
|
||
self.cached_enrs.pop(peer_id)
|
||
}
|
||
|
||
/// This adds a new `FindPeers` query to the queue if one doesn't already exist.
|
||
/// The `target_peers` parameter informs discovery to end the query once the target is found.
|
||
/// The maximum this can be is 16.
|
||
pub fn discover_peers(&mut self, target_peers: usize) {
|
||
// If the discv5 service isn't running or we are in the process of a query, don't bother queuing a new one.
|
||
if !self.started || self.find_peer_active {
|
||
return;
|
||
}
|
||
// Immediately start a FindNode query
|
||
let target_peers = std::cmp::min(FIND_NODE_QUERY_CLOSEST_PEERS, target_peers);
|
||
debug!(self.log, "Starting a peer discovery request"; "target_peers" => target_peers );
|
||
self.find_peer_active = true;
|
||
self.start_query(QueryType::FindPeers, target_peers, |_| true);
|
||
}
|
||
|
||
/// Processes a request to search for more peers on a subnet.
|
||
pub fn discover_subnet_peers(&mut self, subnets_to_discover: Vec<SubnetDiscovery>) {
|
||
// If the discv5 service isn't running, ignore queries
|
||
if !self.started {
|
||
return;
|
||
}
|
||
trace!(
|
||
self.log,
|
||
"Starting discovery query for subnets";
|
||
"subnets" => ?subnets_to_discover.iter().map(|s| s.subnet).collect::<Vec<_>>()
|
||
);
|
||
for subnet in subnets_to_discover {
|
||
self.add_subnet_query(subnet.subnet, subnet.min_ttl, 0);
|
||
}
|
||
}
|
||
|
||
/// Add an ENR to the routing table of the discovery mechanism.
|
||
pub fn add_enr(&mut self, enr: Enr) {
|
||
// add the enr to seen caches
|
||
self.cached_enrs.put(enr.peer_id(), enr.clone());
|
||
|
||
if let Err(e) = self.discv5.add_enr(enr) {
|
||
debug!(
|
||
self.log,
|
||
"Could not add peer to the local routing table";
|
||
"error" => %e
|
||
)
|
||
}
|
||
}
|
||
|
||
/// Returns an iterator over all enr entries in the DHT.
|
||
pub fn table_entries_enr(&self) -> Vec<Enr> {
|
||
self.discv5.table_entries_enr()
|
||
}
|
||
|
||
/// Updates the local ENR TCP port.
|
||
/// There currently isn't a case to update the address here. We opt for discovery to
|
||
/// automatically update the external address.
|
||
///
|
||
/// If the external address needs to be modified, use `update_enr_udp_socket.
|
||
///
|
||
/// This returns Ok(true) if the ENR was updated, otherwise Ok(false) if nothing was done.
|
||
pub fn update_enr_tcp_port(&mut self, port: u16, v6: bool) -> Result<bool, String> {
|
||
let enr_field = if v6 {
|
||
if self.discv5.external_enr().read().tcp6() == Some(port) {
|
||
// The field is already set to the same value, nothing to do
|
||
return Ok(false);
|
||
}
|
||
"tcp6"
|
||
} else {
|
||
if self.discv5.external_enr().read().tcp4() == Some(port) {
|
||
// The field is already set to the same value, nothing to do
|
||
return Ok(false);
|
||
}
|
||
"tcp"
|
||
};
|
||
|
||
self.discv5
|
||
.enr_insert(enr_field, &port)
|
||
.map_err(|e| format!("{:?}", e))?;
|
||
|
||
// replace the global version
|
||
*self.network_globals.local_enr.write() = self.discv5.local_enr();
|
||
// persist modified enr to disk
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &self.local_enr(), &self.log);
|
||
Ok(true)
|
||
}
|
||
|
||
// TODO: Group these functions here once the ENR is shared across discv5 and lighthouse and
|
||
// Lighthouse can modify the ENR directly.
|
||
// This currently doesn't support ipv6. All of these functions should be removed and
|
||
// addressed properly in the following issue.
|
||
// https://github.com/sigp/lighthouse/issues/4706
|
||
pub fn update_enr_quic_port(&mut self, port: u16, v6: bool) -> Result<bool, String> {
|
||
let enr_field = if v6 {
|
||
if self.discv5.external_enr().read().quic6() == Some(port) {
|
||
// The field is already set to the same value, nothing to do
|
||
return Ok(false);
|
||
}
|
||
"quic6"
|
||
} else {
|
||
if self.discv5.external_enr().read().quic4() == Some(port) {
|
||
// The field is already set to the same value, nothing to do
|
||
return Ok(false);
|
||
}
|
||
"quic"
|
||
};
|
||
let current_field = self.discv5.external_enr().read().quic4();
|
||
if current_field == Some(port) {
|
||
// The current field is already set, no need to update.
|
||
return Ok(false);
|
||
}
|
||
|
||
self.discv5
|
||
.enr_insert(enr_field, &port)
|
||
.map_err(|e| format!("{:?}", e))?;
|
||
|
||
// replace the global version
|
||
*self.network_globals.local_enr.write() = self.discv5.local_enr();
|
||
// persist modified enr to disk
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &self.local_enr(), &self.log);
|
||
Ok(true)
|
||
}
|
||
|
||
/// Updates the local ENR UDP socket.
|
||
///
|
||
/// This is with caution. Discovery should automatically maintain this. This should only be
|
||
/// used when automatic discovery is disabled.
|
||
pub fn update_enr_udp_socket(&mut self, socket_addr: SocketAddr) -> Result<(), String> {
|
||
const IS_TCP: bool = false;
|
||
if self.discv5.update_local_enr_socket(socket_addr, IS_TCP) {
|
||
// persist modified enr to disk
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &self.local_enr(), &self.log);
|
||
}
|
||
*self.network_globals.local_enr.write() = self.discv5.local_enr();
|
||
Ok(())
|
||
}
|
||
|
||
/// Adds/Removes a subnet from the ENR attnets/syncnets Bitfield
|
||
pub fn update_enr_bitfield(&mut self, subnet: Subnet, value: bool) -> Result<(), String> {
|
||
let local_enr = self.discv5.local_enr();
|
||
|
||
match subnet {
|
||
Subnet::Attestation(id) => {
|
||
let id = *id as usize;
|
||
let mut current_bitfield = local_enr.attestation_bitfield::<E>()?;
|
||
if id >= current_bitfield.len() {
|
||
return Err(format!(
|
||
"Subnet id: {} is outside the ENR bitfield length: {}",
|
||
id,
|
||
current_bitfield.len()
|
||
));
|
||
}
|
||
|
||
// The bitfield is already set to required value
|
||
if current_bitfield
|
||
.get(id)
|
||
.map_err(|_| String::from("Subnet ID out of bounds"))?
|
||
== value
|
||
{
|
||
return Ok(());
|
||
}
|
||
|
||
// set the subnet bitfield in the ENR
|
||
current_bitfield.set(id, value).map_err(|_| {
|
||
String::from("Subnet ID out of bounds, could not set subnet ID")
|
||
})?;
|
||
|
||
// insert the bitfield into the ENR record
|
||
self.discv5
|
||
.enr_insert::<Bytes>(
|
||
ATTESTATION_BITFIELD_ENR_KEY,
|
||
¤t_bitfield.as_ssz_bytes().into(),
|
||
)
|
||
.map_err(|e| format!("{:?}", e))?;
|
||
}
|
||
Subnet::SyncCommittee(id) => {
|
||
let id = *id as usize;
|
||
let mut current_bitfield = local_enr.sync_committee_bitfield::<E>()?;
|
||
|
||
if id >= current_bitfield.len() {
|
||
return Err(format!(
|
||
"Subnet id: {} is outside the ENR bitfield length: {}",
|
||
id,
|
||
current_bitfield.len()
|
||
));
|
||
}
|
||
|
||
// The bitfield is already set to required value
|
||
if current_bitfield
|
||
.get(id)
|
||
.map_err(|_| String::from("Subnet ID out of bounds"))?
|
||
== value
|
||
{
|
||
return Ok(());
|
||
}
|
||
|
||
// set the subnet bitfield in the ENR
|
||
current_bitfield.set(id, value).map_err(|_| {
|
||
String::from("Subnet ID out of bounds, could not set subnet ID")
|
||
})?;
|
||
|
||
// insert the bitfield into the ENR record
|
||
self.discv5
|
||
.enr_insert::<Bytes>(
|
||
SYNC_COMMITTEE_BITFIELD_ENR_KEY,
|
||
¤t_bitfield.as_ssz_bytes().into(),
|
||
)
|
||
.map_err(|e| format!("{:?}", e))?;
|
||
}
|
||
// Data column subnets are computed from node ID. No subnet bitfield in the ENR.
|
||
Subnet::DataColumn(_) => return Ok(()),
|
||
}
|
||
|
||
// replace the global version
|
||
*self.network_globals.local_enr.write() = self.discv5.local_enr();
|
||
|
||
// persist modified enr to disk
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &self.local_enr(), &self.log);
|
||
Ok(())
|
||
}
|
||
|
||
/// Updates the `eth2` field of our local ENR.
|
||
pub fn update_eth2_enr(&mut self, enr_fork_id: EnrForkId) {
|
||
// to avoid having a reference to the spec constant, for the logging we assume
|
||
// FAR_FUTURE_EPOCH is u64::MAX
|
||
let next_fork_epoch_log = if enr_fork_id.next_fork_epoch == u64::MAX {
|
||
String::from("No other fork")
|
||
} else {
|
||
format!("{:?}", enr_fork_id.next_fork_epoch)
|
||
};
|
||
|
||
info!(self.log, "Updating the ENR fork version";
|
||
"fork_digest" => ?enr_fork_id.fork_digest,
|
||
"next_fork_version" => ?enr_fork_id.next_fork_version,
|
||
"next_fork_epoch" => next_fork_epoch_log,
|
||
);
|
||
|
||
let _ = self
|
||
.discv5
|
||
.enr_insert::<Bytes>(ETH2_ENR_KEY, &enr_fork_id.as_ssz_bytes().into())
|
||
.map_err(|e| {
|
||
warn!(
|
||
self.log,
|
||
"Could not update eth2 ENR field";
|
||
"error" => ?e
|
||
)
|
||
});
|
||
|
||
// replace the global version with discovery version
|
||
*self.network_globals.local_enr.write() = self.discv5.local_enr();
|
||
|
||
// persist modified enr to disk
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &self.local_enr(), &self.log);
|
||
}
|
||
|
||
// Bans a peer and it's associated seen IP addresses.
|
||
pub fn ban_peer(&mut self, peer_id: &PeerId, ip_addresses: Vec<IpAddr>) {
|
||
// first try and convert the peer_id to a node_id.
|
||
if let Ok(node_id) = peer_id_to_node_id(peer_id) {
|
||
// If we could convert this peer id, remove it from the DHT and ban it from discovery.
|
||
self.discv5.ban_node(&node_id, None);
|
||
}
|
||
|
||
for ip_address in ip_addresses {
|
||
self.discv5.ban_ip(ip_address, None);
|
||
}
|
||
}
|
||
|
||
/// Unbans the peer in discovery.
|
||
pub fn unban_peer(&mut self, peer_id: &PeerId, ip_addresses: Vec<IpAddr>) {
|
||
// first try and convert the peer_id to a node_id.
|
||
if let Ok(node_id) = peer_id_to_node_id(peer_id) {
|
||
self.discv5.ban_node_remove(&node_id);
|
||
}
|
||
|
||
for ip_address in ip_addresses {
|
||
self.discv5.ban_ip_remove(&ip_address);
|
||
}
|
||
}
|
||
|
||
/// Marks node as disconnected in the DHT, freeing up space for other nodes, this also removes
|
||
/// nodes from the cached ENR list.
|
||
pub fn disconnect_peer(&mut self, peer_id: &PeerId) {
|
||
if let Ok(node_id) = peer_id_to_node_id(peer_id) {
|
||
self.discv5.disconnect_node(&node_id);
|
||
}
|
||
// Remove the peer from the cached list, to prevent redialing disconnected
|
||
// peers.
|
||
self.cached_enrs.pop(peer_id);
|
||
}
|
||
|
||
/* Internal Functions */
|
||
|
||
/// Adds a subnet query if one doesn't exist. If a subnet query already exists, this
|
||
/// updates the min_ttl field.
|
||
fn add_subnet_query(&mut self, subnet: Subnet, min_ttl: Option<Instant>, retries: usize) {
|
||
// remove the entry and complete the query if greater than the maximum search count
|
||
if retries > MAX_DISCOVERY_RETRY {
|
||
debug!(
|
||
self.log,
|
||
"Subnet peer discovery did not find sufficient peers. Reached max retry limit"
|
||
);
|
||
return;
|
||
}
|
||
|
||
// Search through any queued requests and update the timeout if a query for this subnet
|
||
// already exists
|
||
let mut found = false;
|
||
for subnet_query in self.queued_queries.iter_mut() {
|
||
if subnet_query.subnet == subnet {
|
||
if subnet_query.min_ttl < min_ttl {
|
||
subnet_query.min_ttl = min_ttl;
|
||
}
|
||
// update the number of retries
|
||
subnet_query.retries = retries;
|
||
// mimic an `Iter::Find()` and short-circuit the loop
|
||
found = true;
|
||
break;
|
||
}
|
||
}
|
||
if !found {
|
||
// update the metrics and insert into the queue.
|
||
trace!(self.log, "Queuing subnet query"; "subnet" => ?subnet, "retries" => retries);
|
||
self.queued_queries.push_back(SubnetQuery {
|
||
subnet,
|
||
min_ttl,
|
||
retries,
|
||
});
|
||
metrics::set_gauge(&metrics::DISCOVERY_QUEUE, self.queued_queries.len() as i64);
|
||
}
|
||
}
|
||
|
||
/// Consume the discovery queue and initiate queries when applicable.
|
||
///
|
||
/// This also sanitizes the queue removing out-dated queries.
|
||
/// Returns `true` if any of the queued queries is processed and a subnet discovery
|
||
/// query is started.
|
||
fn process_queue(&mut self) -> bool {
|
||
// Sanitize the queue, removing any out-dated subnet queries
|
||
self.queued_queries.retain(|query| !query.expired());
|
||
|
||
// use this to group subnet queries together for a single discovery request
|
||
let mut subnet_queries: Vec<SubnetQuery> = Vec::new();
|
||
let mut processed = false;
|
||
// Check that we are within our query concurrency limit
|
||
while !self.at_capacity() && !self.queued_queries.is_empty() {
|
||
// consume and process the query queue
|
||
if let Some(subnet_query) = self.queued_queries.pop_front() {
|
||
subnet_queries.push(subnet_query);
|
||
|
||
// We want to start a grouped subnet query if:
|
||
// 1. We've grouped MAX_SUBNETS_IN_QUERY subnets together.
|
||
// 2. There are no more messages in the queue.
|
||
if subnet_queries.len() == MAX_SUBNETS_IN_QUERY || self.queued_queries.is_empty() {
|
||
// This query is for searching for peers of a particular subnet
|
||
// Drain subnet_queries so we can re-use it as we continue to process the queue
|
||
let grouped_queries: Vec<SubnetQuery> = std::mem::take(&mut subnet_queries);
|
||
self.start_subnet_query(grouped_queries);
|
||
processed = true;
|
||
}
|
||
}
|
||
}
|
||
// Update the queue metric
|
||
metrics::set_gauge(&metrics::DISCOVERY_QUEUE, self.queued_queries.len() as i64);
|
||
processed
|
||
}
|
||
|
||
// Returns a boolean indicating if we are currently processing the maximum number of
|
||
// concurrent subnet queries or not.
|
||
fn at_capacity(&self) -> bool {
|
||
self.active_queries
|
||
.len()
|
||
.saturating_sub(self.find_peer_active as usize) // We only count active subnet queries
|
||
>= MAX_CONCURRENT_SUBNET_QUERIES
|
||
}
|
||
|
||
/// Runs a discovery request for a given group of subnets.
|
||
fn start_subnet_query(&mut self, subnet_queries: Vec<SubnetQuery>) {
|
||
let mut filtered_subnets: Vec<Subnet> = Vec::new();
|
||
|
||
// find subnet queries that are still necessary
|
||
let filtered_subnet_queries: Vec<SubnetQuery> = subnet_queries
|
||
.into_iter()
|
||
.filter(|subnet_query| {
|
||
// Determine if we have sufficient peers, which may make this discovery unnecessary.
|
||
let peers_on_subnet = self
|
||
.network_globals
|
||
.peers
|
||
.read()
|
||
.good_peers_on_subnet(subnet_query.subnet)
|
||
.count();
|
||
|
||
if peers_on_subnet >= TARGET_SUBNET_PEERS {
|
||
debug!(self.log, "Discovery ignored";
|
||
"reason" => "Already connected to desired peers",
|
||
"connected_peers_on_subnet" => peers_on_subnet,
|
||
"target_subnet_peers" => TARGET_SUBNET_PEERS,
|
||
);
|
||
return false;
|
||
}
|
||
|
||
let target_peers = TARGET_SUBNET_PEERS.saturating_sub(peers_on_subnet);
|
||
trace!(self.log, "Discovery query started for subnet";
|
||
"subnet_query" => ?subnet_query,
|
||
"connected_peers_on_subnet" => peers_on_subnet,
|
||
"peers_to_find" => target_peers,
|
||
);
|
||
|
||
filtered_subnets.push(subnet_query.subnet);
|
||
true
|
||
})
|
||
.collect();
|
||
|
||
// Only start a discovery query if we have a subnet to look for.
|
||
if !filtered_subnet_queries.is_empty() {
|
||
// build the subnet predicate as a combination of the eth2_fork_predicate and the subnet predicate
|
||
let subnet_predicate =
|
||
subnet_predicate::<E>(filtered_subnets, &self.log, self.spec.clone());
|
||
|
||
debug!(
|
||
self.log,
|
||
"Starting grouped subnet query";
|
||
"subnets" => ?filtered_subnet_queries,
|
||
);
|
||
self.start_query(
|
||
QueryType::Subnet(filtered_subnet_queries),
|
||
TARGET_PEERS_FOR_GROUPED_QUERY,
|
||
subnet_predicate,
|
||
);
|
||
}
|
||
}
|
||
|
||
/// Search for a specified number of new peers using the underlying discovery mechanism.
|
||
///
|
||
/// This can optionally search for peers for a given predicate. Regardless of the predicate
|
||
/// given, this will only search for peers on the same enr_fork_id as specified in the local
|
||
/// ENR.
|
||
fn start_query(
|
||
&mut self,
|
||
query: QueryType,
|
||
target_peers: usize,
|
||
additional_predicate: impl Fn(&Enr) -> bool + Send + 'static,
|
||
) {
|
||
let enr_fork_id = match self.local_enr().eth2() {
|
||
Ok(v) => v,
|
||
Err(e) => {
|
||
crit!(self.log, "Local ENR has no fork id"; "error" => e);
|
||
return;
|
||
}
|
||
};
|
||
// predicate for finding nodes with a matching fork and valid tcp port
|
||
let eth2_fork_predicate = move |enr: &Enr| {
|
||
// `next_fork_epoch` and `next_fork_version` can be different so that
|
||
// we can connect to peers who aren't compatible with an upcoming fork.
|
||
// `fork_digest` **must** be same.
|
||
enr.eth2().map(|e| e.fork_digest) == Ok(enr_fork_id.fork_digest)
|
||
&& (enr.tcp4().is_some() || enr.tcp6().is_some())
|
||
};
|
||
|
||
// General predicate
|
||
let predicate: Box<dyn Fn(&Enr) -> bool + Send> =
|
||
Box::new(move |enr: &Enr| eth2_fork_predicate(enr) && additional_predicate(enr));
|
||
|
||
// Build the future
|
||
let query_future = self
|
||
.discv5
|
||
// Generate a random target node id.
|
||
.find_node_predicate(NodeId::random(), predicate, target_peers)
|
||
.map(|v| QueryResult {
|
||
query_type: query,
|
||
result: v,
|
||
});
|
||
|
||
// Add the future to active queries, to be executed.
|
||
self.active_queries.push(Box::pin(query_future));
|
||
}
|
||
|
||
/// Process the completed QueryResult returned from discv5.
|
||
fn process_completed_queries(
|
||
&mut self,
|
||
query: QueryResult,
|
||
) -> Option<HashMap<Enr, Option<Instant>>> {
|
||
match query.query_type {
|
||
QueryType::FindPeers => {
|
||
self.find_peer_active = false;
|
||
match query.result {
|
||
Ok(r) if r.is_empty() => {
|
||
debug!(self.log, "Discovery query yielded no results.");
|
||
}
|
||
Ok(r) => {
|
||
debug!(self.log, "Discovery query completed"; "peers_found" => r.len());
|
||
let results = r
|
||
.into_iter()
|
||
.map(|enr| {
|
||
// cache the found ENR's
|
||
self.cached_enrs.put(enr.peer_id(), enr.clone());
|
||
(enr, None)
|
||
})
|
||
.collect();
|
||
return Some(results);
|
||
}
|
||
Err(e) => {
|
||
warn!(self.log, "Discovery query failed"; "error" => %e);
|
||
}
|
||
}
|
||
}
|
||
QueryType::Subnet(queries) => {
|
||
let subnets_searched_for: Vec<Subnet> =
|
||
queries.iter().map(|query| query.subnet).collect();
|
||
match query.result {
|
||
Ok(r) if r.is_empty() => {
|
||
debug!(self.log, "Grouped subnet discovery query yielded no results."; "subnets_searched_for" => ?subnets_searched_for);
|
||
queries.iter().for_each(|query| {
|
||
self.add_subnet_query(query.subnet, query.min_ttl, query.retries + 1);
|
||
})
|
||
}
|
||
Ok(r) => {
|
||
debug!(self.log, "Peer grouped subnet discovery request completed"; "peers_found" => r.len(), "subnets_searched_for" => ?subnets_searched_for);
|
||
|
||
let mut mapped_results = HashMap::new();
|
||
|
||
// cache the found ENR's
|
||
for enr in r.iter().cloned() {
|
||
self.cached_enrs.put(enr.peer_id(), enr);
|
||
}
|
||
|
||
// Map each subnet query's min_ttl to the set of ENR's returned for that subnet.
|
||
queries.iter().for_each(|query| {
|
||
let query_str = match query.subnet {
|
||
Subnet::Attestation(_) => "attestation",
|
||
Subnet::SyncCommittee(_) => "sync_committee",
|
||
Subnet::DataColumn(_) => "data_column",
|
||
};
|
||
|
||
if let Some(v) = metrics::get_int_counter(
|
||
&metrics::TOTAL_SUBNET_QUERIES,
|
||
&[query_str],
|
||
) {
|
||
v.inc();
|
||
}
|
||
// A subnet query has completed. Add back to the queue, incrementing retries.
|
||
self.add_subnet_query(query.subnet, query.min_ttl, query.retries + 1);
|
||
|
||
// Check the specific subnet against the enr
|
||
let subnet_predicate = subnet_predicate::<E>(
|
||
vec![query.subnet],
|
||
&self.log,
|
||
self.spec.clone(),
|
||
);
|
||
|
||
r.clone()
|
||
.into_iter()
|
||
.filter(|enr| subnet_predicate(enr))
|
||
.for_each(|enr| {
|
||
if let Some(v) = metrics::get_int_counter(
|
||
&metrics::SUBNET_PEERS_FOUND,
|
||
&[query_str],
|
||
) {
|
||
v.inc();
|
||
}
|
||
let other_min_ttl = mapped_results.get_mut(&enr);
|
||
|
||
// map peer IDs to the min_ttl furthest in the future
|
||
match (query.min_ttl, other_min_ttl) {
|
||
// update the mapping if the min_ttl is greater
|
||
(
|
||
Some(min_ttl_instant),
|
||
Some(Some(other_min_ttl_instant)),
|
||
) => {
|
||
if min_ttl_instant
|
||
.saturating_duration_since(*other_min_ttl_instant)
|
||
> DURATION_DIFFERENCE
|
||
{
|
||
*other_min_ttl_instant = min_ttl_instant;
|
||
}
|
||
}
|
||
// update the mapping if we have a specified min_ttl
|
||
(Some(min_ttl), Some(None)) => {
|
||
mapped_results.insert(enr, Some(min_ttl));
|
||
}
|
||
// first seen min_ttl for this enr
|
||
(min_ttl, None) => {
|
||
mapped_results.insert(enr, min_ttl);
|
||
}
|
||
(None, Some(Some(_))) => {} // Don't replace the existing specific min_ttl
|
||
(None, Some(None)) => {} // No-op because this is a duplicate
|
||
}
|
||
});
|
||
});
|
||
|
||
if mapped_results.is_empty() {
|
||
return None;
|
||
} else {
|
||
return Some(mapped_results);
|
||
}
|
||
}
|
||
Err(e) => {
|
||
warn!(self.log,"Grouped subnet discovery query failed"; "subnets_searched_for" => ?subnets_searched_for, "error" => %e);
|
||
}
|
||
}
|
||
}
|
||
}
|
||
None
|
||
}
|
||
|
||
/// Drives the queries returning any results from completed queries.
|
||
fn poll_queries(&mut self, cx: &mut Context) -> Option<HashMap<Enr, Option<Instant>>> {
|
||
while let Poll::Ready(Some(query_result)) = self.active_queries.poll_next_unpin(cx) {
|
||
let result = self.process_completed_queries(query_result);
|
||
if result.is_some() {
|
||
return result;
|
||
}
|
||
}
|
||
None
|
||
}
|
||
}
|
||
|
||
/* NetworkBehaviour Implementation */
|
||
|
||
impl<E: EthSpec> NetworkBehaviour for Discovery<E> {
|
||
// Discovery is not a real NetworkBehaviour...
|
||
type ConnectionHandler = ConnectionHandler;
|
||
type ToSwarm = DiscoveredPeers;
|
||
|
||
fn handle_established_inbound_connection(
|
||
&mut self,
|
||
_connection_id: ConnectionId,
|
||
_peer: PeerId,
|
||
_local_addr: &Multiaddr,
|
||
_remote_addr: &Multiaddr,
|
||
) -> Result<libp2p::swarm::THandler<Self>, libp2p::swarm::ConnectionDenied> {
|
||
// TODO: we might want to check discovery's banned ips here in the future.
|
||
Ok(ConnectionHandler)
|
||
}
|
||
|
||
fn handle_established_outbound_connection(
|
||
&mut self,
|
||
_connection_id: ConnectionId,
|
||
_peer: PeerId,
|
||
_addr: &Multiaddr,
|
||
_role_override: libp2p::core::Endpoint,
|
||
_port_use: PortUse,
|
||
) -> Result<libp2p::swarm::THandler<Self>, libp2p::swarm::ConnectionDenied> {
|
||
Ok(ConnectionHandler)
|
||
}
|
||
|
||
fn on_connection_handler_event(
|
||
&mut self,
|
||
_peer_id: PeerId,
|
||
_connection_id: ConnectionId,
|
||
_event: void::Void,
|
||
) {
|
||
}
|
||
|
||
// Main execution loop to drive the behaviour
|
||
fn poll(&mut self, cx: &mut Context) -> Poll<ToSwarm<Self::ToSwarm, THandlerInEvent<Self>>> {
|
||
if !self.started {
|
||
return Poll::Pending;
|
||
}
|
||
|
||
// Process the query queue
|
||
self.process_queue();
|
||
|
||
// Drive the queries and return any results from completed queries
|
||
if let Some(peers) = self.poll_queries(cx) {
|
||
// return the result to the peer manager
|
||
return Poll::Ready(ToSwarm::GenerateEvent(DiscoveredPeers { peers }));
|
||
}
|
||
|
||
// Process the server event stream
|
||
match self.event_stream {
|
||
EventStream::Awaiting(ref mut fut) => {
|
||
// Still awaiting the event stream, poll it
|
||
if let Poll::Ready(event_stream) = fut.poll_unpin(cx) {
|
||
match event_stream {
|
||
Ok(stream) => {
|
||
debug!(self.log, "Discv5 event stream ready");
|
||
self.event_stream = EventStream::Present(stream);
|
||
}
|
||
Err(e) => {
|
||
slog::crit!(self.log, "Discv5 event stream failed"; "error" => %e);
|
||
self.event_stream = EventStream::InActive;
|
||
}
|
||
}
|
||
}
|
||
}
|
||
EventStream::InActive => {} // ignore checking the stream
|
||
EventStream::Present(ref mut stream) => {
|
||
while let Poll::Ready(Some(event)) = stream.poll_recv(cx) {
|
||
match event {
|
||
// We filter out unwanted discv5 events here and only propagate useful results to
|
||
// the peer manager.
|
||
discv5::Event::Discovered(_enr) => {
|
||
// Peers that get discovered during a query but are not contactable or
|
||
// don't match a predicate can end up here. For debugging purposes we
|
||
// log these to see if we are unnecessarily dropping discovered peers
|
||
/*
|
||
if enr.eth2() == self.local_enr().eth2() {
|
||
trace!(self.log, "Peer found in process of query"; "peer_id" => format!("{}", enr.peer_id()), "tcp_socket" => enr.tcp_socket());
|
||
} else {
|
||
// this is temporary warning for debugging the DHT
|
||
warn!(self.log, "Found peer during discovery not on correct fork"; "peer_id" => format!("{}", enr.peer_id()), "tcp_socket" => enr.tcp_socket());
|
||
}
|
||
*/
|
||
}
|
||
discv5::Event::SocketUpdated(socket_addr) => {
|
||
info!(self.log, "Address updated"; "ip" => %socket_addr.ip(), "udp_port" => %socket_addr.port());
|
||
metrics::inc_counter(&metrics::ADDRESS_UPDATE_COUNT);
|
||
// Discv5 will have updated our local ENR. We save the updated version
|
||
// to disk.
|
||
|
||
if (self.update_ports.tcp4 && socket_addr.is_ipv4())
|
||
|| (self.update_ports.tcp6 && socket_addr.is_ipv6())
|
||
{
|
||
// Update the TCP port in the ENR
|
||
self.discv5.update_local_enr_socket(socket_addr, true);
|
||
}
|
||
let enr = self.discv5.local_enr();
|
||
enr::save_enr_to_disk(Path::new(&self.enr_dir), &enr, &self.log);
|
||
// update network globals
|
||
*self.network_globals.local_enr.write() = enr;
|
||
// A new UDP socket has been detected.
|
||
// NOTE: We assume libp2p itself can keep track of IP changes and we do
|
||
// not inform it about IP changes found via discovery.
|
||
}
|
||
_ => {} // Ignore all other discv5 server events
|
||
}
|
||
}
|
||
}
|
||
}
|
||
Poll::Pending
|
||
}
|
||
|
||
fn on_swarm_event(&mut self, event: FromSwarm) {
|
||
match event {
|
||
FromSwarm::DialFailure(DialFailure { peer_id, error, .. }) => {
|
||
self.on_dial_failure(peer_id, error)
|
||
}
|
||
FromSwarm::NewListenAddr(ev) => {
|
||
let addr = ev.addr;
|
||
let listener_id = ev.listener_id;
|
||
|
||
trace!(self.log, "Received NewListenAddr event from swarm"; "listener_id" => ?listener_id, "addr" => ?addr);
|
||
|
||
let mut addr_iter = addr.iter();
|
||
|
||
let attempt_enr_update = match addr_iter.next() {
|
||
Some(Protocol::Ip4(_)) => match (addr_iter.next(), addr_iter.next()) {
|
||
(Some(Protocol::Tcp(port)), None) => {
|
||
if !self.update_ports.tcp4 {
|
||
debug!(self.log, "Skipping ENR update"; "multiaddr" => ?addr);
|
||
return;
|
||
}
|
||
|
||
self.update_enr_tcp_port(port, false)
|
||
}
|
||
(Some(Protocol::Udp(port)), Some(Protocol::QuicV1)) => {
|
||
if !self.update_ports.quic4 {
|
||
debug!(self.log, "Skipping ENR update"; "multiaddr" => ?addr);
|
||
return;
|
||
}
|
||
|
||
self.update_enr_quic_port(port, false)
|
||
}
|
||
_ => {
|
||
debug!(self.log, "Encountered unacceptable multiaddr for listening (unsupported transport)"; "addr" => ?addr);
|
||
return;
|
||
}
|
||
},
|
||
Some(Protocol::Ip6(_)) => match (addr_iter.next(), addr_iter.next()) {
|
||
(Some(Protocol::Tcp(port)), None) => {
|
||
if !self.update_ports.tcp6 {
|
||
debug!(self.log, "Skipping ENR update"; "multiaddr" => ?addr);
|
||
return;
|
||
}
|
||
|
||
self.update_enr_tcp_port(port, true)
|
||
}
|
||
(Some(Protocol::Udp(port)), Some(Protocol::QuicV1)) => {
|
||
if !self.update_ports.quic6 {
|
||
debug!(self.log, "Skipping ENR update"; "multiaddr" => ?addr);
|
||
return;
|
||
}
|
||
|
||
self.update_enr_quic_port(port, true)
|
||
}
|
||
_ => {
|
||
debug!(self.log, "Encountered unacceptable multiaddr for listening (unsupported transport)"; "addr" => ?addr);
|
||
return;
|
||
}
|
||
},
|
||
_ => {
|
||
debug!(self.log, "Encountered unacceptable multiaddr for listening (no IP)"; "addr" => ?addr);
|
||
return;
|
||
}
|
||
};
|
||
|
||
let local_enr: Enr = self.discv5.local_enr();
|
||
|
||
match attempt_enr_update {
|
||
Ok(true) => {
|
||
info!(self.log, "Updated local ENR"; "enr" => local_enr.to_base64(), "seq" => local_enr.seq(), "id"=> %local_enr.node_id(), "ip4" => ?local_enr.ip4(), "udp4"=> ?local_enr.udp4(), "tcp4" => ?local_enr.tcp4(), "tcp6" => ?local_enr.tcp6(), "udp6" => ?local_enr.udp6())
|
||
}
|
||
Ok(false) => {} // Nothing to do, ENR already configured
|
||
Err(e) => warn!(self.log, "Failed to update ENR"; "error" => ?e),
|
||
}
|
||
}
|
||
_ => {
|
||
// Ignore events not relevant to discovery
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
impl<E: EthSpec> Discovery<E> {
|
||
fn on_dial_failure(&mut self, peer_id: Option<PeerId>, error: &DialError) {
|
||
if let Some(peer_id) = peer_id {
|
||
match error {
|
||
DialError::Denied { .. } => {
|
||
if self.network_globals.peers.read().is_connected(&peer_id) {
|
||
// There's an active connection, so we don’t disconnect the peer.
|
||
// Lighthouse dials to a peer twice using TCP and QUIC (if QUIC is not
|
||
// disabled). Usually, one establishes a connection, and the other fails
|
||
// because the peer allows only one connection per peer.
|
||
return;
|
||
}
|
||
// set peer as disconnected in discovery DHT
|
||
debug!(self.log, "Marking peer disconnected in DHT"; "peer_id" => %peer_id, "error" => %ClearDialError(error));
|
||
self.disconnect_peer(&peer_id);
|
||
}
|
||
DialError::LocalPeerId { .. }
|
||
| DialError::NoAddresses
|
||
| DialError::Transport(_)
|
||
| DialError::WrongPeerId { .. } => {
|
||
// set peer as disconnected in discovery DHT
|
||
debug!(self.log, "Marking peer disconnected in DHT"; "peer_id" => %peer_id, "error" => %ClearDialError(error));
|
||
self.disconnect_peer(&peer_id);
|
||
}
|
||
DialError::DialPeerConditionFalse(_) | DialError::Aborted => {}
|
||
}
|
||
}
|
||
}
|
||
}
|
||
|
||
#[cfg(test)]
|
||
mod tests {
|
||
use super::*;
|
||
use crate::rpc::methods::{MetaData, MetaDataV2};
|
||
use libp2p::identity::secp256k1;
|
||
use slog::{o, Drain};
|
||
use types::{BitVector, MinimalEthSpec, SubnetId};
|
||
|
||
type E = MinimalEthSpec;
|
||
|
||
pub fn build_log(level: slog::Level, enabled: bool) -> slog::Logger {
|
||
let decorator = slog_term::TermDecorator::new().build();
|
||
let drain = slog_term::FullFormat::new(decorator).build().fuse();
|
||
let drain = slog_async::Async::new(drain).build().fuse();
|
||
|
||
if enabled {
|
||
slog::Logger::root(drain.filter_level(level).fuse(), o!())
|
||
} else {
|
||
slog::Logger::root(drain.filter(|_| false).fuse(), o!())
|
||
}
|
||
}
|
||
|
||
async fn build_discovery() -> Discovery<E> {
|
||
let spec = Arc::new(ChainSpec::default());
|
||
let keypair = secp256k1::Keypair::generate();
|
||
let mut config = NetworkConfig::default();
|
||
config.set_listening_addr(crate::ListenAddress::unused_v4_ports());
|
||
let config = Arc::new(config);
|
||
let enr_key: CombinedKey = CombinedKey::from_secp256k1(&keypair);
|
||
let enr: Enr = build_enr::<E>(&enr_key, &config, &EnrForkId::default(), &spec).unwrap();
|
||
let log = build_log(slog::Level::Debug, false);
|
||
let globals = NetworkGlobals::new(
|
||
enr,
|
||
MetaData::V2(MetaDataV2 {
|
||
seq_number: 0,
|
||
attnets: Default::default(),
|
||
syncnets: Default::default(),
|
||
}),
|
||
vec![],
|
||
false,
|
||
&log,
|
||
config.clone(),
|
||
spec.clone(),
|
||
);
|
||
let keypair = keypair.into();
|
||
Discovery::new(keypair, &config, Arc::new(globals), &log, &spec)
|
||
.await
|
||
.unwrap()
|
||
}
|
||
|
||
#[tokio::test]
|
||
async fn test_add_subnet_query() {
|
||
let mut discovery = build_discovery().await;
|
||
let now = Instant::now();
|
||
let mut subnet_query = SubnetQuery {
|
||
subnet: Subnet::Attestation(SubnetId::new(1)),
|
||
min_ttl: Some(now),
|
||
retries: 0,
|
||
};
|
||
discovery.add_subnet_query(
|
||
subnet_query.subnet,
|
||
subnet_query.min_ttl,
|
||
subnet_query.retries,
|
||
);
|
||
assert_eq!(discovery.queued_queries.back(), Some(&subnet_query));
|
||
|
||
// New query should replace old query
|
||
subnet_query.min_ttl = Some(now + Duration::from_secs(1));
|
||
discovery.add_subnet_query(subnet_query.subnet, subnet_query.min_ttl, 1);
|
||
|
||
subnet_query.retries += 1;
|
||
|
||
assert_eq!(discovery.queued_queries.len(), 1);
|
||
assert_eq!(
|
||
discovery.queued_queries.pop_back(),
|
||
Some(subnet_query.clone())
|
||
);
|
||
|
||
// Retries > MAX_DISCOVERY_RETRY must return immediately without adding
|
||
// anything.
|
||
discovery.add_subnet_query(
|
||
subnet_query.subnet,
|
||
subnet_query.min_ttl,
|
||
MAX_DISCOVERY_RETRY + 1,
|
||
);
|
||
|
||
assert_eq!(discovery.queued_queries.len(), 0);
|
||
}
|
||
|
||
fn make_enr(subnet_ids: Vec<usize>) -> Enr {
|
||
let mut builder = Enr::builder();
|
||
let keypair = secp256k1::Keypair::generate();
|
||
let enr_key: CombinedKey = CombinedKey::from_secp256k1(&keypair);
|
||
|
||
// set the "attnets" field on our ENR
|
||
let mut bitfield = BitVector::<ssz_types::typenum::U64>::new();
|
||
for id in subnet_ids {
|
||
bitfield.set(id, true).unwrap();
|
||
}
|
||
|
||
builder.add_value::<Bytes>(
|
||
ATTESTATION_BITFIELD_ENR_KEY,
|
||
&bitfield.as_ssz_bytes().into(),
|
||
);
|
||
builder.build(&enr_key).unwrap()
|
||
}
|
||
|
||
#[tokio::test]
|
||
async fn test_completed_subnet_queries() {
|
||
let mut discovery = build_discovery().await;
|
||
let now = Instant::now();
|
||
let instant1 = Some(now + Duration::from_secs(10));
|
||
let instant2 = Some(now + Duration::from_secs(5));
|
||
|
||
let query = QueryType::Subnet(vec![
|
||
SubnetQuery {
|
||
subnet: Subnet::Attestation(SubnetId::new(1)),
|
||
min_ttl: instant1,
|
||
retries: 0,
|
||
},
|
||
SubnetQuery {
|
||
subnet: Subnet::Attestation(SubnetId::new(2)),
|
||
min_ttl: instant2,
|
||
retries: 0,
|
||
},
|
||
]);
|
||
|
||
// Create enr which is subscribed to subnets 1 and 2
|
||
let enr1 = make_enr(vec![1, 2]);
|
||
let enr2 = make_enr(vec![2]);
|
||
// Unwanted enr for the given grouped query
|
||
let enr3 = make_enr(vec![3]);
|
||
|
||
let enrs: Vec<Enr> = vec![enr1.clone(), enr2, enr3];
|
||
let results = discovery
|
||
.process_completed_queries(QueryResult {
|
||
query_type: query,
|
||
result: Ok(enrs),
|
||
})
|
||
.unwrap();
|
||
|
||
// enr1 and enr2 are required peers based on the requested subnet ids
|
||
assert_eq!(results.len(), 2);
|
||
|
||
// when a peer belongs to multiple subnet ids, we use the highest ttl.
|
||
assert_eq!(results.get(&enr1).unwrap(), &instant1);
|
||
}
|
||
}
|