mirror of
https://github.com/sigp/lighthouse.git
synced 2026-04-18 05:18:30 +00:00
Reduce number of blobs used in tests to speed up CI (#8194)
`beacon-chain-tests` is now regularly taking 1h+ on CI since Fulu fork was added.
This PR attemtpts to reduce the test time by bringing down the number of blobs generated in tests - instead of generating 0..max_blobs, the generator now generates 0..1 blobs by default, and this can be modified by setting `harness.execution_block_generator.set_min_blob_count(n)`.
Note: The blobs are pre-generated and doesn't require too much CPU to generate however processing a larger number of them on the beacon chain does take a lot of time.
This PR also include a few other small improvements
- Our slowest test (`chain_segment_varying_chunk_size`) runs 3x faster in Fulu just by reusing chain segments
- Avoid re-running fork specific tests on all forks
- Fix a bunch of tests that depends on the harness's existing random blob generation, which is fragile
beacon chain test time on test machine is **~2x** faster:
### `unstable`
```
Summary [ 751.586s] 291 tests run: 291 passed (13 slow), 0 skipped
```
### this branch
```
Summary [ 373.792s] 291 tests run: 291 passed (2 slow), 0 skipped
```
The next set of tests to optimise is the ones that use [`get_chain_segment`](77a9af96de/beacon_node/beacon_chain/tests/block_verification.rs (L45)), as it by default build 320 blocks with supernode - an easy optimisation would be to build these blocks with cgc = 8 for tests that only require fullnodes.
Co-Authored-By: Jimmy Chen <jchen.tc@gmail.com>
Co-Authored-By: Jimmy Chen <jimmy@sigmaprime.io>
This commit is contained in:
@@ -7,11 +7,11 @@ use beacon_chain::custody_context::CUSTODY_CHANGE_DA_EFFECTIVE_DELAY_SECONDS;
|
||||
use beacon_chain::data_availability_checker::AvailableBlock;
|
||||
use beacon_chain::historical_data_columns::HistoricalDataColumnError;
|
||||
use beacon_chain::schema_change::migrate_schema;
|
||||
use beacon_chain::test_utils::SyncCommitteeStrategy;
|
||||
use beacon_chain::test_utils::{
|
||||
AttestationStrategy, BeaconChainHarness, BlockStrategy, DiskHarnessType, get_kzg,
|
||||
mock_execution_layer_from_parts, test_spec,
|
||||
};
|
||||
use beacon_chain::test_utils::{SyncCommitteeStrategy, fork_name_from_env};
|
||||
use beacon_chain::{
|
||||
BeaconChain, BeaconChainError, BeaconChainTypes, BeaconSnapshot, BlockError, ChainConfig,
|
||||
NotifyExecutionLayer, ServerSentEventHandler, WhenSlotSkipped,
|
||||
@@ -3211,12 +3211,13 @@ async fn test_import_historical_data_columns_batch() {
|
||||
for block in block_root_iter {
|
||||
let (block_root, _) = block.unwrap();
|
||||
let data_columns = harness.chain.store.get_data_columns(&block_root).unwrap();
|
||||
assert!(data_columns.is_some());
|
||||
for data_column in data_columns.unwrap() {
|
||||
for data_column in data_columns.unwrap_or_default() {
|
||||
data_columns_list.push(data_column);
|
||||
}
|
||||
}
|
||||
|
||||
assert!(!data_columns_list.is_empty());
|
||||
|
||||
harness
|
||||
.extend_chain(
|
||||
(E::slots_per_epoch() * 4) as usize,
|
||||
@@ -3255,8 +3256,18 @@ async fn test_import_historical_data_columns_batch() {
|
||||
|
||||
for block in block_root_iter {
|
||||
let (block_root, _) = block.unwrap();
|
||||
let data_columns = harness.chain.store.get_data_columns(&block_root).unwrap();
|
||||
assert!(data_columns.is_some())
|
||||
if !harness
|
||||
.get_block(block_root.into())
|
||||
.unwrap()
|
||||
.message()
|
||||
.body()
|
||||
.blob_kzg_commitments()
|
||||
.unwrap()
|
||||
.is_empty()
|
||||
{
|
||||
let data_columns = harness.chain.store.get_data_columns(&block_root).unwrap();
|
||||
assert!(data_columns.is_some())
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
@@ -3290,9 +3301,8 @@ async fn test_import_historical_data_columns_batch_mismatched_block_root() {
|
||||
for block in block_root_iter {
|
||||
let (block_root, _) = block.unwrap();
|
||||
let data_columns = harness.chain.store.get_data_columns(&block_root).unwrap();
|
||||
assert!(data_columns.is_some());
|
||||
|
||||
for data_column in data_columns.unwrap() {
|
||||
for data_column in data_columns.unwrap_or_default() {
|
||||
let mut data_column = (*data_column).clone();
|
||||
if data_column.index % 2 == 0 {
|
||||
data_column.signed_block_header.message.body_root = Hash256::ZERO;
|
||||
@@ -3301,6 +3311,7 @@ async fn test_import_historical_data_columns_batch_mismatched_block_root() {
|
||||
data_columns_list.push(Arc::new(data_column));
|
||||
}
|
||||
}
|
||||
assert!(!data_columns_list.is_empty());
|
||||
|
||||
harness
|
||||
.extend_chain(
|
||||
@@ -3347,7 +3358,11 @@ async fn test_import_historical_data_columns_batch_mismatched_block_root() {
|
||||
// be imported.
|
||||
#[tokio::test]
|
||||
async fn test_import_historical_data_columns_batch_no_block_found() {
|
||||
let spec = ForkName::Fulu.make_genesis_spec(E::default_spec());
|
||||
if fork_name_from_env().is_some_and(|f| !f.fulu_enabled()) {
|
||||
return;
|
||||
};
|
||||
|
||||
let spec = test_spec::<E>();
|
||||
let db_path = tempdir().unwrap();
|
||||
let store = get_store_generic(&db_path, StoreConfig::default(), spec);
|
||||
let start_slot = Slot::new(1);
|
||||
@@ -3374,15 +3389,16 @@ async fn test_import_historical_data_columns_batch_no_block_found() {
|
||||
for block in block_root_iter {
|
||||
let (block_root, _) = block.unwrap();
|
||||
let data_columns = harness.chain.store.get_data_columns(&block_root).unwrap();
|
||||
assert!(data_columns.is_some());
|
||||
|
||||
for data_column in data_columns.unwrap() {
|
||||
for data_column in data_columns.unwrap_or_default() {
|
||||
let mut data_column = (*data_column).clone();
|
||||
data_column.signed_block_header.message.body_root = Hash256::ZERO;
|
||||
data_columns_list.push(Arc::new(data_column));
|
||||
}
|
||||
}
|
||||
|
||||
assert!(!data_columns_list.is_empty());
|
||||
|
||||
harness
|
||||
.extend_chain(
|
||||
(E::slots_per_epoch() * 4) as usize,
|
||||
@@ -4108,6 +4124,12 @@ async fn deneb_prune_blobs_no_finalization() {
|
||||
/// Check that blob pruning does not fail trying to prune across the fork boundary.
|
||||
#[tokio::test]
|
||||
async fn prune_blobs_across_fork_boundary() {
|
||||
// This test covers earlier forks and only need to be executed once.
|
||||
// Note: this test is quite expensive (building a chain to epoch 15) and we should revisit this
|
||||
if fork_name_from_env() != Some(ForkName::latest_stable()) {
|
||||
return;
|
||||
}
|
||||
|
||||
let mut spec = ForkName::Capella.make_genesis_spec(E::default_spec());
|
||||
|
||||
let deneb_fork_epoch = Epoch::new(4);
|
||||
@@ -4124,6 +4146,7 @@ async fn prune_blobs_across_fork_boundary() {
|
||||
let store = get_store_generic(&db_path, StoreConfig::default(), spec);
|
||||
|
||||
let harness = get_harness(store.clone(), LOW_VALIDATOR_COUNT);
|
||||
harness.execution_block_generator().set_min_blob_count(1);
|
||||
|
||||
let blocks_to_deneb_finalization = E::slots_per_epoch() * 7;
|
||||
let blocks_to_electra_finalization = E::slots_per_epoch() * 4;
|
||||
@@ -4279,7 +4302,7 @@ async fn prune_blobs_across_fork_boundary() {
|
||||
// Fulu fork epochs
|
||||
// Pruning should have been triggered
|
||||
assert!(store.get_blob_info().oldest_blob_slot <= Some(oldest_slot));
|
||||
// Oldest blost slot should never be greater than the first fulu slot
|
||||
// Oldest blob slot should never be greater than the first fulu slot
|
||||
let fulu_first_slot = fulu_fork_epoch.start_slot(E::slots_per_epoch());
|
||||
assert!(store.get_blob_info().oldest_blob_slot <= Some(fulu_first_slot));
|
||||
// Blobs should not exist post-Fulu
|
||||
@@ -4764,7 +4787,7 @@ async fn fulu_prune_data_columns_margin_test(margin: u64) {
|
||||
check_data_column_existence(&harness, oldest_data_column_slot, harness.head_slot(), true);
|
||||
}
|
||||
|
||||
/// Check tat there are data column sidecars (or not) at every slot in the range.
|
||||
/// Check that there are data column sidecars (or not) at every slot in the range.
|
||||
fn check_data_column_existence(
|
||||
harness: &TestHarness,
|
||||
start_slot: Slot,
|
||||
|
||||
Reference in New Issue
Block a user