mirror of
https://github.com/sigp/lighthouse.git
synced 2026-03-10 20:22:02 +00:00
There are certain crates which we re-export within `types` which creates a fragmented DevEx, where there are various ways to import the same crates.
```rust
// consensus/types/src/lib.rs
pub use bls::{
AggregatePublicKey, AggregateSignature, Error as BlsError, Keypair, PUBLIC_KEY_BYTES_LEN,
PublicKey, PublicKeyBytes, SIGNATURE_BYTES_LEN, SecretKey, Signature, SignatureBytes,
get_withdrawal_credentials,
};
pub use context_deserialize::{ContextDeserialize, context_deserialize};
pub use fixed_bytes::FixedBytesExtended;
pub use milhouse::{self, List, Vector};
pub use ssz_types::{BitList, BitVector, FixedVector, VariableList, typenum, typenum::Unsigned};
pub use superstruct::superstruct;
```
This PR removes these re-exports and makes it explicit that these types are imported from a non-`consensus/types` crate.
Co-Authored-By: Mac L <mjladson@pm.me>
301 lines
10 KiB
Rust
301 lines
10 KiB
Rust
use crate::Key;
|
|
use crate::hot_cold_store::{BytesKey, HotColdDBError};
|
|
use crate::{
|
|
ColumnIter, ColumnKeyIter, DBColumn, Error, KeyValueStoreOp, get_key_for_col, metrics,
|
|
};
|
|
use fixed_bytes::FixedBytesExtended;
|
|
use leveldb::{
|
|
compaction::Compaction,
|
|
database::{
|
|
Database,
|
|
batch::{Batch, Writebatch},
|
|
kv::KV,
|
|
},
|
|
iterator::{Iterable, LevelDBIterator},
|
|
options::{Options, ReadOptions},
|
|
};
|
|
use std::collections::HashSet;
|
|
use std::marker::PhantomData;
|
|
use std::path::Path;
|
|
use types::{EthSpec, Hash256};
|
|
|
|
use super::interface::WriteOptions;
|
|
|
|
pub struct LevelDB<E: EthSpec> {
|
|
db: Database<BytesKey>,
|
|
_phantom: PhantomData<E>,
|
|
}
|
|
|
|
impl From<WriteOptions> for leveldb::options::WriteOptions {
|
|
fn from(options: WriteOptions) -> Self {
|
|
let mut opts = leveldb::options::WriteOptions::new();
|
|
opts.sync = options.sync;
|
|
opts
|
|
}
|
|
}
|
|
|
|
impl<E: EthSpec> LevelDB<E> {
|
|
pub fn open(path: &Path) -> Result<Self, Error> {
|
|
let mut options = Options::new();
|
|
|
|
options.create_if_missing = true;
|
|
|
|
let db = Database::open(path, options)?;
|
|
|
|
Ok(Self {
|
|
db,
|
|
_phantom: PhantomData,
|
|
})
|
|
}
|
|
|
|
pub fn read_options(&self) -> ReadOptions<'_, BytesKey> {
|
|
ReadOptions::new()
|
|
}
|
|
|
|
pub fn write_options(&self) -> WriteOptions {
|
|
WriteOptions::new()
|
|
}
|
|
|
|
pub fn write_options_sync(&self) -> WriteOptions {
|
|
let mut opts = WriteOptions::new();
|
|
opts.sync = true;
|
|
opts
|
|
}
|
|
|
|
pub fn put_bytes_with_options(
|
|
&self,
|
|
col: DBColumn,
|
|
key: &[u8],
|
|
val: &[u8],
|
|
opts: WriteOptions,
|
|
) -> Result<(), Error> {
|
|
let column_key = get_key_for_col(col, key);
|
|
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_WRITE_COUNT, &[col.into()]);
|
|
metrics::inc_counter_vec_by(
|
|
&metrics::DISK_DB_WRITE_BYTES,
|
|
&[col.into()],
|
|
val.len() as u64,
|
|
);
|
|
let timer = metrics::start_timer(&metrics::DISK_DB_WRITE_TIMES);
|
|
|
|
self.db
|
|
.put(opts.into(), BytesKey::from_vec(column_key), val)
|
|
.map_err(Into::into)
|
|
.map(|()| {
|
|
metrics::stop_timer(timer);
|
|
})
|
|
}
|
|
|
|
/// Store some `value` in `column`, indexed with `key`.
|
|
pub fn put_bytes(&self, col: DBColumn, key: &[u8], val: &[u8]) -> Result<(), Error> {
|
|
self.put_bytes_with_options(col, key, val, self.write_options())
|
|
}
|
|
|
|
pub fn put_bytes_sync(&self, col: DBColumn, key: &[u8], val: &[u8]) -> Result<(), Error> {
|
|
self.put_bytes_with_options(col, key, val, self.write_options_sync())
|
|
}
|
|
|
|
pub fn sync(&self) -> Result<(), Error> {
|
|
self.put_bytes_sync(DBColumn::Dummy, b"sync", b"sync")
|
|
}
|
|
|
|
// Retrieve some bytes in `column` with `key`.
|
|
pub fn get_bytes(&self, col: DBColumn, key: &[u8]) -> Result<Option<Vec<u8>>, Error> {
|
|
let column_key = get_key_for_col(col, key);
|
|
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_READ_COUNT, &[col.into()]);
|
|
let timer = metrics::start_timer(&metrics::DISK_DB_READ_TIMES);
|
|
|
|
self.db
|
|
.get(self.read_options(), BytesKey::from_vec(column_key))
|
|
.map_err(Into::into)
|
|
.map(|opt| {
|
|
opt.inspect(|bytes| {
|
|
metrics::inc_counter_vec_by(
|
|
&metrics::DISK_DB_READ_BYTES,
|
|
&[col.into()],
|
|
bytes.len() as u64,
|
|
);
|
|
metrics::stop_timer(timer);
|
|
})
|
|
})
|
|
}
|
|
|
|
/// Return `true` if `key` exists in `column`.
|
|
pub fn key_exists(&self, col: DBColumn, key: &[u8]) -> Result<bool, Error> {
|
|
let column_key = get_key_for_col(col, key);
|
|
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_EXISTS_COUNT, &[col.into()]);
|
|
|
|
self.db
|
|
.get(self.read_options(), BytesKey::from_vec(column_key))
|
|
.map_err(Into::into)
|
|
.map(|val| val.is_some())
|
|
}
|
|
|
|
/// Removes `key` from `column`.
|
|
pub fn key_delete(&self, col: DBColumn, key: &[u8]) -> Result<(), Error> {
|
|
let column_key = get_key_for_col(col, key);
|
|
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_DELETE_COUNT, &[col.into()]);
|
|
|
|
self.db
|
|
.delete(self.write_options().into(), BytesKey::from_vec(column_key))
|
|
.map_err(Into::into)
|
|
}
|
|
|
|
pub fn do_atomically(&self, ops_batch: Vec<KeyValueStoreOp>) -> Result<(), Error> {
|
|
let mut leveldb_batch = Writebatch::new();
|
|
for op in ops_batch {
|
|
match op {
|
|
KeyValueStoreOp::PutKeyValue(col, key, value) => {
|
|
let _timer = metrics::start_timer(&metrics::DISK_DB_WRITE_TIMES);
|
|
metrics::inc_counter_vec_by(
|
|
&metrics::DISK_DB_WRITE_BYTES,
|
|
&[col.into()],
|
|
value.len() as u64,
|
|
);
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_WRITE_COUNT, &[col.into()]);
|
|
let column_key = get_key_for_col(col, &key);
|
|
leveldb_batch.put(BytesKey::from_vec(column_key), &value);
|
|
}
|
|
|
|
KeyValueStoreOp::DeleteKey(col, key) => {
|
|
let _timer = metrics::start_timer(&metrics::DISK_DB_DELETE_TIMES);
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_DELETE_COUNT, &[col.into()]);
|
|
let column_key = get_key_for_col(col, &key);
|
|
leveldb_batch.delete(BytesKey::from_vec(column_key));
|
|
}
|
|
}
|
|
}
|
|
self.db.write(self.write_options().into(), &leveldb_batch)?;
|
|
Ok(())
|
|
}
|
|
|
|
/// Compact all values in the states and states flag columns.
|
|
pub fn compact(&self) -> Result<(), Error> {
|
|
let _timer = metrics::start_timer(&metrics::DISK_DB_COMPACT_TIMES);
|
|
let endpoints = |column: DBColumn| {
|
|
(
|
|
BytesKey::from_vec(get_key_for_col(column, Hash256::zero().as_slice())),
|
|
BytesKey::from_vec(get_key_for_col(
|
|
column,
|
|
Hash256::repeat_byte(0xff).as_slice(),
|
|
)),
|
|
)
|
|
};
|
|
|
|
for (start_key, end_key) in [
|
|
endpoints(DBColumn::BeaconState),
|
|
endpoints(DBColumn::BeaconStateSummary),
|
|
] {
|
|
self.db.compact(&start_key, &end_key);
|
|
}
|
|
|
|
Ok(())
|
|
}
|
|
|
|
pub fn compact_column(&self, column: DBColumn) -> Result<(), Error> {
|
|
// Use key-size-agnostic keys [] and 0xff..ff with a minimum of 32 bytes to account for
|
|
// columns that may change size between sub-databases or schema versions.
|
|
let start_key = BytesKey::from_vec(get_key_for_col(column, &[]));
|
|
let end_key = BytesKey::from_vec(get_key_for_col(
|
|
column,
|
|
&vec![0xff; std::cmp::max(column.key_size(), 32)],
|
|
));
|
|
self.db.compact(&start_key, &end_key);
|
|
Ok(())
|
|
}
|
|
|
|
pub fn iter_column_from<K: Key>(&self, column: DBColumn, from: &[u8]) -> ColumnIter<'_, K> {
|
|
let start_key = BytesKey::from_vec(get_key_for_col(column, from));
|
|
let iter = self.db.iter(self.read_options());
|
|
iter.seek(&start_key);
|
|
|
|
Box::new(
|
|
iter.take_while(move |(key, _)| key.matches_column(column))
|
|
.map(move |(bytes_key, value)| {
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_READ_COUNT, &[column.into()]);
|
|
metrics::inc_counter_vec_by(
|
|
&metrics::DISK_DB_READ_BYTES,
|
|
&[column.into()],
|
|
value.len() as u64,
|
|
);
|
|
let key = bytes_key.remove_column_variable(column).ok_or_else(|| {
|
|
HotColdDBError::IterationError {
|
|
unexpected_key: bytes_key.clone(),
|
|
}
|
|
})?;
|
|
Ok((K::from_bytes(key)?, value))
|
|
}),
|
|
)
|
|
}
|
|
|
|
pub fn iter_column_keys_from<K: Key>(
|
|
&self,
|
|
column: DBColumn,
|
|
from: &[u8],
|
|
) -> ColumnKeyIter<'_, K> {
|
|
let start_key = BytesKey::from_vec(get_key_for_col(column, from));
|
|
|
|
let iter = self.db.keys_iter(self.read_options());
|
|
iter.seek(&start_key);
|
|
|
|
Box::new(
|
|
iter.take_while(move |key| key.matches_column(column))
|
|
.map(move |bytes_key| {
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_KEY_READ_COUNT, &[column.into()]);
|
|
metrics::inc_counter_vec_by(
|
|
&metrics::DISK_DB_KEY_READ_BYTES,
|
|
&[column.into()],
|
|
bytes_key.key.len() as u64,
|
|
);
|
|
let key = &bytes_key.key[column.as_bytes().len()..];
|
|
K::from_bytes(key)
|
|
}),
|
|
)
|
|
}
|
|
|
|
/// Iterate through all keys and values in a particular column.
|
|
pub fn iter_column_keys<K: Key>(&self, column: DBColumn) -> ColumnKeyIter<'_, K> {
|
|
self.iter_column_keys_from(column, &vec![0; column.key_size()])
|
|
}
|
|
|
|
pub fn iter_column<K: Key>(&self, column: DBColumn) -> ColumnIter<'_, K> {
|
|
self.iter_column_from(column, &vec![0; column.key_size()])
|
|
}
|
|
|
|
pub fn delete_batch(&self, col: DBColumn, ops: HashSet<&[u8]>) -> Result<(), Error> {
|
|
let mut leveldb_batch = Writebatch::new();
|
|
for op in ops {
|
|
let column_key = get_key_for_col(col, op);
|
|
leveldb_batch.delete(BytesKey::from_vec(column_key));
|
|
}
|
|
self.db.write(self.write_options().into(), &leveldb_batch)?;
|
|
Ok(())
|
|
}
|
|
|
|
pub fn delete_if(
|
|
&self,
|
|
column: DBColumn,
|
|
mut f: impl FnMut(&[u8]) -> Result<bool, Error>,
|
|
) -> Result<(), Error> {
|
|
let mut leveldb_batch = Writebatch::new();
|
|
let iter = self.db.iter(self.read_options());
|
|
let start_key = BytesKey::from_vec(column.as_bytes().to_vec());
|
|
iter.seek(&start_key);
|
|
iter.take_while(move |(key, _)| key.matches_column(column))
|
|
.for_each(|(key, value)| {
|
|
if f(&value).unwrap_or(false) {
|
|
let _timer = metrics::start_timer(&metrics::DISK_DB_DELETE_TIMES);
|
|
metrics::inc_counter_vec(&metrics::DISK_DB_DELETE_COUNT, &[column.into()]);
|
|
leveldb_batch.delete(key);
|
|
}
|
|
});
|
|
|
|
self.db.write(self.write_options().into(), &leveldb_batch)?;
|
|
Ok(())
|
|
}
|
|
}
|