parent
63d66ece57
commit
f73d38739a
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,834 @@
|
|||
//! Persistent accounts are stored in below path location:
|
||||
//! <path>/<pid>/data/
|
||||
//!
|
||||
//! The persistent store would allow for this mode of operation:
|
||||
//! - Concurrent single thread append with many concurrent readers.
|
||||
//!
|
||||
//! The underlying memory is memory mapped to a file. The accounts would be
|
||||
//! stored across multiple files and the mappings of file and offset of a
|
||||
//! particular account would be stored in a shared index. This will allow for
|
||||
//! concurrent commits without blocking reads, which will sequentially write
|
||||
//! to memory, ssd or disk, and should be as fast as the hardware allow for.
|
||||
//! The only required in memory data structure with a write lock is the index,
|
||||
//! which should be fast to update.
|
||||
//!
|
||||
//! AppendVec's only store accounts for single forks. To bootstrap the
|
||||
//! index from a persistent store of AppendVec's, the entries include
|
||||
//! a "write_version". A single global atomic `AccountsDB::write_version`
|
||||
//! tracks the number of commits to the entire data store. So the latest
|
||||
//! commit for each fork entry would be indexed.
|
||||
|
||||
use crate::accounts_index::{AccountsIndex, Fork};
|
||||
use crate::append_vec::{AppendVec, StorageMeta, StoredAccount};
|
||||
use hashbrown::{HashMap, HashSet};
|
||||
use log::*;
|
||||
use rand::{thread_rng, Rng};
|
||||
use rayon::prelude::*;
|
||||
use solana_sdk::account::Account;
|
||||
use solana_sdk::pubkey::Pubkey;
|
||||
use std::fs::{create_dir_all, remove_dir_all};
|
||||
use std::path::Path;
|
||||
use std::sync::atomic::{AtomicUsize, Ordering};
|
||||
use std::sync::{Arc, RwLock};
|
||||
|
||||
const ACCOUNT_DATA_FILE_SIZE: u64 = 64 * 1024 * 1024;
|
||||
const ACCOUNT_DATA_FILE: &str = "data";
|
||||
|
||||
#[derive(Debug, Default)]
|
||||
pub struct ErrorCounters {
|
||||
pub account_not_found: usize,
|
||||
pub account_in_use: usize,
|
||||
pub account_loaded_twice: usize,
|
||||
pub blockhash_not_found: usize,
|
||||
pub blockhash_too_old: usize,
|
||||
pub reserve_blockhash: usize,
|
||||
pub insufficient_funds: usize,
|
||||
pub invalid_account_index: usize,
|
||||
pub duplicate_signature: usize,
|
||||
pub call_chain_too_deep: usize,
|
||||
pub missing_signature_for_fee: usize,
|
||||
}
|
||||
|
||||
#[derive(Default, Clone)]
|
||||
pub struct AccountInfo {
|
||||
/// index identifying the append storage
|
||||
id: AppendVecId,
|
||||
|
||||
/// offset into the storage
|
||||
offset: usize,
|
||||
|
||||
/// lamports in the account used when squashing kept for optimization
|
||||
/// purposes to remove accounts with zero balance.
|
||||
lamports: u64,
|
||||
}
|
||||
/// An offset into the AccountsDB::storage vector
|
||||
type AppendVecId = usize;
|
||||
type AccountStorage = Vec<Arc<AccountStorageEntry>>;
|
||||
pub type AccountStorageSlice = [Arc<AccountStorageEntry>];
|
||||
pub type InstructionAccounts = Vec<Account>;
|
||||
pub type InstructionLoaders = Vec<Vec<(Pubkey, Account)>>;
|
||||
|
||||
#[derive(Debug, PartialEq)]
|
||||
pub enum AccountStorageStatus {
|
||||
StorageAvailable = 0,
|
||||
StorageFull = 1,
|
||||
}
|
||||
|
||||
impl From<usize> for AccountStorageStatus {
|
||||
fn from(status: usize) -> Self {
|
||||
use self::AccountStorageStatus::*;
|
||||
match status {
|
||||
0 => StorageAvailable,
|
||||
1 => StorageFull,
|
||||
_ => unreachable!(),
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/// Persistent storage structure holding the accounts
|
||||
pub struct AccountStorageEntry {
|
||||
fork_id: Fork,
|
||||
|
||||
/// storage holding the accounts
|
||||
accounts: AppendVec,
|
||||
|
||||
/// Keeps track of the number of accounts stored in a specific AppendVec.
|
||||
/// This is periodically checked to reuse the stores that do not have
|
||||
/// any accounts in it.
|
||||
count: AtomicUsize,
|
||||
|
||||
/// status corresponding to the storage
|
||||
status: AtomicUsize,
|
||||
}
|
||||
|
||||
impl AccountStorageEntry {
|
||||
pub fn new(path: &str, fork_id: Fork, id: usize, file_size: u64) -> Self {
|
||||
let p = format!("{}/{}", path, id);
|
||||
let path = Path::new(&p);
|
||||
let _ignored = remove_dir_all(path);
|
||||
create_dir_all(path).expect("Create directory failed");
|
||||
let accounts = AppendVec::new(&path.join(ACCOUNT_DATA_FILE), true, file_size as usize);
|
||||
|
||||
AccountStorageEntry {
|
||||
fork_id,
|
||||
accounts,
|
||||
count: AtomicUsize::new(0),
|
||||
status: AtomicUsize::new(AccountStorageStatus::StorageAvailable as usize),
|
||||
}
|
||||
}
|
||||
|
||||
pub fn set_status(&self, status: AccountStorageStatus) {
|
||||
self.status.store(status as usize, Ordering::Relaxed);
|
||||
}
|
||||
|
||||
pub fn get_status(&self) -> AccountStorageStatus {
|
||||
self.status.load(Ordering::Relaxed).into()
|
||||
}
|
||||
|
||||
fn add_account(&self) {
|
||||
self.count.fetch_add(1, Ordering::Relaxed);
|
||||
}
|
||||
|
||||
fn remove_account(&self) -> bool {
|
||||
if self.count.fetch_sub(1, Ordering::Relaxed) == 1 {
|
||||
self.accounts.reset();
|
||||
self.set_status(AccountStorageStatus::StorageAvailable);
|
||||
true
|
||||
} else {
|
||||
false
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// This structure handles the load/store of the accounts
|
||||
#[derive(Default)]
|
||||
pub struct AccountsDB {
|
||||
/// Keeps tracks of index into AppendVec on a per fork basis
|
||||
pub accounts_index: RwLock<AccountsIndex<AccountInfo>>,
|
||||
|
||||
/// Account storage
|
||||
pub storage: RwLock<AccountStorage>,
|
||||
|
||||
/// distribute the accounts across storage lists
|
||||
next_id: AtomicUsize,
|
||||
|
||||
/// write version
|
||||
write_version: AtomicUsize,
|
||||
|
||||
/// Set of storage paths to pick from
|
||||
paths: Vec<String>,
|
||||
|
||||
/// Starting file size of appendvecs
|
||||
file_size: u64,
|
||||
}
|
||||
|
||||
pub fn get_paths_vec(paths: &str) -> Vec<String> {
|
||||
paths.split(',').map(ToString::to_string).collect()
|
||||
}
|
||||
|
||||
impl AccountsDB {
|
||||
pub fn new_with_file_size(paths: &str, file_size: u64) -> Self {
|
||||
let paths = get_paths_vec(&paths);
|
||||
AccountsDB {
|
||||
accounts_index: RwLock::new(AccountsIndex::default()),
|
||||
storage: RwLock::new(vec![]),
|
||||
next_id: AtomicUsize::new(0),
|
||||
write_version: AtomicUsize::new(0),
|
||||
paths,
|
||||
file_size,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn new(paths: &str) -> Self {
|
||||
Self::new_with_file_size(paths, ACCOUNT_DATA_FILE_SIZE)
|
||||
}
|
||||
|
||||
fn new_storage_entry(&self, fork_id: Fork, path: &str) -> AccountStorageEntry {
|
||||
AccountStorageEntry::new(
|
||||
path,
|
||||
fork_id,
|
||||
self.next_id.fetch_add(1, Ordering::Relaxed),
|
||||
self.file_size,
|
||||
)
|
||||
}
|
||||
|
||||
pub fn has_accounts(&self, fork: Fork) -> bool {
|
||||
for x in self.storage.read().unwrap().iter() {
|
||||
if x.fork_id == fork && x.count.load(Ordering::Relaxed) > 0 {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
false
|
||||
}
|
||||
|
||||
/// Scan a specific fork through all the account storage in parallel with sequential read
|
||||
// PERF: Sequentially read each storage entry in parallel
|
||||
pub fn scan_account_storage<F, B>(&self, fork_id: Fork, scan_func: F) -> Vec<B>
|
||||
where
|
||||
F: Fn(&StoredAccount, &mut B) -> (),
|
||||
F: Send + Sync,
|
||||
B: Send + Default,
|
||||
{
|
||||
let storage_maps: Vec<Arc<AccountStorageEntry>> = self
|
||||
.storage
|
||||
.read()
|
||||
.unwrap()
|
||||
.iter()
|
||||
.filter(|store| store.fork_id == fork_id)
|
||||
.cloned()
|
||||
.collect();
|
||||
storage_maps
|
||||
.into_par_iter()
|
||||
.map(|storage| {
|
||||
let accounts = storage.accounts.accounts(0);
|
||||
let mut retval = B::default();
|
||||
accounts
|
||||
.iter()
|
||||
.for_each(|stored_account| scan_func(stored_account, &mut retval));
|
||||
retval
|
||||
})
|
||||
.collect()
|
||||
}
|
||||
|
||||
pub fn load(
|
||||
storage: &AccountStorageSlice,
|
||||
ancestors: &HashMap<Fork, usize>,
|
||||
accounts_index: &AccountsIndex<AccountInfo>,
|
||||
pubkey: &Pubkey,
|
||||
) -> Option<Account> {
|
||||
let info = accounts_index.get(pubkey, ancestors)?;
|
||||
//TODO: thread this as a ref
|
||||
storage
|
||||
.get(info.id)
|
||||
.and_then(|store| Some(store.accounts.get_account(info.offset)?.0.clone_account()))
|
||||
}
|
||||
|
||||
pub fn load_slow(&self, ancestors: &HashMap<Fork, usize>, pubkey: &Pubkey) -> Option<Account> {
|
||||
let accounts_index = self.accounts_index.read().unwrap();
|
||||
let storage = self.storage.read().unwrap();
|
||||
Self::load(&storage, ancestors, &accounts_index, pubkey)
|
||||
}
|
||||
|
||||
fn get_storage_id(&self, fork_id: Fork, start: usize, current: usize) -> usize {
|
||||
let mut id = current;
|
||||
let len: usize;
|
||||
{
|
||||
let stores = self.storage.read().unwrap();
|
||||
len = stores.len();
|
||||
if len > 0 {
|
||||
if id == std::usize::MAX {
|
||||
id = start % len;
|
||||
if stores[id].get_status() == AccountStorageStatus::StorageAvailable {
|
||||
return id;
|
||||
}
|
||||
} else {
|
||||
stores[id].set_status(AccountStorageStatus::StorageFull);
|
||||
}
|
||||
|
||||
loop {
|
||||
id = (id + 1) % len;
|
||||
if fork_id == stores[id].fork_id
|
||||
&& stores[id].get_status() == AccountStorageStatus::StorageAvailable
|
||||
{
|
||||
break;
|
||||
}
|
||||
if id == start % len {
|
||||
break;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
if len == 0 || id == start % len {
|
||||
let mut stores = self.storage.write().unwrap();
|
||||
// check if new store was already created
|
||||
if stores.len() == len {
|
||||
let path_idx = thread_rng().gen_range(0, self.paths.len());
|
||||
let storage = self.new_storage_entry(fork_id, &self.paths[path_idx]);
|
||||
stores.push(Arc::new(storage));
|
||||
}
|
||||
id = stores.len() - 1;
|
||||
}
|
||||
id
|
||||
}
|
||||
|
||||
fn append_account(&self, fork_id: Fork, pubkey: &Pubkey, account: &Account) -> (usize, usize) {
|
||||
let offset: usize;
|
||||
let start = self.next_id.fetch_add(1, Ordering::Relaxed);
|
||||
let mut id = self.get_storage_id(fork_id, start, std::usize::MAX);
|
||||
|
||||
// Even if no lamports, need to preserve the account owner so
|
||||
// we can update the vote_accounts correctly if this account is purged
|
||||
// when squashing.
|
||||
let acc = &mut account.clone();
|
||||
if account.lamports == 0 {
|
||||
acc.data.resize(0, 0);
|
||||
}
|
||||
|
||||
loop {
|
||||
let result: Option<usize>;
|
||||
{
|
||||
let accounts = &self.storage.read().unwrap()[id];
|
||||
let write_version = self.write_version.fetch_add(1, Ordering::Relaxed) as u64;
|
||||
let meta = StorageMeta {
|
||||
write_version,
|
||||
pubkey: *pubkey,
|
||||
data_len: account.data.len() as u64,
|
||||
};
|
||||
result = accounts.accounts.append_account(meta, account);
|
||||
accounts.add_account();
|
||||
}
|
||||
if let Some(val) = result {
|
||||
offset = val;
|
||||
break;
|
||||
} else {
|
||||
id = self.get_storage_id(fork_id, start, id);
|
||||
}
|
||||
}
|
||||
(id, offset)
|
||||
}
|
||||
|
||||
pub fn purge_fork(&self, fork: Fork) {
|
||||
//add_root should be called first
|
||||
let is_root = self.accounts_index.read().unwrap().is_root(fork);
|
||||
trace!("PURGING {} {}", fork, is_root);
|
||||
if !is_root {
|
||||
self.storage.write().unwrap().retain(|x| {
|
||||
trace!("PURGING {} {}", x.fork_id, fork);
|
||||
x.fork_id != fork
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
/// Store the account update.
|
||||
pub fn store(&self, fork_id: Fork, accounts: &[(&Pubkey, &Account)]) {
|
||||
//TODO; these blocks should be separate functions and unit tested
|
||||
let infos: Vec<_> = accounts
|
||||
.iter()
|
||||
.map(|(pubkey, account)| {
|
||||
let (id, offset) = self.append_account(fork_id, pubkey, account);
|
||||
AccountInfo {
|
||||
id,
|
||||
offset,
|
||||
lamports: account.lamports,
|
||||
}
|
||||
})
|
||||
.collect();
|
||||
|
||||
let reclaims: Vec<(Fork, AccountInfo)> = {
|
||||
let mut index = self.accounts_index.write().unwrap();
|
||||
let mut reclaims = vec![];
|
||||
for (i, info) in infos.into_iter().enumerate() {
|
||||
let key = &accounts[i].0;
|
||||
reclaims.extend(index.insert(fork_id, key, info).into_iter())
|
||||
}
|
||||
reclaims
|
||||
};
|
||||
|
||||
let dead_forks: HashSet<Fork> = {
|
||||
let stores = self.storage.read().unwrap();
|
||||
let mut cleared_forks: HashSet<Fork> = HashSet::new();
|
||||
for (fork_id, account_info) in reclaims {
|
||||
let cleared = stores[account_info.id].remove_account();
|
||||
if cleared {
|
||||
cleared_forks.insert(fork_id);
|
||||
}
|
||||
}
|
||||
let live_forks: HashSet<Fork> = stores.iter().map(|x| x.fork_id).collect();
|
||||
cleared_forks.difference(&live_forks).cloned().collect()
|
||||
};
|
||||
{
|
||||
let mut index = self.accounts_index.write().unwrap();
|
||||
for fork in dead_forks {
|
||||
index.cleanup_dead_fork(fork);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
pub fn add_root(&self, fork: Fork) {
|
||||
self.accounts_index.write().unwrap().add_root(fork)
|
||||
}
|
||||
}
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
// TODO: all the bank tests are bank specific, issue: 2194
|
||||
use super::*;
|
||||
use rand::{thread_rng, Rng};
|
||||
use solana_sdk::account::Account;
|
||||
|
||||
fn cleanup_paths(paths: &str) {
|
||||
let paths = get_paths_vec(&paths);
|
||||
paths.iter().for_each(|p| {
|
||||
let _ignored = remove_dir_all(p);
|
||||
});
|
||||
}
|
||||
|
||||
struct TempPaths {
|
||||
pub paths: String,
|
||||
}
|
||||
|
||||
impl Drop for TempPaths {
|
||||
fn drop(&mut self) {
|
||||
cleanup_paths(&self.paths);
|
||||
}
|
||||
}
|
||||
|
||||
fn get_tmp_accounts_path(paths: &str) -> TempPaths {
|
||||
let vpaths = get_paths_vec(paths);
|
||||
let out_dir = std::env::var("OUT_DIR").unwrap_or_else(|_| "target".to_string());
|
||||
let vpaths: Vec<_> = vpaths
|
||||
.iter()
|
||||
.map(|path| format!("{}/{}", out_dir, path))
|
||||
.collect();
|
||||
TempPaths {
|
||||
paths: vpaths.join(","),
|
||||
}
|
||||
}
|
||||
|
||||
#[macro_export]
|
||||
macro_rules! tmp_accounts_name {
|
||||
() => {
|
||||
&format!("{}-{}", file!(), line!())
|
||||
};
|
||||
}
|
||||
|
||||
#[macro_export]
|
||||
macro_rules! get_tmp_accounts_path {
|
||||
() => {
|
||||
get_tmp_accounts_path(tmp_accounts_name!())
|
||||
};
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accountsdb_add_root() {
|
||||
solana_logger::setup();
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
let key = Pubkey::default();
|
||||
let account0 = Account::new(1, 0, &key);
|
||||
|
||||
db.store(0, &[(&key, &account0)]);
|
||||
db.add_root(0);
|
||||
let ancestors = vec![(1, 1)].into_iter().collect();
|
||||
assert_eq!(db.load_slow(&ancestors, &key), Some(account0));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accountsdb_latest_ancestor() {
|
||||
solana_logger::setup();
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
let key = Pubkey::default();
|
||||
let account0 = Account::new(1, 0, &key);
|
||||
|
||||
db.store(0, &[(&key, &account0)]);
|
||||
|
||||
let account1 = Account::new(0, 0, &key);
|
||||
db.store(1, &[(&key, &account1)]);
|
||||
|
||||
let ancestors = vec![(1, 1)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account1);
|
||||
|
||||
let ancestors = vec![(1, 1), (0, 0)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account1);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accountsdb_latest_ancestor_with_root() {
|
||||
solana_logger::setup();
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
let key = Pubkey::default();
|
||||
let account0 = Account::new(1, 0, &key);
|
||||
|
||||
db.store(0, &[(&key, &account0)]);
|
||||
|
||||
let account1 = Account::new(0, 0, &key);
|
||||
db.store(1, &[(&key, &account1)]);
|
||||
db.add_root(0);
|
||||
|
||||
let ancestors = vec![(1, 1)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account1);
|
||||
|
||||
let ancestors = vec![(1, 1), (0, 0)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account1);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accountsdb_root_one_fork() {
|
||||
solana_logger::setup();
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
let key = Pubkey::default();
|
||||
let account0 = Account::new(1, 0, &key);
|
||||
|
||||
// store value 1 in the "root", i.e. db zero
|
||||
db.store(0, &[(&key, &account0)]);
|
||||
|
||||
// now we have:
|
||||
//
|
||||
// root0 -> key.lamports==1
|
||||
// / \
|
||||
// / \
|
||||
// key.lamports==0 <- fork1 \
|
||||
// fork2 -> key.lamports==1
|
||||
// (via root0)
|
||||
|
||||
// store value 0 in one child
|
||||
let account1 = Account::new(0, 0, &key);
|
||||
db.store(1, &[(&key, &account1)]);
|
||||
|
||||
// masking accounts is done at the Accounts level, at accountsDB we see
|
||||
// original account (but could also accept "None", which is implemented
|
||||
// at the Accounts level)
|
||||
let ancestors = vec![(0, 0), (1, 1)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account1);
|
||||
|
||||
// we should see 1 token in fork 2
|
||||
let ancestors = vec![(0, 0), (2, 2)].into_iter().collect();
|
||||
assert_eq!(&db.load_slow(&ancestors, &key).unwrap(), &account0);
|
||||
|
||||
db.add_root(0);
|
||||
|
||||
let ancestors = vec![(1, 1)].into_iter().collect();
|
||||
assert_eq!(db.load_slow(&ancestors, &key), Some(account1));
|
||||
let ancestors = vec![(2, 2)].into_iter().collect();
|
||||
assert_eq!(db.load_slow(&ancestors, &key), Some(account0)); // original value
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accountsdb_add_root_many() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&db, &mut pubkeys, 0, 100, 0, 0);
|
||||
for _ in 1..100 {
|
||||
let idx = thread_rng().gen_range(0, 99);
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
let account = db.load_slow(&ancestors, &pubkeys[idx]).unwrap();
|
||||
let mut default_account = Account::default();
|
||||
default_account.lamports = (idx + 1) as u64;
|
||||
assert_eq!(default_account, account);
|
||||
}
|
||||
|
||||
db.add_root(0);
|
||||
|
||||
// check that all the accounts appear with a new root
|
||||
for _ in 1..100 {
|
||||
let idx = thread_rng().gen_range(0, 99);
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
let account0 = db.load_slow(&ancestors, &pubkeys[idx]).unwrap();
|
||||
let ancestors = vec![(1, 1)].into_iter().collect();
|
||||
let account1 = db.load_slow(&ancestors, &pubkeys[idx]).unwrap();
|
||||
let mut default_account = Account::default();
|
||||
default_account.lamports = (idx + 1) as u64;
|
||||
assert_eq!(&default_account, &account0);
|
||||
assert_eq!(&default_account, &account1);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore]
|
||||
fn test_accountsdb_count_stores() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db = AccountsDB::new(&paths.paths);
|
||||
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(
|
||||
&db,
|
||||
&mut pubkeys,
|
||||
0,
|
||||
2,
|
||||
ACCOUNT_DATA_FILE_SIZE as usize / 3,
|
||||
0,
|
||||
);
|
||||
assert!(check_storage(&db, 2));
|
||||
|
||||
let pubkey = Pubkey::new_rand();
|
||||
let account = Account::new(1, ACCOUNT_DATA_FILE_SIZE as usize / 3, &pubkey);
|
||||
db.store(1, &[(&pubkey, &account)]);
|
||||
db.store(1, &[(&pubkeys[0], &account)]);
|
||||
{
|
||||
let stores = db.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 2);
|
||||
assert_eq!(stores[0].count.load(Ordering::Relaxed), 2);
|
||||
assert_eq!(stores[1].count.load(Ordering::Relaxed), 2);
|
||||
}
|
||||
db.add_root(1);
|
||||
{
|
||||
let stores = db.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 2);
|
||||
assert_eq!(stores[0].count.load(Ordering::Relaxed), 2);
|
||||
assert_eq!(stores[1].count.load(Ordering::Relaxed), 2);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_accounts_unsquashed() {
|
||||
let key = Pubkey::default();
|
||||
|
||||
// 1 token in the "root", i.e. db zero
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let db0 = AccountsDB::new(&paths.paths);
|
||||
let account0 = Account::new(1, 0, &key);
|
||||
db0.store(0, &[(&key, &account0)]);
|
||||
|
||||
// 0 lamports in the child
|
||||
let account1 = Account::new(0, 0, &key);
|
||||
db0.store(1, &[(&key, &account1)]);
|
||||
|
||||
// masking accounts is done at the Accounts level, at accountsDB we see
|
||||
// original account
|
||||
let ancestors = vec![(0, 0), (1, 1)].into_iter().collect();
|
||||
assert_eq!(db0.load_slow(&ancestors, &key), Some(account1));
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
assert_eq!(db0.load_slow(&ancestors, &key), Some(account0));
|
||||
}
|
||||
|
||||
fn create_account(
|
||||
accounts: &AccountsDB,
|
||||
pubkeys: &mut Vec<Pubkey>,
|
||||
fork: Fork,
|
||||
num: usize,
|
||||
space: usize,
|
||||
num_vote: usize,
|
||||
) {
|
||||
for t in 0..num {
|
||||
let pubkey = Pubkey::new_rand();
|
||||
let account = Account::new((t + 1) as u64, space, &Account::default().owner);
|
||||
pubkeys.push(pubkey.clone());
|
||||
let ancestors = vec![(fork, 0)].into_iter().collect();
|
||||
assert!(accounts.load_slow(&ancestors, &pubkey).is_none());
|
||||
accounts.store(fork, &[(&pubkey, &account)]);
|
||||
}
|
||||
for t in 0..num_vote {
|
||||
let pubkey = Pubkey::new_rand();
|
||||
let account = Account::new((num + t + 1) as u64, space, &solana_vote_api::id());
|
||||
pubkeys.push(pubkey.clone());
|
||||
let ancestors = vec![(fork, 0)].into_iter().collect();
|
||||
assert!(accounts.load_slow(&ancestors, &pubkey).is_none());
|
||||
accounts.store(fork, &[(&pubkey, &account)]);
|
||||
}
|
||||
}
|
||||
|
||||
fn update_accounts(accounts: &AccountsDB, pubkeys: &Vec<Pubkey>, fork: Fork, range: usize) {
|
||||
for _ in 1..1000 {
|
||||
let idx = thread_rng().gen_range(0, range);
|
||||
let ancestors = vec![(fork, 0)].into_iter().collect();
|
||||
if let Some(mut account) = accounts.load_slow(&ancestors, &pubkeys[idx]) {
|
||||
account.lamports = account.lamports + 1;
|
||||
accounts.store(fork, &[(&pubkeys[idx], &account)]);
|
||||
if account.lamports == 0 {
|
||||
let ancestors = vec![(fork, 0)].into_iter().collect();
|
||||
assert!(accounts.load_slow(&ancestors, &pubkeys[idx]).is_none());
|
||||
} else {
|
||||
let mut default_account = Account::default();
|
||||
default_account.lamports = account.lamports;
|
||||
assert_eq!(default_account, account);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn check_storage(accounts: &AccountsDB, count: usize) -> bool {
|
||||
let stores = accounts.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 1);
|
||||
assert_eq!(
|
||||
stores[0].get_status(),
|
||||
AccountStorageStatus::StorageAvailable
|
||||
);
|
||||
stores[0].count.load(Ordering::Relaxed) == count
|
||||
}
|
||||
|
||||
fn check_accounts(accounts: &AccountsDB, pubkeys: &Vec<Pubkey>, fork: Fork) {
|
||||
for _ in 1..100 {
|
||||
let idx = thread_rng().gen_range(0, 99);
|
||||
let ancestors = vec![(fork, 0)].into_iter().collect();
|
||||
let account = accounts.load_slow(&ancestors, &pubkeys[idx]).unwrap();
|
||||
let mut default_account = Account::default();
|
||||
default_account.lamports = (idx + 1) as u64;
|
||||
assert_eq!(default_account, account);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_account_one() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&accounts, &mut pubkeys, 0, 1, 0, 0);
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
let account = accounts.load_slow(&ancestors, &pubkeys[0]).unwrap();
|
||||
let mut default_account = Account::default();
|
||||
default_account.lamports = 1;
|
||||
assert_eq!(default_account, account);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_account_many() {
|
||||
let paths = get_tmp_accounts_path("many0,many1");
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&accounts, &mut pubkeys, 0, 100, 0, 0);
|
||||
check_accounts(&accounts, &pubkeys, 0);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_account_update() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&accounts, &mut pubkeys, 0, 100, 0, 0);
|
||||
update_accounts(&accounts, &pubkeys, 0, 99);
|
||||
assert_eq!(check_storage(&accounts, 100), true);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_account_grow_many() {
|
||||
let paths = get_tmp_accounts_path("many2,many3");
|
||||
let size = 4096;
|
||||
let accounts = AccountsDB::new_with_file_size(&paths.paths, size);
|
||||
let mut keys = vec![];
|
||||
for i in 0..9 {
|
||||
let key = Pubkey::new_rand();
|
||||
let account = Account::new(i + 1, size as usize / 4, &key);
|
||||
accounts.store(0, &[(&key, &account)]);
|
||||
keys.push(key);
|
||||
}
|
||||
for (i, key) in keys.iter().enumerate() {
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
assert_eq!(
|
||||
accounts.load_slow(&ancestors, &key).unwrap().lamports,
|
||||
(i as u64) + 1
|
||||
);
|
||||
}
|
||||
|
||||
let mut append_vec_histogram = HashMap::new();
|
||||
for storage in accounts.storage.read().unwrap().iter() {
|
||||
*append_vec_histogram.entry(storage.fork_id).or_insert(0) += 1;
|
||||
}
|
||||
for count in append_vec_histogram.values() {
|
||||
assert!(*count >= 2);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
#[ignore]
|
||||
fn test_account_grow() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let count = [0, 1];
|
||||
let status = [
|
||||
AccountStorageStatus::StorageAvailable,
|
||||
AccountStorageStatus::StorageFull,
|
||||
];
|
||||
let pubkey1 = Pubkey::new_rand();
|
||||
let account1 = Account::new(1, ACCOUNT_DATA_FILE_SIZE as usize / 2, &pubkey1);
|
||||
accounts.store(0, &[(&pubkey1, &account1)]);
|
||||
{
|
||||
let stores = accounts.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 1);
|
||||
assert_eq!(stores[0].count.load(Ordering::Relaxed), 1);
|
||||
assert_eq!(stores[0].get_status(), status[0]);
|
||||
}
|
||||
|
||||
let pubkey2 = Pubkey::new_rand();
|
||||
let account2 = Account::new(1, ACCOUNT_DATA_FILE_SIZE as usize / 2, &pubkey2);
|
||||
accounts.store(0, &[(&pubkey2, &account2)]);
|
||||
{
|
||||
let stores = accounts.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 2);
|
||||
assert_eq!(stores[0].count.load(Ordering::Relaxed), 1);
|
||||
assert_eq!(stores[0].get_status(), status[1]);
|
||||
assert_eq!(stores[1].count.load(Ordering::Relaxed), 1);
|
||||
assert_eq!(stores[1].get_status(), status[0]);
|
||||
}
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
assert_eq!(accounts.load_slow(&ancestors, &pubkey1).unwrap(), account1);
|
||||
assert_eq!(accounts.load_slow(&ancestors, &pubkey2).unwrap(), account2);
|
||||
|
||||
for i in 0..25 {
|
||||
let index = i % 2;
|
||||
accounts.store(0, &[(&pubkey1, &account1)]);
|
||||
{
|
||||
let stores = accounts.storage.read().unwrap();
|
||||
assert_eq!(stores.len(), 3);
|
||||
assert_eq!(stores[0].count.load(Ordering::Relaxed), count[index]);
|
||||
assert_eq!(stores[0].get_status(), status[0]);
|
||||
assert_eq!(stores[1].count.load(Ordering::Relaxed), 1);
|
||||
assert_eq!(stores[1].get_status(), status[1]);
|
||||
assert_eq!(stores[2].count.load(Ordering::Relaxed), count[index ^ 1]);
|
||||
assert_eq!(stores[2].get_status(), status[0]);
|
||||
}
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
assert_eq!(accounts.load_slow(&ancestors, &pubkey1).unwrap(), account1);
|
||||
assert_eq!(accounts.load_slow(&ancestors, &pubkey2).unwrap(), account2);
|
||||
}
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_purge_fork_not_root() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&accounts, &mut pubkeys, 0, 1, 0, 0);
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
assert!(accounts.load_slow(&ancestors, &pubkeys[0]).is_some());;
|
||||
accounts.purge_fork(0);
|
||||
assert!(accounts.load_slow(&ancestors, &pubkeys[0]).is_none());;
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn test_purge_fork_after_root() {
|
||||
let paths = get_tmp_accounts_path!();
|
||||
let accounts = AccountsDB::new(&paths.paths);
|
||||
let mut pubkeys: Vec<Pubkey> = vec![];
|
||||
create_account(&accounts, &mut pubkeys, 0, 1, 0, 0);
|
||||
let ancestors = vec![(0, 0)].into_iter().collect();
|
||||
accounts.add_root(0);
|
||||
accounts.purge_fork(0);
|
||||
assert!(accounts.load_slow(&ancestors, &pubkeys[0]).is_some());
|
||||
}
|
||||
|
||||
}
|
|
@ -3,7 +3,8 @@
|
|||
//! on behalf of the caller, and a low-level API for when they have
|
||||
//! already been signed and verified.
|
||||
|
||||
use crate::accounts::{Accounts, ErrorCounters, InstructionAccounts, InstructionLoaders};
|
||||
use crate::accounts::Accounts;
|
||||
use crate::accounts_db::{ErrorCounters, InstructionAccounts, InstructionLoaders};
|
||||
use crate::blockhash_queue::BlockhashQueue;
|
||||
use crate::locked_accounts_results::LockedAccountsResults;
|
||||
use crate::message_processor::{MessageProcessor, ProcessInstruction};
|
||||
|
|
|
@ -1,4 +1,5 @@
|
|||
mod accounts;
|
||||
mod accounts_db;
|
||||
mod accounts_index;
|
||||
pub mod append_vec;
|
||||
pub mod bank;
|
||||
|
|
Loading…
Reference in New Issue