Move long-running local-cluster tests to local-cluster-slow (#24952)
This commit is contained in:
parent
326e53be97
commit
e1866aacad
|
@ -305,8 +305,12 @@ EOF
|
||||||
". ci/rust-version.sh; ci/docker-run.sh \$\$rust_stable_docker_image ci/test-local-cluster-flakey.sh" \
|
". ci/rust-version.sh; ci/docker-run.sh \$\$rust_stable_docker_image ci/test-local-cluster-flakey.sh" \
|
||||||
10
|
10
|
||||||
|
|
||||||
command_step "local-cluster-slow" \
|
command_step "local-cluster-slow-1" \
|
||||||
". ci/rust-version.sh; ci/docker-run.sh \$\$rust_stable_docker_image ci/test-local-cluster-slow.sh" \
|
". ci/rust-version.sh; ci/docker-run.sh \$\$rust_stable_docker_image ci/test-local-cluster-slow-1.sh" \
|
||||||
|
40
|
||||||
|
|
||||||
|
command_step "local-cluster-slow-2" \
|
||||||
|
". ci/rust-version.sh; ci/docker-run.sh \$\$rust_stable_docker_image ci/test-local-cluster-slow-2.sh" \
|
||||||
40
|
40
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1 @@
|
||||||
|
test-stable.sh
|
|
@ -112,9 +112,14 @@ test-local-cluster-flakey)
|
||||||
_ "$cargo" stable test --release --package solana-local-cluster --test local_cluster_flakey ${V:+--verbose} -- --nocapture --test-threads=1
|
_ "$cargo" stable test --release --package solana-local-cluster --test local_cluster_flakey ${V:+--verbose} -- --nocapture --test-threads=1
|
||||||
exit 0
|
exit 0
|
||||||
;;
|
;;
|
||||||
test-local-cluster-slow)
|
test-local-cluster-slow-1)
|
||||||
_ "$cargo" stable build --release --bins ${V:+--verbose}
|
_ "$cargo" stable build --release --bins ${V:+--verbose}
|
||||||
_ "$cargo" stable test --release --package solana-local-cluster --test local_cluster_slow ${V:+--verbose} -- --nocapture --test-threads=1
|
_ "$cargo" stable test --release --package solana-local-cluster --test local_cluster_slow_1 ${V:+--verbose} -- --nocapture --test-threads=1
|
||||||
|
exit 0
|
||||||
|
;;
|
||||||
|
test-local-cluster-slow-2)
|
||||||
|
_ "$cargo" stable build --release --bins ${V:+--verbose}
|
||||||
|
_ "$cargo" stable test --release --package solana-local-cluster --test local_cluster_slow_2 ${V:+--verbose} -- --nocapture --test-threads=1
|
||||||
exit 0
|
exit 0
|
||||||
;;
|
;;
|
||||||
test-wasm)
|
test-wasm)
|
||||||
|
|
|
@ -5,7 +5,7 @@ use {
|
||||||
solana_core::{
|
solana_core::{
|
||||||
broadcast_stage::BroadcastStageType,
|
broadcast_stage::BroadcastStageType,
|
||||||
consensus::{Tower, SWITCH_FORK_THRESHOLD},
|
consensus::{Tower, SWITCH_FORK_THRESHOLD},
|
||||||
tower_storage::FileTowerStorage,
|
tower_storage::{FileTowerStorage, SavedTower, SavedTowerVersions, TowerStorage},
|
||||||
validator::ValidatorConfig,
|
validator::ValidatorConfig,
|
||||||
},
|
},
|
||||||
solana_gossip::gossip_service::discover_cluster,
|
solana_gossip::gossip_service::discover_cluster,
|
||||||
|
@ -21,6 +21,7 @@ use {
|
||||||
local_cluster::{ClusterConfig, LocalCluster},
|
local_cluster::{ClusterConfig, LocalCluster},
|
||||||
validator_configs::*,
|
validator_configs::*,
|
||||||
},
|
},
|
||||||
|
solana_runtime::snapshot_config::SnapshotConfig,
|
||||||
solana_sdk::{
|
solana_sdk::{
|
||||||
account::AccountSharedData,
|
account::AccountSharedData,
|
||||||
clock::{self, Slot, DEFAULT_MS_PER_SLOT, DEFAULT_TICKS_PER_SLOT},
|
clock::{self, Slot, DEFAULT_MS_PER_SLOT, DEFAULT_TICKS_PER_SLOT},
|
||||||
|
@ -32,7 +33,7 @@ use {
|
||||||
std::{
|
std::{
|
||||||
collections::HashSet,
|
collections::HashSet,
|
||||||
fs, iter,
|
fs, iter,
|
||||||
path::Path,
|
path::{Path, PathBuf},
|
||||||
sync::{
|
sync::{
|
||||||
atomic::{AtomicBool, Ordering},
|
atomic::{AtomicBool, Ordering},
|
||||||
Arc,
|
Arc,
|
||||||
|
@ -40,6 +41,7 @@ use {
|
||||||
thread::sleep,
|
thread::sleep,
|
||||||
time::Duration,
|
time::Duration,
|
||||||
},
|
},
|
||||||
|
tempfile::TempDir,
|
||||||
};
|
};
|
||||||
|
|
||||||
pub const RUST_LOG_FILTER: &str =
|
pub const RUST_LOG_FILTER: &str =
|
||||||
|
@ -415,3 +417,103 @@ pub fn test_faulty_node(
|
||||||
|
|
||||||
(cluster, validator_keys)
|
(cluster, validator_keys)
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn farf_dir() -> PathBuf {
|
||||||
|
std::env::var("FARF_DIR")
|
||||||
|
.unwrap_or_else(|_| "farf".to_string())
|
||||||
|
.into()
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn generate_account_paths(num_account_paths: usize) -> (Vec<TempDir>, Vec<PathBuf>) {
|
||||||
|
let account_storage_dirs: Vec<TempDir> = (0..num_account_paths)
|
||||||
|
.map(|_| tempfile::tempdir_in(farf_dir()).unwrap())
|
||||||
|
.collect();
|
||||||
|
let account_storage_paths: Vec<_> = account_storage_dirs
|
||||||
|
.iter()
|
||||||
|
.map(|a| a.path().to_path_buf())
|
||||||
|
.collect();
|
||||||
|
(account_storage_dirs, account_storage_paths)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub struct SnapshotValidatorConfig {
|
||||||
|
pub bank_snapshots_dir: TempDir,
|
||||||
|
pub snapshot_archives_dir: TempDir,
|
||||||
|
pub account_storage_dirs: Vec<TempDir>,
|
||||||
|
pub validator_config: ValidatorConfig,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl SnapshotValidatorConfig {
|
||||||
|
pub fn new(
|
||||||
|
full_snapshot_archive_interval_slots: Slot,
|
||||||
|
incremental_snapshot_archive_interval_slots: Slot,
|
||||||
|
accounts_hash_interval_slots: Slot,
|
||||||
|
num_account_paths: usize,
|
||||||
|
) -> SnapshotValidatorConfig {
|
||||||
|
assert!(accounts_hash_interval_slots > 0);
|
||||||
|
assert!(full_snapshot_archive_interval_slots > 0);
|
||||||
|
assert!(full_snapshot_archive_interval_slots % accounts_hash_interval_slots == 0);
|
||||||
|
if incremental_snapshot_archive_interval_slots != Slot::MAX {
|
||||||
|
assert!(incremental_snapshot_archive_interval_slots > 0);
|
||||||
|
assert!(
|
||||||
|
incremental_snapshot_archive_interval_slots % accounts_hash_interval_slots == 0
|
||||||
|
);
|
||||||
|
assert!(
|
||||||
|
full_snapshot_archive_interval_slots % incremental_snapshot_archive_interval_slots
|
||||||
|
== 0
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Create the snapshot config
|
||||||
|
let _ = fs::create_dir_all(farf_dir());
|
||||||
|
let bank_snapshots_dir = tempfile::tempdir_in(farf_dir()).unwrap();
|
||||||
|
let snapshot_archives_dir = tempfile::tempdir_in(farf_dir()).unwrap();
|
||||||
|
let snapshot_config = SnapshotConfig {
|
||||||
|
full_snapshot_archive_interval_slots,
|
||||||
|
incremental_snapshot_archive_interval_slots,
|
||||||
|
snapshot_archives_dir: snapshot_archives_dir.path().to_path_buf(),
|
||||||
|
bank_snapshots_dir: bank_snapshots_dir.path().to_path_buf(),
|
||||||
|
maximum_full_snapshot_archives_to_retain: usize::MAX,
|
||||||
|
maximum_incremental_snapshot_archives_to_retain: usize::MAX,
|
||||||
|
..SnapshotConfig::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
// Create the account paths
|
||||||
|
let (account_storage_dirs, account_storage_paths) =
|
||||||
|
generate_account_paths(num_account_paths);
|
||||||
|
|
||||||
|
// Create the validator config
|
||||||
|
let validator_config = ValidatorConfig {
|
||||||
|
snapshot_config: Some(snapshot_config),
|
||||||
|
account_paths: account_storage_paths,
|
||||||
|
accounts_hash_interval_slots,
|
||||||
|
..ValidatorConfig::default_for_test()
|
||||||
|
};
|
||||||
|
|
||||||
|
SnapshotValidatorConfig {
|
||||||
|
bank_snapshots_dir,
|
||||||
|
snapshot_archives_dir,
|
||||||
|
account_storage_dirs,
|
||||||
|
validator_config,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn setup_snapshot_validator_config(
|
||||||
|
snapshot_interval_slots: Slot,
|
||||||
|
num_account_paths: usize,
|
||||||
|
) -> SnapshotValidatorConfig {
|
||||||
|
SnapshotValidatorConfig::new(
|
||||||
|
snapshot_interval_slots,
|
||||||
|
Slot::MAX,
|
||||||
|
snapshot_interval_slots,
|
||||||
|
num_account_paths,
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn save_tower(tower_path: &Path, tower: &Tower, node_keypair: &Keypair) {
|
||||||
|
let file_tower_storage = FileTowerStorage::new(tower_path.to_path_buf());
|
||||||
|
let saved_tower = SavedTower::new(tower, node_keypair).unwrap();
|
||||||
|
file_tower_storage
|
||||||
|
.store(&SavedTowerVersions::from(saved_tower))
|
||||||
|
.unwrap();
|
||||||
|
}
|
||||||
|
|
|
@ -1,12 +1,7 @@
|
||||||
#![allow(clippy::integer_arithmetic)]
|
#![allow(clippy::integer_arithmetic)]
|
||||||
use {
|
use {
|
||||||
assert_matches::assert_matches,
|
assert_matches::assert_matches,
|
||||||
common::{
|
common::*,
|
||||||
copy_blocks, create_custom_leader_schedule_with_random_keys, last_vote_in_tower,
|
|
||||||
ms_for_n_slots, open_blockstore, purge_slots, remove_tower, restore_tower,
|
|
||||||
run_cluster_partition, run_kill_partition_switch_threshold, test_faulty_node,
|
|
||||||
wait_for_last_vote_in_tower_to_land_in_ledger, RUST_LOG_FILTER,
|
|
||||||
},
|
|
||||||
crossbeam_channel::{unbounded, Receiver},
|
crossbeam_channel::{unbounded, Receiver},
|
||||||
gag::BufferRedirect,
|
gag::BufferRedirect,
|
||||||
log::*,
|
log::*,
|
||||||
|
@ -23,7 +18,7 @@ use {
|
||||||
consensus::{Tower, SWITCH_FORK_THRESHOLD, VOTE_THRESHOLD_DEPTH},
|
consensus::{Tower, SWITCH_FORK_THRESHOLD, VOTE_THRESHOLD_DEPTH},
|
||||||
optimistic_confirmation_verifier::OptimisticConfirmationVerifier,
|
optimistic_confirmation_verifier::OptimisticConfirmationVerifier,
|
||||||
replay_stage::DUPLICATE_THRESHOLD,
|
replay_stage::DUPLICATE_THRESHOLD,
|
||||||
tower_storage::{FileTowerStorage, SavedTower, SavedTowerVersions, TowerStorage},
|
tower_storage::FileTowerStorage,
|
||||||
validator::ValidatorConfig,
|
validator::ValidatorConfig,
|
||||||
},
|
},
|
||||||
solana_download_utils::download_snapshot_archive,
|
solana_download_utils::download_snapshot_archive,
|
||||||
|
@ -37,7 +32,6 @@ use {
|
||||||
},
|
},
|
||||||
solana_runtime::{
|
solana_runtime::{
|
||||||
snapshot_archive_info::SnapshotArchiveInfoGetter,
|
snapshot_archive_info::SnapshotArchiveInfoGetter,
|
||||||
snapshot_config::SnapshotConfig,
|
|
||||||
snapshot_package::SnapshotType,
|
snapshot_package::SnapshotType,
|
||||||
snapshot_utils::{self, ArchiveFormat},
|
snapshot_utils::{self, ArchiveFormat},
|
||||||
},
|
},
|
||||||
|
@ -60,7 +54,7 @@ use {
|
||||||
fs,
|
fs,
|
||||||
io::Read,
|
io::Read,
|
||||||
iter,
|
iter,
|
||||||
path::{Path, PathBuf},
|
path::Path,
|
||||||
sync::{
|
sync::{
|
||||||
atomic::{AtomicBool, Ordering},
|
atomic::{AtomicBool, Ordering},
|
||||||
Arc,
|
Arc,
|
||||||
|
@ -68,7 +62,6 @@ use {
|
||||||
thread::{sleep, Builder, JoinHandle},
|
thread::{sleep, Builder, JoinHandle},
|
||||||
time::{Duration, Instant},
|
time::{Duration, Instant},
|
||||||
},
|
},
|
||||||
tempfile::TempDir,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
mod common;
|
mod common;
|
||||||
|
@ -102,48 +95,6 @@ fn test_local_cluster_start_and_exit_with_config() {
|
||||||
assert_eq!(cluster.validators.len(), NUM_NODES);
|
assert_eq!(cluster.validators.len(), NUM_NODES);
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
#[serial]
|
|
||||||
fn test_ledger_cleanup_service() {
|
|
||||||
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
|
||||||
error!("test_ledger_cleanup_service");
|
|
||||||
let num_nodes = 3;
|
|
||||||
let validator_config = ValidatorConfig {
|
|
||||||
max_ledger_shreds: Some(100),
|
|
||||||
..ValidatorConfig::default_for_test()
|
|
||||||
};
|
|
||||||
let mut config = ClusterConfig {
|
|
||||||
cluster_lamports: 10_000,
|
|
||||||
poh_config: PohConfig::new_sleep(Duration::from_millis(50)),
|
|
||||||
node_stakes: vec![100; num_nodes],
|
|
||||||
validator_configs: make_identical_validator_configs(&validator_config, num_nodes),
|
|
||||||
..ClusterConfig::default()
|
|
||||||
};
|
|
||||||
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
|
||||||
// 200ms/per * 100 = 20 seconds, so sleep a little longer than that.
|
|
||||||
sleep(Duration::from_secs(60));
|
|
||||||
|
|
||||||
cluster_tests::spend_and_verify_all_nodes(
|
|
||||||
&cluster.entry_point_info,
|
|
||||||
&cluster.funding_keypair,
|
|
||||||
num_nodes,
|
|
||||||
HashSet::new(),
|
|
||||||
SocketAddrSpace::Unspecified,
|
|
||||||
);
|
|
||||||
cluster.close_preserve_ledgers();
|
|
||||||
//check everyone's ledgers and make sure only ~100 slots are stored
|
|
||||||
for info in cluster.validators.values() {
|
|
||||||
let mut slots = 0;
|
|
||||||
let blockstore = Blockstore::open(&info.info.ledger_path).unwrap();
|
|
||||||
blockstore
|
|
||||||
.slot_meta_iterator(0)
|
|
||||||
.unwrap()
|
|
||||||
.for_each(|_| slots += 1);
|
|
||||||
// with 3 nodes up to 3 slots can be in progress and not complete so max slots in blockstore should be up to 103
|
|
||||||
assert!(slots <= 103, "got {}", slots);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[serial]
|
#[serial]
|
||||||
fn test_spend_and_verify_all_nodes_1() {
|
fn test_spend_and_verify_all_nodes_1() {
|
||||||
|
@ -288,75 +239,6 @@ fn test_spend_and_verify_all_nodes_env_num_nodes() {
|
||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Cluster needs a supermajority to remain, so the minimum size for this test is 4
|
|
||||||
#[test]
|
|
||||||
#[serial]
|
|
||||||
fn test_leader_failure_4() {
|
|
||||||
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
|
||||||
error!("test_leader_failure_4");
|
|
||||||
let num_nodes = 4;
|
|
||||||
let validator_config = ValidatorConfig::default_for_test();
|
|
||||||
let mut config = ClusterConfig {
|
|
||||||
cluster_lamports: 10_000,
|
|
||||||
node_stakes: vec![100; 4],
|
|
||||||
validator_configs: make_identical_validator_configs(&validator_config, num_nodes),
|
|
||||||
..ClusterConfig::default()
|
|
||||||
};
|
|
||||||
let local = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
|
||||||
|
|
||||||
cluster_tests::kill_entry_and_spend_and_verify_rest(
|
|
||||||
&local.entry_point_info,
|
|
||||||
&local
|
|
||||||
.validators
|
|
||||||
.get(&local.entry_point_info.id)
|
|
||||||
.unwrap()
|
|
||||||
.config
|
|
||||||
.validator_exit,
|
|
||||||
&local.funding_keypair,
|
|
||||||
num_nodes,
|
|
||||||
config.ticks_per_slot * config.poh_config.target_tick_duration.as_millis() as u64,
|
|
||||||
SocketAddrSpace::Unspecified,
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
#[serial]
|
|
||||||
fn test_cluster_partition_1_1() {
|
|
||||||
let empty = |_: &mut LocalCluster, _: &mut ()| {};
|
|
||||||
let on_partition_resolved = |cluster: &mut LocalCluster, _: &mut ()| {
|
|
||||||
cluster.check_for_new_roots(16, "PARTITION_TEST", SocketAddrSpace::Unspecified);
|
|
||||||
};
|
|
||||||
run_cluster_partition(
|
|
||||||
&[1, 1],
|
|
||||||
None,
|
|
||||||
(),
|
|
||||||
empty,
|
|
||||||
empty,
|
|
||||||
on_partition_resolved,
|
|
||||||
None,
|
|
||||||
vec![],
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
#[serial]
|
|
||||||
fn test_cluster_partition_1_1_1() {
|
|
||||||
let empty = |_: &mut LocalCluster, _: &mut ()| {};
|
|
||||||
let on_partition_resolved = |cluster: &mut LocalCluster, _: &mut ()| {
|
|
||||||
cluster.check_for_new_roots(16, "PARTITION_TEST", SocketAddrSpace::Unspecified);
|
|
||||||
};
|
|
||||||
run_cluster_partition(
|
|
||||||
&[1, 1, 1],
|
|
||||||
None,
|
|
||||||
(),
|
|
||||||
empty,
|
|
||||||
empty,
|
|
||||||
on_partition_resolved,
|
|
||||||
None,
|
|
||||||
vec![],
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[serial]
|
#[serial]
|
||||||
fn test_two_unbalanced_stakes() {
|
fn test_two_unbalanced_stakes() {
|
||||||
|
@ -534,109 +416,6 @@ fn test_mainnet_beta_cluster_type() {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[test]
|
|
||||||
#[serial]
|
|
||||||
fn test_consistency_halt() {
|
|
||||||
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
|
||||||
let snapshot_interval_slots = 20;
|
|
||||||
let num_account_paths = 1;
|
|
||||||
|
|
||||||
// Create cluster with a leader producing bad snapshot hashes.
|
|
||||||
let mut leader_snapshot_test_config =
|
|
||||||
setup_snapshot_validator_config(snapshot_interval_slots, num_account_paths);
|
|
||||||
leader_snapshot_test_config
|
|
||||||
.validator_config
|
|
||||||
.accounts_hash_fault_injection_slots = 40;
|
|
||||||
|
|
||||||
let validator_stake = 10_000;
|
|
||||||
let mut config = ClusterConfig {
|
|
||||||
node_stakes: vec![validator_stake],
|
|
||||||
cluster_lamports: 100_000,
|
|
||||||
validator_configs: vec![leader_snapshot_test_config.validator_config],
|
|
||||||
..ClusterConfig::default()
|
|
||||||
};
|
|
||||||
|
|
||||||
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
|
||||||
|
|
||||||
sleep(Duration::from_millis(5000));
|
|
||||||
let cluster_nodes = discover_cluster(
|
|
||||||
&cluster.entry_point_info.gossip,
|
|
||||||
1,
|
|
||||||
SocketAddrSpace::Unspecified,
|
|
||||||
)
|
|
||||||
.unwrap();
|
|
||||||
info!("num_nodes: {}", cluster_nodes.len());
|
|
||||||
|
|
||||||
// Add a validator with the leader as trusted, it should halt when it detects
|
|
||||||
// mismatch.
|
|
||||||
let mut validator_snapshot_test_config =
|
|
||||||
setup_snapshot_validator_config(snapshot_interval_slots, num_account_paths);
|
|
||||||
|
|
||||||
let mut known_validators = HashSet::new();
|
|
||||||
known_validators.insert(cluster_nodes[0].id);
|
|
||||||
|
|
||||||
validator_snapshot_test_config
|
|
||||||
.validator_config
|
|
||||||
.known_validators = Some(known_validators);
|
|
||||||
validator_snapshot_test_config
|
|
||||||
.validator_config
|
|
||||||
.halt_on_known_validators_accounts_hash_mismatch = true;
|
|
||||||
|
|
||||||
warn!("adding a validator");
|
|
||||||
cluster.add_validator(
|
|
||||||
&validator_snapshot_test_config.validator_config,
|
|
||||||
validator_stake as u64,
|
|
||||||
Arc::new(Keypair::new()),
|
|
||||||
None,
|
|
||||||
SocketAddrSpace::Unspecified,
|
|
||||||
);
|
|
||||||
let num_nodes = 2;
|
|
||||||
assert_eq!(
|
|
||||||
discover_cluster(
|
|
||||||
&cluster.entry_point_info.gossip,
|
|
||||||
num_nodes,
|
|
||||||
SocketAddrSpace::Unspecified
|
|
||||||
)
|
|
||||||
.unwrap()
|
|
||||||
.len(),
|
|
||||||
num_nodes
|
|
||||||
);
|
|
||||||
|
|
||||||
// Check for only 1 node on the network.
|
|
||||||
let mut encountered_error = false;
|
|
||||||
loop {
|
|
||||||
let discover = discover_cluster(
|
|
||||||
&cluster.entry_point_info.gossip,
|
|
||||||
2,
|
|
||||||
SocketAddrSpace::Unspecified,
|
|
||||||
);
|
|
||||||
match discover {
|
|
||||||
Err(_) => {
|
|
||||||
encountered_error = true;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
Ok(nodes) => {
|
|
||||||
if nodes.len() < 2 {
|
|
||||||
encountered_error = true;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
info!("checking cluster for fewer nodes.. {:?}", nodes.len());
|
|
||||||
}
|
|
||||||
}
|
|
||||||
let client = cluster
|
|
||||||
.get_validator_client(&cluster.entry_point_info.id)
|
|
||||||
.unwrap();
|
|
||||||
if let Ok(slot) = client.get_slot() {
|
|
||||||
if slot > 210 {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
info!("slot: {}", slot);
|
|
||||||
}
|
|
||||||
sleep(Duration::from_millis(1000));
|
|
||||||
}
|
|
||||||
assert!(encountered_error);
|
|
||||||
}
|
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
#[serial]
|
#[serial]
|
||||||
fn test_snapshot_download() {
|
fn test_snapshot_download() {
|
||||||
|
@ -1918,217 +1697,10 @@ fn test_validator_saves_tower() {
|
||||||
assert!(tower4.root() >= new_root);
|
assert!(tower4.root() >= new_root);
|
||||||
}
|
}
|
||||||
|
|
||||||
fn save_tower(tower_path: &Path, tower: &Tower, node_keypair: &Keypair) {
|
|
||||||
let file_tower_storage = FileTowerStorage::new(tower_path.to_path_buf());
|
|
||||||
let saved_tower = SavedTower::new(tower, node_keypair).unwrap();
|
|
||||||
file_tower_storage
|
|
||||||
.store(&SavedTowerVersions::from(saved_tower))
|
|
||||||
.unwrap();
|
|
||||||
}
|
|
||||||
|
|
||||||
fn root_in_tower(tower_path: &Path, node_pubkey: &Pubkey) -> Option<Slot> {
|
fn root_in_tower(tower_path: &Path, node_pubkey: &Pubkey) -> Option<Slot> {
|
||||||
restore_tower(tower_path, node_pubkey).map(|tower| tower.root())
|
restore_tower(tower_path, node_pubkey).map(|tower| tower.root())
|
||||||
}
|
}
|
||||||
|
|
||||||
// This test verifies that even if votes from a validator end up taking too long to land, and thus
|
|
||||||
// some of the referenced slots are slots are no longer present in the slot hashes sysvar,
|
|
||||||
// consensus can still be attained.
|
|
||||||
//
|
|
||||||
// Validator A (60%)
|
|
||||||
// Validator B (40%)
|
|
||||||
// / --- 10 --- [..] --- 16 (B is voting, due to network issues is initally not able to see the other fork at all)
|
|
||||||
// /
|
|
||||||
// 1 - 2 - 3 - 4 - 5 - 6 - 7 - 8 - 9 (A votes 1 - 9 votes are landing normally. B does the same however votes are not landing)
|
|
||||||
// \
|
|
||||||
// \--[..]-- 73 (majority fork)
|
|
||||||
// A is voting on the majority fork and B wants to switch to this fork however in this majority fork
|
|
||||||
// the earlier votes for B (1 - 9) never landed so when B eventually goes to vote on 73, slots in
|
|
||||||
// its local vote state are no longer present in slot hashes.
|
|
||||||
//
|
|
||||||
// 1. Wait for B's tower to see local vote state was updated to new fork
|
|
||||||
// 2. Wait X blocks, check B's vote state on chain has been properly updated
|
|
||||||
//
|
|
||||||
// NOTE: it is not reliable for B to organically have 1 to reach 2^16 lockout, so we simulate the 6
|
|
||||||
// consecutive votes on the minor fork by manually incrementing the confirmation levels for the
|
|
||||||
// common ancestor votes in tower.
|
|
||||||
// To allow this test to run in a reasonable time we change the
|
|
||||||
// slot_hash expiry to 64 slots.
|
|
||||||
|
|
||||||
#[test]
|
|
||||||
fn test_slot_hash_expiry() {
|
|
||||||
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
|
||||||
solana_sdk::slot_hashes::set_entries_for_tests_only(64);
|
|
||||||
|
|
||||||
let slots_per_epoch = 2048;
|
|
||||||
let node_stakes = vec![60, 40];
|
|
||||||
let validator_keys = vec![
|
|
||||||
"28bN3xyvrP4E8LwEgtLjhnkb7cY4amQb6DrYAbAYjgRV4GAGgkVM2K7wnxnAS7WDneuavza7x21MiafLu1HkwQt4",
|
|
||||||
"2saHBBoTkLMmttmPQP8KfBkcCw45S5cwtV3wTdGCscRC8uxdgvHxpHiWXKx4LvJjNJtnNcbSv5NdheokFFqnNDt8",
|
|
||||||
]
|
|
||||||
.iter()
|
|
||||||
.map(|s| (Arc::new(Keypair::from_base58_string(s)), true))
|
|
||||||
.collect::<Vec<_>>();
|
|
||||||
let node_vote_keys = vec![
|
|
||||||
"3NDQ3ud86RTVg8hTy2dDWnS4P8NfjhZ2gDgQAJbr3heaKaUVS1FW3sTLKA1GmDrY9aySzsa4QxpDkbLv47yHxzr3",
|
|
||||||
"46ZHpHE6PEvXYPu3hf9iQqjBk2ZNDaJ9ejqKWHEjxaQjpAGasKaWKbKHbP3646oZhfgDRzx95DH9PCBKKsoCVngk",
|
|
||||||
]
|
|
||||||
.iter()
|
|
||||||
.map(|s| Arc::new(Keypair::from_base58_string(s)))
|
|
||||||
.collect::<Vec<_>>();
|
|
||||||
let vs = validator_keys
|
|
||||||
.iter()
|
|
||||||
.map(|(kp, _)| kp.pubkey())
|
|
||||||
.collect::<Vec<_>>();
|
|
||||||
let (a_pubkey, b_pubkey) = (vs[0], vs[1]);
|
|
||||||
|
|
||||||
// We want B to not vote (we are trying to simulate its votes not landing until it gets to the
|
|
||||||
// minority fork)
|
|
||||||
let mut validator_configs =
|
|
||||||
make_identical_validator_configs(&ValidatorConfig::default_for_test(), node_stakes.len());
|
|
||||||
validator_configs[1].voting_disabled = true;
|
|
||||||
|
|
||||||
let mut config = ClusterConfig {
|
|
||||||
cluster_lamports: 100_000,
|
|
||||||
node_stakes,
|
|
||||||
validator_configs,
|
|
||||||
validator_keys: Some(validator_keys),
|
|
||||||
node_vote_keys: Some(node_vote_keys),
|
|
||||||
slots_per_epoch,
|
|
||||||
stakers_slot_offset: slots_per_epoch,
|
|
||||||
skip_warmup_slots: true,
|
|
||||||
..ClusterConfig::default()
|
|
||||||
};
|
|
||||||
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
|
||||||
|
|
||||||
let mut common_ancestor_slot = 8;
|
|
||||||
|
|
||||||
let a_ledger_path = cluster.ledger_path(&a_pubkey);
|
|
||||||
let b_ledger_path = cluster.ledger_path(&b_pubkey);
|
|
||||||
|
|
||||||
// Immediately kill B (we just needed it for the initial stake distribution)
|
|
||||||
info!("Killing B");
|
|
||||||
let mut b_info = cluster.exit_node(&b_pubkey);
|
|
||||||
|
|
||||||
// Let A run for a while until we get to the common ancestor
|
|
||||||
info!("Letting A run until common_ancestor_slot");
|
|
||||||
loop {
|
|
||||||
if let Some((last_vote, _)) = last_vote_in_tower(&a_ledger_path, &a_pubkey) {
|
|
||||||
if last_vote >= common_ancestor_slot {
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sleep(Duration::from_millis(100));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Keep A running, but setup B so that it thinks it has voted up until common ancestor (but
|
|
||||||
// doesn't know anything past that)
|
|
||||||
{
|
|
||||||
info!("Copying A's ledger to B");
|
|
||||||
std::fs::remove_dir_all(&b_info.info.ledger_path).unwrap();
|
|
||||||
let mut opt = fs_extra::dir::CopyOptions::new();
|
|
||||||
opt.copy_inside = true;
|
|
||||||
fs_extra::dir::copy(&a_ledger_path, &b_ledger_path, &opt).unwrap();
|
|
||||||
|
|
||||||
// remove A's tower in B's new copied ledger
|
|
||||||
info!("Removing A's tower in B's ledger dir");
|
|
||||||
remove_tower(&b_ledger_path, &a_pubkey);
|
|
||||||
|
|
||||||
// load A's tower and save it as B's tower
|
|
||||||
info!("Loading A's tower");
|
|
||||||
if let Some(mut a_tower) = restore_tower(&a_ledger_path, &a_pubkey) {
|
|
||||||
a_tower.node_pubkey = b_pubkey;
|
|
||||||
// Update common_ancestor_slot because A is still running
|
|
||||||
if let Some(s) = a_tower.last_voted_slot() {
|
|
||||||
common_ancestor_slot = s;
|
|
||||||
info!("New common_ancestor_slot {}", common_ancestor_slot);
|
|
||||||
} else {
|
|
||||||
panic!("A's tower has no votes");
|
|
||||||
}
|
|
||||||
info!("Increase lockout by 6 confirmation levels and save as B's tower");
|
|
||||||
a_tower.increase_lockout(6);
|
|
||||||
save_tower(&b_ledger_path, &a_tower, &b_info.info.keypair);
|
|
||||||
info!("B's new tower: {:?}", a_tower.tower_slots());
|
|
||||||
} else {
|
|
||||||
panic!("A's tower is missing");
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get rid of any slots past common_ancestor_slot
|
|
||||||
info!("Removing extra slots from B's blockstore");
|
|
||||||
let blockstore = open_blockstore(&b_ledger_path);
|
|
||||||
purge_slots(&blockstore, common_ancestor_slot + 1, 100);
|
|
||||||
}
|
|
||||||
|
|
||||||
info!(
|
|
||||||
"Run A on majority fork until it reaches slot hash expiry {}",
|
|
||||||
solana_sdk::slot_hashes::get_entries()
|
|
||||||
);
|
|
||||||
let mut last_vote_on_a;
|
|
||||||
// Keep A running for a while longer so the majority fork has some decent size
|
|
||||||
loop {
|
|
||||||
last_vote_on_a = wait_for_last_vote_in_tower_to_land_in_ledger(&a_ledger_path, &a_pubkey);
|
|
||||||
if last_vote_on_a
|
|
||||||
>= common_ancestor_slot + 2 * (solana_sdk::slot_hashes::get_entries() as u64)
|
|
||||||
{
|
|
||||||
let blockstore = open_blockstore(&a_ledger_path);
|
|
||||||
info!(
|
|
||||||
"A majority fork: {:?}",
|
|
||||||
AncestorIterator::new(last_vote_on_a, &blockstore).collect::<Vec<Slot>>()
|
|
||||||
);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
sleep(Duration::from_millis(100));
|
|
||||||
}
|
|
||||||
|
|
||||||
// Kill A and restart B with voting. B should now fork off
|
|
||||||
info!("Killing A");
|
|
||||||
let a_info = cluster.exit_node(&a_pubkey);
|
|
||||||
|
|
||||||
info!("Restarting B");
|
|
||||||
b_info.config.voting_disabled = false;
|
|
||||||
cluster.restart_node(&b_pubkey, b_info, SocketAddrSpace::Unspecified);
|
|
||||||
|
|
||||||
// B will fork off and accumulate enough lockout
|
|
||||||
info!("Allowing B to fork");
|
|
||||||
loop {
|
|
||||||
let blockstore = open_blockstore(&b_ledger_path);
|
|
||||||
let last_vote = wait_for_last_vote_in_tower_to_land_in_ledger(&b_ledger_path, &b_pubkey);
|
|
||||||
let mut ancestors = AncestorIterator::new(last_vote, &blockstore);
|
|
||||||
if let Some(index) = ancestors.position(|x| x == common_ancestor_slot) {
|
|
||||||
if index > 7 {
|
|
||||||
info!(
|
|
||||||
"B has forked for enough lockout: {:?}",
|
|
||||||
AncestorIterator::new(last_vote, &blockstore).collect::<Vec<Slot>>()
|
|
||||||
);
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
sleep(Duration::from_millis(1000));
|
|
||||||
}
|
|
||||||
|
|
||||||
info!("Kill B");
|
|
||||||
b_info = cluster.exit_node(&b_pubkey);
|
|
||||||
|
|
||||||
info!("Resolve the partition");
|
|
||||||
{
|
|
||||||
// Here we let B know about the missing blocks that A had produced on its partition
|
|
||||||
let a_blockstore = open_blockstore(&a_ledger_path);
|
|
||||||
let b_blockstore = open_blockstore(&b_ledger_path);
|
|
||||||
copy_blocks(last_vote_on_a, &a_blockstore, &b_blockstore);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Now restart A and B and see if B is able to eventually switch onto the majority fork
|
|
||||||
info!("Restarting A & B");
|
|
||||||
cluster.restart_node(&a_pubkey, a_info, SocketAddrSpace::Unspecified);
|
|
||||||
cluster.restart_node(&b_pubkey, b_info, SocketAddrSpace::Unspecified);
|
|
||||||
|
|
||||||
info!("Waiting for B to switch to majority fork and make a root");
|
|
||||||
cluster_tests::check_for_new_roots(
|
|
||||||
16,
|
|
||||||
&[cluster.get_contact_info(&a_pubkey).unwrap().clone()],
|
|
||||||
"test_slot_hashes_expiry",
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
enum ClusterMode {
|
enum ClusterMode {
|
||||||
MasterOnly,
|
MasterOnly,
|
||||||
MasterSlave,
|
MasterSlave,
|
||||||
|
@ -2828,95 +2400,3 @@ fn run_test_load_program_accounts(scan_commitment: CommitmentConfig) {
|
||||||
t_update.join().unwrap();
|
t_update.join().unwrap();
|
||||||
t_scan.join().unwrap();
|
t_scan.join().unwrap();
|
||||||
}
|
}
|
||||||
|
|
||||||
fn farf_dir() -> PathBuf {
|
|
||||||
std::env::var("FARF_DIR")
|
|
||||||
.unwrap_or_else(|_| "farf".to_string())
|
|
||||||
.into()
|
|
||||||
}
|
|
||||||
|
|
||||||
fn generate_account_paths(num_account_paths: usize) -> (Vec<TempDir>, Vec<PathBuf>) {
|
|
||||||
let account_storage_dirs: Vec<TempDir> = (0..num_account_paths)
|
|
||||||
.map(|_| tempfile::tempdir_in(farf_dir()).unwrap())
|
|
||||||
.collect();
|
|
||||||
let account_storage_paths: Vec<_> = account_storage_dirs
|
|
||||||
.iter()
|
|
||||||
.map(|a| a.path().to_path_buf())
|
|
||||||
.collect();
|
|
||||||
(account_storage_dirs, account_storage_paths)
|
|
||||||
}
|
|
||||||
|
|
||||||
struct SnapshotValidatorConfig {
|
|
||||||
bank_snapshots_dir: TempDir,
|
|
||||||
snapshot_archives_dir: TempDir,
|
|
||||||
account_storage_dirs: Vec<TempDir>,
|
|
||||||
validator_config: ValidatorConfig,
|
|
||||||
}
|
|
||||||
|
|
||||||
impl SnapshotValidatorConfig {
|
|
||||||
pub fn new(
|
|
||||||
full_snapshot_archive_interval_slots: Slot,
|
|
||||||
incremental_snapshot_archive_interval_slots: Slot,
|
|
||||||
accounts_hash_interval_slots: Slot,
|
|
||||||
num_account_paths: usize,
|
|
||||||
) -> SnapshotValidatorConfig {
|
|
||||||
assert!(accounts_hash_interval_slots > 0);
|
|
||||||
assert!(full_snapshot_archive_interval_slots > 0);
|
|
||||||
assert!(full_snapshot_archive_interval_slots % accounts_hash_interval_slots == 0);
|
|
||||||
if incremental_snapshot_archive_interval_slots != Slot::MAX {
|
|
||||||
assert!(incremental_snapshot_archive_interval_slots > 0);
|
|
||||||
assert!(
|
|
||||||
incremental_snapshot_archive_interval_slots % accounts_hash_interval_slots == 0
|
|
||||||
);
|
|
||||||
assert!(
|
|
||||||
full_snapshot_archive_interval_slots % incremental_snapshot_archive_interval_slots
|
|
||||||
== 0
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
// Create the snapshot config
|
|
||||||
let _ = fs::create_dir_all(farf_dir());
|
|
||||||
let bank_snapshots_dir = tempfile::tempdir_in(farf_dir()).unwrap();
|
|
||||||
let snapshot_archives_dir = tempfile::tempdir_in(farf_dir()).unwrap();
|
|
||||||
let snapshot_config = SnapshotConfig {
|
|
||||||
full_snapshot_archive_interval_slots,
|
|
||||||
incremental_snapshot_archive_interval_slots,
|
|
||||||
snapshot_archives_dir: snapshot_archives_dir.path().to_path_buf(),
|
|
||||||
bank_snapshots_dir: bank_snapshots_dir.path().to_path_buf(),
|
|
||||||
maximum_full_snapshot_archives_to_retain: usize::MAX,
|
|
||||||
maximum_incremental_snapshot_archives_to_retain: usize::MAX,
|
|
||||||
..SnapshotConfig::default()
|
|
||||||
};
|
|
||||||
|
|
||||||
// Create the account paths
|
|
||||||
let (account_storage_dirs, account_storage_paths) =
|
|
||||||
generate_account_paths(num_account_paths);
|
|
||||||
|
|
||||||
// Create the validator config
|
|
||||||
let validator_config = ValidatorConfig {
|
|
||||||
snapshot_config: Some(snapshot_config),
|
|
||||||
account_paths: account_storage_paths,
|
|
||||||
accounts_hash_interval_slots,
|
|
||||||
..ValidatorConfig::default_for_test()
|
|
||||||
};
|
|
||||||
|
|
||||||
SnapshotValidatorConfig {
|
|
||||||
bank_snapshots_dir,
|
|
||||||
snapshot_archives_dir,
|
|
||||||
account_storage_dirs,
|
|
||||||
validator_config,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
fn setup_snapshot_validator_config(
|
|
||||||
snapshot_interval_slots: Slot,
|
|
||||||
num_account_paths: usize,
|
|
||||||
) -> SnapshotValidatorConfig {
|
|
||||||
SnapshotValidatorConfig::new(
|
|
||||||
snapshot_interval_slots,
|
|
||||||
Slot::MAX,
|
|
||||||
snapshot_interval_slots,
|
|
||||||
num_account_paths,
|
|
||||||
)
|
|
||||||
}
|
|
||||||
|
|
|
@ -2,12 +2,7 @@
|
||||||
//! biggest improvements to CI times can be found.
|
//! biggest improvements to CI times can be found.
|
||||||
#![allow(clippy::integer_arithmetic)]
|
#![allow(clippy::integer_arithmetic)]
|
||||||
use {
|
use {
|
||||||
common::{
|
common::*,
|
||||||
copy_blocks, create_custom_leader_schedule_with_random_keys, last_vote_in_tower,
|
|
||||||
ms_for_n_slots, open_blockstore, restore_tower, run_cluster_partition,
|
|
||||||
run_kill_partition_switch_threshold, test_faulty_node,
|
|
||||||
wait_for_last_vote_in_tower_to_land_in_ledger, RUST_LOG_FILTER,
|
|
||||||
},
|
|
||||||
log::*,
|
log::*,
|
||||||
serial_test::serial,
|
serial_test::serial,
|
||||||
solana_core::{
|
solana_core::{
|
|
@ -0,0 +1,440 @@
|
||||||
|
//! If a test takes over 100s to run on CI, move it here so that it's clear where the
|
||||||
|
//! biggest improvements to CI times can be found.
|
||||||
|
#![allow(clippy::integer_arithmetic)]
|
||||||
|
use {
|
||||||
|
common::*,
|
||||||
|
log::*,
|
||||||
|
serial_test::serial,
|
||||||
|
solana_core::validator::ValidatorConfig,
|
||||||
|
solana_gossip::gossip_service::discover_cluster,
|
||||||
|
solana_ledger::{ancestor_iterator::AncestorIterator, blockstore::Blockstore},
|
||||||
|
solana_local_cluster::{
|
||||||
|
cluster::Cluster,
|
||||||
|
cluster_tests,
|
||||||
|
local_cluster::{ClusterConfig, LocalCluster},
|
||||||
|
validator_configs::*,
|
||||||
|
},
|
||||||
|
solana_sdk::{
|
||||||
|
client::SyncClient,
|
||||||
|
clock::Slot,
|
||||||
|
poh_config::PohConfig,
|
||||||
|
signature::{Keypair, Signer},
|
||||||
|
},
|
||||||
|
solana_streamer::socket::SocketAddrSpace,
|
||||||
|
std::{collections::HashSet, sync::Arc, thread::sleep, time::Duration},
|
||||||
|
};
|
||||||
|
|
||||||
|
mod common;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
#[serial]
|
||||||
|
fn test_cluster_partition_1_1() {
|
||||||
|
let empty = |_: &mut LocalCluster, _: &mut ()| {};
|
||||||
|
let on_partition_resolved = |cluster: &mut LocalCluster, _: &mut ()| {
|
||||||
|
cluster.check_for_new_roots(16, "PARTITION_TEST", SocketAddrSpace::Unspecified);
|
||||||
|
};
|
||||||
|
run_cluster_partition(
|
||||||
|
&[1, 1],
|
||||||
|
None,
|
||||||
|
(),
|
||||||
|
empty,
|
||||||
|
empty,
|
||||||
|
on_partition_resolved,
|
||||||
|
None,
|
||||||
|
vec![],
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
#[serial]
|
||||||
|
fn test_cluster_partition_1_1_1() {
|
||||||
|
let empty = |_: &mut LocalCluster, _: &mut ()| {};
|
||||||
|
let on_partition_resolved = |cluster: &mut LocalCluster, _: &mut ()| {
|
||||||
|
cluster.check_for_new_roots(16, "PARTITION_TEST", SocketAddrSpace::Unspecified);
|
||||||
|
};
|
||||||
|
run_cluster_partition(
|
||||||
|
&[1, 1, 1],
|
||||||
|
None,
|
||||||
|
(),
|
||||||
|
empty,
|
||||||
|
empty,
|
||||||
|
on_partition_resolved,
|
||||||
|
None,
|
||||||
|
vec![],
|
||||||
|
)
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
#[serial]
|
||||||
|
fn test_consistency_halt() {
|
||||||
|
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
||||||
|
let snapshot_interval_slots = 20;
|
||||||
|
let num_account_paths = 1;
|
||||||
|
|
||||||
|
// Create cluster with a leader producing bad snapshot hashes.
|
||||||
|
let mut leader_snapshot_test_config =
|
||||||
|
setup_snapshot_validator_config(snapshot_interval_slots, num_account_paths);
|
||||||
|
leader_snapshot_test_config
|
||||||
|
.validator_config
|
||||||
|
.accounts_hash_fault_injection_slots = 40;
|
||||||
|
|
||||||
|
let validator_stake = 10_000;
|
||||||
|
let mut config = ClusterConfig {
|
||||||
|
node_stakes: vec![validator_stake],
|
||||||
|
cluster_lamports: 100_000,
|
||||||
|
validator_configs: vec![leader_snapshot_test_config.validator_config],
|
||||||
|
..ClusterConfig::default()
|
||||||
|
};
|
||||||
|
|
||||||
|
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
||||||
|
|
||||||
|
sleep(Duration::from_millis(5000));
|
||||||
|
let cluster_nodes = discover_cluster(
|
||||||
|
&cluster.entry_point_info.gossip,
|
||||||
|
1,
|
||||||
|
SocketAddrSpace::Unspecified,
|
||||||
|
)
|
||||||
|
.unwrap();
|
||||||
|
info!("num_nodes: {}", cluster_nodes.len());
|
||||||
|
|
||||||
|
// Add a validator with the leader as trusted, it should halt when it detects
|
||||||
|
// mismatch.
|
||||||
|
let mut validator_snapshot_test_config =
|
||||||
|
setup_snapshot_validator_config(snapshot_interval_slots, num_account_paths);
|
||||||
|
|
||||||
|
let mut known_validators = HashSet::new();
|
||||||
|
known_validators.insert(cluster_nodes[0].id);
|
||||||
|
|
||||||
|
validator_snapshot_test_config
|
||||||
|
.validator_config
|
||||||
|
.known_validators = Some(known_validators);
|
||||||
|
validator_snapshot_test_config
|
||||||
|
.validator_config
|
||||||
|
.halt_on_known_validators_accounts_hash_mismatch = true;
|
||||||
|
|
||||||
|
warn!("adding a validator");
|
||||||
|
cluster.add_validator(
|
||||||
|
&validator_snapshot_test_config.validator_config,
|
||||||
|
validator_stake as u64,
|
||||||
|
Arc::new(Keypair::new()),
|
||||||
|
None,
|
||||||
|
SocketAddrSpace::Unspecified,
|
||||||
|
);
|
||||||
|
let num_nodes = 2;
|
||||||
|
assert_eq!(
|
||||||
|
discover_cluster(
|
||||||
|
&cluster.entry_point_info.gossip,
|
||||||
|
num_nodes,
|
||||||
|
SocketAddrSpace::Unspecified
|
||||||
|
)
|
||||||
|
.unwrap()
|
||||||
|
.len(),
|
||||||
|
num_nodes
|
||||||
|
);
|
||||||
|
|
||||||
|
// Check for only 1 node on the network.
|
||||||
|
let mut encountered_error = false;
|
||||||
|
loop {
|
||||||
|
let discover = discover_cluster(
|
||||||
|
&cluster.entry_point_info.gossip,
|
||||||
|
2,
|
||||||
|
SocketAddrSpace::Unspecified,
|
||||||
|
);
|
||||||
|
match discover {
|
||||||
|
Err(_) => {
|
||||||
|
encountered_error = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
Ok(nodes) => {
|
||||||
|
if nodes.len() < 2 {
|
||||||
|
encountered_error = true;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
info!("checking cluster for fewer nodes.. {:?}", nodes.len());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
let client = cluster
|
||||||
|
.get_validator_client(&cluster.entry_point_info.id)
|
||||||
|
.unwrap();
|
||||||
|
if let Ok(slot) = client.get_slot() {
|
||||||
|
if slot > 210 {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
info!("slot: {}", slot);
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(1000));
|
||||||
|
}
|
||||||
|
assert!(encountered_error);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Cluster needs a supermajority to remain, so the minimum size for this test is 4
|
||||||
|
#[test]
|
||||||
|
#[serial]
|
||||||
|
fn test_leader_failure_4() {
|
||||||
|
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
||||||
|
error!("test_leader_failure_4");
|
||||||
|
let num_nodes = 4;
|
||||||
|
let validator_config = ValidatorConfig::default_for_test();
|
||||||
|
let mut config = ClusterConfig {
|
||||||
|
cluster_lamports: 10_000,
|
||||||
|
node_stakes: vec![100; 4],
|
||||||
|
validator_configs: make_identical_validator_configs(&validator_config, num_nodes),
|
||||||
|
..ClusterConfig::default()
|
||||||
|
};
|
||||||
|
let local = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
||||||
|
|
||||||
|
cluster_tests::kill_entry_and_spend_and_verify_rest(
|
||||||
|
&local.entry_point_info,
|
||||||
|
&local
|
||||||
|
.validators
|
||||||
|
.get(&local.entry_point_info.id)
|
||||||
|
.unwrap()
|
||||||
|
.config
|
||||||
|
.validator_exit,
|
||||||
|
&local.funding_keypair,
|
||||||
|
num_nodes,
|
||||||
|
config.ticks_per_slot * config.poh_config.target_tick_duration.as_millis() as u64,
|
||||||
|
SocketAddrSpace::Unspecified,
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
#[serial]
|
||||||
|
fn test_ledger_cleanup_service() {
|
||||||
|
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
||||||
|
error!("test_ledger_cleanup_service");
|
||||||
|
let num_nodes = 3;
|
||||||
|
let validator_config = ValidatorConfig {
|
||||||
|
max_ledger_shreds: Some(100),
|
||||||
|
..ValidatorConfig::default_for_test()
|
||||||
|
};
|
||||||
|
let mut config = ClusterConfig {
|
||||||
|
cluster_lamports: 10_000,
|
||||||
|
poh_config: PohConfig::new_sleep(Duration::from_millis(50)),
|
||||||
|
node_stakes: vec![100; num_nodes],
|
||||||
|
validator_configs: make_identical_validator_configs(&validator_config, num_nodes),
|
||||||
|
..ClusterConfig::default()
|
||||||
|
};
|
||||||
|
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
||||||
|
// 200ms/per * 100 = 20 seconds, so sleep a little longer than that.
|
||||||
|
sleep(Duration::from_secs(60));
|
||||||
|
|
||||||
|
cluster_tests::spend_and_verify_all_nodes(
|
||||||
|
&cluster.entry_point_info,
|
||||||
|
&cluster.funding_keypair,
|
||||||
|
num_nodes,
|
||||||
|
HashSet::new(),
|
||||||
|
SocketAddrSpace::Unspecified,
|
||||||
|
);
|
||||||
|
cluster.close_preserve_ledgers();
|
||||||
|
//check everyone's ledgers and make sure only ~100 slots are stored
|
||||||
|
for info in cluster.validators.values() {
|
||||||
|
let mut slots = 0;
|
||||||
|
let blockstore = Blockstore::open(&info.info.ledger_path).unwrap();
|
||||||
|
blockstore
|
||||||
|
.slot_meta_iterator(0)
|
||||||
|
.unwrap()
|
||||||
|
.for_each(|_| slots += 1);
|
||||||
|
// with 3 nodes up to 3 slots can be in progress and not complete so max slots in blockstore should be up to 103
|
||||||
|
assert!(slots <= 103, "got {}", slots);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// This test verifies that even if votes from a validator end up taking too long to land, and thus
|
||||||
|
// some of the referenced slots are slots are no longer present in the slot hashes sysvar,
|
||||||
|
// consensus can still be attained.
|
||||||
|
//
|
||||||
|
// Validator A (60%)
|
||||||
|
// Validator B (40%)
|
||||||
|
// / --- 10 --- [..] --- 16 (B is voting, due to network issues is initally not able to see the other fork at all)
|
||||||
|
// /
|
||||||
|
// 1 - 2 - 3 - 4 - 5 - 6 - 7 - 8 - 9 (A votes 1 - 9 votes are landing normally. B does the same however votes are not landing)
|
||||||
|
// \
|
||||||
|
// \--[..]-- 73 (majority fork)
|
||||||
|
// A is voting on the majority fork and B wants to switch to this fork however in this majority fork
|
||||||
|
// the earlier votes for B (1 - 9) never landed so when B eventually goes to vote on 73, slots in
|
||||||
|
// its local vote state are no longer present in slot hashes.
|
||||||
|
//
|
||||||
|
// 1. Wait for B's tower to see local vote state was updated to new fork
|
||||||
|
// 2. Wait X blocks, check B's vote state on chain has been properly updated
|
||||||
|
//
|
||||||
|
// NOTE: it is not reliable for B to organically have 1 to reach 2^16 lockout, so we simulate the 6
|
||||||
|
// consecutive votes on the minor fork by manually incrementing the confirmation levels for the
|
||||||
|
// common ancestor votes in tower.
|
||||||
|
// To allow this test to run in a reasonable time we change the
|
||||||
|
// slot_hash expiry to 64 slots.
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn test_slot_hash_expiry() {
|
||||||
|
solana_logger::setup_with_default(RUST_LOG_FILTER);
|
||||||
|
solana_sdk::slot_hashes::set_entries_for_tests_only(64);
|
||||||
|
|
||||||
|
let slots_per_epoch = 2048;
|
||||||
|
let node_stakes = vec![60, 40];
|
||||||
|
let validator_keys = vec![
|
||||||
|
"28bN3xyvrP4E8LwEgtLjhnkb7cY4amQb6DrYAbAYjgRV4GAGgkVM2K7wnxnAS7WDneuavza7x21MiafLu1HkwQt4",
|
||||||
|
"2saHBBoTkLMmttmPQP8KfBkcCw45S5cwtV3wTdGCscRC8uxdgvHxpHiWXKx4LvJjNJtnNcbSv5NdheokFFqnNDt8",
|
||||||
|
]
|
||||||
|
.iter()
|
||||||
|
.map(|s| (Arc::new(Keypair::from_base58_string(s)), true))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let node_vote_keys = vec![
|
||||||
|
"3NDQ3ud86RTVg8hTy2dDWnS4P8NfjhZ2gDgQAJbr3heaKaUVS1FW3sTLKA1GmDrY9aySzsa4QxpDkbLv47yHxzr3",
|
||||||
|
"46ZHpHE6PEvXYPu3hf9iQqjBk2ZNDaJ9ejqKWHEjxaQjpAGasKaWKbKHbP3646oZhfgDRzx95DH9PCBKKsoCVngk",
|
||||||
|
]
|
||||||
|
.iter()
|
||||||
|
.map(|s| Arc::new(Keypair::from_base58_string(s)))
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let vs = validator_keys
|
||||||
|
.iter()
|
||||||
|
.map(|(kp, _)| kp.pubkey())
|
||||||
|
.collect::<Vec<_>>();
|
||||||
|
let (a_pubkey, b_pubkey) = (vs[0], vs[1]);
|
||||||
|
|
||||||
|
// We want B to not vote (we are trying to simulate its votes not landing until it gets to the
|
||||||
|
// minority fork)
|
||||||
|
let mut validator_configs =
|
||||||
|
make_identical_validator_configs(&ValidatorConfig::default_for_test(), node_stakes.len());
|
||||||
|
validator_configs[1].voting_disabled = true;
|
||||||
|
|
||||||
|
let mut config = ClusterConfig {
|
||||||
|
cluster_lamports: 100_000,
|
||||||
|
node_stakes,
|
||||||
|
validator_configs,
|
||||||
|
validator_keys: Some(validator_keys),
|
||||||
|
node_vote_keys: Some(node_vote_keys),
|
||||||
|
slots_per_epoch,
|
||||||
|
stakers_slot_offset: slots_per_epoch,
|
||||||
|
skip_warmup_slots: true,
|
||||||
|
..ClusterConfig::default()
|
||||||
|
};
|
||||||
|
let mut cluster = LocalCluster::new(&mut config, SocketAddrSpace::Unspecified);
|
||||||
|
|
||||||
|
let mut common_ancestor_slot = 8;
|
||||||
|
|
||||||
|
let a_ledger_path = cluster.ledger_path(&a_pubkey);
|
||||||
|
let b_ledger_path = cluster.ledger_path(&b_pubkey);
|
||||||
|
|
||||||
|
// Immediately kill B (we just needed it for the initial stake distribution)
|
||||||
|
info!("Killing B");
|
||||||
|
let mut b_info = cluster.exit_node(&b_pubkey);
|
||||||
|
|
||||||
|
// Let A run for a while until we get to the common ancestor
|
||||||
|
info!("Letting A run until common_ancestor_slot");
|
||||||
|
loop {
|
||||||
|
if let Some((last_vote, _)) = last_vote_in_tower(&a_ledger_path, &a_pubkey) {
|
||||||
|
if last_vote >= common_ancestor_slot {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(100));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Keep A running, but setup B so that it thinks it has voted up until common ancestor (but
|
||||||
|
// doesn't know anything past that)
|
||||||
|
{
|
||||||
|
info!("Copying A's ledger to B");
|
||||||
|
std::fs::remove_dir_all(&b_info.info.ledger_path).unwrap();
|
||||||
|
let mut opt = fs_extra::dir::CopyOptions::new();
|
||||||
|
opt.copy_inside = true;
|
||||||
|
fs_extra::dir::copy(&a_ledger_path, &b_ledger_path, &opt).unwrap();
|
||||||
|
|
||||||
|
// remove A's tower in B's new copied ledger
|
||||||
|
info!("Removing A's tower in B's ledger dir");
|
||||||
|
remove_tower(&b_ledger_path, &a_pubkey);
|
||||||
|
|
||||||
|
// load A's tower and save it as B's tower
|
||||||
|
info!("Loading A's tower");
|
||||||
|
if let Some(mut a_tower) = restore_tower(&a_ledger_path, &a_pubkey) {
|
||||||
|
a_tower.node_pubkey = b_pubkey;
|
||||||
|
// Update common_ancestor_slot because A is still running
|
||||||
|
if let Some(s) = a_tower.last_voted_slot() {
|
||||||
|
common_ancestor_slot = s;
|
||||||
|
info!("New common_ancestor_slot {}", common_ancestor_slot);
|
||||||
|
} else {
|
||||||
|
panic!("A's tower has no votes");
|
||||||
|
}
|
||||||
|
info!("Increase lockout by 6 confirmation levels and save as B's tower");
|
||||||
|
a_tower.increase_lockout(6);
|
||||||
|
save_tower(&b_ledger_path, &a_tower, &b_info.info.keypair);
|
||||||
|
info!("B's new tower: {:?}", a_tower.tower_slots());
|
||||||
|
} else {
|
||||||
|
panic!("A's tower is missing");
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get rid of any slots past common_ancestor_slot
|
||||||
|
info!("Removing extra slots from B's blockstore");
|
||||||
|
let blockstore = open_blockstore(&b_ledger_path);
|
||||||
|
purge_slots(&blockstore, common_ancestor_slot + 1, 100);
|
||||||
|
}
|
||||||
|
|
||||||
|
info!(
|
||||||
|
"Run A on majority fork until it reaches slot hash expiry {}",
|
||||||
|
solana_sdk::slot_hashes::get_entries()
|
||||||
|
);
|
||||||
|
let mut last_vote_on_a;
|
||||||
|
// Keep A running for a while longer so the majority fork has some decent size
|
||||||
|
loop {
|
||||||
|
last_vote_on_a = wait_for_last_vote_in_tower_to_land_in_ledger(&a_ledger_path, &a_pubkey);
|
||||||
|
if last_vote_on_a
|
||||||
|
>= common_ancestor_slot + 2 * (solana_sdk::slot_hashes::get_entries() as u64)
|
||||||
|
{
|
||||||
|
let blockstore = open_blockstore(&a_ledger_path);
|
||||||
|
info!(
|
||||||
|
"A majority fork: {:?}",
|
||||||
|
AncestorIterator::new(last_vote_on_a, &blockstore).collect::<Vec<Slot>>()
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(100));
|
||||||
|
}
|
||||||
|
|
||||||
|
// Kill A and restart B with voting. B should now fork off
|
||||||
|
info!("Killing A");
|
||||||
|
let a_info = cluster.exit_node(&a_pubkey);
|
||||||
|
|
||||||
|
info!("Restarting B");
|
||||||
|
b_info.config.voting_disabled = false;
|
||||||
|
cluster.restart_node(&b_pubkey, b_info, SocketAddrSpace::Unspecified);
|
||||||
|
|
||||||
|
// B will fork off and accumulate enough lockout
|
||||||
|
info!("Allowing B to fork");
|
||||||
|
loop {
|
||||||
|
let blockstore = open_blockstore(&b_ledger_path);
|
||||||
|
let last_vote = wait_for_last_vote_in_tower_to_land_in_ledger(&b_ledger_path, &b_pubkey);
|
||||||
|
let mut ancestors = AncestorIterator::new(last_vote, &blockstore);
|
||||||
|
if let Some(index) = ancestors.position(|x| x == common_ancestor_slot) {
|
||||||
|
if index > 7 {
|
||||||
|
info!(
|
||||||
|
"B has forked for enough lockout: {:?}",
|
||||||
|
AncestorIterator::new(last_vote, &blockstore).collect::<Vec<Slot>>()
|
||||||
|
);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
sleep(Duration::from_millis(1000));
|
||||||
|
}
|
||||||
|
|
||||||
|
info!("Kill B");
|
||||||
|
b_info = cluster.exit_node(&b_pubkey);
|
||||||
|
|
||||||
|
info!("Resolve the partition");
|
||||||
|
{
|
||||||
|
// Here we let B know about the missing blocks that A had produced on its partition
|
||||||
|
let a_blockstore = open_blockstore(&a_ledger_path);
|
||||||
|
let b_blockstore = open_blockstore(&b_ledger_path);
|
||||||
|
copy_blocks(last_vote_on_a, &a_blockstore, &b_blockstore);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Now restart A and B and see if B is able to eventually switch onto the majority fork
|
||||||
|
info!("Restarting A & B");
|
||||||
|
cluster.restart_node(&a_pubkey, a_info, SocketAddrSpace::Unspecified);
|
||||||
|
cluster.restart_node(&b_pubkey, b_info, SocketAddrSpace::Unspecified);
|
||||||
|
|
||||||
|
info!("Waiting for B to switch to majority fork and make a root");
|
||||||
|
cluster_tests::check_for_new_roots(
|
||||||
|
16,
|
||||||
|
&[cluster.get_contact_info(&a_pubkey).unwrap().clone()],
|
||||||
|
"test_slot_hashes_expiry",
|
||||||
|
);
|
||||||
|
}
|
Loading…
Reference in New Issue