Add newly completed slots signal to Blocktree (#4225)
* Add channel to blocktree for communicating when slots are completed * Refactor RepairService options into a RepairStrategy
This commit is contained in:
parent
a031b09190
commit
575a0e318b
|
@ -28,7 +28,7 @@ use std::cmp;
|
|||
use std::fs;
|
||||
use std::io;
|
||||
use std::rc::Rc;
|
||||
use std::sync::mpsc::{sync_channel, Receiver, SyncSender};
|
||||
use std::sync::mpsc::{sync_channel, Receiver, SyncSender, TrySendError};
|
||||
use std::sync::{Arc, RwLock};
|
||||
|
||||
pub use self::meta::*;
|
||||
|
@ -63,6 +63,10 @@ db_imports! {rocks, Rocks, "rocksdb"}
|
|||
#[cfg(feature = "kvstore")]
|
||||
db_imports! {kvs, Kvs, "kvstore"}
|
||||
|
||||
pub const MAX_COMPLETED_SLOTS_IN_CHANNEL: usize = 100_000;
|
||||
|
||||
pub type CompletedSlotsReceiver = Receiver<Vec<u64>>;
|
||||
|
||||
#[derive(Debug)]
|
||||
pub enum BlocktreeError {
|
||||
BlobForIndexExists,
|
||||
|
@ -83,6 +87,7 @@ pub struct Blocktree {
|
|||
batch_processor: Arc<RwLock<BatchProcessor>>,
|
||||
session: Arc<erasure::Session>,
|
||||
pub new_blobs_signals: Vec<SyncSender<bool>>,
|
||||
pub completed_slots_senders: Vec<SyncSender<Vec<u64>>>,
|
||||
}
|
||||
|
||||
// Column family for metadata about a leader slot
|
||||
|
@ -141,15 +146,21 @@ impl Blocktree {
|
|||
session,
|
||||
new_blobs_signals: vec![],
|
||||
batch_processor,
|
||||
completed_slots_senders: vec![],
|
||||
})
|
||||
}
|
||||
|
||||
pub fn open_with_signal(ledger_path: &str) -> Result<(Self, Receiver<bool>)> {
|
||||
pub fn open_with_signal(
|
||||
ledger_path: &str,
|
||||
) -> Result<(Self, Receiver<bool>, CompletedSlotsReceiver)> {
|
||||
let mut blocktree = Self::open(ledger_path)?;
|
||||
let (signal_sender, signal_receiver) = sync_channel(1);
|
||||
let (completed_slots_sender, completed_slots_receiver) =
|
||||
sync_channel(MAX_COMPLETED_SLOTS_IN_CHANNEL);
|
||||
blocktree.new_blobs_signals = vec![signal_sender];
|
||||
blocktree.completed_slots_senders = vec![completed_slots_sender];
|
||||
|
||||
Ok((blocktree, signal_receiver))
|
||||
Ok((blocktree, signal_receiver, completed_slots_receiver))
|
||||
}
|
||||
|
||||
pub fn destroy(ledger_path: &str) -> Result<()> {
|
||||
|
@ -340,11 +351,17 @@ impl Blocktree {
|
|||
// Handle chaining for the working set
|
||||
handle_chaining(&db, &mut write_batch, &slot_meta_working_set)?;
|
||||
let mut should_signal = false;
|
||||
let mut newly_completed_slots = vec![];
|
||||
|
||||
// Check if any metadata was changed, if so, insert the new version of the
|
||||
// metadata into the write batch
|
||||
for (slot, (meta, meta_backup)) in slot_meta_working_set.iter() {
|
||||
let meta: &SlotMeta = &RefCell::borrow(&*meta);
|
||||
if !self.completed_slots_senders.is_empty()
|
||||
&& is_newly_completed_slot(meta, meta_backup)
|
||||
{
|
||||
newly_completed_slots.push(*slot);
|
||||
}
|
||||
// Check if the working copy of the metadata has changed
|
||||
if Some(meta) != meta_backup.as_ref() {
|
||||
should_signal = should_signal || slot_has_updates(meta, &meta_backup);
|
||||
|
@ -356,13 +373,38 @@ impl Blocktree {
|
|||
write_batch.put::<cf::ErasureMeta>((slot, set_index), &erasure_meta)?;
|
||||
}
|
||||
|
||||
batch_processor.write(write_batch)?;
|
||||
|
||||
if should_signal {
|
||||
for signal in self.new_blobs_signals.iter() {
|
||||
for signal in &self.new_blobs_signals {
|
||||
let _ = signal.try_send(true);
|
||||
}
|
||||
}
|
||||
|
||||
batch_processor.write(write_batch)?;
|
||||
if !self.completed_slots_senders.is_empty() && !newly_completed_slots.is_empty() {
|
||||
let mut slots: Vec<_> = (0..self.completed_slots_senders.len() - 1)
|
||||
.map(|_| newly_completed_slots.clone())
|
||||
.collect();
|
||||
|
||||
slots.push(newly_completed_slots);
|
||||
|
||||
for (signal, slots) in self.completed_slots_senders.iter().zip(slots.into_iter()) {
|
||||
let res = signal.try_send(slots);
|
||||
if let Err(TrySendError::Full(_)) = res {
|
||||
solana_metrics::submit(
|
||||
solana_metrics::influxdb::Point::new("blocktree_error")
|
||||
.add_field(
|
||||
"error",
|
||||
solana_metrics::influxdb::Value::String(
|
||||
"Unable to send newly completed slot because channel is full"
|
||||
.to_string(),
|
||||
),
|
||||
)
|
||||
.to_owned(),
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
Ok(())
|
||||
}
|
||||
|
@ -880,7 +922,7 @@ fn insert_data_blob<'a>(
|
|||
slot_meta.received = cmp::max(blob_index + 1, slot_meta.received);
|
||||
slot_meta.consumed = new_consumed;
|
||||
slot_meta.last_index = {
|
||||
// If the last slot hasn't been set before, then
|
||||
// If the last index in the slot hasn't been set before, then
|
||||
// set it to this blob index
|
||||
if slot_meta.last_index == std::u64::MAX {
|
||||
if blob_to_insert.is_last_in_slot() {
|
||||
|
@ -1123,9 +1165,8 @@ fn handle_chaining_for_slot(
|
|||
.expect("Slot must exist in the working_set hashmap");
|
||||
|
||||
{
|
||||
let is_orphaned = meta_backup.is_some() && is_orphan(meta_backup.as_ref().unwrap());
|
||||
|
||||
let mut meta_mut = meta.borrow_mut();
|
||||
let was_orphan_slot = meta_backup.is_some() && is_orphan(meta_backup.as_ref().unwrap());
|
||||
|
||||
// If:
|
||||
// 1) This is a new slot
|
||||
|
@ -1137,27 +1178,32 @@ fn handle_chaining_for_slot(
|
|||
// Check if the slot represented by meta_mut is either a new slot or a orphan.
|
||||
// In both cases we need to run the chaining logic b/c the parent on the slot was
|
||||
// previously unknown.
|
||||
if meta_backup.is_none() || is_orphaned {
|
||||
if meta_backup.is_none() || was_orphan_slot {
|
||||
let prev_slot_meta =
|
||||
find_slot_meta_else_create(db, working_set, new_chained_slots, prev_slot)?;
|
||||
|
||||
// This is a newly inserted slot so run the chaining logic
|
||||
// This is a newly inserted slot/orphan so run the chaining logic to link it to a
|
||||
// newly discovered parent
|
||||
chain_new_slot_to_prev_slot(&mut prev_slot_meta.borrow_mut(), slot, &mut meta_mut);
|
||||
|
||||
// If the parent of `slot` is a newly inserted orphan, insert it into the orphans
|
||||
// column family
|
||||
if is_orphan(&RefCell::borrow(&*prev_slot_meta)) {
|
||||
write_batch.put::<cf::Orphans>(prev_slot, &true)?;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// At this point this slot has received a parent, so no longer a orphan
|
||||
if is_orphaned {
|
||||
// At this point this slot has received a parent, so it's no longer an orphan
|
||||
if was_orphan_slot {
|
||||
write_batch.delete::<cf::Orphans>(slot)?;
|
||||
}
|
||||
}
|
||||
|
||||
// This is a newly inserted slot and slot.is_connected is true, so update all
|
||||
// child slots so that their `is_connected` = true
|
||||
// If this is a newly inserted slot, then we know the children of this slot were not previously
|
||||
// connected to the trunk of the ledger. Thus if slot.is_connected is now true, we need to
|
||||
// update all child slots with `is_connected` = true because these children are also now newly
|
||||
// connected to to trunk of the the ledger
|
||||
let should_propagate_is_connected =
|
||||
is_newly_completed_slot(&RefCell::borrow(&*meta), meta_backup)
|
||||
&& RefCell::borrow(&*meta).is_connected;
|
||||
|
@ -1238,7 +1284,6 @@ fn chain_new_slot_to_prev_slot(
|
|||
fn is_newly_completed_slot(slot_meta: &SlotMeta, backup_slot_meta: &Option<SlotMeta>) -> bool {
|
||||
slot_meta.is_full()
|
||||
&& (backup_slot_meta.is_none()
|
||||
|| is_orphan(&backup_slot_meta.as_ref().unwrap())
|
||||
|| slot_meta.consumed != backup_slot_meta.as_ref().unwrap().consumed)
|
||||
}
|
||||
|
||||
|
@ -2112,7 +2157,7 @@ pub mod tests {
|
|||
pub fn test_new_blobs_signal() {
|
||||
// Initialize ledger
|
||||
let ledger_path = get_tmp_ledger_path("test_new_blobs_signal");
|
||||
let (ledger, recvr) = Blocktree::open_with_signal(&ledger_path).unwrap();
|
||||
let (ledger, recvr, _) = Blocktree::open_with_signal(&ledger_path).unwrap();
|
||||
let ledger = Arc::new(ledger);
|
||||
|
||||
let entries_per_slot = 10;
|
||||
|
@ -2188,6 +2233,98 @@ pub mod tests {
|
|||
Blocktree::destroy(&ledger_path).expect("Expected successful database destruction");
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn test_completed_blobs_signal() {
|
||||
// Initialize ledger
|
||||
let ledger_path = get_tmp_ledger_path("test_completed_blobs_signal");
|
||||
let (ledger, _, recvr) = Blocktree::open_with_signal(&ledger_path).unwrap();
|
||||
let ledger = Arc::new(ledger);
|
||||
|
||||
let entries_per_slot = 10;
|
||||
|
||||
// Create blobs for slot 0
|
||||
let (blobs, _) = make_slot_entries(0, 0, entries_per_slot);
|
||||
|
||||
// Insert all but the first blob in the slot, should not be considered complete
|
||||
ledger
|
||||
.insert_data_blobs(&blobs[1..entries_per_slot as usize])
|
||||
.unwrap();
|
||||
assert!(recvr.try_recv().is_err());
|
||||
|
||||
// Insert first blob, slot should now be considered complete
|
||||
ledger.insert_data_blobs(once(&blobs[0])).unwrap();
|
||||
assert_eq!(recvr.try_recv().unwrap(), vec![0]);
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn test_completed_blobs_signal_orphans() {
|
||||
// Initialize ledger
|
||||
let ledger_path = get_tmp_ledger_path("test_completed_blobs_signal_orphans");
|
||||
let (ledger, _, recvr) = Blocktree::open_with_signal(&ledger_path).unwrap();
|
||||
let ledger = Arc::new(ledger);
|
||||
|
||||
let entries_per_slot = 10;
|
||||
let slots = vec![2, 5, 10];
|
||||
let all_blobs = make_chaining_slot_entries(&slots[..], entries_per_slot);
|
||||
|
||||
// Get the blobs for slot 5 chaining to slot 2
|
||||
let (ref orphan_blobs, _) = all_blobs[1];
|
||||
|
||||
// Get the blobs for slot 10, chaining to slot 5
|
||||
let (ref orphan_child, _) = all_blobs[2];
|
||||
|
||||
// Insert all but the first blob in the slot, should not be considered complete
|
||||
ledger
|
||||
.insert_data_blobs(&orphan_child[1..entries_per_slot as usize])
|
||||
.unwrap();
|
||||
assert!(recvr.try_recv().is_err());
|
||||
|
||||
// Insert first blob, slot should now be considered complete
|
||||
ledger.insert_data_blobs(once(&orphan_child[0])).unwrap();
|
||||
assert_eq!(recvr.try_recv().unwrap(), vec![slots[2]]);
|
||||
|
||||
// Insert the blobs for the orphan_slot
|
||||
ledger
|
||||
.insert_data_blobs(&orphan_blobs[1..entries_per_slot as usize])
|
||||
.unwrap();
|
||||
assert!(recvr.try_recv().is_err());
|
||||
|
||||
// Insert first blob, slot should now be considered complete
|
||||
ledger.insert_data_blobs(once(&orphan_blobs[0])).unwrap();
|
||||
assert_eq!(recvr.try_recv().unwrap(), vec![slots[1]]);
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn test_completed_blobs_signal_many() {
|
||||
// Initialize ledger
|
||||
let ledger_path = get_tmp_ledger_path("test_completed_blobs_signal_many");
|
||||
let (ledger, _, recvr) = Blocktree::open_with_signal(&ledger_path).unwrap();
|
||||
let ledger = Arc::new(ledger);
|
||||
|
||||
let entries_per_slot = 10;
|
||||
let mut slots = vec![2, 5, 10];
|
||||
let all_blobs = make_chaining_slot_entries(&slots[..], entries_per_slot);
|
||||
let disconnected_slot = 4;
|
||||
|
||||
let (ref blobs0, _) = all_blobs[0];
|
||||
let (ref blobs1, _) = all_blobs[1];
|
||||
let (ref blobs2, _) = all_blobs[2];
|
||||
let (ref blobs3, _) = make_slot_entries(disconnected_slot, 1, entries_per_slot);
|
||||
|
||||
let mut all_blobs: Vec<_> = vec![blobs0, blobs1, blobs2, blobs3]
|
||||
.into_iter()
|
||||
.flatten()
|
||||
.collect();
|
||||
|
||||
all_blobs.shuffle(&mut thread_rng());
|
||||
ledger.insert_data_blobs(all_blobs).unwrap();
|
||||
let mut result = recvr.try_recv().unwrap();
|
||||
result.sort();
|
||||
slots.push(disconnected_slot);
|
||||
slots.sort();
|
||||
assert_eq!(result, slots);
|
||||
}
|
||||
|
||||
#[test]
|
||||
pub fn test_handle_chaining_basic() {
|
||||
let blocktree_path = get_tmp_ledger_path("test_handle_chaining_basic");
|
||||
|
@ -3375,4 +3512,28 @@ pub mod tests {
|
|||
|
||||
(blobs, entries)
|
||||
}
|
||||
|
||||
// Create blobs for slots that have a parent-child relationship defined by the input `chain`
|
||||
pub fn make_chaining_slot_entries(
|
||||
chain: &[u64],
|
||||
entries_per_slot: u64,
|
||||
) -> Vec<(Vec<Blob>, Vec<Entry>)> {
|
||||
let mut slots_blobs_and_entries = vec![];
|
||||
for (i, slot) in chain.iter().enumerate() {
|
||||
let parent_slot = {
|
||||
if *slot == 0 {
|
||||
0
|
||||
} else if i == 0 {
|
||||
std::u64::MAX
|
||||
} else {
|
||||
chain[i - 1]
|
||||
}
|
||||
};
|
||||
|
||||
let result = make_slot_entries(*slot, parent_slot, entries_per_slot);
|
||||
slots_blobs_and_entries.push(result);
|
||||
}
|
||||
|
||||
slots_blobs_and_entries
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1,7 +1,7 @@
|
|||
//! The `fullnode` module hosts all the fullnode microservices.
|
||||
|
||||
use crate::bank_forks::BankForks;
|
||||
use crate::blocktree::Blocktree;
|
||||
use crate::blocktree::{Blocktree, CompletedSlotsReceiver};
|
||||
use crate::blocktree_processor::{self, BankForksInfo};
|
||||
use crate::cluster_info::{ClusterInfo, Node};
|
||||
use crate::contact_info::ContactInfo;
|
||||
|
@ -95,8 +95,14 @@ impl Fullnode {
|
|||
let id = keypair.pubkey();
|
||||
assert_eq!(id, node.info.id);
|
||||
|
||||
let (bank_forks, bank_forks_info, blocktree, ledger_signal_receiver, leader_schedule_cache) =
|
||||
new_banks_from_blocktree(ledger_path, config.account_paths.clone());
|
||||
let (
|
||||
bank_forks,
|
||||
bank_forks_info,
|
||||
blocktree,
|
||||
ledger_signal_receiver,
|
||||
completed_slots_receiver,
|
||||
leader_schedule_cache,
|
||||
) = new_banks_from_blocktree(ledger_path, config.account_paths.clone());
|
||||
|
||||
let leader_schedule_cache = Arc::new(leader_schedule_cache);
|
||||
let exit = Arc::new(AtomicBool::new(false));
|
||||
|
@ -236,6 +242,7 @@ impl Fullnode {
|
|||
&leader_schedule_cache,
|
||||
&exit,
|
||||
&genesis_blockhash,
|
||||
completed_slots_receiver,
|
||||
);
|
||||
|
||||
if config.sigverify_disabled {
|
||||
|
@ -290,12 +297,14 @@ pub fn new_banks_from_blocktree(
|
|||
Vec<BankForksInfo>,
|
||||
Blocktree,
|
||||
Receiver<bool>,
|
||||
CompletedSlotsReceiver,
|
||||
LeaderScheduleCache,
|
||||
) {
|
||||
let genesis_block =
|
||||
GenesisBlock::load(blocktree_path).expect("Expected to successfully open genesis block");
|
||||
|
||||
let (blocktree, ledger_signal_receiver) = Blocktree::open_with_signal(blocktree_path)
|
||||
let (blocktree, ledger_signal_receiver, completed_slots_receiver) =
|
||||
Blocktree::open_with_signal(blocktree_path)
|
||||
.expect("Expected to successfully open database ledger");
|
||||
|
||||
let (bank_forks, bank_forks_info, leader_schedule_cache) =
|
||||
|
@ -307,6 +316,7 @@ pub fn new_banks_from_blocktree(
|
|||
bank_forks_info,
|
||||
blocktree,
|
||||
ledger_signal_receiver,
|
||||
completed_slots_receiver,
|
||||
leader_schedule_cache,
|
||||
)
|
||||
}
|
||||
|
|
|
@ -2,7 +2,7 @@
|
|||
//! regularly finds missing blobs in the ledger and sends repair requests for those blobs
|
||||
|
||||
use crate::bank_forks::BankForks;
|
||||
use crate::blocktree::{Blocktree, SlotMeta};
|
||||
use crate::blocktree::{Blocktree, CompletedSlotsReceiver, SlotMeta};
|
||||
use crate::cluster_info::ClusterInfo;
|
||||
use crate::result::Result;
|
||||
use crate::service::Service;
|
||||
|
@ -22,6 +22,14 @@ pub const MAX_REPAIR_TRIES: u64 = 128;
|
|||
pub const NUM_FORKS_TO_REPAIR: usize = 5;
|
||||
pub const MAX_ORPHANS: usize = 5;
|
||||
|
||||
pub enum RepairStrategy {
|
||||
RepairRange(RepairSlotRange),
|
||||
RepairAll {
|
||||
bank_forks: Arc<RwLock<BankForks>>,
|
||||
completed_slots_receiver: CompletedSlotsReceiver,
|
||||
},
|
||||
}
|
||||
|
||||
#[derive(Serialize, Deserialize, Debug, Clone, Copy, PartialEq, Eq)]
|
||||
pub enum RepairType {
|
||||
Orphan(u64),
|
||||
|
@ -68,8 +76,7 @@ impl RepairService {
|
|||
exit: &Arc<AtomicBool>,
|
||||
repair_socket: Arc<UdpSocket>,
|
||||
cluster_info: Arc<RwLock<ClusterInfo>>,
|
||||
bank_forks: Option<Arc<RwLock<BankForks>>>,
|
||||
repair_slot_range: Option<RepairSlotRange>,
|
||||
repair_strategy: RepairStrategy,
|
||||
) -> Self {
|
||||
let exit = exit.clone();
|
||||
let t_repair = Builder::new()
|
||||
|
@ -80,8 +87,7 @@ impl RepairService {
|
|||
exit,
|
||||
&repair_socket,
|
||||
&cluster_info,
|
||||
&bank_forks,
|
||||
repair_slot_range,
|
||||
repair_strategy,
|
||||
)
|
||||
})
|
||||
.unwrap();
|
||||
|
@ -94,8 +100,7 @@ impl RepairService {
|
|||
exit: Arc<AtomicBool>,
|
||||
repair_socket: &Arc<UdpSocket>,
|
||||
cluster_info: &Arc<RwLock<ClusterInfo>>,
|
||||
bank_forks: &Option<Arc<RwLock<BankForks>>>,
|
||||
repair_slot_range: Option<RepairSlotRange>,
|
||||
repair_strategy: RepairStrategy,
|
||||
) {
|
||||
let mut repair_info = RepairInfo::new();
|
||||
let epoch_slots: HashSet<u64> = HashSet::new();
|
||||
|
@ -106,7 +111,8 @@ impl RepairService {
|
|||
}
|
||||
|
||||
let repairs = {
|
||||
if let Some(ref repair_slot_range) = repair_slot_range {
|
||||
match repair_strategy {
|
||||
RepairStrategy::RepairRange(ref repair_slot_range) => {
|
||||
// Strategy used by replicators
|
||||
Self::generate_repairs_in_range(
|
||||
blocktree,
|
||||
|
@ -114,13 +120,22 @@ impl RepairService {
|
|||
&mut repair_info,
|
||||
repair_slot_range,
|
||||
)
|
||||
} else {
|
||||
let bank_forks = bank_forks
|
||||
.as_ref()
|
||||
.expect("Non-replicator repair strategy missing BankForks");
|
||||
Self::update_fast_repair(id, &epoch_slots, &cluster_info, bank_forks);
|
||||
}
|
||||
|
||||
RepairStrategy::RepairAll {
|
||||
ref bank_forks,
|
||||
ref completed_slots_receiver,
|
||||
} => {
|
||||
Self::update_epoch_slots(
|
||||
id,
|
||||
&epoch_slots,
|
||||
&cluster_info,
|
||||
bank_forks,
|
||||
completed_slots_receiver,
|
||||
);
|
||||
Self::generate_repairs(blocktree, MAX_REPAIR_LENGTH)
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
if let Ok(repairs) = repairs {
|
||||
|
@ -278,11 +293,14 @@ impl RepairService {
|
|||
}
|
||||
}
|
||||
|
||||
fn update_fast_repair(
|
||||
// Update the gossiped structure used for the "Repairmen" repair protocol. See book
|
||||
// for details.
|
||||
fn update_epoch_slots(
|
||||
id: Pubkey,
|
||||
slots: &HashSet<u64>,
|
||||
cluster_info: &RwLock<ClusterInfo>,
|
||||
bank_forks: &Arc<RwLock<BankForks>>,
|
||||
_completed_slots_receiver: &CompletedSlotsReceiver,
|
||||
) {
|
||||
let root = bank_forks.read().unwrap().root();
|
||||
cluster_info
|
||||
|
|
|
@ -6,7 +6,7 @@ use crate::cluster_info::{ClusterInfo, Node, FULLNODE_PORT_RANGE};
|
|||
use crate::contact_info::ContactInfo;
|
||||
use crate::gossip_service::GossipService;
|
||||
use crate::packet::to_shared_blob;
|
||||
use crate::repair_service::RepairSlotRange;
|
||||
use crate::repair_service::{RepairSlotRange, RepairStrategy};
|
||||
use crate::result::Result;
|
||||
use crate::service::Service;
|
||||
use crate::storage_stage::SLOTS_PER_SEGMENT;
|
||||
|
@ -232,14 +232,13 @@ impl Replicator {
|
|||
|
||||
let window_service = WindowService::new(
|
||||
None, //TODO: need a way to validate blobs... https://github.com/solana-labs/solana/issues/3924
|
||||
None, //TODO: see above ^
|
||||
blocktree.clone(),
|
||||
cluster_info.clone(),
|
||||
blob_fetch_receiver,
|
||||
retransmit_sender,
|
||||
repair_socket,
|
||||
&exit,
|
||||
Some(repair_slot_range),
|
||||
RepairStrategy::RepairRange(repair_slot_range),
|
||||
&Hash::default(),
|
||||
);
|
||||
|
||||
|
|
|
@ -1,9 +1,10 @@
|
|||
//! The `retransmit_stage` retransmits blobs between validators
|
||||
|
||||
use crate::bank_forks::BankForks;
|
||||
use crate::blocktree::Blocktree;
|
||||
use crate::blocktree::{Blocktree, CompletedSlotsReceiver};
|
||||
use crate::cluster_info::{compute_retransmit_peers, ClusterInfo, DATA_PLANE_FANOUT};
|
||||
use crate::leader_schedule_cache::LeaderScheduleCache;
|
||||
use crate::repair_service::RepairStrategy;
|
||||
use crate::result::{Error, Result};
|
||||
use crate::service::Service;
|
||||
use crate::staking_utils;
|
||||
|
@ -108,6 +109,7 @@ pub struct RetransmitStage {
|
|||
|
||||
impl RetransmitStage {
|
||||
#[allow(clippy::new_ret_no_self)]
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub fn new(
|
||||
bank_forks: Arc<RwLock<BankForks>>,
|
||||
leader_schedule_cache: &Arc<LeaderScheduleCache>,
|
||||
|
@ -118,6 +120,7 @@ impl RetransmitStage {
|
|||
fetch_stage_receiver: BlobReceiver,
|
||||
exit: &Arc<AtomicBool>,
|
||||
genesis_blockhash: &Hash,
|
||||
completed_slots_receiver: CompletedSlotsReceiver,
|
||||
) -> Self {
|
||||
let (retransmit_sender, retransmit_receiver) = channel();
|
||||
|
||||
|
@ -128,8 +131,12 @@ impl RetransmitStage {
|
|||
cluster_info.clone(),
|
||||
retransmit_receiver,
|
||||
);
|
||||
|
||||
let repair_strategy = RepairStrategy::RepairAll {
|
||||
bank_forks,
|
||||
completed_slots_receiver,
|
||||
};
|
||||
let window_service = WindowService::new(
|
||||
Some(bank_forks),
|
||||
Some(leader_schedule_cache.clone()),
|
||||
blocktree,
|
||||
cluster_info.clone(),
|
||||
|
@ -137,7 +144,7 @@ impl RetransmitStage {
|
|||
retransmit_sender,
|
||||
repair_socket,
|
||||
exit,
|
||||
None,
|
||||
repair_strategy,
|
||||
genesis_blockhash,
|
||||
);
|
||||
|
||||
|
|
|
@ -15,7 +15,7 @@
|
|||
use crate::bank_forks::BankForks;
|
||||
use crate::blob_fetch_stage::BlobFetchStage;
|
||||
use crate::blockstream_service::BlockstreamService;
|
||||
use crate::blocktree::Blocktree;
|
||||
use crate::blocktree::{Blocktree, CompletedSlotsReceiver};
|
||||
use crate::cluster_info::ClusterInfo;
|
||||
use crate::leader_schedule_cache::LeaderScheduleCache;
|
||||
use crate::poh_recorder::PohRecorder;
|
||||
|
@ -71,6 +71,7 @@ impl Tvu {
|
|||
leader_schedule_cache: &Arc<LeaderScheduleCache>,
|
||||
exit: &Arc<AtomicBool>,
|
||||
genesis_blockhash: &Hash,
|
||||
completed_slots_receiver: CompletedSlotsReceiver,
|
||||
) -> Self
|
||||
where
|
||||
T: 'static + KeypairUtil + Sync + Send,
|
||||
|
@ -108,6 +109,7 @@ impl Tvu {
|
|||
blob_fetch_receiver,
|
||||
&exit,
|
||||
genesis_blockhash,
|
||||
completed_slots_receiver,
|
||||
);
|
||||
|
||||
let (replay_stage, slot_full_receiver, root_slot_receiver) = ReplayStage::new(
|
||||
|
@ -203,7 +205,8 @@ pub mod tests {
|
|||
let cref1 = Arc::new(RwLock::new(cluster_info1));
|
||||
|
||||
let blocktree_path = get_tmp_ledger_path!();
|
||||
let (blocktree, l_receiver) = Blocktree::open_with_signal(&blocktree_path)
|
||||
let (blocktree, l_receiver, completed_slots_receiver) =
|
||||
Blocktree::open_with_signal(&blocktree_path)
|
||||
.expect("Expected to successfully open ledger");
|
||||
let blocktree = Arc::new(blocktree);
|
||||
let bank = bank_forks.working_bank();
|
||||
|
@ -233,6 +236,7 @@ pub mod tests {
|
|||
&leader_schedule_cache,
|
||||
&exit,
|
||||
&Hash::default(),
|
||||
completed_slots_receiver,
|
||||
);
|
||||
exit.store(true, Ordering::Relaxed);
|
||||
tvu.join().unwrap();
|
||||
|
|
|
@ -7,7 +7,7 @@ use crate::cluster_info::ClusterInfo;
|
|||
use crate::leader_schedule_cache::LeaderScheduleCache;
|
||||
use crate::leader_schedule_utils::slot_leader_at;
|
||||
use crate::packet::{Blob, SharedBlob, BLOB_HEADER_SIZE};
|
||||
use crate::repair_service::{RepairService, RepairSlotRange};
|
||||
use crate::repair_service::{RepairService, RepairStrategy};
|
||||
use crate::result::{Error, Result};
|
||||
use crate::service::Service;
|
||||
use crate::streamer::{BlobReceiver, BlobSender};
|
||||
|
@ -175,7 +175,6 @@ pub struct WindowService {
|
|||
impl WindowService {
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub fn new(
|
||||
bank_forks: Option<Arc<RwLock<BankForks>>>,
|
||||
leader_schedule_cache: Option<Arc<LeaderScheduleCache>>,
|
||||
blocktree: Arc<Blocktree>,
|
||||
cluster_info: Arc<RwLock<ClusterInfo>>,
|
||||
|
@ -183,16 +182,21 @@ impl WindowService {
|
|||
retransmit: BlobSender,
|
||||
repair_socket: Arc<UdpSocket>,
|
||||
exit: &Arc<AtomicBool>,
|
||||
repair_slot_range: Option<RepairSlotRange>,
|
||||
repair_strategy: RepairStrategy,
|
||||
genesis_blockhash: &Hash,
|
||||
) -> WindowService {
|
||||
let bank_forks = match repair_strategy {
|
||||
RepairStrategy::RepairRange(_) => None,
|
||||
|
||||
RepairStrategy::RepairAll { ref bank_forks, .. } => Some(bank_forks.clone()),
|
||||
};
|
||||
|
||||
let repair_service = RepairService::new(
|
||||
blocktree.clone(),
|
||||
exit,
|
||||
repair_socket,
|
||||
cluster_info.clone(),
|
||||
bank_forks.clone(),
|
||||
repair_slot_range,
|
||||
repair_strategy,
|
||||
);
|
||||
let exit = exit.clone();
|
||||
let leader_schedule_cache = leader_schedule_cache.clone();
|
||||
|
@ -207,6 +211,7 @@ impl WindowService {
|
|||
if exit.load(Ordering::Relaxed) {
|
||||
break;
|
||||
}
|
||||
|
||||
if let Err(e) = recv_window(
|
||||
bank_forks.as_ref(),
|
||||
leader_schedule_cache.as_ref(),
|
||||
|
@ -351,15 +356,18 @@ mod test {
|
|||
let t_receiver = blob_receiver(Arc::new(leader_node.sockets.gossip), &exit, s_reader);
|
||||
let (s_retransmit, r_retransmit) = channel();
|
||||
let blocktree_path = get_tmp_ledger_path!();
|
||||
let blocktree = Arc::new(
|
||||
Blocktree::open(&blocktree_path).expect("Expected to be able to open database ledger"),
|
||||
);
|
||||
let (blocktree, _, completed_slots_receiver) = Blocktree::open_with_signal(&blocktree_path)
|
||||
.expect("Expected to be able to open database ledger");
|
||||
let blocktree = Arc::new(blocktree);
|
||||
|
||||
let bank = Bank::new(&create_genesis_block_with_leader(100, &me_id, 10).0);
|
||||
let leader_schedule_cache = Arc::new(LeaderScheduleCache::new_from_bank(&bank));
|
||||
let bank_forks = Some(Arc::new(RwLock::new(BankForks::new(0, bank))));
|
||||
let t_window = WindowService::new(
|
||||
let bank_forks = Arc::new(RwLock::new(BankForks::new(0, bank)));
|
||||
let repair_strategy = RepairStrategy::RepairAll {
|
||||
bank_forks,
|
||||
completed_slots_receiver,
|
||||
};
|
||||
let t_window = WindowService::new(
|
||||
Some(leader_schedule_cache),
|
||||
blocktree,
|
||||
subs,
|
||||
|
@ -367,7 +375,7 @@ mod test {
|
|||
s_retransmit,
|
||||
Arc::new(leader_node.sockets.repair),
|
||||
&exit,
|
||||
None,
|
||||
repair_strategy,
|
||||
&Hash::default(),
|
||||
);
|
||||
let t_responder = {
|
||||
|
@ -430,14 +438,18 @@ mod test {
|
|||
let t_receiver = blob_receiver(Arc::new(leader_node.sockets.gossip), &exit, s_reader);
|
||||
let (s_retransmit, r_retransmit) = channel();
|
||||
let blocktree_path = get_tmp_ledger_path!();
|
||||
let blocktree = Arc::new(
|
||||
Blocktree::open(&blocktree_path).expect("Expected to be able to open database ledger"),
|
||||
);
|
||||
let (blocktree, _, completed_slots_receiver) = Blocktree::open_with_signal(&blocktree_path)
|
||||
.expect("Expected to be able to open database ledger");
|
||||
|
||||
let blocktree = Arc::new(blocktree);
|
||||
let bank = Bank::new(&create_genesis_block_with_leader(100, &me_id, 10).0);
|
||||
let leader_schedule_cache = Arc::new(LeaderScheduleCache::new_from_bank(&bank));
|
||||
let bank_forks = Some(Arc::new(RwLock::new(BankForks::new(0, bank))));
|
||||
let t_window = WindowService::new(
|
||||
let bank_forks = Arc::new(RwLock::new(BankForks::new(0, bank)));
|
||||
let repair_strategy = RepairStrategy::RepairAll {
|
||||
bank_forks,
|
||||
completed_slots_receiver,
|
||||
};
|
||||
let t_window = WindowService::new(
|
||||
Some(leader_schedule_cache),
|
||||
blocktree,
|
||||
subs.clone(),
|
||||
|
@ -445,7 +457,7 @@ mod test {
|
|||
s_retransmit,
|
||||
Arc::new(leader_node.sockets.repair),
|
||||
&exit,
|
||||
None,
|
||||
repair_strategy,
|
||||
&Hash::default(),
|
||||
);
|
||||
let t_responder = {
|
||||
|
|
|
@ -83,8 +83,14 @@ fn test_replay() {
|
|||
|
||||
let tvu_addr = target1.info.tvu;
|
||||
|
||||
let (bank_forks, _bank_forks_info, blocktree, ledger_signal_receiver, leader_schedule_cache) =
|
||||
fullnode::new_banks_from_blocktree(&blocktree_path, None);
|
||||
let (
|
||||
bank_forks,
|
||||
_bank_forks_info,
|
||||
blocktree,
|
||||
ledger_signal_receiver,
|
||||
completed_slots_receiver,
|
||||
leader_schedule_cache,
|
||||
) = fullnode::new_banks_from_blocktree(&blocktree_path, None);
|
||||
let working_bank = bank_forks.working_bank();
|
||||
assert_eq!(
|
||||
working_bank.get_balance(&mint_keypair.pubkey()),
|
||||
|
@ -126,6 +132,7 @@ fn test_replay() {
|
|||
&leader_schedule_cache,
|
||||
&exit,
|
||||
&solana_sdk::hash::Hash::default(),
|
||||
completed_slots_receiver,
|
||||
);
|
||||
|
||||
let mut mint_ref_balance = mint_balance;
|
||||
|
|
Loading…
Reference in New Issue