solana/core/src/tvu.rs

242 lines
8.1 KiB
Rust
Raw Normal View History

//! The `tvu` module implements the Transaction Validation Unit, a
//! multi-stage transaction validation pipeline in software.
2018-06-14 17:05:12 -07:00
//!
//! 1. BlobFetchStage
//! - Incoming blobs are picked up from the TVU sockets and repair socket.
//! 2. RetransmitStage
2018-06-15 14:49:22 -07:00
//! - Blobs are windowed until a contiguous chunk is available. This stage also repairs and
//! retransmits blobs that are in the queue.
//! 3. ReplayStage
2018-06-15 14:49:22 -07:00
//! - Transactions in blobs are processed and applied to the bank.
//! - TODO We need to verify the signatures in the blobs.
//! 4. StorageStage
//! - Generating the keys used to encrypt the ledger and sample it for storage mining.
2019-02-21 11:19:45 -08:00
use crate::bank_forks::BankForks;
2018-12-07 19:16:27 -08:00
use crate::blob_fetch_stage::BlobFetchStage;
2019-02-21 15:16:09 -08:00
use crate::blockstream_service::BlockstreamService;
2019-02-07 20:52:39 -08:00
use crate::blocktree::Blocktree;
2019-02-21 11:37:48 -08:00
use crate::blocktree_processor::BankForksInfo;
2018-12-07 19:16:27 -08:00
use crate::cluster_info::ClusterInfo;
use crate::entry::{EntryReceiver, EntrySender};
use crate::poh_recorder::PohRecorder;
use crate::replay_stage::ReplayStage;
2018-12-07 19:16:27 -08:00
use crate::retransmit_stage::RetransmitStage;
2019-02-18 18:08:54 -08:00
use crate::rpc_subscriptions::RpcSubscriptions;
2018-12-07 19:16:27 -08:00
use crate::service::Service;
use crate::storage_stage::{StorageStage, StorageState};
use solana_sdk::pubkey::Pubkey;
2019-01-30 19:28:48 -08:00
use solana_sdk::signature::{Keypair, KeypairUtil};
use std::net::UdpSocket;
use std::sync::atomic::AtomicBool;
use std::sync::mpsc::{channel, Receiver};
use std::sync::{Arc, Mutex, RwLock};
use std::thread;
pub struct Tvu {
fetch_stage: BlobFetchStage,
retransmit_stage: RetransmitStage,
replay_stage: ReplayStage,
2019-02-21 15:16:09 -08:00
blockstream_service: Option<BlockstreamService>,
storage_stage: StorageStage,
}
pub struct Sockets {
pub fetch: Vec<UdpSocket>,
pub repair: UdpSocket,
pub retransmit: UdpSocket,
}
impl Tvu {
/// This service receives messages from a leader in the network and processes the transactions
2018-05-14 14:33:11 -07:00
/// on the bank state.
/// # Arguments
2018-10-08 19:55:54 -07:00
/// * `cluster_info` - The cluster_info state.
2019-02-21 11:37:48 -08:00
/// * `sockets` - fetch, repair, and retransmit sockets
2019-02-07 20:52:39 -08:00
/// * `blocktree` - the ledger itself
#[allow(clippy::new_ret_no_self, clippy::too_many_arguments)]
2019-02-21 21:43:35 -08:00
pub fn new<T>(
vote_account: &Pubkey,
2019-02-21 21:43:35 -08:00
voting_keypair: Option<Arc<T>>,
2019-02-21 11:19:45 -08:00
bank_forks: &Arc<RwLock<BankForks>>,
2019-02-21 11:37:48 -08:00
bank_forks_info: &[BankForksInfo],
cluster_info: &Arc<RwLock<ClusterInfo>>,
sockets: Sockets,
2019-02-07 20:52:39 -08:00
blocktree: Arc<Blocktree>,
storage_rotate_count: u64,
storage_state: &StorageState,
2019-02-21 15:16:09 -08:00
blockstream: Option<&String>,
ledger_signal_receiver: Receiver<bool>,
2019-02-18 18:08:54 -08:00
subscriptions: &Arc<RpcSubscriptions>,
poh_recorder: &Arc<Mutex<PohRecorder>>,
storage_entry_sender: EntrySender,
storage_entry_receiver: EntryReceiver,
exit: &Arc<AtomicBool>,
2019-02-21 21:43:35 -08:00
) -> Self
where
T: 'static + KeypairUtil + Sync + Send,
{
let keypair: Arc<Keypair> = cluster_info
.read()
.expect("Unable to read from cluster_info during Tvu creation")
.keypair
.clone();
let Sockets {
repair: repair_socket,
fetch: fetch_sockets,
retransmit: retransmit_socket,
} = sockets;
let (blob_fetch_sender, blob_fetch_receiver) = channel();
let repair_socket = Arc::new(repair_socket);
let mut blob_sockets: Vec<Arc<UdpSocket>> =
fetch_sockets.into_iter().map(Arc::new).collect();
blob_sockets.push(repair_socket.clone());
2019-03-04 20:50:02 -08:00
let fetch_stage = BlobFetchStage::new_multi_socket(blob_sockets, &blob_fetch_sender, &exit);
//TODO
//the packets coming out of blob_receiver need to be sent to the GPU and verified
//then sent to the window, which does the erasure coding reconstruction
let retransmit_stage = RetransmitStage::new(
2019-02-21 11:19:45 -08:00
&bank_forks,
2019-02-07 20:52:39 -08:00
blocktree.clone(),
2018-10-08 19:55:54 -07:00
&cluster_info,
Arc::new(retransmit_socket),
repair_socket,
blob_fetch_receiver,
2019-03-04 20:50:02 -08:00
&exit,
2018-05-22 15:17:59 -07:00
);
let (replay_stage, slot_full_receiver) = ReplayStage::new(
&keypair.pubkey(),
vote_account,
voting_keypair,
2019-02-07 20:52:39 -08:00
blocktree.clone(),
2019-02-21 11:37:48 -08:00
&bank_forks,
cluster_info.clone(),
2019-03-04 20:50:02 -08:00
&exit,
ledger_signal_receiver,
2019-02-18 18:08:54 -08:00
subscriptions,
poh_recorder,
storage_entry_sender,
);
2019-02-21 15:16:09 -08:00
let blockstream_service = if blockstream.is_some() {
let blockstream_service = BlockstreamService::new(
slot_full_receiver,
blocktree.clone(),
2019-02-21 15:16:09 -08:00
blockstream.unwrap().to_string(),
2019-03-04 20:50:02 -08:00
&exit,
2019-02-11 13:51:14 -08:00
);
2019-02-21 15:16:09 -08:00
Some(blockstream_service)
2019-02-11 13:51:14 -08:00
} else {
None
};
let storage_stage = StorageStage::new(
storage_state,
storage_entry_receiver,
2019-02-07 20:52:39 -08:00
Some(blocktree),
2019-01-17 14:41:48 -08:00
&keypair,
2019-03-04 20:50:02 -08:00
&exit,
2019-02-21 11:37:48 -08:00
bank_forks_info[0].entry_height, // TODO: StorageStage needs to deal with BankForks somehow still
storage_rotate_count,
2019-01-17 14:41:48 -08:00
&cluster_info,
);
Tvu {
fetch_stage,
retransmit_stage,
replay_stage,
2019-02-21 15:16:09 -08:00
blockstream_service,
storage_stage,
}
}
}
impl Service for Tvu {
2019-02-08 08:06:27 -08:00
type JoinReturnType = ();
2019-02-08 08:06:27 -08:00
fn join(self) -> thread::Result<()> {
Leader scheduler plumbing (#1440) * Added LeaderScheduler module and tests * plumbing for LeaderScheduler in Fullnode + tests. Add vote processing for active set to ReplicateStage and WriteStage * Add LeaderScheduler plumbing for Tvu, window, and tests * Fix bank and switch tests to use new LeaderScheduler * move leader rotation check from window service to replicate stage * Add replicate_stage leader rotation exit test * removed leader scheduler from the window service and associated modules/tests * Corrected is_leader calculation in repair() function in window.rs * Integrate LeaderScheduler with write_stage for leader to validator transitions * Integrated LeaderScheduler with BroadcastStage * Removed gossip leader rotation from crdt * Add multi validator, leader test * Comments and cleanup * Remove unneeded checks from broadcast stage * Fix case where a validator/leader need to immediately transition on startup after reading ledger and seeing they are not in the correct role * Set new leader in validator -> validator transitions * Clean up for PR comments, refactor LeaderScheduler from process_entry/process_ledger_tail * Cleaned out LeaderScheduler options, implemented LeaderScheduler strategy that only picks the bootstrap leader to support existing tests, drone/airdrops * Ignore test_full_leader_validator_network test due to bug where the next leader in line fails to get the last entry before rotation (b/c it hasn't started up yet). Added a test test_dropped_handoff_recovery go track this bug
2018-10-10 16:49:41 -07:00
self.retransmit_stage.join()?;
self.fetch_stage.join()?;
self.storage_stage.join()?;
2019-02-21 15:16:09 -08:00
if self.blockstream_service.is_some() {
self.blockstream_service.unwrap().join()?;
2019-02-11 13:51:14 -08:00
}
2019-02-08 08:06:27 -08:00
self.replay_stage.join()?;
Ok(())
}
}
#[cfg(test)]
2018-05-23 10:49:48 -07:00
pub mod tests {
2019-02-08 08:10:28 -08:00
use super::*;
use crate::banking_stage::create_test_recorder;
use crate::blocktree::get_tmp_ledger_path;
2018-12-07 19:16:27 -08:00
use crate::cluster_info::{ClusterInfo, Node};
2019-02-08 08:10:28 -08:00
use crate::storage_stage::STORAGE_ROTATE_TEST_COUNT;
2019-02-21 11:19:45 -08:00
use solana_runtime::bank::Bank;
use solana_sdk::genesis_block::GenesisBlock;
use std::sync::atomic::Ordering;
#[test]
fn test_tvu_exit() {
solana_logger::setup();
let leader = Node::new_localhost();
let target1_keypair = Keypair::new();
let target1 = Node::new_localhost_with_pubkey(&target1_keypair.pubkey());
let starting_balance = 10_000;
let (genesis_block, _mint_keypair) = GenesisBlock::new(starting_balance);
2019-02-21 11:19:45 -08:00
let bank_forks = BankForks::new(0, Bank::new(&genesis_block));
2019-02-21 11:37:48 -08:00
let bank_forks_info = vec![BankForksInfo {
2019-03-04 16:40:28 -08:00
bank_slot: 0,
2019-02-21 11:37:48 -08:00
entry_height: 0,
}];
//start cluster_info1
let mut cluster_info1 = ClusterInfo::new_with_invalid_keypair(target1.info.clone());
cluster_info1.insert_info(leader.info.clone());
let cref1 = Arc::new(RwLock::new(cluster_info1));
let blocktree_path = get_tmp_ledger_path!();
let (blocktree, l_receiver) = Blocktree::open_with_signal(&blocktree_path)
.expect("Expected to successfully open ledger");
let bank = bank_forks.working_bank();
2019-03-04 20:50:02 -08:00
let (exit, poh_recorder, poh_service, _entry_receiver) = create_test_recorder(&bank);
let voting_keypair = Keypair::new();
let (storage_entry_sender, storage_entry_receiver) = channel();
let tvu = Tvu::new(
&voting_keypair.pubkey(),
Some(Arc::new(voting_keypair)),
2019-02-21 11:19:45 -08:00
&Arc::new(RwLock::new(bank_forks)),
2019-02-21 11:37:48 -08:00
&bank_forks_info,
&cref1,
{
Sockets {
repair: target1.sockets.repair,
retransmit: target1.sockets.retransmit,
fetch: target1.sockets.tvu,
}
},
2019-02-07 20:52:39 -08:00
Arc::new(blocktree),
STORAGE_ROTATE_TEST_COUNT,
&StorageState::default(),
None,
l_receiver,
2019-02-18 18:08:54 -08:00
&Arc::new(RpcSubscriptions::default()),
&poh_recorder,
storage_entry_sender,
storage_entry_receiver,
&exit,
);
exit.store(true, Ordering::Relaxed);
tvu.join().unwrap();
2019-03-04 20:50:02 -08:00
poh_service.join().unwrap();
}
}