//! The `tvu` module implements the Transaction Validation Unit, a multi-stage transaction //! validation pipeline in software. use { crate::{ banking_trace::BankingTracer, cache_block_meta_service::CacheBlockMetaSender, cluster_info_vote_listener::{ GossipDuplicateConfirmedSlotsReceiver, GossipVerifiedVoteHashReceiver, VerifiedVoteReceiver, VoteTracker, }, cluster_slots_service::{cluster_slots::ClusterSlots, ClusterSlotsService}, completed_data_sets_service::CompletedDataSetsSender, consensus::{tower_storage::TowerStorage, Tower}, cost_update_service::CostUpdateService, drop_bank_service::DropBankService, ledger_cleanup_service::LedgerCleanupService, repair::{quic_endpoint::LocalRequest, repair_service::RepairInfo}, replay_stage::{ReplayStage, ReplayStageConfig}, rewards_recorder_service::RewardsRecorderSender, shred_fetch_stage::ShredFetchStage, voting_service::VotingService, warm_quic_cache_service::WarmQuicCacheService, window_service::WindowService, }, bytes::Bytes, crossbeam_channel::{unbounded, Receiver, Sender}, solana_client::connection_cache::ConnectionCache, solana_geyser_plugin_manager::block_metadata_notifier_interface::BlockMetadataNotifierLock, solana_gossip::{ cluster_info::ClusterInfo, duplicate_shred_handler::DuplicateShredHandler, duplicate_shred_listener::DuplicateShredListener, }, solana_ledger::{ blockstore::Blockstore, blockstore_processor::TransactionStatusSender, entry_notifier_service::EntryNotifierSender, leader_schedule_cache::LeaderScheduleCache, }, solana_poh::poh_recorder::PohRecorder, solana_rpc::{ max_slots::MaxSlots, optimistically_confirmed_bank_tracker::BankNotificationSenderConfig, rpc_subscriptions::RpcSubscriptions, }, solana_runtime::{ accounts_background_service::AbsRequestSender, bank_forks::BankForks, commitment::BlockCommitmentCache, prioritization_fee_cache::PrioritizationFeeCache, }, solana_sdk::{clock::Slot, pubkey::Pubkey, signature::Keypair}, solana_turbine::retransmit_stage::RetransmitStage, solana_vote::vote_sender_types::ReplayVoteSender, std::{ collections::HashSet, net::{SocketAddr, UdpSocket}, sync::{atomic::AtomicBool, Arc, RwLock}, thread::{self, JoinHandle}, }, tokio::sync::mpsc::Sender as AsyncSender, }; pub struct Tvu { fetch_stage: ShredFetchStage, shred_sigverify: JoinHandle<()>, retransmit_stage: RetransmitStage, window_service: WindowService, cluster_slots_service: ClusterSlotsService, replay_stage: ReplayStage, ledger_cleanup_service: Option, cost_update_service: CostUpdateService, voting_service: VotingService, warm_quic_cache_service: Option, drop_bank_service: DropBankService, duplicate_shred_listener: DuplicateShredListener, } pub struct TvuSockets { pub fetch: Vec, pub repair: UdpSocket, pub retransmit: Vec, pub ancestor_hashes_requests: UdpSocket, } #[derive(Default)] pub struct TvuConfig { pub max_ledger_shreds: Option, pub shred_version: u16, // Validators from which repairs are requested pub repair_validators: Option>, // Validators which should be given priority when serving repairs pub repair_whitelist: Arc>>, pub wait_for_vote_to_start_leader: bool, pub replay_slots_concurrently: bool, } impl Tvu { /// This service receives messages from a leader in the network and processes the transactions /// on the bank state. /// # Arguments /// * `cluster_info` - The cluster_info state. /// * `sockets` - fetch, repair, and retransmit sockets /// * `blockstore` - the ledger itself #[allow(clippy::too_many_arguments)] pub fn new( vote_account: &Pubkey, authorized_voter_keypairs: Arc>>>, bank_forks: &Arc>, cluster_info: &Arc, sockets: TvuSockets, blockstore: Arc, ledger_signal_receiver: Receiver, rpc_subscriptions: &Arc, poh_recorder: &Arc>, tower: Tower, tower_storage: Arc, leader_schedule_cache: &Arc, exit: Arc, block_commitment_cache: Arc>, turbine_disabled: Arc, transaction_status_sender: Option, rewards_recorder_sender: Option, cache_block_meta_sender: Option, entry_notification_sender: Option, vote_tracker: Arc, retransmit_slots_sender: Sender, gossip_verified_vote_hash_receiver: GossipVerifiedVoteHashReceiver, verified_vote_receiver: VerifiedVoteReceiver, replay_vote_sender: ReplayVoteSender, completed_data_sets_sender: CompletedDataSetsSender, bank_notification_sender: Option, gossip_confirmed_slots_receiver: GossipDuplicateConfirmedSlotsReceiver, tvu_config: TvuConfig, max_slots: &Arc, block_metadata_notifier: Option, wait_to_vote_slot: Option, accounts_background_request_sender: AbsRequestSender, log_messages_bytes_limit: Option, connection_cache: &Arc, prioritization_fee_cache: &Arc, banking_tracer: Arc, turbine_quic_endpoint_sender: AsyncSender<(SocketAddr, Bytes)>, turbine_quic_endpoint_receiver: Receiver<(Pubkey, SocketAddr, Bytes)>, repair_quic_endpoint_sender: AsyncSender, ) -> Result { let TvuSockets { repair: repair_socket, fetch: fetch_sockets, retransmit: retransmit_sockets, ancestor_hashes_requests: ancestor_hashes_socket, } = sockets; let (fetch_sender, fetch_receiver) = unbounded(); let repair_socket = Arc::new(repair_socket); let ancestor_hashes_socket = Arc::new(ancestor_hashes_socket); let fetch_sockets: Vec> = fetch_sockets.into_iter().map(Arc::new).collect(); let (repair_quic_endpoint_response_sender, repair_quic_endpoint_response_receiver) = unbounded(); let fetch_stage = ShredFetchStage::new( fetch_sockets, turbine_quic_endpoint_receiver, repair_socket.clone(), repair_quic_endpoint_response_receiver, fetch_sender, tvu_config.shred_version, bank_forks.clone(), cluster_info.clone(), turbine_disabled, exit.clone(), ); let (verified_sender, verified_receiver) = unbounded(); let (retransmit_sender, retransmit_receiver) = unbounded(); let shred_sigverify = solana_turbine::sigverify_shreds::spawn_shred_sigverify( cluster_info.clone(), bank_forks.clone(), leader_schedule_cache.clone(), fetch_receiver, retransmit_sender.clone(), verified_sender, ); let retransmit_stage = RetransmitStage::new( bank_forks.clone(), leader_schedule_cache.clone(), cluster_info.clone(), Arc::new(retransmit_sockets), turbine_quic_endpoint_sender, retransmit_receiver, max_slots.clone(), Some(rpc_subscriptions.clone()), ); let cluster_slots = Arc::new(ClusterSlots::default()); let (ancestor_duplicate_slots_sender, ancestor_duplicate_slots_receiver) = unbounded(); let (duplicate_slots_sender, duplicate_slots_receiver) = unbounded(); let (ancestor_hashes_replay_update_sender, ancestor_hashes_replay_update_receiver) = unbounded(); let (dumped_slots_sender, dumped_slots_receiver) = unbounded(); let (popular_pruned_forks_sender, popular_pruned_forks_receiver) = unbounded(); let window_service = { let epoch_schedule = *bank_forks.read().unwrap().working_bank().epoch_schedule(); let repair_info = RepairInfo { bank_forks: bank_forks.clone(), epoch_schedule, ancestor_duplicate_slots_sender, repair_validators: tvu_config.repair_validators, repair_whitelist: tvu_config.repair_whitelist, cluster_info: cluster_info.clone(), cluster_slots: cluster_slots.clone(), }; WindowService::new( blockstore.clone(), verified_receiver, retransmit_sender, repair_socket, ancestor_hashes_socket, repair_quic_endpoint_sender, repair_quic_endpoint_response_sender, exit.clone(), repair_info, leader_schedule_cache.clone(), verified_vote_receiver, completed_data_sets_sender, duplicate_slots_sender, ancestor_hashes_replay_update_receiver, dumped_slots_receiver, popular_pruned_forks_sender, ) }; let (cluster_slots_update_sender, cluster_slots_update_receiver) = unbounded(); let cluster_slots_service = ClusterSlotsService::new( blockstore.clone(), cluster_slots.clone(), bank_forks.clone(), cluster_info.clone(), cluster_slots_update_receiver, exit.clone(), ); let (ledger_cleanup_slot_sender, ledger_cleanup_slot_receiver) = unbounded(); let replay_stage_config = ReplayStageConfig { vote_account: *vote_account, authorized_voter_keypairs, exit: exit.clone(), rpc_subscriptions: rpc_subscriptions.clone(), leader_schedule_cache: leader_schedule_cache.clone(), latest_root_senders: vec![ledger_cleanup_slot_sender], accounts_background_request_sender, block_commitment_cache, transaction_status_sender, rewards_recorder_sender, cache_block_meta_sender, entry_notification_sender, bank_notification_sender, wait_for_vote_to_start_leader: tvu_config.wait_for_vote_to_start_leader, ancestor_hashes_replay_update_sender, tower_storage: tower_storage.clone(), wait_to_vote_slot, replay_slots_concurrently: tvu_config.replay_slots_concurrently, }; let (voting_sender, voting_receiver) = unbounded(); let voting_service = VotingService::new( voting_receiver, cluster_info.clone(), poh_recorder.clone(), tower_storage, ); let warm_quic_cache_service = if connection_cache.use_quic() { Some(WarmQuicCacheService::new( connection_cache.clone(), cluster_info.clone(), poh_recorder.clone(), exit.clone(), )) } else { None }; let (cost_update_sender, cost_update_receiver) = unbounded(); let cost_update_service = CostUpdateService::new(blockstore.clone(), cost_update_receiver); let (drop_bank_sender, drop_bank_receiver) = unbounded(); let drop_bank_service = DropBankService::new(drop_bank_receiver); let replay_stage = ReplayStage::new( replay_stage_config, blockstore.clone(), bank_forks.clone(), cluster_info.clone(), ledger_signal_receiver, duplicate_slots_receiver, poh_recorder.clone(), tower, vote_tracker, cluster_slots, retransmit_slots_sender, ancestor_duplicate_slots_receiver, replay_vote_sender, gossip_confirmed_slots_receiver, gossip_verified_vote_hash_receiver, cluster_slots_update_sender, cost_update_sender, voting_sender, drop_bank_sender, block_metadata_notifier, log_messages_bytes_limit, prioritization_fee_cache.clone(), dumped_slots_sender, banking_tracer, popular_pruned_forks_receiver, )?; let ledger_cleanup_service = tvu_config.max_ledger_shreds.map(|max_ledger_shreds| { LedgerCleanupService::new( ledger_cleanup_slot_receiver, blockstore.clone(), max_ledger_shreds, exit.clone(), ) }); let duplicate_shred_listener = DuplicateShredListener::new( exit, cluster_info.clone(), DuplicateShredHandler::new( blockstore, leader_schedule_cache.clone(), bank_forks.clone(), ), ); Ok(Tvu { fetch_stage, shred_sigverify, retransmit_stage, window_service, cluster_slots_service, replay_stage, ledger_cleanup_service, cost_update_service, voting_service, warm_quic_cache_service, drop_bank_service, duplicate_shred_listener, }) } pub fn join(self) -> thread::Result<()> { self.retransmit_stage.join()?; self.window_service.join()?; self.cluster_slots_service.join()?; self.fetch_stage.join()?; self.shred_sigverify.join()?; if self.ledger_cleanup_service.is_some() { self.ledger_cleanup_service.unwrap().join()?; } self.replay_stage.join()?; self.cost_update_service.join()?; self.voting_service.join()?; if let Some(warmup_service) = self.warm_quic_cache_service { warmup_service.join()?; } self.drop_bank_service.join()?; self.duplicate_shred_listener.join()?; Ok(()) } } #[cfg(test)] pub mod tests { use { super::*, crate::consensus::tower_storage::FileTowerStorage, serial_test::serial, solana_gossip::cluster_info::{ClusterInfo, Node}, solana_ledger::{ blockstore::BlockstoreSignals, blockstore_options::BlockstoreOptions, create_new_tmp_ledger, genesis_utils::{create_genesis_config, GenesisConfigInfo}, }, solana_poh::poh_recorder::create_test_recorder, solana_rpc::optimistically_confirmed_bank_tracker::OptimisticallyConfirmedBank, solana_runtime::bank::Bank, solana_sdk::signature::{Keypair, Signer}, solana_streamer::socket::SocketAddrSpace, std::sync::atomic::{AtomicU64, Ordering}, }; #[ignore] #[test] #[serial] fn test_tvu_exit() { solana_logger::setup(); let leader = Node::new_localhost(); let target1_keypair = Keypair::new(); let target1 = Node::new_localhost_with_pubkey(&target1_keypair.pubkey()); let starting_balance = 10_000; let GenesisConfigInfo { genesis_config, .. } = create_genesis_config(starting_balance); let bank_forks = BankForks::new_rw_arc(Bank::new_for_tests(&genesis_config)); let keypair = Arc::new(Keypair::new()); let (turbine_quic_endpoint_sender, _turbine_quic_endpoint_receiver) = tokio::sync::mpsc::channel(/*capacity:*/ 128); let (_turbine_quic_endpoint_sender, turbine_quic_endpoint_receiver) = unbounded(); let (repair_quic_endpoint_sender, _repair_quic_endpoint_receiver) = tokio::sync::mpsc::channel(/*buffer:*/ 128); //start cluster_info1 let cluster_info1 = ClusterInfo::new(target1.info.clone(), keypair, SocketAddrSpace::Unspecified); cluster_info1.insert_info(leader.info); let cref1 = Arc::new(cluster_info1); let (blockstore_path, _) = create_new_tmp_ledger!(&genesis_config); let BlockstoreSignals { blockstore, ledger_signal_receiver, .. } = Blockstore::open_with_signal(&blockstore_path, BlockstoreOptions::default()) .expect("Expected to successfully open ledger"); let blockstore = Arc::new(blockstore); let bank = bank_forks.read().unwrap().working_bank(); let (exit, poh_recorder, poh_service, _entry_receiver) = create_test_recorder(bank.clone(), blockstore.clone(), None, None); let vote_keypair = Keypair::new(); let leader_schedule_cache = Arc::new(LeaderScheduleCache::new_from_bank(&bank)); let block_commitment_cache = Arc::new(RwLock::new(BlockCommitmentCache::default())); let (retransmit_slots_sender, _retransmit_slots_receiver) = unbounded(); let (_gossip_verified_vote_hash_sender, gossip_verified_vote_hash_receiver) = unbounded(); let (_verified_vote_sender, verified_vote_receiver) = unbounded(); let (replay_vote_sender, _replay_vote_receiver) = unbounded(); let (completed_data_sets_sender, _completed_data_sets_receiver) = unbounded(); let (_, gossip_confirmed_slots_receiver) = unbounded(); let max_complete_transaction_status_slot = Arc::new(AtomicU64::default()); let max_complete_rewards_slot = Arc::new(AtomicU64::default()); let ignored_prioritization_fee_cache = Arc::new(PrioritizationFeeCache::new(0u64)); let tvu = Tvu::new( &vote_keypair.pubkey(), Arc::new(RwLock::new(vec![Arc::new(vote_keypair)])), &bank_forks, &cref1, { TvuSockets { repair: target1.sockets.repair, retransmit: target1.sockets.retransmit_sockets, fetch: target1.sockets.tvu, ancestor_hashes_requests: target1.sockets.ancestor_hashes_requests, } }, blockstore, ledger_signal_receiver, &Arc::new(RpcSubscriptions::new_for_tests( exit.clone(), max_complete_transaction_status_slot, max_complete_rewards_slot, bank_forks.clone(), block_commitment_cache.clone(), OptimisticallyConfirmedBank::locked_from_bank_forks_root(&bank_forks), )), &poh_recorder, Tower::default(), Arc::new(FileTowerStorage::default()), &leader_schedule_cache, exit.clone(), block_commitment_cache, Arc::::default(), None, None, None, None, Arc::::default(), retransmit_slots_sender, gossip_verified_vote_hash_receiver, verified_vote_receiver, replay_vote_sender, completed_data_sets_sender, None, gossip_confirmed_slots_receiver, TvuConfig::default(), &Arc::new(MaxSlots::default()), None, None, AbsRequestSender::default(), None, &Arc::new(ConnectionCache::new("connection_cache_test")), &ignored_prioritization_fee_cache, BankingTracer::new_disabled(), turbine_quic_endpoint_sender, turbine_quic_endpoint_receiver, repair_quic_endpoint_sender, ) .expect("assume success"); exit.store(true, Ordering::Relaxed); tvu.join().unwrap(); poh_service.join().unwrap(); } }