2023-02-15 08:52:13 -08:00
|
|
|
use {
|
|
|
|
super::{
|
|
|
|
committer::{CommitTransactionDetails, Committer},
|
|
|
|
BankingStageStats,
|
|
|
|
},
|
|
|
|
crate::{
|
|
|
|
banking_stage::committer::PreBalanceInfo,
|
|
|
|
immutable_deserialized_packet::ImmutableDeserializedPacket,
|
|
|
|
leader_slot_banking_stage_metrics::{LeaderSlotMetricsTracker, ProcessTransactionsSummary},
|
|
|
|
leader_slot_banking_stage_timing_metrics::LeaderExecuteAndCommitTimings,
|
|
|
|
qos_service::QosService,
|
|
|
|
unprocessed_transaction_storage::{ConsumeScannerPayload, UnprocessedTransactionStorage},
|
|
|
|
},
|
|
|
|
itertools::Itertools,
|
|
|
|
solana_ledger::token_balances::collect_token_balances,
|
2023-02-15 17:20:55 -08:00
|
|
|
solana_measure::{measure::Measure, measure_us},
|
2023-02-15 08:52:13 -08:00
|
|
|
solana_poh::poh_recorder::{
|
|
|
|
BankStart, PohRecorderError, RecordTransactionsSummary, RecordTransactionsTimings,
|
|
|
|
TransactionRecorder,
|
|
|
|
},
|
|
|
|
solana_program_runtime::timings::ExecuteTimings,
|
|
|
|
solana_runtime::{
|
|
|
|
bank::{Bank, LoadAndExecuteTransactionsOutput, TransactionCheckResult},
|
|
|
|
transaction_batch::TransactionBatch,
|
|
|
|
transaction_error_metrics::TransactionErrorMetrics,
|
|
|
|
},
|
|
|
|
solana_sdk::{
|
|
|
|
clock::{FORWARD_TRANSACTIONS_TO_LEADER_AT_SLOT_OFFSET, MAX_PROCESSING_AGE},
|
2023-05-17 09:24:59 -07:00
|
|
|
feature_set, saturating_add_assign,
|
2023-02-15 08:52:13 -08:00
|
|
|
timing::timestamp,
|
|
|
|
transaction::{self, SanitizedTransaction, TransactionError},
|
|
|
|
},
|
|
|
|
std::{
|
|
|
|
sync::{atomic::Ordering, Arc},
|
|
|
|
time::Instant,
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
pub const MAX_NUM_TRANSACTIONS_PER_BATCH: usize = 64;
|
|
|
|
|
|
|
|
pub struct ProcessTransactionBatchOutput {
|
|
|
|
// The number of transactions filtered out by the cost model
|
|
|
|
cost_model_throttled_transactions_count: usize,
|
|
|
|
// Amount of time spent running the cost model
|
|
|
|
cost_model_us: u64,
|
2023-05-15 08:56:57 -07:00
|
|
|
pub execute_and_commit_transactions_output: ExecuteAndCommitTransactionsOutput,
|
2023-02-15 08:52:13 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
pub struct ExecuteAndCommitTransactionsOutput {
|
|
|
|
// Total number of transactions that were passed as candidates for execution
|
|
|
|
transactions_attempted_execution_count: usize,
|
|
|
|
// The number of transactions of that were executed. See description of in `ProcessTransactionsSummary`
|
|
|
|
// for possible outcomes of execution.
|
|
|
|
executed_transactions_count: usize,
|
|
|
|
// Total number of the executed transactions that returned success/not
|
|
|
|
// an error.
|
|
|
|
executed_with_successful_result_count: usize,
|
|
|
|
// Transactions that either were not executed, or were executed and failed to be committed due
|
|
|
|
// to the block ending.
|
|
|
|
retryable_transaction_indexes: Vec<usize>,
|
|
|
|
// A result that indicates whether transactions were successfully
|
|
|
|
// committed into the Poh stream.
|
2023-05-15 08:56:57 -07:00
|
|
|
pub commit_transactions_result: Result<Vec<CommitTransactionDetails>, PohRecorderError>,
|
2023-02-15 08:52:13 -08:00
|
|
|
execute_and_commit_timings: LeaderExecuteAndCommitTimings,
|
|
|
|
error_counters: TransactionErrorMetrics,
|
|
|
|
}
|
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
pub struct Consumer {
|
|
|
|
committer: Committer,
|
|
|
|
transaction_recorder: TransactionRecorder,
|
|
|
|
qos_service: QosService,
|
|
|
|
log_messages_bytes_limit: Option<usize>,
|
|
|
|
}
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
impl Consumer {
|
2023-03-06 09:13:28 -08:00
|
|
|
pub fn new(
|
|
|
|
committer: Committer,
|
|
|
|
transaction_recorder: TransactionRecorder,
|
|
|
|
qos_service: QosService,
|
|
|
|
log_messages_bytes_limit: Option<usize>,
|
|
|
|
) -> Self {
|
|
|
|
Self {
|
|
|
|
committer,
|
|
|
|
transaction_recorder,
|
|
|
|
qos_service,
|
|
|
|
log_messages_bytes_limit,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2023-02-15 08:52:13 -08:00
|
|
|
pub fn consume_buffered_packets(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank_start: &BankStart,
|
|
|
|
unprocessed_transaction_storage: &mut UnprocessedTransactionStorage,
|
|
|
|
banking_stage_stats: &BankingStageStats,
|
|
|
|
slot_metrics_tracker: &mut LeaderSlotMetricsTracker,
|
|
|
|
) {
|
|
|
|
let mut rebuffered_packet_count = 0;
|
|
|
|
let mut consumed_buffered_packets_count = 0;
|
|
|
|
let mut proc_start = Measure::start("consume_buffered_process");
|
|
|
|
let num_packets_to_process = unprocessed_transaction_storage.len();
|
|
|
|
|
|
|
|
let reached_end_of_slot = unprocessed_transaction_storage.process_packets(
|
|
|
|
bank_start.working_bank.clone(),
|
|
|
|
banking_stage_stats,
|
|
|
|
slot_metrics_tracker,
|
|
|
|
|packets_to_process, payload| {
|
2023-03-06 09:13:28 -08:00
|
|
|
self.do_process_packets(
|
2023-02-15 08:52:13 -08:00
|
|
|
bank_start,
|
|
|
|
payload,
|
|
|
|
banking_stage_stats,
|
|
|
|
&mut consumed_buffered_packets_count,
|
|
|
|
&mut rebuffered_packet_count,
|
|
|
|
packets_to_process,
|
|
|
|
)
|
|
|
|
},
|
|
|
|
);
|
|
|
|
|
|
|
|
if reached_end_of_slot {
|
|
|
|
slot_metrics_tracker.set_end_of_slot_unprocessed_buffer_len(
|
|
|
|
unprocessed_transaction_storage.len() as u64,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
proc_start.stop();
|
|
|
|
debug!(
|
|
|
|
"@{:?} done processing buffered batches: {} time: {:?}ms tx count: {} tx/s: {}",
|
|
|
|
timestamp(),
|
|
|
|
num_packets_to_process,
|
|
|
|
proc_start.as_ms(),
|
|
|
|
consumed_buffered_packets_count,
|
|
|
|
(consumed_buffered_packets_count as f32) / (proc_start.as_s())
|
|
|
|
);
|
|
|
|
|
|
|
|
banking_stage_stats
|
|
|
|
.consume_buffered_packets_elapsed
|
|
|
|
.fetch_add(proc_start.as_us(), Ordering::Relaxed);
|
|
|
|
banking_stage_stats
|
|
|
|
.rebuffered_packets_count
|
|
|
|
.fetch_add(rebuffered_packet_count, Ordering::Relaxed);
|
|
|
|
banking_stage_stats
|
|
|
|
.consumed_buffered_packets_count
|
|
|
|
.fetch_add(consumed_buffered_packets_count, Ordering::Relaxed);
|
|
|
|
}
|
|
|
|
|
|
|
|
fn do_process_packets(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank_start: &BankStart,
|
|
|
|
payload: &mut ConsumeScannerPayload,
|
|
|
|
banking_stage_stats: &BankingStageStats,
|
|
|
|
consumed_buffered_packets_count: &mut usize,
|
|
|
|
rebuffered_packet_count: &mut usize,
|
|
|
|
packets_to_process: &Vec<Arc<ImmutableDeserializedPacket>>,
|
|
|
|
) -> Option<Vec<usize>> {
|
|
|
|
if payload.reached_end_of_slot {
|
|
|
|
return None;
|
|
|
|
}
|
|
|
|
|
|
|
|
let packets_to_process_len = packets_to_process.len();
|
2023-03-06 09:13:28 -08:00
|
|
|
let (process_transactions_summary, process_packets_transactions_us) = measure_us!(self
|
|
|
|
.process_packets_transactions(
|
2023-02-15 08:52:13 -08:00
|
|
|
&bank_start.working_bank,
|
|
|
|
&bank_start.bank_creation_time,
|
|
|
|
&payload.sanitized_transactions,
|
|
|
|
banking_stage_stats,
|
|
|
|
payload.slot_metrics_tracker,
|
2023-02-15 17:20:55 -08:00
|
|
|
));
|
2023-02-15 08:52:13 -08:00
|
|
|
payload
|
|
|
|
.slot_metrics_tracker
|
2023-02-15 17:20:55 -08:00
|
|
|
.increment_process_packets_transactions_us(process_packets_transactions_us);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// Clear payload for next iteration
|
|
|
|
payload.sanitized_transactions.clear();
|
|
|
|
payload.account_locks.clear();
|
|
|
|
|
|
|
|
let ProcessTransactionsSummary {
|
|
|
|
reached_max_poh_height,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = process_transactions_summary;
|
|
|
|
|
|
|
|
if reached_max_poh_height || !bank_start.should_working_bank_still_be_processing_txs() {
|
|
|
|
payload.reached_end_of_slot = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
// The difference between all transactions passed to execution and the ones that
|
|
|
|
// are retryable were the ones that were either:
|
|
|
|
// 1) Committed into the block
|
|
|
|
// 2) Dropped without being committed because they had some fatal error (too old,
|
|
|
|
// duplicate signature, etc.)
|
|
|
|
//
|
|
|
|
// Note: This assumes that every packet deserializes into one transaction!
|
|
|
|
*consumed_buffered_packets_count +=
|
|
|
|
packets_to_process_len.saturating_sub(retryable_transaction_indexes.len());
|
|
|
|
|
|
|
|
// Out of the buffered packets just retried, collect any still unprocessed
|
|
|
|
// transactions in this batch for forwarding
|
|
|
|
*rebuffered_packet_count += retryable_transaction_indexes.len();
|
|
|
|
|
|
|
|
payload
|
|
|
|
.slot_metrics_tracker
|
|
|
|
.increment_retryable_packets_count(retryable_transaction_indexes.len() as u64);
|
|
|
|
|
|
|
|
Some(retryable_transaction_indexes)
|
|
|
|
}
|
|
|
|
|
|
|
|
fn process_packets_transactions(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank: &Arc<Bank>,
|
|
|
|
bank_creation_time: &Instant,
|
|
|
|
sanitized_transactions: &[SanitizedTransaction],
|
|
|
|
banking_stage_stats: &BankingStageStats,
|
|
|
|
slot_metrics_tracker: &mut LeaderSlotMetricsTracker,
|
|
|
|
) -> ProcessTransactionsSummary {
|
2023-03-06 09:13:28 -08:00
|
|
|
let (mut process_transactions_summary, process_transactions_us) = measure_us!(
|
|
|
|
self.process_transactions(bank, bank_creation_time, sanitized_transactions)
|
|
|
|
);
|
2023-02-15 08:52:13 -08:00
|
|
|
slot_metrics_tracker.increment_process_transactions_us(process_transactions_us);
|
|
|
|
banking_stage_stats
|
|
|
|
.transaction_processing_elapsed
|
|
|
|
.fetch_add(process_transactions_us, Ordering::Relaxed);
|
|
|
|
|
|
|
|
let ProcessTransactionsSummary {
|
|
|
|
ref retryable_transaction_indexes,
|
|
|
|
ref error_counters,
|
|
|
|
..
|
|
|
|
} = process_transactions_summary;
|
|
|
|
|
|
|
|
slot_metrics_tracker.accumulate_process_transactions_summary(&process_transactions_summary);
|
|
|
|
slot_metrics_tracker.accumulate_transaction_errors(error_counters);
|
|
|
|
|
|
|
|
let retryable_tx_count = retryable_transaction_indexes.len();
|
|
|
|
inc_new_counter_info!("banking_stage-unprocessed_transactions", retryable_tx_count);
|
|
|
|
|
|
|
|
// Filter out the retryable transactions that are too old
|
2023-02-15 17:20:55 -08:00
|
|
|
let (filtered_retryable_transaction_indexes, filter_retryable_packets_us) =
|
|
|
|
measure_us!(Self::filter_pending_packets_from_pending_txs(
|
2023-02-15 08:52:13 -08:00
|
|
|
bank,
|
|
|
|
sanitized_transactions,
|
|
|
|
retryable_transaction_indexes,
|
2023-02-15 17:20:55 -08:00
|
|
|
));
|
2023-02-15 08:52:13 -08:00
|
|
|
slot_metrics_tracker.increment_filter_retryable_packets_us(filter_retryable_packets_us);
|
|
|
|
banking_stage_stats
|
|
|
|
.filter_pending_packets_elapsed
|
|
|
|
.fetch_add(filter_retryable_packets_us, Ordering::Relaxed);
|
|
|
|
|
|
|
|
let retryable_packets_filtered_count = retryable_transaction_indexes
|
|
|
|
.len()
|
|
|
|
.saturating_sub(filtered_retryable_transaction_indexes.len());
|
|
|
|
slot_metrics_tracker
|
|
|
|
.increment_retryable_packets_filtered_count(retryable_packets_filtered_count as u64);
|
|
|
|
|
2023-05-10 13:44:42 -07:00
|
|
|
banking_stage_stats
|
|
|
|
.dropped_forward_packets_count
|
|
|
|
.fetch_add(retryable_packets_filtered_count, Ordering::Relaxed);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
process_transactions_summary.retryable_transaction_indexes =
|
|
|
|
filtered_retryable_transaction_indexes;
|
|
|
|
process_transactions_summary
|
|
|
|
}
|
|
|
|
|
|
|
|
/// Sends transactions to the bank.
|
|
|
|
///
|
|
|
|
/// Returns the number of transactions successfully processed by the bank, which may be less
|
|
|
|
/// than the total number if max PoH height was reached and the bank halted
|
|
|
|
fn process_transactions(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank: &Arc<Bank>,
|
|
|
|
bank_creation_time: &Instant,
|
|
|
|
transactions: &[SanitizedTransaction],
|
|
|
|
) -> ProcessTransactionsSummary {
|
|
|
|
let mut chunk_start = 0;
|
|
|
|
let mut all_retryable_tx_indexes = vec![];
|
|
|
|
// All the transactions that attempted execution. See description of
|
|
|
|
// struct ProcessTransactionsSummary above for possible outcomes.
|
|
|
|
let mut total_transactions_attempted_execution_count: usize = 0;
|
|
|
|
// All transactions that were executed and committed
|
|
|
|
let mut total_committed_transactions_count: usize = 0;
|
|
|
|
// All transactions that were executed and committed with a successful result
|
|
|
|
let mut total_committed_transactions_with_successful_result_count: usize = 0;
|
|
|
|
// All transactions that were executed but then failed record because the
|
|
|
|
// slot ended
|
|
|
|
let mut total_failed_commit_count: usize = 0;
|
|
|
|
let mut total_cost_model_throttled_transactions_count: usize = 0;
|
|
|
|
let mut total_cost_model_us: u64 = 0;
|
|
|
|
let mut total_execute_and_commit_timings = LeaderExecuteAndCommitTimings::default();
|
|
|
|
let mut total_error_counters = TransactionErrorMetrics::default();
|
|
|
|
let mut reached_max_poh_height = false;
|
|
|
|
while chunk_start != transactions.len() {
|
|
|
|
let chunk_end = std::cmp::min(
|
|
|
|
transactions.len(),
|
|
|
|
chunk_start + MAX_NUM_TRANSACTIONS_PER_BATCH,
|
|
|
|
);
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transaction_batch_output = self.process_and_record_transactions(
|
2023-02-15 08:52:13 -08:00
|
|
|
bank,
|
|
|
|
&transactions[chunk_start..chunk_end],
|
|
|
|
chunk_start,
|
|
|
|
);
|
|
|
|
|
|
|
|
let ProcessTransactionBatchOutput {
|
|
|
|
cost_model_throttled_transactions_count: new_cost_model_throttled_transactions_count,
|
|
|
|
cost_model_us: new_cost_model_us,
|
|
|
|
execute_and_commit_transactions_output,
|
|
|
|
} = process_transaction_batch_output;
|
2023-02-16 15:19:43 -08:00
|
|
|
saturating_add_assign!(
|
|
|
|
total_cost_model_throttled_transactions_count,
|
|
|
|
new_cost_model_throttled_transactions_count
|
|
|
|
);
|
|
|
|
saturating_add_assign!(total_cost_model_us, new_cost_model_us);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count: new_transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count: new_executed_transactions_count,
|
|
|
|
executed_with_successful_result_count: new_executed_with_successful_result_count,
|
|
|
|
retryable_transaction_indexes: new_retryable_transaction_indexes,
|
|
|
|
commit_transactions_result: new_commit_transactions_result,
|
|
|
|
execute_and_commit_timings: new_execute_and_commit_timings,
|
|
|
|
error_counters: new_error_counters,
|
|
|
|
..
|
|
|
|
} = execute_and_commit_transactions_output;
|
|
|
|
|
|
|
|
total_execute_and_commit_timings.accumulate(&new_execute_and_commit_timings);
|
|
|
|
total_error_counters.accumulate(&new_error_counters);
|
2023-02-16 15:19:43 -08:00
|
|
|
saturating_add_assign!(
|
|
|
|
total_transactions_attempted_execution_count,
|
|
|
|
new_transactions_attempted_execution_count
|
|
|
|
);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
trace!(
|
|
|
|
"process_transactions result: {:?}",
|
|
|
|
new_commit_transactions_result
|
|
|
|
);
|
|
|
|
|
|
|
|
if new_commit_transactions_result.is_ok() {
|
2023-02-16 15:19:43 -08:00
|
|
|
saturating_add_assign!(
|
|
|
|
total_committed_transactions_count,
|
|
|
|
new_executed_transactions_count
|
|
|
|
);
|
|
|
|
saturating_add_assign!(
|
|
|
|
total_committed_transactions_with_successful_result_count,
|
|
|
|
new_executed_with_successful_result_count
|
|
|
|
);
|
2023-02-15 08:52:13 -08:00
|
|
|
} else {
|
2023-02-16 15:19:43 -08:00
|
|
|
saturating_add_assign!(total_failed_commit_count, new_executed_transactions_count);
|
2023-02-15 08:52:13 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
// Add the retryable txs (transactions that errored in a way that warrants a retry)
|
|
|
|
// to the list of unprocessed txs.
|
|
|
|
all_retryable_tx_indexes.extend_from_slice(&new_retryable_transaction_indexes);
|
|
|
|
|
|
|
|
let should_bank_still_be_processing_txs =
|
|
|
|
Bank::should_bank_still_be_processing_txs(bank_creation_time, bank.ns_per_slot);
|
|
|
|
match (
|
|
|
|
new_commit_transactions_result,
|
|
|
|
should_bank_still_be_processing_txs,
|
|
|
|
) {
|
|
|
|
(Err(PohRecorderError::MaxHeightReached), _) | (_, false) => {
|
|
|
|
info!(
|
|
|
|
"process transactions: max height reached slot: {} height: {}",
|
|
|
|
bank.slot(),
|
|
|
|
bank.tick_height()
|
|
|
|
);
|
|
|
|
// process_and_record_transactions has returned all retryable errors in
|
|
|
|
// transactions[chunk_start..chunk_end], so we just need to push the remaining
|
|
|
|
// transactions into the unprocessed queue.
|
|
|
|
all_retryable_tx_indexes.extend(chunk_end..transactions.len());
|
|
|
|
reached_max_poh_height = true;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
_ => (),
|
|
|
|
}
|
|
|
|
// Don't exit early on any other type of error, continue processing...
|
|
|
|
chunk_start = chunk_end;
|
|
|
|
}
|
|
|
|
|
|
|
|
ProcessTransactionsSummary {
|
|
|
|
reached_max_poh_height,
|
|
|
|
transactions_attempted_execution_count: total_transactions_attempted_execution_count,
|
|
|
|
committed_transactions_count: total_committed_transactions_count,
|
|
|
|
committed_transactions_with_successful_result_count:
|
|
|
|
total_committed_transactions_with_successful_result_count,
|
|
|
|
failed_commit_count: total_failed_commit_count,
|
|
|
|
retryable_transaction_indexes: all_retryable_tx_indexes,
|
|
|
|
cost_model_throttled_transactions_count: total_cost_model_throttled_transactions_count,
|
|
|
|
cost_model_us: total_cost_model_us,
|
|
|
|
execute_and_commit_timings: total_execute_and_commit_timings,
|
|
|
|
error_counters: total_error_counters,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
pub fn process_and_record_transactions(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank: &Arc<Bank>,
|
|
|
|
txs: &[SanitizedTransaction],
|
|
|
|
chunk_offset: usize,
|
|
|
|
) -> ProcessTransactionBatchOutput {
|
|
|
|
let (
|
2023-05-12 08:53:22 -07:00
|
|
|
(transaction_qos_cost_results, cost_model_throttled_transactions_count),
|
2023-02-15 17:20:55 -08:00
|
|
|
cost_model_us,
|
2023-05-12 08:53:22 -07:00
|
|
|
) = measure_us!(self.qos_service.select_and_accumulate_transaction_costs(
|
|
|
|
bank,
|
|
|
|
txs,
|
|
|
|
std::iter::repeat(Ok(())) // no filtering before QoS
|
|
|
|
));
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// Only lock accounts for those transactions are selected for the block;
|
|
|
|
// Once accounts are locked, other threads cannot encode transactions that will modify the
|
|
|
|
// same account state
|
2023-05-12 08:53:22 -07:00
|
|
|
let (batch, lock_us) = measure_us!(bank.prepare_sanitized_batch_with_results(
|
|
|
|
txs,
|
|
|
|
transaction_qos_cost_results.iter().map(|r| match r {
|
|
|
|
Ok(_cost) => Ok(()),
|
|
|
|
Err(err) => Err(err.clone()),
|
|
|
|
})
|
|
|
|
));
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// retryable_txs includes AccountInUse, WouldExceedMaxBlockCostLimit
|
|
|
|
// WouldExceedMaxAccountCostLimit, WouldExceedMaxVoteCostLimit
|
|
|
|
// and WouldExceedMaxAccountDataCostLimit
|
|
|
|
let mut execute_and_commit_transactions_output =
|
2023-03-06 09:13:28 -08:00
|
|
|
self.execute_and_commit_transactions_locked(bank, &batch);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// Once the accounts are new transactions can enter the pipeline to process them
|
2023-02-15 17:20:55 -08:00
|
|
|
let (_, unlock_us) = measure_us!(drop(batch));
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
ref mut retryable_transaction_indexes,
|
|
|
|
ref execute_and_commit_timings,
|
|
|
|
ref commit_transactions_result,
|
|
|
|
..
|
|
|
|
} = execute_and_commit_transactions_output;
|
|
|
|
|
2023-05-17 09:24:59 -07:00
|
|
|
// once feature `apply_cost_tracker_during_replay` is activated, leader shall no longer
|
2023-05-18 10:37:09 -07:00
|
|
|
// adjust block with executed cost (a behavior more inline with bankless leader), it
|
|
|
|
// should use requested, or default `compute_unit_limit` as transaction's execution cost.
|
2023-05-17 09:24:59 -07:00
|
|
|
if !bank
|
|
|
|
.feature_set
|
|
|
|
.is_active(&feature_set::apply_cost_tracker_during_replay::id())
|
|
|
|
{
|
|
|
|
QosService::update_or_remove_transaction_costs(
|
|
|
|
transaction_qos_cost_results.iter(),
|
|
|
|
commit_transactions_result.as_ref().ok(),
|
|
|
|
bank,
|
|
|
|
);
|
|
|
|
}
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
retryable_transaction_indexes
|
|
|
|
.iter_mut()
|
|
|
|
.for_each(|x| *x += chunk_offset);
|
|
|
|
|
|
|
|
let (cu, us) =
|
|
|
|
Self::accumulate_execute_units_and_time(&execute_and_commit_timings.execute_timings);
|
2023-03-06 09:13:28 -08:00
|
|
|
self.qos_service.accumulate_actual_execute_cu(cu);
|
|
|
|
self.qos_service.accumulate_actual_execute_time(us);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// reports qos service stats for this batch
|
2023-04-14 13:15:31 -07:00
|
|
|
self.qos_service.report_metrics(bank.slot());
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
debug!(
|
|
|
|
"bank: {} lock: {}us unlock: {}us txs_len: {}",
|
|
|
|
bank.slot(),
|
2023-02-15 17:20:55 -08:00
|
|
|
lock_us,
|
|
|
|
unlock_us,
|
2023-02-15 08:52:13 -08:00
|
|
|
txs.len(),
|
|
|
|
);
|
|
|
|
|
|
|
|
ProcessTransactionBatchOutput {
|
|
|
|
cost_model_throttled_transactions_count,
|
2023-02-15 17:20:55 -08:00
|
|
|
cost_model_us,
|
2023-02-15 08:52:13 -08:00
|
|
|
execute_and_commit_transactions_output,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn execute_and_commit_transactions_locked(
|
2023-03-06 09:13:28 -08:00
|
|
|
&self,
|
2023-02-15 08:52:13 -08:00
|
|
|
bank: &Arc<Bank>,
|
|
|
|
batch: &TransactionBatch,
|
|
|
|
) -> ExecuteAndCommitTransactionsOutput {
|
2023-03-06 09:13:28 -08:00
|
|
|
let transaction_status_sender_enabled = self.committer.transaction_status_sender_enabled();
|
2023-02-15 08:52:13 -08:00
|
|
|
let mut execute_and_commit_timings = LeaderExecuteAndCommitTimings::default();
|
|
|
|
|
|
|
|
let mut pre_balance_info = PreBalanceInfo::default();
|
2023-02-15 17:20:55 -08:00
|
|
|
let (_, collect_balances_us) = measure_us!({
|
|
|
|
// If the extra meta-data services are enabled for RPC, collect the
|
|
|
|
// pre-balances for native and token programs.
|
|
|
|
if transaction_status_sender_enabled {
|
|
|
|
pre_balance_info.native = bank.collect_balances(batch);
|
|
|
|
pre_balance_info.token =
|
|
|
|
collect_token_balances(bank, batch, &mut pre_balance_info.mint_decimals)
|
|
|
|
}
|
|
|
|
});
|
|
|
|
execute_and_commit_timings.collect_balances_us = collect_balances_us;
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-02-15 17:20:55 -08:00
|
|
|
let (load_and_execute_transactions_output, load_execute_us) = measure_us!(bank
|
|
|
|
.load_and_execute_transactions(
|
2023-02-15 08:52:13 -08:00
|
|
|
batch,
|
|
|
|
MAX_PROCESSING_AGE,
|
|
|
|
transaction_status_sender_enabled,
|
|
|
|
transaction_status_sender_enabled,
|
|
|
|
transaction_status_sender_enabled,
|
|
|
|
&mut execute_and_commit_timings.execute_timings,
|
|
|
|
None, // account_overrides
|
2023-03-06 09:13:28 -08:00
|
|
|
self.log_messages_bytes_limit
|
2023-02-15 17:20:55 -08:00
|
|
|
));
|
|
|
|
execute_and_commit_timings.load_execute_us = load_execute_us;
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let LoadAndExecuteTransactionsOutput {
|
|
|
|
mut loaded_transactions,
|
|
|
|
execution_results,
|
|
|
|
mut retryable_transaction_indexes,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_non_vote_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
signature_count,
|
|
|
|
error_counters,
|
|
|
|
..
|
|
|
|
} = load_and_execute_transactions_output;
|
|
|
|
|
|
|
|
let transactions_attempted_execution_count = execution_results.len();
|
2023-02-15 17:20:55 -08:00
|
|
|
let (executed_transactions, execution_results_to_transactions_us) =
|
|
|
|
measure_us!(execution_results
|
2023-02-15 08:52:13 -08:00
|
|
|
.iter()
|
|
|
|
.zip(batch.sanitized_transactions())
|
|
|
|
.filter_map(|(execution_result, tx)| {
|
|
|
|
if execution_result.was_executed() {
|
|
|
|
Some(tx.to_versioned_transaction())
|
|
|
|
} else {
|
|
|
|
None
|
|
|
|
}
|
|
|
|
})
|
2023-02-15 17:20:55 -08:00
|
|
|
.collect_vec());
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-02-15 17:20:55 -08:00
|
|
|
let (freeze_lock, freeze_lock_us) = measure_us!(bank.freeze_lock());
|
|
|
|
execute_and_commit_timings.freeze_lock_us = freeze_lock_us;
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let (record_transactions_summary, record_us) = measure_us!(self
|
|
|
|
.transaction_recorder
|
|
|
|
.record_transactions(bank.slot(), executed_transactions));
|
2023-02-15 17:20:55 -08:00
|
|
|
execute_and_commit_timings.record_us = record_us;
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let RecordTransactionsSummary {
|
|
|
|
result: record_transactions_result,
|
|
|
|
record_transactions_timings,
|
|
|
|
starting_transaction_index,
|
|
|
|
} = record_transactions_summary;
|
|
|
|
execute_and_commit_timings.record_transactions_timings = RecordTransactionsTimings {
|
2023-02-15 17:20:55 -08:00
|
|
|
execution_results_to_transactions_us,
|
2023-02-15 08:52:13 -08:00
|
|
|
..record_transactions_timings
|
|
|
|
};
|
|
|
|
|
|
|
|
if let Err(recorder_err) = record_transactions_result {
|
|
|
|
retryable_transaction_indexes.extend(execution_results.iter().enumerate().filter_map(
|
|
|
|
|(index, execution_result)| execution_result.was_executed().then_some(index),
|
|
|
|
));
|
|
|
|
|
|
|
|
return ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
commit_transactions_result: Err(recorder_err),
|
|
|
|
execute_and_commit_timings,
|
|
|
|
error_counters,
|
|
|
|
};
|
|
|
|
}
|
|
|
|
|
|
|
|
let (commit_time_us, commit_transaction_statuses) = if executed_transactions_count != 0 {
|
2023-03-06 09:13:28 -08:00
|
|
|
self.committer.commit_transactions(
|
2023-02-15 08:52:13 -08:00
|
|
|
batch,
|
|
|
|
&mut loaded_transactions,
|
|
|
|
execution_results,
|
|
|
|
starting_transaction_index,
|
|
|
|
bank,
|
|
|
|
&mut pre_balance_info,
|
|
|
|
&mut execute_and_commit_timings,
|
|
|
|
signature_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_non_vote_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
)
|
|
|
|
} else {
|
|
|
|
(
|
|
|
|
0,
|
|
|
|
vec![CommitTransactionDetails::NotCommitted; execution_results.len()],
|
|
|
|
)
|
|
|
|
};
|
|
|
|
|
|
|
|
drop(freeze_lock);
|
|
|
|
|
|
|
|
debug!(
|
|
|
|
"bank: {} process_and_record_locked: {}us record: {}us commit: {}us txs_len: {}",
|
|
|
|
bank.slot(),
|
2023-02-15 17:20:55 -08:00
|
|
|
load_execute_us,
|
|
|
|
record_us,
|
2023-02-15 08:52:13 -08:00
|
|
|
commit_time_us,
|
|
|
|
batch.sanitized_transactions().len(),
|
|
|
|
);
|
|
|
|
|
|
|
|
debug!(
|
|
|
|
"execute_and_commit_transactions_locked: {:?}",
|
|
|
|
execute_and_commit_timings.execute_timings,
|
|
|
|
);
|
|
|
|
|
|
|
|
debug_assert_eq!(
|
|
|
|
commit_transaction_statuses.len(),
|
|
|
|
transactions_attempted_execution_count
|
|
|
|
);
|
|
|
|
|
|
|
|
ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
commit_transactions_result: Ok(commit_transaction_statuses),
|
|
|
|
execute_and_commit_timings,
|
|
|
|
error_counters,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn accumulate_execute_units_and_time(execute_timings: &ExecuteTimings) -> (u64, u64) {
|
2023-02-15 17:22:24 -08:00
|
|
|
execute_timings.details.per_program_timings.values().fold(
|
|
|
|
(0, 0),
|
|
|
|
|(units, times), program_timings| {
|
2023-02-15 08:52:13 -08:00
|
|
|
(
|
2023-02-16 15:19:43 -08:00
|
|
|
units.saturating_add(program_timings.accumulated_units),
|
|
|
|
times.saturating_add(program_timings.accumulated_us),
|
2023-02-15 08:52:13 -08:00
|
|
|
)
|
2023-02-15 17:22:24 -08:00
|
|
|
},
|
|
|
|
)
|
2023-02-15 08:52:13 -08:00
|
|
|
}
|
|
|
|
|
|
|
|
/// This function filters pending packets that are still valid
|
|
|
|
/// # Arguments
|
|
|
|
/// * `transactions` - a batch of transactions deserialized from packets
|
|
|
|
/// * `pending_indexes` - identifies which indexes in the `transactions` list are still pending
|
|
|
|
fn filter_pending_packets_from_pending_txs(
|
|
|
|
bank: &Arc<Bank>,
|
|
|
|
transactions: &[SanitizedTransaction],
|
|
|
|
pending_indexes: &[usize],
|
|
|
|
) -> Vec<usize> {
|
|
|
|
let filter =
|
|
|
|
Self::prepare_filter_for_pending_transactions(transactions.len(), pending_indexes);
|
|
|
|
|
|
|
|
let results = bank.check_transactions_with_forwarding_delay(
|
|
|
|
transactions,
|
|
|
|
&filter,
|
|
|
|
FORWARD_TRANSACTIONS_TO_LEADER_AT_SLOT_OFFSET,
|
|
|
|
);
|
|
|
|
|
|
|
|
Self::filter_valid_transaction_indexes(&results)
|
|
|
|
}
|
|
|
|
|
|
|
|
/// This function creates a filter of transaction results with Ok() for every pending
|
|
|
|
/// transaction. The non-pending transactions are marked with TransactionError
|
|
|
|
fn prepare_filter_for_pending_transactions(
|
|
|
|
transactions_len: usize,
|
|
|
|
pending_tx_indexes: &[usize],
|
|
|
|
) -> Vec<transaction::Result<()>> {
|
|
|
|
let mut mask = vec![Err(TransactionError::BlockhashNotFound); transactions_len];
|
|
|
|
pending_tx_indexes.iter().for_each(|x| mask[*x] = Ok(()));
|
|
|
|
mask
|
|
|
|
}
|
|
|
|
|
|
|
|
/// This function returns a vector containing index of all valid transactions. A valid
|
|
|
|
/// transaction has result Ok() as the value
|
|
|
|
fn filter_valid_transaction_indexes(valid_txs: &[TransactionCheckResult]) -> Vec<usize> {
|
|
|
|
valid_txs
|
|
|
|
.iter()
|
|
|
|
.enumerate()
|
|
|
|
.filter_map(|(index, (x, _h))| if x.is_ok() { Some(index) } else { None })
|
|
|
|
.collect_vec()
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
#[cfg(test)]
|
|
|
|
mod tests {
|
|
|
|
use {
|
|
|
|
super::*,
|
|
|
|
crate::{
|
|
|
|
banking_stage::tests::{create_slow_genesis_config, simulate_poh},
|
2023-05-09 14:39:23 -07:00
|
|
|
immutable_deserialized_packet::DeserializedPacketError,
|
|
|
|
unprocessed_packet_batches::{DeserializedPacket, UnprocessedPacketBatches},
|
2023-02-15 08:52:13 -08:00
|
|
|
unprocessed_transaction_storage::ThreadType,
|
|
|
|
},
|
|
|
|
crossbeam_channel::{unbounded, Receiver},
|
|
|
|
solana_address_lookup_table_program::state::{AddressLookupTable, LookupTableMeta},
|
|
|
|
solana_entry::entry::{next_entry, next_versioned_entry},
|
|
|
|
solana_ledger::{
|
|
|
|
blockstore::{entries_to_test_shreds, Blockstore},
|
|
|
|
blockstore_processor::TransactionStatusSender,
|
|
|
|
genesis_utils::GenesisConfigInfo,
|
|
|
|
get_tmp_ledger_path_auto_delete,
|
|
|
|
leader_schedule_cache::LeaderScheduleCache,
|
|
|
|
},
|
2023-05-09 14:39:23 -07:00
|
|
|
solana_perf::packet::Packet,
|
2023-02-15 08:52:13 -08:00
|
|
|
solana_poh::poh_recorder::{PohRecorder, WorkingBankEntry},
|
|
|
|
solana_program_runtime::timings::ProgramTiming,
|
|
|
|
solana_rpc::transaction_status_service::TransactionStatusService,
|
2023-05-18 10:37:58 -07:00
|
|
|
solana_runtime::{cost_model::CostModel, prioritization_fee_cache::PrioritizationFeeCache},
|
2023-02-15 08:52:13 -08:00
|
|
|
solana_sdk::{
|
|
|
|
account::AccountSharedData,
|
|
|
|
instruction::InstructionError,
|
|
|
|
message::{v0, v0::MessageAddressTableLookup, MessageHeader, VersionedMessage},
|
|
|
|
poh_config::PohConfig,
|
|
|
|
pubkey::Pubkey,
|
|
|
|
signature::Keypair,
|
|
|
|
signer::Signer,
|
|
|
|
system_transaction,
|
|
|
|
transaction::{MessageHash, Transaction, VersionedTransaction},
|
|
|
|
},
|
|
|
|
solana_transaction_status::{TransactionStatusMeta, VersionedTransactionWithStatusMeta},
|
|
|
|
std::{
|
|
|
|
borrow::Cow,
|
|
|
|
path::Path,
|
|
|
|
sync::{
|
|
|
|
atomic::{AtomicBool, AtomicU64},
|
|
|
|
RwLock,
|
|
|
|
},
|
2023-03-09 09:13:04 -08:00
|
|
|
thread::JoinHandle,
|
2023-02-15 08:52:13 -08:00
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
fn sanitize_transactions(txs: Vec<Transaction>) -> Vec<SanitizedTransaction> {
|
|
|
|
txs.into_iter()
|
|
|
|
.map(SanitizedTransaction::from_transaction_for_tests)
|
|
|
|
.collect()
|
|
|
|
}
|
|
|
|
|
|
|
|
fn execute_transactions_with_dummy_poh_service(
|
|
|
|
bank: Arc<Bank>,
|
|
|
|
transactions: Vec<Transaction>,
|
|
|
|
) -> ProcessTransactionsSummary {
|
|
|
|
let transactions = sanitize_transactions(transactions);
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&Pubkey::new_unique(),
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_summary =
|
|
|
|
consumer.process_transactions(&bank, &Instant::now(), &transactions);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
|
|
|
|
process_transactions_summary
|
|
|
|
}
|
|
|
|
|
|
|
|
fn generate_new_address_lookup_table(
|
|
|
|
authority: Option<Pubkey>,
|
|
|
|
num_addresses: usize,
|
|
|
|
) -> AddressLookupTable<'static> {
|
|
|
|
let mut addresses = Vec::with_capacity(num_addresses);
|
|
|
|
addresses.resize_with(num_addresses, Pubkey::new_unique);
|
|
|
|
AddressLookupTable {
|
|
|
|
meta: LookupTableMeta {
|
|
|
|
authority,
|
|
|
|
..LookupTableMeta::default()
|
|
|
|
},
|
|
|
|
addresses: Cow::Owned(addresses),
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn store_address_lookup_table(
|
|
|
|
bank: &Bank,
|
|
|
|
account_address: Pubkey,
|
|
|
|
address_lookup_table: AddressLookupTable<'static>,
|
|
|
|
) -> AccountSharedData {
|
|
|
|
let data = address_lookup_table.serialize_for_tests().unwrap();
|
|
|
|
let mut account =
|
|
|
|
AccountSharedData::new(1, data.len(), &solana_address_lookup_table_program::id());
|
|
|
|
account.set_data(data);
|
|
|
|
bank.store_account(&account_address, &account);
|
|
|
|
|
|
|
|
account
|
|
|
|
}
|
|
|
|
|
|
|
|
#[allow(clippy::type_complexity)]
|
|
|
|
fn setup_conflicting_transactions(
|
|
|
|
ledger_path: &Path,
|
|
|
|
) -> (
|
|
|
|
Vec<Transaction>,
|
|
|
|
Arc<Bank>,
|
|
|
|
Arc<RwLock<PohRecorder>>,
|
|
|
|
Receiver<WorkingBankEntry>,
|
|
|
|
JoinHandle<()>,
|
|
|
|
) {
|
|
|
|
Blockstore::destroy(ledger_path).unwrap();
|
|
|
|
let genesis_config_info = create_slow_genesis_config(10_000);
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = &genesis_config_info;
|
|
|
|
let blockstore =
|
|
|
|
Blockstore::open(ledger_path).expect("Expected to be able to open database ledger");
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(genesis_config));
|
|
|
|
let exit = Arc::new(AtomicBool::default());
|
|
|
|
let (poh_recorder, entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&solana_sdk::pubkey::new_rand(),
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
exit,
|
|
|
|
);
|
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
// Set up unparallelizable conflicting transactions
|
|
|
|
let pubkey0 = solana_sdk::pubkey::new_rand();
|
|
|
|
let pubkey1 = solana_sdk::pubkey::new_rand();
|
|
|
|
let pubkey2 = solana_sdk::pubkey::new_rand();
|
|
|
|
let transactions = vec![
|
|
|
|
system_transaction::transfer(mint_keypair, &pubkey0, 1, genesis_config.hash()),
|
|
|
|
system_transaction::transfer(mint_keypair, &pubkey1, 1, genesis_config.hash()),
|
|
|
|
system_transaction::transfer(mint_keypair, &pubkey2, 1, genesis_config.hash()),
|
|
|
|
];
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
(
|
|
|
|
transactions,
|
|
|
|
bank,
|
|
|
|
poh_recorder,
|
|
|
|
entry_receiver,
|
|
|
|
poh_simulator,
|
|
|
|
)
|
|
|
|
}
|
|
|
|
|
2023-05-09 14:39:23 -07:00
|
|
|
fn transactions_to_deserialized_packets(
|
|
|
|
transactions: &[Transaction],
|
|
|
|
) -> Result<Vec<DeserializedPacket>, DeserializedPacketError> {
|
|
|
|
transactions
|
|
|
|
.iter()
|
|
|
|
.map(|transaction| {
|
|
|
|
let packet = Packet::from_data(None, transaction)?;
|
|
|
|
DeserializedPacket::new(packet)
|
|
|
|
})
|
|
|
|
.collect()
|
|
|
|
}
|
|
|
|
|
2023-02-15 08:52:13 -08:00
|
|
|
#[test]
|
|
|
|
fn test_bank_process_and_record_transactions() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&pubkey,
|
|
|
|
1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
)]);
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&pubkey,
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
commit_transactions_result,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
|
|
|
|
assert_eq!(transactions_attempted_execution_count, 1);
|
|
|
|
assert_eq!(executed_transactions_count, 1);
|
|
|
|
assert_eq!(executed_with_successful_result_count, 1);
|
|
|
|
assert!(commit_transactions_result.is_ok());
|
|
|
|
|
|
|
|
// Tick up to max tick height
|
|
|
|
while poh_recorder.read().unwrap().tick_height() != bank.max_tick_height() {
|
|
|
|
poh_recorder.write().unwrap().tick();
|
|
|
|
}
|
|
|
|
|
|
|
|
let mut done = false;
|
|
|
|
// read entries until I find mine, might be ticks...
|
|
|
|
while let Ok((_bank, (entry, _tick_height))) = entry_receiver.recv() {
|
|
|
|
if !entry.is_tick() {
|
|
|
|
trace!("got entry");
|
|
|
|
assert_eq!(entry.transactions.len(), transactions.len());
|
|
|
|
assert_eq!(bank.get_balance(&pubkey), 1);
|
|
|
|
done = true;
|
|
|
|
}
|
|
|
|
if done {
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
trace!("done ticking");
|
|
|
|
|
|
|
|
assert!(done);
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&pubkey,
|
|
|
|
2,
|
|
|
|
genesis_config.hash(),
|
|
|
|
)]);
|
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
commit_transactions_result,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
assert_eq!(transactions_attempted_execution_count, 1);
|
|
|
|
// Transactions was still executed, just wasn't committed, so should be counted here.
|
|
|
|
assert_eq!(executed_transactions_count, 1);
|
|
|
|
assert_eq!(executed_with_successful_result_count, 1);
|
|
|
|
assert_eq!(retryable_transaction_indexes, vec![0]);
|
|
|
|
assert_matches!(
|
|
|
|
commit_transactions_result,
|
|
|
|
Err(PohRecorderError::MaxHeightReached)
|
|
|
|
);
|
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
|
|
|
|
assert_eq!(bank.get_balance(&pubkey), 1);
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_bank_process_and_record_transactions_all_unexecuted() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
|
|
|
|
let transactions = {
|
|
|
|
let mut tx =
|
|
|
|
system_transaction::transfer(&mint_keypair, &pubkey, 1, genesis_config.hash());
|
|
|
|
// Add duplicate account key
|
|
|
|
tx.message.account_keys.push(pubkey);
|
|
|
|
sanitize_transactions(vec![tx])
|
|
|
|
};
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&pubkey,
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
commit_transactions_result,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
|
|
|
|
assert_eq!(transactions_attempted_execution_count, 1);
|
|
|
|
assert_eq!(executed_transactions_count, 0);
|
|
|
|
assert_eq!(executed_with_successful_result_count, 0);
|
|
|
|
assert!(retryable_transaction_indexes.is_empty());
|
|
|
|
assert_eq!(
|
|
|
|
commit_transactions_result.ok(),
|
|
|
|
Some(vec![CommitTransactionDetails::NotCommitted; 1])
|
|
|
|
);
|
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_bank_process_and_record_transactions_cost_tracker() {
|
2023-05-18 10:37:58 -07:00
|
|
|
for apply_cost_tracker_during_replay_enabled in [true, false] {
|
|
|
|
bank_process_and_record_transactions_cost_tracker(
|
|
|
|
apply_cost_tracker_during_replay_enabled,
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
fn bank_process_and_record_transactions_cost_tracker(
|
|
|
|
apply_cost_tracker_during_replay_enabled: bool,
|
|
|
|
) {
|
2023-02-15 08:52:13 -08:00
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
2023-05-17 09:24:59 -07:00
|
|
|
let mut bank = Bank::new_no_wallclock_throttle_for_tests(&genesis_config);
|
2023-05-18 10:37:58 -07:00
|
|
|
if !apply_cost_tracker_during_replay_enabled {
|
|
|
|
bank.deactivate_feature(&feature_set::apply_cost_tracker_during_replay::id());
|
|
|
|
}
|
2023-05-17 09:24:59 -07:00
|
|
|
let bank = Arc::new(bank);
|
2023-02-15 08:52:13 -08:00
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&pubkey,
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let get_block_cost = || bank.read_cost_tracker().unwrap().block_cost();
|
|
|
|
let get_tx_count = || bank.read_cost_tracker().unwrap().transaction_count();
|
|
|
|
assert_eq!(get_block_cost(), 0);
|
|
|
|
assert_eq!(get_tx_count(), 0);
|
|
|
|
|
|
|
|
//
|
|
|
|
// TEST: cost tracker's block cost increases when successfully processing a tx
|
|
|
|
//
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&pubkey,
|
|
|
|
1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
)]);
|
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
commit_transactions_result,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
assert_eq!(executed_with_successful_result_count, 1);
|
|
|
|
assert!(commit_transactions_result.is_ok());
|
|
|
|
|
|
|
|
let single_transfer_cost = get_block_cost();
|
|
|
|
assert_ne!(single_transfer_cost, 0);
|
|
|
|
assert_eq!(get_tx_count(), 1);
|
|
|
|
|
|
|
|
//
|
|
|
|
// TEST: When a tx in a batch can't be executed (here because of account
|
2023-05-18 10:37:58 -07:00
|
|
|
// locks), then its cost does not affect the cost tracker only if qos
|
|
|
|
// adjusts it with actual execution cost (when apply_cost_tracker_during_replay
|
|
|
|
// is not enabled).
|
2023-02-15 08:52:13 -08:00
|
|
|
//
|
|
|
|
|
|
|
|
let allocate_keypair = Keypair::new();
|
2023-05-18 10:37:58 -07:00
|
|
|
|
2023-02-15 08:52:13 -08:00
|
|
|
let transactions = sanitize_transactions(vec![
|
|
|
|
system_transaction::transfer(&mint_keypair, &pubkey, 2, genesis_config.hash()),
|
|
|
|
// intentionally use a tx that has a different cost
|
|
|
|
system_transaction::allocate(
|
|
|
|
&mint_keypair,
|
|
|
|
&allocate_keypair,
|
|
|
|
genesis_config.hash(),
|
|
|
|
1,
|
|
|
|
),
|
|
|
|
]);
|
2023-05-18 10:37:58 -07:00
|
|
|
let mut expected_block_cost = 2 * single_transfer_cost;
|
|
|
|
let mut expected_tracked_tx_count = 2;
|
|
|
|
if apply_cost_tracker_during_replay_enabled {
|
|
|
|
expected_block_cost +=
|
|
|
|
CostModel::calculate_cost(&transactions[1], &bank.feature_set).sum();
|
|
|
|
expected_tracked_tx_count += 1;
|
|
|
|
}
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
executed_with_successful_result_count,
|
|
|
|
commit_transactions_result,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
assert_eq!(executed_with_successful_result_count, 1);
|
|
|
|
assert!(commit_transactions_result.is_ok());
|
|
|
|
assert_eq!(retryable_transaction_indexes, vec![1]);
|
|
|
|
|
2023-05-18 10:37:58 -07:00
|
|
|
assert_eq!(get_block_cost(), expected_block_cost);
|
|
|
|
assert_eq!(get_tx_count(), expected_tracked_tx_count);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_bank_process_and_record_transactions_account_in_use() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
let pubkey1 = solana_sdk::pubkey::new_rand();
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![
|
|
|
|
system_transaction::transfer(&mint_keypair, &pubkey, 1, genesis_config.hash()),
|
|
|
|
system_transaction::transfer(&mint_keypair, &pubkey1, 1, genesis_config.hash()),
|
|
|
|
]);
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&pubkey,
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_batch_output =
|
|
|
|
consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
|
|
|
|
let ExecuteAndCommitTransactionsOutput {
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
executed_transactions_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
commit_transactions_result,
|
|
|
|
..
|
|
|
|
} = process_transactions_batch_output.execute_and_commit_transactions_output;
|
|
|
|
|
|
|
|
assert_eq!(transactions_attempted_execution_count, 2);
|
|
|
|
assert_eq!(executed_transactions_count, 1);
|
|
|
|
assert_eq!(retryable_transaction_indexes, vec![1],);
|
|
|
|
assert!(commit_transactions_result.is_ok());
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_process_transactions_instruction_error() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let lamports = 10_000;
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(lamports);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
// set cost tracker limits to MAX so it will not filter out TXs
|
|
|
|
bank.write_cost_tracker()
|
|
|
|
.unwrap()
|
|
|
|
.set_limits(std::u64::MAX, std::u64::MAX, std::u64::MAX);
|
|
|
|
|
|
|
|
// Transfer more than the balance of the mint keypair, should cause a
|
|
|
|
// InstructionError::InsufficientFunds that is then committed. Needs to be
|
|
|
|
// MAX_NUM_TRANSACTIONS_PER_BATCH at least so it doesn't conflict on account locks
|
|
|
|
// with the below transaction
|
|
|
|
let mut transactions = vec![
|
|
|
|
system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&Pubkey::new_unique(),
|
|
|
|
lamports + 1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
);
|
|
|
|
MAX_NUM_TRANSACTIONS_PER_BATCH
|
|
|
|
];
|
|
|
|
|
|
|
|
// Make one transaction that will succeed.
|
|
|
|
transactions.push(system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&Pubkey::new_unique(),
|
|
|
|
1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
));
|
|
|
|
|
|
|
|
let transactions_count = transactions.len();
|
|
|
|
let ProcessTransactionsSummary {
|
|
|
|
reached_max_poh_height,
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
committed_transactions_count,
|
|
|
|
committed_transactions_with_successful_result_count,
|
|
|
|
failed_commit_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = execute_transactions_with_dummy_poh_service(bank, transactions);
|
|
|
|
|
|
|
|
// All the transactions should have been replayed, but only 1 committed
|
|
|
|
assert!(!reached_max_poh_height);
|
|
|
|
assert_eq!(transactions_attempted_execution_count, transactions_count);
|
|
|
|
// Both transactions should have been committed, even though one was an error,
|
|
|
|
// because InstructionErrors are committed
|
|
|
|
assert_eq!(committed_transactions_count, 2);
|
|
|
|
assert_eq!(committed_transactions_with_successful_result_count, 1);
|
|
|
|
assert_eq!(failed_commit_count, 0);
|
|
|
|
assert_eq!(
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
(1..transactions_count - 1).collect::<Vec<usize>>()
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_process_transactions_account_in_use() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
// set cost tracker limits to MAX so it will not filter out TXs
|
|
|
|
bank.write_cost_tracker()
|
|
|
|
.unwrap()
|
|
|
|
.set_limits(std::u64::MAX, std::u64::MAX, std::u64::MAX);
|
|
|
|
|
|
|
|
// Make all repetitive transactions that conflict on the `mint_keypair`, so only 1 should be executed
|
|
|
|
let mut transactions = vec![
|
|
|
|
system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&Pubkey::new_unique(),
|
|
|
|
1,
|
|
|
|
genesis_config.hash()
|
|
|
|
);
|
|
|
|
MAX_NUM_TRANSACTIONS_PER_BATCH
|
|
|
|
];
|
|
|
|
|
|
|
|
// Make one more in separate batch that also conflicts, but because it's in a separate batch, it
|
|
|
|
// should be executed
|
|
|
|
transactions.push(system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&Pubkey::new_unique(),
|
|
|
|
1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
));
|
|
|
|
|
|
|
|
let transactions_count = transactions.len();
|
|
|
|
let ProcessTransactionsSummary {
|
|
|
|
reached_max_poh_height,
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
committed_transactions_count,
|
|
|
|
committed_transactions_with_successful_result_count,
|
|
|
|
failed_commit_count,
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = execute_transactions_with_dummy_poh_service(bank, transactions);
|
|
|
|
|
|
|
|
// All the transactions should have been replayed, but only 2 committed (first and last)
|
|
|
|
assert!(!reached_max_poh_height);
|
|
|
|
assert_eq!(transactions_attempted_execution_count, transactions_count);
|
|
|
|
assert_eq!(committed_transactions_count, 2);
|
|
|
|
assert_eq!(committed_transactions_with_successful_result_count, 2);
|
|
|
|
assert_eq!(failed_commit_count, 0,);
|
|
|
|
|
|
|
|
// Everything except first and last index of the transactions failed and are last retryable
|
|
|
|
assert_eq!(
|
|
|
|
retryable_transaction_indexes,
|
|
|
|
(1..transactions_count - 1).collect::<Vec<usize>>()
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_process_transactions_returns_unprocessed_txs() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
|
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&pubkey,
|
|
|
|
1,
|
|
|
|
genesis_config.hash(),
|
|
|
|
)]);
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&solana_sdk::pubkey::new_rand(),
|
2023-04-19 11:41:18 -07:00
|
|
|
Arc::new(blockstore),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
|
|
|
|
|
|
|
// Poh Recorder has no working bank, so should throw MaxHeightReached error on
|
|
|
|
// record
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &Arc::new(RwLock::new(poh_recorder)));
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder.clone(), QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let process_transactions_summary =
|
|
|
|
consumer.process_transactions(&bank, &Instant::now(), &transactions);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
let ProcessTransactionsSummary {
|
|
|
|
reached_max_poh_height,
|
|
|
|
transactions_attempted_execution_count,
|
|
|
|
committed_transactions_count,
|
|
|
|
committed_transactions_with_successful_result_count,
|
|
|
|
failed_commit_count,
|
|
|
|
mut retryable_transaction_indexes,
|
|
|
|
..
|
|
|
|
} = process_transactions_summary;
|
|
|
|
assert!(reached_max_poh_height);
|
|
|
|
assert_eq!(transactions_attempted_execution_count, 1);
|
|
|
|
assert_eq!(failed_commit_count, 1);
|
|
|
|
// MaxHeightReached error does not commit, should be zero here
|
|
|
|
assert_eq!(committed_transactions_count, 0);
|
|
|
|
assert_eq!(committed_transactions_with_successful_result_count, 0);
|
|
|
|
|
|
|
|
retryable_transaction_indexes.sort_unstable();
|
|
|
|
let expected: Vec<usize> = (0..transactions.len()).collect();
|
|
|
|
assert_eq!(retryable_transaction_indexes, expected);
|
|
|
|
|
|
|
|
recorder.is_exited.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_write_persist_transaction_status() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
mut genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(solana_sdk::native_token::sol_to_lamports(1000.0));
|
|
|
|
genesis_config.rent.lamports_per_byte_year = 50;
|
|
|
|
genesis_config.rent.exemption_threshold = 2.0;
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
let pubkey = solana_sdk::pubkey::new_rand();
|
|
|
|
let pubkey1 = solana_sdk::pubkey::new_rand();
|
|
|
|
let keypair1 = Keypair::new();
|
|
|
|
|
|
|
|
let rent_exempt_amount = bank.get_minimum_balance_for_rent_exemption(0);
|
|
|
|
|
|
|
|
let success_tx = system_transaction::transfer(
|
|
|
|
&mint_keypair,
|
|
|
|
&pubkey,
|
|
|
|
rent_exempt_amount,
|
|
|
|
genesis_config.hash(),
|
|
|
|
);
|
|
|
|
let success_signature = success_tx.signatures[0];
|
|
|
|
let entry_1 = next_entry(&genesis_config.hash(), 1, vec![success_tx.clone()]);
|
|
|
|
let ix_error_tx = system_transaction::transfer(
|
|
|
|
&keypair1,
|
|
|
|
&pubkey1,
|
|
|
|
2 * rent_exempt_amount,
|
|
|
|
genesis_config.hash(),
|
|
|
|
);
|
|
|
|
let ix_error_signature = ix_error_tx.signatures[0];
|
|
|
|
let entry_2 = next_entry(&entry_1.hash, 1, vec![ix_error_tx.clone()]);
|
|
|
|
let entries = vec![entry_1, entry_2];
|
|
|
|
|
|
|
|
let transactions = sanitize_transactions(vec![success_tx, ix_error_tx]);
|
|
|
|
bank.transfer(rent_exempt_amount, &mint_keypair, &keypair1.pubkey())
|
|
|
|
.unwrap();
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let blockstore = Arc::new(blockstore);
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&pubkey,
|
2023-04-19 11:41:18 -07:00
|
|
|
blockstore.clone(),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
|
|
|
|
let shreds = entries_to_test_shreds(
|
|
|
|
&entries,
|
|
|
|
bank.slot(),
|
|
|
|
0, // parent_slot
|
|
|
|
true, // is_full_slot
|
|
|
|
0, // version
|
|
|
|
true, // merkle_variant
|
|
|
|
);
|
|
|
|
blockstore.insert_shreds(shreds, None, false).unwrap();
|
|
|
|
blockstore.set_roots(std::iter::once(&bank.slot())).unwrap();
|
|
|
|
|
|
|
|
let (transaction_status_sender, transaction_status_receiver) = unbounded();
|
|
|
|
let transaction_status_service = TransactionStatusService::new(
|
|
|
|
transaction_status_receiver,
|
|
|
|
Arc::new(AtomicU64::default()),
|
|
|
|
true,
|
|
|
|
None,
|
|
|
|
blockstore.clone(),
|
|
|
|
false,
|
|
|
|
&Arc::new(AtomicBool::new(false)),
|
|
|
|
);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
|
|
|
let committer = Committer::new(
|
|
|
|
Some(TransactionStatusSender {
|
|
|
|
sender: transaction_status_sender,
|
|
|
|
}),
|
|
|
|
replay_vote_sender,
|
2023-03-23 17:05:54 -07:00
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
2023-02-15 08:52:13 -08:00
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let _ = consumer.process_and_record_transactions(&bank, &transactions, 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
drop(consumer); // drop/disconnect transaction_status_sender
|
2023-02-15 08:52:13 -08:00
|
|
|
transaction_status_service.join().unwrap();
|
|
|
|
|
|
|
|
let confirmed_block = blockstore.get_rooted_block(bank.slot(), false).unwrap();
|
|
|
|
let actual_tx_results: Vec<_> = confirmed_block
|
|
|
|
.transactions
|
|
|
|
.into_iter()
|
|
|
|
.map(|VersionedTransactionWithStatusMeta { transaction, meta }| {
|
|
|
|
(transaction.signatures[0], meta.status)
|
|
|
|
})
|
|
|
|
.collect();
|
|
|
|
let expected_tx_results = vec![
|
|
|
|
(success_signature, Ok(())),
|
|
|
|
(
|
|
|
|
ix_error_signature,
|
|
|
|
Err(TransactionError::InstructionError(
|
|
|
|
0,
|
|
|
|
InstructionError::Custom(1),
|
|
|
|
)),
|
|
|
|
),
|
|
|
|
];
|
|
|
|
assert_eq!(actual_tx_results, expected_tx_results);
|
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_write_persist_loaded_addresses() {
|
|
|
|
solana_logger::setup();
|
|
|
|
let GenesisConfigInfo {
|
|
|
|
genesis_config,
|
|
|
|
mint_keypair,
|
|
|
|
..
|
|
|
|
} = create_slow_genesis_config(10_000);
|
|
|
|
let bank = Arc::new(Bank::new_no_wallclock_throttle_for_tests(&genesis_config));
|
|
|
|
let keypair = Keypair::new();
|
|
|
|
|
|
|
|
let address_table_key = Pubkey::new_unique();
|
|
|
|
let address_table_state = generate_new_address_lookup_table(None, 2);
|
|
|
|
store_address_lookup_table(&bank, address_table_key, address_table_state);
|
|
|
|
|
|
|
|
let bank = Arc::new(Bank::new_from_parent(&bank, &Pubkey::new_unique(), 1));
|
|
|
|
let message = VersionedMessage::V0(v0::Message {
|
|
|
|
header: MessageHeader {
|
|
|
|
num_required_signatures: 1,
|
|
|
|
num_readonly_signed_accounts: 0,
|
|
|
|
num_readonly_unsigned_accounts: 0,
|
|
|
|
},
|
|
|
|
recent_blockhash: genesis_config.hash(),
|
|
|
|
account_keys: vec![keypair.pubkey()],
|
|
|
|
address_table_lookups: vec![MessageAddressTableLookup {
|
|
|
|
account_key: address_table_key,
|
|
|
|
writable_indexes: vec![0],
|
|
|
|
readonly_indexes: vec![1],
|
|
|
|
}],
|
|
|
|
instructions: vec![],
|
|
|
|
});
|
|
|
|
|
|
|
|
let tx = VersionedTransaction::try_new(message, &[&keypair]).unwrap();
|
|
|
|
let sanitized_tx = SanitizedTransaction::try_create(
|
|
|
|
tx.clone(),
|
|
|
|
MessageHash::Compute,
|
|
|
|
Some(false),
|
|
|
|
bank.as_ref(),
|
|
|
|
true, // require_static_program_ids
|
|
|
|
)
|
|
|
|
.unwrap();
|
|
|
|
|
|
|
|
let entry = next_versioned_entry(&genesis_config.hash(), 1, vec![tx]);
|
|
|
|
let entries = vec![entry];
|
|
|
|
|
|
|
|
bank.transfer(1, &mint_keypair, &keypair.pubkey()).unwrap();
|
|
|
|
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let blockstore = Blockstore::open(ledger_path.path())
|
|
|
|
.expect("Expected to be able to open database ledger");
|
|
|
|
let blockstore = Arc::new(blockstore);
|
|
|
|
let (poh_recorder, _entry_receiver, record_receiver) = PohRecorder::new(
|
|
|
|
bank.tick_height(),
|
|
|
|
bank.last_blockhash(),
|
|
|
|
bank.clone(),
|
|
|
|
Some((4, 4)),
|
|
|
|
bank.ticks_per_slot(),
|
|
|
|
&Pubkey::new_unique(),
|
2023-04-19 11:41:18 -07:00
|
|
|
blockstore.clone(),
|
2023-02-15 08:52:13 -08:00
|
|
|
&Arc::new(LeaderScheduleCache::new_from_bank(&bank)),
|
|
|
|
&PohConfig::default(),
|
|
|
|
Arc::new(AtomicBool::default()),
|
|
|
|
);
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let poh_recorder = Arc::new(RwLock::new(poh_recorder));
|
|
|
|
|
|
|
|
let poh_simulator = simulate_poh(record_receiver, &poh_recorder);
|
|
|
|
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
|
|
|
|
let shreds = entries_to_test_shreds(
|
|
|
|
&entries,
|
|
|
|
bank.slot(),
|
|
|
|
0, // parent_slot
|
|
|
|
true, // is_full_slot
|
|
|
|
0, // version
|
|
|
|
true, // merkle_variant
|
|
|
|
);
|
|
|
|
blockstore.insert_shreds(shreds, None, false).unwrap();
|
|
|
|
blockstore.set_roots(std::iter::once(&bank.slot())).unwrap();
|
|
|
|
|
|
|
|
let (transaction_status_sender, transaction_status_receiver) = unbounded();
|
|
|
|
let transaction_status_service = TransactionStatusService::new(
|
|
|
|
transaction_status_receiver,
|
|
|
|
Arc::new(AtomicU64::default()),
|
|
|
|
true,
|
|
|
|
None,
|
|
|
|
blockstore.clone(),
|
|
|
|
false,
|
|
|
|
&Arc::new(AtomicBool::new(false)),
|
|
|
|
);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
|
|
|
let committer = Committer::new(
|
|
|
|
Some(TransactionStatusSender {
|
|
|
|
sender: transaction_status_sender,
|
|
|
|
}),
|
|
|
|
replay_vote_sender,
|
2023-03-23 17:05:54 -07:00
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
2023-02-15 08:52:13 -08:00
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
let _ = consumer.process_and_record_transactions(&bank, &[sanitized_tx.clone()], 0);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-06 09:13:28 -08:00
|
|
|
drop(consumer); // drop/disconnect transaction_status_sender
|
2023-02-15 08:52:13 -08:00
|
|
|
transaction_status_service.join().unwrap();
|
|
|
|
|
|
|
|
let mut confirmed_block = blockstore.get_rooted_block(bank.slot(), false).unwrap();
|
|
|
|
assert_eq!(confirmed_block.transactions.len(), 1);
|
|
|
|
|
|
|
|
let recorded_meta = confirmed_block.transactions.pop().unwrap().meta;
|
|
|
|
assert_eq!(
|
|
|
|
recorded_meta,
|
|
|
|
TransactionStatusMeta {
|
|
|
|
status: Ok(()),
|
|
|
|
pre_balances: vec![1, 0, 0],
|
|
|
|
post_balances: vec![1, 0, 0],
|
|
|
|
pre_token_balances: Some(vec![]),
|
|
|
|
post_token_balances: Some(vec![]),
|
|
|
|
rewards: Some(vec![]),
|
|
|
|
loaded_addresses: sanitized_tx.get_loaded_addresses(),
|
|
|
|
compute_units_consumed: Some(0),
|
|
|
|
..TransactionStatusMeta::default()
|
|
|
|
}
|
|
|
|
);
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_consume_buffered_packets() {
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let (transactions, bank, poh_recorder, _entry_receiver, poh_simulator) =
|
|
|
|
setup_conflicting_transactions(ledger_path.path());
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.read().unwrap().new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let num_conflicting_transactions = transactions.len();
|
2023-05-09 14:39:23 -07:00
|
|
|
let deserialized_packets = transactions_to_deserialized_packets(&transactions).unwrap();
|
2023-02-15 08:52:13 -08:00
|
|
|
assert_eq!(deserialized_packets.len(), num_conflicting_transactions);
|
|
|
|
let mut buffered_packet_batches =
|
|
|
|
UnprocessedTransactionStorage::new_transaction_storage(
|
|
|
|
UnprocessedPacketBatches::from_iter(
|
|
|
|
deserialized_packets.into_iter(),
|
|
|
|
num_conflicting_transactions,
|
|
|
|
),
|
|
|
|
ThreadType::Transactions,
|
|
|
|
);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// When the working bank in poh_recorder is None, no packets should be processed (consume will not be called)
|
|
|
|
assert!(!poh_recorder.read().unwrap().has_bank());
|
|
|
|
assert_eq!(buffered_packet_batches.len(), num_conflicting_transactions);
|
|
|
|
// When the working bank in poh_recorder is Some, all packets should be processed.
|
|
|
|
// Multi-Iterator will process them 1-by-1 if all txs are conflicting.
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let bank_start = poh_recorder.read().unwrap().bank_start().unwrap();
|
2023-03-09 09:13:04 -08:00
|
|
|
let banking_stage_stats = BankingStageStats::default();
|
2023-03-06 09:13:28 -08:00
|
|
|
consumer.consume_buffered_packets(
|
2023-02-15 08:52:13 -08:00
|
|
|
&bank_start,
|
|
|
|
&mut buffered_packet_batches,
|
2023-03-09 09:13:04 -08:00
|
|
|
&banking_stage_stats,
|
2023-02-15 08:52:13 -08:00
|
|
|
&mut LeaderSlotMetricsTracker::new(0),
|
|
|
|
);
|
2023-03-09 09:13:04 -08:00
|
|
|
|
|
|
|
// Check that all packets were processed without retrying
|
2023-02-15 08:52:13 -08:00
|
|
|
assert!(buffered_packet_batches.is_empty());
|
2023-03-09 09:13:04 -08:00
|
|
|
assert_eq!(
|
|
|
|
banking_stage_stats
|
|
|
|
.consumed_buffered_packets_count
|
|
|
|
.load(Ordering::Relaxed),
|
|
|
|
num_conflicting_transactions
|
|
|
|
);
|
|
|
|
assert_eq!(
|
|
|
|
banking_stage_stats
|
|
|
|
.rebuffered_packets_count
|
|
|
|
.load(Ordering::Relaxed),
|
|
|
|
0
|
|
|
|
);
|
|
|
|
// Use bank to check the number of entries (batches)
|
|
|
|
assert_eq!(bank_start.working_bank.transactions_per_entry_max(), 1);
|
|
|
|
assert_eq!(
|
|
|
|
bank_start.working_bank.transaction_entries_count(),
|
|
|
|
num_conflicting_transactions as u64
|
|
|
|
);
|
|
|
|
|
2023-02-15 08:52:13 -08:00
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_consume_buffered_packets_sanitization_error() {
|
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let (mut transactions, bank, poh_recorder, _entry_receiver, poh_simulator) =
|
|
|
|
setup_conflicting_transactions(ledger_path.path());
|
|
|
|
let duplicate_account_key = transactions[0].message.account_keys[0];
|
|
|
|
transactions[0]
|
|
|
|
.message
|
|
|
|
.account_keys
|
|
|
|
.push(duplicate_account_key); // corrupt transaction
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.read().unwrap().new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let num_conflicting_transactions = transactions.len();
|
2023-05-09 14:39:23 -07:00
|
|
|
let deserialized_packets = transactions_to_deserialized_packets(&transactions).unwrap();
|
2023-02-15 08:52:13 -08:00
|
|
|
assert_eq!(deserialized_packets.len(), num_conflicting_transactions);
|
|
|
|
let mut buffered_packet_batches =
|
|
|
|
UnprocessedTransactionStorage::new_transaction_storage(
|
|
|
|
UnprocessedPacketBatches::from_iter(
|
|
|
|
deserialized_packets.into_iter(),
|
|
|
|
num_conflicting_transactions,
|
|
|
|
),
|
|
|
|
ThreadType::Transactions,
|
|
|
|
);
|
|
|
|
|
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
|
|
|
// When the working bank in poh_recorder is None, no packets should be processed
|
|
|
|
assert!(!poh_recorder.read().unwrap().has_bank());
|
|
|
|
assert_eq!(buffered_packet_batches.len(), num_conflicting_transactions);
|
|
|
|
// When the working bank in poh_recorder is Some, all packets should be processed.
|
|
|
|
// Multi-Iterator will process them 1-by-1 if all txs are conflicting.
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let bank_start = poh_recorder.read().unwrap().bank_start().unwrap();
|
2023-03-06 09:13:28 -08:00
|
|
|
consumer.consume_buffered_packets(
|
2023-02-15 08:52:13 -08:00
|
|
|
&bank_start,
|
|
|
|
&mut buffered_packet_batches,
|
|
|
|
&BankingStageStats::default(),
|
|
|
|
&mut LeaderSlotMetricsTracker::new(0),
|
|
|
|
);
|
|
|
|
assert!(buffered_packet_batches.is_empty());
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
2023-03-09 09:13:04 -08:00
|
|
|
fn test_consume_buffered_packets_retryable() {
|
2023-02-15 08:52:13 -08:00
|
|
|
let ledger_path = get_tmp_ledger_path_auto_delete!();
|
|
|
|
{
|
|
|
|
let (transactions, bank, poh_recorder, _entry_receiver, poh_simulator) =
|
|
|
|
setup_conflicting_transactions(ledger_path.path());
|
2023-03-21 21:19:20 -07:00
|
|
|
let recorder = poh_recorder.read().unwrap().new_recorder();
|
2023-02-15 08:52:13 -08:00
|
|
|
let num_conflicting_transactions = transactions.len();
|
2023-05-09 14:39:23 -07:00
|
|
|
let deserialized_packets = transactions_to_deserialized_packets(&transactions).unwrap();
|
2023-03-09 09:13:04 -08:00
|
|
|
assert_eq!(deserialized_packets.len(), num_conflicting_transactions);
|
|
|
|
let retryable_packet = deserialized_packets[0].clone();
|
|
|
|
let mut buffered_packet_batches =
|
|
|
|
UnprocessedTransactionStorage::new_transaction_storage(
|
|
|
|
UnprocessedPacketBatches::from_iter(
|
|
|
|
deserialized_packets.into_iter(),
|
|
|
|
num_conflicting_transactions,
|
|
|
|
),
|
|
|
|
ThreadType::Transactions,
|
|
|
|
);
|
|
|
|
|
2023-02-15 08:52:13 -08:00
|
|
|
let (replay_vote_sender, _replay_vote_receiver) = unbounded();
|
2023-03-23 17:05:54 -07:00
|
|
|
let committer = Committer::new(
|
|
|
|
None,
|
|
|
|
replay_vote_sender,
|
|
|
|
Arc::new(PrioritizationFeeCache::new(0u64)),
|
|
|
|
);
|
2023-03-10 09:30:25 -08:00
|
|
|
let consumer = Consumer::new(committer, recorder, QosService::new(1), None);
|
2023-02-15 08:52:13 -08:00
|
|
|
|
2023-03-09 09:13:04 -08:00
|
|
|
// When the working bank in poh_recorder is None, no packets should be processed (consume will not be called)
|
|
|
|
assert!(!poh_recorder.read().unwrap().has_bank());
|
|
|
|
assert_eq!(buffered_packet_batches.len(), num_conflicting_transactions);
|
|
|
|
// When the working bank in poh_recorder is Some, all packets should be processed
|
|
|
|
// except except for retryable errors. Manually take the lock of a transaction to
|
|
|
|
// simulate another thread processing a transaction with that lock.
|
|
|
|
poh_recorder.write().unwrap().set_bank(&bank, false);
|
|
|
|
let bank_start = poh_recorder.read().unwrap().bank_start().unwrap();
|
|
|
|
|
|
|
|
let lock_account = transactions[0].message.account_keys[1];
|
|
|
|
let manual_lock_tx =
|
|
|
|
SanitizedTransaction::from_transaction_for_tests(system_transaction::transfer(
|
|
|
|
&Keypair::new(),
|
|
|
|
&lock_account,
|
|
|
|
1,
|
|
|
|
bank.last_blockhash(),
|
|
|
|
));
|
|
|
|
let _ = bank_start.working_bank.accounts().lock_accounts(
|
|
|
|
std::iter::once(&manual_lock_tx),
|
|
|
|
bank_start.working_bank.get_transaction_account_lock_limit(),
|
|
|
|
);
|
|
|
|
|
|
|
|
let banking_stage_stats = BankingStageStats::default();
|
|
|
|
consumer.consume_buffered_packets(
|
|
|
|
&bank_start,
|
|
|
|
&mut buffered_packet_batches,
|
|
|
|
&banking_stage_stats,
|
|
|
|
&mut LeaderSlotMetricsTracker::new(0),
|
|
|
|
);
|
|
|
|
|
|
|
|
// Check that all but 1 transaction was processed. And that it was rebuffered.
|
|
|
|
assert_eq!(buffered_packet_batches.len(), 1);
|
|
|
|
assert_eq!(
|
|
|
|
buffered_packet_batches.iter().next().unwrap(),
|
|
|
|
&retryable_packet
|
|
|
|
);
|
|
|
|
assert_eq!(
|
|
|
|
banking_stage_stats
|
|
|
|
.consumed_buffered_packets_count
|
|
|
|
.load(Ordering::Relaxed),
|
|
|
|
num_conflicting_transactions - 1,
|
|
|
|
);
|
|
|
|
assert_eq!(
|
|
|
|
banking_stage_stats
|
|
|
|
.rebuffered_packets_count
|
|
|
|
.load(Ordering::Relaxed),
|
|
|
|
1
|
|
|
|
);
|
|
|
|
// Use bank to check the number of entries (batches)
|
|
|
|
assert_eq!(bank_start.working_bank.transactions_per_entry_max(), 1);
|
|
|
|
assert_eq!(
|
|
|
|
bank_start.working_bank.transaction_entries_count(),
|
|
|
|
num_conflicting_transactions as u64 - 1
|
|
|
|
);
|
|
|
|
|
|
|
|
poh_recorder
|
|
|
|
.read()
|
|
|
|
.unwrap()
|
|
|
|
.is_exited
|
|
|
|
.store(true, Ordering::Relaxed);
|
2023-02-15 08:52:13 -08:00
|
|
|
let _ = poh_simulator.join();
|
|
|
|
}
|
|
|
|
Blockstore::destroy(ledger_path.path()).unwrap();
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_accumulate_execute_units_and_time() {
|
|
|
|
let mut execute_timings = ExecuteTimings::default();
|
|
|
|
let mut expected_units = 0;
|
|
|
|
let mut expected_us = 0;
|
|
|
|
|
|
|
|
for n in 0..10 {
|
|
|
|
execute_timings.details.per_program_timings.insert(
|
|
|
|
Pubkey::new_unique(),
|
|
|
|
ProgramTiming {
|
|
|
|
accumulated_us: n * 100,
|
|
|
|
accumulated_units: n * 1000,
|
|
|
|
count: n as u32,
|
|
|
|
errored_txs_compute_consumed: vec![],
|
|
|
|
total_errored_units: 0,
|
|
|
|
},
|
|
|
|
);
|
|
|
|
expected_us += n * 100;
|
|
|
|
expected_units += n * 1000;
|
|
|
|
}
|
|
|
|
|
|
|
|
let (units, us) = Consumer::accumulate_execute_units_and_time(&execute_timings);
|
|
|
|
|
|
|
|
assert_eq!(expected_units, units);
|
|
|
|
assert_eq!(expected_us, us);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_bank_prepare_filter_for_pending_transaction() {
|
|
|
|
assert_eq!(
|
|
|
|
Consumer::prepare_filter_for_pending_transactions(6, &[2, 4, 5]),
|
|
|
|
vec![
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
Ok(()),
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
Ok(()),
|
|
|
|
Ok(())
|
|
|
|
]
|
|
|
|
);
|
|
|
|
|
|
|
|
assert_eq!(
|
|
|
|
Consumer::prepare_filter_for_pending_transactions(6, &[0, 2, 3]),
|
|
|
|
vec![
|
|
|
|
Ok(()),
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
Ok(()),
|
|
|
|
Ok(()),
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
Err(TransactionError::BlockhashNotFound),
|
|
|
|
]
|
|
|
|
);
|
|
|
|
}
|
|
|
|
|
|
|
|
#[test]
|
|
|
|
fn test_bank_filter_valid_transaction_indexes() {
|
|
|
|
assert_eq!(
|
|
|
|
Consumer::filter_valid_transaction_indexes(&[
|
|
|
|
(Err(TransactionError::BlockhashNotFound), None),
|
|
|
|
(Err(TransactionError::BlockhashNotFound), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
(Err(TransactionError::BlockhashNotFound), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
]),
|
|
|
|
[2, 4, 5]
|
|
|
|
);
|
|
|
|
|
|
|
|
assert_eq!(
|
|
|
|
Consumer::filter_valid_transaction_indexes(&[
|
|
|
|
(Ok(()), None),
|
|
|
|
(Err(TransactionError::BlockhashNotFound), None),
|
|
|
|
(Err(TransactionError::BlockhashNotFound), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
(Ok(()), None),
|
|
|
|
]),
|
|
|
|
[0, 3, 4, 5]
|
|
|
|
);
|
|
|
|
}
|
|
|
|
}
|