2018-05-23 23:29:01 -07:00
|
|
|
//! The `banking_stage` processes Transaction messages.
|
2018-05-14 16:31:13 -07:00
|
|
|
|
|
|
|
use bank::Bank;
|
|
|
|
use bincode::deserialize;
|
|
|
|
use packet;
|
|
|
|
use packet::SharedPackets;
|
|
|
|
use rayon::prelude::*;
|
2018-05-23 14:49:59 -07:00
|
|
|
use record_stage::Signal;
|
2018-05-14 16:31:13 -07:00
|
|
|
use result::Result;
|
|
|
|
use std::net::SocketAddr;
|
|
|
|
use std::sync::atomic::{AtomicBool, Ordering};
|
|
|
|
use std::sync::mpsc::{channel, Receiver, Sender};
|
2018-05-29 20:20:28 -07:00
|
|
|
use std::sync::Arc;
|
2018-05-30 13:38:15 -07:00
|
|
|
use std::thread::{Builder, JoinHandle};
|
2018-05-14 16:31:13 -07:00
|
|
|
use std::time::Duration;
|
|
|
|
use std::time::Instant;
|
|
|
|
use timing;
|
2018-05-23 23:29:01 -07:00
|
|
|
use transaction::Transaction;
|
2018-05-14 16:31:13 -07:00
|
|
|
|
|
|
|
pub struct BankingStage {
|
|
|
|
pub thread_hdl: JoinHandle<()>,
|
|
|
|
pub signal_receiver: Receiver<Signal>,
|
|
|
|
}
|
|
|
|
|
|
|
|
impl BankingStage {
|
|
|
|
pub fn new(
|
|
|
|
bank: Arc<Bank>,
|
|
|
|
exit: Arc<AtomicBool>,
|
|
|
|
verified_receiver: Receiver<Vec<(SharedPackets, Vec<u8>)>>,
|
|
|
|
packet_recycler: packet::PacketRecycler,
|
|
|
|
) -> Self {
|
|
|
|
let (signal_sender, signal_receiver) = channel();
|
2018-05-30 13:38:15 -07:00
|
|
|
let thread_hdl = Builder::new()
|
|
|
|
.name("solana-banking-stage".to_string())
|
|
|
|
.spawn(move || loop {
|
|
|
|
let e = Self::process_packets(
|
|
|
|
bank.clone(),
|
|
|
|
&verified_receiver,
|
|
|
|
&signal_sender,
|
|
|
|
&packet_recycler,
|
|
|
|
);
|
|
|
|
if e.is_err() {
|
|
|
|
if exit.load(Ordering::Relaxed) {
|
|
|
|
break;
|
|
|
|
}
|
2018-05-14 16:31:13 -07:00
|
|
|
}
|
2018-05-30 13:38:15 -07:00
|
|
|
})
|
|
|
|
.unwrap();
|
2018-05-14 16:31:13 -07:00
|
|
|
BankingStage {
|
|
|
|
thread_hdl,
|
|
|
|
signal_receiver,
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-05-25 14:51:41 -07:00
|
|
|
fn deserialize_transactions(p: &packet::Packets) -> Vec<Option<(Transaction, SocketAddr)>> {
|
2018-05-14 16:31:13 -07:00
|
|
|
p.packets
|
|
|
|
.par_iter()
|
|
|
|
.map(|x| {
|
|
|
|
deserialize(&x.data[0..x.meta.size])
|
|
|
|
.map(|req| (req, x.meta.addr()))
|
|
|
|
.ok()
|
|
|
|
})
|
|
|
|
.collect()
|
|
|
|
}
|
|
|
|
|
|
|
|
fn process_packets(
|
|
|
|
bank: Arc<Bank>,
|
|
|
|
verified_receiver: &Receiver<Vec<(SharedPackets, Vec<u8>)>>,
|
|
|
|
signal_sender: &Sender<Signal>,
|
|
|
|
packet_recycler: &packet::PacketRecycler,
|
|
|
|
) -> Result<()> {
|
|
|
|
let timer = Duration::new(1, 0);
|
|
|
|
let recv_start = Instant::now();
|
|
|
|
let mms = verified_receiver.recv_timeout(timer)?;
|
|
|
|
let mut reqs_len = 0;
|
|
|
|
let mms_len = mms.len();
|
|
|
|
info!(
|
|
|
|
"@{:?} process start stalled for: {:?}ms batches: {}",
|
|
|
|
timing::timestamp(),
|
|
|
|
timing::duration_as_ms(&recv_start.elapsed()),
|
|
|
|
mms.len(),
|
|
|
|
);
|
|
|
|
let proc_start = Instant::now();
|
|
|
|
for (msgs, vers) in mms {
|
2018-05-25 14:51:41 -07:00
|
|
|
let transactions = Self::deserialize_transactions(&msgs.read().unwrap());
|
|
|
|
reqs_len += transactions.len();
|
|
|
|
let transactions = transactions
|
2018-05-14 16:31:13 -07:00
|
|
|
.into_iter()
|
|
|
|
.zip(vers)
|
2018-05-25 14:51:41 -07:00
|
|
|
.filter_map(|(tx, ver)| match tx {
|
2018-05-14 16:31:13 -07:00
|
|
|
None => None,
|
2018-05-25 14:51:41 -07:00
|
|
|
Some((tx, _addr)) => if tx.verify_plan() && ver != 0 {
|
|
|
|
Some(tx)
|
2018-05-14 16:31:13 -07:00
|
|
|
} else {
|
|
|
|
None
|
|
|
|
},
|
|
|
|
})
|
|
|
|
.collect();
|
|
|
|
|
2018-05-25 14:51:41 -07:00
|
|
|
debug!("process_transactions");
|
2018-05-29 08:52:40 -07:00
|
|
|
let results = bank.process_transactions(transactions);
|
2018-05-25 14:51:41 -07:00
|
|
|
let transactions = results.into_iter().filter_map(|x| x.ok()).collect();
|
|
|
|
signal_sender.send(Signal::Events(transactions))?;
|
|
|
|
debug!("done process_transactions");
|
2018-05-14 16:31:13 -07:00
|
|
|
|
|
|
|
packet_recycler.recycle(msgs);
|
|
|
|
}
|
|
|
|
let total_time_s = timing::duration_as_s(&proc_start.elapsed());
|
|
|
|
let total_time_ms = timing::duration_as_ms(&proc_start.elapsed());
|
|
|
|
info!(
|
2018-05-25 14:51:41 -07:00
|
|
|
"@{:?} done processing transaction batches: {} time: {:?}ms reqs: {} reqs/s: {}",
|
2018-05-14 16:31:13 -07:00
|
|
|
timing::timestamp(),
|
|
|
|
mms_len,
|
|
|
|
total_time_ms,
|
|
|
|
reqs_len,
|
|
|
|
(reqs_len as f32) / (total_time_s)
|
|
|
|
);
|
|
|
|
Ok(())
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// TODO: When banking is pulled out of RequestStage, add this test back in.
|
|
|
|
|
|
|
|
//use bank::Bank;
|
|
|
|
//use entry::Entry;
|
|
|
|
//use hash::Hash;
|
|
|
|
//use record_stage::RecordStage;
|
2018-05-23 14:49:59 -07:00
|
|
|
//use record_stage::Signal;
|
2018-05-14 16:31:13 -07:00
|
|
|
//use result::Result;
|
|
|
|
//use std::sync::mpsc::{channel, Sender};
|
|
|
|
//use std::sync::{Arc, Mutex};
|
|
|
|
//use std::time::Duration;
|
2018-05-23 23:29:01 -07:00
|
|
|
//use transaction::Transaction;
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
//#[cfg(test)]
|
|
|
|
//mod tests {
|
|
|
|
// use bank::Bank;
|
|
|
|
// use mint::Mint;
|
|
|
|
// use signature::{KeyPair, KeyPairUtil};
|
|
|
|
// use transaction::Transaction;
|
|
|
|
//
|
|
|
|
// #[test]
|
2018-05-25 14:51:41 -07:00
|
|
|
// // TODO: Move this test banking_stage. Calling process_transactions() directly
|
2018-05-14 16:31:13 -07:00
|
|
|
// // defeats the purpose of this test.
|
|
|
|
// fn test_banking_sequential_consistency() {
|
|
|
|
// // In this attack we'll demonstrate that a verifier can interpret the ledger
|
|
|
|
// // differently if either the server doesn't signal the ledger to add an
|
|
|
|
// // Entry OR if the verifier tries to parallelize across multiple Entries.
|
|
|
|
// let mint = Mint::new(2);
|
|
|
|
// let bank = Bank::new(&mint);
|
2018-05-25 14:51:41 -07:00
|
|
|
// let banking_stage = EventProcessor::new(bank, &mint.last_id(), None);
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// // Process a batch that includes a transaction that receives two tokens.
|
|
|
|
// let alice = KeyPair::new();
|
2018-05-25 15:05:37 -07:00
|
|
|
// let tx = Transaction::new(&mint.keypair(), alice.pubkey(), 2, mint.last_id());
|
|
|
|
// let transactions = vec![tx];
|
2018-05-25 14:51:41 -07:00
|
|
|
// let entry0 = banking_stage.process_transactions(transactions).unwrap();
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// // Process a second batch that spends one of those tokens.
|
2018-05-25 15:05:37 -07:00
|
|
|
// let tx = Transaction::new(&alice, mint.pubkey(), 1, mint.last_id());
|
|
|
|
// let transactions = vec![tx];
|
2018-05-25 14:51:41 -07:00
|
|
|
// let entry1 = banking_stage.process_transactions(transactions).unwrap();
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// // Collect the ledger and feed it to a new bank.
|
|
|
|
// let entries = vec![entry0, entry1];
|
|
|
|
//
|
|
|
|
// // Assert the user holds one token, not two. If the server only output one
|
|
|
|
// // entry, then the second transaction will be rejected, because it drives
|
|
|
|
// // the account balance below zero before the credit is added.
|
|
|
|
// let bank = Bank::new(&mint);
|
|
|
|
// for entry in entries {
|
|
|
|
// assert!(
|
|
|
|
// bank
|
2018-05-29 08:52:40 -07:00
|
|
|
// .process_transactions(entry.transactions)
|
2018-05-14 16:31:13 -07:00
|
|
|
// .into_iter()
|
|
|
|
// .all(|x| x.is_ok())
|
|
|
|
// );
|
|
|
|
// }
|
|
|
|
// assert_eq!(bank.get_balance(&alice.pubkey()), Some(1));
|
|
|
|
// }
|
|
|
|
//}
|
|
|
|
//
|
|
|
|
//#[cfg(all(feature = "unstable", test))]
|
|
|
|
//mod bench {
|
|
|
|
// extern crate test;
|
|
|
|
// use self::test::Bencher;
|
|
|
|
// use bank::{Bank, MAX_ENTRY_IDS};
|
|
|
|
// use bincode::serialize;
|
|
|
|
// use hash::hash;
|
|
|
|
// use mint::Mint;
|
|
|
|
// use rayon::prelude::*;
|
|
|
|
// use signature::{KeyPair, KeyPairUtil};
|
|
|
|
// use std::collections::HashSet;
|
|
|
|
// use std::time::Instant;
|
|
|
|
// use transaction::Transaction;
|
|
|
|
//
|
|
|
|
// #[bench]
|
2018-05-25 19:56:28 -07:00
|
|
|
// fn bench_process_transactions(_bencher: &mut Bencher) {
|
2018-05-14 16:31:13 -07:00
|
|
|
// let mint = Mint::new(100_000_000);
|
|
|
|
// let bank = Bank::new(&mint);
|
|
|
|
// // Create transactions between unrelated parties.
|
|
|
|
// let txs = 100_000;
|
|
|
|
// let last_ids: Mutex<HashSet<Hash>> = Mutex::new(HashSet::new());
|
|
|
|
// let transactions: Vec<_> = (0..txs)
|
|
|
|
// .into_par_iter()
|
|
|
|
// .map(|i| {
|
|
|
|
// // Seed the 'to' account and a cell for its signature.
|
|
|
|
// let dummy_id = i % (MAX_ENTRY_IDS as i32);
|
|
|
|
// let last_id = hash(&serialize(&dummy_id).unwrap()); // Semi-unique hash
|
|
|
|
// {
|
|
|
|
// let mut last_ids = last_ids.lock().unwrap();
|
|
|
|
// if !last_ids.contains(&last_id) {
|
|
|
|
// last_ids.insert(last_id);
|
|
|
|
// bank.register_entry_id(&last_id);
|
|
|
|
// }
|
|
|
|
// }
|
|
|
|
//
|
|
|
|
// // Seed the 'from' account.
|
|
|
|
// let rando0 = KeyPair::new();
|
2018-05-25 15:05:37 -07:00
|
|
|
// let tx = Transaction::new(&mint.keypair(), rando0.pubkey(), 1_000, last_id);
|
2018-05-29 08:52:40 -07:00
|
|
|
// bank.process_transaction(&tx).unwrap();
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// let rando1 = KeyPair::new();
|
2018-05-25 15:05:37 -07:00
|
|
|
// let tx = Transaction::new(&rando0, rando1.pubkey(), 2, last_id);
|
2018-05-29 08:52:40 -07:00
|
|
|
// bank.process_transaction(&tx).unwrap();
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// // Finally, return a transaction that's unique
|
|
|
|
// Transaction::new(&rando0, rando1.pubkey(), 1, last_id)
|
|
|
|
// })
|
|
|
|
// .collect();
|
|
|
|
//
|
2018-05-25 14:51:41 -07:00
|
|
|
// let banking_stage = EventProcessor::new(bank, &mint.last_id(), None);
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// let now = Instant::now();
|
2018-05-25 14:51:41 -07:00
|
|
|
// assert!(banking_stage.process_transactions(transactions).is_ok());
|
2018-05-14 16:31:13 -07:00
|
|
|
// let duration = now.elapsed();
|
|
|
|
// let sec = duration.as_secs() as f64 + duration.subsec_nanos() as f64 / 1_000_000_000.0;
|
|
|
|
// let tps = txs as f64 / sec;
|
|
|
|
//
|
|
|
|
// // Ensure that all transactions were successfully logged.
|
2018-05-25 14:51:41 -07:00
|
|
|
// drop(banking_stage.historian_input);
|
|
|
|
// let entries: Vec<Entry> = banking_stage.output.lock().unwrap().iter().collect();
|
2018-05-14 16:31:13 -07:00
|
|
|
// assert_eq!(entries.len(), 1);
|
2018-05-23 23:29:01 -07:00
|
|
|
// assert_eq!(entries[0].transactions.len(), txs as usize);
|
2018-05-14 16:31:13 -07:00
|
|
|
//
|
|
|
|
// println!("{} tps", tps);
|
|
|
|
// }
|
|
|
|
//}
|
2018-05-16 15:33:34 -07:00
|
|
|
|
|
|
|
#[cfg(all(feature = "unstable", test))]
|
|
|
|
mod bench {
|
|
|
|
extern crate test;
|
|
|
|
use self::test::Bencher;
|
|
|
|
use bank::*;
|
|
|
|
use banking_stage::BankingStage;
|
|
|
|
use mint::Mint;
|
|
|
|
use packet::{to_packets, PacketRecycler};
|
2018-05-23 14:49:59 -07:00
|
|
|
use record_stage::Signal;
|
2018-05-16 15:33:34 -07:00
|
|
|
use signature::{KeyPair, KeyPairUtil};
|
|
|
|
use std::iter;
|
|
|
|
use std::sync::mpsc::channel;
|
2018-05-29 20:20:28 -07:00
|
|
|
use std::sync::Arc;
|
2018-05-23 23:29:01 -07:00
|
|
|
use transaction::Transaction;
|
2018-05-16 15:33:34 -07:00
|
|
|
|
|
|
|
#[bench]
|
2018-05-25 19:56:28 -07:00
|
|
|
fn bench_stage(bencher: &mut Bencher) {
|
2018-05-16 15:33:34 -07:00
|
|
|
let tx = 100_usize;
|
|
|
|
let mint = Mint::new(1_000_000_000);
|
|
|
|
let pubkey = KeyPair::new().pubkey();
|
|
|
|
|
2018-05-25 14:51:41 -07:00
|
|
|
let transactions: Vec<_> = (0..tx)
|
2018-05-23 23:29:01 -07:00
|
|
|
.map(|i| Transaction::new(&mint.keypair(), pubkey, i as i64, mint.last_id()))
|
2018-05-16 15:33:34 -07:00
|
|
|
.collect();
|
|
|
|
|
|
|
|
let (verified_sender, verified_receiver) = channel();
|
|
|
|
let (signal_sender, signal_receiver) = channel();
|
|
|
|
let packet_recycler = PacketRecycler::default();
|
2018-05-25 14:51:41 -07:00
|
|
|
let verified: Vec<_> = to_packets(&packet_recycler, transactions)
|
2018-05-16 15:33:34 -07:00
|
|
|
.into_iter()
|
|
|
|
.map(|x| {
|
|
|
|
let len = (*x).read().unwrap().packets.len();
|
|
|
|
(x, iter::repeat(1).take(len).collect())
|
|
|
|
})
|
|
|
|
.collect();
|
|
|
|
|
|
|
|
bencher.iter(move || {
|
|
|
|
let bank = Arc::new(Bank::new(&mint));
|
|
|
|
verified_sender.send(verified.clone()).unwrap();
|
|
|
|
BankingStage::process_packets(
|
|
|
|
bank.clone(),
|
|
|
|
&verified_receiver,
|
|
|
|
&signal_sender,
|
|
|
|
&packet_recycler,
|
|
|
|
).unwrap();
|
|
|
|
let signal = signal_receiver.recv().unwrap();
|
2018-06-04 16:17:23 -07:00
|
|
|
if let Signal::Events(transactions) = signal {
|
2018-05-25 14:51:41 -07:00
|
|
|
assert_eq!(transactions.len(), tx);
|
2018-05-16 15:33:34 -07:00
|
|
|
} else {
|
|
|
|
assert!(false);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
}
|
|
|
|
}
|