solana/src/entry.rs

290 lines
9.9 KiB
Rust
Raw Normal View History

2018-03-30 10:43:38 -07:00
//! The `entry` module is a fundamental building block of Proof of History. It contains a
//! unique ID that is the hash of the Entry before it, plus the hash of the
//! transactions within it. Entries cannot be reordered, and its field `num_hashes`
2018-03-29 11:20:54 -07:00
//! represents an approximate amount of time since the last Entry was created.
2018-08-06 12:35:38 -07:00
use bincode::{serialize_into, serialized_size};
use hash::{extend_and_hash, hash, Hash};
2018-08-06 12:35:38 -07:00
use packet::{BlobRecycler, SharedBlob, BLOB_DATA_SIZE};
use rayon::prelude::*;
2018-08-06 12:35:38 -07:00
use signature::PublicKey;
use std::io::Cursor;
2018-05-23 23:29:01 -07:00
use transaction::Transaction;
2018-03-29 11:20:54 -07:00
/// Each Entry contains three pieces of data. The `num_hashes` field is the number
/// of hashes performed since the previous entry. The `id` field is the result
2018-05-25 14:51:41 -07:00
/// of hashing `id` from the previous entry `num_hashes` times. The `transactions`
2018-06-06 10:19:56 -07:00
/// field points to Transactions that took place shortly before `id` was generated.
2018-03-29 11:20:54 -07:00
///
/// If you divide `num_hashes` by the amount of time it takes to generate a new hash, you
/// get a duration estimate since the last Entry. Since processing power increases
/// over time, one should expect the duration `num_hashes` represents to decrease proportionally.
2018-06-06 10:19:56 -07:00
/// An upper bound on Duration can be estimated by assuming each hash was generated by the
/// world's fastest processor at the time the entry was recorded. Or said another way, it
/// is physically not possible for a shorter duration to have occurred if one assumes the
/// hash was computed by the world's fastest processor at that time. The hash chain is both
2018-08-06 23:36:09 -07:00
/// a Verifiable Delay Function (VDF) and a Proof of Work (not to be confused with Proof of
2018-06-06 10:19:56 -07:00
/// Work consensus!)
#[derive(Serialize, Deserialize, Debug, PartialEq, Eq, Clone)]
pub struct Entry {
2018-06-06 10:19:56 -07:00
/// The number of hashes since the previous Entry ID.
pub num_hashes: u64,
2018-06-06 10:19:56 -07:00
/// The SHA-256 hash `num_hashes` after the previous Entry ID.
pub id: Hash,
2018-06-06 10:19:56 -07:00
/// An unordered list of transactions that were observed before the Entry ID was
2018-08-06 23:55:00 -07:00
/// generated. They may have been observed before a previous Entry ID but were
2018-06-06 10:19:56 -07:00
/// pushed back into this list to ensure deterministic interpretation of the ledger.
2018-05-25 14:51:41 -07:00
pub transactions: Vec<Transaction>,
/// Indication that:
/// 1. the next Entry in the ledger has transactions that can potentially
/// be verified in parallel with these transactions
/// 2. this Entry can be left out of the bank's entry_id cache for
/// purposes of duplicate rejection
pub has_more: bool,
2018-06-26 12:35:52 -07:00
/// Erasure requires that Entry be a multiple of 4 bytes in size
pad: [u8; 3],
}
impl Entry {
2018-05-16 16:49:58 -07:00
/// Creates the next Entry `num_hashes` after `start_hash`.
pub fn new(
start_hash: &Hash,
cur_hashes: u64,
transactions: Vec<Transaction>,
has_more: bool,
) -> Self {
2018-05-25 14:51:41 -07:00
let num_hashes = cur_hashes + if transactions.is_empty() { 0 } else { 1 };
let id = next_hash(start_hash, 0, &transactions);
let entry = Entry {
2018-05-16 16:49:58 -07:00
num_hashes,
id,
2018-05-25 14:51:41 -07:00
transactions,
has_more,
2018-06-26 12:35:52 -07:00
pad: [0, 0, 0],
};
2018-07-24 11:14:33 -07:00
let size = serialized_size(&entry).unwrap();
if size > BLOB_DATA_SIZE as u64 {
panic!(
"Serialized entry size too large: {} ({} transactions):",
size,
entry.transactions.len()
);
}
entry
2018-05-16 16:49:58 -07:00
}
2018-08-06 12:35:38 -07:00
pub fn to_blob(
&self,
blob_recycler: &BlobRecycler,
idx: Option<u64>,
id: Option<PublicKey>,
) -> SharedBlob {
let blob = blob_recycler.allocate();
{
let mut blob_w = blob.write().unwrap();
let pos = {
let mut out = Cursor::new(blob_w.data_mut());
serialize_into(&mut out, &self).expect("failed to serialize output");
out.position() as usize
};
blob_w.set_size(pos);
if let Some(idx) = idx {
blob_w.set_index(idx).expect("set_index()");
}
if let Some(id) = id {
blob_w.set_id(id).expect("set_id()");
}
}
blob
}
2018-06-26 12:35:52 -07:00
pub fn will_fit(transactions: Vec<Transaction>) -> bool {
serialized_size(&Entry {
num_hashes: 0,
id: Hash::default(),
transactions,
has_more: false,
pad: [0, 0, 0],
}).unwrap() <= BLOB_DATA_SIZE as u64
}
2018-05-16 16:49:58 -07:00
/// Creates the next Tick Entry `num_hashes` after `start_hash`.
2018-05-25 14:51:41 -07:00
pub fn new_mut(
start_hash: &mut Hash,
cur_hashes: &mut u64,
transactions: Vec<Transaction>,
has_more: bool,
2018-05-25 14:51:41 -07:00
) -> Self {
let entry = Self::new(start_hash, *cur_hashes, transactions, has_more);
2018-05-16 16:49:58 -07:00
*start_hash = entry.id;
*cur_hashes = 0;
assert!(serialized_size(&entry).unwrap() <= BLOB_DATA_SIZE as u64);
2018-05-16 16:49:58 -07:00
entry
}
2018-05-25 14:51:41 -07:00
/// Creates a Entry from the number of hashes `num_hashes` since the previous transaction
2018-03-22 13:40:28 -07:00
/// and that resulting `id`.
pub fn new_tick(num_hashes: u64, id: &Hash) -> Self {
Entry {
num_hashes,
id: *id,
2018-05-25 14:51:41 -07:00
transactions: vec![],
has_more: false,
2018-06-26 12:35:52 -07:00
pad: [0, 0, 0],
}
}
2018-03-22 13:40:28 -07:00
/// Verifies self.id is the result of hashing a `start_hash` `self.num_hashes` times.
2018-05-25 14:51:41 -07:00
/// If the transaction is not a Tick, then hash that as well.
pub fn verify(&self, start_hash: &Hash) -> bool {
2018-07-23 16:13:24 -07:00
let tx_plans_verified = self.transactions.par_iter().all(|tx| {
let r = tx.verify_plan();
if !r {
warn!("tx plan invalid: {:?}", tx);
2018-07-23 16:13:24 -07:00
}
r
});
if !tx_plans_verified {
return false;
}
let ref_hash = next_hash(start_hash, self.num_hashes, &self.transactions);
if self.id != ref_hash {
warn!(
2018-07-23 16:13:24 -07:00
"next_hash is invalid expected: {:?} actual: {:?}",
self.id, ref_hash
);
return false;
}
true
}
}
2018-05-25 15:05:37 -07:00
fn add_transaction_data(hash_data: &mut Vec<u8>, tx: &Transaction) {
2018-05-23 23:29:01 -07:00
hash_data.push(0u8);
hash_data.extend_from_slice(&tx.sig.as_ref());
}
2018-05-25 14:51:41 -07:00
/// Creates the hash `num_hashes` after `start_hash`. If the transaction contains
/// a signature, the final hash will be a hash of both the previous ID and
2018-06-18 13:23:15 -07:00
/// the signature. If num_hashes is zero and there's no transaction data,
2018-06-21 16:09:17 -07:00
/// start_hash is returned.
2018-06-18 13:23:15 -07:00
fn next_hash(start_hash: &Hash, num_hashes: u64, transactions: &[Transaction]) -> Hash {
let mut id = *start_hash;
for _ in 1..num_hashes {
2018-08-01 11:23:52 -07:00
id = hash(&id.as_ref());
}
2018-05-25 14:51:41 -07:00
// Hash all the transaction data
let mut hash_data = vec![];
2018-05-25 14:51:41 -07:00
for tx in transactions {
add_transaction_data(&mut hash_data, tx);
}
if !hash_data.is_empty() {
extend_and_hash(&id, &hash_data)
} else if num_hashes != 0 {
2018-08-01 11:23:52 -07:00
hash(&id.as_ref())
} else {
id
}
}
/// Creates the next Tick or Transaction Entry `num_hashes` after `start_hash`.
2018-05-25 14:51:41 -07:00
pub fn next_entry(start_hash: &Hash, num_hashes: u64, transactions: Vec<Transaction>) -> Entry {
2018-07-11 13:40:46 -07:00
assert!(num_hashes > 0 || transactions.is_empty());
Entry {
num_hashes,
2018-05-25 14:51:41 -07:00
id: next_hash(start_hash, num_hashes, &transactions),
transactions,
has_more: false,
2018-06-26 12:35:52 -07:00
pad: [0, 0, 0],
}
}
#[cfg(test)]
mod tests {
use super::*;
use chrono::prelude::*;
2018-05-16 16:49:58 -07:00
use entry::Entry;
2018-03-09 16:02:17 -08:00
use hash::hash;
use signature::{KeyPair, KeyPairUtil};
use transaction::Transaction;
#[test]
2018-03-09 16:02:17 -08:00
fn test_entry_verify() {
let zero = Hash::default();
2018-08-01 11:23:52 -07:00
let one = hash(&zero.as_ref());
assert!(Entry::new_tick(0, &zero).verify(&zero)); // base case
assert!(!Entry::new_tick(0, &zero).verify(&one)); // base case, bad
assert!(next_entry(&zero, 1, vec![]).verify(&zero)); // inductive step
assert!(!next_entry(&zero, 1, vec![]).verify(&one)); // inductive step, bad
}
2018-03-09 16:02:17 -08:00
#[test]
2018-05-25 14:51:41 -07:00
fn test_transaction_reorder_attack() {
2018-03-09 16:02:17 -08:00
let zero = Hash::default();
// First, verify entries
let keypair = KeyPair::new();
2018-05-29 09:12:27 -07:00
let tx0 = Transaction::new(&keypair, keypair.pubkey(), 0, zero);
let tx1 = Transaction::new(&keypair, keypair.pubkey(), 1, zero);
let mut e0 = Entry::new(&zero, 0, vec![tx0.clone(), tx1.clone()], false);
2018-03-09 16:02:17 -08:00
assert!(e0.verify(&zero));
2018-05-25 14:51:41 -07:00
// Next, swap two transactions and ensure verification fails.
2018-05-29 09:12:27 -07:00
e0.transactions[0] = tx1; // <-- attack
e0.transactions[1] = tx0;
2018-03-09 16:02:17 -08:00
assert!(!e0.verify(&zero));
}
#[test]
fn test_witness_reorder_attack() {
let zero = Hash::default();
// First, verify entries
let keypair = KeyPair::new();
2018-05-29 09:12:27 -07:00
let tx0 = Transaction::new_timestamp(&keypair, Utc::now(), zero);
let tx1 = Transaction::new_signature(&keypair, Default::default(), zero);
let mut e0 = Entry::new(&zero, 0, vec![tx0.clone(), tx1.clone()], false);
assert!(e0.verify(&zero));
2018-05-25 14:51:41 -07:00
// Next, swap two witness transactions and ensure verification fails.
2018-05-29 09:12:27 -07:00
e0.transactions[0] = tx1; // <-- attack
e0.transactions[1] = tx0;
assert!(!e0.verify(&zero));
}
#[test]
fn test_next_entry() {
let zero = Hash::default();
2018-05-11 09:34:46 -07:00
let tick = next_entry(&zero, 1, vec![]);
assert_eq!(tick.num_hashes, 1);
assert_ne!(tick.id, zero);
2018-06-18 13:23:15 -07:00
let tick = next_entry(&zero, 0, vec![]);
assert_eq!(tick.num_hashes, 0);
assert_eq!(tick.id, zero);
let keypair = KeyPair::new();
let tx0 = Transaction::new_timestamp(&keypair, Utc::now(), zero);
let entry0 = next_entry(&zero, 1, vec![tx0.clone()]);
assert_eq!(entry0.num_hashes, 1);
assert_eq!(entry0.id, next_hash(&zero, 1, &vec![tx0]));
}
#[test]
#[should_panic]
fn test_next_entry_panic() {
let zero = Hash::default();
let keypair = KeyPair::new();
let tx = Transaction::new(&keypair, keypair.pubkey(), 0, zero);
next_entry(&zero, 0, vec![tx]);
}
}