From 7283b4bfd079c53370a05ef5f0611dd3910b2a1a Mon Sep 17 00:00:00 2001 From: teor Date: Thu, 10 Mar 2022 09:34:50 +1000 Subject: [PATCH] 4. test(db): add large transaction tests (#3759) * refactor(test/block): rename large single transaction function ```sh fastmod single_transaction_block single_transaction_block_many_inputs ``` * rustfmt * test(block): add a test block with many transparent outputs * doc(db): explain why we can't just get the UTXOs right before they are deleted * refactor(db): split out a block data write method * refactor(block): add a height argument to new_outputs * test(db): add block and transaction round-trip tests Including large blocks and transactions. * test(db): fix large block serialization instability in the tests * doc(block): add TODOs for generating correct blocks * Make transparent output functions which take a height test-only * make sure generated blocks are actually over/under-sized * replace println!() with an error!() log --- zebra-chain/benches/block.rs | 8 +- zebra-chain/src/block/tests/generate.rs | 325 ++++++++++++++---- zebra-chain/src/block/tests/vectors.rs | 8 +- zebra-chain/src/transparent.rs | 12 +- zebra-chain/src/transparent/utxo.rs | 41 ++- zebra-consensus/src/block/tests.rs | 6 +- zebra-state/src/arbitrary.rs | 32 +- zebra-state/src/request.rs | 16 +- zebra-state/src/service/finalized_state.rs | 1 + .../service/finalized_state/zebra_db/block.rs | 47 ++- .../finalized_state/zebra_db/block/tests.rs | 1 + .../zebra_db/block/tests/vectors.rs | 147 ++++++++ 12 files changed, 542 insertions(+), 102 deletions(-) create mode 100644 zebra-state/src/service/finalized_state/zebra_db/block/tests/vectors.rs diff --git a/zebra-chain/benches/block.rs b/zebra-chain/benches/block.rs index 71194e1a6..916890d8f 100644 --- a/zebra-chain/benches/block.rs +++ b/zebra-chain/benches/block.rs @@ -7,7 +7,9 @@ use criterion::{criterion_group, criterion_main, BenchmarkId, Criterion}; use zebra_chain::{ block::{ - tests::generate::{large_multi_transaction_block, large_single_transaction_block}, + tests::generate::{ + large_multi_transaction_block, large_single_transaction_block_many_inputs, + }, Block, }, serialization::{ZcashDeserialize, ZcashSerialize}, @@ -26,8 +28,8 @@ fn block_serialization(c: &mut Criterion) { large_multi_transaction_block(), ), ( - "large_single_transaction_block", - large_single_transaction_block(), + "large_single_transaction_block_many_inputs", + large_single_transaction_block_many_inputs(), ), ]; diff --git a/zebra-chain/src/block/tests/generate.rs b/zebra-chain/src/block/tests/generate.rs index c01286e99..1f93fabc8 100644 --- a/zebra-chain/src/block/tests/generate.rs +++ b/zebra-chain/src/block/tests/generate.rs @@ -1,106 +1,207 @@ -//! Generate blockchain testing constructions +//! Generate large transparent blocks and transactions for testing. + use chrono::{DateTime, NaiveDateTime, Utc}; use std::sync::Arc; use crate::{ + block::{serialize::MAX_BLOCK_BYTES, Block, Header}, serialization::{ZcashDeserialize, ZcashSerialize}, transaction::{LockTime, Transaction}, transparent, }; -use super::super::{serialize::MAX_BLOCK_BYTES, Block, Header}; +/// The minimum size of the blocks produced by this module. +pub const MIN_LARGE_BLOCK_BYTES: u64 = MAX_BLOCK_BYTES - 100; -/// Generate a block header -pub fn block_header() -> Header { - Header::zcash_deserialize(&zebra_test::vectors::DUMMY_HEADER[..]).unwrap() +/// The maximum number of bytes used to serialize a CompactSize, +/// for the transaction, input, and output counts generated by this module. +pub const MAX_COMPACT_SIZE_BYTES: usize = 4; + +/// The number of bytes used to serialize a version 1 transaction header. +pub const TX_V1_HEADER_BYTES: usize = 4; + +/// Returns a generated block header, and its canonical serialized bytes. +pub fn block_header() -> (Header, Vec) { + // Some of the test vectors are in a non-canonical format, + // so we have to round-trip serialize them. + + let block_header = Header::zcash_deserialize(&zebra_test::vectors::DUMMY_HEADER[..]).unwrap(); + let block_header_bytes = block_header.zcash_serialize_to_vec().unwrap(); + + (block_header, block_header_bytes) } -/// Generate a block with multiple transactions just below limit +/// Returns a generated transparent transaction, and its canonical serialized bytes. +pub fn transaction() -> (Transaction, Vec) { + // Some of the test vectors are in a non-canonical format, + // so we have to round-trip serialize them. + + let transaction = Transaction::zcash_deserialize(&zebra_test::vectors::DUMMY_TX1[..]).unwrap(); + let transaction_bytes = transaction.zcash_serialize_to_vec().unwrap(); + + (transaction, transaction_bytes) +} + +/// Returns a generated transparent lock time, and its canonical serialized bytes. +pub fn lock_time() -> (LockTime, Vec) { + let lock_time = LockTime::Time(DateTime::::from_utc( + NaiveDateTime::from_timestamp(61, 0), + Utc, + )); + let lock_time_bytes = lock_time.zcash_serialize_to_vec().unwrap(); + + (lock_time, lock_time_bytes) +} + +/// Returns a generated transparent input, and its canonical serialized bytes. +pub fn input() -> (transparent::Input, Vec) { + // Some of the test vectors are in a non-canonical format, + // so we have to round-trip serialize them. + + let input = + transparent::Input::zcash_deserialize(&zebra_test::vectors::DUMMY_INPUT1[..]).unwrap(); + let input_bytes = input.zcash_serialize_to_vec().unwrap(); + + (input, input_bytes) +} + +/// Returns a generated transparent output, and its canonical serialized bytes. +pub fn output() -> (transparent::Output, Vec) { + // Some of the test vectors are in a non-canonical format, + // so we have to round-trip serialize them. + + let output = + transparent::Output::zcash_deserialize(&zebra_test::vectors::DUMMY_OUTPUT1[..]).unwrap(); + let output_bytes = output.zcash_serialize_to_vec().unwrap(); + + (output, output_bytes) +} + +/// Generate a block with multiple transparent transactions just below limit +/// +/// TODO: add a coinbase height to the returned block pub fn large_multi_transaction_block() -> Block { multi_transaction_block(false) } -/// Generate a block with one transaction and multiple inputs just below limit -pub fn large_single_transaction_block() -> Block { - single_transaction_block(false) +/// Generate a block with one transaction and multiple transparent inputs just below limit +/// +/// TODO: add a coinbase height to the returned block +/// make the returned block stable under round-trip serialization +pub fn large_single_transaction_block_many_inputs() -> Block { + single_transaction_block_many_inputs(false) } -/// Generate a block with multiple transactions just above limit +/// Generate a block with one transaction and multiple transparent outputs just below limit +/// +/// TODO: add a coinbase height to the returned block +/// make the returned block stable under round-trip serialization +pub fn large_single_transaction_block_many_outputs() -> Block { + single_transaction_block_many_outputs(false) +} + +/// Generate a block with multiple transparent transactions just above limit +/// +/// TODO: add a coinbase height to the returned block pub fn oversized_multi_transaction_block() -> Block { multi_transaction_block(true) } -/// Generate a block with one transaction and multiple inputs just above limit -pub fn oversized_single_transaction_block() -> Block { - single_transaction_block(true) +/// Generate a block with one transaction and multiple transparent inputs just above limit +/// +/// TODO: add a coinbase height to the returned block +/// make the returned block stable under round-trip serialization +pub fn oversized_single_transaction_block_many_inputs() -> Block { + single_transaction_block_many_inputs(true) } -// Implementation of block generation with multiple transactions +/// Generate a block with one transaction and multiple transparent outputs just above limit +/// +/// TODO: add a coinbase height to the returned block +/// make the returned block stable under round-trip serialization +pub fn oversized_single_transaction_block_many_outputs() -> Block { + single_transaction_block_many_outputs(true) +} + +/// Implementation of block generation with multiple transparent transactions fn multi_transaction_block(oversized: bool) -> Block { // A dummy transaction - let tx = Transaction::zcash_deserialize(&zebra_test::vectors::DUMMY_TX1[..]).unwrap(); + let (transaction, transaction_bytes) = transaction(); // A block header - let header = block_header(); + let (block_header, block_header_bytes) = block_header(); - // Serialize header - let mut data_header = Vec::new(); - header - .zcash_serialize(&mut data_header) - .expect("Block header should serialize"); - - // Calculate the number of transactions we need - let mut max_transactions_in_block = - (MAX_BLOCK_BYTES as usize - data_header.len()) / zebra_test::vectors::DUMMY_TX1[..].len(); + // Calculate the number of transactions we need, + // subtracting the bytes used to serialize the expected transaction count. + let mut max_transactions_in_block = (usize::try_from(MAX_BLOCK_BYTES).unwrap() + - block_header_bytes.len() + - MAX_COMPACT_SIZE_BYTES) + / transaction_bytes.len(); if oversized { max_transactions_in_block += 1; } // Create transactions to be just below or just above the limit - let transactions = std::iter::repeat(Arc::new(tx)) + let transactions = std::iter::repeat(Arc::new(transaction)) .take(max_transactions_in_block) .collect::>(); // Add the transactions into a block - Block { - header, + let block = Block { + header: block_header, transactions, - } + }; + + let serialized_len = block.zcash_serialize_to_vec().unwrap().len(); + assert_eq!( + oversized, + serialized_len > MAX_BLOCK_BYTES.try_into().unwrap(), + "block is over-sized if requested:\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MAX_BLOCK_BYTES: {},", + oversized, + serialized_len, + MAX_BLOCK_BYTES, + ); + assert!( + serialized_len > MIN_LARGE_BLOCK_BYTES.try_into().unwrap(), + "block is large\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MIN_LARGE_BLOCK_BYTES: {},", + oversized, + serialized_len, + MIN_LARGE_BLOCK_BYTES, + ); + + block } -// Implementation of block generation with one transaction and multiple inputs -fn single_transaction_block(oversized: bool) -> Block { +/// Implementation of block generation with one transaction and multiple transparent inputs +fn single_transaction_block_many_inputs(oversized: bool) -> Block { // Dummy input and output - let input = - transparent::Input::zcash_deserialize(&zebra_test::vectors::DUMMY_INPUT1[..]).unwrap(); - let output = - transparent::Output::zcash_deserialize(&zebra_test::vectors::DUMMY_OUTPUT1[..]).unwrap(); + let (input, input_bytes) = input(); + let (output, output_bytes) = output(); // A block header - let header = block_header(); + let (block_header, block_header_bytes) = block_header(); - // Serialize header - let mut data_header = Vec::new(); - header - .zcash_serialize(&mut data_header) - .expect("Block header should serialize"); + // A LockTime + let (lock_time, lock_time_bytes) = lock_time(); - // Serialize a LockTime - let lock_time = LockTime::Time(DateTime::::from_utc( - NaiveDateTime::from_timestamp(61, 0), - Utc, - )); - let mut data_locktime = Vec::new(); - lock_time - .zcash_serialize(&mut data_locktime) - .expect("LockTime should serialize"); - - // Calculate the number of inputs we need - let mut max_inputs_in_tx = (MAX_BLOCK_BYTES as usize - - data_header.len() - - zebra_test::vectors::DUMMY_OUTPUT1[..].len() - - data_locktime.len()) - / (zebra_test::vectors::DUMMY_INPUT1[..].len() - 1); + // Calculate the number of inputs we need, + // subtracting the bytes used to serialize the expected input count, + // transaction count, and output count. + let mut max_inputs_in_tx = (usize::try_from(MAX_BLOCK_BYTES).unwrap() + - block_header_bytes.len() + - 1 + - TX_V1_HEADER_BYTES + - lock_time_bytes.len() + - MAX_COMPACT_SIZE_BYTES + - 1 + - output_bytes.len()) + / input_bytes.len(); if oversized { max_inputs_in_tx += 1; @@ -125,8 +226,112 @@ fn single_transaction_block(oversized: bool) -> Block { // Put the big transaction into a block let transactions = vec![Arc::new(big_transaction)]; - Block { - header, + + let block = Block { + header: block_header, transactions, - } + }; + + let serialized_len = block.zcash_serialize_to_vec().unwrap().len(); + assert_eq!( + oversized, + serialized_len > MAX_BLOCK_BYTES.try_into().unwrap(), + "block is over-sized if requested:\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MAX_BLOCK_BYTES: {},", + oversized, + serialized_len, + MAX_BLOCK_BYTES, + ); + assert!( + serialized_len > MIN_LARGE_BLOCK_BYTES.try_into().unwrap(), + "block is large\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MIN_LARGE_BLOCK_BYTES: {},", + oversized, + serialized_len, + MIN_LARGE_BLOCK_BYTES, + ); + + block +} + +/// Implementation of block generation with one transaction and multiple transparent outputs +fn single_transaction_block_many_outputs(oversized: bool) -> Block { + // Dummy input and output + let (input, input_bytes) = input(); + let (output, output_bytes) = output(); + + // A block header + let (block_header, block_header_bytes) = block_header(); + + // A LockTime + let (lock_time, lock_time_bytes) = lock_time(); + + // Calculate the number of outputs we need, + // subtracting the bytes used to serialize the expected output count, + // transaction count, and input count. + let mut max_outputs_in_tx = (usize::try_from(MAX_BLOCK_BYTES).unwrap() + - block_header_bytes.len() + - 1 + - TX_V1_HEADER_BYTES + - lock_time_bytes.len() + - 1 + - input_bytes.len() + - MAX_COMPACT_SIZE_BYTES) + / output_bytes.len(); + + if oversized { + max_outputs_in_tx += 1; + } + + // 1 single input + let inputs = vec![input]; + + // Create outputs to be just below the limit + let outputs = std::iter::repeat(output) + .take(max_outputs_in_tx) + .collect::>(); + + // Create a big transaction + let big_transaction = Transaction::V1 { + inputs, + outputs, + lock_time, + }; + + // Put the big transaction into a block + let transactions = vec![Arc::new(big_transaction)]; + + let block = Block { + header: block_header, + transactions, + }; + + let serialized_len = block.zcash_serialize_to_vec().unwrap().len(); + assert_eq!( + oversized, + serialized_len > MAX_BLOCK_BYTES.try_into().unwrap(), + "block is over-sized if requested:\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MAX_BLOCK_BYTES: {},", + oversized, + serialized_len, + MAX_BLOCK_BYTES, + ); + assert!( + serialized_len > MIN_LARGE_BLOCK_BYTES.try_into().unwrap(), + "block is large\n\ + oversized: {},\n\ + serialized_len: {},\n\ + MIN_LARGE_BLOCK_BYTES: {},", + oversized, + serialized_len, + MIN_LARGE_BLOCK_BYTES, + ); + + block } diff --git a/zebra-chain/src/block/tests/vectors.rs b/zebra-chain/src/block/tests/vectors.rs index 6ab780548..524c3d3c3 100644 --- a/zebra-chain/src/block/tests/vectors.rs +++ b/zebra-chain/src/block/tests/vectors.rs @@ -39,7 +39,7 @@ fn blockheaderhash_debug() { fn blockheaderhash_from_blockheader() { zebra_test::init(); - let blockheader = generate::block_header(); + let (blockheader, _blockheader_bytes) = generate::block_header(); let hash = Hash::from(&blockheader); @@ -290,7 +290,7 @@ fn block_limits_single_tx() { // Test block limit with a big single transaction // Create a block just below the limit - let mut block = generate::large_single_transaction_block(); + let mut block = generate::large_single_transaction_block_many_inputs(); // Serialize the block let mut data = Vec::new(); @@ -305,7 +305,7 @@ fn block_limits_single_tx() { .expect("block should deserialize as we are just below limit"); // Add 1 more input to the transaction, limit will be reached - block = generate::oversized_single_transaction_block(); + block = generate::oversized_single_transaction_block_many_inputs(); let mut data = Vec::new(); block @@ -326,7 +326,7 @@ fn node_time_check( block_header_time: DateTime, now: DateTime, ) -> Result<(), BlockTimeError> { - let mut header = generate::block_header(); + let (mut header, _header_bytes) = generate::block_header(); header.time = block_header_time; // pass a zero height and hash - they are only used in the returned error header.time_is_valid_at(now, &Height(0), &Hash([0; 32])) diff --git a/zebra-chain/src/transparent.rs b/zebra-chain/src/transparent.rs index 891e384bc..4e998d021 100644 --- a/zebra-chain/src/transparent.rs +++ b/zebra-chain/src/transparent.rs @@ -1,5 +1,4 @@ //! Transparent-related (Bitcoin-inherited) functionality. -#![allow(clippy::unit_arg)] mod address; mod keys; @@ -11,20 +10,21 @@ pub use address::Address; pub use script::Script; pub use serialize::GENESIS_COINBASE_DATA; pub use utxo::{ - new_ordered_outputs, new_outputs, utxos_from_ordered_utxos, CoinbaseSpendRestriction, - OrderedUtxo, Utxo, + new_ordered_outputs, new_outputs, outputs_from_utxos, utxos_from_ordered_utxos, + CoinbaseSpendRestriction, OrderedUtxo, Utxo, }; -pub(crate) use utxo::outputs_from_utxos; - #[cfg(any(test, feature = "proptest-impl"))] -pub(crate) use utxo::new_transaction_ordered_outputs; +pub use utxo::{ + new_ordered_outputs_with_height, new_outputs_with_height, new_transaction_ordered_outputs, +}; #[cfg(any(test, feature = "proptest-impl"))] use proptest_derive::Arbitrary; #[cfg(any(test, feature = "proptest-impl"))] mod arbitrary; + #[cfg(test)] mod tests; diff --git a/zebra-chain/src/transparent/utxo.rs b/zebra-chain/src/transparent/utxo.rs index 9362266df..e7bfda4a2 100644 --- a/zebra-chain/src/transparent/utxo.rs +++ b/zebra-chain/src/transparent/utxo.rs @@ -3,7 +3,7 @@ use std::{collections::HashMap, convert::TryInto}; use crate::{ - block::{self, Block}, + block::{self, Block, Height}, transaction::{self, Transaction}, transparent, }; @@ -100,7 +100,7 @@ pub fn utxos_from_ordered_utxos( } /// Compute an index of [`Output`]s, given an index of [`Utxo`]s. -pub(crate) fn outputs_from_utxos( +pub fn outputs_from_utxos( utxos: HashMap, ) -> HashMap { utxos @@ -118,15 +118,46 @@ pub fn new_outputs( utxos_from_ordered_utxos(new_ordered_outputs(block, transaction_hashes)) } +/// Compute an index of newly created [`Utxo`]s, given a block and a +/// list of precomputed transaction hashes. +/// +/// This is a test-only function, prefer [`new_outputs`]. +#[cfg(any(test, feature = "proptest-impl"))] +pub fn new_outputs_with_height( + block: &Block, + height: Height, + transaction_hashes: &[transaction::Hash], +) -> HashMap { + utxos_from_ordered_utxos(new_ordered_outputs_with_height( + block, + height, + transaction_hashes, + )) +} + /// Compute an index of newly created [`OrderedUtxo`]s, given a block and a /// list of precomputed transaction hashes. pub fn new_ordered_outputs( block: &Block, transaction_hashes: &[transaction::Hash], ) -> HashMap { - let mut new_ordered_outputs = HashMap::new(); let height = block.coinbase_height().expect("block has coinbase height"); + new_ordered_outputs_with_height(block, height, transaction_hashes) +} + +/// Compute an index of newly created [`OrderedUtxo`]s, given a block and a +/// list of precomputed transaction hashes. +/// +/// This function is intended for use in this module, and in tests. +/// Prefer [`new_ordered_outputs`]. +pub fn new_ordered_outputs_with_height( + block: &Block, + height: Height, + transaction_hashes: &[transaction::Hash], +) -> HashMap { + let mut new_ordered_outputs = HashMap::new(); + for (tx_index_in_block, (transaction, hash)) in block .transactions .iter() @@ -148,8 +179,8 @@ pub fn new_ordered_outputs( /// its precomputed transaction hash, the transaction's index in its block, /// and the block's height. /// -/// This function is only intended for use in tests. -pub(crate) fn new_transaction_ordered_outputs( +/// This function is only for use in this module, and in tests. +pub fn new_transaction_ordered_outputs( transaction: &Transaction, hash: transaction::Hash, tx_index_in_block: usize, diff --git a/zebra-consensus/src/block/tests.rs b/zebra-consensus/src/block/tests.rs index cc7332b68..db9a901d1 100644 --- a/zebra-consensus/src/block/tests.rs +++ b/zebra-consensus/src/block/tests.rs @@ -11,7 +11,9 @@ use zebra_chain::{ amount::{Amount, MAX_MONEY}, block::{ self, - tests::generate::{large_multi_transaction_block, large_single_transaction_block}, + tests::generate::{ + large_multi_transaction_block, large_single_transaction_block_many_inputs, + }, Block, Height, }, parameters::{Network, NetworkUpgrade}, @@ -631,7 +633,7 @@ fn legacy_sigops_count_for_large_generated_blocks() { // We can't test sigops using the transaction verifier, because it looks up UTXOs. - let block = large_single_transaction_block(); + let block = large_single_transaction_block_many_inputs(); let mut legacy_sigop_count = 0; for transaction in block.transactions { let cached_ffi_transaction = diff --git a/zebra-state/src/arbitrary.rs b/zebra-state/src/arbitrary.rs index 6e382c642..1366f2d6d 100644 --- a/zebra-state/src/arbitrary.rs +++ b/zebra-state/src/arbitrary.rs @@ -8,7 +8,10 @@ use zebra_chain::{ value_balance::ValueBalance, }; -use crate::{request::ContextuallyValidBlock, service::chain_tip::ChainTipBlock, PreparedBlock}; +use crate::{ + request::ContextuallyValidBlock, service::chain_tip::ChainTipBlock, FinalizedBlock, + PreparedBlock, +}; /// Mocks computation done during semantic validation pub trait Prepare { @@ -21,7 +24,8 @@ impl Prepare for Arc { let hash = block.hash(); let height = block.coinbase_height().unwrap(); let transaction_hashes: Arc<[_]> = block.transactions.iter().map(|tx| tx.hash()).collect(); - let new_outputs = transparent::new_ordered_outputs(&block, &transaction_hashes); + let new_outputs = + transparent::new_ordered_outputs_with_height(&block, height, &transaction_hashes); PreparedBlock { block, @@ -155,3 +159,27 @@ impl ContextuallyValidBlock { Self::test_with_chain_pool_change(block, ValueBalance::zero()) } } + +impl FinalizedBlock { + /// Create a block that's ready to be committed to the finalized state, + /// using a precalculated [`block::Hash`] and [`block::Height`]. + /// + /// This is a test-only method, prefer [`FinalizedBlock::with_hash`]. + #[cfg(any(test, feature = "proptest-impl"))] + pub fn with_hash_and_height( + block: Arc, + hash: block::Hash, + height: block::Height, + ) -> Self { + let transaction_hashes: Arc<[_]> = block.transactions.iter().map(|tx| tx.hash()).collect(); + let new_outputs = transparent::new_outputs_with_height(&block, height, &transaction_hashes); + + Self { + block, + hash, + height, + new_outputs, + transaction_hashes, + } + } +} diff --git a/zebra-state/src/request.rs b/zebra-state/src/request.rs index 39d8dda6c..5f886290c 100644 --- a/zebra-state/src/request.rs +++ b/zebra-state/src/request.rs @@ -179,15 +179,14 @@ impl ContextuallyValidBlock { impl FinalizedBlock { /// Create a block that's ready to be committed to the finalized state, - /// using a precalculated [`block::Hash`] and [`block::Height`]. + /// using a precalculated [`block::Hash`]. /// /// Note: a [`FinalizedBlock`] isn't actually finalized /// until [`Request::CommitFinalizedBlock`] returns success. - pub fn with_hash_and_height( - block: Arc, - hash: block::Hash, - height: block::Height, - ) -> Self { + pub fn with_hash(block: Arc, hash: block::Hash) -> Self { + let height = block + .coinbase_height() + .expect("coinbase height was already checked"); let transaction_hashes: Arc<[_]> = block.transactions.iter().map(|tx| tx.hash()).collect(); let new_outputs = transparent::new_outputs(&block, &transaction_hashes); @@ -204,11 +203,8 @@ impl FinalizedBlock { impl From> for FinalizedBlock { fn from(block: Arc) -> Self { let hash = block.hash(); - let height = block - .coinbase_height() - .expect("finalized blocks must have a valid coinbase height"); - FinalizedBlock::with_hash_and_height(block, hash, height) + FinalizedBlock::with_hash(block, hash) } } diff --git a/zebra-state/src/service/finalized_state.rs b/zebra-state/src/service/finalized_state.rs index 3d08e991d..f7a8fc4f6 100644 --- a/zebra-state/src/service/finalized_state.rs +++ b/zebra-state/src/service/finalized_state.rs @@ -327,6 +327,7 @@ impl FinalizedState { ) -> Result { let finalized_hash = finalized.hash; + // Get a list of the spent UTXOs, before we delete any from the database let all_utxos_spent_by_block = finalized .block .transactions diff --git a/zebra-state/src/service/finalized_state/zebra_db/block.rs b/zebra-state/src/service/finalized_state/zebra_db/block.rs index 974fd1450..2f1873607 100644 --- a/zebra-state/src/service/finalized_state/zebra_db/block.rs +++ b/zebra-state/src/service/finalized_state/zebra_db/block.rs @@ -120,10 +120,6 @@ impl DiskWriteBatch { history_tree: HistoryTree, value_pool: ValueBalance, ) -> Result<(), BoxError> { - let hash_by_height = db.cf_handle("hash_by_height").unwrap(); - let height_by_hash = db.cf_handle("height_by_hash").unwrap(); - let block_by_height = db.cf_handle("block_by_height").unwrap(); - let FinalizedBlock { block, hash, @@ -131,12 +127,9 @@ impl DiskWriteBatch { .. } = &finalized; - // Index the block - self.zs_insert(hash_by_height, height, hash); - self.zs_insert(height_by_hash, hash, height); - - // TODO: as part of ticket #3151, commit transaction data, but not UTXOs or address indexes - self.zs_insert(block_by_height, height, block); + // Commit block and transaction data, + // but not transaction indexes, note commitments, or UTXOs. + self.prepare_block_header_transactions_batch(db, &finalized)?; // # Consensus // @@ -151,6 +144,7 @@ impl DiskWriteBatch { return Ok(()); } + // Commit transaction indexes self.prepare_transaction_index_batch(db, &finalized, &mut note_commitment_trees)?; self.prepare_note_commitment_batch( @@ -161,6 +155,7 @@ impl DiskWriteBatch { history_tree, )?; + // Commit UTXOs and value pools self.prepare_chain_value_pools_batch(db, &finalized, all_utxos_spent_by_block, value_pool)?; // The block has passed contextual validation, so update the metrics @@ -169,6 +164,38 @@ impl DiskWriteBatch { Ok(()) } + /// Prepare a database batch containing the block header and transactions + /// from `finalized.block`, and return it (without actually writing anything). + /// + /// # Errors + /// + /// - This method does not currently return any errors. + pub fn prepare_block_header_transactions_batch( + &mut self, + db: &DiskDb, + finalized: &FinalizedBlock, + ) -> Result<(), BoxError> { + let hash_by_height = db.cf_handle("hash_by_height").unwrap(); + let height_by_hash = db.cf_handle("height_by_hash").unwrap(); + let block_by_height = db.cf_handle("block_by_height").unwrap(); + + let FinalizedBlock { + block, + hash, + height, + .. + } = finalized; + + // Index the block + self.zs_insert(hash_by_height, height, hash); + self.zs_insert(height_by_hash, hash, height); + + // Commit block and transaction data, but not UTXOs or address indexes + self.zs_insert(block_by_height, height, block); + + Ok(()) + } + /// If `finalized.block` is a genesis block, /// prepare a database batch that finishes initializing the database, /// and return `true` (without actually writing anything). diff --git a/zebra-state/src/service/finalized_state/zebra_db/block/tests.rs b/zebra-state/src/service/finalized_state/zebra_db/block/tests.rs index 87af1565f..6824423e7 100644 --- a/zebra-state/src/service/finalized_state/zebra_db/block/tests.rs +++ b/zebra-state/src/service/finalized_state/zebra_db/block/tests.rs @@ -1,3 +1,4 @@ //! Tests for finalized database blocks and transactions. mod snapshot; +mod vectors; diff --git a/zebra-state/src/service/finalized_state/zebra_db/block/tests/vectors.rs b/zebra-state/src/service/finalized_state/zebra_db/block/tests/vectors.rs new file mode 100644 index 000000000..b053aca0d --- /dev/null +++ b/zebra-state/src/service/finalized_state/zebra_db/block/tests/vectors.rs @@ -0,0 +1,147 @@ +//! Fixed database test vectors for blocks and transactions. +//! +//! These tests check that the database correctly serializes +//! and deserializes large heights, blocks and transactions. +//! +//! # TODO +//! +//! Test large blocks and transactions with shielded data, +//! including data activated in Overwinter and later network upgrades. +//! +//! Check transparent address indexes, UTXOs, etc. + +use std::{iter, sync::Arc}; + +use zebra_chain::{ + block::{ + tests::generate::{ + large_multi_transaction_block, large_single_transaction_block_many_inputs, + large_single_transaction_block_many_outputs, + }, + Block, Height, + }, + parameters::Network::{self, *}, + serialization::{ZcashDeserializeInto, ZcashSerialize}, +}; +use zebra_test::vectors::{MAINNET_BLOCKS, TESTNET_BLOCKS}; + +use crate::{ + service::finalized_state::{disk_db::DiskWriteBatch, disk_format::IntoDisk, FinalizedState}, + Config, FinalizedBlock, +}; + +/// Storage round-trip test for block and transaction data in the finalized state database. +#[test] +fn test_block_db_round_trip() { + let mainnet_test_cases = MAINNET_BLOCKS + .iter() + .map(|(_height, block)| block.zcash_deserialize_into().unwrap()); + let testnet_test_cases = TESTNET_BLOCKS + .iter() + .map(|(_height, block)| block.zcash_deserialize_into().unwrap()); + + test_block_db_round_trip_with(Mainnet, mainnet_test_cases); + test_block_db_round_trip_with(Testnet, testnet_test_cases); + + // It doesn't matter if these blocks are mainnet or testnet, + // because there is no validation at this level of the database. + // + // These blocks have the same height and header hash, so they each need a new state. + test_block_db_round_trip_with(Mainnet, iter::once(large_multi_transaction_block())); + + // These blocks are unstable under serialization, so we apply a round-trip first. + // + // TODO: fix the bug in the generated test vectors. + let block = large_single_transaction_block_many_inputs(); + let block_data = block + .zcash_serialize_to_vec() + .expect("serialization to vec never fails"); + let block: Block = block_data + .zcash_deserialize_into() + .expect("deserialization of valid serialized block never fails"); + test_block_db_round_trip_with(Mainnet, iter::once(block)); + + let block = large_single_transaction_block_many_outputs(); + let block_data = block + .zcash_serialize_to_vec() + .expect("serialization to vec never fails"); + let block: Block = block_data + .zcash_deserialize_into() + .expect("deserialization of valid serialized block never fails"); + test_block_db_round_trip_with(Mainnet, iter::once(block)); +} + +fn test_block_db_round_trip_with( + network: Network, + block_test_cases: impl IntoIterator, +) { + zebra_test::init(); + + let state = FinalizedState::new(&Config::ephemeral(), network); + + // Check that each block round-trips to the database + for original_block in block_test_cases.into_iter() { + // First, check that the block round-trips without using the database + let block_data = original_block + .zcash_serialize_to_vec() + .expect("serialization to vec never fails"); + let round_trip_block: Block = block_data + .zcash_deserialize_into() + .expect("deserialization of valid serialized block never fails"); + let round_trip_data = round_trip_block + .zcash_serialize_to_vec() + .expect("serialization to vec never fails"); + + assert_eq!( + original_block, round_trip_block, + "test block structure must round-trip", + ); + assert_eq!( + block_data, round_trip_data, + "test block data must round-trip", + ); + + // Now, use the database + let original_block = Arc::new(original_block); + let finalized = if original_block.coinbase_height().is_some() { + original_block.clone().into() + } else { + // Fake a zero height + FinalizedBlock::with_hash_and_height( + original_block.clone(), + original_block.hash(), + Height(0), + ) + }; + + // Skip validation by writing the block directly to the database + let mut batch = DiskWriteBatch::new(); + batch + .prepare_block_header_transactions_batch(&state.db, &finalized) + .expect("block is valid for batch"); + state.db.write(batch).expect("block is valid for writing"); + + // Now read it back from the state + let stored_block = state + .block(finalized.height.into()) + .expect("block was stored at height"); + + if stored_block != original_block { + error!( + " + detailed block mismatch report: + original: {:?}\n\ + original data: {:?}\n\ + stored: {:?}\n\ + stored data: {:?}\n\ + ", + original_block, + hex::encode(original_block.as_bytes()), + stored_block, + hex::encode(stored_block.as_bytes()), + ); + } + + assert_eq!(stored_block, original_block); + } +}