Merge pull request #872 from nuttycom/feature/pre_dag_sync-suggest_scan_ranges

Implement `suggest_scan_ranges` and updated chain validation.
This commit is contained in:
str4d 2023-07-19 17:37:43 +01:00 committed by GitHub
commit f4221889fe
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
19 changed files with 2027 additions and 200 deletions

View File

@ -19,6 +19,6 @@ panic = 'abort'
codegen-units = 1 codegen-units = 1
[patch.crates-io] [patch.crates-io]
incrementalmerkletree = { git = "https://github.com/zcash/incrementalmerkletree.git", rev = "67111e29403c33f2e36d6924167f1d5f04ad0fc2" } incrementalmerkletree = { git = "https://github.com/zcash/incrementalmerkletree.git", rev = "bae25ad89c0c192bee625252d2d419bd56638e48" }
shardtree = { git = "https://github.com/zcash/incrementalmerkletree.git", rev = "67111e29403c33f2e36d6924167f1d5f04ad0fc2" } shardtree = { git = "https://github.com/zcash/incrementalmerkletree.git", rev = "bae25ad89c0c192bee625252d2d419bd56638e48" }
orchard = { git = "https://github.com/zcash/orchard.git", rev = "6ef89d5f154de2cf7b7dd87edb8d8c49158beebb" } orchard = { git = "https://github.com/zcash/orchard.git", rev = "6ef89d5f154de2cf7b7dd87edb8d8c49158beebb" }

View File

@ -17,8 +17,9 @@ and this library adheres to Rust's notion of
- `ShieldedProtocol` - `ShieldedProtocol`
- `WalletCommitmentTrees` - `WalletCommitmentTrees`
- `WalletRead::{block_metadata, block_fully_scanned, suggest_scan_ranges}` - `WalletRead::{block_metadata, block_fully_scanned, suggest_scan_ranges}`
- `WalletWrite::put_blocks` - `WalletWrite::{put_blocks, update_chain_tip}`
- `chain::CommitmentTreeRoot` - `chain::CommitmentTreeRoot`
- `scanning` A new module containing types required for `suggest_scan_ranges`
- `testing::MockWalletDb::new` - `testing::MockWalletDb::new`
- `wallet::input_sellection::Proposal::{min_target_height, min_anchor_height}`: - `wallet::input_sellection::Proposal::{min_target_height, min_anchor_height}`:
- `zcash_client_backend::wallet::WalletSaplingOutput::note_commitment_tree_position` - `zcash_client_backend::wallet::WalletSaplingOutput::note_commitment_tree_position`
@ -36,8 +37,10 @@ and this library adheres to Rust's notion of
and its signature has changed; it now subsumes the removed `WalletRead::get_all_nullifiers`. and its signature has changed; it now subsumes the removed `WalletRead::get_all_nullifiers`.
- `WalletRead::get_target_and_anchor_heights` now takes its argument as a `NonZeroU32` - `WalletRead::get_target_and_anchor_heights` now takes its argument as a `NonZeroU32`
- `chain::scan_cached_blocks` now takes a `from_height` argument that - `chain::scan_cached_blocks` now takes a `from_height` argument that
permits the caller to control the starting position of the scan range permits the caller to control the starting position of the scan range.
In addition, the `limit` parameter is now required. In addition, the `limit` parameter is now required and has type `usize`.
- `chain::BlockSource::with_blocks` now takes its limit as an `Option<usize>`
instead of `Option<u32>`.
- A new `CommitmentTree` variant has been added to `data_api::error::Error` - A new `CommitmentTree` variant has been added to `data_api::error::Error`
- `data_api::wallet::{create_spend_to_address, create_proposed_transaction, - `data_api::wallet::{create_spend_to_address, create_proposed_transaction,
shield_transparent_funds}` all now require that `WalletCommitmentTrees` be shield_transparent_funds}` all now require that `WalletCommitmentTrees` be
@ -67,7 +70,8 @@ and this library adheres to Rust's notion of
method now takes an optional `BlockMetadata` argument instead of a base commitment method now takes an optional `BlockMetadata` argument instead of a base commitment
tree and incremental witnesses for each previously-known note. In addition, the tree and incremental witnesses for each previously-known note. In addition, the
return type has now been updated to return a `Result<ScannedBlock, ScanError>`. return type has now been updated to return a `Result<ScannedBlock, ScanError>`.
- `proto/service.proto` has been updated to include the new GRPC endpoints
supported by lightwalletd v0.4.15
### Removed ### Removed
- `zcash_client_backend::data_api`: - `zcash_client_backend::data_api`:
@ -81,8 +85,8 @@ and this library adheres to Rust's notion of
feature flag, has been modified by the addition of a `sapling_tree` property. feature flag, has been modified by the addition of a `sapling_tree` property.
- `wallet::input_selection`: - `wallet::input_selection`:
- `Proposal::target_height` (use `Proposal::min_target_height` instead). - `Proposal::target_height` (use `Proposal::min_target_height` instead).
- `zcash_client_backend::data_api::chain::validate_chain` TODO: document how - `zcash_client_backend::data_api::chain::validate_chain` (logic merged into
to handle validation given out-of-order blocks. `chain::scan_cached_blocks`.
- `zcash_client_backend::data_api::chain::error::{ChainError, Cause}` have been - `zcash_client_backend::data_api::chain::error::{ChainError, Cause}` have been
replaced by `zcash_client_backend::scanning::ScanError` replaced by `zcash_client_backend::scanning::ScanError`
- `zcash_client_backend::wallet::WalletSaplingOutput::{witness, witness_mut}` - `zcash_client_backend::wallet::WalletSaplingOutput::{witness, witness_mut}`

View File

@ -118,6 +118,22 @@ message TreeState {
string orchardTree = 6; // orchard commitment tree state string orchardTree = 6; // orchard commitment tree state
} }
enum ShieldedProtocol {
sapling = 0;
orchard = 1;
}
message GetSubtreeRootsArg {
uint32 startIndex = 1; // Index identifying where to start returning subtree roots
ShieldedProtocol shieldedProtocol = 2; // Shielded protocol to return subtree roots for
uint32 maxEntries = 3; // Maximum number of entries to return, or 0 for all entries.
}
message SubtreeRoot {
bytes rootHash = 2; // The 32-byte Merkle root of the subtree.
bytes completingBlockHash = 3; // The hash of the block that completed this subtree.
uint64 completingBlockHeight = 4; // The height of the block that completed this subtree in the main chain.
}
// Results are sorted by height, which makes it easy to issue another // Results are sorted by height, which makes it easy to issue another
// request that picks up from where the previous left off. // request that picks up from where the previous left off.
message GetAddressUtxosArg { message GetAddressUtxosArg {
@ -142,8 +158,12 @@ service CompactTxStreamer {
rpc GetLatestBlock(ChainSpec) returns (BlockID) {} rpc GetLatestBlock(ChainSpec) returns (BlockID) {}
// Return the compact block corresponding to the given block identifier // Return the compact block corresponding to the given block identifier
rpc GetBlock(BlockID) returns (CompactBlock) {} rpc GetBlock(BlockID) returns (CompactBlock) {}
// Same as GetBlock except actions contain only nullifiers
rpc GetBlockNullifiers(BlockID) returns (CompactBlock) {}
// Return a list of consecutive compact blocks // Return a list of consecutive compact blocks
rpc GetBlockRange(BlockRange) returns (stream CompactBlock) {} rpc GetBlockRange(BlockRange) returns (stream CompactBlock) {}
// Same as GetBlockRange except actions contain only nullifiers
rpc GetBlockRangeNullifiers(BlockRange) returns (stream CompactBlock) {}
// Return the requested full (not compact) transaction (as from zcashd) // Return the requested full (not compact) transaction (as from zcashd)
rpc GetTransaction(TxFilter) returns (RawTransaction) {} rpc GetTransaction(TxFilter) returns (RawTransaction) {}
@ -177,6 +197,10 @@ service CompactTxStreamer {
rpc GetTreeState(BlockID) returns (TreeState) {} rpc GetTreeState(BlockID) returns (TreeState) {}
rpc GetLatestTreeState(Empty) returns (TreeState) {} rpc GetLatestTreeState(Empty) returns (TreeState) {}
// Returns a stream of information about roots of subtrees of the Sapling and Orchard
// note commitment trees.
rpc GetSubtreeRoots(GetSubtreeRootsArg) returns (stream SubtreeRoot) {}
rpc GetAddressUtxos(GetAddressUtxosArg) returns (GetAddressUtxosReplyList) {} rpc GetAddressUtxos(GetAddressUtxosArg) returns (GetAddressUtxosReplyList) {}
rpc GetAddressUtxosStream(GetAddressUtxosArg) returns (stream GetAddressUtxosReply) {} rpc GetAddressUtxosStream(GetAddressUtxosArg) returns (stream GetAddressUtxosReply) {}

View File

@ -1,9 +1,9 @@
//! Interfaces for wallet data persistence & low-level wallet utilities. //! Interfaces for wallet data persistence & low-level wallet utilities.
use std::cmp;
use std::collections::HashMap; use std::collections::HashMap;
use std::fmt::Debug; use std::fmt::Debug;
use std::num::NonZeroU32; use std::num::NonZeroU32;
use std::{cmp, ops::Range};
use incrementalmerkletree::Retention; use incrementalmerkletree::Retention;
use secrecy::SecretVec; use secrecy::SecretVec;
@ -29,9 +29,11 @@ use crate::{
}; };
use self::chain::CommitmentTreeRoot; use self::chain::CommitmentTreeRoot;
use self::scanning::ScanRange;
pub mod chain; pub mod chain;
pub mod error; pub mod error;
pub mod scanning;
pub mod wallet; pub mod wallet;
pub const SAPLING_SHARD_HEIGHT: u8 = sapling::NOTE_COMMITMENT_TREE_DEPTH / 2; pub const SAPLING_SHARD_HEIGHT: u8 = sapling::NOTE_COMMITMENT_TREE_DEPTH / 2;
@ -87,12 +89,10 @@ pub trait WalletRead {
/// tree size information for each block; or else the scan is likely to fail if notes belonging /// tree size information for each block; or else the scan is likely to fail if notes belonging
/// to the wallet are detected. /// to the wallet are detected.
/// ///
/// The returned range(s) may include block heights beyond the current chain tip.
///
/// [`CompactBlock`]: crate::proto::compact_formats::CompactBlock /// [`CompactBlock`]: crate::proto::compact_formats::CompactBlock
fn suggest_scan_ranges( fn suggest_scan_ranges(&self) -> Result<Vec<ScanRange>, Self::Error>;
&self,
batch_size: usize,
limit: usize,
) -> Result<Vec<Range<BlockHeight>>, Self::Error>;
/// Returns the default target height (for the block in which a new /// Returns the default target height (for the block in which a new
/// transaction would be mined) and anchor height (to use for a new /// transaction would be mined) and anchor height (to use for a new
@ -501,6 +501,15 @@ pub trait WalletWrite: WalletRead {
block: Vec<ScannedBlock<sapling::Nullifier>>, block: Vec<ScannedBlock<sapling::Nullifier>>,
) -> Result<Vec<Self::NoteRef>, Self::Error>; ) -> Result<Vec<Self::NoteRef>, Self::Error>;
/// Updates the wallet's view of the blockchain.
///
/// This method is used to provide the wallet with information about the state of the
/// blockchain, and detect any previously scanned data that needs to be re-validated
/// before proceeding with scanning. It should be called at wallet startup prior to calling
/// [`WalletRead::suggest_scan_ranges`] in order to provide the wallet with the information it
/// needs to correctly prioritize scanning operations.
fn update_chain_tip(&mut self, tip_height: BlockHeight) -> Result<(), Self::Error>;
/// Caches a decrypted transaction in the persistent wallet store. /// Caches a decrypted transaction in the persistent wallet store.
fn store_decrypted_tx( fn store_decrypted_tx(
&mut self, &mut self,
@ -569,7 +578,7 @@ pub mod testing {
use incrementalmerkletree::Address; use incrementalmerkletree::Address;
use secrecy::{ExposeSecret, SecretVec}; use secrecy::{ExposeSecret, SecretVec};
use shardtree::{memory::MemoryShardStore, ShardTree, ShardTreeError}; use shardtree::{memory::MemoryShardStore, ShardTree, ShardTreeError};
use std::{collections::HashMap, convert::Infallible, ops::Range}; use std::{collections::HashMap, convert::Infallible};
use zcash_primitives::{ use zcash_primitives::{
block::BlockHash, block::BlockHash,
@ -591,9 +600,9 @@ pub mod testing {
}; };
use super::{ use super::{
chain::CommitmentTreeRoot, BlockMetadata, DecryptedTransaction, NullifierQuery, chain::CommitmentTreeRoot, scanning::ScanRange, BlockMetadata, DecryptedTransaction,
ScannedBlock, SentTransaction, WalletCommitmentTrees, WalletRead, WalletWrite, NullifierQuery, ScannedBlock, SentTransaction, WalletCommitmentTrees, WalletRead,
SAPLING_SHARD_HEIGHT, WalletWrite, SAPLING_SHARD_HEIGHT,
}; };
pub struct MockWalletDb { pub struct MockWalletDb {
@ -634,11 +643,7 @@ pub mod testing {
Ok(None) Ok(None)
} }
fn suggest_scan_ranges( fn suggest_scan_ranges(&self) -> Result<Vec<ScanRange>, Self::Error> {
&self,
_batch_size: usize,
_limit: usize,
) -> Result<Vec<Range<BlockHeight>>, Self::Error> {
Ok(vec![]) Ok(vec![])
} }
@ -780,6 +785,10 @@ pub mod testing {
Ok(vec![]) Ok(vec![])
} }
fn update_chain_tip(&mut self, _tip_height: BlockHeight) -> Result<(), Self::Error> {
Ok(())
}
fn store_decrypted_tx( fn store_decrypted_tx(
&mut self, &mut self,
_received_tx: DecryptedTransaction, _received_tx: DecryptedTransaction,

View File

@ -7,18 +7,21 @@
//! # #[cfg(feature = "test-dependencies")] //! # #[cfg(feature = "test-dependencies")]
//! # { //! # {
//! use zcash_primitives::{ //! use zcash_primitives::{
//! consensus::{BlockHeight, Network, Parameters} //! consensus::{BlockHeight, Network, Parameters},
//! sapling
//! }; //! };
//! //!
//! use zcash_client_backend::{ //! use zcash_client_backend::{
//! data_api::{ //! data_api::{
//! WalletRead, WalletWrite, //! WalletRead, WalletWrite, WalletCommitmentTrees,
//! chain::{ //! chain::{
//! BlockSource, //! BlockSource,
//! CommitmentTreeRoot,
//! error::Error, //! error::Error,
//! scan_cached_blocks, //! scan_cached_blocks,
//! testing as chain_testing, //! testing as chain_testing,
//! }, //! },
//! scanning::ScanPriority,
//! testing, //! testing,
//! }, //! },
//! }; //! };
@ -32,20 +35,111 @@
//! # fn test() -> Result<(), Error<(), Infallible>> { //! # fn test() -> Result<(), Error<(), Infallible>> {
//! let network = Network::TestNetwork; //! let network = Network::TestNetwork;
//! let block_source = chain_testing::MockBlockSource; //! let block_source = chain_testing::MockBlockSource;
//! let mut db_data = testing::MockWalletDb::new(Network::TestNetwork); //! let mut wallet_db = testing::MockWalletDb::new(Network::TestNetwork);
//! //!
//! // 1) Download new CompactBlocks into block_source. //! // 1) Download note commitment tree data from lightwalletd
//! // //! let roots: Vec<CommitmentTreeRoot<sapling::Node>> = unimplemented!();
//! // 2) FIXME: Obtain necessary block metadata for continuity checking? //!
//! // //! // 2) Pass the commitment tree data to the database.
//! // 3) Scan cached blocks. //! wallet_db.put_sapling_subtree_roots(0, &roots).unwrap();
//! // //!
//! // FIXME: update documentation on how to detect when a rewind is required. //! // 3) Download chain tip metadata from lightwalletd
//! // //! let tip_height: BlockHeight = unimplemented!();
//! // At this point, the cache and scanned data are locally consistent (though not //!
//! // necessarily consistent with the latest chain tip - this would be discovered the //! // 4) Notify the wallet of the updated chain tip.
//! // next time this codepath is executed after new blocks are received). //! wallet_db.update_chain_tip(tip_height).map_err(Error::Wallet)?;
//! scan_cached_blocks(&network, &block_source, &mut db_data, BlockHeight::from(0), 10) //!
//! // 5) Get the suggested scan ranges from the wallet database
//! let mut scan_ranges = wallet_db.suggest_scan_ranges().map_err(Error::Wallet)?;
//!
//! // 6) Run the following loop until the wallet's view of the chain tip as of the previous wallet
//! // session is valid.
//! loop {
//! // If there is a range of blocks that needs to be verified, it will always be returned as
//! // the first element of the vector of suggested ranges.
//! match scan_ranges.first() {
//! Some(scan_range) if scan_range.priority() == ScanPriority::Verify => {
//! // Download the blocks in `scan_range` into the block source, overwriting any
//! // existing blocks in this range.
//! unimplemented!();
//!
//! // Scan the downloaded blocks
//! let scan_result = scan_cached_blocks(
//! &network,
//! &block_source,
//! &mut wallet_db,
//! scan_range.block_range().start,
//! scan_range.len()
//! );
//!
//! // Check for scanning errors that indicate that the wallet's chain tip is out of
//! // sync with blockchain history.
//! match scan_result {
//! Ok(_) => {
//! // At this point, the cache and scanned data are locally consistent (though
//! // not necessarily consistent with the latest chain tip - this would be
//! // discovered the next time this codepath is executed after new blocks are
//! // received) so we can break out of the loop.
//! break;
//! }
//! Err(Error::Scan(err)) if err.is_continuity_error() => {
//! // Pick a height to rewind to, which must be at least one block before
//! // the height at which the error occurred, but may be an earlier height
//! // determined based on heuristics such as the platform, available bandwidth,
//! // size of recent CompactBlocks, etc.
//! let rewind_height = err.at_height().saturating_sub(10);
//!
//! // Rewind to the chosen height.
//! wallet_db.truncate_to_height(rewind_height).map_err(Error::Wallet)?;
//!
//! // Delete cached blocks from rewind_height onwards.
//! //
//! // This does imply that assumed-valid blocks will be re-downloaded, but it
//! // is also possible that in the intervening time, a chain reorg has
//! // occurred that orphaned some of those blocks.
//! unimplemented!();
//! }
//! Err(other) => {
//! // Handle or return other errors
//! }
//! }
//!
//! // In case we updated the suggested scan ranges, now re-request.
//! scan_ranges = wallet_db.suggest_scan_ranges().map_err(Error::Wallet)?;
//! }
//! _ => {
//! // Nothing to verify; break out of the loop
//! break;
//! }
//! }
//! }
//!
//! // 7) Loop over the remaining suggested scan ranges, retrieving the requested data and calling
//! // `scan_cached_blocks` on each range. Periodically, or if a continuity error is
//! // encountered, this process should be repeated starting at step (3).
//! let scan_ranges = wallet_db.suggest_scan_ranges().map_err(Error::Wallet)?;
//! for scan_range in scan_ranges {
//! // Download the blocks in `scan_range` into the block source. While in this example this
//! // step is performed in-line, it's fine for the download of scan ranges to be asynchronous
//! // and for the scanner to process the downloaded ranges as they become available in a
//! // separate thread. The scan ranges should also be broken down into smaller chunks as
//! // appropriate, and for ranges with priority `Historic` it can be useful to download and
//! // scan the range in reverse order (to discover more recent unspent notes sooner), or from
//! // the start and end of the range inwards.
//! unimplemented!();
//!
//! // Scan the downloaded blocks.
//! let scan_result = scan_cached_blocks(
//! &network,
//! &block_source,
//! &mut wallet_db,
//! scan_range.block_range().start,
//! scan_range.len()
//! )?;
//!
//! // Handle scan errors, etc.
//! }
//! # Ok(())
//! # } //! # }
//! # } //! # }
//! ``` //! ```
@ -58,14 +152,12 @@ use zcash_primitives::{
}; };
use crate::{ use crate::{
data_api::{NullifierQuery, WalletWrite}, data_api::{BlockMetadata, NullifierQuery, WalletWrite},
proto::compact_formats::CompactBlock, proto::compact_formats::CompactBlock,
scan::BatchRunner, scan::BatchRunner,
scanning::{add_block_to_runner, check_continuity, scan_block_with_runner}, scanning::{add_block_to_runner, check_continuity, scan_block_with_runner},
}; };
use super::BlockMetadata;
pub mod error; pub mod error;
use error::Error; use error::Error;
@ -114,7 +206,7 @@ pub trait BlockSource {
fn with_blocks<F, WalletErrT>( fn with_blocks<F, WalletErrT>(
&self, &self,
from_height: Option<BlockHeight>, from_height: Option<BlockHeight>,
limit: Option<u32>, limit: Option<usize>,
with_row: F, with_row: F,
) -> Result<(), error::Error<WalletErrT, Self::Error>> ) -> Result<(), error::Error<WalletErrT, Self::Error>>
where where
@ -145,7 +237,7 @@ pub fn scan_cached_blocks<ParamsT, DbT, BlockSourceT>(
block_source: &BlockSourceT, block_source: &BlockSourceT,
data_db: &mut DbT, data_db: &mut DbT,
from_height: BlockHeight, from_height: BlockHeight,
limit: u32, limit: usize,
) -> Result<(), Error<DbT::Error, BlockSourceT::Error>> ) -> Result<(), Error<DbT::Error, BlockSourceT::Error>>
where where
ParamsT: consensus::Parameters + Send + 'static, ParamsT: consensus::Parameters + Send + 'static,
@ -292,7 +384,7 @@ pub mod testing {
fn with_blocks<F, DbErrT>( fn with_blocks<F, DbErrT>(
&self, &self,
_from_height: Option<BlockHeight>, _from_height: Option<BlockHeight>,
_limit: Option<u32>, _limit: Option<usize>,
_with_row: F, _with_row: F,
) -> Result<(), Error<DbErrT, Infallible>> ) -> Result<(), Error<DbErrT, Infallible>>
where where

View File

@ -0,0 +1,186 @@
use std::fmt;
use std::ops::Range;
use zcash_primitives::consensus::BlockHeight;
/// Scanning range priority levels.
#[derive(Debug, Clone, Copy, PartialEq, Eq, PartialOrd, Ord)]
pub enum ScanPriority {
/// Block ranges that have already been scanned have lowest priority.
Scanned,
/// Block ranges to be scanned to advance the fully-scanned height.
Historic,
/// Block ranges adjacent to heights at which the user opened the wallet.
OpenAdjacent,
/// Blocks that must be scanned to complete note commitment tree shards adjacent to found notes.
FoundNote,
/// Blocks that must be scanned to complete the latest note commitment tree shard.
ChainTip,
/// A previously scanned range that must be verified to check it is still in the
/// main chain, has highest priority.
Verify,
}
/// A range of blocks to be scanned, along with its associated priority.
#[derive(Debug, Clone, PartialEq, Eq)]
pub struct ScanRange {
block_range: Range<BlockHeight>,
priority: ScanPriority,
}
impl fmt::Display for ScanRange {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
write!(
f,
"{:?}({}..{})",
self.priority, self.block_range.start, self.block_range.end,
)
}
}
impl ScanRange {
/// Constructs a scan range from its constituent parts.
pub fn from_parts(block_range: Range<BlockHeight>, priority: ScanPriority) -> Self {
assert!(
block_range.end >= block_range.start,
"{:?} is invalid for ScanRange({:?})",
block_range,
priority,
);
ScanRange {
block_range,
priority,
}
}
/// Returns the range of block heights to be scanned.
pub fn block_range(&self) -> &Range<BlockHeight> {
&self.block_range
}
/// Returns the priority with which the scan range should be scanned.
pub fn priority(&self) -> ScanPriority {
self.priority
}
/// Returns whether or not the scan range is empty.
pub fn is_empty(&self) -> bool {
self.block_range.is_empty()
}
/// Returns the number of blocks in the scan range.
pub fn len(&self) -> usize {
usize::try_from(u32::from(self.block_range.end) - u32::from(self.block_range.start))
.unwrap()
}
/// Shifts the start of the block range to the right if `block_height >
/// self.block_range().start`. Returns `None` if the resulting range would
/// be empty (or the range was already empty).
pub fn truncate_start(&self, block_height: BlockHeight) -> Option<Self> {
if block_height >= self.block_range.end || self.is_empty() {
None
} else {
Some(ScanRange {
block_range: self.block_range.start.max(block_height)..self.block_range.end,
priority: self.priority,
})
}
}
/// Shifts the end of the block range to the left if `block_height <
/// self.block_range().end`. Returns `None` if the resulting range would
/// be empty (or the range was already empty).
pub fn truncate_end(&self, block_height: BlockHeight) -> Option<Self> {
if block_height <= self.block_range.start || self.is_empty() {
None
} else {
Some(ScanRange {
block_range: self.block_range.start..self.block_range.end.min(block_height),
priority: self.priority,
})
}
}
/// Splits this scan range at the specified height, such that the provided height becomes the
/// end of the first range returned and the start of the second. Returns `None` if
/// `p <= self.block_range().start || p >= self.block_range().end`.
pub fn split_at(&self, p: BlockHeight) -> Option<(Self, Self)> {
(p > self.block_range.start && p < self.block_range.end).then_some((
ScanRange {
block_range: self.block_range.start..p,
priority: self.priority,
},
ScanRange {
block_range: p..self.block_range.end,
priority: self.priority,
},
))
}
}
#[cfg(test)]
mod tests {
use super::{ScanPriority, ScanRange};
fn scan_range(start: u32, end: u32) -> ScanRange {
ScanRange::from_parts((start.into())..(end.into()), ScanPriority::Scanned)
}
#[test]
fn truncate_start() {
let r = scan_range(5, 8);
assert_eq!(r.truncate_start(4.into()), Some(scan_range(5, 8)));
assert_eq!(r.truncate_start(5.into()), Some(scan_range(5, 8)));
assert_eq!(r.truncate_start(6.into()), Some(scan_range(6, 8)));
assert_eq!(r.truncate_start(7.into()), Some(scan_range(7, 8)));
assert_eq!(r.truncate_start(8.into()), None);
assert_eq!(r.truncate_start(9.into()), None);
let empty = scan_range(5, 5);
assert_eq!(empty.truncate_start(4.into()), None);
assert_eq!(empty.truncate_start(5.into()), None);
assert_eq!(empty.truncate_start(6.into()), None);
}
#[test]
fn truncate_end() {
let r = scan_range(5, 8);
assert_eq!(r.truncate_end(9.into()), Some(scan_range(5, 8)));
assert_eq!(r.truncate_end(8.into()), Some(scan_range(5, 8)));
assert_eq!(r.truncate_end(7.into()), Some(scan_range(5, 7)));
assert_eq!(r.truncate_end(6.into()), Some(scan_range(5, 6)));
assert_eq!(r.truncate_end(5.into()), None);
assert_eq!(r.truncate_end(4.into()), None);
let empty = scan_range(5, 5);
assert_eq!(empty.truncate_end(4.into()), None);
assert_eq!(empty.truncate_end(5.into()), None);
assert_eq!(empty.truncate_end(6.into()), None);
}
#[test]
fn split_at() {
let r = scan_range(5, 8);
assert_eq!(r.split_at(4.into()), None);
assert_eq!(r.split_at(5.into()), None);
assert_eq!(
r.split_at(6.into()),
Some((scan_range(5, 6), scan_range(6, 8)))
);
assert_eq!(
r.split_at(7.into()),
Some((scan_range(5, 7), scan_range(7, 8)))
);
assert_eq!(r.split_at(8.into()), None);
assert_eq!(r.split_at(9.into()), None);
let empty = scan_range(5, 5);
assert_eq!(empty.split_at(4.into()), None);
assert_eq!(empty.split_at(5.into()), None);
assert_eq!(empty.split_at(6.into()), None);
}
}

View File

@ -187,6 +187,32 @@ pub struct TreeState {
#[prost(string, tag = "6")] #[prost(string, tag = "6")]
pub orchard_tree: ::prost::alloc::string::String, pub orchard_tree: ::prost::alloc::string::String,
} }
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetSubtreeRootsArg {
/// Index identifying where to start returning subtree roots
#[prost(uint32, tag = "1")]
pub start_index: u32,
/// Shielded protocol to return subtree roots for
#[prost(enumeration = "ShieldedProtocol", tag = "2")]
pub shielded_protocol: i32,
/// Maximum number of entries to return, or 0 for all entries.
#[prost(uint32, tag = "3")]
pub max_entries: u32,
}
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SubtreeRoot {
/// The 32-byte Merkle root of the subtree.
#[prost(bytes = "vec", tag = "2")]
pub root_hash: ::prost::alloc::vec::Vec<u8>,
/// The hash of the block that completed this subtree.
#[prost(bytes = "vec", tag = "3")]
pub completing_block_hash: ::prost::alloc::vec::Vec<u8>,
/// The height of the block that completed this subtree in the main chain.
#[prost(uint64, tag = "4")]
pub completing_block_height: u64,
}
/// Results are sorted by height, which makes it easy to issue another /// Results are sorted by height, which makes it easy to issue another
/// request that picks up from where the previous left off. /// request that picks up from where the previous left off.
#[allow(clippy::derive_partial_eq_without_eq)] #[allow(clippy::derive_partial_eq_without_eq)]
@ -222,6 +248,32 @@ pub struct GetAddressUtxosReplyList {
#[prost(message, repeated, tag = "1")] #[prost(message, repeated, tag = "1")]
pub address_utxos: ::prost::alloc::vec::Vec<GetAddressUtxosReply>, pub address_utxos: ::prost::alloc::vec::Vec<GetAddressUtxosReply>,
} }
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum ShieldedProtocol {
Sapling = 0,
Orchard = 1,
}
impl ShieldedProtocol {
/// String value of the enum field names used in the ProtoBuf definition.
///
/// The values are not transformed in any way and thus are considered stable
/// (if the ProtoBuf definition does not change) and safe for programmatic use.
pub fn as_str_name(&self) -> &'static str {
match self {
ShieldedProtocol::Sapling => "sapling",
ShieldedProtocol::Orchard => "orchard",
}
}
/// Creates an enum from field names used in the ProtoBuf definition.
pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
match value {
"sapling" => Some(Self::Sapling),
"orchard" => Some(Self::Orchard),
_ => None,
}
}
}
/// Generated client implementations. /// Generated client implementations.
pub mod compact_tx_streamer_client { pub mod compact_tx_streamer_client {
#![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)] #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
@ -366,6 +418,37 @@ pub mod compact_tx_streamer_client {
); );
self.inner.unary(req, path, codec).await self.inner.unary(req, path, codec).await
} }
/// Same as GetBlock except actions contain only nullifiers
pub async fn get_block_nullifiers(
&mut self,
request: impl tonic::IntoRequest<super::BlockId>,
) -> std::result::Result<
tonic::Response<crate::proto::compact_formats::CompactBlock>,
tonic::Status,
> {
self.inner
.ready()
.await
.map_err(|e| {
tonic::Status::new(
tonic::Code::Unknown,
format!("Service was not ready: {}", e.into()),
)
})?;
let codec = tonic::codec::ProstCodec::default();
let path = http::uri::PathAndQuery::from_static(
"/cash.z.wallet.sdk.rpc.CompactTxStreamer/GetBlockNullifiers",
);
let mut req = request.into_request();
req.extensions_mut()
.insert(
GrpcMethod::new(
"cash.z.wallet.sdk.rpc.CompactTxStreamer",
"GetBlockNullifiers",
),
);
self.inner.unary(req, path, codec).await
}
/// Return a list of consecutive compact blocks /// Return a list of consecutive compact blocks
pub async fn get_block_range( pub async fn get_block_range(
&mut self, &mut self,
@ -399,6 +482,39 @@ pub mod compact_tx_streamer_client {
); );
self.inner.server_streaming(req, path, codec).await self.inner.server_streaming(req, path, codec).await
} }
/// Same as GetBlockRange except actions contain only nullifiers
pub async fn get_block_range_nullifiers(
&mut self,
request: impl tonic::IntoRequest<super::BlockRange>,
) -> std::result::Result<
tonic::Response<
tonic::codec::Streaming<crate::proto::compact_formats::CompactBlock>,
>,
tonic::Status,
> {
self.inner
.ready()
.await
.map_err(|e| {
tonic::Status::new(
tonic::Code::Unknown,
format!("Service was not ready: {}", e.into()),
)
})?;
let codec = tonic::codec::ProstCodec::default();
let path = http::uri::PathAndQuery::from_static(
"/cash.z.wallet.sdk.rpc.CompactTxStreamer/GetBlockRangeNullifiers",
);
let mut req = request.into_request();
req.extensions_mut()
.insert(
GrpcMethod::new(
"cash.z.wallet.sdk.rpc.CompactTxStreamer",
"GetBlockRangeNullifiers",
),
);
self.inner.server_streaming(req, path, codec).await
}
/// Return the requested full (not compact) transaction (as from zcashd) /// Return the requested full (not compact) transaction (as from zcashd)
pub async fn get_transaction( pub async fn get_transaction(
&mut self, &mut self,
@ -671,6 +787,38 @@ pub mod compact_tx_streamer_client {
); );
self.inner.unary(req, path, codec).await self.inner.unary(req, path, codec).await
} }
/// Returns a stream of information about roots of subtrees of the Sapling and Orchard
/// note commitment trees.
pub async fn get_subtree_roots(
&mut self,
request: impl tonic::IntoRequest<super::GetSubtreeRootsArg>,
) -> std::result::Result<
tonic::Response<tonic::codec::Streaming<super::SubtreeRoot>>,
tonic::Status,
> {
self.inner
.ready()
.await
.map_err(|e| {
tonic::Status::new(
tonic::Code::Unknown,
format!("Service was not ready: {}", e.into()),
)
})?;
let codec = tonic::codec::ProstCodec::default();
let path = http::uri::PathAndQuery::from_static(
"/cash.z.wallet.sdk.rpc.CompactTxStreamer/GetSubtreeRoots",
);
let mut req = request.into_request();
req.extensions_mut()
.insert(
GrpcMethod::new(
"cash.z.wallet.sdk.rpc.CompactTxStreamer",
"GetSubtreeRoots",
),
);
self.inner.server_streaming(req, path, codec).await
}
pub async fn get_address_utxos( pub async fn get_address_utxos(
&mut self, &mut self,
request: impl tonic::IntoRequest<super::GetAddressUtxosArg>, request: impl tonic::IntoRequest<super::GetAddressUtxosArg>,

View File

@ -110,7 +110,7 @@ impl ScanningKey for SaplingIvk {
} }
/// Errors that may occur in chain scanning /// Errors that may occur in chain scanning
#[derive(Copy, Clone, Debug)] #[derive(Clone, Debug)]
pub enum ScanError { pub enum ScanError {
/// The hash of the parent block given by a proposed new chain tip does not match the hash of /// The hash of the parent block given by a proposed new chain tip does not match the hash of
/// the current chain tip. /// the current chain tip.
@ -141,21 +141,46 @@ pub enum ScanError {
}, },
} }
impl ScanError {
/// Returns whether this error is the result of a failed continuity check
pub fn is_continuity_error(&self) -> bool {
use ScanError::*;
match self {
PrevHashMismatch { .. } => true,
BlockHeightDiscontinuity { .. } => true,
TreeSizeMismatch { .. } => true,
TreeSizeUnknown { .. } => false,
}
}
/// Returns the block height at which the scan error occurred
pub fn at_height(&self) -> BlockHeight {
use ScanError::*;
match self {
PrevHashMismatch { at_height } => *at_height,
BlockHeightDiscontinuity { new_height, .. } => *new_height,
TreeSizeMismatch { at_height, .. } => *at_height,
TreeSizeUnknown { at_height, .. } => *at_height,
}
}
}
impl fmt::Display for ScanError { impl fmt::Display for ScanError {
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result { fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
use ScanError::*;
match &self { match &self {
ScanError::PrevHashMismatch { at_height } => write!( PrevHashMismatch { at_height } => write!(
f, f,
"The parent hash of proposed block does not correspond to the block hash at height {}.", "The parent hash of proposed block does not correspond to the block hash at height {}.",
at_height at_height
), ),
ScanError::BlockHeightDiscontinuity { prev_height, new_height } => { BlockHeightDiscontinuity { prev_height, new_height } => {
write!(f, "Block height discontinuity at height {}; next height is : {}", prev_height, new_height) write!(f, "Block height discontinuity at height {}; previous height was: {}", new_height, prev_height)
} }
ScanError::TreeSizeMismatch { protocol, at_height, given, computed } => { TreeSizeMismatch { protocol, at_height, given, computed } => {
write!(f, "The {:?} note commitment tree size provided by a compact block did not match the expected size at height {}; given {}, expected {}", protocol, at_height, given, computed) write!(f, "The {:?} note commitment tree size provided by a compact block did not match the expected size at height {}; given {}, expected {}", protocol, at_height, given, computed)
} }
ScanError::TreeSizeUnknown { protocol, at_height } => { TreeSizeUnknown { protocol, at_height } => {
write!(f, "Unable to determine {:?} note commitment tree size at height {}", protocol, at_height) write!(f, "Unable to determine {:?} note commitment tree size at height {}", protocol, at_height)
} }
} }

View File

@ -29,7 +29,7 @@ pub mod migrations;
pub(crate) fn blockdb_with_blocks<F, DbErrT>( pub(crate) fn blockdb_with_blocks<F, DbErrT>(
block_source: &BlockDb, block_source: &BlockDb,
from_height: Option<BlockHeight>, from_height: Option<BlockHeight>,
limit: Option<u32>, limit: Option<usize>,
mut with_row: F, mut with_row: F,
) -> Result<(), Error<DbErrT, SqliteClientError>> ) -> Result<(), Error<DbErrT, SqliteClientError>>
where where
@ -52,7 +52,9 @@ where
let mut rows = stmt_blocks let mut rows = stmt_blocks
.query(params![ .query(params![
from_height.map_or(0u32, u32::from), from_height.map_or(0u32, u32::from),
limit.unwrap_or(u32::max_value()), limit
.and_then(|l| u32::try_from(l).ok())
.unwrap_or(u32::MAX)
]) ])
.map_err(to_chain_error)?; .map_err(to_chain_error)?;
@ -198,7 +200,7 @@ pub(crate) fn blockmetadb_find_block(
pub(crate) fn fsblockdb_with_blocks<F, DbErrT>( pub(crate) fn fsblockdb_with_blocks<F, DbErrT>(
cache: &FsBlockDb, cache: &FsBlockDb,
from_height: Option<BlockHeight>, from_height: Option<BlockHeight>,
limit: Option<u32>, limit: Option<usize>,
mut with_block: F, mut with_block: F,
) -> Result<(), Error<DbErrT, FsBlockDbError>> ) -> Result<(), Error<DbErrT, FsBlockDbError>>
where where
@ -223,7 +225,9 @@ where
.query_map( .query_map(
params![ params![
from_height.map_or(0u32, u32::from), from_height.map_or(0u32, u32::from),
limit.unwrap_or(u32::max_value()), limit
.and_then(|l| u32::try_from(l).ok())
.unwrap_or(u32::MAX)
], ],
|row| { |row| {
Ok(BlockMeta { Ok(BlockMeta {
@ -279,11 +283,12 @@ mod tests {
use zcash_client_backend::{ use zcash_client_backend::{
address::RecipientAddress, address::RecipientAddress,
data_api::{ data_api::{
chain::scan_cached_blocks, chain::{error::Error, scan_cached_blocks},
wallet::{input_selection::GreedyInputSelector, spend}, wallet::{input_selection::GreedyInputSelector, spend},
WalletRead, WalletWrite, WalletRead, WalletWrite,
}, },
fees::{zip317::SingleOutputChangeStrategy, DustOutputPolicy}, fees::{zip317::SingleOutputChangeStrategy, DustOutputPolicy},
scanning::ScanError,
wallet::OvkPolicy, wallet::OvkPolicy,
zip321::{Payment, TransactionRequest}, zip321::{Payment, TransactionRequest},
}; };
@ -315,12 +320,9 @@ mod tests {
assert_matches!(db_data.get_max_height_hash(), Ok(None)); assert_matches!(db_data.get_max_height_hash(), Ok(None));
// Create a fake CompactBlock sending value to the address // Create a fake CompactBlock sending value to the address
let fake_block_hash = BlockHash([0; 32]);
let fake_block_height = sapling_activation_height();
let (cb, _) = fake_compact_block( let (cb, _) = fake_compact_block(
fake_block_height, sapling_activation_height(),
fake_block_hash, BlockHash([0; 32]),
&dfvk, &dfvk,
AddressType::DefaultExternal, AddressType::DefaultExternal,
Amount::from_u64(5).unwrap(), Amount::from_u64(5).unwrap(),
@ -348,17 +350,20 @@ mod tests {
Amount::from_u64(7).unwrap(), Amount::from_u64(7).unwrap(),
1, 1,
); );
insert_into_cache(&db_cache, &cb2); insert_into_cache(&db_cache, &cb2);
// Scan the cache again // Scanning should detect no inconsistencies
scan_cached_blocks( assert_matches!(
&tests::network(), scan_cached_blocks(
&db_cache, &tests::network(),
&mut db_data, &db_cache,
sapling_activation_height() + 1, &mut db_data,
1, sapling_activation_height() + 1,
) 1,
.unwrap(); ),
Ok(())
);
} }
#[test] #[test]
@ -394,15 +399,17 @@ mod tests {
insert_into_cache(&db_cache, &cb); insert_into_cache(&db_cache, &cb);
insert_into_cache(&db_cache, &cb2); insert_into_cache(&db_cache, &cb2);
// Scan the cache // Scanning the cache should find no inconsistencies
scan_cached_blocks( assert_matches!(
&tests::network(), scan_cached_blocks(
&db_cache, &tests::network(),
&mut db_data, &db_cache,
sapling_activation_height(), &mut db_data,
2, sapling_activation_height(),
) 2,
.unwrap(); ),
Ok(())
);
// Create more fake CompactBlocks that don't connect to the scanned ones // Create more fake CompactBlocks that don't connect to the scanned ones
let (cb3, _) = fake_compact_block( let (cb3, _) = fake_compact_block(
@ -433,83 +440,8 @@ mod tests {
sapling_activation_height() + 2, sapling_activation_height() + 2,
2 2
), ),
Err(_) // FIXME: check error result more closely Err(Error::Scan(ScanError::PrevHashMismatch { at_height }))
); if at_height == sapling_activation_height() + 2
}
#[test]
fn invalid_chain_cache_reorg() {
let cache_file = NamedTempFile::new().unwrap();
let db_cache = BlockDb::for_path(cache_file.path()).unwrap();
init_cache_database(&db_cache).unwrap();
let data_file = NamedTempFile::new().unwrap();
let mut db_data = WalletDb::for_path(data_file.path(), tests::network()).unwrap();
init_wallet_db(&mut db_data, Some(Secret::new(vec![]))).unwrap();
// Add an account to the wallet
let (dfvk, _taddr) = init_test_accounts_table(&mut db_data);
// Create some fake CompactBlocks
let (cb, _) = fake_compact_block(
sapling_activation_height(),
BlockHash([0; 32]),
&dfvk,
AddressType::DefaultExternal,
Amount::from_u64(5).unwrap(),
0,
);
let (cb2, _) = fake_compact_block(
sapling_activation_height() + 1,
cb.hash(),
&dfvk,
AddressType::DefaultExternal,
Amount::from_u64(7).unwrap(),
1,
);
insert_into_cache(&db_cache, &cb);
insert_into_cache(&db_cache, &cb2);
// Scan the cache
scan_cached_blocks(
&tests::network(),
&db_cache,
&mut db_data,
sapling_activation_height(),
2,
)
.unwrap();
// Create more fake CompactBlocks that contain a reorg
let (cb3, _) = fake_compact_block(
sapling_activation_height() + 2,
cb2.hash(),
&dfvk,
AddressType::DefaultExternal,
Amount::from_u64(8).unwrap(),
2,
);
let (cb4, _) = fake_compact_block(
sapling_activation_height() + 3,
BlockHash([1; 32]),
&dfvk,
AddressType::DefaultExternal,
Amount::from_u64(3).unwrap(),
3,
);
insert_into_cache(&db_cache, &cb3);
insert_into_cache(&db_cache, &cb4);
// Data+cache chain should be invalid inside the cache
assert_matches!(
scan_cached_blocks(
&tests::network(),
&db_cache,
&mut db_data,
sapling_activation_height() + 2,
2
),
Err(_) // FIXME: check error result more closely
); );
} }

View File

@ -36,7 +36,6 @@ use either::Either;
use rusqlite::{self, Connection}; use rusqlite::{self, Connection};
use secrecy::{ExposeSecret, SecretVec}; use secrecy::{ExposeSecret, SecretVec};
use std::{borrow::Borrow, collections::HashMap, convert::AsRef, fmt, io, ops::Range, path::Path}; use std::{borrow::Borrow, collections::HashMap, convert::AsRef, fmt, io, ops::Range, path::Path};
use wallet::commitment_tree::put_shard_roots;
use incrementalmerkletree::Position; use incrementalmerkletree::Position;
use shardtree::{ShardTree, ShardTreeError}; use shardtree::{ShardTree, ShardTreeError};
@ -58,6 +57,7 @@ use zcash_client_backend::{
data_api::{ data_api::{
self, self,
chain::{BlockSource, CommitmentTreeRoot}, chain::{BlockSource, CommitmentTreeRoot},
scanning::{ScanPriority, ScanRange},
BlockMetadata, DecryptedTransaction, NullifierQuery, PoolType, Recipient, ScannedBlock, BlockMetadata, DecryptedTransaction, NullifierQuery, PoolType, Recipient, ScannedBlock,
SentTransaction, ShieldedProtocol, WalletCommitmentTrees, WalletRead, WalletWrite, SentTransaction, ShieldedProtocol, WalletCommitmentTrees, WalletRead, WalletWrite,
SAPLING_SHARD_HEIGHT, SAPLING_SHARD_HEIGHT,
@ -80,13 +80,18 @@ use {
pub mod chain; pub mod chain;
pub mod error; pub mod error;
pub mod serialization; pub mod serialization;
pub mod wallet; pub mod wallet;
use wallet::commitment_tree::put_shard_roots;
/// The maximum number of blocks the wallet is allowed to rewind. This is /// The maximum number of blocks the wallet is allowed to rewind. This is
/// consistent with the bound in zcashd, and allows block data deeper than /// consistent with the bound in zcashd, and allows block data deeper than
/// this delta from the chain tip to be pruned. /// this delta from the chain tip to be pruned.
pub(crate) const PRUNING_DEPTH: u32 = 100; pub(crate) const PRUNING_DEPTH: u32 = 100;
/// The number of blocks to verify ahead when the chain tip is updated.
pub(crate) const VERIFY_LOOKAHEAD: u32 = 10;
pub(crate) const SAPLING_TABLES_PREFIX: &str = "sapling"; pub(crate) const SAPLING_TABLES_PREFIX: &str = "sapling";
/// A newtype wrapper for sqlite primary key values for the notes /// A newtype wrapper for sqlite primary key values for the notes
@ -167,12 +172,9 @@ impl<C: Borrow<rusqlite::Connection>, P: consensus::Parameters> WalletRead for W
wallet::block_fully_scanned(self.conn.borrow()) wallet::block_fully_scanned(self.conn.borrow())
} }
fn suggest_scan_ranges( fn suggest_scan_ranges(&self) -> Result<Vec<ScanRange>, Self::Error> {
&self, wallet::scanning::suggest_scan_ranges(self.conn.borrow(), ScanPriority::Historic)
_batch_size: usize, .map_err(SqliteClientError::from)
_limit: usize,
) -> Result<Vec<Range<BlockHeight>>, Self::Error> {
todo!()
} }
fn get_min_unspent_height(&self) -> Result<Option<BlockHeight>, Self::Error> { fn get_min_unspent_height(&self) -> Result<Option<BlockHeight>, Self::Error> {
@ -401,18 +403,24 @@ impl<P: consensus::Parameters> WalletWrite for WalletDb<rusqlite::Connection, P>
blocks: Vec<ScannedBlock<sapling::Nullifier>>, blocks: Vec<ScannedBlock<sapling::Nullifier>>,
) -> Result<Vec<Self::NoteRef>, Self::Error> { ) -> Result<Vec<Self::NoteRef>, Self::Error> {
self.transactionally(|wdb| { self.transactionally(|wdb| {
let start_position = blocks.first().map(|block| { let start_positions = blocks.first().map(|block| {
Position::from( (
u64::from(block.metadata().sapling_tree_size()) block.height(),
- u64::try_from(block.sapling_commitments().len()).unwrap(), Position::from(
u64::from(block.metadata().sapling_tree_size())
- u64::try_from(block.sapling_commitments().len()).unwrap(),
),
) )
}); });
let mut wallet_note_ids = vec![]; let mut wallet_note_ids = vec![];
let mut sapling_commitments = vec![]; let mut sapling_commitments = vec![];
let mut end_height = None; let mut last_scanned_height = None;
let mut note_positions = vec![];
for block in blocks.into_iter() { for block in blocks.into_iter() {
if end_height.iter().any(|prev| block.height() != *prev + 1) { if last_scanned_height
.iter()
.any(|prev| block.height() != *prev + 1)
{
return Err(SqliteClientError::NonSequentialBlocks); return Err(SqliteClientError::NonSequentialBlocks);
} }
@ -442,13 +450,21 @@ impl<P: consensus::Parameters> WalletWrite for WalletDb<rusqlite::Connection, P>
} }
} }
end_height = Some(block.height()); note_positions.extend(block.transactions().iter().flat_map(|wtx| {
wtx.sapling_outputs
.iter()
.map(|out| out.note_commitment_tree_position())
}));
last_scanned_height = Some(block.height());
sapling_commitments.extend(block.into_sapling_commitments().into_iter()); sapling_commitments.extend(block.into_sapling_commitments().into_iter());
} }
// We will have a start position and an end height in all cases where `blocks` is // We will have a start position and a last scanned height in all cases where
// non-empty. // `blocks` is non-empty.
if let Some((start_position, end_height)) = start_position.zip(end_height) { if let Some(((start_height, start_position), last_scanned_height)) =
start_positions.zip(last_scanned_height)
{
// Update the Sapling note commitment tree with all newly read note commitments // Update the Sapling note commitment tree with all newly read note commitments
let mut sapling_commitments = sapling_commitments.into_iter(); let mut sapling_commitments = sapling_commitments.into_iter();
wdb.with_sapling_tree_mut::<_, _, SqliteClientError>(move |sapling_tree| { wdb.with_sapling_tree_mut::<_, _, SqliteClientError>(move |sapling_tree| {
@ -457,13 +473,30 @@ impl<P: consensus::Parameters> WalletWrite for WalletDb<rusqlite::Connection, P>
})?; })?;
// Update now-expired transactions that didn't get mined. // Update now-expired transactions that didn't get mined.
wallet::update_expired_notes(wdb.conn.0, end_height)?; wallet::update_expired_notes(wdb.conn.0, last_scanned_height)?;
wallet::scanning::scan_complete(
wdb.conn.0,
&wdb.params,
Range {
start: start_height,
end: last_scanned_height + 1,
},
&note_positions,
)?;
} }
Ok(wallet_note_ids) Ok(wallet_note_ids)
}) })
} }
fn update_chain_tip(&mut self, tip_height: BlockHeight) -> Result<(), Self::Error> {
let tx = self.conn.transaction()?;
wallet::scanning::update_chain_tip(&tx, &self.params, tip_height)?;
tx.commit()?;
Ok(())
}
fn store_decrypted_tx( fn store_decrypted_tx(
&mut self, &mut self,
d_tx: DecryptedTransaction, d_tx: DecryptedTransaction,
@ -750,7 +783,7 @@ impl BlockSource for BlockDb {
fn with_blocks<F, DbErrT>( fn with_blocks<F, DbErrT>(
&self, &self,
from_height: Option<BlockHeight>, from_height: Option<BlockHeight>,
limit: Option<u32>, limit: Option<usize>,
with_row: F, with_row: F,
) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>> ) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>
where where
@ -928,7 +961,7 @@ impl BlockSource for FsBlockDb {
fn with_blocks<F, DbErrT>( fn with_blocks<F, DbErrT>(
&self, &self,
from_height: Option<BlockHeight>, from_height: Option<BlockHeight>,
limit: Option<u32>, limit: Option<usize>,
with_row: F, with_row: F,
) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>> ) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>
where where

View File

@ -4,7 +4,7 @@ use byteorder::{ReadBytesExt, WriteBytesExt};
use core::ops::Deref; use core::ops::Deref;
use shardtree::{Node, PrunableTree, RetentionFlags, Tree}; use shardtree::{Node, PrunableTree, RetentionFlags, Tree};
use std::io::{self, Read, Write}; use std::io::{self, Read, Write};
use std::rc::Rc; use std::sync::Arc;
use zcash_encoding::Optional; use zcash_encoding::Optional;
use zcash_primitives::merkle_tree::HashSer; use zcash_primitives::merkle_tree::HashSer;
@ -53,7 +53,7 @@ pub fn write_shard<H: HashSer, W: Write>(writer: &mut W, tree: &PrunableTree<H>)
fn read_shard_v1<H: HashSer, R: Read>(mut reader: &mut R) -> io::Result<PrunableTree<H>> { fn read_shard_v1<H: HashSer, R: Read>(mut reader: &mut R) -> io::Result<PrunableTree<H>> {
match reader.read_u8()? { match reader.read_u8()? {
PARENT_TAG => { PARENT_TAG => {
let ann = Optional::read(&mut reader, <H as HashSer>::read)?.map(Rc::new); let ann = Optional::read(&mut reader, <H as HashSer>::read)?.map(Arc::new);
let left = read_shard_v1(reader)?; let left = read_shard_v1(reader)?;
let right = read_shard_v1(reader)?; let right = read_shard_v1(reader)?;
Ok(Tree::parent(ann, left, right)) Ok(Tree::parent(ann, left, right))

View File

@ -68,6 +68,8 @@ use rusqlite::{self, named_params, OptionalExtension, ToSql};
use std::collections::HashMap; use std::collections::HashMap;
use std::convert::TryFrom; use std::convert::TryFrom;
use std::io::{self, Cursor}; use std::io::{self, Cursor};
use zcash_client_backend::data_api::scanning::{ScanPriority, ScanRange};
use zcash_client_backend::data_api::ShieldedProtocol; use zcash_client_backend::data_api::ShieldedProtocol;
use zcash_primitives::transaction::TransactionData; use zcash_primitives::transaction::TransactionData;
@ -91,10 +93,13 @@ use zcash_client_backend::{
wallet::WalletTx, wallet::WalletTx,
}; };
use crate::VERIFY_LOOKAHEAD;
use crate::{ use crate::{
error::SqliteClientError, SqlTransaction, WalletCommitmentTrees, WalletDb, PRUNING_DEPTH, error::SqliteClientError, SqlTransaction, WalletCommitmentTrees, WalletDb, PRUNING_DEPTH,
}; };
use self::scanning::replace_queue_entries;
#[cfg(feature = "transparent-inputs")] #[cfg(feature = "transparent-inputs")]
use { use {
crate::UtxoId, crate::UtxoId,
@ -109,6 +114,7 @@ use {
pub(crate) mod commitment_tree; pub(crate) mod commitment_tree;
pub mod init; pub mod init;
pub(crate) mod sapling; pub(crate) mod sapling;
pub(crate) mod scanning;
pub(crate) const BLOCK_SAPLING_FRONTIER_ABSENT: &[u8] = &[0x0]; pub(crate) const BLOCK_SAPLING_FRONTIER_ABSENT: &[u8] = &[0x0];
@ -629,8 +635,8 @@ pub(crate) fn block_metadata(
) -> Result<Option<BlockMetadata>, SqliteClientError> { ) -> Result<Option<BlockMetadata>, SqliteClientError> {
conn.query_row( conn.query_row(
"SELECT height, hash, sapling_commitment_tree_size, sapling_tree "SELECT height, hash, sapling_commitment_tree_size, sapling_tree
FROM blocks FROM blocks
WHERE height = :block_height", WHERE height = :block_height",
named_params![":block_height": u32::from(block_height)], named_params![":block_height": u32::from(block_height)],
|row| { |row| {
let height: u32 = row.get(0)?; let height: u32 = row.get(0)?;
@ -742,7 +748,7 @@ pub(crate) fn truncate_to_height<P: consensus::Parameters>(
) -> Result<(), SqliteClientError> { ) -> Result<(), SqliteClientError> {
let sapling_activation_height = params let sapling_activation_height = params
.activation_height(NetworkUpgrade::Sapling) .activation_height(NetworkUpgrade::Sapling)
.expect("Sapling activation height mutst be available."); .expect("Sapling activation height must be available.");
// Recall where we synced up to previously. // Recall where we synced up to previously.
let last_scanned_height = conn.query_row("SELECT MAX(height) FROM blocks", [], |row| { let last_scanned_height = conn.query_row("SELECT MAX(height) FROM blocks", [], |row| {
@ -800,7 +806,8 @@ pub(crate) fn truncate_to_height<P: consensus::Parameters>(
// Un-mine transactions. // Un-mine transactions.
conn.execute( conn.execute(
"UPDATE transactions SET block = NULL, tx_index = NULL WHERE block IS NOT NULL AND block > ?", "UPDATE transactions SET block = NULL, tx_index = NULL
WHERE block IS NOT NULL AND block > ?",
[u32::from(block_height)], [u32::from(block_height)],
)?; )?;
@ -809,6 +816,27 @@ pub(crate) fn truncate_to_height<P: consensus::Parameters>(
"DELETE FROM blocks WHERE height > ?", "DELETE FROM blocks WHERE height > ?",
[u32::from(block_height)], [u32::from(block_height)],
)?; )?;
// Delete from the scanning queue any range with a start height greater than the
// truncation height, and then truncate any remaining range by setting the end
// equal to the truncation height + 1.
conn.execute(
"DELETE FROM scan_queue
WHERE block_range_start > :block_height",
named_params![":block_height": u32::from(block_height)],
)?;
conn.execute(
"UPDATE scan_queue
SET block_range_end = :end_height
WHERE block_range_end > :end_height",
named_params![":end_height": u32::from(block_height + 1)],
)?;
// Prioritize the range starting at the height we just rewound to for verification
let query_range = block_height..(block_height + VERIFY_LOOKAHEAD);
let scan_range = ScanRange::from_parts(query_range.clone(), ScanPriority::Verify);
replace_queue_entries(conn, &query_range, Some(scan_range).into_iter())?;
} }
Ok(()) Ok(())

View File

@ -4,7 +4,7 @@ use std::{
collections::BTreeSet, collections::BTreeSet,
io::{self, Cursor}, io::{self, Cursor},
marker::PhantomData, marker::PhantomData,
rc::Rc, sync::Arc,
}; };
use zcash_client_backend::data_api::chain::CommitmentTreeRoot; use zcash_client_backend::data_api::chain::CommitmentTreeRoot;
@ -281,7 +281,7 @@ pub(crate) fn get_shard<H: HashSer>(
let located_tree = LocatedPrunableTree::from_parts(shard_root_addr, shard_tree); let located_tree = LocatedPrunableTree::from_parts(shard_root_addr, shard_tree);
if let Some(root_hash_data) = root_hash { if let Some(root_hash_data) = root_hash {
let root_hash = H::read(Cursor::new(root_hash_data)).map_err(Either::Left)?; let root_hash = H::read(Cursor::new(root_hash_data)).map_err(Either::Left)?;
Ok(located_tree.reannotate_root(Some(Rc::new(root_hash)))) Ok(located_tree.reannotate_root(Some(Arc::new(root_hash))))
} else { } else {
Ok(located_tree) Ok(located_tree)
} }
@ -772,12 +772,44 @@ pub(crate) fn put_shard_roots<
return Ok(()); return Ok(());
} }
// We treat the cap as a DEPTH-SHARD_HEIGHT tree so that we can make a batch insertion of // We treat the cap as a tree with `DEPTH - SHARD_HEIGHT` levels, so that we can make a
// root data using `Position::from(start_index)` as the starting position and treating the // batch insertion of root data using `Position::from(start_index)` as the starting position
// roots as level-0 leaves. // and treating the roots as level-0 leaves.
#[derive(Clone, Debug, PartialEq, Eq)]
struct LevelShifter<H, const SHARD_HEIGHT: u8>(H);
impl<H: Hashable, const SHARD_HEIGHT: u8> Hashable for LevelShifter<H, SHARD_HEIGHT> {
fn empty_leaf() -> Self {
Self(H::empty_root(SHARD_HEIGHT.into()))
}
fn combine(level: Level, a: &Self, b: &Self) -> Self {
Self(H::combine(level + SHARD_HEIGHT, &a.0, &b.0))
}
fn empty_root(level: Level) -> Self
where
Self: Sized,
{
Self(H::empty_root(level + SHARD_HEIGHT))
}
}
impl<H: HashSer, const SHARD_HEIGHT: u8> HashSer for LevelShifter<H, SHARD_HEIGHT> {
fn read<R: io::Read>(reader: R) -> io::Result<Self>
where
Self: Sized,
{
H::read(reader).map(Self)
}
fn write<W: io::Write>(&self, writer: W) -> io::Result<()> {
self.0.write(writer)
}
}
let cap = LocatedTree::from_parts( let cap = LocatedTree::from_parts(
Address::from_parts((DEPTH - SHARD_HEIGHT).into(), 0), Address::from_parts((DEPTH - SHARD_HEIGHT).into(), 0),
get_cap(conn, table_prefix).map_err(ShardTreeError::Storage)?, get_cap::<LevelShifter<H, SHARD_HEIGHT>>(conn, table_prefix)
.map_err(ShardTreeError::Storage)?,
); );
let cap_result = cap let cap_result = cap
@ -785,7 +817,7 @@ pub(crate) fn put_shard_roots<
Position::from(start_index), Position::from(start_index),
roots.iter().map(|r| { roots.iter().map(|r| {
( (
r.root_hash().clone(), LevelShifter(r.root_hash().clone()),
Retention::Checkpoint { Retention::Checkpoint {
id: (), id: (),
is_marked: false, is_marked: false,

View File

@ -2,6 +2,7 @@
use either::Either; use either::Either;
use incrementalmerkletree::Retention; use incrementalmerkletree::Retention;
use std::{collections::HashMap, fmt, io}; use std::{collections::HashMap, fmt, io};
use tracing::debug;
use rusqlite::{self, types::ToSql}; use rusqlite::{self, types::ToSql};
use schemer::{Migrator, MigratorError}; use schemer::{Migrator, MigratorError};
@ -318,6 +319,7 @@ pub fn init_blocks_table<P: consensus::Parameters>(
)?; )?;
if let Some(nonempty_frontier) = block_end_tree.to_frontier().value() { if let Some(nonempty_frontier) = block_end_tree.to_frontier().value() {
debug!("Inserting frontier into ShardTree: {:?}", nonempty_frontier);
let shard_store = let shard_store =
SqliteShardStore::<_, sapling::Node, SAPLING_SHARD_HEIGHT>::from_connection( SqliteShardStore::<_, sapling::Node, SAPLING_SHARD_HEIGHT>::from_connection(
wdb.conn.0, wdb.conn.0,
@ -463,6 +465,16 @@ mod tests {
FOREIGN KEY (block) REFERENCES blocks(height), FOREIGN KEY (block) REFERENCES blocks(height),
CONSTRAINT witness_height UNIQUE (note, block) CONSTRAINT witness_height UNIQUE (note, block)
)", )",
"CREATE TABLE scan_queue (
block_range_start INTEGER NOT NULL,
block_range_end INTEGER NOT NULL,
priority INTEGER NOT NULL,
CONSTRAINT range_start_uniq UNIQUE (block_range_start),
CONSTRAINT range_end_uniq UNIQUE (block_range_end),
CONSTRAINT range_bounds_order CHECK (
block_range_start < block_range_end
)
)",
"CREATE TABLE schemer_migrations ( "CREATE TABLE schemer_migrations (
id blob PRIMARY KEY id blob PRIMARY KEY
)", )",

View File

@ -8,10 +8,14 @@ use incrementalmerkletree::Retention;
use rusqlite::{self, named_params, params}; use rusqlite::{self, named_params, params};
use schemer; use schemer;
use schemer_rusqlite::RusqliteMigration; use schemer_rusqlite::RusqliteMigration;
use shardtree::ShardTree; use shardtree::{caching::CachingShardStore, ShardTree, ShardTreeError};
use tracing::{debug, trace};
use uuid::Uuid; use uuid::Uuid;
use zcash_client_backend::data_api::SAPLING_SHARD_HEIGHT; use zcash_client_backend::data_api::{
scanning::{ScanPriority, ScanRange},
SAPLING_SHARD_HEIGHT,
};
use zcash_primitives::{ use zcash_primitives::{
consensus::BlockHeight, consensus::BlockHeight,
merkle_tree::{read_commitment_tree, read_incremental_witness}, merkle_tree::{read_commitment_tree, read_incremental_witness},
@ -20,8 +24,10 @@ use zcash_primitives::{
use crate::{ use crate::{
wallet::{ wallet::{
block_height_extrema,
commitment_tree::SqliteShardStore, commitment_tree::SqliteShardStore,
init::{migrations::received_notes_nullable_nf, WalletMigrationError}, init::{migrations::received_notes_nullable_nf, WalletMigrationError},
scanning::insert_queue_entries,
}, },
PRUNING_DEPTH, SAPLING_TABLES_PREFIX, PRUNING_DEPTH, SAPLING_TABLES_PREFIX,
}; };
@ -56,6 +62,7 @@ impl RusqliteMigration for Migration {
fn up(&self, transaction: &rusqlite::Transaction) -> Result<(), WalletMigrationError> { fn up(&self, transaction: &rusqlite::Transaction) -> Result<(), WalletMigrationError> {
// Add commitment tree sizes to block metadata. // Add commitment tree sizes to block metadata.
debug!("Adding new columns");
transaction.execute_batch( transaction.execute_batch(
"ALTER TABLE blocks ADD COLUMN sapling_commitment_tree_size INTEGER; "ALTER TABLE blocks ADD COLUMN sapling_commitment_tree_size INTEGER;
ALTER TABLE blocks ADD COLUMN orchard_commitment_tree_size INTEGER; ALTER TABLE blocks ADD COLUMN orchard_commitment_tree_size INTEGER;
@ -63,6 +70,7 @@ impl RusqliteMigration for Migration {
)?; )?;
// Add shard persistence // Add shard persistence
debug!("Creating tables for shard persistence");
transaction.execute_batch( transaction.execute_batch(
"CREATE TABLE sapling_tree_shards ( "CREATE TABLE sapling_tree_shards (
shard_index INTEGER PRIMARY KEY, shard_index INTEGER PRIMARY KEY,
@ -81,6 +89,7 @@ impl RusqliteMigration for Migration {
)?; )?;
// Add checkpoint persistence // Add checkpoint persistence
debug!("Creating tables for checkpoint persistence");
transaction.execute_batch( transaction.execute_batch(
"CREATE TABLE sapling_tree_checkpoints ( "CREATE TABLE sapling_tree_checkpoints (
checkpoint_id INTEGER PRIMARY KEY, checkpoint_id INTEGER PRIMARY KEY,
@ -99,6 +108,7 @@ impl RusqliteMigration for Migration {
transaction, transaction,
SAPLING_TABLES_PREFIX, SAPLING_TABLES_PREFIX,
)?; )?;
let shard_store = CachingShardStore::load(shard_store).map_err(ShardTreeError::Storage)?;
let mut shard_tree: ShardTree< let mut shard_tree: ShardTree<
_, _,
{ sapling::NOTE_COMMITMENT_TREE_DEPTH }, { sapling::NOTE_COMMITMENT_TREE_DEPTH },
@ -128,22 +138,42 @@ impl RusqliteMigration for Migration {
) )
})?; })?;
if block_height % 1000 == 0 {
debug!(height = block_height, "Migrating tree data to shardtree");
}
trace!(
height = block_height,
size = block_end_tree.size(),
"Storing Sapling commitment tree size"
);
stmt_update_block_sapling_tree_size stmt_update_block_sapling_tree_size
.execute(params![block_end_tree.size(), block_height])?; .execute(params![block_end_tree.size(), block_height])?;
if let Some(nonempty_frontier) = block_end_tree.to_frontier().value() { if let Some(nonempty_frontier) = block_end_tree.to_frontier().value() {
shard_tree.insert_frontier_nodes( trace!(
nonempty_frontier.clone(), height = block_height,
Retention::Checkpoint { frontier = ?nonempty_frontier,
id: BlockHeight::from(block_height), "Inserting frontier nodes",
is_marked: false, );
}, shard_tree
)?; .insert_frontier_nodes(
nonempty_frontier.clone(),
Retention::Checkpoint {
id: BlockHeight::from(block_height),
is_marked: false,
},
)
.map_err(|e| match e {
ShardTreeError::Query(e) => ShardTreeError::Query(e),
ShardTreeError::Insert(e) => ShardTreeError::Insert(e),
ShardTreeError::Storage(_) => unreachable!(),
})?
} }
} }
} }
// Insert all the tree information that we can get from existing incremental witnesses // Insert all the tree information that we can get from existing incremental witnesses
debug!("Migrating witness data to shardtree");
{ {
let mut stmt_blocks = let mut stmt_blocks =
transaction.prepare("SELECT note, block, witness FROM sapling_witnesses")?; transaction.prepare("SELECT note, block, witness FROM sapling_witnesses")?;
@ -180,10 +210,50 @@ impl RusqliteMigration for Migration {
updated_note_positions.insert(witnessed_position); updated_note_positions.insert(witnessed_position);
} }
shard_tree.insert_witness_nodes(witness, BlockHeight::from(block_height))?; shard_tree
.insert_witness_nodes(witness, BlockHeight::from(block_height))
.map_err(|e| match e {
ShardTreeError::Query(e) => ShardTreeError::Query(e),
ShardTreeError::Insert(e) => ShardTreeError::Insert(e),
ShardTreeError::Storage(_) => unreachable!(),
})?;
} }
} }
shard_tree
.into_store()
.flush()
.map_err(ShardTreeError::Storage)?;
// Establish the scan queue & wallet history table.
// block_range_end is exclusive.
debug!("Creating table for scan queue");
transaction.execute_batch(
"CREATE TABLE scan_queue (
block_range_start INTEGER NOT NULL,
block_range_end INTEGER NOT NULL,
priority INTEGER NOT NULL,
CONSTRAINT range_start_uniq UNIQUE (block_range_start),
CONSTRAINT range_end_uniq UNIQUE (block_range_end),
CONSTRAINT range_bounds_order CHECK (
block_range_start < block_range_end
)
);",
)?;
if let Some((start, end)) = block_height_extrema(transaction)? {
// `ScanRange` uses an exclusive upper bound.
let chain_end = end + 1;
insert_queue_entries(
transaction,
Some(ScanRange::from_parts(
start..chain_end,
ScanPriority::Historic,
))
.iter(),
)?;
}
Ok(()) Ok(())
} }

File diff suppressed because it is too large Load Diff

View File

@ -8,6 +8,7 @@ and this library adheres to Rust's notion of
## [Unreleased] ## [Unreleased]
### Added ### Added
- `zcash_primitives::consensus::BlockHeight::saturating_sub`
- `zcash_primitives::transaction::builder`: - `zcash_primitives::transaction::builder`:
- `Builder::add_orchard_spend` - `Builder::add_orchard_spend`
- `Builder::add_orchard_output` - `Builder::add_orchard_output`

View File

@ -23,6 +23,12 @@ impl BlockHeight {
pub const fn from_u32(v: u32) -> BlockHeight { pub const fn from_u32(v: u32) -> BlockHeight {
BlockHeight(v) BlockHeight(v)
} }
/// Subtracts the provided value from this height, returning `H0` if this would result in
/// underflow of the wrapped `u32`.
pub fn saturating_sub(self, v: u32) -> BlockHeight {
BlockHeight(self.0.saturating_sub(v))
}
} }
impl fmt::Display for BlockHeight { impl fmt::Display for BlockHeight {

View File

@ -2,6 +2,8 @@ use bitvec::{order::Lsb0, view::AsBits};
use group::{ff::PrimeField, Curve}; use group::{ff::PrimeField, Curve};
use incrementalmerkletree::{Hashable, Level}; use incrementalmerkletree::{Hashable, Level};
use lazy_static::lazy_static; use lazy_static::lazy_static;
use std::fmt;
use std::io::{self, Read, Write}; use std::io::{self, Read, Write};
use super::{ use super::{
@ -64,11 +66,19 @@ pub fn merkle_hash(depth: usize, lhs: &[u8; 32], rhs: &[u8; 32]) -> [u8; 32] {
} }
/// A node within the Sapling commitment tree. /// A node within the Sapling commitment tree.
#[derive(Clone, Copy, Debug, PartialEq, Eq)] #[derive(Clone, Copy, PartialEq, Eq)]
pub struct Node { pub struct Node {
pub(super) repr: [u8; 32], pub(super) repr: [u8; 32],
} }
impl fmt::Debug for Node {
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
f.debug_struct("Node")
.field("repr", &hex::encode(self.repr))
.finish()
}
}
impl Node { impl Node {
#[cfg(test)] #[cfg(test)]
pub(crate) fn new(repr: [u8; 32]) -> Self { pub(crate) fn new(repr: [u8; 32]) -> Self {