1223 lines
46 KiB
Rust
1223 lines
46 KiB
Rust
//! *An SQLite-based Zcash light client.*
|
|
//!
|
|
//! `zcash_client_sqlite` contains complete SQLite-based implementations of the [`WalletRead`],
|
|
//! [`WalletWrite`], and [`BlockSource`] traits from the [`zcash_client_backend`] crate. In
|
|
//! combination with [`zcash_client_backend`], it provides a full implementation of a SQLite-backed
|
|
//! client for the Zcash network.
|
|
//!
|
|
//! # Design
|
|
//!
|
|
//! The light client is built around two SQLite databases:
|
|
//!
|
|
//! - A cache database, used to inform the light client about new [`CompactBlock`]s. It is
|
|
//! read-only within all light client APIs *except* for [`init_cache_database`] which
|
|
//! can be used to initialize the database.
|
|
//!
|
|
//! - A data database, where the light client's state is stored. It is read-write within
|
|
//! the light client APIs, and **assumed to be read-only outside these APIs**. Callers
|
|
//! **MUST NOT** write to the database without using these APIs. Callers **MAY** read
|
|
//! the database directly in order to extract information for display to users.
|
|
//!
|
|
//! ## Feature flags
|
|
#![doc = document_features::document_features!()]
|
|
//!
|
|
//! [`WalletRead`]: zcash_client_backend::data_api::WalletRead
|
|
//! [`WalletWrite`]: zcash_client_backend::data_api::WalletWrite
|
|
//! [`BlockSource`]: zcash_client_backend::data_api::chain::BlockSource
|
|
//! [`CompactBlock`]: zcash_client_backend::proto::compact_formats::CompactBlock
|
|
//! [`init_cache_database`]: crate::chain::init::init_cache_database
|
|
|
|
#![cfg_attr(docsrs, feature(doc_cfg))]
|
|
#![cfg_attr(docsrs, feature(doc_auto_cfg))]
|
|
// Catch documentation errors caused by code changes.
|
|
#![deny(rustdoc::broken_intra_doc_links)]
|
|
|
|
use maybe_rayon::{
|
|
prelude::{IndexedParallelIterator, ParallelIterator},
|
|
slice::ParallelSliceMut,
|
|
};
|
|
use rusqlite::{self, Connection};
|
|
use secrecy::{ExposeSecret, SecretVec};
|
|
use std::{
|
|
borrow::Borrow, collections::HashMap, convert::AsRef, fmt, num::NonZeroU32, ops::Range,
|
|
path::Path,
|
|
};
|
|
|
|
use incrementalmerkletree::Position;
|
|
use shardtree::{error::ShardTreeError, ShardTree};
|
|
use zcash_primitives::{
|
|
block::BlockHash,
|
|
consensus::{self, BlockHeight},
|
|
memo::{Memo, MemoBytes},
|
|
transaction::{
|
|
components::amount::{Amount, NonNegativeAmount},
|
|
Transaction, TxId,
|
|
},
|
|
zip32::{AccountId, DiversifierIndex, Scope},
|
|
};
|
|
|
|
use zcash_client_backend::{
|
|
address::UnifiedAddress,
|
|
data_api::{
|
|
self,
|
|
chain::{BlockSource, CommitmentTreeRoot},
|
|
scanning::{ScanPriority, ScanRange},
|
|
AccountBirthday, BlockMetadata, DecryptedTransaction, InputSource, NullifierQuery,
|
|
ScannedBlock, SentTransaction, WalletCommitmentTrees, WalletRead, WalletSummary,
|
|
WalletWrite, SAPLING_SHARD_HEIGHT,
|
|
},
|
|
keys::{UnifiedAddressRequest, UnifiedFullViewingKey, UnifiedSpendingKey},
|
|
proto::compact_formats::CompactBlock,
|
|
wallet::{Note, NoteId, ReceivedNote, Recipient, WalletTransparentOutput},
|
|
DecryptedOutput, PoolType, ShieldedProtocol, TransferType,
|
|
};
|
|
|
|
use crate::{error::SqliteClientError, wallet::commitment_tree::SqliteShardStore};
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
use {
|
|
zcash_client_backend::wallet::TransparentAddressMetadata,
|
|
zcash_primitives::{legacy::TransparentAddress, transaction::components::OutPoint},
|
|
};
|
|
|
|
#[cfg(feature = "unstable")]
|
|
use {
|
|
crate::chain::{fsblockdb_with_blocks, BlockMeta},
|
|
std::path::PathBuf,
|
|
std::{fs, io},
|
|
};
|
|
|
|
pub mod chain;
|
|
pub mod error;
|
|
|
|
pub mod wallet;
|
|
use wallet::{
|
|
commitment_tree::{self, put_shard_roots},
|
|
SubtreeScanProgress,
|
|
};
|
|
|
|
#[cfg(test)]
|
|
mod testing;
|
|
|
|
/// The maximum number of blocks the wallet is allowed to rewind. This is
|
|
/// consistent with the bound in zcashd, and allows block data deeper than
|
|
/// this delta from the chain tip to be pruned.
|
|
pub(crate) const PRUNING_DEPTH: u32 = 100;
|
|
|
|
/// The number of blocks to verify ahead when the chain tip is updated.
|
|
pub(crate) const VERIFY_LOOKAHEAD: u32 = 10;
|
|
|
|
pub(crate) const SAPLING_TABLES_PREFIX: &str = "sapling";
|
|
|
|
#[cfg(not(feature = "transparent-inputs"))]
|
|
pub(crate) const UA_TRANSPARENT: bool = false;
|
|
#[cfg(feature = "transparent-inputs")]
|
|
pub(crate) const UA_TRANSPARENT: bool = true;
|
|
|
|
pub(crate) const DEFAULT_UA_REQUEST: UnifiedAddressRequest =
|
|
UnifiedAddressRequest::unsafe_new(false, true, UA_TRANSPARENT);
|
|
|
|
/// A newtype wrapper for received note identifiers.
|
|
#[derive(Debug, Copy, Clone, PartialEq, Eq, PartialOrd, Ord)]
|
|
pub struct ReceivedNoteId(pub(crate) i64);
|
|
|
|
impl fmt::Display for ReceivedNoteId {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
match self {
|
|
ReceivedNoteId(id) => write!(f, "Received Note {}", id),
|
|
}
|
|
}
|
|
}
|
|
|
|
/// A newtype wrapper for sqlite primary key values for the utxos
|
|
/// table.
|
|
#[derive(Debug, Copy, Clone, PartialEq, Eq)]
|
|
pub struct UtxoId(pub i64);
|
|
|
|
/// A wrapper for the SQLite connection to the wallet database.
|
|
pub struct WalletDb<C, P> {
|
|
conn: C,
|
|
params: P,
|
|
}
|
|
|
|
/// A wrapper for a SQLite transaction affecting the wallet database.
|
|
pub struct SqlTransaction<'conn>(pub(crate) &'conn rusqlite::Transaction<'conn>);
|
|
|
|
impl Borrow<rusqlite::Connection> for SqlTransaction<'_> {
|
|
fn borrow(&self) -> &rusqlite::Connection {
|
|
self.0
|
|
}
|
|
}
|
|
|
|
impl<P: consensus::Parameters + Clone> WalletDb<Connection, P> {
|
|
/// Construct a connection to the wallet database stored at the specified path.
|
|
pub fn for_path<F: AsRef<Path>>(path: F, params: P) -> Result<Self, rusqlite::Error> {
|
|
Connection::open(path).and_then(move |conn| {
|
|
rusqlite::vtab::array::load_module(&conn)?;
|
|
Ok(WalletDb { conn, params })
|
|
})
|
|
}
|
|
|
|
pub fn transactionally<F, A, E: From<rusqlite::Error>>(&mut self, f: F) -> Result<A, E>
|
|
where
|
|
F: FnOnce(&mut WalletDb<SqlTransaction<'_>, P>) -> Result<A, E>,
|
|
{
|
|
let tx = self.conn.transaction()?;
|
|
let mut wdb = WalletDb {
|
|
conn: SqlTransaction(&tx),
|
|
params: self.params.clone(),
|
|
};
|
|
let result = f(&mut wdb)?;
|
|
tx.commit()?;
|
|
Ok(result)
|
|
}
|
|
}
|
|
|
|
impl<C: Borrow<rusqlite::Connection>, P: consensus::Parameters> InputSource for WalletDb<C, P> {
|
|
type Error = SqliteClientError;
|
|
type NoteRef = ReceivedNoteId;
|
|
|
|
fn get_spendable_note(
|
|
&self,
|
|
txid: &TxId,
|
|
_protocol: ShieldedProtocol,
|
|
index: u32,
|
|
) -> Result<Option<ReceivedNote<Self::NoteRef, Note>>, Self::Error> {
|
|
wallet::sapling::get_spendable_sapling_note(self.conn.borrow(), &self.params, txid, index)
|
|
}
|
|
|
|
fn select_spendable_notes(
|
|
&self,
|
|
account: AccountId,
|
|
target_value: Amount,
|
|
_sources: &[ShieldedProtocol],
|
|
anchor_height: BlockHeight,
|
|
exclude: &[Self::NoteRef],
|
|
) -> Result<Vec<ReceivedNote<Self::NoteRef, Note>>, Self::Error> {
|
|
wallet::sapling::select_spendable_sapling_notes(
|
|
self.conn.borrow(),
|
|
&self.params,
|
|
account,
|
|
target_value,
|
|
anchor_height,
|
|
exclude,
|
|
)
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
fn get_unspent_transparent_output(
|
|
&self,
|
|
outpoint: &OutPoint,
|
|
) -> Result<Option<WalletTransparentOutput>, Self::Error> {
|
|
wallet::get_unspent_transparent_output(self.conn.borrow(), outpoint)
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
fn get_unspent_transparent_outputs(
|
|
&self,
|
|
address: &TransparentAddress,
|
|
max_height: BlockHeight,
|
|
exclude: &[OutPoint],
|
|
) -> Result<Vec<WalletTransparentOutput>, Self::Error> {
|
|
wallet::get_unspent_transparent_outputs(
|
|
self.conn.borrow(),
|
|
&self.params,
|
|
address,
|
|
max_height,
|
|
exclude,
|
|
)
|
|
}
|
|
}
|
|
|
|
impl<C: Borrow<rusqlite::Connection>, P: consensus::Parameters> WalletRead for WalletDb<C, P> {
|
|
type Error = SqliteClientError;
|
|
|
|
fn chain_height(&self) -> Result<Option<BlockHeight>, Self::Error> {
|
|
wallet::scan_queue_extrema(self.conn.borrow())
|
|
.map(|h| h.map(|range| *range.end()))
|
|
.map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn block_metadata(&self, height: BlockHeight) -> Result<Option<BlockMetadata>, Self::Error> {
|
|
wallet::block_metadata(self.conn.borrow(), &self.params, height)
|
|
}
|
|
|
|
fn block_fully_scanned(&self) -> Result<Option<BlockMetadata>, Self::Error> {
|
|
wallet::block_fully_scanned(self.conn.borrow(), &self.params)
|
|
}
|
|
|
|
fn block_max_scanned(&self) -> Result<Option<BlockMetadata>, Self::Error> {
|
|
wallet::block_max_scanned(self.conn.borrow(), &self.params)
|
|
}
|
|
|
|
fn suggest_scan_ranges(&self) -> Result<Vec<ScanRange>, Self::Error> {
|
|
wallet::scanning::suggest_scan_ranges(self.conn.borrow(), ScanPriority::Historic)
|
|
.map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_target_and_anchor_heights(
|
|
&self,
|
|
min_confirmations: NonZeroU32,
|
|
) -> Result<Option<(BlockHeight, BlockHeight)>, Self::Error> {
|
|
wallet::get_target_and_anchor_heights(self.conn.borrow(), min_confirmations)
|
|
.map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_min_unspent_height(&self) -> Result<Option<BlockHeight>, Self::Error> {
|
|
wallet::get_min_unspent_height(self.conn.borrow()).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_block_hash(&self, block_height: BlockHeight) -> Result<Option<BlockHash>, Self::Error> {
|
|
wallet::get_block_hash(self.conn.borrow(), block_height).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_max_height_hash(&self) -> Result<Option<(BlockHeight, BlockHash)>, Self::Error> {
|
|
wallet::get_max_height_hash(self.conn.borrow()).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_tx_height(&self, txid: TxId) -> Result<Option<BlockHeight>, Self::Error> {
|
|
wallet::get_tx_height(self.conn.borrow(), txid).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_wallet_birthday(&self) -> Result<Option<BlockHeight>, Self::Error> {
|
|
wallet::wallet_birthday(self.conn.borrow()).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_account_birthday(&self, account: AccountId) -> Result<BlockHeight, Self::Error> {
|
|
wallet::account_birthday(self.conn.borrow(), account).map_err(SqliteClientError::from)
|
|
}
|
|
|
|
fn get_current_address(
|
|
&self,
|
|
account: AccountId,
|
|
) -> Result<Option<UnifiedAddress>, Self::Error> {
|
|
wallet::get_current_address(self.conn.borrow(), &self.params, account)
|
|
.map(|res| res.map(|(addr, _)| addr))
|
|
}
|
|
|
|
fn get_unified_full_viewing_keys(
|
|
&self,
|
|
) -> Result<HashMap<AccountId, UnifiedFullViewingKey>, Self::Error> {
|
|
wallet::get_unified_full_viewing_keys(self.conn.borrow(), &self.params)
|
|
}
|
|
|
|
fn get_account_for_ufvk(
|
|
&self,
|
|
ufvk: &UnifiedFullViewingKey,
|
|
) -> Result<Option<AccountId>, Self::Error> {
|
|
wallet::get_account_for_ufvk(self.conn.borrow(), &self.params, ufvk)
|
|
}
|
|
|
|
fn get_wallet_summary(
|
|
&self,
|
|
min_confirmations: u32,
|
|
) -> Result<Option<WalletSummary>, Self::Error> {
|
|
// This will return a runtime error if we call `get_wallet_summary` from two
|
|
// threads at the same time, as transactions cannot nest.
|
|
wallet::get_wallet_summary(
|
|
&self.conn.borrow().unchecked_transaction()?,
|
|
&self.params,
|
|
min_confirmations,
|
|
&SubtreeScanProgress,
|
|
)
|
|
}
|
|
|
|
fn get_memo(&self, note_id: NoteId) -> Result<Option<Memo>, Self::Error> {
|
|
let sent_memo = wallet::get_sent_memo(self.conn.borrow(), note_id)?;
|
|
if sent_memo.is_some() {
|
|
Ok(sent_memo)
|
|
} else {
|
|
wallet::get_received_memo(self.conn.borrow(), note_id)
|
|
}
|
|
}
|
|
|
|
fn get_transaction(&self, txid: TxId) -> Result<Transaction, Self::Error> {
|
|
wallet::get_transaction(self.conn.borrow(), &self.params, txid).map(|(_, tx)| tx)
|
|
}
|
|
|
|
fn get_sapling_nullifiers(
|
|
&self,
|
|
query: NullifierQuery,
|
|
) -> Result<Vec<(AccountId, sapling::Nullifier)>, Self::Error> {
|
|
match query {
|
|
NullifierQuery::Unspent => wallet::sapling::get_sapling_nullifiers(self.conn.borrow()),
|
|
NullifierQuery::All => wallet::sapling::get_all_sapling_nullifiers(self.conn.borrow()),
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
fn get_transparent_receivers(
|
|
&self,
|
|
_account: AccountId,
|
|
) -> Result<HashMap<TransparentAddress, Option<TransparentAddressMetadata>>, Self::Error> {
|
|
wallet::get_transparent_receivers(self.conn.borrow(), &self.params, _account)
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
fn get_transparent_balances(
|
|
&self,
|
|
_account: AccountId,
|
|
_max_height: BlockHeight,
|
|
) -> Result<HashMap<TransparentAddress, Amount>, Self::Error> {
|
|
wallet::get_transparent_balances(self.conn.borrow(), &self.params, _account, _max_height)
|
|
}
|
|
|
|
#[cfg(feature = "orchard")]
|
|
fn get_orchard_nullifiers(
|
|
&self,
|
|
_query: NullifierQuery,
|
|
) -> Result<Vec<(AccountId, orchard::note::Nullifier)>, Self::Error> {
|
|
todo!()
|
|
}
|
|
|
|
fn get_account_ids(&self) -> Result<Vec<AccountId>, Self::Error> {
|
|
wallet::get_account_ids(self.conn.borrow())
|
|
}
|
|
}
|
|
|
|
impl<P: consensus::Parameters> WalletWrite for WalletDb<rusqlite::Connection, P> {
|
|
type UtxoRef = UtxoId;
|
|
|
|
fn create_account(
|
|
&mut self,
|
|
seed: &SecretVec<u8>,
|
|
birthday: AccountBirthday,
|
|
) -> Result<(AccountId, UnifiedSpendingKey), Self::Error> {
|
|
self.transactionally(|wdb| {
|
|
let account = wallet::get_max_account_id(wdb.conn.0)?
|
|
.map(|a| a.next().ok_or(SqliteClientError::AccountIdOutOfRange))
|
|
.transpose()?
|
|
.unwrap_or(AccountId::ZERO);
|
|
|
|
let usk = UnifiedSpendingKey::from_seed(&wdb.params, seed.expose_secret(), account)
|
|
.map_err(|_| SqliteClientError::KeyDerivationError(account))?;
|
|
let ufvk = usk.to_unified_full_viewing_key();
|
|
|
|
wallet::add_account(wdb.conn.0, &wdb.params, account, &ufvk, birthday)?;
|
|
|
|
Ok((account, usk))
|
|
})
|
|
}
|
|
|
|
fn get_next_available_address(
|
|
&mut self,
|
|
account: AccountId,
|
|
request: UnifiedAddressRequest,
|
|
) -> Result<Option<UnifiedAddress>, Self::Error> {
|
|
self.transactionally(
|
|
|wdb| match wdb.get_unified_full_viewing_keys()?.get(&account) {
|
|
Some(ufvk) => {
|
|
let search_from =
|
|
match wallet::get_current_address(wdb.conn.0, &wdb.params, account)? {
|
|
Some((_, mut last_diversifier_index)) => {
|
|
last_diversifier_index
|
|
.increment()
|
|
.map_err(|_| SqliteClientError::DiversifierIndexOutOfRange)?;
|
|
last_diversifier_index
|
|
}
|
|
None => DiversifierIndex::default(),
|
|
};
|
|
|
|
let (addr, diversifier_index) = ufvk
|
|
.find_address(search_from, request)
|
|
.ok_or(SqliteClientError::DiversifierIndexOutOfRange)?;
|
|
|
|
wallet::insert_address(
|
|
wdb.conn.0,
|
|
&wdb.params,
|
|
account,
|
|
diversifier_index,
|
|
&addr,
|
|
)?;
|
|
|
|
Ok(Some(addr))
|
|
}
|
|
None => Ok(None),
|
|
},
|
|
)
|
|
}
|
|
|
|
#[tracing::instrument(skip_all, fields(height = blocks.first().map(|b| u32::from(b.height()))))]
|
|
#[allow(clippy::type_complexity)]
|
|
fn put_blocks(
|
|
&mut self,
|
|
blocks: Vec<ScannedBlock<sapling::Nullifier, Scope>>,
|
|
) -> Result<(), Self::Error> {
|
|
self.transactionally(|wdb| {
|
|
let start_positions = blocks.first().map(|block| {
|
|
(
|
|
block.height(),
|
|
Position::from(
|
|
u64::from(block.sapling().final_tree_size())
|
|
- u64::try_from(block.sapling().commitments().len()).unwrap(),
|
|
),
|
|
)
|
|
});
|
|
let mut sapling_commitments = vec![];
|
|
let mut last_scanned_height = None;
|
|
let mut note_positions = vec![];
|
|
for block in blocks.into_iter() {
|
|
if last_scanned_height
|
|
.iter()
|
|
.any(|prev| block.height() != *prev + 1)
|
|
{
|
|
return Err(SqliteClientError::NonSequentialBlocks);
|
|
}
|
|
|
|
// Insert the block into the database.
|
|
wallet::put_block(
|
|
wdb.conn.0,
|
|
block.height(),
|
|
block.block_hash(),
|
|
block.block_time(),
|
|
block.sapling().final_tree_size(),
|
|
block.sapling().commitments().len().try_into().unwrap(),
|
|
)?;
|
|
|
|
for tx in block.transactions() {
|
|
let tx_row = wallet::put_tx_meta(wdb.conn.0, tx, block.height())?;
|
|
|
|
// Mark notes as spent and remove them from the scanning cache
|
|
for spend in &tx.sapling_spends {
|
|
wallet::sapling::mark_sapling_note_spent(wdb.conn.0, tx_row, spend.nf())?;
|
|
}
|
|
|
|
for output in &tx.sapling_outputs {
|
|
// Check whether this note was spent in a later block range that
|
|
// we previously scanned.
|
|
let spent_in = wallet::query_nullifier_map::<_, Scope>(
|
|
wdb.conn.0,
|
|
ShieldedProtocol::Sapling,
|
|
output.nf(),
|
|
)?;
|
|
|
|
wallet::sapling::put_received_note(wdb.conn.0, output, tx_row, spent_in)?;
|
|
}
|
|
}
|
|
|
|
// Insert the new nullifiers from this block into the nullifier map.
|
|
wallet::insert_nullifier_map(
|
|
wdb.conn.0,
|
|
block.height(),
|
|
ShieldedProtocol::Sapling,
|
|
block.sapling().nullifier_map(),
|
|
)?;
|
|
|
|
note_positions.extend(block.transactions().iter().flat_map(|wtx| {
|
|
wtx.sapling_outputs
|
|
.iter()
|
|
.map(|out| out.note_commitment_tree_position())
|
|
}));
|
|
|
|
last_scanned_height = Some(block.height());
|
|
let block_commitments = block.into_commitments();
|
|
sapling_commitments.extend(block_commitments.sapling.into_iter().map(Some));
|
|
}
|
|
|
|
// Prune the nullifier map of entries we no longer need.
|
|
if let Some(meta) = wdb.block_fully_scanned()? {
|
|
wallet::prune_nullifier_map(
|
|
wdb.conn.0,
|
|
meta.block_height().saturating_sub(PRUNING_DEPTH),
|
|
)?;
|
|
}
|
|
|
|
// We will have a start position and a last scanned height in all cases where
|
|
// `blocks` is non-empty.
|
|
if let Some(((start_height, start_position), last_scanned_height)) =
|
|
start_positions.zip(last_scanned_height)
|
|
{
|
|
// Create subtrees from the note commitments in parallel.
|
|
const CHUNK_SIZE: usize = 1024;
|
|
let subtrees = sapling_commitments
|
|
.par_chunks_mut(CHUNK_SIZE)
|
|
.enumerate()
|
|
.filter_map(|(i, chunk)| {
|
|
let start = start_position + (i * CHUNK_SIZE) as u64;
|
|
let end = start + chunk.len() as u64;
|
|
|
|
shardtree::LocatedTree::from_iter(
|
|
start..end,
|
|
SAPLING_SHARD_HEIGHT.into(),
|
|
chunk.iter_mut().map(|n| n.take().expect("always Some")),
|
|
)
|
|
})
|
|
.map(|res| (res.subtree, res.checkpoints))
|
|
.collect::<Vec<_>>();
|
|
|
|
// Update the Sapling note commitment tree with all newly read note commitments
|
|
let mut subtrees = subtrees.into_iter();
|
|
wdb.with_sapling_tree_mut::<_, _, Self::Error>(move |sapling_tree| {
|
|
for (tree, checkpoints) in &mut subtrees {
|
|
sapling_tree.insert_tree(tree, checkpoints)?;
|
|
}
|
|
|
|
Ok(())
|
|
})?;
|
|
|
|
// Update now-expired transactions that didn't get mined.
|
|
wallet::update_expired_notes(wdb.conn.0, last_scanned_height)?;
|
|
|
|
wallet::scanning::scan_complete(
|
|
wdb.conn.0,
|
|
&wdb.params,
|
|
Range {
|
|
start: start_height,
|
|
end: last_scanned_height + 1,
|
|
},
|
|
¬e_positions,
|
|
)?;
|
|
}
|
|
|
|
Ok(())
|
|
})
|
|
}
|
|
|
|
fn update_chain_tip(&mut self, tip_height: BlockHeight) -> Result<(), Self::Error> {
|
|
let tx = self.conn.transaction()?;
|
|
wallet::scanning::update_chain_tip(&tx, &self.params, tip_height)?;
|
|
tx.commit()?;
|
|
Ok(())
|
|
}
|
|
|
|
fn store_decrypted_tx(&mut self, d_tx: DecryptedTransaction) -> Result<(), Self::Error> {
|
|
self.transactionally(|wdb| {
|
|
let tx_ref = wallet::put_tx_data(wdb.conn.0, d_tx.tx, None, None)?;
|
|
|
|
let mut spending_account_id: Option<AccountId> = None;
|
|
for output in d_tx.sapling_outputs {
|
|
match output.transfer_type {
|
|
TransferType::Outgoing | TransferType::WalletInternal => {
|
|
let recipient = if output.transfer_type == TransferType::Outgoing {
|
|
Recipient::Sapling(output.note.recipient())
|
|
} else {
|
|
Recipient::InternalAccount(
|
|
output.account,
|
|
PoolType::Shielded(ShieldedProtocol::Sapling)
|
|
)
|
|
};
|
|
|
|
wallet::put_sent_output(
|
|
wdb.conn.0,
|
|
&wdb.params,
|
|
output.account,
|
|
tx_ref,
|
|
output.index,
|
|
&recipient,
|
|
NonNegativeAmount::from_u64(output.note.value().inner()).map_err(|_| {
|
|
SqliteClientError::CorruptedData(
|
|
"Note value is not a valid Zcash amount.".to_string(),
|
|
)
|
|
})?,
|
|
Some(&output.memo),
|
|
)?;
|
|
|
|
if matches!(recipient, Recipient::InternalAccount(_, _)) {
|
|
wallet::sapling::put_received_note(wdb.conn.0, output, tx_ref, None)?;
|
|
}
|
|
}
|
|
TransferType::Incoming => {
|
|
match spending_account_id {
|
|
Some(id) => {
|
|
if id != output.account {
|
|
panic!("Unable to determine a unique account identifier for z->t spend.");
|
|
}
|
|
}
|
|
None => {
|
|
spending_account_id = Some(output.account);
|
|
}
|
|
}
|
|
|
|
wallet::sapling::put_received_note(wdb.conn.0, output, tx_ref, None)?;
|
|
}
|
|
}
|
|
}
|
|
|
|
// If any of the utxos spent in the transaction are ours, mark them as spent.
|
|
#[cfg(feature = "transparent-inputs")]
|
|
for txin in d_tx.tx.transparent_bundle().iter().flat_map(|b| b.vin.iter()) {
|
|
wallet::mark_transparent_utxo_spent(wdb.conn.0, tx_ref, &txin.prevout)?;
|
|
}
|
|
|
|
// If we have some transparent outputs:
|
|
if d_tx.tx.transparent_bundle().iter().any(|b| !b.vout.is_empty()) {
|
|
let nullifiers = wdb.get_sapling_nullifiers(NullifierQuery::All)?;
|
|
// If the transaction contains shielded spends from our wallet, we will store z->t
|
|
// transactions we observe in the same way they would be stored by
|
|
// create_spend_to_address.
|
|
if let Some((account_id, _)) = nullifiers.iter().find(
|
|
|(_, nf)|
|
|
d_tx.tx.sapling_bundle().iter().flat_map(|b| b.shielded_spends().iter())
|
|
.any(|input| nf == input.nullifier())
|
|
) {
|
|
for (output_index, txout) in d_tx.tx.transparent_bundle().iter().flat_map(|b| b.vout.iter()).enumerate() {
|
|
if let Some(address) = txout.recipient_address() {
|
|
wallet::put_sent_output(
|
|
wdb.conn.0,
|
|
&wdb.params,
|
|
*account_id,
|
|
tx_ref,
|
|
output_index,
|
|
&Recipient::Transparent(address),
|
|
txout.value,
|
|
None
|
|
)?;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
})
|
|
}
|
|
|
|
fn store_sent_tx(&mut self, sent_tx: &SentTransaction) -> Result<(), Self::Error> {
|
|
self.transactionally(|wdb| {
|
|
let tx_ref = wallet::put_tx_data(
|
|
wdb.conn.0,
|
|
sent_tx.tx,
|
|
Some(sent_tx.fee_amount),
|
|
Some(sent_tx.created),
|
|
)?;
|
|
|
|
// Mark notes as spent.
|
|
//
|
|
// This locks the notes so they aren't selected again by a subsequent call to
|
|
// create_spend_to_address() before this transaction has been mined (at which point the notes
|
|
// get re-marked as spent).
|
|
//
|
|
// Assumes that create_spend_to_address() will never be called in parallel, which is a
|
|
// reasonable assumption for a light client such as a mobile phone.
|
|
if let Some(bundle) = sent_tx.tx.sapling_bundle() {
|
|
for spend in bundle.shielded_spends() {
|
|
wallet::sapling::mark_sapling_note_spent(
|
|
wdb.conn.0,
|
|
tx_ref,
|
|
spend.nullifier(),
|
|
)?;
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
for utxo_outpoint in &sent_tx.utxos_spent {
|
|
wallet::mark_transparent_utxo_spent(wdb.conn.0, tx_ref, utxo_outpoint)?;
|
|
}
|
|
|
|
for output in &sent_tx.outputs {
|
|
wallet::insert_sent_output(
|
|
wdb.conn.0,
|
|
&wdb.params,
|
|
tx_ref,
|
|
sent_tx.account,
|
|
output,
|
|
)?;
|
|
|
|
if let Some((account, note)) = output.sapling_change_to() {
|
|
wallet::sapling::put_received_note(
|
|
wdb.conn.0,
|
|
&DecryptedOutput {
|
|
index: output.output_index(),
|
|
note: note.clone(),
|
|
account: *account,
|
|
memo: output
|
|
.memo()
|
|
.map_or_else(MemoBytes::empty, |memo| memo.clone()),
|
|
transfer_type: TransferType::WalletInternal,
|
|
},
|
|
tx_ref,
|
|
None,
|
|
)?;
|
|
}
|
|
}
|
|
|
|
Ok(())
|
|
})
|
|
}
|
|
|
|
fn truncate_to_height(&mut self, block_height: BlockHeight) -> Result<(), Self::Error> {
|
|
self.transactionally(|wdb| {
|
|
wallet::truncate_to_height(wdb.conn.0, &wdb.params, block_height)
|
|
})
|
|
}
|
|
|
|
fn put_received_transparent_utxo(
|
|
&mut self,
|
|
_output: &WalletTransparentOutput,
|
|
) -> Result<Self::UtxoRef, Self::Error> {
|
|
#[cfg(feature = "transparent-inputs")]
|
|
return wallet::put_received_transparent_utxo(&self.conn, &self.params, _output);
|
|
|
|
#[cfg(not(feature = "transparent-inputs"))]
|
|
panic!(
|
|
"The wallet must be compiled with the transparent-inputs feature to use this method."
|
|
);
|
|
}
|
|
}
|
|
|
|
impl<P: consensus::Parameters> WalletCommitmentTrees for WalletDb<rusqlite::Connection, P> {
|
|
type Error = commitment_tree::Error;
|
|
type SaplingShardStore<'a> =
|
|
SqliteShardStore<&'a rusqlite::Transaction<'a>, sapling::Node, SAPLING_SHARD_HEIGHT>;
|
|
|
|
fn with_sapling_tree_mut<F, A, E>(&mut self, mut callback: F) -> Result<A, E>
|
|
where
|
|
for<'a> F: FnMut(
|
|
&'a mut ShardTree<
|
|
Self::SaplingShardStore<'a>,
|
|
{ sapling::NOTE_COMMITMENT_TREE_DEPTH },
|
|
SAPLING_SHARD_HEIGHT,
|
|
>,
|
|
) -> Result<A, E>,
|
|
E: From<ShardTreeError<Self::Error>>,
|
|
{
|
|
let tx = self
|
|
.conn
|
|
.transaction()
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?;
|
|
let shard_store = SqliteShardStore::from_connection(&tx, SAPLING_TABLES_PREFIX)
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?;
|
|
let result = {
|
|
let mut shardtree = ShardTree::new(shard_store, PRUNING_DEPTH.try_into().unwrap());
|
|
callback(&mut shardtree)?
|
|
};
|
|
|
|
tx.commit()
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?;
|
|
Ok(result)
|
|
}
|
|
|
|
fn put_sapling_subtree_roots(
|
|
&mut self,
|
|
start_index: u64,
|
|
roots: &[CommitmentTreeRoot<sapling::Node>],
|
|
) -> Result<(), ShardTreeError<Self::Error>> {
|
|
let tx = self
|
|
.conn
|
|
.transaction()
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?;
|
|
put_shard_roots::<_, { sapling::NOTE_COMMITMENT_TREE_DEPTH }, SAPLING_SHARD_HEIGHT>(
|
|
&tx,
|
|
SAPLING_TABLES_PREFIX,
|
|
start_index,
|
|
roots,
|
|
)?;
|
|
tx.commit()
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?;
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
impl<'conn, P: consensus::Parameters> WalletCommitmentTrees for WalletDb<SqlTransaction<'conn>, P> {
|
|
type Error = commitment_tree::Error;
|
|
type SaplingShardStore<'a> =
|
|
SqliteShardStore<&'a rusqlite::Transaction<'a>, sapling::Node, SAPLING_SHARD_HEIGHT>;
|
|
|
|
fn with_sapling_tree_mut<F, A, E>(&mut self, mut callback: F) -> Result<A, E>
|
|
where
|
|
for<'a> F: FnMut(
|
|
&'a mut ShardTree<
|
|
Self::SaplingShardStore<'a>,
|
|
{ sapling::NOTE_COMMITMENT_TREE_DEPTH },
|
|
SAPLING_SHARD_HEIGHT,
|
|
>,
|
|
) -> Result<A, E>,
|
|
E: From<ShardTreeError<commitment_tree::Error>>,
|
|
{
|
|
let mut shardtree = ShardTree::new(
|
|
SqliteShardStore::from_connection(self.conn.0, SAPLING_TABLES_PREFIX)
|
|
.map_err(|e| ShardTreeError::Storage(commitment_tree::Error::Query(e)))?,
|
|
PRUNING_DEPTH.try_into().unwrap(),
|
|
);
|
|
let result = callback(&mut shardtree)?;
|
|
|
|
Ok(result)
|
|
}
|
|
|
|
fn put_sapling_subtree_roots(
|
|
&mut self,
|
|
start_index: u64,
|
|
roots: &[CommitmentTreeRoot<sapling::Node>],
|
|
) -> Result<(), ShardTreeError<Self::Error>> {
|
|
put_shard_roots::<_, { sapling::NOTE_COMMITMENT_TREE_DEPTH }, SAPLING_SHARD_HEIGHT>(
|
|
self.conn.0,
|
|
SAPLING_TABLES_PREFIX,
|
|
start_index,
|
|
roots,
|
|
)
|
|
}
|
|
}
|
|
|
|
/// A handle for the SQLite block source.
|
|
pub struct BlockDb(Connection);
|
|
|
|
impl BlockDb {
|
|
/// Opens a connection to the wallet database stored at the specified path.
|
|
pub fn for_path<P: AsRef<Path>>(path: P) -> Result<Self, rusqlite::Error> {
|
|
Connection::open(path).map(BlockDb)
|
|
}
|
|
}
|
|
|
|
impl BlockSource for BlockDb {
|
|
type Error = SqliteClientError;
|
|
|
|
fn with_blocks<F, DbErrT>(
|
|
&self,
|
|
from_height: Option<BlockHeight>,
|
|
limit: Option<usize>,
|
|
with_row: F,
|
|
) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>
|
|
where
|
|
F: FnMut(CompactBlock) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>,
|
|
{
|
|
chain::blockdb_with_blocks(self, from_height, limit, with_row)
|
|
}
|
|
}
|
|
|
|
/// A block source that reads block data from disk and block metadata from a SQLite database.
|
|
///
|
|
/// This block source expects each compact block to be stored on disk in the `blocks` subdirectory
|
|
/// of the `blockstore_root` path provided at the time of construction. Each block should be
|
|
/// written, as the serialized bytes of its protobuf representation, where the path for each block
|
|
/// has the pattern:
|
|
///
|
|
/// `<blockstore_root>/blocks/<block_height>-<block_hash>-compactblock`
|
|
///
|
|
/// where `<block_height>` is the decimal value of the height at which the block was mined, and
|
|
/// `<block_hash>` is the hexadecimal representation of the block hash, as produced by the
|
|
/// [`fmt::Display`] implementation for [`zcash_primitives::block::BlockHash`].
|
|
///
|
|
/// This block source is intended to be used with the following data flow:
|
|
/// * When the cache is being filled:
|
|
/// * The caller requests the current maximum height height at which cached data is available
|
|
/// using [`FsBlockDb::get_max_cached_height`]. If no cached data is available, the caller
|
|
/// can use the wallet's synced-to height for the following operations instead.
|
|
/// * (recommended for privacy) the caller should round the returned height down to some 100- /
|
|
/// 1000-block boundary.
|
|
/// * The caller uses the lightwalletd's `getblock` gRPC method to obtain a stream of blocks.
|
|
/// For each block returned, the caller writes the compact block to `blocks_dir` using the
|
|
/// path format specified above. It is fine to overwrite an existing block, since block hashes
|
|
/// are immutable and collision-resistant.
|
|
/// * Once a caller-determined number of blocks have been successfully written to disk, the
|
|
/// caller should invoke [`FsBlockDb::write_block_metadata`] with the metadata for each block
|
|
/// written to disk.
|
|
/// * The cache can then be scanned using the [`BlockSource`] implementation, providing the
|
|
/// wallet's synced-to-height as a starting point.
|
|
/// * When part of the cache is no longer needed:
|
|
/// * The caller determines some height `H` that is the earliest block data it needs to preserve.
|
|
/// This might be determined based on where the wallet is fully-synced to, or other heuristics.
|
|
/// * The caller searches the defined filesystem folder for all files beginning in `HEIGHT-*` where
|
|
/// `HEIGHT < H`, and deletes those files.
|
|
///
|
|
/// Note: This API is unstable, and may change in the future. In particular, the [`BlockSource`]
|
|
/// API and the above description currently assume that scanning is performed in linear block
|
|
/// order; this assumption is likely to be weakened and/or removed in a future update.
|
|
#[cfg(feature = "unstable")]
|
|
pub struct FsBlockDb {
|
|
conn: Connection,
|
|
blocks_dir: PathBuf,
|
|
}
|
|
|
|
/// Errors that can be generated by the filesystem/sqlite-backed
|
|
/// block source.
|
|
#[derive(Debug)]
|
|
#[cfg(feature = "unstable")]
|
|
pub enum FsBlockDbError {
|
|
Fs(io::Error),
|
|
Db(rusqlite::Error),
|
|
Protobuf(prost::DecodeError),
|
|
MissingBlockPath(PathBuf),
|
|
InvalidBlockstoreRoot(PathBuf),
|
|
InvalidBlockPath(PathBuf),
|
|
CorruptedData(String),
|
|
CacheMiss(BlockHeight),
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl From<io::Error> for FsBlockDbError {
|
|
fn from(err: io::Error) -> Self {
|
|
FsBlockDbError::Fs(err)
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl From<rusqlite::Error> for FsBlockDbError {
|
|
fn from(err: rusqlite::Error) -> Self {
|
|
FsBlockDbError::Db(err)
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl From<prost::DecodeError> for FsBlockDbError {
|
|
fn from(e: prost::DecodeError) -> Self {
|
|
FsBlockDbError::Protobuf(e)
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl FsBlockDb {
|
|
/// Creates a filesystem-backed block store at the given path.
|
|
///
|
|
/// This will construct or open a SQLite database at the path
|
|
/// `<fsblockdb_root>/blockmeta.sqlite` and will ensure that a directory exists at
|
|
/// `<fsblockdb_root>/blocks` where this block store will expect to find serialized block
|
|
/// files as described for [`FsBlockDb`].
|
|
///
|
|
/// An application using this constructor should ensure that they call
|
|
/// [`crate::chain::init::init_blockmeta_db`] at application startup to ensure
|
|
/// that the resulting metadata database is properly initialized and has had all required
|
|
/// migrations applied before use.
|
|
pub fn for_path<P: AsRef<Path>>(fsblockdb_root: P) -> Result<Self, FsBlockDbError> {
|
|
let meta = fs::metadata(&fsblockdb_root).map_err(FsBlockDbError::Fs)?;
|
|
if meta.is_dir() {
|
|
let db_path = fsblockdb_root.as_ref().join("blockmeta.sqlite");
|
|
let blocks_dir = fsblockdb_root.as_ref().join("blocks");
|
|
fs::create_dir_all(&blocks_dir)?;
|
|
Ok(FsBlockDb {
|
|
conn: Connection::open(db_path).map_err(FsBlockDbError::Db)?,
|
|
blocks_dir,
|
|
})
|
|
} else {
|
|
Err(FsBlockDbError::InvalidBlockstoreRoot(
|
|
fsblockdb_root.as_ref().to_path_buf(),
|
|
))
|
|
}
|
|
}
|
|
|
|
/// Returns the maximum height of blocks known to the block metadata database.
|
|
pub fn get_max_cached_height(&self) -> Result<Option<BlockHeight>, FsBlockDbError> {
|
|
Ok(chain::blockmetadb_get_max_cached_height(&self.conn)?)
|
|
}
|
|
|
|
/// Adds a set of block metadata entries to the metadata database, overwriting any
|
|
/// existing entries at the given block heights.
|
|
///
|
|
/// This will return an error if any block file corresponding to one of these metadata records
|
|
/// is absent from the blocks directory.
|
|
pub fn write_block_metadata(&self, block_meta: &[BlockMeta]) -> Result<(), FsBlockDbError> {
|
|
for m in block_meta {
|
|
let block_path = m.block_file_path(&self.blocks_dir);
|
|
match fs::metadata(&block_path) {
|
|
Err(e) => {
|
|
return Err(match e.kind() {
|
|
io::ErrorKind::NotFound => FsBlockDbError::MissingBlockPath(block_path),
|
|
_ => FsBlockDbError::Fs(e),
|
|
});
|
|
}
|
|
Ok(meta) => {
|
|
if !meta.is_file() {
|
|
return Err(FsBlockDbError::InvalidBlockPath(block_path));
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
Ok(chain::blockmetadb_insert(&self.conn, block_meta)?)
|
|
}
|
|
|
|
/// Returns the metadata for the block with the given height, if it exists in the
|
|
/// database.
|
|
pub fn find_block(&self, height: BlockHeight) -> Result<Option<BlockMeta>, FsBlockDbError> {
|
|
Ok(chain::blockmetadb_find_block(&self.conn, height)?)
|
|
}
|
|
|
|
/// Rewinds the BlockMeta Db to the `block_height` provided.
|
|
///
|
|
/// This doesn't delete any files referenced by the records
|
|
/// stored in BlockMeta.
|
|
///
|
|
/// If the requested height is greater than or equal to the height
|
|
/// of the last scanned block, or if the DB is empty, this function
|
|
/// does nothing.
|
|
pub fn truncate_to_height(&self, block_height: BlockHeight) -> Result<(), FsBlockDbError> {
|
|
Ok(chain::blockmetadb_truncate_to_height(
|
|
&self.conn,
|
|
block_height,
|
|
)?)
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl BlockSource for FsBlockDb {
|
|
type Error = FsBlockDbError;
|
|
|
|
fn with_blocks<F, DbErrT>(
|
|
&self,
|
|
from_height: Option<BlockHeight>,
|
|
limit: Option<usize>,
|
|
with_row: F,
|
|
) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>
|
|
where
|
|
F: FnMut(CompactBlock) -> Result<(), data_api::chain::error::Error<DbErrT, Self::Error>>,
|
|
{
|
|
fsblockdb_with_blocks(self, from_height, limit, with_row)
|
|
}
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
impl std::fmt::Display for FsBlockDbError {
|
|
fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result {
|
|
match self {
|
|
FsBlockDbError::Fs(io_error) => {
|
|
write!(f, "Failed to access the file system: {}", io_error)
|
|
}
|
|
FsBlockDbError::Db(e) => {
|
|
write!(f, "There was a problem with the sqlite db: {}", e)
|
|
}
|
|
FsBlockDbError::Protobuf(e) => {
|
|
write!(f, "Failed to parse protobuf-encoded record: {}", e)
|
|
}
|
|
FsBlockDbError::MissingBlockPath(block_path) => {
|
|
write!(
|
|
f,
|
|
"CompactBlock file expected but not found at {}",
|
|
block_path.display(),
|
|
)
|
|
}
|
|
FsBlockDbError::InvalidBlockstoreRoot(fsblockdb_root) => {
|
|
write!(
|
|
f,
|
|
"The block storage root {} is not a directory",
|
|
fsblockdb_root.display(),
|
|
)
|
|
}
|
|
FsBlockDbError::InvalidBlockPath(block_path) => {
|
|
write!(
|
|
f,
|
|
"CompactBlock path {} is not a file",
|
|
block_path.display(),
|
|
)
|
|
}
|
|
FsBlockDbError::CorruptedData(e) => {
|
|
write!(
|
|
f,
|
|
"The block cache has corrupted data and this caused an error: {}",
|
|
e,
|
|
)
|
|
}
|
|
FsBlockDbError::CacheMiss(height) => {
|
|
write!(
|
|
f,
|
|
"Requested height {} does not exist in the block cache",
|
|
height
|
|
)
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
#[macro_use]
|
|
extern crate assert_matches;
|
|
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use zcash_client_backend::data_api::{AccountBirthday, WalletRead, WalletWrite};
|
|
|
|
use crate::{testing::TestBuilder, AccountId, DEFAULT_UA_REQUEST};
|
|
|
|
#[cfg(feature = "unstable")]
|
|
use {
|
|
crate::testing::AddressType,
|
|
zcash_client_backend::keys::sapling,
|
|
zcash_primitives::{
|
|
consensus::Parameters, transaction::components::amount::NonNegativeAmount,
|
|
},
|
|
};
|
|
|
|
#[test]
|
|
pub(crate) fn get_next_available_address() {
|
|
let mut st = TestBuilder::new()
|
|
.with_test_account(AccountBirthday::from_sapling_activation)
|
|
.build();
|
|
|
|
let account = AccountId::ZERO;
|
|
let current_addr = st.wallet().get_current_address(account).unwrap();
|
|
assert!(current_addr.is_some());
|
|
|
|
// TODO: Add Orchard
|
|
let addr2 = st
|
|
.wallet_mut()
|
|
.get_next_available_address(account, DEFAULT_UA_REQUEST)
|
|
.unwrap();
|
|
assert!(addr2.is_some());
|
|
assert_ne!(current_addr, addr2);
|
|
|
|
let addr2_cur = st.wallet().get_current_address(account).unwrap();
|
|
assert_eq!(addr2, addr2_cur);
|
|
}
|
|
|
|
#[cfg(feature = "transparent-inputs")]
|
|
#[test]
|
|
fn transparent_receivers() {
|
|
// Add an account to the wallet.
|
|
let st = TestBuilder::new()
|
|
.with_block_cache()
|
|
.with_test_account(AccountBirthday::from_sapling_activation)
|
|
.build();
|
|
|
|
let (_, usk, _) = st.test_account().unwrap();
|
|
let ufvk = usk.to_unified_full_viewing_key();
|
|
let (taddr, _) = usk.default_transparent_address();
|
|
|
|
let receivers = st
|
|
.wallet()
|
|
.get_transparent_receivers(AccountId::ZERO)
|
|
.unwrap();
|
|
|
|
// The receiver for the default UA should be in the set.
|
|
assert!(receivers.contains_key(
|
|
ufvk.default_address(DEFAULT_UA_REQUEST)
|
|
.0
|
|
.transparent()
|
|
.unwrap()
|
|
));
|
|
|
|
// The default t-addr should be in the set.
|
|
assert!(receivers.contains_key(&taddr));
|
|
}
|
|
|
|
#[cfg(feature = "unstable")]
|
|
#[test]
|
|
pub(crate) fn fsblockdb_api() {
|
|
let mut st = TestBuilder::new().with_fs_block_cache().build();
|
|
|
|
// The BlockMeta DB starts off empty.
|
|
assert_eq!(st.cache().get_max_cached_height().unwrap(), None);
|
|
|
|
// Generate some fake CompactBlocks.
|
|
let seed = [0u8; 32];
|
|
let account = AccountId::ZERO;
|
|
let extsk = sapling::spending_key(&seed, st.wallet().params.coin_type(), account);
|
|
let dfvk = extsk.to_diversifiable_full_viewing_key();
|
|
let (h1, meta1, _) = st.generate_next_block(
|
|
&dfvk,
|
|
AddressType::DefaultExternal,
|
|
NonNegativeAmount::const_from_u64(5),
|
|
);
|
|
let (h2, meta2, _) = st.generate_next_block(
|
|
&dfvk,
|
|
AddressType::DefaultExternal,
|
|
NonNegativeAmount::const_from_u64(10),
|
|
);
|
|
|
|
// The BlockMeta DB is not updated until we do so explicitly.
|
|
assert_eq!(st.cache().get_max_cached_height().unwrap(), None);
|
|
|
|
// Inform the BlockMeta DB about the newly-persisted CompactBlocks.
|
|
st.cache().write_block_metadata(&[meta1, meta2]).unwrap();
|
|
|
|
// The BlockMeta DB now sees blocks up to height 2.
|
|
assert_eq!(st.cache().get_max_cached_height().unwrap(), Some(h2),);
|
|
assert_eq!(st.cache().find_block(h1).unwrap(), Some(meta1));
|
|
assert_eq!(st.cache().find_block(h2).unwrap(), Some(meta2));
|
|
assert_eq!(st.cache().find_block(h2 + 1).unwrap(), None);
|
|
|
|
// Rewinding to height 1 should cause the metadata for height 2 to be deleted.
|
|
st.cache().truncate_to_height(h1).unwrap();
|
|
assert_eq!(st.cache().get_max_cached_height().unwrap(), Some(h1));
|
|
assert_eq!(st.cache().find_block(h1).unwrap(), Some(meta1));
|
|
assert_eq!(st.cache().find_block(h2).unwrap(), None);
|
|
assert_eq!(st.cache().find_block(h2 + 1).unwrap(), None);
|
|
}
|
|
}
|