create Downloads for Inbound
This commit is contained in:
parent
31eb0a5126
commit
486e55104a
|
@ -67,7 +67,7 @@ impl StartCmd {
|
||||||
let inbound = ServiceBuilder::new()
|
let inbound = ServiceBuilder::new()
|
||||||
.load_shed()
|
.load_shed()
|
||||||
.buffer(20)
|
.buffer(20)
|
||||||
.service(Inbound::new(setup_rx, state.clone()));
|
.service(Inbound::new(setup_rx, state.clone(), verifier.clone()));
|
||||||
|
|
||||||
let (peer_set, address_book) = zebra_network::init(config.network.clone(), inbound).await;
|
let (peer_set, address_book) = zebra_network::init(config.network.clone(), inbound).await;
|
||||||
setup_tx
|
setup_tx
|
||||||
|
|
|
@ -12,12 +12,21 @@ use futures::{
|
||||||
use tokio::sync::oneshot;
|
use tokio::sync::oneshot;
|
||||||
use tower::{buffer::Buffer, util::BoxService, Service, ServiceExt};
|
use tower::{buffer::Buffer, util::BoxService, Service, ServiceExt};
|
||||||
|
|
||||||
|
use zebra_chain as zc;
|
||||||
|
use zebra_consensus as zcon;
|
||||||
use zebra_network as zn;
|
use zebra_network as zn;
|
||||||
use zebra_network::AddressBook;
|
use zebra_network::AddressBook;
|
||||||
use zebra_state as zs;
|
use zebra_state as zs;
|
||||||
|
|
||||||
|
mod downloads;
|
||||||
|
use downloads::Downloads;
|
||||||
|
|
||||||
type Outbound = Buffer<BoxService<zn::Request, zn::Response, zn::BoxError>, zn::Request>;
|
type Outbound = Buffer<BoxService<zn::Request, zn::Response, zn::BoxError>, zn::Request>;
|
||||||
type State = Buffer<BoxService<zs::Request, zs::Response, zs::BoxError>, zs::Request>;
|
type State = Buffer<BoxService<zs::Request, zs::Response, zs::BoxError>, zs::Request>;
|
||||||
|
type Verifier = Buffer<
|
||||||
|
BoxService<Arc<zc::block::Block>, zc::block::Hash, zcon::chain::VerifyChainError>,
|
||||||
|
Arc<zc::block::Block>,
|
||||||
|
>;
|
||||||
|
|
||||||
pub type SetupData = (Outbound, Arc<Mutex<AddressBook>>);
|
pub type SetupData = (Outbound, Arc<Mutex<AddressBook>>);
|
||||||
|
|
||||||
|
@ -52,15 +61,23 @@ pub struct Inbound {
|
||||||
outbound: Option<Outbound>,
|
outbound: Option<Outbound>,
|
||||||
address_book: Option<Arc<Mutex<zn::AddressBook>>>,
|
address_book: Option<Arc<Mutex<zn::AddressBook>>>,
|
||||||
state: State,
|
state: State,
|
||||||
|
verifier: Verifier,
|
||||||
|
downloads: Option<Downloads<Outbound, Verifier, State>>,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Inbound {
|
impl Inbound {
|
||||||
pub fn new(network_setup: oneshot::Receiver<SetupData>, state: State) -> Self {
|
pub fn new(
|
||||||
|
network_setup: oneshot::Receiver<SetupData>,
|
||||||
|
state: State,
|
||||||
|
verifier: Verifier,
|
||||||
|
) -> Self {
|
||||||
Self {
|
Self {
|
||||||
network_setup: Some(network_setup),
|
network_setup: Some(network_setup),
|
||||||
outbound: None,
|
outbound: None,
|
||||||
address_book: None,
|
address_book: None,
|
||||||
state,
|
state,
|
||||||
|
verifier,
|
||||||
|
downloads: None,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -85,6 +102,11 @@ impl Service<zn::Request> for Inbound {
|
||||||
self.outbound = Some(outbound);
|
self.outbound = Some(outbound);
|
||||||
self.address_book = Some(address_book);
|
self.address_book = Some(address_book);
|
||||||
self.network_setup = None;
|
self.network_setup = None;
|
||||||
|
self.downloads = Some(Downloads::new(
|
||||||
|
self.outbound.clone().unwrap(),
|
||||||
|
self.verifier.clone(),
|
||||||
|
self.state.clone(),
|
||||||
|
));
|
||||||
}
|
}
|
||||||
Err(TryRecvError::Empty) => {
|
Err(TryRecvError::Empty) => {
|
||||||
self.network_setup = Some(rx);
|
self.network_setup = Some(rx);
|
||||||
|
@ -98,6 +120,10 @@ impl Service<zn::Request> for Inbound {
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Clean up completed download tasks
|
||||||
|
while let Poll::Ready(Some(_)) = self.downloads.unwrap().poll_next(cx) {}
|
||||||
|
|
||||||
// Now report readiness based on readiness of the inner services, if they're available.
|
// Now report readiness based on readiness of the inner services, if they're available.
|
||||||
// XXX do we want to propagate backpressure from the network here?
|
// XXX do we want to propagate backpressure from the network here?
|
||||||
match (
|
match (
|
||||||
|
@ -171,9 +197,25 @@ impl Service<zn::Request> for Inbound {
|
||||||
debug!("ignoring unimplemented request");
|
debug!("ignoring unimplemented request");
|
||||||
async { Ok(zn::Response::Nil) }.boxed()
|
async { Ok(zn::Response::Nil) }.boxed()
|
||||||
}
|
}
|
||||||
zn::Request::AdvertiseBlock(_block) => {
|
zn::Request::AdvertiseBlock(hash) => {
|
||||||
debug!("ignoring unimplemented request");
|
// this sucks
|
||||||
async { Ok(zn::Response::Nil) }.boxed()
|
let mut downloads = self.downloads.take().unwrap();
|
||||||
|
self.downloads = Some(Downloads::new(
|
||||||
|
self.outbound.as_ref().unwrap().clone(),
|
||||||
|
self.verifier.clone(),
|
||||||
|
self.state.clone(),
|
||||||
|
));
|
||||||
|
|
||||||
|
async move {
|
||||||
|
if downloads.download_and_verify(hash).await? {
|
||||||
|
tracing::info!(?hash, "queued download and verification of gossiped block");
|
||||||
|
} else {
|
||||||
|
tracing::debug!(?hash, "gossiped block already queued or verified");
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(zn::Response::Nil)
|
||||||
|
}
|
||||||
|
.boxed()
|
||||||
}
|
}
|
||||||
zn::Request::MempoolTransactions => {
|
zn::Request::MempoolTransactions => {
|
||||||
debug!("ignoring unimplemented request");
|
debug!("ignoring unimplemented request");
|
||||||
|
|
|
@ -0,0 +1,206 @@
|
||||||
|
use std::{
|
||||||
|
collections::HashMap,
|
||||||
|
pin::Pin,
|
||||||
|
sync::Arc,
|
||||||
|
task::{Context, Poll},
|
||||||
|
};
|
||||||
|
|
||||||
|
use color_eyre::eyre::{eyre, Report};
|
||||||
|
use futures::{
|
||||||
|
future::TryFutureExt,
|
||||||
|
ready,
|
||||||
|
stream::{FuturesUnordered, Stream},
|
||||||
|
};
|
||||||
|
use pin_project::pin_project;
|
||||||
|
use tokio::{sync::oneshot, task::JoinHandle};
|
||||||
|
use tower::{Service, ServiceExt};
|
||||||
|
use tracing_futures::Instrument;
|
||||||
|
|
||||||
|
use zebra_chain::block::{self, Block};
|
||||||
|
use zebra_network as zn;
|
||||||
|
use zebra_state as zs;
|
||||||
|
|
||||||
|
type BoxError = Box<dyn std::error::Error + Send + Sync + 'static>;
|
||||||
|
|
||||||
|
/// Manages download and verification of blocks gossiped to this peer.
|
||||||
|
#[pin_project]
|
||||||
|
#[derive(Debug)]
|
||||||
|
pub struct Downloads<ZN, ZV, ZS>
|
||||||
|
where
|
||||||
|
ZN: Service<zn::Request, Response = zn::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZN::Future: Send,
|
||||||
|
ZV: Service<Arc<Block>, Response = block::Hash, Error = BoxError> + Send + Clone + 'static,
|
||||||
|
ZV::Future: Send,
|
||||||
|
ZS: Service<zs::Request, Response = zs::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZS::Future: Send,
|
||||||
|
{
|
||||||
|
network: ZN,
|
||||||
|
verifier: ZV,
|
||||||
|
state: ZS,
|
||||||
|
#[pin]
|
||||||
|
pending: FuturesUnordered<JoinHandle<Result<block::Hash, (BoxError, block::Hash)>>>,
|
||||||
|
cancel_handles: HashMap<block::Hash, oneshot::Sender<()>>,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<ZN, ZV, ZS> Stream for Downloads<ZN, ZV, ZS>
|
||||||
|
where
|
||||||
|
ZN: Service<zn::Request, Response = zn::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZN::Future: Send,
|
||||||
|
ZV: Service<Arc<Block>, Response = block::Hash, Error = BoxError> + Send + Clone + 'static,
|
||||||
|
ZV::Future: Send,
|
||||||
|
ZS: Service<zs::Request, Response = zs::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZS::Future: Send,
|
||||||
|
{
|
||||||
|
type Item = Result<block::Hash, BoxError>;
|
||||||
|
|
||||||
|
fn poll_next(self: Pin<&mut Self>, cx: &mut Context) -> Poll<Option<Self::Item>> {
|
||||||
|
let this = self.project();
|
||||||
|
// This would be cleaner with poll_map #63514, but that's nightly only.
|
||||||
|
if let Some(join_result) = ready!(this.pending.poll_next(cx)) {
|
||||||
|
match join_result.expect("block download and verify tasks must not panic") {
|
||||||
|
Ok(hash) => {
|
||||||
|
this.cancel_handles.remove(&hash);
|
||||||
|
Poll::Ready(Some(Ok(hash)))
|
||||||
|
}
|
||||||
|
Err((e, hash)) => {
|
||||||
|
this.cancel_handles.remove(&hash);
|
||||||
|
Poll::Ready(Some(Err(e)))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Poll::Ready(None)
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn size_hint(&self) -> (usize, Option<usize>) {
|
||||||
|
self.pending.size_hint()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
impl<ZN, ZV, ZS> Downloads<ZN, ZV, ZS>
|
||||||
|
where
|
||||||
|
ZN: Service<zn::Request, Response = zn::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZN::Future: Send,
|
||||||
|
ZV: Service<Arc<Block>, Response = block::Hash, Error = BoxError> + Send + Clone + 'static,
|
||||||
|
ZV::Future: Send,
|
||||||
|
ZS: Service<zs::Request, Response = zs::Response, Error = BoxError> + Send + 'static,
|
||||||
|
ZS::Future: Send,
|
||||||
|
{
|
||||||
|
/// Initialize a new download stream with the provided `network` and
|
||||||
|
/// `verifier` services.
|
||||||
|
///
|
||||||
|
/// The [`Downloads`] stream is agnostic to the network policy, so retry and
|
||||||
|
/// timeout limits should be applied to the `network` service passed into
|
||||||
|
/// this constructor.
|
||||||
|
pub fn new(network: ZN, verifier: ZV, state: ZS) -> Self {
|
||||||
|
Self {
|
||||||
|
network,
|
||||||
|
verifier,
|
||||||
|
state,
|
||||||
|
pending: FuturesUnordered::new(),
|
||||||
|
cancel_handles: HashMap::new(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
/// Queue a block for download and verification.
|
||||||
|
///
|
||||||
|
/// This method waits for the network to become ready, and returns an error
|
||||||
|
/// only if the network service fails. It returns immediately after queuing
|
||||||
|
/// the request.
|
||||||
|
#[instrument(skip(self))]
|
||||||
|
pub async fn download_and_verify(&mut self, hash: block::Hash) -> Result<bool, Report> {
|
||||||
|
if self.cancel_handles.contains_key(&hash) {
|
||||||
|
return Ok(false);
|
||||||
|
}
|
||||||
|
|
||||||
|
// Check if the block is already in the state.
|
||||||
|
let block_state_req = self
|
||||||
|
.state
|
||||||
|
.ready_and()
|
||||||
|
.await
|
||||||
|
.map_err(|e| eyre!(e))?
|
||||||
|
.call(zs::Request::Block(zs::HashOrHeight::from(hash)))
|
||||||
|
.await;
|
||||||
|
|
||||||
|
let block_in_chain = match block_state_req {
|
||||||
|
Ok(zs::Response::Block(block)) => block,
|
||||||
|
_ => None,
|
||||||
|
};
|
||||||
|
|
||||||
|
// Block already in state, get out.
|
||||||
|
if block_in_chain.is_some() {
|
||||||
|
return Ok(false);
|
||||||
|
}
|
||||||
|
|
||||||
|
// We construct the block requests sequentially, waiting for the peer
|
||||||
|
// set to be ready to process each request. This ensures that we start
|
||||||
|
// block downloads in the order we want them (though they may resolve
|
||||||
|
// out of order), and it means that we respect backpressure. Otherwise,
|
||||||
|
// if we waited for readiness and did the service call in the spawned
|
||||||
|
// tasks, all of the spawned tasks would race each other waiting for the
|
||||||
|
// network to become ready.
|
||||||
|
tracing::debug!("waiting to request block");
|
||||||
|
let block_req = self
|
||||||
|
.network
|
||||||
|
.ready_and()
|
||||||
|
.await
|
||||||
|
.map_err(|e| eyre!(e))?
|
||||||
|
.call(zn::Request::BlocksByHash(std::iter::once(hash).collect()));
|
||||||
|
tracing::debug!("requested block");
|
||||||
|
|
||||||
|
// This oneshot is used to signal cancellation to the download task.
|
||||||
|
let (cancel_tx, mut cancel_rx) = oneshot::channel::<()>();
|
||||||
|
|
||||||
|
let mut verifier = self.verifier.clone();
|
||||||
|
let task = tokio::spawn(
|
||||||
|
async move {
|
||||||
|
let rsp = tokio::select! {
|
||||||
|
_ = &mut cancel_rx => {
|
||||||
|
tracing::trace!("task cancelled prior to download completion");
|
||||||
|
metrics::counter!("gossip.cancelled.download.count", 1);
|
||||||
|
return Err("canceled block_fetch_verify".into())
|
||||||
|
}
|
||||||
|
rsp = block_req => rsp?,
|
||||||
|
};
|
||||||
|
|
||||||
|
let block = if let zn::Response::Blocks(blocks) = rsp {
|
||||||
|
blocks
|
||||||
|
.into_iter()
|
||||||
|
.next()
|
||||||
|
.expect("successful response has the block in it")
|
||||||
|
} else {
|
||||||
|
unreachable!("wrong response to block request");
|
||||||
|
};
|
||||||
|
metrics::counter!("gossip.downloaded.block.count", 1);
|
||||||
|
|
||||||
|
let rsp = verifier.ready_and().await?.call(block);
|
||||||
|
let verification = tokio::select! {
|
||||||
|
_ = &mut cancel_rx => {
|
||||||
|
tracing::trace!("task cancelled prior to verification");
|
||||||
|
metrics::counter!("gossip.cancelled.verify.count", 1);
|
||||||
|
return Err("canceled block_fetch_verify".into())
|
||||||
|
}
|
||||||
|
verification = rsp => verification,
|
||||||
|
};
|
||||||
|
if verification.is_ok() {
|
||||||
|
metrics::counter!("sync.verified.block.count", 1);
|
||||||
|
}
|
||||||
|
|
||||||
|
verification
|
||||||
|
}
|
||||||
|
.in_current_span()
|
||||||
|
// Tack the hash onto the error so we can remove the cancel handle
|
||||||
|
// on failure as well as on success.
|
||||||
|
.map_err(move |e| (e, hash)),
|
||||||
|
);
|
||||||
|
|
||||||
|
self.pending.push(task);
|
||||||
|
// XXX replace with expect_none when stable
|
||||||
|
assert!(
|
||||||
|
self.cancel_handles.insert(hash, cancel_tx).is_none(),
|
||||||
|
"blocks are only queued once"
|
||||||
|
);
|
||||||
|
|
||||||
|
Ok(true)
|
||||||
|
}
|
||||||
|
}
|
Loading…
Reference in New Issue