824 lines
30 KiB
Rust
824 lines
30 KiB
Rust
//! A Tokio codec mapping byte streams to Bitcoin message streams.
|
|
|
|
use std::fmt;
|
|
use std::io::{Cursor, Read, Write};
|
|
|
|
use byteorder::{LittleEndian, ReadBytesExt, WriteBytesExt};
|
|
use bytes::{BufMut, BytesMut};
|
|
use chrono::{TimeZone, Utc};
|
|
use tokio_util::codec::{Decoder, Encoder};
|
|
|
|
use zebra_chain::{
|
|
block::{self, Block},
|
|
parameters::Network,
|
|
serialization::{
|
|
sha256d, ReadZcashExt, SerializationError as Error, WriteZcashExt, ZcashDeserialize,
|
|
ZcashSerialize,
|
|
},
|
|
transaction::Transaction,
|
|
};
|
|
|
|
use crate::constants;
|
|
|
|
use super::{
|
|
message::{Message, RejectReason},
|
|
types::*,
|
|
};
|
|
|
|
/// The length of a Bitcoin message header.
|
|
const HEADER_LEN: usize = 24usize;
|
|
|
|
/// Maximum size of a protocol message body.
|
|
const MAX_PROTOCOL_MESSAGE_LEN: usize = 2 * 1024 * 1024;
|
|
|
|
/// A codec which produces Bitcoin messages from byte streams and vice versa.
|
|
pub struct Codec {
|
|
builder: Builder,
|
|
state: DecodeState,
|
|
}
|
|
|
|
/// A builder for specifying [`Codec`] options.
|
|
pub struct Builder {
|
|
/// The network magic to use in encoding.
|
|
network: Network,
|
|
/// The protocol version to speak when encoding/decoding.
|
|
version: Version,
|
|
/// The maximum allowable message length.
|
|
max_len: usize,
|
|
/// An optional label to use for reporting metrics.
|
|
metrics_label: Option<String>,
|
|
}
|
|
|
|
impl Codec {
|
|
/// Return a builder for constructing a [`Codec`].
|
|
pub fn builder() -> Builder {
|
|
Builder {
|
|
network: Network::Mainnet,
|
|
version: constants::CURRENT_VERSION,
|
|
max_len: MAX_PROTOCOL_MESSAGE_LEN,
|
|
metrics_label: None,
|
|
}
|
|
}
|
|
|
|
/// Reconfigure the version used by the codec, e.g., after completing a handshake.
|
|
pub fn reconfigure_version(&mut self, version: Version) {
|
|
self.builder.version = version;
|
|
}
|
|
}
|
|
|
|
impl Builder {
|
|
/// Finalize the builder and return a [`Codec`].
|
|
pub fn finish(self) -> Codec {
|
|
Codec {
|
|
builder: self,
|
|
state: DecodeState::Head,
|
|
}
|
|
}
|
|
|
|
/// Configure the codec for the given [`Network`].
|
|
pub fn for_network(mut self, network: Network) -> Self {
|
|
self.network = network;
|
|
self
|
|
}
|
|
|
|
/// Configure the codec for the given [`Version`].
|
|
#[allow(dead_code)]
|
|
pub fn for_version(mut self, version: Version) -> Self {
|
|
self.version = version;
|
|
self
|
|
}
|
|
|
|
/// Configure the codec's maximum accepted payload size, in bytes.
|
|
#[allow(dead_code)]
|
|
pub fn with_max_body_len(mut self, len: usize) -> Self {
|
|
self.max_len = len;
|
|
self
|
|
}
|
|
|
|
/// Configure the codec for the given peer address.
|
|
pub fn with_metrics_label(mut self, metrics_label: String) -> Self {
|
|
self.metrics_label = Some(metrics_label);
|
|
self
|
|
}
|
|
}
|
|
|
|
// ======== Encoding =========
|
|
|
|
impl Encoder<Message> for Codec {
|
|
type Error = Error;
|
|
|
|
fn encode(&mut self, item: Message, dst: &mut BytesMut) -> Result<(), Self::Error> {
|
|
use Error::Parse;
|
|
|
|
let body_length = self.body_length(&item);
|
|
|
|
if body_length > self.builder.max_len {
|
|
return Err(Parse("body length exceeded maximum size"));
|
|
}
|
|
|
|
if let Some(label) = self.builder.metrics_label.clone() {
|
|
metrics::counter!("zcash.net.out.bytes.total", (body_length + HEADER_LEN) as u64, "addr" => label);
|
|
}
|
|
|
|
use Message::*;
|
|
// Note: because all match arms must have
|
|
// the same type, and the array length is
|
|
// part of the type, having at least one
|
|
// of length 12 checks that they are all
|
|
// of length 12, as they must be &[u8; 12].
|
|
let command = match item {
|
|
Version { .. } => b"version\0\0\0\0\0",
|
|
Verack { .. } => b"verack\0\0\0\0\0\0",
|
|
Ping { .. } => b"ping\0\0\0\0\0\0\0\0",
|
|
Pong { .. } => b"pong\0\0\0\0\0\0\0\0",
|
|
Reject { .. } => b"reject\0\0\0\0\0\0",
|
|
Addr { .. } => b"addr\0\0\0\0\0\0\0\0",
|
|
GetAddr { .. } => b"getaddr\0\0\0\0\0",
|
|
Block { .. } => b"block\0\0\0\0\0\0\0",
|
|
GetBlocks { .. } => b"getblocks\0\0\0",
|
|
Headers { .. } => b"headers\0\0\0\0\0",
|
|
GetHeaders { .. } => b"getheaders\0\0",
|
|
Inv { .. } => b"inv\0\0\0\0\0\0\0\0\0",
|
|
GetData { .. } => b"getdata\0\0\0\0\0",
|
|
NotFound { .. } => b"notfound\0\0\0\0",
|
|
Tx { .. } => b"tx\0\0\0\0\0\0\0\0\0\0",
|
|
Mempool { .. } => b"mempool\0\0\0\0\0",
|
|
FilterLoad { .. } => b"filterload\0\0",
|
|
FilterAdd { .. } => b"filteradd\0\0\0",
|
|
FilterClear { .. } => b"filterclear\0",
|
|
};
|
|
trace!(?item, len = body_length);
|
|
|
|
dst.reserve(HEADER_LEN + body_length);
|
|
let start_len = dst.len();
|
|
{
|
|
let dst = &mut dst.writer();
|
|
dst.write_all(&Magic::from(self.builder.network).0[..])?;
|
|
dst.write_all(command)?;
|
|
dst.write_u32::<LittleEndian>(body_length as u32)?;
|
|
|
|
// We zero the checksum at first, and compute it later
|
|
// after the body has been written.
|
|
dst.write_u32::<LittleEndian>(0)?;
|
|
|
|
self.write_body(&item, dst)?;
|
|
}
|
|
let checksum = sha256d::Checksum::from(&dst[start_len + HEADER_LEN..]);
|
|
dst[start_len + 20..][..4].copy_from_slice(&checksum.0);
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
impl Codec {
|
|
/// Obtain the size of the body of a given message. This will match the
|
|
/// number of bytes written to the writer provided to `write_body` for the
|
|
/// same message.
|
|
///
|
|
/// TODO: Replace with a size estimate, to avoid multiple serializations
|
|
/// for large data structures like lists, blocks, and transactions.
|
|
/// See #1774.
|
|
fn body_length(&self, msg: &Message) -> usize {
|
|
struct FakeWriter(usize);
|
|
|
|
impl std::io::Write for FakeWriter {
|
|
fn write(&mut self, buf: &[u8]) -> std::io::Result<usize> {
|
|
self.0 += buf.len();
|
|
|
|
Ok(buf.len())
|
|
}
|
|
|
|
fn flush(&mut self) -> std::io::Result<()> {
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
let mut writer = FakeWriter(0);
|
|
self.write_body(msg, &mut writer)
|
|
.expect("writer should never fail");
|
|
writer.0
|
|
}
|
|
|
|
/// Write the body of the message into the given writer. This allows writing
|
|
/// the message body prior to writing the header, so that the header can
|
|
/// contain a checksum of the message body.
|
|
fn write_body<W: Write>(&self, msg: &Message, mut writer: W) -> Result<(), Error> {
|
|
match msg {
|
|
Message::Version {
|
|
version,
|
|
services,
|
|
timestamp,
|
|
address_recv,
|
|
address_from,
|
|
nonce,
|
|
user_agent,
|
|
start_height,
|
|
relay,
|
|
} => {
|
|
writer.write_u32::<LittleEndian>(version.0)?;
|
|
writer.write_u64::<LittleEndian>(services.bits())?;
|
|
writer.write_i64::<LittleEndian>(timestamp.timestamp())?;
|
|
|
|
let (recv_services, recv_addr) = address_recv;
|
|
writer.write_u64::<LittleEndian>(recv_services.bits())?;
|
|
writer.write_socket_addr(*recv_addr)?;
|
|
|
|
let (from_services, from_addr) = address_from;
|
|
writer.write_u64::<LittleEndian>(from_services.bits())?;
|
|
writer.write_socket_addr(*from_addr)?;
|
|
|
|
writer.write_u64::<LittleEndian>(nonce.0)?;
|
|
writer.write_string(&user_agent)?;
|
|
writer.write_u32::<LittleEndian>(start_height.0)?;
|
|
writer.write_u8(*relay as u8)?;
|
|
}
|
|
Message::Verack => { /* Empty payload -- no-op */ }
|
|
Message::Ping(nonce) => {
|
|
writer.write_u64::<LittleEndian>(nonce.0)?;
|
|
}
|
|
Message::Pong(nonce) => {
|
|
writer.write_u64::<LittleEndian>(nonce.0)?;
|
|
}
|
|
Message::Reject {
|
|
message,
|
|
ccode,
|
|
reason,
|
|
data,
|
|
} => {
|
|
writer.write_string(&message)?;
|
|
writer.write_u8(*ccode as u8)?;
|
|
writer.write_string(&reason)?;
|
|
writer.write_all(&data.unwrap())?;
|
|
}
|
|
Message::Addr(addrs) => addrs.zcash_serialize(&mut writer)?,
|
|
Message::GetAddr => { /* Empty payload -- no-op */ }
|
|
Message::Block(block) => block.zcash_serialize(&mut writer)?,
|
|
Message::GetBlocks { known_blocks, stop } => {
|
|
writer.write_u32::<LittleEndian>(self.builder.version.0)?;
|
|
known_blocks.zcash_serialize(&mut writer)?;
|
|
stop.unwrap_or(block::Hash([0; 32]))
|
|
.zcash_serialize(&mut writer)?;
|
|
}
|
|
Message::GetHeaders { known_blocks, stop } => {
|
|
writer.write_u32::<LittleEndian>(self.builder.version.0)?;
|
|
known_blocks.zcash_serialize(&mut writer)?;
|
|
stop.unwrap_or(block::Hash([0; 32]))
|
|
.zcash_serialize(&mut writer)?;
|
|
}
|
|
Message::Headers(headers) => headers.zcash_serialize(&mut writer)?,
|
|
Message::Inv(hashes) => hashes.zcash_serialize(&mut writer)?,
|
|
Message::GetData(hashes) => hashes.zcash_serialize(&mut writer)?,
|
|
Message::NotFound(hashes) => hashes.zcash_serialize(&mut writer)?,
|
|
Message::Tx(transaction) => transaction.zcash_serialize(&mut writer)?,
|
|
Message::Mempool => { /* Empty payload -- no-op */ }
|
|
Message::FilterLoad {
|
|
filter,
|
|
hash_functions_count,
|
|
tweak,
|
|
flags,
|
|
} => {
|
|
writer.write_all(&filter.0)?;
|
|
writer.write_u32::<LittleEndian>(*hash_functions_count)?;
|
|
writer.write_u32::<LittleEndian>(tweak.0)?;
|
|
writer.write_u8(*flags)?;
|
|
}
|
|
Message::FilterAdd { data } => {
|
|
writer.write_all(data)?;
|
|
}
|
|
Message::FilterClear => { /* Empty payload -- no-op */ }
|
|
}
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
// ======== Decoding =========
|
|
|
|
enum DecodeState {
|
|
Head,
|
|
Body {
|
|
body_len: usize,
|
|
command: [u8; 12],
|
|
checksum: sha256d::Checksum,
|
|
},
|
|
}
|
|
|
|
impl fmt::Debug for DecodeState {
|
|
fn fmt(&self, f: &mut fmt::Formatter) -> fmt::Result {
|
|
match self {
|
|
DecodeState::Head => write!(f, "DecodeState::Head"),
|
|
DecodeState::Body {
|
|
body_len,
|
|
command,
|
|
checksum,
|
|
} => f
|
|
.debug_struct("DecodeState::Body")
|
|
.field("body_len", &body_len)
|
|
.field("command", &String::from_utf8_lossy(command))
|
|
.field("checksum", &checksum)
|
|
.finish(),
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Decoder for Codec {
|
|
type Item = Message;
|
|
type Error = Error;
|
|
|
|
fn decode(&mut self, src: &mut BytesMut) -> Result<Option<Self::Item>, Self::Error> {
|
|
use Error::Parse;
|
|
match self.state {
|
|
DecodeState::Head => {
|
|
// First check that the src buffer contains an entire header.
|
|
if src.len() < HEADER_LEN {
|
|
trace!(?self.state, "src buffer does not have an entire header, waiting");
|
|
// Signal that decoding requires more data.
|
|
return Ok(None);
|
|
}
|
|
|
|
// Now that we know that src contains a header, split off the header section.
|
|
let header = src.split_to(HEADER_LEN);
|
|
|
|
// Create a cursor over the header and parse its fields.
|
|
let mut header_reader = Cursor::new(&header);
|
|
let magic = Magic(header_reader.read_4_bytes()?);
|
|
let command = header_reader.read_12_bytes()?;
|
|
let body_len = header_reader.read_u32::<LittleEndian>()? as usize;
|
|
let checksum = sha256d::Checksum(header_reader.read_4_bytes()?);
|
|
trace!(
|
|
?self.state,
|
|
?magic,
|
|
command = %String::from_utf8(
|
|
command.iter()
|
|
.cloned()
|
|
.flat_map(std::ascii::escape_default)
|
|
.collect()
|
|
).unwrap(),
|
|
body_len,
|
|
?checksum,
|
|
"read header from src buffer"
|
|
);
|
|
|
|
if magic != Magic::from(self.builder.network) {
|
|
return Err(Parse("supplied magic did not meet expectations"));
|
|
}
|
|
if body_len > self.builder.max_len {
|
|
return Err(Parse("body length exceeded maximum size"));
|
|
}
|
|
|
|
if let Some(label) = self.builder.metrics_label.clone() {
|
|
metrics::counter!("zcash.net.in.bytes.total", (body_len + HEADER_LEN) as u64, "addr" => label);
|
|
}
|
|
|
|
// Reserve buffer space for the expected body and the following header.
|
|
src.reserve(body_len + HEADER_LEN);
|
|
|
|
self.state = DecodeState::Body {
|
|
body_len,
|
|
command,
|
|
checksum,
|
|
};
|
|
|
|
// Now that the state is updated, recurse to attempt body decoding.
|
|
self.decode(src)
|
|
}
|
|
DecodeState::Body {
|
|
body_len,
|
|
command,
|
|
checksum,
|
|
} => {
|
|
if src.len() < body_len {
|
|
// Need to wait for the full body
|
|
trace!(?self.state, len = src.len(), "src buffer does not have an entire body, waiting");
|
|
return Ok(None);
|
|
}
|
|
|
|
// Now that we know we have the full body, split off the body,
|
|
// and reset the decoder state for the next message. Otherwise
|
|
// we will attempt to read the next header as the current body.
|
|
let body = src.split_to(body_len);
|
|
self.state = DecodeState::Head;
|
|
|
|
if checksum != sha256d::Checksum::from(&body[..]) {
|
|
return Err(Parse(
|
|
"supplied message checksum does not match computed checksum",
|
|
));
|
|
}
|
|
|
|
let mut body_reader = Cursor::new(&body);
|
|
match &command {
|
|
b"version\0\0\0\0\0" => self.read_version(&mut body_reader),
|
|
b"verack\0\0\0\0\0\0" => self.read_verack(&mut body_reader),
|
|
b"ping\0\0\0\0\0\0\0\0" => self.read_ping(&mut body_reader),
|
|
b"pong\0\0\0\0\0\0\0\0" => self.read_pong(&mut body_reader),
|
|
b"reject\0\0\0\0\0\0" => self.read_reject(&mut body_reader),
|
|
b"addr\0\0\0\0\0\0\0\0" => self.read_addr(&mut body_reader),
|
|
b"getaddr\0\0\0\0\0" => self.read_getaddr(&mut body_reader),
|
|
b"block\0\0\0\0\0\0\0" => self.read_block(&mut body_reader),
|
|
b"getblocks\0\0\0" => self.read_getblocks(&mut body_reader),
|
|
b"headers\0\0\0\0\0" => self.read_headers(&mut body_reader),
|
|
b"getheaders\0\0" => self.read_getheaders(&mut body_reader),
|
|
b"inv\0\0\0\0\0\0\0\0\0" => self.read_inv(&mut body_reader),
|
|
b"getdata\0\0\0\0\0" => self.read_getdata(&mut body_reader),
|
|
b"notfound\0\0\0\0" => self.read_notfound(&mut body_reader),
|
|
b"tx\0\0\0\0\0\0\0\0\0\0" => self.read_tx(&mut body_reader),
|
|
b"mempool\0\0\0\0\0" => self.read_mempool(&mut body_reader),
|
|
b"filterload\0\0" => self.read_filterload(&mut body_reader, body_len),
|
|
b"filteradd\0\0\0" => self.read_filteradd(&mut body_reader),
|
|
b"filterclear\0" => self.read_filterclear(&mut body_reader),
|
|
_ => return Err(Parse("unknown command")),
|
|
}
|
|
// We need Ok(Some(msg)) to signal that we're done decoding.
|
|
// This is also convenient for tracing the parse result.
|
|
.map(|msg| {
|
|
// bitcoin allows extra data at the end of most messages,
|
|
// so that old nodes can still read newer message formats,
|
|
// and ignore any extra fields
|
|
let extra_bytes = body.len() as u64 - body_reader.position();
|
|
if extra_bytes == 0 {
|
|
trace!(?extra_bytes, %msg, "finished message decoding");
|
|
} else {
|
|
// log when there are extra bytes, so we know when we need to
|
|
// upgrade message formats
|
|
debug!(?extra_bytes, %msg, "extra data after decoding message");
|
|
}
|
|
Some(msg)
|
|
})
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
impl Codec {
|
|
fn read_version<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Version {
|
|
version: Version(reader.read_u32::<LittleEndian>()?),
|
|
// Use from_bits_truncate to discard unknown service bits.
|
|
services: PeerServices::from_bits_truncate(reader.read_u64::<LittleEndian>()?),
|
|
timestamp: Utc.timestamp(reader.read_i64::<LittleEndian>()?, 0),
|
|
address_recv: (
|
|
PeerServices::from_bits_truncate(reader.read_u64::<LittleEndian>()?),
|
|
reader.read_socket_addr()?,
|
|
),
|
|
address_from: (
|
|
PeerServices::from_bits_truncate(reader.read_u64::<LittleEndian>()?),
|
|
reader.read_socket_addr()?,
|
|
),
|
|
nonce: Nonce(reader.read_u64::<LittleEndian>()?),
|
|
user_agent: String::zcash_deserialize(&mut reader)?,
|
|
start_height: block::Height(reader.read_u32::<LittleEndian>()?),
|
|
relay: match reader.read_u8()? {
|
|
0 => false,
|
|
1 => true,
|
|
_ => return Err(Error::Parse("non-bool value supplied in relay field")),
|
|
},
|
|
})
|
|
}
|
|
|
|
fn read_verack<R: Read>(&self, mut _reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Verack)
|
|
}
|
|
|
|
fn read_ping<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Ping(Nonce(reader.read_u64::<LittleEndian>()?)))
|
|
}
|
|
|
|
fn read_pong<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Pong(Nonce(reader.read_u64::<LittleEndian>()?)))
|
|
}
|
|
|
|
fn read_reject<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Reject {
|
|
message: String::zcash_deserialize(&mut reader)?,
|
|
ccode: match reader.read_u8()? {
|
|
0x01 => RejectReason::Malformed,
|
|
0x10 => RejectReason::Invalid,
|
|
0x11 => RejectReason::Obsolete,
|
|
0x12 => RejectReason::Duplicate,
|
|
0x40 => RejectReason::Nonstandard,
|
|
0x41 => RejectReason::Dust,
|
|
0x42 => RejectReason::InsufficientFee,
|
|
0x43 => RejectReason::Checkpoint,
|
|
0x50 => RejectReason::Other,
|
|
_ => return Err(Error::Parse("invalid RejectReason value in ccode field")),
|
|
},
|
|
reason: String::zcash_deserialize(&mut reader)?,
|
|
// Sometimes there's data, sometimes there isn't. There's no length
|
|
// field, this is just implicitly encoded by the body_len.
|
|
// Apparently all existing implementations only supply 32 bytes of
|
|
// data (hash identifying the rejected object) or none (and we model
|
|
// the Reject message that way), so instead of passing in the
|
|
// body_len separately and calculating remaining bytes, just try to
|
|
// read 32 bytes and ignore any failures.
|
|
data: reader.read_32_bytes().ok(),
|
|
})
|
|
}
|
|
|
|
fn read_addr<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Addr(Vec::zcash_deserialize(reader)?))
|
|
}
|
|
|
|
fn read_getaddr<R: Read>(&self, mut _reader: R) -> Result<Message, Error> {
|
|
Ok(Message::GetAddr)
|
|
}
|
|
|
|
fn read_block<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Block(Block::zcash_deserialize(reader)?.into()))
|
|
}
|
|
|
|
fn read_getblocks<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
if self.builder.version == Version(reader.read_u32::<LittleEndian>()?) {
|
|
let known_blocks = Vec::zcash_deserialize(&mut reader)?;
|
|
let stop_hash = block::Hash::zcash_deserialize(&mut reader)?;
|
|
let stop = if stop_hash != block::Hash([0; 32]) {
|
|
Some(stop_hash)
|
|
} else {
|
|
None
|
|
};
|
|
Ok(Message::GetBlocks { known_blocks, stop })
|
|
} else {
|
|
Err(Error::Parse("getblocks version did not match negotiation"))
|
|
}
|
|
}
|
|
|
|
/// Deserialize a `headers` message.
|
|
///
|
|
/// See [Zcash block header] for the enumeration of these fields.
|
|
///
|
|
/// [Zcash block header](https://zips.z.cash/protocol/protocol.pdf#page=84)
|
|
fn read_headers<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Headers(Vec::zcash_deserialize(&mut reader)?))
|
|
}
|
|
|
|
fn read_getheaders<R: Read>(&self, mut reader: R) -> Result<Message, Error> {
|
|
if self.builder.version == Version(reader.read_u32::<LittleEndian>()?) {
|
|
let known_blocks = Vec::zcash_deserialize(&mut reader)?;
|
|
let stop_hash = block::Hash::zcash_deserialize(&mut reader)?;
|
|
let stop = if stop_hash != block::Hash([0; 32]) {
|
|
Some(stop_hash)
|
|
} else {
|
|
None
|
|
};
|
|
Ok(Message::GetHeaders { known_blocks, stop })
|
|
} else {
|
|
Err(Error::Parse("getblocks version did not match negotiation"))
|
|
}
|
|
}
|
|
|
|
fn read_inv<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Inv(Vec::zcash_deserialize(reader)?))
|
|
}
|
|
|
|
fn read_getdata<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::GetData(Vec::zcash_deserialize(reader)?))
|
|
}
|
|
|
|
fn read_notfound<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::NotFound(Vec::zcash_deserialize(reader)?))
|
|
}
|
|
|
|
fn read_tx<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Tx(Transaction::zcash_deserialize(reader)?.into()))
|
|
}
|
|
|
|
fn read_mempool<R: Read>(&self, mut _reader: R) -> Result<Message, Error> {
|
|
Ok(Message::Mempool)
|
|
}
|
|
|
|
fn read_filterload<R: Read>(&self, mut reader: R, body_len: usize) -> Result<Message, Error> {
|
|
if !(FILTERLOAD_REMAINDER_LENGTH <= body_len
|
|
&& body_len <= FILTERLOAD_REMAINDER_LENGTH + MAX_FILTER_LENGTH)
|
|
{
|
|
return Err(Error::Parse("Invalid filterload message body length."));
|
|
}
|
|
|
|
const MAX_FILTER_LENGTH: usize = 36000;
|
|
const FILTERLOAD_REMAINDER_LENGTH: usize = 4 + 4 + 1;
|
|
|
|
let filter_length: usize = body_len - FILTERLOAD_REMAINDER_LENGTH;
|
|
|
|
let mut filter_bytes = vec![0; filter_length];
|
|
reader.read_exact(&mut filter_bytes)?;
|
|
|
|
Ok(Message::FilterLoad {
|
|
filter: Filter(filter_bytes),
|
|
hash_functions_count: reader.read_u32::<LittleEndian>()?,
|
|
tweak: Tweak(reader.read_u32::<LittleEndian>()?),
|
|
flags: reader.read_u8()?,
|
|
})
|
|
}
|
|
|
|
fn read_filteradd<R: Read>(&self, reader: R) -> Result<Message, Error> {
|
|
let mut bytes = Vec::new();
|
|
|
|
// Maximum size of data is 520 bytes.
|
|
reader.take(520).read_exact(&mut bytes)?;
|
|
|
|
Ok(Message::FilterAdd { data: bytes })
|
|
}
|
|
|
|
fn read_filterclear<R: Read>(&self, mut _reader: R) -> Result<Message, Error> {
|
|
Ok(Message::FilterClear)
|
|
}
|
|
}
|
|
|
|
// XXX replace these interior unit tests with exterior integration tests + proptest
|
|
#[cfg(test)]
|
|
mod tests {
|
|
use super::*;
|
|
use futures::prelude::*;
|
|
use tokio::runtime::Runtime;
|
|
|
|
#[test]
|
|
fn version_message_round_trip() {
|
|
zebra_test::init();
|
|
use std::net::{IpAddr, Ipv4Addr, SocketAddr};
|
|
let services = PeerServices::NODE_NETWORK;
|
|
let timestamp = Utc.timestamp(1_568_000_000, 0);
|
|
|
|
let rt = Runtime::new().unwrap();
|
|
|
|
let v = Message::Version {
|
|
version: crate::constants::CURRENT_VERSION,
|
|
services,
|
|
timestamp,
|
|
address_recv: (
|
|
services,
|
|
SocketAddr::new(IpAddr::V4(Ipv4Addr::new(203, 0, 113, 6)), 8233),
|
|
),
|
|
address_from: (
|
|
services,
|
|
SocketAddr::new(IpAddr::V4(Ipv4Addr::new(203, 0, 113, 6)), 8233),
|
|
),
|
|
nonce: Nonce(0x9082_4908_8927_9238),
|
|
user_agent: "Zebra".to_owned(),
|
|
start_height: block::Height(540_000),
|
|
relay: true,
|
|
};
|
|
|
|
use tokio_util::codec::{FramedRead, FramedWrite};
|
|
let v_bytes = rt.block_on(async {
|
|
let mut bytes = Vec::new();
|
|
{
|
|
let mut fw = FramedWrite::new(&mut bytes, Codec::builder().finish());
|
|
fw.send(v.clone())
|
|
.await
|
|
.expect("message should be serialized");
|
|
}
|
|
bytes
|
|
});
|
|
|
|
let v_parsed = rt.block_on(async {
|
|
let mut fr = FramedRead::new(Cursor::new(&v_bytes), Codec::builder().finish());
|
|
fr.next()
|
|
.await
|
|
.expect("a next message should be available")
|
|
.expect("that message should deserialize")
|
|
});
|
|
|
|
assert_eq!(v, v_parsed);
|
|
}
|
|
|
|
#[test]
|
|
fn filterload_message_round_trip() {
|
|
zebra_test::init();
|
|
|
|
let rt = Runtime::new().unwrap();
|
|
|
|
let v = Message::FilterLoad {
|
|
filter: Filter(vec![0; 35999]),
|
|
hash_functions_count: 0,
|
|
tweak: Tweak(0),
|
|
flags: 0,
|
|
};
|
|
|
|
use tokio_util::codec::{FramedRead, FramedWrite};
|
|
let v_bytes = rt.block_on(async {
|
|
let mut bytes = Vec::new();
|
|
{
|
|
let mut fw = FramedWrite::new(&mut bytes, Codec::builder().finish());
|
|
fw.send(v.clone())
|
|
.await
|
|
.expect("message should be serialized");
|
|
}
|
|
bytes
|
|
});
|
|
|
|
let v_parsed = rt.block_on(async {
|
|
let mut fr = FramedRead::new(Cursor::new(&v_bytes), Codec::builder().finish());
|
|
fr.next()
|
|
.await
|
|
.expect("a next message should be available")
|
|
.expect("that message should deserialize")
|
|
});
|
|
|
|
assert_eq!(v, v_parsed);
|
|
}
|
|
|
|
#[test]
|
|
fn filterload_message_too_large_round_trip() {
|
|
zebra_test::init();
|
|
|
|
let rt = Runtime::new().unwrap();
|
|
|
|
let v = Message::FilterLoad {
|
|
filter: Filter(vec![0; 40000]),
|
|
hash_functions_count: 0,
|
|
tweak: Tweak(0),
|
|
flags: 0,
|
|
};
|
|
|
|
use tokio_util::codec::{FramedRead, FramedWrite};
|
|
let v_bytes = rt.block_on(async {
|
|
let mut bytes = Vec::new();
|
|
{
|
|
let mut fw = FramedWrite::new(&mut bytes, Codec::builder().finish());
|
|
fw.send(v.clone())
|
|
.await
|
|
.expect("message should be serialized");
|
|
}
|
|
bytes
|
|
});
|
|
|
|
rt.block_on(async {
|
|
let mut fr = FramedRead::new(Cursor::new(&v_bytes), Codec::builder().finish());
|
|
fr.next()
|
|
.await
|
|
.expect("a next message should be available")
|
|
.expect_err("that message should not deserialize")
|
|
});
|
|
}
|
|
|
|
#[test]
|
|
fn max_msg_size_round_trip() {
|
|
use std::sync::Arc;
|
|
use zebra_chain::serialization::ZcashDeserializeInto;
|
|
zebra_test::init();
|
|
|
|
let rt = Runtime::new().unwrap();
|
|
|
|
// make tests with a Tx message
|
|
let tx = zebra_test::vectors::DUMMY_TX1
|
|
.zcash_deserialize_into()
|
|
.unwrap();
|
|
let msg = Message::Tx(Arc::new(tx));
|
|
|
|
use tokio_util::codec::{FramedRead, FramedWrite};
|
|
|
|
// i know the above msg has a body of 85 bytes
|
|
let size = 85;
|
|
|
|
// reducing the max size to body size - 1
|
|
rt.block_on(async {
|
|
let mut bytes = Vec::new();
|
|
{
|
|
let mut fw = FramedWrite::new(
|
|
&mut bytes,
|
|
Codec::builder().with_max_body_len(size - 1).finish(),
|
|
);
|
|
fw.send(msg.clone()).await.expect_err(
|
|
"message should not encode as it is bigger than the max allowed value",
|
|
);
|
|
}
|
|
});
|
|
|
|
// send again with the msg body size as max size
|
|
let msg_bytes = rt.block_on(async {
|
|
let mut bytes = Vec::new();
|
|
{
|
|
let mut fw = FramedWrite::new(
|
|
&mut bytes,
|
|
Codec::builder().with_max_body_len(size).finish(),
|
|
);
|
|
fw.send(msg.clone())
|
|
.await
|
|
.expect("message should encode with the msg body size as max allowed value");
|
|
}
|
|
bytes
|
|
});
|
|
|
|
// receive with a reduced max size
|
|
rt.block_on(async {
|
|
let mut fr = FramedRead::new(
|
|
Cursor::new(&msg_bytes),
|
|
Codec::builder().with_max_body_len(size - 1).finish(),
|
|
);
|
|
fr.next()
|
|
.await
|
|
.expect("a next message should be available")
|
|
.expect_err("message should not decode as it is bigger than the max allowed value")
|
|
});
|
|
|
|
// receive again with the tx size as max size
|
|
rt.block_on(async {
|
|
let mut fr = FramedRead::new(
|
|
Cursor::new(&msg_bytes),
|
|
Codec::builder().with_max_body_len(size).finish(),
|
|
);
|
|
fr.next()
|
|
.await
|
|
.expect("a next message should be available")
|
|
.expect("message should decode with the msg body size as max allowed value")
|
|
});
|
|
}
|
|
}
|