buffer: Move buffer to service.rs and layer to layer.rs (#246)

This commit is contained in:
Lucio Franco 2019-04-10 15:35:05 -04:00 committed by Carl Lerche
parent 1012b20737
commit 23ce9133b5
3 changed files with 167 additions and 156 deletions

46
tower-buffer/src/layer.rs Normal file
View File

@ -0,0 +1,46 @@
use crate::{error::Error, service::Buffer, worker::WorkerExecutor};
use tokio_executor::DefaultExecutor;
use tower_layer::Layer;
use tower_service::Service;
/// Buffer requests with a bounded buffer
pub struct BufferLayer<E = DefaultExecutor> {
bound: usize,
executor: E,
}
impl BufferLayer<DefaultExecutor> {
pub fn new(bound: usize) -> Self {
BufferLayer {
bound,
executor: DefaultExecutor::current(),
}
}
}
impl<E> BufferLayer<E> {
pub fn with_executor<S, Request>(bound: usize, executor: E) -> Self
where
S: Service<Request>,
S::Error: Into<Error>,
E: WorkerExecutor<S, Request> + Clone,
{
BufferLayer { bound, executor }
}
}
impl<E, S, Request> Layer<S, Request> for BufferLayer<E>
where
S: Service<Request>,
S::Error: Into<Error>,
E: WorkerExecutor<S, Request> + Clone,
{
type Response = S::Response;
type Error = Error;
type LayerError = Error;
type Service = Buffer<S, Request>;
fn layer(&self, service: S) -> Result<Self::Service, Self::LayerError> {
Buffer::with_executor(service, self.bound, &mut self.executor.clone())
}
}

View File

@ -10,163 +10,11 @@
pub mod error;
pub mod future;
mod layer;
mod message;
mod service;
mod worker;
pub use crate::layer::BufferLayer;
pub use crate::service::Buffer;
pub use crate::worker::WorkerExecutor;
use crate::{error::Error, future::ResponseFuture, message::Message, worker::Worker};
use futures::Poll;
use tokio_executor::DefaultExecutor;
use tokio_sync::{mpsc, oneshot};
use tower_layer::Layer;
use tower_service::Service;
/// Adds a buffer in front of an inner service.
///
/// See crate level documentation for more details.
pub struct Buffer<T, Request>
where
T: Service<Request>,
{
tx: mpsc::Sender<Message<Request, T::Future>>,
worker: worker::Handle,
}
/// Buffer requests with a bounded buffer
pub struct BufferLayer<E = DefaultExecutor> {
bound: usize,
executor: E,
}
impl BufferLayer<DefaultExecutor> {
pub fn new(bound: usize) -> Self {
BufferLayer {
bound,
executor: DefaultExecutor::current(),
}
}
}
impl<E> BufferLayer<E> {
pub fn with_executor<S, Request>(bound: usize, executor: E) -> Self
where
S: Service<Request>,
S::Error: Into<Error>,
E: WorkerExecutor<S, Request> + Clone,
{
BufferLayer { bound, executor }
}
}
impl<E, S, Request> Layer<S, Request> for BufferLayer<E>
where
S: Service<Request>,
S::Error: Into<Error>,
E: WorkerExecutor<S, Request> + Clone,
{
type Response = S::Response;
type Error = Error;
type LayerError = Error;
type Service = Buffer<S, Request>;
fn layer(&self, service: S) -> Result<Self::Service, Self::LayerError> {
Buffer::with_executor(service, self.bound, &mut self.executor.clone())
}
}
impl<T, Request> Buffer<T, Request>
where
T: Service<Request>,
T::Error: Into<Error>,
{
/// Creates a new `Buffer` wrapping `service`.
///
/// `bound` gives the maximal number of requests that can be queued for the service before
/// backpressure is applied to callers.
///
/// The default Tokio executor is used to run the given service, which means that this method
/// must be called while on the Tokio runtime.
pub fn new(service: T, bound: usize) -> Result<Self, Error>
where
T: Send + 'static,
T::Future: Send,
T::Error: Send + Sync,
Request: Send + 'static,
{
Self::with_executor(service, bound, &mut DefaultExecutor::current())
}
/// Creates a new `Buffer` wrapping `service`.
///
/// `executor` is used to spawn a new `Worker` task that is dedicated to
/// draining the buffer and dispatching the requests to the internal
/// service.
///
/// `bound` gives the maximal number of requests that can be queued for the service before
/// backpressure is applied to callers.
pub fn with_executor<E>(service: T, bound: usize, executor: &mut E) -> Result<Self, Error>
where
E: WorkerExecutor<T, Request>,
{
let (tx, rx) = mpsc::channel(bound);
Worker::spawn(service, rx, executor).map(|worker| Buffer { tx, worker })
}
}
impl<T, Request> Service<Request> for Buffer<T, Request>
where
T: Service<Request>,
T::Error: Into<Error>,
{
type Response = T::Response;
type Error = Error;
type Future = ResponseFuture<T::Future>;
fn poll_ready(&mut self) -> Poll<(), Self::Error> {
// If the inner service has errored, then we error here.
self.tx
.poll_ready()
.map_err(|_| self.worker.get_error_on_closed())
}
fn call(&mut self, request: Request) -> Self::Future {
// TODO:
// ideally we'd poll_ready again here so we don't allocate the oneshot
// if the try_send is about to fail, but sadly we can't call poll_ready
// outside of task context.
let (tx, rx) = oneshot::channel();
match self.tx.try_send(Message { request, tx }) {
Err(e) => {
if e.is_closed() {
ResponseFuture::failed(self.worker.get_error_on_closed())
} else {
// When `mpsc::Sender::poll_ready` returns `Ready`, a slot
// in the channel is reserved for the handle. Other `Sender`
// handles may not send a message using that slot. This
// guarantees capacity for `request`.
//
// Given this, the only way to hit this code path is if
// `poll_ready` has not been called & `Ready` returned.
panic!("buffer full; poll_ready must be called first");
}
}
Ok(_) => ResponseFuture::new(rx),
}
}
}
impl<T, Request> Clone for Buffer<T, Request>
where
T: Service<Request>,
{
fn clone(&self) -> Self {
Self {
tx: self.tx.clone(),
worker: self.worker.clone(),
}
}
}

117
tower-buffer/src/service.rs Normal file
View File

@ -0,0 +1,117 @@
use crate::{
error::Error,
future::ResponseFuture,
message::Message,
worker::{Handle, Worker, WorkerExecutor},
};
use futures::Poll;
use tokio_executor::DefaultExecutor;
use tokio_sync::{mpsc, oneshot};
use tower_service::Service;
/// Adds a buffer in front of an inner service.
///
/// See crate level documentation for more details.
pub struct Buffer<T, Request>
where
T: Service<Request>,
{
tx: mpsc::Sender<Message<Request, T::Future>>,
worker: Handle,
}
impl<T, Request> Buffer<T, Request>
where
T: Service<Request>,
T::Error: Into<Error>,
{
/// Creates a new `Buffer` wrapping `service`.
///
/// `bound` gives the maximal number of requests that can be queued for the service before
/// backpressure is applied to callers.
///
/// The default Tokio executor is used to run the given service, which means that this method
/// must be called while on the Tokio runtime.
pub fn new(service: T, bound: usize) -> Result<Self, Error>
where
T: Send + 'static,
T::Future: Send,
T::Error: Send + Sync,
Request: Send + 'static,
{
Self::with_executor(service, bound, &mut DefaultExecutor::current())
}
/// Creates a new `Buffer` wrapping `service`.
///
/// `executor` is used to spawn a new `Worker` task that is dedicated to
/// draining the buffer and dispatching the requests to the internal
/// service.
///
/// `bound` gives the maximal number of requests that can be queued for the service before
/// backpressure is applied to callers.
pub fn with_executor<E>(service: T, bound: usize, executor: &mut E) -> Result<Self, Error>
where
E: WorkerExecutor<T, Request>,
{
let (tx, rx) = mpsc::channel(bound);
Worker::spawn(service, rx, executor).map(|worker| Buffer { tx, worker })
}
}
impl<T, Request> Service<Request> for Buffer<T, Request>
where
T: Service<Request>,
T::Error: Into<Error>,
{
type Response = T::Response;
type Error = Error;
type Future = ResponseFuture<T::Future>;
fn poll_ready(&mut self) -> Poll<(), Self::Error> {
// If the inner service has errored, then we error here.
self.tx
.poll_ready()
.map_err(|_| self.worker.get_error_on_closed())
}
fn call(&mut self, request: Request) -> Self::Future {
// TODO:
// ideally we'd poll_ready again here so we don't allocate the oneshot
// if the try_send is about to fail, but sadly we can't call poll_ready
// outside of task context.
let (tx, rx) = oneshot::channel();
match self.tx.try_send(Message { request, tx }) {
Err(e) => {
if e.is_closed() {
ResponseFuture::failed(self.worker.get_error_on_closed())
} else {
// When `mpsc::Sender::poll_ready` returns `Ready`, a slot
// in the channel is reserved for the handle. Other `Sender`
// handles may not send a message using that slot. This
// guarantees capacity for `request`.
//
// Given this, the only way to hit this code path is if
// `poll_ready` has not been called & `Ready` returned.
panic!("buffer full; poll_ready must be called first");
}
}
Ok(_) => ResponseFuture::new(rx),
}
}
}
impl<T, Request> Clone for Buffer<T, Request>
where
T: Service<Request>,
{
fn clone(&self) -> Self {
Self {
tx: self.tx.clone(),
worker: self.worker.clone(),
}
}
}