refactor(hermes): removed `impl Store`

This commit is contained in:
Reisen 2023-09-13 13:32:27 +00:00 committed by Ali Behjati
parent f2ca88176a
commit e144dd2bd0
11 changed files with 336 additions and 331 deletions

View File

@ -65,14 +65,13 @@ pub async fn get_price_feed(
) -> Result<Json<RpcPriceFeed>, RestError> {
let price_id: PriceIdentifier = params.id.into();
let price_feeds_with_update_data = state
.store
.get_price_feeds_with_update_data(
vec![price_id],
RequestTime::FirstAfter(params.publish_time),
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let price_feeds_with_update_data = crate::store::get_price_feeds_with_update_data(
&state.store,
vec![price_id],
RequestTime::FirstAfter(params.publish_time),
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let mut price_feed = price_feeds_with_update_data
.price_feeds

View File

@ -5,9 +5,12 @@ use {
types::PriceIdInput,
},
doc_examples,
store::types::{
RequestTime,
UnixTimestamp,
store::{
self,
types::{
RequestTime,
UnixTimestamp,
},
},
},
anyhow::Result,
@ -72,14 +75,13 @@ pub async fn get_vaa(
) -> Result<Json<GetVaaResponse>, RestError> {
let price_id: PriceIdentifier = params.id.into();
let price_feeds_with_update_data = state
.store
.get_price_feeds_with_update_data(
vec![price_id],
RequestTime::FirstAfter(params.publish_time),
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let price_feeds_with_update_data = store::get_price_feeds_with_update_data(
&state.store,
vec![price_id],
RequestTime::FirstAfter(params.publish_time),
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let vaa = price_feeds_with_update_data
.update_data

View File

@ -68,11 +68,13 @@ pub async fn get_vaa_ccip(
.map_err(|_| RestError::InvalidCCIPInput)?,
);
let price_feeds_with_update_data = state
.store
.get_price_feeds_with_update_data(vec![price_id], RequestTime::FirstAfter(publish_time))
.await
.map_err(|_| RestError::CcipUpdateDataNotFound)?;
let price_feeds_with_update_data = crate::store::get_price_feeds_with_update_data(
&state.store,
vec![price_id],
RequestTime::FirstAfter(publish_time),
)
.await
.map_err(|_| RestError::CcipUpdateDataNotFound)?;
let bytes = price_feeds_with_update_data
.update_data

View File

@ -63,11 +63,13 @@ pub async fn latest_price_feeds(
QsQuery(params): QsQuery<LatestPriceFeedsQueryParams>,
) -> Result<Json<Vec<RpcPriceFeed>>, RestError> {
let price_ids: Vec<PriceIdentifier> = params.ids.into_iter().map(|id| id.into()).collect();
let price_feeds_with_update_data = state
.store
.get_price_feeds_with_update_data(price_ids, RequestTime::Latest)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let price_feeds_with_update_data = crate::store::get_price_feeds_with_update_data(
&state.store,
price_ids,
RequestTime::Latest,
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
Ok(Json(
price_feeds_with_update_data

View File

@ -58,11 +58,13 @@ pub async fn latest_vaas(
QsQuery(params): QsQuery<LatestVaasQueryParams>,
) -> Result<Json<Vec<String>>, RestError> {
let price_ids: Vec<PriceIdentifier> = params.ids.into_iter().map(|id| id.into()).collect();
let price_feeds_with_update_data = state
.store
.get_price_feeds_with_update_data(price_ids, RequestTime::Latest)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
let price_feeds_with_update_data = crate::store::get_price_feeds_with_update_data(
&state.store,
price_ids,
RequestTime::Latest,
)
.await
.map_err(|_| RestError::UpdateDataNotFound)?;
Ok(Json(
price_feeds_with_update_data

View File

@ -24,9 +24,7 @@ use {
pub async fn price_feed_ids(
State(state): State<crate::api::State>,
) -> Result<Json<Vec<RpcPriceIdentifier>>, RestError> {
let price_feed_ids = state
.store
.get_price_feed_ids()
let price_feed_ids = crate::store::get_price_feed_ids(&state.store)
.await
.iter()
.map(RpcPriceIdentifier::from)

View File

@ -8,7 +8,7 @@ use axum::{
};
pub async fn ready(State(state): State<crate::api::State>) -> Response {
match state.store.is_ready().await {
match crate::store::is_ready(&state.store).await {
true => (StatusCode::OK, "OK").into_response(),
false => (StatusCode::SERVICE_UNAVAILABLE, "Service Unavailable").into_response(),
}

View File

@ -149,11 +149,13 @@ impl Subscriber {
async fn handle_price_feeds_update(&mut self) -> Result<()> {
let price_feed_ids = self.price_feeds_with_config.keys().cloned().collect();
for update in self
.store
.get_price_feeds_with_update_data(price_feed_ids, RequestTime::Latest)
.await?
.price_feeds
for update in crate::store::get_price_feeds_with_update_data(
&self.store,
price_feed_ids,
RequestTime::Latest,
)
.await?
.price_feeds
{
let config = self
.price_feeds_with_config
@ -231,7 +233,7 @@ impl Subscriber {
binary,
}) => {
let price_ids: Vec<PriceIdentifier> = ids.into_iter().map(|id| id.into()).collect();
let available_price_ids = self.store.get_price_feed_ids().await;
let available_price_ids = crate::store::get_price_feed_ids(&self.store).await;
let not_found_price_ids: Vec<&PriceIdentifier> = price_ids
.iter()

View File

@ -213,7 +213,7 @@ pub async fn spawn(opts: RunOptions, store: Arc<Store>) -> Result<()> {
let store = store.clone();
tokio::spawn(async move {
if let Err(e) = store.store_update(Update::Vaa(vaa)).await {
if let Err(e) = crate::store::store_update(&store, Update::Vaa(vaa)).await {
tracing::error!(error = ?e, "Failed to process VAA.");
}
});

View File

@ -175,9 +175,11 @@ pub async fn run(store: Arc<Store>, pythnet_ws_endpoint: String) -> Result<()> {
if candidate.to_string() == update.value.pubkey {
let store = store.clone();
tokio::spawn(async move {
if let Err(err) = store
.store_update(Update::AccumulatorMessages(accumulator_messages))
.await
if let Err(err) = crate::store::store_update(
&store,
Update::AccumulatorMessages(accumulator_messages),
)
.await
{
tracing::error!(error = ?err, "Failed to store accumulator messages.");
}
@ -228,9 +230,7 @@ async fn fetch_existing_guardian_sets(
"Retrieved Current GuardianSet.",
);
store
.update_guardian_set(bridge.guardian_set_index, current)
.await;
crate::store::update_guardian_set(&store, bridge.guardian_set_index, current).await;
// If there are more than one guardian set, we want to fetch the previous one as well as it
// may still be in transition phase if a guardian upgrade has just occurred.
@ -248,9 +248,7 @@ async fn fetch_existing_guardian_sets(
"Retrieved Previous GuardianSet.",
);
store
.update_guardian_set(bridge.guardian_set_index - 1, previous)
.await;
crate::store::update_guardian_set(&store, bridge.guardian_set_index - 1, previous).await;
}
Ok(())

View File

@ -91,20 +91,17 @@ const OBSERVED_CACHE_SIZE: usize = 1000;
const READINESS_STALENESS_THRESHOLD: Duration = Duration::from_secs(30);
pub struct Store {
/// Storage is a short-lived cache of the state of all the updates
/// that have been passed to the store.
/// Storage is a short-lived cache of the state of all the updates that have been passed to the
/// store.
pub storage: Storage,
/// Sequence numbers of lately observed Vaas. Store uses this set
/// to ignore the previously observed Vaas as a performance boost.
pub observed_vaa_seqs: RwLock<BTreeSet<u64>>,
/// Wormhole guardian sets. It is used to verify Vaas before using
/// them.
/// Wormhole guardian sets. It is used to verify Vaas before using them.
pub guardian_set: RwLock<BTreeMap<u32, GuardianSet>>,
/// The sender to the channel between Store and Api to notify
/// completed updates.
/// The sender to the channel between Store and Api to notify completed updates.
pub update_tx: Sender<()>,
/// Time of the last completed update. This is used for the health
/// probes.
/// Time of the last completed update. This is used for the health probes.
pub last_completed_update_at: RwLock<Option<Instant>>,
/// Benchmarks endpoint
pub benchmarks_endpoint: Option<Url>,
@ -125,243 +122,246 @@ impl Store {
benchmarks_endpoint,
})
}
}
/// Stores the update data in the store
#[tracing::instrument(skip(self, update))]
pub async fn store_update(&self, update: Update) -> Result<()> {
// The slot that the update is originating from. It should be available
// in all the updates.
let slot = match update {
Update::Vaa(update_vaa) => {
// FIXME: Move to wormhole.rs
let vaa =
serde_wormhole::from_slice::<Vaa<&serde_wormhole::RawMessage>>(&update_vaa)?;
/// Stores the update data in the store
#[tracing::instrument(skip(store, update))]
pub async fn store_update(store: &Store, update: Update) -> Result<()> {
// The slot that the update is originating from. It should be available
// in all the updates.
let slot = match update {
Update::Vaa(update_vaa) => {
// FIXME: Move to wormhole.rs
let vaa = serde_wormhole::from_slice::<Vaa<&serde_wormhole::RawMessage>>(&update_vaa)?;
if self.observed_vaa_seqs.read().await.contains(&vaa.sequence) {
return Ok(()); // Ignore VAA if we have already seen it
}
let vaa = verify_vaa(self, vaa).await;
let vaa = match vaa {
Ok(vaa) => vaa,
Err(err) => {
tracing::warn!(error = ?err, "Ignoring invalid VAA.");
return Ok(());
}
};
{
let mut observed_vaa_seqs = self.observed_vaa_seqs.write().await;
if observed_vaa_seqs.contains(&vaa.sequence) {
return Ok(()); // Ignore VAA if we have already seen it
}
observed_vaa_seqs.insert(vaa.sequence);
while observed_vaa_seqs.len() > OBSERVED_CACHE_SIZE {
observed_vaa_seqs.pop_first();
}
}
match WormholeMessage::try_from_bytes(vaa.payload)?.payload {
WormholePayload::Merkle(proof) => {
tracing::info!(slot = proof.slot, "Storing VAA Merkle Proof.");
store_wormhole_merkle_verified_message(
self,
proof.clone(),
update_vaa.to_owned(),
)
.await?;
proof.slot
}
}
if store.observed_vaa_seqs.read().await.contains(&vaa.sequence) {
return Ok(()); // Ignore VAA if we have already seen it
}
Update::AccumulatorMessages(accumulator_messages) => {
let slot = accumulator_messages.slot;
tracing::info!(slot = slot, "Storing Accumulator Messages.");
let vaa = verify_vaa(store, vaa).await;
self.storage
.store_accumulator_messages(accumulator_messages)
.await?;
slot
}
};
let accumulator_messages = self.storage.fetch_accumulator_messages(slot).await?;
let wormhole_merkle_state = self.storage.fetch_wormhole_merkle_state(slot).await?;
let (accumulator_messages, wormhole_merkle_state) =
match (accumulator_messages, wormhole_merkle_state) {
(Some(accumulator_messages), Some(wormhole_merkle_state)) => {
(accumulator_messages, wormhole_merkle_state)
let vaa = match vaa {
Ok(vaa) => vaa,
Err(err) => {
tracing::warn!(error = ?err, "Ignoring invalid VAA.");
return Ok(());
}
_ => return Ok(()),
};
tracing::info!(slot = wormhole_merkle_state.root.slot, "Completed Update.");
// Once the accumulator reaches a complete state for a specific slot
// we can build the message states
self.build_message_states(accumulator_messages, wormhole_merkle_state)
.await?;
self.update_tx.send(()).await?;
self.last_completed_update_at
.write()
.await
.replace(Instant::now());
Ok(())
}
#[tracing::instrument(skip(self, accumulator_messages, wormhole_merkle_state))]
async fn build_message_states(
&self,
accumulator_messages: AccumulatorMessages,
wormhole_merkle_state: WormholeMerkleState,
) -> Result<()> {
let wormhole_merkle_message_states_proofs =
construct_message_states_proofs(&accumulator_messages, &wormhole_merkle_state)?;
let current_time: UnixTimestamp =
SystemTime::now().duration_since(UNIX_EPOCH)?.as_secs() as _;
let message_states = accumulator_messages
.raw_messages
.into_iter()
.enumerate()
.map(|(idx, raw_message)| {
Ok(MessageState::new(
from_slice::<BigEndian, _>(raw_message.as_ref())
.map_err(|e| anyhow!("Failed to deserialize message: {:?}", e))?,
raw_message,
ProofSet {
wormhole_merkle_proof: wormhole_merkle_message_states_proofs
.get(idx)
.ok_or(anyhow!("Missing proof for message"))?
.clone(),
},
accumulator_messages.slot,
current_time,
))
})
.collect::<Result<Vec<_>>>()?;
tracing::info!(len = message_states.len(), "Storing Message States.");
self.storage.store_message_states(message_states).await?;
Ok(())
}
pub async fn update_guardian_set(&self, id: u32, guardian_set: GuardianSet) {
let mut guardian_sets = self.guardian_set.write().await;
guardian_sets.insert(id, guardian_set);
}
async fn get_price_feeds_with_update_data_from_storage(
&self,
price_ids: Vec<PriceIdentifier>,
request_time: RequestTime,
) -> Result<PriceFeedsWithUpdateData> {
let messages = self
.storage
.fetch_message_states(
price_ids
.iter()
.map(|price_id| price_id.to_bytes())
.collect(),
request_time,
MessageStateFilter::Only(MessageType::PriceFeedMessage),
)
.await?;
let price_feeds = messages
.iter()
.map(|message_state| match message_state.message {
Message::PriceFeedMessage(price_feed) => Ok(PriceFeedUpdate {
price_feed: PriceFeed::new(
PriceIdentifier::new(price_feed.feed_id),
Price {
price: price_feed.price,
conf: price_feed.conf,
expo: price_feed.exponent,
publish_time: price_feed.publish_time,
},
Price {
price: price_feed.ema_price,
conf: price_feed.ema_conf,
expo: price_feed.exponent,
publish_time: price_feed.publish_time,
},
),
received_at: Some(message_state.received_at),
slot: Some(message_state.slot),
update_data: Some(
construct_update_data(vec![message_state.clone().into()])?
.into_iter()
.next()
.ok_or(anyhow!("Missing update data for message"))?,
),
}),
_ => Err(anyhow!("Invalid message state type")),
})
.collect::<Result<Vec<_>>>()?;
let update_data = construct_update_data(messages.into_iter().map(|m| m.into()).collect())?;
Ok(PriceFeedsWithUpdateData {
price_feeds,
update_data,
})
}
pub async fn get_price_feeds_with_update_data(
&self,
price_ids: Vec<PriceIdentifier>,
request_time: RequestTime,
) -> Result<PriceFeedsWithUpdateData> {
match self
.get_price_feeds_with_update_data_from_storage(price_ids.clone(), request_time.clone())
.await
{
Ok(price_feeds_with_update_data) => Ok(price_feeds_with_update_data),
Err(e) => {
if let RequestTime::FirstAfter(publish_time) = request_time {
if let Some(endpoint) = &self.benchmarks_endpoint {
return benchmarks::get_price_feeds_with_update_data_from_benchmarks(
endpoint.clone(),
price_ids,
publish_time,
)
.await;
}
{
let mut observed_vaa_seqs = store.observed_vaa_seqs.write().await;
if observed_vaa_seqs.contains(&vaa.sequence) {
return Ok(()); // Ignore VAA if we have already seen it
}
observed_vaa_seqs.insert(vaa.sequence);
while observed_vaa_seqs.len() > OBSERVED_CACHE_SIZE {
observed_vaa_seqs.pop_first();
}
}
match WormholeMessage::try_from_bytes(vaa.payload)?.payload {
WormholePayload::Merkle(proof) => {
tracing::info!(slot = proof.slot, "Storing VAA Merkle Proof.");
store_wormhole_merkle_verified_message(
store,
proof.clone(),
update_vaa.to_owned(),
)
.await?;
proof.slot
}
Err(e)
}
}
}
pub async fn get_price_feed_ids(&self) -> HashSet<PriceIdentifier> {
self.storage
.message_state_keys()
.await
.iter()
.map(|key| PriceIdentifier::new(key.feed_id))
.collect()
}
Update::AccumulatorMessages(accumulator_messages) => {
let slot = accumulator_messages.slot;
tracing::info!(slot = slot, "Storing Accumulator Messages.");
pub async fn is_ready(&self) -> bool {
let last_completed_update_at = self.last_completed_update_at.read().await;
match last_completed_update_at.as_ref() {
Some(last_completed_update_at) => {
last_completed_update_at.elapsed() < READINESS_STALENESS_THRESHOLD
}
None => false,
store
.storage
.store_accumulator_messages(accumulator_messages)
.await?;
slot
}
};
let accumulator_messages = store.storage.fetch_accumulator_messages(slot).await?;
let wormhole_merkle_state = store.storage.fetch_wormhole_merkle_state(slot).await?;
let (accumulator_messages, wormhole_merkle_state) =
match (accumulator_messages, wormhole_merkle_state) {
(Some(accumulator_messages), Some(wormhole_merkle_state)) => {
(accumulator_messages, wormhole_merkle_state)
}
_ => return Ok(()),
};
tracing::info!(slot = wormhole_merkle_state.root.slot, "Completed Update.");
// Once the accumulator reaches a complete state for a specific slot
// we can build the message states
build_message_states(store, accumulator_messages, wormhole_merkle_state).await?;
store.update_tx.send(()).await?;
store
.last_completed_update_at
.write()
.await
.replace(Instant::now());
Ok(())
}
#[tracing::instrument(skip(store, accumulator_messages, wormhole_merkle_state))]
async fn build_message_states(
store: &Store,
accumulator_messages: AccumulatorMessages,
wormhole_merkle_state: WormholeMerkleState,
) -> Result<()> {
let wormhole_merkle_message_states_proofs =
construct_message_states_proofs(&accumulator_messages, &wormhole_merkle_state)?;
let current_time: UnixTimestamp = SystemTime::now().duration_since(UNIX_EPOCH)?.as_secs() as _;
let message_states = accumulator_messages
.raw_messages
.into_iter()
.enumerate()
.map(|(idx, raw_message)| {
Ok(MessageState::new(
from_slice::<BigEndian, _>(raw_message.as_ref())
.map_err(|e| anyhow!("Failed to deserialize message: {:?}", e))?,
raw_message,
ProofSet {
wormhole_merkle_proof: wormhole_merkle_message_states_proofs
.get(idx)
.ok_or(anyhow!("Missing proof for message"))?
.clone(),
},
accumulator_messages.slot,
current_time,
))
})
.collect::<Result<Vec<_>>>()?;
tracing::info!(len = message_states.len(), "Storing Message States.");
store.storage.store_message_states(message_states).await?;
Ok(())
}
pub async fn update_guardian_set(store: &Store, id: u32, guardian_set: GuardianSet) {
let mut guardian_sets = store.guardian_set.write().await;
guardian_sets.insert(id, guardian_set);
}
async fn get_price_feeds_with_update_data_from_storage(
store: &Store,
price_ids: Vec<PriceIdentifier>,
request_time: RequestTime,
) -> Result<PriceFeedsWithUpdateData> {
let messages = store
.storage
.fetch_message_states(
price_ids
.iter()
.map(|price_id| price_id.to_bytes())
.collect(),
request_time,
MessageStateFilter::Only(MessageType::PriceFeedMessage),
)
.await?;
let price_feeds = messages
.iter()
.map(|message_state| match message_state.message {
Message::PriceFeedMessage(price_feed) => Ok(PriceFeedUpdate {
price_feed: PriceFeed::new(
PriceIdentifier::new(price_feed.feed_id),
Price {
price: price_feed.price,
conf: price_feed.conf,
expo: price_feed.exponent,
publish_time: price_feed.publish_time,
},
Price {
price: price_feed.ema_price,
conf: price_feed.ema_conf,
expo: price_feed.exponent,
publish_time: price_feed.publish_time,
},
),
received_at: Some(message_state.received_at),
slot: Some(message_state.slot),
update_data: Some(
construct_update_data(vec![message_state.clone().into()])?
.into_iter()
.next()
.ok_or(anyhow!("Missing update data for message"))?,
),
}),
_ => Err(anyhow!("Invalid message state type")),
})
.collect::<Result<Vec<_>>>()?;
let update_data = construct_update_data(messages.into_iter().map(|m| m.into()).collect())?;
Ok(PriceFeedsWithUpdateData {
price_feeds,
update_data,
})
}
pub async fn get_price_feeds_with_update_data(
store: &Store,
price_ids: Vec<PriceIdentifier>,
request_time: RequestTime,
) -> Result<PriceFeedsWithUpdateData> {
match get_price_feeds_with_update_data_from_storage(
store,
price_ids.clone(),
request_time.clone(),
)
.await
{
Ok(price_feeds_with_update_data) => Ok(price_feeds_with_update_data),
Err(e) => {
if let RequestTime::FirstAfter(publish_time) = request_time {
if let Some(endpoint) = &store.benchmarks_endpoint {
return benchmarks::get_price_feeds_with_update_data_from_benchmarks(
endpoint.clone(),
price_ids,
publish_time,
)
.await;
}
}
Err(e)
}
}
}
pub async fn get_price_feed_ids(store: &Store) -> HashSet<PriceIdentifier> {
store
.storage
.message_state_keys()
.await
.iter()
.map(|key| PriceIdentifier::new(key.feed_id))
.collect()
}
pub async fn is_ready(store: &Store) -> bool {
let last_completed_update_at = store.last_completed_update_at.read().await;
match last_completed_update_at.as_ref() {
Some(last_completed_update_at) => {
last_completed_update_at.elapsed() < READINESS_STALENESS_THRESHOLD
}
None => false,
}
}
@ -478,20 +478,20 @@ mod test {
let store = Store::new(update_tx, cache_size, None);
// Add an initial guardian set with public key 0
store
.update_guardian_set(
0,
GuardianSet {
keys: vec![[0; 20]],
},
)
.await;
update_guardian_set(
&store,
0,
GuardianSet {
keys: vec![[0; 20]],
},
)
.await;
(store, update_rx)
}
pub async fn store_multiple_concurrent_valid_updates(store: Arc<Store>, updates: Vec<Update>) {
let res = join_all(updates.into_iter().map(|u| store.store_update(u))).await;
let res = join_all(updates.into_iter().map(|u| store_update(&store, u))).await;
// Check that all store_update calls succeeded
assert!(res.into_iter().all(|r| r.is_ok()));
}
@ -514,19 +514,19 @@ mod test {
// Check the price ids are stored correctly
assert_eq!(
store.get_price_feed_ids().await,
get_price_feed_ids(&store).await,
vec![PriceIdentifier::new([100; 32])].into_iter().collect()
);
// Check get_price_feeds_with_update_data retrieves the correct
// price feed with correct update data.
let price_feeds_with_update_data = store
.get_price_feeds_with_update_data(
vec![PriceIdentifier::new([100; 32])],
RequestTime::Latest,
)
.await
.unwrap();
let price_feeds_with_update_data = get_price_feeds_with_update_data(
&store,
vec![PriceIdentifier::new([100; 32])],
RequestTime::Latest,
)
.await
.unwrap();
assert_eq!(
price_feeds_with_update_data.price_feeds,
@ -640,13 +640,13 @@ mod test {
MockClock::advance(Duration::from_secs(1));
// Get the price feeds with update data
let price_feeds_with_update_data = store
.get_price_feeds_with_update_data(
vec![PriceIdentifier::new([100; 32])],
RequestTime::Latest,
)
.await
.unwrap();
let price_feeds_with_update_data = get_price_feeds_with_update_data(
&store,
vec![PriceIdentifier::new([100; 32])],
RequestTime::Latest,
)
.await
.unwrap();
// check received_at is correct
assert_eq!(price_feeds_with_update_data.price_feeds.len(), 1);
@ -656,13 +656,13 @@ mod test {
);
// Check the store is ready
assert!(store.is_ready().await);
assert!(is_ready(&store).await);
// Advance the clock to make the prices stale
MockClock::advance_system_time(READINESS_STALENESS_THRESHOLD);
MockClock::advance(READINESS_STALENESS_THRESHOLD);
// Check the store is not ready
assert!(!store.is_ready().await);
assert!(!is_ready(&store).await);
}
/// Test that the store retains the latest slots upon cache eviction.
@ -705,16 +705,16 @@ mod test {
// Check the last 100 slots are retained
for slot in 900..1000 {
let price_feeds_with_update_data = store
.get_price_feeds_with_update_data(
vec![
PriceIdentifier::new([100; 32]),
PriceIdentifier::new([200; 32]),
],
RequestTime::FirstAfter(slot as i64),
)
.await
.unwrap();
let price_feeds_with_update_data = get_price_feeds_with_update_data(
&store,
vec![
PriceIdentifier::new([100; 32]),
PriceIdentifier::new([200; 32]),
],
RequestTime::FirstAfter(slot as i64),
)
.await
.unwrap();
assert_eq!(price_feeds_with_update_data.price_feeds.len(), 2);
assert_eq!(price_feeds_with_update_data.price_feeds[0].slot, Some(slot));
assert_eq!(price_feeds_with_update_data.price_feeds[1].slot, Some(slot));
@ -722,16 +722,16 @@ mod test {
// Check nothing else is retained
for slot in 0..900 {
assert!(store
.get_price_feeds_with_update_data(
vec![
PriceIdentifier::new([100; 32]),
PriceIdentifier::new([200; 32]),
],
RequestTime::FirstAfter(slot as i64),
)
.await
.is_err());
assert!(get_price_feeds_with_update_data(
&store,
vec![
PriceIdentifier::new([100; 32]),
PriceIdentifier::new([200; 32]),
],
RequestTime::FirstAfter(slot as i64),
)
.await
.is_err());
}
}
}