typos, clippy and saturating_add_assign (#30272)

* typo

* remove clippy allow

---------

Co-authored-by: haoran <haoran@mbook>
This commit is contained in:
HaoranYi 2023-02-14 10:03:16 -06:00 committed by GitHub
parent 253517cba3
commit 27f18f96e8
No known key found for this signature in database
GPG Key ID: 4AEE18F83AFDEB23
2 changed files with 5 additions and 6 deletions

View File

@ -1659,7 +1659,7 @@ impl SplitAncientStorages {
}
/// create once ancient slots have been identified
/// This is easier to test, removing SortedStorges as a type to deal with here.
/// This is easier to test, removing SortedStorages as a type to deal with here.
fn new_with_ancient_info(
range: &Range<Slot>,
ancient_slots: Vec<Slot>,
@ -7218,7 +7218,6 @@ impl AccountsDb {
}
}
#[allow(clippy::too_many_arguments)]
fn calculate_accounts_hash_with_verify(
&self,
data_source: CalcAccountsHashDataSource,

View File

@ -102,7 +102,7 @@ impl AncientSlotInfos {
alive_bytes,
should_shrink,
});
self.total_alive_bytes += alive_bytes;
saturating_add_assign!(self.total_alive_bytes, alive_bytes);
}
}
@ -136,10 +136,10 @@ impl AncientSlotInfos {
// shrink enough slots to write 'percent_of_alive_shrunk_data'% of the total alive data
// from slots that exceeded the shrink threshold.
// The goal is to limit overall i/o in this pass while making progress.
let threhold_bytes = self.total_alive_bytes_shrink * percent_of_alive_shrunk_data / 100;
let threshold_bytes = self.total_alive_bytes_shrink * percent_of_alive_shrunk_data / 100;
for info_index in &self.shrink_indexes {
let info = &mut self.all_infos[*info_index];
if bytes_to_shrink_due_to_ratio >= threhold_bytes {
if bytes_to_shrink_due_to_ratio >= threshold_bytes {
// we exceeded the amount to shrink due to alive ratio, so don't shrink this one just due to 'should_shrink'
// It MAY be shrunk based on total capacity still.
// Mark it as false for 'should_shrink' so it gets evaluated solely based on # of files.
@ -681,7 +681,7 @@ pub enum StorageSelector {
/// The accounts may have to be split between 2 storages (primary and overflow) if there is not enough room in the primary storage.
/// The 'store' functions need data stored in a slice of specific type.
/// We need 1-2 of these slices constructed based on available bytes and individual account sizes.
/// The slice arithmetic accross both hashes and account data gets messy. So, this struct abstracts that.
/// The slice arithmetic across both hashes and account data gets messy. So, this struct abstracts that.
pub struct AccountsToStore<'a> {
accounts: &'a [&'a StoredAccountMeta<'a>],
/// if 'accounts' contains more items than can be contained in the primary storage, then we have to split these accounts.