Upgrade to Rust v1.49.0

This commit is contained in:
Michael Vines 2021-01-23 11:55:15 -08:00
parent 7604edb16f
commit cbffab7850
36 changed files with 156 additions and 132 deletions

View File

@ -1,4 +1,4 @@
FROM solanalabs/rust:1.48.0 FROM solanalabs/rust:1.49.0
ARG date ARG date
RUN set -x \ RUN set -x \

View File

@ -1,6 +1,6 @@
# Note: when the rust version is changed also modify # Note: when the rust version is changed also modify
# ci/rust-version.sh to pick up the new image tag # ci/rust-version.sh to pick up the new image tag
FROM rust:1.48.0 FROM rust:1.49.0
# Add Google Protocol Buffers for Libra's metrics library. # Add Google Protocol Buffers for Libra's metrics library.
ENV PROTOC_VERSION 3.8.0 ENV PROTOC_VERSION 3.8.0

View File

@ -18,13 +18,13 @@
if [[ -n $RUST_STABLE_VERSION ]]; then if [[ -n $RUST_STABLE_VERSION ]]; then
stable_version="$RUST_STABLE_VERSION" stable_version="$RUST_STABLE_VERSION"
else else
stable_version=1.48.0 stable_version=1.49.0
fi fi
if [[ -n $RUST_NIGHTLY_VERSION ]]; then if [[ -n $RUST_NIGHTLY_VERSION ]]; then
nightly_version="$RUST_NIGHTLY_VERSION" nightly_version="$RUST_NIGHTLY_VERSION"
else else
nightly_version=2020-12-13 nightly_version=2021-01-23
fi fi

View File

@ -167,12 +167,12 @@ pub fn resolve_signer(
name: &str, name: &str,
wallet_manager: &mut Option<Arc<RemoteWalletManager>>, wallet_manager: &mut Option<Arc<RemoteWalletManager>>,
) -> Result<Option<String>, Box<dyn std::error::Error>> { ) -> Result<Option<String>, Box<dyn std::error::Error>> {
Ok(resolve_signer_from_path( resolve_signer_from_path(
matches, matches,
matches.value_of(name).unwrap(), matches.value_of(name).unwrap(),
name, name,
wallet_manager, wallet_manager,
)?) )
} }
pub fn lamports_of_sol(matches: &ArgMatches<'_>, name: &str) -> Option<u64> { pub fn lamports_of_sol(matches: &ArgMatches<'_>, name: &str) -> Option<u64> {

View File

@ -35,16 +35,16 @@ impl From<TransportError> for ClientErrorKind {
} }
} }
impl Into<TransportError> for ClientErrorKind { impl From<ClientErrorKind> for TransportError {
fn into(self) -> TransportError { fn from(client_error_kind: ClientErrorKind) -> Self {
match self { match client_error_kind {
Self::Io(err) => TransportError::IoError(err), ClientErrorKind::Io(err) => Self::IoError(err),
Self::TransactionError(err) => TransportError::TransactionError(err), ClientErrorKind::TransactionError(err) => Self::TransactionError(err),
Self::Reqwest(err) => TransportError::Custom(format!("{:?}", err)), ClientErrorKind::Reqwest(err) => Self::Custom(format!("{:?}", err)),
Self::RpcError(err) => TransportError::Custom(format!("{:?}", err)), ClientErrorKind::RpcError(err) => Self::Custom(format!("{:?}", err)),
Self::SerdeJson(err) => TransportError::Custom(format!("{:?}", err)), ClientErrorKind::SerdeJson(err) => Self::Custom(format!("{:?}", err)),
Self::SigningError(err) => TransportError::Custom(format!("{:?}", err)), ClientErrorKind::SigningError(err) => Self::Custom(format!("{:?}", err)),
Self::Custom(err) => TransportError::Custom(format!("{:?}", err)), ClientErrorKind::Custom(err) => Self::Custom(format!("{:?}", err)),
} }
} }
} }
@ -100,9 +100,9 @@ impl From<TransportError> for ClientError {
} }
} }
impl Into<TransportError> for ClientError { impl From<ClientError> for TransportError {
fn into(self) -> TransportError { fn from(client_error: ClientError) -> Self {
self.kind.into() client_error.kind.into()
} }
} }

View File

@ -347,6 +347,7 @@ fn update_peer_stats(
) { ) {
let now = timestamp(); let now = timestamp();
let last = last_datapoint_submit.load(Ordering::Relaxed); let last = last_datapoint_submit.load(Ordering::Relaxed);
#[allow(deprecated)]
if now.saturating_sub(last) > 1000 if now.saturating_sub(last) > 1000
&& last_datapoint_submit.compare_and_swap(last, now, Ordering::Relaxed) == last && last_datapoint_submit.compare_and_swap(last, now, Ordering::Relaxed) == last
{ {

View File

@ -329,6 +329,7 @@ impl StandardBroadcastRun {
let mut get_peers_time = Measure::start("broadcast::get_peers"); let mut get_peers_time = Measure::start("broadcast::get_peers");
let now = timestamp(); let now = timestamp();
let last = self.last_peer_update.load(Ordering::Relaxed); let last = self.last_peer_update.load(Ordering::Relaxed);
#[allow(deprecated)]
if now - last > BROADCAST_PEER_UPDATE_INTERVAL_MS if now - last > BROADCAST_PEER_UPDATE_INTERVAL_MS
&& self && self
.last_peer_update .last_peer_update

View File

@ -403,6 +403,7 @@ impl ClusterInfoVoteListener {
let last_version = bank.last_vote_sync.load(Ordering::Relaxed); let last_version = bank.last_vote_sync.load(Ordering::Relaxed);
let (new_version, msgs) = verified_vote_packets.get_latest_votes(last_version); let (new_version, msgs) = verified_vote_packets.get_latest_votes(last_version);
verified_packets_sender.send(msgs)?; verified_packets_sender.send(msgs)?;
#[allow(deprecated)]
bank.last_vote_sync.compare_and_swap( bank.last_vote_sync.compare_and_swap(
last_version, last_version,
new_version, new_version,

View File

@ -263,15 +263,9 @@ mod tests {
#[test] #[test]
fn test_get_highest_confirmed_root() { fn test_get_highest_confirmed_root() {
assert_eq!(get_highest_confirmed_root(vec![], 10), 0); assert_eq!(get_highest_confirmed_root(vec![], 10), 0);
let mut rooted_stake = vec![]; let rooted_stake = vec![(0, 5), (1, 5)];
rooted_stake.push((0, 5));
rooted_stake.push((1, 5));
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 0); assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 0);
let mut rooted_stake = vec![]; let rooted_stake = vec![(1, 5), (0, 10), (2, 5), (1, 4)];
rooted_stake.push((1, 5));
rooted_stake.push((0, 10));
rooted_stake.push((2, 5));
rooted_stake.push((1, 4));
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 1); assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 1);
} }

View File

@ -145,10 +145,10 @@ impl CrdsFilterSet {
} }
} }
impl Into<Vec<CrdsFilter>> for CrdsFilterSet { impl From<CrdsFilterSet> for Vec<CrdsFilter> {
fn into(self) -> Vec<CrdsFilter> { fn from(cfs: CrdsFilterSet) -> Self {
let mask_bits = self.mask_bits; let mask_bits = cfs.mask_bits;
self.filters cfs.filters
.into_iter() .into_iter()
.enumerate() .enumerate()
.map(|(seed, filter)| CrdsFilter { .map(|(seed, filter)| CrdsFilter {

View File

@ -303,7 +303,7 @@ impl<'de> Deserialize<'de> for Vote {
from: Pubkey, from: Pubkey,
transaction: Transaction, transaction: Transaction,
wallclock: u64, wallclock: u64,
}; }
let vote = Vote::deserialize(deserializer)?; let vote = Vote::deserialize(deserializer)?;
let vote = match vote.transaction.sanitize() { let vote = match vote.transaction.sanitize() {
Ok(_) => Self::new(vote.from, vote.transaction, vote.wallclock), Ok(_) => Self::new(vote.from, vote.transaction, vote.wallclock),

View File

@ -125,7 +125,7 @@ impl PohService {
// sleep is not accurate enough to get a predictable time. // sleep is not accurate enough to get a predictable time.
// Kernel can not schedule the thread for a while. // Kernel can not schedule the thread for a while.
while (now.elapsed().as_nanos() as u64) < target_tick_ns { while (now.elapsed().as_nanos() as u64) < target_tick_ns {
std::sync::atomic::spin_loop_hint(); std::hint::spin_loop();
} }
total_sleep_us += (now.elapsed().as_nanos() as u64 - elapsed_ns) / 1000; total_sleep_us += (now.elapsed().as_nanos() as u64 - elapsed_ns) / 1000;
now = Instant::now(); now = Instant::now();

View File

@ -121,6 +121,7 @@ fn update_retransmit_stats(
let now = timestamp(); let now = timestamp();
let last = stats.last_ts.load(Ordering::Relaxed); let last = stats.last_ts.load(Ordering::Relaxed);
#[allow(deprecated)]
if now.saturating_sub(last) > 2000 if now.saturating_sub(last) > 2000
&& stats.last_ts.compare_and_swap(last, now, Ordering::Relaxed) == last && stats.last_ts.compare_and_swap(last, now, Ordering::Relaxed) == last
{ {
@ -310,6 +311,7 @@ fn retransmit(
let now = timestamp(); let now = timestamp();
let last = last_peer_update.load(Ordering::Relaxed); let last = last_peer_update.load(Ordering::Relaxed);
#[allow(deprecated)]
if now.saturating_sub(last) > 1000 if now.saturating_sub(last) > 1000
&& last_peer_update.compare_and_swap(last, now, Ordering::Relaxed) == last && last_peer_update.compare_and_swap(last, now, Ordering::Relaxed) == last
{ {

View File

@ -749,7 +749,7 @@ impl JsonRpcRequestProcessor {
// If the starting slot is lower than what's available in blockstore assume the entire // If the starting slot is lower than what's available in blockstore assume the entire
// [start_slot..end_slot] can be fetched from BigTable. // [start_slot..end_slot] can be fetched from BigTable.
if let Some(bigtable_ledger_storage) = &self.bigtable_ledger_storage { if let Some(bigtable_ledger_storage) = &self.bigtable_ledger_storage {
return Ok(self return self
.runtime_handle .runtime_handle
.block_on( .block_on(
bigtable_ledger_storage bigtable_ledger_storage
@ -764,7 +764,7 @@ impl JsonRpcRequestProcessor {
"BigTable query failed (maybe timeout due to too large range?)" "BigTable query failed (maybe timeout due to too large range?)"
.to_string(), .to_string(),
) )
})?); });
} }
} }

View File

@ -119,11 +119,12 @@ impl Tpu {
} }
pub fn join(self) -> thread::Result<()> { pub fn join(self) -> thread::Result<()> {
let mut results = vec![]; let results = vec![
results.push(self.fetch_stage.join()); self.fetch_stage.join(),
results.push(self.sigverify_stage.join()); self.sigverify_stage.join(),
results.push(self.cluster_info_vote_listener.join()); self.cluster_info_vote_listener.join(),
results.push(self.banking_stage.join()); self.banking_stage.join(),
];
let broadcast_result = self.broadcast_stage.join(); let broadcast_result = self.broadcast_stage.join();
for result in results { for result in results {
result?; result?;

View File

@ -71,6 +71,7 @@ fn filter_allow_attrs(attrs: &mut Vec<Attribute>) {
}); });
} }
#[allow(deprecated)]
#[cfg(RUSTC_WITH_SPECIALIZATION)] #[cfg(RUSTC_WITH_SPECIALIZATION)]
fn quote_for_specialization_detection() -> TokenStream2 { fn quote_for_specialization_detection() -> TokenStream2 {
lazy_static! { lazy_static! {

View File

@ -645,7 +645,7 @@ fn hardforks_of(matches: &ArgMatches<'_>, name: &str) -> Option<Vec<Slot>> {
fn load_bank_forks( fn load_bank_forks(
arg_matches: &ArgMatches, arg_matches: &ArgMatches,
ledger_path: &PathBuf, ledger_path: &Path,
genesis_config: &GenesisConfig, genesis_config: &GenesisConfig,
process_options: ProcessOptions, process_options: ProcessOptions,
access_type: AccessType, access_type: AccessType,
@ -653,7 +653,7 @@ fn load_bank_forks(
snapshot_archive_path: Option<PathBuf>, snapshot_archive_path: Option<PathBuf>,
) -> bank_forks_utils::LoadResult { ) -> bank_forks_utils::LoadResult {
let blockstore = open_blockstore(&ledger_path, access_type, wal_recovery_mode); let blockstore = open_blockstore(&ledger_path, access_type, wal_recovery_mode);
let snapshot_path = ledger_path.clone().join(if blockstore.is_primary_access() { let snapshot_path = ledger_path.join(if blockstore.is_primary_access() {
"snapshot" "snapshot"
} else { } else {
"snapshot.ledger-tool" "snapshot.ledger-tool"
@ -662,7 +662,7 @@ fn load_bank_forks(
None None
} else { } else {
let snapshot_package_output_path = let snapshot_package_output_path =
snapshot_archive_path.unwrap_or_else(|| ledger_path.clone()); snapshot_archive_path.unwrap_or_else(|| ledger_path.to_path_buf());
Some(SnapshotConfig { Some(SnapshotConfig {
snapshot_interval_slots: 0, // Value doesn't matter snapshot_interval_slots: 0, // Value doesn't matter
snapshot_package_output_path, snapshot_package_output_path,
@ -1443,7 +1443,7 @@ fn main() {
last_in_slot: bool, last_in_slot: bool,
data_complete: bool, data_complete: bool,
shred: &'a Shred, shred: &'a Shred,
}; }
let starting_slot = value_t_or_exit!(arg_matches, "starting_slot", Slot); let starting_slot = value_t_or_exit!(arg_matches, "starting_slot", Slot);
let ending_slot = value_t!(arg_matches, "ending_slot", Slot).unwrap_or(Slot::MAX); let ending_slot = value_t!(arg_matches, "ending_slot", Slot).unwrap_or(Slot::MAX);
let ledger = open_blockstore(&ledger_path, AccessType::TryPrimaryThenSecondary, None); let ledger = open_blockstore(&ledger_path, AccessType::TryPrimaryThenSecondary, None);
@ -2442,13 +2442,13 @@ fn main() {
cluster_points: String, cluster_points: String,
old_capitalization: u64, old_capitalization: u64,
new_capitalization: u64, new_capitalization: u64,
}; }
fn format_or_na<T: std::fmt::Display>( fn format_or_na<T: std::fmt::Display>(
data: Option<T>, data: Option<T>,
) -> String { ) -> String {
data.map(|data| format!("{}", data)) data.map(|data| format!("{}", data))
.unwrap_or_else(|| "N/A".to_owned()) .unwrap_or_else(|| "N/A".to_owned())
}; }
let mut point_details = detail let mut point_details = detail
.map(|d| d.points.iter().map(Some).collect::<Vec<_>>()) .map(|d| d.points.iter().map(Some).collect::<Vec<_>>())
.unwrap_or_default(); .unwrap_or_default();

View File

@ -185,9 +185,10 @@ impl From<&str> for BlockstoreRecoveryMode {
} }
} }
} }
impl Into<DBRecoveryMode> for BlockstoreRecoveryMode {
fn into(self) -> DBRecoveryMode { impl From<BlockstoreRecoveryMode> for DBRecoveryMode {
match self { fn from(brm: BlockstoreRecoveryMode) -> Self {
match brm {
BlockstoreRecoveryMode::TolerateCorruptedTailRecords => { BlockstoreRecoveryMode::TolerateCorruptedTailRecords => {
DBRecoveryMode::TolerateCorruptedTailRecords DBRecoveryMode::TolerateCorruptedTailRecords
} }
@ -404,6 +405,7 @@ pub trait Column {
fn key(index: Self::Index) -> Vec<u8>; fn key(index: Self::Index) -> Vec<u8>;
fn index(key: &[u8]) -> Self::Index; fn index(key: &[u8]) -> Self::Index;
fn primary_index(index: Self::Index) -> Slot; fn primary_index(index: Self::Index) -> Slot;
#[allow(clippy::wrong_self_convention)]
fn as_index(slot: Slot) -> Self::Index; fn as_index(slot: Slot) -> Self::Index;
} }

View File

@ -164,6 +164,7 @@ impl Counter {
} }
} }
pub fn init(&mut self) { pub fn init(&mut self) {
#![allow(deprecated)]
self.lograte self.lograte
.compare_and_swap(0, Self::default_log_rate(), Ordering::Relaxed); .compare_and_swap(0, Self::default_log_rate(), Ordering::Relaxed);
self.metricsrate self.metricsrate
@ -188,6 +189,7 @@ impl Counter {
} }
let lastlog = self.lastlog.load(Ordering::Relaxed); let lastlog = self.lastlog.load(Ordering::Relaxed);
#[allow(deprecated)]
let prev = self let prev = self
.lastlog .lastlog
.compare_and_swap(lastlog, counts, Ordering::Relaxed); .compare_and_swap(lastlog, counts, Ordering::Relaxed);

View File

@ -19,11 +19,11 @@ use std::{
type CounterMap = HashMap<(&'static str, u64), CounterPoint>; type CounterMap = HashMap<(&'static str, u64), CounterPoint>;
impl Into<DataPoint> for CounterPoint { impl From<CounterPoint> for DataPoint {
fn into(self) -> DataPoint { fn from(counter_point: CounterPoint) -> Self {
let mut point = DataPoint::new(self.name); let mut point = Self::new(counter_point.name);
point.timestamp = self.timestamp; point.timestamp = counter_point.timestamp;
point.add_field_i64("count", self.count); point.add_field_i64("count", counter_point.count);
point point
} }
} }

View File

@ -89,18 +89,17 @@ impl<T: Clone + Default + Sized> Default for PinnedVec<T> {
} }
} }
impl<T: Clone + Default + Sized> Into<Vec<T>> for PinnedVec<T> { impl<T: Clone + Default + Sized> From<PinnedVec<T>> for Vec<T> {
fn into(mut self) -> Vec<T> { fn from(mut pinned_vec: PinnedVec<T>) -> Self {
if self.pinned { if pinned_vec.pinned {
unpin(self.x.as_mut_ptr()); unpin(pinned_vec.x.as_mut_ptr());
self.pinned = false; pinned_vec.pinned = false;
} }
self.pinnable = false; pinned_vec.pinnable = false;
self.recycler = None; pinned_vec.recycler = None;
std::mem::take(&mut self.x) std::mem::take(&mut pinned_vec.x)
} }
} }
pub struct PinnedIter<'a, T>(std::slice::Iter<'a, T>); pub struct PinnedIter<'a, T>(std::slice::Iter<'a, T>);
pub struct PinnedIterMut<'a, T>(std::slice::IterMut<'a, T>); pub struct PinnedIterMut<'a, T>(std::slice::IterMut<'a, T>);

View File

@ -114,9 +114,12 @@ impl<T: Default + Reset> RecyclerX<T> {
let max_gc = self.stats.max_gc.load(Ordering::Relaxed); let max_gc = self.stats.max_gc.load(Ordering::Relaxed);
if len > max_gc { if len > max_gc {
// this is not completely accurate, but for most cases should be fine. // this is not completely accurate, but for most cases should be fine.
self.stats let _ = self.stats.max_gc.compare_exchange(
.max_gc max_gc,
.compare_and_swap(max_gc, len, Ordering::Relaxed); len,
Ordering::Relaxed,
Ordering::Relaxed,
);
} }
let total = self.stats.total.load(Ordering::Relaxed); let total = self.stats.total.load(Ordering::Relaxed);
let reuse = self.stats.reuse.load(Ordering::Relaxed); let reuse = self.stats.reuse.load(Ordering::Relaxed);

View File

@ -3908,6 +3908,8 @@ impl AccountsDB {
fn report_store_timings(&self) { fn report_store_timings(&self) {
let last = self.stats.last_store_report.load(Ordering::Relaxed); let last = self.stats.last_store_report.load(Ordering::Relaxed);
let now = solana_sdk::timing::timestamp(); let now = solana_sdk::timing::timestamp();
#[allow(deprecated)]
if now.saturating_sub(last) > 1000 if now.saturating_sub(last) > 1000
&& self && self
.stats .stats

View File

@ -196,14 +196,18 @@ impl<T: BloomHashIndex> AtomicBloom<T> {
} }
} }
impl<T: BloomHashIndex> Into<Bloom<T>> for AtomicBloom<T> { impl<T: BloomHashIndex> From<AtomicBloom<T>> for Bloom<T> {
fn into(self) -> Bloom<T> { fn from(atomic_bloom: AtomicBloom<T>) -> Self {
let bits: Vec<_> = self.bits.into_iter().map(AtomicU64::into_inner).collect(); let bits: Vec<_> = atomic_bloom
.bits
.into_iter()
.map(AtomicU64::into_inner)
.collect();
let num_bits_set = bits.iter().map(|x| x.count_ones() as u64).sum(); let num_bits_set = bits.iter().map(|x| x.count_ones() as u64).sum();
let mut bits: BitVec<u64> = bits.into(); let mut bits: BitVec<u64> = bits.into();
bits.truncate(self.num_bits); bits.truncate(atomic_bloom.num_bits);
Bloom { Bloom {
keys: self.keys, keys: atomic_bloom.keys,
bits, bits,
num_bits_set, num_bits_set,
_phantom: PhantomData::default(), _phantom: PhantomData::default(),

View File

@ -30,9 +30,9 @@ impl LogCollector {
} }
} }
impl Into<Vec<String>> for LogCollector { impl From<LogCollector> for Vec<String> {
fn into(self) -> Vec<String> { fn from(log_collector: LogCollector) -> Self {
self.inner.into_inner().messages log_collector.inner.into_inner().messages
} }
} }

View File

@ -14,7 +14,13 @@ use solana_sdk::{
process_instruction::{InvokeContext, LoaderEntrypoint}, process_instruction::{InvokeContext, LoaderEntrypoint},
pubkey::Pubkey, pubkey::Pubkey,
}; };
use std::{collections::HashMap, env, path::PathBuf, str, sync::RwLock}; use std::{
collections::HashMap,
env,
path::{Path, PathBuf},
str,
sync::RwLock,
};
use thiserror::Error; use thiserror::Error;
#[derive(Error, Debug, Serialize, Clone, PartialEq, FromPrimitive, ToPrimitive)] #[derive(Error, Debug, Serialize, Clone, PartialEq, FromPrimitive, ToPrimitive)]
@ -85,12 +91,12 @@ impl NativeLoader {
} }
#[cfg(windows)] #[cfg(windows)]
fn library_open(path: &PathBuf) -> Result<Library, libloading::Error> { fn library_open(path: &Path) -> Result<Library, libloading::Error> {
Library::new(path) Library::new(path)
} }
#[cfg(not(windows))] #[cfg(not(windows))]
fn library_open(path: &PathBuf) -> Result<Library, libloading::Error> { fn library_open(path: &Path) -> Result<Library, libloading::Error> {
// Linux tls bug can cause crash on dlclose(), workaround by never unloading // Linux tls bug can cause crash on dlclose(), workaround by never unloading
Library::open(Some(path), libc::RTLD_NODELETE | libc::RTLD_NOW) Library::open(Some(path), libc::RTLD_NODELETE | libc::RTLD_NOW)
} }

View File

@ -24,9 +24,9 @@ impl From<&AccountStorageEntry> for SerializableAccountStorageEntry {
} }
} }
impl Into<AccountStorageEntry> for SerializableAccountStorageEntry { impl From<SerializableAccountStorageEntry> for AccountStorageEntry {
fn into(self) -> AccountStorageEntry { fn from(s: SerializableAccountStorageEntry) -> Self {
AccountStorageEntry::new_empty_map(self.id, self.accounts_current_len) AccountStorageEntry::new_empty_map(s.id, s.accounts_current_len)
} }
} }
@ -71,43 +71,44 @@ pub(crate) struct DeserializableVersionedBank {
pub(crate) message_processor: MessageProcessor, pub(crate) message_processor: MessageProcessor,
} }
impl Into<BankFieldsToDeserialize> for DeserializableVersionedBank { impl From<DeserializableVersionedBank> for BankFieldsToDeserialize {
fn into(self) -> BankFieldsToDeserialize { fn from(dvb: DeserializableVersionedBank) -> Self {
BankFieldsToDeserialize { BankFieldsToDeserialize {
blockhash_queue: self.blockhash_queue, blockhash_queue: dvb.blockhash_queue,
ancestors: self.ancestors, ancestors: dvb.ancestors,
hash: self.hash, hash: dvb.hash,
parent_hash: self.parent_hash, parent_hash: dvb.parent_hash,
parent_slot: self.parent_slot, parent_slot: dvb.parent_slot,
hard_forks: self.hard_forks, hard_forks: dvb.hard_forks,
transaction_count: self.transaction_count, transaction_count: dvb.transaction_count,
tick_height: self.tick_height, tick_height: dvb.tick_height,
signature_count: self.signature_count, signature_count: dvb.signature_count,
capitalization: self.capitalization, capitalization: dvb.capitalization,
max_tick_height: self.max_tick_height, max_tick_height: dvb.max_tick_height,
hashes_per_tick: self.hashes_per_tick, hashes_per_tick: dvb.hashes_per_tick,
ticks_per_slot: self.ticks_per_slot, ticks_per_slot: dvb.ticks_per_slot,
ns_per_slot: self.ns_per_slot, ns_per_slot: dvb.ns_per_slot,
genesis_creation_time: self.genesis_creation_time, genesis_creation_time: dvb.genesis_creation_time,
slots_per_year: self.slots_per_year, slots_per_year: dvb.slots_per_year,
unused: self.unused, unused: dvb.unused,
slot: self.slot, slot: dvb.slot,
epoch: self.epoch, epoch: dvb.epoch,
block_height: self.block_height, block_height: dvb.block_height,
collector_id: self.collector_id, collector_id: dvb.collector_id,
collector_fees: self.collector_fees, collector_fees: dvb.collector_fees,
fee_calculator: self.fee_calculator, fee_calculator: dvb.fee_calculator,
fee_rate_governor: self.fee_rate_governor, fee_rate_governor: dvb.fee_rate_governor,
collected_rent: self.collected_rent, collected_rent: dvb.collected_rent,
rent_collector: self.rent_collector, rent_collector: dvb.rent_collector,
epoch_schedule: self.epoch_schedule, epoch_schedule: dvb.epoch_schedule,
inflation: self.inflation, inflation: dvb.inflation,
stakes: self.stakes, stakes: dvb.stakes,
epoch_stakes: self.epoch_stakes, epoch_stakes: dvb.epoch_stakes,
is_delta: self.is_delta, is_delta: dvb.is_delta,
} }
} }
} }
// Serializable version of Bank, not Deserializable to avoid cloning by using refs. // Serializable version of Bank, not Deserializable to avoid cloning by using refs.
// Sync fields with DeserializableVersionedBank! // Sync fields with DeserializableVersionedBank!
#[derive(Serialize)] #[derive(Serialize)]

View File

@ -585,7 +585,7 @@ pub fn remove_snapshot<P: AsRef<Path>>(slot: Slot, snapshot_path: P) -> Result<(
pub fn bank_from_archive<P: AsRef<Path>>( pub fn bank_from_archive<P: AsRef<Path>>(
account_paths: &[PathBuf], account_paths: &[PathBuf],
frozen_account_pubkeys: &[Pubkey], frozen_account_pubkeys: &[Pubkey],
snapshot_path: &PathBuf, snapshot_path: &Path,
snapshot_tar: P, snapshot_tar: P,
archive_format: ArchiveFormat, archive_format: ArchiveFormat,
genesis_config: &GenesisConfig, genesis_config: &GenesisConfig,
@ -760,7 +760,7 @@ fn rebuild_bank_from_snapshots<P>(
snapshot_version: &str, snapshot_version: &str,
account_paths: &[PathBuf], account_paths: &[PathBuf],
frozen_account_pubkeys: &[Pubkey], frozen_account_pubkeys: &[Pubkey],
unpacked_snapshots_dir: &PathBuf, unpacked_snapshots_dir: &Path,
append_vecs_path: P, append_vecs_path: P,
genesis_config: &GenesisConfig, genesis_config: &GenesisConfig,
debug_keys: Option<Arc<HashSet<Pubkey>>>, debug_keys: Option<Arc<HashSet<Pubkey>>>,

View File

@ -10,7 +10,7 @@ if ! command -v grcov; then
exit 1 exit 1
fi fi
if [[ ! "$(grcov --version)" =~ 0.6.1 ]]; then if [[ ! "$(grcov --version)" =~ 0.[67].1 ]]; then
echo Error: Required grcov version not installed echo Error: Required grcov version not installed
echo "Installed version: $(grcov --version)" echo "Installed version: $(grcov --version)"

View File

@ -70,11 +70,7 @@ where
} }
} }
fn build_bpf_package( fn build_bpf_package(config: &Config, target_directory: &Path, package: &cargo_metadata::Package) {
config: &Config,
target_directory: &PathBuf,
package: &cargo_metadata::Package,
) {
let program_name = { let program_name = {
let cdylib_targets = package let cdylib_targets = package
.targets .targets

View File

@ -67,11 +67,7 @@ where
} }
} }
fn test_bpf_package( fn test_bpf_package(config: &Config, target_directory: &Path, package: &cargo_metadata::Package) {
config: &Config,
target_directory: &PathBuf,
package: &cargo_metadata::Package,
) {
let set_test_bpf_feature = package.features.contains_key("test-bpf"); let set_test_bpf_feature = package.features.contains_key("test-bpf");
let bpf_out_dir = config let bpf_out_dir = config

View File

@ -950,6 +950,17 @@ impl<T> From<Option<T>> for COption<T> {
} }
} }
#[rustversion::since(1.49.0)]
impl<T> From<COption<T>> for Option<T> {
fn from(coption: COption<T>) -> Self {
match coption {
COption::Some(value) => Some(value),
COption::None => None,
}
}
}
#[rustversion::before(1.49.0)] // Remove `Into` once the BPF toolchain upgrades to 1.49.0 or newer
impl<T> Into<Option<T>> for COption<T> { impl<T> Into<Option<T>> for COption<T> {
fn into(self) -> Option<T> { fn into(self) -> Option<T> {
match self { match self {

View File

@ -43,7 +43,7 @@ pub trait Pack: Sealed {
if input.len() != Self::LEN { if input.len() != Self::LEN {
return Err(ProgramError::InvalidAccountData); return Err(ProgramError::InvalidAccountData);
} }
Ok(Self::unpack_from_slice(input)?) Self::unpack_from_slice(input)
} }
/// Pack into slice /// Pack into slice

View File

@ -125,9 +125,9 @@ impl fmt::Display for Signature {
} }
} }
impl Into<[u8; 64]> for Signature { impl From<Signature> for [u8; 64] {
fn into(self) -> [u8; 64] { fn from(signature: Signature) -> Self {
<GenericArray<u8, U64> as Into<[u8; 64]>>::into(self.0) signature.0.into()
} }
} }

View File

@ -89,6 +89,7 @@ impl AccessToken {
return; return;
} }
#[allow(deprecated)]
if self if self
.refresh_active .refresh_active
.compare_and_swap(false, true, Ordering::Relaxed) .compare_and_swap(false, true, Ordering::Relaxed)

View File

@ -768,11 +768,11 @@ impl TryFrom<tx_by_addr::TransactionByAddr> for Vec<TransactionByAddrInfo> {
type Error = &'static str; type Error = &'static str;
fn try_from(collection: tx_by_addr::TransactionByAddr) -> Result<Self, Self::Error> { fn try_from(collection: tx_by_addr::TransactionByAddr) -> Result<Self, Self::Error> {
Ok(collection collection
.tx_by_addrs .tx_by_addrs
.into_iter() .into_iter()
.map(|tx_by_addr| tx_by_addr.try_into()) .map(|tx_by_addr| tx_by_addr.try_into())
.collect::<Result<Vec<TransactionByAddrInfo>, Self::Error>>()?) .collect::<Result<Vec<TransactionByAddrInfo>, Self::Error>>()
} }
} }