Upgrade to Rust v1.49.0
This commit is contained in:
parent
7604edb16f
commit
cbffab7850
|
@ -1,4 +1,4 @@
|
||||||
FROM solanalabs/rust:1.48.0
|
FROM solanalabs/rust:1.49.0
|
||||||
ARG date
|
ARG date
|
||||||
|
|
||||||
RUN set -x \
|
RUN set -x \
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
# Note: when the rust version is changed also modify
|
# Note: when the rust version is changed also modify
|
||||||
# ci/rust-version.sh to pick up the new image tag
|
# ci/rust-version.sh to pick up the new image tag
|
||||||
FROM rust:1.48.0
|
FROM rust:1.49.0
|
||||||
|
|
||||||
# Add Google Protocol Buffers for Libra's metrics library.
|
# Add Google Protocol Buffers for Libra's metrics library.
|
||||||
ENV PROTOC_VERSION 3.8.0
|
ENV PROTOC_VERSION 3.8.0
|
||||||
|
|
|
@ -18,13 +18,13 @@
|
||||||
if [[ -n $RUST_STABLE_VERSION ]]; then
|
if [[ -n $RUST_STABLE_VERSION ]]; then
|
||||||
stable_version="$RUST_STABLE_VERSION"
|
stable_version="$RUST_STABLE_VERSION"
|
||||||
else
|
else
|
||||||
stable_version=1.48.0
|
stable_version=1.49.0
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ -n $RUST_NIGHTLY_VERSION ]]; then
|
if [[ -n $RUST_NIGHTLY_VERSION ]]; then
|
||||||
nightly_version="$RUST_NIGHTLY_VERSION"
|
nightly_version="$RUST_NIGHTLY_VERSION"
|
||||||
else
|
else
|
||||||
nightly_version=2020-12-13
|
nightly_version=2021-01-23
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -167,12 +167,12 @@ pub fn resolve_signer(
|
||||||
name: &str,
|
name: &str,
|
||||||
wallet_manager: &mut Option<Arc<RemoteWalletManager>>,
|
wallet_manager: &mut Option<Arc<RemoteWalletManager>>,
|
||||||
) -> Result<Option<String>, Box<dyn std::error::Error>> {
|
) -> Result<Option<String>, Box<dyn std::error::Error>> {
|
||||||
Ok(resolve_signer_from_path(
|
resolve_signer_from_path(
|
||||||
matches,
|
matches,
|
||||||
matches.value_of(name).unwrap(),
|
matches.value_of(name).unwrap(),
|
||||||
name,
|
name,
|
||||||
wallet_manager,
|
wallet_manager,
|
||||||
)?)
|
)
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn lamports_of_sol(matches: &ArgMatches<'_>, name: &str) -> Option<u64> {
|
pub fn lamports_of_sol(matches: &ArgMatches<'_>, name: &str) -> Option<u64> {
|
||||||
|
|
|
@ -35,16 +35,16 @@ impl From<TransportError> for ClientErrorKind {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<TransportError> for ClientErrorKind {
|
impl From<ClientErrorKind> for TransportError {
|
||||||
fn into(self) -> TransportError {
|
fn from(client_error_kind: ClientErrorKind) -> Self {
|
||||||
match self {
|
match client_error_kind {
|
||||||
Self::Io(err) => TransportError::IoError(err),
|
ClientErrorKind::Io(err) => Self::IoError(err),
|
||||||
Self::TransactionError(err) => TransportError::TransactionError(err),
|
ClientErrorKind::TransactionError(err) => Self::TransactionError(err),
|
||||||
Self::Reqwest(err) => TransportError::Custom(format!("{:?}", err)),
|
ClientErrorKind::Reqwest(err) => Self::Custom(format!("{:?}", err)),
|
||||||
Self::RpcError(err) => TransportError::Custom(format!("{:?}", err)),
|
ClientErrorKind::RpcError(err) => Self::Custom(format!("{:?}", err)),
|
||||||
Self::SerdeJson(err) => TransportError::Custom(format!("{:?}", err)),
|
ClientErrorKind::SerdeJson(err) => Self::Custom(format!("{:?}", err)),
|
||||||
Self::SigningError(err) => TransportError::Custom(format!("{:?}", err)),
|
ClientErrorKind::SigningError(err) => Self::Custom(format!("{:?}", err)),
|
||||||
Self::Custom(err) => TransportError::Custom(format!("{:?}", err)),
|
ClientErrorKind::Custom(err) => Self::Custom(format!("{:?}", err)),
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -100,9 +100,9 @@ impl From<TransportError> for ClientError {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<TransportError> for ClientError {
|
impl From<ClientError> for TransportError {
|
||||||
fn into(self) -> TransportError {
|
fn from(client_error: ClientError) -> Self {
|
||||||
self.kind.into()
|
client_error.kind.into()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -347,6 +347,7 @@ fn update_peer_stats(
|
||||||
) {
|
) {
|
||||||
let now = timestamp();
|
let now = timestamp();
|
||||||
let last = last_datapoint_submit.load(Ordering::Relaxed);
|
let last = last_datapoint_submit.load(Ordering::Relaxed);
|
||||||
|
#[allow(deprecated)]
|
||||||
if now.saturating_sub(last) > 1000
|
if now.saturating_sub(last) > 1000
|
||||||
&& last_datapoint_submit.compare_and_swap(last, now, Ordering::Relaxed) == last
|
&& last_datapoint_submit.compare_and_swap(last, now, Ordering::Relaxed) == last
|
||||||
{
|
{
|
||||||
|
|
|
@ -329,6 +329,7 @@ impl StandardBroadcastRun {
|
||||||
let mut get_peers_time = Measure::start("broadcast::get_peers");
|
let mut get_peers_time = Measure::start("broadcast::get_peers");
|
||||||
let now = timestamp();
|
let now = timestamp();
|
||||||
let last = self.last_peer_update.load(Ordering::Relaxed);
|
let last = self.last_peer_update.load(Ordering::Relaxed);
|
||||||
|
#[allow(deprecated)]
|
||||||
if now - last > BROADCAST_PEER_UPDATE_INTERVAL_MS
|
if now - last > BROADCAST_PEER_UPDATE_INTERVAL_MS
|
||||||
&& self
|
&& self
|
||||||
.last_peer_update
|
.last_peer_update
|
||||||
|
|
|
@ -403,6 +403,7 @@ impl ClusterInfoVoteListener {
|
||||||
let last_version = bank.last_vote_sync.load(Ordering::Relaxed);
|
let last_version = bank.last_vote_sync.load(Ordering::Relaxed);
|
||||||
let (new_version, msgs) = verified_vote_packets.get_latest_votes(last_version);
|
let (new_version, msgs) = verified_vote_packets.get_latest_votes(last_version);
|
||||||
verified_packets_sender.send(msgs)?;
|
verified_packets_sender.send(msgs)?;
|
||||||
|
#[allow(deprecated)]
|
||||||
bank.last_vote_sync.compare_and_swap(
|
bank.last_vote_sync.compare_and_swap(
|
||||||
last_version,
|
last_version,
|
||||||
new_version,
|
new_version,
|
||||||
|
|
|
@ -263,15 +263,9 @@ mod tests {
|
||||||
#[test]
|
#[test]
|
||||||
fn test_get_highest_confirmed_root() {
|
fn test_get_highest_confirmed_root() {
|
||||||
assert_eq!(get_highest_confirmed_root(vec![], 10), 0);
|
assert_eq!(get_highest_confirmed_root(vec![], 10), 0);
|
||||||
let mut rooted_stake = vec![];
|
let rooted_stake = vec![(0, 5), (1, 5)];
|
||||||
rooted_stake.push((0, 5));
|
|
||||||
rooted_stake.push((1, 5));
|
|
||||||
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 0);
|
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 0);
|
||||||
let mut rooted_stake = vec![];
|
let rooted_stake = vec![(1, 5), (0, 10), (2, 5), (1, 4)];
|
||||||
rooted_stake.push((1, 5));
|
|
||||||
rooted_stake.push((0, 10));
|
|
||||||
rooted_stake.push((2, 5));
|
|
||||||
rooted_stake.push((1, 4));
|
|
||||||
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 1);
|
assert_eq!(get_highest_confirmed_root(rooted_stake, 10), 1);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -145,10 +145,10 @@ impl CrdsFilterSet {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<Vec<CrdsFilter>> for CrdsFilterSet {
|
impl From<CrdsFilterSet> for Vec<CrdsFilter> {
|
||||||
fn into(self) -> Vec<CrdsFilter> {
|
fn from(cfs: CrdsFilterSet) -> Self {
|
||||||
let mask_bits = self.mask_bits;
|
let mask_bits = cfs.mask_bits;
|
||||||
self.filters
|
cfs.filters
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.enumerate()
|
.enumerate()
|
||||||
.map(|(seed, filter)| CrdsFilter {
|
.map(|(seed, filter)| CrdsFilter {
|
||||||
|
|
|
@ -303,7 +303,7 @@ impl<'de> Deserialize<'de> for Vote {
|
||||||
from: Pubkey,
|
from: Pubkey,
|
||||||
transaction: Transaction,
|
transaction: Transaction,
|
||||||
wallclock: u64,
|
wallclock: u64,
|
||||||
};
|
}
|
||||||
let vote = Vote::deserialize(deserializer)?;
|
let vote = Vote::deserialize(deserializer)?;
|
||||||
let vote = match vote.transaction.sanitize() {
|
let vote = match vote.transaction.sanitize() {
|
||||||
Ok(_) => Self::new(vote.from, vote.transaction, vote.wallclock),
|
Ok(_) => Self::new(vote.from, vote.transaction, vote.wallclock),
|
||||||
|
|
|
@ -125,7 +125,7 @@ impl PohService {
|
||||||
// sleep is not accurate enough to get a predictable time.
|
// sleep is not accurate enough to get a predictable time.
|
||||||
// Kernel can not schedule the thread for a while.
|
// Kernel can not schedule the thread for a while.
|
||||||
while (now.elapsed().as_nanos() as u64) < target_tick_ns {
|
while (now.elapsed().as_nanos() as u64) < target_tick_ns {
|
||||||
std::sync::atomic::spin_loop_hint();
|
std::hint::spin_loop();
|
||||||
}
|
}
|
||||||
total_sleep_us += (now.elapsed().as_nanos() as u64 - elapsed_ns) / 1000;
|
total_sleep_us += (now.elapsed().as_nanos() as u64 - elapsed_ns) / 1000;
|
||||||
now = Instant::now();
|
now = Instant::now();
|
||||||
|
|
|
@ -121,6 +121,7 @@ fn update_retransmit_stats(
|
||||||
|
|
||||||
let now = timestamp();
|
let now = timestamp();
|
||||||
let last = stats.last_ts.load(Ordering::Relaxed);
|
let last = stats.last_ts.load(Ordering::Relaxed);
|
||||||
|
#[allow(deprecated)]
|
||||||
if now.saturating_sub(last) > 2000
|
if now.saturating_sub(last) > 2000
|
||||||
&& stats.last_ts.compare_and_swap(last, now, Ordering::Relaxed) == last
|
&& stats.last_ts.compare_and_swap(last, now, Ordering::Relaxed) == last
|
||||||
{
|
{
|
||||||
|
@ -310,6 +311,7 @@ fn retransmit(
|
||||||
|
|
||||||
let now = timestamp();
|
let now = timestamp();
|
||||||
let last = last_peer_update.load(Ordering::Relaxed);
|
let last = last_peer_update.load(Ordering::Relaxed);
|
||||||
|
#[allow(deprecated)]
|
||||||
if now.saturating_sub(last) > 1000
|
if now.saturating_sub(last) > 1000
|
||||||
&& last_peer_update.compare_and_swap(last, now, Ordering::Relaxed) == last
|
&& last_peer_update.compare_and_swap(last, now, Ordering::Relaxed) == last
|
||||||
{
|
{
|
||||||
|
|
|
@ -749,7 +749,7 @@ impl JsonRpcRequestProcessor {
|
||||||
// If the starting slot is lower than what's available in blockstore assume the entire
|
// If the starting slot is lower than what's available in blockstore assume the entire
|
||||||
// [start_slot..end_slot] can be fetched from BigTable.
|
// [start_slot..end_slot] can be fetched from BigTable.
|
||||||
if let Some(bigtable_ledger_storage) = &self.bigtable_ledger_storage {
|
if let Some(bigtable_ledger_storage) = &self.bigtable_ledger_storage {
|
||||||
return Ok(self
|
return self
|
||||||
.runtime_handle
|
.runtime_handle
|
||||||
.block_on(
|
.block_on(
|
||||||
bigtable_ledger_storage
|
bigtable_ledger_storage
|
||||||
|
@ -764,7 +764,7 @@ impl JsonRpcRequestProcessor {
|
||||||
"BigTable query failed (maybe timeout due to too large range?)"
|
"BigTable query failed (maybe timeout due to too large range?)"
|
||||||
.to_string(),
|
.to_string(),
|
||||||
)
|
)
|
||||||
})?);
|
});
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -119,11 +119,12 @@ impl Tpu {
|
||||||
}
|
}
|
||||||
|
|
||||||
pub fn join(self) -> thread::Result<()> {
|
pub fn join(self) -> thread::Result<()> {
|
||||||
let mut results = vec![];
|
let results = vec![
|
||||||
results.push(self.fetch_stage.join());
|
self.fetch_stage.join(),
|
||||||
results.push(self.sigverify_stage.join());
|
self.sigverify_stage.join(),
|
||||||
results.push(self.cluster_info_vote_listener.join());
|
self.cluster_info_vote_listener.join(),
|
||||||
results.push(self.banking_stage.join());
|
self.banking_stage.join(),
|
||||||
|
];
|
||||||
let broadcast_result = self.broadcast_stage.join();
|
let broadcast_result = self.broadcast_stage.join();
|
||||||
for result in results {
|
for result in results {
|
||||||
result?;
|
result?;
|
||||||
|
|
|
@ -71,6 +71,7 @@ fn filter_allow_attrs(attrs: &mut Vec<Attribute>) {
|
||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[allow(deprecated)]
|
||||||
#[cfg(RUSTC_WITH_SPECIALIZATION)]
|
#[cfg(RUSTC_WITH_SPECIALIZATION)]
|
||||||
fn quote_for_specialization_detection() -> TokenStream2 {
|
fn quote_for_specialization_detection() -> TokenStream2 {
|
||||||
lazy_static! {
|
lazy_static! {
|
||||||
|
|
|
@ -645,7 +645,7 @@ fn hardforks_of(matches: &ArgMatches<'_>, name: &str) -> Option<Vec<Slot>> {
|
||||||
|
|
||||||
fn load_bank_forks(
|
fn load_bank_forks(
|
||||||
arg_matches: &ArgMatches,
|
arg_matches: &ArgMatches,
|
||||||
ledger_path: &PathBuf,
|
ledger_path: &Path,
|
||||||
genesis_config: &GenesisConfig,
|
genesis_config: &GenesisConfig,
|
||||||
process_options: ProcessOptions,
|
process_options: ProcessOptions,
|
||||||
access_type: AccessType,
|
access_type: AccessType,
|
||||||
|
@ -653,7 +653,7 @@ fn load_bank_forks(
|
||||||
snapshot_archive_path: Option<PathBuf>,
|
snapshot_archive_path: Option<PathBuf>,
|
||||||
) -> bank_forks_utils::LoadResult {
|
) -> bank_forks_utils::LoadResult {
|
||||||
let blockstore = open_blockstore(&ledger_path, access_type, wal_recovery_mode);
|
let blockstore = open_blockstore(&ledger_path, access_type, wal_recovery_mode);
|
||||||
let snapshot_path = ledger_path.clone().join(if blockstore.is_primary_access() {
|
let snapshot_path = ledger_path.join(if blockstore.is_primary_access() {
|
||||||
"snapshot"
|
"snapshot"
|
||||||
} else {
|
} else {
|
||||||
"snapshot.ledger-tool"
|
"snapshot.ledger-tool"
|
||||||
|
@ -662,7 +662,7 @@ fn load_bank_forks(
|
||||||
None
|
None
|
||||||
} else {
|
} else {
|
||||||
let snapshot_package_output_path =
|
let snapshot_package_output_path =
|
||||||
snapshot_archive_path.unwrap_or_else(|| ledger_path.clone());
|
snapshot_archive_path.unwrap_or_else(|| ledger_path.to_path_buf());
|
||||||
Some(SnapshotConfig {
|
Some(SnapshotConfig {
|
||||||
snapshot_interval_slots: 0, // Value doesn't matter
|
snapshot_interval_slots: 0, // Value doesn't matter
|
||||||
snapshot_package_output_path,
|
snapshot_package_output_path,
|
||||||
|
@ -1443,7 +1443,7 @@ fn main() {
|
||||||
last_in_slot: bool,
|
last_in_slot: bool,
|
||||||
data_complete: bool,
|
data_complete: bool,
|
||||||
shred: &'a Shred,
|
shred: &'a Shred,
|
||||||
};
|
}
|
||||||
let starting_slot = value_t_or_exit!(arg_matches, "starting_slot", Slot);
|
let starting_slot = value_t_or_exit!(arg_matches, "starting_slot", Slot);
|
||||||
let ending_slot = value_t!(arg_matches, "ending_slot", Slot).unwrap_or(Slot::MAX);
|
let ending_slot = value_t!(arg_matches, "ending_slot", Slot).unwrap_or(Slot::MAX);
|
||||||
let ledger = open_blockstore(&ledger_path, AccessType::TryPrimaryThenSecondary, None);
|
let ledger = open_blockstore(&ledger_path, AccessType::TryPrimaryThenSecondary, None);
|
||||||
|
@ -2442,13 +2442,13 @@ fn main() {
|
||||||
cluster_points: String,
|
cluster_points: String,
|
||||||
old_capitalization: u64,
|
old_capitalization: u64,
|
||||||
new_capitalization: u64,
|
new_capitalization: u64,
|
||||||
};
|
}
|
||||||
fn format_or_na<T: std::fmt::Display>(
|
fn format_or_na<T: std::fmt::Display>(
|
||||||
data: Option<T>,
|
data: Option<T>,
|
||||||
) -> String {
|
) -> String {
|
||||||
data.map(|data| format!("{}", data))
|
data.map(|data| format!("{}", data))
|
||||||
.unwrap_or_else(|| "N/A".to_owned())
|
.unwrap_or_else(|| "N/A".to_owned())
|
||||||
};
|
}
|
||||||
let mut point_details = detail
|
let mut point_details = detail
|
||||||
.map(|d| d.points.iter().map(Some).collect::<Vec<_>>())
|
.map(|d| d.points.iter().map(Some).collect::<Vec<_>>())
|
||||||
.unwrap_or_default();
|
.unwrap_or_default();
|
||||||
|
|
|
@ -185,9 +185,10 @@ impl From<&str> for BlockstoreRecoveryMode {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
impl Into<DBRecoveryMode> for BlockstoreRecoveryMode {
|
|
||||||
fn into(self) -> DBRecoveryMode {
|
impl From<BlockstoreRecoveryMode> for DBRecoveryMode {
|
||||||
match self {
|
fn from(brm: BlockstoreRecoveryMode) -> Self {
|
||||||
|
match brm {
|
||||||
BlockstoreRecoveryMode::TolerateCorruptedTailRecords => {
|
BlockstoreRecoveryMode::TolerateCorruptedTailRecords => {
|
||||||
DBRecoveryMode::TolerateCorruptedTailRecords
|
DBRecoveryMode::TolerateCorruptedTailRecords
|
||||||
}
|
}
|
||||||
|
@ -404,6 +405,7 @@ pub trait Column {
|
||||||
fn key(index: Self::Index) -> Vec<u8>;
|
fn key(index: Self::Index) -> Vec<u8>;
|
||||||
fn index(key: &[u8]) -> Self::Index;
|
fn index(key: &[u8]) -> Self::Index;
|
||||||
fn primary_index(index: Self::Index) -> Slot;
|
fn primary_index(index: Self::Index) -> Slot;
|
||||||
|
#[allow(clippy::wrong_self_convention)]
|
||||||
fn as_index(slot: Slot) -> Self::Index;
|
fn as_index(slot: Slot) -> Self::Index;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -164,6 +164,7 @@ impl Counter {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
pub fn init(&mut self) {
|
pub fn init(&mut self) {
|
||||||
|
#![allow(deprecated)]
|
||||||
self.lograte
|
self.lograte
|
||||||
.compare_and_swap(0, Self::default_log_rate(), Ordering::Relaxed);
|
.compare_and_swap(0, Self::default_log_rate(), Ordering::Relaxed);
|
||||||
self.metricsrate
|
self.metricsrate
|
||||||
|
@ -188,6 +189,7 @@ impl Counter {
|
||||||
}
|
}
|
||||||
|
|
||||||
let lastlog = self.lastlog.load(Ordering::Relaxed);
|
let lastlog = self.lastlog.load(Ordering::Relaxed);
|
||||||
|
#[allow(deprecated)]
|
||||||
let prev = self
|
let prev = self
|
||||||
.lastlog
|
.lastlog
|
||||||
.compare_and_swap(lastlog, counts, Ordering::Relaxed);
|
.compare_and_swap(lastlog, counts, Ordering::Relaxed);
|
||||||
|
|
|
@ -19,11 +19,11 @@ use std::{
|
||||||
|
|
||||||
type CounterMap = HashMap<(&'static str, u64), CounterPoint>;
|
type CounterMap = HashMap<(&'static str, u64), CounterPoint>;
|
||||||
|
|
||||||
impl Into<DataPoint> for CounterPoint {
|
impl From<CounterPoint> for DataPoint {
|
||||||
fn into(self) -> DataPoint {
|
fn from(counter_point: CounterPoint) -> Self {
|
||||||
let mut point = DataPoint::new(self.name);
|
let mut point = Self::new(counter_point.name);
|
||||||
point.timestamp = self.timestamp;
|
point.timestamp = counter_point.timestamp;
|
||||||
point.add_field_i64("count", self.count);
|
point.add_field_i64("count", counter_point.count);
|
||||||
point
|
point
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -89,18 +89,17 @@ impl<T: Clone + Default + Sized> Default for PinnedVec<T> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: Clone + Default + Sized> Into<Vec<T>> for PinnedVec<T> {
|
impl<T: Clone + Default + Sized> From<PinnedVec<T>> for Vec<T> {
|
||||||
fn into(mut self) -> Vec<T> {
|
fn from(mut pinned_vec: PinnedVec<T>) -> Self {
|
||||||
if self.pinned {
|
if pinned_vec.pinned {
|
||||||
unpin(self.x.as_mut_ptr());
|
unpin(pinned_vec.x.as_mut_ptr());
|
||||||
self.pinned = false;
|
pinned_vec.pinned = false;
|
||||||
}
|
}
|
||||||
self.pinnable = false;
|
pinned_vec.pinnable = false;
|
||||||
self.recycler = None;
|
pinned_vec.recycler = None;
|
||||||
std::mem::take(&mut self.x)
|
std::mem::take(&mut pinned_vec.x)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
pub struct PinnedIter<'a, T>(std::slice::Iter<'a, T>);
|
pub struct PinnedIter<'a, T>(std::slice::Iter<'a, T>);
|
||||||
|
|
||||||
pub struct PinnedIterMut<'a, T>(std::slice::IterMut<'a, T>);
|
pub struct PinnedIterMut<'a, T>(std::slice::IterMut<'a, T>);
|
||||||
|
|
|
@ -114,9 +114,12 @@ impl<T: Default + Reset> RecyclerX<T> {
|
||||||
let max_gc = self.stats.max_gc.load(Ordering::Relaxed);
|
let max_gc = self.stats.max_gc.load(Ordering::Relaxed);
|
||||||
if len > max_gc {
|
if len > max_gc {
|
||||||
// this is not completely accurate, but for most cases should be fine.
|
// this is not completely accurate, but for most cases should be fine.
|
||||||
self.stats
|
let _ = self.stats.max_gc.compare_exchange(
|
||||||
.max_gc
|
max_gc,
|
||||||
.compare_and_swap(max_gc, len, Ordering::Relaxed);
|
len,
|
||||||
|
Ordering::Relaxed,
|
||||||
|
Ordering::Relaxed,
|
||||||
|
);
|
||||||
}
|
}
|
||||||
let total = self.stats.total.load(Ordering::Relaxed);
|
let total = self.stats.total.load(Ordering::Relaxed);
|
||||||
let reuse = self.stats.reuse.load(Ordering::Relaxed);
|
let reuse = self.stats.reuse.load(Ordering::Relaxed);
|
||||||
|
|
|
@ -3908,6 +3908,8 @@ impl AccountsDB {
|
||||||
fn report_store_timings(&self) {
|
fn report_store_timings(&self) {
|
||||||
let last = self.stats.last_store_report.load(Ordering::Relaxed);
|
let last = self.stats.last_store_report.load(Ordering::Relaxed);
|
||||||
let now = solana_sdk::timing::timestamp();
|
let now = solana_sdk::timing::timestamp();
|
||||||
|
|
||||||
|
#[allow(deprecated)]
|
||||||
if now.saturating_sub(last) > 1000
|
if now.saturating_sub(last) > 1000
|
||||||
&& self
|
&& self
|
||||||
.stats
|
.stats
|
||||||
|
|
|
@ -196,14 +196,18 @@ impl<T: BloomHashIndex> AtomicBloom<T> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<T: BloomHashIndex> Into<Bloom<T>> for AtomicBloom<T> {
|
impl<T: BloomHashIndex> From<AtomicBloom<T>> for Bloom<T> {
|
||||||
fn into(self) -> Bloom<T> {
|
fn from(atomic_bloom: AtomicBloom<T>) -> Self {
|
||||||
let bits: Vec<_> = self.bits.into_iter().map(AtomicU64::into_inner).collect();
|
let bits: Vec<_> = atomic_bloom
|
||||||
|
.bits
|
||||||
|
.into_iter()
|
||||||
|
.map(AtomicU64::into_inner)
|
||||||
|
.collect();
|
||||||
let num_bits_set = bits.iter().map(|x| x.count_ones() as u64).sum();
|
let num_bits_set = bits.iter().map(|x| x.count_ones() as u64).sum();
|
||||||
let mut bits: BitVec<u64> = bits.into();
|
let mut bits: BitVec<u64> = bits.into();
|
||||||
bits.truncate(self.num_bits);
|
bits.truncate(atomic_bloom.num_bits);
|
||||||
Bloom {
|
Bloom {
|
||||||
keys: self.keys,
|
keys: atomic_bloom.keys,
|
||||||
bits,
|
bits,
|
||||||
num_bits_set,
|
num_bits_set,
|
||||||
_phantom: PhantomData::default(),
|
_phantom: PhantomData::default(),
|
||||||
|
|
|
@ -30,9 +30,9 @@ impl LogCollector {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<Vec<String>> for LogCollector {
|
impl From<LogCollector> for Vec<String> {
|
||||||
fn into(self) -> Vec<String> {
|
fn from(log_collector: LogCollector) -> Self {
|
||||||
self.inner.into_inner().messages
|
log_collector.inner.into_inner().messages
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -14,7 +14,13 @@ use solana_sdk::{
|
||||||
process_instruction::{InvokeContext, LoaderEntrypoint},
|
process_instruction::{InvokeContext, LoaderEntrypoint},
|
||||||
pubkey::Pubkey,
|
pubkey::Pubkey,
|
||||||
};
|
};
|
||||||
use std::{collections::HashMap, env, path::PathBuf, str, sync::RwLock};
|
use std::{
|
||||||
|
collections::HashMap,
|
||||||
|
env,
|
||||||
|
path::{Path, PathBuf},
|
||||||
|
str,
|
||||||
|
sync::RwLock,
|
||||||
|
};
|
||||||
use thiserror::Error;
|
use thiserror::Error;
|
||||||
|
|
||||||
#[derive(Error, Debug, Serialize, Clone, PartialEq, FromPrimitive, ToPrimitive)]
|
#[derive(Error, Debug, Serialize, Clone, PartialEq, FromPrimitive, ToPrimitive)]
|
||||||
|
@ -85,12 +91,12 @@ impl NativeLoader {
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(windows)]
|
#[cfg(windows)]
|
||||||
fn library_open(path: &PathBuf) -> Result<Library, libloading::Error> {
|
fn library_open(path: &Path) -> Result<Library, libloading::Error> {
|
||||||
Library::new(path)
|
Library::new(path)
|
||||||
}
|
}
|
||||||
|
|
||||||
#[cfg(not(windows))]
|
#[cfg(not(windows))]
|
||||||
fn library_open(path: &PathBuf) -> Result<Library, libloading::Error> {
|
fn library_open(path: &Path) -> Result<Library, libloading::Error> {
|
||||||
// Linux tls bug can cause crash on dlclose(), workaround by never unloading
|
// Linux tls bug can cause crash on dlclose(), workaround by never unloading
|
||||||
Library::open(Some(path), libc::RTLD_NODELETE | libc::RTLD_NOW)
|
Library::open(Some(path), libc::RTLD_NODELETE | libc::RTLD_NOW)
|
||||||
}
|
}
|
||||||
|
|
|
@ -24,9 +24,9 @@ impl From<&AccountStorageEntry> for SerializableAccountStorageEntry {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<AccountStorageEntry> for SerializableAccountStorageEntry {
|
impl From<SerializableAccountStorageEntry> for AccountStorageEntry {
|
||||||
fn into(self) -> AccountStorageEntry {
|
fn from(s: SerializableAccountStorageEntry) -> Self {
|
||||||
AccountStorageEntry::new_empty_map(self.id, self.accounts_current_len)
|
AccountStorageEntry::new_empty_map(s.id, s.accounts_current_len)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -71,43 +71,44 @@ pub(crate) struct DeserializableVersionedBank {
|
||||||
pub(crate) message_processor: MessageProcessor,
|
pub(crate) message_processor: MessageProcessor,
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<BankFieldsToDeserialize> for DeserializableVersionedBank {
|
impl From<DeserializableVersionedBank> for BankFieldsToDeserialize {
|
||||||
fn into(self) -> BankFieldsToDeserialize {
|
fn from(dvb: DeserializableVersionedBank) -> Self {
|
||||||
BankFieldsToDeserialize {
|
BankFieldsToDeserialize {
|
||||||
blockhash_queue: self.blockhash_queue,
|
blockhash_queue: dvb.blockhash_queue,
|
||||||
ancestors: self.ancestors,
|
ancestors: dvb.ancestors,
|
||||||
hash: self.hash,
|
hash: dvb.hash,
|
||||||
parent_hash: self.parent_hash,
|
parent_hash: dvb.parent_hash,
|
||||||
parent_slot: self.parent_slot,
|
parent_slot: dvb.parent_slot,
|
||||||
hard_forks: self.hard_forks,
|
hard_forks: dvb.hard_forks,
|
||||||
transaction_count: self.transaction_count,
|
transaction_count: dvb.transaction_count,
|
||||||
tick_height: self.tick_height,
|
tick_height: dvb.tick_height,
|
||||||
signature_count: self.signature_count,
|
signature_count: dvb.signature_count,
|
||||||
capitalization: self.capitalization,
|
capitalization: dvb.capitalization,
|
||||||
max_tick_height: self.max_tick_height,
|
max_tick_height: dvb.max_tick_height,
|
||||||
hashes_per_tick: self.hashes_per_tick,
|
hashes_per_tick: dvb.hashes_per_tick,
|
||||||
ticks_per_slot: self.ticks_per_slot,
|
ticks_per_slot: dvb.ticks_per_slot,
|
||||||
ns_per_slot: self.ns_per_slot,
|
ns_per_slot: dvb.ns_per_slot,
|
||||||
genesis_creation_time: self.genesis_creation_time,
|
genesis_creation_time: dvb.genesis_creation_time,
|
||||||
slots_per_year: self.slots_per_year,
|
slots_per_year: dvb.slots_per_year,
|
||||||
unused: self.unused,
|
unused: dvb.unused,
|
||||||
slot: self.slot,
|
slot: dvb.slot,
|
||||||
epoch: self.epoch,
|
epoch: dvb.epoch,
|
||||||
block_height: self.block_height,
|
block_height: dvb.block_height,
|
||||||
collector_id: self.collector_id,
|
collector_id: dvb.collector_id,
|
||||||
collector_fees: self.collector_fees,
|
collector_fees: dvb.collector_fees,
|
||||||
fee_calculator: self.fee_calculator,
|
fee_calculator: dvb.fee_calculator,
|
||||||
fee_rate_governor: self.fee_rate_governor,
|
fee_rate_governor: dvb.fee_rate_governor,
|
||||||
collected_rent: self.collected_rent,
|
collected_rent: dvb.collected_rent,
|
||||||
rent_collector: self.rent_collector,
|
rent_collector: dvb.rent_collector,
|
||||||
epoch_schedule: self.epoch_schedule,
|
epoch_schedule: dvb.epoch_schedule,
|
||||||
inflation: self.inflation,
|
inflation: dvb.inflation,
|
||||||
stakes: self.stakes,
|
stakes: dvb.stakes,
|
||||||
epoch_stakes: self.epoch_stakes,
|
epoch_stakes: dvb.epoch_stakes,
|
||||||
is_delta: self.is_delta,
|
is_delta: dvb.is_delta,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
// Serializable version of Bank, not Deserializable to avoid cloning by using refs.
|
// Serializable version of Bank, not Deserializable to avoid cloning by using refs.
|
||||||
// Sync fields with DeserializableVersionedBank!
|
// Sync fields with DeserializableVersionedBank!
|
||||||
#[derive(Serialize)]
|
#[derive(Serialize)]
|
||||||
|
|
|
@ -585,7 +585,7 @@ pub fn remove_snapshot<P: AsRef<Path>>(slot: Slot, snapshot_path: P) -> Result<(
|
||||||
pub fn bank_from_archive<P: AsRef<Path>>(
|
pub fn bank_from_archive<P: AsRef<Path>>(
|
||||||
account_paths: &[PathBuf],
|
account_paths: &[PathBuf],
|
||||||
frozen_account_pubkeys: &[Pubkey],
|
frozen_account_pubkeys: &[Pubkey],
|
||||||
snapshot_path: &PathBuf,
|
snapshot_path: &Path,
|
||||||
snapshot_tar: P,
|
snapshot_tar: P,
|
||||||
archive_format: ArchiveFormat,
|
archive_format: ArchiveFormat,
|
||||||
genesis_config: &GenesisConfig,
|
genesis_config: &GenesisConfig,
|
||||||
|
@ -760,7 +760,7 @@ fn rebuild_bank_from_snapshots<P>(
|
||||||
snapshot_version: &str,
|
snapshot_version: &str,
|
||||||
account_paths: &[PathBuf],
|
account_paths: &[PathBuf],
|
||||||
frozen_account_pubkeys: &[Pubkey],
|
frozen_account_pubkeys: &[Pubkey],
|
||||||
unpacked_snapshots_dir: &PathBuf,
|
unpacked_snapshots_dir: &Path,
|
||||||
append_vecs_path: P,
|
append_vecs_path: P,
|
||||||
genesis_config: &GenesisConfig,
|
genesis_config: &GenesisConfig,
|
||||||
debug_keys: Option<Arc<HashSet<Pubkey>>>,
|
debug_keys: Option<Arc<HashSet<Pubkey>>>,
|
||||||
|
|
|
@ -10,7 +10,7 @@ if ! command -v grcov; then
|
||||||
exit 1
|
exit 1
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [[ ! "$(grcov --version)" =~ 0.6.1 ]]; then
|
if [[ ! "$(grcov --version)" =~ 0.[67].1 ]]; then
|
||||||
echo Error: Required grcov version not installed
|
echo Error: Required grcov version not installed
|
||||||
|
|
||||||
echo "Installed version: $(grcov --version)"
|
echo "Installed version: $(grcov --version)"
|
||||||
|
|
|
@ -70,11 +70,7 @@ where
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn build_bpf_package(
|
fn build_bpf_package(config: &Config, target_directory: &Path, package: &cargo_metadata::Package) {
|
||||||
config: &Config,
|
|
||||||
target_directory: &PathBuf,
|
|
||||||
package: &cargo_metadata::Package,
|
|
||||||
) {
|
|
||||||
let program_name = {
|
let program_name = {
|
||||||
let cdylib_targets = package
|
let cdylib_targets = package
|
||||||
.targets
|
.targets
|
||||||
|
|
|
@ -67,11 +67,7 @@ where
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
fn test_bpf_package(
|
fn test_bpf_package(config: &Config, target_directory: &Path, package: &cargo_metadata::Package) {
|
||||||
config: &Config,
|
|
||||||
target_directory: &PathBuf,
|
|
||||||
package: &cargo_metadata::Package,
|
|
||||||
) {
|
|
||||||
let set_test_bpf_feature = package.features.contains_key("test-bpf");
|
let set_test_bpf_feature = package.features.contains_key("test-bpf");
|
||||||
|
|
||||||
let bpf_out_dir = config
|
let bpf_out_dir = config
|
||||||
|
|
|
@ -950,6 +950,17 @@ impl<T> From<Option<T>> for COption<T> {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[rustversion::since(1.49.0)]
|
||||||
|
impl<T> From<COption<T>> for Option<T> {
|
||||||
|
fn from(coption: COption<T>) -> Self {
|
||||||
|
match coption {
|
||||||
|
COption::Some(value) => Some(value),
|
||||||
|
COption::None => None,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[rustversion::before(1.49.0)] // Remove `Into` once the BPF toolchain upgrades to 1.49.0 or newer
|
||||||
impl<T> Into<Option<T>> for COption<T> {
|
impl<T> Into<Option<T>> for COption<T> {
|
||||||
fn into(self) -> Option<T> {
|
fn into(self) -> Option<T> {
|
||||||
match self {
|
match self {
|
||||||
|
|
|
@ -43,7 +43,7 @@ pub trait Pack: Sealed {
|
||||||
if input.len() != Self::LEN {
|
if input.len() != Self::LEN {
|
||||||
return Err(ProgramError::InvalidAccountData);
|
return Err(ProgramError::InvalidAccountData);
|
||||||
}
|
}
|
||||||
Ok(Self::unpack_from_slice(input)?)
|
Self::unpack_from_slice(input)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Pack into slice
|
/// Pack into slice
|
||||||
|
|
|
@ -125,9 +125,9 @@ impl fmt::Display for Signature {
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Into<[u8; 64]> for Signature {
|
impl From<Signature> for [u8; 64] {
|
||||||
fn into(self) -> [u8; 64] {
|
fn from(signature: Signature) -> Self {
|
||||||
<GenericArray<u8, U64> as Into<[u8; 64]>>::into(self.0)
|
signature.0.into()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -89,6 +89,7 @@ impl AccessToken {
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[allow(deprecated)]
|
||||||
if self
|
if self
|
||||||
.refresh_active
|
.refresh_active
|
||||||
.compare_and_swap(false, true, Ordering::Relaxed)
|
.compare_and_swap(false, true, Ordering::Relaxed)
|
||||||
|
|
|
@ -768,11 +768,11 @@ impl TryFrom<tx_by_addr::TransactionByAddr> for Vec<TransactionByAddrInfo> {
|
||||||
type Error = &'static str;
|
type Error = &'static str;
|
||||||
|
|
||||||
fn try_from(collection: tx_by_addr::TransactionByAddr) -> Result<Self, Self::Error> {
|
fn try_from(collection: tx_by_addr::TransactionByAddr) -> Result<Self, Self::Error> {
|
||||||
Ok(collection
|
collection
|
||||||
.tx_by_addrs
|
.tx_by_addrs
|
||||||
.into_iter()
|
.into_iter()
|
||||||
.map(|tx_by_addr| tx_by_addr.try_into())
|
.map(|tx_by_addr| tx_by_addr.try_into())
|
||||||
.collect::<Result<Vec<TransactionByAddrInfo>, Self::Error>>()?)
|
.collect::<Result<Vec<TransactionByAddrInfo>, Self::Error>>()
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue