Skip to content

Commit

Permalink
Fix merge issues
Browse files Browse the repository at this point in the history
  • Loading branch information
joncinque committed Jan 12, 2023
1 parent 7a27de6 commit 41653fe
Show file tree
Hide file tree
Showing 4 changed files with 25 additions and 24 deletions.
30 changes: 16 additions & 14 deletions client/src/connection_cache.rs
Original file line number Diff line number Diff line change
Expand Up @@ -465,6 +465,7 @@ mod tests {
rand::{Rng, SeedableRng},
rand_chacha::ChaChaRng,
solana_sdk::{
packet::TransactionPacket,
pubkey::Pubkey,
quic::{
QUIC_MAX_UNSTAKED_CONCURRENT_STREAMS, QUIC_MIN_STAKED_CONCURRENT_STREAMS,
Expand Down Expand Up @@ -657,20 +658,21 @@ mod tests {

let staked_nodes = Arc::new(RwLock::new(StakedNodes::default()));

let (response_recv_endpoint, response_recv_thread) = solana_streamer::quic::spawn_server(
response_recv_socket,
&keypair2,
response_recv_ip,
sender2,
response_recv_exit.clone(),
1,
staked_nodes,
10,
10,
response_recv_stats,
DEFAULT_WAIT_FOR_CHUNK_TIMEOUT_MS,
)
.unwrap();
let (response_recv_endpoint, response_recv_thread) =
solana_streamer::quic::spawn_server::<{ TransactionPacket::DATA_SIZE }>(
response_recv_socket,
&keypair2,
response_recv_ip,
sender2,
response_recv_exit.clone(),
1,
staked_nodes,
10,
10,
response_recv_stats,
DEFAULT_WAIT_FOR_CHUNK_TIMEOUT_MS,
)
.unwrap();

let connection_cache = ConnectionCache::new_with_endpoint(1, response_recv_endpoint);

Expand Down
8 changes: 2 additions & 6 deletions core/src/ancestor_hashes_service.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1224,9 +1224,7 @@ mod test {
.recv_timeout(Duration::from_millis(10_000))
.unwrap();
let packet = &mut response_packet[0];
packet
.meta_mut()
.set_socket_addr(&responder_info.serve_repair);
packet.meta.set_socket_addr(&responder_info.serve_repair);
let decision = AncestorHashesService::verify_and_process_ancestor_response(
packet,
&ancestor_hashes_request_statuses,
Expand Down Expand Up @@ -1603,9 +1601,7 @@ mod test {
.recv_timeout(Duration::from_millis(10_000))
.unwrap();
let packet = &mut response_packet[0];
packet
.meta_mut()
.set_socket_addr(&responder_info.serve_repair);
packet.meta.set_socket_addr(&responder_info.serve_repair);
let decision = AncestorHashesService::verify_and_process_ancestor_response(
packet,
&ancestor_hashes_request_statuses,
Expand Down
9 changes: 6 additions & 3 deletions ledger/src/sigverify_shreds.rs
Original file line number Diff line number Diff line change
Expand Up @@ -148,7 +148,7 @@ where

// Recovers merkle roots from shreds binary.
fn get_merkle_roots(
packets: &[PacketBatch],
packets: &[PacketBatch<{ Packet::DATA_SIZE }>],
recycler_cache: &RecyclerCache,
) -> (
PinnedVec<u8>, // Merkle roots
Expand All @@ -159,7 +159,7 @@ fn get_merkle_roots(
.par_iter()
.flat_map(|packets| {
packets.par_iter().map(|packet| {
if packet.meta().discard() {
if packet.meta.discard() {
return None;
}
let shred = shred::layout::get_shred(packet)?;
Expand Down Expand Up @@ -814,7 +814,10 @@ mod tests {
shreds
}

fn make_packets<R: Rng>(rng: &mut R, shreds: &[Shred]) -> Vec<PacketBatch> {
fn make_packets<R: Rng>(
rng: &mut R,
shreds: &[Shred],
) -> Vec<PacketBatch<{ Packet::DATA_SIZE }>> {
let mut packets = shreds.iter().map(|shred| {
let mut packet = Packet::default();
shred.copy_to_packet(&mut packet);
Expand Down
2 changes: 1 addition & 1 deletion streamer/src/nonblocking/quic.rs
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@ pub fn spawn_server<const N: usize>(
}

pub async fn run_server<const N: usize>(
incoming: Incoming,
incoming: Endpoint,
packet_sender: Sender<PacketBatch<N>>,
exit: Arc<AtomicBool>,
max_connections_per_peer: usize,
Expand Down

0 comments on commit 41653fe

Please sign in to comment.