Use PohRecorder to synchronize instead of rotate. (#3080)
This commit is contained in:
committed by
GitHub
parent
2ec9bc9f05
commit
1654199b23
196
core/src/tpu.rs
196
core/src/tpu.rs
@@ -1,29 +1,22 @@
|
||||
//! The `tpu` module implements the Transaction Processing Unit, a
|
||||
//! multi-stage transaction processing pipeline in software.
|
||||
|
||||
use crate::banking_stage::{BankingStage, UnprocessedPackets};
|
||||
use crate::banking_stage::BankingStage;
|
||||
use crate::blocktree::Blocktree;
|
||||
use crate::broadcast_stage::BroadcastStage;
|
||||
use crate::cluster_info::ClusterInfo;
|
||||
use crate::cluster_info_vote_listener::ClusterInfoVoteListener;
|
||||
use crate::fetch_stage::FetchStage;
|
||||
use crate::poh_recorder::PohRecorder;
|
||||
use crate::poh_recorder::{PohRecorder, WorkingBankEntries};
|
||||
use crate::service::Service;
|
||||
use crate::sigverify_stage::SigVerifyStage;
|
||||
use crate::tpu_forwarder::TpuForwarder;
|
||||
use solana_runtime::bank::Bank;
|
||||
use solana_sdk::pubkey::Pubkey;
|
||||
use std::net::UdpSocket;
|
||||
use std::sync::atomic::{AtomicBool, Ordering};
|
||||
use std::sync::mpsc::channel;
|
||||
use std::sync::mpsc::{channel, Receiver};
|
||||
use std::sync::{Arc, Mutex, RwLock};
|
||||
use std::thread;
|
||||
|
||||
pub enum TpuMode {
|
||||
Leader(LeaderServices),
|
||||
Forwarder(ForwarderServices),
|
||||
}
|
||||
|
||||
pub struct LeaderServices {
|
||||
fetch_stage: FetchStage,
|
||||
sigverify_stage: SigVerifyStage,
|
||||
@@ -49,7 +42,7 @@ impl LeaderServices {
|
||||
}
|
||||
}
|
||||
|
||||
fn exit(&self) {
|
||||
pub fn exit(&self) {
|
||||
self.fetch_stage.close();
|
||||
}
|
||||
|
||||
@@ -67,187 +60,63 @@ impl LeaderServices {
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn close(self) -> thread::Result<()> {
|
||||
self.exit();
|
||||
self.join()
|
||||
}
|
||||
}
|
||||
|
||||
pub struct ForwarderServices {
|
||||
tpu_forwarder: TpuForwarder,
|
||||
}
|
||||
|
||||
impl ForwarderServices {
|
||||
fn new(tpu_forwarder: TpuForwarder) -> Self {
|
||||
ForwarderServices { tpu_forwarder }
|
||||
}
|
||||
|
||||
fn exit(&self) {
|
||||
self.tpu_forwarder.close();
|
||||
}
|
||||
|
||||
fn join(self) -> thread::Result<()> {
|
||||
self.tpu_forwarder.join()
|
||||
}
|
||||
|
||||
fn close(self) -> thread::Result<()> {
|
||||
pub fn close(self) -> thread::Result<()> {
|
||||
self.exit();
|
||||
self.join()
|
||||
}
|
||||
}
|
||||
|
||||
pub struct Tpu {
|
||||
tpu_mode: Option<TpuMode>,
|
||||
leader_services: LeaderServices,
|
||||
exit: Arc<AtomicBool>,
|
||||
id: Pubkey,
|
||||
cluster_info: Arc<RwLock<ClusterInfo>>,
|
||||
pub id: Pubkey,
|
||||
}
|
||||
|
||||
impl Tpu {
|
||||
pub fn new(id: Pubkey, cluster_info: &Arc<RwLock<ClusterInfo>>) -> Self {
|
||||
Self {
|
||||
tpu_mode: None,
|
||||
exit: Arc::new(AtomicBool::new(false)),
|
||||
id,
|
||||
cluster_info: cluster_info.clone(),
|
||||
}
|
||||
}
|
||||
|
||||
fn mode_exit(&mut self) {
|
||||
match &mut self.tpu_mode {
|
||||
Some(TpuMode::Leader(svcs)) => {
|
||||
svcs.exit();
|
||||
}
|
||||
Some(TpuMode::Forwarder(svcs)) => {
|
||||
svcs.exit();
|
||||
}
|
||||
None => (),
|
||||
}
|
||||
}
|
||||
|
||||
fn mode_close(&mut self) {
|
||||
let tpu_mode = self.tpu_mode.take();
|
||||
if let Some(tpu_mode) = tpu_mode {
|
||||
match tpu_mode {
|
||||
TpuMode::Leader(svcs) => {
|
||||
let _ = svcs.close();
|
||||
}
|
||||
TpuMode::Forwarder(svcs) => {
|
||||
let _ = svcs.close();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
fn forward_unprocessed_packets(
|
||||
tpu: &std::net::SocketAddr,
|
||||
unprocessed_packets: UnprocessedPackets,
|
||||
) -> std::io::Result<()> {
|
||||
let socket = UdpSocket::bind("0.0.0.0:0")?;
|
||||
for (packets, start_index) in unprocessed_packets {
|
||||
let packets = packets.read().unwrap();
|
||||
for packet in packets.packets.iter().skip(start_index) {
|
||||
socket.send_to(&packet.data[..packet.meta.size], tpu)?;
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
fn close_and_forward_unprocessed_packets(&mut self) {
|
||||
self.mode_exit();
|
||||
|
||||
let unprocessed_packets = match self.tpu_mode.as_mut() {
|
||||
Some(TpuMode::Leader(svcs)) => {
|
||||
svcs.banking_stage.join_and_collect_unprocessed_packets()
|
||||
}
|
||||
Some(TpuMode::Forwarder(svcs)) => {
|
||||
svcs.tpu_forwarder.join_and_collect_unprocessed_packets()
|
||||
}
|
||||
None => vec![],
|
||||
};
|
||||
|
||||
if !unprocessed_packets.is_empty() {
|
||||
let tpu = self.cluster_info.read().unwrap().leader_data().unwrap().tpu;
|
||||
info!("forwarding unprocessed packets to new leader at {:?}", tpu);
|
||||
Tpu::forward_unprocessed_packets(&tpu, unprocessed_packets).unwrap_or_else(|err| {
|
||||
warn!("Failed to forward unprocessed transactions: {:?}", err)
|
||||
});
|
||||
}
|
||||
|
||||
self.mode_close();
|
||||
}
|
||||
|
||||
pub fn switch_to_forwarder(&mut self, leader_id: Pubkey, transactions_sockets: Vec<UdpSocket>) {
|
||||
self.close_and_forward_unprocessed_packets();
|
||||
|
||||
self.cluster_info.write().unwrap().set_leader(leader_id);
|
||||
|
||||
let tpu_forwarder = TpuForwarder::new(transactions_sockets, self.cluster_info.clone());
|
||||
self.tpu_mode = Some(TpuMode::Forwarder(ForwarderServices::new(tpu_forwarder)));
|
||||
}
|
||||
|
||||
#[allow(clippy::too_many_arguments)]
|
||||
pub fn switch_to_leader(
|
||||
&mut self,
|
||||
bank: &Arc<Bank>,
|
||||
pub fn new(
|
||||
id: Pubkey,
|
||||
cluster_info: &Arc<RwLock<ClusterInfo>>,
|
||||
poh_recorder: &Arc<Mutex<PohRecorder>>,
|
||||
entry_receiver: Receiver<WorkingBankEntries>,
|
||||
transactions_sockets: Vec<UdpSocket>,
|
||||
broadcast_socket: UdpSocket,
|
||||
sigverify_disabled: bool,
|
||||
slot: u64,
|
||||
blocktree: &Arc<Blocktree>,
|
||||
) {
|
||||
self.close_and_forward_unprocessed_packets();
|
||||
) -> Self {
|
||||
cluster_info.write().unwrap().set_leader(id);
|
||||
|
||||
self.cluster_info.write().unwrap().set_leader(self.id);
|
||||
|
||||
self.exit = Arc::new(AtomicBool::new(false));
|
||||
let exit = Arc::new(AtomicBool::new(false));
|
||||
let (packet_sender, packet_receiver) = channel();
|
||||
let fetch_stage = FetchStage::new_with_sender(
|
||||
transactions_sockets,
|
||||
self.exit.clone(),
|
||||
&packet_sender.clone(),
|
||||
);
|
||||
let cluster_info_vote_listener = ClusterInfoVoteListener::new(
|
||||
self.exit.clone(),
|
||||
self.cluster_info.clone(),
|
||||
packet_sender,
|
||||
);
|
||||
let fetch_stage =
|
||||
FetchStage::new_with_sender(transactions_sockets, exit.clone(), &packet_sender.clone());
|
||||
let cluster_info_vote_listener =
|
||||
ClusterInfoVoteListener::new(exit.clone(), cluster_info.clone(), packet_sender);
|
||||
|
||||
let (sigverify_stage, verified_receiver) =
|
||||
SigVerifyStage::new(packet_receiver, sigverify_disabled);
|
||||
|
||||
// TODO: Fix BankingStage/BroadcastStage to operate on `slot` directly instead of
|
||||
// `max_tick_height`
|
||||
let max_tick_height = (slot + 1) * bank.ticks_per_slot() - 1;
|
||||
let blob_index = blocktree
|
||||
.meta(slot)
|
||||
.expect("Database error")
|
||||
.map(|meta| meta.consumed)
|
||||
.unwrap_or(0);
|
||||
|
||||
let (banking_stage, entry_receiver) =
|
||||
BankingStage::new(&bank, poh_recorder, verified_receiver, max_tick_height);
|
||||
let banking_stage = BankingStage::new(&cluster_info, poh_recorder, verified_receiver);
|
||||
|
||||
let broadcast_stage = BroadcastStage::new(
|
||||
slot,
|
||||
bank,
|
||||
broadcast_socket,
|
||||
self.cluster_info.clone(),
|
||||
blob_index,
|
||||
cluster_info.clone(),
|
||||
entry_receiver,
|
||||
self.exit.clone(),
|
||||
exit.clone(),
|
||||
blocktree,
|
||||
);
|
||||
|
||||
let svcs = LeaderServices::new(
|
||||
let leader_services = LeaderServices::new(
|
||||
fetch_stage,
|
||||
sigverify_stage,
|
||||
banking_stage,
|
||||
cluster_info_vote_listener,
|
||||
broadcast_stage,
|
||||
);
|
||||
self.tpu_mode = Some(TpuMode::Leader(svcs));
|
||||
Self {
|
||||
leader_services,
|
||||
exit,
|
||||
id,
|
||||
}
|
||||
}
|
||||
|
||||
pub fn exit(&self) {
|
||||
@@ -258,8 +127,8 @@ impl Tpu {
|
||||
self.exit.load(Ordering::Relaxed)
|
||||
}
|
||||
|
||||
pub fn close(mut self) -> thread::Result<()> {
|
||||
self.mode_close();
|
||||
pub fn close(self) -> thread::Result<()> {
|
||||
self.exit();
|
||||
self.join()
|
||||
}
|
||||
}
|
||||
@@ -268,11 +137,6 @@ impl Service for Tpu {
|
||||
type JoinReturnType = ();
|
||||
|
||||
fn join(self) -> thread::Result<()> {
|
||||
match self.tpu_mode {
|
||||
Some(TpuMode::Leader(svcs)) => svcs.join()?,
|
||||
Some(TpuMode::Forwarder(svcs)) => svcs.join()?,
|
||||
None => (),
|
||||
}
|
||||
Ok(())
|
||||
self.leader_services.join()
|
||||
}
|
||||
}
|
||||
|
Reference in New Issue
Block a user