mirror of
https://github.com/neondatabase/neon.git
synced 2026-01-06 04:52:55 +00:00
418 lines
14 KiB
Rust
418 lines
14 KiB
Rust
use anyhow::{Context, Result};
|
|
use tokio::task::JoinHandle;
|
|
|
|
use std::cmp::min;
|
|
use std::collections::HashMap;
|
|
use std::path::{Path, PathBuf};
|
|
use std::sync::Arc;
|
|
use std::time::Duration;
|
|
|
|
use postgres_ffi::xlog_utils::{XLogFileName, XLogSegNo, XLogSegNoOffsetToRecPtr, PG_TLI};
|
|
use remote_storage::{GenericRemoteStorage, RemoteStorage};
|
|
use tokio::fs::File;
|
|
use tokio::runtime::Builder;
|
|
|
|
use tokio::select;
|
|
use tokio::sync::mpsc::{self, Receiver, Sender};
|
|
use tokio::sync::watch;
|
|
use tokio::time::sleep;
|
|
use tracing::*;
|
|
|
|
use utils::{lsn::Lsn, zid::ZTenantTimelineId};
|
|
|
|
use crate::broker::{Election, ElectionLeader};
|
|
use crate::timeline::{GlobalTimelines, Timeline};
|
|
use crate::{broker, SafeKeeperConf};
|
|
|
|
use once_cell::sync::OnceCell;
|
|
|
|
const BACKUP_ELECTION_NAME: &str = "WAL_BACKUP";
|
|
|
|
const BROKER_CONNECTION_RETRY_DELAY_MS: u64 = 1000;
|
|
|
|
const UPLOAD_FAILURE_RETRY_MIN_MS: u64 = 10;
|
|
const UPLOAD_FAILURE_RETRY_MAX_MS: u64 = 5000;
|
|
|
|
pub fn wal_backup_launcher_thread_main(
|
|
conf: SafeKeeperConf,
|
|
wal_backup_launcher_rx: Receiver<ZTenantTimelineId>,
|
|
) {
|
|
let rt = Builder::new_multi_thread()
|
|
.worker_threads(conf.backup_runtime_threads)
|
|
.enable_all()
|
|
.build()
|
|
.expect("failed to create wal backup runtime");
|
|
|
|
rt.block_on(async {
|
|
wal_backup_launcher_main_loop(conf, wal_backup_launcher_rx).await;
|
|
});
|
|
}
|
|
|
|
/// Check whether wal backup is required for timeline and mark that launcher is
|
|
/// aware of current status (if timeline exists).
|
|
fn is_wal_backup_required(zttid: ZTenantTimelineId) -> bool {
|
|
if let Some(tli) = GlobalTimelines::get_loaded(zttid) {
|
|
tli.wal_backup_attend()
|
|
} else {
|
|
false
|
|
}
|
|
}
|
|
|
|
struct WalBackupTaskHandle {
|
|
shutdown_tx: Sender<()>,
|
|
handle: JoinHandle<()>,
|
|
}
|
|
|
|
/// Sits on wal_backup_launcher_rx and starts/stops per timeline wal backup
|
|
/// tasks. Having this in separate task simplifies locking, allows to reap
|
|
/// panics and separate elections from offloading itself.
|
|
async fn wal_backup_launcher_main_loop(
|
|
conf: SafeKeeperConf,
|
|
mut wal_backup_launcher_rx: Receiver<ZTenantTimelineId>,
|
|
) {
|
|
info!(
|
|
"WAL backup launcher: started, remote config {:?}",
|
|
conf.remote_storage
|
|
);
|
|
|
|
let conf_ = conf.clone();
|
|
REMOTE_STORAGE.get_or_init(|| {
|
|
conf_.remote_storage.as_ref().map(|c| {
|
|
GenericRemoteStorage::new(conf_.workdir, c).expect("failed to create remote storage")
|
|
})
|
|
});
|
|
|
|
let mut tasks: HashMap<ZTenantTimelineId, WalBackupTaskHandle> = HashMap::new();
|
|
|
|
loop {
|
|
// channel is never expected to get closed
|
|
let zttid = wal_backup_launcher_rx.recv().await.unwrap();
|
|
let is_wal_backup_required = is_wal_backup_required(zttid);
|
|
if conf.remote_storage.is_none() || !conf.wal_backup_enabled {
|
|
continue; /* just drain the channel and do nothing */
|
|
}
|
|
// do we need to do anything at all?
|
|
if is_wal_backup_required != tasks.contains_key(&zttid) {
|
|
if is_wal_backup_required {
|
|
// need to start the task
|
|
info!("starting WAL backup task for {}", zttid);
|
|
|
|
// TODO: decide who should offload in launcher itself by simply checking current state
|
|
let election_name = broker::get_campaign_name(
|
|
BACKUP_ELECTION_NAME.to_string(),
|
|
conf.broker_etcd_prefix.clone(),
|
|
&zttid,
|
|
);
|
|
let my_candidate_name = broker::get_candiate_name(conf.my_id);
|
|
let election = broker::Election::new(
|
|
election_name,
|
|
my_candidate_name,
|
|
conf.broker_endpoints.clone(),
|
|
);
|
|
|
|
let (shutdown_tx, shutdown_rx) = mpsc::channel(1);
|
|
let timeline_dir = conf.timeline_dir(&zttid);
|
|
|
|
let handle = tokio::spawn(
|
|
backup_task_main(zttid, timeline_dir, shutdown_rx, election)
|
|
.instrument(info_span!("WAL backup task", zttid = %zttid)),
|
|
);
|
|
|
|
tasks.insert(
|
|
zttid,
|
|
WalBackupTaskHandle {
|
|
shutdown_tx,
|
|
handle,
|
|
},
|
|
);
|
|
} else {
|
|
// need to stop the task
|
|
info!("stopping WAL backup task for {}", zttid);
|
|
|
|
let wb_handle = tasks.remove(&zttid).unwrap();
|
|
// Tell the task to shutdown. Error means task exited earlier, that's ok.
|
|
let _ = wb_handle.shutdown_tx.send(()).await;
|
|
// Await the task itself. TODO: restart panicked tasks earlier.
|
|
// Hm, why I can't await on reference to handle?
|
|
if let Err(e) = wb_handle.handle.await {
|
|
warn!("WAL backup task for {} panicked: {}", zttid, e);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
struct WalBackupTask {
|
|
timeline: Arc<Timeline>,
|
|
timeline_dir: PathBuf,
|
|
wal_seg_size: usize,
|
|
commit_lsn_watch_rx: watch::Receiver<Lsn>,
|
|
leader: Option<ElectionLeader>,
|
|
election: Election,
|
|
}
|
|
|
|
/// Offload single timeline.
|
|
async fn backup_task_main(
|
|
zttid: ZTenantTimelineId,
|
|
timeline_dir: PathBuf,
|
|
mut shutdown_rx: Receiver<()>,
|
|
election: Election,
|
|
) {
|
|
info!("started");
|
|
let timeline: Arc<Timeline> = if let Some(tli) = GlobalTimelines::get_loaded(zttid) {
|
|
tli
|
|
} else {
|
|
/* Timeline could get deleted while task was starting, just exit then. */
|
|
info!("no timeline, exiting");
|
|
return;
|
|
};
|
|
|
|
let mut wb = WalBackupTask {
|
|
wal_seg_size: timeline.get_wal_seg_size(),
|
|
commit_lsn_watch_rx: timeline.get_commit_lsn_watch_rx(),
|
|
timeline,
|
|
timeline_dir,
|
|
leader: None,
|
|
election,
|
|
};
|
|
|
|
// task is spinned up only when wal_seg_size already initialized
|
|
assert!(wb.wal_seg_size > 0);
|
|
|
|
let mut canceled = false;
|
|
select! {
|
|
_ = wb.run() => {}
|
|
_ = shutdown_rx.recv() => {
|
|
canceled = true;
|
|
}
|
|
}
|
|
if let Some(l) = wb.leader {
|
|
l.give_up().await;
|
|
}
|
|
info!("task {}", if canceled { "canceled" } else { "terminated" });
|
|
}
|
|
|
|
impl WalBackupTask {
|
|
async fn run(&mut self) {
|
|
let mut backup_lsn = Lsn(0);
|
|
|
|
// election loop
|
|
loop {
|
|
let mut retry_attempt = 0u32;
|
|
|
|
if let Some(l) = self.leader.take() {
|
|
l.give_up().await;
|
|
}
|
|
|
|
match broker::get_leader(&self.election).await {
|
|
Ok(l) => {
|
|
self.leader = Some(l);
|
|
}
|
|
Err(e) => {
|
|
error!("error during leader election {:?}", e);
|
|
sleep(Duration::from_millis(BROKER_CONNECTION_RETRY_DELAY_MS)).await;
|
|
continue;
|
|
}
|
|
}
|
|
|
|
// offload loop
|
|
loop {
|
|
if retry_attempt == 0 {
|
|
// wait for new WAL to arrive
|
|
if let Err(e) = self.commit_lsn_watch_rx.changed().await {
|
|
// should never happen, as we hold Arc to timeline.
|
|
error!("commit_lsn watch shut down: {:?}", e);
|
|
return;
|
|
}
|
|
} else {
|
|
// or just sleep if we errored previously
|
|
let mut retry_delay = UPLOAD_FAILURE_RETRY_MAX_MS;
|
|
if let Some(backoff_delay) =
|
|
UPLOAD_FAILURE_RETRY_MIN_MS.checked_shl(retry_attempt)
|
|
{
|
|
retry_delay = min(retry_delay, backoff_delay);
|
|
}
|
|
sleep(Duration::from_millis(retry_delay)).await;
|
|
}
|
|
|
|
let commit_lsn = *self.commit_lsn_watch_rx.borrow();
|
|
|
|
// Note that backup_lsn can be higher than commit_lsn if we
|
|
// don't have much local WAL and others already uploaded
|
|
// segments we don't even have.
|
|
if backup_lsn.segment_number(self.wal_seg_size)
|
|
>= commit_lsn.segment_number(self.wal_seg_size)
|
|
{
|
|
continue; /* nothing to do, common case as we wake up on every commit_lsn bump */
|
|
}
|
|
// Perhaps peers advanced the position, check shmem value.
|
|
backup_lsn = self.timeline.get_wal_backup_lsn();
|
|
if backup_lsn.segment_number(self.wal_seg_size)
|
|
>= commit_lsn.segment_number(self.wal_seg_size)
|
|
{
|
|
continue;
|
|
}
|
|
|
|
if let Some(l) = self.leader.as_mut() {
|
|
// Optimization idea for later:
|
|
// Avoid checking election leader every time by returning current lease grant expiration time
|
|
// Re-check leadership only after expiration time,
|
|
// such approach would reduce overhead on write-intensive workloads
|
|
|
|
match l
|
|
.check_am_i(
|
|
self.election.election_name.clone(),
|
|
self.election.candidate_name.clone(),
|
|
)
|
|
.await
|
|
{
|
|
Ok(leader) => {
|
|
if !leader {
|
|
info!("leader has changed");
|
|
break;
|
|
}
|
|
}
|
|
Err(e) => {
|
|
warn!("error validating leader, {:?}", e);
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
|
|
match backup_lsn_range(
|
|
backup_lsn,
|
|
commit_lsn,
|
|
self.wal_seg_size,
|
|
&self.timeline_dir,
|
|
)
|
|
.await
|
|
{
|
|
Ok(backup_lsn_result) => {
|
|
backup_lsn = backup_lsn_result;
|
|
self.timeline.set_wal_backup_lsn(backup_lsn_result);
|
|
retry_attempt = 0;
|
|
}
|
|
Err(e) => {
|
|
error!(
|
|
"failed while offloading range {}-{}: {:?}",
|
|
backup_lsn, commit_lsn, e
|
|
);
|
|
|
|
retry_attempt = min(retry_attempt + 1, u32::MAX);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
pub async fn backup_lsn_range(
|
|
start_lsn: Lsn,
|
|
end_lsn: Lsn,
|
|
wal_seg_size: usize,
|
|
timeline_dir: &Path,
|
|
) -> Result<Lsn> {
|
|
let mut res = start_lsn;
|
|
let segments = get_segments(start_lsn, end_lsn, wal_seg_size);
|
|
for s in &segments {
|
|
backup_single_segment(s, timeline_dir)
|
|
.await
|
|
.with_context(|| format!("offloading segno {}", s.seg_no))?;
|
|
|
|
res = s.end_lsn;
|
|
}
|
|
info!(
|
|
"offloaded segnos {:?} up to {}, previous backup_lsn {}",
|
|
segments.iter().map(|&s| s.seg_no).collect::<Vec<_>>(),
|
|
end_lsn,
|
|
start_lsn,
|
|
);
|
|
Ok(res)
|
|
}
|
|
|
|
async fn backup_single_segment(seg: &Segment, timeline_dir: &Path) -> Result<()> {
|
|
let segment_file_name = seg.file_path(timeline_dir)?;
|
|
|
|
backup_object(&segment_file_name, seg.size()).await?;
|
|
debug!("Backup of {} done", segment_file_name.display());
|
|
|
|
Ok(())
|
|
}
|
|
|
|
#[derive(Debug, Copy, Clone)]
|
|
pub struct Segment {
|
|
seg_no: XLogSegNo,
|
|
start_lsn: Lsn,
|
|
end_lsn: Lsn,
|
|
}
|
|
|
|
impl Segment {
|
|
pub fn new(seg_no: u64, start_lsn: Lsn, end_lsn: Lsn) -> Self {
|
|
Self {
|
|
seg_no,
|
|
start_lsn,
|
|
end_lsn,
|
|
}
|
|
}
|
|
|
|
pub fn object_name(self) -> String {
|
|
XLogFileName(PG_TLI, self.seg_no, self.size())
|
|
}
|
|
|
|
pub fn file_path(self, timeline_dir: &Path) -> Result<PathBuf> {
|
|
Ok(timeline_dir.join(self.object_name()))
|
|
}
|
|
|
|
pub fn size(self) -> usize {
|
|
(u64::from(self.end_lsn) - u64::from(self.start_lsn)) as usize
|
|
}
|
|
}
|
|
|
|
fn get_segments(start: Lsn, end: Lsn, seg_size: usize) -> Vec<Segment> {
|
|
let first_seg = start.segment_number(seg_size);
|
|
let last_seg = end.segment_number(seg_size);
|
|
|
|
let res: Vec<Segment> = (first_seg..last_seg)
|
|
.map(|s| {
|
|
let start_lsn = XLogSegNoOffsetToRecPtr(s, 0, seg_size);
|
|
let end_lsn = XLogSegNoOffsetToRecPtr(s + 1, 0, seg_size);
|
|
Segment::new(s, Lsn::from(start_lsn), Lsn::from(end_lsn))
|
|
})
|
|
.collect();
|
|
res
|
|
}
|
|
|
|
static REMOTE_STORAGE: OnceCell<Option<GenericRemoteStorage>> = OnceCell::new();
|
|
|
|
async fn backup_object(source_file: &Path, size: usize) -> Result<()> {
|
|
let storage = REMOTE_STORAGE.get().expect("failed to get remote storage");
|
|
|
|
let file = File::open(&source_file).await?;
|
|
|
|
// Storage is initialized by launcher at this point.
|
|
match storage.as_ref().unwrap() {
|
|
GenericRemoteStorage::Local(local_storage) => {
|
|
let destination = local_storage.remote_object_id(source_file)?;
|
|
|
|
debug!(
|
|
"local upload about to start from {} to {}",
|
|
source_file.display(),
|
|
destination.display()
|
|
);
|
|
local_storage.upload(file, size, &destination, None).await
|
|
}
|
|
GenericRemoteStorage::S3(s3_storage) => {
|
|
let s3key = s3_storage.remote_object_id(source_file)?;
|
|
|
|
debug!(
|
|
"S3 upload about to start from {} to {:?}",
|
|
source_file.display(),
|
|
s3key
|
|
);
|
|
s3_storage.upload(file, size, &s3key, None).await
|
|
}
|
|
}?;
|
|
|
|
Ok(())
|
|
}
|