mirror of
https://github.com/neondatabase/neon.git
synced 2026-01-13 08:22:55 +00:00
145 lines
4.4 KiB
Rust
145 lines
4.4 KiB
Rust
pub mod basebackup;
|
|
pub mod config;
|
|
pub mod http;
|
|
pub mod import_datadir;
|
|
pub mod keyspace;
|
|
pub mod layered_repository;
|
|
pub mod page_cache;
|
|
pub mod page_service;
|
|
pub mod pgdatadir_mapping;
|
|
pub mod profiling;
|
|
pub mod reltag;
|
|
pub mod repository;
|
|
pub mod storage_sync;
|
|
pub mod tenant_config;
|
|
pub mod tenant_mgr;
|
|
pub mod tenant_tasks;
|
|
pub mod thread_mgr;
|
|
pub mod timelines;
|
|
pub mod virtual_file;
|
|
pub mod walingest;
|
|
pub mod walreceiver;
|
|
pub mod walrecord;
|
|
pub mod walredo;
|
|
|
|
use once_cell::sync::Lazy;
|
|
use tracing::info;
|
|
|
|
use crate::thread_mgr::ThreadKind;
|
|
use metrics::{register_int_gauge_vec, IntGaugeVec};
|
|
|
|
use pgdatadir_mapping::DatadirTimeline;
|
|
|
|
/// Current storage format version
|
|
///
|
|
/// This is embedded in the metadata file, and also in the header of all the
|
|
/// layer files. If you make any backwards-incompatible changes to the storage
|
|
/// format, bump this!
|
|
pub const STORAGE_FORMAT_VERSION: u16 = 3;
|
|
|
|
// Magic constants used to identify different kinds of files
|
|
pub const IMAGE_FILE_MAGIC: u16 = 0x5A60;
|
|
pub const DELTA_FILE_MAGIC: u16 = 0x5A61;
|
|
|
|
static LIVE_CONNECTIONS_COUNT: Lazy<IntGaugeVec> = Lazy::new(|| {
|
|
register_int_gauge_vec!(
|
|
"pageserver_live_connections",
|
|
"Number of live network connections",
|
|
&["pageserver_connection_kind"]
|
|
)
|
|
.expect("failed to define a metric")
|
|
});
|
|
|
|
pub const LOG_FILE_NAME: &str = "pageserver.log";
|
|
|
|
/// Config for the Repository checkpointer
|
|
#[derive(Debug, Clone, Copy)]
|
|
pub enum CheckpointConfig {
|
|
// Flush all in-memory data
|
|
Flush,
|
|
// Flush all in-memory data and reconstruct all page images
|
|
Forced,
|
|
}
|
|
|
|
pub fn shutdown_pageserver(exit_code: i32) {
|
|
// Shut down the libpq endpoint thread. This prevents new connections from
|
|
// being accepted.
|
|
thread_mgr::shutdown_threads(Some(ThreadKind::LibpqEndpointListener), None, None);
|
|
|
|
// Shut down any page service threads.
|
|
thread_mgr::shutdown_threads(Some(ThreadKind::PageRequestHandler), None, None);
|
|
|
|
// Shut down all the tenants. This flushes everything to disk and kills
|
|
// the checkpoint and GC threads.
|
|
tenant_mgr::shutdown_all_tenants();
|
|
|
|
// Stop syncing with remote storage.
|
|
//
|
|
// FIXME: Does this wait for the sync thread to finish syncing what's queued up?
|
|
// Should it?
|
|
thread_mgr::shutdown_threads(Some(ThreadKind::StorageSync), None, None);
|
|
|
|
// Shut down the HTTP endpoint last, so that you can still check the server's
|
|
// status while it's shutting down.
|
|
thread_mgr::shutdown_threads(Some(ThreadKind::HttpEndpointListener), None, None);
|
|
|
|
// There should be nothing left, but let's be sure
|
|
thread_mgr::shutdown_threads(None, None, None);
|
|
|
|
info!("Shut down successfully completed");
|
|
std::process::exit(exit_code);
|
|
}
|
|
|
|
const DEFAULT_BASE_BACKOFF_SECONDS: f64 = 0.1;
|
|
const DEFAULT_MAX_BACKOFF_SECONDS: f64 = 3.0;
|
|
|
|
async fn exponential_backoff(n: u32, base_increment: f64, max_seconds: f64) {
|
|
let backoff_duration_seconds =
|
|
exponential_backoff_duration_seconds(n, base_increment, max_seconds);
|
|
if backoff_duration_seconds > 0.0 {
|
|
info!(
|
|
"Backoff: waiting {backoff_duration_seconds} seconds before processing with the task",
|
|
);
|
|
tokio::time::sleep(std::time::Duration::from_secs_f64(backoff_duration_seconds)).await;
|
|
}
|
|
}
|
|
|
|
fn exponential_backoff_duration_seconds(n: u32, base_increment: f64, max_seconds: f64) -> f64 {
|
|
if n == 0 {
|
|
0.0
|
|
} else {
|
|
(1.0 + base_increment).powf(f64::from(n)).min(max_seconds)
|
|
}
|
|
}
|
|
|
|
#[cfg(test)]
|
|
mod backoff_defaults_tests {
|
|
use super::*;
|
|
|
|
#[test]
|
|
fn backoff_defaults_produce_growing_backoff_sequence() {
|
|
let mut current_backoff_value = None;
|
|
|
|
for i in 0..10_000 {
|
|
let new_backoff_value = exponential_backoff_duration_seconds(
|
|
i,
|
|
DEFAULT_BASE_BACKOFF_SECONDS,
|
|
DEFAULT_MAX_BACKOFF_SECONDS,
|
|
);
|
|
|
|
if let Some(old_backoff_value) = current_backoff_value.replace(new_backoff_value) {
|
|
assert!(
|
|
old_backoff_value <= new_backoff_value,
|
|
"{i}th backoff value {new_backoff_value} is smaller than the previous one {old_backoff_value}"
|
|
)
|
|
}
|
|
}
|
|
|
|
assert_eq!(
|
|
current_backoff_value.expect("Should have produced backoff values to compare"),
|
|
DEFAULT_MAX_BACKOFF_SECONDS,
|
|
"Given big enough of retries, backoff should reach its allowed max value"
|
|
);
|
|
}
|
|
}
|