repo
stringlengths
6
65
file_url
stringlengths
81
311
file_path
stringlengths
6
227
content
stringlengths
0
32.8k
language
stringclasses
1 value
license
stringclasses
7 values
commit_sha
stringlengths
40
40
retrieved_at
stringdate
2026-01-04 15:31:58
2026-01-04 20:25:31
truncated
bool
2 classes
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/bin/compute_ctl.rs
compute_tools/src/bin/compute_ctl.rs
//! //! Postgres wrapper (`compute_ctl`) is intended to be run as a Docker entrypoint or as a `systemd` //! `ExecStart` option. It will handle all the `Neon` specifics during compute node //! initialization: //! - `compute_ctl` accepts cluster (compute node) specification as a JSON file. //! - Every start is a fresh start, so the data directory is removed and //! initialized again on each run. //! - If remote_extension_config is provided, it will be used to fetch extensions list //! and download `shared_preload_libraries` from the remote storage. //! - Next it will put configuration files into the `PGDATA` directory. //! - Sync safekeepers and get commit LSN. //! - Get `basebackup` from pageserver using the returned on the previous step LSN. //! - Try to start `postgres` and wait until it is ready to accept connections. //! - Check and alter/drop/create roles and databases. //! - Hang waiting on the `postmaster` process to exit. //! //! Also `compute_ctl` spawns two separate service threads: //! - `compute-monitor` checks the last Postgres activity timestamp and saves it //! into the shared `ComputeNode`; //! - `http-endpoint` runs a Hyper HTTP API server, which serves readiness and the //! last activity requests. //! //! If `AUTOSCALING` environment variable is set, `compute_ctl` will start the //! `vm-monitor` located in [`neon/libs/vm_monitor`]. For VM compute nodes, //! `vm-monitor` communicates with the VM autoscaling system. It coordinates //! downscaling and requests immediate upscaling under resource pressure. //! //! Usage example: //! ```sh //! compute_ctl -D /var/db/postgres/compute \ //! -C 'postgresql://cloud_admin@localhost/postgres' \ //! -c /var/db/postgres/configs/config.json \ //! -b /usr/local/bin/postgres \ //! -r http://pg-ext-s3-gateway \ //! ``` use std::ffi::OsString; use std::fs::File; use std::process::exit; use std::sync::Arc; use std::sync::atomic::AtomicU64; use std::sync::mpsc; use std::thread; use std::time::Duration; use anyhow::{Context, Result, bail}; use clap::Parser; use compute_api::responses::ComputeConfig; use compute_tools::compute::{ BUILD_TAG, ComputeNode, ComputeNodeParams, forward_termination_signal, }; use compute_tools::extension_server::get_pg_version_string; use compute_tools::params::*; use compute_tools::pg_isready::get_pg_isready_bin; use compute_tools::spec::*; use compute_tools::{hadron_metrics, installed_extensions, logger::*}; use rlimit::{Resource, setrlimit}; use signal_hook::consts::{SIGINT, SIGQUIT, SIGTERM}; use signal_hook::iterator::Signals; use tracing::{error, info}; use url::Url; use utils::failpoint_support; #[derive(Debug, Parser)] #[command(rename_all = "kebab-case")] struct Cli { #[arg(short = 'b', long, default_value = "postgres", env = "POSTGRES_PATH")] pub pgbin: String, /// The base URL for the remote extension storage proxy gateway. #[arg(short = 'r', long, value_parser = Self::parse_remote_ext_base_url)] pub remote_ext_base_url: Option<Url>, /// The port to bind the external listening HTTP server to. Clients running /// outside the compute will talk to the compute through this port. Keep /// the previous name for this argument around for a smoother release /// with the control plane. #[arg(long, default_value_t = 3080)] pub external_http_port: u16, /// The port to bind the internal listening HTTP server to. Clients include /// the neon extension (for installing remote extensions) and local_proxy. #[arg(long, default_value_t = 3081)] pub internal_http_port: u16, /// Backwards-compatible --http-port for Hadron deployments. Functionally the /// same as --external-http-port. #[arg( long, conflicts_with = "external_http_port", conflicts_with = "internal_http_port" )] pub http_port: Option<u16>, #[arg(short = 'D', long, value_name = "DATADIR")] pub pgdata: String, #[arg(short = 'C', long, value_name = "DATABASE_URL")] pub connstr: String, #[arg( long, default_value = "neon_superuser", value_name = "PRIVILEGED_ROLE_NAME", value_parser = Self::parse_privileged_role_name )] pub privileged_role_name: String, #[cfg(target_os = "linux")] #[arg(long, default_value = "neon-postgres")] pub cgroup: String, #[cfg(target_os = "linux")] #[arg( long, default_value = "host=localhost port=5432 dbname=postgres user=cloud_admin sslmode=disable application_name=vm-monitor" )] pub filecache_connstr: String, #[cfg(target_os = "linux")] #[arg(long, default_value = "0.0.0.0:10301")] pub vm_monitor_addr: String, #[arg(long, action = clap::ArgAction::SetTrue)] pub resize_swap_on_bind: bool, #[arg(long)] pub set_disk_quota_for_fs: Option<String>, #[arg(short = 'c', long)] pub config: Option<OsString>, #[arg(short = 'i', long, group = "compute-id")] pub compute_id: String, #[arg( short = 'p', long, conflicts_with = "config", value_name = "CONTROL_PLANE_API_BASE_URL", requires = "compute-id" )] pub control_plane_uri: Option<String>, /// Interval in seconds for collecting installed extensions statistics #[arg(long, default_value = "3600")] pub installed_extensions_collection_interval: u64, /// Run in development mode, skipping VM-specific operations like process termination #[arg(long, action = clap::ArgAction::SetTrue)] pub dev: bool, #[arg(long)] pub pg_init_timeout: Option<u64>, #[arg(long, default_value_t = false, action = clap::ArgAction::Set)] pub lakebase_mode: bool, } impl Cli { /// Parse a URL from an argument. By default, this isn't necessary, but we /// want to do some sanity checking. fn parse_remote_ext_base_url(value: &str) -> Result<Url> { // Remove extra trailing slashes, and add one. We use Url::join() later // when downloading remote extensions. If the base URL is something like // http://example.com/pg-ext-s3-gateway, and join() is called with // something like "xyz", the resulting URL is http://example.com/xyz. let value = value.trim_end_matches('/').to_owned() + "/"; let url = Url::parse(&value)?; if url.query_pairs().count() != 0 { bail!("parameters detected in remote extensions base URL") } Ok(url) } /// For simplicity, we do not escape `privileged_role_name` anywhere in the code. /// Since it's a system role, which we fully control, that's fine. Still, let's /// validate it to avoid any surprises. fn parse_privileged_role_name(value: &str) -> Result<String> { use regex::Regex; let pattern = Regex::new(r"^[a-z_]+$").unwrap(); if !pattern.is_match(value) { bail!("--privileged-role-name can only contain lowercase letters and underscores") } Ok(value.to_string()) } } // Hadron helpers to get compatible compute_ctl http ports from Cli. The old `--http-port` // arg is used and acts the same as `--external-http-port`. The internal http port is defined // to be http_port + 1. Hadron runs in the dblet environment which uses the host network, so // we need to be careful with the ports to choose. fn get_external_http_port(cli: &Cli) -> u16 { if cli.lakebase_mode { return cli.http_port.unwrap_or(cli.external_http_port); } cli.external_http_port } fn get_internal_http_port(cli: &Cli) -> u16 { if cli.lakebase_mode { return cli .http_port .map(|p| p + 1) .unwrap_or(cli.internal_http_port); } cli.internal_http_port } fn main() -> Result<()> { let cli = Cli::parse(); let scenario = failpoint_support::init(); // For historical reasons, the main thread that processes the config and launches postgres // is synchronous, but we always have this tokio runtime available and we "enter" it so // that you can use tokio::spawn() and tokio::runtime::Handle::current().block_on(...) // from all parts of compute_ctl. let runtime = tokio::runtime::Builder::new_multi_thread() .enable_all() .build()?; let _rt_guard = runtime.enter(); let mut log_dir = None; if cli.lakebase_mode { log_dir = std::env::var("COMPUTE_CTL_LOG_DIRECTORY").ok(); } let (tracing_provider, _file_logs_guard) = init(cli.dev, log_dir)?; // enable core dumping for all child processes setrlimit(Resource::CORE, rlimit::INFINITY, rlimit::INFINITY)?; if cli.lakebase_mode { installed_extensions::initialize_metrics(); hadron_metrics::initialize_metrics(); } let connstr = Url::parse(&cli.connstr).context("cannot parse connstr as a URL")?; let config = get_config(&cli)?; let external_http_port = get_external_http_port(&cli); let internal_http_port = get_internal_http_port(&cli); let compute_node = ComputeNode::new( ComputeNodeParams { compute_id: cli.compute_id, connstr, privileged_role_name: cli.privileged_role_name.clone(), pgdata: cli.pgdata.clone(), pgbin: cli.pgbin.clone(), pgversion: get_pg_version_string(&cli.pgbin), external_http_port, internal_http_port, remote_ext_base_url: cli.remote_ext_base_url.clone(), resize_swap_on_bind: cli.resize_swap_on_bind, set_disk_quota_for_fs: cli.set_disk_quota_for_fs, #[cfg(target_os = "linux")] filecache_connstr: cli.filecache_connstr, #[cfg(target_os = "linux")] cgroup: cli.cgroup, #[cfg(target_os = "linux")] vm_monitor_addr: cli.vm_monitor_addr, installed_extensions_collection_interval: Arc::new(AtomicU64::new( cli.installed_extensions_collection_interval, )), pg_init_timeout: cli.pg_init_timeout.map(Duration::from_secs), pg_isready_bin: get_pg_isready_bin(&cli.pgbin), instance_id: std::env::var("INSTANCE_ID").ok(), lakebase_mode: cli.lakebase_mode, build_tag: BUILD_TAG.to_string(), control_plane_uri: cli.control_plane_uri, config_path_test_only: cli.config, }, config, )?; let exit_code = compute_node.run().context("running compute node")?; scenario.teardown(); deinit_and_exit(tracing_provider, exit_code); } fn init( dev_mode: bool, log_dir: Option<String>, ) -> Result<( Option<tracing_utils::Provider>, Option<tracing_appender::non_blocking::WorkerGuard>, )> { let (provider, file_logs_guard) = init_tracing_and_logging(DEFAULT_LOG_LEVEL, &log_dir)?; let mut signals = Signals::new([SIGINT, SIGTERM, SIGQUIT])?; thread::spawn(move || { for sig in signals.forever() { handle_exit_signal(sig, dev_mode); } }); info!("compute build_tag: {}", &BUILD_TAG.to_string()); Ok((provider, file_logs_guard)) } fn get_config(cli: &Cli) -> Result<ComputeConfig> { // First, read the config from the path if provided if let Some(ref config) = cli.config { let file = File::open(config)?; return Ok(serde_json::from_reader(&file)?); } // If the config wasn't provided in the CLI arguments, then retrieve it from // the control plane match get_config_from_control_plane(cli.control_plane_uri.as_ref().unwrap(), &cli.compute_id) { Ok(config) => Ok(config), Err(e) => { error!( "cannot get response from control plane: {}\n\ neither spec nor confirmation that compute is in the Empty state was received", e ); Err(e) } } } fn deinit_and_exit(tracing_provider: Option<tracing_utils::Provider>, exit_code: Option<i32>) -> ! { if let Some(p) = tracing_provider { // Shutdown trace pipeline gracefully, so that it has a chance to send any // pending traces before we exit. Shutting down OTEL tracing provider may // hang for quite some time, see, for example: // - https://github.com/open-telemetry/opentelemetry-rust/issues/868 // - and our problems with staging https://github.com/neondatabase/cloud/issues/3707#issuecomment-1493983636 // // Yet, we want computes to shut down fast enough, as we may need a new one // for the same timeline ASAP. So wait no longer than 2s for the shutdown to // complete, then just error out and exit the main thread. info!("shutting down tracing"); let (sender, receiver) = mpsc::channel(); let _ = thread::spawn(move || { _ = p.shutdown(); sender.send(()).ok() }); let shutdown_res = receiver.recv_timeout(Duration::from_millis(2000)); if shutdown_res.is_err() { error!("timed out while shutting down tracing, exiting anyway"); } } info!("shutting down"); exit(exit_code.unwrap_or(1)) } /// When compute_ctl is killed, send also termination signal to sync-safekeepers /// to prevent leakage. TODO: it is better to convert compute_ctl to async and /// wait for termination which would be easy then. fn handle_exit_signal(sig: i32, dev_mode: bool) { info!("received {sig} termination signal"); forward_termination_signal(dev_mode); exit(1); } #[cfg(test)] mod test { use clap::{CommandFactory, Parser}; use url::Url; use super::Cli; #[test] fn verify_cli() { Cli::command().debug_assert() } #[test] fn verify_remote_ext_base_url() { let cli = Cli::parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--remote-ext-base-url", "https://example.com/subpath", ]); assert_eq!( cli.remote_ext_base_url.unwrap(), Url::parse("https://example.com/subpath/").unwrap() ); let cli = Cli::parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--remote-ext-base-url", "https://example.com//", ]); assert_eq!( cli.remote_ext_base_url.unwrap(), Url::parse("https://example.com").unwrap() ); Cli::try_parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--remote-ext-base-url", "https://example.com?hello=world", ]) .expect_err("URL parameters are not allowed"); } #[test] fn verify_privileged_role_name() { // Valid name let cli = Cli::parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--privileged-role-name", "my_superuser", ]); assert_eq!(cli.privileged_role_name, "my_superuser"); // Invalid names Cli::try_parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--privileged-role-name", "NeonSuperuser", ]) .expect_err("uppercase letters are not allowed"); Cli::try_parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--privileged-role-name", "$'neon_superuser", ]) .expect_err("special characters are not allowed"); Cli::try_parse_from([ "compute_ctl", "--pgdata=test", "--connstr=test", "--compute-id=test", "--privileged-role-name", "", ]) .expect_err("empty name is not allowed"); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/bin/fast_import/s3_uri.rs
compute_tools/src/bin/fast_import/s3_uri.rs
use std::str::FromStr; use anyhow::Result; /// Struct to hold parsed S3 components #[derive(Debug, Clone, PartialEq, Eq)] pub struct S3Uri { pub bucket: String, pub key: String, } impl FromStr for S3Uri { type Err = anyhow::Error; /// Parse an S3 URI into a bucket and key fn from_str(uri: &str) -> Result<Self> { // Ensure the URI starts with "s3://" if !uri.starts_with("s3://") { return Err(anyhow::anyhow!("Invalid S3 URI scheme")); } // Remove the "s3://" prefix let stripped_uri = &uri[5..]; // Split the remaining string into bucket and key parts if let Some((bucket, key)) = stripped_uri.split_once('/') { Ok(S3Uri { bucket: bucket.to_string(), key: key.to_string(), }) } else { Err(anyhow::anyhow!( "Invalid S3 URI format, missing bucket or key" )) } } } impl S3Uri { pub fn append(&self, suffix: &str) -> Self { Self { bucket: self.bucket.clone(), key: format!("{}{}", self.key, suffix), } } } impl std::fmt::Display for S3Uri { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "s3://{}/{}", self.bucket, self.key) } } impl clap::builder::TypedValueParser for S3Uri { type Value = Self; fn parse_ref( &self, _cmd: &clap::Command, _arg: Option<&clap::Arg>, value: &std::ffi::OsStr, ) -> Result<Self::Value, clap::Error> { let value_str = value.to_str().ok_or_else(|| { clap::Error::raw( clap::error::ErrorKind::InvalidUtf8, "Invalid UTF-8 sequence", ) })?; S3Uri::from_str(value_str).map_err(|e| { clap::Error::raw( clap::error::ErrorKind::InvalidValue, format!("Failed to parse S3 URI: {e}"), ) }) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/bin/fast_import/aws_s3_sync.rs
compute_tools/src/bin/fast_import/aws_s3_sync.rs
use camino::{Utf8Path, Utf8PathBuf}; use tokio::task::JoinSet; use tracing::{info, warn}; use walkdir::WalkDir; use super::s3_uri::S3Uri; const MAX_PARALLEL_UPLOADS: usize = 10; /// Upload all files from 'local' to 'remote' pub(crate) async fn upload_dir_recursive( s3_client: &aws_sdk_s3::Client, local: &Utf8Path, remote: &S3Uri, ) -> anyhow::Result<()> { // Recursively scan directory let mut dirwalker = WalkDir::new(local) .into_iter() .map(|entry| { let entry = entry?; let file_type = entry.file_type(); let path = <&Utf8Path>::try_from(entry.path())?.to_path_buf(); Ok((file_type, path)) }) .filter_map(|e: anyhow::Result<(std::fs::FileType, Utf8PathBuf)>| { match e { Ok((file_type, path)) if file_type.is_file() => Some(Ok(path)), Ok((file_type, _path)) if file_type.is_dir() => { // The WalkDir iterator will recurse into directories, but we don't want // to do anything with directories as such. There's no concept of uploading // an empty directory to S3. None } Ok((file_type, path)) if file_type.is_symlink() => { // huh, didn't expect a symlink. Can't upload that to S3. Warn and skip. warn!("cannot upload symlink ({})", path); None } Ok((_file_type, path)) => { // should not happen warn!("directory entry has unexpected type ({})", path); None } Err(e) => Some(Err(e)), } }); // Spawn upload tasks for each file, keeping MAX_PARALLEL_UPLOADS active in // parallel. let mut joinset = JoinSet::new(); loop { // Could we upload more? while joinset.len() < MAX_PARALLEL_UPLOADS { if let Some(full_local_path) = dirwalker.next() { let full_local_path = full_local_path?; let relative_local_path = full_local_path .strip_prefix(local) .expect("all paths start from the walkdir root"); let remote_path = remote.append(relative_local_path.as_str()); info!( "starting upload of {} to {}", &full_local_path, &remote_path ); let upload_task = upload_file(s3_client.clone(), full_local_path, remote_path); joinset.spawn(upload_task); } else { info!("draining upload tasks"); break; } } // Wait for an upload to complete if let Some(res) = joinset.join_next().await { let _ = res?; } else { // all done! break; } } Ok(()) } pub(crate) async fn upload_file( s3_client: aws_sdk_s3::Client, local_path: Utf8PathBuf, remote: S3Uri, ) -> anyhow::Result<()> { use aws_smithy_types::byte_stream::ByteStream; let stream = ByteStream::from_path(&local_path).await?; let _result = s3_client .put_object() .bucket(remote.bucket) .key(&remote.key) .body(stream) .send() .await?; info!("upload of {} to {} finished", &local_path, &remote.key); Ok(()) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/bin/fast_import/child_stdio_to_log.rs
compute_tools/src/bin/fast_import/child_stdio_to_log.rs
use tokio::io::{AsyncBufReadExt, BufReader}; use tokio::process::{ChildStderr, ChildStdout}; use tracing::info; /// Asynchronously relays the output from a child process's `stdout` and `stderr` to the tracing log. /// Each line is read and logged individually, with lossy UTF-8 conversion. /// /// # Arguments /// /// * `stdout`: An `Option<ChildStdout>` from the child process. /// * `stderr`: An `Option<ChildStderr>` from the child process. /// pub(crate) async fn relay_process_output(stdout: Option<ChildStdout>, stderr: Option<ChildStderr>) { let stdout_fut = async { if let Some(stdout) = stdout { let reader = BufReader::new(stdout); let mut lines = reader.lines(); while let Ok(Some(line)) = lines.next_line().await { info!(fd = "stdout", "{}", line); } } }; let stderr_fut = async { if let Some(stderr) = stderr { let reader = BufReader::new(stderr); let mut lines = reader.lines(); while let Ok(Some(line)) = lines.next_line().await { info!(fd = "stderr", "{}", line); } } }; tokio::join!(stdout_fut, stderr_fut); }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/headers.rs
compute_tools/src/http/headers.rs
/// Constant for `X-Request-Id` header. pub const X_REQUEST_ID: &str = "x-request-id";
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/mod.rs
compute_tools/src/http/mod.rs
use axum::body::Body; use axum::response::Response; use compute_api::responses::{ComputeStatus, GenericAPIError}; use http::StatusCode; use http::header::CONTENT_TYPE; use serde::Serialize; use tracing::error; mod extract; mod headers; mod middleware; mod routes; pub mod server; /// Convenience response builder for JSON responses struct JsonResponse; impl JsonResponse { /// Helper for actually creating a response fn create_response(code: StatusCode, body: impl Serialize) -> Response { Response::builder() .status(code) .header(CONTENT_TYPE.as_str(), "application/json") .body(Body::from(serde_json::to_string(&body).unwrap())) .unwrap() } /// Create a successful error response pub(self) fn success(code: StatusCode, body: impl Serialize) -> Response { assert!({ let code = code.as_u16(); (200..300).contains(&code) }); Self::create_response(code, body) } /// Create an error response pub(self) fn error(code: StatusCode, error: impl ToString) -> Response { assert!(code.as_u16() >= 400); let message = error.to_string(); error!(message); Self::create_response(code, &GenericAPIError { error: message }) } /// Create an error response related to the compute being in an invalid state pub(self) fn invalid_status(status: ComputeStatus) -> Response { Self::error( StatusCode::PRECONDITION_FAILED, format!("invalid compute status: {status}"), ) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/server.rs
compute_tools/src/http/server.rs
use std::fmt::Display; use std::net::{IpAddr, Ipv6Addr, SocketAddr}; use std::sync::Arc; use std::time::Duration; use anyhow::Result; use axum::Router; use axum::middleware::{self}; use axum::response::IntoResponse; use axum::routing::{get, post}; use compute_api::responses::ComputeCtlConfig; use http::StatusCode; use tokio::net::TcpListener; use tower::ServiceBuilder; use tower_http::{ auth::AsyncRequireAuthorizationLayer, request_id::PropagateRequestIdLayer, trace::TraceLayer, }; use tracing::{Span, error, info}; use super::middleware::request_id::maybe_add_request_id_header; use super::{ headers::X_REQUEST_ID, middleware::authorize::Authorize, routes::{ check_writability, configure, database_schema, dbs_and_roles, extension_server, extensions, grants, hadron_liveness_probe, insights, lfc, metrics, metrics_json, promote, refresh_configuration, status, terminate, }, }; use crate::compute::ComputeNode; /// `compute_ctl` has two servers: internal and external. The internal server /// binds to the loopback interface and handles communication from clients on /// the compute. The external server is what receives communication from the /// control plane, the metrics scraper, etc. We make the distinction because /// certain routes in `compute_ctl` only need to be exposed to local processes /// like Postgres via the neon extension and local_proxy. #[derive(Clone, Debug)] pub enum Server { Internal { port: u16, }, External { port: u16, config: ComputeCtlConfig, compute_id: String, instance_id: Option<String>, }, } impl Display for Server { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match self { Server::Internal { .. } => f.write_str("internal"), Server::External { .. } => f.write_str("external"), } } } impl From<&Server> for Router<Arc<ComputeNode>> { fn from(server: &Server) -> Self { let mut router = Router::<Arc<ComputeNode>>::new(); router = match server { Server::Internal { .. } => { router = router .route( "/extension_server/{*filename}", post(extension_server::download_extension), ) .route("/extensions", post(extensions::install_extension)) .route("/grants", post(grants::add_grant)) // Hadron: Compute-initiated configuration refresh .route( "/refresh_configuration", post(refresh_configuration::refresh_configuration), ); // Add in any testing support if cfg!(feature = "testing") { use super::routes::failpoints; router = router.route("/failpoints", post(failpoints::configure_failpoints)); } router } Server::External { config, compute_id, instance_id, .. } => { let unauthenticated_router = Router::<Arc<ComputeNode>>::new() .route("/metrics", get(metrics::get_metrics)) .route( "/autoscaling_metrics", get(metrics::get_autoscaling_metrics), ); let authenticated_router = Router::<Arc<ComputeNode>>::new() .route( "/lfc/prewarm", get(lfc::prewarm_state) .post(lfc::prewarm) .delete(lfc::cancel_prewarm), ) .route("/lfc/offload", get(lfc::offload_state).post(lfc::offload)) .route("/promote", post(promote::promote)) .route("/check_writability", post(check_writability::is_writable)) .route("/configure", post(configure::configure)) .route("/database_schema", get(database_schema::get_schema_dump)) .route("/dbs_and_roles", get(dbs_and_roles::get_catalog_objects)) .route("/insights", get(insights::get_insights)) .route("/metrics.json", get(metrics_json::get_metrics)) .route("/status", get(status::get_status)) .route("/terminate", post(terminate::terminate)) .route( "/hadron_liveness_probe", get(hadron_liveness_probe::hadron_liveness_probe), ) .layer(AsyncRequireAuthorizationLayer::new(Authorize::new( compute_id.clone(), instance_id.clone(), config.jwks.clone(), ))); router .merge(unauthenticated_router) .merge(authenticated_router) } }; router .fallback(Server::handle_404) .method_not_allowed_fallback(Server::handle_405) .layer( ServiceBuilder::new() .layer(tower_otel::trace::HttpLayer::server(tracing::Level::INFO)) // Add this middleware since we assume the request ID exists .layer(middleware::from_fn(maybe_add_request_id_header)) .layer( TraceLayer::new_for_http() .on_request(|request: &http::Request<_>, _span: &Span| { let request_id = request .headers() .get(X_REQUEST_ID) .unwrap() .to_str() .unwrap(); info!(%request_id, "{} {}", request.method(), request.uri()); }) .on_response( |response: &http::Response<_>, latency: Duration, _span: &Span| { let request_id = response .headers() .get(X_REQUEST_ID) .unwrap() .to_str() .unwrap(); info!( %request_id, code = response.status().as_u16(), latency = latency.as_millis() ); }, ), ) .layer(PropagateRequestIdLayer::x_request_id()), ) } } impl Server { async fn handle_404() -> impl IntoResponse { StatusCode::NOT_FOUND } async fn handle_405() -> impl IntoResponse { StatusCode::METHOD_NOT_ALLOWED } async fn listener(&self) -> Result<TcpListener> { let addr = SocketAddr::new(self.ip(), self.port()); let listener = TcpListener::bind(&addr).await?; Ok(listener) } fn ip(&self) -> IpAddr { match self { // TODO: Change this to Ipv6Addr::LOCALHOST when the GitHub runners // allow binding to localhost Server::Internal { .. } => IpAddr::from(Ipv6Addr::UNSPECIFIED), Server::External { .. } => IpAddr::from(Ipv6Addr::UNSPECIFIED), } } fn port(&self) -> u16 { match self { Server::Internal { port, .. } => *port, Server::External { port, .. } => *port, } } async fn serve(self, compute: Arc<ComputeNode>) { let listener = self.listener().await.unwrap_or_else(|e| { // If we can't bind, the compute cannot operate correctly panic!( "failed to bind the compute_ctl {} HTTP server to {}: {}", self, SocketAddr::new(self.ip(), self.port()), e ); }); if tracing::enabled!(tracing::Level::INFO) { let local_addr = match listener.local_addr() { Ok(local_addr) => local_addr, Err(_) => SocketAddr::new(self.ip(), self.port()), }; info!( "compute_ctl {} HTTP server listening at {}", self, local_addr ); } let router = Router::from(&self) .with_state(compute) .into_make_service_with_connect_info::<SocketAddr>(); if let Err(e) = axum::serve(listener, router).await { error!("compute_ctl {} HTTP server error: {}", self, e); } } pub fn launch(self, compute: &Arc<ComputeNode>) { let state = Arc::clone(compute); info!("Launching the {} server", self); tokio::spawn(self.serve(state)); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/extract/path.rs
compute_tools/src/http/extract/path.rs
use std::ops::{Deref, DerefMut}; use axum::extract::FromRequestParts; use axum::extract::rejection::PathRejection; use compute_api::responses::GenericAPIError; use http::StatusCode; use http::request::Parts; /// Custom `Path` extractor, so that we can format errors into /// `JsonResponse<GenericAPIError>`. #[derive(Debug, Clone, Copy, Default)] pub(crate) struct Path<T>(pub T); impl<S, T> FromRequestParts<S> for Path<T> where axum::extract::Path<T>: FromRequestParts<S, Rejection = PathRejection>, S: Send + Sync, { type Rejection = (StatusCode, axum::Json<GenericAPIError>); async fn from_request_parts(parts: &mut Parts, state: &S) -> Result<Self, Self::Rejection> { match axum::extract::Path::<T>::from_request_parts(parts, state).await { Ok(value) => Ok(Self(value.0)), Err(rejection) => Err(( rejection.status(), axum::Json(GenericAPIError { error: rejection.body_text().to_ascii_lowercase(), }), )), } } } impl<T> Deref for Path<T> { type Target = T; fn deref(&self) -> &Self::Target { &self.0 } } impl<T> DerefMut for Path<T> { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/extract/request_id.rs
compute_tools/src/http/extract/request_id.rs
use std::{ fmt::Display, ops::{Deref, DerefMut}, }; use axum::{extract::FromRequestParts, response::IntoResponse}; use http::{StatusCode, request::Parts}; use crate::http::{JsonResponse, headers::X_REQUEST_ID}; /// Extract the request ID from the `X-Request-Id` header. #[derive(Debug, Clone, Default)] pub(crate) struct RequestId(pub String); #[derive(Debug)] /// Rejection used for [`RequestId`]. /// /// Contains one variant for each way the [`RequestId`] extractor can /// fail. pub(crate) enum RequestIdRejection { /// The request is missing the header. MissingRequestId, /// The value of the header is invalid UTF-8. InvalidUtf8, } impl RequestIdRejection { pub fn status(&self) -> StatusCode { match self { RequestIdRejection::MissingRequestId => StatusCode::INTERNAL_SERVER_ERROR, RequestIdRejection::InvalidUtf8 => StatusCode::BAD_REQUEST, } } pub fn message(&self) -> String { match self { RequestIdRejection::MissingRequestId => "request ID is missing", RequestIdRejection::InvalidUtf8 => "request ID is invalid UTF-8", } .to_string() } } impl IntoResponse for RequestIdRejection { fn into_response(self) -> axum::response::Response { JsonResponse::error(self.status(), self.message()) } } impl<S> FromRequestParts<S> for RequestId where S: Send + Sync, { type Rejection = RequestIdRejection; async fn from_request_parts(parts: &mut Parts, _state: &S) -> Result<Self, Self::Rejection> { match parts.headers.get(X_REQUEST_ID) { Some(value) => match value.to_str() { Ok(request_id) => Ok(Self(request_id.to_string())), Err(_) => Err(RequestIdRejection::InvalidUtf8), }, None => Err(RequestIdRejection::MissingRequestId), } } } impl Deref for RequestId { type Target = String; fn deref(&self) -> &Self::Target { &self.0 } } impl DerefMut for RequestId { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } impl Display for RequestId { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { f.write_str(&self.0) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/extract/json.rs
compute_tools/src/http/extract/json.rs
use std::ops::{Deref, DerefMut}; use axum::extract::rejection::JsonRejection; use axum::extract::{FromRequest, Request}; use compute_api::responses::GenericAPIError; use http::StatusCode; /// Custom `Json` extractor, so that we can format errors into /// `JsonResponse<GenericAPIError>`. #[derive(Debug, Clone, Copy, Default)] pub(crate) struct Json<T>(pub T); impl<S, T> FromRequest<S> for Json<T> where axum::Json<T>: FromRequest<S, Rejection = JsonRejection>, S: Send + Sync, { type Rejection = (StatusCode, axum::Json<GenericAPIError>); async fn from_request(req: Request, state: &S) -> Result<Self, Self::Rejection> { match axum::Json::<T>::from_request(req, state).await { Ok(value) => Ok(Self(value.0)), Err(rejection) => Err(( rejection.status(), axum::Json(GenericAPIError { error: rejection.body_text().to_lowercase(), }), )), } } } impl<T> Deref for Json<T> { type Target = T; fn deref(&self) -> &Self::Target { &self.0 } } impl<T> DerefMut for Json<T> { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/extract/mod.rs
compute_tools/src/http/extract/mod.rs
pub(crate) mod json; pub(crate) mod path; pub(crate) mod query; pub(crate) mod request_id; pub(crate) use json::Json; pub(crate) use path::Path; pub(crate) use query::Query; #[allow(unused)] pub(crate) use request_id::RequestId;
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/extract/query.rs
compute_tools/src/http/extract/query.rs
use std::ops::{Deref, DerefMut}; use axum::extract::FromRequestParts; use axum::extract::rejection::QueryRejection; use compute_api::responses::GenericAPIError; use http::StatusCode; use http::request::Parts; /// Custom `Query` extractor, so that we can format errors into /// `JsonResponse<GenericAPIError>`. #[derive(Debug, Clone, Copy, Default)] pub(crate) struct Query<T>(pub T); impl<S, T> FromRequestParts<S> for Query<T> where axum::extract::Query<T>: FromRequestParts<S, Rejection = QueryRejection>, S: Send + Sync, { type Rejection = (StatusCode, axum::Json<GenericAPIError>); async fn from_request_parts(parts: &mut Parts, state: &S) -> Result<Self, Self::Rejection> { match axum::extract::Query::<T>::from_request_parts(parts, state).await { Ok(value) => Ok(Self(value.0)), Err(rejection) => Err(( rejection.status(), axum::Json(GenericAPIError { error: rejection.body_text().to_ascii_lowercase(), }), )), } } } impl<T> Deref for Query<T> { type Target = T; fn deref(&self) -> &Self::Target { &self.0 } } impl<T> DerefMut for Query<T> { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/middleware/authorize.rs
compute_tools/src/http/middleware/authorize.rs
use anyhow::{Result, anyhow}; use axum::{RequestExt, body::Body}; use axum_extra::{ TypedHeader, headers::{Authorization, authorization::Bearer}, }; use compute_api::requests::{COMPUTE_AUDIENCE, ComputeClaims, ComputeClaimsScope}; use futures::future::BoxFuture; use http::{Request, Response, StatusCode}; use jsonwebtoken::{Algorithm, DecodingKey, TokenData, Validation, jwk::JwkSet}; use tower_http::auth::AsyncAuthorizeRequest; use tracing::{debug, warn}; use crate::http::JsonResponse; #[derive(Clone, Debug)] pub(in crate::http) struct Authorize { compute_id: String, // BEGIN HADRON // Hadron instance ID. Only set if it's a Lakebase V1 a.k.a. Hadron instance. instance_id: Option<String>, // END HADRON jwks: JwkSet, validation: Validation, } impl Authorize { pub fn new(compute_id: String, instance_id: Option<String>, jwks: JwkSet) -> Self { let mut validation = Validation::new(Algorithm::EdDSA); // BEGIN HADRON let use_rsa = jwks.keys.iter().any(|jwk| { jwk.common .key_algorithm .is_some_and(|alg| alg == jsonwebtoken::jwk::KeyAlgorithm::RS256) }); if use_rsa { validation = Validation::new(Algorithm::RS256); } // END HADRON validation.validate_exp = true; // Unused by the control plane validation.validate_nbf = false; // Unused by the control plane validation.validate_aud = false; validation.set_audience(&[COMPUTE_AUDIENCE]); // Nothing is currently required validation.set_required_spec_claims(&[] as &[&str; 0]); Self { compute_id, instance_id, jwks, validation, } } } impl AsyncAuthorizeRequest<Body> for Authorize { type RequestBody = Body; type ResponseBody = Body; type Future = BoxFuture<'static, Result<Request<Body>, Response<Self::ResponseBody>>>; fn authorize(&mut self, mut request: Request<Body>) -> Self::Future { let compute_id = self.compute_id.clone(); let is_hadron_instance = self.instance_id.is_some(); let jwks = self.jwks.clone(); let validation = self.validation.clone(); Box::pin(async move { // BEGIN HADRON // In Hadron deployments the "external" HTTP endpoint on compute_ctl can only be // accessed by trusted components (enforced by dblet network policy), so we can bypass // all auth here. if is_hadron_instance { return Ok(request); } // END HADRON let TypedHeader(Authorization(bearer)) = request .extract_parts::<TypedHeader<Authorization<Bearer>>>() .await .map_err(|_| { JsonResponse::error(StatusCode::BAD_REQUEST, "invalid authorization token") })?; let data = match Self::verify(&jwks, bearer.token(), &validation) { Ok(claims) => claims, Err(e) => return Err(JsonResponse::error(StatusCode::UNAUTHORIZED, e)), }; match data.claims.scope { // TODO: We should validate audience for every token, but // instead of this ad-hoc validation, we should turn // [`Validation::validate_aud`] on. This is merely a stopgap // while we roll out `aud` deployment. We return a 401 // Unauthorized because when we eventually do use // [`Validation`], we will hit the above `Err` match arm which // returns 401 Unauthorized. Some(ComputeClaimsScope::Admin) => { let Some(ref audience) = data.claims.audience else { return Err(JsonResponse::error( StatusCode::UNAUTHORIZED, "missing audience in authorization token claims", )); }; if !audience.iter().any(|a| a == COMPUTE_AUDIENCE) { return Err(JsonResponse::error( StatusCode::UNAUTHORIZED, "invalid audience in authorization token claims", )); } } // If the scope is not [`ComputeClaimsScope::Admin`], then we // must validate the compute_id _ => { let Some(ref claimed_compute_id) = data.claims.compute_id else { return Err(JsonResponse::error( StatusCode::FORBIDDEN, "missing compute_id in authorization token claims", )); }; if *claimed_compute_id != compute_id { return Err(JsonResponse::error( StatusCode::FORBIDDEN, "invalid compute ID in authorization token claims", )); } } } // Make claims available to any subsequent middleware or request // handlers request.extensions_mut().insert(data.claims); Ok(request) }) } } impl Authorize { /// Verify the token using the JSON Web Key set and return the token data. fn verify( jwks: &JwkSet, token: &str, validation: &Validation, ) -> Result<TokenData<ComputeClaims>> { debug_assert!(!jwks.keys.is_empty()); debug!("verifying token {}", token); for jwk in jwks.keys.iter() { let decoding_key = match DecodingKey::from_jwk(jwk) { Ok(key) => key, Err(e) => { warn!( "failed to construct decoding key from {}: {}", jwk.common.key_id.as_ref().unwrap(), e ); continue; } }; match jsonwebtoken::decode::<ComputeClaims>(token, &decoding_key, validation) { Ok(data) => return Ok(data), Err(e) => { warn!( "failed to decode authorization token using {}: {}", jwk.common.key_id.as_ref().unwrap(), e ); continue; } } } Err(anyhow!("failed to verify authorization token")) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/middleware/request_id.rs
compute_tools/src/http/middleware/request_id.rs
use axum::{extract::Request, middleware::Next, response::Response}; use uuid::Uuid; use crate::http::headers::X_REQUEST_ID; /// This middleware function allows compute_ctl to generate its own request ID /// if one isn't supplied. The control plane will always send one as a UUID. The /// neon Postgres extension on the other hand does not send one. pub async fn maybe_add_request_id_header(mut request: Request, next: Next) -> Response { let headers = request.headers_mut(); if !headers.contains_key(X_REQUEST_ID) { headers.append(X_REQUEST_ID, Uuid::new_v4().to_string().parse().unwrap()); } next.run(request).await }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/middleware/mod.rs
compute_tools/src/http/middleware/mod.rs
pub(in crate::http) mod authorize; pub(in crate::http) mod request_id;
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/promote.rs
compute_tools/src/http/routes/promote.rs
use crate::http::JsonResponse; use axum::extract::Json; use compute_api::responses::PromoteConfig; use http::StatusCode; pub(in crate::http) async fn promote( compute: axum::extract::State<std::sync::Arc<crate::compute::ComputeNode>>, Json(cfg): Json<PromoteConfig>, ) -> axum::response::Response { // Return early at the cost of extra parsing spec let pspec = match crate::compute::ParsedSpec::try_from(cfg.spec) { Ok(p) => p, Err(e) => return JsonResponse::error(StatusCode::BAD_REQUEST, e), }; let cfg = PromoteConfig { spec: pspec.spec, wal_flush_lsn: cfg.wal_flush_lsn, }; let state = compute.promote(cfg).await; if let compute_api::responses::PromoteState::Failed { error: _ } = state { return JsonResponse::create_response(StatusCode::INTERNAL_SERVER_ERROR, state); } JsonResponse::success(StatusCode::OK, state) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/terminate.rs
compute_tools/src/http/routes/terminate.rs
use crate::compute::{ComputeNode, forward_termination_signal}; use crate::http::JsonResponse; use axum::extract::State; use axum::response::{IntoResponse, Response}; use axum_extra::extract::OptionalQuery; use compute_api::responses::{ComputeStatus, TerminateMode, TerminateResponse}; use http::StatusCode; use serde::Deserialize; use std::sync::Arc; use tokio::task; use tracing::info; #[derive(Deserialize, Default)] pub struct TerminateQuery { mode: TerminateMode, } /// Terminate the compute. pub(in crate::http) async fn terminate( State(compute): State<Arc<ComputeNode>>, OptionalQuery(terminate): OptionalQuery<TerminateQuery>, ) -> Response { let mode = terminate.unwrap_or_default().mode; { let mut state = compute.state.lock().unwrap(); if state.status == ComputeStatus::Terminated { let response = TerminateResponse { lsn: state.terminate_flush_lsn, }; return JsonResponse::success(StatusCode::CREATED, response); } if !matches!(state.status, ComputeStatus::Empty | ComputeStatus::Running) { return JsonResponse::invalid_status(state.status); } // If compute is Empty, there's no Postgres to terminate. The regular compute_ctl termination path // assumes Postgres to be configured and running, so we just special-handle this case by exiting // the process directly. if compute.params.lakebase_mode && state.status == ComputeStatus::Empty { drop(state); info!("terminating empty compute - will exit process"); // Queue a task to exit the process after 5 seconds. The 5-second delay aims to // give enough time for the HTTP response to be sent so that HCM doesn't get an abrupt // connection termination. tokio::spawn(async { tokio::time::sleep(tokio::time::Duration::from_secs(5)).await; info!("exiting process after terminating empty compute"); std::process::exit(0); }); return StatusCode::OK.into_response(); } // For Running status, proceed with normal termination state.set_status(mode.into(), &compute.state_changed); drop(state); } forward_termination_signal(false); info!("sent signal and notified waiters"); // Spawn a blocking thread to wait for compute to become Terminated. // This is needed to do not block the main pool of workers and // be able to serve other requests while some particular request // is waiting for compute to finish configuration. let c = compute.clone(); let lsn = task::spawn_blocking(move || { let mut state = c.state.lock().unwrap(); while state.status != ComputeStatus::Terminated { state = c.state_changed.wait(state).unwrap(); info!( "waiting for compute to become {}, current status: {:?}", ComputeStatus::Terminated, state.status ); } state.terminate_flush_lsn }) .await .unwrap(); info!("terminated Postgres"); JsonResponse::success(StatusCode::OK, TerminateResponse { lsn }) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/hadron_liveness_probe.rs
compute_tools/src/http/routes/hadron_liveness_probe.rs
use crate::pg_isready::pg_isready; use crate::{compute::ComputeNode, http::JsonResponse}; use axum::{extract::State, http::StatusCode, response::Response}; use std::sync::Arc; /// NOTE: NOT ENABLED YET /// Detect if the compute is alive. /// Called by the liveness probe of the compute container. pub(in crate::http) async fn hadron_liveness_probe( State(compute): State<Arc<ComputeNode>>, ) -> Response { let port = match compute.params.connstr.port() { Some(port) => port, None => { return JsonResponse::error( StatusCode::INTERNAL_SERVER_ERROR, "Failed to get the port from the connection string", ); } }; match pg_isready(&compute.params.pg_isready_bin, port) { Ok(_) => { // The connection is successful, so the compute is alive. // Return a 200 OK response. JsonResponse::success(StatusCode::OK, "ok") } Err(e) => { tracing::error!("Hadron liveness probe failed: {}", e); // The connection failed, so the compute is not alive. // Return a 500 Internal Server Error response. JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e) } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/failpoints.rs
compute_tools/src/http/routes/failpoints.rs
use axum::response::{IntoResponse, Response}; use http::StatusCode; use serde::{Deserialize, Serialize}; use tracing::info; use utils::failpoint_support::apply_failpoint; pub type ConfigureFailpointsRequest = Vec<FailpointConfig>; /// Information for configuring a single fail point #[derive(Debug, Serialize, Deserialize)] pub struct FailpointConfig { /// Name of the fail point pub name: String, /// List of actions to take, using the format described in `fail::cfg` /// /// We also support `actions = "exit"` to cause the fail point to immediately exit. pub actions: String, } use crate::http::JsonResponse; use crate::http::extract::Json; /// Configure failpoints for testing purposes. pub(in crate::http) async fn configure_failpoints( failpoints: Json<ConfigureFailpointsRequest>, ) -> Response { if !fail::has_failpoints() { return JsonResponse::error( StatusCode::PRECONDITION_FAILED, "Cannot manage failpoints because neon was compiled without failpoints support", ); } for fp in &*failpoints { info!("cfg failpoint: {} {}", fp.name, fp.actions); // We recognize one extra "action" that's not natively recognized // by the failpoints crate: exit, to immediately kill the process let cfg_result = apply_failpoint(&fp.name, &fp.actions); if let Err(e) = cfg_result { return JsonResponse::error( StatusCode::BAD_REQUEST, format!("failed to configure failpoints: {e}"), ); } } StatusCode::OK.into_response() }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/extension_server.rs
compute_tools/src/http/routes/extension_server.rs
use std::sync::Arc; use axum::extract::State; use axum::response::{IntoResponse, Response}; use http::StatusCode; use serde::Deserialize; use crate::compute::{BUILD_TAG, ComputeNode}; use crate::http::JsonResponse; use crate::http::extract::{Path, Query}; #[derive(Debug, Clone, Deserialize)] pub(in crate::http) struct ExtensionServerParams { #[serde(default)] is_library: bool, } /// Download a remote extension. pub(in crate::http) async fn download_extension( Path(filename): Path<String>, ext_server_params: Query<ExtensionServerParams>, State(compute): State<Arc<ComputeNode>>, ) -> Response { // Don't even try to download extensions if no remote storage is configured if compute.params.remote_ext_base_url.is_none() { return JsonResponse::error( StatusCode::PRECONDITION_FAILED, "remote storage is not configured", ); } let ext = { let state = compute.state.lock().unwrap(); let pspec = state.pspec.as_ref().unwrap(); let spec = &pspec.spec; let remote_extensions = match spec.remote_extensions.as_ref() { Some(r) => r, None => { return JsonResponse::error( StatusCode::CONFLICT, "information about remote extensions is unavailable", ); } }; remote_extensions.get_ext( &filename, ext_server_params.is_library, &BUILD_TAG, &compute.params.pgversion, ) }; match ext { Ok((ext_name, ext_path)) => match compute.download_extension(ext_name, ext_path).await { Ok(_) => StatusCode::OK.into_response(), Err(e) => JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e), }, Err(e) => JsonResponse::error(StatusCode::NOT_FOUND, e), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/lfc.rs
compute_tools/src/http/routes/lfc.rs
use crate::http::JsonResponse; use axum::response::{IntoResponse, Response}; use axum::{Json, http::StatusCode}; use axum_extra::extract::OptionalQuery; use compute_api::responses::{LfcOffloadState, LfcPrewarmState}; type Compute = axum::extract::State<std::sync::Arc<crate::compute::ComputeNode>>; pub(in crate::http) async fn prewarm_state(compute: Compute) -> Json<LfcPrewarmState> { Json(compute.lfc_prewarm_state().await) } // Following functions are marked async for axum, as it's more convenient than wrapping these // in async lambdas at call site pub(in crate::http) async fn offload_state(compute: Compute) -> Json<LfcOffloadState> { Json(compute.lfc_offload_state()) } #[derive(serde::Deserialize)] pub struct PrewarmQuery { pub from_endpoint: String, } pub(in crate::http) async fn prewarm( compute: Compute, OptionalQuery(query): OptionalQuery<PrewarmQuery>, ) -> Response { if compute.prewarm_lfc(query.map(|q| q.from_endpoint)) { StatusCode::ACCEPTED.into_response() } else { JsonResponse::error( StatusCode::TOO_MANY_REQUESTS, "Multiple requests for prewarm are not allowed", ) } } pub(in crate::http) async fn offload(compute: Compute) -> Response { if compute.offload_lfc() { StatusCode::ACCEPTED.into_response() } else { JsonResponse::error( StatusCode::TOO_MANY_REQUESTS, "Multiple requests for prewarm offload are not allowed", ) } } pub(in crate::http) async fn cancel_prewarm(compute: Compute) -> StatusCode { compute.cancel_prewarm(); StatusCode::ACCEPTED }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/status.rs
compute_tools/src/http/routes/status.rs
use std::ops::Deref; use std::sync::Arc; use axum::extract::State; use axum::http::StatusCode; use axum::response::Response; use compute_api::responses::ComputeStatusResponse; use crate::compute::ComputeNode; use crate::http::JsonResponse; /// Retrieve the state of the comute. pub(in crate::http) async fn get_status(State(compute): State<Arc<ComputeNode>>) -> Response { let state = compute.state.lock().unwrap(); let body = ComputeStatusResponse::from(state.deref()); JsonResponse::success(StatusCode::OK, body) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/metrics_json.rs
compute_tools/src/http/routes/metrics_json.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use http::StatusCode; use crate::compute::ComputeNode; use crate::http::JsonResponse; /// Get startup metrics. pub(in crate::http) async fn get_metrics(State(compute): State<Arc<ComputeNode>>) -> Response { let metrics = compute.state.lock().unwrap().metrics.clone(); JsonResponse::success(StatusCode::OK, metrics) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/grants.rs
compute_tools/src/http/routes/grants.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use compute_api::requests::SetRoleGrantsRequest; use compute_api::responses::{ComputeStatus, SetRoleGrantsResponse}; use http::StatusCode; use crate::compute::ComputeNode; use crate::http::JsonResponse; use crate::http::extract::Json; /// Add grants for a role. pub(in crate::http) async fn add_grant( State(compute): State<Arc<ComputeNode>>, request: Json<SetRoleGrantsRequest>, ) -> Response { let status = compute.get_status(); if status != ComputeStatus::Running { return JsonResponse::invalid_status(status); } match compute .set_role_grants( &request.database, &request.schema, &request.privileges, &request.role, ) .await { Ok(()) => JsonResponse::success( StatusCode::CREATED, Some(SetRoleGrantsResponse { database: request.database.clone(), schema: request.schema.clone(), role: request.role.clone(), privileges: request.privileges.clone(), }), ), Err(e) => JsonResponse::error( StatusCode::INTERNAL_SERVER_ERROR, format!("failed to grant role privileges to the schema: {e}"), ), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/database_schema.rs
compute_tools/src/http/routes/database_schema.rs
use std::sync::Arc; use axum::body::Body; use axum::extract::State; use axum::response::Response; use http::StatusCode; use http::header::CONTENT_TYPE; use serde::Deserialize; use crate::catalog::{SchemaDumpError, get_database_schema}; use crate::compute::ComputeNode; use crate::http::JsonResponse; use crate::http::extract::Query; #[derive(Debug, Clone, Deserialize)] pub(in crate::http) struct DatabaseSchemaParams { database: String, } /// Get a schema dump of the requested database. pub(in crate::http) async fn get_schema_dump( params: Query<DatabaseSchemaParams>, State(compute): State<Arc<ComputeNode>>, ) -> Response { match get_database_schema(&compute, &params.database).await { Ok(schema) => Response::builder() .status(StatusCode::OK) .header(CONTENT_TYPE.as_str(), "application/json") .body(Body::from_stream(schema)) .unwrap(), Err(SchemaDumpError::DatabaseDoesNotExist) => { JsonResponse::error(StatusCode::NOT_FOUND, SchemaDumpError::DatabaseDoesNotExist) } Err(e) => JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/extensions.rs
compute_tools/src/http/routes/extensions.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use compute_api::requests::ExtensionInstallRequest; use compute_api::responses::{ComputeStatus, ExtensionInstallResponse}; use http::StatusCode; use crate::compute::ComputeNode; use crate::http::JsonResponse; use crate::http::extract::Json; /// Install a extension. pub(in crate::http) async fn install_extension( State(compute): State<Arc<ComputeNode>>, request: Json<ExtensionInstallRequest>, ) -> Response { let status = compute.get_status(); if status != ComputeStatus::Running { return JsonResponse::invalid_status(status); } match compute .install_extension( &request.extension, &request.database, request.version.to_string(), ) .await { Ok(version) => JsonResponse::success( StatusCode::CREATED, Some(ExtensionInstallResponse { extension: request.extension.clone(), version, }), ), Err(e) => JsonResponse::error( StatusCode::INTERNAL_SERVER_ERROR, format!("failed to install extension: {e}"), ), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/mod.rs
compute_tools/src/http/routes/mod.rs
use compute_api::responses::ComputeStatusResponse; use crate::compute::ComputeState; pub(in crate::http) mod check_writability; pub(in crate::http) mod configure; pub(in crate::http) mod database_schema; pub(in crate::http) mod dbs_and_roles; pub(in crate::http) mod extension_server; pub(in crate::http) mod extensions; pub(in crate::http) mod failpoints; pub(in crate::http) mod grants; pub(in crate::http) mod hadron_liveness_probe; pub(in crate::http) mod insights; pub(in crate::http) mod lfc; pub(in crate::http) mod metrics; pub(in crate::http) mod metrics_json; pub(in crate::http) mod promote; pub(in crate::http) mod refresh_configuration; pub(in crate::http) mod status; pub(in crate::http) mod terminate; impl From<&ComputeState> for ComputeStatusResponse { fn from(state: &ComputeState) -> Self { ComputeStatusResponse { start_time: state.start_time, tenant: state .pspec .as_ref() .map(|pspec| pspec.tenant_id.to_string()), timeline: state .pspec .as_ref() .map(|pspec| pspec.timeline_id.to_string()), status: state.status, last_active: state.last_active, error: state.error.clone(), } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/dbs_and_roles.rs
compute_tools/src/http/routes/dbs_and_roles.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use http::StatusCode; use crate::catalog::get_dbs_and_roles; use crate::compute::ComputeNode; use crate::http::JsonResponse; /// Get the databases and roles from the compute. pub(in crate::http) async fn get_catalog_objects( State(compute): State<Arc<ComputeNode>>, ) -> Response { match get_dbs_and_roles(&compute).await { Ok(catalog_objects) => JsonResponse::success(StatusCode::OK, catalog_objects), Err(e) => JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/metrics.rs
compute_tools/src/http/routes/metrics.rs
use std::path::Path; use std::sync::Arc; use anyhow::Context; use axum::body::Body; use axum::extract::State; use axum::response::Response; use http::header::CONTENT_TYPE; use http_body_util::BodyExt; use hyper::{Request, StatusCode}; use metrics::proto::MetricFamily; use metrics::{Encoder, TextEncoder}; use crate::communicator_socket_client::connect_communicator_socket; use crate::compute::ComputeNode; use crate::hadron_metrics; use crate::http::JsonResponse; use crate::metrics::collect; /// Expose Prometheus metrics. pub(in crate::http) async fn get_metrics() -> Response { // When we call TextEncoder::encode() below, it will immediately return an // error if a metric family has no metrics, so we need to preemptively // filter out metric families with no metrics. let mut metrics = collect() .into_iter() .filter(|m| !m.get_metric().is_empty()) .collect::<Vec<MetricFamily>>(); // Add Hadron metrics. let hadron_metrics: Vec<MetricFamily> = hadron_metrics::collect() .into_iter() .filter(|m| !m.get_metric().is_empty()) .collect(); metrics.extend(hadron_metrics); let encoder = TextEncoder::new(); let mut buffer = vec![]; if let Err(e) = encoder.encode(&metrics, &mut buffer) { return JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e); } Response::builder() .status(StatusCode::OK) .header(CONTENT_TYPE, encoder.format_type()) .body(Body::from(buffer)) .unwrap() } /// Fetch and forward metrics from the Postgres neon extension's metrics /// exporter that are used by autoscaling-agent. /// /// The neon extension exposes these metrics over a Unix domain socket /// in the data directory. That's not accessible directly from the outside /// world, so we have this endpoint in compute_ctl to expose it pub(in crate::http) async fn get_autoscaling_metrics( State(compute): State<Arc<ComputeNode>>, ) -> Result<Response, Response> { let pgdata = Path::new(&compute.params.pgdata); // Connect to the communicator process's metrics socket let mut metrics_client = connect_communicator_socket(pgdata) .await .map_err(|e| JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, format!("{e:#}")))?; // Make a request for /autoscaling_metrics let request = Request::builder() .method("GET") .uri("/autoscaling_metrics") .header("Host", "localhost") // hyper requires Host, even though the server won't care .body(Body::from("")) .unwrap(); let resp = metrics_client .send_request(request) .await .context("fetching metrics from Postgres metrics service") .map_err(|e| JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, format!("{e:#}")))?; // Build a response that just forwards the response we got. let mut response = Response::builder(); response = response.status(resp.status()); if let Some(content_type) = resp.headers().get(CONTENT_TYPE) { response = response.header(CONTENT_TYPE, content_type); } let body = tonic::service::AxumBody::from_stream(resp.into_body().into_data_stream()); Ok(response.body(body).unwrap()) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/insights.rs
compute_tools/src/http/routes/insights.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use compute_api::responses::ComputeStatus; use http::StatusCode; use crate::compute::ComputeNode; use crate::http::JsonResponse; /// Collect current Postgres usage insights. pub(in crate::http) async fn get_insights(State(compute): State<Arc<ComputeNode>>) -> Response { let status = compute.get_status(); if status != ComputeStatus::Running { return JsonResponse::invalid_status(status); } let insights = compute.collect_insights().await; JsonResponse::success(StatusCode::OK, insights) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/configure.rs
compute_tools/src/http/routes/configure.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use compute_api::requests::ConfigurationRequest; use compute_api::responses::{ComputeStatus, ComputeStatusResponse}; use http::StatusCode; use tokio::task; use tracing::info; use crate::compute::{ComputeNode, ParsedSpec}; use crate::http::JsonResponse; use crate::http::extract::Json; // Accept spec in JSON format and request compute configuration. If anything // goes wrong after we set the compute status to `ConfigurationPending` and // update compute state with new spec, we basically leave compute in the // potentially wrong state. That said, it's control-plane's responsibility to // watch compute state after reconfiguration request and to clean restart in // case of errors. pub(in crate::http) async fn configure( State(compute): State<Arc<ComputeNode>>, request: Json<ConfigurationRequest>, ) -> Response { let pspec = match ParsedSpec::try_from(request.0.spec) { Ok(p) => p, Err(e) => return JsonResponse::error(StatusCode::BAD_REQUEST, e), }; // XXX: wrap state update under lock in a code block. Otherwise, we will try // to `Send` `mut state` into the spawned thread bellow, which will cause // the following rustc error: // // error: future cannot be sent between threads safely { let mut state = compute.state.lock().unwrap(); if !matches!(state.status, ComputeStatus::Empty | ComputeStatus::Running) { return JsonResponse::invalid_status(state.status); } // Pass the tracing span to the main thread that performs the startup, // so that the start_compute operation is considered a child of this // configure request for tracing purposes. state.startup_span = Some(tracing::Span::current()); if compute.params.lakebase_mode { ComputeNode::set_spec(&compute.params, &mut state, pspec); } else { state.pspec = Some(pspec); } state.set_status(ComputeStatus::ConfigurationPending, &compute.state_changed); drop(state); } // Spawn a blocking thread to wait for compute to become Running. This is // needed to not block the main pool of workers and to be able to serve // other requests while some particular request is waiting for compute to // finish configuration. let c = compute.clone(); let completed = task::spawn_blocking(move || { let mut state = c.state.lock().unwrap(); while state.status != ComputeStatus::Running { state = c.state_changed.wait(state).unwrap(); info!( "waiting for compute to become {}, current status: {}", ComputeStatus::Running, state.status ); if state.status == ComputeStatus::Failed { let err = state.error.as_ref().map_or("unknown error", |x| x); let msg = format!("compute configuration failed: {err:?}"); return Err(msg); } } Ok(()) }) .await .unwrap(); if let Err(e) = completed { return JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e); } // Return current compute state if everything went well. let state = compute.state.lock().unwrap().clone(); let body = ComputeStatusResponse::from(&state); JsonResponse::success(StatusCode::OK, body) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/check_writability.rs
compute_tools/src/http/routes/check_writability.rs
use std::sync::Arc; use axum::extract::State; use axum::response::Response; use compute_api::responses::ComputeStatus; use http::StatusCode; use crate::checker::check_writability; use crate::compute::ComputeNode; use crate::http::JsonResponse; /// Check that the compute is currently running. pub(in crate::http) async fn is_writable(State(compute): State<Arc<ComputeNode>>) -> Response { let status = compute.get_status(); if status != ComputeStatus::Running { return JsonResponse::invalid_status(status); } match check_writability(&compute).await { Ok(_) => JsonResponse::success(StatusCode::OK, true), Err(e) => JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/src/http/routes/refresh_configuration.rs
compute_tools/src/http/routes/refresh_configuration.rs
// This file is added by Hadron use std::sync::Arc; use axum::{ extract::State, response::{IntoResponse, Response}, }; use http::StatusCode; use crate::compute::ComputeNode; use crate::hadron_metrics::POSTGRES_PAGESTREAM_REQUEST_ERRORS; use crate::http::JsonResponse; /// The /refresh_configuration POST method is used to nudge compute_ctl to pull a new spec /// from the HCC and attempt to reconfigure Postgres with the new spec. The method does not wait /// for the reconfiguration to complete. Rather, it simply delivers a signal that will cause /// configuration to be reloaded in a best effort manner. Invocation of this method does not /// guarantee that a reconfiguration will occur. The caller should consider keep sending this /// request while it believes that the compute configuration is out of date. pub(in crate::http) async fn refresh_configuration( State(compute): State<Arc<ComputeNode>>, ) -> Response { POSTGRES_PAGESTREAM_REQUEST_ERRORS.inc(); match compute.signal_refresh_configuration().await { Ok(_) => StatusCode::OK.into_response(), Err(e) => JsonResponse::error(StatusCode::INTERNAL_SERVER_ERROR, e), } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/tests/config_test.rs
compute_tools/tests/config_test.rs
#[cfg(test)] mod config_tests { use std::fs::{File, remove_file}; use std::io::{Read, Write}; use std::path::Path; use compute_tools::config::*; fn write_test_file(path: &Path, content: &str) { let mut file = File::create(path).unwrap(); file.write_all(content.as_bytes()).unwrap(); } fn check_file_content(path: &Path, expected_content: &str) { let mut file = File::open(path).unwrap(); let mut content = String::new(); file.read_to_string(&mut content).unwrap(); assert_eq!(content, expected_content); } #[test] fn test_line_in_file() { let path = Path::new("./tests/tmp/config_test.txt"); write_test_file(path, "line1\nline2.1\t line2.2\nline3"); let line = "line2.1\t line2.2"; let result = line_in_file(path, line).unwrap(); assert!(!result); check_file_content(path, "line1\nline2.1\t line2.2\nline3"); let line = "line4"; let result = line_in_file(path, line).unwrap(); assert!(result); check_file_content(path, "line1\nline2.1\t line2.2\nline3\nline4"); remove_file(path).unwrap(); let path = Path::new("./tests/tmp/new_config_test.txt"); let line = "line4"; let result = line_in_file(path, line).unwrap(); assert!(result); check_file_content(path, "line4"); remove_file(path).unwrap(); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/compute_tools/tests/pg_helpers_tests.rs
compute_tools/tests/pg_helpers_tests.rs
#[cfg(test)] mod pg_helpers_tests { use std::fs::File; use compute_api::spec::{ComputeSpec, GenericOption, GenericOptions, PgIdent}; use compute_tools::pg_helpers::*; #[test] fn params_serialize() { let file = File::open("../libs/compute_api/tests/cluster_spec.json").unwrap(); let spec: ComputeSpec = serde_json::from_reader(file).unwrap(); assert_eq!( spec.cluster.databases.first().unwrap().to_pg_options(), "LC_COLLATE 'C' LC_CTYPE 'C' TEMPLATE template0 OWNER \"alexk\"" ); assert_eq!( spec.cluster.roles.first().unwrap().to_pg_options(), " LOGIN PASSWORD 'md56b1d16b78004bbd51fa06af9eda75972'" ); } #[test] fn settings_serialize() { let file = File::open("../libs/compute_api/tests/cluster_spec.json").unwrap(); let spec: ComputeSpec = serde_json::from_reader(file).unwrap(); assert_eq!( spec.cluster.settings.as_pg_settings(), r#"fsync = off wal_level = logical hot_standby = on autoprewarm = off offload_lfc_interval_seconds = 20 neon.safekeepers = '127.0.0.1:6502,127.0.0.1:6503,127.0.0.1:6501' wal_log_hints = on log_connections = on shared_buffers = 32768 port = 55432 max_connections = 100 max_wal_senders = 10 listen_addresses = '0.0.0.0' wal_sender_timeout = 0 password_encryption = md5 maintenance_work_mem = 65536 max_parallel_workers = 8 max_worker_processes = 8 neon.tenant_id = 'b0554b632bd4d547a63b86c3630317e8' max_replication_slots = 10 neon.timeline_id = '2414a61ffc94e428f14b5758fe308e13' shared_preload_libraries = 'neon' synchronous_standby_names = 'walproposer' neon.pageserver_connstring = 'host=127.0.0.1 port=6400' test.escaping = 'here''s a backslash \\ and a quote '' and a double-quote " hooray' "# ); } #[test] fn ident_pg_quote() { let ident: PgIdent = PgIdent::from("\"name\";\\n select 1;"); assert_eq!(ident.pg_quote(), "\"\"\"name\"\";\\n select 1;\""); } #[test] fn ident_pg_quote_dollar() { let test_cases = vec![ ("name", ("$x$name$x$", "xx")), ("name$", ("$x$name$$x$", "xx")), ("name$$", ("$x$name$$$x$", "xx")), ("name$$$", ("$x$name$$$$x$", "xx")), ("name$$$$", ("$x$name$$$$$x$", "xx")), ("name$x$", ("$xx$name$x$$xx$", "xxx")), ("x", ("$xx$x$xx$", "xxx")), ("xx", ("$xxx$xx$xxx$", "xxxx")), ("$x", ("$xx$$x$xx$", "xxx")), ("x$", ("$xx$x$$xx$", "xxx")), ("$x$", ("$xx$$x$$xx$", "xxx")), ("xx$", ("$xxx$xx$$xxx$", "xxxx")), ("$xx", ("$xxx$$xx$xxx$", "xxxx")), ("$xx$", ("$xxx$$xx$$xxx$", "xxxx")), ]; for (input, expected) in test_cases { let (escaped, tag) = PgIdent::from(input).pg_quote_dollar(); assert_eq!(escaped, expected.0); assert_eq!(tag, expected.1); } } #[test] fn generic_options_search() { let generic_options: GenericOptions = Some(vec![ GenericOption { name: "present_value".into(), value: Some("value".into()), vartype: "string".into(), }, GenericOption { name: "missed_value".into(), value: None, vartype: "int".into(), }, ]); assert_eq!(generic_options.find("present_value"), Some("value".into())); assert_eq!(generic_options.find("missed_value"), None); assert_eq!(generic_options.find("invalid_value"), None); let empty_generic_options: GenericOptions = Some(vec![]); assert_eq!(empty_generic_options.find("present_value"), None); assert_eq!(empty_generic_options.find("missed_value"), None); assert_eq!(empty_generic_options.find("invalid_value"), None); let none_generic_options: GenericOptions = None; assert_eq!(none_generic_options.find("present_value"), None); assert_eq!(none_generic_options.find("missed_value"), None); assert_eq!(none_generic_options.find("invalid_value"), None); } #[test] fn test_escape_literal() { assert_eq!(escape_literal("test"), "'test'"); assert_eq!(escape_literal("test'"), "'test'''"); assert_eq!(escape_literal("test\\'"), "E'test\\\\'''"); assert_eq!(escape_literal("test\\'\\'"), "E'test\\\\''\\\\'''"); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/safekeeper_client.rs
storage_controller/src/safekeeper_client.rs
use safekeeper_api::models::{ self, PullTimelineRequest, PullTimelineResponse, SafekeeperUtilization, TimelineCreateRequest, }; use safekeeper_client::mgmt_api::{Client, Result}; use utils::id::{NodeId, TenantId, TimelineId}; use utils::logging::SecretString; use crate::metrics::SafekeeperRequestLabelGroup; /// Thin wrapper around [`safekeeper_client::mgmt_api::Client`]. It allows the storage /// controller to collect metrics in a non-intrusive manner. /// /// Analogous to [`crate::pageserver_client::PageserverClient`]. #[derive(Debug, Clone)] pub(crate) struct SafekeeperClient { inner: Client, node_id_label: String, } macro_rules! measured_request { ($name:literal, $method:expr, $node_id: expr, $invoke:expr) => {{ let labels = SafekeeperRequestLabelGroup { safekeeper_id: $node_id, path: $name, method: $method, }; let latency = &crate::metrics::METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_request_latency; let _timer_guard = latency.start_timer(labels.clone()); let res = $invoke; if res.is_err() { let error_counters = &crate::metrics::METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_request_error; error_counters.inc(labels) } res }}; } impl SafekeeperClient { pub(crate) fn new( node_id: NodeId, raw_client: reqwest::Client, mgmt_api_endpoint: String, jwt: Option<SecretString>, ) -> Self { Self { inner: Client::new(raw_client, mgmt_api_endpoint, jwt), node_id_label: node_id.0.to_string(), } } pub(crate) fn node_id_label(&self) -> &str { &self.node_id_label } pub(crate) async fn create_timeline( &self, req: &TimelineCreateRequest, ) -> Result<reqwest::Response> { measured_request!( "create_timeline", crate::metrics::Method::Post, &self.node_id_label, self.inner.create_timeline(req).await ) } #[allow(unused)] pub(crate) async fn exclude_timeline( &self, tenant_id: TenantId, timeline_id: TimelineId, req: &models::TimelineMembershipSwitchRequest, ) -> Result<models::TimelineDeleteResult> { measured_request!( "exclude_timeline", crate::metrics::Method::Post, &self.node_id_label, self.inner .exclude_timeline(tenant_id, timeline_id, req) .await ) } pub(crate) async fn delete_timeline( &self, tenant_id: TenantId, timeline_id: TimelineId, ) -> Result<models::TimelineDeleteResult> { measured_request!( "delete_timeline", crate::metrics::Method::Delete, &self.node_id_label, self.inner.delete_timeline(tenant_id, timeline_id).await ) } #[allow(unused)] pub(crate) async fn switch_timeline_membership( &self, tenant_id: TenantId, timeline_id: TimelineId, req: &models::TimelineMembershipSwitchRequest, ) -> Result<models::TimelineMembershipSwitchResponse> { measured_request!( "switch_timeline_membership", crate::metrics::Method::Put, &self.node_id_label, self.inner .switch_timeline_membership(tenant_id, timeline_id, req) .await ) } pub(crate) async fn delete_tenant( &self, tenant_id: TenantId, ) -> Result<models::TenantDeleteResult> { measured_request!( "delete_tenant", crate::metrics::Method::Delete, &self.node_id_label, self.inner.delete_tenant(tenant_id).await ) } pub(crate) async fn pull_timeline( &self, req: &PullTimelineRequest, ) -> Result<PullTimelineResponse> { measured_request!( "pull_timeline", crate::metrics::Method::Post, &self.node_id_label, self.inner.pull_timeline(req).await ) } #[allow(unused)] pub(crate) async fn bump_timeline_term( &self, tenant_id: TenantId, timeline_id: TimelineId, req: &models::TimelineTermBumpRequest, ) -> Result<models::TimelineTermBumpResponse> { measured_request!( "term_bump", crate::metrics::Method::Post, &self.node_id_label, self.inner .bump_timeline_term(tenant_id, timeline_id, req) .await ) } pub(crate) async fn get_utilization(&self) -> Result<SafekeeperUtilization> { measured_request!( "utilization", crate::metrics::Method::Get, &self.node_id_label, self.inner.utilization().await ) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/node.rs
storage_controller/src/node.rs
use std::str::FromStr; use std::time::Duration; use pageserver_api::controller_api::{ AvailabilityZone, NodeAvailability, NodeDescribeResponse, NodeLifecycle, NodeRegisterRequest, NodeSchedulingPolicy, TenantLocateResponseShard, }; use pageserver_api::shard::TenantShardId; use pageserver_client::mgmt_api; use reqwest::StatusCode; use serde::Serialize; use tokio_util::sync::CancellationToken; use utils::backoff; use utils::id::NodeId; use crate::pageserver_client::PageserverClient; use crate::persistence::NodePersistence; use crate::scheduler::MaySchedule; /// Represents the in-memory description of a Node. /// /// Scheduling statistics are maintened separately in [`crate::scheduler`]. /// /// The persistent subset of the Node is defined in [`crate::persistence::NodePersistence`]: the /// implementation of serialization on this type is only for debug dumps. #[derive(Clone, Serialize)] pub(crate) struct Node { id: NodeId, availability: NodeAvailability, scheduling: NodeSchedulingPolicy, lifecycle: NodeLifecycle, listen_http_addr: String, listen_http_port: u16, listen_https_port: Option<u16>, listen_pg_addr: String, listen_pg_port: u16, listen_grpc_addr: Option<String>, listen_grpc_port: Option<u16>, availability_zone_id: AvailabilityZone, // Flag from storcon's config to use https for pageserver admin API. // Invariant: if |true|, listen_https_port should contain a value. use_https: bool, // This cancellation token means "stop any RPCs in flight to this node, and don't start // any more". It is not related to process shutdown. #[serde(skip)] cancel: CancellationToken, } #[allow(dead_code)] const ONE_MILLION: i64 = 1000000; // Converts a pool ID to a large number that can be used to assign unique IDs to pods in StatefulSets. /// For example, if pool_id is 1, then the pods have NodeIds 1000000, 1000001, 1000002, etc. /// If pool_id is None, then the pods have NodeIds 0, 1, 2, etc. #[allow(dead_code)] pub fn transform_pool_id(pool_id: Option<i32>) -> i64 { match pool_id { Some(id) => (id as i64) * ONE_MILLION, None => 0, } } #[allow(dead_code)] pub fn get_pool_id_from_node_id(node_id: i64) -> i32 { (node_id / ONE_MILLION) as i32 } /// Example pod name: page-server-0-1, safe-keeper-1-0 #[allow(dead_code)] pub fn get_node_id_from_pod_name(pod_name: &str) -> anyhow::Result<NodeId> { let parts: Vec<&str> = pod_name.split('-').collect(); if parts.len() != 4 { return Err(anyhow::anyhow!("Invalid pod name: {}", pod_name)); } let pool_id = parts[2].parse::<i32>()?; let node_offset = parts[3].parse::<i64>()?; let node_id = transform_pool_id(Some(pool_id)) + node_offset; Ok(NodeId(node_id as u64)) } /// When updating [`Node::availability`] we use this type to indicate to the caller /// whether/how they changed it. pub(crate) enum AvailabilityTransition { ToActive, ToWarmingUpFromActive, ToWarmingUpFromOffline, ToOffline, Unchanged, } impl Node { pub(crate) fn base_url(&self) -> String { if self.use_https { format!( "https://{}:{}", self.listen_http_addr, self.listen_https_port .expect("https port should be specified if use_https is on") ) } else { format!("http://{}:{}", self.listen_http_addr, self.listen_http_port) } } pub(crate) fn get_id(&self) -> NodeId { self.id } #[allow(unused)] pub(crate) fn get_availability_zone_id(&self) -> &AvailabilityZone { &self.availability_zone_id } pub(crate) fn get_scheduling(&self) -> NodeSchedulingPolicy { self.scheduling } pub(crate) fn set_scheduling(&mut self, scheduling: NodeSchedulingPolicy) { self.scheduling = scheduling } pub(crate) fn has_https_port(&self) -> bool { self.listen_https_port.is_some() } /// Does this registration request match `self`? This is used when deciding whether a registration /// request should be allowed to update an existing record with the same node ID. pub(crate) fn registration_match(&self, register_req: &NodeRegisterRequest) -> bool { self.id == register_req.node_id && self.listen_http_addr == register_req.listen_http_addr && self.listen_http_port == register_req.listen_http_port // Note: HTTPS and gRPC addresses may change, to allow for migrations. See // [`Self::need_update`] for more details. && self.listen_pg_addr == register_req.listen_pg_addr && self.listen_pg_port == register_req.listen_pg_port && self.availability_zone_id == register_req.availability_zone_id } // Do we need to update an existing record in DB on this registration request? pub(crate) fn need_update(&self, register_req: &NodeRegisterRequest) -> bool { // These are checked here, since they may change before we're fully migrated. self.listen_https_port != register_req.listen_https_port || self.listen_grpc_addr != register_req.listen_grpc_addr || self.listen_grpc_port != register_req.listen_grpc_port } /// For a shard located on this node, populate a response object /// with this node's address information. pub(crate) fn shard_location(&self, shard_id: TenantShardId) -> TenantLocateResponseShard { TenantLocateResponseShard { shard_id, node_id: self.id, listen_http_addr: self.listen_http_addr.clone(), listen_http_port: self.listen_http_port, listen_https_port: self.listen_https_port, listen_pg_addr: self.listen_pg_addr.clone(), listen_pg_port: self.listen_pg_port, listen_grpc_addr: self.listen_grpc_addr.clone(), listen_grpc_port: self.listen_grpc_port, } } pub(crate) fn get_availability(&self) -> &NodeAvailability { &self.availability } pub(crate) fn set_availability(&mut self, availability: NodeAvailability) { use AvailabilityTransition::*; use NodeAvailability::WarmingUp; match self.get_availability_transition(&availability) { ToActive => { // Give the node a new cancellation token, effectively resetting it to un-cancelled. Any // users of previously-cloned copies of the node will still see the old cancellation // state. For example, Reconcilers in flight will have to complete and be spawned // again to realize that the node has become available. self.cancel = CancellationToken::new(); } ToOffline | ToWarmingUpFromActive => { // Fire the node's cancellation token to cancel any in-flight API requests to it self.cancel.cancel(); } Unchanged | ToWarmingUpFromOffline => {} } if let (WarmingUp(crnt), WarmingUp(proposed)) = (&self.availability, &availability) { self.availability = WarmingUp(std::cmp::max(*crnt, *proposed)); } else { self.availability = availability; } } /// Without modifying the availability of the node, convert the intended availability /// into a description of the transition. pub(crate) fn get_availability_transition( &self, availability: &NodeAvailability, ) -> AvailabilityTransition { use AvailabilityTransition::*; use NodeAvailability::*; match (&self.availability, availability) { (Offline, Active(_)) => ToActive, (Active(_), Offline) => ToOffline, (Active(_), WarmingUp(_)) => ToWarmingUpFromActive, (WarmingUp(_), Offline) => ToOffline, (WarmingUp(_), Active(_)) => ToActive, (Offline, WarmingUp(_)) => ToWarmingUpFromOffline, _ => Unchanged, } } /// Whether we may send API requests to this node. pub(crate) fn is_available(&self) -> bool { // When we clone a node, [`Self::availability`] is a snapshot, but [`Self::cancel`] holds // a reference to the original Node's cancellation status. Checking both of these results // in a "pessimistic" check where we will consider a Node instance unavailable if it was unavailable // when we cloned it, or if the original Node instance's cancellation token was fired. matches!(self.availability, NodeAvailability::Active(_)) && !self.cancel.is_cancelled() } /// Is this node elegible to have work scheduled onto it? pub(crate) fn may_schedule(&self) -> MaySchedule { let utilization = match &self.availability { NodeAvailability::Active(u) => u.clone(), NodeAvailability::Offline | NodeAvailability::WarmingUp(_) => return MaySchedule::No, }; match self.scheduling { NodeSchedulingPolicy::Active => MaySchedule::Yes(utilization), NodeSchedulingPolicy::Deleting => MaySchedule::No, NodeSchedulingPolicy::Draining => MaySchedule::No, NodeSchedulingPolicy::Filling => MaySchedule::Yes(utilization), NodeSchedulingPolicy::Pause => MaySchedule::No, NodeSchedulingPolicy::PauseForRestart => MaySchedule::No, } } #[allow(clippy::too_many_arguments)] pub(crate) fn new( id: NodeId, listen_http_addr: String, listen_http_port: u16, listen_https_port: Option<u16>, listen_pg_addr: String, listen_pg_port: u16, listen_grpc_addr: Option<String>, listen_grpc_port: Option<u16>, availability_zone_id: AvailabilityZone, use_https: bool, ) -> anyhow::Result<Self> { if use_https && listen_https_port.is_none() { anyhow::bail!( "cannot create node {id}: \ https is enabled, but https port is not specified" ); } if listen_grpc_addr.is_some() != listen_grpc_port.is_some() { anyhow::bail!("cannot create node {id}: must specify both gRPC address and port"); } Ok(Self { id, listen_http_addr, listen_http_port, listen_https_port, listen_pg_addr, listen_pg_port, listen_grpc_addr, listen_grpc_port, scheduling: NodeSchedulingPolicy::Active, lifecycle: NodeLifecycle::Active, availability: NodeAvailability::Offline, availability_zone_id, use_https, cancel: CancellationToken::new(), }) } pub(crate) fn to_persistent(&self) -> NodePersistence { NodePersistence { node_id: self.id.0 as i64, scheduling_policy: self.scheduling.into(), lifecycle: self.lifecycle.into(), listen_http_addr: self.listen_http_addr.clone(), listen_http_port: self.listen_http_port as i32, listen_https_port: self.listen_https_port.map(|x| x as i32), listen_pg_addr: self.listen_pg_addr.clone(), listen_pg_port: self.listen_pg_port as i32, listen_grpc_addr: self.listen_grpc_addr.clone(), listen_grpc_port: self.listen_grpc_port.map(|port| port as i32), availability_zone_id: self.availability_zone_id.0.clone(), } } pub(crate) fn from_persistent(np: NodePersistence, use_https: bool) -> anyhow::Result<Self> { if use_https && np.listen_https_port.is_none() { anyhow::bail!( "cannot load node {} from persistent: \ https is enabled, but https port is not specified", np.node_id, ); } if np.listen_grpc_addr.is_some() != np.listen_grpc_port.is_some() { anyhow::bail!( "can't load node {}: must specify both gRPC address and port", np.node_id ); } Ok(Self { id: NodeId(np.node_id as u64), // At startup we consider a node offline until proven otherwise. availability: NodeAvailability::Offline, scheduling: NodeSchedulingPolicy::from_str(&np.scheduling_policy) .expect("Bad scheduling policy in DB"), lifecycle: NodeLifecycle::from_str(&np.lifecycle).expect("Bad lifecycle in DB"), listen_http_addr: np.listen_http_addr, listen_http_port: np.listen_http_port as u16, listen_https_port: np.listen_https_port.map(|x| x as u16), listen_pg_addr: np.listen_pg_addr, listen_pg_port: np.listen_pg_port as u16, listen_grpc_addr: np.listen_grpc_addr, listen_grpc_port: np.listen_grpc_port.map(|port| port as u16), availability_zone_id: AvailabilityZone(np.availability_zone_id), use_https, cancel: CancellationToken::new(), }) } /// Wrapper for issuing requests to pageserver management API: takes care of generic /// retry/backoff for retryable HTTP status codes. /// /// This will return None to indicate cancellation. Cancellation may happen from /// the cancellation token passed in, or from Self's cancellation token (i.e. node /// going offline). #[allow(clippy::too_many_arguments)] pub(crate) async fn with_client_retries<T, O, F>( &self, mut op: O, http_client: &reqwest::Client, jwt: &Option<String>, warn_threshold: u32, max_retries: u32, timeout: Duration, cancel: &CancellationToken, ) -> Option<mgmt_api::Result<T>> where O: FnMut(PageserverClient) -> F, F: std::future::Future<Output = mgmt_api::Result<T>>, { fn is_fatal(e: &mgmt_api::Error) -> bool { use mgmt_api::Error::*; match e { SendRequest(_) | ReceiveBody(_) | ReceiveErrorBody(_) => false, ApiError(StatusCode::SERVICE_UNAVAILABLE, _) | ApiError(StatusCode::GATEWAY_TIMEOUT, _) | ApiError(StatusCode::REQUEST_TIMEOUT, _) => false, ApiError(_, _) => true, Cancelled => true, Timeout(_) => false, } } backoff::retry( || { let client = PageserverClient::new( self.get_id(), http_client.clone(), self.base_url(), jwt.as_deref(), ); let node_cancel_fut = self.cancel.cancelled(); let op_fut = tokio::time::timeout(timeout, op(client)); async { tokio::select! { r = op_fut => match r { Ok(r) => r, Err(e) => Err(mgmt_api::Error::Timeout(format!("{e}"))), }, _ = node_cancel_fut => { Err(mgmt_api::Error::Cancelled) }} } }, is_fatal, warn_threshold, max_retries, &format!( "Call to node {} ({}) management API", self.id, self.base_url(), ), cancel, ) .await } /// Generate the simplified API-friendly description of a node's state pub(crate) fn describe(&self) -> NodeDescribeResponse { NodeDescribeResponse { id: self.id, availability: self.availability.clone().into(), scheduling: self.scheduling, availability_zone_id: self.availability_zone_id.0.clone(), listen_http_addr: self.listen_http_addr.clone(), listen_http_port: self.listen_http_port, listen_https_port: self.listen_https_port, listen_pg_addr: self.listen_pg_addr.clone(), listen_pg_port: self.listen_pg_port, listen_grpc_addr: self.listen_grpc_addr.clone(), listen_grpc_port: self.listen_grpc_port, } } } impl std::fmt::Display for Node { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!(f, "{} ({})", self.id, self.listen_http_addr) } } impl std::fmt::Debug for Node { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!(f, "{} ({})", self.id, self.listen_http_addr) } } #[cfg(test)] mod tests { use utils::id::NodeId; use crate::node::get_node_id_from_pod_name; #[test] fn test_get_node_id_from_pod_name() { let pod_name = "page-server-3-12"; let node_id = get_node_id_from_pod_name(pod_name).unwrap(); assert_eq!(node_id, NodeId(3000012)); let pod_name = "safe-keeper-1-0"; let node_id = get_node_id_from_pod_name(pod_name).unwrap(); assert_eq!(node_id, NodeId(1000000)); let pod_name = "invalid-pod-name"; let result = get_node_id_from_pod_name(pod_name); assert!(result.is_err()); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/reconciler.rs
storage_controller/src/reconciler.rs
use std::borrow::Cow; use std::collections::HashMap; use std::sync::Arc; use std::time::{Duration, Instant}; use json_structural_diff::JsonDiff; use pageserver_api::controller_api::{AvailabilityZone, MigrationConfig, PlacementPolicy}; use pageserver_api::models::{ LocationConfig, LocationConfigMode, LocationConfigSecondary, TenantConfig, TenantWaitLsnRequest, }; use pageserver_api::shard::{ShardIdentity, TenantShardId}; use pageserver_client::mgmt_api; use reqwest::StatusCode; use tokio_util::sync::CancellationToken; use utils::backoff::exponential_backoff; use utils::generation::Generation; use utils::id::{NodeId, TimelineId}; use utils::lsn::Lsn; use utils::pausable_failpoint; use utils::sync::gate::GateGuard; use crate::compute_hook::{ComputeHook, NotifyError}; use crate::node::Node; use crate::pageserver_client::PageserverClient; use crate::persistence::Persistence; use crate::tenant_shard::{IntentState, ObservedState, ObservedStateDelta, ObservedStateLocation}; use crate::{compute_hook, service}; const DEFAULT_HEATMAP_PERIOD: Duration = Duration::from_secs(60); /// Object with the lifetime of the background reconcile task that is created /// for tenants which have a difference between their intent and observed states. pub(super) struct Reconciler { /// See [`crate::tenant_shard::TenantShard`] for the meanings of these fields: they are a snapshot /// of a tenant's state from when we spawned a reconcile task. pub(super) tenant_shard_id: TenantShardId, pub(crate) shard: ShardIdentity, pub(crate) placement_policy: PlacementPolicy, pub(crate) generation: Option<Generation>, pub(crate) intent: TargetState, /// Nodes not referenced by [`Self::intent`], from which we should try /// to detach this tenant shard. pub(crate) detach: Vec<Node>, /// Configuration specific to this reconciler pub(crate) reconciler_config: ReconcilerConfig, pub(crate) config: TenantConfig, pub(crate) preferred_az: Option<AvailabilityZone>, /// Observed state from the point of view of the reconciler. /// This gets updated as the reconciliation makes progress. pub(crate) observed: ObservedState, /// Snapshot of the observed state at the point when the reconciler /// was spawned. pub(crate) original_observed: ObservedState, pub(crate) service_config: service::Config, /// A hook to notify the running postgres instances when we change the location /// of a tenant. Use this via [`Self::compute_notify`] to update our failure flag /// and guarantee eventual retries. pub(crate) compute_hook: Arc<ComputeHook>, /// To avoid stalling if the cloud control plane is unavailable, we may proceed /// past failures in [`ComputeHook::notify_attach`], but we _must_ remember that we failed /// so that we can set [`crate::tenant_shard::TenantShard::pending_compute_notification`] to ensure a later retry. pub(crate) compute_notify_failure: bool, /// Reconciler is responsible for keeping alive semaphore units that limit concurrency on how many /// we will spawn. pub(crate) _resource_units: ReconcileUnits, /// A means to abort background reconciliation: it is essential to /// call this when something changes in the original TenantShard that /// will make this reconciliation impossible or unnecessary, for /// example when a pageserver node goes offline, or the PlacementPolicy for /// the tenant is changed. pub(crate) cancel: CancellationToken, /// Reconcilers are registered with a Gate so that during a graceful shutdown we /// can wait for all the reconcilers to respond to their cancellation tokens. pub(crate) _gate_guard: GateGuard, /// Access to persistent storage for updating generation numbers pub(crate) persistence: Arc<Persistence>, /// HTTP client with proper CA certs. pub(crate) http_client: reqwest::Client, } pub(crate) struct ReconcilerConfigBuilder { config: ReconcilerConfig, } impl ReconcilerConfigBuilder { /// Priority is special: you must pick one thoughtfully, do not just use 'normal' as the default pub(crate) fn new(priority: ReconcilerPriority) -> Self { Self { config: ReconcilerConfig::new(priority), } } pub(crate) fn secondary_warmup_timeout(self, value: Duration) -> Self { Self { config: ReconcilerConfig { secondary_warmup_timeout: Some(value), ..self.config }, } } pub(crate) fn secondary_download_request_timeout(self, value: Duration) -> Self { Self { config: ReconcilerConfig { secondary_download_request_timeout: Some(value), ..self.config }, } } pub(crate) fn tenant_creation_hint(self, hint: bool) -> Self { Self { config: ReconcilerConfig { tenant_creation_hint: hint, ..self.config }, } } pub(crate) fn build(self) -> ReconcilerConfig { self.config } } // Higher priorities are used for user-facing tasks, so that a long backlog of housekeeping work (e.g. reconciling on startup, rescheduling // things on node changes) does not starve user-facing tasks. #[derive(Debug, Copy, Clone)] pub(crate) enum ReconcilerPriority { Normal, High, } #[derive(Debug, Copy, Clone)] pub(crate) struct ReconcilerConfig { pub(crate) priority: ReconcilerPriority, // During live migration give up on warming-up the secondary // after this timeout. secondary_warmup_timeout: Option<Duration>, // During live migrations this is the amount of time that // the pagserver will hold our poll. secondary_download_request_timeout: Option<Duration>, // A hint indicating whether this reconciliation is done on the // creation of a new tenant. This only informs logging behaviour. tenant_creation_hint: bool, } impl ReconcilerConfig { /// Configs are always constructed with an explicit priority, to force callers to think about whether /// the operation they're scheduling is high-priority or not. Normal priority is not a safe default, because /// scheduling something user-facing at normal priority can result in it getting starved out by background work. pub(crate) fn new(priority: ReconcilerPriority) -> Self { Self { priority, secondary_warmup_timeout: None, secondary_download_request_timeout: None, tenant_creation_hint: false, } } pub(crate) fn get_secondary_warmup_timeout(&self) -> Duration { const SECONDARY_WARMUP_TIMEOUT_DEFAULT: Duration = Duration::from_secs(300); self.secondary_warmup_timeout .unwrap_or(SECONDARY_WARMUP_TIMEOUT_DEFAULT) } pub(crate) fn get_secondary_download_request_timeout(&self) -> Duration { const SECONDARY_DOWNLOAD_REQUEST_TIMEOUT_DEFAULT: Duration = Duration::from_secs(20); self.secondary_download_request_timeout .unwrap_or(SECONDARY_DOWNLOAD_REQUEST_TIMEOUT_DEFAULT) } pub(crate) fn tenant_creation_hint(&self) -> bool { self.tenant_creation_hint } } impl From<&MigrationConfig> for ReconcilerConfig { fn from(value: &MigrationConfig) -> Self { // Run reconciler at high priority because MigrationConfig comes from human requests that should // be presumed urgent. let mut builder = ReconcilerConfigBuilder::new(ReconcilerPriority::High); if let Some(timeout) = value.secondary_warmup_timeout { builder = builder.secondary_warmup_timeout(timeout) } if let Some(timeout) = value.secondary_download_request_timeout { builder = builder.secondary_download_request_timeout(timeout) } builder.build() } } /// RAII resource units granted to a Reconciler, which it should keep alive until it finishes doing I/O pub(crate) struct ReconcileUnits { _sem_units: tokio::sync::OwnedSemaphorePermit, } impl ReconcileUnits { pub(crate) fn new(sem_units: tokio::sync::OwnedSemaphorePermit) -> Self { Self { _sem_units: sem_units, } } } /// This is a snapshot of [`crate::tenant_shard::IntentState`], but it does not do any /// reference counting for Scheduler. The IntentState is what the scheduler works with, /// and the TargetState is just the instruction for a particular Reconciler run. #[derive(Debug)] pub(crate) struct TargetState { pub(crate) attached: Option<Node>, pub(crate) secondary: Vec<Node>, } impl TargetState { pub(crate) fn from_intent(nodes: &HashMap<NodeId, Node>, intent: &IntentState) -> Self { Self { attached: intent.get_attached().map(|n| { nodes .get(&n) .expect("Intent attached referenced non-existent node") .clone() }), secondary: intent .get_secondary() .iter() .map(|n| { nodes .get(n) .expect("Intent secondary referenced non-existent node") .clone() }) .collect(), } } } #[derive(thiserror::Error, Debug)] pub(crate) enum ReconcileError { #[error(transparent)] Remote(#[from] mgmt_api::Error), #[error(transparent)] Notify(#[from] NotifyError), #[error("Cancelled")] Cancel, #[error(transparent)] Other(#[from] anyhow::Error), } impl Reconciler { async fn location_config( &mut self, node: &Node, config: LocationConfig, flush_ms: Option<Duration>, lazy: bool, ) -> Result<(), ReconcileError> { if !node.is_available() && config.mode == LocationConfigMode::Detached { // [`crate::service::Service::node_activate_reconcile`] will update the observed state // when the node comes back online. At that point, the intent and observed states will // be mismatched and a background reconciliation will detach. tracing::info!( "Node {node} is unavailable during detach: proceeding anyway, it will be detached via background reconciliation" ); return Ok(()); } self.observed .locations .insert(node.get_id(), ObservedStateLocation { conf: None }); // TODO: amend locations that use long-polling: they will hit this timeout. let timeout = Duration::from_secs(25); tracing::info!("location_config({node}) calling: {:?}", config); let tenant_shard_id = self.tenant_shard_id; let config_ref = &config; match node .with_client_retries( |client| async move { let config = config_ref.clone(); client .location_config(tenant_shard_id, config.clone(), flush_ms, lazy) .await }, &self.http_client, &self.service_config.pageserver_jwt_token, 1, 3, timeout, &self.cancel, ) .await { Some(Ok(_)) => {} Some(Err(e)) => return Err(e.into()), None => return Err(ReconcileError::Cancel), }; tracing::info!("location_config({node}) complete: {:?}", config); match config.mode { LocationConfigMode::Detached => { self.observed.locations.remove(&node.get_id()); } _ => { self.observed .locations .insert(node.get_id(), ObservedStateLocation { conf: Some(config) }); } } Ok(()) } fn get_node(&self, node_id: &NodeId) -> Option<&Node> { if let Some(node) = self.intent.attached.as_ref() { if node.get_id() == *node_id { return Some(node); } } if let Some(node) = self .intent .secondary .iter() .find(|n| n.get_id() == *node_id) { return Some(node); } if let Some(node) = self.detach.iter().find(|n| n.get_id() == *node_id) { return Some(node); } None } async fn maybe_live_migrate(&mut self) -> Result<(), ReconcileError> { let destination = if let Some(node) = &self.intent.attached { match self.observed.locations.get(&node.get_id()) { Some(conf) => { // We will do a live migration only if the intended destination is not // currently in an attached state. match &conf.conf { Some(conf) if conf.mode == LocationConfigMode::Secondary => { // Fall through to do a live migration node } None | Some(_) => { // Attached or uncertain: don't do a live migration, proceed // with a general-case reconciliation tracing::info!("maybe_live_migrate: destination is None or attached"); return Ok(()); } } } None => { // Our destination is not attached: maybe live migrate if some other // node is currently attached. Fall through. node } } } else { // No intent to be attached tracing::info!("maybe_live_migrate: no attached intent"); return Ok(()); }; let mut origin = None; for (node_id, state) in &self.observed.locations { if let Some(observed_conf) = &state.conf { if observed_conf.mode == LocationConfigMode::AttachedSingle { // We will only attempt live migration if the origin is not offline: this // avoids trying to do it while reconciling after responding to an HA failover. if let Some(node) = self.get_node(node_id) { if node.is_available() { origin = Some(node.clone()); break; } } } } } let Some(origin) = origin else { tracing::info!("maybe_live_migrate: no origin found"); return Ok(()); }; // We have an origin and a destination: proceed to do the live migration tracing::info!("Live migrating {}->{}", origin, destination); self.live_migrate(origin, destination.clone()).await?; Ok(()) } async fn wait_lsn( &self, node: &Node, tenant_shard_id: TenantShardId, timelines: HashMap<TimelineId, Lsn>, ) -> Result<StatusCode, ReconcileError> { const TIMEOUT: Duration = Duration::from_secs(10); let client = PageserverClient::new( node.get_id(), self.http_client.clone(), node.base_url(), self.service_config.pageserver_jwt_token.as_deref(), ); client .wait_lsn( tenant_shard_id, TenantWaitLsnRequest { timelines, timeout: TIMEOUT, }, ) .await .map_err(|e| e.into()) } async fn get_lsns( &self, tenant_shard_id: TenantShardId, node: &Node, ) -> anyhow::Result<HashMap<TimelineId, Lsn>> { let client = PageserverClient::new( node.get_id(), self.http_client.clone(), node.base_url(), self.service_config.pageserver_jwt_token.as_deref(), ); let timelines = client.timeline_list(&tenant_shard_id).await?; Ok(timelines .into_iter() .map(|t| (t.timeline_id, t.last_record_lsn)) .collect()) } async fn secondary_download( &self, tenant_shard_id: TenantShardId, node: &Node, ) -> Result<(), ReconcileError> { // This is not the timeout for a request, but the total amount of time we're willing to wait // for a secondary location to get up to date before let total_download_timeout = self.reconciler_config.get_secondary_warmup_timeout(); // This the long-polling interval for the secondary download requests we send to destination pageserver // during a migration. let request_download_timeout = self .reconciler_config .get_secondary_download_request_timeout(); let started_at = Instant::now(); loop { let (status, progress) = match node .with_client_retries( |client| async move { client .tenant_secondary_download( tenant_shard_id, Some(request_download_timeout), ) .await }, &self.http_client, &self.service_config.pageserver_jwt_token, 1, 3, request_download_timeout * 2, &self.cancel, ) .await { None => Err(ReconcileError::Cancel), Some(Ok(v)) => Ok(v), Some(Err(e)) => { // Give up, but proceed: it's unfortunate if we couldn't freshen the destination before // attaching, but we should not let an issue with a secondary location stop us proceeding // with a live migration. tracing::warn!("Failed to prepare by downloading layers on node {node}: {e})"); return Ok(()); } }?; if status == StatusCode::OK { tracing::info!( "Downloads to {} complete: {}/{} layers, {}/{} bytes", node, progress.layers_downloaded, progress.layers_total, progress.bytes_downloaded, progress.bytes_total ); return Ok(()); } else if status == StatusCode::ACCEPTED { let total_runtime = started_at.elapsed(); if total_runtime > total_download_timeout { tracing::warn!( "Timed out after {}ms downloading layers to {node}. Progress so far: {}/{} layers, {}/{} bytes", total_runtime.as_millis(), progress.layers_downloaded, progress.layers_total, progress.bytes_downloaded, progress.bytes_total ); // Give up, but proceed: an incompletely warmed destination doesn't prevent migration working, // it just makes the I/O performance for users less good. return Ok(()); } // Log and proceed around the loop to retry. We don't sleep between requests, because our HTTP call // to the pageserver is a long-poll. tracing::info!( "Downloads to {} not yet complete: {}/{} layers, {}/{} bytes", node, progress.layers_downloaded, progress.layers_total, progress.bytes_downloaded, progress.bytes_total ); } } } /// This function does _not_ mutate any state, so it is cancellation safe. /// /// This function does not respect [`Self::cancel`], callers should handle that. async fn await_lsn( &self, tenant_shard_id: TenantShardId, node: &Node, baseline: HashMap<TimelineId, Lsn>, ) -> anyhow::Result<()> { // Signal to the pageserver that it should ingest up to the baseline LSNs. loop { match self.wait_lsn(node, tenant_shard_id, baseline.clone()).await { Ok(StatusCode::OK) => { // Everything is caught up return Ok(()); } Ok(StatusCode::ACCEPTED) => { // Some timelines are not caught up yet. // They'll be polled below. break; } Ok(StatusCode::NOT_FOUND) => { // None of the timelines are present on the pageserver. // This is correct if they've all been deleted, but // let let the polling loop below cross check. break; } Ok(status_code) => { tracing::warn!( "Unexpected status code ({status_code}) returned by wait_lsn endpoint" ); break; } Err(e) => { tracing::info!("🕑 Can't trigger LSN wait on {node} yet, waiting ({e})",); tokio::time::sleep(Duration::from_millis(500)).await; continue; } } } // Poll the LSNs until they catch up loop { let latest = match self.get_lsns(tenant_shard_id, node).await { Ok(l) => l, Err(e) => { tracing::info!("🕑 Can't get LSNs on node {node} yet, waiting ({e})",); tokio::time::sleep(Duration::from_millis(500)).await; continue; } }; let mut any_behind: bool = false; for (timeline_id, baseline_lsn) in &baseline { match latest.get(timeline_id) { Some(latest_lsn) => { tracing::info!(timeline_id = %timeline_id, "🕑 LSN origin {baseline_lsn} vs destination {latest_lsn}"); if latest_lsn < baseline_lsn { any_behind = true; } } None => { // Timeline was deleted in the meantime - ignore it } } } if !any_behind { tracing::info!("✅ LSN caught up. Proceeding..."); break; } else { tokio::time::sleep(Duration::from_millis(500)).await; } } Ok(()) } pub async fn live_migrate( &mut self, origin_ps: Node, dest_ps: Node, ) -> Result<(), ReconcileError> { // `maybe_live_migrate` is responsibble for sanity of inputs assert!(origin_ps.get_id() != dest_ps.get_id()); fn build_location_config( shard: &ShardIdentity, config: &TenantConfig, mode: LocationConfigMode, generation: Option<Generation>, secondary_conf: Option<LocationConfigSecondary>, ) -> LocationConfig { LocationConfig { mode, generation: generation.map(|g| g.into().unwrap()), secondary_conf, tenant_conf: config.clone(), shard_number: shard.number.0, shard_count: shard.count.literal(), shard_stripe_size: shard.stripe_size.0, } } tracing::info!("🔁 Switching origin node {origin_ps} to stale mode",); // FIXME: it is incorrect to use self.generation here, we should use the generation // from the ObservedState of the origin pageserver (it might be older than self.generation) let stale_conf = build_location_config( &self.shard, &self.config, LocationConfigMode::AttachedStale, self.generation, None, ); self.location_config(&origin_ps, stale_conf, Some(Duration::from_secs(10)), false) .await?; let baseline_lsns = Some(self.get_lsns(self.tenant_shard_id, &origin_ps).await?); // If we are migrating to a destination that has a secondary location, warm it up first if let Some(destination_conf) = self.observed.locations.get(&dest_ps.get_id()) { if let Some(destination_conf) = &destination_conf.conf { if destination_conf.mode == LocationConfigMode::Secondary { tracing::info!("🔁 Downloading latest layers to destination node {dest_ps}",); self.secondary_download(self.tenant_shard_id, &dest_ps) .await?; } } } pausable_failpoint!("reconciler-live-migrate-pre-generation-inc"); // Increment generation before attaching to new pageserver self.generation = Some( self.persistence .increment_generation(self.tenant_shard_id, dest_ps.get_id()) .await?, ); pausable_failpoint!("reconciler-live-migrate-post-generation-inc"); let dest_conf = build_location_config( &self.shard, &self.config, LocationConfigMode::AttachedMulti, self.generation, None, ); tracing::info!("🔁 Attaching to pageserver {dest_ps}"); self.location_config(&dest_ps, dest_conf, None, false) .await?; pausable_failpoint!("reconciler-live-migrate-pre-await-lsn"); if let Some(baseline) = baseline_lsns { tracing::info!("🕑 Waiting for LSN to catch up..."); tokio::select! { r = self.await_lsn(self.tenant_shard_id, &dest_ps, baseline) => {r?;} _ = self.cancel.cancelled() => {return Err(ReconcileError::Cancel)} }; } tracing::info!("🔁 Notifying compute to use pageserver {dest_ps}"); // During a live migration it is unhelpful to proceed if we couldn't notify compute: if we detach // the origin without notifying compute, we will render the tenant unavailable. self.compute_notify_blocking(&origin_ps).await?; pausable_failpoint!("reconciler-live-migrate-post-notify"); // Downgrade the origin to secondary. If the tenant's policy is PlacementPolicy::Attached(0), then // this location will be deleted in the general case reconciliation that runs after this. let origin_secondary_conf = build_location_config( &self.shard, &self.config, LocationConfigMode::Secondary, None, Some(LocationConfigSecondary { warm: true }), ); self.location_config(&origin_ps, origin_secondary_conf.clone(), None, false) .await?; // TODO: we should also be setting the ObservedState on earlier API calls, in case we fail // partway through. In fact, all location conf API calls should be in a wrapper that sets // the observed state to None, then runs, then sets it to what we wrote. self.observed.locations.insert( origin_ps.get_id(), ObservedStateLocation { conf: Some(origin_secondary_conf), }, ); pausable_failpoint!("reconciler-live-migrate-post-detach"); tracing::info!("🔁 Switching to AttachedSingle mode on node {dest_ps}",); let dest_final_conf = build_location_config( &self.shard, &self.config, LocationConfigMode::AttachedSingle, self.generation, None, ); self.location_config(&dest_ps, dest_final_conf.clone(), None, false) .await?; self.observed.locations.insert( dest_ps.get_id(), ObservedStateLocation { conf: Some(dest_final_conf), }, ); tracing::info!("✅ Migration complete"); Ok(()) } /// Returns true if the observed state of the attached location was refreshed /// and false otherwise. async fn maybe_refresh_observed(&mut self) -> Result<bool, ReconcileError> { // If the attached node has uncertain state, read it from the pageserver before proceeding: this // is important to avoid spurious generation increments. // // We don't need to do this for secondary/detach locations because it's harmless to just PUT their // location conf, whereas for attached locations it can interrupt clients if we spuriously destroy/recreate // the `Timeline` object in the pageserver. let Some(attached_node) = self.intent.attached.as_ref() else { // Nothing to do return Ok(false); }; if matches!( self.observed.locations.get(&attached_node.get_id()), Some(ObservedStateLocation { conf: None }) ) { let tenant_shard_id = self.tenant_shard_id; let observed_conf = match attached_node .with_client_retries( |client| async move { client.get_location_config(tenant_shard_id).await }, &self.http_client, &self.service_config.pageserver_jwt_token, 1, 1, Duration::from_secs(5), &self.cancel, ) .await { Some(Ok(observed)) => Some(observed), Some(Err(mgmt_api::Error::ApiError(status, _msg))) if status == StatusCode::NOT_FOUND => { None } Some(Err(e)) => return Err(e.into()), None => return Err(ReconcileError::Cancel), }; tracing::info!("Scanned location configuration on {attached_node}: {observed_conf:?}"); match observed_conf { Some(conf) => { // Pageserver returned a state: update it in observed. This may still be an indeterminate (None) state, // if internally the pageserver's TenantSlot was being mutated (e.g. some long running API call is still running) self.observed .locations .insert(attached_node.get_id(), ObservedStateLocation { conf }); } None => { // Pageserver returned 404: we have confirmation that there is no state for this shard on that pageserver. self.observed.locations.remove(&attached_node.get_id()); } } } Ok(true) } /// Reconciling a tenant makes API calls to pageservers until the observed state /// matches the intended state. /// /// First we apply special case handling (e.g. for live migrations), and then a /// general case reconciliation where we walk through the intent by pageserver /// and call out to the pageserver to apply the desired state. /// /// An Ok(()) result indicates that we successfully attached the tenant, but _not_ that /// all locations for the tenant are in the expected state. When nodes that are to be detached /// or configured as secondary are unavailable, we may return Ok(()) but leave the shard in a /// state where it still requires later reconciliation. pub(crate) async fn reconcile(&mut self) -> Result<(), ReconcileError> { // Prepare: if we have uncertain `observed` state for our would-be attachement location, then refresh it let refreshed = self.maybe_refresh_observed().await?; // Special case: live migration self.maybe_live_migrate().await?; // If the attached pageserver is not attached, do so now. if let Some(node) = self.intent.attached.as_ref() { // If we are in an attached policy, then generation must have been set (null generations // are only present when a tenant is initially loaded with a secondary policy)
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/lib.rs
storage_controller/src/lib.rs
use serde::Serialize; use utils::seqwait::MonotonicCounter; extern crate hyper0 as hyper; mod auth; mod background_node_operations; mod compute_hook; pub mod hadron_utils; mod heartbeater; pub mod http; mod id_lock_map; mod leadership; pub mod metrics; mod node; mod operation_utils; mod pageserver_client; mod peer_client; pub mod persistence; mod reconciler; mod safekeeper; mod safekeeper_client; mod scheduler; mod schema; pub mod service; mod tenant_shard; mod timeline_import; #[derive(Ord, PartialOrd, Eq, PartialEq, Copy, Clone, Serialize)] struct Sequence(u64); impl Sequence { fn initial() -> Self { Self(0) } } impl std::fmt::Display for Sequence { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "{}", self.0) } } impl std::fmt::Debug for Sequence { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "{}", self.0) } } impl MonotonicCounter<Sequence> for Sequence { fn cnt_advance(&mut self, v: Sequence) { assert!(*self <= v); *self = v; } fn cnt_value(&self) -> Sequence { *self } } impl Sequence { fn next(&self) -> Sequence { Sequence(self.0 + 1) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/background_node_operations.rs
storage_controller/src/background_node_operations.rs
use std::borrow::Cow; use std::fmt::{Debug, Display}; use tokio_util::sync::CancellationToken; use utils::id::NodeId; pub(crate) const MAX_RECONCILES_PER_OPERATION: usize = 64; #[derive(Copy, Clone)] pub(crate) struct Delete { pub(crate) node_id: NodeId, } #[derive(Copy, Clone)] pub(crate) struct Drain { pub(crate) node_id: NodeId, } #[derive(Copy, Clone)] pub(crate) struct Fill { pub(crate) node_id: NodeId, } #[derive(Copy, Clone)] pub(crate) enum Operation { Delete(Delete), Drain(Drain), Fill(Fill), } #[derive(Debug, thiserror::Error)] pub(crate) enum OperationError { #[error("Node state changed during operation: {0}")] NodeStateChanged(Cow<'static, str>), #[error("Operation finalize error: {0}")] FinalizeError(Cow<'static, str>), #[error("Operation cancelled")] Cancelled, #[error("Impossible constraint error: {0}")] ImpossibleConstraint(Cow<'static, str>), } pub(crate) struct OperationHandler { pub(crate) operation: Operation, #[allow(unused)] pub(crate) cancel: CancellationToken, } impl Display for Delete { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "delete {}", self.node_id) } } impl Display for Drain { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "drain {}", self.node_id) } } impl Display for Fill { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { write!(f, "fill {}", self.node_id) } } impl Display for Operation { fn fmt(&self, f: &mut std::fmt::Formatter) -> std::fmt::Result { match self { Operation::Delete(op) => write!(f, "{op}"), Operation::Drain(op) => write!(f, "{op}"), Operation::Fill(op) => write!(f, "{op}"), } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/http.rs
storage_controller/src/http.rs
use std::str::FromStr; use std::sync::{Arc, LazyLock}; use std::time::{Duration, Instant}; use anyhow::Context; use control_plane::storage_controller::{AttachHookRequest, InspectRequest}; use futures::Future; use http_utils::endpoint::{ self, auth_middleware, check_permission_with, profile_cpu_handler, profile_heap_handler, request_span, }; use http_utils::error::ApiError; use http_utils::failpoints::failpoints_handler; use http_utils::json::{json_request, json_response}; use http_utils::request::{must_get_query_param, parse_query_param, parse_request_param}; use http_utils::{RequestExt, RouterBuilder}; use hyper::header::CONTENT_TYPE; use hyper::{Body, Request, Response, StatusCode, Uri}; use metrics::{BuildInfo, NeonMetrics}; use pageserver_api::controller_api::{ MetadataHealthListOutdatedRequest, MetadataHealthListOutdatedResponse, MetadataHealthListUnhealthyResponse, MetadataHealthUpdateRequest, MetadataHealthUpdateResponse, NodeAvailability, NodeConfigureRequest, NodeRegisterRequest, SafekeeperSchedulingPolicyRequest, ShardsPreferredAzsRequest, TenantCreateRequest, TenantPolicyRequest, TenantShardMigrateRequest, TimelineImportRequest, TimelineSafekeeperMigrateRequest, }; use pageserver_api::models::{ DetachBehavior, LsnLeaseRequest, TenantConfigPatchRequest, TenantConfigRequest, TenantLocationConfigRequest, TenantShardSplitRequest, TenantTimeTravelRequest, TimelineArchivalConfigRequest, TimelineCreateRequest, }; use pageserver_api::shard::TenantShardId; use pageserver_api::upcall_api::{ PutTimelineImportStatusRequest, ReAttachRequest, TimelineImportStatusRequest, ValidateRequest, }; use pageserver_client::{BlockUnblock, mgmt_api}; use routerify::Middleware; use tokio_util::sync::CancellationToken; use tracing::warn; use utils::auth::{Scope, SwappableJwtAuth}; use utils::id::{NodeId, TenantId, TimelineId}; use crate::http; use crate::metrics::{ HttpRequestLatencyLabelGroup, HttpRequestStatusLabelGroup, METRICS_REGISTRY, PageserverRequestLabelGroup, }; use crate::persistence::SafekeeperUpsert; use crate::reconciler::ReconcileError; use crate::service::{ LeadershipStatus, RECONCILE_TIMEOUT, STARTUP_RECONCILE_TIMEOUT, Service, TenantMutationLocations, }; /// State available to HTTP request handlers pub struct HttpState { service: Arc<crate::service::Service>, auth: Option<Arc<SwappableJwtAuth>>, rate_limiter: governor::DefaultKeyedRateLimiter<TenantId>, neon_metrics: NeonMetrics, allowlist_routes: &'static [&'static str], } impl HttpState { pub fn new( service: Arc<crate::service::Service>, auth: Option<Arc<SwappableJwtAuth>>, build_info: BuildInfo, ) -> Self { let quota = governor::Quota::per_second(service.get_config().tenant_rate_limit); Self { service, auth, rate_limiter: governor::RateLimiter::keyed(quota), neon_metrics: NeonMetrics::new(build_info), allowlist_routes: &[ "/status", "/live", "/ready", "/metrics", "/profile/cpu", "/profile/heap", ], } } } #[inline(always)] fn get_state(request: &Request<Body>) -> &HttpState { request .data::<Arc<HttpState>>() .expect("unknown state type") .as_ref() } /// Rate limits tenant requests. /// /// TODO: this should be a request middleware, but requires us to extract the tenant ID from /// different URLs in a systematic way. /// /// TODO: consider returning a 429 response if these start piling up. async fn maybe_rate_limit(request: &Request<Body>, tenant_id: TenantId) { // Check if the tenant should be rate-limited. let rate_limiter = &get_state(request).rate_limiter; if rate_limiter.check_key(&tenant_id).is_ok() { return; } // Measure the rate limiting delay. let _timer = METRICS_REGISTRY .metrics_group .storage_controller_http_request_rate_limited .start_timer(); // Log rate limited tenants once every 10 seconds. static LOG_RATE_LIMITER: LazyLock<governor::DefaultKeyedRateLimiter<TenantId>> = LazyLock::new(|| { let quota = governor::Quota::with_period(Duration::from_secs(10)).unwrap(); governor::RateLimiter::keyed(quota) }); if LOG_RATE_LIMITER.check_key(&tenant_id).is_ok() { warn!("tenant {tenant_id} is rate limited") } // Wait for quota. rate_limiter.until_key_ready(&tenant_id).await; } /// Pageserver calls into this on startup, to learn which tenants it should attach async fn handle_re_attach(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::GenerationsApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let reattach_req = json_request::<ReAttachRequest>(&mut req).await?; let state = get_state(&req); json_response(StatusCode::OK, state.service.re_attach(reattach_req).await?) } /// Pageserver calls into this before doing deletions, to confirm that it still /// holds the latest generation for the tenants with deletions enqueued async fn handle_validate(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::GenerationsApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let validate_req = json_request::<ValidateRequest>(&mut req).await?; let state = get_state(&req); json_response(StatusCode::OK, state.service.validate(validate_req).await?) } async fn handle_get_timeline_import_status(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::GenerationsApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let get_req = json_request::<TimelineImportStatusRequest>(&mut req).await?; let state = get_state(&req); json_response( StatusCode::OK, state .service .handle_timeline_shard_import_progress(get_req) .await?, ) } async fn handle_put_timeline_import_status(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::GenerationsApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let put_req = json_request::<PutTimelineImportStatusRequest>(&mut req).await?; let state = get_state(&req); json_response( StatusCode::OK, state .service .handle_timeline_shard_import_progress_upcall(put_req) .await?, ) } /// Call into this before attaching a tenant to a pageserver, to acquire a generation number /// (in the real control plane this is unnecessary, because the same program is managing /// generation numbers and doing attachments). async fn handle_attach_hook(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::Admin)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let attach_req = json_request::<AttachHookRequest>(&mut req).await?; let state = get_state(&req); json_response( StatusCode::OK, state .service .attach_hook(attach_req) .await .map_err(ApiError::InternalServerError)?, ) } async fn handle_inspect(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::Admin)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let inspect_req = json_request::<InspectRequest>(&mut req).await?; let state = get_state(&req); json_response(StatusCode::OK, state.service.inspect(inspect_req)) } async fn handle_tenant_create( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::PageServerApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let create_req = json_request::<TenantCreateRequest>(&mut req).await?; json_response( StatusCode::CREATED, service.tenant_create(create_req).await?, ) } async fn handle_tenant_location_config( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_shard_id: TenantShardId = parse_request_param(&req, "tenant_shard_id")?; check_permissions(&req, Scope::PageServerApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let config_req = json_request::<TenantLocationConfigRequest>(&mut req).await?; json_response( StatusCode::OK, service .tenant_location_config(tenant_shard_id, config_req) .await?, ) } async fn handle_tenant_config_patch( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::PageServerApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let config_req = json_request::<TenantConfigPatchRequest>(&mut req).await?; json_response( StatusCode::OK, service.tenant_config_patch(config_req).await?, ) } async fn handle_tenant_config_set( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::PageServerApi)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let config_req = json_request::<TenantConfigRequest>(&mut req).await?; json_response(StatusCode::OK, service.tenant_config_set(config_req).await?) } async fn handle_tenant_config_get( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; json_response(StatusCode::OK, service.tenant_config_get(tenant_id)?) } async fn handle_tenant_time_travel_remote_storage( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let time_travel_req = json_request::<TenantTimeTravelRequest>(&mut req).await?; let timestamp_raw = must_get_query_param(&req, "travel_to")?; let _timestamp = humantime::parse_rfc3339(&timestamp_raw).map_err(|_e| { ApiError::BadRequest(anyhow::anyhow!( "Invalid time for travel_to: {timestamp_raw:?}" )) })?; let done_if_after_raw = must_get_query_param(&req, "done_if_after")?; let _done_if_after = humantime::parse_rfc3339(&done_if_after_raw).map_err(|_e| { ApiError::BadRequest(anyhow::anyhow!( "Invalid time for done_if_after: {done_if_after_raw:?}" )) })?; service .tenant_time_travel_remote_storage( &time_travel_req, tenant_id, timestamp_raw, done_if_after_raw, ) .await?; json_response(StatusCode::OK, ()) } fn map_reqwest_hyper_status(status: reqwest::StatusCode) -> Result<hyper::StatusCode, ApiError> { hyper::StatusCode::from_u16(status.as_u16()) .context("invalid status code") .map_err(ApiError::InternalServerError) } async fn handle_tenant_secondary_download( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let wait = parse_query_param(&req, "wait_ms")?.map(Duration::from_millis); maybe_rate_limit(&req, tenant_id).await; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; let (status, progress) = service.tenant_secondary_download(tenant_id, wait).await?; json_response(map_reqwest_hyper_status(status)?, progress) } async fn handle_tenant_delete( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; let status_code = service .tenant_delete(tenant_id) .await .and_then(map_reqwest_hyper_status)?; if status_code == StatusCode::NOT_FOUND { // The pageserver uses 404 for successful deletion, but we use 200 json_response(StatusCode::OK, ()) } else { json_response(status_code, ()) } } async fn handle_tenant_timeline_create( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let create_req = json_request::<TimelineCreateRequest>(&mut req).await?; json_response( StatusCode::CREATED, service .tenant_timeline_create(tenant_id, create_req) .await?, ) } async fn handle_tenant_timeline_delete( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; service .maybe_delete_timeline_import(tenant_id, timeline_id) .await?; // For timeline deletions, which both implement an "initially return 202, then 404 once // we're done" semantic, we wrap with a retry loop to expose a simpler API upstream. async fn deletion_wrapper<R, F>(service: Arc<Service>, f: F) -> Result<Response<Body>, ApiError> where R: std::future::Future<Output = Result<StatusCode, ApiError>> + Send + 'static, F: Fn(Arc<Service>) -> R + Send + Sync + 'static, { // On subsequent retries, wait longer. // Enable callers with a 25 second request timeout to reliably get a response const MAX_WAIT: Duration = Duration::from_secs(25); const MAX_RETRY_PERIOD: Duration = Duration::from_secs(5); let started_at = Instant::now(); // To keep deletion reasonably snappy for small tenants, initially check after 1 second if deletion // completed. let mut retry_period = Duration::from_secs(1); loop { let status = f(service.clone()).await?; match status { StatusCode::ACCEPTED => { tracing::info!("Deletion accepted, waiting to try again..."); tokio::time::sleep(retry_period).await; retry_period = MAX_RETRY_PERIOD; } StatusCode::CONFLICT => { tracing::info!("Deletion already in progress, waiting to try again..."); tokio::time::sleep(retry_period).await; } StatusCode::NOT_FOUND => { tracing::info!("Deletion complete"); return json_response(StatusCode::OK, ()); } _ => { tracing::warn!("Unexpected status {status}"); return json_response(status, ()); } } let now = Instant::now(); if now + retry_period > started_at + MAX_WAIT { tracing::info!("Deletion timed out waiting for 404"); // REQUEST_TIMEOUT would be more appropriate, but CONFLICT is already part of // the pageserver's swagger definition for this endpoint, and has the same desired // effect of causing the control plane to retry later. return json_response(StatusCode::CONFLICT, ()); } } } deletion_wrapper(service, move |service| async move { service .tenant_timeline_delete(tenant_id, timeline_id) .await .and_then(map_reqwest_hyper_status) }) .await } async fn handle_tenant_timeline_archival_config( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let create_req = json_request::<TimelineArchivalConfigRequest>(&mut req).await?; service .tenant_timeline_archival_config(tenant_id, timeline_id, create_req) .await?; json_response(StatusCode::OK, ()) } async fn handle_tenant_timeline_detach_ancestor( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; let behavior: Option<DetachBehavior> = parse_query_param(&req, "detach_behavior")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; let res = service .tenant_timeline_detach_ancestor(tenant_id, timeline_id, behavior) .await?; json_response(StatusCode::OK, res) } async fn handle_tenant_timeline_block_unblock_gc( service: Arc<Service>, req: Request<Body>, dir: BlockUnblock, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; service .tenant_timeline_block_unblock_gc(tenant_id, timeline_id, dir) .await?; json_response(StatusCode::OK, ()) } async fn handle_tenant_timeline_download_heatmap_layers( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_shard_id: TenantShardId = parse_request_param(&req, "tenant_shard_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_shard_id.tenant_id).await; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; let concurrency: Option<usize> = parse_query_param(&req, "concurrency")?; let recurse = parse_query_param(&req, "recurse")?.unwrap_or(false); service .tenant_timeline_download_heatmap_layers(tenant_shard_id, timeline_id, concurrency, recurse) .await?; json_response(StatusCode::OK, ()) } async fn handle_tenant_timeline_safekeeper_migrate( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; // TODO(diko): it's not PS operation, there should be a different permission scope. check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let migrate_req = json_request::<TimelineSafekeeperMigrateRequest>(&mut req).await?; service .tenant_timeline_safekeeper_migrate(tenant_id, timeline_id, migrate_req) .await?; json_response(StatusCode::OK, ()) } async fn handle_tenant_timeline_safekeeper_migrate_abort( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; // TODO(diko): it's not PS operation, there should be a different permission scope. check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; service .tenant_timeline_safekeeper_migrate_abort(tenant_id, timeline_id) .await?; json_response(StatusCode::OK, ()) } async fn handle_tenant_timeline_lsn_lease( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_id).await; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let lsn_lease_request = json_request::<LsnLeaseRequest>(&mut req).await?; service .tenant_timeline_lsn_lease(tenant_id, timeline_id, lsn_lease_request.lsn) .await?; json_response(StatusCode::OK, ()) } // For metric labels where we would like to include the approximate path, but exclude high-cardinality fields like query parameters // and tenant/timeline IDs. Since we are proxying to arbitrary paths, we don't have routing templates to // compare to, so we can just filter out our well known ID format with regexes. fn path_without_ids(path: &str) -> String { static ID_REGEX: std::sync::OnceLock<regex::Regex> = std::sync::OnceLock::new(); ID_REGEX .get_or_init(|| regex::Regex::new(r"([0-9a-fA-F]{32}(-[0-9]{4})?|\?.*)").unwrap()) .replace_all(path, "") .to_string() } async fn handle_tenant_timeline_passthrough( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_or_shard_id: TenantShardId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::PageServerApi)?; maybe_rate_limit(&req, tenant_or_shard_id.tenant_id).await; let req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, }; let Some(path) = req.uri().path_and_query() else { // This should never happen, our request router only calls us if there is a path return Err(ApiError::BadRequest(anyhow::anyhow!("Missing path"))); }; let method = match *req.method() { hyper::Method::GET => reqwest::Method::GET, hyper::Method::POST => reqwest::Method::POST, hyper::Method::PUT => reqwest::Method::PUT, hyper::Method::DELETE => reqwest::Method::DELETE, hyper::Method::PATCH => reqwest::Method::PATCH, _ => return Err(ApiError::BadRequest(anyhow::anyhow!("Unsupported method"))), }; tracing::info!( "Proxying request for tenant {} ({})", tenant_or_shard_id.tenant_id, path ); let tenant_shard_id = if tenant_or_shard_id.is_unsharded() { // If the request contains only tenant ID, find the node that holds shard zero let (_, shard_id) = service .tenant_shard0_node(tenant_or_shard_id.tenant_id) .await?; shard_id } else { tenant_or_shard_id }; let service_inner = service.clone(); service.tenant_shard_remote_mutation(tenant_shard_id, |locations| async move { let TenantMutationLocations(locations) = locations; if locations.is_empty() { return Err(ApiError::NotFound(anyhow::anyhow!("Tenant {} not found", tenant_or_shard_id.tenant_id).into())); } let (tenant_or_shard_id, locations) = locations.into_iter().next().unwrap(); let node = locations.latest.node; // Callers will always pass an unsharded tenant ID. Before proxying, we must // rewrite this to a shard-aware shard zero ID. let path = format!("{path}"); let tenant_str = tenant_or_shard_id.tenant_id.to_string(); let tenant_shard_str = format!("{tenant_shard_id}"); let path = path.replace(&tenant_str, &tenant_shard_str); let latency = &METRICS_REGISTRY .metrics_group .storage_controller_passthrough_request_latency; let path_label = path_without_ids(&path) .split('/') .filter(|token| !token.is_empty()) .collect::<Vec<_>>() .join("_"); let labels = PageserverRequestLabelGroup { pageserver_id: &node.get_id().to_string(), path: &path_label, method: crate::metrics::Method::Get, }; let _timer = latency.start_timer(labels.clone()); let client = mgmt_api::Client::new( service_inner.get_http_client().clone(), node.base_url(), service_inner.get_config().pageserver_jwt_token.as_deref(), ); let resp = client.op_raw(method, path).await.map_err(|e| // We return 503 here because if we can't successfully send a request to the pageserver, // either we aren't available or the pageserver is unavailable. ApiError::ResourceUnavailable(format!("Error sending pageserver API request to {node}: {e}").into()))?; if !resp.status().is_success() { let error_counter = &METRICS_REGISTRY .metrics_group .storage_controller_passthrough_request_error; error_counter.inc(labels); } let resp_staus = resp.status(); // We have a reqest::Response, would like a http::Response let mut builder = hyper::Response::builder().status(map_reqwest_hyper_status(resp_staus)?); for (k, v) in resp.headers() { builder = builder.header(k.as_str(), v.as_bytes()); } let resp_bytes = resp .bytes() .await .map_err(|e| ApiError::InternalServerError(e.into()))?; // Inspect 404 errors: at this point, we know that the tenant exists, but the pageserver we route // the request to might not yet be ready. Therefore, if it is a _tenant_ not found error, we can // convert it into a 503. TODO: we should make this part of the check in `tenant_shard_remote_mutation`. // However, `tenant_shard_remote_mutation` currently cannot inspect the HTTP error response body, // so we have to do it here instead. if resp_staus == reqwest::StatusCode::NOT_FOUND { let resp_str = std::str::from_utf8(&resp_bytes) .map_err(|e| ApiError::InternalServerError(e.into()))?; // We only handle "tenant not found" errors; other 404s like timeline not found should // be forwarded as-is. if Service::is_tenant_not_found_error(resp_str, tenant_or_shard_id.tenant_id) { // Rather than retry here, send the client a 503 to prompt a retry: this matches // the pageserver's use of 503, and all clients calling this API should retry on 503. return Err(ApiError::ResourceUnavailable( format!( "Pageserver {node} returned tenant 404 due to ongoing migration, retry later" ) .into(), )); } } let response = builder .body(Body::from(resp_bytes)) .map_err(|e| ApiError::InternalServerError(e.into()))?; Ok(response) }).await? } async fn handle_tenant_locate( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::Admin)?; // NB: don't rate limit: admin operation. match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; json_response(StatusCode::OK, service.tenant_locate(tenant_id)?) } async fn handle_tenant_describe( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; check_permissions(&req, Scope::Scrubber)?; // NB: don't rate limit: scrubber operation. match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; json_response(StatusCode::OK, service.tenant_describe(tenant_id)?) } /* BEGIN_HADRON */ async fn handle_tenant_timeline_describe( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::Scrubber)?; let tenant_id: TenantId = parse_request_param(&req, "tenant_id")?; let timeline_id: TimelineId = parse_request_param(&req, "timeline_id")?; match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; json_response( StatusCode::OK, service .tenant_timeline_describe(tenant_id, timeline_id) .await?, ) } /* END_HADRON */ async fn handle_tenant_list( service: Arc<Service>, req: Request<Body>, ) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::Admin)?; let limit: Option<usize> = parse_query_param(&req, "limit")?; let start_after: Option<TenantId> = parse_query_param(&req, "start_after")?; tracing::info!("start_after: {:?}", start_after); match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(_req) => {} }; json_response(StatusCode::OK, service.tenant_list(limit, start_after)) } async fn handle_node_register(req: Request<Body>) -> Result<Response<Body>, ApiError> { check_permissions(&req, Scope::Infra)?; let mut req = match maybe_forward(req).await { ForwardOutcome::Forwarded(res) => { return res; } ForwardOutcome::NotForwarded(req) => req, };
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/compute_hook.rs
storage_controller/src/compute_hook.rs
use std::borrow::Cow; use std::collections::HashMap; use std::error::Error as _; use std::sync::Arc; use std::time::Duration; use anyhow::Context; use compute_api::spec::PageserverProtocol; use compute_api::spec::PageserverShardInfo; use control_plane::endpoint::{ ComputeControlPlane, EndpointStatus, PageserverConnectionInfo, PageserverShardConnectionInfo, }; use control_plane::local_env::LocalEnv; use futures::StreamExt; use hyper::StatusCode; use pageserver_api::config::DEFAULT_GRPC_LISTEN_PORT; use pageserver_api::controller_api::AvailabilityZone; use pageserver_api::shard::{ShardCount, ShardIndex, ShardNumber, ShardStripeSize, TenantShardId}; use postgres_connection::parse_host_port; use safekeeper_api::membership::SafekeeperGeneration; use serde::{Deserialize, Serialize}; use tokio_util::sync::CancellationToken; use tracing::{Instrument, info_span}; use utils::backoff::{self}; use utils::id::{NodeId, TenantId, TenantTimelineId, TimelineId}; use crate::service::Config; const SLOWDOWN_DELAY: Duration = Duration::from_secs(5); const NOTIFY_REQUEST_TIMEOUT: Duration = Duration::from_secs(10); pub(crate) const API_CONCURRENCY: usize = 32; struct UnshardedComputeHookTenant { // Which node is this tenant attached to node_id: NodeId, // The tenant's preferred AZ, so that we may pass this on to the control plane preferred_az: Option<AvailabilityZone>, // Must hold this lock to send a notification. send_lock: Arc<tokio::sync::Mutex<Option<ComputeRemoteTenantState>>>, } struct ShardedComputeHookTenant { stripe_size: ShardStripeSize, shard_count: ShardCount, shards: Vec<(ShardNumber, NodeId)>, // The tenant's preferred AZ, so that we may pass this on to the control plane preferred_az: Option<AvailabilityZone>, // Must hold this lock to send a notification. The contents represent // the last successfully sent notification, and are used to coalesce multiple // updates by only sending when there is a chance since our last successful send. send_lock: Arc<tokio::sync::Mutex<Option<ComputeRemoteTenantState>>>, } /// Represents our knowledge of the compute's state: we can update this when we get a /// response from a notify API call, which tells us what has been applied. /// /// Should be wrapped in an Option<>, as we cannot always know the remote state. #[derive(PartialEq, Eq, Debug)] struct ComputeRemoteState<R> { // The request body which was acked by the compute request: R, // Whether the cplane indicated that the state was applied to running computes, or just // persisted. In the Neon control plane, this is the difference between a 423 response (meaning // persisted but not applied), and a 2xx response (both persisted and applied) applied: bool, } type ComputeRemoteTenantState = ComputeRemoteState<NotifyAttachRequest>; type ComputeRemoteTimelineState = ComputeRemoteState<NotifySafekeepersRequest>; /// The trait which define the handler-specific types and methods. /// We have two implementations of this trait so far: /// - [`ComputeHookTenant`] for tenant attach notifications ("/notify-attach") /// - [`ComputeHookTimeline`] for safekeeper change notifications ("/notify-safekeepers") trait ApiMethod { /// Type of the key which identifies the resource. /// It's either TenantId for tenant attach notifications, /// or TenantTimelineId for safekeeper change notifications. type Key: std::cmp::Eq + std::hash::Hash + Clone; type Request: serde::Serialize + std::fmt::Debug; const API_PATH: &'static str; fn maybe_send( &self, key: Self::Key, lock: Option<tokio::sync::OwnedMutexGuard<Option<ComputeRemoteState<Self::Request>>>>, ) -> MaybeSendResult<Self::Request, Self::Key>; async fn notify_local( env: &LocalEnv, cplane: &ComputeControlPlane, req: &Self::Request, ) -> Result<(), NotifyError>; } enum ComputeHookTenant { Unsharded(UnshardedComputeHookTenant), Sharded(ShardedComputeHookTenant), } impl ComputeHookTenant { /// Construct with at least one shard's information fn new( tenant_shard_id: TenantShardId, stripe_size: ShardStripeSize, preferred_az: Option<AvailabilityZone>, node_id: NodeId, ) -> Self { if tenant_shard_id.shard_count.count() > 1 { Self::Sharded(ShardedComputeHookTenant { shards: vec![(tenant_shard_id.shard_number, node_id)], stripe_size, shard_count: tenant_shard_id.shard_count, preferred_az, send_lock: Arc::default(), }) } else { Self::Unsharded(UnshardedComputeHookTenant { node_id, preferred_az, send_lock: Arc::default(), }) } } fn get_send_lock(&self) -> &Arc<tokio::sync::Mutex<Option<ComputeRemoteTenantState>>> { match self { Self::Unsharded(unsharded_tenant) => &unsharded_tenant.send_lock, Self::Sharded(sharded_tenant) => &sharded_tenant.send_lock, } } fn is_sharded(&self) -> bool { matches!(self, ComputeHookTenant::Sharded(_)) } /// Clear compute hook state for the specified shard. /// Only valid for [`ComputeHookTenant::Sharded`] instances. fn remove_shard(&mut self, tenant_shard_id: TenantShardId, stripe_size: ShardStripeSize) { match self { ComputeHookTenant::Sharded(sharded) => { if sharded.stripe_size != stripe_size || sharded.shard_count != tenant_shard_id.shard_count { tracing::warn!("Shard split detected while handling detach") } let shard_idx = sharded.shards.iter().position(|(shard_number, _node_id)| { *shard_number == tenant_shard_id.shard_number }); if let Some(shard_idx) = shard_idx { sharded.shards.remove(shard_idx); } else { // This is a valid but niche case, where the tenant was previously attached // as a Secondary location and then detached, so has no previously notified // state. tracing::info!("Shard not found while handling detach") } } ComputeHookTenant::Unsharded(_) => { unreachable!("Detach of unsharded tenants is handled externally"); } } } /// Set one shard's location. If stripe size or shard count have changed, Self is reset /// and drops existing content. fn update(&mut self, shard_update: ShardUpdate) { let tenant_shard_id = shard_update.tenant_shard_id; let node_id = shard_update.node_id; let stripe_size = shard_update.stripe_size; let preferred_az = shard_update.preferred_az; match self { Self::Unsharded(unsharded_tenant) if tenant_shard_id.shard_count.count() == 1 => { unsharded_tenant.node_id = node_id; if unsharded_tenant.preferred_az.as_ref() != preferred_az.as_ref().map(|az| az.as_ref()) { unsharded_tenant.preferred_az = preferred_az.map(|az| az.as_ref().clone()); } } Self::Sharded(sharded_tenant) if sharded_tenant.stripe_size == stripe_size && sharded_tenant.shard_count == tenant_shard_id.shard_count => { if let Some(existing) = sharded_tenant .shards .iter() .position(|s| s.0 == tenant_shard_id.shard_number) { sharded_tenant.shards.get_mut(existing).unwrap().1 = node_id; } else { sharded_tenant .shards .push((tenant_shard_id.shard_number, node_id)); sharded_tenant.shards.sort_by_key(|s| s.0) } if sharded_tenant.preferred_az.as_ref() != preferred_az.as_ref().map(|az| az.as_ref()) { sharded_tenant.preferred_az = preferred_az.map(|az| az.as_ref().clone()); } } _ => { // Shard count changed: reset struct. *self = Self::new( tenant_shard_id, stripe_size, preferred_az.map(|az| az.into_owned()), node_id, ); } } } } /// The state of a timeline we need to notify the compute about. struct ComputeHookTimeline { generation: SafekeeperGeneration, safekeepers: Vec<SafekeeperInfo>, send_lock: Arc<tokio::sync::Mutex<Option<ComputeRemoteTimelineState>>>, } impl ComputeHookTimeline { /// Construct a new ComputeHookTimeline with the given safekeepers and generation. fn new(generation: SafekeeperGeneration, safekeepers: Vec<SafekeeperInfo>) -> Self { Self { generation, safekeepers, send_lock: Arc::default(), } } /// Update the state with a new SafekeepersUpdate. /// Noop if the update generation is not greater than the current generation. fn update(&mut self, sk_update: SafekeepersUpdate) { if sk_update.generation > self.generation { self.generation = sk_update.generation; self.safekeepers = sk_update.safekeepers; } } } impl ApiMethod for ComputeHookTimeline { type Key = TenantTimelineId; type Request = NotifySafekeepersRequest; const API_PATH: &'static str = "notify-safekeepers"; fn maybe_send( &self, ttid: TenantTimelineId, lock: Option<tokio::sync::OwnedMutexGuard<Option<ComputeRemoteTimelineState>>>, ) -> MaybeSendNotifySafekeepersResult { let locked = match lock { Some(already_locked) => already_locked, None => { // Lock order: this _must_ be only a try_lock, because we are called inside of the [`ComputeHook::timelines`] lock. let Ok(locked) = self.send_lock.clone().try_lock_owned() else { return MaybeSendResult::AwaitLock((ttid, self.send_lock.clone())); }; locked } }; if locked .as_ref() .is_some_and(|s| s.request.generation >= self.generation) { return MaybeSendResult::Noop; } MaybeSendResult::Transmit(( NotifySafekeepersRequest { tenant_id: ttid.tenant_id, timeline_id: ttid.timeline_id, generation: self.generation, safekeepers: self.safekeepers.clone(), }, locked, )) } async fn notify_local( _env: &LocalEnv, cplane: &ComputeControlPlane, req: &NotifySafekeepersRequest, ) -> Result<(), NotifyError> { let NotifySafekeepersRequest { tenant_id, timeline_id, generation, safekeepers, } = req; for (endpoint_name, endpoint) in &cplane.endpoints { if endpoint.tenant_id == *tenant_id && endpoint.timeline_id == *timeline_id && endpoint.status() == EndpointStatus::Running { tracing::info!("Reconfiguring safekeepers for endpoint {endpoint_name}"); let safekeepers = safekeepers.iter().map(|sk| sk.id).collect::<Vec<_>>(); endpoint .reconfigure_safekeepers(safekeepers, *generation) .await .map_err(NotifyError::NeonLocal)?; } } Ok(()) } } #[derive(Serialize, Deserialize, Debug, Eq, PartialEq)] struct NotifyAttachRequestShard { node_id: NodeId, shard_number: ShardNumber, } /// Request body that we send to the control plane to notify it of where a tenant is attached #[derive(Serialize, Deserialize, Debug, Eq, PartialEq)] struct NotifyAttachRequest { tenant_id: TenantId, preferred_az: Option<String>, stripe_size: Option<ShardStripeSize>, shards: Vec<NotifyAttachRequestShard>, } #[derive(Serialize, Deserialize, Debug, Eq, PartialEq, Clone)] pub(crate) struct SafekeeperInfo { pub id: NodeId, /// Hostname of the safekeeper. /// It exists for better debuggability. Might be missing. /// Should not be used for anything else. pub hostname: Option<String>, } #[derive(Serialize, Deserialize, Debug, Eq, PartialEq)] struct NotifySafekeepersRequest { tenant_id: TenantId, timeline_id: TimelineId, generation: SafekeeperGeneration, safekeepers: Vec<SafekeeperInfo>, } /// Error type for attempts to call into the control plane compute notification hook #[derive(thiserror::Error, Debug)] pub(crate) enum NotifyError { // Request was not send successfully, e.g. transport error #[error("Sending request: {0}{}", .0.source().map(|e| format!(": {e}")).unwrap_or_default())] Request(#[from] reqwest::Error), // Request could not be serviced right now due to ongoing Operation in control plane, but should be possible soon. #[error("Control plane tenant busy")] Busy, // Explicit 429 response asking us to retry less frequently #[error("Control plane overloaded")] SlowDown, // A 503 response indicates the control plane can't handle the request right now #[error("Control plane unavailable (status {0})")] Unavailable(StatusCode), // API returned unexpected non-success status. We will retry, but log a warning. #[error("Control plane returned unexpected status {0}")] Unexpected(StatusCode), // We shutdown while sending #[error("Shutting down")] ShuttingDown, // A response indicates we will never succeed, such as 400 or 403 #[error("Non-retryable error {0}")] Fatal(StatusCode), #[error("neon_local error: {0}")] NeonLocal(anyhow::Error), } enum MaybeSendResult<R, K> { // Please send this request while holding the lock, and if you succeed then write // the request into the lock. Transmit( ( R, tokio::sync::OwnedMutexGuard<Option<ComputeRemoteState<R>>>, ), ), // Something requires sending, but you must wait for a current sender then call again AwaitLock((K, Arc<tokio::sync::Mutex<Option<ComputeRemoteState<R>>>>)), // Nothing requires sending Noop, } type MaybeSendNotifyAttachResult = MaybeSendResult<NotifyAttachRequest, TenantId>; type MaybeSendNotifySafekeepersResult = MaybeSendResult<NotifySafekeepersRequest, TenantTimelineId>; impl ApiMethod for ComputeHookTenant { type Key = TenantId; type Request = NotifyAttachRequest; const API_PATH: &'static str = "notify-attach"; fn maybe_send( &self, tenant_id: TenantId, lock: Option<tokio::sync::OwnedMutexGuard<Option<ComputeRemoteTenantState>>>, ) -> MaybeSendNotifyAttachResult { let locked = match lock { Some(already_locked) => already_locked, None => { // Lock order: this _must_ be only a try_lock, because we are called inside of the [`ComputeHook::tenants`] lock. let Ok(locked) = self.get_send_lock().clone().try_lock_owned() else { return MaybeSendResult::AwaitLock((tenant_id, self.get_send_lock().clone())); }; locked } }; let request = match self { Self::Unsharded(unsharded_tenant) => Some(NotifyAttachRequest { tenant_id, shards: vec![NotifyAttachRequestShard { shard_number: ShardNumber(0), node_id: unsharded_tenant.node_id, }], stripe_size: None, preferred_az: unsharded_tenant .preferred_az .as_ref() .map(|az| az.0.clone()), }), Self::Sharded(sharded_tenant) if sharded_tenant.shards.len() == sharded_tenant.shard_count.count() as usize => { Some(NotifyAttachRequest { tenant_id, shards: sharded_tenant .shards .iter() .map(|(shard_number, node_id)| NotifyAttachRequestShard { shard_number: *shard_number, node_id: *node_id, }) .collect(), stripe_size: Some(sharded_tenant.stripe_size), preferred_az: sharded_tenant.preferred_az.as_ref().map(|az| az.0.clone()), }) } Self::Sharded(sharded_tenant) => { // Sharded tenant doesn't yet have information for all its shards tracing::info!( "ComputeHookTenant::maybe_send: not enough shards ({}/{})", sharded_tenant.shards.len(), sharded_tenant.shard_count.count() ); None } }; match request { None => { // Not yet ready to emit a notification tracing::info!("Tenant isn't yet ready to emit a notification"); MaybeSendResult::Noop } Some(request) if Some(&request) == locked.as_ref().map(|s| &s.request) && locked.as_ref().map(|s| s.applied).unwrap_or(false) => { tracing::info!( "Skipping notification because remote state already matches ({:?})", &request ); // No change from the last value successfully sent, and our state indicates that the last // value sent was fully applied on the control plane side. MaybeSendResult::Noop } Some(request) => { // Our request differs from the last one sent, or the last one sent was not fully applied on the compute side MaybeSendResult::Transmit((request, locked)) } } } async fn notify_local( env: &LocalEnv, cplane: &ComputeControlPlane, req: &NotifyAttachRequest, ) -> Result<(), NotifyError> { let NotifyAttachRequest { tenant_id, shards, stripe_size, preferred_az: _preferred_az, } = req; for (endpoint_name, endpoint) in &cplane.endpoints { if endpoint.tenant_id == *tenant_id && endpoint.status() == EndpointStatus::Running { tracing::info!("Reconfiguring pageservers for endpoint {endpoint_name}"); let shard_count = match shards.len() { 1 => ShardCount::unsharded(), n => ShardCount(n.try_into().expect("too many shards")), }; let mut shard_infos: HashMap<ShardIndex, PageserverShardInfo> = HashMap::new(); let prefer_protocol = if endpoint.grpc { PageserverProtocol::Grpc } else { PageserverProtocol::Libpq }; for shard in shards.iter() { let ps_conf = env .get_pageserver_conf(shard.node_id) .expect("Unknown pageserver"); let libpq_url = Some({ let (host, port) = parse_host_port(&ps_conf.listen_pg_addr) .expect("Unable to parse listen_pg_addr"); let port = port.unwrap_or(5432); format!("postgres://no_user@{host}:{port}") }); let grpc_url = if let Some(grpc_addr) = &ps_conf.listen_grpc_addr { let (host, port) = parse_host_port(grpc_addr).expect("invalid gRPC address"); let port = port.unwrap_or(DEFAULT_GRPC_LISTEN_PORT); Some(format!("grpc://no_user@{host}:{port}")) } else { None }; let pageserver = PageserverShardConnectionInfo { id: Some(shard.node_id), libpq_url, grpc_url, }; let shard_info = PageserverShardInfo { pageservers: vec![pageserver], }; shard_infos.insert( ShardIndex { shard_number: shard.shard_number, shard_count, }, shard_info, ); } let pageserver_conninfo = PageserverConnectionInfo { shard_count, stripe_size: stripe_size.map(|val| ShardStripeSize(val.0)), shards: shard_infos, prefer_protocol, }; endpoint .reconfigure_pageservers(&pageserver_conninfo) .await .map_err(NotifyError::NeonLocal)?; } } Ok(()) } } /// The compute hook is a destination for notifications about changes to tenant:pageserver /// mapping. It aggregates updates for the shards in a tenant, and when appropriate reconfigures /// the compute connection string. pub(super) struct ComputeHook { config: Config, tenants: std::sync::Mutex<HashMap<TenantId, ComputeHookTenant>>, timelines: std::sync::Mutex<HashMap<TenantTimelineId, ComputeHookTimeline>>, authorization_header: Option<String>, // Concurrency limiter, so that we do not overload the cloud control plane when updating // large numbers of tenants (e.g. when failing over after a node failure) api_concurrency: tokio::sync::Semaphore, // This lock is only used in testing enviroments, to serialize calls into neon_local neon_local_lock: tokio::sync::Mutex<()>, // We share a client across all notifications to enable connection re-use etc when // sending large numbers of notifications client: reqwest::Client, } /// Callers may give us a list of these when asking us to send a bulk batch /// of notifications in the background. This is a 'notification' in the sense of /// other code notifying us of a shard's status, rather than being the final notification /// that we send upwards to the control plane for the whole tenant. pub(crate) struct ShardUpdate<'a> { pub(crate) tenant_shard_id: TenantShardId, pub(crate) node_id: NodeId, pub(crate) stripe_size: ShardStripeSize, pub(crate) preferred_az: Option<Cow<'a, AvailabilityZone>>, } pub(crate) struct SafekeepersUpdate { pub(crate) tenant_id: TenantId, pub(crate) timeline_id: TimelineId, pub(crate) generation: SafekeeperGeneration, pub(crate) safekeepers: Vec<SafekeeperInfo>, } impl ComputeHook { pub(super) fn new(config: Config) -> anyhow::Result<Self> { let authorization_header = config .control_plane_jwt_token .clone() .map(|jwt| format!("Bearer {jwt}")); let mut client = reqwest::ClientBuilder::new().timeout(NOTIFY_REQUEST_TIMEOUT); for cert in &config.ssl_ca_certs { client = client.add_root_certificate(cert.clone()); } let client = client .build() .context("Failed to build http client for compute hook")?; Ok(Self { tenants: Default::default(), timelines: Default::default(), config, authorization_header, neon_local_lock: Default::default(), api_concurrency: tokio::sync::Semaphore::new(API_CONCURRENCY), client, }) } /// For test environments: use neon_local's LocalEnv to update compute async fn do_notify_local<M: ApiMethod>(&self, req: &M::Request) -> Result<(), NotifyError> { // neon_local updates are not safe to call concurrently, use a lock to serialize // all calls to this function let _locked = self.neon_local_lock.lock().await; let Some(repo_dir) = self.config.neon_local_repo_dir.as_deref() else { tracing::warn!( "neon_local_repo_dir not set, likely a bug in neon_local; skipping compute update" ); return Ok(()); }; let env = match LocalEnv::load_config(repo_dir) { Ok(e) => e, Err(e) => { tracing::warn!("Couldn't load neon_local config, skipping compute update ({e})"); return Ok(()); } }; let cplane = ComputeControlPlane::load(env.clone()).expect("Error loading compute control plane"); M::notify_local(&env, &cplane, req).await } async fn do_notify_iteration<Req: serde::Serialize + std::fmt::Debug>( &self, url: &String, reconfigure_request: &Req, cancel: &CancellationToken, ) -> Result<(), NotifyError> { let req = self.client.request(reqwest::Method::PUT, url); let req = if let Some(value) = &self.authorization_header { req.header(reqwest::header::AUTHORIZATION, value) } else { req }; tracing::info!( "Sending notify request to {} ({:?})", url, reconfigure_request ); let send_result = req.json(&reconfigure_request).send().await; let response = match send_result { Ok(r) => r, Err(e) => return Err(e.into()), }; // Treat all 2xx responses as success if response.status().is_success() { if response.status() != reqwest::StatusCode::OK { // Non-200 2xx response: it doesn't make sense to retry, but this is unexpected, so // log a warning. tracing::warn!( "Unexpected 2xx response code {} from control plane", response.status() ); } return Ok(()); } // Error response codes match response.status() { reqwest::StatusCode::TOO_MANY_REQUESTS => { // TODO: 429 handling should be global: set some state visible to other requests // so that they will delay before starting, rather than all notifications trying // once before backing off. tokio::time::timeout(SLOWDOWN_DELAY, cancel.cancelled()) .await .ok(); Err(NotifyError::SlowDown) } reqwest::StatusCode::LOCKED => { // We consider this fatal, because it's possible that the operation blocking the control one is // also the one that is waiting for this reconcile. We should let the reconciler calling // this hook fail, to give control plane a chance to un-lock. tracing::info!("Control plane reports tenant is locked, dropping out of notify"); Err(NotifyError::Busy) } reqwest::StatusCode::SERVICE_UNAVAILABLE => { Err(NotifyError::Unavailable(StatusCode::SERVICE_UNAVAILABLE)) } reqwest::StatusCode::GATEWAY_TIMEOUT => { Err(NotifyError::Unavailable(StatusCode::GATEWAY_TIMEOUT)) } reqwest::StatusCode::BAD_GATEWAY => { Err(NotifyError::Unavailable(StatusCode::BAD_GATEWAY)) } reqwest::StatusCode::BAD_REQUEST => Err(NotifyError::Fatal(StatusCode::BAD_REQUEST)), reqwest::StatusCode::UNAUTHORIZED => Err(NotifyError::Fatal(StatusCode::UNAUTHORIZED)), reqwest::StatusCode::FORBIDDEN => Err(NotifyError::Fatal(StatusCode::FORBIDDEN)), status => Err(NotifyError::Unexpected( hyper::StatusCode::from_u16(status.as_u16()) .unwrap_or(StatusCode::INTERNAL_SERVER_ERROR), )), } } async fn do_notify<R: serde::Serialize + std::fmt::Debug>( &self, url: &String, reconfigure_request: &R, cancel: &CancellationToken, ) -> Result<(), NotifyError> { // We hold these semaphore units across all retries, rather than only across each // HTTP request: this is to preserve fairness and avoid a situation where a retry might // time out waiting for a semaphore. let _units = self .api_concurrency .acquire() .await // Interpret closed semaphore as shutdown .map_err(|_| NotifyError::ShuttingDown)?; backoff::retry( || self.do_notify_iteration(url, reconfigure_request, cancel), |e| { matches!( e, NotifyError::Fatal(_) | NotifyError::Unexpected(_) | NotifyError::Busy ) }, 3, 10, "Send compute notification", cancel, ) .await .ok_or_else(|| NotifyError::ShuttingDown) .and_then(|x| x) } /// Synchronous phase: update the per-tenant state for the next intended notification fn notify_attach_prepare(&self, shard_update: ShardUpdate) -> MaybeSendNotifyAttachResult { let mut tenants_locked = self.tenants.lock().unwrap(); use std::collections::hash_map::Entry; let tenant_shard_id = shard_update.tenant_shard_id; let tenant = match tenants_locked.entry(tenant_shard_id.tenant_id) { Entry::Vacant(e) => { let ShardUpdate { tenant_shard_id, node_id, stripe_size, preferred_az, } = shard_update; e.insert(ComputeHookTenant::new( tenant_shard_id, stripe_size, preferred_az.map(|az| az.into_owned()), node_id, )) } Entry::Occupied(e) => { let tenant = e.into_mut(); tenant.update(shard_update); tenant } }; tenant.maybe_send(tenant_shard_id.tenant_id, None) } fn notify_safekeepers_prepare( &self, safekeepers_update: SafekeepersUpdate, ) -> MaybeSendNotifySafekeepersResult { let mut timelines_locked = self.timelines.lock().unwrap(); let ttid = TenantTimelineId { tenant_id: safekeepers_update.tenant_id, timeline_id: safekeepers_update.timeline_id, }; use std::collections::hash_map::Entry; let timeline = match timelines_locked.entry(ttid) { Entry::Vacant(e) => e.insert(ComputeHookTimeline::new( safekeepers_update.generation, safekeepers_update.safekeepers, )), Entry::Occupied(e) => { let timeline = e.into_mut(); timeline.update(safekeepers_update); timeline } }; timeline.maybe_send(ttid, None) } async fn notify_execute<M: ApiMethod>( &self, state: &std::sync::Mutex<HashMap<M::Key, M>>, maybe_send_result: MaybeSendResult<M::Request, M::Key>, cancel: &CancellationToken, ) -> Result<(), NotifyError> { // Process result: we may get an update to send, or we may have to wait for a lock // before trying again. let (request, mut send_lock_guard) = match maybe_send_result { MaybeSendResult::Noop => { return Ok(()); } MaybeSendResult::AwaitLock((key, send_lock)) => { let send_locked = tokio::select! { guard = send_lock.lock_owned() => {guard}, _ = cancel.cancelled() => { tracing::info!("Notification cancelled while waiting for lock"); return Err(NotifyError::ShuttingDown) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/tenant_shard.rs
storage_controller/src/tenant_shard.rs
use std::collections::{HashMap, HashSet}; use std::sync::Arc; use std::time::Duration; use futures::future::{self, Either}; use itertools::Itertools; use pageserver_api::controller_api::{AvailabilityZone, PlacementPolicy, ShardSchedulingPolicy}; use pageserver_api::models::{LocationConfig, LocationConfigMode, TenantConfig}; use pageserver_api::shard::{ShardIdentity, TenantShardId}; use serde::{Deserialize, Serialize}; use tokio::task::JoinHandle; use tokio_util::sync::CancellationToken; use tracing::{Instrument, instrument}; use utils::generation::Generation; use utils::id::NodeId; use utils::seqwait::{SeqWait, SeqWaitError}; use utils::shard::ShardCount; use utils::sync::gate::GateGuard; use crate::compute_hook::ComputeHook; use crate::metrics::{ self, ReconcileCompleteLabelGroup, ReconcileLongRunningLabelGroup, ReconcileOutcome, }; use crate::node::Node; use crate::persistence::split_state::SplitState; use crate::persistence::{Persistence, TenantShardPersistence}; use crate::reconciler::{ ReconcileError, ReconcileUnits, Reconciler, ReconcilerConfig, TargetState, attached_location_conf, secondary_location_conf, }; use crate::scheduler::{ AffinityScore, AttachedShardTag, NodeSchedulingScore, NodeSecondarySchedulingScore, RefCountUpdate, ScheduleContext, ScheduleError, Scheduler, SecondaryShardTag, ShardTag, }; use crate::service::ReconcileResultRequest; use crate::timeline_import::TimelineImportState; use crate::{Sequence, service}; /// Serialization helper fn read_last_error<S, T>(v: &std::sync::Mutex<Option<T>>, serializer: S) -> Result<S::Ok, S::Error> where S: serde::ser::Serializer, T: std::fmt::Display, { serializer.collect_str( &v.lock() .unwrap() .as_ref() .map(|e| format!("{e}")) .unwrap_or("".to_string()), ) } /// In-memory state for a particular tenant shard. /// /// This struct implement Serialize for debugging purposes, but is _not_ persisted /// itself: see [`crate::persistence`] for the subset of tenant shard state that is persisted. #[derive(Serialize)] pub(crate) struct TenantShard { pub(crate) tenant_shard_id: TenantShardId, pub(crate) shard: ShardIdentity, // Runtime only: sequence used to coordinate when updating this object while // with background reconcilers may be running. A reconciler runs to a particular // sequence. pub(crate) sequence: Sequence, // Latest generation number: next time we attach, increment this // and use the incremented number when attaching. // // None represents an incompletely onboarded tenant via the [`Service::location_config`] // API, where this tenant may only run in PlacementPolicy::Secondary. pub(crate) generation: Option<Generation>, // High level description of how the tenant should be set up. Provided // externally. pub(crate) policy: PlacementPolicy, // Low level description of exactly which pageservers should fulfil // which role. Generated by `Self::schedule`. pub(crate) intent: IntentState, // Low level description of how the tenant is configured on pageservers: // if this does not match `Self::intent` then the tenant needs reconciliation // with `Self::reconcile`. pub(crate) observed: ObservedState, // Tenant configuration, passed through opaquely to the pageserver. Identical // for all shards in a tenant. pub(crate) config: TenantConfig, /// If a reconcile task is currently in flight, it may be joined here (it is /// only safe to join if either the result has been received or the reconciler's /// cancellation token has been fired) #[serde(skip)] pub(crate) reconciler: Option<ReconcilerHandle>, /// If a tenant is being split, then all shards with that TenantId will have a /// SplitState set, this acts as a guard against other operations such as background /// reconciliation, and timeline creation. pub(crate) splitting: SplitState, /// Flag indicating whether the tenant has an in-progress timeline import. /// Used to disallow shard splits while an import is in progress. pub(crate) importing: TimelineImportState, /// If a tenant was enqueued for later reconcile due to hitting concurrency limit, this flag /// is set. This flag is cleared when the tenant is popped off the delay queue. pub(crate) delayed_reconcile: bool, /// Optionally wait for reconciliation to complete up to a particular /// sequence number. #[serde(skip)] pub(crate) waiter: std::sync::Arc<SeqWait<Sequence, Sequence>>, /// Indicates sequence number for which we have encountered an error reconciling. If /// this advances ahead of [`Self::waiter`] then a reconciliation error has occurred, /// and callers should stop waiting for `waiter` and propagate the error. #[serde(skip)] pub(crate) error_waiter: std::sync::Arc<SeqWait<Sequence, Sequence>>, /// The most recent error from a reconcile on this tenant. This is a nested Arc /// because: /// - ReconcileWaiters need to Arc-clone the overall object to read it later /// - ReconcileWaitError needs to use an `Arc<ReconcileError>` because we can construct /// many waiters for one shard, and the underlying error types are not Clone. /// /// TODO: generalize to an array of recent events /// TOOD: use a ArcSwap instead of mutex for faster reads? #[serde(serialize_with = "read_last_error")] pub(crate) last_error: std::sync::Arc<std::sync::Mutex<Option<Arc<ReconcileError>>>>, /// Amount of consecutive [`crate::service::Service::reconcile_all`] iterations that have been /// scheduled a reconciliation for this shard. /// /// If this reaches `MAX_CONSECUTIVE_RECONCILES`, the shard is considered "stuck" and will be /// ignored when deciding whether optimizations can run. This includes both successful and failed /// reconciliations. /// /// Incremented in [`crate::service::Service::process_result`], and reset to 0 when /// [`crate::service::Service::reconcile_all`] determines no reconciliation is needed for this shard. pub(crate) consecutive_reconciles_count: usize, /// If we have a pending compute notification that for some reason we weren't able to send, /// set this to true. If this is set, calls to [`Self::get_reconcile_needed`] will return Yes /// and trigger a Reconciler run. This is the mechanism by which compute notifications are included in the scope /// of state that we publish externally in an eventually consistent way. pub(crate) pending_compute_notification: bool, /// To do a graceful migration, set this field to the destination pageserver, and optimization /// functions will consider this node the best location and react appropriately. preferred_node: Option<NodeId>, // Support/debug tool: if something is going wrong or flapping with scheduling, this may // be set to a non-active state to avoid making changes while the issue is fixed. scheduling_policy: ShardSchedulingPolicy, } #[derive(Clone, Debug, Serialize)] pub(crate) struct IntentState { attached: Option<NodeId>, secondary: Vec<NodeId>, // We should attempt to schedule this shard in the provided AZ to // decrease chances of cross-AZ compute. preferred_az_id: Option<AvailabilityZone>, } impl IntentState { pub(crate) fn new(preferred_az_id: Option<AvailabilityZone>) -> Self { Self { attached: None, secondary: vec![], preferred_az_id, } } pub(crate) fn single( scheduler: &mut Scheduler, node_id: Option<NodeId>, preferred_az_id: Option<AvailabilityZone>, ) -> Self { if let Some(node_id) = node_id { scheduler.update_node_ref_counts( node_id, preferred_az_id.as_ref(), RefCountUpdate::Attach, ); } Self { attached: node_id, secondary: vec![], preferred_az_id, } } pub(crate) fn set_attached(&mut self, scheduler: &mut Scheduler, new_attached: Option<NodeId>) { if self.attached != new_attached { if let Some(old_attached) = self.attached.take() { scheduler.update_node_ref_counts( old_attached, self.preferred_az_id.as_ref(), RefCountUpdate::Detach, ); } if let Some(new_attached) = &new_attached { scheduler.update_node_ref_counts( *new_attached, self.preferred_az_id.as_ref(), RefCountUpdate::Attach, ); } self.attached = new_attached; } if let Some(new_attached) = &new_attached { assert!(!self.secondary.contains(new_attached)); } } /// Like set_attached, but the node is from [`Self::secondary`]. This swaps the node from /// secondary to attached while maintaining the scheduler's reference counts. pub(crate) fn promote_attached( &mut self, scheduler: &mut Scheduler, promote_secondary: NodeId, ) { // If we call this with a node that isn't in secondary, it would cause incorrect // scheduler reference counting, since we assume the node is already referenced as a secondary. debug_assert!(self.secondary.contains(&promote_secondary)); self.secondary.retain(|n| n != &promote_secondary); let demoted = self.attached; self.attached = Some(promote_secondary); scheduler.update_node_ref_counts( promote_secondary, self.preferred_az_id.as_ref(), RefCountUpdate::PromoteSecondary, ); if let Some(demoted) = demoted { scheduler.update_node_ref_counts( demoted, self.preferred_az_id.as_ref(), RefCountUpdate::DemoteAttached, ); } } pub(crate) fn push_secondary(&mut self, scheduler: &mut Scheduler, new_secondary: NodeId) { // Every assertion here should probably have a corresponding check in // `validate_optimization` unless it is an invariant that should never be violated. Note // that the lock is not held between planning optimizations and applying them so you have to // assume any valid state transition of the intent state may have occurred assert!(!self.secondary.contains(&new_secondary)); assert!(self.attached != Some(new_secondary)); scheduler.update_node_ref_counts( new_secondary, self.preferred_az_id.as_ref(), RefCountUpdate::AddSecondary, ); self.secondary.push(new_secondary); } /// It is legal to call this with a node that is not currently a secondary: that is a no-op pub(crate) fn remove_secondary(&mut self, scheduler: &mut Scheduler, node_id: NodeId) { let index = self.secondary.iter().position(|n| *n == node_id); if let Some(index) = index { scheduler.update_node_ref_counts( node_id, self.preferred_az_id.as_ref(), RefCountUpdate::RemoveSecondary, ); self.secondary.remove(index); } } pub(crate) fn clear_secondary(&mut self, scheduler: &mut Scheduler) { for secondary in self.secondary.drain(..) { scheduler.update_node_ref_counts( secondary, self.preferred_az_id.as_ref(), RefCountUpdate::RemoveSecondary, ); } } /// Remove the last secondary node from the list of secondaries pub(crate) fn pop_secondary(&mut self, scheduler: &mut Scheduler) { if let Some(node_id) = self.secondary.pop() { scheduler.update_node_ref_counts( node_id, self.preferred_az_id.as_ref(), RefCountUpdate::RemoveSecondary, ); } } pub(crate) fn clear(&mut self, scheduler: &mut Scheduler) { if let Some(old_attached) = self.attached.take() { scheduler.update_node_ref_counts( old_attached, self.preferred_az_id.as_ref(), RefCountUpdate::Detach, ); } self.clear_secondary(scheduler); } pub(crate) fn all_pageservers(&self) -> Vec<NodeId> { let mut result = Vec::new(); if let Some(p) = self.attached { result.push(p) } result.extend(self.secondary.iter().copied()); result } pub(crate) fn get_attached(&self) -> &Option<NodeId> { &self.attached } pub(crate) fn get_secondary(&self) -> &Vec<NodeId> { &self.secondary } /// If the node is in use as the attached location, demote it into /// the list of secondary locations. This is used when a node goes offline, /// and we want to use a different node for attachment, but not permanently /// forget the location on the offline node. /// /// Returns true if a change was made pub(crate) fn demote_attached(&mut self, scheduler: &mut Scheduler, node_id: NodeId) -> bool { if self.attached == Some(node_id) { self.attached = None; self.secondary.push(node_id); scheduler.update_node_ref_counts( node_id, self.preferred_az_id.as_ref(), RefCountUpdate::DemoteAttached, ); true } else { false } } pub(crate) fn set_preferred_az( &mut self, scheduler: &mut Scheduler, preferred_az: Option<AvailabilityZone>, ) { let new_az = preferred_az.as_ref(); let old_az = self.preferred_az_id.as_ref(); if old_az != new_az { if let Some(node_id) = self.attached { scheduler.update_node_ref_counts( node_id, new_az, RefCountUpdate::ChangePreferredAzFrom(old_az), ); } for node_id in &self.secondary { scheduler.update_node_ref_counts( *node_id, new_az, RefCountUpdate::ChangePreferredAzFrom(old_az), ); } self.preferred_az_id = preferred_az; } } pub(crate) fn get_preferred_az(&self) -> Option<&AvailabilityZone> { self.preferred_az_id.as_ref() } } impl Drop for IntentState { fn drop(&mut self) { // Must clear before dropping, to avoid leaving stale refcounts in the Scheduler. // We do not check this while panicking, to avoid polluting unit test failures or // other assertions with this assertion's output. It's still wrong to leak these, // but if we already have a panic then we don't need to independently flag this case. if !(std::thread::panicking()) { debug_assert!(self.attached.is_none() && self.secondary.is_empty()); } } } #[derive(Default, Clone, Serialize, Deserialize, Debug)] pub(crate) struct ObservedState { pub(crate) locations: HashMap<NodeId, ObservedStateLocation>, } /// Our latest knowledge of how this tenant is configured in the outside world. /// /// Meaning: /// * No instance of this type exists for a node: we are certain that we have nothing configured on that /// node for this shard. /// * Instance exists with conf==None: we *might* have some state on that node, but we don't know /// what it is (e.g. we failed partway through configuring it) /// * Instance exists with conf==Some: this tells us what we last successfully configured on this node, /// and that configuration will still be present unless something external interfered. #[derive(Clone, Serialize, Deserialize, Debug)] pub(crate) struct ObservedStateLocation { /// If None, it means we do not know the status of this shard's location on this node, but /// we know that we might have some state on this node. pub(crate) conf: Option<LocationConfig>, } pub(crate) struct ReconcilerWaiter { // For observability purposes, remember the ID of the shard we're // waiting for. pub(crate) tenant_shard_id: TenantShardId, seq_wait: std::sync::Arc<SeqWait<Sequence, Sequence>>, error_seq_wait: std::sync::Arc<SeqWait<Sequence, Sequence>>, error: std::sync::Arc<std::sync::Mutex<Option<Arc<ReconcileError>>>>, seq: Sequence, } pub(crate) enum ReconcilerStatus { Done, Failed, InProgress, } #[derive(thiserror::Error, Debug)] pub(crate) enum ReconcileWaitError { #[error("Timeout waiting for shard {0}")] Timeout(TenantShardId), #[error("shutting down")] Shutdown, #[error("Reconcile error on shard {0}: {1}")] Failed(TenantShardId, Arc<ReconcileError>), } #[derive(Eq, PartialEq, Debug, Clone)] pub(crate) struct ReplaceSecondary { old_node_id: NodeId, new_node_id: NodeId, } #[derive(Eq, PartialEq, Debug, Clone)] pub(crate) struct MigrateAttachment { pub(crate) old_attached_node_id: NodeId, pub(crate) new_attached_node_id: NodeId, } #[derive(Eq, PartialEq, Debug, Clone)] pub(crate) enum ScheduleOptimizationAction { // Replace one of our secondary locations with a different node ReplaceSecondary(ReplaceSecondary), // Migrate attachment to an existing secondary location MigrateAttachment(MigrateAttachment), // Create a secondary location, with the intent of later migrating to it CreateSecondary(NodeId), // Remove a secondary location that we previously created to facilitate a migration RemoveSecondary(NodeId), } #[derive(Eq, PartialEq, Debug, Clone)] pub(crate) struct ScheduleOptimization { // What was the reconcile sequence when we generated this optimization? The optimization // should only be applied if the shard's sequence is still at this value, in case other changes // happened between planning the optimization and applying it. sequence: Sequence, pub(crate) action: ScheduleOptimizationAction, } impl ReconcilerWaiter { pub(crate) async fn wait_timeout(&self, timeout: Duration) -> Result<(), ReconcileWaitError> { tokio::select! { result = self.seq_wait.wait_for_timeout(self.seq, timeout)=> { result.map_err(|e| match e { SeqWaitError::Timeout => ReconcileWaitError::Timeout(self.tenant_shard_id), SeqWaitError::Shutdown => ReconcileWaitError::Shutdown })?; }, result = self.error_seq_wait.wait_for(self.seq) => { result.map_err(|e| match e { SeqWaitError::Shutdown => ReconcileWaitError::Shutdown, SeqWaitError::Timeout => unreachable!() })?; return Err(ReconcileWaitError::Failed(self.tenant_shard_id, self.error.lock().unwrap().clone().expect("If error_seq_wait was advanced error was set").clone())) } } Ok(()) } pub(crate) fn get_status(&self) -> ReconcilerStatus { if self.seq_wait.would_wait_for(self.seq).is_ok() { ReconcilerStatus::Done } else if self.error_seq_wait.would_wait_for(self.seq).is_ok() { ReconcilerStatus::Failed } else { ReconcilerStatus::InProgress } } } /// Having spawned a reconciler task, the tenant shard's state will carry enough /// information to optionally cancel & await it later. pub(crate) struct ReconcilerHandle { sequence: Sequence, handle: JoinHandle<()>, cancel: CancellationToken, } pub(crate) enum ReconcileNeeded { /// shard either doesn't need reconciliation, or is forbidden from spawning a reconciler /// in its current state (e.g. shard split in progress, or ShardSchedulingPolicy forbids it) No, /// shard has a reconciler running, and its intent hasn't changed since that one was /// spawned: wait for the existing reconciler rather than spawning a new one. WaitExisting(ReconcilerWaiter), /// shard needs reconciliation: call into [`TenantShard::spawn_reconciler`] Yes(ReconcileReason), } #[derive(Debug)] pub(crate) enum ReconcileReason { ActiveNodesDirty, UnknownLocation, PendingComputeNotification, } /// Pending modification to the observed state of a tenant shard. /// Produced by [`Reconciler::observed_deltas`] and applied in [`crate::service::Service::process_result`]. pub(crate) enum ObservedStateDelta { Upsert(Box<(NodeId, ObservedStateLocation)>), Delete(NodeId), } impl ObservedStateDelta { pub(crate) fn node_id(&self) -> &NodeId { match self { Self::Upsert(up) => &up.0, Self::Delete(nid) => nid, } } } /// When a reconcile task completes, it sends this result object /// to be applied to the primary TenantShard. pub(crate) struct ReconcileResult { pub(crate) sequence: Sequence, /// On errors, `observed` should be treated as an incompleted description /// of state (i.e. any nodes present in the result should override nodes /// present in the parent tenant state, but any unmentioned nodes should /// not be removed from parent tenant state) pub(crate) result: Result<(), ReconcileError>, pub(crate) tenant_shard_id: TenantShardId, pub(crate) generation: Option<Generation>, pub(crate) observed_deltas: Vec<ObservedStateDelta>, /// Set [`TenantShard::pending_compute_notification`] from this flag pub(crate) pending_compute_notification: bool, } impl ObservedState { pub(crate) fn new() -> Self { Self { locations: HashMap::new(), } } pub(crate) fn is_empty(&self) -> bool { self.locations.is_empty() } } impl TenantShard { pub(crate) fn new( tenant_shard_id: TenantShardId, shard: ShardIdentity, policy: PlacementPolicy, preferred_az_id: Option<AvailabilityZone>, ) -> Self { metrics::METRICS_REGISTRY .metrics_group .storage_controller_tenant_shards .inc(); Self { tenant_shard_id, policy, intent: IntentState::new(preferred_az_id), generation: Some(Generation::new(0)), shard, observed: ObservedState::default(), config: TenantConfig::default(), reconciler: None, splitting: SplitState::Idle, importing: TimelineImportState::Idle, sequence: Sequence(1), delayed_reconcile: false, waiter: Arc::new(SeqWait::new(Sequence(0))), error_waiter: Arc::new(SeqWait::new(Sequence(0))), last_error: Arc::default(), consecutive_reconciles_count: 0, pending_compute_notification: false, scheduling_policy: ShardSchedulingPolicy::default(), preferred_node: None, } } /// For use on startup when learning state from pageservers: generate my [`IntentState`] from my /// [`ObservedState`], even if it violates my [`PlacementPolicy`]. Call [`Self::schedule`] next, /// to get an intent state that complies with placement policy. The overall goal is to do scheduling /// in a way that makes use of any configured locations that already exist in the outside world. pub(crate) fn intent_from_observed(&mut self, scheduler: &mut Scheduler) { // Choose an attached location by filtering observed locations, and then sorting to get the highest // generation let mut attached_locs = self .observed .locations .iter() .filter_map(|(node_id, l)| { if let Some(conf) = &l.conf { if conf.mode == LocationConfigMode::AttachedMulti || conf.mode == LocationConfigMode::AttachedSingle || conf.mode == LocationConfigMode::AttachedStale { Some((node_id, conf.generation)) } else { None } } else { None } }) .collect::<Vec<_>>(); attached_locs.sort_by_key(|i| i.1); if let Some((node_id, _gen)) = attached_locs.into_iter().next_back() { self.intent.set_attached(scheduler, Some(*node_id)); } // All remaining observed locations generate secondary intents. This includes None // observations, as these may well have some local content on disk that is usable (this // is an edge case that might occur if we restarted during a migration or other change) // // We may leave intent.attached empty if we didn't find any attached locations: [`Self::schedule`] // will take care of promoting one of these secondaries to be attached. self.observed.locations.keys().for_each(|node_id| { if Some(*node_id) != self.intent.attached { self.intent.push_secondary(scheduler, *node_id); } }); } /// Part of [`Self::schedule`] that is used to choose exactly one node to act as the /// attached pageserver for a shard. /// /// Returns whether we modified it, and the NodeId selected. fn schedule_attached( &mut self, scheduler: &mut Scheduler, context: &ScheduleContext, ) -> Result<(bool, NodeId), ScheduleError> { // No work to do if we already have an attached tenant if let Some(node_id) = self.intent.attached { return Ok((false, node_id)); } if let Some(promote_secondary) = self.preferred_secondary(scheduler) { // Promote a secondary tracing::debug!("Promoted secondary {} to attached", promote_secondary); self.intent.promote_attached(scheduler, promote_secondary); Ok((true, promote_secondary)) } else { // Pick a fresh node: either we had no secondaries or none were schedulable let node_id = scheduler.schedule_shard::<AttachedShardTag>( &self.intent.secondary, &self.intent.preferred_az_id, context, )?; tracing::debug!("Selected {} as attached", node_id); self.intent.set_attached(scheduler, Some(node_id)); Ok((true, node_id)) } } #[instrument(skip_all, fields( tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), sequence=%self.sequence ))] pub(crate) fn schedule( &mut self, scheduler: &mut Scheduler, context: &mut ScheduleContext, ) -> Result<(), ScheduleError> { let r = self.do_schedule(scheduler, context); context.avoid(&self.intent.all_pageservers()); r } pub(crate) fn do_schedule( &mut self, scheduler: &mut Scheduler, context: &ScheduleContext, ) -> Result<(), ScheduleError> { // TODO: before scheduling new nodes, check if any existing content in // self.intent refers to pageservers that are offline, and pick other // pageservers if so. // TODO: respect the splitting bit on tenants: if they are currently splitting then we may not // change their attach location. match self.scheduling_policy { ShardSchedulingPolicy::Active | ShardSchedulingPolicy::Essential => {} ShardSchedulingPolicy::Pause | ShardSchedulingPolicy::Stop => { // Warn to make it obvious why other things aren't happening/working, if we skip scheduling tracing::warn!(tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), "Scheduling is disabled by policy {:?}", self.scheduling_policy); return Ok(()); } } // Build the set of pageservers already in use by this tenant, to avoid scheduling // more work on the same pageservers we're already using. let mut modified = false; // Add/remove nodes to fulfil policy use PlacementPolicy::*; match self.policy { Attached(secondary_count) => { // Should have exactly one attached, and at least N secondaries let (modified_attached, attached_node_id) = self.schedule_attached(scheduler, context)?; modified |= modified_attached; let mut used_pageservers = vec![attached_node_id]; while self.intent.secondary.len() < secondary_count { let node_id = scheduler.schedule_shard::<SecondaryShardTag>( &used_pageservers, &self.intent.preferred_az_id, context, )?; self.intent.push_secondary(scheduler, node_id); used_pageservers.push(node_id); modified = true; } } Secondary => { if let Some(node_id) = self.intent.get_attached() { // Populate secondary by demoting the attached node self.intent.demote_attached(scheduler, *node_id); modified = true; } else if self.intent.secondary.is_empty() { // Populate secondary by scheduling a fresh node // // We use [`AttachedShardTag`] because when a secondary location is the only one // a shard has, we expect that its next use will be as an attached location: we want // the tenant to be ready to warm up and run fast in their preferred AZ. let node_id = scheduler.schedule_shard::<AttachedShardTag>( &[], &self.intent.preferred_az_id, context, )?; self.intent.push_secondary(scheduler, node_id); modified = true; } while self.intent.secondary.len() > 1 { // If we have multiple secondaries (e.g. when transitioning from Attached to Secondary and // having just demoted our attached location), then we should prefer to keep the location // in our preferred AZ. Tenants in Secondary mode want to be in the preferred AZ so that // they have a warm location to become attached when transitioning back into Attached. let mut candidates = self.intent.get_secondary().clone(); // Sort to get secondaries outside preferred AZ last candidates .sort_by_key(|n| scheduler.get_node_az(n).as_ref() != self.preferred_az()); let secondary_to_remove = candidates.pop().unwrap(); self.intent.remove_secondary(scheduler, secondary_to_remove); modified = true; } } Detached => { // Never add locations in this mode if self.intent.get_attached().is_some() || !self.intent.get_secondary().is_empty() { self.intent.clear(scheduler); modified = true; } } } if modified { self.sequence.0 += 1; } Ok(()) } /// Reschedule this tenant shard to one of its secondary locations. Returns a scheduling error /// if the swap is not possible and leaves the intent state in its original state. /// /// Arguments: /// `promote_to`: an optional secondary location of this tenant shard. If set to None, we ask /// the scheduler to recommend a node pub(crate) fn reschedule_to_secondary( &mut self, promote_to: Option<NodeId>, scheduler: &mut Scheduler, ) -> Result<(), ScheduleError> { let promote_to = match promote_to { Some(node) => node, None => match self.preferred_secondary(scheduler) { Some(node) => node, None => { return Err(ScheduleError::ImpossibleConstraint); }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/timeline_import.rs
storage_controller/src/timeline_import.rs
use std::time::Duration; use std::{collections::HashMap, str::FromStr}; use http_utils::error::ApiError; use reqwest::Method; use serde::{Deserialize, Serialize}; use pageserver_api::models::{ShardImportProgress, ShardImportStatus}; use tokio_util::sync::CancellationToken; use utils::sync::gate::Gate; use utils::{ id::{TenantId, TimelineId}, shard::ShardIndex, }; use crate::{persistence::TimelineImportPersistence, service::Config}; #[derive(Deserialize, Serialize, PartialEq, Eq)] pub(crate) enum TimelineImportState { Importing, Idle, } #[derive(Serialize, Deserialize, Clone, Debug)] pub(crate) struct ShardImportStatuses(pub(crate) HashMap<ShardIndex, ShardImportStatus>); impl ShardImportStatuses { pub(crate) fn new(shards: Vec<ShardIndex>) -> Self { ShardImportStatuses( shards .into_iter() .map(|ts_id| { ( ts_id, ShardImportStatus::InProgress(None::<ShardImportProgress>), ) }) .collect(), ) } } #[derive(Debug)] pub(crate) struct TimelineImport { pub(crate) tenant_id: TenantId, pub(crate) timeline_id: TimelineId, pub(crate) shard_statuses: ShardImportStatuses, } pub(crate) enum TimelineImportUpdateFollowUp { Persist, None, } #[derive(thiserror::Error, Debug)] pub(crate) enum TimelineImportFinalizeError { #[error("Shut down interrupted import finalize")] ShuttingDown, #[error("Import finalization was cancelled")] Cancelled, #[error("Mismatched shard detected during import finalize: {0}")] MismatchedShards(ShardIndex), } pub(crate) enum TimelineImportUpdateError { ImportNotFound { tenant_id: TenantId, timeline_id: TimelineId, }, MismatchedShards, UnexpectedUpdate, } impl From<TimelineImportUpdateError> for ApiError { fn from(err: TimelineImportUpdateError) -> ApiError { match err { TimelineImportUpdateError::ImportNotFound { tenant_id, timeline_id, } => ApiError::NotFound( anyhow::anyhow!("Import for {tenant_id}/{timeline_id} not found").into(), ), TimelineImportUpdateError::MismatchedShards => { ApiError::InternalServerError(anyhow::anyhow!( "Import shards do not match update request, likely a shard split happened during import, this is a bug" )) } TimelineImportUpdateError::UnexpectedUpdate => { ApiError::InternalServerError(anyhow::anyhow!("Update request is unexpected")) } } } } impl TimelineImport { pub(crate) fn from_persistent(persistent: TimelineImportPersistence) -> anyhow::Result<Self> { let tenant_id = TenantId::from_str(persistent.tenant_id.as_str())?; let timeline_id = TimelineId::from_str(persistent.timeline_id.as_str())?; let shard_statuses = serde_json::from_value(persistent.shard_statuses)?; Ok(TimelineImport { tenant_id, timeline_id, shard_statuses, }) } pub(crate) fn to_persistent(&self) -> TimelineImportPersistence { TimelineImportPersistence { tenant_id: self.tenant_id.to_string(), timeline_id: self.timeline_id.to_string(), shard_statuses: serde_json::to_value(self.shard_statuses.clone()).unwrap(), } } pub(crate) fn update( &mut self, shard: ShardIndex, status: ShardImportStatus, ) -> Result<TimelineImportUpdateFollowUp, TimelineImportUpdateError> { use std::collections::hash_map::Entry::*; match self.shard_statuses.0.entry(shard) { Occupied(mut occ) => { let crnt = occ.get_mut(); if *crnt == status { Ok(TimelineImportUpdateFollowUp::None) } else if crnt.is_terminal() && *crnt != status { Err(TimelineImportUpdateError::UnexpectedUpdate) } else { *crnt = status; Ok(TimelineImportUpdateFollowUp::Persist) } } Vacant(_) => Err(TimelineImportUpdateError::MismatchedShards), } } pub(crate) fn is_complete(&self) -> bool { self.shard_statuses .0 .values() .all(|status| status.is_terminal()) } pub(crate) fn completion_error(&self) -> Option<String> { assert!(self.is_complete()); let shard_errors: HashMap<_, _> = self .shard_statuses .0 .iter() .filter_map(|(shard, status)| { if let ShardImportStatus::Error(err) = status { Some((*shard, err.clone())) } else { None } }) .collect(); if shard_errors.is_empty() { None } else { Some(serde_json::to_string(&shard_errors).unwrap()) } } } pub(crate) struct FinalizingImport { pub(crate) gate: Gate, pub(crate) cancel: CancellationToken, } pub(crate) type ImportResult = Result<(), String>; pub(crate) struct UpcallClient { authorization_header: Option<String>, client: reqwest::Client, cancel: CancellationToken, base_url: String, } const IMPORT_COMPLETE_REQUEST_TIMEOUT: Duration = Duration::from_secs(10); #[derive(Serialize, Deserialize, Debug)] struct ImportCompleteRequest { tenant_id: TenantId, timeline_id: TimelineId, error: Option<String>, } impl UpcallClient { pub(crate) fn new(config: &Config, cancel: CancellationToken) -> Self { let authorization_header = config .control_plane_jwt_token .clone() .map(|jwt| format!("Bearer {jwt}")); let client = reqwest::ClientBuilder::new() .timeout(IMPORT_COMPLETE_REQUEST_TIMEOUT) .build() .expect("Failed to construct HTTP client"); let base_url = config .control_plane_url .clone() .expect("must be configured"); Self { authorization_header, client, cancel, base_url, } } /// Notify control plane of a completed import /// /// This method guarantees at least once delivery semantics assuming /// eventual cplane availability. The cplane API is idempotent. pub(crate) async fn notify_import_complete( &self, tenant_id: TenantId, timeline_id: TimelineId, import_result: ImportResult, ) -> anyhow::Result<()> { let endpoint = if self.base_url.ends_with('/') { format!("{}import_complete", self.base_url) } else { format!("{}/import_complete", self.base_url) }; let request = self .client .request(Method::PUT, endpoint) .json(&ImportCompleteRequest { tenant_id, timeline_id, error: import_result.err(), }) .timeout(IMPORT_COMPLETE_REQUEST_TIMEOUT); let request = if let Some(auth) = &self.authorization_header { request.header(reqwest::header::AUTHORIZATION, auth) } else { request }; const RETRY_DELAY: Duration = Duration::from_secs(1); let mut attempt = 1; loop { if self.cancel.is_cancelled() { return Err(anyhow::anyhow!( "Shutting down while notifying cplane of import completion" )); } match request.try_clone().unwrap().send().await { Ok(response) if response.status().is_success() => { return Ok(()); } Ok(response) => { tracing::warn!( "Import complete notification failed with status {}, attempt {}", response.status(), attempt ); } Err(e) => { tracing::warn!( "Import complete notification failed with error: {}, attempt {}", e, attempt ); } } tokio::select! { _ = tokio::time::sleep(RETRY_DELAY) => {} _ = self.cancel.cancelled() => { return Err(anyhow::anyhow!("Shutting down while notifying cplane of import completion")); } } attempt += 1; } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/leadership.rs
storage_controller/src/leadership.rs
use std::sync::Arc; use hyper::Uri; use tokio_util::sync::CancellationToken; use crate::peer_client::{GlobalObservedState, PeerClient}; use crate::persistence::{ControllerPersistence, DatabaseError, DatabaseResult, Persistence}; use crate::service::Config; /// Helper for storage controller leadership acquisition pub(crate) struct Leadership { persistence: Arc<Persistence>, config: Config, cancel: CancellationToken, } #[derive(thiserror::Error, Debug)] pub(crate) enum Error { #[error(transparent)] Database(#[from] DatabaseError), } pub(crate) type Result<T> = std::result::Result<T, Error>; impl Leadership { pub(crate) fn new( persistence: Arc<Persistence>, config: Config, cancel: CancellationToken, ) -> Self { Self { persistence, config, cancel, } } /// Find the current leader in the database and request it to step down if required. /// Should be called early on in within the start-up sequence. /// /// Returns a tuple of two optionals: the current leader and its observed state pub(crate) async fn step_down_current_leader( &self, ) -> Result<(Option<ControllerPersistence>, Option<GlobalObservedState>)> { let leader = self.current_leader().await?; if leader.as_ref().map(|l| &l.address) == self .config .address_for_peers .as_ref() .map(Uri::to_string) .as_ref() { // We already are the current leader. This is a restart. return Ok((leader, None)); } let leader_step_down_state = if let Some(ref leader) = leader { if self.config.start_as_candidate { self.request_step_down(leader).await } else { None } } else { tracing::info!("No leader found to request step down from. Will build observed state."); None }; Ok((leader, leader_step_down_state)) } /// Mark the current storage controller instance as the leader in the database pub(crate) async fn become_leader( &self, current_leader: Option<ControllerPersistence>, ) -> Result<()> { if let Some(address_for_peers) = &self.config.address_for_peers { // TODO: `address-for-peers` can become a mandatory cli arg // after we update the k8s setup let proposed_leader = ControllerPersistence { address: address_for_peers.to_string(), started_at: chrono::Utc::now(), }; self.persistence .update_leader(current_leader, proposed_leader) .await .map_err(Error::Database) } else { tracing::info!("No address-for-peers provided. Skipping leader persistence."); Ok(()) } } async fn current_leader(&self) -> DatabaseResult<Option<ControllerPersistence>> { let res = self.persistence.get_leader().await; if let Err(DatabaseError::Query(diesel::result::Error::DatabaseError(_kind, ref err))) = res { const REL_NOT_FOUND_MSG: &str = "relation \"controllers\" does not exist"; if err.message().trim() == REL_NOT_FOUND_MSG { // Special case: if this is a brand new storage controller, migrations will not // have run at this point yet, and, hence, the controllers table does not exist. // Detect this case via the error string (diesel doesn't type it) and allow it. tracing::info!( "Detected first storage controller start-up. Allowing missing controllers table ..." ); return Ok(None); } } res } /// Request step down from the currently registered leader in the database /// /// If such an entry is persisted, the success path returns the observed /// state and details of the leader. Otherwise, None is returned indicating /// there is no leader currently. async fn request_step_down( &self, leader: &ControllerPersistence, ) -> Option<GlobalObservedState> { tracing::info!("Sending step down request to {leader:?}"); let mut http_client = reqwest::Client::builder(); for cert in &self.config.ssl_ca_certs { http_client = http_client.add_root_certificate(cert.clone()); } let http_client = match http_client.build() { Ok(http_client) => http_client, Err(err) => { tracing::error!("Failed to build client for leader step-down request: {err}"); return None; } }; let client = PeerClient::new( http_client, Uri::try_from(leader.address.as_str()).expect("Failed to build leader URI"), self.config.peer_jwt_token.clone(), ); let state = client.step_down(&self.cancel).await; match state { Ok(state) => Some(state), Err(err) => { // TODO: Make leaders periodically update a timestamp field in the // database and, if the leader is not reachable from the current instance, // but inferred as alive from the timestamp, abort start-up. This avoids // a potential scenario in which we have two controllers acting as leaders. tracing::error!( "Leader ({}) did not respond to step-down request: {}", leader.address, err ); None } } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/heartbeater.rs
storage_controller/src/heartbeater.rs
use std::collections::HashMap; use std::fmt::Debug; use std::future::Future; use std::sync::Arc; use std::time::{Duration, Instant}; use futures::StreamExt; use futures::stream::FuturesUnordered; use pageserver_api::controller_api::{NodeAvailability, SkSchedulingPolicy}; use pageserver_api::models::PageserverUtilization; use safekeeper_api::models::SafekeeperUtilization; use safekeeper_client::mgmt_api; use thiserror::Error; use tokio_util::sync::CancellationToken; use tracing::Instrument; use utils::id::NodeId; use utils::logging::SecretString; use crate::node::Node; use crate::safekeeper::Safekeeper; struct HeartbeaterTask<Server, State> { receiver: tokio::sync::mpsc::UnboundedReceiver<HeartbeatRequest<Server, State>>, cancel: CancellationToken, state: HashMap<NodeId, State>, max_offline_interval: Duration, max_warming_up_interval: Duration, http_client: reqwest::Client, jwt_token: Option<String>, } #[derive(Debug, Clone)] pub(crate) enum PageserverState { Available { last_seen_at: Instant, utilization: PageserverUtilization, }, WarmingUp { started_at: Instant, }, Offline, } #[derive(Debug, Clone)] pub(crate) enum SafekeeperState { Available { last_seen_at: Instant, utilization: SafekeeperUtilization, }, Offline, } #[derive(Debug)] pub(crate) struct AvailablityDeltas<State>(pub Vec<(NodeId, State)>); #[derive(Debug, Error)] pub(crate) enum HeartbeaterError { #[error("Cancelled")] Cancel, } struct HeartbeatRequest<Server, State> { servers: Arc<HashMap<NodeId, Server>>, reply: tokio::sync::oneshot::Sender<Result<AvailablityDeltas<State>, HeartbeaterError>>, } pub(crate) struct Heartbeater<Server, State> { sender: tokio::sync::mpsc::UnboundedSender<HeartbeatRequest<Server, State>>, } #[allow(private_bounds)] impl<Server: Send + Sync + 'static, State: Debug + Send + 'static> Heartbeater<Server, State> where HeartbeaterTask<Server, State>: HeartBeat<Server, State>, { pub(crate) fn new( http_client: reqwest::Client, jwt_token: Option<String>, max_offline_interval: Duration, max_warming_up_interval: Duration, cancel: CancellationToken, ) -> Self { let (sender, receiver) = tokio::sync::mpsc::unbounded_channel::<HeartbeatRequest<Server, State>>(); let mut heartbeater = HeartbeaterTask::new( receiver, http_client, jwt_token, max_offline_interval, max_warming_up_interval, cancel, ); tokio::task::spawn(async move { heartbeater.run().await }); Self { sender } } pub(crate) async fn heartbeat( &self, servers: Arc<HashMap<NodeId, Server>>, ) -> Result<AvailablityDeltas<State>, HeartbeaterError> { let (sender, receiver) = tokio::sync::oneshot::channel(); self.sender .send(HeartbeatRequest { servers, reply: sender, }) .map_err(|_| HeartbeaterError::Cancel)?; receiver .await .map_err(|_| HeartbeaterError::Cancel) .and_then(|x| x) } } impl<Server, State: Debug> HeartbeaterTask<Server, State> where HeartbeaterTask<Server, State>: HeartBeat<Server, State>, { fn new( receiver: tokio::sync::mpsc::UnboundedReceiver<HeartbeatRequest<Server, State>>, http_client: reqwest::Client, jwt_token: Option<String>, max_offline_interval: Duration, max_warming_up_interval: Duration, cancel: CancellationToken, ) -> Self { Self { receiver, cancel, state: HashMap::new(), max_offline_interval, max_warming_up_interval, http_client, jwt_token, } } async fn run(&mut self) { loop { tokio::select! { request = self.receiver.recv() => { match request { Some(req) => { if req.reply.is_closed() { // Prevent a possibly infinite buildup of the receiver channel, if requests arrive faster than we can handle them continue; } let res = self.heartbeat(req.servers).await; // Ignore the return value in order to not panic if the heartbeat function's future was cancelled _ = req.reply.send(res); }, None => { return; } } }, _ = self.cancel.cancelled() => return } } } } pub(crate) trait HeartBeat<Server, State> { fn heartbeat( &mut self, pageservers: Arc<HashMap<NodeId, Server>>, ) -> impl Future<Output = Result<AvailablityDeltas<State>, HeartbeaterError>> + Send; } impl HeartBeat<Node, PageserverState> for HeartbeaterTask<Node, PageserverState> { async fn heartbeat( &mut self, pageservers: Arc<HashMap<NodeId, Node>>, ) -> Result<AvailablityDeltas<PageserverState>, HeartbeaterError> { let mut new_state = HashMap::new(); let mut heartbeat_futs = FuturesUnordered::new(); for (node_id, node) in &*pageservers { heartbeat_futs.push({ let http_client = self.http_client.clone(); let jwt_token = self.jwt_token.clone(); let cancel = self.cancel.clone(); // Clone the node and mark it as available such that the request // goes through to the pageserver even when the node is marked offline. // This doesn't impact the availability observed by [`crate::service::Service`]. let mut node_clone = node.clone(); node_clone .set_availability(NodeAvailability::Active(PageserverUtilization::full())); async move { let response = node_clone .with_client_retries( |client| async move { client.get_utilization().await }, &http_client, &jwt_token, 3, 3, Duration::from_secs(1), &cancel, ) .await; let response = match response { Some(r) => r, None => { // This indicates cancellation of the request. // We ignore the node in this case. return None; } }; let status = if let Ok(utilization) = response { PageserverState::Available { last_seen_at: Instant::now(), utilization, } } else if let NodeAvailability::WarmingUp(last_seen_at) = node.get_availability() { PageserverState::WarmingUp { started_at: *last_seen_at, } } else { PageserverState::Offline }; Some((*node_id, status)) } .instrument(tracing::info_span!("heartbeat_ps", %node_id)) }); } loop { let maybe_status = tokio::select! { next = heartbeat_futs.next() => { match next { Some(result) => result, None => { break; } } }, _ = self.cancel.cancelled() => { return Err(HeartbeaterError::Cancel); } }; if let Some((node_id, status)) = maybe_status { new_state.insert(node_id, status); } } let mut warming_up = 0; let mut offline = 0; for state in new_state.values() { match state { PageserverState::WarmingUp { .. } => { warming_up += 1; } PageserverState::Offline => offline += 1, PageserverState::Available { .. } => {} } } tracing::info!( "Heartbeat round complete for {} nodes, {} warming-up, {} offline", new_state.len(), warming_up, offline ); let mut deltas = Vec::new(); let now = Instant::now(); for (node_id, ps_state) in new_state.iter_mut() { use std::collections::hash_map::Entry::*; let entry = self.state.entry(*node_id); let mut needs_update = false; match entry { Occupied(ref occ) => match (occ.get(), &ps_state) { (PageserverState::Offline, PageserverState::Offline) => {} (PageserverState::Available { last_seen_at, .. }, PageserverState::Offline) => { if now - *last_seen_at >= self.max_offline_interval { deltas.push((*node_id, ps_state.clone())); needs_update = true; } } (_, PageserverState::WarmingUp { started_at }) => { if now - *started_at >= self.max_warming_up_interval { *ps_state = PageserverState::Offline; } deltas.push((*node_id, ps_state.clone())); needs_update = true; } _ => { deltas.push((*node_id, ps_state.clone())); needs_update = true; } }, Vacant(_) => { // This is a new node. Don't generate a delta for it. deltas.push((*node_id, ps_state.clone())); } } match entry { Occupied(mut occ) if needs_update => { (*occ.get_mut()) = ps_state.clone(); } Vacant(vac) => { vac.insert(ps_state.clone()); } _ => {} } } Ok(AvailablityDeltas(deltas)) } } impl HeartBeat<Safekeeper, SafekeeperState> for HeartbeaterTask<Safekeeper, SafekeeperState> { async fn heartbeat( &mut self, safekeepers: Arc<HashMap<NodeId, Safekeeper>>, ) -> Result<AvailablityDeltas<SafekeeperState>, HeartbeaterError> { let mut new_state = HashMap::new(); let mut heartbeat_futs = FuturesUnordered::new(); for (node_id, sk) in &*safekeepers { if sk.scheduling_policy() == SkSchedulingPolicy::Decomissioned { continue; } heartbeat_futs.push({ let http_client = self.http_client.clone(); let jwt_token = self .jwt_token .as_ref() .map(|t| SecretString::from(t.to_owned())); let cancel = self.cancel.clone(); async move { let response = sk .with_client_retries( |client| async move { client.get_utilization().await }, &http_client, &jwt_token, 3, 3, Duration::from_secs(1), &cancel, ) .await; let status = match response { Ok(utilization) => SafekeeperState::Available { last_seen_at: Instant::now(), utilization, }, Err(mgmt_api::Error::Cancelled) => { // This indicates cancellation of the request. // We ignore the node in this case. return None; } Err(e) => { tracing::info!( "Marking safekeeper {} at as offline: {e}", sk.base_url() ); SafekeeperState::Offline } }; Some((*node_id, status)) } .instrument(tracing::info_span!("heartbeat_sk", %node_id)) }); } loop { let maybe_status = tokio::select! { next = heartbeat_futs.next() => { match next { Some(result) => result, None => { break; } } }, _ = self.cancel.cancelled() => { return Err(HeartbeaterError::Cancel); } }; if let Some((node_id, status)) = maybe_status { new_state.insert(node_id, status); } } let mut offline = 0; for state in new_state.values() { match state { SafekeeperState::Offline => offline += 1, SafekeeperState::Available { .. } => {} } } tracing::info!( "Heartbeat round complete for {} safekeepers, {} offline", new_state.len(), offline ); let mut deltas = Vec::new(); let now = Instant::now(); for (node_id, sk_state) in new_state.iter_mut() { use std::collections::hash_map::Entry::*; let entry = self.state.entry(*node_id); let mut needs_update = false; match entry { Occupied(ref occ) => match (occ.get(), &sk_state) { (SafekeeperState::Offline, SafekeeperState::Offline) => {} (SafekeeperState::Available { last_seen_at, .. }, SafekeeperState::Offline) => { if now - *last_seen_at >= self.max_offline_interval { deltas.push((*node_id, sk_state.clone())); needs_update = true; } } _ => { deltas.push((*node_id, sk_state.clone())); needs_update = true; } }, Vacant(_) => { // This is a new node. Don't generate a delta for it. deltas.push((*node_id, sk_state.clone())); } } match entry { Occupied(mut occ) if needs_update => { (*occ.get_mut()) = sk_state.clone(); } Vacant(vac) => { vac.insert(sk_state.clone()); } _ => {} } } Ok(AvailablityDeltas(deltas)) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/persistence.rs
storage_controller/src/persistence.rs
pub(crate) mod split_state; use std::collections::HashMap; use std::io::Write; use std::str::FromStr; use std::sync::Arc; use std::time::{Duration, Instant}; use diesel::deserialize::{FromSql, FromSqlRow}; use diesel::expression::AsExpression; use diesel::pg::Pg; use diesel::prelude::*; use diesel::serialize::{IsNull, ToSql}; use diesel_async::async_connection_wrapper::AsyncConnectionWrapper; use diesel_async::pooled_connection::bb8::Pool; use diesel_async::pooled_connection::{AsyncDieselConnectionManager, ManagerConfig}; use diesel_async::{AsyncPgConnection, RunQueryDsl}; use diesel_migrations::{EmbeddedMigrations, embed_migrations}; use futures::FutureExt; use futures::future::BoxFuture; use itertools::Itertools; use pageserver_api::controller_api::{ AvailabilityZone, MetadataHealthRecord, NodeLifecycle, NodeSchedulingPolicy, PlacementPolicy, SafekeeperDescribeResponse, ShardSchedulingPolicy, SkSchedulingPolicy, }; use pageserver_api::models::{ShardImportStatus, TenantConfig}; use pageserver_api::shard::{ ShardConfigError, ShardCount, ShardIdentity, ShardNumber, ShardStripeSize, TenantShardId, }; use rustls::client::WebPkiServerVerifier; use rustls::client::danger::{ServerCertVerified, ServerCertVerifier}; use rustls::crypto::ring; use safekeeper_api::membership::SafekeeperGeneration; use scoped_futures::ScopedBoxFuture; use serde::{Deserialize, Serialize}; use utils::generation::Generation; use utils::id::{NodeId, TenantId, TimelineId}; use utils::lsn::Lsn; use self::split_state::SplitState; use crate::metrics::{ DatabaseQueryErrorLabelGroup, DatabaseQueryLatencyLabelGroup, METRICS_REGISTRY, }; use crate::node::Node; use crate::timeline_import::{ TimelineImport, TimelineImportUpdateError, TimelineImportUpdateFollowUp, }; const MIGRATIONS: EmbeddedMigrations = embed_migrations!("./migrations"); /// ## What do we store? /// /// The storage controller service does not store most of its state durably. /// /// The essential things to store durably are: /// - generation numbers, as these must always advance monotonically to ensure data safety. /// - Tenant's PlacementPolicy and TenantConfig, as the source of truth for these is something external. /// - Node's scheduling policies, as the source of truth for these is something external. /// /// Other things we store durably as an implementation detail: /// - Node's host/port: this could be avoided it we made nodes emit a self-registering heartbeat, /// but it is operationally simpler to make this service the authority for which nodes /// it talks to. /// /// ## Performance/efficiency /// /// The storage controller service does not go via the database for most things: there are /// a couple of places where we must, and where efficiency matters: /// - Incrementing generation numbers: the Reconciler has to wait for this to complete /// before it can attach a tenant, so this acts as a bound on how fast things like /// failover can happen. /// - Pageserver re-attach: we will increment many shards' generations when this happens, /// so it is important to avoid e.g. issuing O(N) queries. /// /// Database calls relating to nodes have low performance requirements, as they are very rarely /// updated, and reads of nodes are always from memory, not the database. We only require that /// we can UPDATE a node's scheduling mode reasonably quickly to mark a bad node offline. pub struct Persistence { connection_pool: Pool<AsyncPgConnection>, } /// Legacy format, for use in JSON compat objects in test environment #[derive(Serialize, Deserialize)] struct JsonPersistence { tenants: HashMap<TenantShardId, TenantShardPersistence>, } #[derive(thiserror::Error, Debug)] pub(crate) enum DatabaseError { #[error(transparent)] Query(#[from] diesel::result::Error), #[error(transparent)] Connection(#[from] diesel::result::ConnectionError), #[error(transparent)] ConnectionPool(#[from] diesel_async::pooled_connection::bb8::RunError), #[error("Logical error: {0}")] Logical(String), #[error("Migration error: {0}")] Migration(String), #[error("CAS error: {0}")] Cas(String), } #[derive(measured::FixedCardinalityLabel, Copy, Clone)] pub(crate) enum DatabaseOperation { InsertNode, UpdateNode, DeleteNode, ListNodes, ListTombstones, BeginShardSplit, CompleteShardSplit, AbortShardSplit, Detach, ReAttach, IncrementGeneration, TenantGenerations, ShardGenerations, ListTenantShards, LoadTenant, InsertTenantShards, UpdateTenantShard, DeleteTenant, UpdateTenantConfig, UpdateMetadataHealth, ListMetadataHealth, ListMetadataHealthUnhealthy, ListMetadataHealthOutdated, ListSafekeepers, GetLeader, UpdateLeader, SetPreferredAzs, InsertTimeline, UpdateTimeline, UpdateTimelineMembership, UpdateCplaneNotifiedGeneration, UpdateSkSetNotifiedGeneration, GetTimeline, InsertTimelineReconcile, RemoveTimelineReconcile, ListTimelineReconcile, ListTimelineReconcileStartup, InsertTimelineImport, UpdateTimelineImport, DeleteTimelineImport, ListTimelineImports, IsTenantImportingTimeline, } #[must_use] pub(crate) enum AbortShardSplitStatus { /// We aborted the split in the database by reverting to the parent shards Aborted, /// The split had already been persisted. Complete, } pub(crate) type DatabaseResult<T> = Result<T, DatabaseError>; /// Some methods can operate on either a whole tenant or a single shard #[derive(Clone)] pub(crate) enum TenantFilter { Tenant(TenantId), Shard(TenantShardId), } /// Represents the results of looking up generation+pageserver for the shards of a tenant pub(crate) struct ShardGenerationState { pub(crate) tenant_shard_id: TenantShardId, pub(crate) generation: Option<Generation>, pub(crate) generation_pageserver: Option<NodeId>, } // A generous allowance for how many times we may retry serializable transactions // before giving up. This is not expected to be hit: it is a defensive measure in case we // somehow engineer a situation where duelling transactions might otherwise live-lock. const MAX_RETRIES: usize = 128; impl Persistence { // The default postgres connection limit is 100. We use up to 99, to leave one free for a human admin under // normal circumstances. This assumes we have exclusive use of the database cluster to which we connect. pub const MAX_CONNECTIONS: u32 = 99; // We don't want to keep a lot of connections alive: close them down promptly if they aren't being used. const IDLE_CONNECTION_TIMEOUT: Duration = Duration::from_secs(10); const MAX_CONNECTION_LIFETIME: Duration = Duration::from_secs(60); pub async fn new(database_url: String) -> Self { let mut mgr_config = ManagerConfig::default(); mgr_config.custom_setup = Box::new(establish_connection_rustls); let manager = AsyncDieselConnectionManager::<AsyncPgConnection>::new_with_config( database_url, mgr_config, ); // We will use a connection pool: this is primarily to _limit_ our connection count, rather than to optimize time // to execute queries (database queries are not generally on latency-sensitive paths). let connection_pool = Pool::builder() .max_size(Self::MAX_CONNECTIONS) .max_lifetime(Some(Self::MAX_CONNECTION_LIFETIME)) .idle_timeout(Some(Self::IDLE_CONNECTION_TIMEOUT)) // Always keep at least one connection ready to go .min_idle(Some(1)) .test_on_check_out(true) .build(manager) .await .expect("Could not build connection pool"); Self { connection_pool } } /// A helper for use during startup, where we would like to tolerate concurrent restarts of the /// database and the storage controller, therefore the database might not be available right away pub async fn await_connection( database_url: &str, timeout: Duration, ) -> Result<(), diesel::ConnectionError> { let started_at = Instant::now(); log_postgres_connstr_info(database_url) .map_err(|e| diesel::ConnectionError::InvalidConnectionUrl(e.to_string()))?; loop { match establish_connection_rustls(database_url).await { Ok(_) => { tracing::info!("Connected to database."); return Ok(()); } Err(e) => { if started_at.elapsed() > timeout { return Err(e); } else { tracing::info!("Database not yet available, waiting... ({e})"); tokio::time::sleep(Duration::from_millis(100)).await; } } } } } /// Execute the diesel migrations that are built into this binary pub(crate) async fn migration_run(&self) -> DatabaseResult<()> { use diesel_migrations::{HarnessWithOutput, MigrationHarness}; // Can't use self.with_conn here as we do spawn_blocking which requires static. let conn = self .connection_pool .dedicated_connection() .await .map_err(|e| DatabaseError::Migration(e.to_string()))?; let mut async_wrapper: AsyncConnectionWrapper<AsyncPgConnection> = AsyncConnectionWrapper::from(conn); tokio::task::spawn_blocking(move || { let mut retry_count = 0; loop { let result = HarnessWithOutput::write_to_stdout(&mut async_wrapper) .run_pending_migrations(MIGRATIONS) .map(|_| ()) .map_err(|e| DatabaseError::Migration(e.to_string())); match result { Ok(r) => break Ok(r), Err( err @ DatabaseError::Query(diesel::result::Error::DatabaseError( diesel::result::DatabaseErrorKind::SerializationFailure, _, )), ) => { retry_count += 1; if retry_count > MAX_RETRIES { tracing::error!( "Exceeded max retries on SerializationFailure errors: {err:?}" ); break Err(err); } else { // Retry on serialization errors: these are expected, because even though our // transactions don't fight for the same rows, they will occasionally collide // on index pages (e.g. increment_generation for unrelated shards can collide) tracing::debug!( "Retrying transaction on serialization failure {err:?}" ); continue; } } Err(e) => break Err(e), } } }) .await .map_err(|e| DatabaseError::Migration(e.to_string()))??; Ok(()) } /// Wraps `with_conn` in order to collect latency and error metrics async fn with_measured_conn<'a, 'b, F, R>( &self, op: DatabaseOperation, func: F, ) -> DatabaseResult<R> where F: for<'r> Fn(&'r mut AsyncPgConnection) -> ScopedBoxFuture<'b, 'r, DatabaseResult<R>> + Send + std::marker::Sync + 'a, R: Send + 'b, { let latency = &METRICS_REGISTRY .metrics_group .storage_controller_database_query_latency; let _timer = latency.start_timer(DatabaseQueryLatencyLabelGroup { operation: op }); let res = self.with_conn(func).await; if let Err(err) = &res { let error_counter = &METRICS_REGISTRY .metrics_group .storage_controller_database_query_error; error_counter.inc(DatabaseQueryErrorLabelGroup { error_type: err.error_label(), operation: op, }) } res } /// Call the provided function with a Diesel database connection in a retry loop async fn with_conn<'a, 'b, F, R>(&self, func: F) -> DatabaseResult<R> where F: for<'r> Fn(&'r mut AsyncPgConnection) -> ScopedBoxFuture<'b, 'r, DatabaseResult<R>> + Send + std::marker::Sync + 'a, R: Send + 'b, { let mut retry_count = 0; loop { let mut conn = self.connection_pool.get().await?; match conn .build_transaction() .serializable() .run(|c| func(c)) .await { Ok(r) => break Ok(r), Err( err @ DatabaseError::Query(diesel::result::Error::DatabaseError( diesel::result::DatabaseErrorKind::SerializationFailure, _, )), ) => { retry_count += 1; if retry_count > MAX_RETRIES { tracing::error!( "Exceeded max retries on SerializationFailure errors: {err:?}" ); break Err(err); } else { // Retry on serialization errors: these are expected, because even though our // transactions don't fight for the same rows, they will occasionally collide // on index pages (e.g. increment_generation for unrelated shards can collide) tracing::debug!("Retrying transaction on serialization failure {err:?}"); continue; } } Err(e) => break Err(e), } } } /// When a node is first registered, persist it before using it for anything /// If the provided node_id already exists, it will be error. /// The common case is when a node marked for deletion wants to register. pub(crate) async fn insert_node(&self, node: &Node) -> DatabaseResult<()> { let np = &node.to_persistent(); self.with_measured_conn(DatabaseOperation::InsertNode, move |conn| { Box::pin(async move { diesel::insert_into(crate::schema::nodes::table) .values(np) .execute(conn) .await?; Ok(()) }) }) .await } /// At startup, populate the list of nodes which our shards may be placed on pub(crate) async fn list_nodes(&self) -> DatabaseResult<Vec<NodePersistence>> { use crate::schema::nodes::dsl::*; let result: Vec<NodePersistence> = self .with_measured_conn(DatabaseOperation::ListNodes, move |conn| { Box::pin(async move { Ok(crate::schema::nodes::table .filter(lifecycle.ne(String::from(NodeLifecycle::Deleted))) .load::<NodePersistence>(conn) .await?) }) }) .await?; tracing::info!("list_nodes: loaded {} nodes", result.len()); Ok(result) } pub(crate) async fn list_tombstones(&self) -> DatabaseResult<Vec<NodePersistence>> { use crate::schema::nodes::dsl::*; let result: Vec<NodePersistence> = self .with_measured_conn(DatabaseOperation::ListTombstones, move |conn| { Box::pin(async move { Ok(crate::schema::nodes::table .filter(lifecycle.eq(String::from(NodeLifecycle::Deleted))) .load::<NodePersistence>(conn) .await?) }) }) .await?; tracing::info!("list_tombstones: loaded {} nodes", result.len()); Ok(result) } pub(crate) async fn update_node<V>( &self, input_node_id: NodeId, values: V, ) -> DatabaseResult<()> where V: diesel::AsChangeset<Target = crate::schema::nodes::table> + Clone + Send + Sync, V::Changeset: diesel::query_builder::QueryFragment<diesel::pg::Pg> + Send, // valid Postgres SQL { use crate::schema::nodes::dsl::*; let updated = self .with_measured_conn(DatabaseOperation::UpdateNode, move |conn| { let values = values.clone(); Box::pin(async move { let updated = diesel::update(nodes) .filter(node_id.eq(input_node_id.0 as i64)) .filter(lifecycle.ne(String::from(NodeLifecycle::Deleted))) .set(values) .execute(conn) .await?; Ok(updated) }) }) .await?; if updated != 1 { Err(DatabaseError::Logical(format!( "Node {node_id:?} not found for update", ))) } else { Ok(()) } } pub(crate) async fn update_node_scheduling_policy( &self, input_node_id: NodeId, input_scheduling: NodeSchedulingPolicy, ) -> DatabaseResult<()> { use crate::schema::nodes::dsl::*; self.update_node( input_node_id, scheduling_policy.eq(String::from(input_scheduling)), ) .await } pub(crate) async fn update_node_on_registration( &self, input_node_id: NodeId, input_https_port: Option<u16>, input_grpc_addr: Option<String>, input_grpc_port: Option<u16>, ) -> DatabaseResult<()> { use crate::schema::nodes::dsl::*; self.update_node( input_node_id, ( listen_https_port.eq(input_https_port.map(|x| x as i32)), listen_grpc_addr.eq(input_grpc_addr), listen_grpc_port.eq(input_grpc_port.map(|x| x as i32)), ), ) .await } /// Tombstone is a special state where the node is not deleted from the database, /// but it is not available for usage. /// The main reason for it is to prevent the flaky node to register. pub(crate) async fn set_tombstone(&self, del_node_id: NodeId) -> DatabaseResult<()> { use crate::schema::nodes::dsl::*; self.update_node( del_node_id, lifecycle.eq(String::from(NodeLifecycle::Deleted)), ) .await } pub(crate) async fn delete_node(&self, del_node_id: NodeId) -> DatabaseResult<()> { use crate::schema::nodes::dsl::*; self.with_measured_conn(DatabaseOperation::DeleteNode, move |conn| { Box::pin(async move { // You can hard delete a node only if it has a tombstone. // So we need to check if the node has lifecycle set to deleted. let node_to_delete = nodes .filter(node_id.eq(del_node_id.0 as i64)) .first::<NodePersistence>(conn) .await .optional()?; if let Some(np) = node_to_delete { let lc = NodeLifecycle::from_str(&np.lifecycle).map_err(|e| { DatabaseError::Logical(format!( "Node {del_node_id} has invalid lifecycle: {e}" )) })?; if lc != NodeLifecycle::Deleted { return Err(DatabaseError::Logical(format!( "Node {del_node_id} was not soft deleted before, cannot hard delete it" ))); } diesel::delete(nodes) .filter(node_id.eq(del_node_id.0 as i64)) .execute(conn) .await?; } Ok(()) }) }) .await } /// At startup, load the high level state for shards, such as their config + policy. This will /// be enriched at runtime with state discovered on pageservers. /// /// We exclude shards configured to be detached. During startup, if we see any attached locations /// for such shards, they will automatically be detached as 'orphans'. pub(crate) async fn load_active_tenant_shards( &self, ) -> DatabaseResult<Vec<TenantShardPersistence>> { use crate::schema::tenant_shards::dsl::*; self.with_measured_conn(DatabaseOperation::ListTenantShards, move |conn| { Box::pin(async move { let query = tenant_shards.filter( placement_policy.ne(serde_json::to_string(&PlacementPolicy::Detached).unwrap()), ); let result = query.load::<TenantShardPersistence>(conn).await?; Ok(result) }) }) .await } /// When restoring a previously detached tenant into memory, load it from the database pub(crate) async fn load_tenant( &self, filter_tenant_id: TenantId, ) -> DatabaseResult<Vec<TenantShardPersistence>> { use crate::schema::tenant_shards::dsl::*; self.with_measured_conn(DatabaseOperation::LoadTenant, move |conn| { Box::pin(async move { let query = tenant_shards.filter(tenant_id.eq(filter_tenant_id.to_string())); let result = query.load::<TenantShardPersistence>(conn).await?; Ok(result) }) }) .await } /// Tenants must be persisted before we schedule them for the first time. This enables us /// to correctly retain generation monotonicity, and the externally provided placement policy & config. pub(crate) async fn insert_tenant_shards( &self, shards: Vec<TenantShardPersistence>, ) -> DatabaseResult<()> { use crate::schema::{metadata_health, tenant_shards}; let now = chrono::Utc::now(); let metadata_health_records = shards .iter() .map(|t| MetadataHealthPersistence { tenant_id: t.tenant_id.clone(), shard_number: t.shard_number, shard_count: t.shard_count, healthy: true, last_scrubbed_at: now, }) .collect::<Vec<_>>(); let shards = &shards; let metadata_health_records = &metadata_health_records; self.with_measured_conn(DatabaseOperation::InsertTenantShards, move |conn| { Box::pin(async move { diesel::insert_into(tenant_shards::table) .values(shards) .execute(conn) .await?; diesel::insert_into(metadata_health::table) .values(metadata_health_records) .execute(conn) .await?; Ok(()) }) }) .await } /// Ordering: call this _after_ deleting the tenant on pageservers, but _before_ dropping state for /// the tenant from memory on this server. pub(crate) async fn delete_tenant(&self, del_tenant_id: TenantId) -> DatabaseResult<()> { use crate::schema::tenant_shards::dsl::*; self.with_measured_conn(DatabaseOperation::DeleteTenant, move |conn| { Box::pin(async move { // `metadata_health` status (if exists) is also deleted based on the cascade behavior. diesel::delete(tenant_shards) .filter(tenant_id.eq(del_tenant_id.to_string())) .execute(conn) .await?; Ok(()) }) }) .await } /// When a tenant invokes the /re-attach API, this function is responsible for doing an efficient /// batched increment of the generations of all tenants whose generation_pageserver is equal to /// the node that called /re-attach. #[tracing::instrument(skip_all, fields(node_id))] pub(crate) async fn re_attach( &self, input_node_id: NodeId, ) -> DatabaseResult<HashMap<TenantShardId, Generation>> { use crate::schema::nodes::dsl::{scheduling_policy, *}; use crate::schema::tenant_shards::dsl::*; let updated = self .with_measured_conn(DatabaseOperation::ReAttach, move |conn| { Box::pin(async move { let node: Option<NodePersistence> = nodes .filter(node_id.eq(input_node_id.0 as i64)) .first::<NodePersistence>(conn) .await .optional()?; // Check if the node is not marked as deleted match node { Some(node) if matches!(NodeLifecycle::from_str(&node.lifecycle), Ok(NodeLifecycle::Deleted)) => { return Err(DatabaseError::Logical(format!( "Node {input_node_id} is marked as deleted, re-attach is not allowed" ))); } _ => { // go through } }; let rows_updated = diesel::update(tenant_shards) .filter(generation_pageserver.eq(input_node_id.0 as i64)) .set(generation.eq(generation + 1)) .execute(conn) .await?; tracing::info!("Incremented {} tenants' generations", rows_updated); // TODO: UPDATE+SELECT in one query let updated = tenant_shards .filter(generation_pageserver.eq(input_node_id.0 as i64)) .select(TenantShardPersistence::as_select()) .load(conn) .await?; if let Some(node) = node { let old_scheduling_policy = NodeSchedulingPolicy::from_str(&node.scheduling_policy).unwrap(); let new_scheduling_policy = match old_scheduling_policy { NodeSchedulingPolicy::Active => NodeSchedulingPolicy::Active, NodeSchedulingPolicy::PauseForRestart => NodeSchedulingPolicy::Active, NodeSchedulingPolicy::Draining => NodeSchedulingPolicy::Active, NodeSchedulingPolicy::Filling => NodeSchedulingPolicy::Active, NodeSchedulingPolicy::Pause => NodeSchedulingPolicy::Pause, NodeSchedulingPolicy::Deleting => NodeSchedulingPolicy::Pause, }; diesel::update(nodes) .filter(node_id.eq(input_node_id.0 as i64)) .set(scheduling_policy.eq(String::from(new_scheduling_policy))) .execute(conn) .await?; } Ok(updated) }) }) .await?; let mut result = HashMap::new(); for tsp in updated { let tenant_shard_id = TenantShardId { tenant_id: TenantId::from_str(tsp.tenant_id.as_str()) .map_err(|e| DatabaseError::Logical(format!("Malformed tenant id: {e}")))?, shard_number: ShardNumber(tsp.shard_number as u8), shard_count: ShardCount::new(tsp.shard_count as u8), }; let Some(g) = tsp.generation else { // If the generation_pageserver column was non-NULL, then the generation column should also be non-NULL: // we only set generation_pageserver when setting generation. return Err(DatabaseError::Logical( "Generation should always be set after incrementing".to_string(), )); }; result.insert(tenant_shard_id, Generation::new(g as u32)); } Ok(result) } /// Reconciler calls this immediately before attaching to a new pageserver, to acquire a unique, monotonically /// advancing generation number. We also store the NodeId for which the generation was issued, so that in /// [`Self::re_attach`] we can do a bulk UPDATE on the generations for that node. pub(crate) async fn increment_generation( &self, tenant_shard_id: TenantShardId, node_id: NodeId, ) -> anyhow::Result<Generation> { use crate::schema::tenant_shards::dsl::*; let updated = self .with_measured_conn(DatabaseOperation::IncrementGeneration, move |conn| { Box::pin(async move { let updated = diesel::update(tenant_shards) .filter(tenant_id.eq(tenant_shard_id.tenant_id.to_string())) .filter(shard_number.eq(tenant_shard_id.shard_number.0 as i32)) .filter(shard_count.eq(tenant_shard_id.shard_count.literal() as i32)) .set(( generation.eq(generation + 1), generation_pageserver.eq(node_id.0 as i64), )) // TODO: only returning() the generation column .returning(TenantShardPersistence::as_returning()) .get_result(conn) .await?; Ok(updated) }) }) .await?; // Generation is always non-null in the rseult: if the generation column had been NULL, then we // should have experienced an SQL Confilict error while executing a query that tries to increment it. debug_assert!(updated.generation.is_some()); let Some(g) = updated.generation else { return Err(DatabaseError::Logical( "Generation should always be set after incrementing".to_string(), ) .into()); }; Ok(Generation::new(g as u32)) } /// When we want to call out to the running shards for a tenant, e.g. during timeline CRUD operations, /// we need to know where the shard is attached, _and_ the generation, so that we can re-check the generation /// afterwards to confirm that our timeline CRUD operation is truly persistent (it must have happened in the /// latest generation) /// /// If the tenant doesn't exist, an empty vector is returned. /// /// Output is sorted by shard number pub(crate) async fn tenant_generations( &self, filter_tenant_id: TenantId, ) -> Result<Vec<ShardGenerationState>, DatabaseError> { use crate::schema::tenant_shards::dsl::*; let rows = self .with_measured_conn(DatabaseOperation::TenantGenerations, move |conn| { Box::pin(async move { let result = tenant_shards .filter(tenant_id.eq(filter_tenant_id.to_string())) .select(TenantShardPersistence::as_select()) .order(shard_number) .load(conn) .await?; Ok(result) }) }) .await?; Ok(rows .into_iter() .map(|p| ShardGenerationState { tenant_shard_id: p .get_tenant_shard_id() .expect("Corrupt tenant shard id in database"), generation: p.generation.map(|g| Generation::new(g as u32)), generation_pageserver: p.generation_pageserver.map(|n| NodeId(n as u64)), }) .collect()) } /// Read the generation number of specific tenant shards /// /// Output is unsorted. Output may not include values for all inputs, if they are missing in the database. pub(crate) async fn shard_generations( &self, mut tenant_shard_ids: impl Iterator<Item = &TenantShardId>, ) -> Result<Vec<(TenantShardId, Option<Generation>)>, DatabaseError> {
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/safekeeper.rs
storage_controller/src/safekeeper.rs
use std::time::Duration; use pageserver_api::controller_api::{SafekeeperDescribeResponse, SkSchedulingPolicy}; use reqwest::StatusCode; use safekeeper_api::membership::SafekeeperId; use safekeeper_client::mgmt_api; use tokio_util::sync::CancellationToken; use utils::backoff; use utils::id::NodeId; use utils::logging::SecretString; use crate::heartbeater::SafekeeperState; use crate::persistence::{DatabaseError, SafekeeperPersistence}; use crate::safekeeper_client::SafekeeperClient; #[derive(Clone)] pub struct Safekeeper { pub(crate) skp: SafekeeperPersistence, cancel: CancellationToken, listen_http_addr: String, listen_http_port: u16, listen_https_port: Option<u16>, scheduling_policy: SkSchedulingPolicy, id: NodeId, /// Heartbeating result. availability: SafekeeperState, // Flag from storcon's config to use https for safekeeper API. // Invariant: if |true|, listen_https_port should contain a value. use_https: bool, } impl Safekeeper { pub(crate) fn from_persistence( skp: SafekeeperPersistence, cancel: CancellationToken, use_https: bool, ) -> anyhow::Result<Self> { if use_https && skp.https_port.is_none() { anyhow::bail!( "cannot load safekeeper {} from persistence: \ https is enabled, but https port is not specified", skp.id, ); } let scheduling_policy = skp.scheduling_policy.0; Ok(Self { cancel, listen_http_addr: skp.host.clone(), listen_http_port: skp.http_port as u16, listen_https_port: skp.https_port.map(|x| x as u16), id: NodeId(skp.id as u64), skp, availability: SafekeeperState::Offline, scheduling_policy, use_https, }) } pub(crate) fn base_url(&self) -> String { if self.use_https { format!( "https://{}:{}", self.listen_http_addr, self.listen_https_port .expect("https port should be specified if use_https is on"), ) } else { format!("http://{}:{}", self.listen_http_addr, self.listen_http_port) } } pub(crate) fn get_id(&self) -> NodeId { self.id } pub(crate) fn describe_response(&self) -> Result<SafekeeperDescribeResponse, DatabaseError> { self.skp.as_describe_response() } pub(crate) fn set_availability(&mut self, availability: SafekeeperState) { self.availability = availability; } pub(crate) fn scheduling_policy(&self) -> SkSchedulingPolicy { self.scheduling_policy } pub(crate) fn set_scheduling_policy(&mut self, scheduling_policy: SkSchedulingPolicy) { self.scheduling_policy = scheduling_policy; self.skp.scheduling_policy = scheduling_policy.into(); } pub(crate) fn availability(&self) -> SafekeeperState { self.availability.clone() } pub(crate) fn has_https_port(&self) -> bool { self.listen_https_port.is_some() } pub(crate) fn get_safekeeper_id(&self) -> SafekeeperId { SafekeeperId { id: self.id, host: self.skp.host.clone(), pg_port: self.skp.port as u16, } } /// Perform an operation (which is given a [`SafekeeperClient`]) with retries #[allow(clippy::too_many_arguments)] pub(crate) async fn with_client_retries<T, O, F>( &self, mut op: O, http_client: &reqwest::Client, jwt: &Option<SecretString>, warn_threshold: u32, max_retries: u32, timeout: Duration, cancel: &CancellationToken, ) -> mgmt_api::Result<T> where O: FnMut(SafekeeperClient) -> F, F: std::future::Future<Output = mgmt_api::Result<T>>, { fn is_fatal(e: &mgmt_api::Error) -> bool { use mgmt_api::Error::*; match e { ReceiveBody(_) | ReceiveErrorBody(_) => false, ApiError(StatusCode::SERVICE_UNAVAILABLE, _) | ApiError(StatusCode::GATEWAY_TIMEOUT, _) | ApiError(StatusCode::REQUEST_TIMEOUT, _) => false, ApiError(_, _) => true, Cancelled => true, Timeout(_) => false, } } backoff::retry( || { let client = SafekeeperClient::new( self.get_id(), http_client.clone(), self.base_url(), jwt.clone(), ); let node_cancel_fut = self.cancel.cancelled(); let op_fut = tokio::time::timeout(timeout, op(client)); async { tokio::select! { r = op_fut => match r { Ok(r) => r, Err(e) => Err(mgmt_api::Error::Timeout(format!("{e}"))), }, _ = node_cancel_fut => { Err(mgmt_api::Error::Cancelled) }} } }, is_fatal, warn_threshold, max_retries, &format!( "Call to safekeeper {} ({}) management API", self.id, self.base_url(), ), cancel, ) .await .unwrap_or(Err(mgmt_api::Error::Cancelled)) } pub(crate) fn update_from_record( &mut self, record: crate::persistence::SafekeeperUpsert, ) -> anyhow::Result<()> { let crate::persistence::SafekeeperUpsert { active: _, availability_zone_id: _, host, http_port, https_port, id, port: _, region_id: _, version: _, } = record.clone(); if id != self.id.0 as i64 { // The way the function is called ensures this. If we regress on that, it's a bug. panic!( "id can't be changed via update_from_record function: {id} != {}", self.id.0 ); } if self.use_https && https_port.is_none() { anyhow::bail!( "cannot update safekeeper {id}: \ https is enabled, but https port is not specified" ); } self.skp = crate::persistence::SafekeeperPersistence::from_upsert(record, self.scheduling_policy); self.listen_http_port = http_port as u16; self.listen_https_port = https_port.map(|x| x as u16); self.listen_http_addr = host; Ok(()) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service.rs
storage_controller/src/service.rs
pub mod chaos_injector; pub mod feature_flag; pub(crate) mod safekeeper_reconciler; mod safekeeper_service; mod tenant_shard_iterator; use std::borrow::Cow; use std::cmp::Ordering; use std::collections::{BTreeMap, HashMap, HashSet}; use std::error::Error; use std::num::NonZeroU32; use std::ops::{Deref, DerefMut}; use std::path::PathBuf; use std::str::FromStr; use std::sync::{Arc, OnceLock}; use std::time::{Duration, Instant, SystemTime}; use anyhow::Context; use control_plane::storage_controller::{ AttachHookRequest, AttachHookResponse, InspectRequest, InspectResponse, }; use diesel::result::DatabaseErrorKind; use futures::StreamExt; use futures::stream::FuturesUnordered; use http_utils::error::ApiError; use hyper::Uri; use itertools::Itertools; use pageserver_api::config::PostHogConfig; use pageserver_api::controller_api::{ AvailabilityZone, MetadataHealthRecord, MetadataHealthUpdateRequest, NodeAvailability, NodeRegisterRequest, NodeSchedulingPolicy, NodeShard, NodeShardResponse, PlacementPolicy, ShardSchedulingPolicy, ShardsPreferredAzsRequest, ShardsPreferredAzsResponse, SkSchedulingPolicy, TenantCreateRequest, TenantCreateResponse, TenantCreateResponseShard, TenantDescribeResponse, TenantDescribeResponseShard, TenantLocateResponse, TenantPolicyRequest, TenantShardMigrateRequest, TenantShardMigrateResponse, TenantTimelineDescribeResponse, }; use pageserver_api::models::{ self, DetachBehavior, LocationConfig, LocationConfigListResponse, LocationConfigMode, LsnLease, PageserverUtilization, SecondaryProgress, ShardImportStatus, ShardParameters, TenantConfig, TenantConfigPatchRequest, TenantConfigRequest, TenantLocationConfigRequest, TenantLocationConfigResponse, TenantShardLocation, TenantShardSplitRequest, TenantShardSplitResponse, TenantSorting, TenantTimeTravelRequest, TimelineArchivalConfigRequest, TimelineCreateRequest, TimelineCreateResponseStorcon, TimelineInfo, TopTenantShardItem, TopTenantShardsRequest, }; use pageserver_api::shard::{ DEFAULT_STRIPE_SIZE, ShardCount, ShardIdentity, ShardNumber, ShardStripeSize, TenantShardId, }; use pageserver_api::upcall_api::{ PutTimelineImportStatusRequest, ReAttachRequest, ReAttachResponse, ReAttachResponseTenant, TimelineImportStatusRequest, ValidateRequest, ValidateResponse, ValidateResponseTenant, }; use pageserver_client::{BlockUnblock, mgmt_api}; use reqwest::{Certificate, StatusCode}; use safekeeper_api::models::SafekeeperUtilization; use safekeeper_reconciler::SafekeeperReconcilers; use tenant_shard_iterator::{TenantShardExclusiveIterator, create_shared_shard_iterator}; use tokio::sync::TryAcquireError; use tokio::sync::mpsc::error::TrySendError; use tokio_util::sync::CancellationToken; use tracing::{Instrument, debug, error, info, info_span, instrument, warn}; use utils::completion::Barrier; use utils::env; use utils::generation::Generation; use utils::id::{NodeId, TenantId, TimelineId}; use utils::lsn::Lsn; use utils::shard::ShardIndex; use utils::sync::gate::{Gate, GateGuard}; use utils::{failpoint_support, pausable_failpoint}; use crate::background_node_operations::{ Delete, Drain, Fill, MAX_RECONCILES_PER_OPERATION, Operation, OperationError, OperationHandler, }; use crate::compute_hook::{self, ComputeHook, NotifyError}; use crate::heartbeater::{Heartbeater, PageserverState, SafekeeperState}; use crate::id_lock_map::{ IdLockMap, TracingExclusiveGuard, trace_exclusive_lock, trace_shared_lock, }; use crate::leadership::Leadership; use crate::metrics; use crate::node::{AvailabilityTransition, Node}; use crate::operation_utils::{self, TenantShardDrain, TenantShardDrainAction}; use crate::pageserver_client::PageserverClient; use crate::peer_client::GlobalObservedState; use crate::persistence::split_state::SplitState; use crate::persistence::{ AbortShardSplitStatus, ControllerPersistence, DatabaseError, DatabaseResult, MetadataHealthPersistence, Persistence, ShardGenerationState, TenantFilter, TenantShardPersistence, }; use crate::reconciler::{ ReconcileError, ReconcileUnits, ReconcilerConfig, ReconcilerConfigBuilder, ReconcilerPriority, attached_location_conf, }; use crate::safekeeper::Safekeeper; use crate::scheduler::{ AttachedShardTag, MaySchedule, ScheduleContext, ScheduleError, ScheduleMode, Scheduler, }; use crate::tenant_shard::{ IntentState, MigrateAttachment, ObservedState, ObservedStateDelta, ObservedStateLocation, ReconcileNeeded, ReconcileResult, ReconcileWaitError, ReconcilerStatus, ReconcilerWaiter, ScheduleOptimization, ScheduleOptimizationAction, TenantShard, }; use crate::timeline_import::{ FinalizingImport, ImportResult, ShardImportStatuses, TimelineImport, TimelineImportFinalizeError, TimelineImportState, UpcallClient, }; const WAITER_OPERATION_POLL_TIMEOUT: Duration = Duration::from_millis(500); // For operations that should be quick, like attaching a new tenant const SHORT_RECONCILE_TIMEOUT: Duration = Duration::from_secs(5); // For operations that might be slow, like migrating a tenant with // some data in it. pub const RECONCILE_TIMEOUT: Duration = Duration::from_secs(30); // If we receive a call using Secondary mode initially, it will omit generation. We will initialize // tenant shards into this generation, and as long as it remains in this generation, we will accept // input generation from future requests as authoritative. const INITIAL_GENERATION: Generation = Generation::new(0); /// How long [`Service::startup_reconcile`] is allowed to take before it should give /// up on unresponsive pageservers and proceed. pub(crate) const STARTUP_RECONCILE_TIMEOUT: Duration = Duration::from_secs(30); /// How long a node may be unresponsive to heartbeats before we declare it offline. /// This must be long enough to cover node restarts as well as normal operations: in future pub const MAX_OFFLINE_INTERVAL_DEFAULT: Duration = Duration::from_secs(30); /// How long a node may be unresponsive to heartbeats during start up before we declare it /// offline. /// /// This is much more lenient than [`MAX_OFFLINE_INTERVAL_DEFAULT`] since the pageserver's /// handling of the re-attach response may take a long time and blocks heartbeats from /// being handled on the pageserver side. pub const MAX_WARMING_UP_INTERVAL_DEFAULT: Duration = Duration::from_secs(300); /// How often to send heartbeats to registered nodes? pub const HEARTBEAT_INTERVAL_DEFAULT: Duration = Duration::from_secs(5); /// How long is too long for a reconciliation? pub const LONG_RECONCILE_THRESHOLD_DEFAULT: Duration = Duration::from_secs(120); #[derive(Clone, strum_macros::Display)] enum TenantOperations { Create, LocationConfig, ConfigSet, ConfigPatch, TimeTravelRemoteStorage, Delete, UpdatePolicy, ShardSplit, SecondaryDownload, TimelineCreate, TimelineDelete, AttachHook, TimelineArchivalConfig, TimelineDetachAncestor, TimelineGcBlockUnblock, DropDetached, DownloadHeatmapLayers, TimelineLsnLease, TimelineSafekeeperMigrate, } #[derive(Clone, strum_macros::Display)] enum NodeOperations { Register, Configure, Delete, DeleteTombstone, } /// The leadership status for the storage controller process. /// Allowed transitions are: /// 1. Leader -> SteppedDown /// 2. Candidate -> Leader #[derive( Eq, PartialEq, Copy, Clone, strum_macros::Display, strum_macros::EnumIter, measured::FixedCardinalityLabel, )] #[strum(serialize_all = "snake_case")] pub(crate) enum LeadershipStatus { /// This is the steady state where the storage controller can produce /// side effects in the cluster. Leader, /// We've been notified to step down by another candidate. No reconciliations /// take place in this state. SteppedDown, /// Initial state for a new storage controller instance. Will attempt to assume leadership. #[allow(unused)] Candidate, } enum ShardGenerationValidity { Valid, Mismatched { claimed: Generation, actual: Option<Generation>, }, } pub const RECONCILER_CONCURRENCY_DEFAULT: usize = 128; pub const PRIORITY_RECONCILER_CONCURRENCY_DEFAULT: usize = 256; pub const SAFEKEEPER_RECONCILER_CONCURRENCY_DEFAULT: usize = 32; // Number of consecutive reconciliations that have occurred for one shard, // after which the shard is ignored when considering to run optimizations. const MAX_CONSECUTIVE_RECONCILES: usize = 10; // Depth of the channel used to enqueue shards for reconciliation when they can't do it immediately. // This channel is finite-size to avoid using excessive memory if we get into a state where reconciles are finishing more slowly // than they're being pushed onto the queue. const MAX_DELAYED_RECONCILES: usize = 10000; // Top level state available to all HTTP handlers struct ServiceState { leadership_status: LeadershipStatus, tenants: BTreeMap<TenantShardId, TenantShard>, nodes: Arc<HashMap<NodeId, Node>>, safekeepers: Arc<HashMap<NodeId, Safekeeper>>, safekeeper_reconcilers: SafekeeperReconcilers, scheduler: Scheduler, /// Ongoing background operation on the cluster if any is running. /// Note that only one such operation may run at any given time, /// hence the type choice. ongoing_operation: Option<OperationHandler>, /// Queue of tenants who are waiting for concurrency limits to permit them to reconcile delayed_reconcile_rx: tokio::sync::mpsc::Receiver<TenantShardId>, /// Tracks ongoing timeline import finalization tasks imports_finalizing: BTreeMap<(TenantId, TimelineId), FinalizingImport>, } /// Transform an error from a pageserver into an error to return to callers of a storage /// controller API. fn passthrough_api_error(node: &Node, e: mgmt_api::Error) -> ApiError { match e { mgmt_api::Error::SendRequest(e) => { // Presume errors sending requests are connectivity/availability issues ApiError::ResourceUnavailable(format!("{node} error sending request: {e}").into()) } mgmt_api::Error::ReceiveErrorBody(str) => { // Presume errors receiving body are connectivity/availability issues ApiError::ResourceUnavailable( format!("{node} error receiving error body: {str}").into(), ) } mgmt_api::Error::ReceiveBody(err) if err.is_decode() => { // Return 500 for decoding errors. ApiError::InternalServerError(anyhow::Error::from(err).context("error decoding body")) } mgmt_api::Error::ReceiveBody(err) => { // Presume errors receiving body are connectivity/availability issues except for decoding errors let src_str = err.source().map(|e| e.to_string()).unwrap_or_default(); ApiError::ResourceUnavailable( format!("{node} error receiving error body: {err} {src_str}").into(), ) } mgmt_api::Error::ApiError(StatusCode::NOT_FOUND, msg) => { ApiError::NotFound(anyhow::anyhow!(format!("{node}: {msg}")).into()) } mgmt_api::Error::ApiError(StatusCode::SERVICE_UNAVAILABLE, msg) => { ApiError::ResourceUnavailable(format!("{node}: {msg}").into()) } mgmt_api::Error::ApiError(status @ StatusCode::UNAUTHORIZED, msg) | mgmt_api::Error::ApiError(status @ StatusCode::FORBIDDEN, msg) => { // Auth errors talking to a pageserver are not auth errors for the caller: they are // internal server errors, showing that something is wrong with the pageserver or // storage controller's auth configuration. ApiError::InternalServerError(anyhow::anyhow!("{node} {status}: {msg}")) } mgmt_api::Error::ApiError(status @ StatusCode::TOO_MANY_REQUESTS, msg) => { // Pass through 429 errors: if pageserver is asking us to wait + retry, we in // turn ask our clients to wait + retry ApiError::Conflict(format!("{node} {status}: {status} {msg}")) } mgmt_api::Error::ApiError(status, msg) => { // Presume general case of pageserver API errors is that we tried to do something // that can't be done right now. ApiError::Conflict(format!("{node} {status}: {status} {msg}")) } mgmt_api::Error::Cancelled => ApiError::ShuttingDown, mgmt_api::Error::Timeout(e) => ApiError::Timeout(e.into()), } } impl ServiceState { fn new( nodes: HashMap<NodeId, Node>, safekeepers: HashMap<NodeId, Safekeeper>, tenants: BTreeMap<TenantShardId, TenantShard>, scheduler: Scheduler, delayed_reconcile_rx: tokio::sync::mpsc::Receiver<TenantShardId>, initial_leadership_status: LeadershipStatus, reconcilers_cancel: CancellationToken, ) -> Self { metrics::update_leadership_status(initial_leadership_status); Self { leadership_status: initial_leadership_status, tenants, nodes: Arc::new(nodes), safekeepers: Arc::new(safekeepers), safekeeper_reconcilers: SafekeeperReconcilers::new(reconcilers_cancel), scheduler, ongoing_operation: None, delayed_reconcile_rx, imports_finalizing: Default::default(), } } fn parts_mut( &mut self, ) -> ( &mut Arc<HashMap<NodeId, Node>>, &mut BTreeMap<TenantShardId, TenantShard>, &mut Scheduler, ) { (&mut self.nodes, &mut self.tenants, &mut self.scheduler) } #[allow(clippy::type_complexity)] fn parts_mut_sk( &mut self, ) -> ( &mut Arc<HashMap<NodeId, Node>>, &mut Arc<HashMap<NodeId, Safekeeper>>, &mut BTreeMap<TenantShardId, TenantShard>, &mut Scheduler, ) { ( &mut self.nodes, &mut self.safekeepers, &mut self.tenants, &mut self.scheduler, ) } fn get_leadership_status(&self) -> LeadershipStatus { self.leadership_status } fn step_down(&mut self) { self.leadership_status = LeadershipStatus::SteppedDown; metrics::update_leadership_status(self.leadership_status); } fn become_leader(&mut self) { self.leadership_status = LeadershipStatus::Leader; metrics::update_leadership_status(self.leadership_status); } } #[derive(Clone)] pub struct Config { // All pageservers managed by one instance of this service must have // the same public key. This JWT token will be used to authenticate // this service to the pageservers it manages. pub pageserver_jwt_token: Option<String>, // All safekeepers managed by one instance of this service must have // the same public key. This JWT token will be used to authenticate // this service to the safekeepers it manages. pub safekeeper_jwt_token: Option<String>, // This JWT token will be used to authenticate this service to the control plane. pub control_plane_jwt_token: Option<String>, // This JWT token will be used to authenticate with other storage controller instances pub peer_jwt_token: Option<String>, /// Prefix for storage API endpoints of the control plane. We use this prefix to compute /// URLs that we use to send pageserver and safekeeper attachment locations. /// If this is None, the compute hook will assume it is running in a test environment /// and try to invoke neon_local instead. pub control_plane_url: Option<String>, /// Grace period within which a pageserver does not respond to heartbeats, but is still /// considered active. Once the grace period elapses, the next heartbeat failure will /// mark the pagseserver offline. pub max_offline_interval: Duration, /// Extended grace period within which pageserver may not respond to heartbeats. /// This extended grace period kicks in after the node has been drained for restart /// and/or upon handling the re-attach request from a node. pub max_warming_up_interval: Duration, /// How many normal-priority Reconcilers may be spawned concurrently pub reconciler_concurrency: usize, /// How many high-priority Reconcilers may be spawned concurrently pub priority_reconciler_concurrency: usize, /// How many safekeeper reconciles may happen concurrently (per safekeeper) pub safekeeper_reconciler_concurrency: usize, /// How many API requests per second to allow per tenant, across all /// tenant-scoped API endpoints. Further API requests queue until ready. pub tenant_rate_limit: NonZeroU32, /// If a tenant shard's largest timeline (max_logical_size) exceeds this value, all tenant /// shards will be split in 2 until they fall below split_threshold (up to max_split_shards). /// /// This will greedily split into as many shards as necessary to fall below split_threshold, as /// powers of 2: if a tenant shard is 7 times larger than split_threshold, it will split into 8 /// immediately, rather than first 2 then 4 then 8. /// /// None or 0 disables auto-splitting. /// /// TODO: consider using total logical size of all timelines instead. pub split_threshold: Option<u64>, /// The maximum number of shards a tenant can be split into during autosplits. Does not affect /// manual split requests. 0 or 1 disables autosplits, as we already have 1 shard. pub max_split_shards: u8, /// The size at which an unsharded tenant should initially split. Ingestion is significantly /// faster with multiple shards, so eagerly splitting below split_threshold will typically speed /// up initial ingestion of large tenants. /// /// This should be below split_threshold, but it is not required. If both split_threshold and /// initial_split_threshold qualify, the largest number of target shards will be used. /// /// Does not apply to already sharded tenants: changing initial_split_threshold or /// initial_split_shards is not retroactive for already-sharded tenants. /// /// None or 0 disables initial splits. pub initial_split_threshold: Option<u64>, /// The number of shards to split into when reaching initial_split_threshold. Will /// be clamped to max_split_shards. /// /// 0 or 1 disables initial splits. Has no effect if initial_split_threshold is disabled. pub initial_split_shards: u8, // TODO: make this cfg(feature = "testing") pub neon_local_repo_dir: Option<PathBuf>, // Maximum acceptable download lag for the secondary location // while draining a node. If the secondary location is lagging // by more than the configured amount, then the secondary is not // upgraded to primary. pub max_secondary_lag_bytes: Option<u64>, pub heartbeat_interval: Duration, pub address_for_peers: Option<Uri>, pub start_as_candidate: bool, pub long_reconcile_threshold: Duration, pub use_https_pageserver_api: bool, pub use_https_safekeeper_api: bool, pub ssl_ca_certs: Vec<Certificate>, pub timelines_onto_safekeepers: bool, pub use_local_compute_notifications: bool, /// Number of safekeepers to choose for a timeline when creating it. /// Safekeepers will be choosen from different availability zones. pub timeline_safekeeper_count: usize, /// PostHog integration config pub posthog_config: Option<PostHogConfig>, /// When set, actively checks and initiates heatmap downloads/uploads. pub kick_secondary_downloads: bool, /// Timeout used for HTTP client of split requests. [`Duration::MAX`] if None. pub shard_split_request_timeout: Duration, // Feature flag: Whether the storage controller should act to rectify pageserver-reported local disk loss. pub handle_ps_local_disk_loss: bool, } impl From<DatabaseError> for ApiError { fn from(err: DatabaseError) -> ApiError { match err { DatabaseError::Query(e) => ApiError::InternalServerError(e.into()), // FIXME: ApiError doesn't have an Unavailable variant, but ShuttingDown maps to 503. DatabaseError::Connection(_) | DatabaseError::ConnectionPool(_) => { ApiError::ShuttingDown } DatabaseError::Logical(reason) | DatabaseError::Migration(reason) => { ApiError::InternalServerError(anyhow::anyhow!(reason)) } DatabaseError::Cas(reason) => ApiError::Conflict(reason), } } } enum InitialShardScheduleOutcome { Scheduled(TenantCreateResponseShard), NotScheduled, ShardScheduleError(ScheduleError), } pub struct Service { inner: Arc<std::sync::RwLock<ServiceState>>, config: Config, persistence: Arc<Persistence>, compute_hook: Arc<ComputeHook>, result_tx: tokio::sync::mpsc::UnboundedSender<ReconcileResultRequest>, heartbeater_ps: Heartbeater<Node, PageserverState>, heartbeater_sk: Heartbeater<Safekeeper, SafekeeperState>, // Channel for background cleanup from failed operations that require cleanup, such as shard split abort_tx: tokio::sync::mpsc::UnboundedSender<TenantShardSplitAbort>, // Locking on a tenant granularity (covers all shards in the tenant): // - Take exclusively for rare operations that mutate the tenant's persistent state (e.g. create/delete/split) // - Take in shared mode for operations that need the set of shards to stay the same to complete reliably (e.g. timeline CRUD) tenant_op_locks: IdLockMap<TenantId, TenantOperations>, // Locking for node-mutating operations: take exclusively for operations that modify the node's persistent state, or // that transition it to/from Active. node_op_locks: IdLockMap<NodeId, NodeOperations>, // Limit how many Reconcilers we will spawn concurrently for normal-priority tasks such as background reconciliations // and reconciliation on startup. reconciler_concurrency: Arc<tokio::sync::Semaphore>, // Limit how many Reconcilers we will spawn concurrently for high-priority tasks such as tenant/timeline CRUD, which // a human user might be waiting for. priority_reconciler_concurrency: Arc<tokio::sync::Semaphore>, /// Queue of tenants who are waiting for concurrency limits to permit them to reconcile /// Send into this queue to promptly attempt to reconcile this shard next time units are available. /// /// Note that this state logically lives inside ServiceState, but carrying Sender here makes the code simpler /// by avoiding needing a &mut ref to something inside the ServiceState. This could be optimized to /// use a VecDeque instead of a channel to reduce synchronization overhead, at the cost of some code complexity. delayed_reconcile_tx: tokio::sync::mpsc::Sender<TenantShardId>, // Process shutdown will fire this token cancel: CancellationToken, // Child token of [`Service::cancel`] used by reconcilers reconcilers_cancel: CancellationToken, // Background tasks will hold this gate gate: Gate, // Reconcilers background tasks will hold this gate reconcilers_gate: Gate, /// This waits for initial reconciliation with pageservers to complete. Until this barrier /// passes, it isn't safe to do any actions that mutate tenants. pub(crate) startup_complete: Barrier, /// HTTP client with proper CA certs. http_client: reqwest::Client, /// Handle for the step down background task if one was ever requested step_down_barrier: OnceLock<tokio::sync::watch::Receiver<Option<GlobalObservedState>>>, } impl From<ReconcileWaitError> for ApiError { fn from(value: ReconcileWaitError) -> Self { match value { ReconcileWaitError::Shutdown => ApiError::ShuttingDown, e @ ReconcileWaitError::Timeout(_) => ApiError::Timeout(format!("{e}").into()), e @ ReconcileWaitError::Failed(..) => ApiError::InternalServerError(anyhow::anyhow!(e)), } } } impl From<OperationError> for ApiError { fn from(value: OperationError) -> Self { match value { OperationError::NodeStateChanged(err) | OperationError::FinalizeError(err) | OperationError::ImpossibleConstraint(err) => { ApiError::InternalServerError(anyhow::anyhow!(err)) } OperationError::Cancelled => ApiError::Conflict("Operation was cancelled".into()), } } } #[allow(clippy::large_enum_variant)] enum TenantCreateOrUpdate { Create(TenantCreateRequest), Update(Vec<ShardUpdate>), } struct ShardSplitParams { old_shard_count: ShardCount, new_shard_count: ShardCount, new_stripe_size: Option<ShardStripeSize>, targets: Vec<ShardSplitTarget>, policy: PlacementPolicy, config: TenantConfig, shard_ident: ShardIdentity, preferred_az_id: Option<AvailabilityZone>, } // When preparing for a shard split, we may either choose to proceed with the split, // or find that the work is already done and return NoOp. enum ShardSplitAction { Split(Box<ShardSplitParams>), NoOp(TenantShardSplitResponse), } // A parent shard which will be split struct ShardSplitTarget { parent_id: TenantShardId, node: Node, child_ids: Vec<TenantShardId>, } /// When we tenant shard split operation fails, we may not be able to clean up immediately, because nodes /// might not be available. We therefore use a queue of abort operations processed in the background. struct TenantShardSplitAbort { tenant_id: TenantId, /// The target values from the request that failed new_shard_count: ShardCount, new_stripe_size: Option<ShardStripeSize>, /// Until this abort op is complete, no other operations may be done on the tenant _tenant_lock: TracingExclusiveGuard<TenantOperations>, /// The reconciler gate for the duration of the split operation, and any included abort. _gate: GateGuard, } #[derive(thiserror::Error, Debug)] enum TenantShardSplitAbortError { #[error(transparent)] Database(#[from] DatabaseError), #[error(transparent)] Remote(#[from] mgmt_api::Error), #[error("Unavailable")] Unavailable, } /// Inputs for computing a target shard count for a tenant. struct ShardSplitInputs { /// Current shard count. shard_count: ShardCount, /// Total size of largest timeline summed across all shards. max_logical_size: u64, /// Size-based split threshold. Zero if size-based splits are disabled. split_threshold: u64, /// Upper bound on target shards. 0 or 1 disables splits. max_split_shards: u8, /// Initial split threshold. Zero if initial splits are disabled. initial_split_threshold: u64, /// Number of shards for initial splits. 0 or 1 disables initial splits. initial_split_shards: u8, } struct ShardUpdate { tenant_shard_id: TenantShardId, placement_policy: PlacementPolicy, tenant_config: TenantConfig, /// If this is None, generation is not updated. generation: Option<Generation>, /// If this is None, scheduling policy is not updated. scheduling_policy: Option<ShardSchedulingPolicy>, } enum StopReconciliationsReason { ShuttingDown, SteppingDown, } impl std::fmt::Display for StopReconciliationsReason { fn fmt(&self, writer: &mut std::fmt::Formatter) -> std::fmt::Result { let s = match self { Self::ShuttingDown => "Shutting down", Self::SteppingDown => "Stepping down", }; write!(writer, "{s}") } } pub(crate) enum ReconcileResultRequest { ReconcileResult(ReconcileResult), Stop, } #[derive(Clone)] pub(crate) struct MutationLocation { pub(crate) node: Node, pub(crate) generation: Generation, } #[derive(Clone)] pub(crate) struct ShardMutationLocations { pub(crate) latest: MutationLocation, pub(crate) other: Vec<MutationLocation>, } #[derive(Default, Clone)] pub(crate) struct TenantMutationLocations(pub BTreeMap<TenantShardId, ShardMutationLocations>); struct ReconcileAllResult { spawned_reconciles: usize, stuck_reconciles: usize, has_delayed_reconciles: bool, } impl ReconcileAllResult { fn new( spawned_reconciles: usize, stuck_reconciles: usize, has_delayed_reconciles: bool, ) -> Self { assert!( spawned_reconciles >= stuck_reconciles, "It is impossible to have less spawned reconciles than stuck reconciles" ); Self { spawned_reconciles, stuck_reconciles, has_delayed_reconciles, } } /// We can run optimizations only if we don't have any delayed reconciles and /// all spawned reconciles are also stuck reconciles. fn can_run_optimizations(&self) -> bool { !self.has_delayed_reconciles && self.spawned_reconciles == self.stuck_reconciles } } enum TenantIdOrShardId { TenantId(TenantId), TenantShardId(TenantShardId), } impl TenantIdOrShardId { fn tenant_id(&self) -> TenantId { match self { TenantIdOrShardId::TenantId(tenant_id) => *tenant_id, TenantIdOrShardId::TenantShardId(tenant_shard_id) => tenant_shard_id.tenant_id, } } fn matches(&self, tenant_shard_id: &TenantShardId) -> bool { match self { TenantIdOrShardId::TenantId(tenant_id) => tenant_shard_id.tenant_id == *tenant_id, TenantIdOrShardId::TenantShardId(this_tenant_shard_id) => { this_tenant_shard_id == tenant_shard_id } } } } impl Service { pub fn get_config(&self) -> &Config { &self.config } pub fn get_http_client(&self) -> &reqwest::Client { &self.http_client } /// Called once on startup, this function attempts to contact all pageservers to build an up-to-date /// view of the world, and determine which pageservers are responsive. #[instrument(skip_all)] async fn startup_reconcile( self: &Arc<Service>, current_leader: Option<ControllerPersistence>, leader_step_down_state: Option<GlobalObservedState>, bg_compute_notify_result_tx: tokio::sync::mpsc::Sender< Result<(), (TenantShardId, NotifyError)>, >, ) { // Startup reconciliation does I/O to other services: whether they // are responsive or not, we should aim to finish within our deadline, because: // - If we don't, a k8s readiness hook watching /ready will kill us. // - While we're waiting for startup reconciliation, we are not fully // available for end user operations like creating/deleting tenants and timelines. // // We set multiple deadlines to break up the time available between the phases of work: this is // arbitrary, but avoids a situation where the first phase could burn our entire timeout period. let start_at = Instant::now(); let node_scan_deadline = start_at .checked_add(STARTUP_RECONCILE_TIMEOUT / 2) .expect("Reconcile timeout is a modest constant"); let observed = if let Some(state) = leader_step_down_state { tracing::info!( "Using observed state received from leader at {}", current_leader.as_ref().unwrap().address ); state } else { self.build_global_observed_state(node_scan_deadline).await }; // Accumulate a list of any tenant locations that ought to be detached let mut cleanup = Vec::new(); // Send initial heartbeat requests to all nodes loaded from the database let all_nodes = { let locked = self.inner.read().unwrap(); locked.nodes.clone() }; let (mut nodes_online, mut sks_online) = self.initial_heartbeat_round(all_nodes.keys()).await; // List of tenants for which we will attempt to notify compute of their location at startup let mut compute_notifications = Vec::new(); // Populate intent and observed states for all tenants, based on reported state on pageservers tracing::info!("Populating tenant shards' states from initial pageserver scan..."); let shard_count = { let mut locked = self.inner.write().unwrap(); let (nodes, safekeepers, tenants, scheduler) = locked.parts_mut_sk(); // Mark nodes online if they responded to us: nodes are offline by default after a restart. let mut new_nodes = (**nodes).clone();
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/hadron_utils.rs
storage_controller/src/hadron_utils.rs
use std::collections::BTreeMap; use rand::Rng; use utils::shard::TenantShardId; static CHARSET: &[u8] = b"ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789!@#$%^&*()"; /// Generate a random string of `length` that can be used as a password. The generated string /// contains alphanumeric characters and special characters (!@#$%^&*()) pub fn generate_random_password(length: usize) -> String { let mut rng = rand::rng(); (0..length) .map(|_| { let idx = rng.random_range(0..CHARSET.len()); CHARSET[idx] as char }) .collect() } pub(crate) struct TenantShardSizeMap { #[expect(dead_code)] pub map: BTreeMap<TenantShardId, u64>, } impl TenantShardSizeMap { pub fn new(map: BTreeMap<TenantShardId, u64>) -> Self { Self { map } } } #[cfg(test)] mod test { use super::*; #[test] fn test_generate_random_password() { let pwd1 = generate_random_password(10); assert_eq!(pwd1.len(), 10); let pwd2 = generate_random_password(10); assert_ne!(pwd1, pwd2); assert!(pwd1.chars().all(|c| CHARSET.contains(&(c as u8)))); assert!(pwd2.chars().all(|c| CHARSET.contains(&(c as u8)))); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/schema.rs
storage_controller/src/schema.rs
// @generated automatically by Diesel CLI. pub mod sql_types { #[derive(diesel::query_builder::QueryId, diesel::sql_types::SqlType)] #[diesel(postgres_type(name = "pg_lsn", schema = "pg_catalog"))] pub struct PgLsn; } diesel::table! { controllers (address, started_at) { address -> Varchar, started_at -> Timestamptz, } } diesel::table! { hadron_safekeepers (sk_node_id) { sk_node_id -> Int8, listen_http_addr -> Varchar, listen_http_port -> Int4, listen_pg_addr -> Varchar, listen_pg_port -> Int4, } } diesel::table! { hadron_timeline_safekeepers (timeline_id, sk_node_id) { timeline_id -> Varchar, sk_node_id -> Int8, legacy_endpoint_id -> Nullable<Uuid>, } } diesel::table! { metadata_health (tenant_id, shard_number, shard_count) { tenant_id -> Varchar, shard_number -> Int4, shard_count -> Int4, healthy -> Bool, last_scrubbed_at -> Timestamptz, } } diesel::table! { nodes (node_id) { node_id -> Int8, scheduling_policy -> Varchar, listen_http_addr -> Varchar, listen_http_port -> Int4, listen_pg_addr -> Varchar, listen_pg_port -> Int4, availability_zone_id -> Varchar, listen_https_port -> Nullable<Int4>, lifecycle -> Varchar, listen_grpc_addr -> Nullable<Varchar>, listen_grpc_port -> Nullable<Int4>, } } diesel::table! { safekeeper_timeline_pending_ops (tenant_id, timeline_id, sk_id) { sk_id -> Int8, tenant_id -> Varchar, timeline_id -> Varchar, generation -> Int4, op_kind -> Varchar, } } diesel::table! { safekeepers (id) { id -> Int8, region_id -> Text, version -> Int8, host -> Text, port -> Int4, http_port -> Int4, availability_zone_id -> Text, scheduling_policy -> Varchar, https_port -> Nullable<Int4>, } } diesel::table! { tenant_shards (tenant_id, shard_number, shard_count) { tenant_id -> Varchar, shard_number -> Int4, shard_count -> Int4, shard_stripe_size -> Int4, generation -> Nullable<Int4>, generation_pageserver -> Nullable<Int8>, placement_policy -> Varchar, splitting -> Int2, config -> Text, scheduling_policy -> Varchar, preferred_az_id -> Nullable<Varchar>, } } diesel::table! { timeline_imports (tenant_id, timeline_id) { tenant_id -> Varchar, timeline_id -> Varchar, shard_statuses -> Jsonb, } } diesel::table! { use diesel::sql_types::*; use super::sql_types::PgLsn; timelines (tenant_id, timeline_id) { tenant_id -> Varchar, timeline_id -> Varchar, start_lsn -> PgLsn, generation -> Int4, sk_set -> Array<Nullable<Int8>>, new_sk_set -> Nullable<Array<Nullable<Int8>>>, cplane_notified_generation -> Int4, deleted_at -> Nullable<Timestamptz>, sk_set_notified_generation -> Int4, } } diesel::allow_tables_to_appear_in_same_query!( controllers, hadron_safekeepers, hadron_timeline_safekeepers, metadata_health, nodes, safekeeper_timeline_pending_ops, safekeepers, tenant_shards, timeline_imports, timelines, );
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/auth.rs
storage_controller/src/auth.rs
use utils::auth::{AuthError, Claims, Scope}; use uuid::Uuid; pub fn check_permission(claims: &Claims, required_scope: Scope) -> Result<(), AuthError> { if claims.scope != required_scope { return Err(AuthError("Scope mismatch. Permission denied".into())); } Ok(()) } #[allow(dead_code)] pub fn check_endpoint_permission(claims: &Claims, endpoint_id: Uuid) -> Result<(), AuthError> { if claims.scope != Scope::TenantEndpoint { return Err(AuthError("Scope mismatch. Permission denied".into())); } if claims.endpoint_id != Some(endpoint_id) { return Err(AuthError("Endpoint id mismatch. Permission denied".into())); } Ok(()) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/pageserver_client.rs
storage_controller/src/pageserver_client.rs
use std::time::Duration; use pageserver_api::models::detach_ancestor::AncestorDetached; use pageserver_api::models::{ DetachBehavior, LocationConfig, LocationConfigListResponse, LsnLease, PageserverUtilization, SecondaryProgress, TenantScanRemoteStorageResponse, TenantShardSplitRequest, TenantShardSplitResponse, TenantWaitLsnRequest, TimelineArchivalConfigRequest, TimelineCreateRequest, TimelineInfo, TopTenantShardsRequest, TopTenantShardsResponse, }; use pageserver_api::shard::TenantShardId; use pageserver_client::BlockUnblock; use pageserver_client::mgmt_api::{Client, Result}; use reqwest::StatusCode; use utils::id::{NodeId, TenantId, TimelineId}; use utils::lsn::Lsn; use crate::hadron_utils::TenantShardSizeMap; /// Thin wrapper around [`pageserver_client::mgmt_api::Client`]. It allows the storage /// controller to collect metrics in a non-intrusive manner. #[derive(Debug, Clone)] pub(crate) struct PageserverClient { inner: Client, node_id_label: String, } macro_rules! measured_request { ($name:literal, $method:expr, $node_id: expr, $invoke:expr) => {{ let labels = crate::metrics::PageserverRequestLabelGroup { pageserver_id: $node_id, path: $name, method: $method, }; let latency = &crate::metrics::METRICS_REGISTRY .metrics_group .storage_controller_pageserver_request_latency; let _timer_guard = latency.start_timer(labels.clone()); let res = $invoke; if res.is_err() { let error_counters = &crate::metrics::METRICS_REGISTRY .metrics_group .storage_controller_pageserver_request_error; error_counters.inc(labels) } res }}; } impl PageserverClient { pub(crate) fn new( node_id: NodeId, raw_client: reqwest::Client, mgmt_api_endpoint: String, jwt: Option<&str>, ) -> Self { Self { inner: Client::new(raw_client, mgmt_api_endpoint, jwt), node_id_label: node_id.0.to_string(), } } pub(crate) async fn tenant_delete(&self, tenant_shard_id: TenantShardId) -> Result<StatusCode> { measured_request!( "tenant", crate::metrics::Method::Delete, &self.node_id_label, self.inner.tenant_delete(tenant_shard_id).await ) } pub(crate) async fn tenant_time_travel_remote_storage( &self, tenant_shard_id: TenantShardId, timestamp: &str, done_if_after: &str, ) -> Result<()> { measured_request!( "tenant_time_travel_remote_storage", crate::metrics::Method::Put, &self.node_id_label, self.inner .tenant_time_travel_remote_storage(tenant_shard_id, timestamp, done_if_after) .await ) } #[expect(dead_code)] pub(crate) async fn tenant_timeline_compact( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, force_image_layer_creation: bool, wait_until_done: bool, ) -> Result<()> { measured_request!( "tenant_timeline_compact", crate::metrics::Method::Put, &self.node_id_label, self.inner .tenant_timeline_compact( tenant_shard_id, timeline_id, force_image_layer_creation, true, false, wait_until_done, ) .await ) } /* BEGIN_HADRON */ pub(crate) async fn tenant_timeline_describe( &self, tenant_shard_id: &TenantShardId, timeline_id: &TimelineId, ) -> Result<TimelineInfo> { measured_request!( "tenant_timeline_describe", crate::metrics::Method::Get, &self.node_id_label, self.inner .tenant_timeline_describe(tenant_shard_id, timeline_id,) .await ) } #[expect(dead_code)] pub(crate) async fn list_tenant_visible_size(&self) -> Result<TenantShardSizeMap> { measured_request!( "list_tenant_visible_size", crate::metrics::Method::Get, &self.node_id_label, self.inner.list_tenant_visible_size().await ) .map(TenantShardSizeMap::new) } /* END_HADRON */ pub(crate) async fn tenant_scan_remote_storage( &self, tenant_id: TenantId, ) -> Result<TenantScanRemoteStorageResponse> { measured_request!( "tenant_scan_remote_storage", crate::metrics::Method::Get, &self.node_id_label, self.inner.tenant_scan_remote_storage(tenant_id).await ) } pub(crate) async fn tenant_secondary_download( &self, tenant_id: TenantShardId, wait: Option<std::time::Duration>, ) -> Result<(StatusCode, SecondaryProgress)> { measured_request!( "tenant_secondary_download", crate::metrics::Method::Post, &self.node_id_label, self.inner.tenant_secondary_download(tenant_id, wait).await ) } pub(crate) async fn tenant_secondary_status( &self, tenant_shard_id: TenantShardId, ) -> Result<SecondaryProgress> { measured_request!( "tenant_secondary_status", crate::metrics::Method::Get, &self.node_id_label, self.inner.tenant_secondary_status(tenant_shard_id).await ) } pub(crate) async fn tenant_heatmap_upload(&self, tenant_id: TenantShardId) -> Result<()> { measured_request!( "tenant_heatmap_upload", crate::metrics::Method::Post, &self.node_id_label, self.inner.tenant_heatmap_upload(tenant_id).await ) } pub(crate) async fn location_config( &self, tenant_shard_id: TenantShardId, config: LocationConfig, flush_ms: Option<std::time::Duration>, lazy: bool, ) -> Result<()> { measured_request!( "location_config", crate::metrics::Method::Put, &self.node_id_label, self.inner .location_config(tenant_shard_id, config, flush_ms, lazy) .await ) } pub(crate) async fn list_location_config(&self) -> Result<LocationConfigListResponse> { measured_request!( "location_configs", crate::metrics::Method::Get, &self.node_id_label, self.inner.list_location_config().await ) } pub(crate) async fn get_location_config( &self, tenant_shard_id: TenantShardId, ) -> Result<Option<LocationConfig>> { measured_request!( "location_config", crate::metrics::Method::Get, &self.node_id_label, self.inner.get_location_config(tenant_shard_id).await ) } pub(crate) async fn timeline_create( &self, tenant_shard_id: TenantShardId, req: &TimelineCreateRequest, ) -> Result<TimelineInfo> { measured_request!( "timeline", crate::metrics::Method::Post, &self.node_id_label, self.inner.timeline_create(tenant_shard_id, req).await ) } pub(crate) async fn timeline_delete( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, ) -> Result<StatusCode> { measured_request!( "timeline", crate::metrics::Method::Delete, &self.node_id_label, self.inner .timeline_delete(tenant_shard_id, timeline_id) .await ) } pub(crate) async fn timeline_lease_lsn( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, lsn: Lsn, ) -> Result<LsnLease> { measured_request!( "timeline_lease_lsn", crate::metrics::Method::Post, &self.node_id_label, self.inner .timeline_init_lsn_lease(tenant_shard_id, timeline_id, lsn) .await ) } #[allow(unused)] pub(crate) async fn timeline_detail( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, ) -> Result<TimelineInfo> { measured_request!( "timeline_detail", crate::metrics::Method::Get, &self.node_id_label, self.inner .timeline_detail(tenant_shard_id, timeline_id) .await ) } pub(crate) async fn tenant_shard_split( &self, tenant_shard_id: TenantShardId, req: TenantShardSplitRequest, ) -> Result<TenantShardSplitResponse> { measured_request!( "tenant_shard_split", crate::metrics::Method::Put, &self.node_id_label, self.inner.tenant_shard_split(tenant_shard_id, req).await ) } pub(crate) async fn timeline_list( &self, tenant_shard_id: &TenantShardId, ) -> Result<Vec<TimelineInfo>> { measured_request!( "timelines", crate::metrics::Method::Get, &self.node_id_label, self.inner.timeline_list(tenant_shard_id).await ) } pub(crate) async fn timeline_archival_config( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, req: &TimelineArchivalConfigRequest, ) -> Result<()> { measured_request!( "timeline_archival_config", crate::metrics::Method::Put, &self.node_id_label, self.inner .timeline_archival_config(tenant_shard_id, timeline_id, req) .await ) } pub(crate) async fn timeline_detach_ancestor( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, behavior: Option<DetachBehavior>, ) -> Result<AncestorDetached> { measured_request!( "timeline_detach_ancestor", crate::metrics::Method::Put, &self.node_id_label, self.inner .timeline_detach_ancestor(tenant_shard_id, timeline_id, behavior) .await ) } pub(crate) async fn timeline_block_unblock_gc( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, dir: BlockUnblock, ) -> Result<()> { // measuring these makes no sense because we synchronize with the gc loop and remote // storage on block_gc so there should be huge outliers measured_request!( "timeline_block_unblock_gc", crate::metrics::Method::Post, &self.node_id_label, self.inner .timeline_block_unblock_gc(tenant_shard_id, timeline_id, dir) .await ) } pub(crate) async fn timeline_download_heatmap_layers( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, concurrency: Option<usize>, recurse: bool, ) -> Result<()> { measured_request!( "download_heatmap_layers", crate::metrics::Method::Post, &self.node_id_label, self.inner .timeline_download_heatmap_layers( tenant_shard_id, timeline_id, concurrency, recurse ) .await ) } pub(crate) async fn get_utilization(&self) -> Result<PageserverUtilization> { measured_request!( "utilization", crate::metrics::Method::Get, &self.node_id_label, self.inner.get_utilization().await ) } pub(crate) async fn top_tenant_shards( &self, request: TopTenantShardsRequest, ) -> Result<TopTenantShardsResponse> { measured_request!( "top_tenants", crate::metrics::Method::Post, &self.node_id_label, self.inner.top_tenant_shards(request).await ) } #[expect(dead_code)] pub(crate) async fn reset_alert_gauges(&self) -> Result<()> { measured_request!( "reset_alert_gauges", crate::metrics::Method::Post, &self.node_id_label, self.inner.reset_alert_gauges().await ) } pub(crate) async fn wait_lsn( &self, tenant_shard_id: TenantShardId, request: TenantWaitLsnRequest, ) -> Result<StatusCode> { measured_request!( "wait_lsn", crate::metrics::Method::Post, &self.node_id_label, self.inner.wait_lsn(tenant_shard_id, request).await ) } pub(crate) async fn activate_post_import( &self, tenant_shard_id: TenantShardId, timeline_id: TimelineId, timeline_activate_timeout: Duration, ) -> Result<TimelineInfo> { measured_request!( "activate_post_import", crate::metrics::Method::Put, &self.node_id_label, self.inner .activate_post_import(tenant_shard_id, timeline_id, timeline_activate_timeout) .await ) } pub(crate) async fn update_feature_flag_spec(&self, spec: String) -> Result<()> { measured_request!( "update_feature_flag_spec", crate::metrics::Method::Post, &self.node_id_label, self.inner.update_feature_flag_spec(spec).await ) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/id_lock_map.rs
storage_controller/src/id_lock_map.rs
use std::collections::HashMap; use std::fmt::Display; use std::sync::Arc; use std::time::{Duration, Instant}; use crate::service::RECONCILE_TIMEOUT; const LOCK_TIMEOUT_ALERT_THRESHOLD: Duration = RECONCILE_TIMEOUT; /// A wrapper around `OwnedRwLockWriteGuard` used for tracking the /// operation that holds the lock, and print a warning if it exceeds /// the LOCK_TIMEOUT_ALERT_THRESHOLD time pub struct TracingExclusiveGuard<T: Display> { guard: tokio::sync::OwnedRwLockWriteGuard<Option<T>>, start: Instant, } impl<T: Display> TracingExclusiveGuard<T> { pub fn new(guard: tokio::sync::OwnedRwLockWriteGuard<Option<T>>) -> Self { Self { guard, start: Instant::now(), } } } impl<T: Display> Drop for TracingExclusiveGuard<T> { fn drop(&mut self) { let duration = self.start.elapsed(); if duration > LOCK_TIMEOUT_ALERT_THRESHOLD { tracing::warn!( "Exclusive lock by {} was held for {:?}", self.guard.as_ref().unwrap(), duration ); } *self.guard = None; } } // A wrapper around `OwnedRwLockReadGuard` used for tracking the /// operation that holds the lock, and print a warning if it exceeds /// the LOCK_TIMEOUT_ALERT_THRESHOLD time pub struct TracingSharedGuard<T: Display> { _guard: tokio::sync::OwnedRwLockReadGuard<Option<T>>, operation: T, start: Instant, } impl<T: Display> TracingSharedGuard<T> { pub fn new(guard: tokio::sync::OwnedRwLockReadGuard<Option<T>>, operation: T) -> Self { Self { _guard: guard, operation, start: Instant::now(), } } } impl<T: Display> Drop for TracingSharedGuard<T> { fn drop(&mut self) { let duration = self.start.elapsed(); if duration > LOCK_TIMEOUT_ALERT_THRESHOLD { tracing::warn!( "Shared lock by {} was held for {:?}", self.operation, duration ); } } } /// A map of locks covering some arbitrary identifiers. Useful if you have a collection of objects but don't /// want to embed a lock in each one, or if your locking granularity is different to your object granularity. /// For example, used in the storage controller where the objects are tenant shards, but sometimes locking /// is needed at a tenant-wide granularity. pub(crate) struct IdLockMap<T, I> where T: Eq + PartialEq + std::hash::Hash, { /// A synchronous lock for getting/setting the async locks that our callers will wait on. entities: std::sync::Mutex<std::collections::HashMap<T, Arc<tokio::sync::RwLock<Option<I>>>>>, } impl<T, I> IdLockMap<T, I> where T: Eq + PartialEq + std::hash::Hash, I: Display, { pub(crate) fn shared( &self, key: T, operation: I, ) -> impl std::future::Future<Output = TracingSharedGuard<I>> { let mut locked = self.entities.lock().unwrap(); let entry = locked.entry(key).or_default().clone(); async move { TracingSharedGuard::new(entry.read_owned().await, operation) } } pub(crate) fn exclusive( &self, key: T, operation: I, ) -> impl std::future::Future<Output = TracingExclusiveGuard<I>> { let mut locked = self.entities.lock().unwrap(); let entry = locked.entry(key).or_default().clone(); async move { let mut guard = TracingExclusiveGuard::new(entry.write_owned().await); *guard.guard = Some(operation); guard } } pub(crate) fn try_exclusive(&self, key: T, operation: I) -> Option<TracingExclusiveGuard<I>> { let mut locked = self.entities.lock().unwrap(); let entry = locked.entry(key).or_default().clone(); let mut guard = TracingExclusiveGuard::new(entry.try_write_owned().ok()?); *guard.guard = Some(operation); Some(guard) } /// Rather than building a lock guard that re-takes the [`Self::entities`] lock, we just do /// periodic housekeeping to avoid the map growing indefinitely pub(crate) fn housekeeping(&self) { let mut locked = self.entities.lock().unwrap(); locked.retain(|_k, entry| entry.try_write().is_err()) } } impl<T, I> Default for IdLockMap<T, I> where T: Eq + PartialEq + std::hash::Hash, { fn default() -> Self { Self { entities: std::sync::Mutex::new(HashMap::new()), } } } pub async fn trace_exclusive_lock< T: Clone + Display + Eq + PartialEq + std::hash::Hash, I: Clone + Display, >( op_locks: &IdLockMap<T, I>, key: T, operation: I, ) -> TracingExclusiveGuard<I> { let start = Instant::now(); let guard = op_locks.exclusive(key.clone(), operation.clone()).await; let duration = start.elapsed(); if duration > LOCK_TIMEOUT_ALERT_THRESHOLD { tracing::warn!( "Operation {} on key {} has waited {:?} for exclusive lock", operation, key, duration ); } guard } pub async fn trace_shared_lock< T: Clone + Display + Eq + PartialEq + std::hash::Hash, I: Clone + Display, >( op_locks: &IdLockMap<T, I>, key: T, operation: I, ) -> TracingSharedGuard<I> { let start = Instant::now(); let guard = op_locks.shared(key.clone(), operation.clone()).await; let duration = start.elapsed(); if duration > LOCK_TIMEOUT_ALERT_THRESHOLD { tracing::warn!( "Operation {} on key {} has waited {:?} for shared lock", operation, key, duration ); } guard } #[cfg(test)] mod tests { use super::IdLockMap; #[derive(Clone, Debug, strum_macros::Display, PartialEq)] enum Operations { Op1, Op2, } #[tokio::test] async fn multiple_shared_locks() { let id_lock_map: IdLockMap<i32, Operations> = IdLockMap::default(); let shared_lock_1 = id_lock_map.shared(1, Operations::Op1).await; let shared_lock_2 = id_lock_map.shared(1, Operations::Op2).await; assert_eq!(shared_lock_1.operation, Operations::Op1); assert_eq!(shared_lock_2.operation, Operations::Op2); } #[tokio::test] async fn exclusive_locks() { let id_lock_map = IdLockMap::default(); let resource_id = 1; { let _ex_lock = id_lock_map.exclusive(resource_id, Operations::Op1).await; assert_eq!(_ex_lock.guard.clone().unwrap(), Operations::Op1); let _ex_lock_2 = tokio::time::timeout( tokio::time::Duration::from_millis(1), id_lock_map.exclusive(resource_id, Operations::Op2), ) .await; assert!(_ex_lock_2.is_err()); } let shared_lock_1 = id_lock_map.shared(resource_id, Operations::Op1).await; assert_eq!(shared_lock_1.operation, Operations::Op1); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/main.rs
storage_controller/src/main.rs
use std::num::NonZeroU32; use std::path::PathBuf; use std::sync::Arc; use std::time::Duration; use anyhow::{Context, anyhow}; use camino::Utf8PathBuf; use clap::{ArgAction, Parser}; use futures::future::OptionFuture; use http_utils::tls_certs::ReloadingCertificateResolver; use hyper0::Uri; use metrics::BuildInfo; use metrics::launch_timestamp::LaunchTimestamp; use pageserver_api::config::PostHogConfig; use reqwest::Certificate; use storage_controller::http::make_router; use storage_controller::metrics::preinitialize_metrics; use storage_controller::persistence::Persistence; use storage_controller::service::chaos_injector::ChaosInjector; use storage_controller::service::feature_flag::FeatureFlagService; use storage_controller::service::{ Config, HEARTBEAT_INTERVAL_DEFAULT, LONG_RECONCILE_THRESHOLD_DEFAULT, MAX_OFFLINE_INTERVAL_DEFAULT, MAX_WARMING_UP_INTERVAL_DEFAULT, PRIORITY_RECONCILER_CONCURRENCY_DEFAULT, RECONCILER_CONCURRENCY_DEFAULT, SAFEKEEPER_RECONCILER_CONCURRENCY_DEFAULT, Service, }; use tokio::signal::unix::SignalKind; use tokio_util::sync::CancellationToken; use tracing::Instrument; use utils::auth::{JwtAuth, SwappableJwtAuth}; use utils::logging::{self, LogFormat}; use utils::sentry_init::init_sentry; use utils::{project_build_tag, project_git_version, tcp_listener}; project_git_version!(GIT_VERSION); project_build_tag!(BUILD_TAG); #[global_allocator] static GLOBAL: tikv_jemallocator::Jemalloc = tikv_jemallocator::Jemalloc; /// Configure jemalloc to profile heap allocations by sampling stack traces every 2 MB (1 << 21). /// This adds roughly 3% overhead for allocations on average, which is acceptable considering /// performance-sensitive code will avoid allocations as far as possible anyway. #[allow(non_upper_case_globals)] #[unsafe(export_name = "malloc_conf")] pub static malloc_conf: &[u8] = b"prof:true,prof_active:true,lg_prof_sample:21\0"; const DEFAULT_SSL_KEY_FILE: &str = "server.key"; const DEFAULT_SSL_CERT_FILE: &str = "server.crt"; const DEFAULT_SSL_CERT_RELOAD_PERIOD: &str = "60s"; #[derive(Parser)] #[command(author, version, about, long_about = None)] #[command(arg_required_else_help(true))] #[clap(group( clap::ArgGroup::new("listen-addresses") .required(true) .multiple(true) .args(&["listen", "listen_https"]), ))] struct Cli { /// Host and port to listen HTTP on, like `127.0.0.1:1234`. /// At least one of ["listen", "listen_https"] should be specified. // TODO: Make this option dev-only when https is out everywhere. #[arg(short, long)] listen: Option<std::net::SocketAddr>, /// Host and port to listen HTTPS on, like `127.0.0.1:1234`. /// At least one of ["listen", "listen_https"] should be specified. #[arg(long)] listen_https: Option<std::net::SocketAddr>, /// Public key for JWT authentication of clients #[arg(long)] public_key: Option<String>, /// Token for authenticating this service with the pageservers it controls #[arg(long)] jwt_token: Option<String>, /// Token for authenticating this service with the safekeepers it controls #[arg(long)] safekeeper_jwt_token: Option<String>, /// Token for authenticating this service with the control plane, when calling /// the compute notification endpoint #[arg(long)] control_plane_jwt_token: Option<String>, #[arg(long)] peer_jwt_token: Option<String>, /// URL to control plane storage API prefix #[arg(long)] control_plane_url: Option<String>, /// URL to connect to postgres, like postgresql://localhost:1234/storage_controller #[arg(long)] database_url: Option<String>, /// Flag to enable dev mode, which permits running without auth #[arg(long, default_value = "false")] dev: bool, /// Grace period before marking unresponsive pageserver offline #[arg(long)] max_offline_interval: Option<humantime::Duration>, /// More tolerant grace period before marking unresponsive pagserver offline used /// around pageserver restarts #[arg(long)] max_warming_up_interval: Option<humantime::Duration>, /// Size threshold for automatically splitting shards (disabled by default) #[arg(long)] split_threshold: Option<u64>, /// Maximum number of shards during autosplits. 0 disables autosplits. Defaults /// to 16 as a safety to avoid too many shards by accident. #[arg(long, default_value = "16")] max_split_shards: u8, /// Size threshold for initial shard splits of unsharded tenants. 0 disables initial splits. #[arg(long)] initial_split_threshold: Option<u64>, /// Number of target shards for initial splits. 0 or 1 disables initial splits. Defaults to 2. #[arg(long, default_value = "2")] initial_split_shards: u8, /// Maximum number of normal-priority reconcilers that may run in parallel #[arg(long)] reconciler_concurrency: Option<usize>, /// Maximum number of high-priority reconcilers that may run in parallel #[arg(long)] priority_reconciler_concurrency: Option<usize>, /// Maximum number of safekeeper reconciliations that may run in parallel (per safekeeper) #[arg(long)] safekeeper_reconciler_concurrency: Option<usize>, /// Tenant API rate limit, as requests per second per tenant. #[arg(long, default_value = "10")] tenant_rate_limit: NonZeroU32, /// How long to wait for the initial database connection to be available. #[arg(long, default_value = "5s")] db_connect_timeout: humantime::Duration, #[arg(long, default_value = "false")] start_as_candidate: bool, // TODO: make this mandatory once the helm chart gets updated #[arg(long)] address_for_peers: Option<Uri>, /// `neon_local` sets this to the path of the neon_local repo dir. /// Only relevant for testing. // TODO: make `cfg(feature = "testing")` #[arg(long)] neon_local_repo_dir: Option<PathBuf>, /// Chaos testing: exercise tenant migrations #[arg(long)] chaos_interval: Option<humantime::Duration>, /// Chaos testing: exercise an immediate exit #[arg(long)] chaos_exit_crontab: Option<cron::Schedule>, /// Maximum acceptable lag for the secondary location while draining /// a pageserver #[arg(long)] max_secondary_lag_bytes: Option<u64>, /// Period with which to send heartbeats to registered nodes #[arg(long)] heartbeat_interval: Option<humantime::Duration>, #[arg(long)] long_reconcile_threshold: Option<humantime::Duration>, /// Flag to use https for requests to pageserver API. #[arg(long, default_value = "false")] use_https_pageserver_api: bool, // Whether to put timelines onto safekeepers #[arg(long, default_value = "false")] timelines_onto_safekeepers: bool, /// Flag to use https for requests to safekeeper API. #[arg(long, default_value = "false")] use_https_safekeeper_api: bool, /// Path to a file with certificate's private key for https API. #[arg(long, default_value = DEFAULT_SSL_KEY_FILE)] ssl_key_file: Utf8PathBuf, /// Path to a file with a X509 certificate for https API. #[arg(long, default_value = DEFAULT_SSL_CERT_FILE)] ssl_cert_file: Utf8PathBuf, /// Period to reload certificate and private key from files. #[arg(long, default_value = DEFAULT_SSL_CERT_RELOAD_PERIOD)] ssl_cert_reload_period: humantime::Duration, /// Trusted root CA certificates to use in https APIs. #[arg(long)] ssl_ca_file: Option<Utf8PathBuf>, /// Neon local specific flag. When set, ignore [`Cli::control_plane_url`] and deliver /// the compute notification directly (instead of via control plane). #[arg(long, default_value = "false")] use_local_compute_notifications: bool, /// Number of safekeepers to choose for a timeline when creating it. /// Safekeepers will be choosen from different availability zones. /// This option exists primarily for testing purposes. #[arg(long, default_value = "3", value_parser = clap::builder::RangedU64ValueParser::<usize>::new().range(1..))] timeline_safekeeper_count: usize, /// When set, actively checks and initiates heatmap downloads/uploads during reconciliation. /// This speed up migrations by avoiding the default wait for the heatmap download interval. /// Primarily useful for testing to reduce test execution time. #[arg(long, default_value = "false", action=ArgAction::Set)] kick_secondary_downloads: bool, #[arg(long)] shard_split_request_timeout: Option<humantime::Duration>, /// **Feature Flag** Whether the storage controller should act to rectify pageserver-reported local disk loss. #[arg(long, default_value = "false")] handle_ps_local_disk_loss: bool, } enum StrictMode { /// In strict mode, we will require that all secrets are loaded, i.e. security features /// may not be implicitly turned off by omitting secrets in the environment. Strict, /// In dev mode, secrets are optional, and omitting a particular secret will implicitly /// disable the auth related to it (e.g. no pageserver jwt key -> send unauthenticated /// requests, no public key -> don't authenticate incoming requests). Dev, } impl Default for StrictMode { fn default() -> Self { Self::Strict } } /// Secrets may either be provided on the command line (for testing), or loaded from AWS SecretManager: this /// type encapsulates the logic to decide which and do the loading. struct Secrets { database_url: String, public_key: Option<JwtAuth>, pageserver_jwt_token: Option<String>, safekeeper_jwt_token: Option<String>, control_plane_jwt_token: Option<String>, peer_jwt_token: Option<String>, } const POSTHOG_CONFIG_ENV: &str = "POSTHOG_CONFIG"; impl Secrets { const DATABASE_URL_ENV: &'static str = "DATABASE_URL"; const PAGESERVER_JWT_TOKEN_ENV: &'static str = "PAGESERVER_JWT_TOKEN"; const SAFEKEEPER_JWT_TOKEN_ENV: &'static str = "SAFEKEEPER_JWT_TOKEN"; const CONTROL_PLANE_JWT_TOKEN_ENV: &'static str = "CONTROL_PLANE_JWT_TOKEN"; const PEER_JWT_TOKEN_ENV: &'static str = "PEER_JWT_TOKEN"; const PUBLIC_KEY_ENV: &'static str = "PUBLIC_KEY"; /// Load secrets from, in order of preference: /// - CLI args if database URL is provided on the CLI /// - Environment variables if DATABASE_URL is set. async fn load(args: &Cli) -> anyhow::Result<Self> { let Some(database_url) = Self::load_secret(&args.database_url, Self::DATABASE_URL_ENV) else { anyhow::bail!( "Database URL is not set (set `--database-url`, or `DATABASE_URL` environment)" ) }; let public_key = match Self::load_secret(&args.public_key, Self::PUBLIC_KEY_ENV) { Some(v) => Some(JwtAuth::from_key(v).context("Loading public key")?), None => None, }; let this = Self { database_url, public_key, pageserver_jwt_token: Self::load_secret( &args.jwt_token, Self::PAGESERVER_JWT_TOKEN_ENV, ), safekeeper_jwt_token: Self::load_secret( &args.safekeeper_jwt_token, Self::SAFEKEEPER_JWT_TOKEN_ENV, ), control_plane_jwt_token: Self::load_secret( &args.control_plane_jwt_token, Self::CONTROL_PLANE_JWT_TOKEN_ENV, ), peer_jwt_token: Self::load_secret(&args.peer_jwt_token, Self::PEER_JWT_TOKEN_ENV), }; Ok(this) } fn load_secret(cli: &Option<String>, env_name: &str) -> Option<String> { if let Some(v) = cli { Some(v.clone()) } else { std::env::var(env_name).ok() } } } fn main() -> anyhow::Result<()> { logging::init( LogFormat::Plain, logging::TracingErrorLayerEnablement::Disabled, logging::Output::Stdout, )?; // log using tracing so we don't get confused output by default hook writing to stderr utils::logging::replace_panic_hook_with_tracing_panic_hook().forget(); let _sentry_guard = init_sentry(Some(GIT_VERSION.into()), &[]); let hook = std::panic::take_hook(); std::panic::set_hook(Box::new(move |info| { // let sentry send a message (and flush) // and trace the error hook(info); std::process::exit(1); })); tokio::runtime::Builder::new_current_thread() // We use spawn_blocking for database operations, so require approximately // as many blocking threads as we will open database connections. .max_blocking_threads(Persistence::MAX_CONNECTIONS as usize) .enable_all() .build() .unwrap() .block_on(async_main()) } async fn async_main() -> anyhow::Result<()> { let launch_ts = Box::leak(Box::new(LaunchTimestamp::generate())); preinitialize_metrics(); let args = Cli::parse(); tracing::info!( "version: {}, launch_timestamp: {}, build_tag {}", GIT_VERSION, launch_ts.to_string(), BUILD_TAG, ); let build_info = BuildInfo { revision: GIT_VERSION, build_tag: BUILD_TAG, }; let strict_mode = if args.dev { StrictMode::Dev } else { StrictMode::Strict }; let secrets = Secrets::load(&args).await?; // Validate required secrets and arguments are provided in strict mode match strict_mode { StrictMode::Strict if (secrets.public_key.is_none() || secrets.pageserver_jwt_token.is_none() || secrets.control_plane_jwt_token.is_none() || secrets.safekeeper_jwt_token.is_none()) => { // Production systems should always have secrets configured: if public_key was not set // then we would implicitly disable auth. anyhow::bail!( "Insecure config! One or more secrets is not set. This is only permitted in `--dev` mode" ); } StrictMode::Strict if args.control_plane_url.is_none() => { // Production systems should always have a control plane URL set, to prevent falling // back to trying to use neon_local. anyhow::bail!( "`--control-plane-url` is not set: this is only permitted in `--dev` mode" ); } StrictMode::Strict if args.use_local_compute_notifications => { anyhow::bail!("`--use-local-compute-notifications` is only permitted in `--dev` mode"); } StrictMode::Strict if args.timeline_safekeeper_count < 3 => { anyhow::bail!( "Running with less than 3 safekeepers per timeline is only permitted in `--dev` mode" ); } StrictMode::Strict => { tracing::info!("Starting in strict mode: configuration is OK.") } StrictMode::Dev => { tracing::warn!("Starting in dev mode: this may be an insecure configuration.") } } let ssl_ca_certs = match args.ssl_ca_file.as_ref() { Some(ssl_ca_file) => { tracing::info!("Using ssl root CA file: {ssl_ca_file:?}"); let buf = tokio::fs::read(ssl_ca_file).await?; Certificate::from_pem_bundle(&buf)? } None => Vec::new(), }; let posthog_config = if let Ok(json) = std::env::var(POSTHOG_CONFIG_ENV) { let res: Result<PostHogConfig, _> = serde_json::from_str(&json); if let Ok(config) = res { Some(config) } else { tracing::warn!("Invalid posthog config: {json}"); None } } else { None }; let config = Config { pageserver_jwt_token: secrets.pageserver_jwt_token, safekeeper_jwt_token: secrets.safekeeper_jwt_token, control_plane_jwt_token: secrets.control_plane_jwt_token, peer_jwt_token: secrets.peer_jwt_token, control_plane_url: args.control_plane_url, max_offline_interval: args .max_offline_interval .map(humantime::Duration::into) .unwrap_or(MAX_OFFLINE_INTERVAL_DEFAULT), max_warming_up_interval: args .max_warming_up_interval .map(humantime::Duration::into) .unwrap_or(MAX_WARMING_UP_INTERVAL_DEFAULT), reconciler_concurrency: args .reconciler_concurrency .unwrap_or(RECONCILER_CONCURRENCY_DEFAULT), priority_reconciler_concurrency: args .priority_reconciler_concurrency .unwrap_or(PRIORITY_RECONCILER_CONCURRENCY_DEFAULT), safekeeper_reconciler_concurrency: args .safekeeper_reconciler_concurrency .unwrap_or(SAFEKEEPER_RECONCILER_CONCURRENCY_DEFAULT), tenant_rate_limit: args.tenant_rate_limit, split_threshold: args.split_threshold, max_split_shards: args.max_split_shards, initial_split_threshold: args.initial_split_threshold, initial_split_shards: args.initial_split_shards, neon_local_repo_dir: args.neon_local_repo_dir, max_secondary_lag_bytes: args.max_secondary_lag_bytes, heartbeat_interval: args .heartbeat_interval .map(humantime::Duration::into) .unwrap_or(HEARTBEAT_INTERVAL_DEFAULT), long_reconcile_threshold: args .long_reconcile_threshold .map(humantime::Duration::into) .unwrap_or(LONG_RECONCILE_THRESHOLD_DEFAULT), address_for_peers: args.address_for_peers, start_as_candidate: args.start_as_candidate, use_https_pageserver_api: args.use_https_pageserver_api, use_https_safekeeper_api: args.use_https_safekeeper_api, ssl_ca_certs, timelines_onto_safekeepers: args.timelines_onto_safekeepers, use_local_compute_notifications: args.use_local_compute_notifications, timeline_safekeeper_count: args.timeline_safekeeper_count, posthog_config: posthog_config.clone(), kick_secondary_downloads: args.kick_secondary_downloads, shard_split_request_timeout: args .shard_split_request_timeout .map(humantime::Duration::into) .unwrap_or(Duration::MAX), handle_ps_local_disk_loss: args.handle_ps_local_disk_loss, }; // Validate that we can connect to the database Persistence::await_connection(&secrets.database_url, args.db_connect_timeout.into()).await?; let persistence = Arc::new(Persistence::new(secrets.database_url).await); let service = Service::spawn(config, persistence.clone()).await?; let auth = secrets .public_key .map(|jwt_auth| Arc::new(SwappableJwtAuth::new(jwt_auth))); let router = make_router(service.clone(), auth, build_info) .build() .map_err(|err| anyhow!(err))?; let http_service = Arc::new(http_utils::RequestServiceBuilder::new(router).map_err(|err| anyhow!(err))?); let api_shutdown = CancellationToken::new(); // Start HTTP server let http_server_task: OptionFuture<_> = match args.listen { Some(http_addr) => { let http_listener = tcp_listener::bind(http_addr)?; let http_server = http_utils::server::Server::new(Arc::clone(&http_service), http_listener, None)?; tracing::info!("Serving HTTP on {}", http_addr); Some(tokio::task::spawn(http_server.serve(api_shutdown.clone()))) } None => None, } .into(); // Start HTTPS server let https_server_task: OptionFuture<_> = match args.listen_https { Some(https_addr) => { let https_listener = tcp_listener::bind(https_addr)?; let resolver = ReloadingCertificateResolver::new( "main", &args.ssl_key_file, &args.ssl_cert_file, *args.ssl_cert_reload_period, ) .await?; let server_config = rustls::ServerConfig::builder() .with_no_client_auth() .with_cert_resolver(resolver); let tls_acceptor = tokio_rustls::TlsAcceptor::from(Arc::new(server_config)); let https_server = http_utils::server::Server::new(http_service, https_listener, Some(tls_acceptor))?; tracing::info!("Serving HTTPS on {}", https_addr); Some(tokio::task::spawn(https_server.serve(api_shutdown.clone()))) } None => None, } .into(); let chaos_task = args.chaos_interval.map(|interval| { let service = service.clone(); let cancel = CancellationToken::new(); let cancel_bg = cancel.clone(); let chaos_exit_crontab = args.chaos_exit_crontab; ( tokio::task::spawn( async move { let mut chaos_injector = ChaosInjector::new(service, interval.into(), chaos_exit_crontab); chaos_injector.run(cancel_bg).await } .instrument(tracing::info_span!("chaos_injector")), ), cancel, ) }); let feature_flag_task = if let Some(posthog_config) = posthog_config { let service = service.clone(); let cancel = CancellationToken::new(); let cancel_bg = cancel.clone(); let task = tokio::task::spawn( async move { match FeatureFlagService::new(service, posthog_config) { Ok(feature_flag_service) => { let feature_flag_service = Arc::new(feature_flag_service); feature_flag_service.run(cancel_bg).await } Err(e) => { tracing::warn!("Failed to create feature flag service: {}", e); } }; } .instrument(tracing::info_span!("feature_flag_service")), ); Some((task, cancel)) } else { None }; // Wait until we receive a signal let mut sigint = tokio::signal::unix::signal(SignalKind::interrupt())?; let mut sigquit = tokio::signal::unix::signal(SignalKind::quit())?; let mut sigterm = tokio::signal::unix::signal(SignalKind::terminate())?; tokio::pin!(http_server_task, https_server_task); tokio::select! { _ = sigint.recv() => {}, _ = sigterm.recv() => {}, _ = sigquit.recv() => {}, Some(err) = &mut http_server_task => { panic!("HTTP server task failed: {err:#?}"); } Some(err) = &mut https_server_task => { panic!("HTTPS server task failed: {err:#?}"); } } tracing::info!("Terminating on signal"); // Stop HTTP and HTTPS servers first, so that we don't have to service requests // while shutting down Service. api_shutdown.cancel(); // If the deadline is exceeded, we will fall through and shut down the service anyway, // any request handlers in flight will experience cancellation & their clients will // see a torn connection. let deadline = tokio::time::Instant::now() + Duration::from_secs(5); match tokio::time::timeout_at(deadline, http_server_task).await { Ok(Some(Ok(_))) => tracing::info!("Joined HTTP server task"), Ok(Some(Err(e))) => tracing::error!("Error joining HTTP server task: {e}"), Ok(None) => {} // HTTP is disabled. Err(_) => tracing::warn!("Timed out joining HTTP server task"), } match tokio::time::timeout_at(deadline, https_server_task).await { Ok(Some(Ok(_))) => tracing::info!("Joined HTTPS server task"), Ok(Some(Err(e))) => tracing::error!("Error joining HTTPS server task: {e}"), Ok(None) => {} // HTTPS is disabled. Err(_) => tracing::warn!("Timed out joining HTTPS server task"), } // If we were injecting chaos, stop that so that we're not calling into Service while it shuts down if let Some((chaos_jh, chaos_cancel)) = chaos_task { chaos_cancel.cancel(); chaos_jh.await.ok(); } // If we were running the feature flag service, stop that so that we're not calling into Service while it shuts down if let Some((feature_flag_task, feature_flag_cancel)) = feature_flag_task { feature_flag_cancel.cancel(); feature_flag_task.await.ok(); } service.shutdown().await; tracing::info!("Service shutdown complete"); std::process::exit(0); }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/metrics.rs
storage_controller/src/metrics.rs
//! //! This module provides metric definitions for the storage controller. //! //! All metrics are grouped in [`StorageControllerMetricGroup`]. [`StorageControllerMetrics`] holds //! the mentioned metrics and their encoder. It's globally available via the [`METRICS_REGISTRY`] //! constant. //! //! The rest of the code defines label group types and deals with converting outer types to labels. //! use std::sync::Mutex; use bytes::Bytes; use measured::label::LabelValue; use measured::metric::histogram; use measured::{FixedCardinalityLabel, MetricGroup}; use metrics::NeonMetrics; use once_cell::sync::Lazy; use strum::IntoEnumIterator; use crate::persistence::{DatabaseError, DatabaseOperation}; use crate::service::LeadershipStatus; pub(crate) static METRICS_REGISTRY: Lazy<StorageControllerMetrics> = Lazy::new(StorageControllerMetrics::default); pub fn preinitialize_metrics() { Lazy::force(&METRICS_REGISTRY); } pub(crate) struct StorageControllerMetrics { pub(crate) metrics_group: StorageControllerMetricGroup, encoder: Mutex<measured::text::BufferedTextEncoder>, } #[derive(measured::MetricGroup)] #[metric(new())] pub(crate) struct StorageControllerMetricGroup { /// Count of how many times we spawn a reconcile task pub(crate) storage_controller_reconcile_spawn: measured::Counter, /// Size of the in-memory map of tenant shards pub(crate) storage_controller_tenant_shards: measured::Gauge, /// Size of the in-memory map of pageserver_nodes pub(crate) storage_controller_pageserver_nodes: measured::Gauge, /// Count of how many pageserver nodes from in-memory map have https configured pub(crate) storage_controller_https_pageserver_nodes: measured::Gauge, /// Size of the in-memory map of safekeeper_nodes pub(crate) storage_controller_safekeeper_nodes: measured::Gauge, /// Count of how many safekeeper nodes from in-memory map have https configured pub(crate) storage_controller_https_safekeeper_nodes: measured::Gauge, /// Reconciler tasks completed, broken down by success/failure/cancelled pub(crate) storage_controller_reconcile_complete: measured::CounterVec<ReconcileCompleteLabelGroupSet>, /// Count of how many times we make an optimization change to a tenant's scheduling pub(crate) storage_controller_schedule_optimization: measured::Counter, /// How many shards are not scheduled into their preferred AZ pub(crate) storage_controller_schedule_az_violation: measured::Gauge, /// How many shard locations (secondary or attached) on each node pub(crate) storage_controller_node_shards: measured::GaugeVec<NodeLabelGroupSet>, /// How many _attached_ shard locations on each node pub(crate) storage_controller_node_attached_shards: measured::GaugeVec<NodeLabelGroupSet>, /// How many _home_ shard locations on each node (i.e. the node's AZ matches the shard's /// preferred AZ) pub(crate) storage_controller_node_home_shards: measured::GaugeVec<NodeLabelGroupSet>, /// How many shards would like to reconcile but were blocked by concurrency limits pub(crate) storage_controller_pending_reconciles: measured::Gauge, /// How many shards are stuck and will be ignored when considering to run optimizations pub(crate) storage_controller_stuck_reconciles: measured::Gauge, /// HTTP request status counters for handled requests pub(crate) storage_controller_http_request_status: measured::CounterVec<HttpRequestStatusLabelGroupSet>, /// HTTP request handler latency across all status codes #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 2.0))] pub(crate) storage_controller_http_request_latency: measured::HistogramVec<HttpRequestLatencyLabelGroupSet, 5>, /// HTTP rate limiting latency across all tenants and endpoints #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 10.0))] pub(crate) storage_controller_http_request_rate_limited: measured::Histogram<10>, /// Count of HTTP requests to the pageserver that resulted in an error, /// broken down by the pageserver node id, request name and method pub(crate) storage_controller_pageserver_request_error: measured::CounterVec<PageserverRequestLabelGroupSet>, /// Count of HTTP requests to the safekeeper that resulted in an error, /// broken down by the safekeeper node id, request name and method pub(crate) storage_controller_safekeeper_request_error: measured::CounterVec<SafekeeperRequestLabelGroupSet>, /// Latency of HTTP requests to the pageserver, broken down by pageserver /// node id, request name and method. This include both successful and unsuccessful /// requests. #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 2.0))] pub(crate) storage_controller_pageserver_request_latency: measured::HistogramVec<PageserverRequestLabelGroupSet, 5>, /// Latency of HTTP requests to the safekeeper, broken down by safekeeper /// node id, request name and method. This include both successful and unsuccessful /// requests. #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 2.0))] pub(crate) storage_controller_safekeeper_request_latency: measured::HistogramVec<SafekeeperRequestLabelGroupSet, 5>, /// Count of pass-through HTTP requests to the pageserver that resulted in an error, /// broken down by the pageserver node id, request name and method pub(crate) storage_controller_passthrough_request_error: measured::CounterVec<PageserverRequestLabelGroupSet>, /// Latency of pass-through HTTP requests to the pageserver, broken down by pageserver /// node id, request name and method. This include both successful and unsuccessful /// requests. #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 2.0))] pub(crate) storage_controller_passthrough_request_latency: measured::HistogramVec<PageserverRequestLabelGroupSet, 5>, /// Count of errors in database queries, broken down by error type and operation. pub(crate) storage_controller_database_query_error: measured::CounterVec<DatabaseQueryErrorLabelGroupSet>, /// Latency of database queries, broken down by operation. #[metric(metadata = histogram::Thresholds::exponential_buckets(0.1, 2.0))] pub(crate) storage_controller_database_query_latency: measured::HistogramVec<DatabaseQueryLatencyLabelGroupSet, 5>, pub(crate) storage_controller_leadership_status: measured::GaugeVec<LeadershipStatusGroupSet>, /// Indicator of stucked (long-running) reconciles, broken down by tenant, shard and sequence. /// The metric is automatically removed once the reconciliation completes. pub(crate) storage_controller_reconcile_long_running: measured::CounterVec<ReconcileLongRunningLabelGroupSet>, /// Indicator of safekeeper reconciler queue depth, broken down by safekeeper, excluding ongoing reconciles. pub(crate) storage_controller_safekeeper_reconciles_queued: measured::GaugeVec<SafekeeperReconcilerLabelGroupSet>, /// Indicator of completed safekeeper reconciles, broken down by safekeeper. pub(crate) storage_controller_safekeeper_reconciles_complete: measured::CounterVec<SafekeeperReconcilerLabelGroupSet>, /* BEGIN HADRON */ /// Hadron `config_watcher` reconciliation runs completed, broken down by success/failure. pub(crate) storage_controller_config_watcher_complete: measured::CounterVec<ConfigWatcherCompleteLabelGroupSet>, /// Hadron long waits for node state changes during drain and fill. pub(crate) storage_controller_drain_and_fill_long_waits: measured::Counter, /// Set to 1 if we detect any page server pods with pending node pool rotation annotations. /// Requires manual reset after oncall investigation. pub(crate) storage_controller_ps_node_pool_rotation_pending: measured::Gauge, /// Hadron storage scrubber status. pub(crate) storage_controller_storage_scrub_status: measured::CounterVec<StorageScrubberLabelGroupSet>, /// Desired number of pageservers managed by the storage controller pub(crate) storage_controller_num_pageservers_desired: measured::Gauge, /// Desired number of safekeepers managed by the storage controller pub(crate) storage_controller_num_safekeeper_desired: measured::Gauge, /* END HADRON */ } impl StorageControllerMetrics { pub(crate) fn encode(&self, neon_metrics: &NeonMetrics) -> Bytes { let mut encoder = self.encoder.lock().unwrap(); neon_metrics .collect_group_into(&mut *encoder) .unwrap_or_else(|infallible| match infallible {}); self.metrics_group .collect_group_into(&mut *encoder) .unwrap_or_else(|infallible| match infallible {}); encoder.finish() } } impl Default for StorageControllerMetrics { fn default() -> Self { let mut metrics_group = StorageControllerMetricGroup::new(); metrics_group .storage_controller_reconcile_complete .init_all_dense(); metrics_group .storage_controller_config_watcher_complete .init_all_dense(); Self { metrics_group, encoder: Mutex::new(measured::text::BufferedTextEncoder::new()), } } } #[derive(measured::LabelGroup, Clone)] #[label(set = NodeLabelGroupSet)] pub(crate) struct NodeLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) az: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) node_id: &'a str, } #[derive(measured::LabelGroup)] #[label(set = ReconcileCompleteLabelGroupSet)] pub(crate) struct ReconcileCompleteLabelGroup { pub(crate) status: ReconcileOutcome, } #[derive(measured::LabelGroup)] #[label(set = HttpRequestStatusLabelGroupSet)] pub(crate) struct HttpRequestStatusLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) path: &'a str, pub(crate) method: Method, pub(crate) status: StatusCode, } #[derive(measured::LabelGroup)] #[label(set = HttpRequestLatencyLabelGroupSet)] pub(crate) struct HttpRequestLatencyLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) path: &'a str, pub(crate) method: Method, } #[derive(measured::LabelGroup, Clone)] #[label(set = PageserverRequestLabelGroupSet)] pub(crate) struct PageserverRequestLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) pageserver_id: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) path: &'a str, pub(crate) method: Method, } #[derive(measured::LabelGroup, Clone)] #[label(set = SafekeeperRequestLabelGroupSet)] pub(crate) struct SafekeeperRequestLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) safekeeper_id: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) path: &'a str, pub(crate) method: Method, } #[derive(measured::LabelGroup)] #[label(set = DatabaseQueryErrorLabelGroupSet)] pub(crate) struct DatabaseQueryErrorLabelGroup { pub(crate) error_type: DatabaseErrorLabel, pub(crate) operation: DatabaseOperation, } #[derive(measured::LabelGroup)] #[label(set = DatabaseQueryLatencyLabelGroupSet)] pub(crate) struct DatabaseQueryLatencyLabelGroup { pub(crate) operation: DatabaseOperation, } #[derive(measured::LabelGroup)] #[label(set = LeadershipStatusGroupSet)] pub(crate) struct LeadershipStatusGroup { pub(crate) status: LeadershipStatus, } #[derive(measured::LabelGroup, Clone)] #[label(set = ReconcileLongRunningLabelGroupSet)] pub(crate) struct ReconcileLongRunningLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) tenant_id: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) shard_number: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) sequence: &'a str, } #[derive(measured::LabelGroup, Clone)] #[label(set = StorageScrubberLabelGroupSet)] pub(crate) struct StorageScrubberLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) tenant_id: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) shard_number: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) timeline_id: &'a str, pub(crate) outcome: StorageScrubberOutcome, } #[derive(FixedCardinalityLabel, Clone, Copy)] pub(crate) enum StorageScrubberOutcome { PSOk, PSWarning, PSError, PSOrphan, SKOk, SKError, } #[derive(measured::LabelGroup)] #[label(set = ConfigWatcherCompleteLabelGroupSet)] pub(crate) struct ConfigWatcherCompleteLabelGroup { // Reuse the ReconcileOutcome from the SC's reconciliation metrics. pub(crate) status: ReconcileOutcome, } #[derive(FixedCardinalityLabel, Clone, Copy)] pub(crate) enum ReconcileOutcome { // Successfully reconciled everything. #[label(rename = "ok")] Success, // Used by tenant-shard reconciler only. Reconciled pageserver state successfully, // but failed to delivery the compute notificiation. This error is typically transient // but if its occurance keeps increasing, it should be investigated. #[label(rename = "ok_no_notify")] SuccessNoNotify, // We failed to reconcile some state and the reconcilation will be retried. Error, // Reconciliation was cancelled. Cancel, } #[derive(FixedCardinalityLabel, Copy, Clone)] pub(crate) enum Method { Get, Put, Post, Delete, Other, } #[derive(measured::LabelGroup, Clone)] #[label(set = SafekeeperReconcilerLabelGroupSet)] pub(crate) struct SafekeeperReconcilerLabelGroup<'a> { #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) sk_az: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) sk_node_id: &'a str, #[label(dynamic_with = lasso::ThreadedRodeo, default)] pub(crate) sk_hostname: &'a str, } impl From<hyper::Method> for Method { fn from(value: hyper::Method) -> Self { if value == hyper::Method::GET { Method::Get } else if value == hyper::Method::PUT { Method::Put } else if value == hyper::Method::POST { Method::Post } else if value == hyper::Method::DELETE { Method::Delete } else { Method::Other } } } #[derive(Clone, Copy)] pub(crate) struct StatusCode(pub(crate) hyper::http::StatusCode); impl LabelValue for StatusCode { fn visit<V: measured::label::LabelVisitor>(&self, v: V) -> V::Output { v.write_int(self.0.as_u16() as i64) } } impl FixedCardinalityLabel for StatusCode { fn cardinality() -> usize { (100..1000).len() } fn encode(&self) -> usize { self.0.as_u16() as usize } fn decode(value: usize) -> Self { Self(hyper::http::StatusCode::from_u16(u16::try_from(value).unwrap()).unwrap()) } } #[derive(FixedCardinalityLabel, Clone, Copy)] pub(crate) enum DatabaseErrorLabel { Query, Connection, ConnectionPool, Logical, Migration, Cas, } impl DatabaseError { pub(crate) fn error_label(&self) -> DatabaseErrorLabel { match self { Self::Query(_) => DatabaseErrorLabel::Query, Self::Connection(_) => DatabaseErrorLabel::Connection, Self::ConnectionPool(_) => DatabaseErrorLabel::ConnectionPool, Self::Logical(_) => DatabaseErrorLabel::Logical, Self::Migration(_) => DatabaseErrorLabel::Migration, Self::Cas(_) => DatabaseErrorLabel::Cas, } } } /// Update the leadership status metric gauges to reflect the requested status pub(crate) fn update_leadership_status(status: LeadershipStatus) { let status_metric = &METRICS_REGISTRY .metrics_group .storage_controller_leadership_status; for s in LeadershipStatus::iter() { if s == status { status_metric.set(LeadershipStatusGroup { status: s }, 1); } else { status_metric.set(LeadershipStatusGroup { status: s }, 0); } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/peer_client.rs
storage_controller/src/peer_client.rs
use std::collections::HashMap; use std::error::Error as _; use std::time::Duration; use http_utils::error::HttpErrorBody; use hyper::Uri; use pageserver_api::shard::TenantShardId; use reqwest::{StatusCode, Url}; use serde::{Deserialize, Serialize}; use tokio_util::sync::CancellationToken; use utils::backoff; use crate::tenant_shard::ObservedState; #[derive(Debug, Clone)] pub(crate) struct PeerClient { uri: Uri, jwt: Option<String>, client: reqwest::Client, } #[derive(thiserror::Error, Debug)] pub(crate) enum StorageControllerPeerError { #[error( "failed to deserialize error response with status code {0} at {1}: {2}{}", .2.source().map(|e| format!(": {e}")).unwrap_or_default() )] DeserializationError(StatusCode, Url, reqwest::Error), #[error("storage controller peer API error ({0}): {1}")] ApiError(StatusCode, String), #[error("failed to send HTTP request: {0}{}", .0.source().map(|e| format!(": {e}")).unwrap_or_default())] SendError(reqwest::Error), #[error("Cancelled")] Cancelled, } pub(crate) type Result<T> = std::result::Result<T, StorageControllerPeerError>; pub(crate) trait ResponseErrorMessageExt: Sized { fn error_from_body(self) -> impl std::future::Future<Output = Result<Self>> + Send; } impl ResponseErrorMessageExt for reqwest::Response { async fn error_from_body(self) -> Result<Self> { let status = self.status(); if !(status.is_client_error() || status.is_server_error()) { return Ok(self); } let url = self.url().to_owned(); Err(match self.json::<HttpErrorBody>().await { Ok(HttpErrorBody { msg }) => StorageControllerPeerError::ApiError(status, msg), Err(err) => StorageControllerPeerError::DeserializationError(status, url, err), }) } } #[derive(Serialize, Deserialize, Debug, Default, Clone)] pub(crate) struct GlobalObservedState(pub(crate) HashMap<TenantShardId, ObservedState>); const STEP_DOWN_RETRIES: u32 = 8; const STEP_DOWN_TIMEOUT: Duration = Duration::from_secs(1); impl PeerClient { pub(crate) fn new(http_client: reqwest::Client, uri: Uri, jwt: Option<String>) -> Self { Self { uri, jwt, client: http_client, } } async fn request_step_down(&self) -> Result<GlobalObservedState> { let step_down_path = format!("{}control/v1/step_down", self.uri); let req = self.client.put(step_down_path); let req = if let Some(jwt) = &self.jwt { req.header(reqwest::header::AUTHORIZATION, format!("Bearer {jwt}")) } else { req }; let req = req.timeout(STEP_DOWN_TIMEOUT); let res = req .send() .await .map_err(StorageControllerPeerError::SendError)?; let response = res.error_from_body().await?; let status = response.status(); let url = response.url().to_owned(); response .json() .await .map_err(|err| StorageControllerPeerError::DeserializationError(status, url, err)) } /// Request the peer to step down and return its current observed state /// All errors are re-tried pub(crate) async fn step_down( &self, cancel: &CancellationToken, ) -> Result<GlobalObservedState> { backoff::retry( || self.request_step_down(), |_e| false, 2, STEP_DOWN_RETRIES, "Send step down request", cancel, ) .await .ok_or_else(|| StorageControllerPeerError::Cancelled) .and_then(|x| x) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/operation_utils.rs
storage_controller/src/operation_utils.rs
use std::collections::{BTreeMap, HashMap}; use std::sync::Arc; use pageserver_api::controller_api::{NodeSchedulingPolicy, ShardSchedulingPolicy}; use utils::id::NodeId; use utils::shard::TenantShardId; use crate::background_node_operations::OperationError; use crate::node::Node; use crate::scheduler::Scheduler; use crate::tenant_shard::TenantShard; /// Check that the state of the node being drained is as expected: /// node is present in memory and scheduling policy is set to expected_policy pub(crate) fn validate_node_state( node_id: &NodeId, nodes: Arc<HashMap<NodeId, Node>>, expected_policy: NodeSchedulingPolicy, ) -> Result<(), OperationError> { let node = nodes.get(node_id).ok_or(OperationError::NodeStateChanged( format!("node {node_id} was removed").into(), ))?; let current_policy = node.get_scheduling(); if current_policy != expected_policy { // TODO(vlad): maybe cancel pending reconciles before erroring out. need to think // about it return Err(OperationError::NodeStateChanged( format!("node {node_id} changed state to {current_policy:?}").into(), )); } Ok(()) } /// Struct that houses a few utility methods for draining pageserver nodes pub(crate) struct TenantShardDrain { pub(crate) drained_node: NodeId, pub(crate) tenant_shard_id: TenantShardId, } impl TenantShardDrain { /// Check if the tenant shard under question is eligible for drainining: /// it's primary attachment is on the node being drained pub(crate) fn tenant_shard_eligible_for_drain( &self, tenants: &BTreeMap<TenantShardId, TenantShard>, scheduler: &Scheduler, ) -> TenantShardDrainAction { let Some(tenant_shard) = tenants.get(&self.tenant_shard_id) else { return TenantShardDrainAction::Skip; }; if *tenant_shard.intent.get_attached() != Some(self.drained_node) { // If the intent attached node is not the drained node, check the observed state // of the shard on the drained node. If it is Attached*, it means the shard is // beeing migrated from the drained node. The drain loop needs to wait for the // reconciliation to complete for a smooth draining. use pageserver_api::models::LocationConfigMode::*; let attach_mode = tenant_shard .observed .locations .get(&self.drained_node) .and_then(|observed| observed.conf.as_ref().map(|conf| conf.mode)); return match (attach_mode, tenant_shard.intent.get_attached()) { (Some(AttachedSingle | AttachedMulti | AttachedStale), Some(intent_node_id)) => { TenantShardDrainAction::Reconcile(*intent_node_id) } _ => TenantShardDrainAction::Skip, }; } // Only tenants with a normal (Active) scheduling policy are proactively moved // around during a node drain. Shards which have been manually configured to a different // policy are only rescheduled by manual intervention. match tenant_shard.get_scheduling_policy() { ShardSchedulingPolicy::Active | ShardSchedulingPolicy::Essential => { // A migration during drain is classed as 'essential' because it is required to // uphold our availability goals for the tenant: this shard is elegible for migration. } ShardSchedulingPolicy::Pause | ShardSchedulingPolicy::Stop => { // If we have been asked to avoid rescheduling this shard, then do not migrate it during a drain return TenantShardDrainAction::Skip; } } match tenant_shard.preferred_secondary(scheduler) { Some(node) => TenantShardDrainAction::RescheduleToSecondary(node), None => { tracing::warn!( tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), "No eligible secondary while draining {}", self.drained_node ); TenantShardDrainAction::Skip } } } /// Attempt to reschedule the tenant shard under question to one of its secondary locations /// Returns an Err when the operation should be aborted and Ok(None) when the tenant shard /// should be skipped. pub(crate) fn reschedule_to_secondary<'a>( &self, destination: NodeId, tenants: &'a mut BTreeMap<TenantShardId, TenantShard>, scheduler: &mut Scheduler, nodes: &Arc<HashMap<NodeId, Node>>, ) -> Result<Option<&'a mut TenantShard>, OperationError> { let tenant_shard = match tenants.get_mut(&self.tenant_shard_id) { Some(some) => some, None => { // Tenant shard was removed in the meantime. // Skip to the next one, but don't fail the overall operation return Ok(None); } }; if !nodes.contains_key(&destination) { return Err(OperationError::NodeStateChanged( format!("node {destination} was removed").into(), )); } if !tenant_shard.intent.get_secondary().contains(&destination) { tracing::info!( tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), "Secondary moved away from {destination} during drain" ); return Ok(None); } match tenant_shard.reschedule_to_secondary(Some(destination), scheduler) { Err(e) => { tracing::warn!( tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), "Scheduling error when draining pageserver {} : {}", self.drained_node, e ); Ok(None) } Ok(()) => { let scheduled_to = tenant_shard.intent.get_attached(); tracing::info!( tenant_id=%self.tenant_shard_id.tenant_id, shard_id=%self.tenant_shard_id.shard_slug(), "Rescheduled shard while draining node {}: {} -> {:?}", self.drained_node, self.drained_node, scheduled_to ); Ok(Some(tenant_shard)) } } } } /// Action to take when draining a tenant shard. pub(crate) enum TenantShardDrainAction { /// The tenant shard is on the draining node. /// Reschedule the tenant shard to a secondary location. /// Holds a destination node id to reschedule to. RescheduleToSecondary(NodeId), /// The tenant shard is beeing migrated from the draining node. /// Wait for the reconciliation to complete. /// Holds the intent attached node id. Reconcile(NodeId), /// The tenant shard is not eligible for drainining, skip it. Skip, }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/scheduler.rs
storage_controller/src/scheduler.rs
use std::collections::HashMap; use std::fmt::Debug; use http_utils::error::ApiError; use itertools::Itertools; use pageserver_api::controller_api::AvailabilityZone; use pageserver_api::models::PageserverUtilization; use serde::Serialize; use utils::id::NodeId; use crate::metrics::NodeLabelGroup; use crate::node::Node; use crate::tenant_shard::TenantShard; /// Scenarios in which we cannot find a suitable location for a tenant shard #[derive(thiserror::Error, Debug)] pub enum ScheduleError { #[error("No pageservers found")] NoPageservers, #[error("No pageserver found matching constraint")] ImpossibleConstraint, } impl From<ScheduleError> for ApiError { fn from(value: ScheduleError) -> Self { ApiError::Conflict(format!("Scheduling error: {value}")) } } #[derive(Serialize)] pub enum MaySchedule { Yes(PageserverUtilization), No, } #[derive(Serialize)] pub(crate) struct SchedulerNode { /// How many shards are currently scheduled on this node, via their [`crate::tenant_shard::IntentState`]. shard_count: usize, /// How many shards are currently attached on this node, via their [`crate::tenant_shard::IntentState`]. attached_shard_count: usize, /// How many shards have a location on this node (via [`crate::tenant_shard::IntentState`]) _and_ this node /// is in their preferred AZ (i.e. this is their 'home' location) home_shard_count: usize, /// Availability zone id in which the node resides az: AvailabilityZone, /// Whether this node is currently elegible to have new shards scheduled (this is derived /// from a node's availability state and scheduling policy). may_schedule: MaySchedule, } pub(crate) trait NodeSchedulingScore: Debug + Ord + Copy + Sized { fn generate( node_id: &NodeId, node: &mut SchedulerNode, preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Option<Self>; /// Return a score that drops any components based on node utilization: this is useful /// for finding scores for scheduling optimisation, when we want to avoid rescheduling /// shards due to e.g. disk usage, to avoid flapping. fn for_optimization(&self) -> Self; fn is_overloaded(&self) -> bool; fn node_id(&self) -> NodeId; } pub(crate) trait ShardTag { type Score: NodeSchedulingScore; } pub(crate) struct AttachedShardTag {} impl ShardTag for AttachedShardTag { type Score = NodeAttachmentSchedulingScore; } pub(crate) struct SecondaryShardTag {} impl ShardTag for SecondaryShardTag { type Score = NodeSecondarySchedulingScore; } #[derive(PartialEq, Eq, Debug, Clone, Copy)] enum AzMatch { Yes, No, Unknown, } impl AzMatch { fn new(node_az: &AvailabilityZone, shard_preferred_az: Option<&AvailabilityZone>) -> Self { match shard_preferred_az { Some(preferred_az) if preferred_az == node_az => Self::Yes, Some(_preferred_az) => Self::No, None => Self::Unknown, } } } #[derive(PartialEq, Eq, Debug, Clone, Copy)] struct AttachmentAzMatch(AzMatch); impl Ord for AttachmentAzMatch { fn cmp(&self, other: &Self) -> std::cmp::Ordering { // Lower scores indicate a more suitable node. // Note that we prefer a node for which we don't have // info to a node which we are certain doesn't match the // preferred AZ of the shard. let az_match_score = |az_match: &AzMatch| match az_match { AzMatch::Yes => 0, AzMatch::Unknown => 1, AzMatch::No => 2, }; az_match_score(&self.0).cmp(&az_match_score(&other.0)) } } impl PartialOrd for AttachmentAzMatch { fn partial_cmp(&self, other: &Self) -> Option<std::cmp::Ordering> { Some(self.cmp(other)) } } #[derive(PartialEq, Eq, Debug, Clone, Copy)] struct SecondaryAzMatch(AzMatch); impl Ord for SecondaryAzMatch { fn cmp(&self, other: &Self) -> std::cmp::Ordering { // Lower scores indicate a more suitable node. // For secondary locations we wish to avoid the preferred AZ // of the shard. let az_match_score = |az_match: &AzMatch| match az_match { AzMatch::No => 0, AzMatch::Unknown => 1, AzMatch::Yes => 2, }; az_match_score(&self.0).cmp(&az_match_score(&other.0)) } } impl PartialOrd for SecondaryAzMatch { fn partial_cmp(&self, other: &Self) -> Option<std::cmp::Ordering> { Some(self.cmp(other)) } } /// Scheduling score of a given node for shard attachments. /// Lower scores indicate more suitable nodes. /// Ordering is given by member declaration order (top to bottom). #[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone, Copy)] pub(crate) struct NodeAttachmentSchedulingScore { /// Flag indicating whether this node matches the preferred AZ /// of the shard. For equal affinity scores, nodes in the matching AZ /// are considered first. az_match: AttachmentAzMatch, /// The number of shards belonging to the tenant currently being /// scheduled that are attached to this node. affinity_score: AffinityScore, /// Utilisation score that combines shard count and disk utilisation utilization_score: u64, /// Total number of shards attached to this node. When nodes have identical utilisation, this /// acts as an anti-affinity between attached shards. total_attached_shard_count: usize, /// Convenience to make selection deterministic in tests and empty systems node_id: NodeId, } impl NodeSchedulingScore for NodeAttachmentSchedulingScore { fn generate( node_id: &NodeId, node: &mut SchedulerNode, preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Option<Self> { let utilization = match &mut node.may_schedule { MaySchedule::Yes(u) => u, MaySchedule::No => { return None; } }; Some(Self { affinity_score: context .nodes .get(node_id) .copied() .unwrap_or(AffinityScore::FREE), az_match: AttachmentAzMatch(AzMatch::new(&node.az, preferred_az.as_ref())), utilization_score: utilization.cached_score(), total_attached_shard_count: node.attached_shard_count, node_id: *node_id, }) } /// For use in scheduling optimisation, where we only want to consider the aspects /// of the score that can only be resolved by moving things (such as inter-shard affinity /// and AZ affinity), and ignore aspects that reflect the total utilization of a node (which /// can fluctuate for other reasons) fn for_optimization(&self) -> Self { Self { utilization_score: 0, total_attached_shard_count: 0, node_id: NodeId(0), ..*self } } fn is_overloaded(&self) -> bool { PageserverUtilization::is_overloaded(self.utilization_score) } fn node_id(&self) -> NodeId { self.node_id } } /// Scheduling score of a given node for shard secondaries. /// Lower scores indicate more suitable nodes. /// Ordering is given by member declaration order (top to bottom). #[derive(Debug, PartialEq, Eq, PartialOrd, Ord, Clone, Copy)] pub(crate) struct NodeSecondarySchedulingScore { /// Flag indicating whether this node matches the preferred AZ /// of the shard. For secondary locations we wish to avoid nodes in. /// the preferred AZ of the shard, since that's where the attached location /// should be scheduled and having the secondary in the same AZ is bad for HA. az_match: SecondaryAzMatch, /// The number of shards belonging to the tenant currently being /// scheduled that are attached to this node. affinity_score: AffinityScore, /// Utilisation score that combines shard count and disk utilisation utilization_score: u64, /// Anti-affinity with other non-home locations: this gives the behavior that secondaries /// will spread out across the nodes in an AZ. total_non_home_shard_count: usize, /// Convenience to make selection deterministic in tests and empty systems node_id: NodeId, } impl NodeSchedulingScore for NodeSecondarySchedulingScore { fn generate( node_id: &NodeId, node: &mut SchedulerNode, preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Option<Self> { let utilization = match &mut node.may_schedule { MaySchedule::Yes(u) => u, MaySchedule::No => { return None; } }; Some(Self { az_match: SecondaryAzMatch(AzMatch::new(&node.az, preferred_az.as_ref())), affinity_score: context .nodes .get(node_id) .copied() .unwrap_or(AffinityScore::FREE), utilization_score: utilization.cached_score(), total_non_home_shard_count: (node.shard_count - node.home_shard_count), node_id: *node_id, }) } fn for_optimization(&self) -> Self { Self { utilization_score: 0, total_non_home_shard_count: 0, node_id: NodeId(0), ..*self } } fn is_overloaded(&self) -> bool { PageserverUtilization::is_overloaded(self.utilization_score) } fn node_id(&self) -> NodeId { self.node_id } } impl PartialEq for SchedulerNode { fn eq(&self, other: &Self) -> bool { let may_schedule_matches = matches!( (&self.may_schedule, &other.may_schedule), (MaySchedule::Yes(_), MaySchedule::Yes(_)) | (MaySchedule::No, MaySchedule::No) ); may_schedule_matches && self.shard_count == other.shard_count && self.attached_shard_count == other.attached_shard_count && self.az == other.az } } impl Eq for SchedulerNode {} /// This type is responsible for selecting which node is used when a tenant shard needs to choose a pageserver /// on which to run. /// /// The type has no persistent state of its own: this is all populated at startup. The Serialize /// impl is only for debug dumps. #[derive(Serialize)] pub(crate) struct Scheduler { nodes: HashMap<NodeId, SchedulerNode>, } /// Score for soft constraint scheduling: lower scores are preferred to higher scores. /// /// For example, we may set an affinity score based on the number of shards from the same /// tenant already on a node, to implicitly prefer to balance out shards. #[derive(Copy, Clone, Debug, Eq, PartialEq, PartialOrd, Ord)] pub(crate) struct AffinityScore(pub(crate) usize); impl AffinityScore { /// If we have no anti-affinity at all toward a node, this is its score. It means /// the scheduler has a free choice amongst nodes with this score, and may pick a node /// based on other information such as total utilization. pub(crate) const FREE: Self = Self(0); pub(crate) fn inc(&mut self) { self.0 += 1; } pub(crate) fn dec(&mut self) { self.0 -= 1; } } impl std::ops::Add for AffinityScore { type Output = Self; fn add(self, rhs: Self) -> Self::Output { Self(self.0 + rhs.0) } } /// Hint for whether this is a sincere attempt to schedule, or a speculative /// check for where we _would_ schedule (done during optimization) #[derive(Debug, Clone)] pub(crate) enum ScheduleMode { Normal, Speculative, } impl Default for ScheduleMode { fn default() -> Self { Self::Normal } } // For carrying state between multiple calls to [`TenantShard::schedule`], e.g. when calling // it for many shards in the same tenant. #[derive(Debug, Default, Clone)] pub(crate) struct ScheduleContext { /// Sparse map of nodes: omitting a node implicitly makes its affinity [`AffinityScore::FREE`] pub(crate) nodes: HashMap<NodeId, AffinityScore>, pub(crate) mode: ScheduleMode, } impl ScheduleContext { pub(crate) fn new(mode: ScheduleMode) -> Self { Self { nodes: HashMap::new(), mode, } } /// Input is a list of nodes we would like to avoid using again within this context. The more /// times a node is passed into this call, the less inclined we are to use it. pub(crate) fn avoid(&mut self, nodes: &[NodeId]) { for node_id in nodes { let entry = self.nodes.entry(*node_id).or_insert(AffinityScore::FREE); entry.inc() } } /// Remove `shard`'s contributions to this context. This is useful when considering scheduling /// this shard afresh, where we don't want it to e.g. experience anti-affinity to its current location. pub(crate) fn project_detach(&self, shard: &TenantShard) -> Self { let mut new_context = self.clone(); if let Some(attached) = shard.intent.get_attached() { if let Some(score) = new_context.nodes.get_mut(attached) { score.dec(); } } for secondary in shard.intent.get_secondary() { if let Some(score) = new_context.nodes.get_mut(secondary) { score.dec(); } } new_context } /// For test, track the sum of AffinityScore values, which is effectively how many /// attached or secondary locations have been registered with this context. #[cfg(test)] pub(crate) fn location_count(&self) -> usize { self.nodes.values().map(|i| i.0).sum() } } pub(crate) enum RefCountUpdate<'a> { PromoteSecondary, Attach, Detach, DemoteAttached, AddSecondary, RemoveSecondary, ChangePreferredAzFrom(Option<&'a AvailabilityZone>), } impl Scheduler { pub(crate) fn new<'a>(nodes: impl Iterator<Item = &'a Node>) -> Self { let mut scheduler_nodes = HashMap::new(); for node in nodes { scheduler_nodes.insert( node.get_id(), SchedulerNode { shard_count: 0, attached_shard_count: 0, home_shard_count: 0, may_schedule: node.may_schedule(), az: node.get_availability_zone_id().clone(), }, ); } Self { nodes: scheduler_nodes, } } /// For debug/support: check that our internal statistics are in sync with the state of /// the nodes & tenant shards. /// /// If anything is inconsistent, log details and return an error. pub(crate) fn consistency_check<'a>( &self, nodes: impl Iterator<Item = &'a Node>, shards: impl Iterator<Item = &'a TenantShard>, ) -> anyhow::Result<()> { let mut expect_nodes: HashMap<NodeId, SchedulerNode> = HashMap::new(); for node in nodes { expect_nodes.insert( node.get_id(), SchedulerNode { shard_count: 0, attached_shard_count: 0, home_shard_count: 0, may_schedule: node.may_schedule(), az: node.get_availability_zone_id().clone(), }, ); } for shard in shards { if let Some(node_id) = shard.intent.get_attached() { match expect_nodes.get_mut(node_id) { Some(node) => { node.shard_count += 1; node.attached_shard_count += 1; if Some(&node.az) == shard.preferred_az() { node.home_shard_count += 1; } } None => anyhow::bail!( "Tenant {} references nonexistent node {}", shard.tenant_shard_id, node_id ), } } for node_id in shard.intent.get_secondary() { match expect_nodes.get_mut(node_id) { Some(node) => { node.shard_count += 1; if Some(&node.az) == shard.preferred_az() { node.home_shard_count += 1; } } None => anyhow::bail!( "Tenant {} references nonexistent node {}", shard.tenant_shard_id, node_id ), } } } for (node_id, expect_node) in &expect_nodes { let Some(self_node) = self.nodes.get(node_id) else { anyhow::bail!("Node {node_id} not found in Self") }; if self_node != expect_node { tracing::error!("Inconsistency detected in scheduling state for node {node_id}"); tracing::error!("Expected state: {}", serde_json::to_string(expect_node)?); tracing::error!("Self state: {}", serde_json::to_string(self_node)?); anyhow::bail!("Inconsistent state on {node_id}"); } } if expect_nodes.len() != self.nodes.len() { // We just checked that all the expected nodes are present. If the lengths don't match, // it means that we have nodes in Self that are unexpected. for node_id in self.nodes.keys() { if !expect_nodes.contains_key(node_id) { anyhow::bail!("Node {node_id} found in Self but not in expected nodes"); } } } Ok(()) } /// Update the reference counts of a node. These reference counts are used to guide scheduling /// decisions, not for memory management: they represent the number of tenant shard whose IntentState /// targets this node and the number of tenants shars whose IntentState is attached to this /// node. /// /// It is an error to call this for a node that is not known to the scheduler (i.e. passed into /// [`Self::new`] or [`Self::node_upsert`]) pub(crate) fn update_node_ref_counts( &mut self, node_id: NodeId, preferred_az: Option<&AvailabilityZone>, update: RefCountUpdate, ) { let Some(node) = self.nodes.get_mut(&node_id) else { debug_assert!(false); tracing::error!("Scheduler missing node {node_id}"); return; }; let is_home_az = Some(&node.az) == preferred_az; match update { RefCountUpdate::PromoteSecondary => { node.attached_shard_count += 1; } RefCountUpdate::Attach => { node.shard_count += 1; node.attached_shard_count += 1; if is_home_az { node.home_shard_count += 1; } } RefCountUpdate::Detach => { node.shard_count -= 1; node.attached_shard_count -= 1; if is_home_az { node.home_shard_count -= 1; } } RefCountUpdate::DemoteAttached => { node.attached_shard_count -= 1; } RefCountUpdate::AddSecondary => { node.shard_count += 1; if is_home_az { node.home_shard_count += 1; } } RefCountUpdate::RemoveSecondary => { node.shard_count -= 1; if is_home_az { node.home_shard_count -= 1; } } RefCountUpdate::ChangePreferredAzFrom(old_az) => { if Some(&node.az) == old_az { node.home_shard_count -= 1; } if is_home_az { node.home_shard_count += 1; } } } // Maybe update PageserverUtilization match update { RefCountUpdate::AddSecondary | RefCountUpdate::Attach => { // Referencing the node: if this takes our shard_count above the utilzation structure's // shard count, then artifically bump it: this ensures that the scheduler immediately // recognizes that this node has more work on it, without waiting for the next heartbeat // to update the utilization. if let MaySchedule::Yes(utilization) = &mut node.may_schedule { utilization.adjust_shard_count_max(node.shard_count as u32); } } RefCountUpdate::PromoteSecondary | RefCountUpdate::Detach | RefCountUpdate::RemoveSecondary | RefCountUpdate::DemoteAttached | RefCountUpdate::ChangePreferredAzFrom(_) => { // De-referencing the node: leave the utilization's shard_count at a stale higher // value until some future heartbeat after we have physically removed this shard // from the node: this prevents the scheduler over-optimistically trying to schedule // more work onto the node before earlier detaches are done. } } } // Check if the number of shards attached to a given node is lagging below // the cluster average. If that's the case, the node should be filled. pub(crate) fn compute_fill_requirement(&self, node_id: NodeId) -> usize { let Some(node) = self.nodes.get(&node_id) else { debug_assert!(false); tracing::error!("Scheduler missing node {node_id}"); return 0; }; assert!(!self.nodes.is_empty()); let expected_attached_shards_per_node = self.expected_attached_shard_count(); for (node_id, node) in self.nodes.iter() { tracing::trace!(%node_id, "attached_shard_count={} shard_count={} expected={}", node.attached_shard_count, node.shard_count, expected_attached_shards_per_node); } expected_attached_shards_per_node.saturating_sub(node.attached_shard_count) } pub(crate) fn expected_attached_shard_count(&self) -> usize { let total_attached_shards: usize = self.nodes.values().map(|n| n.attached_shard_count).sum(); assert!(!self.nodes.is_empty()); total_attached_shards / self.nodes.len() } pub(crate) fn nodes_by_attached_shard_count(&self) -> Vec<(NodeId, usize)> { self.nodes .iter() .map(|(node_id, stats)| (*node_id, stats.attached_shard_count)) .sorted_by(|lhs, rhs| Ord::cmp(&lhs.1, &rhs.1).reverse()) .collect() } pub(crate) fn node_upsert(&mut self, node: &Node) { use std::collections::hash_map::Entry::*; match self.nodes.entry(node.get_id()) { Occupied(mut entry) => { // Updates to MaySchedule are how we receive updated PageserverUtilization: adjust these values // to account for any shards scheduled on the controller but not yet visible to the pageserver. let mut may_schedule = node.may_schedule(); match &mut may_schedule { MaySchedule::Yes(utilization) => { utilization.adjust_shard_count_max(entry.get().shard_count as u32); } MaySchedule::No => { // Nothing to tweak } } entry.get_mut().may_schedule = may_schedule; } Vacant(entry) => { entry.insert(SchedulerNode { shard_count: 0, attached_shard_count: 0, home_shard_count: 0, may_schedule: node.may_schedule(), az: node.get_availability_zone_id().clone(), }); } } } pub(crate) fn node_remove(&mut self, node_id: NodeId) { if self.nodes.remove(&node_id).is_none() { tracing::warn!(node_id=%node_id, "Removed non-existent node from scheduler"); } } /// Calculate a single node's score, used in optimizer logic to compare specific /// nodes' scores. pub(crate) fn compute_node_score<Score>( &mut self, node_id: NodeId, preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Option<Score> where Score: NodeSchedulingScore, { self.nodes .get_mut(&node_id) .and_then(|node| Score::generate(&node_id, node, preferred_az, context)) } /// Compute a schedulling score for each node that the scheduler knows of /// minus a set of hard excluded nodes. fn compute_node_scores<Score>( &mut self, hard_exclude: &[NodeId], preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Vec<Score> where Score: NodeSchedulingScore, { self.nodes .iter_mut() .filter_map(|(k, v)| { if hard_exclude.contains(k) { None } else { Score::generate(k, v, preferred_az, context) } }) .collect() } /// hard_exclude: it is forbidden to use nodes in this list, typically becacuse they /// are already in use by this shard -- we use this to avoid picking the same node /// as both attached and secondary location. This is a hard constraint: if we cannot /// find any nodes that aren't in this list, then we will return a [`ScheduleError::ImpossibleConstraint`]. /// /// context: we prefer to avoid using nodes identified in the context, according /// to their anti-affinity score. We use this to prefeer to avoid placing shards in /// the same tenant on the same node. This is a soft constraint: the context will never /// cause us to fail to schedule a shard. pub(crate) fn schedule_shard<Tag: ShardTag>( &mut self, hard_exclude: &[NodeId], preferred_az: &Option<AvailabilityZone>, context: &ScheduleContext, ) -> Result<NodeId, ScheduleError> { if self.nodes.is_empty() { return Err(ScheduleError::NoPageservers); } let mut scores = self.compute_node_scores::<Tag::Score>(hard_exclude, preferred_az, context); // Exclude nodes whose utilization is critically high, if there are alternatives available. This will // cause us to violate affinity rules if it is necessary to avoid critically overloading nodes: for example // we may place shards in the same tenant together on the same pageserver if all other pageservers are // overloaded. let non_overloaded_scores = scores .iter() .filter(|i| !i.is_overloaded()) .copied() .collect::<Vec<_>>(); if !non_overloaded_scores.is_empty() { scores = non_overloaded_scores; } // Sort the nodes by score. The one with the lowest scores will be the preferred node. // Refer to [`NodeAttachmentSchedulingScore`] for attached locations and // [`NodeSecondarySchedulingScore`] for secondary locations to understand how the nodes // are ranked. scores.sort(); if scores.is_empty() { // After applying constraints, no pageservers were left. if !matches!(context.mode, ScheduleMode::Speculative) { // If this was not a speculative attempt, log details to understand why we couldn't // schedule: this may help an engineer understand if some nodes are marked offline // in a way that's preventing progress. tracing::info!( "Scheduling failure, while excluding {hard_exclude:?}, node states:" ); for (node_id, node) in &self.nodes { tracing::info!( "Node {node_id}: may_schedule={} shards={}", !matches!(node.may_schedule, MaySchedule::No), node.shard_count ); } } return Err(ScheduleError::ImpossibleConstraint); } // Lowest score wins let node_id = scores.first().unwrap().node_id(); if !matches!(context.mode, ScheduleMode::Speculative) { tracing::info!( "scheduler selected node {node_id} (elegible nodes {:?}, hard exclude: {hard_exclude:?}, soft exclude: {context:?}, preferred_az: {:?})", scores.iter().map(|i| i.node_id().0).collect::<Vec<_>>(), preferred_az, ); } // Note that we do not update shard count here to reflect the scheduling: that // is IntentState's job when the scheduled location is used. Ok(node_id) } /// Selects any available node. This is suitable for performing background work (e.g. S3 /// deletions). pub(crate) fn any_available_node(&mut self) -> Result<NodeId, ScheduleError> { self.schedule_shard::<AttachedShardTag>(&[], &None, &ScheduleContext::default()) } /// For choosing which AZ to schedule a new shard into, use this. It will return the /// AZ with the the lowest number of shards currently scheduled in this AZ as their home /// location. /// /// We use an AZ-wide measure rather than simply selecting the AZ of the least-loaded /// node, because while tenants start out single sharded, when they grow and undergo /// shard-split, they will occupy space on many nodes within an AZ. It is important /// that we pick the AZ in a way that balances this _future_ load. /// /// Once we've picked an AZ, subsequent scheduling within that AZ will be driven by /// nodes' utilization scores. pub(crate) fn get_az_for_new_tenant(&self) -> Option<AvailabilityZone> { if self.nodes.is_empty() { return None; } #[derive(Default)] struct AzScore { home_shard_count: usize, scheduleable: bool, node_count: usize, } let mut azs: HashMap<&AvailabilityZone, AzScore> = HashMap::new(); for node in self.nodes.values() { let az = azs.entry(&node.az).or_default(); az.home_shard_count += node.home_shard_count; az.scheduleable |= matches!(node.may_schedule, MaySchedule::Yes(_)); az.node_count += 1; } // If any AZs are schedulable, then filter out the non-schedulable ones (i.e. AZs where // all nodes are overloaded or otherwise unschedulable). if azs.values().any(|i| i.scheduleable) { azs.retain(|_, i| i.scheduleable); } // We will multiply up shard counts by the max node count for scoring, before dividing // by per-node max node count, to get a normalized score that doesn't collapse to zero // when the absolute shard count is less than the node count. let max_node_count = azs.values().map(|i| i.node_count).max().unwrap_or(0); // Find the AZ with the lowest number of shards currently allocated Some( azs.into_iter() .min_by_key(|i| { ( (i.1.home_shard_count * max_node_count) / i.1.node_count, i.0, ) }) .unwrap() .0 .clone(), ) } pub(crate) fn get_node_az(&self, node_id: &NodeId) -> Option<AvailabilityZone> { self.nodes.get(node_id).map(|n| n.az.clone()) } /// For use when choosing a preferred secondary location: filter out nodes that are not /// available, and gather their AZs. pub(crate) fn filter_usable_nodes( &self, nodes: &[NodeId], ) -> Vec<(NodeId, Option<AvailabilityZone>)> { nodes .iter() .filter_map(|node_id| { let node = self .nodes .get(node_id) .expect("Referenced nodes always exist"); if matches!(node.may_schedule, MaySchedule::Yes(_)) { Some((*node_id, Some(node.az.clone()))) } else { None } }) .collect() }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service/safekeeper_service.rs
storage_controller/src/service/safekeeper_service.rs
use std::collections::HashSet; use std::str::FromStr; use std::sync::Arc; use std::time::Duration; use super::safekeeper_reconciler::ScheduleRequest; use crate::compute_hook; use crate::heartbeater::SafekeeperState; use crate::id_lock_map::trace_shared_lock; use crate::metrics; use crate::persistence::{ DatabaseError, SafekeeperTimelineOpKind, TimelinePendingOpPersistence, TimelinePersistence, TimelineUpdate, }; use crate::safekeeper::Safekeeper; use crate::safekeeper_client::SafekeeperClient; use crate::service::TenantOperations; use crate::timeline_import::TimelineImportFinalizeError; use anyhow::Context; use http_utils::error::ApiError; use pageserver_api::controller_api::{ SafekeeperDescribeResponse, SkSchedulingPolicy, TimelineImportRequest, TimelineSafekeeperMigrateRequest, }; use pageserver_api::models::{SafekeeperInfo, SafekeepersInfo, TimelineInfo}; use safekeeper_api::PgVersionId; use safekeeper_api::Term; use safekeeper_api::membership::{self, MemberSet, SafekeeperGeneration}; use safekeeper_api::models::{ PullTimelineRequest, TimelineLocateResponse, TimelineMembershipSwitchRequest, TimelineMembershipSwitchResponse, }; use safekeeper_client::mgmt_api; use tokio::task::JoinSet; use tokio_util::sync::CancellationToken; use utils::id::{NodeId, TenantId, TimelineId}; use utils::logging::SecretString; use utils::lsn::Lsn; use super::Service; impl Service { fn make_member_set(safekeepers: &[Safekeeper]) -> Result<MemberSet, anyhow::Error> { let members = safekeepers .iter() .map(|sk| sk.get_safekeeper_id()) .collect::<Vec<_>>(); MemberSet::new(members) } fn get_safekeepers(&self, ids: &[i64]) -> Result<Vec<Safekeeper>, ApiError> { let safekeepers = { let locked = self.inner.read().unwrap(); locked.safekeepers.clone() }; ids.iter() .map(|&id| { let node_id = NodeId(id as u64); safekeepers.get(&node_id).cloned().ok_or_else(|| { ApiError::InternalServerError(anyhow::anyhow!( "safekeeper {node_id} is not registered" )) }) }) .collect::<Result<Vec<_>, _>>() } /// Timeline creation on safekeepers /// /// Returns `Ok(left)` if the timeline has been created on a quorum of safekeepers, /// where `left` contains the list of safekeepers that didn't have a successful response. /// Assumes tenant lock is held while calling this function. pub(super) async fn tenant_timeline_create_safekeepers_quorum( &self, tenant_id: TenantId, timeline_id: TimelineId, pg_version: PgVersionId, timeline_persistence: &TimelinePersistence, ) -> Result<Vec<NodeId>, ApiError> { let safekeepers = self.get_safekeepers(&timeline_persistence.sk_set)?; let mset = Self::make_member_set(&safekeepers).map_err(ApiError::InternalServerError)?; let mconf = safekeeper_api::membership::Configuration::new(mset); let req = safekeeper_api::models::TimelineCreateRequest { commit_lsn: None, mconf, pg_version, start_lsn: timeline_persistence.start_lsn.0, system_id: None, tenant_id, timeline_id, wal_seg_size: None, }; const SK_CREATE_TIMELINE_RECONCILE_TIMEOUT: Duration = Duration::from_secs(30); let results = self .tenant_timeline_safekeeper_op_quorum( &safekeepers, move |client| { let req = req.clone(); async move { client.create_timeline(&req).await } }, SK_CREATE_TIMELINE_RECONCILE_TIMEOUT, ) .await?; Ok(results .into_iter() .enumerate() .filter_map(|(idx, res)| { if res.is_ok() { None // Success, don't return this safekeeper } else { Some(safekeepers[idx].get_id()) // Failure, return this safekeeper } }) .collect::<Vec<_>>()) } /// Perform an operation on a list of safekeepers in parallel with retries. /// /// Return the results of the operation on each safekeeper in the input order. async fn tenant_timeline_safekeeper_op<T, O, F>( &self, safekeepers: &[Safekeeper], op: O, timeout: Duration, ) -> Result<Vec<mgmt_api::Result<T>>, ApiError> where O: FnMut(SafekeeperClient) -> F + Send + 'static, O: Clone, F: std::future::Future<Output = mgmt_api::Result<T>> + Send + 'static, T: Sync + Send + 'static, { let jwt = self .config .safekeeper_jwt_token .clone() .map(SecretString::from); let mut joinset = JoinSet::new(); for (idx, sk) in safekeepers.iter().enumerate() { let sk = sk.clone(); let http_client = self.http_client.clone(); let jwt = jwt.clone(); let op = op.clone(); joinset.spawn(async move { let res = sk .with_client_retries( op, &http_client, &jwt, 3, 3, // TODO(diko): This is a wrong timeout. // It should be scaled to the retry count. timeout, &CancellationToken::new(), ) .await; (idx, res) }); } // Initialize results with timeout errors in case we never get a response. let mut results: Vec<mgmt_api::Result<T>> = safekeepers .iter() .map(|_| { Err(mgmt_api::Error::Timeout( "safekeeper operation timed out".to_string(), )) }) .collect(); // After we have built the joinset, we now wait for the tasks to complete, // but with a specified timeout to make sure we return swiftly, either with // a failure or success. let reconcile_deadline = tokio::time::Instant::now() + timeout; // Wait until all tasks finish or timeout is hit, whichever occurs // first. let mut result_count = 0; loop { if let Ok(res) = tokio::time::timeout_at(reconcile_deadline, joinset.join_next()).await { let Some(res) = res else { break }; match res { Ok((idx, res)) => { let sk = &safekeepers[idx]; tracing::info!( "response from safekeeper id:{} at {}: {:?}", sk.get_id(), sk.skp.host, // Only print errors, as there is no Debug trait for T. res.as_ref().map(|_| ()), ); results[idx] = res; result_count += 1; } Err(join_err) => { tracing::info!("join_err for task in joinset: {join_err}"); } } } else { tracing::info!("timeout for operation call after {result_count} responses",); break; } } Ok(results) } /// Perform an operation on a list of safekeepers in parallel with retries, /// and validates that we reach a quorum of successful responses. /// /// Return the results of the operation on each safekeeper in the input order. /// It's guaranteed that at least a quorum of the responses are successful. async fn tenant_timeline_safekeeper_op_quorum<T, O, F>( &self, safekeepers: &[Safekeeper], op: O, timeout: Duration, ) -> Result<Vec<mgmt_api::Result<T>>, ApiError> where O: FnMut(SafekeeperClient) -> F, O: Clone + Send + 'static, F: std::future::Future<Output = mgmt_api::Result<T>> + Send + 'static, T: Sync + Send + 'static, { let target_sk_count = safekeepers.len(); if target_sk_count == 0 { return Err(ApiError::InternalServerError(anyhow::anyhow!( "timeline configured without any safekeepers" ))); } if target_sk_count < self.config.timeline_safekeeper_count { tracing::warn!( "running a quorum operation with {} safekeepers, which is less than configured {} safekeepers per timeline", target_sk_count, self.config.timeline_safekeeper_count ); } let results = self .tenant_timeline_safekeeper_op(safekeepers, op, timeout) .await?; // Now check if quorum was reached in results. let quorum_size = target_sk_count / 2 + 1; let success_count = results.iter().filter(|res| res.is_ok()).count(); if success_count < quorum_size { // Failure return Err(ApiError::InternalServerError(anyhow::anyhow!( "not enough successful reconciliations to reach quorum size: {success_count} of {quorum_size} of total {target_sk_count}" ))); } Ok(results) } /// Create timeline in controller database and on safekeepers. /// `timeline_info` is result of timeline creation on pageserver. /// /// All actions must be idempotent as the call is retried until success. It /// tries to create timeline in the db and on at least majority of /// safekeepers + queue creation for safekeepers which missed it in the db /// for infinite retries; after that, call returns Ok. /// /// The idea is that once this is reached as long as we have alive majority /// of safekeepers it is expected to get eventually operational as storcon /// will be able to seed timeline on nodes which missed creation by making /// pull_timeline from peers. On the other hand we don't want to fail /// timeline creation if one safekeeper is down. pub(super) async fn tenant_timeline_create_safekeepers( self: &Arc<Self>, tenant_id: TenantId, timeline_info: &TimelineInfo, read_only: bool, ) -> Result<SafekeepersInfo, ApiError> { let timeline_id = timeline_info.timeline_id; let pg_version = PgVersionId::from(timeline_info.pg_version); // Initially start_lsn is determined by last_record_lsn in pageserver // response as it does initdb. However, later we persist it and in sk // creation calls replace with the value from the timeline row if it // previously existed as on retries in theory endpoint might have // already written some data and advanced last_record_lsn, while we want // safekeepers to have consistent start_lsn. let start_lsn = timeline_info.last_record_lsn; // Choose initial set of safekeepers respecting affinity let sks = if !read_only { self.safekeepers_for_new_timeline().await? } else { Vec::new() }; let sks_persistence = sks.iter().map(|sk| sk.id.0 as i64).collect::<Vec<_>>(); // Add timeline to db let mut timeline_persist = TimelinePersistence { tenant_id: tenant_id.to_string(), timeline_id: timeline_id.to_string(), start_lsn: start_lsn.into(), generation: 1, sk_set: sks_persistence.clone(), new_sk_set: None, cplane_notified_generation: 0, deleted_at: None, sk_set_notified_generation: 0, }; let inserted = self .persistence .insert_timeline(timeline_persist.clone()) .await?; if !inserted { if let Some(existent_persist) = self .persistence .get_timeline(tenant_id, timeline_id) .await? { // Replace with what we have in the db, to get stuff like the generation right. // We do still repeat the http calls to the safekeepers. After all, we could have // crashed right after the wrote to the DB. timeline_persist = existent_persist; } else { return Err(ApiError::InternalServerError(anyhow::anyhow!( "insertion said timeline already in db, but looking it up, it was gone" ))); } } let ret = SafekeepersInfo { generation: timeline_persist.generation as u32, safekeepers: sks.clone(), tenant_id, timeline_id, }; if read_only { return Ok(ret); } // Create the timeline on a quorum of safekeepers let remaining = self .tenant_timeline_create_safekeepers_quorum( tenant_id, timeline_id, pg_version, &timeline_persist, ) .await?; // For the remaining safekeepers, take care of their reconciliation asynchronously for &remaining_id in remaining.iter() { let pending_op = TimelinePendingOpPersistence { tenant_id: tenant_id.to_string(), timeline_id: timeline_id.to_string(), generation: timeline_persist.generation, op_kind: crate::persistence::SafekeeperTimelineOpKind::Pull, sk_id: remaining_id.0 as i64, }; tracing::info!("writing pending op for sk id {remaining_id}"); self.persistence.insert_pending_op(pending_op).await?; } if !remaining.is_empty() { let locked = self.inner.read().unwrap(); for remaining_id in remaining { let Some(sk) = locked.safekeepers.get(&remaining_id) else { return Err(ApiError::InternalServerError(anyhow::anyhow!( "Couldn't find safekeeper with id {remaining_id}" ))); }; let Ok(host_list) = sks .iter() .map(|sk| { Ok(( sk.id, locked .safekeepers .get(&sk.id) .ok_or_else(|| { ApiError::InternalServerError(anyhow::anyhow!( "Couldn't find safekeeper with id {} to pull from", sk.id )) })? .base_url(), )) }) .collect::<Result<_, ApiError>>() else { continue; }; let req = ScheduleRequest { safekeeper: Box::new(sk.clone()), host_list, tenant_id, timeline_id: Some(timeline_id), generation: timeline_persist.generation as u32, kind: crate::persistence::SafekeeperTimelineOpKind::Pull, }; locked.safekeeper_reconcilers.schedule_request(req); } } Ok(ret) } pub(crate) async fn tenant_timeline_create_safekeepers_until_success( self: &Arc<Self>, tenant_id: TenantId, timeline_info: TimelineInfo, ) -> Result<(), TimelineImportFinalizeError> { const BACKOFF: Duration = Duration::from_secs(5); loop { if self.cancel.is_cancelled() { return Err(TimelineImportFinalizeError::ShuttingDown); } // This function is only used in non-read-only scenarios let read_only = false; let res = self .tenant_timeline_create_safekeepers(tenant_id, &timeline_info, read_only) .await; match res { Ok(_) => { tracing::info!("Timeline created on safekeepers"); break; } Err(err) => { tracing::error!("Failed to create timeline on safekeepers: {err}"); tokio::select! { _ = self.cancel.cancelled() => { return Err(TimelineImportFinalizeError::ShuttingDown); }, _ = tokio::time::sleep(BACKOFF) => {} }; } } } Ok(()) } /// Directly insert the timeline into the database without reconciling it with safekeepers. /// /// Useful if the timeline already exists on the specified safekeepers, /// but we want to make it storage controller managed. pub(crate) async fn timeline_import(&self, req: TimelineImportRequest) -> Result<(), ApiError> { let persistence = TimelinePersistence { tenant_id: req.tenant_id.to_string(), timeline_id: req.timeline_id.to_string(), start_lsn: req.start_lsn.into(), generation: 1, sk_set: req.sk_set.iter().map(|sk_id| sk_id.0 as i64).collect(), new_sk_set: None, cplane_notified_generation: 1, deleted_at: None, sk_set_notified_generation: 1, }; let inserted = self .persistence .insert_timeline(persistence.clone()) .await?; if inserted { tracing::info!("imported timeline into db"); return Ok(()); } tracing::info!("timeline already present in db, updating"); let update = TimelineUpdate { tenant_id: persistence.tenant_id, timeline_id: persistence.timeline_id, start_lsn: persistence.start_lsn, sk_set: persistence.sk_set, new_sk_set: persistence.new_sk_set, }; self.persistence.update_timeline_unsafe(update).await?; tracing::info!("timeline updated"); Ok(()) } /// Locate safekeepers for a timeline. /// Return the generation, sk_set and new_sk_set if present. /// If the timeline is not storcon-managed, return NotFound. pub(crate) async fn tenant_timeline_locate( &self, tenant_id: TenantId, timeline_id: TimelineId, ) -> Result<TimelineLocateResponse, ApiError> { let timeline = self .persistence .get_timeline(tenant_id, timeline_id) .await?; let Some(timeline) = timeline else { return Err(ApiError::NotFound( anyhow::anyhow!("Timeline {}/{} not found", tenant_id, timeline_id).into(), )); }; Ok(TimelineLocateResponse { generation: SafekeeperGeneration::new(timeline.generation as u32), sk_set: timeline .sk_set .iter() .map(|id| NodeId(*id as u64)) .collect(), new_sk_set: timeline .new_sk_set .map(|sk_set| sk_set.iter().map(|id| NodeId(*id as u64)).collect()), }) } /// Perform timeline deletion on safekeepers. Will return success: we persist the deletion into the reconciler. pub(super) async fn tenant_timeline_delete_safekeepers( self: &Arc<Self>, tenant_id: TenantId, timeline_id: TimelineId, ) -> Result<(), ApiError> { let tl = self .persistence .get_timeline(tenant_id, timeline_id) .await?; let Some(tl) = tl else { tracing::info!( "timeline {tenant_id}/{timeline_id} doesn't exist in timelines table, no deletions on safekeepers needed" ); return Ok(()); }; self.persistence .timeline_set_deleted_at(tenant_id, timeline_id) .await?; let all_sks = tl .new_sk_set .iter() .flatten() .chain(tl.sk_set.iter()) .collect::<HashSet<_>>(); // The timeline has no safekeepers: we need to delete it from the db manually, // as no safekeeper reconciler will get to it if all_sks.is_empty() { if let Err(err) = self .persistence .delete_timeline(tenant_id, timeline_id) .await { tracing::warn!(%tenant_id, %timeline_id, "couldn't delete timeline from db: {err}"); } } // Schedule reconciliations for &sk_id in all_sks.iter() { let pending_op = TimelinePendingOpPersistence { tenant_id: tenant_id.to_string(), timeline_id: timeline_id.to_string(), generation: i32::MAX, op_kind: SafekeeperTimelineOpKind::Delete, sk_id: *sk_id, }; tracing::info!("writing pending op for sk id {sk_id}"); self.persistence.insert_pending_op(pending_op).await?; } { let locked = self.inner.read().unwrap(); for sk_id in all_sks { let sk_id = NodeId(*sk_id as u64); let Some(sk) = locked.safekeepers.get(&sk_id) else { return Err(ApiError::InternalServerError(anyhow::anyhow!( "Couldn't find safekeeper with id {sk_id}" ))); }; let req = ScheduleRequest { safekeeper: Box::new(sk.clone()), // we don't use this for this kind, put a dummy value host_list: Vec::new(), tenant_id, timeline_id: Some(timeline_id), generation: tl.generation as u32, kind: SafekeeperTimelineOpKind::Delete, }; locked.safekeeper_reconcilers.schedule_request(req); } } Ok(()) } /// Perform tenant deletion on safekeepers. pub(super) async fn tenant_delete_safekeepers( self: &Arc<Self>, tenant_id: TenantId, ) -> Result<(), ApiError> { let timeline_list = self .persistence .list_timelines_for_tenant(tenant_id) .await?; if timeline_list.is_empty() { // Early exit: the tenant is either empty or not migrated to the storcon yet tracing::info!("Skipping tenant delete as the timeline doesn't exist in db"); return Ok(()); } let timeline_list = timeline_list .into_iter() .map(|timeline| { let timeline_id = TimelineId::from_str(&timeline.timeline_id) .context("timeline id loaded from db") .map_err(ApiError::InternalServerError)?; Ok((timeline_id, timeline)) }) .collect::<Result<Vec<_>, ApiError>>()?; // Remove pending ops from db, and set `deleted_at`. // We cancel them in a later iteration once we hold the state lock. for (timeline_id, _timeline) in timeline_list.iter() { self.persistence .remove_pending_ops_for_timeline(tenant_id, Some(*timeline_id)) .await?; self.persistence .timeline_set_deleted_at(tenant_id, *timeline_id) .await?; } // The list of safekeepers that have any of the timelines let mut sk_list = HashSet::new(); // List all pending ops for all timelines, cancel them for (_timeline_id, timeline) in timeline_list.iter() { let sk_iter = timeline .sk_set .iter() .chain(timeline.new_sk_set.iter().flatten()) .map(|id| NodeId(*id as u64)); sk_list.extend(sk_iter); } for &sk_id in sk_list.iter() { let pending_op = TimelinePendingOpPersistence { tenant_id: tenant_id.to_string(), timeline_id: String::new(), generation: i32::MAX, op_kind: SafekeeperTimelineOpKind::Delete, sk_id: sk_id.0 as i64, }; tracing::info!("writing pending op for sk id {sk_id}"); self.persistence.insert_pending_op(pending_op).await?; } let mut locked = self.inner.write().unwrap(); for (timeline_id, _timeline) in timeline_list.iter() { for sk_id in sk_list.iter() { locked .safekeeper_reconcilers .cancel_reconciles_for_timeline(*sk_id, tenant_id, Some(*timeline_id)); } } // unwrap is safe: we return above for an empty timeline list let max_generation = timeline_list .iter() .map(|(_tl_id, tl)| tl.generation as u32) .max() .unwrap(); for sk_id in sk_list { let Some(safekeeper) = locked.safekeepers.get(&sk_id) else { tracing::warn!("Couldn't find safekeeper with id {sk_id}"); continue; }; // Add pending op for tenant deletion let req = ScheduleRequest { generation: max_generation, host_list: Vec::new(), kind: SafekeeperTimelineOpKind::Delete, safekeeper: Box::new(safekeeper.clone()), tenant_id, timeline_id: None, }; locked.safekeeper_reconcilers.schedule_request(req); } Ok(()) } /// Choose safekeepers for the new timeline in different azs. /// 3 are choosen by default, but may be configured via config (for testing). pub(crate) async fn safekeepers_for_new_timeline( &self, ) -> Result<Vec<SafekeeperInfo>, ApiError> { let mut all_safekeepers = { let locked = self.inner.read().unwrap(); locked .safekeepers .iter() .filter_map(|sk| { if sk.1.scheduling_policy() != SkSchedulingPolicy::Active { // If we don't want to schedule stuff onto the safekeeper, respect that. return None; } let utilization_opt = if let SafekeeperState::Available { last_seen_at: _, utilization, } = sk.1.availability() { Some(utilization) } else { // non-available safekeepers still get a chance for new timelines, // but put them last in the list. None }; let info = SafekeeperInfo { hostname: sk.1.skp.host.clone(), id: NodeId(sk.1.skp.id as u64), }; Some((utilization_opt, info, sk.1.skp.availability_zone_id.clone())) }) .collect::<Vec<_>>() }; all_safekeepers.sort_by_key(|sk| { ( sk.0.as_ref() .map(|ut| ut.timeline_count) .unwrap_or(u64::MAX), // Use the id to decide on equal scores for reliability sk.1.id.0, ) }); // Number of safekeepers in different AZs we are looking for let wanted_count = self.config.timeline_safekeeper_count; let mut sks = Vec::new(); let mut azs = HashSet::new(); for (_sk_util, sk_info, az_id) in all_safekeepers.iter() { if !azs.insert(az_id) { continue; } sks.push(sk_info.clone()); if sks.len() == wanted_count { break; } } if sks.len() == wanted_count { Ok(sks) } else { Err(ApiError::InternalServerError(anyhow::anyhow!( "couldn't find {wanted_count} safekeepers in different AZs for new timeline (found: {}, total active: {})", sks.len(), all_safekeepers.len(), ))) } } pub(crate) async fn safekeepers_list( &self, ) -> Result<Vec<SafekeeperDescribeResponse>, DatabaseError> { let locked = self.inner.read().unwrap(); let mut list = locked .safekeepers .iter() .map(|sk| sk.1.describe_response()) .collect::<Result<Vec<_>, _>>()?; list.sort_by_key(|v| v.id); Ok(list) } pub(crate) async fn get_safekeeper( &self, id: i64, ) -> Result<SafekeeperDescribeResponse, DatabaseError> { let locked = self.inner.read().unwrap(); let sk = locked .safekeepers .get(&NodeId(id as u64)) .ok_or(diesel::result::Error::NotFound)?; sk.describe_response() } pub(crate) async fn upsert_safekeeper( self: &Arc<Service>, record: crate::persistence::SafekeeperUpsert, ) -> Result<(), ApiError> { let node_id = NodeId(record.id as u64); let use_https = self.config.use_https_safekeeper_api; if use_https && record.https_port.is_none() { return Err(ApiError::PreconditionFailed( format!( "cannot upsert safekeeper {node_id}: \ https is enabled, but https port is not specified" ) .into(), )); } self.persistence.safekeeper_upsert(record.clone()).await?; { let mut locked = self.inner.write().unwrap(); let mut safekeepers = (*locked.safekeepers).clone(); match safekeepers.entry(node_id) { std::collections::hash_map::Entry::Occupied(mut entry) => entry .get_mut() .update_from_record(record) .expect("all preconditions should be checked before upsert to database"), std::collections::hash_map::Entry::Vacant(entry) => { entry.insert( Safekeeper::from_persistence( crate::persistence::SafekeeperPersistence::from_upsert( record, SkSchedulingPolicy::Activating, ), CancellationToken::new(), use_https, ) .expect("all preconditions should be checked before upsert to database"), ); } } locked .safekeeper_reconcilers .start_reconciler(node_id, self); locked.safekeepers = Arc::new(safekeepers); metrics::METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_nodes .set(locked.safekeepers.len() as i64); metrics::METRICS_REGISTRY .metrics_group .storage_controller_https_safekeeper_nodes .set( locked .safekeepers .values() .filter(|s| s.has_https_port()) .count() as i64, ); } Ok(()) } pub(crate) async fn set_safekeeper_scheduling_policy( self: &Arc<Service>, id: i64, scheduling_policy: SkSchedulingPolicy, ) -> Result<(), DatabaseError> { self.persistence .set_safekeeper_scheduling_policy(id, scheduling_policy) .await?; let node_id = NodeId(id as u64); // After the change has been persisted successfully, update the in-memory state self.set_safekeeper_scheduling_policy_in_mem(node_id, scheduling_policy)
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service/feature_flag.rs
storage_controller/src/service/feature_flag.rs
use std::{sync::Arc, time::Duration}; use futures::StreamExt; use pageserver_api::config::PostHogConfig; use pageserver_client::mgmt_api; use posthog_client_lite::PostHogClient; use reqwest::StatusCode; use tokio::time::MissedTickBehavior; use tokio_util::sync::CancellationToken; use crate::{pageserver_client::PageserverClient, service::Service}; pub struct FeatureFlagService { service: Arc<Service>, config: PostHogConfig, client: PostHogClient, http_client: reqwest::Client, } const DEFAULT_POSTHOG_REFRESH_INTERVAL: Duration = Duration::from_secs(30); impl FeatureFlagService { pub fn new(service: Arc<Service>, config: PostHogConfig) -> Result<Self, &'static str> { let client = PostHogClient::new(config.clone().try_into_posthog_config()?); Ok(Self { service, config, client, http_client: reqwest::Client::new(), }) } async fn refresh(self: Arc<Self>, cancel: CancellationToken) -> Result<(), anyhow::Error> { let nodes = { let inner = self.service.inner.read().unwrap(); inner.nodes.clone() }; let feature_flag_spec = self.client.get_feature_flags_local_evaluation_raw().await?; let stream = futures::stream::iter(nodes.values().cloned()).map(|node| { let this = self.clone(); let feature_flag_spec = feature_flag_spec.clone(); async move { let res = async { let client = PageserverClient::new( node.get_id(), this.http_client.clone(), node.base_url(), // TODO: what if we rotate the token during storcon lifetime? this.service.config.pageserver_jwt_token.as_deref(), ); client.update_feature_flag_spec(feature_flag_spec).await?; tracing::info!( "Updated {}({}) with feature flag spec", node.get_id(), node.base_url() ); Ok::<_, mgmt_api::Error>(()) }; if let Err(e) = res.await { if let mgmt_api::Error::ApiError(status, _) = e { if status == StatusCode::NOT_FOUND { // This is expected during deployments where the API is not available, so we can ignore it return; } } tracing::warn!( "Failed to update feature flag spec for {}: {e}", node.get_id() ); } } }); let mut stream = stream.buffer_unordered(8); while stream.next().await.is_some() { if cancel.is_cancelled() { return Ok(()); } } Ok(()) } pub async fn run(self: Arc<Self>, cancel: CancellationToken) { let refresh_interval = self .config .refresh_interval .unwrap_or(DEFAULT_POSTHOG_REFRESH_INTERVAL); let mut interval = tokio::time::interval(refresh_interval); interval.set_missed_tick_behavior(MissedTickBehavior::Skip); tracing::info!( "Starting feature flag service with refresh interval: {:?}", refresh_interval ); loop { tokio::select! { _ = interval.tick() => {} _ = cancel.cancelled() => { break; } } let res = self.clone().refresh(cancel.clone()).await; if let Err(e) = res { tracing::error!("Failed to refresh feature flags: {e:#?}"); } } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service/chaos_injector.rs
storage_controller/src/service/chaos_injector.rs
use std::collections::{BTreeMap, HashMap}; use std::sync::Arc; use std::time::Duration; use pageserver_api::controller_api::ShardSchedulingPolicy; use rand::Rng; use rand::seq::{IndexedRandom, SliceRandom}; use tokio_util::sync::CancellationToken; use utils::id::NodeId; use utils::shard::TenantShardId; use super::{Node, Scheduler, Service, TenantShard}; pub struct ChaosInjector { service: Arc<Service>, interval: Duration, chaos_exit_crontab: Option<cron::Schedule>, } fn cron_to_next_duration(cron: &cron::Schedule) -> anyhow::Result<tokio::time::Sleep> { use chrono::Utc; let next = cron.upcoming(Utc).next().unwrap(); let duration = (next - Utc::now()).to_std()?; Ok(tokio::time::sleep(duration)) } async fn maybe_sleep(sleep: Option<tokio::time::Sleep>) -> Option<()> { if let Some(sleep) = sleep { sleep.await; Some(()) } else { None } } impl ChaosInjector { pub fn new( service: Arc<Service>, interval: Duration, chaos_exit_crontab: Option<cron::Schedule>, ) -> Self { Self { service, interval, chaos_exit_crontab, } } fn get_cron_interval_sleep_future(&self) -> Option<tokio::time::Sleep> { if let Some(ref chaos_exit_crontab) = self.chaos_exit_crontab { match cron_to_next_duration(chaos_exit_crontab) { Ok(interval_exit) => Some(interval_exit), Err(e) => { tracing::error!("Error processing the cron schedule: {e}"); None } } } else { None } } pub async fn run(&mut self, cancel: CancellationToken) { let mut interval = tokio::time::interval(self.interval); #[derive(Debug)] enum ChaosEvent { MigrationsToSecondary, ForceKillController, GracefulMigrationsAnywhere, } loop { let cron_interval = self.get_cron_interval_sleep_future(); let chaos_type = tokio::select! { _ = interval.tick() => { if rand::rng().random_bool(0.5) { ChaosEvent::MigrationsToSecondary } else { ChaosEvent::GracefulMigrationsAnywhere } } Some(_) = maybe_sleep(cron_interval) => { ChaosEvent::ForceKillController } _ = cancel.cancelled() => { tracing::info!("Shutting down"); return; } }; tracing::info!("Chaos iteration: {chaos_type:?}..."); match chaos_type { ChaosEvent::MigrationsToSecondary => { self.inject_migrations_to_secondary(); } ChaosEvent::GracefulMigrationsAnywhere => { self.inject_graceful_migrations_anywhere(); } ChaosEvent::ForceKillController => { self.force_kill().await; } } } } fn is_shard_eligible_for_chaos(&self, shard: &TenantShard) -> bool { // - Skip non-active scheduling policies, so that a shard with a policy like Pause can // be pinned without being disrupted by us. // - Skip shards doing a graceful migration already, so that we allow these to run to // completion rather than only exercising the first part and then cancelling with // some other chaos. matches!(shard.get_scheduling_policy(), ShardSchedulingPolicy::Active) && shard.get_preferred_node().is_none() } /// If a shard has a secondary and attached location, then re-assign the secondary to be /// attached and the attached to be secondary. /// /// Only modifies tenants if they're in Active scheduling policy. fn maybe_migrate_to_secondary( &self, tenant_shard_id: TenantShardId, nodes: &Arc<HashMap<NodeId, Node>>, tenants: &mut BTreeMap<TenantShardId, TenantShard>, scheduler: &mut Scheduler, ) { let shard = tenants .get_mut(&tenant_shard_id) .expect("Held lock between choosing ID and this get"); if !self.is_shard_eligible_for_chaos(shard) { return; } // Pick a secondary to promote let Some(new_location) = shard .intent .get_secondary() .choose(&mut rand::rng()) .cloned() else { tracing::info!( "Skipping shard {tenant_shard_id}: no secondary location, can't migrate" ); return; }; let Some(old_location) = *shard.intent.get_attached() else { tracing::info!("Skipping shard {tenant_shard_id}: currently has no attached location"); return; }; tracing::info!("Injecting chaos: migrate {tenant_shard_id} {old_location}->{new_location}"); shard.intent.demote_attached(scheduler, old_location); shard.intent.promote_attached(scheduler, new_location); self.service.maybe_reconcile_shard( shard, nodes, crate::reconciler::ReconcilerPriority::Normal, ); } async fn force_kill(&mut self) { tracing::warn!("Injecting chaos: force kill"); std::process::exit(1); } // Unlike [`Self::inject_migrations_to_secondary`], this function will not only cut over to secondary, it // will migrate a tenant to a random node in its home AZ using a graceful migration of the same type // that my be initiated by an API caller using prewarm=true. // // This is a much more expensive operation in terms of I/O and time, as we will fully warm up // some new location in order to migrate the tenant there. For that reason we do far fewer of these. fn inject_graceful_migrations_anywhere(&mut self) { let batch_size = 1; let mut inner = self.service.inner.write().unwrap(); let (nodes, tenants, _scheduler) = inner.parts_mut(); let mut candidates = tenants .values_mut() .filter(|shard| self.is_shard_eligible_for_chaos(shard)) .collect::<Vec<_>>(); tracing::info!( "Injecting chaos: found {} candidates for graceful migrations anywhere", candidates.len() ); let mut victims: Vec<&mut TenantShard> = Vec::new(); // Pick our victims: use a hand-rolled loop rather than choose_multiple() because we want // to take the mutable refs from our candidates rather than ref'ing them. while !candidates.is_empty() && victims.len() < batch_size { let i = rand::rng().random_range(0..candidates.len()); victims.push(candidates.swap_remove(i)); } for victim in victims.into_iter() { // Find a node in the same AZ as the shard, or if the shard has no AZ preference, which // is not where they are currently attached. let candidate_nodes = nodes .values() .filter(|node| { if let Some(preferred_az) = victim.preferred_az() { node.get_availability_zone_id() == preferred_az } else if let Some(attached) = *victim.intent.get_attached() { node.get_id() != attached } else { true } }) .collect::<Vec<_>>(); let Some(victim_node) = candidate_nodes.choose(&mut rand::rng()) else { // This can happen if e.g. we are in a small region with only one pageserver per AZ. tracing::info!( "no candidate nodes found for migrating shard {tenant_shard_id} within its home AZ", tenant_shard_id = victim.tenant_shard_id ); continue; }; // This doesn't change intent immediately: next iteration of Service::optimize_all should do that. We avoid // doing it here because applying optimizations requires dropping lock to do some async work to check the optimisation // is valid given remote state, and it would be a shame to duplicate that dance here. tracing::info!( "Injecting chaos: migrate {} to {}", victim.tenant_shard_id, victim_node ); victim.set_preferred_node(Some(victim_node.get_id())); } } /// Migrations of attached locations to their secondary location. This exercises reconciliation in general, /// live migration in particular, and the pageserver code for cleanly shutting down and starting up tenants /// during such migrations. fn inject_migrations_to_secondary(&mut self) { // Pick some shards to interfere with let batch_size = 128; let mut inner = self.service.inner.write().unwrap(); let (nodes, tenants, scheduler) = inner.parts_mut(); // Prefer to migrate tenants that are currently outside their home AZ. This avoids the chaos injector // continuously pushing tenants outside their home AZ: instead, we'll tend to cycle between picking some // random tenants to move, and then on next chaos iteration moving them back, then picking some new // random tenants on the next iteration. let (out_of_home_az, in_home_az): (Vec<_>, Vec<_>) = tenants .values() .map(|shard| { ( shard.tenant_shard_id, shard.is_attached_outside_preferred_az(nodes), ) }) .partition(|(_id, is_outside)| *is_outside); let mut out_of_home_az: Vec<_> = out_of_home_az.into_iter().map(|(id, _)| id).collect(); let mut in_home_az: Vec<_> = in_home_az.into_iter().map(|(id, _)| id).collect(); let mut victims = Vec::with_capacity(batch_size); if out_of_home_az.len() >= batch_size { tracing::info!( "Injecting chaos: found {batch_size} shards to migrate back to home AZ (total {} out of home AZ)", out_of_home_az.len() ); out_of_home_az.shuffle(&mut rand::rng()); victims.extend(out_of_home_az.into_iter().take(batch_size)); } else { tracing::info!( "Injecting chaos: found {} shards to migrate back to home AZ, picking {} random shards to migrate", out_of_home_az.len(), std::cmp::min(batch_size - out_of_home_az.len(), in_home_az.len()) ); victims.extend(out_of_home_az); in_home_az.shuffle(&mut rand::rng()); victims.extend(in_home_az.into_iter().take(batch_size - victims.len())); } for victim in victims { self.maybe_migrate_to_secondary(victim, nodes, tenants, scheduler); } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service/tenant_shard_iterator.rs
storage_controller/src/service/tenant_shard_iterator.rs
use std::collections::BTreeMap; use std::sync::Arc; use utils::id::TenantId; use utils::shard::TenantShardId; use crate::scheduler::{ScheduleContext, ScheduleMode}; use crate::tenant_shard::TenantShard; use super::Service; /// Exclusive iterator over all tenant shards. /// It is used to iterate over consistent tenants state at specific point in time. /// /// When making scheduling decisions, it is useful to have the ScheduleContext for a whole /// tenant while considering the individual shards within it. This iterator is a helper /// that gathers all the shards in a tenant and then yields them together with a ScheduleContext /// for the tenant. pub(super) struct TenantShardExclusiveIterator<'a> { schedule_mode: ScheduleMode, inner: std::collections::btree_map::IterMut<'a, TenantShardId, TenantShard>, } impl<'a> TenantShardExclusiveIterator<'a> { pub(super) fn new( tenants: &'a mut BTreeMap<TenantShardId, TenantShard>, schedule_mode: ScheduleMode, ) -> Self { Self { schedule_mode, inner: tenants.iter_mut(), } } } impl<'a> Iterator for TenantShardExclusiveIterator<'a> { type Item = (TenantId, ScheduleContext, Vec<&'a mut TenantShard>); fn next(&mut self) -> Option<Self::Item> { let mut tenant_shards = Vec::new(); let mut schedule_context = ScheduleContext::new(self.schedule_mode.clone()); loop { let (tenant_shard_id, shard) = self.inner.next()?; if tenant_shard_id.is_shard_zero() { // Cleared on last shard of previous tenant assert!(tenant_shards.is_empty()); } // Accumulate the schedule context for all the shards in a tenant schedule_context.avoid(&shard.intent.all_pageservers()); tenant_shards.push(shard); if tenant_shard_id.shard_number.0 == tenant_shard_id.shard_count.count() - 1 { return Some((tenant_shard_id.tenant_id, schedule_context, tenant_shards)); } } } } /// Shared iterator over all tenant shards. /// It is used to iterate over all tenants without blocking another code, working with tenants /// /// A simple iterator which can be used in tandem with [`crate::service::Service`] /// to iterate over all known tenant shard ids without holding the lock on the /// service state at all times. pub(crate) struct TenantShardSharedIterator<F> { tenants_accessor: F, inspected_all_shards: bool, last_inspected_shard: Option<TenantShardId>, } impl<F> TenantShardSharedIterator<F> where F: Fn(Option<TenantShardId>) -> Option<TenantShardId>, { pub(crate) fn new(tenants_accessor: F) -> Self { Self { tenants_accessor, inspected_all_shards: false, last_inspected_shard: None, } } pub(crate) fn finished(&self) -> bool { self.inspected_all_shards } } impl<F> Iterator for TenantShardSharedIterator<F> where F: Fn(Option<TenantShardId>) -> Option<TenantShardId>, { // TODO(ephemeralsad): consider adding schedule context to the iterator type Item = TenantShardId; /// Returns the next tenant shard id if one exists fn next(&mut self) -> Option<Self::Item> { if self.inspected_all_shards { return None; } match (self.tenants_accessor)(self.last_inspected_shard) { Some(tid) => { self.last_inspected_shard = Some(tid); Some(tid) } None => { self.inspected_all_shards = true; None } } } } pub(crate) fn create_shared_shard_iterator( service: Arc<Service>, ) -> TenantShardSharedIterator<impl Fn(Option<TenantShardId>) -> Option<TenantShardId>> { let tenants_accessor = move |last_inspected_shard: Option<TenantShardId>| { let locked = &service.inner.read().unwrap(); let tenants = &locked.tenants; let entry = match last_inspected_shard { Some(skip_past) => { // Skip to the last seen tenant shard id let mut cursor = tenants.iter().skip_while(|(tid, _)| **tid != skip_past); // Skip past the last seen cursor.nth(1) } None => tenants.first_key_value(), }; entry.map(|(tid, _)| tid).copied() }; TenantShardSharedIterator::new(tenants_accessor) } #[cfg(test)] mod tests { use std::collections::BTreeMap; use std::str::FromStr; use std::sync::Arc; use pageserver_api::controller_api::PlacementPolicy; use utils::id::TenantId; use utils::shard::{ShardCount, ShardNumber, TenantShardId}; use super::*; use crate::scheduler::test_utils::make_test_nodes; use crate::service::Scheduler; use crate::tenant_shard::tests::make_test_tenant_with_id; #[test] fn test_exclusive_shard_iterator() { // Hand-crafted tenant IDs to ensure they appear in the expected order when put into // a btreemap & iterated let mut t_1_shards = make_test_tenant_with_id( TenantId::from_str("af0480929707ee75372337efaa5ecf96").unwrap(), PlacementPolicy::Attached(1), ShardCount(1), None, ); let t_2_shards = make_test_tenant_with_id( TenantId::from_str("bf0480929707ee75372337efaa5ecf96").unwrap(), PlacementPolicy::Attached(1), ShardCount(4), None, ); let mut t_3_shards = make_test_tenant_with_id( TenantId::from_str("cf0480929707ee75372337efaa5ecf96").unwrap(), PlacementPolicy::Attached(1), ShardCount(1), None, ); let t1_id = t_1_shards[0].tenant_shard_id.tenant_id; let t2_id = t_2_shards[0].tenant_shard_id.tenant_id; let t3_id = t_3_shards[0].tenant_shard_id.tenant_id; let mut tenants = BTreeMap::new(); tenants.insert(t_1_shards[0].tenant_shard_id, t_1_shards.pop().unwrap()); for shard in t_2_shards { tenants.insert(shard.tenant_shard_id, shard); } tenants.insert(t_3_shards[0].tenant_shard_id, t_3_shards.pop().unwrap()); let nodes = make_test_nodes(3, &[]); let mut scheduler = Scheduler::new(nodes.values()); let mut context = ScheduleContext::default(); for shard in tenants.values_mut() { shard.schedule(&mut scheduler, &mut context).unwrap(); } let mut iter = TenantShardExclusiveIterator::new(&mut tenants, ScheduleMode::Speculative); let (tenant_id, context, shards) = iter.next().unwrap(); assert_eq!(tenant_id, t1_id); assert_eq!(shards[0].tenant_shard_id.shard_number, ShardNumber(0)); assert_eq!(shards.len(), 1); assert_eq!(context.location_count(), 2); let (tenant_id, context, shards) = iter.next().unwrap(); assert_eq!(tenant_id, t2_id); assert_eq!(shards[0].tenant_shard_id.shard_number, ShardNumber(0)); assert_eq!(shards[1].tenant_shard_id.shard_number, ShardNumber(1)); assert_eq!(shards[2].tenant_shard_id.shard_number, ShardNumber(2)); assert_eq!(shards[3].tenant_shard_id.shard_number, ShardNumber(3)); assert_eq!(shards.len(), 4); assert_eq!(context.location_count(), 8); let (tenant_id, context, shards) = iter.next().unwrap(); assert_eq!(tenant_id, t3_id); assert_eq!(shards[0].tenant_shard_id.shard_number, ShardNumber(0)); assert_eq!(shards.len(), 1); assert_eq!(context.location_count(), 2); for shard in tenants.values_mut() { shard.intent.clear(&mut scheduler); } } #[test] fn test_shared_shard_iterator() { let tenant_id = TenantId::generate(); let shard_count = ShardCount(8); let mut tenant_shards = Vec::default(); for i in 0..shard_count.0 { tenant_shards.push(( TenantShardId { tenant_id, shard_number: ShardNumber(i), shard_count, }, (), )) } let tenant_shards = Arc::new(tenant_shards); let tid_iter = TenantShardSharedIterator::new({ let tenants = tenant_shards.clone(); move |last_inspected_shard: Option<TenantShardId>| { let entry = match last_inspected_shard { Some(skip_past) => { let mut cursor = tenants.iter().skip_while(|(tid, _)| *tid != skip_past); cursor.nth(1) } None => tenants.first(), }; entry.map(|(tid, _)| tid).copied() } }); let mut iterated_over = Vec::default(); for tid in tid_iter { iterated_over.push((tid, ())); } assert_eq!(iterated_over, *tenant_shards); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/service/safekeeper_reconciler.rs
storage_controller/src/service/safekeeper_reconciler.rs
use std::{ collections::HashMap, str::FromStr, sync::{Arc, atomic::AtomicU64}, time::Duration, }; use clashmap::{ClashMap, Entry}; use safekeeper_api::models::PullTimelineRequest; use safekeeper_client::mgmt_api; use tokio::sync::{ Semaphore, mpsc::{self, UnboundedReceiver, UnboundedSender}, }; use tokio_util::sync::CancellationToken; use tracing::Instrument; use utils::{ id::{NodeId, TenantId, TimelineId}, logging::SecretString, }; use crate::{ metrics::{METRICS_REGISTRY, SafekeeperReconcilerLabelGroup}, persistence::SafekeeperTimelineOpKind, safekeeper::Safekeeper, safekeeper_client::SafekeeperClient, }; use super::Service; pub(crate) struct SafekeeperReconcilers { cancel: CancellationToken, reconcilers: HashMap<NodeId, ReconcilerHandle>, } impl SafekeeperReconcilers { pub fn new(cancel: CancellationToken) -> Self { SafekeeperReconcilers { cancel, reconcilers: HashMap::new(), } } /// Adds a safekeeper-specific reconciler. /// Can be called multiple times, but it needs to be called at least once /// for every new safekeeper added. pub(crate) fn start_reconciler(&mut self, node_id: NodeId, service: &Arc<Service>) { self.reconcilers.entry(node_id).or_insert_with(|| { SafekeeperReconciler::spawn(self.cancel.child_token(), service.clone()) }); } /// Stop a safekeeper-specific reconciler. /// Stops the reconciler, cancelling all ongoing tasks. pub(crate) fn stop_reconciler(&mut self, node_id: NodeId) { if let Some(handle) = self.reconcilers.remove(&node_id) { handle.cancel.cancel(); } } pub(crate) fn schedule_request_vec(&self, reqs: Vec<ScheduleRequest>) { tracing::info!( "Scheduling {} pending safekeeper ops loaded from db", reqs.len() ); for req in reqs { self.schedule_request(req); } } pub(crate) fn schedule_request(&self, req: ScheduleRequest) { let node_id = req.safekeeper.get_id(); let reconciler_handle = self.reconcilers.get(&node_id).unwrap(); reconciler_handle.schedule_reconcile(req); } /// Cancel ongoing reconciles for the given timeline /// /// Specifying `None` here only removes reconciles for the tenant-global reconciliation, /// instead of doing this for all timelines of the tenant. /// /// Callers must remove the reconciles from the db manually pub(crate) fn cancel_reconciles_for_timeline( &mut self, node_id: NodeId, tenant_id: TenantId, timeline_id: Option<TimelineId>, ) { if let Some(handle) = self.reconcilers.get(&node_id) { handle.cancel_reconciliation(tenant_id, timeline_id); } } } /// Initial load of the pending operations from the db pub(crate) async fn load_schedule_requests( service: &Arc<Service>, safekeepers: &HashMap<NodeId, Safekeeper>, ) -> anyhow::Result<Vec<ScheduleRequest>> { let pending_ops_timelines = service .persistence .list_pending_ops_with_timelines() .await?; let mut res = Vec::with_capacity(pending_ops_timelines.len()); for (op_persist, timeline_persist) in pending_ops_timelines { let node_id = NodeId(op_persist.sk_id as u64); let Some(sk) = safekeepers.get(&node_id) else { // This shouldn't happen, at least the safekeeper should exist as decomissioned. tracing::warn!( tenant_id = op_persist.tenant_id, timeline_id = op_persist.timeline_id, "couldn't find safekeeper with pending op id {node_id} in list of stored safekeepers" ); continue; }; let sk = Box::new(sk.clone()); let tenant_id = TenantId::from_str(&op_persist.tenant_id)?; let timeline_id = if !op_persist.timeline_id.is_empty() { Some(TimelineId::from_str(&op_persist.timeline_id)?) } else { None }; let host_list = match op_persist.op_kind { SafekeeperTimelineOpKind::Delete => Vec::new(), SafekeeperTimelineOpKind::Exclude => Vec::new(), SafekeeperTimelineOpKind::Pull => { if timeline_id.is_none() { // We only do this extra check (outside of timeline_persist check) to give better error msgs anyhow::bail!( "timeline_id is empty for `pull` schedule request for {tenant_id}" ); }; let Some(timeline_persist) = timeline_persist else { // This shouldn't happen, the timeline should still exist tracing::warn!( tenant_id = op_persist.tenant_id, timeline_id = op_persist.timeline_id, "couldn't find timeline for corresponding pull op" ); continue; }; timeline_persist .sk_set .iter() .filter_map(|sk_id| { let other_node_id = NodeId(*sk_id as u64); if node_id == other_node_id { // We obviously don't want to pull from ourselves return None; } let Some(sk) = safekeepers.get(&other_node_id) else { tracing::warn!( "couldn't find safekeeper with pending op id {other_node_id}, not pulling from it" ); return None; }; Some((other_node_id, sk.base_url())) }) .collect::<Vec<_>>() } }; let req = ScheduleRequest { safekeeper: sk, host_list, tenant_id, timeline_id, generation: op_persist.generation as u32, kind: op_persist.op_kind, }; res.push(req); } Ok(res) } pub(crate) struct ScheduleRequest { pub(crate) safekeeper: Box<Safekeeper>, pub(crate) host_list: Vec<(NodeId, String)>, pub(crate) tenant_id: TenantId, pub(crate) timeline_id: Option<TimelineId>, pub(crate) generation: u32, pub(crate) kind: SafekeeperTimelineOpKind, } /// A way to keep ongoing/queued reconcile requests apart #[derive(Copy, Clone, PartialEq, Eq)] struct TokenId(u64); type OngoingTokens = ClashMap<(TenantId, Option<TimelineId>), (CancellationToken, TokenId)>; /// Handle to per safekeeper reconciler. struct ReconcilerHandle { tx: UnboundedSender<(ScheduleRequest, CancellationToken, TokenId)>, ongoing_tokens: Arc<OngoingTokens>, token_id_counter: AtomicU64, cancel: CancellationToken, } impl ReconcilerHandle { /// Obtain a new token slot, cancelling any existing reconciliations for /// that timeline. It is not useful to have >1 operation per <tenant_id, /// timeline_id, safekeeper>, hence scheduling op cancels current one if it /// exists. fn new_token_slot( &self, tenant_id: TenantId, timeline_id: Option<TimelineId>, ) -> (CancellationToken, TokenId) { let token_id = self .token_id_counter .fetch_add(1, std::sync::atomic::Ordering::Relaxed); let token_id = TokenId(token_id); let entry = self.ongoing_tokens.entry((tenant_id, timeline_id)); if let Entry::Occupied(entry) = &entry { let (cancel, _) = entry.get(); cancel.cancel(); } entry.insert((self.cancel.child_token(), token_id)).clone() } /// Cancel an ongoing reconciliation fn cancel_reconciliation(&self, tenant_id: TenantId, timeline_id: Option<TimelineId>) { if let Some((_, (cancel, _id))) = self.ongoing_tokens.remove(&(tenant_id, timeline_id)) { cancel.cancel(); } } fn schedule_reconcile(&self, req: ScheduleRequest) { let (cancel, token_id) = self.new_token_slot(req.tenant_id, req.timeline_id); let hostname = req.safekeeper.skp.host.clone(); let sk_az = req.safekeeper.skp.availability_zone_id.clone(); let sk_node_id = req.safekeeper.get_id().to_string(); // We don't have direct access to the queue depth here, so increase it blindly by 1. // We know that putting into the queue increases the queue depth. The receiver will // update with the correct value once it processes the next item. To avoid races where we // reduce before we increase, leaving the gauge with a 1 value for a long time, we // increase it before putting into the queue. let queued_gauge = &METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_reconciles_queued; let label_group = SafekeeperReconcilerLabelGroup { sk_az: &sk_az, sk_node_id: &sk_node_id, sk_hostname: &hostname, }; queued_gauge.inc(label_group.clone()); if let Err(err) = self.tx.send((req, cancel, token_id)) { queued_gauge.set(label_group, 0); tracing::info!("scheduling request onto {hostname} returned error: {err}"); } } } pub(crate) struct SafekeeperReconciler { inner: SafekeeperReconcilerInner, concurrency_limiter: Arc<Semaphore>, rx: UnboundedReceiver<(ScheduleRequest, CancellationToken, TokenId)>, cancel: CancellationToken, } /// Thin wrapper over `Service` to not clutter its inherent functions #[derive(Clone)] struct SafekeeperReconcilerInner { ongoing_tokens: Arc<OngoingTokens>, service: Arc<Service>, } impl SafekeeperReconciler { fn spawn(cancel: CancellationToken, service: Arc<Service>) -> ReconcilerHandle { // We hold the ServiceInner lock so we don't want to make sending to the reconciler channel to be blocking. let (tx, rx) = mpsc::unbounded_channel(); let concurrency = service.config.safekeeper_reconciler_concurrency; let ongoing_tokens = Arc::new(ClashMap::new()); let mut reconciler = SafekeeperReconciler { inner: SafekeeperReconcilerInner { service, ongoing_tokens: ongoing_tokens.clone(), }, rx, concurrency_limiter: Arc::new(Semaphore::new(concurrency)), cancel: cancel.clone(), }; let handle = ReconcilerHandle { tx, ongoing_tokens, token_id_counter: AtomicU64::new(0), cancel, }; tokio::spawn(async move { reconciler.run().await }); handle } async fn run(&mut self) { loop { let req = tokio::select! { req = self.rx.recv() => req, _ = self.cancel.cancelled() => break, }; let Some((req, req_cancel, req_token_id)) = req else { break; }; let permit_res = tokio::select! { req = self.concurrency_limiter.clone().acquire_owned() => req, _ = self.cancel.cancelled() => break, }; let Ok(_permit) = permit_res else { return }; let inner = self.inner.clone(); if req_cancel.is_cancelled() { continue; } let queued_gauge = &METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_reconciles_queued; queued_gauge.set( SafekeeperReconcilerLabelGroup { sk_az: &req.safekeeper.skp.availability_zone_id, sk_node_id: &req.safekeeper.get_id().to_string(), sk_hostname: &req.safekeeper.skp.host, }, self.rx.len() as i64, ); tokio::task::spawn(async move { let kind = req.kind; let tenant_id = req.tenant_id; let timeline_id = req.timeline_id; let node_id = req.safekeeper.skp.id; inner .reconcile_one(req, req_cancel, req_token_id) .instrument(tracing::info_span!( "reconcile_one", ?kind, %tenant_id, ?timeline_id, %node_id, )) .await; }); } } } impl SafekeeperReconcilerInner { async fn reconcile_one( &self, req: ScheduleRequest, req_cancel: CancellationToken, req_token_id: TokenId, ) { let req_host = req.safekeeper.skp.host.clone(); let success; match req.kind { SafekeeperTimelineOpKind::Pull => { let Some(timeline_id) = req.timeline_id else { tracing::warn!( "ignoring invalid schedule request: timeline_id is empty for `pull`" ); return; }; let our_id = req.safekeeper.get_id(); let http_hosts = req .host_list .iter() .filter(|(node_id, _hostname)| *node_id != our_id) .map(|(_, hostname)| hostname.clone()) .collect::<Vec<_>>(); let pull_req = PullTimelineRequest { http_hosts, tenant_id: req.tenant_id, timeline_id, // TODO(diko): get mconf from "timelines" table and pass it here. // Now we use pull_timeline reconciliation only for the timeline creation, // so it's not critical right now. // It could be fixed together with other reconciliation issues: // https://github.com/neondatabase/neon/issues/12189 mconf: None, }; success = self .reconcile_inner( &req, async |client| client.pull_timeline(&pull_req).await, |resp| { if let Some(host) = resp.safekeeper_host { tracing::info!("pulled timeline from {host} onto {req_host}"); } else { tracing::info!( "timeline already present on safekeeper on {req_host}" ); } }, req_cancel, ) .await; } SafekeeperTimelineOpKind::Exclude => { // TODO actually exclude instead of delete here let tenant_id = req.tenant_id; let Some(timeline_id) = req.timeline_id else { tracing::warn!( "ignoring invalid schedule request: timeline_id is empty for `exclude`" ); return; }; success = self .reconcile_inner( &req, async |client| client.delete_timeline(tenant_id, timeline_id).await, |_resp| { tracing::info!("deleted timeline from {req_host}"); }, req_cancel, ) .await; } SafekeeperTimelineOpKind::Delete => { let tenant_id = req.tenant_id; if let Some(timeline_id) = req.timeline_id { success = self .reconcile_inner( &req, async |client| client.delete_timeline(tenant_id, timeline_id).await, |_resp| { tracing::info!("deleted timeline from {req_host}"); }, req_cancel, ) .await; if success { self.delete_timeline_from_db(tenant_id, timeline_id).await; } } else { success = self .reconcile_inner( &req, async |client| client.delete_tenant(tenant_id).await, |_resp| { tracing::info!(%tenant_id, "deleted tenant from {req_host}"); }, req_cancel, ) .await; if success { self.delete_tenant_timelines_from_db(tenant_id).await; } } } } if success { self.ongoing_tokens.remove_if( &(req.tenant_id, req.timeline_id), |_ttid, (_cancel, token_id)| { // Ensure that this request is indeed the request we just finished and not a new one req_token_id == *token_id }, ); } } async fn delete_timeline_from_db(&self, tenant_id: TenantId, timeline_id: TimelineId) { match self .service .persistence .list_pending_ops_for_timeline(tenant_id, timeline_id) .await { Ok(list) => { if !list.is_empty() { // duplicate the timeline_id here because it might be None in the reconcile context tracing::info!(%timeline_id, "not deleting timeline from db as there is {} open reconciles", list.len()); return; } } Err(e) => { tracing::warn!(%timeline_id, "couldn't query pending ops: {e}"); return; } } tracing::info!(%tenant_id, %timeline_id, "deleting timeline from db after all reconciles succeeded"); // In theory we could crash right after deleting the op from the db and right before reaching this, // but then we'll boot up with a timeline that has deleted_at set, so hopefully we'll issue deletion ops for it again. if let Err(err) = self .service .persistence .delete_timeline(tenant_id, timeline_id) .await { tracing::warn!(%tenant_id, %timeline_id, "couldn't delete timeline from db: {err}"); } } async fn delete_tenant_timelines_from_db(&self, tenant_id: TenantId) { let timeline_list = match self .service .persistence .list_timelines_for_tenant(tenant_id) .await { Ok(timeline_list) => timeline_list, Err(e) => { tracing::warn!(%tenant_id, "couldn't query timelines: {e}"); return; } }; for timeline in timeline_list { let Ok(timeline_id) = TimelineId::from_str(&timeline.timeline_id) else { tracing::warn!("Invalid timeline ID in database {}", timeline.timeline_id); continue; }; self.delete_timeline_from_db(tenant_id, timeline_id).await; } } /// Returns whether the reconciliation happened successfully (or we got cancelled) async fn reconcile_inner<T, F, U>( &self, req: &ScheduleRequest, closure: impl Fn(SafekeeperClient) -> F, log_success: impl FnOnce(T) -> U, req_cancel: CancellationToken, ) -> bool where F: Future<Output = Result<T, safekeeper_client::mgmt_api::Error>>, { let jwt = self .service .config .safekeeper_jwt_token .clone() .map(SecretString::from); loop { let res = req .safekeeper .with_client_retries( |client| { let closure = &closure; async move { closure(client).await } }, self.service.get_http_client(), &jwt, 3, 10, Duration::from_secs(10), &req_cancel, ) .await; match res { Ok(resp) => { log_success(resp); let res = self .service .persistence .remove_pending_op( req.tenant_id, req.timeline_id, req.safekeeper.get_id(), req.generation, ) .await; let complete_counter = &METRICS_REGISTRY .metrics_group .storage_controller_safekeeper_reconciles_complete; complete_counter.inc(SafekeeperReconcilerLabelGroup { sk_az: &req.safekeeper.skp.availability_zone_id, sk_node_id: &req.safekeeper.get_id().to_string(), sk_hostname: &req.safekeeper.skp.host, }); if let Err(err) = res { tracing::info!( "couldn't remove reconciliation request onto {} from persistence: {err:?}", req.safekeeper.skp.host ); } return true; } Err(mgmt_api::Error::Cancelled) => { // On cancellation, the code that issued it will take care of removing db entries (if needed) return false; } Err(e) => { tracing::info!( "Reconcile attempt for safekeeper {} failed, retrying after sleep: {e:?}", req.safekeeper.skp.host ); const SLEEP_TIME: Duration = Duration::from_secs(1); tokio::time::sleep(SLEEP_TIME).await; } } } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/src/persistence/split_state.rs
storage_controller/src/persistence/split_state.rs
use diesel::deserialize::{FromSql, FromSqlRow}; use diesel::expression::AsExpression; use diesel::pg::{Pg, PgValue}; use diesel::serialize::ToSql; use diesel::sql_types::Int2; use serde::{Deserialize, Serialize}; #[derive(Debug, Clone, Copy, PartialEq, Eq, Hash, PartialOrd, Ord, FromSqlRow, AsExpression)] #[diesel(sql_type = SplitStateSQLRepr)] #[derive(Deserialize, Serialize)] pub enum SplitState { Idle = 0, Splitting = 1, } impl Default for SplitState { fn default() -> Self { Self::Idle } } type SplitStateSQLRepr = Int2; impl ToSql<SplitStateSQLRepr, Pg> for SplitState { fn to_sql<'a>( &'a self, out: &'a mut diesel::serialize::Output<Pg>, ) -> diesel::serialize::Result { let raw_value: i16 = *self as i16; let mut new_out = out.reborrow(); ToSql::<SplitStateSQLRepr, Pg>::to_sql(&raw_value, &mut new_out) } } impl FromSql<SplitStateSQLRepr, Pg> for SplitState { fn from_sql(pg_value: PgValue) -> diesel::deserialize::Result<Self> { match FromSql::<SplitStateSQLRepr, Pg>::from_sql(pg_value).map(|v| match v { 0 => Some(Self::Idle), 1 => Some(Self::Splitting), _ => None, })? { Some(v) => Ok(v), None => Err(format!("Invalid SplitState value, was: {:?}", pg_value.as_bytes()).into()), } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/client/src/lib.rs
storage_controller/client/src/lib.rs
pub mod control_api;
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/storage_controller/client/src/control_api.rs
storage_controller/client/src/control_api.rs
use pageserver_client::mgmt_api::{self, ResponseErrorMessageExt}; use reqwest::{Method, Url}; use serde::Serialize; use serde::de::DeserializeOwned; pub struct Client { base_url: Url, jwt_token: Option<String>, client: reqwest::Client, } impl Client { pub fn new(http_client: reqwest::Client, base_url: Url, jwt_token: Option<String>) -> Self { Self { base_url, jwt_token, client: http_client, } } /// Simple HTTP request wrapper for calling into storage controller pub async fn dispatch<RQ, RS>( &self, method: Method, path: String, body: Option<RQ>, ) -> mgmt_api::Result<RS> where RQ: Serialize + Sized, RS: DeserializeOwned + Sized, { let request_path = self .base_url .join(&path) .expect("Failed to build request path"); let mut builder = self.client.request(method, request_path); if let Some(body) = body { builder = builder.json(&body) } if let Some(jwt_token) = &self.jwt_token { builder = builder.header( reqwest::header::AUTHORIZATION, format!("Bearer {jwt_token}"), ); } let response = builder.send().await.map_err(mgmt_api::Error::ReceiveBody)?; let response = response.error_from_body().await?; response .json() .await .map_err(pageserver_client::mgmt_api::Error::ReceiveBody) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/config.rs
proxy/src/config.rs
use std::str::FromStr; use std::sync::Arc; use std::time::Duration; use anyhow::{Context, Ok, bail, ensure}; use arc_swap::ArcSwapOption; use camino::{Utf8Path, Utf8PathBuf}; use clap::ValueEnum; use compute_api::spec::LocalProxySpec; use remote_storage::RemoteStorageConfig; use thiserror::Error; use tokio::sync::Notify; use tracing::{debug, error, info, warn}; use crate::auth::backend::jwt::JwkCache; use crate::auth::backend::local::JWKS_ROLE_MAP; use crate::control_plane::locks::ApiLocks; use crate::control_plane::messages::{EndpointJwksResponse, JwksSettings}; use crate::ext::TaskExt; use crate::intern::RoleNameInt; use crate::rate_limiter::{RateLimitAlgorithm, RateLimiterConfig}; use crate::scram; use crate::serverless::GlobalConnPoolOptions; use crate::serverless::cancel_set::CancelSet; #[cfg(feature = "rest_broker")] use crate::serverless::rest::DbSchemaCache; pub use crate::tls::server_config::{TlsConfig, configure_tls}; use crate::types::{Host, RoleName}; pub struct ProxyConfig { pub tls_config: ArcSwapOption<TlsConfig>, pub metric_collection: Option<MetricCollectionConfig>, pub http_config: HttpConfig, pub authentication_config: AuthenticationConfig, #[cfg(feature = "rest_broker")] pub rest_config: RestConfig, pub proxy_protocol_v2: ProxyProtocolV2, pub handshake_timeout: Duration, pub wake_compute_retry_config: RetryConfig, pub connect_compute_locks: ApiLocks<Host>, pub connect_to_compute: ComputeConfig, pub greetings: String, // Greeting message sent to the client after connection establishment and contains session_id. #[cfg(feature = "testing")] pub disable_pg_session_jwt: bool, } pub struct ComputeConfig { pub retry: RetryConfig, pub tls: Arc<rustls::ClientConfig>, pub timeout: Duration, } #[derive(Copy, Clone, Debug, ValueEnum, PartialEq)] pub enum ProxyProtocolV2 { /// Connection will error if PROXY protocol v2 header is missing Required, /// Connection will error if PROXY protocol v2 header is provided Rejected, } #[derive(Debug)] pub struct MetricCollectionConfig { pub endpoint: reqwest::Url, pub interval: Duration, pub backup_metric_collection_config: MetricBackupCollectionConfig, } pub struct HttpConfig { pub accept_websockets: bool, pub pool_options: GlobalConnPoolOptions, pub cancel_set: CancelSet, pub client_conn_threshold: u64, pub max_request_size_bytes: usize, pub max_response_size_bytes: usize, } pub struct AuthenticationConfig { pub scram_thread_pool: Arc<scram::threadpool::ThreadPool>, pub scram_protocol_timeout: tokio::time::Duration, pub ip_allowlist_check_enabled: bool, pub is_vpc_acccess_proxy: bool, pub jwks_cache: JwkCache, pub is_auth_broker: bool, pub accept_jwts: bool, pub console_redirect_confirmation_timeout: tokio::time::Duration, } #[cfg(feature = "rest_broker")] pub struct RestConfig { pub is_rest_broker: bool, pub db_schema_cache: Option<DbSchemaCache>, pub max_schema_size: usize, pub hostname_prefix: String, } #[derive(Debug)] pub struct MetricBackupCollectionConfig { pub remote_storage_config: Option<RemoteStorageConfig>, pub chunk_size: usize, } pub fn remote_storage_from_toml(s: &str) -> anyhow::Result<RemoteStorageConfig> { RemoteStorageConfig::from_toml(&s.parse()?) } /// Helper for cmdline cache options parsing. #[derive(Debug)] pub struct CacheOptions { /// Max number of entries. pub size: Option<u64>, /// Entry's time-to-live. pub absolute_ttl: Option<Duration>, /// Entry's time-to-idle. pub idle_ttl: Option<Duration>, } impl CacheOptions { /// Default options for [`crate::cache::node_info::NodeInfoCache`]. pub const CACHE_DEFAULT_OPTIONS: &'static str = "size=4000,idle_ttl=4m"; /// Parse cache options passed via cmdline. /// Example: [`Self::CACHE_DEFAULT_OPTIONS`]. fn parse(options: &str) -> anyhow::Result<Self> { let mut size = None; let mut absolute_ttl = None; let mut idle_ttl = None; for option in options.split(',') { let (key, value) = option .split_once('=') .with_context(|| format!("bad key-value pair: {option}"))?; match key { "size" => size = Some(value.parse()?), "absolute_ttl" | "ttl" => absolute_ttl = Some(humantime::parse_duration(value)?), "idle_ttl" | "tti" => idle_ttl = Some(humantime::parse_duration(value)?), unknown => bail!("unknown key: {unknown}"), } } Ok(Self { size, absolute_ttl, idle_ttl, }) } pub fn moka<K, V, C>( &self, mut builder: moka::sync::CacheBuilder<K, V, C>, ) -> moka::sync::CacheBuilder<K, V, C> { if let Some(size) = self.size { builder = builder.max_capacity(size); } if let Some(ttl) = self.absolute_ttl { builder = builder.time_to_live(ttl); } if let Some(tti) = self.idle_ttl { builder = builder.time_to_idle(tti); } builder } } impl FromStr for CacheOptions { type Err = anyhow::Error; fn from_str(options: &str) -> Result<Self, Self::Err> { let error = || format!("failed to parse cache options '{options}'"); Self::parse(options).with_context(error) } } /// Helper for cmdline cache options parsing. #[derive(Debug)] pub struct ProjectInfoCacheOptions { /// Max number of entries. pub size: u64, /// Entry's time-to-live. pub ttl: Duration, /// Max number of roles per endpoint. pub max_roles: u64, /// Gc interval. pub gc_interval: Duration, } impl ProjectInfoCacheOptions { /// Default options for [`crate::cache::project_info::ProjectInfoCache`]. pub const CACHE_DEFAULT_OPTIONS: &'static str = "size=10000,ttl=4m,max_roles=10,gc_interval=60m"; /// Parse cache options passed via cmdline. /// Example: [`Self::CACHE_DEFAULT_OPTIONS`]. fn parse(options: &str) -> anyhow::Result<Self> { let mut size = None; let mut ttl = None; let mut max_roles = None; let mut gc_interval = None; for option in options.split(',') { let (key, value) = option .split_once('=') .with_context(|| format!("bad key-value pair: {option}"))?; match key { "size" => size = Some(value.parse()?), "ttl" => ttl = Some(humantime::parse_duration(value)?), "max_roles" => max_roles = Some(value.parse()?), "gc_interval" => gc_interval = Some(humantime::parse_duration(value)?), unknown => bail!("unknown key: {unknown}"), } } // TTL doesn't matter if cache is always empty. if let Some(0) = size { ttl.get_or_insert(Duration::default()); } Ok(Self { size: size.context("missing `size`")?, ttl: ttl.context("missing `ttl`")?, max_roles: max_roles.context("missing `max_roles`")?, gc_interval: gc_interval.context("missing `gc_interval`")?, }) } } impl FromStr for ProjectInfoCacheOptions { type Err = anyhow::Error; fn from_str(options: &str) -> Result<Self, Self::Err> { let error = || format!("failed to parse cache options '{options}'"); Self::parse(options).with_context(error) } } /// This is a config for connect to compute and wake compute. #[derive(Clone, Copy, Debug)] pub struct RetryConfig { /// Number of times we should retry. pub max_retries: u32, /// Retry duration is base_delay * backoff_factor ^ n, where n starts at 0 pub base_delay: tokio::time::Duration, /// Exponential base for retry wait duration pub backoff_factor: f64, } impl RetryConfig { // Default options for RetryConfig. /// Total delay for 5 retries with 200ms base delay and 2 backoff factor is about 6s. pub const CONNECT_TO_COMPUTE_DEFAULT_VALUES: &'static str = "num_retries=5,base_retry_wait_duration=200ms,retry_wait_exponent_base=2"; /// Total delay for 8 retries with 100ms base delay and 1.6 backoff factor is about 7s. /// Cplane has timeout of 60s on each request. 8m7s in total. pub const WAKE_COMPUTE_DEFAULT_VALUES: &'static str = "num_retries=8,base_retry_wait_duration=100ms,retry_wait_exponent_base=1.6"; /// Parse retry options passed via cmdline. /// Example: [`Self::CONNECT_TO_COMPUTE_DEFAULT_VALUES`]. pub fn parse(options: &str) -> anyhow::Result<Self> { let mut num_retries = None; let mut base_retry_wait_duration = None; let mut retry_wait_exponent_base = None; for option in options.split(',') { let (key, value) = option .split_once('=') .with_context(|| format!("bad key-value pair: {option}"))?; match key { "num_retries" => num_retries = Some(value.parse()?), "base_retry_wait_duration" => { base_retry_wait_duration = Some(humantime::parse_duration(value)?); } "retry_wait_exponent_base" => retry_wait_exponent_base = Some(value.parse()?), unknown => bail!("unknown key: {unknown}"), } } Ok(Self { max_retries: num_retries.context("missing `num_retries`")?, base_delay: base_retry_wait_duration.context("missing `base_retry_wait_duration`")?, backoff_factor: retry_wait_exponent_base .context("missing `retry_wait_exponent_base`")?, }) } } /// Helper for cmdline cache options parsing. #[derive(serde::Deserialize)] pub struct ConcurrencyLockOptions { /// The number of shards the lock map should have pub shards: usize, /// The number of allowed concurrent requests for each endpoitn #[serde(flatten)] pub limiter: RateLimiterConfig, /// Garbage collection epoch #[serde(deserialize_with = "humantime_serde::deserialize")] pub epoch: Duration, /// Lock timeout #[serde(deserialize_with = "humantime_serde::deserialize")] pub timeout: Duration, } impl ConcurrencyLockOptions { /// Default options for [`crate::control_plane::client::ApiLocks`]. pub const DEFAULT_OPTIONS_WAKE_COMPUTE_LOCK: &'static str = "permits=0"; /// Default options for [`crate::control_plane::client::ApiLocks`]. pub const DEFAULT_OPTIONS_CONNECT_COMPUTE_LOCK: &'static str = "shards=64,permits=100,epoch=10m,timeout=10ms"; // pub const DEFAULT_OPTIONS_WAKE_COMPUTE_LOCK: &'static str = "shards=32,permits=4,epoch=10m,timeout=1s"; /// Parse lock options passed via cmdline. /// Example: [`Self::DEFAULT_OPTIONS_WAKE_COMPUTE_LOCK`]. fn parse(options: &str) -> anyhow::Result<Self> { let options = options.trim(); if options.starts_with('{') && options.ends_with('}') { return Ok(serde_json::from_str(options)?); } let mut shards = None; let mut permits = None; let mut epoch = None; let mut timeout = None; for option in options.split(',') { let (key, value) = option .split_once('=') .with_context(|| format!("bad key-value pair: {option}"))?; match key { "shards" => shards = Some(value.parse()?), "permits" => permits = Some(value.parse()?), "epoch" => epoch = Some(humantime::parse_duration(value)?), "timeout" => timeout = Some(humantime::parse_duration(value)?), unknown => bail!("unknown key: {unknown}"), } } // these dont matter if lock is disabled if let Some(0) = permits { timeout = Some(Duration::default()); epoch = Some(Duration::default()); shards = Some(2); } let permits = permits.context("missing `permits`")?; let out = Self { shards: shards.context("missing `shards`")?, limiter: RateLimiterConfig { algorithm: RateLimitAlgorithm::Fixed, initial_limit: permits, }, epoch: epoch.context("missing `epoch`")?, timeout: timeout.context("missing `timeout`")?, }; ensure!(out.shards > 1, "shard count must be > 1"); ensure!( out.shards.is_power_of_two(), "shard count must be a power of two" ); Ok(out) } } impl FromStr for ConcurrencyLockOptions { type Err = anyhow::Error; fn from_str(options: &str) -> Result<Self, Self::Err> { let error = || format!("failed to parse cache lock options '{options}'"); Self::parse(options).with_context(error) } } #[derive(Error, Debug)] pub(crate) enum RefreshConfigError { #[error(transparent)] Read(#[from] std::io::Error), #[error(transparent)] Parse(#[from] serde_json::Error), #[error(transparent)] Validate(anyhow::Error), #[error(transparent)] Tls(anyhow::Error), } pub(crate) async fn refresh_config_loop(config: &ProxyConfig, path: Utf8PathBuf, rx: Arc<Notify>) { let mut init = true; loop { rx.notified().await; match refresh_config_inner(config, &path).await { std::result::Result::Ok(()) => {} // don't log for file not found errors if this is the first time we are checking // for computes that don't use local_proxy, this is not an error. Err(RefreshConfigError::Read(e)) if init && e.kind() == std::io::ErrorKind::NotFound => { debug!(error=?e, ?path, "could not read config file"); } Err(RefreshConfigError::Tls(e)) => { error!(error=?e, ?path, "could not read TLS certificates"); } Err(e) => { error!(error=?e, ?path, "could not read config file"); } } init = false; } } pub(crate) async fn refresh_config_inner( config: &ProxyConfig, path: &Utf8Path, ) -> Result<(), RefreshConfigError> { let bytes = tokio::fs::read(&path).await?; let data: LocalProxySpec = serde_json::from_slice(&bytes)?; let mut jwks_set = vec![]; fn parse_jwks_settings(jwks: compute_api::spec::JwksSettings) -> anyhow::Result<JwksSettings> { let mut jwks_url = url::Url::from_str(&jwks.jwks_url).context("parsing JWKS url")?; ensure!( jwks_url.has_authority() && (jwks_url.scheme() == "http" || jwks_url.scheme() == "https"), "Invalid JWKS url. Must be HTTP", ); ensure!( jwks_url.host().is_some_and(|h| h != url::Host::Domain("")), "Invalid JWKS url. No domain listed", ); // clear username, password and ports jwks_url .set_username("") .expect("url can be a base and has a valid host and is not a file. should not error"); jwks_url .set_password(None) .expect("url can be a base and has a valid host and is not a file. should not error"); // local testing is hard if we need to have a specific restricted port if cfg!(not(feature = "testing")) { jwks_url.set_port(None).expect( "url can be a base and has a valid host and is not a file. should not error", ); } // clear query params jwks_url.set_fragment(None); jwks_url.query_pairs_mut().clear().finish(); if jwks_url.scheme() != "https" { // local testing is hard if we need to set up https support. if cfg!(not(feature = "testing")) { jwks_url .set_scheme("https") .expect("should not error to set the scheme to https if it was http"); } else { warn!(scheme = jwks_url.scheme(), "JWKS url is not HTTPS"); } } Ok(JwksSettings { id: jwks.id, jwks_url, _provider_name: jwks.provider_name, jwt_audience: jwks.jwt_audience, role_names: jwks .role_names .into_iter() .map(RoleName::from) .map(|s| RoleNameInt::from(&s)) .collect(), }) } for jwks in data.jwks.into_iter().flatten() { jwks_set.push(parse_jwks_settings(jwks).map_err(RefreshConfigError::Validate)?); } info!("successfully loaded new config"); JWKS_ROLE_MAP.store(Some(Arc::new(EndpointJwksResponse { jwks: jwks_set }))); if let Some(tls_config) = data.tls { let tls_config = tokio::task::spawn_blocking(move || { crate::tls::server_config::configure_tls( tls_config.key_path.as_ref(), tls_config.cert_path.as_ref(), None, false, ) }) .await .propagate_task_panic() .map_err(RefreshConfigError::Tls)?; config.tls_config.store(Some(Arc::new(tls_config))); } std::result::Result::Ok(()) } #[cfg(test)] mod tests { use super::*; use crate::rate_limiter::Aimd; #[test] fn test_parse_cache_options() -> anyhow::Result<()> { let CacheOptions { size, absolute_ttl, idle_ttl: _, } = "size=4096,ttl=5min".parse()?; assert_eq!(size, Some(4096)); assert_eq!(absolute_ttl, Some(Duration::from_secs(5 * 60))); let CacheOptions { size, absolute_ttl, idle_ttl: _, } = "ttl=4m,size=2".parse()?; assert_eq!(size, Some(2)); assert_eq!(absolute_ttl, Some(Duration::from_secs(4 * 60))); let CacheOptions { size, absolute_ttl, idle_ttl: _, } = "size=0,ttl=1s".parse()?; assert_eq!(size, Some(0)); assert_eq!(absolute_ttl, Some(Duration::from_secs(1))); let CacheOptions { size, absolute_ttl, idle_ttl: _, } = "size=0".parse()?; assert_eq!(size, Some(0)); assert_eq!(absolute_ttl, None); Ok(()) } #[test] fn test_parse_lock_options() -> anyhow::Result<()> { let ConcurrencyLockOptions { epoch, limiter, shards, timeout, } = "shards=32,permits=4,epoch=10m,timeout=1s".parse()?; assert_eq!(epoch, Duration::from_secs(10 * 60)); assert_eq!(timeout, Duration::from_secs(1)); assert_eq!(shards, 32); assert_eq!(limiter.initial_limit, 4); assert_eq!(limiter.algorithm, RateLimitAlgorithm::Fixed); let ConcurrencyLockOptions { epoch, limiter, shards, timeout, } = "epoch=60s,shards=16,timeout=100ms,permits=8".parse()?; assert_eq!(epoch, Duration::from_secs(60)); assert_eq!(timeout, Duration::from_millis(100)); assert_eq!(shards, 16); assert_eq!(limiter.initial_limit, 8); assert_eq!(limiter.algorithm, RateLimitAlgorithm::Fixed); let ConcurrencyLockOptions { epoch, limiter, shards, timeout, } = "permits=0".parse()?; assert_eq!(epoch, Duration::ZERO); assert_eq!(timeout, Duration::ZERO); assert_eq!(shards, 2); assert_eq!(limiter.initial_limit, 0); assert_eq!(limiter.algorithm, RateLimitAlgorithm::Fixed); Ok(()) } #[test] fn test_parse_json_lock_options() -> anyhow::Result<()> { let ConcurrencyLockOptions { epoch, limiter, shards, timeout, } = r#"{"shards":32,"initial_limit":44,"aimd":{"min":5,"max":500,"inc":10,"dec":0.9,"utilisation":0.8},"epoch":"10m","timeout":"1s"}"# .parse()?; assert_eq!(epoch, Duration::from_secs(10 * 60)); assert_eq!(timeout, Duration::from_secs(1)); assert_eq!(shards, 32); assert_eq!(limiter.initial_limit, 44); assert_eq!( limiter.algorithm, RateLimitAlgorithm::Aimd { conf: Aimd { min: 5, max: 500, dec: 0.9, inc: 10, utilisation: 0.8 } }, ); Ok(()) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/stream.rs
proxy/src/stream.rs
use std::pin::Pin; use std::sync::Arc; use std::{io, task}; use rustls::ServerConfig; use thiserror::Error; use tokio::io::{AsyncRead, AsyncWrite, AsyncWriteExt, ReadBuf}; use tokio_rustls::server::TlsStream; use crate::error::{ErrorKind, ReportableError, UserFacingError}; use crate::metrics::Metrics; use crate::pqproto::{ BeMessage, FE_PASSWORD_MESSAGE, FeStartupPacket, SQLSTATE_INTERNAL_ERROR, WriteBuf, read_message, read_startup, }; use crate::tls::TlsServerEndPoint; /// Stream wrapper which implements libpq's protocol. /// /// NOTE: This object deliberately doesn't implement [`AsyncRead`] /// or [`AsyncWrite`] to prevent subtle errors (e.g. trying /// to pass random malformed bytes through the connection). pub struct PqStream<S> { stream: S, read: Vec<u8>, write: WriteBuf, } impl<S> PqStream<S> { pub fn get_ref(&self) -> &S { &self.stream } /// Construct a new libpq protocol wrapper over a stream without the first startup message. #[cfg(test)] pub fn new_skip_handshake(stream: S) -> Self { Self { stream, read: Vec::new(), write: WriteBuf::new(), } } } impl<S: AsyncRead + AsyncWrite + Unpin> PqStream<S> { /// Construct a new libpq protocol wrapper and read the first startup message. /// /// This is not cancel safe. pub async fn parse_startup(mut stream: S) -> io::Result<(Self, FeStartupPacket)> { let startup = read_startup(&mut stream).await?; Ok(( Self { stream, read: Vec::new(), write: WriteBuf::new(), }, startup, )) } /// Tell the client that encryption is not supported. /// /// This is not cancel safe pub async fn reject_encryption(&mut self) -> io::Result<FeStartupPacket> { // N for No. self.write.encryption(b'N'); self.flush().await?; read_startup(&mut self.stream).await } } impl<S: AsyncRead + Unpin> PqStream<S> { /// Read a raw postgres packet, which will respect the max length requested. /// This is not cancel safe. async fn read_raw_expect(&mut self, tag: u8, max: u32) -> io::Result<&mut [u8]> { let (actual_tag, msg) = read_message(&mut self.stream, &mut self.read, max).await?; if actual_tag != tag { return Err(io::Error::other(format!( "incorrect message tag, expected {:?}, got {:?}", tag as char, actual_tag as char, ))); } Ok(msg) } /// Read a postgres password message, which will respect the max length requested. /// This is not cancel safe. pub async fn read_password_message(&mut self) -> io::Result<&mut [u8]> { // passwords are usually pretty short // and SASL SCRAM messages are no longer than 256 bytes in my testing // (a few hashes and random bytes, encoded into base64). const MAX_PASSWORD_LENGTH: u32 = 512; self.read_raw_expect(FE_PASSWORD_MESSAGE, MAX_PASSWORD_LENGTH) .await } } #[derive(Debug)] pub struct ReportedError { source: anyhow::Error, error_kind: ErrorKind, } impl ReportedError { pub fn new(e: impl UserFacingError + Into<anyhow::Error>) -> Self { let error_kind = e.get_error_kind(); Self { source: e.into(), error_kind, } } } impl std::fmt::Display for ReportedError { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { self.source.fmt(f) } } impl std::error::Error for ReportedError { fn source(&self) -> Option<&(dyn std::error::Error + 'static)> { self.source.source() } } impl ReportableError for ReportedError { fn get_error_kind(&self) -> ErrorKind { self.error_kind } } impl<S: AsyncWrite + Unpin> PqStream<S> { /// Tell the client that we are willing to accept SSL. /// This is not cancel safe pub async fn accept_tls(mut self) -> io::Result<S> { // S for SSL. self.write.encryption(b'S'); self.flush().await?; Ok(self.stream) } /// Assert that we are using direct TLS. pub fn accept_direct_tls(self) -> S { self.stream } /// Write a raw message to the internal buffer. pub fn write_raw(&mut self, size_hint: usize, tag: u8, f: impl FnOnce(&mut Vec<u8>)) { self.write.write_raw(size_hint, tag, f); } /// Write the message into an internal buffer pub fn write_message(&mut self, message: BeMessage<'_>) { message.write_message(&mut self.write); } /// Write the buffer to the socket until we have some more space again. pub async fn write_if_full(&mut self) -> io::Result<()> { while self.write.occupied_len() > 2048 { self.stream.write_buf(&mut self.write).await?; } Ok(()) } /// Flush the output buffer into the underlying stream. /// /// This is cancel safe. pub async fn flush(&mut self) -> io::Result<()> { self.stream.write_all_buf(&mut self.write).await?; self.write.reset(); self.stream.flush().await?; Ok(()) } /// Flush the output buffer into the underlying stream. /// /// This is cancel safe. pub async fn flush_and_into_inner(mut self) -> io::Result<S> { self.flush().await?; Ok(self.stream) } /// Write the error message to the client, then re-throw it. /// /// Trait [`UserFacingError`] acts as an allowlist for error types. /// If `ctx` is provided and has testodrome_id set, error messages will be prefixed according to error kind. pub(crate) async fn throw_error<E>( &mut self, error: E, ctx: Option<&crate::context::RequestContext>, ) -> ReportedError where E: UserFacingError + Into<anyhow::Error>, { let error_kind = error.get_error_kind(); let msg = error.to_string_client(); if error_kind != ErrorKind::RateLimit && error_kind != ErrorKind::User { tracing::info!( kind = error_kind.to_metric_label(), msg, "forwarding error to user" ); } let probe_msg; let mut msg = &*msg; if let Some(ctx) = ctx && ctx.get_testodrome_id().is_some() { let tag = match error_kind { ErrorKind::User => "client", ErrorKind::ClientDisconnect => "client", ErrorKind::RateLimit => "proxy", ErrorKind::ServiceRateLimit => "proxy", ErrorKind::Quota => "proxy", ErrorKind::Service => "proxy", ErrorKind::ControlPlane => "controlplane", ErrorKind::Postgres => "other", ErrorKind::Compute => "compute", }; probe_msg = typed_json::json!({ "tag": tag, "msg": msg, "cold_start_info": ctx.cold_start_info(), }) .to_string(); msg = &probe_msg; } // TODO: either preserve the error code from postgres, or assign error codes to proxy errors. self.write.write_error(msg, SQLSTATE_INTERNAL_ERROR); self.flush() .await .unwrap_or_else(|e| tracing::debug!("write_message failed: {e}")); ReportedError::new(error) } } /// Wrapper for upgrading raw streams into secure streams. pub enum Stream<S> { /// We always begin with a raw stream, /// which may then be upgraded into a secure stream. Raw { raw: S }, Tls { /// We box [`TlsStream`] since it can be quite large. tls: Box<TlsStream<S>>, /// Channel binding parameter tls_server_end_point: TlsServerEndPoint, }, } impl<S: Unpin> Unpin for Stream<S> {} impl<S> Stream<S> { /// Construct a new instance from a raw stream. pub fn from_raw(raw: S) -> Self { Self::Raw { raw } } /// Return SNI hostname when it's available. pub fn sni_hostname(&self) -> Option<&str> { match self { Stream::Raw { .. } => None, Stream::Tls { tls, .. } => tls.get_ref().1.server_name(), } } pub(crate) fn tls_server_end_point(&self) -> TlsServerEndPoint { match self { Stream::Raw { .. } => TlsServerEndPoint::Undefined, Stream::Tls { tls_server_end_point, .. } => *tls_server_end_point, } } } #[derive(Debug, Error)] #[error("Can't upgrade TLS stream")] pub enum StreamUpgradeError { #[error("Bad state reached: can't upgrade TLS stream")] AlreadyTls, #[error("Can't upgrade stream: IO error: {0}")] Io(#[from] io::Error), } impl<S: AsyncRead + AsyncWrite + Unpin> Stream<S> { /// If possible, upgrade raw stream into a secure TLS-based stream. pub async fn upgrade( self, cfg: Arc<ServerConfig>, record_handshake_error: bool, ) -> Result<TlsStream<S>, StreamUpgradeError> { match self { Stream::Raw { raw } => Ok(tokio_rustls::TlsAcceptor::from(cfg) .accept(raw) .await .inspect_err(|_| { if record_handshake_error { Metrics::get().proxy.tls_handshake_failures.inc(); } })?), Stream::Tls { .. } => Err(StreamUpgradeError::AlreadyTls), } } } impl<S: AsyncRead + AsyncWrite + Unpin> AsyncRead for Stream<S> { fn poll_read( mut self: Pin<&mut Self>, context: &mut task::Context<'_>, buf: &mut ReadBuf<'_>, ) -> task::Poll<io::Result<()>> { match &mut *self { Self::Raw { raw } => Pin::new(raw).poll_read(context, buf), Self::Tls { tls, .. } => Pin::new(tls).poll_read(context, buf), } } } impl<S: AsyncRead + AsyncWrite + Unpin> AsyncWrite for Stream<S> { fn poll_write( mut self: Pin<&mut Self>, context: &mut task::Context<'_>, buf: &[u8], ) -> task::Poll<io::Result<usize>> { match &mut *self { Self::Raw { raw } => Pin::new(raw).poll_write(context, buf), Self::Tls { tls, .. } => Pin::new(tls).poll_write(context, buf), } } fn poll_flush( mut self: Pin<&mut Self>, context: &mut task::Context<'_>, ) -> task::Poll<io::Result<()>> { match &mut *self { Self::Raw { raw } => Pin::new(raw).poll_flush(context), Self::Tls { tls, .. } => Pin::new(tls).poll_flush(context), } } fn poll_shutdown( mut self: Pin<&mut Self>, context: &mut task::Context<'_>, ) -> task::Poll<io::Result<()>> { match &mut *self { Self::Raw { raw } => Pin::new(raw).poll_shutdown(context), Self::Tls { tls, .. } => Pin::new(tls).poll_shutdown(context), } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/jemalloc.rs
proxy/src/jemalloc.rs
use std::marker::PhantomData; use measured::label::NoLabels; use measured::metric::gauge::GaugeState; use measured::metric::group::Encoding; use measured::metric::name::MetricNameEncoder; use measured::metric::{MetricEncoding, MetricFamilyEncoding, MetricType}; use measured::text::TextEncoder; use measured::{LabelGroup, MetricGroup}; use tikv_jemalloc_ctl::{config, epoch, epoch_mib, stats, version}; pub struct MetricRecorder { epoch: epoch_mib, inner: Metrics, } #[derive(MetricGroup)] struct Metrics { active_bytes: JemallocGaugeFamily<stats::active_mib>, allocated_bytes: JemallocGaugeFamily<stats::allocated_mib>, mapped_bytes: JemallocGaugeFamily<stats::mapped_mib>, metadata_bytes: JemallocGaugeFamily<stats::metadata_mib>, resident_bytes: JemallocGaugeFamily<stats::resident_mib>, retained_bytes: JemallocGaugeFamily<stats::retained_mib>, } impl<Enc: Encoding> MetricGroup<Enc> for MetricRecorder where Metrics: MetricGroup<Enc>, { fn collect_group_into(&self, enc: &mut Enc) -> Result<(), Enc::Err> { if self.epoch.advance().is_ok() { self.inner.collect_group_into(enc)?; } Ok(()) } } impl MetricRecorder { pub fn new() -> Result<Self, anyhow::Error> { tracing::debug!( config = config::malloc_conf::read()?, version = version::read()?, "starting jemalloc recorder" ); Ok(Self { epoch: epoch::mib()?, inner: Metrics { active_bytes: JemallocGaugeFamily(stats::active::mib()?), allocated_bytes: JemallocGaugeFamily(stats::allocated::mib()?), mapped_bytes: JemallocGaugeFamily(stats::mapped::mib()?), metadata_bytes: JemallocGaugeFamily(stats::metadata::mib()?), resident_bytes: JemallocGaugeFamily(stats::resident::mib()?), retained_bytes: JemallocGaugeFamily(stats::retained::mib()?), }, }) } } struct JemallocGauge<T>(PhantomData<T>); impl<T> Default for JemallocGauge<T> { fn default() -> Self { JemallocGauge(PhantomData) } } impl<T> MetricType for JemallocGauge<T> { type Metadata = T; } struct JemallocGaugeFamily<T>(T); impl<M, T: Encoding> MetricFamilyEncoding<T> for JemallocGaugeFamily<M> where JemallocGauge<M>: MetricEncoding<T, Metadata = M>, { fn collect_family_into(&self, name: impl MetricNameEncoder, enc: &mut T) -> Result<(), T::Err> { JemallocGauge::write_type(&name, enc)?; JemallocGauge(PhantomData).collect_into(&self.0, NoLabels, name, enc) } } macro_rules! jemalloc_gauge { ($stat:ident, $mib:ident) => { impl<W: std::io::Write> MetricEncoding<TextEncoder<W>> for JemallocGauge<stats::$mib> { fn write_type( name: impl MetricNameEncoder, enc: &mut TextEncoder<W>, ) -> Result<(), std::io::Error> { GaugeState::write_type(name, enc) } fn collect_into( &self, mib: &stats::$mib, labels: impl LabelGroup, name: impl MetricNameEncoder, enc: &mut TextEncoder<W>, ) -> Result<(), std::io::Error> { if let Ok(v) = mib.read() { GaugeState::new(v as i64).collect_into(&(), labels, name, enc)?; } Ok(()) } } }; } jemalloc_gauge!(active, active_mib); jemalloc_gauge!(allocated, allocated_mib); jemalloc_gauge!(mapped, mapped_mib); jemalloc_gauge!(metadata, metadata_mib); jemalloc_gauge!(resident, resident_mib); jemalloc_gauge!(retained, retained_mib);
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/lib.rs
proxy/src/lib.rs
// rustc lints/lint groups // https://doc.rust-lang.org/rustc/lints/groups.html #![deny(deprecated, future_incompatible, let_underscore, nonstandard_style)] #![warn(clippy::all, clippy::pedantic, clippy::cargo)] // List of denied lints from the clippy::restriction group. // https://rust-lang.github.io/rust-clippy/master/index.html#?groups=restriction #![warn( clippy::undocumented_unsafe_blocks, // TODO: Enable once all individual checks are enabled. //clippy::as_conversions, clippy::dbg_macro, clippy::empty_enum_variants_with_brackets, clippy::exit, clippy::float_cmp_const, clippy::lossy_float_literal, clippy::macro_use_imports, clippy::manual_ok_or, // TODO: consider clippy::map_err_ignore // TODO: consider clippy::mem_forget clippy::rc_mutex, clippy::rest_pat_in_fully_bound_structs, clippy::string_add, clippy::string_to_string, clippy::todo, clippy::unimplemented, clippy::unwrap_used, )] // List of permanently allowed lints. #![allow( // It's ok to cast bool to u8, etc. clippy::cast_lossless, // Seems unavoidable. clippy::multiple_crate_versions, // While #[must_use] is a great feature this check is too noisy. clippy::must_use_candidate, // Inline consts, structs, fns, imports, etc. are ok if they're used by // the following statement(s). clippy::items_after_statements, )] // List of temporarily allowed lints. // TODO: fix code and reduce list or move to permanent list above. #![expect( clippy::cargo_common_metadata, clippy::cast_possible_truncation, clippy::cast_possible_wrap, clippy::cast_precision_loss, clippy::cast_sign_loss, clippy::doc_markdown, clippy::inline_always, clippy::match_same_arms, clippy::match_wild_err_arm, clippy::missing_errors_doc, clippy::missing_panics_doc, clippy::module_name_repetitions, clippy::needless_pass_by_value, clippy::redundant_closure_for_method_calls, clippy::similar_names, clippy::single_match_else, clippy::struct_excessive_bools, clippy::struct_field_names, clippy::too_many_lines, clippy::unused_self )] #![allow( clippy::unsafe_derive_deserialize, reason = "false positive: https://github.com/rust-lang/rust-clippy/issues/15120" )] #![cfg_attr( any(test, feature = "testing"), allow( clippy::needless_raw_string_hashes, clippy::unreadable_literal, clippy::unused_async, ) )] // List of temporarily allowed lints to unblock beta/nightly. #![allow(unknown_lints)] pub mod binary; mod auth; mod batch; mod cache; mod cancellation; mod compute; mod compute_ctl; mod config; mod console_redirect_proxy; mod context; mod control_plane; mod error; mod ext; mod http; mod intern; mod jemalloc; mod logging; mod metrics; mod parse; mod pglb; mod pqproto; mod protocol2; mod proxy; mod rate_limiter; mod redis; mod sasl; mod scram; mod serverless; mod signals; mod stream; mod tls; mod types; mod url; mod usage_metrics; mod util; mod waiters;
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/waiters.rs
proxy/src/waiters.rs
use std::pin::Pin; use std::task; use hashbrown::HashMap; use parking_lot::Mutex; use pin_project_lite::pin_project; use thiserror::Error; use tokio::sync::oneshot; #[derive(Debug, Error)] pub(crate) enum RegisterError { #[error("Waiter `{0}` already registered")] Occupied(String), } #[derive(Debug, Error)] pub(crate) enum NotifyError { #[error("Notify failed: waiter `{0}` not registered")] NotFound(String), #[error("Notify failed: channel hangup")] Hangup, } #[derive(Debug, Error)] pub(crate) enum WaitError { #[error("Wait failed: channel hangup")] Hangup, } pub(crate) struct Waiters<T>(pub(self) Mutex<HashMap<String, oneshot::Sender<T>>>); impl<T> Default for Waiters<T> { fn default() -> Self { Waiters(Mutex::default()) } } impl<T> Waiters<T> { pub(crate) fn register(&self, key: String) -> Result<Waiter<'_, T>, RegisterError> { let (tx, rx) = oneshot::channel(); self.0 .lock() .try_insert(key.clone(), tx) .map_err(|e| RegisterError::Occupied(e.entry.key().clone()))?; Ok(Waiter { receiver: rx, guard: DropKey { registry: self, key, }, }) } pub(crate) fn notify(&self, key: &str, value: T) -> Result<(), NotifyError> where T: Send + Sync, { let tx = self .0 .lock() .remove(key) .ok_or_else(|| NotifyError::NotFound(key.to_string()))?; tx.send(value).map_err(|_| NotifyError::Hangup) } } struct DropKey<'a, T> { key: String, registry: &'a Waiters<T>, } impl<T> Drop for DropKey<'_, T> { fn drop(&mut self) { self.registry.0.lock().remove(&self.key); } } pin_project! { pub(crate) struct Waiter<'a, T> { #[pin] receiver: oneshot::Receiver<T>, guard: DropKey<'a, T>, } } impl<T> std::future::Future for Waiter<'_, T> { type Output = Result<T, WaitError>; fn poll(self: Pin<&mut Self>, cx: &mut task::Context<'_>) -> task::Poll<Self::Output> { self.project() .receiver .poll(cx) .map_err(|_| WaitError::Hangup) } } #[cfg(test)] mod tests { use std::sync::Arc; use super::*; #[tokio::test] async fn test_waiter() -> anyhow::Result<()> { let waiters = Arc::new(Waiters::default()); let key = "Key"; let waiter = waiters.register(key.to_owned())?; let waiters = Arc::clone(&waiters); let notifier = tokio::spawn(async move { waiters.notify(key, ())?; Ok(()) }); waiter.await?; notifier.await? } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/ext.rs
proxy/src/ext.rs
use std::panic::resume_unwind; use std::sync::{Mutex, MutexGuard}; use tokio::task::JoinError; pub(crate) trait LockExt<T> { fn lock_propagate_poison(&self) -> MutexGuard<'_, T>; } impl<T> LockExt<T> for Mutex<T> { /// Lock the mutex and panic if the mutex was poisoned. #[track_caller] fn lock_propagate_poison(&self) -> MutexGuard<'_, T> { match self.lock() { Ok(guard) => guard, // poison occurs when another thread panicked while holding the lock guard. // since panicking is often unrecoverable, propagating the poison panic is reasonable. Err(poison) => panic!("{poison}"), } } } pub(crate) trait TaskExt<T> { fn propagate_task_panic(self) -> T; } impl<T> TaskExt<T> for Result<T, JoinError> { /// Unwrap the result and panic if the inner task panicked. /// Also panics if the task was cancelled #[track_caller] fn propagate_task_panic(self) -> T { match self { Ok(t) => t, // Using resume_unwind prevents the panic hook being called twice. // Since we use this for structured concurrency, there is only // 1 logical panic, so this is more correct. Err(e) if e.is_panic() => resume_unwind(e.into_panic()), Err(e) => panic!("unexpected task error: {e}"), } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/parse.rs
proxy/src/parse.rs
//! Small parsing helpers. use std::ffi::CStr; pub(crate) fn split_cstr(bytes: &[u8]) -> Option<(&CStr, &[u8])> { let cstr = CStr::from_bytes_until_nul(bytes).ok()?; let (_, other) = bytes.split_at(cstr.to_bytes_with_nul().len()); Some((cstr, other)) } #[cfg(test)] mod tests { use super::*; #[test] fn test_split_cstr() { assert!(split_cstr(b"").is_none()); assert!(split_cstr(b"foo").is_none()); let (cstr, rest) = split_cstr(b"\0").expect("uh-oh"); assert_eq!(cstr.to_bytes(), b""); assert_eq!(rest, b""); let (cstr, rest) = split_cstr(b"foo\0bar").expect("uh-oh"); assert_eq!(cstr.to_bytes(), b"foo"); assert_eq!(rest, b"bar"); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/cancellation.rs
proxy/src/cancellation.rs
use std::convert::Infallible; use std::net::{IpAddr, SocketAddr}; use std::pin::pin; use std::sync::{Arc, OnceLock}; use std::time::Duration; use futures::FutureExt; use ipnet::{IpNet, Ipv4Net, Ipv6Net}; use postgres_client::RawCancelToken; use postgres_client::tls::MakeTlsConnect; use redis::{Cmd, FromRedisValue, SetExpiry, SetOptions, Value}; use serde::{Deserialize, Serialize}; use thiserror::Error; use tokio::net::TcpStream; use tokio::time::timeout; use tracing::{debug, error, info}; use crate::auth::AuthError; use crate::auth::backend::ComputeUserInfo; use crate::batch::{BatchQueue, BatchQueueError, QueueProcessing}; use crate::config::ComputeConfig; use crate::context::RequestContext; use crate::control_plane::ControlPlaneApi; use crate::error::ReportableError; use crate::ext::LockExt; use crate::metrics::{CancelChannelSizeGuard, CancellationRequest, Metrics, RedisMsgKind}; use crate::pqproto::CancelKeyData; use crate::rate_limiter::LeakyBucketRateLimiter; use crate::redis::keys::KeyPrefix; use crate::redis::kv_ops::{RedisKVClient, RedisKVClientError}; use crate::util::run_until; type IpSubnetKey = IpNet; /// Initial period and TTL is shorter to clear keys of short-lived connections faster. const CANCEL_KEY_INITIAL_PERIOD: Duration = Duration::from_secs(60); const CANCEL_KEY_REFRESH_PERIOD: Duration = Duration::from_secs(10 * 60); /// `CANCEL_KEY_TTL_SLACK` is added to the periods to determine the actual TTL. const CANCEL_KEY_TTL_SLACK: Duration = Duration::from_secs(30); // Message types for sending through mpsc channel pub enum CancelKeyOp { Store { key: CancelKeyData, value: Box<str>, expire: Duration, }, Refresh { key: CancelKeyData, expire: Duration, }, Get { key: CancelKeyData, }, GetOld { key: CancelKeyData, }, } impl CancelKeyOp { const fn redis_msg_kind(&self) -> RedisMsgKind { match self { CancelKeyOp::Store { .. } => RedisMsgKind::Set, CancelKeyOp::Refresh { .. } => RedisMsgKind::Expire, CancelKeyOp::Get { .. } => RedisMsgKind::Get, CancelKeyOp::GetOld { .. } => RedisMsgKind::HGet, } } fn cancel_channel_metric_guard(&self) -> CancelChannelSizeGuard<'static> { Metrics::get() .proxy .cancel_channel_size .guard(self.redis_msg_kind()) } } #[derive(thiserror::Error, Debug, Clone)] pub enum PipelineError { #[error("could not send cmd to redis: {0}")] RedisKVClient(Arc<RedisKVClientError>), #[error("incorrect number of responses from redis")] IncorrectNumberOfResponses, } pub struct Pipeline { inner: redis::Pipeline, replies: usize, } impl Pipeline { fn with_capacity(n: usize) -> Self { Self { inner: redis::Pipeline::with_capacity(n), replies: 0, } } async fn execute(self, client: &mut RedisKVClient) -> Result<Vec<Value>, PipelineError> { let responses = self.replies; let batch_size = self.inner.len(); if !client.credentials_refreshed() { tracing::debug!( "Redis credentials are not refreshed. Sleeping for 5 seconds before retrying..." ); tokio::time::sleep(Duration::from_secs(5)).await; } match client.query(&self.inner).await { // for each reply, we expect that many values. Ok(Value::Array(values)) if values.len() == responses => { debug!( batch_size, responses, "successfully completed cancellation jobs", ); Ok(values.into_iter().collect()) } Ok(value) => { error!(batch_size, ?value, "unexpected redis return value"); Err(PipelineError::IncorrectNumberOfResponses) } Err(err) => Err(PipelineError::RedisKVClient(Arc::new(err))), } } fn add_command(&mut self, cmd: Cmd) { self.inner.add_command(cmd); self.replies += 1; } } impl CancelKeyOp { fn register(&self, pipe: &mut Pipeline) { match self { CancelKeyOp::Store { key, value, expire } => { let key = KeyPrefix::Cancel(*key).build_redis_key(); pipe.add_command(Cmd::set_options( &key, &**value, SetOptions::default().with_expiration(SetExpiry::EX(expire.as_secs())), )); } CancelKeyOp::Refresh { key, expire } => { let key = KeyPrefix::Cancel(*key).build_redis_key(); pipe.add_command(Cmd::expire(&key, expire.as_secs() as i64)); } CancelKeyOp::GetOld { key } => { let key = KeyPrefix::Cancel(*key).build_redis_key(); pipe.add_command(Cmd::hget(key, "data")); } CancelKeyOp::Get { key } => { let key = KeyPrefix::Cancel(*key).build_redis_key(); pipe.add_command(Cmd::get(key)); } } } } pub struct CancellationProcessor { pub client: RedisKVClient, pub batch_size: usize, } impl QueueProcessing for CancellationProcessor { type Req = (CancelChannelSizeGuard<'static>, CancelKeyOp); type Res = redis::Value; type Err = PipelineError; fn batch_size(&self, _queue_size: usize) -> usize { self.batch_size } async fn apply(&mut self, batch: Vec<Self::Req>) -> Result<Vec<Self::Res>, Self::Err> { if !self.client.credentials_refreshed() { // this will cause a timeout for cancellation operations tracing::debug!( "Redis credentials are not refreshed. Sleeping for 5 seconds before retrying..." ); tokio::time::sleep(Duration::from_secs(5)).await; } let mut pipeline = Pipeline::with_capacity(batch.len()); let batch_size = batch.len(); debug!(batch_size, "running cancellation jobs"); for (_, op) in &batch { op.register(&mut pipeline); } pipeline.execute(&mut self.client).await } } /// Enables serving `CancelRequest`s. /// /// If `CancellationPublisher` is available, cancel request will be used to publish the cancellation key to other proxy instances. pub struct CancellationHandler { compute_config: &'static ComputeConfig, // rate limiter of cancellation requests limiter: Arc<std::sync::Mutex<LeakyBucketRateLimiter<IpSubnetKey>>>, tx: OnceLock<BatchQueue<CancellationProcessor>>, // send messages to the redis KV client task } #[derive(Debug, Error)] pub(crate) enum CancelError { #[error("{0}")] IO(#[from] std::io::Error), #[error("{0}")] Postgres(#[from] postgres_client::Error), #[error("rate limit exceeded")] RateLimit, #[error("Authentication error")] AuthError(#[from] AuthError), #[error("key not found")] NotFound, #[error("proxy service error")] InternalError, } impl ReportableError for CancelError { fn get_error_kind(&self) -> crate::error::ErrorKind { match self { CancelError::IO(_) => crate::error::ErrorKind::Compute, CancelError::Postgres(e) if e.as_db_error().is_some() => { crate::error::ErrorKind::Postgres } CancelError::Postgres(_) => crate::error::ErrorKind::Compute, CancelError::RateLimit => crate::error::ErrorKind::RateLimit, CancelError::NotFound | CancelError::AuthError(_) => crate::error::ErrorKind::User, CancelError::InternalError => crate::error::ErrorKind::Service, } } } impl CancellationHandler { pub fn new(compute_config: &'static ComputeConfig) -> Self { Self { compute_config, tx: OnceLock::new(), limiter: Arc::new(std::sync::Mutex::new( LeakyBucketRateLimiter::<IpSubnetKey>::new_with_shards( LeakyBucketRateLimiter::<IpSubnetKey>::DEFAULT, 64, ), )), } } pub fn init_tx(&self, queue: BatchQueue<CancellationProcessor>) { self.tx .set(queue) .map_err(|_| {}) .expect("cancellation queue should be registered once"); } pub(crate) fn get_key(self: Arc<Self>) -> Session { // we intentionally generate a random "backend pid" and "secret key" here. // we use the corresponding u64 as an identifier for the // actual endpoint+pid+secret for postgres/pgbouncer. // // if we forwarded the backend_pid from postgres to the client, there would be a lot // of overlap between our computes as most pids are small (~100). let key: CancelKeyData = rand::random(); debug!("registered new query cancellation key {key}"); Session { key, cancellation_handler: self, } } /// This is not cancel safe async fn get_cancel_key( &self, key: CancelKeyData, ) -> Result<Option<CancelClosure>, CancelError> { const TIMEOUT: Duration = Duration::from_secs(5); let Some(tx) = self.tx.get() else { tracing::warn!("cancellation handler is not available"); return Err(CancelError::InternalError); }; let guard = Metrics::get() .proxy .cancel_channel_size .guard(RedisMsgKind::Get); let op = CancelKeyOp::Get { key }; let result = timeout( TIMEOUT, tx.call((guard, op), std::future::pending::<Infallible>()), ) .await .map_err(|_| { tracing::warn!("timed out waiting to receive GetCancelData response"); CancelError::RateLimit })?; // We may still have cancel keys set with HSET <key> "data". // Check error type and retry with HGET. // TODO: remove code after HSET is not used anymore. let result = if let Err(err) = result.as_ref() && let BatchQueueError::Result(err) = err && let PipelineError::RedisKVClient(err) = err && let RedisKVClientError::Redis(err) = &**err && let Some(errcode) = err.code() && errcode == "WRONGTYPE" { let guard = Metrics::get() .proxy .cancel_channel_size .guard(RedisMsgKind::HGet); let op = CancelKeyOp::GetOld { key }; timeout( TIMEOUT, tx.call((guard, op), std::future::pending::<Infallible>()), ) .await .map_err(|_| { tracing::warn!("timed out waiting to receive GetCancelData response"); CancelError::RateLimit })? } else { result }; let result = result.map_err(|e| { tracing::warn!("failed to receive GetCancelData response: {e}"); CancelError::InternalError })?; let cancel_state_str = String::from_owned_redis_value(result).map_err(|e| { tracing::warn!("failed to receive GetCancelData response: {e}"); CancelError::InternalError })?; let cancel_closure: CancelClosure = serde_json::from_str(&cancel_state_str).map_err(|e| { tracing::warn!("failed to deserialize cancel state: {e}"); CancelError::InternalError })?; Ok(Some(cancel_closure)) } /// Try to cancel a running query for the corresponding connection. /// If the cancellation key is not found, it will be published to Redis. /// check_allowed - if true, check if the IP is allowed to cancel the query. /// Will fetch IP allowlist internally. /// /// return Result primarily for tests /// /// This is not cancel safe pub(crate) async fn cancel_session<T: ControlPlaneApi>( &self, key: CancelKeyData, ctx: RequestContext, check_ip_allowed: bool, check_vpc_allowed: bool, auth_backend: &T, ) -> Result<(), CancelError> { let subnet_key = match ctx.peer_addr() { IpAddr::V4(ip) => IpNet::V4(Ipv4Net::new_assert(ip, 24).trunc()), // use defaut mask here IpAddr::V6(ip) => IpNet::V6(Ipv6Net::new_assert(ip, 64).trunc()), }; let allowed = { let rate_limit_config = None; let limiter = self.limiter.lock_propagate_poison(); limiter.check(subnet_key, rate_limit_config, 1) }; if !allowed { // log only the subnet part of the IP address to know which subnet is rate limited tracing::warn!("Rate limit exceeded. Skipping cancellation message, {subnet_key}"); Metrics::get() .proxy .cancellation_requests_total .inc(CancellationRequest { kind: crate::metrics::CancellationOutcome::RateLimitExceeded, }); return Err(CancelError::RateLimit); } let cancel_state = self.get_cancel_key(key).await.map_err(|e| { tracing::warn!("failed to receive RedisOp response: {e}"); CancelError::InternalError })?; let Some(cancel_closure) = cancel_state else { tracing::warn!("query cancellation key not found: {key}"); Metrics::get() .proxy .cancellation_requests_total .inc(CancellationRequest { kind: crate::metrics::CancellationOutcome::NotFound, }); return Err(CancelError::NotFound); }; let info = &cancel_closure.user_info; let access_controls = auth_backend .get_endpoint_access_control(&ctx, &info.endpoint, &info.user) .await .map_err(|e| CancelError::AuthError(e.into()))?; access_controls.check(&ctx, check_ip_allowed, check_vpc_allowed)?; Metrics::get() .proxy .cancellation_requests_total .inc(CancellationRequest { kind: crate::metrics::CancellationOutcome::Found, }); info!("cancelling query per user's request using key {key}"); cancel_closure.try_cancel_query(self.compute_config).await } } /// This should've been a [`std::future::Future`], but /// it's impossible to name a type of an unboxed future /// (we'd need something like `#![feature(type_alias_impl_trait)]`). #[derive(Debug, Clone, Serialize, Deserialize)] pub struct CancelClosure { pub socket_addr: SocketAddr, pub cancel_token: RawCancelToken, pub hostname: String, // for pg_sni router pub user_info: ComputeUserInfo, } impl CancelClosure { /// Cancels the query running on user's compute node. pub(crate) async fn try_cancel_query( &self, compute_config: &ComputeConfig, ) -> Result<(), CancelError> { let socket = TcpStream::connect(self.socket_addr).await?; let tls = <_ as MakeTlsConnect<tokio::net::TcpStream>>::make_tls_connect( compute_config, &self.hostname, ) .map_err(|e| CancelError::IO(std::io::Error::other(e.to_string())))?; self.cancel_token.cancel_query_raw(socket, tls).await?; debug!("query was cancelled"); Ok(()) } } /// Helper for registering query cancellation tokens. pub(crate) struct Session { /// The user-facing key identifying this session. key: CancelKeyData, cancellation_handler: Arc<CancellationHandler>, } impl Session { pub(crate) fn key(&self) -> &CancelKeyData { &self.key } /// Ensure the cancel key is continously refreshed, /// but stop when the channel is dropped. /// /// This is not cancel safe pub(crate) async fn maintain_cancel_key( &self, session_id: uuid::Uuid, cancel: tokio::sync::oneshot::Receiver<Infallible>, cancel_closure: &CancelClosure, compute_config: &ComputeConfig, ) { let Some(tx) = self.cancellation_handler.tx.get() else { tracing::warn!("cancellation handler is not available"); // don't exit, as we only want to exit if cancelled externally. std::future::pending().await }; let closure_json = serde_json::to_string(&cancel_closure) .expect("serialising to json string should not fail") .into_boxed_str(); let mut cancel = pin!(cancel); enum State { Init, Refresh, } let mut state = State::Init; loop { let (op, mut wait_interval) = match state { State::Init => { tracing::debug!( src=%self.key, dest=?cancel_closure.cancel_token, "registering cancellation key" ); ( CancelKeyOp::Store { key: self.key, value: closure_json.clone(), expire: CANCEL_KEY_INITIAL_PERIOD + CANCEL_KEY_TTL_SLACK, }, CANCEL_KEY_INITIAL_PERIOD, ) } State::Refresh => { tracing::debug!( src=%self.key, dest=?cancel_closure.cancel_token, "refreshing cancellation key" ); ( CancelKeyOp::Refresh { key: self.key, expire: CANCEL_KEY_REFRESH_PERIOD + CANCEL_KEY_TTL_SLACK, }, CANCEL_KEY_REFRESH_PERIOD, ) } }; match tx .call((op.cancel_channel_metric_guard(), op), cancel.as_mut()) .await { // SET returns OK Ok(Value::Okay) => { tracing::debug!( src=%self.key, dest=?cancel_closure.cancel_token, "registered cancellation key" ); state = State::Refresh; } // EXPIRE returns 1 Ok(Value::Int(1)) => { tracing::debug!( src=%self.key, dest=?cancel_closure.cancel_token, "refreshed cancellation key" ); } Ok(_) => { // Any other response likely means the key expired. tracing::warn!(src=%self.key, "refreshing cancellation key failed"); // Re-enter the SET loop quickly to repush full data. state = State::Init; wait_interval = Duration::ZERO; } // retry immediately. Err(BatchQueueError::Result(error)) => { tracing::warn!(?error, "error refreshing cancellation key"); // Small delay to prevent busy loop with high cpu and logging. wait_interval = Duration::from_millis(10); } Err(BatchQueueError::Cancelled(Err(_cancelled))) => break, } // wait before continuing. break immediately if cancelled. if run_until(tokio::time::sleep(wait_interval), cancel.as_mut()) .await .is_err() { break; } } if let Err(err) = cancel_closure .try_cancel_query(compute_config) .boxed() .await { tracing::warn!( ?session_id, ?err, "could not cancel the query in the database" ); } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/console_redirect_proxy.rs
proxy/src/console_redirect_proxy.rs
use std::sync::Arc; use futures::{FutureExt, TryFutureExt}; use postgres_client::RawCancelToken; use tokio::io::{AsyncRead, AsyncWrite}; use tokio_util::sync::CancellationToken; use tracing::{Instrument, debug, error, info}; use crate::auth::backend::ConsoleRedirectBackend; use crate::cancellation::{CancelClosure, CancellationHandler}; use crate::config::{ProxyConfig, ProxyProtocolV2}; use crate::context::RequestContext; use crate::error::ReportableError; use crate::metrics::{Metrics, NumClientConnectionsGuard}; use crate::pglb::ClientRequestError; use crate::pglb::handshake::{HandshakeData, handshake}; use crate::pglb::passthrough::ProxyPassthrough; use crate::protocol2::{ConnectHeader, ConnectionInfo, read_proxy_protocol}; use crate::proxy::{ ErrorSource, connect_compute, forward_compute_params_to_client, send_client_greeting, }; use crate::util::run_until_cancelled; pub async fn task_main( config: &'static ProxyConfig, backend: &'static ConsoleRedirectBackend, listener: tokio::net::TcpListener, cancellation_token: CancellationToken, cancellation_handler: Arc<CancellationHandler>, ) -> anyhow::Result<()> { scopeguard::defer! { info!("proxy has shut down"); } // When set for the server socket, the keepalive setting // will be inherited by all accepted client sockets. socket2::SockRef::from(&listener).set_keepalive(true)?; let connections = tokio_util::task::task_tracker::TaskTracker::new(); let cancellations = tokio_util::task::task_tracker::TaskTracker::new(); while let Some(accept_result) = run_until_cancelled(listener.accept(), &cancellation_token).await { let (socket, peer_addr) = accept_result?; let conn_gauge = Metrics::get() .proxy .client_connections .guard(crate::metrics::Protocol::Tcp); let session_id = uuid::Uuid::new_v4(); let cancellation_handler = Arc::clone(&cancellation_handler); let cancellations = cancellations.clone(); debug!(protocol = "tcp", %session_id, "accepted new TCP connection"); connections.spawn(async move { let (socket, conn_info) = match config.proxy_protocol_v2 { ProxyProtocolV2::Required => { match read_proxy_protocol(socket).await { Err(e) => { error!("per-client task finished with an error: {e:#}"); return; } // our load balancers will not send any more data. let's just exit immediately Ok((_socket, ConnectHeader::Local)) => { debug!("healthcheck received"); return; } Ok((socket, ConnectHeader::Proxy(info))) => (socket, info), } } // ignore the header - it cannot be confused for a postgres or http connection so will // error later. ProxyProtocolV2::Rejected => ( socket, ConnectionInfo { addr: peer_addr, extra: None, }, ), }; match socket.set_nodelay(true) { Ok(()) => {} Err(e) => { error!( "per-client task finished with an error: failed to set socket option: {e:#}" ); return; } } let ctx = RequestContext::new(session_id, conn_info, crate::metrics::Protocol::Tcp); let res = handle_client( config, backend, &ctx, cancellation_handler, socket, conn_gauge, cancellations, ) .instrument(ctx.span()) .boxed() .await; match res { Err(e) => { ctx.set_error_kind(e.get_error_kind()); error!(parent: &ctx.span(), "per-client task finished with an error: {e:#}"); } Ok(None) => { ctx.set_success(); } Ok(Some(p)) => { ctx.set_success(); let _disconnect = ctx.log_connect(); match p.proxy_pass().await { Ok(()) => {} Err(ErrorSource::Client(e)) => { error!( ?session_id, "per-client task finished with an IO error from the client: {e:#}" ); } Err(ErrorSource::Compute(e)) => { error!( ?session_id, "per-client task finished with an IO error from the compute: {e:#}" ); } } } } }); } connections.close(); cancellations.close(); drop(listener); // Drain connections connections.wait().await; cancellations.wait().await; Ok(()) } #[allow(clippy::too_many_arguments)] pub(crate) async fn handle_client<S: AsyncRead + AsyncWrite + Unpin + Send>( config: &'static ProxyConfig, backend: &'static ConsoleRedirectBackend, ctx: &RequestContext, cancellation_handler: Arc<CancellationHandler>, stream: S, conn_gauge: NumClientConnectionsGuard<'static>, cancellations: tokio_util::task::task_tracker::TaskTracker, ) -> Result<Option<ProxyPassthrough<S>>, ClientRequestError> { debug!( protocol = %ctx.protocol(), "handling interactive connection from client" ); let metrics = &Metrics::get().proxy; let proto = ctx.protocol(); let request_gauge = metrics.connection_requests.guard(proto); let tls = config.tls_config.load(); let tls = tls.as_deref(); let record_handshake_error = !ctx.has_private_peer_addr(); let pause = ctx.latency_timer_pause(crate::metrics::Waiting::Client); let do_handshake = handshake(ctx, stream, tls, record_handshake_error); let (mut stream, params) = match tokio::time::timeout(config.handshake_timeout, do_handshake) .await?? { HandshakeData::Startup(stream, params) => (stream, params), HandshakeData::Cancel(cancel_key_data) => { // spawn a task to cancel the session, but don't wait for it cancellations.spawn({ let cancellation_handler_clone = Arc::clone(&cancellation_handler); let ctx = ctx.clone(); let cancel_span = tracing::span!(parent: None, tracing::Level::INFO, "cancel_session", session_id = ?ctx.session_id()); cancel_span.follows_from(tracing::Span::current()); async move { cancellation_handler_clone .cancel_session( cancel_key_data, ctx, config.authentication_config.ip_allowlist_check_enabled, config.authentication_config.is_vpc_acccess_proxy, backend.get_api(), ) .await .inspect_err(|e | debug!(error = ?e, "cancel_session failed")).ok(); }.instrument(cancel_span) }); return Ok(None); } }; drop(pause); ctx.set_db_options(params.clone()); let (node_info, mut auth_info, user_info) = match backend .authenticate(ctx, &config.authentication_config, &mut stream) .await { Ok(auth_result) => auth_result, Err(e) => Err(stream.throw_error(e, Some(ctx)).await)?, }; auth_info.set_startup_params(&params, true); let mut node = connect_compute::connect_to_compute( ctx, config, &node_info, connect_compute::TlsNegotiation::Postgres, ) .or_else(|e| async { Err(stream.throw_error(e, Some(ctx)).await) }) .await?; auth_info .authenticate(ctx, &mut node) .or_else(|e| async { Err(stream.throw_error(e, Some(ctx)).await) }) .await?; send_client_greeting(ctx, &config.greetings, &mut stream); let session = cancellation_handler.get_key(); let (process_id, secret_key) = forward_compute_params_to_client(ctx, *session.key(), &mut stream, &mut node.stream) .await?; let stream = stream.flush_and_into_inner().await?; let hostname = node.hostname.to_string(); let session_id = ctx.session_id(); let (cancel_on_shutdown, cancel) = tokio::sync::oneshot::channel(); tokio::spawn(async move { session .maintain_cancel_key( session_id, cancel, &CancelClosure { socket_addr: node.socket_addr, cancel_token: RawCancelToken { ssl_mode: node.ssl_mode, process_id, secret_key, }, hostname, user_info, }, &config.connect_to_compute, ) .await; }); Ok(Some(ProxyPassthrough { client: stream, compute: node.stream.into_framed().into_inner(), aux: node.aux, private_link_id: None, _cancel_on_shutdown: cancel_on_shutdown, _req: request_gauge, _conn: conn_gauge, _db_conn: node.guage, })) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/protocol2.rs
proxy/src/protocol2.rs
//! Proxy Protocol V2 implementation //! Compatible with <https://www.haproxy.org/download/3.1/doc/proxy-protocol.txt> use core::fmt; use std::io; use std::net::{Ipv4Addr, Ipv6Addr, SocketAddr}; use bytes::Buf; use smol_str::SmolStr; use strum_macros::FromRepr; use tokio::io::{AsyncRead, AsyncReadExt}; use zerocopy::{FromBytes, Immutable, KnownLayout, Unaligned, network_endian}; /// Proxy Protocol Version 2 Header const SIGNATURE: [u8; 12] = [ 0x0D, 0x0A, 0x0D, 0x0A, 0x00, 0x0D, 0x0A, 0x51, 0x55, 0x49, 0x54, 0x0A, ]; const LOCAL_V2: u8 = 0x20; const PROXY_V2: u8 = 0x21; const TCP_OVER_IPV4: u8 = 0x11; const UDP_OVER_IPV4: u8 = 0x12; const TCP_OVER_IPV6: u8 = 0x21; const UDP_OVER_IPV6: u8 = 0x22; #[derive(PartialEq, Eq, Clone, Debug)] pub struct ConnectionInfo { pub addr: SocketAddr, pub extra: Option<ConnectionInfoExtra>, } #[derive(PartialEq, Eq, Clone, Debug)] pub enum ConnectHeader { Local, Proxy(ConnectionInfo), } impl fmt::Display for ConnectionInfo { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match &self.extra { None => self.addr.ip().fmt(f), Some(ConnectionInfoExtra::Aws { vpce_id }) => { write!(f, "vpce_id[{vpce_id:?}]:addr[{}]", self.addr.ip()) } Some(ConnectionInfoExtra::Azure { link_id }) => { write!(f, "link_id[{link_id}]:addr[{}]", self.addr.ip()) } } } } #[derive(PartialEq, Eq, Clone, Debug)] pub enum ConnectionInfoExtra { Aws { vpce_id: SmolStr }, Azure { link_id: u32 }, } pub(crate) async fn read_proxy_protocol<T: AsyncRead + Unpin>( mut read: T, ) -> std::io::Result<(T, ConnectHeader)> { let mut header = [0; size_of::<ProxyProtocolV2Header>()]; read.read_exact(&mut header).await?; let header: ProxyProtocolV2Header = zerocopy::transmute!(header); if header.signature != SIGNATURE { return Err(std::io::Error::other("invalid proxy protocol header")); } let mut payload = vec![0; usize::from(header.len.get())]; read.read_exact(&mut payload).await?; let res = process_proxy_payload(header, &payload)?; Ok((read, res)) } fn process_proxy_payload( header: ProxyProtocolV2Header, mut payload: &[u8], ) -> std::io::Result<ConnectHeader> { match header.version_and_command { // the connection was established on purpose by the proxy // without being relayed. The connection endpoints are the sender and the // receiver. Such connections exist when the proxy sends health-checks to the // server. The receiver must accept this connection as valid and must use the // real connection endpoints and discard the protocol block including the // family which is ignored. LOCAL_V2 => return Ok(ConnectHeader::Local), // the connection was established on behalf of another node, // and reflects the original connection endpoints. The receiver must then use // the information provided in the protocol block to get original the address. PROXY_V2 => {} // other values are unassigned and must not be emitted by senders. Receivers // must drop connections presenting unexpected values here. _ => { return Err(io::Error::other(format!( "invalid proxy protocol command 0x{:02X}. expected local (0x20) or proxy (0x21)", header.version_and_command ))); } } let size_err = "invalid proxy protocol length. payload not large enough to fit requested IP addresses"; let addr = match header.protocol_and_family { TCP_OVER_IPV4 | UDP_OVER_IPV4 => { let addr = payload .try_get::<ProxyProtocolV2HeaderV4>() .ok_or_else(|| io::Error::other(size_err))?; SocketAddr::from((addr.src_addr.get(), addr.src_port.get())) } TCP_OVER_IPV6 | UDP_OVER_IPV6 => { let addr = payload .try_get::<ProxyProtocolV2HeaderV6>() .ok_or_else(|| io::Error::other(size_err))?; SocketAddr::from((addr.src_addr.get(), addr.src_port.get())) } // unspecified or unix stream. ignore the addresses _ => { return Err(io::Error::other( "invalid proxy protocol address family/transport protocol.", )); } }; let mut extra = None; while let Some(mut tlv) = read_tlv(&mut payload) { match Pp2Kind::from_repr(tlv.kind) { Some(Pp2Kind::Aws) => { if tlv.value.is_empty() { tracing::warn!("invalid aws tlv: no subtype"); } let subtype = tlv.value.get_u8(); match Pp2AwsType::from_repr(subtype) { Some(Pp2AwsType::VpceId) => match std::str::from_utf8(tlv.value) { Ok(s) => { extra = Some(ConnectionInfoExtra::Aws { vpce_id: s.into() }); } Err(e) => { tracing::warn!("invalid aws vpce id: {e}"); } }, None => { tracing::warn!("unknown aws tlv: subtype={subtype}"); } } } Some(Pp2Kind::Azure) => { if tlv.value.is_empty() { tracing::warn!("invalid azure tlv: no subtype"); } let subtype = tlv.value.get_u8(); match Pp2AzureType::from_repr(subtype) { Some(Pp2AzureType::PrivateEndpointLinkId) => { if tlv.value.len() != 4 { tracing::warn!("invalid azure link_id: {:?}", tlv.value); } extra = Some(ConnectionInfoExtra::Azure { link_id: tlv.value.get_u32_le(), }); } None => { tracing::warn!("unknown azure tlv: subtype={subtype}"); } } } Some(kind) => { tracing::debug!("unused tlv[{kind:?}]: {:?}", tlv.value); } None => { tracing::debug!("unknown tlv: {tlv:?}"); } } } Ok(ConnectHeader::Proxy(ConnectionInfo { addr, extra })) } #[derive(FromRepr, Debug, Copy, Clone)] #[repr(u8)] enum Pp2Kind { // The following are defined by https://www.haproxy.org/download/3.1/doc/proxy-protocol.txt // we don't use these but it would be interesting to know what's available Alpn = 0x01, Authority = 0x02, Crc32C = 0x03, Noop = 0x04, UniqueId = 0x05, Ssl = 0x20, NetNs = 0x30, /// <https://docs.aws.amazon.com/elasticloadbalancing/latest/network/edit-target-group-attributes.html#proxy-protocol> Aws = 0xEA, /// <https://learn.microsoft.com/en-us/azure/private-link/private-link-service-overview#getting-connection-information-using-tcp-proxy-v2> Azure = 0xEE, } #[derive(FromRepr, Debug, Copy, Clone)] #[repr(u8)] enum Pp2AwsType { VpceId = 0x01, } #[derive(FromRepr, Debug, Copy, Clone)] #[repr(u8)] enum Pp2AzureType { PrivateEndpointLinkId = 0x01, } #[derive(Debug)] struct Tlv<'a> { kind: u8, value: &'a [u8], } fn read_tlv<'a>(b: &mut &'a [u8]) -> Option<Tlv<'a>> { let tlv_header = b.try_get::<TlvHeader>()?; let len = usize::from(tlv_header.len.get()); Some(Tlv { kind: tlv_header.kind, value: b.split_off(..len)?, }) } trait BufExt: Sized { fn try_get<T: FromBytes>(&mut self) -> Option<T>; } impl BufExt for &[u8] { fn try_get<T: FromBytes>(&mut self) -> Option<T> { let (res, rest) = T::read_from_prefix(self).ok()?; *self = rest; Some(res) } } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(C, packed)] struct ProxyProtocolV2Header { signature: [u8; 12], version_and_command: u8, protocol_and_family: u8, len: network_endian::U16, } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(C, packed)] struct ProxyProtocolV2HeaderV4 { src_addr: NetworkEndianIpv4, dst_addr: NetworkEndianIpv4, src_port: network_endian::U16, dst_port: network_endian::U16, } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(C, packed)] struct ProxyProtocolV2HeaderV6 { src_addr: NetworkEndianIpv6, dst_addr: NetworkEndianIpv6, src_port: network_endian::U16, dst_port: network_endian::U16, } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(C, packed)] struct TlvHeader { kind: u8, len: network_endian::U16, } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(transparent)] struct NetworkEndianIpv4(network_endian::U32); impl NetworkEndianIpv4 { #[inline] fn get(self) -> Ipv4Addr { Ipv4Addr::from_bits(self.0.get()) } } #[derive(FromBytes, KnownLayout, Immutable, Unaligned, Copy, Clone)] #[repr(transparent)] struct NetworkEndianIpv6(network_endian::U128); impl NetworkEndianIpv6 { #[inline] fn get(self) -> Ipv6Addr { Ipv6Addr::from_bits(self.0.get()) } } #[cfg(test)] mod tests { use tokio::io::AsyncReadExt; use crate::protocol2::{ ConnectHeader, LOCAL_V2, PROXY_V2, TCP_OVER_IPV4, UDP_OVER_IPV6, read_proxy_protocol, }; #[tokio::test] async fn test_ipv4() { let header = super::SIGNATURE // Proxy command, IPV4 | TCP .chain([(2 << 4) | 1, (1 << 4) | 1].as_slice()) // 12 + 3 bytes .chain([0, 15].as_slice()) // src ip .chain([127, 0, 0, 1].as_slice()) // dst ip .chain([192, 168, 0, 1].as_slice()) // src port .chain([255, 255].as_slice()) // dst port .chain([1, 1].as_slice()) // TLV .chain([1, 2, 3].as_slice()); let extra_data = [0x55; 256]; let (mut read, info) = read_proxy_protocol(header.chain(extra_data.as_slice())) .await .unwrap(); let mut bytes = vec![]; read.read_to_end(&mut bytes).await.unwrap(); assert_eq!(bytes, extra_data); let ConnectHeader::Proxy(info) = info else { panic!() }; assert_eq!(info.addr, ([127, 0, 0, 1], 65535).into()); } #[tokio::test] async fn test_ipv6() { let header = super::SIGNATURE // Proxy command, IPV6 | UDP .chain([PROXY_V2, UDP_OVER_IPV6].as_slice()) // 36 + 3 bytes .chain([0, 39].as_slice()) // src ip .chain([15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0].as_slice()) // dst ip .chain([0, 15, 1, 14, 2, 13, 3, 12, 4, 11, 5, 10, 6, 9, 7, 8].as_slice()) // src port .chain([1, 1].as_slice()) // dst port .chain([255, 255].as_slice()) // TLV .chain([1, 2, 3].as_slice()); let extra_data = [0x55; 256]; let (mut read, info) = read_proxy_protocol(header.chain(extra_data.as_slice())) .await .unwrap(); let mut bytes = vec![]; read.read_to_end(&mut bytes).await.unwrap(); assert_eq!(bytes, extra_data); let ConnectHeader::Proxy(info) = info else { panic!() }; assert_eq!( info.addr, ([15, 14, 13, 12, 11, 10, 9, 8, 7, 6, 5, 4, 3, 2, 1, 0], 257).into() ); } #[tokio::test] #[should_panic = "invalid proxy protocol header"] async fn test_invalid() { let data = [0x55; 256]; read_proxy_protocol(data.as_slice()).await.unwrap(); } #[tokio::test] #[should_panic = "early eof"] async fn test_short() { let data = [0x55; 10]; read_proxy_protocol(data.as_slice()).await.unwrap(); } #[tokio::test] async fn test_large_tlv() { let tlv = vec![0x55; 32768]; let tlv_len = (tlv.len() as u16).to_be_bytes(); let len = (12 + 3 + tlv.len() as u16).to_be_bytes(); let header = super::SIGNATURE // Proxy command, Inet << 4 | Stream .chain([PROXY_V2, TCP_OVER_IPV4].as_slice()) // 12 + 3 bytes .chain(len.as_slice()) // src ip .chain([55, 56, 57, 58].as_slice()) // dst ip .chain([192, 168, 0, 1].as_slice()) // src port .chain([255, 255].as_slice()) // dst port .chain([1, 1].as_slice()) // TLV .chain([255].as_slice()) .chain(tlv_len.as_slice()) .chain(tlv.as_slice()); let extra_data = [0xaa; 256]; let (mut read, info) = read_proxy_protocol(header.chain(extra_data.as_slice())) .await .unwrap(); let mut bytes = vec![]; read.read_to_end(&mut bytes).await.unwrap(); assert_eq!(bytes, extra_data); let ConnectHeader::Proxy(info) = info else { panic!() }; assert_eq!(info.addr, ([55, 56, 57, 58], 65535).into()); } #[tokio::test] async fn test_local() { let len = 0u16.to_be_bytes(); let header = super::SIGNATURE .chain([LOCAL_V2, 0x00].as_slice()) .chain(len.as_slice()); let extra_data = [0xaa; 256]; let (mut read, info) = read_proxy_protocol(header.chain(extra_data.as_slice())) .await .unwrap(); let mut bytes = vec![]; read.read_to_end(&mut bytes).await.unwrap(); assert_eq!(bytes, extra_data); let ConnectHeader::Local = info else { panic!() }; } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/batch.rs
proxy/src/batch.rs
//! Batch processing system based on intrusive linked lists. //! //! Enqueuing a batch job requires no allocations, with //! direct support for cancelling jobs early. use std::collections::BTreeMap; use std::pin::pin; use std::sync::Mutex; use scopeguard::ScopeGuard; use tokio::sync::oneshot; use tokio::sync::oneshot::error::TryRecvError; use crate::ext::LockExt; type ProcResult<P> = Result<<P as QueueProcessing>::Res, <P as QueueProcessing>::Err>; pub trait QueueProcessing: Send + 'static { type Req: Send + 'static; type Res: Send; type Err: Send + Clone; /// Get the desired batch size. fn batch_size(&self, queue_size: usize) -> usize; /// This applies a full batch of events. /// Must respond with a full batch of replies. /// /// If this apply can error, it's expected that errors be forwarded to each Self::Res. /// /// Batching does not need to happen atomically. fn apply( &mut self, req: Vec<Self::Req>, ) -> impl Future<Output = Result<Vec<Self::Res>, Self::Err>> + Send; } #[derive(thiserror::Error)] pub enum BatchQueueError<E: Clone, C> { #[error(transparent)] Result(E), #[error(transparent)] Cancelled(C), } pub struct BatchQueue<P: QueueProcessing> { processor: tokio::sync::Mutex<P>, inner: Mutex<BatchQueueInner<P>>, } struct BatchJob<P: QueueProcessing> { req: P::Req, res: tokio::sync::oneshot::Sender<Result<P::Res, P::Err>>, } impl<P: QueueProcessing> BatchQueue<P> { pub fn new(p: P) -> Self { Self { processor: tokio::sync::Mutex::new(p), inner: Mutex::new(BatchQueueInner { version: 0, queue: BTreeMap::new(), }), } } /// Perform a single request-response process, this may be batched internally. /// /// This function is not cancel safe. pub async fn call<R>( &self, req: P::Req, cancelled: impl Future<Output = R>, ) -> Result<P::Res, BatchQueueError<P::Err, R>> { let (id, mut rx) = self.inner.lock_propagate_poison().register_job(req); let mut cancelled = pin!(cancelled); let resp: Option<Result<P::Res, P::Err>> = loop { // try become the leader, or try wait for success. let mut processor = tokio::select! { // try become leader. p = self.processor.lock() => p, // wait for success. resp = &mut rx => break resp.ok(), // wait for cancellation. cancel = cancelled.as_mut() => { let mut inner = self.inner.lock_propagate_poison(); if inner.queue.remove(&id).is_some() { tracing::warn!("batched task cancelled before completion"); } return Err(BatchQueueError::Cancelled(cancel)); }, }; tracing::debug!(id, "batch: became leader"); let (reqs, resps) = self.inner.lock_propagate_poison().get_batch(&processor); // snitch incase the task gets cancelled. let cancel_safety = scopeguard::guard((), |()| { if !std::thread::panicking() { tracing::error!( id, "batch: leader cancelled, despite not being cancellation safe" ); } }); // apply a batch. // if this is cancelled, jobs will not be completed and will panic. let values = processor.apply(reqs).await; // good: we didn't get cancelled. ScopeGuard::into_inner(cancel_safety); match values { Ok(values) => { if values.len() != resps.len() { tracing::error!( "batch: invalid response size, expected={}, got={}", resps.len(), values.len() ); } // send response values. for (tx, value) in std::iter::zip(resps, values) { if tx.send(Ok(value)).is_err() { // receiver hung up but that's fine. } } } Err(err) => { for tx in resps { if tx.send(Err(err.clone())).is_err() { // receiver hung up but that's fine. } } } } match rx.try_recv() { Ok(resp) => break Some(resp), Err(TryRecvError::Closed) => break None, // edge case - there was a race condition where // we became the leader but were not in the batch. // // Example: // thread 1: register job id=1 // thread 2: register job id=2 // thread 2: processor.lock().await // thread 1: processor.lock().await // thread 2: becomes leader, batch_size=1, jobs=[1]. Err(TryRecvError::Empty) => {} } }; tracing::debug!(id, "batch: job completed"); resp.expect("no response found. batch processer should not panic") .map_err(BatchQueueError::Result) } } struct BatchQueueInner<P: QueueProcessing> { version: u64, queue: BTreeMap<u64, BatchJob<P>>, } impl<P: QueueProcessing> BatchQueueInner<P> { fn register_job(&mut self, req: P::Req) -> (u64, oneshot::Receiver<ProcResult<P>>) { let (tx, rx) = oneshot::channel(); let id = self.version; // Overflow concern: // This is a u64, and we might enqueue 2^16 tasks per second. // This gives us 2^48 seconds (9 million years). // Even if this does overflow, it will not break, but some // jobs with the higher version might never get prioritised. self.version += 1; self.queue.insert(id, BatchJob { req, res: tx }); tracing::debug!(id, "batch: registered job in the queue"); (id, rx) } fn get_batch(&mut self, p: &P) -> (Vec<P::Req>, Vec<oneshot::Sender<ProcResult<P>>>) { let batch_size = p.batch_size(self.queue.len()); let mut reqs = Vec::with_capacity(batch_size); let mut resps = Vec::with_capacity(batch_size); let mut ids = Vec::with_capacity(batch_size); while reqs.len() < batch_size { let Some((id, job)) = self.queue.pop_first() else { break; }; reqs.push(job.req); resps.push(job.res); ids.push(id); } tracing::debug!(ids=?ids, "batch: acquired jobs"); (reqs, resps) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/url.rs
proxy/src/url.rs
use anyhow::bail; /// A [url](url::Url) type with additional guarantees. #[repr(transparent)] #[derive(Debug, Clone, PartialEq, Eq)] pub struct ApiUrl(url::Url); impl ApiUrl { /// Consume the wrapper and return inner [url](url::Url). pub(crate) fn into_inner(self) -> url::Url { self.0 } /// See [`url::Url::path_segments_mut`]. pub(crate) fn path_segments_mut(&mut self) -> url::PathSegmentsMut<'_> { // We've already verified that it works during construction. self.0.path_segments_mut().expect("bad API url") } } /// This instance imposes additional requirements on the url. impl std::str::FromStr for ApiUrl { type Err = anyhow::Error; fn from_str(s: &str) -> anyhow::Result<Self> { let mut url: url::Url = s.parse()?; // Make sure that we can build upon this URL. if url.path_segments_mut().is_err() { bail!("bad API url provided"); } Ok(Self(url)) } } /// This instance is safe because it doesn't allow us to modify the object. impl std::ops::Deref for ApiUrl { type Target = url::Url; fn deref(&self) -> &Self::Target { &self.0 } } impl std::ops::DerefMut for ApiUrl { fn deref_mut(&mut self) -> &mut Self::Target { &mut self.0 } } impl std::fmt::Display for ApiUrl { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { self.0.fmt(f) } } #[cfg(test)] mod tests { use super::*; #[test] fn bad_url() { let url = "test:foobar"; url.parse::<url::Url>().expect("unexpected parsing failure"); url.parse::<ApiUrl>().expect_err("should not parse"); } #[test] fn good_url() { let url = "test://foobar"; let mut a = url.parse::<url::Url>().expect("unexpected parsing failure"); let mut b = url.parse::<ApiUrl>().expect("unexpected parsing failure"); a.path_segments_mut().unwrap().push("method"); b.path_segments_mut().push("method"); assert_eq!(a, b.into_inner()); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/error.rs
proxy/src/error.rs
use std::fmt; use anyhow::Context; use measured::FixedCardinalityLabel; use tokio::task::JoinError; /// Marks errors that may be safely shown to a client. /// This trait can be seen as a specialized version of [`ToString`]. /// /// NOTE: This trait should not be implemented for [`anyhow::Error`], since it /// is way too convenient and tends to proliferate all across the codebase, /// ultimately leading to accidental leaks of sensitive data. pub(crate) trait UserFacingError: ReportableError { /// Format the error for client, stripping all sensitive info. /// /// Although this might be a no-op for many types, it's highly /// recommended to override the default impl in case error type /// contains anything sensitive: various IDs, IP addresses etc. #[inline(always)] fn to_string_client(&self) -> String { self.to_string() } } #[derive(Copy, Clone, Debug, Eq, PartialEq, FixedCardinalityLabel)] #[label(singleton = "type")] pub enum ErrorKind { /// Wrong password, unknown endpoint, protocol violation, etc... User, /// Network error between user and proxy. Not necessarily user error #[label(rename = "clientdisconnect")] ClientDisconnect, /// Proxy self-imposed user rate limits #[label(rename = "ratelimit")] RateLimit, /// Proxy self-imposed service-wise rate limits #[label(rename = "serviceratelimit")] ServiceRateLimit, /// Proxy quota limit violation #[label(rename = "quota")] Quota, /// internal errors Service, /// Error communicating with control plane #[label(rename = "controlplane")] ControlPlane, /// Postgres error Postgres, /// Error communicating with compute Compute, } impl ErrorKind { pub(crate) fn to_metric_label(self) -> &'static str { match self { ErrorKind::User => "user", ErrorKind::ClientDisconnect => "clientdisconnect", ErrorKind::RateLimit => "ratelimit", ErrorKind::ServiceRateLimit => "serviceratelimit", ErrorKind::Quota => "quota", ErrorKind::Service => "service", ErrorKind::ControlPlane => "controlplane", ErrorKind::Postgres => "postgres", ErrorKind::Compute => "compute", } } } pub(crate) trait ReportableError: fmt::Display + Send + 'static { fn get_error_kind(&self) -> ErrorKind; } /// Flattens `Result<Result<T>>` into `Result<T>`. pub fn flatten_err<T>(r: Result<anyhow::Result<T>, JoinError>) -> anyhow::Result<T> { r.context("join error").and_then(|x| x) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/usage_metrics.rs
proxy/src/usage_metrics.rs
//! Periodically collect proxy consumption metrics //! and push them to a HTTP endpoint. use std::borrow::Cow; use std::convert::Infallible; use std::sync::Arc; use std::sync::atomic::{AtomicU64, AtomicUsize, Ordering}; use std::time::Duration; use anyhow::{Context, bail}; use async_compression::tokio::write::GzipEncoder; use bytes::Bytes; use chrono::{DateTime, Datelike, Timelike, Utc}; use clashmap::ClashMap; use clashmap::mapref::entry::Entry; use consumption_metrics::{CHUNK_SIZE, Event, EventChunk, EventType, idempotency_key}; use once_cell::sync::Lazy; use remote_storage::{GenericRemoteStorage, RemotePath, TimeoutOrCancel}; use serde::{Deserialize, Serialize}; use smol_str::SmolStr; use tokio::io::AsyncWriteExt; use tokio_util::sync::CancellationToken; use tracing::{error, info, instrument, trace, warn}; use utils::backoff; use uuid::{NoContext, Timestamp}; use crate::config::MetricCollectionConfig; use crate::context::parquet::{FAILED_UPLOAD_MAX_RETRIES, FAILED_UPLOAD_WARN_THRESHOLD}; use crate::http; use crate::intern::{BranchIdInt, EndpointIdInt}; const PROXY_IO_BYTES_PER_CLIENT: &str = "proxy_io_bytes_per_client"; const HTTP_REPORTING_REQUEST_TIMEOUT: Duration = Duration::from_secs(10); const HTTP_REPORTING_RETRY_DURATION: Duration = Duration::from_secs(60); /// Key that uniquely identifies the object, this metric describes. /// Currently, endpoint_id is enough, but this may change later, /// so keep it in a named struct. /// /// Both the proxy and the ingestion endpoint will live in the same region (or cell) /// so while the project-id is unique across regions the whole pipeline will work correctly /// because we enrich the event with project_id in the control-plane endpoint. #[derive(Eq, Hash, PartialEq, Serialize, Deserialize, Debug, Clone)] pub(crate) struct Ids { pub(crate) endpoint_id: EndpointIdInt, pub(crate) branch_id: BranchIdInt, #[serde(with = "none_as_empty_string")] pub(crate) private_link_id: Option<SmolStr>, } #[derive(Eq, Hash, PartialEq, Serialize, Deserialize, Debug, Clone)] struct Extra { #[serde(flatten)] ids: Ids, direction: TrafficDirection, } mod none_as_empty_string { use serde::Deserialize; use smol_str::SmolStr; #[allow(clippy::ref_option)] pub fn serialize<S: serde::Serializer>(t: &Option<SmolStr>, s: S) -> Result<S::Ok, S::Error> { s.serialize_str(t.as_deref().unwrap_or("")) } pub fn deserialize<'de, D: serde::Deserializer<'de>>( d: D, ) -> Result<Option<SmolStr>, D::Error> { let s = SmolStr::deserialize(d)?; if s.is_empty() { Ok(None) } else { Ok(Some(s)) } } } #[derive(Eq, Hash, PartialEq, Serialize, Deserialize, Debug, Clone)] #[serde(rename_all = "lowercase")] pub(crate) enum TrafficDirection { Ingress, Egress, } pub(crate) trait MetricCounterRecorder { /// Record that some bytes were sent from the proxy to the client fn record_egress(&self, bytes: u64); /// Record that some bytes were sent from the client to the proxy fn record_ingress(&self, bytes: u64); /// Record that some connections were opened fn record_connection(&self, count: usize); } trait MetricCounterReporter { fn get_metrics(&mut self) -> MetricsData; fn move_metrics(&self) -> MetricsData; } #[derive(Debug)] pub(crate) struct MetricCounter { transmitted: AtomicU64, received: AtomicU64, opened_connections: AtomicUsize, } impl MetricCounterRecorder for MetricCounter { /// Record that some bytes were sent from the proxy to the client fn record_egress(&self, bytes: u64) { self.transmitted.fetch_add(bytes, Ordering::Relaxed); } /// Record that some bytes were sent from the proxy to the client fn record_ingress(&self, bytes: u64) { self.received.fetch_add(bytes, Ordering::Relaxed); } /// Record that some connections were opened fn record_connection(&self, count: usize) { self.opened_connections.fetch_add(count, Ordering::Relaxed); } } impl MetricCounterReporter for MetricCounter { fn get_metrics(&mut self) -> MetricsData { MetricsData { received: *self.received.get_mut(), transmitted: *self.transmitted.get_mut(), connections: *self.opened_connections.get_mut(), } } fn move_metrics(&self) -> MetricsData { MetricsData { received: self.received.swap(0, Ordering::Relaxed), transmitted: self.transmitted.swap(0, Ordering::Relaxed), connections: self.opened_connections.swap(0, Ordering::Relaxed), } } } struct MetricsData { transmitted: u64, received: u64, connections: usize, } struct BytesSent { transmitted: u64, received: u64, } trait Clearable { /// extract the value that should be reported fn should_report(self: &Arc<Self>) -> Option<BytesSent>; /// Determine whether the counter should be cleared from the global map. fn should_clear(self: &mut Arc<Self>) -> bool; } impl<C: MetricCounterReporter> Clearable for C { fn should_report(self: &Arc<Self>) -> Option<BytesSent> { // heuristic to see if the branch is still open // if a clone happens while we are observing, the heuristic will be incorrect. // // Worst case is that we won't report an event for this endpoint. // However, for the strong count to be 1 it must have occured that at one instant // all the endpoints were closed, so missing a report because the endpoints are closed is valid. let is_open = Arc::strong_count(self) > 1; // update cached metrics eagerly, even if they can't get sent // (to avoid sending the same metrics twice) // see the relevant discussion on why to do so even if the status is not success: // https://github.com/neondatabase/neon/pull/4563#discussion_r1246710956 let MetricsData { transmitted, received, connections, } = self.move_metrics(); // Our only requirement is that we report in every interval if there was an open connection // if there were no opened connections since, then we don't need to report if transmitted == 0 && received == 0 && !is_open && connections == 0 { None } else { Some(BytesSent { transmitted, received, }) } } fn should_clear(self: &mut Arc<Self>) -> bool { // we can't clear this entry if it's acquired elsewhere let Some(counter) = Arc::get_mut(self) else { return false; }; let MetricsData { transmitted, received, connections, } = counter.get_metrics(); // clear if there's no data to report transmitted == 0 && received == 0 && connections == 0 } } // endpoint and branch IDs are not user generated so we don't run the risk of hash-dos type FastHasher = std::hash::BuildHasherDefault<rustc_hash::FxHasher>; #[derive(Default)] pub(crate) struct Metrics { endpoints: ClashMap<Ids, Arc<MetricCounter>, FastHasher>, } impl Metrics { /// Register a new byte metrics counter for this endpoint pub(crate) fn register(&self, ids: Ids) -> Arc<MetricCounter> { let entry = if let Some(entry) = self.endpoints.get(&ids) { entry.clone() } else { self.endpoints .entry(ids) .or_insert_with(|| { Arc::new(MetricCounter { received: AtomicU64::new(0), transmitted: AtomicU64::new(0), opened_connections: AtomicUsize::new(0), }) }) .clone() }; entry.record_connection(1); entry } } pub(crate) static USAGE_METRICS: Lazy<Metrics> = Lazy::new(Metrics::default); pub async fn task_main(config: &MetricCollectionConfig) -> anyhow::Result<Infallible> { info!("metrics collector config: {config:?}"); scopeguard::defer! { info!("metrics collector has shut down"); } let http_client = http::new_client_with_timeout( HTTP_REPORTING_REQUEST_TIMEOUT, HTTP_REPORTING_RETRY_DURATION, ); let hostname = hostname::get()?.as_os_str().to_string_lossy().into_owned(); // Even if the remote storage is not configured, we still want to clear the metrics. let storage = if let Some(config) = config .backup_metric_collection_config .remote_storage_config .as_ref() { Some( GenericRemoteStorage::from_config(config) .await .context("remote storage init")?, ) } else { None }; let mut prev = Utc::now(); let mut ticker = tokio::time::interval(config.interval); loop { ticker.tick().await; let now = Utc::now(); collect_metrics_iteration( &USAGE_METRICS.endpoints, &http_client, &config.endpoint, storage.as_ref(), config.backup_metric_collection_config.chunk_size, &hostname, prev, now, ) .await; prev = now; } } fn collect_and_clear_metrics<C: Clearable>( endpoints: &ClashMap<Ids, Arc<C>, FastHasher>, ) -> Vec<(Ids, BytesSent)> { let mut metrics_to_clear = Vec::new(); let metrics_to_send: Vec<(Ids, BytesSent)> = endpoints .iter() .filter_map(|counter| { let key = counter.key().clone(); let Some(value) = counter.should_report() else { metrics_to_clear.push(key); return None; }; Some((key, value)) }) .collect(); for metric in metrics_to_clear { match endpoints.entry(metric) { Entry::Occupied(mut counter) => { if counter.get_mut().should_clear() { counter.remove_entry(); } } Entry::Vacant(_) => {} } } metrics_to_send } fn create_event_chunks<'a>( metrics_to_send: &'a [(Ids, BytesSent)], hostname: &'a str, prev: DateTime<Utc>, now: DateTime<Utc>, chunk_size: usize, ) -> impl Iterator<Item = EventChunk<'a, Event<Extra, &'static str>>> + 'a { metrics_to_send .chunks(chunk_size) .map(move |chunk| EventChunk { events: chunk .iter() .flat_map(|(ids, bytes)| { [ Event { kind: EventType::Incremental { start_time: prev, stop_time: now, }, metric: PROXY_IO_BYTES_PER_CLIENT, idempotency_key: idempotency_key(hostname), value: bytes.transmitted, extra: Extra { ids: ids.clone(), direction: TrafficDirection::Egress, }, }, Event { kind: EventType::Incremental { start_time: prev, stop_time: now, }, metric: PROXY_IO_BYTES_PER_CLIENT, idempotency_key: idempotency_key(hostname), value: bytes.received, extra: Extra { ids: ids.clone(), direction: TrafficDirection::Ingress, }, }, ] }) .collect(), }) } #[expect(clippy::too_many_arguments)] #[instrument(skip_all)] async fn collect_metrics_iteration( endpoints: &ClashMap<Ids, Arc<MetricCounter>, FastHasher>, client: &http::ClientWithMiddleware, metric_collection_endpoint: &reqwest::Url, storage: Option<&GenericRemoteStorage>, outer_chunk_size: usize, hostname: &str, prev: DateTime<Utc>, now: DateTime<Utc>, ) { info!( "starting collect_metrics_iteration. metric_collection_endpoint: {}", metric_collection_endpoint ); let metrics_to_send = collect_and_clear_metrics(endpoints); if metrics_to_send.is_empty() { trace!("no new metrics to send"); } let cancel = CancellationToken::new(); let path_prefix = create_remote_path_prefix(now); // Send metrics. for chunk in create_event_chunks(&metrics_to_send, hostname, prev, now, outer_chunk_size) { tokio::join!( upload_main_events_chunked(client, metric_collection_endpoint, &chunk, CHUNK_SIZE), async { if let Err(e) = upload_backup_events(storage, &chunk, &path_prefix, &cancel).await { error!("failed to upload consumption events to remote storage: {e:?}"); } } ); } } fn create_remote_path_prefix(now: DateTime<Utc>) -> String { format!( "year={year:04}/month={month:02}/day={day:02}/hour={hour:02}/{hour:02}:{minute:02}:{second:02}Z", year = now.year(), month = now.month(), day = now.day(), hour = now.hour(), minute = now.minute(), second = now.second(), ) } async fn upload_main_events_chunked( client: &http::ClientWithMiddleware, metric_collection_endpoint: &reqwest::Url, chunk: &EventChunk<'_, Event<Extra, &str>>, subchunk_size: usize, ) { // Split into smaller chunks to avoid exceeding the max request size for subchunk in chunk.events.chunks(subchunk_size).map(|c| EventChunk { events: Cow::Borrowed(c), }) { let res = client .post(metric_collection_endpoint.clone()) .json(&subchunk) .send() .await; let res = match res { Ok(x) => x, Err(err) => { // TODO: retry? error!("failed to send metrics: {:?}", err); continue; } }; if !res.status().is_success() { error!("metrics endpoint refused the sent metrics: {:?}", res); for metric in subchunk.events.iter().filter(|e| e.value > (1u64 << 40)) { // Report if the metric value is suspiciously large warn!("potentially abnormal metric value: {:?}", metric); } } } } async fn upload_backup_events( storage: Option<&GenericRemoteStorage>, chunk: &EventChunk<'_, Event<Extra, &'static str>>, path_prefix: &str, cancel: &CancellationToken, ) -> anyhow::Result<()> { let Some(storage) = storage else { warn!("no remote storage configured"); return Ok(()); }; let real_now = Utc::now(); let id = uuid::Uuid::new_v7(Timestamp::from_unix( NoContext, real_now.second().into(), real_now.nanosecond(), )); let path = format!("{path_prefix}_{id}.ndjson.gz"); let remote_path = match RemotePath::from_string(&path) { Ok(remote_path) => remote_path, Err(e) => { bail!("failed to create remote path from str {path}: {:?}", e); } }; // TODO: This is async compression from Vec to Vec. Rewrite as byte stream. // Use sync compression in blocking threadpool. let mut encoder = GzipEncoder::new(Vec::new()); for event in chunk.events.iter() { let data = serde_json::to_vec(event).context("serialize metrics")?; encoder.write_all(&data).await.context("compress metrics")?; encoder.write_all(b"\n").await.context("compress metrics")?; } encoder.shutdown().await.context("compress metrics")?; let compressed_data: Bytes = encoder.get_ref().clone().into(); backoff::retry( || async { let stream = futures::stream::once(futures::future::ready(Ok(compressed_data.clone()))); storage .upload(stream, compressed_data.len(), &remote_path, None, cancel) .await }, TimeoutOrCancel::caused_by_cancel, FAILED_UPLOAD_WARN_THRESHOLD, FAILED_UPLOAD_MAX_RETRIES, "usage_metrics_upload", cancel, ) .await .ok_or_else(|| anyhow::Error::new(TimeoutOrCancel::Cancel)) .and_then(|x| x) .with_context(|| format!("usage_metrics_upload: path={remote_path}"))?; Ok(()) } #[cfg(test)] mod tests { use std::fs; use std::io::{BufRead, BufReader}; use std::sync::{Arc, Mutex}; use anyhow::Error; use camino_tempfile::tempdir; use chrono::Utc; use consumption_metrics::{Event, EventChunk}; use http_body_util::BodyExt; use hyper::body::Incoming; use hyper::server::conn::http1; use hyper::service::service_fn; use hyper::{Request, Response}; use hyper_util::rt::TokioIo; use remote_storage::{RemoteStorageConfig, RemoteStorageKind}; use tokio::net::TcpListener; use url::Url; use super::*; use crate::http; use crate::types::{BranchId, EndpointId}; #[tokio::test] async fn metrics() { type Report = EventChunk<'static, Event<Extra, String>>; let reports: Arc<Mutex<Vec<Report>>> = Arc::default(); let listener = TcpListener::bind("127.0.0.1:0").await.unwrap(); let addr = listener.local_addr().unwrap(); tokio::spawn({ let reports = reports.clone(); async move { loop { if let Ok((stream, _addr)) = listener.accept().await { let reports = reports.clone(); http1::Builder::new() .serve_connection( TokioIo::new(stream), service_fn(move |req: Request<Incoming>| { let reports = reports.clone(); async move { let bytes = req.into_body().collect().await?.to_bytes(); let events = serde_json::from_slice(&bytes)?; reports.lock().unwrap().push(events); Ok::<_, Error>(Response::new(String::new())) } }), ) .await .unwrap(); } } } }); let metrics = Metrics::default(); let client = http::new_client(); let endpoint = Url::parse(&format!("http://{addr}")).unwrap(); let now = Utc::now(); let storage_test_dir = tempdir().unwrap(); let local_fs_path = storage_test_dir.path().join("usage_metrics"); fs::create_dir_all(&local_fs_path).unwrap(); let storage = GenericRemoteStorage::from_config(&RemoteStorageConfig { storage: RemoteStorageKind::LocalFs { local_path: local_fs_path.clone(), }, timeout: Duration::from_secs(10), small_timeout: Duration::from_secs(1), }) .await .unwrap(); let mut pushed_chunks: Vec<Report> = Vec::new(); let mut stored_chunks: Vec<Report> = Vec::new(); // no counters have been registered collect_metrics_iteration( &metrics.endpoints, &client, &endpoint, Some(&storage), 1000, "foo", now, now, ) .await; let r = std::mem::take(&mut *reports.lock().unwrap()); assert!(r.is_empty()); // register a new counter let counter = metrics.register(Ids { endpoint_id: (&EndpointId::from("e1")).into(), branch_id: (&BranchId::from("b1")).into(), private_link_id: None, }); // the counter should be observed despite 0 egress collect_metrics_iteration( &metrics.endpoints, &client, &endpoint, Some(&storage), 1000, "foo", now, now, ) .await; let r = std::mem::take(&mut *reports.lock().unwrap()); assert_eq!(r.len(), 1); assert_eq!(r[0].events.len(), 2); assert_eq!(r[0].events[0].value, 0); assert_eq!(r[0].events[0].extra.direction, TrafficDirection::Egress); assert_eq!(r[0].events[1].value, 0); assert_eq!(r[0].events[1].extra.direction, TrafficDirection::Ingress); pushed_chunks.extend(r); // record egress counter.record_egress(1); // record ingress counter.record_ingress(2); // egress should be observered collect_metrics_iteration( &metrics.endpoints, &client, &endpoint, Some(&storage), 1000, "foo", now, now, ) .await; let r = std::mem::take(&mut *reports.lock().unwrap()); assert_eq!(r.len(), 1); assert_eq!(r[0].events.len(), 2); assert_eq!(r[0].events[0].value, 1); assert_eq!(r[0].events[0].extra.direction, TrafficDirection::Egress); assert_eq!(r[0].events[1].value, 2); assert_eq!(r[0].events[1].extra.direction, TrafficDirection::Ingress); pushed_chunks.extend(r); // release counter drop(counter); // we do not observe the counter collect_metrics_iteration( &metrics.endpoints, &client, &endpoint, Some(&storage), 1000, "foo", now, now, ) .await; let r = std::mem::take(&mut *reports.lock().unwrap()); assert!(r.is_empty()); // counter is unregistered assert!(metrics.endpoints.is_empty()); let path_prefix = create_remote_path_prefix(now); for entry in walkdir::WalkDir::new(&local_fs_path) .into_iter() .filter_map(|e| e.ok()) { let path = local_fs_path.join(&path_prefix).to_string(); if entry.path().to_str().unwrap().starts_with(&path) { let file = fs::File::open(entry.into_path()).unwrap(); let decoder = flate2::bufread::GzDecoder::new(BufReader::new(file)); let reader = BufReader::new(decoder); let mut events: Vec<Event<Extra, String>> = Vec::new(); for line in reader.lines() { let line = line.unwrap(); let event: Event<Extra, String> = serde_json::from_str(&line).unwrap(); events.push(event); } let report = Report { events: Cow::Owned(events), }; stored_chunks.push(report); } } storage_test_dir.close().ok(); // sort by first event's idempotency key because the order of files is nondeterministic pushed_chunks.sort_by_cached_key(|c| c.events[0].idempotency_key.clone()); stored_chunks.sort_by_cached_key(|c| c.events[0].idempotency_key.clone()); assert_eq!(pushed_chunks, stored_chunks); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/types.rs
proxy/src/types.rs
use crate::intern::{EndpointIdInt, EndpointIdTag, InternId}; macro_rules! smol_str_wrapper { ($name:ident) => { #[derive(Clone, Debug, PartialEq, Eq, PartialOrd, Ord, Hash, Default)] pub struct $name(smol_str::SmolStr); impl $name { #[allow(unused)] pub(crate) fn as_str(&self) -> &str { self.0.as_str() } } impl std::fmt::Display for $name { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { self.0.fmt(f) } } impl<T> std::cmp::PartialEq<T> for $name where smol_str::SmolStr: std::cmp::PartialEq<T>, { fn eq(&self, other: &T) -> bool { self.0.eq(other) } } impl<T> From<T> for $name where smol_str::SmolStr: From<T>, { fn from(x: T) -> Self { Self(x.into()) } } impl AsRef<str> for $name { fn as_ref(&self) -> &str { self.0.as_ref() } } impl std::ops::Deref for $name { type Target = str; fn deref(&self) -> &str { &*self.0 } } impl<'de> serde::de::Deserialize<'de> for $name { fn deserialize<D: serde::de::Deserializer<'de>>(d: D) -> Result<Self, D::Error> { <smol_str::SmolStr as serde::de::Deserialize<'de>>::deserialize(d).map(Self) } } impl serde::Serialize for $name { fn serialize<S: serde::Serializer>(&self, s: S) -> Result<S::Ok, S::Error> { self.0.serialize(s) } } }; } const POOLER_SUFFIX: &str = "-pooler"; pub(crate) const LOCAL_PROXY_SUFFIX: &str = "-local-proxy"; impl EndpointId { #[must_use] fn normalize_str(&self) -> &str { if let Some(stripped) = self.as_ref().strip_suffix(POOLER_SUFFIX) { stripped } else if let Some(stripped) = self.as_ref().strip_suffix(LOCAL_PROXY_SUFFIX) { stripped } else { self } } #[must_use] pub fn normalize(&self) -> Self { self.normalize_str().into() } #[must_use] pub fn normalize_intern(&self) -> EndpointIdInt { EndpointIdTag::get_interner().get_or_intern(self.normalize_str()) } } // 90% of role name strings are 20 characters or less. smol_str_wrapper!(RoleName); // 50% of endpoint strings are 23 characters or less. smol_str_wrapper!(EndpointId); // 50% of branch strings are 23 characters or less. smol_str_wrapper!(BranchId); // 90% of project strings are 23 characters or less. smol_str_wrapper!(ProjectId); // 90% of account strings are 23 characters or less. smol_str_wrapper!(AccountId); // will usually equal endpoint ID smol_str_wrapper!(EndpointCacheKey); smol_str_wrapper!(DbName); // postgres hostname, will likely be a port:ip addr smol_str_wrapper!(Host);
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/util.rs
proxy/src/util.rs
use std::pin::pin; use futures::future::{Either, select}; use tokio_util::sync::CancellationToken; pub async fn run_until_cancelled<F: Future>( f: F, cancellation_token: &CancellationToken, ) -> Option<F::Output> { run_until(f, cancellation_token.cancelled()).await.ok() } /// Runs the future `f` unless interrupted by future `condition`. pub async fn run_until<F1: Future, F2: Future>( f: F1, condition: F2, ) -> Result<F1::Output, F2::Output> { match select(pin!(f), pin!(condition)).await { Either::Left((f1, _)) => Ok(f1), Either::Right((f2, _)) => Err(f2), } } pub fn deserialize_json_string<'de, D, T>(deserializer: D) -> Result<T, D::Error> where T: for<'de2> serde::Deserialize<'de2>, D: serde::Deserializer<'de>, { use serde::Deserialize; let s = String::deserialize(deserializer)?; serde_json::from_str(&s).map_err(<D::Error as serde::de::Error>::custom) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/pqproto.rs
proxy/src/pqproto.rs
//! Postgres protocol codec //! //! <https://www.postgresql.org/docs/current/protocol-message-formats.html> use std::fmt; use std::io::{self, Cursor}; use bytes::{Buf, BufMut}; use itertools::Itertools; use rand::distr::{Distribution, StandardUniform}; use tokio::io::{AsyncRead, AsyncReadExt, AsyncWrite, AsyncWriteExt}; use zerocopy::{FromBytes, Immutable, IntoBytes, big_endian}; pub type ErrorCode = [u8; 5]; pub const FE_PASSWORD_MESSAGE: u8 = b'p'; pub const SQLSTATE_INTERNAL_ERROR: [u8; 5] = *b"XX000"; /// The protocol version number. /// /// The most significant 16 bits are the major version number (3 for the protocol described here). /// The least significant 16 bits are the minor version number (0 for the protocol described here). /// <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-STARTUPMESSAGE> #[derive(Clone, Copy, PartialEq, PartialOrd, FromBytes, IntoBytes, Immutable)] #[repr(C)] pub struct ProtocolVersion { major: big_endian::U16, minor: big_endian::U16, } impl ProtocolVersion { pub const fn new(major: u16, minor: u16) -> Self { Self { major: big_endian::U16::new(major), minor: big_endian::U16::new(minor), } } pub const fn minor(self) -> u16 { self.minor.get() } pub const fn major(self) -> u16 { self.major.get() } } impl fmt::Debug for ProtocolVersion { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { f.debug_list() .entry(&self.major()) .entry(&self.minor()) .finish() } } /// <https://github.com/postgres/postgres/blob/ca481d3c9ab7bf69ff0c8d71ad3951d407f6a33c/src/include/libpq/pqcomm.h#L118> const MAX_STARTUP_PACKET_LENGTH: usize = 10000; const RESERVED_INVALID_MAJOR_VERSION: u16 = 1234; /// <https://github.com/postgres/postgres/blob/ca481d3c9ab7bf69ff0c8d71ad3951d407f6a33c/src/include/libpq/pqcomm.h#L132> const CANCEL_REQUEST_CODE: ProtocolVersion = ProtocolVersion::new(1234, 5678); /// <https://github.com/postgres/postgres/blob/ca481d3c9ab7bf69ff0c8d71ad3951d407f6a33c/src/include/libpq/pqcomm.h#L166> const NEGOTIATE_SSL_CODE: ProtocolVersion = ProtocolVersion::new(1234, 5679); /// <https://github.com/postgres/postgres/blob/ca481d3c9ab7bf69ff0c8d71ad3951d407f6a33c/src/include/libpq/pqcomm.h#L167> const NEGOTIATE_GSS_CODE: ProtocolVersion = ProtocolVersion::new(1234, 5680); /// This first reads the startup message header, is 8 bytes. /// The first 4 bytes is a big-endian message length, and the next 4 bytes is a version number. /// /// The length value is inclusive of the header. For example, /// an empty message will always have length 8. #[derive(Clone, Copy, FromBytes, IntoBytes, Immutable)] #[repr(C)] struct StartupHeader { len: big_endian::U32, version: ProtocolVersion, } /// read the type from the stream using zerocopy. /// /// not cancel safe. macro_rules! read { ($s:expr => $t:ty) => {{ // cannot be implemented as a function due to lack of const-generic-expr let mut buf = [0; size_of::<$t>()]; $s.read_exact(&mut buf).await?; let res: $t = zerocopy::transmute!(buf); res }}; } /// Returns true if TLS is supported. /// /// This is not cancel safe. pub async fn request_tls<S>(stream: &mut S) -> io::Result<bool> where S: AsyncRead + AsyncWrite + Unpin, { let payload = StartupHeader { len: 8.into(), version: NEGOTIATE_SSL_CODE, }; stream.write_all(payload.as_bytes()).await?; stream.flush().await?; // we expect back either `S` or `N` as a single byte. let mut res = *b"0"; stream.read_exact(&mut res).await?; debug_assert!( res == *b"S" || res == *b"N", "unexpected SSL negotiation response: {}", char::from(res[0]), ); // S for SSL. Ok(res == *b"S") } pub async fn read_startup<S>(stream: &mut S) -> io::Result<FeStartupPacket> where S: AsyncRead + Unpin, { let header = read!(stream => StartupHeader); // <https://github.com/postgres/postgres/blob/04bcf9e19a4261fe9c7df37c777592c2e10c32a7/src/backend/tcop/backend_startup.c#L378-L382> // First byte indicates standard SSL handshake message // (It can't be a Postgres startup length because in network byte order // that would be a startup packet hundreds of megabytes long) if header.as_bytes()[0] == 0x16 { return Ok(FeStartupPacket::SslRequest { // The bytes we read for the header are actually part of a TLS ClientHello. // In theory, if the ClientHello was < 8 bytes we would fail with EOF before we get here. // In practice though, I see no world where a ClientHello is less than 8 bytes // since it includes ephemeral keys etc. direct: Some(zerocopy::transmute!(header)), }); } let Some(len) = (header.len.get() as usize).checked_sub(8) else { return Err(io::Error::other(format!( "invalid startup message length {}, must be at least 8.", header.len, ))); }; // TODO: add a histogram for startup packet lengths if len > MAX_STARTUP_PACKET_LENGTH { tracing::warn!("large startup message detected: {len} bytes"); return Err(io::Error::other(format!( "invalid startup message length {len}" ))); } match header.version { // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-CANCELREQUEST> CANCEL_REQUEST_CODE => { if len != 8 { return Err(io::Error::other( "CancelRequest message is malformed, backend PID / secret key missing", )); } Ok(FeStartupPacket::CancelRequest( read!(stream => CancelKeyData), )) } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-SSLREQUEST> NEGOTIATE_SSL_CODE => { // Requested upgrade to SSL (aka TLS) Ok(FeStartupPacket::SslRequest { direct: None }) } NEGOTIATE_GSS_CODE => { // Requested upgrade to GSSAPI Ok(FeStartupPacket::GssEncRequest) } version if version.major() == RESERVED_INVALID_MAJOR_VERSION => Err(io::Error::other( format!("Unrecognized request code {version:?}"), )), // StartupMessage version => { // The protocol version number is followed by one or more pairs of parameter name and value strings. // A zero byte is required as a terminator after the last name/value pair. // Parameters can appear in any order. user is required, others are optional. let mut buf = vec![0; len]; stream.read_exact(&mut buf).await?; if buf.pop() != Some(b'\0') { return Err(io::Error::other( "StartupMessage params: missing null terminator", )); } // TODO: Don't do this. // There's no guarantee that these messages are utf8, // but they usually happen to be simple ascii. let params = String::from_utf8(buf) .map_err(|_| io::Error::other("StartupMessage params: invalid utf-8"))?; Ok(FeStartupPacket::StartupMessage { version, params: StartupMessageParams { params }, }) } } } /// Read a raw postgres packet, which will respect the max length requested. /// /// This returns the message tag, as well as the message body. The message /// body is written into `buf`, and it is otherwise completely overwritten. /// /// This is not cancel safe. pub async fn read_message<'a, S>( stream: &mut S, buf: &'a mut Vec<u8>, max: u32, ) -> io::Result<(u8, &'a mut [u8])> where S: AsyncRead + Unpin, { /// This first reads the header, which for regular messages in the 3.0 protocol is 5 bytes. /// The first byte is a message tag, and the next 4 bytes is a big-endian length. /// /// Awkwardly, the length value is inclusive of itself, but not of the tag. For example, /// an empty message will always have length 4. #[derive(Clone, Copy, FromBytes)] #[repr(C)] struct Header { tag: u8, len: big_endian::U32, } let header = read!(stream => Header); // as described above, the length must be at least 4. let Some(len) = header.len.get().checked_sub(4) else { return Err(io::Error::other(format!( "invalid startup message length {}, must be at least 4.", header.len, ))); }; // TODO: add a histogram for message lengths // check if the message exceeds our desired max. if len > max { tracing::warn!("large postgres message detected: {len} bytes"); return Err(io::Error::other(format!("invalid message length {len}"))); } // read in our entire message. buf.resize(len as usize, 0); stream.read_exact(buf).await?; Ok((header.tag, buf)) } pub struct WriteBuf(Cursor<Vec<u8>>); impl Buf for WriteBuf { #[inline] fn remaining(&self) -> usize { self.0.remaining() } #[inline] fn chunk(&self) -> &[u8] { self.0.chunk() } #[inline] fn advance(&mut self, cnt: usize) { self.0.advance(cnt); } } impl WriteBuf { pub const fn new() -> Self { Self(Cursor::new(Vec::new())) } /// Use a heuristic to determine if we should shrink the write buffer. #[inline] fn should_shrink(&self) -> bool { let n = self.0.position() as usize; let len = self.0.get_ref().len(); // the unused space at the front of our buffer is 2x the size of our filled portion. n + n > len } /// Shrink the write buffer so that subsequent writes have more spare capacity. #[cold] fn shrink(&mut self) { let n = self.0.position() as usize; let buf = self.0.get_mut(); // buf repr: // [----unused------|-----filled-----|-----uninit-----] // ^ n ^ buf.len() ^ buf.capacity() let filled = n..buf.len(); let filled_len = filled.len(); buf.copy_within(filled, 0); buf.truncate(filled_len); self.0.set_position(0); } /// clear the write buffer. pub fn reset(&mut self) { let buf = self.0.get_mut(); buf.clear(); self.0.set_position(0); } /// Shrinks the buffer if efficient to do so, and returns the remaining size. pub fn occupied_len(&mut self) -> usize { if self.should_shrink() { self.shrink(); } self.0.get_mut().len() } /// Write a raw message to the internal buffer. /// /// The size_hint value is only a hint for reserving space. It's ok if it's incorrect, since /// we calculate the length after the fact. pub fn write_raw(&mut self, size_hint: usize, tag: u8, f: impl FnOnce(&mut Vec<u8>)) { if self.should_shrink() { self.shrink(); } let buf = self.0.get_mut(); buf.reserve(5 + size_hint); buf.push(tag); let start = buf.len(); buf.extend_from_slice(&[0, 0, 0, 0]); f(buf); let end = buf.len(); let len = (end - start) as u32; buf[start..start + 4].copy_from_slice(&len.to_be_bytes()); } /// Write an encryption response message. pub fn encryption(&mut self, m: u8) { self.0.get_mut().push(m); } pub fn write_error(&mut self, msg: &str, error_code: ErrorCode) { self.shrink(); // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-ERRORRESPONSE> // <https://www.postgresql.org/docs/current/protocol-error-fields.html> // "SERROR\0CXXXXX\0M\0\0".len() == 17 self.write_raw(17 + msg.len(), b'E', |buf| { // Severity: ERROR buf.put_slice(b"SERROR\0"); // Code: error_code buf.put_u8(b'C'); buf.put_slice(&error_code); buf.put_u8(0); // Message: msg buf.put_u8(b'M'); buf.put_slice(msg.as_bytes()); buf.put_u8(0); // End. buf.put_u8(0); }); } } #[derive(Debug)] pub enum FeStartupPacket { CancelRequest(CancelKeyData), SslRequest { direct: Option<[u8; 8]>, }, GssEncRequest, StartupMessage { version: ProtocolVersion, params: StartupMessageParams, }, } #[derive(Debug, Clone, Default)] pub struct StartupMessageParams { pub params: String, } impl StartupMessageParams { /// Get parameter's value by its name. pub fn get(&self, name: &str) -> Option<&str> { self.iter().find_map(|(k, v)| (k == name).then_some(v)) } /// Split command-line options according to PostgreSQL's logic, /// taking into account all escape sequences but leaving them as-is. /// [`None`] means that there's no `options` in [`Self`]. pub fn options_raw(&self) -> Option<impl Iterator<Item = &str>> { self.get("options").map(Self::parse_options_raw) } /// Split command-line options according to PostgreSQL's logic, /// taking into account all escape sequences but leaving them as-is. pub fn parse_options_raw(input: &str) -> impl Iterator<Item = &str> { // See `postgres: pg_split_opts`. let mut last_was_escape = false; input .split(move |c: char| { // We split by non-escaped whitespace symbols. let should_split = c.is_ascii_whitespace() && !last_was_escape; last_was_escape = c == '\\' && !last_was_escape; should_split }) .filter(|s| !s.is_empty()) } /// Iterate through key-value pairs in an arbitrary order. pub fn iter(&self) -> impl Iterator<Item = (&str, &str)> { self.params.split_terminator('\0').tuples() } // This function is mostly useful in tests. #[cfg(test)] pub fn new<'a, const N: usize>(pairs: [(&'a str, &'a str); N]) -> Self { let mut b = Self { params: String::new(), }; for (k, v) in pairs { b.insert(k, v); } b } /// Set parameter's value by its name. /// name and value must not contain a \0 byte pub fn insert(&mut self, name: &str, value: &str) { self.params.reserve(name.len() + value.len() + 2); self.params.push_str(name); self.params.push('\0'); self.params.push_str(value); self.params.push('\0'); } } /// Cancel keys usually are represented as PID+SecretKey, but to proxy they're just /// opaque bytes. #[derive(Debug, Hash, PartialEq, Eq, Clone, Copy, FromBytes, IntoBytes, Immutable)] pub struct CancelKeyData(pub big_endian::U64); pub fn id_to_cancel_key(id: u64) -> CancelKeyData { CancelKeyData(big_endian::U64::new(id)) } impl fmt::Display for CancelKeyData { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { let id = self.0; f.debug_tuple("CancelKeyData") .field(&format_args!("{id:x}")) .finish() } } impl Distribution<CancelKeyData> for StandardUniform { fn sample<R: rand::Rng + ?Sized>(&self, rng: &mut R) -> CancelKeyData { id_to_cancel_key(rng.random()) } } pub enum BeMessage<'a> { AuthenticationOk, AuthenticationSasl(BeAuthenticationSaslMessage<'a>), AuthenticationCleartextPassword, BackendKeyData(CancelKeyData), ParameterStatus { name: &'a [u8], value: &'a [u8], }, ReadyForQuery, NoticeResponse(&'a str), NegotiateProtocolVersion { version: ProtocolVersion, options: &'a [&'a str], }, } #[derive(Debug)] pub enum BeAuthenticationSaslMessage<'a> { Methods(&'a [&'a str]), Continue(&'a [u8]), Final(&'a [u8]), } impl BeMessage<'_> { /// Write the message into an internal buffer pub fn write_message(self, buf: &mut WriteBuf) { match self { // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-AUTHENTICATIONCLEARTEXTPASSWORD> BeMessage::AuthenticationOk => { buf.write_raw(1, b'R', |buf| buf.put_i32(0)); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-AUTHENTICATIONCLEARTEXTPASSWORD> BeMessage::AuthenticationCleartextPassword => { buf.write_raw(1, b'R', |buf| buf.put_i32(3)); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-AUTHENTICATIONSASL> BeMessage::AuthenticationSasl(BeAuthenticationSaslMessage::Methods(methods)) => { let len: usize = methods.iter().map(|m| m.len() + 1).sum(); buf.write_raw(len + 2, b'R', |buf| { buf.put_i32(10); // Specifies that SASL auth method is used. for method in methods { buf.put_slice(method.as_bytes()); buf.put_u8(0); } buf.put_u8(0); // zero terminator for the list }); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-AUTHENTICATIONSASL> BeMessage::AuthenticationSasl(BeAuthenticationSaslMessage::Continue(extra)) => { buf.write_raw(extra.len() + 1, b'R', |buf| { buf.put_i32(11); // Continue SASL auth. buf.put_slice(extra); }); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-AUTHENTICATIONSASL> BeMessage::AuthenticationSasl(BeAuthenticationSaslMessage::Final(extra)) => { buf.write_raw(extra.len() + 1, b'R', |buf| { buf.put_i32(12); // Send final SASL message. buf.put_slice(extra); }); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-BACKENDKEYDATA> BeMessage::BackendKeyData(key_data) => { buf.write_raw(8, b'K', |buf| buf.put_slice(key_data.as_bytes())); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-NOTICERESPONSE> // <https://www.postgresql.org/docs/current/protocol-error-fields.html> BeMessage::NoticeResponse(msg) => { // 'N' signalizes NoticeResponse messages buf.write_raw(18 + msg.len(), b'N', |buf| { // Severity: NOTICE buf.put_slice(b"SNOTICE\0"); // Code: XX000 (ignored for notice, but still required) buf.put_slice(b"CXX000\0"); // Message: msg buf.put_u8(b'M'); buf.put_slice(msg.as_bytes()); buf.put_u8(0); // End notice. buf.put_u8(0); }); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-PARAMETERSTATUS> BeMessage::ParameterStatus { name, value } => { buf.write_raw(name.len() + value.len() + 2, b'S', |buf| { buf.put_slice(name.as_bytes()); buf.put_u8(0); buf.put_slice(value.as_bytes()); buf.put_u8(0); }); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-NEGOTIATEPROTOCOLVERSION> BeMessage::ReadyForQuery => { buf.write_raw(1, b'Z', |buf| buf.put_u8(b'I')); } // <https://www.postgresql.org/docs/current/protocol-message-formats.html#PROTOCOL-MESSAGE-FORMATS-NEGOTIATEPROTOCOLVERSION> BeMessage::NegotiateProtocolVersion { version, options } => { let len: usize = options.iter().map(|o| o.len() + 1).sum(); buf.write_raw(8 + len, b'v', |buf| { buf.put_slice(version.as_bytes()); buf.put_u32(options.len() as u32); for option in options { buf.put_slice(option.as_bytes()); buf.put_u8(0); } }); } } } } #[cfg(test)] mod tests { use std::io::Cursor; use tokio::io::{AsyncWriteExt, duplex}; use zerocopy::IntoBytes; use super::ProtocolVersion; use crate::pqproto::{FeStartupPacket, read_message, read_startup}; #[tokio::test] async fn reject_large_startup() { // we're going to define a v3.0 startup message with far too many parameters. let mut payload = vec![]; // 10001 + 8 bytes. payload.extend_from_slice(&10009_u32.to_be_bytes()); payload.extend_from_slice(ProtocolVersion::new(3, 0).as_bytes()); payload.resize(10009, b'a'); let (mut server, mut client) = duplex(128); #[rustfmt::skip] let (server, client) = tokio::join!( async move { read_startup(&mut server).await.unwrap_err() }, async move { client.write_all(&payload).await.unwrap_err() }, ); assert_eq!(server.to_string(), "invalid startup message length 10001"); assert_eq!(client.to_string(), "broken pipe"); } #[tokio::test] async fn reject_large_password() { // we're going to define a password message that is far too long. let mut payload = vec![]; payload.push(b'p'); payload.extend_from_slice(&517_u32.to_be_bytes()); payload.resize(518, b'a'); let (mut server, mut client) = duplex(128); #[rustfmt::skip] let (server, client) = tokio::join!( async move { read_message(&mut server, &mut vec![], 512).await.unwrap_err() }, async move { client.write_all(&payload).await.unwrap_err() }, ); assert_eq!(server.to_string(), "invalid message length 513"); assert_eq!(client.to_string(), "broken pipe"); } #[tokio::test] async fn read_startup_message() { let mut payload = vec![]; payload.extend_from_slice(&17_u32.to_be_bytes()); payload.extend_from_slice(ProtocolVersion::new(3, 0).as_bytes()); payload.extend_from_slice(b"abc\0def\0\0"); let startup = read_startup(&mut Cursor::new(&payload)).await.unwrap(); let FeStartupPacket::StartupMessage { version, params } = startup else { panic!("unexpected startup message: {startup:?}"); }; assert_eq!(version.major(), 3); assert_eq!(version.minor(), 0); assert_eq!(params.params, "abc\0def\0"); } #[tokio::test] async fn read_ssl_message() { let mut payload = vec![]; payload.extend_from_slice(&8_u32.to_be_bytes()); payload.extend_from_slice(ProtocolVersion::new(1234, 5679).as_bytes()); let startup = read_startup(&mut Cursor::new(&payload)).await.unwrap(); let FeStartupPacket::SslRequest { direct: None } = startup else { panic!("unexpected startup message: {startup:?}"); }; } #[tokio::test] async fn read_tls_message() { // sample client hello taken from <https://tls13.xargs.org/#client-hello> let client_hello = [ 0x16, 0x03, 0x01, 0x00, 0xf8, 0x01, 0x00, 0x00, 0xf4, 0x03, 0x03, 0x00, 0x01, 0x02, 0x03, 0x04, 0x05, 0x06, 0x07, 0x08, 0x09, 0x0a, 0x0b, 0x0c, 0x0d, 0x0e, 0x0f, 0x10, 0x11, 0x12, 0x13, 0x14, 0x15, 0x16, 0x17, 0x18, 0x19, 0x1a, 0x1b, 0x1c, 0x1d, 0x1e, 0x1f, 0x20, 0xe0, 0xe1, 0xe2, 0xe3, 0xe4, 0xe5, 0xe6, 0xe7, 0xe8, 0xe9, 0xea, 0xeb, 0xec, 0xed, 0xee, 0xef, 0xf0, 0xf1, 0xf2, 0xf3, 0xf4, 0xf5, 0xf6, 0xf7, 0xf8, 0xf9, 0xfa, 0xfb, 0xfc, 0xfd, 0xfe, 0xff, 0x00, 0x08, 0x13, 0x02, 0x13, 0x03, 0x13, 0x01, 0x00, 0xff, 0x01, 0x00, 0x00, 0xa3, 0x00, 0x00, 0x00, 0x18, 0x00, 0x16, 0x00, 0x00, 0x13, 0x65, 0x78, 0x61, 0x6d, 0x70, 0x6c, 0x65, 0x2e, 0x75, 0x6c, 0x66, 0x68, 0x65, 0x69, 0x6d, 0x2e, 0x6e, 0x65, 0x74, 0x00, 0x0b, 0x00, 0x04, 0x03, 0x00, 0x01, 0x02, 0x00, 0x0a, 0x00, 0x16, 0x00, 0x14, 0x00, 0x1d, 0x00, 0x17, 0x00, 0x1e, 0x00, 0x19, 0x00, 0x18, 0x01, 0x00, 0x01, 0x01, 0x01, 0x02, 0x01, 0x03, 0x01, 0x04, 0x00, 0x23, 0x00, 0x00, 0x00, 0x16, 0x00, 0x00, 0x00, 0x17, 0x00, 0x00, 0x00, 0x0d, 0x00, 0x1e, 0x00, 0x1c, 0x04, 0x03, 0x05, 0x03, 0x06, 0x03, 0x08, 0x07, 0x08, 0x08, 0x08, 0x09, 0x08, 0x0a, 0x08, 0x0b, 0x08, 0x04, 0x08, 0x05, 0x08, 0x06, 0x04, 0x01, 0x05, 0x01, 0x06, 0x01, 0x00, 0x2b, 0x00, 0x03, 0x02, 0x03, 0x04, 0x00, 0x2d, 0x00, 0x02, 0x01, 0x01, 0x00, 0x33, 0x00, 0x26, 0x00, 0x24, 0x00, 0x1d, 0x00, 0x20, 0x35, 0x80, 0x72, 0xd6, 0x36, 0x58, 0x80, 0xd1, 0xae, 0xea, 0x32, 0x9a, 0xdf, 0x91, 0x21, 0x38, 0x38, 0x51, 0xed, 0x21, 0xa2, 0x8e, 0x3b, 0x75, 0xe9, 0x65, 0xd0, 0xd2, 0xcd, 0x16, 0x62, 0x54, ]; let mut cursor = Cursor::new(&client_hello); let startup = read_startup(&mut cursor).await.unwrap(); let FeStartupPacket::SslRequest { direct: Some(prefix), } = startup else { panic!("unexpected startup message: {startup:?}"); }; // check that no data is lost. assert_eq!(prefix, [0x16, 0x03, 0x01, 0x00, 0xf8, 0x01, 0x00, 0x00]); assert_eq!(cursor.position(), 8); } #[tokio::test] async fn read_message_success() { let query = b"Q\0\0\0\x0cSELECT 1Q\0\0\0\x0cSELECT 2"; let mut cursor = Cursor::new(&query); let mut buf = vec![]; let (tag, message) = read_message(&mut cursor, &mut buf, 100).await.unwrap(); assert_eq!(tag, b'Q'); assert_eq!(message, b"SELECT 1"); let (tag, message) = read_message(&mut cursor, &mut buf, 100).await.unwrap(); assert_eq!(tag, b'Q'); assert_eq!(message, b"SELECT 2"); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/intern.rs
proxy/src/intern.rs
use std::hash::BuildHasherDefault; use std::marker::PhantomData; use std::num::NonZeroUsize; use std::ops::Index; use std::sync::OnceLock; use lasso::{Capacity, MemoryLimits, Spur, ThreadedRodeo}; use rustc_hash::FxHasher; use crate::types::{AccountId, BranchId, EndpointId, ProjectId, RoleName}; pub trait InternId: Sized + 'static { fn get_interner() -> &'static StringInterner<Self>; } pub struct StringInterner<Id> { inner: ThreadedRodeo<Spur, BuildHasherDefault<FxHasher>>, _id: PhantomData<Id>, } #[derive(PartialEq, Debug, Clone, Copy, Eq, Hash)] pub struct InternedString<Id> { inner: Spur, _id: PhantomData<Id>, } impl<Id: InternId> std::fmt::Display for InternedString<Id> { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { self.as_str().fmt(f) } } impl<Id: InternId> InternedString<Id> { pub(crate) fn as_str(&self) -> &'static str { Id::get_interner().inner.resolve(&self.inner) } pub(crate) fn get(s: &str) -> Option<Self> { Id::get_interner().get(s) } } impl<Id: InternId> AsRef<str> for InternedString<Id> { fn as_ref(&self) -> &str { self.as_str() } } impl<Id: InternId> std::ops::Deref for InternedString<Id> { type Target = str; fn deref(&self) -> &str { self.as_str() } } impl<'de, Id: InternId> serde::de::Deserialize<'de> for InternedString<Id> { fn deserialize<D: serde::de::Deserializer<'de>>(d: D) -> Result<Self, D::Error> { struct Visitor<Id>(PhantomData<Id>); impl<Id: InternId> serde::de::Visitor<'_> for Visitor<Id> { type Value = InternedString<Id>; fn expecting(&self, formatter: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { formatter.write_str("a string") } fn visit_str<E>(self, v: &str) -> Result<Self::Value, E> where E: serde::de::Error, { Ok(Id::get_interner().get_or_intern(v)) } } d.deserialize_str(Visitor::<Id>(PhantomData)) } } impl<Id: InternId> serde::Serialize for InternedString<Id> { fn serialize<S: serde::Serializer>(&self, s: S) -> Result<S::Ok, S::Error> { self.as_str().serialize(s) } } impl<Id: InternId> StringInterner<Id> { pub(crate) fn new() -> Self { StringInterner { inner: ThreadedRodeo::with_capacity_memory_limits_and_hasher( Capacity::new(2500, NonZeroUsize::new(1 << 16).expect("value is nonzero")), // unbounded MemoryLimits::for_memory_usage(usize::MAX), BuildHasherDefault::<FxHasher>::default(), ), _id: PhantomData, } } #[cfg(test)] fn len(&self) -> usize { self.inner.len() } #[cfg(test)] fn current_memory_usage(&self) -> usize { self.inner.current_memory_usage() } pub(crate) fn get_or_intern(&self, s: &str) -> InternedString<Id> { InternedString { inner: self.inner.get_or_intern(s), _id: PhantomData, } } pub(crate) fn get(&self, s: &str) -> Option<InternedString<Id>> { Some(InternedString { inner: self.inner.get(s)?, _id: PhantomData, }) } } impl<Id: InternId> Index<InternedString<Id>> for StringInterner<Id> { type Output = str; fn index(&self, index: InternedString<Id>) -> &Self::Output { self.inner.resolve(&index.inner) } } impl<Id: InternId> Default for StringInterner<Id> { fn default() -> Self { Self::new() } } #[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)] pub struct RoleNameTag; impl InternId for RoleNameTag { fn get_interner() -> &'static StringInterner<Self> { static ROLE_NAMES: OnceLock<StringInterner<RoleNameTag>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } pub type RoleNameInt = InternedString<RoleNameTag>; impl From<&RoleName> for RoleNameInt { fn from(value: &RoleName) -> Self { RoleNameTag::get_interner().get_or_intern(value) } } #[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)] pub struct EndpointIdTag; impl InternId for EndpointIdTag { fn get_interner() -> &'static StringInterner<Self> { static ROLE_NAMES: OnceLock<StringInterner<EndpointIdTag>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } pub type EndpointIdInt = InternedString<EndpointIdTag>; impl From<&EndpointId> for EndpointIdInt { fn from(value: &EndpointId) -> Self { EndpointIdTag::get_interner().get_or_intern(value) } } impl From<EndpointId> for EndpointIdInt { fn from(value: EndpointId) -> Self { EndpointIdTag::get_interner().get_or_intern(&value) } } #[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)] pub struct BranchIdTag; impl InternId for BranchIdTag { fn get_interner() -> &'static StringInterner<Self> { static ROLE_NAMES: OnceLock<StringInterner<BranchIdTag>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } pub type BranchIdInt = InternedString<BranchIdTag>; impl From<&BranchId> for BranchIdInt { fn from(value: &BranchId) -> Self { BranchIdTag::get_interner().get_or_intern(value) } } impl From<BranchId> for BranchIdInt { fn from(value: BranchId) -> Self { BranchIdTag::get_interner().get_or_intern(&value) } } #[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)] pub struct ProjectIdTag; impl InternId for ProjectIdTag { fn get_interner() -> &'static StringInterner<Self> { static ROLE_NAMES: OnceLock<StringInterner<ProjectIdTag>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } pub type ProjectIdInt = InternedString<ProjectIdTag>; impl From<&ProjectId> for ProjectIdInt { fn from(value: &ProjectId) -> Self { ProjectIdTag::get_interner().get_or_intern(value) } } impl From<ProjectId> for ProjectIdInt { fn from(value: ProjectId) -> Self { ProjectIdTag::get_interner().get_or_intern(&value) } } #[derive(Copy, Clone, Debug, PartialEq, Eq, Hash)] pub struct AccountIdTag; impl InternId for AccountIdTag { fn get_interner() -> &'static StringInterner<Self> { static ROLE_NAMES: OnceLock<StringInterner<AccountIdTag>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } pub type AccountIdInt = InternedString<AccountIdTag>; impl From<&AccountId> for AccountIdInt { fn from(value: &AccountId) -> Self { AccountIdTag::get_interner().get_or_intern(value) } } impl From<AccountId> for AccountIdInt { fn from(value: AccountId) -> Self { AccountIdTag::get_interner().get_or_intern(&value) } } #[cfg(test)] mod tests { use std::sync::OnceLock; use super::InternId; use crate::intern::StringInterner; struct MyId; impl InternId for MyId { fn get_interner() -> &'static StringInterner<Self> { pub(crate) static ROLE_NAMES: OnceLock<StringInterner<MyId>> = OnceLock::new(); ROLE_NAMES.get_or_init(Default::default) } } #[test] fn push_many_strings() { use rand::rngs::StdRng; use rand::{Rng, SeedableRng}; use rand_distr::Zipf; let endpoint_dist = Zipf::new(500000.0, 0.8).unwrap(); let endpoints = StdRng::seed_from_u64(272488357).sample_iter(endpoint_dist); let interner = MyId::get_interner(); const N: usize = 100_000; let mut verify = Vec::with_capacity(N); for endpoint in endpoints.take(N) { let endpoint = format!("ep-string-interning-{endpoint}"); let key = interner.get_or_intern(&endpoint); verify.push((endpoint, key)); } for (s, key) in verify { assert_eq!(interner[key], s); } // 2031616/59861 = 34 bytes per string assert_eq!(interner.len(), 59_861); // will have other overhead for the internal hashmaps that are not accounted for. assert_eq!(interner.current_memory_usage(), 2_031_616); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/logging.rs
proxy/src/logging.rs
use std::cell::RefCell; use std::collections::HashMap; use std::sync::Arc; use std::{env, io}; use chrono::{DateTime, Utc}; use opentelemetry::trace::TraceContextExt; use tracing::subscriber::Interest; use tracing::{Event, Metadata, Span, Subscriber, callsite, span}; use tracing_opentelemetry::OpenTelemetrySpanExt; use tracing_subscriber::filter::{EnvFilter, LevelFilter}; use tracing_subscriber::fmt::format::{Format, Full}; use tracing_subscriber::fmt::time::SystemTime; use tracing_subscriber::fmt::{FormatEvent, FormatFields}; use tracing_subscriber::layer::{Context, Layer}; use tracing_subscriber::prelude::*; use tracing_subscriber::registry::LookupSpan; use crate::metrics::Metrics; /// Initialize logging and OpenTelemetry tracing and exporter. /// /// Logging can be configured using `RUST_LOG` environment variable. /// /// OpenTelemetry is configured with OTLP/HTTP exporter. It picks up /// configuration from environment variables. For example, to change the /// destination, set `OTEL_EXPORTER_OTLP_ENDPOINT=http://jaeger:4318`. /// See <https://opentelemetry.io/docs/reference/specification/sdk-environment-variables> pub fn init() -> anyhow::Result<LoggingGuard> { let logfmt = LogFormat::from_env()?; let env_filter = EnvFilter::builder() .with_default_directive(LevelFilter::INFO.into()) .from_env_lossy() .add_directive( "aws_config=info" .parse() .expect("this should be a valid filter directive"), ) .add_directive( "azure_core::policies::transport=off" .parse() .expect("this should be a valid filter directive"), ); let provider = tracing_utils::init_tracing("proxy", tracing_utils::ExportConfig::default()); let otlp_layer = provider.as_ref().map(tracing_utils::layer); let json_log_layer = if logfmt == LogFormat::Json { Some(JsonLoggingLayer::new( RealClock, StderrWriter { stderr: std::io::stderr(), }, &["conn_id", "ep", "query_id", "request_id", "session_id"], )) } else { None }; let text_log_layer = if logfmt == LogFormat::Text { Some( tracing_subscriber::fmt::layer() .with_ansi(false) .with_writer(std::io::stderr) .with_target(false), ) } else { None }; tracing_subscriber::registry() .with(env_filter) .with(otlp_layer) .with(json_log_layer) .with(text_log_layer) .try_init()?; Ok(LoggingGuard(provider)) } /// Initialize logging for local_proxy with log prefix and no opentelemetry. /// /// Logging can be configured using `RUST_LOG` environment variable. pub fn init_local_proxy() -> anyhow::Result<LoggingGuard> { let env_filter = EnvFilter::builder() .with_default_directive(LevelFilter::INFO.into()) .from_env_lossy(); let fmt_layer = tracing_subscriber::fmt::layer() .with_ansi(false) .with_writer(std::io::stderr) .event_format(LocalProxyFormatter(Format::default().with_target(false))); tracing_subscriber::registry() .with(env_filter) .with(fmt_layer) .try_init()?; Ok(LoggingGuard(None)) } pub struct LocalProxyFormatter(Format<Full, SystemTime>); impl<S, N> FormatEvent<S, N> for LocalProxyFormatter where S: Subscriber + for<'a> LookupSpan<'a>, N: for<'a> FormatFields<'a> + 'static, { fn format_event( &self, ctx: &tracing_subscriber::fmt::FmtContext<'_, S, N>, mut writer: tracing_subscriber::fmt::format::Writer<'_>, event: &tracing::Event<'_>, ) -> std::fmt::Result { writer.write_str("[local_proxy] ")?; self.0.format_event(ctx, writer, event) } } pub struct LoggingGuard(Option<tracing_utils::Provider>); impl Drop for LoggingGuard { fn drop(&mut self) { if let Some(p) = &self.0 { // Shutdown trace pipeline gracefully, so that it has a chance to send any // pending traces before we exit. tracing::info!("shutting down the tracing machinery"); drop(p.shutdown()); } } } #[derive(Copy, Clone, PartialEq, Eq, Default, Debug)] enum LogFormat { Text, #[default] Json, } impl LogFormat { fn from_env() -> anyhow::Result<Self> { let logfmt = env::var("LOGFMT"); Ok(match logfmt.as_deref() { Err(_) => LogFormat::default(), Ok("text") => LogFormat::Text, Ok("json") => LogFormat::Json, Ok(logfmt) => anyhow::bail!("unknown log format: {logfmt}"), }) } } trait MakeWriter { fn make_writer(&self) -> impl io::Write; } struct StderrWriter { stderr: io::Stderr, } impl MakeWriter for StderrWriter { #[inline] fn make_writer(&self) -> impl io::Write { self.stderr.lock() } } // TODO: move into separate module or even separate crate. trait Clock { fn now(&self) -> DateTime<Utc>; } struct RealClock; impl Clock for RealClock { #[inline] fn now(&self) -> DateTime<Utc> { Utc::now() } } /// Name of the field used by tracing crate to store the event message. const MESSAGE_FIELD: &str = "message"; /// Tracing used to enforce that spans/events have no more than 32 fields. /// It seems this is no longer the case, but it's still documented in some places. /// Generally, we shouldn't expect more than 32 fields anyway, so we can try and /// rely on it for some (minor) performance gains. const MAX_TRACING_FIELDS: usize = 32; thread_local! { /// Thread-local instance with per-thread buffer for log writing. static EVENT_FORMATTER: RefCell<EventFormatter> = const { RefCell::new(EventFormatter::new()) }; /// Cached OS thread ID. static THREAD_ID: u64 = gettid::gettid(); } /// Map for values fixed at callsite registration. // We use papaya here because registration rarely happens post-startup. // papaya is good for read-heavy workloads. // // We use rustc_hash here because callsite::Identifier will always be an integer with low-bit entropy, // since it's always a pointer to static mutable data. rustc_hash was designed for low-bit entropy. type CallsiteMap<T> = papaya::HashMap<callsite::Identifier, T, std::hash::BuildHasherDefault<rustc_hash::FxHasher>>; /// Implements tracing layer to handle events specific to logging. struct JsonLoggingLayer<C: Clock, W: MakeWriter> { clock: C, writer: W, /// tracks which fields of each **event** are duplicates skipped_field_indices: CallsiteMap<SkippedFieldIndices>, /// tracks callsite names to an ID. callsite_name_ids: papaya::HashMap<&'static str, u32, ahash::RandomState>, span_info: CallsiteMap<CallsiteSpanInfo>, /// Fields we want to keep track of in a separate json object. extract_fields: &'static [&'static str], } impl<C: Clock, W: MakeWriter> JsonLoggingLayer<C, W> { fn new(clock: C, writer: W, extract_fields: &'static [&'static str]) -> Self { JsonLoggingLayer { clock, skipped_field_indices: CallsiteMap::default(), span_info: CallsiteMap::default(), callsite_name_ids: papaya::HashMap::default(), writer, extract_fields, } } #[inline] fn span_info(&self, metadata: &'static Metadata<'static>) -> CallsiteSpanInfo { self.span_info .pin() .get_or_insert_with(metadata.callsite(), || { CallsiteSpanInfo::new(&self.callsite_name_ids, metadata, self.extract_fields) }) .clone() } } impl<S, C: Clock + 'static, W: MakeWriter + 'static> Layer<S> for JsonLoggingLayer<C, W> where S: Subscriber + for<'a> LookupSpan<'a>, { fn on_event(&self, event: &Event<'_>, ctx: Context<'_, S>) { use std::io::Write; // TODO: consider special tracing subscriber to grab timestamp very // early, before OTel machinery, and add as event extension. let now = self.clock.now(); EVENT_FORMATTER.with(|f| { let mut borrow = f.try_borrow_mut(); let formatter = match borrow.as_deref_mut() { Ok(formatter) => formatter, // If the thread local formatter is borrowed, // then we likely hit an edge case were we panicked during formatting. // We allow the logging to proceed with an uncached formatter. Err(_) => &mut EventFormatter::new(), }; formatter.format( now, event, &ctx, &self.skipped_field_indices, self.extract_fields, ); let mut writer = self.writer.make_writer(); if writer.write_all(formatter.buffer()).is_err() { Metrics::get().proxy.logging_errors_count.inc(); } }); } /// Registers a SpanFields instance as span extension. fn on_new_span(&self, attrs: &span::Attributes<'_>, id: &span::Id, ctx: Context<'_, S>) { let span = ctx.span(id).expect("span must exist"); let mut fields = SpanFields::new(self.span_info(span.metadata())); attrs.record(&mut fields); // This is a new span: the extensions should not be locked // unless some layer spawned a thread to process this span. // I don't think any layers do that. span.extensions_mut().insert(fields); } fn on_record(&self, id: &span::Id, values: &span::Record<'_>, ctx: Context<'_, S>) { let span = ctx.span(id).expect("span must exist"); // assumption: `on_record` is rarely called. // assumption: a span being updated by one thread, // and formatted by another thread is even rarer. let mut ext = span.extensions_mut(); if let Some(fields) = ext.get_mut::<SpanFields>() { values.record(fields); } } /// Called (lazily) roughly once per event/span instance. We quickly check /// for duplicate field names and record duplicates as skippable. Last field wins. fn register_callsite(&self, metadata: &'static Metadata<'static>) -> Interest { debug_assert!( metadata.fields().len() <= MAX_TRACING_FIELDS, "callsite {metadata:?} has too many fields." ); if !metadata.is_event() { // register the span info. self.span_info(metadata); // Must not be never because we wouldn't get trace and span data. return Interest::always(); } let mut field_indices = SkippedFieldIndices::default(); let mut seen_fields = HashMap::new(); for field in metadata.fields() { if let Some(old_index) = seen_fields.insert(field.name(), field.index()) { field_indices.set(old_index); } } if !field_indices.is_empty() { self.skipped_field_indices .pin() .insert(metadata.callsite(), field_indices); } Interest::always() } } /// Any span info that is fixed to a particular callsite. Not variable between span instances. #[derive(Clone)] struct CallsiteSpanInfo { /// index of each field to extract. usize::MAX if not found. extract: Arc<[usize]>, /// tracks the fixed "callsite ID" for each span. /// note: this is not stable between runs. normalized_name: Arc<str>, } impl CallsiteSpanInfo { fn new( callsite_name_ids: &papaya::HashMap<&'static str, u32, ahash::RandomState>, metadata: &'static Metadata<'static>, extract_fields: &[&'static str], ) -> Self { let names: Vec<&'static str> = metadata.fields().iter().map(|f| f.name()).collect(); // get all the indices of span fields we want to focus let extract = extract_fields .iter() // use rposition, since we want last match wins. .map(|f1| names.iter().rposition(|f2| f1 == f2).unwrap_or(usize::MAX)) .collect(); // normalized_name is unique for each callsite, but it is not // unified across separate proxy instances. // todo: can we do better here? let cid = *callsite_name_ids .pin() .update_or_insert(metadata.name(), |&cid| cid + 1, 0); // we hope that most span names are unique, in which case this will always be 0 let normalized_name = if cid == 0 { metadata.name().into() } else { // if the span name is not unique, add the numeric ID to span name to distinguish it. // sadly this is non-determinstic, across restarts but we should fix it by disambiguating re-used span names instead. format!("{}#{cid}", metadata.name()).into() }; Self { extract, normalized_name, } } } #[derive(Clone)] struct RawValue(Box<[u8]>); impl RawValue { fn new(v: impl json::ValueEncoder) -> Self { Self(json::value_to_vec!(|val| v.encode(val)).into_boxed_slice()) } } impl json::ValueEncoder for &RawValue { fn encode(self, v: json::ValueSer<'_>) { v.write_raw_json(&self.0); } } /// Stores span field values recorded during the spans lifetime. struct SpanFields { values: [Option<RawValue>; MAX_TRACING_FIELDS], /// cached span info so we can avoid extra hashmap lookups in the hot path. span_info: CallsiteSpanInfo, } impl SpanFields { fn new(span_info: CallsiteSpanInfo) -> Self { Self { span_info, values: [const { None }; MAX_TRACING_FIELDS], } } } impl tracing::field::Visit for SpanFields { #[inline] fn record_f64(&mut self, field: &tracing::field::Field, value: f64) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_i64(&mut self, field: &tracing::field::Field, value: i64) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_u64(&mut self, field: &tracing::field::Field, value: u64) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_i128(&mut self, field: &tracing::field::Field, value: i128) { if let Ok(value) = i64::try_from(value) { self.values[field.index()] = Some(RawValue::new(value)); } else { self.values[field.index()] = Some(RawValue::new(format_args!("{value}"))); } } #[inline] fn record_u128(&mut self, field: &tracing::field::Field, value: u128) { if let Ok(value) = u64::try_from(value) { self.values[field.index()] = Some(RawValue::new(value)); } else { self.values[field.index()] = Some(RawValue::new(format_args!("{value}"))); } } #[inline] fn record_bool(&mut self, field: &tracing::field::Field, value: bool) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_bytes(&mut self, field: &tracing::field::Field, value: &[u8]) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_str(&mut self, field: &tracing::field::Field, value: &str) { self.values[field.index()] = Some(RawValue::new(value)); } #[inline] fn record_debug(&mut self, field: &tracing::field::Field, value: &dyn std::fmt::Debug) { self.values[field.index()] = Some(RawValue::new(format_args!("{value:?}"))); } #[inline] fn record_error( &mut self, field: &tracing::field::Field, value: &(dyn std::error::Error + 'static), ) { self.values[field.index()] = Some(RawValue::new(format_args!("{value}"))); } } /// List of field indices skipped during logging. Can list duplicate fields or /// metafields not meant to be logged. #[derive(Copy, Clone, Default)] struct SkippedFieldIndices { // 32-bits is large enough for `MAX_TRACING_FIELDS` bits: u32, } impl SkippedFieldIndices { #[inline] fn is_empty(self) -> bool { self.bits == 0 } #[inline] fn set(&mut self, index: usize) { debug_assert!(index <= 32, "index out of bounds of 32-bit set"); self.bits |= 1 << index; } #[inline] fn contains(self, index: usize) -> bool { self.bits & (1 << index) != 0 } } /// Formats a tracing event and writes JSON to its internal buffer including a newline. // TODO: buffer capacity management, truncate if too large struct EventFormatter { logline_buffer: Vec<u8>, } impl EventFormatter { #[inline] const fn new() -> Self { EventFormatter { logline_buffer: Vec::new(), } } #[inline] fn buffer(&self) -> &[u8] { &self.logline_buffer } fn format<S>( &mut self, now: DateTime<Utc>, event: &Event<'_>, ctx: &Context<'_, S>, skipped_field_indices: &CallsiteMap<SkippedFieldIndices>, extract_fields: &'static [&'static str], ) where S: Subscriber + for<'a> LookupSpan<'a>, { let timestamp = now.to_rfc3339_opts(chrono::SecondsFormat::Micros, true); use tracing_log::NormalizeEvent; let normalized_meta = event.normalized_metadata(); let meta = normalized_meta.as_ref().unwrap_or_else(|| event.metadata()); let skipped_field_indices = skipped_field_indices .pin() .get(&meta.callsite()) .copied() .unwrap_or_default(); self.logline_buffer.clear(); let serializer = json::ValueSer::new(&mut self.logline_buffer); json::value_as_object!(|serializer| { // Timestamp comes first, so raw lines can be sorted by timestamp. serializer.entry("timestamp", &*timestamp); // Level next. serializer.entry("level", meta.level().as_str()); // Message next. let mut message_extractor = MessageFieldExtractor::new(serializer.key("message"), skipped_field_indices); event.record(&mut message_extractor); message_extractor.finish(); // Direct message fields. { let mut message_skipper = MessageFieldSkipper::new( serializer.key("fields").object(), skipped_field_indices, ); event.record(&mut message_skipper); // rollback if no fields are present. if message_skipper.present { message_skipper.serializer.finish(); } } let mut extracted = ExtractedSpanFields::new(extract_fields); let spans = serializer.key("spans"); json::value_as_object!(|spans| { let parent_spans = ctx .event_span(event) .map_or(vec![], |parent| parent.scope().collect()); for span in parent_spans.iter().rev() { let ext = span.extensions(); // all spans should have this extension. let Some(fields) = ext.get() else { continue }; extracted.layer_span(fields); let SpanFields { values, span_info } = fields; let span_fields = spans.key(&*span_info.normalized_name); json::value_as_object!(|span_fields| { for (field, value) in std::iter::zip(span.metadata().fields(), values) { if let Some(value) = value { span_fields.entry(field.name(), value); } } }); } }); // TODO: thread-local cache? let pid = std::process::id(); // Skip adding pid 1 to reduce noise for services running in containers. if pid != 1 { serializer.entry("process_id", pid); } THREAD_ID.with(|tid| serializer.entry("thread_id", tid)); // TODO: tls cache? name could change if let Some(thread_name) = std::thread::current().name() && !thread_name.is_empty() && thread_name != "tokio-runtime-worker" { serializer.entry("thread_name", thread_name); } if let Some(task_id) = tokio::task::try_id() { serializer.entry("task_id", format_args!("{task_id}")); } serializer.entry("target", meta.target()); // Skip adding module if it's the same as target. if let Some(module) = meta.module_path() && module != meta.target() { serializer.entry("module", module); } if let Some(file) = meta.file() { if let Some(line) = meta.line() { serializer.entry("src", format_args!("{file}:{line}")); } else { serializer.entry("src", file); } } { let otel_context = Span::current().context(); let otel_spanref = otel_context.span(); let span_context = otel_spanref.span_context(); if span_context.is_valid() { serializer.entry("trace_id", format_args!("{}", span_context.trace_id())); } } if extracted.has_values() { // TODO: add fields from event, too? let extract = serializer.key("extract"); json::value_as_object!(|extract| { for (key, value) in std::iter::zip(extracted.names, extracted.values) { if let Some(value) = value { extract.entry(*key, &value); } } }); } }); self.logline_buffer.push(b'\n'); } } /// Extracts the message field that's mixed will other fields. struct MessageFieldExtractor<'buf> { serializer: Option<json::ValueSer<'buf>>, skipped_field_indices: SkippedFieldIndices, } impl<'buf> MessageFieldExtractor<'buf> { #[inline] fn new(serializer: json::ValueSer<'buf>, skipped_field_indices: SkippedFieldIndices) -> Self { Self { serializer: Some(serializer), skipped_field_indices, } } #[inline] fn finish(self) { if let Some(ser) = self.serializer { ser.value(""); } } #[inline] fn record_field(&mut self, field: &tracing::field::Field, v: impl json::ValueEncoder) { if field.name() == MESSAGE_FIELD && !self.skipped_field_indices.contains(field.index()) && let Some(ser) = self.serializer.take() { ser.value(v); } } } impl tracing::field::Visit for MessageFieldExtractor<'_> { #[inline] fn record_f64(&mut self, field: &tracing::field::Field, value: f64) { self.record_field(field, value); } #[inline] fn record_i64(&mut self, field: &tracing::field::Field, value: i64) { self.record_field(field, value); } #[inline] fn record_u64(&mut self, field: &tracing::field::Field, value: u64) { self.record_field(field, value); } #[inline] fn record_i128(&mut self, field: &tracing::field::Field, value: i128) { self.record_field(field, value); } #[inline] fn record_u128(&mut self, field: &tracing::field::Field, value: u128) { self.record_field(field, value); } #[inline] fn record_bool(&mut self, field: &tracing::field::Field, value: bool) { self.record_field(field, value); } #[inline] fn record_bytes(&mut self, field: &tracing::field::Field, value: &[u8]) { self.record_field(field, format_args!("{value:x?}")); } #[inline] fn record_str(&mut self, field: &tracing::field::Field, value: &str) { self.record_field(field, value); } #[inline] fn record_debug(&mut self, field: &tracing::field::Field, value: &dyn std::fmt::Debug) { self.record_field(field, format_args!("{value:?}")); } #[inline] fn record_error( &mut self, field: &tracing::field::Field, value: &(dyn std::error::Error + 'static), ) { self.record_field(field, format_args!("{value}")); } } /// A tracing field visitor that skips the message field. struct MessageFieldSkipper<'buf> { serializer: json::ObjectSer<'buf>, skipped_field_indices: SkippedFieldIndices, present: bool, } impl<'buf> MessageFieldSkipper<'buf> { #[inline] fn new(serializer: json::ObjectSer<'buf>, skipped_field_indices: SkippedFieldIndices) -> Self { Self { serializer, skipped_field_indices, present: false, } } #[inline] fn record_field(&mut self, field: &tracing::field::Field, v: impl json::ValueEncoder) { if field.name() != MESSAGE_FIELD && !field.name().starts_with("log.") && !self.skipped_field_indices.contains(field.index()) { self.serializer.entry(field.name(), v); self.present |= true; } } } impl tracing::field::Visit for MessageFieldSkipper<'_> { #[inline] fn record_f64(&mut self, field: &tracing::field::Field, value: f64) { self.record_field(field, value); } #[inline] fn record_i64(&mut self, field: &tracing::field::Field, value: i64) { self.record_field(field, value); } #[inline] fn record_u64(&mut self, field: &tracing::field::Field, value: u64) { self.record_field(field, value); } #[inline] fn record_i128(&mut self, field: &tracing::field::Field, value: i128) { self.record_field(field, value); } #[inline] fn record_u128(&mut self, field: &tracing::field::Field, value: u128) { self.record_field(field, value); } #[inline] fn record_bool(&mut self, field: &tracing::field::Field, value: bool) { self.record_field(field, value); } #[inline] fn record_bytes(&mut self, field: &tracing::field::Field, value: &[u8]) { self.record_field(field, format_args!("{value:x?}")); } #[inline] fn record_str(&mut self, field: &tracing::field::Field, value: &str) { self.record_field(field, value); } #[inline] fn record_debug(&mut self, field: &tracing::field::Field, value: &dyn std::fmt::Debug) { self.record_field(field, format_args!("{value:?}")); } #[inline] fn record_error( &mut self, field: &tracing::field::Field, value: &(dyn std::error::Error + 'static), ) { self.record_field(field, format_args!("{value}")); } } struct ExtractedSpanFields { names: &'static [&'static str], values: Vec<Option<RawValue>>, } impl ExtractedSpanFields { fn new(names: &'static [&'static str]) -> Self { ExtractedSpanFields { names, values: vec![None; names.len()], } } fn layer_span(&mut self, fields: &SpanFields) { let SpanFields { values, span_info } = fields; // extract the fields for (i, &j) in span_info.extract.iter().enumerate() { let Some(Some(value)) = values.get(j) else { continue; }; // TODO: replace clone with reference, if possible. self.values[i] = Some(value.clone()); } } #[inline] fn has_values(&self) -> bool { self.values.iter().any(|v| v.is_some()) } } #[cfg(test)] mod tests { use std::sync::{Arc, Mutex, MutexGuard}; use assert_json_diff::assert_json_eq; use tracing::info_span; use super::*; struct TestClock { current_time: Mutex<DateTime<Utc>>, } impl Clock for Arc<TestClock> { fn now(&self) -> DateTime<Utc> { *self.current_time.lock().expect("poisoned") } } struct VecWriter<'a> { buffer: MutexGuard<'a, Vec<u8>>, } impl MakeWriter for Arc<Mutex<Vec<u8>>> { fn make_writer(&self) -> impl io::Write { VecWriter { buffer: self.lock().expect("poisoned"), } } } impl io::Write for VecWriter<'_> { fn write(&mut self, buf: &[u8]) -> io::Result<usize> { self.buffer.write(buf) } fn flush(&mut self) -> io::Result<()> { Ok(()) } } #[test] fn test_field_collection() { let clock = Arc::new(TestClock { current_time: Mutex::new(Utc::now()), }); let buffer = Arc::new(Mutex::new(Vec::new())); let log_layer = JsonLoggingLayer { clock: clock.clone(), skipped_field_indices: papaya::HashMap::default(), span_info: papaya::HashMap::default(), callsite_name_ids: papaya::HashMap::default(), writer: buffer.clone(), extract_fields: &["x"], }; let registry = tracing_subscriber::Registry::default().with(log_layer); tracing::subscriber::with_default(registry, || { info_span!("some_span", x = 24).in_scope(|| { info_span!("some_other_span", y = 30).in_scope(|| { info_span!("some_span", x = 40, x = 41, x = 42).in_scope(|| { tracing::error!( a = 1, a = 2, a = 3, message = "explicit message field", "implicit message field" ); }); }); }); }); let buffer = Arc::try_unwrap(buffer) .expect("no other reference") .into_inner() .expect("poisoned"); let actual: serde_json::Value = serde_json::from_slice(&buffer).expect("valid JSON"); let expected: serde_json::Value = serde_json::json!( { "timestamp": clock.now().to_rfc3339_opts(chrono::SecondsFormat::Micros, true), "level": "ERROR", "message": "explicit message field", "fields": { "a": 3, }, "spans": { "some_span":{ "x": 24, }, "some_other_span": { "y": 30, }, "some_span#1": { "x": 42, }, }, "extract": { "x": 42, }, "src": actual.as_object().unwrap().get("src").unwrap().as_str().unwrap(), "target": "proxy::logging::tests", "process_id": actual.as_object().unwrap().get("process_id").unwrap().as_number().unwrap(), "thread_id": actual.as_object().unwrap().get("thread_id").unwrap().as_number().unwrap(), "thread_name": "logging::tests::test_field_collection", } ); assert_json_eq!(actual, expected); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/metrics.rs
proxy/src/metrics.rs
use std::sync::{Arc, OnceLock}; use lasso::ThreadedRodeo; use measured::label::{ FixedCardinalitySet, LabelGroupSet, LabelGroupVisitor, LabelName, LabelSet, LabelValue, StaticLabelSet, }; use measured::metric::group::Encoding; use measured::metric::histogram::Thresholds; use measured::metric::name::MetricName; use measured::{ Counter, CounterVec, FixedCardinalityLabel, Gauge, GaugeVec, Histogram, HistogramVec, LabelGroup, MetricGroup, }; use metrics::{CounterPairAssoc, CounterPairVec, HyperLogLogVec, InfoMetric}; use tokio::time::{self, Instant}; use crate::control_plane::messages::ColdStartInfo; use crate::error::ErrorKind; #[derive(MetricGroup)] #[metric(new())] pub struct Metrics { #[metric(namespace = "proxy")] #[metric(init = ProxyMetrics::new())] pub proxy: ProxyMetrics, #[metric(namespace = "wake_compute_lock")] pub wake_compute_lock: ApiLockMetrics, #[metric(namespace = "service")] pub service: ServiceMetrics, #[metric(namespace = "cache")] pub cache: CacheMetrics, } impl Metrics { #[track_caller] pub fn get() -> &'static Self { static SELF: OnceLock<Metrics> = OnceLock::new(); SELF.get_or_init(|| { let mut metrics = Metrics::new(); metrics.proxy.errors_total.init_all_dense(); metrics.proxy.redis_errors_total.init_all_dense(); metrics.proxy.redis_events_count.init_all_dense(); metrics.proxy.retries_metric.init_all_dense(); metrics.proxy.connection_failures_total.init_all_dense(); metrics }) } } #[derive(MetricGroup)] #[metric(new())] pub struct ProxyMetrics { #[metric(flatten)] pub db_connections: CounterPairVec<NumDbConnectionsGauge>, #[metric(flatten)] pub client_connections: CounterPairVec<NumClientConnectionsGauge>, #[metric(flatten)] pub connection_requests: CounterPairVec<NumConnectionRequestsGauge>, #[metric(flatten)] pub http_endpoint_pools: HttpEndpointPools, #[metric(flatten)] pub cancel_channel_size: CounterPairVec<CancelChannelSizeGauge>, /// Time it took for proxy to establish a connection to the compute endpoint. // largest bucket = 2^16 * 0.5ms = 32s #[metric(metadata = Thresholds::exponential_buckets(0.0005, 2.0))] pub compute_connection_latency_seconds: HistogramVec<ComputeConnectionLatencySet, 16>, /// Time it took for proxy to receive a response from control plane. #[metric( // largest bucket = 2^16 * 0.2ms = 13s metadata = Thresholds::exponential_buckets(0.0002, 2.0), )] pub console_request_latency: HistogramVec<ConsoleRequestSet, 16>, /// Size of the HTTP request body lengths. // smallest bucket = 16 bytes // largest bucket = 4^12 * 16 bytes = 256MB #[metric(metadata = Thresholds::exponential_buckets(16.0, 4.0))] pub http_conn_content_length_bytes: HistogramVec<StaticLabelSet<HttpDirection>, 12>, /// Time it takes to reclaim unused connection pools. #[metric(metadata = Thresholds::exponential_buckets(1e-6, 2.0))] pub http_pool_reclaimation_lag_seconds: Histogram<16>, /// Number of opened connections to a database. pub http_pool_opened_connections: Gauge, /// Number of allowed ips #[metric(metadata = Thresholds::with_buckets([0.0, 1.0, 2.0, 3.0, 4.0, 5.0, 10.0, 20.0, 50.0, 100.0]))] pub allowed_ips_number: Histogram<10>, /// Number of allowed VPC endpoints IDs #[metric(metadata = Thresholds::with_buckets([0.0, 1.0, 2.0, 3.0, 4.0, 5.0, 10.0, 20.0, 50.0, 100.0]))] pub allowed_vpc_endpoint_ids: Histogram<10>, /// Number of connections, by the method we used to determine the endpoint. pub accepted_connections_by_sni: CounterVec<SniSet>, /// Number of connection failures (per kind). pub connection_failures_total: CounterVec<StaticLabelSet<ConnectionFailureKind>>, /// Number of wake-up failures (per kind). pub connection_failures_breakdown: CounterVec<ConnectionFailuresBreakdownSet>, /// Number of bytes sent/received between all clients and backends. pub io_bytes: CounterVec<StaticLabelSet<Direction>>, /// Number of IO errors while logging. pub logging_errors_count: Counter, /// Number of errors by a given classification. pub errors_total: CounterVec<StaticLabelSet<crate::error::ErrorKind>>, /// Number of cancellation requests (per found/not_found). pub cancellation_requests_total: CounterVec<CancellationRequestSet>, /// Number of errors by a given classification pub redis_errors_total: CounterVec<RedisErrorsSet>, /// Number of TLS handshake failures pub tls_handshake_failures: Counter, /// Number of SHA 256 rounds executed. pub sha_rounds: Counter, /// HLL approximate cardinality of endpoints that are connecting pub connecting_endpoints: HyperLogLogVec<StaticLabelSet<Protocol>, 32>, /// Number of endpoints affected by errors of a given classification pub endpoints_affected_by_errors: HyperLogLogVec<StaticLabelSet<crate::error::ErrorKind>, 32>, /// Number of retries (per outcome, per retry_type). #[metric(metadata = Thresholds::with_buckets([0.0, 1.0, 2.0, 3.0, 4.0, 5.0, 6.0, 7.0, 8.0]))] pub retries_metric: HistogramVec<RetriesMetricSet, 9>, /// Number of events consumed from redis (per event type). pub redis_events_count: CounterVec<StaticLabelSet<RedisEventsCount>>, #[metric(namespace = "connect_compute_lock")] pub connect_compute_lock: ApiLockMetrics, #[metric(namespace = "scram_pool")] pub scram_pool: OnceLockWrapper<Arc<ThreadPoolMetrics>>, } /// A Wrapper over [`OnceLock`] to implement [`MetricGroup`]. pub struct OnceLockWrapper<T>(pub OnceLock<T>); impl<T> Default for OnceLockWrapper<T> { fn default() -> Self { Self(OnceLock::new()) } } impl<Enc: Encoding, T: MetricGroup<Enc>> MetricGroup<Enc> for OnceLockWrapper<T> { fn collect_group_into(&self, enc: &mut Enc) -> Result<(), Enc::Err> { if let Some(inner) = self.0.get() { inner.collect_group_into(enc)?; } Ok(()) } } #[derive(MetricGroup)] #[metric(new())] pub struct ApiLockMetrics { /// Number of semaphores registered in this api lock pub semaphores_registered: Counter, /// Number of semaphores unregistered in this api lock pub semaphores_unregistered: Counter, /// Time it takes to reclaim unused semaphores in the api lock #[metric(metadata = Thresholds::exponential_buckets(1e-6, 2.0))] pub reclamation_lag_seconds: Histogram<16>, /// Time it takes to acquire a semaphore lock #[metric(metadata = Thresholds::exponential_buckets(1e-4, 2.0))] pub semaphore_acquire_seconds: Histogram<16>, } impl Default for ApiLockMetrics { fn default() -> Self { Self::new() } } #[derive(FixedCardinalityLabel, Copy, Clone)] #[label(singleton = "direction")] pub enum HttpDirection { Request, Response, } #[derive(FixedCardinalityLabel, Copy, Clone)] #[label(singleton = "direction")] pub enum Direction { Tx, Rx, } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] #[label(singleton = "protocol")] pub enum Protocol { Http, Ws, Tcp, SniRouter, } impl Protocol { pub fn as_str(self) -> &'static str { match self { Protocol::Http => "http", Protocol::Ws => "ws", Protocol::Tcp => "tcp", Protocol::SniRouter => "sni_router", } } } impl std::fmt::Display for Protocol { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { f.write_str(self.as_str()) } } #[derive(FixedCardinalityLabel, Copy, Clone)] pub enum Bool { True, False, } #[derive(LabelGroup)] #[label(set = ConsoleRequestSet)] pub struct ConsoleRequest<'a> { #[label(dynamic_with = ThreadedRodeo, default)] pub request: &'a str, } #[derive(MetricGroup, Default)] pub struct HttpEndpointPools { /// Number of endpoints we have registered pools for pub http_pool_endpoints_registered_total: Counter, /// Number of endpoints we have unregistered pools for pub http_pool_endpoints_unregistered_total: Counter, } pub struct HttpEndpointPoolsGuard<'a> { dec: &'a Counter, } impl Drop for HttpEndpointPoolsGuard<'_> { fn drop(&mut self) { self.dec.inc(); } } impl HttpEndpointPools { pub fn guard(&self) -> HttpEndpointPoolsGuard<'_> { self.http_pool_endpoints_registered_total.inc(); HttpEndpointPoolsGuard { dec: &self.http_pool_endpoints_unregistered_total, } } } pub struct NumDbConnectionsGauge; impl CounterPairAssoc for NumDbConnectionsGauge { const INC_NAME: &'static MetricName = MetricName::from_str("opened_db_connections_total"); const DEC_NAME: &'static MetricName = MetricName::from_str("closed_db_connections_total"); const INC_HELP: &'static str = "Number of opened connections to a database."; const DEC_HELP: &'static str = "Number of closed connections to a database."; type LabelGroupSet = StaticLabelSet<Protocol>; } pub type NumDbConnectionsGuard<'a> = metrics::MeasuredCounterPairGuard<'a, NumDbConnectionsGauge>; pub struct NumClientConnectionsGauge; impl CounterPairAssoc for NumClientConnectionsGauge { const INC_NAME: &'static MetricName = MetricName::from_str("opened_client_connections_total"); const DEC_NAME: &'static MetricName = MetricName::from_str("closed_client_connections_total"); const INC_HELP: &'static str = "Number of opened connections from a client."; const DEC_HELP: &'static str = "Number of closed connections from a client."; type LabelGroupSet = StaticLabelSet<Protocol>; } pub type NumClientConnectionsGuard<'a> = metrics::MeasuredCounterPairGuard<'a, NumClientConnectionsGauge>; pub struct NumConnectionRequestsGauge; impl CounterPairAssoc for NumConnectionRequestsGauge { const INC_NAME: &'static MetricName = MetricName::from_str("accepted_connections_total"); const DEC_NAME: &'static MetricName = MetricName::from_str("closed_connections_total"); const INC_HELP: &'static str = "Number of client connections accepted."; const DEC_HELP: &'static str = "Number of client connections closed."; type LabelGroupSet = StaticLabelSet<Protocol>; } pub type NumConnectionRequestsGuard<'a> = metrics::MeasuredCounterPairGuard<'a, NumConnectionRequestsGauge>; pub struct CancelChannelSizeGauge; impl CounterPairAssoc for CancelChannelSizeGauge { const INC_NAME: &'static MetricName = MetricName::from_str("opened_msgs_cancel_channel_total"); const DEC_NAME: &'static MetricName = MetricName::from_str("closed_msgs_cancel_channel_total"); const INC_HELP: &'static str = "Number of processing messages in the cancellation channel."; const DEC_HELP: &'static str = "Number of closed messages in the cancellation channel."; type LabelGroupSet = StaticLabelSet<RedisMsgKind>; } pub type CancelChannelSizeGuard<'a> = metrics::MeasuredCounterPairGuard<'a, CancelChannelSizeGauge>; #[derive(LabelGroup)] #[label(set = ComputeConnectionLatencySet)] pub struct ComputeConnectionLatencyGroup { protocol: Protocol, cold_start_info: ColdStartInfo, outcome: ConnectOutcome, excluded: LatencyExclusions, } #[derive(FixedCardinalityLabel, Copy, Clone)] pub enum LatencyExclusions { Client, ClientAndCplane, ClientCplaneCompute, ClientCplaneComputeRetry, } #[derive(LabelGroup)] #[label(set = SniSet)] pub struct SniGroup { pub protocol: Protocol, pub kind: SniKind, } #[derive(FixedCardinalityLabel, Copy, Clone)] pub enum SniKind { /// Domain name based routing. SNI for libpq/websockets. Host for HTTP Sni, /// Metadata based routing. `options` for libpq/websockets. Header for HTTP NoSni, /// Metadata based routing, using the password field. PasswordHack, } #[derive(FixedCardinalityLabel, Copy, Clone)] #[label(singleton = "kind")] pub enum ConnectionFailureKind { ComputeCached, ComputeUncached, } #[derive(LabelGroup)] #[label(set = ConnectionFailuresBreakdownSet)] pub struct ConnectionFailuresBreakdownGroup { pub kind: ErrorKind, pub retry: Bool, } #[derive(LabelGroup, Copy, Clone)] #[label(set = RedisErrorsSet)] pub struct RedisErrors<'a> { #[label(dynamic_with = ThreadedRodeo, default)] pub channel: &'a str, } #[derive(FixedCardinalityLabel, Copy, Clone)] pub enum CancellationOutcome { NotFound, Found, RateLimitExceeded, } #[derive(LabelGroup)] #[label(set = CancellationRequestSet)] pub struct CancellationRequest { pub kind: CancellationOutcome, } #[derive(Clone, Copy)] pub enum Waiting { Cplane, Client, Compute, RetryTimeout, } #[derive(FixedCardinalityLabel, Copy, Clone)] #[label(singleton = "kind")] #[allow(clippy::enum_variant_names)] pub enum RedisMsgKind { Set, Get, Expire, HGet, } #[derive(Default, Clone)] pub struct LatencyAccumulated { pub cplane: time::Duration, pub client: time::Duration, pub compute: time::Duration, pub retry: time::Duration, } impl std::fmt::Display for LatencyAccumulated { fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!( f, "client: {}, cplane: {}, compute: {}, retry: {}", self.client.as_micros(), self.cplane.as_micros(), self.compute.as_micros(), self.retry.as_micros() ) } } pub struct LatencyTimer { // time since the stopwatch was started start: time::Instant, // time since the stopwatch was stopped stop: Option<time::Instant>, // accumulated time on the stopwatch accumulated: LatencyAccumulated, // label data protocol: Protocol, cold_start_info: ColdStartInfo, outcome: ConnectOutcome, skip_reporting: bool, } impl LatencyTimer { pub fn new(protocol: Protocol) -> Self { Self { start: time::Instant::now(), stop: None, accumulated: LatencyAccumulated::default(), protocol, cold_start_info: ColdStartInfo::Unknown, // assume failed unless otherwise specified outcome: ConnectOutcome::Failed, skip_reporting: false, } } pub(crate) fn noop(protocol: Protocol) -> Self { Self { start: time::Instant::now(), stop: None, accumulated: LatencyAccumulated::default(), protocol, cold_start_info: ColdStartInfo::Unknown, // assume failed unless otherwise specified outcome: ConnectOutcome::Failed, skip_reporting: true, } } pub fn unpause(&mut self, start: Instant, waiting_for: Waiting) { let dur = start.elapsed(); match waiting_for { Waiting::Cplane => self.accumulated.cplane += dur, Waiting::Client => self.accumulated.client += dur, Waiting::Compute => self.accumulated.compute += dur, Waiting::RetryTimeout => self.accumulated.retry += dur, } } pub fn cold_start_info(&mut self, cold_start_info: ColdStartInfo) { self.cold_start_info = cold_start_info; } pub fn success(&mut self) { // stop the stopwatch and record the time that we have accumulated self.stop = Some(time::Instant::now()); // success self.outcome = ConnectOutcome::Success; } pub fn accumulated(&self) -> LatencyAccumulated { self.accumulated.clone() } } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] pub enum ConnectOutcome { Success, Failed, } impl Drop for LatencyTimer { fn drop(&mut self) { if self.skip_reporting { return; } let duration = self .stop .unwrap_or_else(time::Instant::now) .duration_since(self.start); let metric = &Metrics::get().proxy.compute_connection_latency_seconds; // Excluding client communication from the accumulated time. metric.observe( ComputeConnectionLatencyGroup { protocol: self.protocol, cold_start_info: self.cold_start_info, outcome: self.outcome, excluded: LatencyExclusions::Client, }, duration .saturating_sub(self.accumulated.client) .as_secs_f64(), ); // Exclude client and cplane communication from the accumulated time. let accumulated_total = self.accumulated.client + self.accumulated.cplane; metric.observe( ComputeConnectionLatencyGroup { protocol: self.protocol, cold_start_info: self.cold_start_info, outcome: self.outcome, excluded: LatencyExclusions::ClientAndCplane, }, duration.saturating_sub(accumulated_total).as_secs_f64(), ); // Exclude client, cplane, compute communication from the accumulated time. let accumulated_total = self.accumulated.client + self.accumulated.cplane + self.accumulated.compute; metric.observe( ComputeConnectionLatencyGroup { protocol: self.protocol, cold_start_info: self.cold_start_info, outcome: self.outcome, excluded: LatencyExclusions::ClientCplaneCompute, }, duration.saturating_sub(accumulated_total).as_secs_f64(), ); // Exclude client, cplane, compute, retry communication from the accumulated time. let accumulated_total = self.accumulated.client + self.accumulated.cplane + self.accumulated.compute + self.accumulated.retry; metric.observe( ComputeConnectionLatencyGroup { protocol: self.protocol, cold_start_info: self.cold_start_info, outcome: self.outcome, excluded: LatencyExclusions::ClientCplaneComputeRetry, }, duration.saturating_sub(accumulated_total).as_secs_f64(), ); } } impl From<bool> for Bool { fn from(value: bool) -> Self { if value { Bool::True } else { Bool::False } } } #[derive(LabelGroup)] #[label(set = RetriesMetricSet)] pub struct RetriesMetricGroup { pub outcome: ConnectOutcome, pub retry_type: RetryType, } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] pub enum RetryType { WakeCompute, ConnectToCompute, } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] #[label(singleton = "event")] pub enum RedisEventsCount { EndpointCreated, BranchCreated, ProjectCreated, CancelSession, InvalidateRole, InvalidateEndpoint, InvalidateProject, InvalidateProjects, InvalidateOrg, } pub struct ThreadPoolWorkers(usize); #[derive(Copy, Clone)] pub struct ThreadPoolWorkerId(pub usize); impl LabelValue for ThreadPoolWorkerId { fn visit<V: measured::label::LabelVisitor>(&self, v: V) -> V::Output { v.write_int(self.0 as i64) } } impl LabelGroup for ThreadPoolWorkerId { fn visit_values(&self, v: &mut impl measured::label::LabelGroupVisitor) { v.write_value(LabelName::from_str("worker"), self); } } impl LabelGroupSet for ThreadPoolWorkers { type Group<'a> = ThreadPoolWorkerId; fn cardinality(&self) -> Option<usize> { Some(self.0) } fn encode_dense(&self, value: Self::Unique) -> Option<usize> { Some(value) } fn decode_dense(&self, value: usize) -> Self::Group<'_> { ThreadPoolWorkerId(value) } type Unique = usize; fn encode(&self, value: Self::Group<'_>) -> Option<Self::Unique> { Some(value.0) } fn decode(&self, value: &Self::Unique) -> Self::Group<'_> { ThreadPoolWorkerId(*value) } } impl LabelSet for ThreadPoolWorkers { type Value<'a> = ThreadPoolWorkerId; fn dynamic_cardinality(&self) -> Option<usize> { Some(self.0) } fn encode(&self, value: Self::Value<'_>) -> Option<usize> { (value.0 < self.0).then_some(value.0) } fn decode(&self, value: usize) -> Self::Value<'_> { ThreadPoolWorkerId(value) } } impl FixedCardinalitySet for ThreadPoolWorkers { fn cardinality(&self) -> usize { self.0 } } #[derive(MetricGroup)] #[metric(new(workers: usize))] pub struct ThreadPoolMetrics { #[metric(init = CounterVec::with_label_set(ThreadPoolWorkers(workers)))] pub worker_task_turns_total: CounterVec<ThreadPoolWorkers>, #[metric(init = CounterVec::with_label_set(ThreadPoolWorkers(workers)))] pub worker_task_skips_total: CounterVec<ThreadPoolWorkers>, } #[derive(MetricGroup, Default)] pub struct ServiceMetrics { pub info: InfoMetric<ServiceInfo>, } #[derive(Default)] pub struct ServiceInfo { pub state: ServiceState, } impl ServiceInfo { pub const fn running() -> Self { ServiceInfo { state: ServiceState::Running, } } pub const fn terminating() -> Self { ServiceInfo { state: ServiceState::Terminating, } } } impl LabelGroup for ServiceInfo { fn visit_values(&self, v: &mut impl LabelGroupVisitor) { const STATE: &LabelName = LabelName::from_str("state"); v.write_value(STATE, &self.state); } } #[derive(FixedCardinalityLabel, Clone, Copy, Debug, Default)] #[label(singleton = "state")] pub enum ServiceState { #[default] Init, Running, Terminating, } #[derive(MetricGroup)] #[metric(new())] pub struct CacheMetrics { /// The capacity of the cache pub capacity: GaugeVec<StaticLabelSet<CacheKind>>, /// The total number of entries inserted into the cache pub inserted_total: CounterVec<StaticLabelSet<CacheKind>>, /// The total number of entries removed from the cache pub evicted_total: CounterVec<CacheEvictionSet>, /// The total number of cache requests pub request_total: CounterVec<CacheOutcomeSet>, } impl Default for CacheMetrics { fn default() -> Self { Self::new() } } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] #[label(singleton = "cache")] pub enum CacheKind { NodeInfo, ProjectInfoEndpoints, ProjectInfoRoles, Schema, Pbkdf2, } #[derive(FixedCardinalityLabel, Clone, Copy, Debug)] pub enum CacheRemovalCause { Expired, Explicit, Replaced, Size, } #[derive(LabelGroup)] #[label(set = CacheEvictionSet)] pub struct CacheEviction { pub cache: CacheKind, pub cause: CacheRemovalCause, } #[derive(FixedCardinalityLabel, Copy, Clone)] pub enum CacheOutcome { Hit, Miss, } #[derive(LabelGroup)] #[label(set = CacheOutcomeSet)] pub struct CacheOutcomeGroup { pub cache: CacheKind, pub outcome: CacheOutcome, }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/signals.rs
proxy/src/signals.rs
use std::convert::Infallible; use anyhow::bail; use tokio_util::sync::CancellationToken; use tracing::{info, warn}; use crate::metrics::{Metrics, ServiceInfo}; /// Handle unix signals appropriately. pub async fn handle<F>( token: CancellationToken, mut refresh_config: F, ) -> anyhow::Result<Infallible> where F: FnMut(), { use tokio::signal::unix::{SignalKind, signal}; let mut hangup = signal(SignalKind::hangup())?; let mut interrupt = signal(SignalKind::interrupt())?; let mut terminate = signal(SignalKind::terminate())?; loop { tokio::select! { // Hangup is commonly used for config reload. _ = hangup.recv() => { info!("received SIGHUP"); refresh_config(); } // Shut down the whole application. _ = interrupt.recv() => { warn!("received SIGINT, exiting immediately"); Metrics::get().service.info.set_label(ServiceInfo::terminating()); bail!("interrupted"); } _ = terminate.recv() => { warn!("received SIGTERM, shutting down once all existing connections have closed"); Metrics::get().service.info.set_label(ServiceInfo::terminating()); token.cancel(); } } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/compute_ctl/mod.rs
proxy/src/compute_ctl/mod.rs
use compute_api::responses::GenericAPIError; use hyper::{Method, StatusCode}; use serde::de::DeserializeOwned; use serde::{Deserialize, Serialize}; use thiserror::Error; use crate::http; use crate::types::{DbName, RoleName}; use crate::url::ApiUrl; pub struct ComputeCtlApi { pub(crate) api: http::Endpoint, } #[derive(Serialize, Debug)] pub struct ExtensionInstallRequest { pub extension: &'static str, pub database: DbName, pub version: &'static str, } #[derive(Serialize, Debug)] pub struct SetRoleGrantsRequest { pub database: DbName, pub schema: &'static str, pub privileges: Vec<Privilege>, pub role: RoleName, } #[derive(Clone, Debug, Deserialize)] pub struct ExtensionInstallResponse {} #[derive(Clone, Debug, Deserialize)] pub struct SetRoleGrantsResponse {} #[derive(Debug, Serialize, Deserialize, Clone, Copy)] #[serde(rename_all = "UPPERCASE")] pub enum Privilege { Usage, } #[derive(Error, Debug)] pub enum ComputeCtlError { #[error("connection error: {0}")] Connection(#[source] reqwest_middleware::Error), #[error("request error [{status}]: {body:?}")] Request { status: StatusCode, body: Option<GenericAPIError>, }, #[error("response parsing error: {0}")] Response(#[source] reqwest::Error), } impl ComputeCtlApi { pub async fn install_extension( &self, req: &ExtensionInstallRequest, ) -> Result<ExtensionInstallResponse, ComputeCtlError> { self.generic_request(req, Method::POST, |url| { url.path_segments_mut().push("extensions"); }) .await } pub async fn grant_role( &self, req: &SetRoleGrantsRequest, ) -> Result<SetRoleGrantsResponse, ComputeCtlError> { self.generic_request(req, Method::POST, |url| { url.path_segments_mut().push("grants"); }) .await } async fn generic_request<Req, Resp>( &self, req: &Req, method: Method, url: impl for<'a> FnOnce(&'a mut ApiUrl), ) -> Result<Resp, ComputeCtlError> where Req: Serialize, Resp: DeserializeOwned, { let resp = self .api .request_with_url(method, url) .json(req) .send() .await .map_err(ComputeCtlError::Connection)?; let status = resp.status(); if status.is_client_error() || status.is_server_error() { let body = resp.json().await.ok(); return Err(ComputeCtlError::Request { status, body }); } resp.json().await.map_err(ComputeCtlError::Response) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/flow.rs
proxy/src/auth/flow.rs
//! Main authentication flow. use std::sync::Arc; use postgres_protocol::authentication::sasl::{SCRAM_SHA_256, SCRAM_SHA_256_PLUS}; use tokio::io::{AsyncRead, AsyncWrite}; use tracing::info; use super::backend::ComputeCredentialKeys; use super::{AuthError, PasswordHackPayload}; use crate::context::RequestContext; use crate::control_plane::AuthSecret; use crate::intern::{EndpointIdInt, RoleNameInt}; use crate::pqproto::{BeAuthenticationSaslMessage, BeMessage}; use crate::sasl; use crate::scram::threadpool::ThreadPool; use crate::scram::{self}; use crate::stream::{PqStream, Stream}; use crate::tls::TlsServerEndPoint; /// Use [SCRAM](crate::scram)-based auth in [`AuthFlow`]. pub(crate) struct Scram<'a>( pub(crate) &'a scram::ServerSecret, pub(crate) &'a RequestContext, ); impl Scram<'_> { #[inline(always)] fn first_message(&self, channel_binding: bool) -> BeMessage<'_> { if channel_binding { BeMessage::AuthenticationSasl(BeAuthenticationSaslMessage::Methods(scram::METHODS)) } else { BeMessage::AuthenticationSasl(BeAuthenticationSaslMessage::Methods( scram::METHODS_WITHOUT_PLUS, )) } } } /// Use an ad hoc auth flow (for clients which don't support SNI) proposed in /// <https://github.com/neondatabase/cloud/issues/1620#issuecomment-1165332290>. pub(crate) struct PasswordHack; /// Use clear-text password auth called `password` in docs /// <https://www.postgresql.org/docs/current/auth-password.html> pub(crate) struct CleartextPassword { pub(crate) pool: Arc<ThreadPool>, pub(crate) endpoint: EndpointIdInt, pub(crate) role: RoleNameInt, pub(crate) secret: AuthSecret, } /// This wrapper for [`PqStream`] performs client authentication. #[must_use] pub(crate) struct AuthFlow<'a, S, State> { /// The underlying stream which implements libpq's protocol. stream: &'a mut PqStream<Stream<S>>, /// State might contain ancillary data. state: State, tls_server_end_point: TlsServerEndPoint, } /// Initial state of the stream wrapper. impl<'a, S: AsyncRead + AsyncWrite + Unpin, M> AuthFlow<'a, S, M> { /// Create a new wrapper for client authentication. pub(crate) fn new(stream: &'a mut PqStream<Stream<S>>, method: M) -> Self { let tls_server_end_point = stream.get_ref().tls_server_end_point(); Self { stream, state: method, tls_server_end_point, } } } impl<S: AsyncRead + AsyncWrite + Unpin> AuthFlow<'_, S, PasswordHack> { /// Perform user authentication. Raise an error in case authentication failed. pub(crate) async fn get_password(self) -> super::Result<PasswordHackPayload> { self.stream .write_message(BeMessage::AuthenticationCleartextPassword); self.stream.flush().await?; let msg = self.stream.read_password_message().await?; let password = msg .strip_suffix(&[0]) .ok_or(AuthError::MalformedPassword("missing terminator"))?; let payload = PasswordHackPayload::parse(password) // If we ended up here and the payload is malformed, it means that // the user neither enabled SNI nor resorted to any other method // for passing the project name we rely on. We should show them // the most helpful error message and point to the documentation. .ok_or(AuthError::MissingEndpointName)?; Ok(payload) } } impl<S: AsyncRead + AsyncWrite + Unpin> AuthFlow<'_, S, CleartextPassword> { /// Perform user authentication. Raise an error in case authentication failed. pub(crate) async fn authenticate(self) -> super::Result<sasl::Outcome<ComputeCredentialKeys>> { self.stream .write_message(BeMessage::AuthenticationCleartextPassword); self.stream.flush().await?; let msg = self.stream.read_password_message().await?; let password = msg .strip_suffix(&[0]) .ok_or(AuthError::MalformedPassword("missing terminator"))?; let outcome = validate_password_and_exchange( &self.state.pool, self.state.endpoint, self.state.role, password, self.state.secret, ) .await?; if let sasl::Outcome::Success(_) = &outcome { self.stream.write_message(BeMessage::AuthenticationOk); } Ok(outcome) } } /// Stream wrapper for handling [SCRAM](crate::scram) auth. impl<S: AsyncRead + AsyncWrite + Unpin> AuthFlow<'_, S, Scram<'_>> { /// Perform user authentication. Raise an error in case authentication failed. pub(crate) async fn authenticate(self) -> super::Result<sasl::Outcome<scram::ScramKey>> { let Scram(secret, ctx) = self.state; let channel_binding = self.tls_server_end_point; // send sasl message. { // pause the timer while we communicate with the client let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Client); let sasl = self.state.first_message(channel_binding.supported()); self.stream.write_message(sasl); self.stream.flush().await?; } // complete sasl handshake. sasl::authenticate(ctx, self.stream, |method| { // Currently, the only supported SASL method is SCRAM. match method { SCRAM_SHA_256 => ctx.set_auth_method(crate::context::AuthMethod::ScramSha256), SCRAM_SHA_256_PLUS => { ctx.set_auth_method(crate::context::AuthMethod::ScramSha256Plus); } method => return Err(sasl::Error::BadAuthMethod(method.into())), } // TODO: make this a metric instead info!("client chooses {}", method); Ok(scram::Exchange::new(secret, rand::random, channel_binding)) }) .await .map_err(AuthError::Sasl) } } pub(crate) async fn validate_password_and_exchange( pool: &ThreadPool, endpoint: EndpointIdInt, role: RoleNameInt, password: &[u8], secret: AuthSecret, ) -> super::Result<sasl::Outcome<ComputeCredentialKeys>> { match secret { // perform scram authentication as both client and server to validate the keys AuthSecret::Scram(scram_secret) => { let outcome = crate::scram::exchange(pool, endpoint, role, &scram_secret, password).await?; let client_key = match outcome { sasl::Outcome::Success(client_key) => client_key, sasl::Outcome::Failure(reason) => return Ok(sasl::Outcome::Failure(reason)), }; let keys = crate::compute::ScramKeys { client_key: client_key.as_bytes(), server_key: scram_secret.server_key.as_bytes(), }; Ok(sasl::Outcome::Success(ComputeCredentialKeys::AuthKeys( postgres_client::config::AuthKeys::ScramSha256(keys), ))) } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/password_hack.rs
proxy/src/auth/password_hack.rs
//! Payload for ad hoc authentication method for clients that don't support SNI. //! See the `impl` for [`super::backend::Backend<ClientCredentials>`]. //! Read more: <https://github.com/neondatabase/cloud/issues/1620#issuecomment-1165332290>. //! UPDATE (Mon Aug 8 13:20:34 UTC 2022): the payload format has been simplified. use bstr::ByteSlice; use crate::types::EndpointId; pub(crate) struct PasswordHackPayload { pub(crate) endpoint: EndpointId, pub(crate) password: Vec<u8>, } impl PasswordHackPayload { pub(crate) fn parse(bytes: &[u8]) -> Option<Self> { // The format is `project=<utf-8>;<password-bytes>` or `project=<utf-8>$<password-bytes>`. let separators = [";", "$"]; for sep in separators { if let Some((endpoint, password)) = bytes.split_once_str(sep) { let endpoint = endpoint.to_str().ok()?; return Some(Self { endpoint: parse_endpoint_param(endpoint)?.into(), password: password.to_owned(), }); } } None } } pub(crate) fn parse_endpoint_param(bytes: &str) -> Option<&str> { bytes .strip_prefix("project=") .or_else(|| bytes.strip_prefix("endpoint=")) } #[cfg(test)] mod tests { use super::*; #[test] fn parse_endpoint_param_fn() { let input = ""; assert!(parse_endpoint_param(input).is_none()); let input = "project="; assert_eq!(parse_endpoint_param(input), Some("")); let input = "project=foobar"; assert_eq!(parse_endpoint_param(input), Some("foobar")); let input = "endpoint="; assert_eq!(parse_endpoint_param(input), Some("")); let input = "endpoint=foobar"; assert_eq!(parse_endpoint_param(input), Some("foobar")); let input = "other_option=foobar"; assert!(parse_endpoint_param(input).is_none()); } #[test] fn parse_password_hack_payload_project() { let bytes = b""; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"project="; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"project=;"; let payload: PasswordHackPayload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, ""); assert_eq!(payload.password, b""); let bytes = b"project=foobar;pass;word"; let payload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, "foobar"); assert_eq!(payload.password, b"pass;word"); } #[test] fn parse_password_hack_payload_endpoint() { let bytes = b""; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"endpoint="; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"endpoint=;"; let payload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, ""); assert_eq!(payload.password, b""); let bytes = b"endpoint=foobar;pass;word"; let payload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, "foobar"); assert_eq!(payload.password, b"pass;word"); } #[test] fn parse_password_hack_payload_dollar() { let bytes = b""; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"endpoint="; assert!(PasswordHackPayload::parse(bytes).is_none()); let bytes = b"endpoint=$"; let payload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, ""); assert_eq!(payload.password, b""); let bytes = b"endpoint=foobar$pass$word"; let payload = PasswordHackPayload::parse(bytes).expect("parsing failed"); assert_eq!(payload.endpoint, "foobar"); assert_eq!(payload.password, b"pass$word"); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/mod.rs
proxy/src/auth/mod.rs
//! Client authentication mechanisms. pub mod backend; pub use backend::Backend; mod credentials; pub(crate) use credentials::{ ComputeUserInfoMaybeEndpoint, ComputeUserInfoParseError, IpPattern, check_peer_addr_is_in_list, endpoint_sni, }; mod password_hack; use password_hack::PasswordHackPayload; pub(crate) use password_hack::parse_endpoint_param; mod flow; use std::io; use std::net::IpAddr; pub(crate) use flow::*; use thiserror::Error; use tokio::time::error::Elapsed; use crate::auth::backend::jwt::JwtError; use crate::control_plane; use crate::error::{ReportableError, UserFacingError}; /// Convenience wrapper for the authentication error. pub(crate) type Result<T> = std::result::Result<T, AuthError>; /// Common authentication error. #[derive(Debug, Error)] pub(crate) enum AuthError { #[error(transparent)] ConsoleRedirect(#[from] backend::ConsoleRedirectError), #[error(transparent)] GetAuthInfo(#[from] control_plane::errors::GetAuthInfoError), /// SASL protocol errors (includes [SCRAM](crate::scram)). #[error(transparent)] Sasl(#[from] crate::sasl::Error), #[error("Unsupported authentication method: {0}")] BadAuthMethod(Box<str>), #[error("Malformed password message: {0}")] MalformedPassword(&'static str), #[error( "Endpoint ID is not specified. \ Either please upgrade the postgres client library (libpq) for SNI support \ or pass the endpoint ID (first part of the domain name) as a parameter: '?options=endpoint%3D<endpoint-id>'. \ See more at https://neon.tech/sni" )] MissingEndpointName, #[error( "VPC endpoint ID is not specified. \ This endpoint requires a VPC endpoint ID to connect." )] MissingVPCEndpointId, #[error("password authentication failed for user '{0}'")] PasswordFailed(Box<str>), /// Errors produced by e.g. [`crate::stream::PqStream`]. #[error(transparent)] Io(#[from] io::Error), #[error( "This IP address {0} is not allowed to connect to this endpoint. \ Please add it to the allowed list in the Neon console. \ Make sure to check for IPv4 or IPv6 addresses." )] IpAddressNotAllowed(IpAddr), #[error("This connection is trying to access this endpoint from a blocked network.")] NetworkNotAllowed, #[error( "This VPC endpoint id {0} is not allowed to connect to this endpoint. \ Please add it to the allowed list in the Neon console." )] VpcEndpointIdNotAllowed(String), #[error("Too many connections to this endpoint. Please try again later.")] TooManyConnections, #[error("Authentication timed out")] UserTimeout(Elapsed), #[error("Disconnected due to inactivity after {0}.")] ConfirmationTimeout(humantime::Duration), #[error(transparent)] Jwt(#[from] JwtError), } impl AuthError { pub(crate) fn bad_auth_method(name: impl Into<Box<str>>) -> Self { AuthError::BadAuthMethod(name.into()) } pub(crate) fn password_failed(user: impl Into<Box<str>>) -> Self { AuthError::PasswordFailed(user.into()) } pub(crate) fn ip_address_not_allowed(ip: IpAddr) -> Self { AuthError::IpAddressNotAllowed(ip) } pub(crate) fn vpc_endpoint_id_not_allowed(id: String) -> Self { AuthError::VpcEndpointIdNotAllowed(id) } pub(crate) fn too_many_connections() -> Self { AuthError::TooManyConnections } pub(crate) fn is_password_failed(&self) -> bool { matches!(self, AuthError::PasswordFailed(_)) } pub(crate) fn user_timeout(elapsed: Elapsed) -> Self { AuthError::UserTimeout(elapsed) } pub(crate) fn confirmation_timeout(timeout: humantime::Duration) -> Self { AuthError::ConfirmationTimeout(timeout) } } impl UserFacingError for AuthError { fn to_string_client(&self) -> String { match self { Self::ConsoleRedirect(e) => e.to_string_client(), Self::GetAuthInfo(e) => e.to_string_client(), Self::Sasl(e) => e.to_string_client(), Self::PasswordFailed(_) => self.to_string(), Self::BadAuthMethod(_) => self.to_string(), Self::MalformedPassword(_) => self.to_string(), Self::MissingEndpointName => self.to_string(), Self::MissingVPCEndpointId => self.to_string(), Self::Io(_) => "Internal error".to_string(), Self::IpAddressNotAllowed(_) => self.to_string(), Self::NetworkNotAllowed => self.to_string(), Self::VpcEndpointIdNotAllowed(_) => self.to_string(), Self::TooManyConnections => self.to_string(), Self::UserTimeout(_) => self.to_string(), Self::ConfirmationTimeout(_) => self.to_string(), Self::Jwt(_) => self.to_string(), } } } impl ReportableError for AuthError { fn get_error_kind(&self) -> crate::error::ErrorKind { match self { Self::ConsoleRedirect(e) => e.get_error_kind(), Self::GetAuthInfo(e) => e.get_error_kind(), Self::Sasl(e) => e.get_error_kind(), Self::PasswordFailed(_) => crate::error::ErrorKind::User, Self::BadAuthMethod(_) => crate::error::ErrorKind::User, Self::MalformedPassword(_) => crate::error::ErrorKind::User, Self::MissingEndpointName => crate::error::ErrorKind::User, Self::MissingVPCEndpointId => crate::error::ErrorKind::User, Self::Io(_) => crate::error::ErrorKind::ClientDisconnect, Self::IpAddressNotAllowed(_) => crate::error::ErrorKind::User, Self::NetworkNotAllowed => crate::error::ErrorKind::User, Self::VpcEndpointIdNotAllowed(_) => crate::error::ErrorKind::User, Self::TooManyConnections => crate::error::ErrorKind::RateLimit, Self::UserTimeout(_) => crate::error::ErrorKind::User, Self::ConfirmationTimeout(_) => crate::error::ErrorKind::User, Self::Jwt(_) => crate::error::ErrorKind::User, } } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/credentials.rs
proxy/src/auth/credentials.rs
//! User credentials used in authentication. use std::collections::HashSet; use std::net::IpAddr; use std::str::FromStr; use itertools::Itertools; use thiserror::Error; use tracing::{debug, warn}; use crate::auth::password_hack::parse_endpoint_param; use crate::context::RequestContext; use crate::error::{ReportableError, UserFacingError}; use crate::metrics::{Metrics, SniGroup, SniKind}; use crate::pqproto::StartupMessageParams; use crate::proxy::NeonOptions; use crate::serverless::{AUTH_BROKER_SNI, SERVERLESS_DRIVER_SNI}; use crate::types::{EndpointId, RoleName}; #[derive(Debug, Error, PartialEq, Eq, Clone)] pub(crate) enum ComputeUserInfoParseError { #[error("Parameter '{0}' is missing in startup packet.")] MissingKey(&'static str), #[error( "Inconsistent project name inferred from \ SNI ('{}') and project option ('{}').", .domain, .option, )] InconsistentProjectNames { domain: EndpointId, option: EndpointId, }, #[error("Project name ('{0}') must contain only alphanumeric characters and hyphen.")] MalformedProjectName(EndpointId), } impl UserFacingError for ComputeUserInfoParseError {} impl ReportableError for ComputeUserInfoParseError { fn get_error_kind(&self) -> crate::error::ErrorKind { crate::error::ErrorKind::User } } /// Various client credentials which we use for authentication. /// Note that we don't store any kind of client key or password here. #[derive(Debug, Clone, PartialEq, Eq)] pub(crate) struct ComputeUserInfoMaybeEndpoint { pub(crate) user: RoleName, pub(crate) endpoint_id: Option<EndpointId>, pub(crate) options: NeonOptions, } impl ComputeUserInfoMaybeEndpoint { #[inline] pub(crate) fn endpoint(&self) -> Option<&str> { self.endpoint_id.as_deref() } } pub(crate) fn endpoint_sni(sni: &str, common_names: &HashSet<String>) -> Option<EndpointId> { let (subdomain, common_name) = sni.split_once('.')?; if !common_names.contains(common_name) { return None; } if subdomain == SERVERLESS_DRIVER_SNI || subdomain == AUTH_BROKER_SNI { return None; } Some(EndpointId::from(subdomain)) } impl ComputeUserInfoMaybeEndpoint { pub(crate) fn parse( ctx: &RequestContext, params: &StartupMessageParams, sni: Option<&str>, common_names: Option<&HashSet<String>>, ) -> Result<Self, ComputeUserInfoParseError> { // Some parameters are stored in the startup message. let get_param = |key| { params .get(key) .ok_or(ComputeUserInfoParseError::MissingKey(key)) }; let user: RoleName = get_param("user")?.into(); // Project name might be passed via PG's command-line options. let endpoint_option = params .options_raw() .and_then(|options| { // We support both `project` (deprecated) and `endpoint` options for backward compatibility. // However, if both are present, we don't exactly know which one to use. // Therefore we require that only one of them is present. options .filter_map(parse_endpoint_param) .at_most_one() .ok()? }) .map(|name| name.into()); let endpoint_from_domain = sni.and_then(|sni_str| common_names.and_then(|cn| endpoint_sni(sni_str, cn))); let endpoint = match (endpoint_option, endpoint_from_domain) { // Invariant: if we have both project name variants, they should match. (Some(option), Some(domain)) if option != domain => { Some(Err(ComputeUserInfoParseError::InconsistentProjectNames { domain, option, })) } // Invariant: project name may not contain certain characters. (a, b) => a.or(b).map(|name| { if project_name_valid(name.as_ref()) { Ok(name) } else { Err(ComputeUserInfoParseError::MalformedProjectName(name)) } }), } .transpose()?; if let Some(ep) = &endpoint { ctx.set_endpoint_id(ep.clone()); } let metrics = Metrics::get(); debug!(%user, "credentials"); let protocol = ctx.protocol(); let kind = if sni.is_some() { debug!("Connection with sni"); SniKind::Sni } else if endpoint.is_some() { debug!("Connection without sni"); SniKind::NoSni } else { debug!("Connection with password hack"); SniKind::PasswordHack }; metrics .proxy .accepted_connections_by_sni .inc(SniGroup { protocol, kind }); let options = NeonOptions::parse_params(params); Ok(Self { user, endpoint_id: endpoint, options, }) } } pub(crate) fn check_peer_addr_is_in_list(peer_addr: &IpAddr, ip_list: &[IpPattern]) -> bool { ip_list.is_empty() || ip_list.iter().any(|pattern| check_ip(peer_addr, pattern)) } #[derive(Debug, Clone, Eq, PartialEq)] pub(crate) enum IpPattern { Subnet(ipnet::IpNet), Range(IpAddr, IpAddr), Single(IpAddr), None, } impl<'de> serde::de::Deserialize<'de> for IpPattern { fn deserialize<D>(deserializer: D) -> Result<Self, D::Error> where D: serde::Deserializer<'de>, { struct StrVisitor; impl serde::de::Visitor<'_> for StrVisitor { type Value = IpPattern; fn expecting(&self, formatter: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { write!( formatter, "comma separated list with ip address, ip address range, or ip address subnet mask" ) } fn visit_str<E>(self, v: &str) -> Result<Self::Value, E> where E: serde::de::Error, { Ok(parse_ip_pattern(v).unwrap_or_else(|e| { warn!("Cannot parse ip pattern {v}: {e}"); IpPattern::None })) } } deserializer.deserialize_str(StrVisitor) } } impl FromStr for IpPattern { type Err = anyhow::Error; fn from_str(s: &str) -> Result<Self, Self::Err> { parse_ip_pattern(s) } } fn parse_ip_pattern(pattern: &str) -> anyhow::Result<IpPattern> { if pattern.contains('/') { let subnet: ipnet::IpNet = pattern.parse()?; return Ok(IpPattern::Subnet(subnet)); } if let Some((start, end)) = pattern.split_once('-') { let start: IpAddr = start.parse()?; let end: IpAddr = end.parse()?; return Ok(IpPattern::Range(start, end)); } let addr: IpAddr = pattern.parse()?; Ok(IpPattern::Single(addr)) } fn check_ip(ip: &IpAddr, pattern: &IpPattern) -> bool { match pattern { IpPattern::Subnet(subnet) => subnet.contains(ip), IpPattern::Range(start, end) => start <= ip && ip <= end, IpPattern::Single(addr) => addr == ip, IpPattern::None => false, } } fn project_name_valid(name: &str) -> bool { name.chars().all(|c| c.is_alphanumeric() || c == '-') } #[cfg(test)] mod tests { use ComputeUserInfoParseError::*; use serde_json::json; use super::*; #[test] fn parse_bare_minimum() -> anyhow::Result<()> { // According to postgresql, only `user` should be required. let options = StartupMessageParams::new([("user", "john_doe")]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id, None); Ok(()) } #[test] fn parse_excessive() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ("database", "world"), // should be ignored ("foo", "bar"), // should be ignored ]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id, None); Ok(()) } #[test] fn parse_project_from_sni() -> anyhow::Result<()> { let options = StartupMessageParams::new([("user", "john_doe")]); let sni = Some("foo.localhost"); let common_names = Some(["localhost".into()].into()); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref())?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id.as_deref(), Some("foo")); assert_eq!(user_info.options.get_cache_key("foo"), "foo"); Ok(()) } #[test] fn parse_project_from_options() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ("options", "-ckey=1 project=bar -c geqo=off"), ]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id.as_deref(), Some("bar")); Ok(()) } #[test] fn parse_endpoint_from_options() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ("options", "-ckey=1 endpoint=bar -c geqo=off"), ]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id.as_deref(), Some("bar")); Ok(()) } #[test] fn parse_three_endpoints_from_options() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ( "options", "-ckey=1 endpoint=one endpoint=two endpoint=three -c geqo=off", ), ]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert!(user_info.endpoint_id.is_none()); Ok(()) } #[test] fn parse_when_endpoint_and_project_are_in_options() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ("options", "-ckey=1 endpoint=bar project=foo -c geqo=off"), ]); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, None, None)?; assert_eq!(user_info.user, "john_doe"); assert!(user_info.endpoint_id.is_none()); Ok(()) } #[test] fn parse_projects_identical() -> anyhow::Result<()> { let options = StartupMessageParams::new([("user", "john_doe"), ("options", "project=baz")]); let sni = Some("baz.localhost"); let common_names = Some(["localhost".into()].into()); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref())?; assert_eq!(user_info.user, "john_doe"); assert_eq!(user_info.endpoint_id.as_deref(), Some("baz")); Ok(()) } #[test] fn parse_multi_common_names() -> anyhow::Result<()> { let options = StartupMessageParams::new([("user", "john_doe")]); let common_names = Some(["a.com".into(), "b.com".into()].into()); let sni = Some("p1.a.com"); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref())?; assert_eq!(user_info.endpoint_id.as_deref(), Some("p1")); let common_names = Some(["a.com".into(), "b.com".into()].into()); let sni = Some("p1.b.com"); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref())?; assert_eq!(user_info.endpoint_id.as_deref(), Some("p1")); Ok(()) } #[test] fn parse_projects_different() { let options = StartupMessageParams::new([("user", "john_doe"), ("options", "project=first")]); let sni = Some("second.localhost"); let common_names = Some(["localhost".into()].into()); let ctx = RequestContext::test(); let err = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref()) .expect_err("should fail"); match err { InconsistentProjectNames { domain, option } => { assert_eq!(option, "first"); assert_eq!(domain, "second"); } _ => panic!("bad error: {err:?}"), } } #[test] fn parse_unknown_sni() { let options = StartupMessageParams::new([("user", "john_doe")]); let sni = Some("project.localhost"); let common_names = Some(["example.com".into()].into()); let ctx = RequestContext::test(); let info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref()) .unwrap(); assert!(info.endpoint_id.is_none()); } #[test] fn parse_unknown_sni_with_options() { let options = StartupMessageParams::new([ ("user", "john_doe"), ("options", "endpoint=foo-bar-baz-1234"), ]); let sni = Some("project.localhost"); let common_names = Some(["example.com".into()].into()); let ctx = RequestContext::test(); let info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref()) .unwrap(); assert_eq!(info.endpoint_id.as_deref(), Some("foo-bar-baz-1234")); } #[test] fn parse_neon_options() -> anyhow::Result<()> { let options = StartupMessageParams::new([ ("user", "john_doe"), ("options", "neon_lsn:0/2 neon_endpoint_type:read_write"), ]); let sni = Some("project.localhost"); let common_names = Some(["localhost".into()].into()); let ctx = RequestContext::test(); let user_info = ComputeUserInfoMaybeEndpoint::parse(&ctx, &options, sni, common_names.as_ref())?; assert_eq!(user_info.endpoint_id.as_deref(), Some("project")); assert_eq!( user_info.options.get_cache_key("project"), "project endpoint_type:read_write lsn:0/2" ); Ok(()) } #[test] fn test_check_peer_addr_is_in_list() { fn check(v: serde_json::Value) -> bool { let peer_addr = IpAddr::from([127, 0, 0, 1]); let ip_list: Vec<IpPattern> = serde_json::from_value(v).unwrap(); check_peer_addr_is_in_list(&peer_addr, &ip_list) } assert!(check(json!([]))); assert!(check(json!(["127.0.0.1"]))); assert!(!check(json!(["8.8.8.8"]))); // If there is an incorrect address, it will be skipped. assert!(check(json!(["88.8.8", "127.0.0.1"]))); } #[test] fn test_parse_ip_v4() -> anyhow::Result<()> { let peer_addr = IpAddr::from([127, 0, 0, 1]); // Ok assert_eq!(parse_ip_pattern("127.0.0.1")?, IpPattern::Single(peer_addr)); assert_eq!( parse_ip_pattern("127.0.0.1/31")?, IpPattern::Subnet(ipnet::IpNet::new(peer_addr, 31)?) ); assert_eq!( parse_ip_pattern("0.0.0.0-200.0.1.2")?, IpPattern::Range(IpAddr::from([0, 0, 0, 0]), IpAddr::from([200, 0, 1, 2])) ); // Error assert!(parse_ip_pattern("300.0.1.2").is_err()); assert!(parse_ip_pattern("30.1.2").is_err()); assert!(parse_ip_pattern("127.0.0.1/33").is_err()); assert!(parse_ip_pattern("127.0.0.1-127.0.3").is_err()); assert!(parse_ip_pattern("1234.0.0.1-127.0.3.0").is_err()); Ok(()) } #[test] fn test_check_ipv4() -> anyhow::Result<()> { let peer_addr = IpAddr::from([127, 0, 0, 1]); let peer_addr_next = IpAddr::from([127, 0, 0, 2]); let peer_addr_prev = IpAddr::from([127, 0, 0, 0]); // Success assert!(check_ip(&peer_addr, &IpPattern::Single(peer_addr))); assert!(check_ip( &peer_addr, &IpPattern::Subnet(ipnet::IpNet::new(peer_addr_prev, 31)?) )); assert!(check_ip( &peer_addr, &IpPattern::Subnet(ipnet::IpNet::new(peer_addr_next, 30)?) )); assert!(check_ip( &peer_addr, &IpPattern::Range(IpAddr::from([0, 0, 0, 0]), IpAddr::from([200, 0, 1, 2])) )); assert!(check_ip( &peer_addr, &IpPattern::Range(peer_addr, peer_addr) )); // Not success assert!(!check_ip(&peer_addr, &IpPattern::Single(peer_addr_prev))); assert!(!check_ip( &peer_addr, &IpPattern::Subnet(ipnet::IpNet::new(peer_addr_next, 31)?) )); assert!(!check_ip( &peer_addr, &IpPattern::Range(IpAddr::from([0, 0, 0, 0]), peer_addr_prev) )); assert!(!check_ip( &peer_addr, &IpPattern::Range(peer_addr_next, IpAddr::from([128, 0, 0, 0])) )); // There is no check that for range start <= end. But it's fine as long as for all this cases the result is false. assert!(!check_ip( &peer_addr, &IpPattern::Range(peer_addr, peer_addr_prev) )); Ok(()) } #[test] fn test_connection_blocker() { fn check(v: serde_json::Value) -> bool { let peer_addr = IpAddr::from([127, 0, 0, 1]); let ip_list: Vec<IpPattern> = serde_json::from_value(v).unwrap(); check_peer_addr_is_in_list(&peer_addr, &ip_list) } assert!(check(json!([]))); assert!(check(json!(["127.0.0.1"]))); assert!(!check(json!(["255.255.255.255"]))); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/local.rs
proxy/src/auth/backend/local.rs
use std::net::SocketAddr; use arc_swap::ArcSwapOption; use postgres_client::config::SslMode; use tokio::sync::Semaphore; use super::jwt::{AuthRule, FetchAuthRules}; use crate::auth::backend::jwt::FetchAuthRulesError; use crate::compute::ConnectInfo; use crate::compute_ctl::ComputeCtlApi; use crate::context::RequestContext; use crate::control_plane::NodeInfo; use crate::control_plane::messages::{ColdStartInfo, EndpointJwksResponse, MetricsAuxInfo}; use crate::http; use crate::intern::{BranchIdTag, EndpointIdTag, InternId, ProjectIdTag}; use crate::types::EndpointId; use crate::url::ApiUrl; pub struct LocalBackend { pub(crate) initialize: Semaphore, pub(crate) compute_ctl: ComputeCtlApi, pub(crate) node_info: NodeInfo, } impl LocalBackend { pub fn new(postgres_addr: SocketAddr, compute_ctl: ApiUrl) -> Self { LocalBackend { initialize: Semaphore::new(1), compute_ctl: ComputeCtlApi { api: http::Endpoint::new(compute_ctl, http::new_client()), }, node_info: NodeInfo { conn_info: ConnectInfo { host_addr: Some(postgres_addr.ip()), host: postgres_addr.ip().to_string().into(), port: postgres_addr.port(), ssl_mode: SslMode::Disable, }, // TODO(conrad): make this better reflect compute info rather than endpoint info. aux: MetricsAuxInfo { endpoint_id: EndpointIdTag::get_interner().get_or_intern("local"), project_id: ProjectIdTag::get_interner().get_or_intern("local"), branch_id: BranchIdTag::get_interner().get_or_intern("local"), compute_id: "local".into(), cold_start_info: ColdStartInfo::WarmCached, }, }, } } } #[derive(Clone, Copy)] pub(crate) struct StaticAuthRules; pub static JWKS_ROLE_MAP: ArcSwapOption<EndpointJwksResponse> = ArcSwapOption::const_empty(); impl FetchAuthRules for StaticAuthRules { async fn fetch_auth_rules( &self, _ctx: &RequestContext, _endpoint: EndpointId, ) -> Result<Vec<AuthRule>, FetchAuthRulesError> { let mappings = JWKS_ROLE_MAP.load(); let role_mappings = mappings .as_deref() .ok_or(FetchAuthRulesError::RoleJwksNotConfigured)?; let mut rules = vec![]; for setting in &role_mappings.jwks { rules.push(AuthRule { id: setting.id.clone(), jwks_url: setting.jwks_url.clone(), audience: setting.jwt_audience.clone(), role_names: setting.role_names.clone(), }); } Ok(rules) } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/jwt.rs
proxy/src/auth/backend/jwt.rs
use std::borrow::Cow; use std::future::Future; use std::sync::Arc; use std::time::{Duration, SystemTime}; use arc_swap::ArcSwapOption; use base64::Engine as _; use base64::prelude::BASE64_URL_SAFE_NO_PAD; use clashmap::ClashMap; use jose_jwk::crypto::KeyInfo; use reqwest::{Client, redirect}; use reqwest_retry::RetryTransientMiddleware; use reqwest_retry::policies::ExponentialBackoff; use serde::de::Visitor; use serde::{Deserialize, Deserializer}; use serde_json::value::RawValue; use signature::Verifier; use thiserror::Error; use tokio::time::Instant; use crate::auth::backend::ComputeCredentialKeys; use crate::context::RequestContext; use crate::control_plane::errors::GetEndpointJwksError; use crate::http::read_body_with_limit; use crate::intern::RoleNameInt; use crate::types::{EndpointId, RoleName}; // TODO(conrad): make these configurable. const CLOCK_SKEW_LEEWAY: Duration = Duration::from_secs(30); const MIN_RENEW: Duration = Duration::from_secs(30); const AUTO_RENEW: Duration = Duration::from_secs(300); const MAX_RENEW: Duration = Duration::from_secs(3600); const MAX_JWK_BODY_SIZE: usize = 64 * 1024; const JWKS_USER_AGENT: &str = "neon-proxy"; const JWKS_CONNECT_TIMEOUT: Duration = Duration::from_secs(2); const JWKS_FETCH_TIMEOUT: Duration = Duration::from_secs(5); const JWKS_FETCH_RETRIES: u32 = 3; /// How to get the JWT auth rules pub(crate) trait FetchAuthRules: Clone + Send + Sync + 'static { fn fetch_auth_rules( &self, ctx: &RequestContext, endpoint: EndpointId, ) -> impl Future<Output = Result<Vec<AuthRule>, FetchAuthRulesError>> + Send; } #[derive(Error, Debug)] pub(crate) enum FetchAuthRulesError { #[error(transparent)] GetEndpointJwks(#[from] GetEndpointJwksError), #[error("JWKs settings for this role were not configured")] RoleJwksNotConfigured, } #[derive(Clone)] pub(crate) struct AuthRule { pub(crate) id: String, pub(crate) jwks_url: url::Url, pub(crate) audience: Option<String>, pub(crate) role_names: Vec<RoleNameInt>, } pub struct JwkCache { client: reqwest_middleware::ClientWithMiddleware, map: ClashMap<(EndpointId, RoleName), Arc<JwkCacheEntryLock>>, } pub(crate) struct JwkCacheEntry { /// Should refetch at least every hour to verify when old keys have been removed. /// Should refetch when new key IDs are seen only every 5 minutes or so last_retrieved: Instant, /// cplane will return multiple JWKs urls that we need to scrape. key_sets: ahash::HashMap<String, KeySet>, } impl JwkCacheEntry { fn find_jwk_and_audience( &self, key_id: &str, role_name: &RoleName, ) -> Option<(&jose_jwk::Jwk, Option<&str>)> { self.key_sets .values() // make sure our requested role has access to the key set .filter(|key_set| key_set.role_names.iter().any(|role| **role == **role_name)) // try and find the requested key-id in the key set .find_map(|key_set| { key_set .find_key(key_id) .map(|jwk| (jwk, key_set.audience.as_deref())) }) } } struct KeySet { jwks: jose_jwk::JwkSet, audience: Option<String>, role_names: Vec<RoleNameInt>, } impl KeySet { fn find_key(&self, key_id: &str) -> Option<&jose_jwk::Jwk> { self.jwks .keys .iter() .find(|jwk| jwk.prm.kid.as_deref() == Some(key_id)) } } pub(crate) struct JwkCacheEntryLock { cached: ArcSwapOption<JwkCacheEntry>, lookup: tokio::sync::Semaphore, } impl Default for JwkCacheEntryLock { fn default() -> Self { JwkCacheEntryLock { cached: ArcSwapOption::empty(), lookup: tokio::sync::Semaphore::new(1), } } } #[derive(Deserialize)] struct JwkSet<'a> { /// we parse into raw-value because not all keys in a JWKS are ones /// we can parse directly, so we parse them lazily. #[serde(borrow)] keys: Vec<&'a RawValue>, } /// Given a jwks_url, fetch the JWKS and parse out all the signing JWKs. /// Returns `None` and log a warning if there are any errors. async fn fetch_jwks( client: &reqwest_middleware::ClientWithMiddleware, jwks_url: url::Url, ) -> Option<jose_jwk::JwkSet> { let req = client.get(jwks_url.clone()); // TODO(conrad): We need to filter out URLs that point to local resources. Public internet only. let resp = req.send().await.and_then(|r| { r.error_for_status() .map_err(reqwest_middleware::Error::Reqwest) }); let resp = match resp { Ok(r) => r, // TODO: should we re-insert JWKs if we want to keep this JWKs URL? // I expect these failures would be quite sparse. Err(e) => { tracing::warn!(url=?jwks_url, error=?e, "could not fetch JWKs"); return None; } }; let resp: http::Response<reqwest::Body> = resp.into(); let bytes = match read_body_with_limit(resp.into_body(), MAX_JWK_BODY_SIZE).await { Ok(bytes) => bytes, Err(e) => { tracing::warn!(url=?jwks_url, error=?e, "could not decode JWKs"); return None; } }; let jwks = match serde_json::from_slice::<JwkSet>(&bytes) { Ok(jwks) => jwks, Err(e) => { tracing::warn!(url=?jwks_url, error=?e, "could not decode JWKs"); return None; } }; // `jose_jwk::Jwk` is quite large (288 bytes). Let's not pre-allocate for what we don't need. // // Even though we limit our responses to 64KiB, we could still receive a payload like // `{"keys":[` + repeat(`0`).take(30000).join(`,`) + `]}`. Parsing this as `RawValue` uses 468KiB. // Pre-allocating the corresponding `Vec::<jose_jwk::Jwk>::with_capacity(30000)` uses 8.2MiB. let mut keys = vec![]; let mut failed = 0; for key in jwks.keys { let key = match serde_json::from_str::<jose_jwk::Jwk>(key.get()) { Ok(key) => key, Err(e) => { tracing::debug!(url=?jwks_url, failed=?e, "could not decode JWK"); failed += 1; continue; } }; // if `use` (called `cls` in rust) is specified to be something other than signing, // we can skip storing it. if key .prm .cls .as_ref() .is_some_and(|c| *c != jose_jwk::Class::Signing) { continue; } keys.push(key); } keys.shrink_to_fit(); if failed > 0 { tracing::warn!(url=?jwks_url, failed, "could not decode JWKs"); } if keys.is_empty() { tracing::warn!(url=?jwks_url, "no valid JWKs found inside the response body"); return None; } Some(jose_jwk::JwkSet { keys }) } impl JwkCacheEntryLock { async fn acquire_permit(self: &Arc<Self>) -> JwkRenewalPermit<'_> { JwkRenewalPermit::acquire_permit(self).await } fn try_acquire_permit(self: &Arc<Self>) -> Option<JwkRenewalPermit<'_>> { JwkRenewalPermit::try_acquire_permit(self) } async fn renew_jwks<F: FetchAuthRules>( &self, _permit: JwkRenewalPermit<'_>, ctx: &RequestContext, client: &reqwest_middleware::ClientWithMiddleware, endpoint: EndpointId, auth_rules: &F, ) -> Result<Arc<JwkCacheEntry>, JwtError> { // double check that no one beat us to updating the cache. let now = Instant::now(); let guard = self.cached.load_full(); if let Some(cached) = guard { let last_update = now.duration_since(cached.last_retrieved); if last_update < Duration::from_secs(300) { return Ok(cached); } } let rules = auth_rules.fetch_auth_rules(ctx, endpoint).await?; let mut key_sets = ahash::HashMap::with_capacity_and_hasher(rules.len(), ahash::RandomState::new()); // TODO(conrad): run concurrently // TODO(conrad): strip the JWKs urls (should be checked by cplane as well - cloud#16284) for rule in rules { if let Some(jwks) = fetch_jwks(client, rule.jwks_url).await { key_sets.insert( rule.id, KeySet { jwks, audience: rule.audience, role_names: rule.role_names, }, ); } } let entry = Arc::new(JwkCacheEntry { last_retrieved: now, key_sets, }); self.cached.swap(Some(Arc::clone(&entry))); Ok(entry) } async fn get_or_update_jwk_cache<F: FetchAuthRules>( self: &Arc<Self>, ctx: &RequestContext, client: &reqwest_middleware::ClientWithMiddleware, endpoint: EndpointId, fetch: &F, ) -> Result<Arc<JwkCacheEntry>, JwtError> { let now = Instant::now(); let guard = self.cached.load_full(); // if we have no cached JWKs, try and get some let Some(cached) = guard else { let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Compute); let permit = self.acquire_permit().await; return self.renew_jwks(permit, ctx, client, endpoint, fetch).await; }; let last_update = now.duration_since(cached.last_retrieved); // check if the cached JWKs need updating. if last_update > MAX_RENEW { let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Compute); let permit = self.acquire_permit().await; // it's been too long since we checked the keys. wait for them to update. return self.renew_jwks(permit, ctx, client, endpoint, fetch).await; } // every 5 minutes we should spawn a job to eagerly update the token. if last_update > AUTO_RENEW { if let Some(permit) = self.try_acquire_permit() { tracing::debug!("JWKs should be renewed. Renewal permit acquired"); let permit = permit.into_owned(); let entry = self.clone(); let client = client.clone(); let fetch = fetch.clone(); let ctx = ctx.clone(); tokio::spawn(async move { if let Err(e) = entry .renew_jwks(permit, &ctx, &client, endpoint, &fetch) .await { tracing::warn!(error=?e, "could not fetch JWKs in background job"); } }); } else { tracing::debug!("JWKs should be renewed. Renewal permit already taken, skipping"); } } Ok(cached) } async fn check_jwt<F: FetchAuthRules>( self: &Arc<Self>, ctx: &RequestContext, jwt: &str, client: &reqwest_middleware::ClientWithMiddleware, endpoint: EndpointId, role_name: &RoleName, fetch: &F, ) -> Result<ComputeCredentialKeys, JwtError> { // JWT compact form is defined to be // <B64(Header)> || . || <B64(Payload)> || . || <B64(Signature)> // where Signature = alg(<B64(Header)> || . || <B64(Payload)>); let (header_payload, signature) = jwt .rsplit_once('.') .ok_or(JwtEncodingError::InvalidCompactForm)?; let (header, payload) = header_payload .split_once('.') .ok_or(JwtEncodingError::InvalidCompactForm)?; let header = BASE64_URL_SAFE_NO_PAD.decode(header)?; let header = serde_json::from_slice::<JwtHeader<'_>>(&header)?; let payloadb = BASE64_URL_SAFE_NO_PAD.decode(payload)?; let payload = serde_json::from_slice::<JwtPayload<'_>>(&payloadb)?; if let Some(iss) = &payload.issuer { ctx.set_jwt_issuer(iss.as_ref().to_owned()); } let sig = BASE64_URL_SAFE_NO_PAD.decode(signature)?; let kid = header.key_id.ok_or(JwtError::MissingKeyId)?; let mut guard = self .get_or_update_jwk_cache(ctx, client, endpoint.clone(), fetch) .await?; // get the key from the JWKs if possible. If not, wait for the keys to update. let (jwk, expected_audience) = loop { match guard.find_jwk_and_audience(&kid, role_name) { Some(jwk) => break jwk, None if guard.last_retrieved.elapsed() > MIN_RENEW => { let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Compute); let permit = self.acquire_permit().await; guard = self .renew_jwks(permit, ctx, client, endpoint.clone(), fetch) .await?; } _ => return Err(JwtError::JwkNotFound), } }; if !jwk.is_supported(&header.algorithm) { return Err(JwtError::SignatureAlgorithmNotSupported); } match &jwk.key { jose_jwk::Key::Ec(key) => { verify_ec_signature(header_payload.as_bytes(), &sig, key)?; } jose_jwk::Key::Rsa(key) => { verify_rsa_signature(header_payload.as_bytes(), &sig, key, &header.algorithm)?; } key => return Err(JwtError::UnsupportedKeyType(key.into())), } tracing::debug!(?payload, "JWT signature valid with claims"); if let Some(aud) = expected_audience && payload.audience.0.iter().all(|s| s != aud) { return Err(JwtError::InvalidClaims( JwtClaimsError::InvalidJwtTokenAudience, )); } let now = SystemTime::now(); if let Some(exp) = payload.expiration && now >= exp + CLOCK_SKEW_LEEWAY { return Err(JwtError::InvalidClaims(JwtClaimsError::JwtTokenHasExpired( exp.duration_since(SystemTime::UNIX_EPOCH) .unwrap_or_default() .as_secs(), ))); } if let Some(nbf) = payload.not_before && nbf >= now + CLOCK_SKEW_LEEWAY { return Err(JwtError::InvalidClaims( JwtClaimsError::JwtTokenNotYetReadyToUse( nbf.duration_since(SystemTime::UNIX_EPOCH) .unwrap_or_default() .as_secs(), ), )); } Ok(ComputeCredentialKeys::JwtPayload(payloadb)) } } impl JwkCache { pub(crate) async fn check_jwt<F: FetchAuthRules>( &self, ctx: &RequestContext, endpoint: EndpointId, role_name: &RoleName, fetch: &F, jwt: &str, ) -> Result<ComputeCredentialKeys, JwtError> { // try with just a read lock first let key = (endpoint.clone(), role_name.clone()); let entry = self.map.get(&key).as_deref().map(Arc::clone); let entry = entry.unwrap_or_else(|| { // acquire a write lock after to insert. let entry = self.map.entry(key).or_default(); Arc::clone(&*entry) }); entry .check_jwt(ctx, jwt, &self.client, endpoint, role_name, fetch) .await } } impl Default for JwkCache { fn default() -> Self { let client = Client::builder() .user_agent(JWKS_USER_AGENT) .redirect(redirect::Policy::none()) .tls_built_in_native_certs(true) .connect_timeout(JWKS_CONNECT_TIMEOUT) .timeout(JWKS_FETCH_TIMEOUT) .build() .expect("client config should be valid"); // Retry up to 3 times with increasing intervals between attempts. let retry_policy = ExponentialBackoff::builder().build_with_max_retries(JWKS_FETCH_RETRIES); let client = reqwest_middleware::ClientBuilder::new(client) .with(RetryTransientMiddleware::new_with_policy(retry_policy)) .build(); JwkCache { client, map: ClashMap::default(), } } } fn verify_ec_signature(data: &[u8], sig: &[u8], key: &jose_jwk::Ec) -> Result<(), JwtError> { use ecdsa::Signature; use signature::Verifier; match key.crv { jose_jwk::EcCurves::P256 => { let pk = p256::PublicKey::try_from(key).map_err(JwtError::InvalidP256Key)?; let key = p256::ecdsa::VerifyingKey::from(&pk); let sig = Signature::from_slice(sig)?; key.verify(data, &sig)?; } key => return Err(JwtError::UnsupportedEcKeyType(key)), } Ok(()) } fn verify_rsa_signature( data: &[u8], sig: &[u8], key: &jose_jwk::Rsa, alg: &jose_jwa::Algorithm, ) -> Result<(), JwtError> { use jose_jwa::{Algorithm, Signing}; use rsa::RsaPublicKey; use rsa::pkcs1v15::{Signature, VerifyingKey}; let key = RsaPublicKey::try_from(key).map_err(JwtError::InvalidRsaKey)?; match alg { Algorithm::Signing(Signing::Rs256) => { let key = VerifyingKey::<sha2::Sha256>::new(key); let sig = Signature::try_from(sig)?; key.verify(data, &sig)?; } _ => return Err(JwtError::InvalidRsaSigningAlgorithm), } Ok(()) } /// <https://datatracker.ietf.org/doc/html/rfc7515#section-4.1> #[derive(serde::Deserialize, serde::Serialize)] struct JwtHeader<'a> { /// must be a supported alg #[serde(rename = "alg")] algorithm: jose_jwa::Algorithm, /// key id, must be provided for our usecase #[serde(rename = "kid", borrow)] key_id: Option<Cow<'a, str>>, } /// <https://datatracker.ietf.org/doc/html/rfc7519#section-4.1> #[derive(serde::Deserialize, Debug)] #[allow(dead_code)] struct JwtPayload<'a> { /// Audience - Recipient for which the JWT is intended #[serde(rename = "aud", default)] audience: OneOrMany, /// Expiration - Time after which the JWT expires #[serde(rename = "exp", deserialize_with = "numeric_date_opt", default)] expiration: Option<SystemTime>, /// Not before - Time before which the JWT is not valid #[serde(rename = "nbf", deserialize_with = "numeric_date_opt", default)] not_before: Option<SystemTime>, // the following entries are only extracted for the sake of debug logging. /// Issuer of the JWT #[serde(rename = "iss", borrow)] issuer: Option<Cow<'a, str>>, /// Subject of the JWT (the user) #[serde(rename = "sub", borrow)] subject: Option<Cow<'a, str>>, /// Unique token identifier #[serde(rename = "jti", borrow)] jwt_id: Option<Cow<'a, str>>, /// Unique session identifier #[serde(rename = "sid", borrow)] session_id: Option<Cow<'a, str>>, } /// `OneOrMany` supports parsing either a single item or an array of items. /// /// Needed for <https://datatracker.ietf.org/doc/html/rfc7519#section-4.1.3> /// /// > The "aud" (audience) claim identifies the recipients that the JWT is /// > intended for. Each principal intended to process the JWT MUST /// > identify itself with a value in the audience claim. If the principal /// > processing the claim does not identify itself with a value in the /// > "aud" claim when this claim is present, then the JWT MUST be /// > rejected. In the general case, the "aud" value is **an array of case- /// > sensitive strings**, each containing a StringOrURI value. In the /// > special case when the JWT has one audience, the "aud" value MAY be a /// > **single case-sensitive string** containing a StringOrURI value. The /// > interpretation of audience values is generally application specific. /// > Use of this claim is OPTIONAL. #[derive(Default, Debug)] struct OneOrMany(Vec<String>); impl<'de> Deserialize<'de> for OneOrMany { fn deserialize<D>(deserializer: D) -> Result<Self, D::Error> where D: Deserializer<'de>, { struct OneOrManyVisitor; impl<'de> Visitor<'de> for OneOrManyVisitor { type Value = OneOrMany; fn expecting(&self, formatter: &mut std::fmt::Formatter) -> std::fmt::Result { formatter.write_str("a single string or an array of strings") } fn visit_str<E>(self, v: &str) -> Result<Self::Value, E> where E: serde::de::Error, { Ok(OneOrMany(vec![v.to_owned()])) } fn visit_seq<A>(self, mut seq: A) -> Result<Self::Value, A::Error> where A: serde::de::SeqAccess<'de>, { let mut v = vec![]; while let Some(s) = seq.next_element()? { v.push(s); } Ok(OneOrMany(v)) } } deserializer.deserialize_any(OneOrManyVisitor) } } fn numeric_date_opt<'de, D: Deserializer<'de>>(d: D) -> Result<Option<SystemTime>, D::Error> { <Option<u64>>::deserialize(d)? .map(|t| { SystemTime::UNIX_EPOCH .checked_add(Duration::from_secs(t)) .ok_or_else(|| { serde::de::Error::custom(format_args!("timestamp out of bounds: {t}")) }) }) .transpose() } struct JwkRenewalPermit<'a> { inner: Option<JwkRenewalPermitInner<'a>>, } enum JwkRenewalPermitInner<'a> { Owned(Arc<JwkCacheEntryLock>), Borrowed(&'a Arc<JwkCacheEntryLock>), } impl JwkRenewalPermit<'_> { fn into_owned(mut self) -> JwkRenewalPermit<'static> { JwkRenewalPermit { inner: self.inner.take().map(JwkRenewalPermitInner::into_owned), } } async fn acquire_permit(from: &Arc<JwkCacheEntryLock>) -> JwkRenewalPermit<'_> { match from.lookup.acquire().await { Ok(permit) => { permit.forget(); JwkRenewalPermit { inner: Some(JwkRenewalPermitInner::Borrowed(from)), } } Err(_) => panic!("semaphore should not be closed"), } } fn try_acquire_permit(from: &Arc<JwkCacheEntryLock>) -> Option<JwkRenewalPermit<'_>> { match from.lookup.try_acquire() { Ok(permit) => { permit.forget(); Some(JwkRenewalPermit { inner: Some(JwkRenewalPermitInner::Borrowed(from)), }) } Err(tokio::sync::TryAcquireError::NoPermits) => None, Err(tokio::sync::TryAcquireError::Closed) => panic!("semaphore should not be closed"), } } } impl JwkRenewalPermitInner<'_> { fn into_owned(self) -> JwkRenewalPermitInner<'static> { match self { JwkRenewalPermitInner::Owned(p) => JwkRenewalPermitInner::Owned(p), JwkRenewalPermitInner::Borrowed(p) => JwkRenewalPermitInner::Owned(Arc::clone(p)), } } } impl Drop for JwkRenewalPermit<'_> { fn drop(&mut self) { let entry = match &self.inner { None => return, Some(JwkRenewalPermitInner::Owned(p)) => p, Some(JwkRenewalPermitInner::Borrowed(p)) => *p, }; entry.lookup.add_permits(1); } } #[derive(Error, Debug)] #[non_exhaustive] pub(crate) enum JwtError { #[error("jwk not found")] JwkNotFound, #[error("missing key id")] MissingKeyId, #[error("Provided authentication token is not a valid JWT encoding")] JwtEncoding(#[from] JwtEncodingError), #[error(transparent)] InvalidClaims(#[from] JwtClaimsError), #[error("invalid P256 key")] InvalidP256Key(jose_jwk::crypto::Error), #[error("invalid RSA key")] InvalidRsaKey(jose_jwk::crypto::Error), #[error("invalid RSA signing algorithm")] InvalidRsaSigningAlgorithm, #[error("unsupported EC key type {0:?}")] UnsupportedEcKeyType(jose_jwk::EcCurves), #[error("unsupported key type {0:?}")] UnsupportedKeyType(KeyType), #[error("signature algorithm not supported")] SignatureAlgorithmNotSupported, #[error("signature error: {0}")] Signature(#[from] signature::Error), #[error("failed to fetch auth rules: {0}")] FetchAuthRules(#[from] FetchAuthRulesError), } impl From<base64::DecodeError> for JwtError { fn from(err: base64::DecodeError) -> Self { JwtEncodingError::Base64Decode(err).into() } } impl From<serde_json::Error> for JwtError { fn from(err: serde_json::Error) -> Self { JwtEncodingError::SerdeJson(err).into() } } #[derive(Error, Debug)] #[non_exhaustive] pub enum JwtEncodingError { #[error(transparent)] Base64Decode(#[from] base64::DecodeError), #[error(transparent)] SerdeJson(#[from] serde_json::Error), #[error("invalid compact form")] InvalidCompactForm, } #[derive(Error, Debug, PartialEq)] #[non_exhaustive] pub enum JwtClaimsError { #[error("invalid JWT token audience")] InvalidJwtTokenAudience, #[error("JWT token has expired (exp={0})")] JwtTokenHasExpired(u64), #[error("JWT token is not yet ready to use (nbf={0})")] JwtTokenNotYetReadyToUse(u64), } #[allow(dead_code, reason = "Debug use only")] #[derive(Debug)] pub(crate) enum KeyType { Ec(jose_jwk::EcCurves), Rsa, Oct, Okp(jose_jwk::OkpCurves), Unknown, } impl From<&jose_jwk::Key> for KeyType { fn from(key: &jose_jwk::Key) -> Self { match key { jose_jwk::Key::Ec(ec) => Self::Ec(ec.crv), jose_jwk::Key::Rsa(_rsa) => Self::Rsa, jose_jwk::Key::Oct(_oct) => Self::Oct, jose_jwk::Key::Okp(okp) => Self::Okp(okp.crv), _ => Self::Unknown, } } } #[cfg(test)] mod tests { use std::future::IntoFuture; use std::net::SocketAddr; use std::time::SystemTime; use bytes::Bytes; use http::Response; use http_body_util::Full; use hyper::service::service_fn; use hyper_util::rt::TokioIo; use rand_core::OsRng; use rsa::pkcs8::DecodePrivateKey; use serde::Serialize; use serde_json::json; use signature::Signer; use tokio::net::TcpListener; use super::*; use crate::types::RoleName; fn new_ec_jwk(kid: String) -> (p256::SecretKey, jose_jwk::Jwk) { let sk = p256::SecretKey::random(&mut OsRng); let pk = sk.public_key().into(); let jwk = jose_jwk::Jwk { key: jose_jwk::Key::Ec(pk), prm: jose_jwk::Parameters { kid: Some(kid), alg: Some(jose_jwa::Algorithm::Signing(jose_jwa::Signing::Es256)), ..Default::default() }, }; (sk, jwk) } fn new_rsa_jwk(key: &str, kid: String) -> (rsa::RsaPrivateKey, jose_jwk::Jwk) { let sk = rsa::RsaPrivateKey::from_pkcs8_pem(key).unwrap(); let pk = sk.to_public_key().into(); let jwk = jose_jwk::Jwk { key: jose_jwk::Key::Rsa(pk), prm: jose_jwk::Parameters { kid: Some(kid), alg: Some(jose_jwa::Algorithm::Signing(jose_jwa::Signing::Rs256)), ..Default::default() }, }; (sk, jwk) } fn now() -> u64 { SystemTime::now() .duration_since(SystemTime::UNIX_EPOCH) .unwrap() .as_secs() } fn build_jwt_payload(kid: String, sig: jose_jwa::Signing) -> String { let now = now(); let body = typed_json::json! {{ "exp": now + 3600, "nbf": now, "aud": ["audience1", "neon", "audience2"], "sub": "user1", "sid": "session1", "jti": "token1", "iss": "neon-testing", }}; build_custom_jwt_payload(kid, body, sig) } fn build_custom_jwt_payload( kid: String, body: impl Serialize, sig: jose_jwa::Signing, ) -> String { let header = JwtHeader { algorithm: jose_jwa::Algorithm::Signing(sig), key_id: Some(Cow::Owned(kid)), }; let header = BASE64_URL_SAFE_NO_PAD.encode(serde_json::to_string(&header).unwrap()); let body = BASE64_URL_SAFE_NO_PAD.encode(serde_json::to_string(&body).unwrap()); format!("{header}.{body}") } fn new_ec_jwt(kid: String, key: &p256::SecretKey) -> String { use p256::ecdsa::{Signature, SigningKey}; let payload = build_jwt_payload(kid, jose_jwa::Signing::Es256); let sig: Signature = SigningKey::from(key).sign(payload.as_bytes()); let sig = BASE64_URL_SAFE_NO_PAD.encode(sig.to_bytes()); format!("{payload}.{sig}") } fn new_custom_ec_jwt(kid: String, key: &p256::SecretKey, body: impl Serialize) -> String { use p256::ecdsa::{Signature, SigningKey}; let payload = build_custom_jwt_payload(kid, body, jose_jwa::Signing::Es256); let sig: Signature = SigningKey::from(key).sign(payload.as_bytes()); let sig = BASE64_URL_SAFE_NO_PAD.encode(sig.to_bytes()); format!("{payload}.{sig}") } fn new_rsa_jwt(kid: String, key: rsa::RsaPrivateKey) -> String { use rsa::pkcs1v15::SigningKey; use rsa::signature::SignatureEncoding; let payload = build_jwt_payload(kid, jose_jwa::Signing::Rs256); let sig = SigningKey::<sha2::Sha256>::new(key).sign(payload.as_bytes()); let sig = BASE64_URL_SAFE_NO_PAD.encode(sig.to_bytes()); format!("{payload}.{sig}") } // RSA key gen is slow.... const RS1: &str = "-----BEGIN PRIVATE KEY----- MIIEvwIBADANBgkqhkiG9w0BAQEFAASCBKkwggSlAgEAAoIBAQDNuWBIWTlo+54Y aifpGInIrpv6LlsbI/2/2CC81Arlx4RsABORklgA9XSGwaCbHTshHsfd1S916JwA SpjyPQYWfqo6iAV8a4MhjIeJIkRr74prDCSzOGZvIc6VaGeCIb9clf3HSrPHm3hA cfLMB8/p5MgoxERPDOIn3XYoS9SEEuP7l0LkmEZMerg6W6lDjQRDny0Lb50Jky9X mDqnYXBhs99ranbwL5vjy0ba6OIeCWFJme5u+rv5C/P0BOYrJfGxIcEoKa8Ukw5s PlM+qrz9ope1eOuXMNNdyFDReNBUyaM1AwBAayU5rz57crer7K/UIofaJ42T4cMM nx/SWfBNAgMBAAECggEACqdpBxYn1PoC6/zDaFzu9celKEWyTiuE/qRwvZa1ocS9 ZOJ0IPvVNud/S2NHsADJiSOQ8joSJScQvSsf1Ju4bv3MTw+wSQtAVUJz2nQ92uEi 5/xPAkEPfP3hNvebNLAOuvrBk8qYmOPCTIQaMNrOt6wzeXkAmJ9wLuRXNCsJLHW+ KLpf2WdgTYxqK06ZiJERFgJ2r1MsC2IgTydzjOAdEIrtMarerTLqqCpwFrk/l0cz 1O2OAb17ZxmhuzMhjNMin81c8F2fZAGMeOjn92Jl5kUsYw/pG+0S8QKlbveR/fdP We2tJsgXw2zD0q7OJpp8NXS2yddrZGyysYsof983wQKBgQD2McqNJqo+eWL5zony UbL19loYw0M15EjhzIuzW1Jk0rPj65yQyzpJ6pqicRuWr34MvzCx+ZHM2b3jSiNu GES2fnC7xLIKyeRxfqsXF71xz+6UStEGRQX27r1YWEtyQVuBhvlqB+AGWP3PYAC+ HecZecnZ+vcihJ2K3+l5O3paVQKBgQDV6vKH5h2SY9vgO8obx0P7XSS+djHhmPuU f8C/Fq6AuRbIA1g04pzuLU2WS9T26eIjgM173uVNg2TuqJveWzz+CAAp6nCR6l24 DBg49lMGCWrMo4FqPG46QkUqvK8uSj42GkX/e5Rut1Gyu0209emeM6h2d2K15SvY 9563tYSmGQKBgQDwcH5WTi20KA7e07TroJi8GKWzS3gneNUpGQBS4VxdtV4UuXXF /4TkzafJ/9cm2iurvUmMd6XKP9lw0mY5zp/E70WgTCBp4vUlVsU3H2tYbO+filYL 3ntNx6nKTykX4/a/UJfj0t8as+zli+gNxNx/h+734V9dKdFG4Rl+2fTLpQKBgQCE qJkTEe+Q0wCOBEYICADupwqcWqwAXWDW7IrZdfVtulqYWwqecVIkmk+dPxWosc4d ekjz4nyNH0i+gC15LVebqdaAJ/T7aD4KXuW+nXNLMRfcJCGjgipRUruWD0EMEdqW rqBuGXMpXeH6VxGPgVkJVLvKC6tZZe9VM+pnvteuMQKBgQC8GaL+Lz+al4biyZBf JE8ekWrIotq/gfUBLP7x70+PB9bNtXtlgmTvjgYg4jiu3KR/ZIYYQ8vfVgkb6tDI rWGZw86Pzuoi1ppg/pYhKk9qrmCIT4HPEXbHl7ATahu2BOCIU3hybjTh2lB6LbX9 8LMFlz1QPqSZYN/A/kOcLBfa3A== -----END PRIVATE KEY----- "; const RS2: &str = "-----BEGIN PRIVATE KEY----- MIIEvgIBADANBgkqhkiG9w0BAQEFAASCBKgwggSkAgEAAoIBAQDipm6FIKSRab3J HwmK18t7hp+pohllxIDUSPi7S5mIhN/JG2Plq2Lp746E/fuT8dcBF2R4sJlG2L0J zmxOvBU/i/sQF9s1i4CEfg05k2//gKENIEsF3pMMmrH+mcZi0TTD6rezHpdVxPHk qWxSyOCtIJV29X+wxPwAB59kQFHzy2ooPB1isZcpE8tO0KthAM+oZ3KuCwE0++cO IWLeq9aPwyKhtip/xjTMxd1kzdKh592mGSyzr9D0QSWOYFGvgJXANDdiPdhSSOLt ECWPNPlm2FQvGGvYYBafUqz7VumKHE6x8J6lKdYa2J0ZdDzCIo2IHzlxe+RZNgwy uAD2jhVxAgMBAAECggEAbsZHWBu3MzcKQiVARbLoygvnN0J5xUqAaMDtiKUPejDv K1yOu67DXnDuKEP2VL2rhuYG/hHaKE1AP227c9PrUq6424m9YvM2sgrlrdFIuQkG LeMtp8W7+zoUasp/ssZrUqICfLIj5xCl5UuFHQT/Ar7dLlIYwa3VOLKBDb9+Dnfe QH5/So4uMXG6vw34JN9jf+eAc8Yt0PeIz62ycvRwdpTJQ0MxZN9ZKpCAQp+VTuXT zlzNvDMilabEdqUvAyGyz8lBLNl0wdaVrqPqAEWM5U45QXsdFZknWammP7/tijeX 0z+Bi0J0uSEU5X502zm7GArj/NNIiWMcjmDjwUUhwQKBgQD9C2GoqxOxuVPYqwYR +Jz7f2qMjlSP8adA5Lzuh8UKXDp8JCEQC8ryweLzaOKS9C5MAw+W4W2wd4nJoQI1 P1dgGvBlfvEeRHMgqWtq7FuTsjSe7e0uSEkC4ngDb4sc0QOpv15cMuEz+4+aFLPL x29EcHWAaBX+rkid3zpQHFU4eQKBgQDlTCEqRuXwwa3V+Sq+mNWzD9QIGtD87TH/ FPO/Ij/cK2+GISgFDqhetiGTH4qrvPL0psPT+iH5zGFYcoFmTtwLdWQJdxhxz0bg iX/AceyX5e1Bm+ThT36sU83NrxKPkrdk6jNmr2iUF1OTzTwUKOYdHOPZqdMPfF4M 4XAaWVT2uQKBgQD4nKcNdU+7LE9Rr+4d1/o8Klp/0BMK/ayK2HE7lc8kt6qKb2DA iCWUTqPw7Fq3cQrPia5WWhNP7pJEtFkcAaiR9sW7onW5fBz0uR+dhK0QtmR2xWJj N4fsOp8ZGQ0/eae0rh1CTobucLkM9EwV6VLLlgYL67e4anlUCo8bSEr+WQKBgQCB
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
true
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/mod.rs
proxy/src/auth/backend/mod.rs
mod classic; mod console_redirect; mod hacks; pub mod jwt; pub mod local; use std::sync::Arc; pub use console_redirect::ConsoleRedirectBackend; pub(crate) use console_redirect::ConsoleRedirectError; use local::LocalBackend; use postgres_client::config::AuthKeys; use serde::{Deserialize, Serialize}; use tokio::io::{AsyncRead, AsyncWrite}; use tracing::{debug, info}; use crate::auth::{self, ComputeUserInfoMaybeEndpoint, validate_password_and_exchange}; use crate::cache::Cached; use crate::cache::node_info::CachedNodeInfo; use crate::config::AuthenticationConfig; use crate::context::RequestContext; use crate::control_plane::client::ControlPlaneClient; use crate::control_plane::errors::GetAuthInfoError; use crate::control_plane::messages::EndpointRateLimitConfig; use crate::control_plane::{ self, AccessBlockerFlags, AuthSecret, ControlPlaneApi, EndpointAccessControl, RoleAccessControl, }; use crate::intern::{EndpointIdInt, RoleNameInt}; use crate::pqproto::BeMessage; use crate::proxy::NeonOptions; use crate::proxy::wake_compute::WakeComputeBackend; use crate::rate_limiter::EndpointRateLimiter; use crate::stream::Stream; use crate::types::{EndpointCacheKey, EndpointId, RoleName}; use crate::{scram, stream}; /// Alternative to [`std::borrow::Cow`] but doesn't need `T: ToOwned` as we don't need that functionality pub enum MaybeOwned<'a, T> { Owned(T), Borrowed(&'a T), } impl<T> std::ops::Deref for MaybeOwned<'_, T> { type Target = T; fn deref(&self) -> &Self::Target { match self { MaybeOwned::Owned(t) => t, MaybeOwned::Borrowed(t) => t, } } } /// This type serves two purposes: /// /// * When `T` is `()`, it's just a regular auth backend selector /// which we use in [`crate::config::ProxyConfig`]. /// /// * However, when we substitute `T` with [`ComputeUserInfoMaybeEndpoint`], /// this helps us provide the credentials only to those auth /// backends which require them for the authentication process. pub enum Backend<'a, T> { /// Cloud API (V2). ControlPlane(MaybeOwned<'a, ControlPlaneClient>, T), /// Local proxy uses configured auth credentials and does not wake compute Local(MaybeOwned<'a, LocalBackend>), } impl std::fmt::Display for Backend<'_, ()> { fn fmt(&self, fmt: &mut std::fmt::Formatter<'_>) -> std::fmt::Result { match self { Self::ControlPlane(api, ()) => match &**api { ControlPlaneClient::ProxyV1(endpoint) => fmt .debug_tuple("ControlPlane::ProxyV1") .field(&endpoint.url()) .finish(), #[cfg(any(test, feature = "testing"))] ControlPlaneClient::PostgresMock(endpoint) => { let url = endpoint.url(); match url::Url::parse(url) { Ok(mut url) => { let _ = url.set_password(Some("_redacted_")); let url = url.as_str(); fmt.debug_tuple("ControlPlane::PostgresMock") .field(&url) .finish() } Err(_) => fmt .debug_tuple("ControlPlane::PostgresMock") .field(&url) .finish(), } } #[cfg(test)] ControlPlaneClient::Test(_) => fmt.debug_tuple("ControlPlane::Test").finish(), }, Self::Local(_) => fmt.debug_tuple("Local").finish(), } } } impl<T> Backend<'_, T> { /// Very similar to [`std::option::Option::as_ref`]. /// This helps us pass structured config to async tasks. pub(crate) fn as_ref(&self) -> Backend<'_, &T> { match self { Self::ControlPlane(c, x) => Backend::ControlPlane(MaybeOwned::Borrowed(c), x), Self::Local(l) => Backend::Local(MaybeOwned::Borrowed(l)), } } pub(crate) fn get_api(&self) -> &ControlPlaneClient { match self { Self::ControlPlane(api, _) => api, Self::Local(_) => panic!("Local backend has no API"), } } pub(crate) fn is_local_proxy(&self) -> bool { matches!(self, Self::Local(_)) } } impl<'a, T> Backend<'a, T> { /// Very similar to [`std::option::Option::map`]. /// Maps [`Backend<T>`] to [`Backend<R>`] by applying /// a function to a contained value. pub(crate) fn map<R>(self, f: impl FnOnce(T) -> R) -> Backend<'a, R> { match self { Self::ControlPlane(c, x) => Backend::ControlPlane(c, f(x)), Self::Local(l) => Backend::Local(l), } } } impl<'a, T, E> Backend<'a, Result<T, E>> { /// Very similar to [`std::option::Option::transpose`]. /// This is most useful for error handling. pub(crate) fn transpose(self) -> Result<Backend<'a, T>, E> { match self { Self::ControlPlane(c, x) => x.map(|x| Backend::ControlPlane(c, x)), Self::Local(l) => Ok(Backend::Local(l)), } } } pub(crate) struct ComputeCredentials { pub(crate) info: ComputeUserInfo, pub(crate) keys: ComputeCredentialKeys, } #[derive(Debug, Clone)] pub(crate) struct ComputeUserInfoNoEndpoint { pub(crate) user: RoleName, pub(crate) options: NeonOptions, } #[derive(Debug, Clone, Default, Serialize, Deserialize)] pub(crate) struct ComputeUserInfo { pub(crate) endpoint: EndpointId, pub(crate) user: RoleName, pub(crate) options: NeonOptions, } impl ComputeUserInfo { pub(crate) fn endpoint_cache_key(&self) -> EndpointCacheKey { self.options.get_cache_key(&self.endpoint) } } #[cfg_attr(test, derive(Debug))] pub(crate) enum ComputeCredentialKeys { AuthKeys(AuthKeys), JwtPayload(Vec<u8>), } impl TryFrom<ComputeUserInfoMaybeEndpoint> for ComputeUserInfo { // user name type Error = ComputeUserInfoNoEndpoint; fn try_from(user_info: ComputeUserInfoMaybeEndpoint) -> Result<Self, Self::Error> { match user_info.endpoint_id { None => Err(ComputeUserInfoNoEndpoint { user: user_info.user, options: user_info.options, }), Some(endpoint) => Ok(ComputeUserInfo { endpoint, user: user_info.user, options: user_info.options, }), } } } /// True to its name, this function encapsulates our current auth trade-offs. /// Here, we choose the appropriate auth flow based on circumstances. /// /// All authentication flows will emit an AuthenticationOk message if successful. async fn auth_quirks( ctx: &RequestContext, api: &impl control_plane::ControlPlaneApi, user_info: ComputeUserInfoMaybeEndpoint, client: &mut stream::PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, allow_cleartext: bool, config: &'static AuthenticationConfig, endpoint_rate_limiter: Arc<EndpointRateLimiter>, ) -> auth::Result<ComputeCredentials> { // If there's no project so far, that entails that client doesn't // support SNI or other means of passing the endpoint (project) name. // We now expect to see a very specific payload in the place of password. let (info, unauthenticated_password) = match user_info.try_into() { Err(info) => { let (info, password) = hacks::password_hack_no_authentication(ctx, info, client).await?; ctx.set_endpoint_id(info.endpoint.clone()); (info, Some(password)) } Ok(info) => (info, None), }; debug!("fetching authentication info and allowlists"); let access_controls = api .get_endpoint_access_control(ctx, &info.endpoint, &info.user) .await?; access_controls.check( ctx, config.ip_allowlist_check_enabled, config.is_vpc_acccess_proxy, )?; access_controls.connection_attempt_rate_limit(ctx, &info.endpoint, &endpoint_rate_limiter)?; let role_access = api .get_role_access_control(ctx, &info.endpoint, &info.user) .await?; let secret = if let Some(secret) = role_access.secret { secret } else { // If we don't have an authentication secret, we mock one to // prevent malicious probing (possible due to missing protocol steps). // This mocked secret will never lead to successful authentication. info!("authentication info not found, mocking it"); AuthSecret::Scram(scram::ServerSecret::mock(rand::random())) }; match authenticate_with_secret( ctx, secret, info, client, unauthenticated_password, allow_cleartext, config, ) .await { Ok(keys) => Ok(keys), Err(e) => Err(e), } } async fn authenticate_with_secret( ctx: &RequestContext, secret: AuthSecret, info: ComputeUserInfo, client: &mut stream::PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, unauthenticated_password: Option<Vec<u8>>, allow_cleartext: bool, config: &'static AuthenticationConfig, ) -> auth::Result<ComputeCredentials> { if let Some(password) = unauthenticated_password { let ep = EndpointIdInt::from(&info.endpoint); let role = RoleNameInt::from(&info.user); let auth_outcome = validate_password_and_exchange(&config.scram_thread_pool, ep, role, &password, secret) .await?; let keys = match auth_outcome { crate::sasl::Outcome::Success(key) => key, crate::sasl::Outcome::Failure(reason) => { info!("auth backend failed with an error: {reason}"); return Err(auth::AuthError::password_failed(&*info.user)); } }; // we have authenticated the password client.write_message(BeMessage::AuthenticationOk); return Ok(ComputeCredentials { info, keys }); } // -- the remaining flows are self-authenticating -- // Perform cleartext auth if we're allowed to do that. // Currently, we use it for websocket connections (latency). if allow_cleartext { ctx.set_auth_method(crate::context::AuthMethod::Cleartext); return hacks::authenticate_cleartext(ctx, info, client, secret, config).await; } // Finally, proceed with the main auth flow (SCRAM-based). classic::authenticate(ctx, info, client, config, secret).await } impl<'a> Backend<'a, ComputeUserInfoMaybeEndpoint> { /// Get username from the credentials. pub(crate) fn get_user(&self) -> &str { match self { Self::ControlPlane(_, user_info) => &user_info.user, Self::Local(_) => "local", } } /// Authenticate the client via the requested backend, possibly using credentials. #[tracing::instrument(fields(allow_cleartext = allow_cleartext), skip_all)] pub(crate) async fn authenticate( self, ctx: &RequestContext, client: &mut stream::PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, allow_cleartext: bool, config: &'static AuthenticationConfig, endpoint_rate_limiter: Arc<EndpointRateLimiter>, ) -> auth::Result<Backend<'a, ComputeCredentials>> { let res = match self { Self::ControlPlane(api, user_info) => { debug!( user = &*user_info.user, project = user_info.endpoint(), "performing authentication using the console" ); let auth_res = auth_quirks( ctx, &*api, user_info.clone(), client, allow_cleartext, config, endpoint_rate_limiter, ) .await; match auth_res { Ok(credentials) => Ok(Backend::ControlPlane(api, credentials)), Err(e) => { // The password could have been changed, so we invalidate the cache. // We should only invalidate the cache if the TTL might have expired. if e.is_password_failed() && let ControlPlaneClient::ProxyV1(api) = &*api && let Some(ep) = &user_info.endpoint_id { api.caches .project_info .maybe_invalidate_role_secret(ep, &user_info.user); } Err(e) } } } Self::Local(_) => { return Err(auth::AuthError::bad_auth_method("invalid for local proxy")); } }; // TODO: replace with some metric info!("user successfully authenticated"); res } } impl Backend<'_, ComputeUserInfo> { pub(crate) async fn get_role_secret( &self, ctx: &RequestContext, ) -> Result<RoleAccessControl, GetAuthInfoError> { match self { Self::ControlPlane(api, user_info) => { api.get_role_access_control(ctx, &user_info.endpoint, &user_info.user) .await } Self::Local(_) => Ok(RoleAccessControl { secret: None }), } } pub(crate) async fn get_endpoint_access_control( &self, ctx: &RequestContext, ) -> Result<EndpointAccessControl, GetAuthInfoError> { match self { Self::ControlPlane(api, user_info) => { api.get_endpoint_access_control(ctx, &user_info.endpoint, &user_info.user) .await } Self::Local(_) => Ok(EndpointAccessControl { allowed_ips: Arc::new(vec![]), allowed_vpce: Arc::new(vec![]), flags: AccessBlockerFlags::default(), rate_limits: EndpointRateLimitConfig::default(), }), } } } #[async_trait::async_trait] impl WakeComputeBackend for Backend<'_, ComputeUserInfo> { async fn wake_compute( &self, ctx: &RequestContext, ) -> Result<CachedNodeInfo, control_plane::errors::WakeComputeError> { match self { Self::ControlPlane(api, info) => api.wake_compute(ctx, info).await, Self::Local(local) => Ok(Cached::new_uncached(local.node_info.clone())), } } } #[cfg(test)] mod tests { #![allow(clippy::unimplemented, clippy::unwrap_used)] use std::sync::Arc; use bytes::BytesMut; use control_plane::AuthSecret; use fallible_iterator::FallibleIterator; use once_cell::sync::Lazy; use postgres_protocol::authentication::sasl::{ChannelBinding, ScramSha256}; use postgres_protocol::message::backend::Message as PgMessage; use postgres_protocol::message::frontend; use tokio::io::{AsyncRead, AsyncReadExt, AsyncWriteExt}; use super::auth_quirks; use super::jwt::JwkCache; use crate::auth::{ComputeUserInfoMaybeEndpoint, IpPattern}; use crate::cache::node_info::CachedNodeInfo; use crate::config::AuthenticationConfig; use crate::context::RequestContext; use crate::control_plane::messages::EndpointRateLimitConfig; use crate::control_plane::{ self, AccessBlockerFlags, EndpointAccessControl, RoleAccessControl, }; use crate::proxy::NeonOptions; use crate::rate_limiter::EndpointRateLimiter; use crate::scram::ServerSecret; use crate::scram::threadpool::ThreadPool; use crate::stream::{PqStream, Stream}; struct Auth { ips: Vec<IpPattern>, vpc_endpoint_ids: Vec<String>, access_blocker_flags: AccessBlockerFlags, secret: AuthSecret, } impl control_plane::ControlPlaneApi for Auth { async fn get_role_access_control( &self, _ctx: &RequestContext, _endpoint: &crate::types::EndpointId, _role: &crate::types::RoleName, ) -> Result<RoleAccessControl, control_plane::errors::GetAuthInfoError> { Ok(RoleAccessControl { secret: Some(self.secret.clone()), }) } async fn get_endpoint_access_control( &self, _ctx: &RequestContext, _endpoint: &crate::types::EndpointId, _role: &crate::types::RoleName, ) -> Result<EndpointAccessControl, control_plane::errors::GetAuthInfoError> { Ok(EndpointAccessControl { allowed_ips: Arc::new(self.ips.clone()), allowed_vpce: Arc::new(self.vpc_endpoint_ids.clone()), flags: self.access_blocker_flags, rate_limits: EndpointRateLimitConfig::default(), }) } async fn get_endpoint_jwks( &self, _ctx: &RequestContext, _endpoint: &crate::types::EndpointId, ) -> Result<Vec<super::jwt::AuthRule>, control_plane::errors::GetEndpointJwksError> { unimplemented!() } async fn wake_compute( &self, _ctx: &RequestContext, _user_info: &super::ComputeUserInfo, ) -> Result<CachedNodeInfo, control_plane::errors::WakeComputeError> { unimplemented!() } } static CONFIG: Lazy<AuthenticationConfig> = Lazy::new(|| AuthenticationConfig { jwks_cache: JwkCache::default(), scram_thread_pool: ThreadPool::new(1), scram_protocol_timeout: std::time::Duration::from_secs(5), ip_allowlist_check_enabled: true, is_vpc_acccess_proxy: false, is_auth_broker: false, accept_jwts: false, console_redirect_confirmation_timeout: std::time::Duration::from_secs(5), }); async fn read_message(r: &mut (impl AsyncRead + Unpin), b: &mut BytesMut) -> PgMessage { loop { r.read_buf(&mut *b).await.unwrap(); if let Some(m) = PgMessage::parse(&mut *b).unwrap() { break m; } } } #[tokio::test] async fn auth_quirks_scram() { let (mut client, server) = tokio::io::duplex(1024); let mut stream = PqStream::new_skip_handshake(Stream::from_raw(server)); let ctx = RequestContext::test(); let api = Auth { ips: vec![], vpc_endpoint_ids: vec![], access_blocker_flags: AccessBlockerFlags::default(), secret: AuthSecret::Scram(ServerSecret::build("my-secret-password").await.unwrap()), }; let user_info = ComputeUserInfoMaybeEndpoint { user: "conrad".into(), endpoint_id: Some("endpoint".into()), options: NeonOptions::default(), }; let handle = tokio::spawn(async move { let mut scram = ScramSha256::new(b"my-secret-password", ChannelBinding::unsupported()); let mut read = BytesMut::new(); // server should offer scram match read_message(&mut client, &mut read).await { PgMessage::AuthenticationSasl(a) => { let options: Vec<&str> = a.mechanisms().collect().unwrap(); assert_eq!(options, ["SCRAM-SHA-256"]); } _ => panic!("wrong message"), } // client sends client-first-message let mut write = BytesMut::new(); frontend::sasl_initial_response("SCRAM-SHA-256", scram.message(), &mut write).unwrap(); client.write_all(&write).await.unwrap(); // server response with server-first-message match read_message(&mut client, &mut read).await { PgMessage::AuthenticationSaslContinue(a) => { scram.update(a.data()).await.unwrap(); } _ => panic!("wrong message"), } // client response with client-final-message write.clear(); frontend::sasl_response(scram.message(), &mut write).unwrap(); client.write_all(&write).await.unwrap(); // server response with server-final-message match read_message(&mut client, &mut read).await { PgMessage::AuthenticationSaslFinal(a) => { scram.finish(a.data()).unwrap(); } _ => panic!("wrong message"), } }); let endpoint_rate_limiter = Arc::new(EndpointRateLimiter::new_with_shards( EndpointRateLimiter::DEFAULT, 64, )); let _creds = auth_quirks( &ctx, &api, user_info, &mut stream, false, &CONFIG, endpoint_rate_limiter, ) .await .unwrap(); // flush the final server message stream.flush().await.unwrap(); handle.await.unwrap(); } #[tokio::test] async fn auth_quirks_cleartext() { let (mut client, server) = tokio::io::duplex(1024); let mut stream = PqStream::new_skip_handshake(Stream::from_raw(server)); let ctx = RequestContext::test(); let api = Auth { ips: vec![], vpc_endpoint_ids: vec![], access_blocker_flags: AccessBlockerFlags::default(), secret: AuthSecret::Scram(ServerSecret::build("my-secret-password").await.unwrap()), }; let user_info = ComputeUserInfoMaybeEndpoint { user: "conrad".into(), endpoint_id: Some("endpoint".into()), options: NeonOptions::default(), }; let handle = tokio::spawn(async move { let mut read = BytesMut::new(); let mut write = BytesMut::new(); // server should offer cleartext match read_message(&mut client, &mut read).await { PgMessage::AuthenticationCleartextPassword => {} _ => panic!("wrong message"), } // client responds with password write.clear(); frontend::password_message(b"my-secret-password", &mut write).unwrap(); client.write_all(&write).await.unwrap(); }); let endpoint_rate_limiter = Arc::new(EndpointRateLimiter::new_with_shards( EndpointRateLimiter::DEFAULT, 64, )); let _creds = auth_quirks( &ctx, &api, user_info, &mut stream, true, &CONFIG, endpoint_rate_limiter, ) .await .unwrap(); handle.await.unwrap(); } #[tokio::test] async fn auth_quirks_password_hack() { let (mut client, server) = tokio::io::duplex(1024); let mut stream = PqStream::new_skip_handshake(Stream::from_raw(server)); let ctx = RequestContext::test(); let api = Auth { ips: vec![], vpc_endpoint_ids: vec![], access_blocker_flags: AccessBlockerFlags::default(), secret: AuthSecret::Scram(ServerSecret::build("my-secret-password").await.unwrap()), }; let user_info = ComputeUserInfoMaybeEndpoint { user: "conrad".into(), endpoint_id: None, options: NeonOptions::default(), }; let handle = tokio::spawn(async move { let mut read = BytesMut::new(); // server should offer cleartext match read_message(&mut client, &mut read).await { PgMessage::AuthenticationCleartextPassword => {} _ => panic!("wrong message"), } // client responds with password let mut write = BytesMut::new(); frontend::password_message(b"endpoint=my-endpoint;my-secret-password", &mut write) .unwrap(); client.write_all(&write).await.unwrap(); }); let endpoint_rate_limiter = Arc::new(EndpointRateLimiter::new_with_shards( EndpointRateLimiter::DEFAULT, 64, )); let creds = auth_quirks( &ctx, &api, user_info, &mut stream, true, &CONFIG, endpoint_rate_limiter, ) .await .unwrap(); assert_eq!(creds.info.endpoint, "my-endpoint"); handle.await.unwrap(); } }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/console_redirect.rs
proxy/src/auth/backend/console_redirect.rs
use std::fmt; use async_trait::async_trait; use postgres_client::config::SslMode; use thiserror::Error; use tokio::io::{AsyncRead, AsyncWrite}; use tracing::{info, info_span}; use crate::auth::backend::ComputeUserInfo; use crate::cache::Cached; use crate::cache::node_info::CachedNodeInfo; use crate::compute::AuthInfo; use crate::config::AuthenticationConfig; use crate::context::RequestContext; use crate::control_plane::client::cplane_proxy_v1; use crate::control_plane::{self, NodeInfo}; use crate::error::{ReportableError, UserFacingError}; use crate::pqproto::BeMessage; use crate::proxy::NeonOptions; use crate::proxy::wake_compute::WakeComputeBackend; use crate::stream::PqStream; use crate::types::RoleName; use crate::{auth, compute, waiters}; #[derive(Debug, Error)] pub(crate) enum ConsoleRedirectError { #[error(transparent)] WaiterRegister(#[from] waiters::RegisterError), #[error(transparent)] WaiterWait(#[from] waiters::WaitError), #[error(transparent)] Io(#[from] std::io::Error), } #[derive(Debug)] pub struct ConsoleRedirectBackend { console_uri: reqwest::Url, api: cplane_proxy_v1::NeonControlPlaneClient, } impl fmt::Debug for cplane_proxy_v1::NeonControlPlaneClient { fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result { write!(f, "NeonControlPlaneClient") } } impl UserFacingError for ConsoleRedirectError { fn to_string_client(&self) -> String { "Internal error".to_string() } } impl ReportableError for ConsoleRedirectError { fn get_error_kind(&self) -> crate::error::ErrorKind { match self { Self::WaiterRegister(_) => crate::error::ErrorKind::Service, Self::WaiterWait(_) => crate::error::ErrorKind::Service, Self::Io(_) => crate::error::ErrorKind::ClientDisconnect, } } } fn hello_message( redirect_uri: &reqwest::Url, session_id: &str, duration: std::time::Duration, ) -> String { let formatted_duration = humantime::format_duration(duration).to_string(); format!( concat![ "Welcome to Neon!\n", "Authenticate by visiting (will expire in {duration}):\n", " {redirect_uri}{session_id}\n\n", ], duration = formatted_duration, redirect_uri = redirect_uri, session_id = session_id, ) } pub(crate) fn new_psql_session_id() -> String { hex::encode(rand::random::<[u8; 8]>()) } impl ConsoleRedirectBackend { pub fn new(console_uri: reqwest::Url, api: cplane_proxy_v1::NeonControlPlaneClient) -> Self { Self { console_uri, api } } pub(crate) fn get_api(&self) -> &cplane_proxy_v1::NeonControlPlaneClient { &self.api } pub(crate) async fn authenticate( &self, ctx: &RequestContext, auth_config: &'static AuthenticationConfig, client: &mut PqStream<impl AsyncRead + AsyncWrite + Unpin>, ) -> auth::Result<(ConsoleRedirectNodeInfo, AuthInfo, ComputeUserInfo)> { authenticate(ctx, auth_config, &self.console_uri, client) .await .map(|(node_info, auth_info, user_info)| { (ConsoleRedirectNodeInfo(node_info), auth_info, user_info) }) } } pub struct ConsoleRedirectNodeInfo(pub(super) NodeInfo); #[async_trait] impl WakeComputeBackend for ConsoleRedirectNodeInfo { async fn wake_compute( &self, _ctx: &RequestContext, ) -> Result<CachedNodeInfo, control_plane::errors::WakeComputeError> { Ok(Cached::new_uncached(self.0.clone())) } } async fn authenticate( ctx: &RequestContext, auth_config: &'static AuthenticationConfig, link_uri: &reqwest::Url, client: &mut PqStream<impl AsyncRead + AsyncWrite + Unpin>, ) -> auth::Result<(NodeInfo, AuthInfo, ComputeUserInfo)> { ctx.set_auth_method(crate::context::AuthMethod::ConsoleRedirect); // registering waiter can fail if we get unlucky with rng. // just try again. let (psql_session_id, waiter) = loop { let psql_session_id = new_psql_session_id(); if let Ok(waiter) = control_plane::mgmt::get_waiter(&psql_session_id) { break (psql_session_id, waiter); } }; let span = info_span!("console_redirect", psql_session_id = &psql_session_id); let greeting = hello_message( link_uri, &psql_session_id, auth_config.console_redirect_confirmation_timeout, ); // Give user a URL to spawn a new database. info!(parent: &span, "sending the auth URL to the user"); client.write_message(BeMessage::AuthenticationOk); client.write_message(BeMessage::ParameterStatus { name: b"client_encoding", value: b"UTF8", }); client.write_message(BeMessage::NoticeResponse(&greeting)); client.flush().await?; // Wait for console response via control plane (see `mgmt`). info!(parent: &span, "waiting for console's reply..."); let db_info = tokio::time::timeout(auth_config.console_redirect_confirmation_timeout, waiter) .await .map_err(|_elapsed| { auth::AuthError::confirmation_timeout( auth_config.console_redirect_confirmation_timeout.into(), ) })? .map_err(ConsoleRedirectError::from)?; if auth_config.ip_allowlist_check_enabled && let Some(allowed_ips) = &db_info.allowed_ips && !auth::check_peer_addr_is_in_list(&ctx.peer_addr(), allowed_ips) { return Err(auth::AuthError::ip_address_not_allowed(ctx.peer_addr())); } // Check if the access over the public internet is allowed, otherwise block. Note that // the console redirect is not behind the VPC service endpoint, so we don't need to check // the VPC endpoint ID. if let Some(public_access_allowed) = db_info.public_access_allowed && !public_access_allowed { return Err(auth::AuthError::NetworkNotAllowed); } // Backwards compatibility. pg_sni_proxy uses "--" in domain names // while direct connections do not. Once we migrate to pg_sni_proxy // everywhere, we can remove this. let ssl_mode = if db_info.host.contains("--") { // we need TLS connection with SNI info to properly route it SslMode::Require } else { SslMode::Disable }; let conn_info = compute::ConnectInfo { host: db_info.host.into(), port: db_info.port, ssl_mode, host_addr: None, }; let auth_info = AuthInfo::for_console_redirect(&db_info.dbname, &db_info.user, db_info.password.as_deref()); let user: RoleName = db_info.user.into(); let user_info = ComputeUserInfo { endpoint: db_info.aux.endpoint_id.as_str().into(), user: user.clone(), options: NeonOptions::default(), }; ctx.set_dbname(db_info.dbname.into()); ctx.set_user(user); ctx.set_project(db_info.aux.clone()); info!("woken up a compute node"); Ok(( NodeInfo { conn_info, aux: db_info.aux, }, auth_info, user_info, )) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/classic.rs
proxy/src/auth/backend/classic.rs
use tokio::io::{AsyncRead, AsyncWrite}; use tracing::{debug, info, warn}; use super::{ComputeCredentials, ComputeUserInfo}; use crate::auth::backend::ComputeCredentialKeys; use crate::auth::{self, AuthFlow}; use crate::config::AuthenticationConfig; use crate::context::RequestContext; use crate::control_plane::AuthSecret; use crate::stream::{PqStream, Stream}; use crate::{compute, sasl}; pub(super) async fn authenticate( ctx: &RequestContext, creds: ComputeUserInfo, client: &mut PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, config: &'static AuthenticationConfig, secret: AuthSecret, ) -> auth::Result<ComputeCredentials> { let scram_keys = match secret { AuthSecret::Scram(secret) => { debug!("auth endpoint chooses SCRAM"); let auth_outcome = tokio::time::timeout( config.scram_protocol_timeout, AuthFlow::new(client, auth::Scram(&secret, ctx)).authenticate(), ) .await .inspect_err(|_| warn!("error processing scram messages error = authentication timed out, execution time exceeded {} seconds", config.scram_protocol_timeout.as_secs())) .map_err(auth::AuthError::user_timeout)? .inspect_err(|error| warn!(?error, "error processing scram messages"))?; let client_key = match auth_outcome { sasl::Outcome::Success(key) => key, sasl::Outcome::Failure(reason) => { // TODO: warnings? // TODO: should we get rid of this because double logging? info!("auth backend failed with an error: {reason}"); return Err(auth::AuthError::password_failed(&*creds.user)); } }; compute::ScramKeys { client_key: client_key.as_bytes(), server_key: secret.server_key.as_bytes(), } } }; Ok(ComputeCredentials { info: creds, keys: ComputeCredentialKeys::AuthKeys(postgres_client::config::AuthKeys::ScramSha256( scram_keys, )), }) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/auth/backend/hacks.rs
proxy/src/auth/backend/hacks.rs
use tokio::io::{AsyncRead, AsyncWrite}; use tracing::{debug, info}; use super::{ComputeCredentials, ComputeUserInfo, ComputeUserInfoNoEndpoint}; use crate::auth::{self, AuthFlow}; use crate::config::AuthenticationConfig; use crate::context::RequestContext; use crate::control_plane::AuthSecret; use crate::intern::{EndpointIdInt, RoleNameInt}; use crate::sasl; use crate::stream::{self, Stream}; /// Compared to [SCRAM](crate::scram), cleartext password auth saves /// one round trip and *expensive* computations (>= 4096 HMAC iterations). /// These properties are benefical for serverless JS workers, so we /// use this mechanism for websocket connections. pub(crate) async fn authenticate_cleartext( ctx: &RequestContext, info: ComputeUserInfo, client: &mut stream::PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, secret: AuthSecret, config: &'static AuthenticationConfig, ) -> auth::Result<ComputeCredentials> { debug!("cleartext auth flow override is enabled, proceeding"); ctx.set_auth_method(crate::context::AuthMethod::Cleartext); let ep = EndpointIdInt::from(&info.endpoint); let role = RoleNameInt::from(&info.user); let auth_flow = AuthFlow::new( client, auth::CleartextPassword { secret, endpoint: ep, role, pool: config.scram_thread_pool.clone(), }, ); let auth_outcome = { // pause the timer while we communicate with the client let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Client); // cleartext auth is only allowed to the ws/http protocol. // If we're here, we already received the password in the first message. // Scram protocol will be executed on the proxy side. auth_flow.authenticate().await? }; let keys = match auth_outcome { sasl::Outcome::Success(key) => key, sasl::Outcome::Failure(reason) => { info!("auth backend failed with an error: {reason}"); return Err(auth::AuthError::password_failed(&*info.user)); } }; Ok(ComputeCredentials { info, keys }) } /// Workaround for clients which don't provide an endpoint (project) name. /// Similar to [`authenticate_cleartext`], but there's a specific password format, /// and passwords are not yet validated (we don't know how to validate them!) pub(crate) async fn password_hack_no_authentication( ctx: &RequestContext, info: ComputeUserInfoNoEndpoint, client: &mut stream::PqStream<Stream<impl AsyncRead + AsyncWrite + Unpin>>, ) -> auth::Result<(ComputeUserInfo, Vec<u8>)> { debug!("project not specified, resorting to the password hack auth flow"); ctx.set_auth_method(crate::context::AuthMethod::Cleartext); // pause the timer while we communicate with the client let _paused = ctx.latency_timer_pause(crate::metrics::Waiting::Client); let payload = AuthFlow::new(client, auth::PasswordHack) .get_password() .await?; debug!(project = &*payload.endpoint, "received missing parameter"); // Report tentative success; compute node will check the password anyway. Ok(( ComputeUserInfo { user: info.user, options: info.options, endpoint: payload.endpoint, }, payload.password, )) }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false
neondatabase/neon
https://github.com/neondatabase/neon/blob/015b1c7cb3259a6fcd5039bc2bd46a462e163ae8/proxy/src/tls/client_config.rs
proxy/src/tls/client_config.rs
use std::env; use std::io::Cursor; use std::path::PathBuf; use std::sync::Arc; use anyhow::{Context, bail}; use rustls::crypto::ring; /// We use an internal certificate authority when establishing a TLS connection with compute. fn load_internal_certs(store: &mut rustls::RootCertStore) -> anyhow::Result<()> { let Some(ca_file) = env::var_os("NEON_INTERNAL_CA_FILE") else { return Ok(()); }; let ca_file = PathBuf::from(ca_file); let ca = std::fs::read(&ca_file) .with_context(|| format!("could not read CA from {}", ca_file.display()))?; for cert in rustls_pemfile::certs(&mut Cursor::new(&*ca)) { store .add(cert.context("could not parse internal CA certificate")?) .context("could not parse internal CA certificate")?; } Ok(()) } /// For console redirect proxy, we need to establish a connection to compute via pg-sni-router. /// pg-sni-router needs TLS and uses a Let's Encrypt signed certificate, so we /// load certificates from our native store. fn load_native_certs(store: &mut rustls::RootCertStore) -> anyhow::Result<()> { let der_certs = rustls_native_certs::load_native_certs(); if !der_certs.errors.is_empty() { bail!("could not parse certificates: {:?}", der_certs.errors); } store.add_parsable_certificates(der_certs.certs); Ok(()) } fn load_compute_certs() -> anyhow::Result<Arc<rustls::RootCertStore>> { let mut store = rustls::RootCertStore::empty(); load_native_certs(&mut store)?; load_internal_certs(&mut store)?; Ok(Arc::new(store)) } /// Loads the root certificates and constructs a client config suitable for connecting to the neon compute. /// This function is blocking. pub fn compute_client_config_with_root_certs() -> anyhow::Result<rustls::ClientConfig> { Ok( rustls::ClientConfig::builder_with_provider(Arc::new(ring::default_provider())) .with_safe_default_protocol_versions() .expect("ring should support the default protocol versions") .with_root_certificates(load_compute_certs()?) .with_no_client_auth(), ) } #[cfg(test)] pub fn compute_client_config_with_certs( certs: impl IntoIterator<Item = rustls::pki_types::CertificateDer<'static>>, ) -> rustls::ClientConfig { let mut store = rustls::RootCertStore::empty(); store.add_parsable_certificates(certs); rustls::ClientConfig::builder_with_provider(Arc::new(ring::default_provider())) .with_safe_default_protocol_versions() .expect("ring should support the default protocol versions") .with_root_certificates(store) .with_no_client_auth() }
rust
Apache-2.0
015b1c7cb3259a6fcd5039bc2bd46a462e163ae8
2026-01-04T15:40:24.223849Z
false